Next Article in Journal
A Compact Annular Ring Microstrip Antenna for WSN Applications
Previous Article in Journal
Recent Progress in Distributed Fiber Optic Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial Uncertainty Model for Visual Features Using a Kinect™ Sensor

Robot Convergence R & D Group, Korea Institute of Industrial Technology (KITECH), 1271-18, Sa-3-dong, Sangrok-gu, Ansan 426791, Korea
*
Author to whom correspondence should be addressed.
Sensors 2012, 12(7), 8640-8662; https://doi.org/10.3390/s120708640
Submission received: 18 May 2012 / Revised: 12 June 2012 / Accepted: 13 June 2012 / Published: 26 June 2012
(This article belongs to the Section Physical Sensors)

Abstract

: This study proposes a mathematical uncertainty model for the spatial measurement of visual features using Kinect™ sensors. This model can provide qualitative and quantitative analysis for the utilization of Kinect™ sensors as 3D perception sensors. In order to achieve this objective, we derived the propagation relationship of the uncertainties between the disparity image space and the real Cartesian space with the mapping function between the two spaces. Using this propagation relationship, we obtained the mathematical model for the covariance matrix of the measurement error, which represents the uncertainty for spatial position of visual features from Kinect™ sensors. In order to derive the quantitative model of spatial uncertainty for visual features, we estimated the covariance matrix in the disparity image space using collected visual feature data. Further, we computed the spatial uncertainty information by applying the covariance matrix in the disparity image space and the calibrated sensor parameters to the proposed mathematical model. This spatial uncertainty model was verified by comparing the uncertainty ellipsoids for spatial covariance matrices and the distribution of scattered matching visual features. We expect that this spatial uncertainty model and its analyses will be useful in various Kinect™ sensor applications.

1. Introduction

On 4 November 2010, Kinect™ was launched as a non-contact motion sensing device by Microsoft for the Xbox 360 video game console [1]. However, the remarkable ability of a Kinect™ sensor lies in the important functionality that it can provide after acquisition of high quality 3D scan information in real time at a relatively low cost. Therefore, in addition to motion sensing for gaming, the use of Kinect™ sensors in various applications has been actively investigated in many research areas such as robotics, human-computer interface (HCI), and geospatial information. In the domain of robotics, in particular, many studies are trying to utilize Kinect™ sensors as 3D sensors for perception functionality of intelligent robots [27].

Kinect™ sensors provide disparity image and RGB image information simultaneously. Hence, the colored 3D point cloud information could be acquired by fusing the disparity and RGB information from a Kinect™ sensor. However, a calibration process is required for utilizing a Kinect™ sensor as a 3D sensor. For Kinect™ sensor calibration, certain parameters are required: the pin-hole projection and lens distortion parameters of the disparity and RGB cameras, the homogeneous matrix of the two-camera coordinate frame, and the depth calibration parameter, which can transform disparity image data into actual distance. The pin-hole projection and lens distortion parameters of the depth and RGB cameras can be obtained with the existing calibration solution [8,9]. Further, the homogeneous matrix parameters between the depth camera and the RGB camera coordinates can be obtained by the stereo camera calibration method [10,11] or the point cloud matching method [1214]. Some recent studies have presented results related to depth calibration methods and analyses for acquiring accurate 3D data using the disparity image from a Kinect™ sensor [1518].

Recently, Kinect™ sensors have been widely utilized as 3D perception sensors in various robotic applications such as 3D mapping, object pose estimation, and Simultaneous Localization and Mapping (SLAM) [36]. In these applications, extraction of visual features, matching, and estimation of the 3D position are essential functionalities. Kinect™ sensors are very suitable for these applications because the essential functionalities can be achieved easily using the disparity and the RGB information. These problems can be solved by stochastic optimization methods, which contain measurement error and uncertainties. In this phase, quantitative information about the measurement error and uncertainties of visual features are essential for a reliable estimation result. For example, the covariance matrix of the input noises and errors is the key design parameter for optimal estimation problem using a Kalman filter. In general, all sensors have static and dynamic errors. Static errors, representing the bias of the estimation results, can be corrected by calibration. Dynamic errors, representing the variance of the estimation results, can be improved by filtering methods. However, results for a mathematical uncertainty model representing the covariance matrix form for the spatial measurements of visual features using Kinect™ sensors are unavailable. Khoshelham and Elberink [16] presented an error model and its analysis results; however, these results were represented as an independent error model with respect to the X, Y and Z axis, and not as a covariance matrix. In the Cartesian space, the errors in the X, Y and Z axis data are correlated with each other; thus, the covariance matrix is not in a diagonal form. Therefore, we would like to derive the spatial uncertainty model of visual features using Kinect™ sensors, which is represented by the covariance matrix for 3D measurement errors in the actual Cartesian space.

To achieve this objective, we derive the propagation relationship of the uncertainties between the disparity image space and the real Cartesian space with the mapping function between the two spaces. Then, we obtain the mathematical model for the covariance matrix of the spatial measurement error by using the propagation relationship. Finally, a quantitative analysis of the spatial measurement of Kinect™ sensors is performed by applying the covariance matrix in the disparity image space and the calibrated sensor parameters to the proposed mathematical model.

2. 3D Reconstruction from Kinect™ Sensor Data

Kinect™ sensors provide disparity image and RGB image information. The disparity image represents the spatial information, and the RGB image represents the color information. 3D point cloud data, which contains color information, can be obtained by fusing the disparity image and the RGB image information. Figure 1 shows the disparity image, the RGB image, and the colored 3D point cloud information that was reconstructed from a Kinect™ sensor. Disparity image data, containing information about the distance of the location of each pixel, is expressed as an integer from 0 to 2,047. This data contains relative distance information, which does not represent metric information. In addition, the relationship between distance and disparity image data is non-linear, as shown in the graph in Figure 2. Thus, the depth calibration function, which can transform disparity image data into actual distance information, is needed in order to reconstruct 3D information using Kinect sensors.

The mathematical model between disparity image data d and real depth is represented by Equation (1) [16]. In this equation, Zo, fo, and b indicate the distance of the reference pattern, focal length, and base length respectively. For depth calibration, two parameters, 1/Zo and 1/(fo b), are determined by the least square fitting method [17]:

f ( d ) = Z o 1 + Z o f o b d = 1 1 Z o + 1 f o b d

In our experiment, the maximum detection range of the Kinect™ sensor was 17.3 m at disparity data 1,069, and the distribution of data changed rapidly beyond a distance of approximately 5 m, as shown in Figure 2. For performing data fitting, the depth calibration model of Equation (1) has only two-degrees-of-freedom for the optimization variables; therefore, it has limitations in representing the curvature of our measurement data. Hence, we proposed an extended depth calibration model using a rational function, which contains higher degree-of-freedom in the optimization variable space [19]. Equation (2) shows the rational function model that is applied to the depth calibration of the Kinect™ sensor:

f ( d ) = P ( d ) Q ( d ) ( P ( d ) = i = 1 m θ i d i 1 , Q ( d ) = i = 1 n θ m + i d i 1 )

where P(d) is the numerator polynomial and Q(d) is the denominator polynomial.

To perform depth calibration with the rational function model, a non-linear optimization method such as the Levenberg-Marquardt algorithm can be used. We obtained the depth calibration function with fourth-order polynomials of the numerator and denominator, which can transform disparity data into a real distance of up to approximately 15 m. The depth calibration parameters for the fourth-order rational function model are shown in Table 1. Figure 3 shows the fitting results and the fitting residual results for the depth calibration function in Equations (1) and (2), respectively. In Figure 3(a), both calibration functions seemed to fit the measurement data well. However, as seen in Figure 3(b), the residual error of the rational function model appeared to be nearer to the X-axis than the model represented by Equation (1). This implies that the rational function model with a higher degree-of-freedom of the optimization variables can be fitted more precisely in the depth calibration problem. The norm of residual vector for Equation (1) and the rational function were computed to be 1.045495 and 1.034060, respectively.

After performing depth calibration, the disparity image data can be transformed into the actual distance information by the depth calibration function. Using this actual distance information, the 3D spatial position information can be reconstructed with the pin-hole camera projection model. Equation (3) shows the mapping relationship between the disparity image space data u = [u v d]T and the spatial position information x = [x y z]T in the Cartesian space. u and v are the horizontal and vertical coordinates, respectively, of the disparity image, and d is the disparity data, expressed as an integer from 0 to 2,047. f(d) is the actual distance information that is calculated by the depth calibration function:

x = F ( u ) [ x y z ] = f ( d ) [ 1 f Depth , x 0 C Depth , x f Depth , x 0 1 f Depth , y C Depth , y f Depth , y 0 0 1 ] [ u v 1 ]

In the pin-hole camera projection model, fDepth,x and fDepth,y are focal length parameters, while CDepth,x and CDepth,y are optical axis parameters of the depth camera. These parameters can be obtained using various general camera calibration methods. The pin-hole camera projection parameters are shown in Table 2. We obtained these parameters using the Matlab camera calibration toolbox developed by Bouguet [20].

3. Spatial Uncertainty Model of Kinect™ Sensor

The disparity image data from Kinect™ sensors can be converted into the 3D spatial point cloud data using the depth calibration function and the pin-hole camera projection model. However, the 3D position information of the visual features using Kinect™ sensors contains some errors caused by various sources such as inaccurate measurement of disparity, lighting condition, properties of the object surfaces in the disparity data, and image processing and matching errors in the image coordinates. In order to utilize sensor data in actual applications, the information about reliability or uncertainty of the sensor is very important. In this study, we would like to propose a mathematical model for the 3D measurement information, which can provide qualitative and quantitative analysis for Kinect™ sensors.

3.1. Qualitative Analysis of Spatial Uncertainty

The reliability of the measured 3D information can be represented by the multi-dimensional Gaussian model in the Cartesian space, as shown in Equation (4) and Figure 4. In the Gaussian model, random variables are modeled using the mean vector and the covariance matrix. The error of the mean vector with respect to the measured data is estimation bias, which should be corrected by calibration. The variance parameter of the Gaussian model represents uncertainties of the measurements, and it can be represented as an uncertainty ellipsoid related to the covariance matrix. Thus, we tried to derive a mathematical model of the covariance matrix that describes the spatial uncertainties:

g xyz ( x ) = ( 2 π ) 3 | Q | 1 2 exp { 1 2 ( x m x ) T Q 1 ( x m x ) } ( x = [ x , y , z ] T , m x = [ m x , m y , m z ] T )

To derive the spatial uncertainty model, the mapping relationship between the disparity image space and the real Cartesian space should be considered. Figure 5 shows this mapping relationship. Owing to the absence of correlations between the elements of vector u in the disparity image space, the covariance matrix R of vector u has a diagonal form, as shown in Equation (5). This deduction can be confirmed from experimental data. The symbols σu and σv represent the variance corresponding to the visual feature position image co-ordinates u and v, respectively. The variance is caused by image processing errors such as image pixel quantization and key point localization. The symbol σd represents the variance of disparity measurements, which result from inaccuracy, lighting condition, and properties of the object surfaces [16]. Thus, the elements of vector [u, v, d]T are unrelated, and the diagonal elements of the covariance matrix can be obtained independently. In addition, the causes of errors are independent of vector u, and hence, the covariance matrix R can be assumed to be the same in the entire disparity image space:

R = [ σ u 2 0 0 0 σ v 2 0 0 0 σ d 2 ]

Uncertainty in the actual space appears as the propagation of uncertainty in the disparity image space by mapping relations. If the relationship between the two spaces is a linear mapping such as y = Ax, the propagated output covariance matrix Q is determined as Q = ARAT for the input covariance matrix R [21]. However, as shown in Equation (3), the relationship between the two spaces is a non-linear mapping. Therefore, we can obtain the covariance matrix in the actual space by a linearized approximation of the mapping function using Jacobian matrix, as shown in Equation (6).

J ( u ) = F ( u ) u = [ f ( d ) f Depth , x 0 f ( d ) d ( u f Depth , x C Depth , x f Depth , x ) 0 f ( d ) f Depth , y f ( d ) d ( v f Depth , y C Depth , y f Depth , y ) 0 0 f ( d ) d ] f ( d ) = P ( d ) Q ( d ) , f ( d ) d = P ˙ ( d ) Q ( d ) P ( d ) Q ˙ ( d ) Q ( d ) 2

Thus, we can obtain the mathematical model of spatial uncertainty shown in Equation (7), and the uncertainty ellipsoid for Kinect™ measurement can be estimated in the entire measurable space using this covariance matrix model:

Q = J ( u ) R J ( u ) T Q = [ σ u 2 A 2 + σ d 2 C 2 D 2 σ d 2 C 2 D E σ d 2 C 2 D σ d 2 C 2 D E σ v 2 B 2 + σ d 2 C 2 E 2 σ d 2 C 2 E σ d 2 C 2 D σ d 2 C 2 E σ d 2 C 2 ]

where ( A = f ( d ) f Depth , x , B = f ( d ) f Depth , y , C = f ( d ) d , D = u C Depth , x f Depth , x , E = v C Depth , y f Depth , y J ( u , v , d ) = [ A 0 C D 0 B C E 0 0 C ] , R = [ σ u 2 0 0 0 σ v 2 0 0 0 σ d 2 ] )

3.2. Quantitative Analysis of Spatial Uncertainty

In order to perform a quantitative analysis of the uncertainty model, quantitative data obtained from the real sensor is needed. Hence, the actual sensor parameters such as the depth calibration parameters, pin-hole projection parameters of the depth camera, and covariance matrix information in the disparity image space are required. The depth calibration parameter and the pin-hole camera parameters are shown in Tables 1 and 2, respectively. The diagonal elements of covariance matrix Ru, σv, and σd) are obtained from the disparity image data of visual features by tracking 850 feature points in the real experimental environment. The SURF [22] algorithm was used to detect and match visual features for tracking the trajectory of feature points. The variances σu, σv, and σd were estimated as 1.051, 0.801, and 1.266, respectively, from experimental data. Equation (8) represents the Jacobian matrix and the input covariance matrix R, which was obtained from real sensor parameters and experimental data:

J ( u ) = [ 0.0017 f ( d ) 0 f ( d ) d ( 0.0017 u 0.549 ) 0 0.0017 f ( d ) f ( d ) d ( 0.0017 v 0.443 ) 0 0 f ( d ) d ] , R = [ 1.051 2 0 0 0 0.801 2 0 0 0 1.266 2 ]

Table 3 shows the Gaussian parameters (mean vector and covariance matrix), square root value of covariance matrix norm (maximum standard deviation), vector of maximum direction, and the uncertainty ellipsoid, for cases (a)–(h). The symbol ε in the covariance matrix represents a very small number with a near-zero value. From the results in various cases, it is observed that the spatial uncertainties vary with the distance and the image coordinates of the measurement position. Figure 6 shows all the uncertainty ellipsoids for cases (a)–(h) in the Cartesian space. From the test cases, it can be seen that the volume and direction of the uncertainty ellipsoids are closely related to the measurement position.

Figure 7 shows the uncertainty ellipsoid map for the Kinect™ sensor in the entire measurable Cartesian space. This uncertainty map is constructed by drawing the 3D ellipsoid for xTQ–1x = k, and calculating each spatial covariance matrix Q by using Equation (7) with increment steps of 40 for u, v, and d. The uncertainty ellipsoid map represents the distribution in volume and direction of the longest axis of uncertainty ellipsoids in the entire space. From this uncertainty map, it can be concluded that the volume of the uncertainty ellipsoid is greatly influenced by the distance of the measured point and its maximum direction is related to the direction of the optical axis of the sensor.

Figure 8 shows the distribution of maximum standard deviation (square root of norm for the covariance matrix Q) of the spatial uncertainties by varying u and v, and by keeping d fixed. The results showed a quadratic distribution in the Cartesian space when the depth remains the same. The measurement point is farther from the center of the optical axis in the image coordinate, and hence, the maximum standard deviation attains a higher value. Figure 9 shows the distribution of the maximum standard deviation obtained by varying u and d, and by keeping v fixed. Its distribution resembled a fan type plane in the Cartesian space when the horizontal measure remains the same. From the distribution, it can be observed that the maximum standard deviation increases with an increase in the depth. Further, an increase in depth causes a steeper gradient because it is farther from the center of the image coordinate. Figure 10 shows the integrated distribution of the maximum standard deviation for (a) various values of u and v, and three values of d and (b) various values of u and d, and three values of v. Further, Figure 11 shows the volume distribution of the maximum standard deviation for most of the disparity image space. From these analyses, it can be confirmed that spatial uncertainty varies with the distance and the image coordinates of the measurement position.

4. Experiments and Results

4.1. Estimation of the Input Covariance Matrix R

The covariance matrix in the disparity image is necessary for the calculation of the spatial uncertainty model. We tried to estimate the input covariance matrix R in the disparity image space from real experiments. Figure 12 shows the overall experimental environment for estimation of the matrix R. In this experimental environment, objects were placed at various locations and orientations in order to obtain visual feature information with a uniform distribution at various conditions and in the entire measurable space. Figure 12(b) shows the visual feature detection results obtained by using the SURF algorithm in this experimental environment. As shown in Figure 12(b), 850 visual features were obtained. Each visual feature is tracked continuously by the SURF matching function, and the corresponding trajectory information is recorded. The index of each visual feature is assigned randomly during the first detection phase. Figure 13(a,b) shows the distribution of the 850 detected visual features in the disparity image space and the real Cartesian space, respectively. Figure 14 shows a histogram representation of the distribution of the visual features in the image space with respect to the u, v, and d axes. The distribution of the histogram for u, v, and d axes confirmed that the visual features were uniformly distributed in the entire image space.

Each visual feature was obtained from 100 data measurements by matching and tracking, and the mean mi and the covariance matrix Ri in Equation (9) were calculated from the measurement data. The covariance matrices were calculated differently owing to various reasons. However, the mean covariance matrix was computed in order to characterize the representative covariance matrix. Based on Equation (10), the mean covariance matrix was computed from the covariance matrix of 850 visual features.

It was confirmed that the mean covariance matrix in Equation (10) was similar to the diagonal matrix form. Thus, the covariance matrix in the disparity image space can be assumed to be a diagonal matrix as in Equation (5), and the diagonal elements of the covariance matrix can be computed independently. Figure 11 shows the standard deviation data and its histogram for each of the 850 visual features, corresponding to the values of u, v, and d. As shown in Figure 15, the deviations for visual features were observed as random variables. Further, the statistical parameters should be obtained by taking most of the element data into consideration. From the data in Figure 15, the mean of standard deviation for u-axis and its deviation were 0.118 and 0.311, respectively. The mean of standard deviation for v-axis and its deviation were 0.072 and 0.243, respectively. The mean of standard deviation for d-axis and its deviation were 0.477 and 0.263, respectively. The input covariance matrix R should be determined by utilizing most of the covariance matrix Ri for each feature. Hence, the 3σ level threshold (99.7%), which can include most of the covariance matrices, was used for determining the elements of the covariance matrix R, as shown in Equation (11). Therefore, the estimated covariance matrix R represents the statistically worst case of measurement at the 3σ level. Figure 16 shows the uncertainty ellipsoids for all the visual features and the estimated covariance matrix in the disparity image space. From the result, it can be confirmed that the ellipsoid for the estimated covariance matrix includes most of the ellipsoids for the visual features:

m i = 1 100 k = 1 100 u i , k = 1 100 k = 1 100 [ u i v i d i ] k , R i = 1 100 k = 1 100 [ ( u i , k m i ) ( u i , k m i ) T ] ( i : Feature index 1 ~ 850 )
R ¯ = 1 850 i = 1 850 R i = [ 0.1564 0.0107 0.0145 0.0107 0.1099 0.0153 0.0145 0.0153 0.4761 ]
σ u = mean ( σ u ( i ) ) + 3 dev ( σ u ( i ) ) = 1.051 σ v = mean ( σ v ( i ) ) + 3 dev ( σ v ( i ) ) = 0.801 σ d = mean ( σ d ( i ) ) + 3 dev ( σ d ( i ) ) = 1.266 , R = [ 1.051 2 0 0 0 0.801 2 0 0 0 1.266 2 ]

4.2. Comparison with the Uncertainty Model and the Distribution of Real Visual Features

Given the covariance matrix in the disparity image space and the Kinect™ calibration parameters, the spatial covariance matrix can be calculated by using the uncertainty model proposed in this study. Using this spatial uncertainty model, we drew the uncertainty ellipsoid map shown in Figure 7, and we could identify its shape, volume, and direction in the Cartesian space. However, it must be confirmed that the calculated uncertainty model can represent the distribution of scattered measurement data. This can be verified by comparing the uncertainty ellipsoid with the distribution of the 3D position for tracked visual features. To confirm that the uncertainty model met all the requirements, some representative features were selected from the 850 visual features, and the uncertainty ellipsoid and the distribution of tracked measurement for the visual features were compared. Figure 17 shows the 20 selected visual features highlighted among all the visual features. As shown in Figure 17(a), the representative visual features were selected to ensure maximum possible uniform distribution in the Cartesian space. Figure 17(b) shows the 3D measurements for the visual features and the calculated uncertainty matrices, represented by the red symbol (*) and the cyan ellipsoids, respectively, for 20 visual features in one frame. However, owing to the difficulties in representing the scale corresponding to each feature, Figure 17(b) is not suitable for performing a detailed analysis. Hence, the results of features (a)–(d) in Figure 17(b) were represented again in Figure 18 by modifying the scale to obtain more detailed results. Then, the uncertainty ellipsoid was compared with the distribution for 3D measurement of visual features.

Figure 18(a) shows the results for the visual feature with id “100”, acquired at u = [331.6, 68, 1,023.6]T in the disparity image space. In the result, the measurements represented by the red symbol (*) appear at seven points clustered around the point x = [0.1, –1.7, 5.2]T in the Cartesian space. The input data in the disparity image space is discrete, and hence, the transformed 3D measurement must also be discrete. Therefore, the 3D measurement distribution is observed as a discrete distribution, and 100 measurements for this visual feature overlap at seven points. Further, it is verified that the cyan uncertainty ellipsoid includes all the 3D measurements corresponding to the visual feature. Figure 18(b) shows the results for the visual feature with id “220”, acquired at u = [36.4, 233.8, 963.8]T in the disparity image space. In the result, the symbols representing the measurements are seen at various points clustered around the point x = [–1.3, –0.1, 2.8]T in the Cartesian space. Further, the uncertainty ellipsoid includes most of the 3D measurements corresponding to the visual feature. In this distribution, 3 measurements are located slightly outside the ellipsoid boundary, but their distances from the boundary are extremely small. Figure 18(c) shows the results for the visual feature with id “282”, acquired at u = [297.6, 300.0, 937.0]T in the disparity image space. In the result, the measurements appear at 7 points clustered around the point x = [–0.1, 0.2, 2.3]T in the Cartesian space, and the uncertainty ellipsoid includes all the 3D measurements corresponding to the visual feature. Figure 18(d) shows the results for the visual feature with id “614”, acquired at u = [490.2, 188.0, 986.6]T in the disparity image space. In the result, the measurements appear at 5 points clustered around the point x = [1.0, –0.4, 3.4]T in the Cartesian space, and the uncertainty ellipsoid includes all the 3D measurements corresponding to the visual feature. The results for all the 20 visual features are represented in Figure 19 by modifying the scale to obtain more detailed results. The overall results show that the simple equation for the proposed spatial uncertainty model represents the worst case model for image space uncertainties; however, it was confirmed that the spatial uncertainty model provided a sufficiently good description of the discrete distribution for most of the 3D measurements of the visual features.

5. Conclusions

In this study, we proposed a mathematical model for spatial measurement uncertainty, which can provide qualitative and quantitative analysis for Kinect™ sensors. To achieve this objective, we derived the spatial covariance matrix model using the mapping function between the disparity image space and the actual Cartesian space. Next, we performed a quantitative analysis of the spatial measurement errors using actual sensor parameters. In order to derive the quantitative model of the spatial uncertainty for the visual features, we estimated the covariance matrix in the disparity image space using the collected visual feature data. Further, we computed the spatial uncertainty information by applying the covariance matrix in the disparity image space and the calibrated sensor parameters to the proposed mathematical model. This spatial uncertainty model was verified by comparing the uncertainty ellipsoids for spatial covariance matrices and the distribution of scattered matching visual features. Quantitative analysis of a Kinect™ sensor facilitates the availability of concrete information about the sensor, rather than abstract information. For example, abstract information, such as “If the measurement distance increases, the uncertainty will be increased”, could be transformed into concrete information, such as “Maximum error at a measurement distance of 1.2 m is 1.68 cm at the level 3σ.”

Recently, Kinect™ sensors have been widely utilized as 3D perception sensors for intelligent robots to solve various problems such as 3D mapping, object pose estimation, and SLAM. In these actual applications, information about the reliability and the uncertainty of the visual features for 3D measurements is very important. Hence, we expect that the uncertainty model presented in this paper will be useful in many applications that employ Kinect™ sensors.

References

  1. Kinect—Wikipedia, the Free Encyclopedia. Available online: http://en.wikipedia.org/wiki/Kinect (accessed on 27 March 2012).
  2. Top 10 Robotic Kinect Hacks. Available online: http://spectrum.ieee.org/automaton/robotics/diy/top-10-robotic-kinect-hacks/ (accessed on 21 March 2012).
  3. Huang, A.S.; Bachrach, A.; Henry, P.; Krainin, M.; Maturana, D.; Fox, D.; Roy, N. Visual Odometry and Mapping for Autonomous Flight Using an RGB-D Camera. Proceedings of ISRR, Flagstaff, AZ, USA, 28 August–1 September 2011.
  4. Taylor, C.J.; Cowley, A. Segmentation and Analysis of RGB-D Data. Proceedings of Robotics Science and Systems (RSS), Indianapolis, IN, USA, 14–16 September 2011.
  5. Sung, J.; Ponce, C.; Selman, B.; Saxena, A. Human Activity Detection from RGBD Images. Proceedings of AAAI Workshop on Pattern, Activity and Intent Recognition (PAIR), San Francisco, CA, USA, 7–8 August 2011.
  6. Benavidez, P.; Jamshidi, M. Mobile Robot Navigation and Target Tracking System. Proceedings of 6th International Conference on System of Systems Engineering, Albuquerque, NM, USA, 27–30 June 2011; pp. 299–304.
  7. Tolgyessy, M.; Hubinsky, P. The Kinect Sensor in Robotics Education. Proceedings of 2nd International Conference on Robotics in Education, Vienna, Austria, 15–16 September 2011.
  8. Zhang, Z. A flexible new technique for camera calibration. IEEE Transact. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar]
  9. Bouguet, J.Y.; Perona, P. Camera Calibration from Points and Lines in Dual-Space Geometry. Proceedings of European Conference on Computer Vision (ECCV), Freiburg, Germany, 2–6 June 1998.
  10. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: Cambridge, UK, 2003; pp. 493–497. [Google Scholar]
  11. House, B.; Nickels, K. Increased automation in stereo camera calibration techniques. J. Syst. Cyber. Inform. 2006, 4, 48–51. [Google Scholar]
  12. Park, J.H.; Shin, Y.D.; Park, K.W.; Baeg, S.H.; Baeg, M.H. Extracting Extrinsic Parameters of a Laser Scanner and a Camera Using EM. Proceedings of ICROS-SICE International Joint Conference, Fukuoka, Japan, 18–21 August 2009; pp. 5269–5272.
  13. Shin, Y.D.; Park, J.H.; Bae, J.H.; Baeg, M.H. A Study on Reliability Enhancement for Laser and Camera Calibration. Int. J. Contr. Autom. Syst. 2012, 10, 109–116. [Google Scholar]
  14. Zhang, Q.; Pless, R. Extrinsic Calibration of a Camera and Laser Range Finder. Proceedings of IEEE International Conference on Intelligent Robots and Systems(IROS), St. Louis, MO, USA, 28 September–2 October 2004; pp. 2301–2306.
  15. Herrera, C.D.; Kannala, J.; Heikkila, J. Accurate and Practical Calibration of a Depth and Color Camera Pair. Proceedings of International Conference on Computer Analysis of Images and Patterns(CAIP), Seville, Spain, 29–31 August 2011; pp. 437–445.
  16. Khoshelham, K.; Elberink, S.O. Accuracy and resolution of Kinect depth data for indoor mapping applications. Sensors 2012, 12, 1437–1454. [Google Scholar]
  17. Menna, F.; Remondino, F.; Battisti, R.; Nocerino, E. Geometric Investigation of a Gaming Active Device. Proceedings of SPIE, Prague, Czech Republic, 19–22 September 2011.
  18. Smisek, J.; Jancosek, M.; Pajdla, T. 3D with Kinect. IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 1154–1160.
  19. Park, J.H.; Bae, J.H.; Baeg, M.H. A data fitting technique for rational function models using the LM optimization algorithm. J. Inst. Contr. Robot. Syst. 2011, 17, 768–776. [Google Scholar]
  20. Camera Calibration Toolbox for Matlab. Available online: http://www.vision.caltech.edu/bouguetj/calib_doc/ (accessed on 15 October 2011).
  21. Leon-Garcia, A. Probability and Random Processes for Electrical Engineering, 2nd ed.; Addison-Wesley: Boston, MA, USA, 1994; pp. 242–246. [Google Scholar]
  22. Bay, H.; Ess, A.; Tuytelaars, T.; Gool, L.V. SURF: Speeded-up robust features. Comput. Vis. Image Understand. 2008, 110, 346–359. [Google Scholar]
Figure 1. Information from Kinect™ sensor. (a) Disparity map image; (b) RGB image; (c) Colored 3D point cloud data.
Figure 1. Information from Kinect™ sensor. (a) Disparity map image; (b) RGB image; (c) Colored 3D point cloud data.
Sensors 12 08640f1 1024
Figure 2. Relationship between the disparity image and the real depth information (disparity: 400–1,069, real depth: 0.5–17.3 m).
Figure 2. Relationship between the disparity image and the real depth information (disparity: 400–1,069, real depth: 0.5–17.3 m).
Sensors 12 08640f2 1024
Figure 3. Depth calibration results. (a) Fitting results (measurement data, Equation (1) model, fourth-order rational function model); (b) Residual (fitting error) results (Equation (1) model, fourth-order rational function model).
Figure 3. Depth calibration results. (a) Fitting results (measurement data, Equation (1) model, fourth-order rational function model); (b) Residual (fitting error) results (Equation (1) model, fourth-order rational function model).
Sensors 12 08640f3 1024
Figure 4. Visualization of a 3D Gaussian model. (a) Uncertainty ellipsoid for (xmx)TQ–1(xmx) = k. (b) Density of p.d.f. g(x). where ( m x = [ 0 0 0 ] , Q = [ 1.0 2 0 0 0 0.6 2 0 0 0 0.5 2 ] ).
Figure 4. Visualization of a 3D Gaussian model. (a) Uncertainty ellipsoid for (xmx)TQ–1(xmx) = k. (b) Density of p.d.f. g(x). where ( m x = [ 0 0 0 ] , Q = [ 1.0 2 0 0 0 0.6 2 0 0 0 0.5 2 ] ).
Sensors 12 08640f4 1024
Figure 5. Mapping relationship of uncertainty between the disparity image space and the real Cartesian space.
Figure 5. Mapping relationship of uncertainty between the disparity image space and the real Cartesian space.
Sensors 12 08640f5 1024
Figure 6. Uncertainty ellipsoids for all test cases.
Figure 6. Uncertainty ellipsoids for all test cases.
Sensors 12 08640f6 1024
Figure 7. (a) Uncertainty ellipsoid map in the entire measurable Cartesian space; (b) View of x-y plane direction; (c) View of y-z plane direction.
Figure 7. (a) Uncertainty ellipsoid map in the entire measurable Cartesian space; (b) View of x-y plane direction; (c) View of y-z plane direction.
Sensors 12 08640f7 1024
Figure 8. (a) Distribution of maximum standard deviation for various values of u and v, and fixed value of d (d = 800, real distance = 1.2 m); (b) View of x-y direction; (c) View of x-z direction; (d) View of y-z direction.
Figure 8. (a) Distribution of maximum standard deviation for various values of u and v, and fixed value of d (d = 800, real distance = 1.2 m); (b) View of x-y direction; (c) View of x-z direction; (d) View of y-z direction.
Sensors 12 08640f8 1024
Figure 9. (a) Distribution of maximum standard deviation for various values of u and d, and fixed value of v (v = 240, Vertical center of disparity image); (b) View of x-y direction; (c) View of x-z direction; (d) View of y-z direction.
Figure 9. (a) Distribution of maximum standard deviation for various values of u and d, and fixed value of v (v = 240, Vertical center of disparity image); (b) View of x-y direction; (c) View of x-z direction; (d) View of y-z direction.
Sensors 12 08640f9 1024
Figure 10. (a) Integrated distribution of the maximum standard deviation for various values of u and v, and three values of d (750, 800, 850); (b) Integrated distribution of the maximum standard deviation for various values of u and d, and three values of v (1, 140, 240).
Figure 10. (a) Integrated distribution of the maximum standard deviation for various values of u and v, and three values of d (750, 800, 850); (b) Integrated distribution of the maximum standard deviation for various values of u and d, and three values of v (1, 140, 240).
Sensors 12 08640f10 1024
Figure 11. Volume distribution of the maximum standard deviation for most of the disparity image space. (a) View of (AZ = –135°, EL = 45°); (b) View of (AZ = 120°, EL = 45°).
Figure 11. Volume distribution of the maximum standard deviation for most of the disparity image space. (a) View of (AZ = –135°, EL = 45°); (b) View of (AZ = 120°, EL = 45°).
Sensors 12 08640f11 1024
Figure 12. Experimental environment. (a) Colored point cloud representation; (b) Detected SURF visual features.
Figure 12. Experimental environment. (a) Colored point cloud representation; (b) Detected SURF visual features.
Sensors 12 08640f12 1024
Figure 13. Detected visual features in the experimental environment. (a) View of the disparity image space; (b) View of the real Cartesian space.
Figure 13. Detected visual features in the experimental environment. (a) View of the disparity image space; (b) View of the real Cartesian space.
Sensors 12 08640f13 1024
Figure 14. Histogram distribution of detected visual features in the disparity image space. (a) u data of visual features. (b) v data of visual features. (c) d data of visual features.
Figure 14. Histogram distribution of detected visual features in the disparity image space. (a) u data of visual features. (b) v data of visual features. (c) d data of visual features.
Sensors 12 08640f14 1024
Figure 15. Standard deviations of visual features and histogram; (a) Standard deviations for u of visual features; (b) Standard deviations for v of visual features; (c) Standard deviations for d of visual features; (d) Histogram distribution for (a); (e) Histogram distribution for (b); (f) Histogram distribution for (c).
Figure 15. Standard deviations of visual features and histogram; (a) Standard deviations for u of visual features; (b) Standard deviations for v of visual features; (c) Standard deviations for d of visual features; (d) Histogram distribution for (a); (e) Histogram distribution for (b); (f) Histogram distribution for (c).
Sensors 12 08640f15 1024
Figure 16. Uncertainty ellipsoids for all the visual features and the estimated covariance matrix in the disparity image space. (Cyan: ellipsoids for covariance matrix Ri, Magenta: ellipsoid for the estimated covariance matrix R).
Figure 16. Uncertainty ellipsoids for all the visual features and the estimated covariance matrix in the disparity image space. (Cyan: ellipsoids for covariance matrix Ri, Magenta: ellipsoid for the estimated covariance matrix R).
Sensors 12 08640f16 1024
Figure 17. 20 representative visual features in the Cartesian space. (a) 20 representative visual features highlighted among all the visual features; (b) 20 representative visual features and their uncertainty ellipsoids.
Figure 17. 20 representative visual features in the Cartesian space. (a) 20 representative visual features highlighted among all the visual features; (b) 20 representative visual features and their uncertainty ellipsoids.
Sensors 12 08640f17 1024
Figure 18. Distribution of 3D measurements for 4 visual features and their uncertainty ellipsoids. (a) id: 100; (b) id: 220; (c) id: 282; (b) id: 614.
Figure 18. Distribution of 3D measurements for 4 visual features and their uncertainty ellipsoids. (a) id: 100; (b) id: 220; (c) id: 282; (b) id: 614.
Sensors 12 08640f18 1024
Figure 19. Distribution of 3D measurements for the 20 selected visual features.
Figure 19. Distribution of 3D measurements for the 20 selected visual features.
Sensors 12 08640f19 1024
Table 1. Depth calibration parameters (4th-order rational function model).
Table 1. Depth calibration parameters (4th-order rational function model).
Polynomial orderConst. (d0)1st (d1)2nd (d2)3rd (d3)4th (d4)
Numerator P(d)452.705−611.068255.254−7.2957.346
Denominator Q(d)−326.149588.446−548.754340.178−47.175
Table 2. Pin-hole camera projection parameters of the depth camera.
Table 2. Pin-hole camera projection parameters of the depth camera.
ParameterfDepth,xfDepth,yCDepth,xCDepth,y
Value582.64586.97320.17260.00
Table 3. Covariance matrix Q, maximum deviation and maximum vector, uncertainty ellipsoids.
Table 3. Covariance matrix Q, maximum deviation and maximum vector, uncertainty ellipsoids.
Test caseux (mx) (mean)Q (Covariance matrix) Q 2 (max. deviation)v (max. vector)Uncertainty ellipsoid xT Q−1x = k
(a) [ 320 240 500 ] [ 0.0002 0.0199 0.5837 ] 10 5 × [ 0.1109 ε ε ε 0.0637 0.0057 ε 0.0629 0.1684 ]0.0013 [ 0.0008 0.0545 0.9985 ]Sensors 12 08640t1
(b) [ 330 200 700 ] [ 0.0149 0.0906 0.8861 ] 10 5 × [ 0.2557 0.0014 0.0139 0.0014 0.1548 0.0842 0.0139 0.0842 0.8236 ]0.0029 [ 0.0241 0.1230 0.9921 ]Sensors 12 08640t2
(c) [ 300 250 900 ] [ 0.0631 0.0311 1.8227 ] 10 3 × [ 0.0110 0.0001 0.0052 0.0001 0.0062 0.0025 0.0052 0.0025 0.1491 ]0.0122 [ 0.0373 0.0178 0.9991 ]Sensors 12 08640t3
(d) [ 100 150 600 ] [ 0.2666 0.1322 0.7054 ] 10 5 × [ 0.2098 0.0238 0.1268 0.0238 0.1044 0.0629 0.1268 0.0629 0.3354 ]0.0021 [ 0.5091 0.2003 0.8371 ]Sensors 12 08640t4
(e) [ 490 400 800 ] [ 0.3474 0.2842 1.1917 ] 10 4 × [ 0.0693 0.0189 0.0793 0.0189 0.0419 0.0649 0.0793 0.0649 0.2722 ]0.0056 [ 0.3137 0.2392 0.9189 ]Sensors 12 08640t5
(f) [ 80 360 920 ] [ 0.8402 0.3473 2.0383 ] 10 3 × [ 0.0531 0.0163 0.0959 0.0163 0.0145 0.0397 0.0959 0.0397 0.2327 ]0.0168 [ 0.3923 0.1587 0.9060 ]Sensors 12 08640t6
(g) [ 600 80 450 ] [ 0.2577 0.1646 0.5366 ] 10 5 × [ 0.1209 0.0174 0.0566 0.0174 0.0647 0.0361 0.0566 0.0361 0.1178 ]0.0014 [ 0.6613 0.2960 0.6893 ]Sensors 12 08640t7
(h) [ 180 450 700 ] [ 0.2132 0.2868 0.8861 ] 10 5 × [ 0.2557 0.0014 0.0139 0.0014 0.1548 0.0842 0.0139 0.0842 0.8236 ]0.0031 [ 0.2902 0.3398 0.8946 ]Sensors 12 08640t8

Share and Cite

MDPI and ACS Style

Park, J.-H.; Shin, Y.-D.; Bae, J.-H.; Baeg, M.-H. Spatial Uncertainty Model for Visual Features Using a Kinect™ Sensor. Sensors 2012, 12, 8640-8662. https://doi.org/10.3390/s120708640

AMA Style

Park J-H, Shin Y-D, Bae J-H, Baeg M-H. Spatial Uncertainty Model for Visual Features Using a Kinect™ Sensor. Sensors. 2012; 12(7):8640-8662. https://doi.org/10.3390/s120708640

Chicago/Turabian Style

Park, Jae-Han, Yong-Deuk Shin, Ji-Hun Bae, and Moon-Hong Baeg. 2012. "Spatial Uncertainty Model for Visual Features Using a Kinect™ Sensor" Sensors 12, no. 7: 8640-8662. https://doi.org/10.3390/s120708640

Article Metrics

Back to TopTop