A Rapid Coordinate Transformation Method Applied in Industrial Robot Calibration Based on Characteristic Line Coincidence

Coordinate transformation plays an indispensable role in industrial measurements, including photogrammetry, geodesy, laser 3-D measurement and robotics. The widely applied methods of coordinate transformation are generally based on solving the equations of point clouds. Despite the high accuracy, this might result in no solution due to the use of ill conditioned matrices. In this paper, a novel coordinate transformation method is proposed, not based on the equation solution but based on the geometric transformation. We construct characteristic lines to represent the coordinate systems. According to the space geometry relation, the characteristic line scan is made to coincide by a series of rotations and translations. The transformation matrix can be obtained using matrix transformation theory. Experiments are designed to compare the proposed method with other methods. The results show that the proposed method has the same high accuracy, but the operation is more convenient and flexible. A multi-sensor combined measurement system is also presented to improve the position accuracy of a robot with the calibration of the robot kinematic parameters. Experimental verification shows that the position accuracy of robot manipulator is improved by 45.8% with the proposed method and robot calibration.


Introduction
Multi-sensor measurement systems usually have different coordinate systems. The original data must be transformed to a common coordinate system for the convenience of the subsequent data acquisition, comparison and fusion [1,2]. The transformation of coordinate systems is applied in many fields, especially vision measurement and robotics. For example, two images need to have a unified coordinate system for image matching [3]. In camera calibration, the coordinate systems of the image plane and the object plane need to be unified for the inner parameter calculation [4]. In robot systems, the coordinate system of the robot twist needs to be transformed to the tool center position (TCP) to obtain the correct pose of robot manipulators [5,6]. A minor error introduced by an imprecise coordinate transformation could cause problems such as the failure of image matching and track breaking [1]. Especially in an error accumulating system such as series industry robots, the coordinate transformation error would accumulate in each step and thereby decrease the position accuracy of the robot manipulator. Therefore, research on coordinate transformation has been of interest to researchers in recent years.

Online Calibration System of Robot
Industrial robots have the characteristics of high repeatability positioning accuracy and low absolute positioning accuracy. This is due to the structure of the robot, manufacturing errors, kinematic parameter error and environmental influence [10]. To improve the absolute positioning accuracy of the robot, the use of an external sensor to measure the position of the robot manipulator it any effective approach. This paper proposes an on-line calibration system for the kinematic parameters of the robot using a laser tracker and a close-range photogrammetric system, as Figure 1 shows. According to the differential equations constructed by the kinematic parameters of each robot axis, the final mathematic model of kinematic parameters of the robot is established. The position errors of the robot manipulator are obtained by comparing the coordinates in the robot base coordinate system and the measurement sensor system. Then, the errors, including the coordinate transformation error, target installation error and position and angle errors of the robot kinematic parameters, are separately corrected. In the robot calibration, on the one hand, the coordinate transformation error directly affects the final error correction of the kinematic parameters. On the other hand, the coordinate systems of sensors are often required to transform in the on-line combined measurement system. Therefore, the premise of obtaining the position errors of a robot manipulator is to unify the coordinate systems of the various measurement sensors by an accurate, fast and stable coordinate transformation algorithm.
Sensors 2016, 16,239 3 of 15 coordinate system and the measurement sensor system. Then, the errors, including the coordinate transformation error, target installation error and position and angle errors of the robot kinematic parameters, are separately corrected. In the robot calibration, on the one hand, the coordinate transformation error directly affects the final error correction of the kinematic parameters. On the other hand, the coordinate systems of sensors are often required to transform in the on-line combined measurement system. Therefore, the premise of obtaining the position errors of a robot manipulator is to unify the coordinate systems of the various measurement sensors by an accurate, fast and stable coordinate transformation algorithm. In combination with the characteristics of the robot, we propose a practical coordinate transformation method. It extracts the characteristic lines from the point clouds in different coordinate systems. According to the theory of space analytic geometry, the rotation and translation parameters needed for the coincidence of the characteristic line scan be calculated. Then, the coordinate transformation matrix is calculated. The coincidence of the characteristic lines represents the coincidence of the point clouds as well as the coincidence of the two coordinate systems. This method has some advantages. First, it does not require the solution of equations and complex calculations. Second, because the transformation matrix is obtained from the space geometry relationships, it would not be affected by robot errors or other environmental factors. The result is accurate and stable. Third, it does not require a sensor with a large field of view. Fourth, the algorithm is small and fast without occupying processor time and resources, and can be integrated into the host computer program. It could be applied easily in measurement coordinate systems that often need to change.

Method of Coordinate Transformation
Suppose that S is a cubic point cloud in space. Point cloud M is the form of S located in the coordinate system of the sensor OSXSYSZS. N is the form of S located in the robot base coordinate system OrXrYrZr. M' represents the point cloud M transformed from the coordinate system of the sensor OSXSYSZS to the robot base coordinate system OrXrYrZr with the transformation matrix TS r . The difference between N and M' is the transformation error caused by the transfer matrix TS r . Then, the coincidence of the coordinate systems OSXSYSZS and OrXrYrZr can be expressed as the coincidence of the two point clouds N and M'. For simplifying this mathematical model of the transformation process, we establish several characteristic lines instead of each point cloud. As verified by experiment, at least two characteristic lines are required to ensure the transformation accuracy. In combination with the characteristics of the robot, we propose a practical coordinate transformation method. It extracts the characteristic lines from the point clouds in different coordinate systems. According to the theory of space analytic geometry, the rotation and translation parameters needed for the coincidence of the characteristic line scan be calculated. Then, the coordinate transformation matrix is calculated. The coincidence of the characteristic lines represents the coincidence of the point clouds as well as the coincidence of the two coordinate systems.
This method has some advantages. First, it does not require the solution of equations and complex calculations. Second, because the transformation matrix is obtained from the space geometry relationships, it would not be affected by robot errors or other environmental factors. The result is accurate and stable. Third, it does not require a sensor with a large field of view. Fourth, the algorithm is small and fast without occupying processor time and resources, and can be integrated into the host computer program. It could be applied easily in measurement coordinate systems that often need to change.

Method of Coordinate Transformation
Suppose that S is a cubic point cloud in space. Point cloud M is the form of S located in the coordinate system of the sensor O S X S Y S Z S . N is the form of S located in the robot base coordinate system O r X r Y r Z r . M' represents the point cloud M transformed from the coordinate system of the sensor O S X S Y S Z S to the robot base coordinate system O r X r Y r Z r with the transformation matrix T S r .
The difference between N and M' is the transformation error caused by the transfer matrix T S r . Then, the coincidence of the coordinate systems O S X S Y S Z S and O r X r Y r Z r can be expressed as the coincidence of the two point clouds N and M'. For simplifying this mathematical model of the transformation process, we establish several characteristic lines instead of each point cloud. As verified by experiment, at least two characteristic lines are required to ensure the transformation accuracy. In Figure 2, two points A 1 and A 2 are chosen to be linked to the characteristic line A. Points B 1 and B 2 form characteristic line B. Similarly, in point cloud N, the corresponding points A 1 'and A 2 ' form line A', and points B 1 ' and B 2 ' form line B'. To achieve the coincidence of lines A and A', line A must be rotated around an axis in space. The rotated axis is the vector C which is perpendicular to the plane constructed by lines A and A'. As Figure 3 shows, the process of a vector rotating around an arbitrary axis can be divided into a series of rotations around the axis X, Y, Z. The following are the decomposition steps. In Figure 2, two points A1 and A2 are chosen to be linked to the characteristic line A. Points B1 and B2 form characteristic line B. Similarly, in point cloud N, the corresponding points A1'and A2' form line A', and points B1' and B2' form line B'. To achieve the coincidence of lines A and A', line A must be rotated around an axis in space. The rotated axis is the vector C which is perpendicular to the plane constructed by lines A and A'. As Figure 3 shows, the process of a vector rotating around an arbitrary axis can be divided into a series of rotations around the axis X, Y, Z. The following are the decomposition steps. Take the first coincidence of Lines A and A' as an example: (a) Translate the rotation axis to the coordinate origin. The corresponding transformation matrix can be calculated as: In Figure 2, two points A1 and A2 are chosen to be linked to the characteristic line A. Points B1 and B2 form characteristic line B. Similarly, in point cloud N, the corresponding points A1'and A2' form line A', and points B1' and B2' form line B'. To achieve the coincidence of lines A and A', line A must be rotated around an axis in space. The rotated axis is the vector C which is perpendicular to the plane constructed by lines A and A'. As Figure 3 shows, the process of a vector rotating around an arbitrary axis can be divided into a series of rotations around the axis X, Y, Z. The following are the decomposition steps. Take the first coincidence of Lines A and A' as an example: (a) Translate the rotation axis to the coordinate origin. The corresponding transformation matrix can be calculated as: 1 0 0 a 0 0 1 0 b 0 ( ,y ,z ) 1 1 1 0 0 1 c 0 where, (a0, b0, c0) is the coordinates of the center point of line A.
(b) Rotate the axis α1 degrees to Plane XOZ. Take the first coincidence of Lines A and A' as an example: (a) Translate the rotation axis to the coordinate origin. The corresponding transformation matrix can be calculated as: (2) α 1 is the angle between the axis and plane XOZ. It can be obtained by cosα 1 " , where, (a 1 , b 1 , c 1 ) are the coordinates of vector C, as Figure 3b shows.
(c) Rotate the axis β 1 degrees to coincide with Axis Z.
where, β 1 is the angle between the rotation axis and axis Z. It can be obtained by Rotate the axis θ 1 degrees around Axis Z, as shown in Figure 3d.
where θ 1 is the angle between lines A and A', which can be obtained by .
(e) Rotate the axis by reversing the process of Step (c) where, β 1 is as the same as in step (c). (f) Rotate the axis by reversing the process of Step (b).
where, α 1 is as the same as in step (b).
(g) Rotate the axis by reversing the process of Step (a) where, (a 0 , b 0 , c 0 ) is as the same as in step (a).
Combining all of the previous steps, the final transformation matrix Trt 1 of the first parallel (lines A and A') is expressed as: Through the rotation matrix Trt 1 calculated by Equation (8), the points P i (x, y, z) in point cloud M can generate a new point cloud M 1 by Equation (9).
Then, the characteristic line A of the new point cloud M 1 is parallel with the characteristic line A' of point cloud N, as Figure 4a shows. (g) Rotate the axis by reversing the process of Step (a) where, (a0, b0, c0) is as the same as in step (a). Combining all of the previous steps, the final transformation matrix 1 Trt of the first parallel (lines A and A') is expressed as: Through the rotation matrix Trt1 calculated by Equation (8), the points Pi(x, y, z) in point cloud M can generate a new point cloud M1 by Equation (9).
Then, the characteristic line A of the new point cloud M1 is parallel with the characteristic line A' of point cloud N, as Figure 4a shows.
Through the rotation matrix Trt2, the points Pi(x, y, z) in point cloud M1 can generate a new point cloud M2 again by Equation (9). Then, the characteristic line B of the new point cloud M2 is parallel with the characteristic line B' of point cloud N, as Figure 4b shows.
Since  Trt 2 " Tp´x 2 ,´y 2 ,´z 2 q¨Rxp´α 2 q¨Rypβ 2 q¨Rzpθ 2 q¨Ryp´β 2 q¨Rxpα 2 q¨Tpx 2 , y 2 , z 2 q Through the rotation matrix Trt 2 , the points P i (x, y, z) in point cloud M 1 can generate a new point cloud M 2 again by Equation (9) (1)-(8): The points P i (x,y,z) in point cloud M 2 can generate a new point cloud M 3 by Equation (9), which are parallel with the point cloud N, as Figure 4c shows. In order to make coincident the point cloud M 3 and point cloud N, the translation matrix Tr needs to be calculated by the two center points of Line A and A'. The new point cloud M' can be generated after translated by Tr. Therefore, through a series of simple rotations and translation, the two point clouds N and M' are coincident, as Figure 4d shows. The final transformation matrix is shown as Equation (12). The result, as a necessary preparation step, can then be used in robot calibration:

Method of Robot Calibration
The actual kinematic parameters of the robot deviate from their nominal values, which is referred to as kinematic errors [10]. The kinematic parameter calibration of a robot is an effective way to improve the absolute position accuracy of the robot manipulator. A simple robot self-calibration method based on the D-H model is described as follows. Reference [20] gives a more detailed description.
is the pose of a certain point in the coordinate system of the photogrammetric system, where r 1p " r 9 p are the attitude parameters and p xp " p zp are the position parameters. Through transformation from the coordinate system of the measurement sensor can be obtained by Equation (12): where, Trt is the transformation matrix, which can be obtained by the method described in Section 3.1.
Given the six DOF robot in the lab, the transformation matrix from the robot tool coordinate system to the robot base coordinate system is expressed as: In this system, the cooperation target of the measurement sensor, which is set up at the end axis of the robot, should be considered as an additional axis, Axis 7. Then, the transformation matrix from Axis 6 to Axis 7 is: where t x , t y , t z are the translation vectors, which can be measured previously. Therefore, according to the kinematic model of the robot, the typical coordinates of the robot manipulator in the robot base where B t is the point pose in the robot tool coordinate system, and B o is the point pose from the robot tool coordinate system to the robot base coordinate system. In the robot calibration, the kinematic parameters are the most significant impact factors, which usually means the link parameters of the robot. In the D-H model, the link parameters include the length of the link a, the link angle α, the joint displacement d and the rotation angle of the joint θ. With the disturbances of the four link parameters, the position error matrix for adjacent robot axes dT i i´1 can be expressed as: where ∆θ i ,∆α i ,∆a i and ∆d i are the small errors of link parameters. Suppose that A qi " pT i i´1 q´1¨B where, q represents the link parameters (a, d, α, θ). If every two adjacent axes are influenced by the link parameters, the transformation matrix from the robot base coordinate system to the coordinate system of the robot manipulator can be expressed as: where, T N 0 is the typical transformation matrix from the robot base coordinate system to the coordinate system of the robot manipulator and dT N 0 is the error matrix caused by the link parameters. Through expanding dT N 0 and performing a large number of simplifications and combinations, Equation (18) can be simplified as: Suppose that k iq " T i 0 A qi T N 1 , where, q represents the four link parameters. The position error of the robot manipulator can be simplified as given in Equation (20): In this paper, because the DOF of the series robot is 6, ∆q i " r∆θ 1 " ∆t z s includes 24 kinematics parameters of the robot a 1 " a 6 , d 1 " d 6 , α 1 " α 6 , θ 1 " θ 6 and three translation error variables of T 6 7 . Therefore, there are 27 parameters of the robot that need to be calibrated. In Equation (20), the left side of equation is the position error at each point, as measured by the measurement sensor, and the right side is the kinematics errors that need to be corrected. These errors can be revised by the least squares method in the generalized inverse matrix sense.

Experiments and Analysis
Through the designed experiments, we show how to use the proposed coordinate transformation method to achieve the coordinate transformation of the on-line robot calibration system. Using verification experiments, we determine the result of the robot calibration using the proposed method. For evaluating the performance of the proposed method, it is compared with four other common methods of coordinate transformation under the same experimental conditions.

Coordinate Transformationin an On-line Robot Calibration System
The on-line robot calibration system we constructed includes an industrial robot, a photographic system and a laser tracker as shown in Figure 4. The model of the robot in lab is the KR 5 arc from KUKA Co. Ltd. (Augsburg, Germany), one of the world's top robotic companies. Its arm length is 1.4 m and the working envelope is 8.4 m 3 . For covering most of the robot working range, the close range photogrammetric system in the lab, TENYOUN 3DMoCap-GC130 (Beijing, China), requires a field of view of more than 1 mˆ1 mˆ1 m without any dead angle. To achieve the goal of on-line measurement, a multi-camera system is needed. We used a multi-camera system symmetrically formed by four CMOS cameras with fixed focal lengths of 6 mm. The laser tracker in the lab, FARO Xi from FARO Co, Ltd. (Lake Mary, FL, USA) is a well-known high accuracy instrument whose absolute distance measurement (ADM) is 10 µm˘1.1 µm/mL. The laser beam can easily be lost in tracking because of barriers or the acceleration of the target, which would cause minor errors. Therefore, we combine the laser tracker with the photographic system to improve the measurement accuracy and stability and thereby make full use of the advantages of the high accuracy of the laser tracker and the free light-of-sight of the photographic system. After proper data fusion, the two types of data from the photographic system and the laser tracker can be gathered together. The method of data fusion and the verified experimental result are detailed in reference [21]. In the experiment, 80 points in the public field of the robot and the photogrammetric system are picked to build a cube of 200 mm2 00 mmˆ200 mm. The reason for building a cube is to facilitate the selection of characteristic lines and the calculation of coincidence parameters. The two targets of the photogrammetric system and laser tracker are installed together with the end axis of the robot by a multi-faced fixture. To obtain accurate and stable data, the robot stops for 7 s at each location, and the sensors measure each point 20 times, providing an adequate measurement time for the photographic system and laser tracker. The experimental parameters of the photogrammetric system are an exposure time of 15 us, a frequency of 10 fps and a gain of 40, based on experience. According to Equations (1)-(7) and the experimental data, we can obtain the parameters of the transformation matrix shown in Table 1, where a i -θ i . are the parameters for the coincidence of characteristic lines in Equations (1)-(7).  where, Trt rp is the transformation matrix from the robot base coordinate system to the coordinate system of the photogrammetric system. Trt rl is the transformation matrix from the robot base coordinate system tothe coordinate system of the laser tracker. By means of the above transformation matrix, we can obtain the point cloud coordinates transformed from the coordinate system of the robot to that of the sensors by Equation (12). Both the origin coordinates before and after transformation as well as the transformation error are shown in Table 2, where, P x ,P y ,P z and R x ,R y ,R z are three components of the original coordinates in two different coordinate systems. T x ,T y ,T z are the coordinates of points transformed from the robot base coordinate system to the sensor coordinate system, and ∆x, ∆y, ∆ z are the three components of the transformation error. It is can be calculated from Table 2 that the average values of the transformation error between the coordinate systems of the robot and photogrammetric system are ∆ x = 0.106 mm, ∆y =´0.062 mm and ∆z = 0.013 mm. The average values of the transformation error between the coordinate systems of the robot and laser tracker are ∆x =´0.015 mm, ∆y = 0.041 mm and ∆z =0.023 mm. Figures 5 and 6 show that the transformation error of the photogrammetric system is approximately 10 times greater than that of the laser tracker. As in the earlier presentation, the nominal measurement accuracy of the photogrammetric system is 10´2 mm and that of the laser tracker is 10´3 mm. The results illustrate that the transformation accuracy has the same order of magnitude as that of the measurement sensor. This indicates that the transformation error is so small that it would not influence the accuracy of the sensors. The transformation method can also make the error distribution of the low precision sensor more uniform to improve the transformation accuracy and the accuracy of the robot calibration.
the transformation accuracy has the same order of magnitude as that of the measurement sensor. This indicates that the transformation error is so small that it would not influence the accuracy of the sensors. The transformation method can also make the error distribution of the low precision sensor more uniform to improve the transformation accuracy and the accuracy of the robot calibration.

Position Error of Robot after Coordinate Transformation and Calibration
Experiments are designed to calibrate the kinematic parameters of the robot. The measurement system is shown in Figure 7. Sixty points in space are used for the calibration. After the coordinate transformation, the position errors between the sensor and the robot manipulator are obtained. A constraint method based on the minimum distance error is adopted to calibrate the robot kinematic parameters [20,21]. Twenty seven kinematic parameters, including 24 link parameters and three parameters of the fixture, are corrected. Then, 60 correct positions are calculated using the calibrated robot kinematic parameters. To evaluate the performance of the proposed method, we use a group of calibration results, which adopts a different coordinate transformation method and the same robot calibration algorithm, as a comparison. The position errors of the robot after the coordinate transformation and calibration are shown in Figure 8. δ1 is the position error after the coordinate transformation with the other method [19], and δ2 is the position error after the coordinate transformation with the proposed method (Characteristic Line method). the transformation accuracy has the same order of magnitude as that of the measurement sensor. This indicates that the transformation error is so small that it would not influence the accuracy of the sensors. The transformation method can also make the error distribution of the low precision sensor more uniform to improve the transformation accuracy and the accuracy of the robot calibration.

Position Error of Robot after Coordinate Transformation and Calibration
Experiments are designed to calibrate the kinematic parameters of the robot. The measurement system is shown in Figure 7. Sixty points in space are used for the calibration. After the coordinate transformation, the position errors between the sensor and the robot manipulator are obtained. A constraint method based on the minimum distance error is adopted to calibrate the robot kinematic parameters [20,21]. Twenty seven kinematic parameters, including 24 link parameters and three parameters of the fixture, are corrected. Then, 60 correct positions are calculated using the calibrated robot kinematic parameters. To evaluate the performance of the proposed method, we use a group of calibration results, which adopts a different coordinate transformation method and the same robot calibration algorithm, as a comparison. The position errors of the robot after the coordinate transformation and calibration are shown in Figure 8. δ1 is the position error after the coordinate transformation with the other method [19], and δ2 is the position error after the coordinate transformation with the proposed method (Characteristic Line method).

Position Error of Robot after Coordinate Transformation and Calibration
Experiments are designed to calibrate the kinematic parameters of the robot. The measurement system is shown in Figure 7. Sixty points in space are used for the calibration. After the coordinate transformation, the position errors between the sensor and the robot manipulator are obtained. A constraint method based on the minimum distance error is adopted to calibrate the robot kinematic parameters [20,21]. Twenty seven kinematic parameters, including 24 link parameters and three parameters of the fixture, are corrected. Then, 60 correct positions are calculated using the calibrated robot kinematic parameters. To evaluate the performance of the proposed method, we use a group of calibration results, which adopts a different coordinate transformation method and the same robot calibration algorithm, as a comparison. The position errors of the robot after the coordinate transformation and calibration are shown in Figure 8. δ 1 is the position error after the coordinate transformation with the other method [19], and δ 2 is the position error after the coordinate transformation with the proposed method (Characteristic Line method).  In position measurement, the distance between two points is often used to evaluate the position accuracy, which is called the root mean square (RMS) error expressed by: It can be indicated from Figure 5    In position measurement, the distance between two points is often used to evaluate the position accuracy, which is called the root mean square (RMS) error expressed by: It can be indicated from Figure   In position measurement, the distance between two points is often used to evaluate the position accuracy, which is called the root mean square (RMS) error expressed by: It can be indicated from Figure 5 that the average RMS using the other coordinate transformation method is δ 1 " 0.436 mm. while the average RMS using the proposed method is δ 2 " 0.200 mm. The position accuracy is improved by 45.8% using the Characteristic Line method.
To evaluate the accuracy distribution of the robot for different areas of the working range, a new set of testing data are utilized in a demonstration experiment. The coordinates of the center of robot calibration region O is (750, 0, 1000) in the robot base coordinate system. Taking Table 3. It is indicated from Table 3 that the average RMS of robot position error within the calibration region is 0.200 mm. The position error outside the calibration region is about 0.323 mm. It is proved that the calibration accuracy isn't consistent in the whole working range of robot. Therefore, this calibration method is more applicable to a smaller working range of the robot.

Accuracy of Coordinate Transformation Method
To obtain the accuracy of the proposed coordinate transformation method, an experiment is designed using the laser tracker. The laser tracker is placed at two different stations to measure five common points. These common points should be collinear; otherwise, the Jacobi matrix will have a rank defect of one and be singular. The choice of five points is for improving the transformation accuracy. After the unification of the coordinate systems by the proposed method, the measurement results of the laser tracker are compared. Then, the accuracy of the proposed transformation method (Characteristic Line method) can be calculated, as Table 3 shows.
In addition, for evaluating the performance of the proposed method, we calculate the transformation accuracy of four methods on the same points: the Three Point method, Rodrigo Matrix method, Singular Value Decomposition method and Quaternion method. The measurement data of five public points measured by the laser tracker from two different stations are substituted into the four algorithms. Then, the rotation matrix R and translation matrix T can be calculated, and the coordinate transformation matrix Trt are also obtained. According to Equation (22), the new coordinates after transformation by the Trt can be generated as below: "¨r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33‹ where (x i , y i , z i ) s1 is the coordinate of the ith point measured by the laser tracker at Station 1. (x i ', y i ', z i ') s2 are the new coordinate at Station 2 generated by the coordinate transformation matrix Trt. R is the rotation matrix, r 11~r33 are the rotation parameters, T is the translation matrix, and (x 0 , y 0 , z 0 ) are the translation parameters. Compared with the actual data measured by the laser tracker in Station 2, the error of the coordinate transformation can be obtained. We also use the root mean square (RMS) of the transformation error of the five public points to describe the transformation accuracy. The experimental results are shown in Table 4 and Figure 9.  The following conclusions can be drawn by the comparison of the results with different algorithms. The accuracy of the Three-Point method is the lowest and its solution depends on the choice of public points. But its algorithm is simple and has the lowest time cost. The Rodrigo Matrix method has the highest accuracy, but the computation of the matrix might be the most complex. It also takes a long time for the calculation. The accuracies of the Singular Value Decomposition method and the Quaternion method are relatively high. The calculation of matrix in the two algorithms are simple and time saving. But they may not be able to work out the rotation matrix when the points are dense. The Characteristic Line method has the same level of accuracy as the Singular Value Decomposition method and the Quaternion method. Its algorithm are simple and stable, and its execution time is short as well. In addition, as Figure 9 shows, it can suppress the large error terms of the other four methods, and does not cumulate errors with the increase of the number of the common points like the other four methods do.

Conclusions
This paper proposes a simple method of coordinate transformation in a multi-sensor combination measurement system for use in the field of industrial robot calibration. It does not require a large amount of computation or a large field of view and is not affected by the errors of the robot. It operates fast, therefore, it can be integrated into the host computer program. It can be applied in cases where the coordinate systems change often. As verified by experiments, the accuracy of the transformation method is 10 −3 mm. It reduces the cumulative error of the coordinate transformation in the robot calibration, thereby improving the accuracy of the robot calibration. To evaluate its performance, the accuracy is compared with four common coordinate transformation methods. The experimental results show that the proposed method has the same high accuracy as the Singular The following conclusions can be drawn by the comparison of the results with different algorithms. The accuracy of the Three-Point method is the lowest and its solution depends on the choice of public points. But its algorithm is simple and has the lowest time cost. The Rodrigo Matrix method has the highest accuracy, but the computation of the matrix might be the most complex. It also takes a long time for the calculation. The accuracies of the Singular Value Decomposition method and the Quaternion method are relatively high. The calculation of matrix in the two algorithms are simple and time saving. But they may not be able to work out the rotation matrix when the points are dense. The Characteristic Line method has the same level of accuracy as the Singular Value Decomposition method and the Quaternion method. Its algorithm are simple and stable, and its execution time is short as well. In addition, as Figure 9 shows, it can suppress the large error terms of the other four methods, and does not cumulate errors with the increase of the number of the common points like the other four methods do.

Conclusions
This paper proposes a simple method of coordinate transformation in a multi-sensor combination measurement system for use in the field of industrial robot calibration. It does not require a large amount of computation or a large field of view and is not affected by the errors of the robot. It operates fast, therefore, it can be integrated into the host computer program. It can be applied in cases where the coordinate systems change often. As verified by experiments, the accuracy of the transformation method is 10´3 mm. It reduces the cumulative error of the coordinate transformation in the robot calibration, thereby improving the accuracy of the robot calibration. To evaluate its performance, the accuracy is compared with four common coordinate transformation methods. The experimental results show that the proposed method has the same high accuracy as the Singular Value Decomposition method and the Quaternion method. It can suppress large error terms and does not cumulate errors. Therefore, this method has the advantages of being simple and fast and exhibiting high accuracy and stability. It should have a wide range of application in the field of industrial combination measurement.