Abstract
As one of the key components for active compliance control and human–robot collaboration, a six-axis force sensor is often used for a robot to obtain contact forces. However, a significant problem is the distortion between the contact forces and the data conveyed by the six-axis force sensor because of its zero drift, system error, and gravity of robot end-effector. To eliminate the above disturbances, an integrated compensation method is proposed, which uses a deep learning network and the least squares method to realize the zero-point prediction and tool load identification, respectively. After that, the proposed method can automatically complete compensation for the six-axis force sensor in complex manufacturing scenarios. Additionally, the experimental results demonstrate that the proposed method can provide effective and robust compensation for force disturbance and achieve high measurement accuracy.
1. Introduction
Nowadays, robots have become indispensable equipment in industrial manufacturing systems [1]. With the urgent demand for flexible manufacturing and fast development of sensor technology, higher requirements are put forward for the intelligence and self-adaptability of robots [2]. Two prominent issues are robot-environment interaction [3] and human–robot collaboration [4,5]. The development of a multi-axis force sensor, especially a six-axis force sensor, makes it possible to obtain force information in robot operation environments. Additionally, six-axis force sensors have been applied to robotic systems that perform contact tasks, such as grinding robots, surgical robots, etc. [6,7].
In order to detect six-axis forces and moments in space, a cylindrical six-axis force sensor [8,9] is often used, which can measure three-axis orthogonal forces and moments relative to the sensor frame, i.e., the force components along the axes of the sensor and the moments around the axes of the sensor. When external forces are applied to the sensor, the internal elastomers are deformed and the corresponding strain signal according to its electrical characteristics are output [10,11,12]. Typically, a six-axis force sensor is mounted between the end flange of an industrial robot and the end-effector to perceive the external forces [13,14]. In this case, the sensor generally measures a combination of forces including the contact forces, the gravity force acting on the tool, self-gravity, system error, and other force disturbances. The tool gravity and self-gravity continuously affect measurement output as the robot pose changes. The self-gravity refers to the gravitational force acting on the six-axis force sensor. To accurately obtain the contact forces applied to the six-axis force sensor, it is necessary to compensate for force disturbance.
For several decades, researchers have been conducting wide studies to compensate for force disturbances of a six-axis force sensor at the end of a robot. Shetty B R et al. [15,16,17]. showed that the output of the sensor consisted of load gravity and interaction force, and derived an algorithm to estimate the effect of load gravity on the sensor output as the robot pose changed in real time. However, the algorithm was proposed under the assumption that the load gravity and the load centre of gravity are known, which is not always the case in practice. Loske et al. [18]. proposed an algorithm that considers the self-gravity of the sensor to be known and calculated the centre of self-gravity from the data of several poses in order to accomplish the compensation of the gravity. Lin et al. [19,20]. considered that the effect of gravity on the moment values can be ignored. Then, the output of the sensor in the six special attitudes were measured, and some parameters of the force sensor were initialized accordingly. Eventually, the gravity was decomposed when the sensor attitude changed, and thus the initial force and moment values were compensated. Li et al. [21]. completed the gravity compensation for the force interaction device using Lagrange’s equation. Vougioukas [22] assumed that the Z-axis of the robot base frame was in the same direction as gravity, and the gravity applied to one of the sensor’s axes was cancelled out by some special poses to calculate the bias of the force sensor. Finally, the load gravity and load centre of gravity positions were calculated by the force coordinate transformation and the least squares method. Taking into account the load gravity, the zero-point of the force sensor and the robot mounting inclination. Zhang et al. [23] proposed a method that used sensor data from no less than three robot poses to obtain the required parameters at once. However, the method assumed the zero-point as a constant and did not fully take into account the systematic error of the force sensor. Zhang et al. [24] used a deep learning algorithm to obtain the mapping relationship between the robot’s pose and the sensor output to complete the numerical compensation of a tandem force sensor. However, this method considers the tool gravity and sensor self-gravity as a whole, which requires data re-collection to train the model where the end-effector needs to be replaced. In addition, the method gradually fails when the end-effector wears out. This has been shown to be time-consuming and inefficient. Dine et al. [25]. presented a recurrent neural network observer to estimate the force disturbance due to gravity, inertia, centrifugal, and Coriolis forces. The method can detect external contact force-moment in a variety of highly dynamic motions. However, the result of the observer is unreliable when the robot is motionless, as the authors pointed out.
In order to solve the above problem, this paper proposes an integrated compensation method for the force disturbance of a six-axis force sensor, which combines deep learning and least squares. This algorithm separates the six-axis force sensor from the end-effector, and fully considers the self-gravity, drift, and system error of the six-axis force sensor. Firstly, the zero-point of the sensor is estimated based on deep learning to eliminate the interference caused by the above factors. Secondly, the load of the end-effector is identified using the least squares method. Finally, the influence on the output due to factors such as the gravity force acting on the tool and the robot mounting inclination can be eliminated based on model derivation, which is convenient for the cases where the end-effector needs to be replaced under complex manufacturing scenarios.
2. Problem Statement
A six-axis force sensor is often used for a robot to obtain contact force information. However, the self-gravity of the sensor and the gravity force acting on the end-effector will have an effect on the output of the sensor, and the influence changes continuously with the change of the sensor attitude. In addition, due to the inherent characteristics of electronic components, drift currents are changing, which affects the output of the sensor. Based on the above factors, the output of the six-axis force sensor differs from the pure contact force applied to the end-effector, and its output can be described as:
where is the drift of the sensor, is the system error, is the effect caused by the self-gravity of the sensor, is the effect caused by the gravity force acting on the tool, is the force output due to robot vibration and inertia, is the pure contact force applied to the end-effector. In addition to the pure contact force, , , and are parameters related to the sensor itself, relies on the selected execution tool. Since the robot is in low speed, the effect of is not considered in this paper. Let the effects caused by the sensor itself be attributed to . Therefore, Equation (1) is rewritten as:
where is the output of the sensor without any external force, which is defined as the zero-point. The goal of the method is to calculate the zero-point of the sensor for different poses of the robot and the effect of the tool gravity on the output. Based on the above calculation results, the effect of force interference factors on the contact force perception is eliminated.
3. Integrated Compensation Method
In view of the effects of drift and system error, as well as the interference of the force sensor self-gravity and end-effector gravity, we need to compensate for the actual output of the force sensor in order to accurately perceive the contact forces. Therefore, we separate the force sensor and the end-effector, use deep learning to obtain the mapping relationship between the attitude and the zero-point of the force sensor, and use the least squares method to estimate the tool load. Then, the compensation of the force sensor is completed, and the influence of force interference is eliminated.
3.1. Zero-Point Estimation of a Six-Axis Force Sensor Based on Deep Learning
The defined zero-point includes the drift, system error, and the influence of its self-gravity. Generally, for simplicity, the drift is often regarded as a constant, and a linear function is established based on the relationship between the self-gravity and the posture. However, in real-life application scenarios of robots, the established linear relationship does not fully reflect the relationship between the posture and output as well as the effects of drift and system error due to the installation and positioning of force sensors, etc. Deep learning is a machine learning technique that has made significant progress in the past decade. With sufficient training data and suitable network structures, neural networks (NNs) can approximate arbitrary nonlinear functions, and this powerful fitting ability has been widely applied to natural language processing and image recognition [26,27,28,29]. Studies have shown that neural networks have a better effect on error compensation [30,31]. Given its capability described above, NN is used in this section to estimate the zero-point of the force sensor.
Thus, it is necessary to establish a mapping between the pose of the force sensor and the zero-point in the absence of any load on the force sensor, i.e.,
where is the model parameter.
The pose of a rigid body can generally be described by rotation vector, rotation matrix, quaternion or Euler angles, etc. To reduce the computational complexity, the literature [15,16,22,23] chose a rotation matrix to derive the relationship between the effect of the self-gravity and the rotation matrix in the ideal case. According to the existing relationship, this paper designs an NN as shown in Figure 1 to estimate the zero-point.
Figure 1.
The neural network for zero-point estimation of a six-axis force sensor. A [0] is the input layer; A [1] is the first hidden layer; A [2] is the second hidden layer; A [3] is the output layer.
The input to the NN is the nine elements of the rotation matrix , and the general robot control system can directly obtain the data related to the pose and thus the rotation matrix can be calculated. Considering the symmetry of the force sensor, the theoretical output of should be 0 in the non-load condition. Additionally, its output is small and fluctuates irregularly during the experiment. To reduce its unreliable effect on model training, the output data of the NN are the five outputs of the force sensor, and the fluctuating output of caused by incidental factors is not estimated. Theoretically, the accuracy of approximation can be improved by increasing the number of neurons by a moderate amount [32]. However, too many neurons require higher computation power, which inevitably affects the response time of the sensor in practice. Therefore, with a combination of fitting accuracy and sensor response speed, we chose a neural network with two hidden layers based on the experiments, the first with 45 nodes and the second with 90, as the regressor to fit the mapping function . Currently, is often recommended as the activation function of NN, but in the force sensor application scenario, this activation function may cause some neurons to never be activated, resulting in the parameters not being updated. Therefore, this paper chooses the derivative of this function as the activation function, where . The activation function of the output layer uses the constant function. The parameter in the model is learned by minimizing the following loss function:
3.2. Tool Load Identification Based on Least Squares Method
Generally, different end-effectors are attached to the end of the force sensor in order to achieve various tasks. As the pose of the robot changes, the pose of the tool changes accordingly, and its influence on the sensor changes as well. In order to achieve a quick compensation of the gravitational influence of the end-effector after its replacement, this section uses the least squares method to estimate the tool load.
Let the robot base frame be and the world frame be , and assume that is reversed with respect to gravity. Since the error in the robot installation will make and not completely coincide, assuming that can be obtained by rotating around the by angle and around the by angle , then, the rotation matrix of the robot base frame relative to the world frame can be written as:
The frame of the force sensor is , and the relationship between the frames is shown in Figure 2.
Figure 2.
Relationship between the frames. is the world frame; is the robot base frame; is the force sensor frame. O is the origin of the frame.
Assuming that the gravity force acting on the tool is G, then it is expressed in as . Through force transformation, the tool gravity in the ith attitude can be converted from the world frame to the force sensor frame:
which can be equated to:
where , , and is the rotation matrix of the sensor frame at the ith pose with respect to the robot base frame.
The gravity force acting on the tool in the force sensor frame is shown in Figure 3.
Figure 3.
The gravity force acting on the tool in the force sensor frame. g indicates the direction of gravity; Gx, Gy, Gz are components of gravity; Mx, My, Mz are the component of the gravitational moment.
The frame of the six-axis force sensor is a spatial Cartesian rectangular frame. Assuming the coordinate of the tool’s centre of gravity in the sensor frame as , according to the relationship between forces and moments, it is obtained that:
where , , and .
Stacking all forces as , moments as , rotation matrix as , and matrix as in n robot poses, we have:
Therefore, the gravity force acting on the tool and centre of gravity can be estimated as:
where is the matrix calculated from the optimal value of the tool gravity. Therefore, according to the least squares method, the optimal estimate of the load gravity and centre of gravity coordinates is:
3.3. Compensation of Force Disturbance
According to the proposed method, the zero-point of the force sensor is predicted by using the NN model when the robot is in an arbitrary attitude . In addition, the optimal value of the tool gravity in the robot base frame and the tool centre of gravity in the force sensor frame can be quickly estimated in a small number of poses. Further, when the robot is operated to an arbitrary attitude , the effect of the tool gravity on the sensor output is calculated as follows:
According to Equation (2), the pure contact force applied to the end-effector is:
Finally, the compensation of force disturbance can be completed according to Equation (16). See Appendix A for details of the pseudo-code of the proposed integrated compensation method.
4. Experimental Results
In order to verify the feasibility of the method, this paper designs the force interference compensation experiment of the six-axis force sensor at the end of the robot. The flow chart of the experiment is shown in Figure 4. The robot used in the experiment is the UR10 collaborative robot. The pose of the six-axis force sensor can be provided from the robot control system by positioning the installation of the six-axis force sensor and setting the position of the robot’s TCP (Tool Centre Point). The force sensor is a contact force sensor in a tandem force sensor [33] independently developed by our laboratory, using self-developed acquisition box and software for data collection. The technical parameters of the force sensor are shown in Table 1. A total of 200 sets of data are collected in each posture to ensure the stability and accuracy of the data, and eliminate the deviation caused by the robot vibration and inertia. Finally, the median value is taken as the response value in that posture.
Figure 4.
Flow chart of the experiment.
Table 1.
The technical parameters of the force sensor.
4.1. Zero-Point Estimation of a Six-Axis Force Sensor Based on Deep Learning
It is necessary to collect sufficient output of the force sensor in different poses to learn the parameter , however, there is no smooth continuous trajectory that allows the robot to traverse all possible poses in the workspace. Xiong [34] et al. pointed out that the first three joints of the robot mainly determine the location of the wrist, and the latter three joints mainly determine the posture of the wrist. Considering the time required to collect the data so that the robot can traverse as many poses as possible while moving at a small angle each time, the following experimental scheme is designed. Keeping the first three joints of the robot fixed and considering the interference problem of the force sensor during the motion and the common working range, the data acquisition range of the fourth, fifth and sixth joints of the robot is restricted to , , respectively, as shown in Figure 5. The step size of the three joints during data collection is , so a total of sets of data are collected. Each set of collected data includes the robot’s pose and the output of the force sensor in the current pose. After the original data collection, the order of the data is disordered randomly, and 9966 sets of data are selected as the training set, while the remaining 474 sets are used as the test set.
Figure 5.
Illustration of the rotation of the joints. The red arrow represents the rotation of the fourth joint; the green arrow represents the rotation of the fifth joint; the blue arrow represents the rotation of the sixth joint.
The experimental environment is as follows: CPU is Intel(R) Core (TM) i7-9750H CPU@2.60 GHz, graphics card is GTX1660Ti, operating system is Windows 10, and Pytorch deep learning framework is used. The complete training data set of 9966 is used in each iteration, and the optimizer uses Adam with 100,000 iterations. The learning rate settings for the iterations are shown in Table 2. To prevent overfitting, L2 regularization is added to the loss function. Figure 6 shows the change of loss function for the 50,000th–100,000th iteration during the training.
Table 2.
Learning Rate.
Figure 6.
Loss function change curve of 50,000th–100,000th iteration.
After training, the model is tested using the test set data, and the results are shown in Figure 7, where the unit of forces is Newton (N) and the unit of moments is Newton-cm (Ncm). To illustrate the distribution of the zero-point estimation errors, a kernel density distribution of the errors was made, as shown in Figure 8. For comparison, a zero-point estimation method of the force sensor based on a dynamical model (DM) is conducted, and the results are shown in Figure 9. The mean (), standard deviation (), and maximum value of the absolute value of the error () obtained from the data are shown in Table 3.
Figure 7.
Zero-point estimation errors of the force sensor based on deep learning. (a) the estimation error of Fx, (b) the estimation error of Fy, (c) the estimation error of Fz, (d) the estimation error of Mx, (e) the estimation error of My. The blue lines represent the response values, and the orange lines represent the error.
Figure 8.
Kernel density plot of errors.
Figure 9.
Zero-point estimation errors of the force sensor based on DM. (a) the estimation error of Fx, (b) the estimation error of Fy, (c) the estimation error of Fz, (d) the estimation error of Mx, (e) the estimation error of My. The blue lines represent the response values, and the orange lines represent the error.
Table 3.
Force sensor zero-point estimation error data.
Compared to Figure 9, the mean value of the error data on the test set in Figure 7 is close to 0. As shown in Table 3, of the test dataset of the NN method is much less than that of the DM method, except for Mx, but the errors are extremely similar. In particular, the of the NN method is much smaller than that of the DM method. The data shows that the error data distribution of the NN method is more concentrated and uniformly disordered on both sides of 0, as can also be illustrated by Figure 8.
To verify that the NN method outperforms the DM method, we reduce the size of the training set for experiments. The experimental results are presented in Appendix B. The bar graphs of are shown in Figure 10 for training data of 1000, 3000, 5000, and 7000 sets, respectively. Figure 10 shows that the DM method still performs much worse than the NN method even with the same training samples. Furthermore, the result shows that the NN model generalizes well and the errors obtained after fitting are caused by noise, so the deep learning-based sensor zero-point estimation method is feasible.
Figure 10.
for different training data sizes. (a) The results of 1000 sets of training data. (b) The results of 3000 sets of training data. (c) The results of 5000 sets of training data. (d) The results of 7000 sets of training data.
4.2. Tool Load Identification by Least Squares Method
In order to verify the feasibility of the least squares-based tool load identification algorithm, a vacuum chuck is installed at the end of the six-axis force sensor to perform the experiment. The experimental site is shown in Figure 11. The robot is controlled by the host computer to adjust to 15 attitudes (rotation vectors) to collect data, respectively. The attitude data are shown in Appendix C.
Figure 11.
Data acquisition site with tool load for least squares method.
The collected data are pre-processed and then calculated using the proposed method in Section 3.2 settlement results show , , calculated as expected. In reality, the mass of the vacuum chuck is 571 g, the gravity is 5.592 N. The gravity obtained from the identification is N, the error of the identification result is less than 8.5%. However, because the true value of the position of the centre of gravity cannot be measured accurately, and it is only used as the intermediate value in the compensation process, the error of the obtained result is not analysed. Therefore, we used Equations (15) and (16) in Section 3.3 to calculate the compensation results to analyse the feasibility of the method.
4.3. Compensation Results
When the tool at the end of the six-axis force sensor is not in contact with any object, the sensor does not receive any external force except for gravity, that is . According to Formula (16), the error after the compensation is:
According to the proposed method, the robot is adjusted to a series of random postures in the work area and collect force sensor data to verify the feasibility of the algorithm. The robot is controlled to move to 215 different random poses in the work area, and the corresponding robot pose data and force sensor output data are collected. The error between the response values of the force sensor and the predicted values of the algorithm , i.e., the compensation error of the disturbance force, is calculated and the box plot of the error is shown in Figure 12. The figure shows that the error distribution of the test data is relatively concentrated, and the percentage of abnormal data is less than 0.9%. However, the error values of Mx and My are large relative to the forces, which is due to the fact that they are influenced by both the forces and the load centre of gravity, resulting in an unavoidable accumulation of errors.
Figure 12.
The box plot of test data.
Table 4 shows the maximum absolute error (MAX) and the mean absolute error (MAE) of the compensation, where Bias+LSM is the compensation algorithm proposed in the literature [22], Double-LSM is the method proposed in the literature [23], which considers the sensor and the tool load as a whole, and NN + LSM is the proposed method. Table 4 shows that, compared with Bias + LSM and Double-LSM, the proposed method in this paper can effectively reduce the MAX and the MAE of after the compensation. The compensation error level of is comparable to that of the Double-LSM, which is due to the fact that the zero-point estimation in this algorithm does not compensate for and only compensates for it in the load gravity identification stage. However, in general, the present algorithm can accomplish the compensation of robot end force disturbance with higher accuracy.
Table 4.
Comparison of compensation error results.
According to the technical parameters of the force sensor provided in Table 1, the errors after the compensation is completed by this method are 0.83%F.S., 1.39%F.S., 0.71%F.S., 0.45%F.S., 0.84%F.S., and 0.19%F.S., respectively.
To quantify the degree of error more accurately, the disturbance force compensation degree factor is defined as:
where is the output of the ith axis in the test. Therefore, this method reduces the effects of force sensor self-gravity, drift, system error, and tool gravity on the sensor output by 89.33%, 83.83%, 83.83%, 92.91%, 82.89%, and 76.48% in each dimension of the output, respectively. Through analysis, we conclude that the factors leading to incomplete compensation are: robot control error, the influence of the collection cable during the collection process, the calibration error of the six-axis force sensor, and the random error during collection. However, the compensated force error is still controlled within 1.50 N (1.5%F.S.) and the moment error is controlled within 4.3 Ncm (0.86%F.S.) by this algorithm. Compared to the literature [24], the present algorithm is able to perform the compensation task efficiently while meeting the accuracy requirements. Although the test was conducted for the particular robot and end-effector, the method is general.
In conclusion, compared to existing methods, the proposed method can significantly reduce the influence of the end of the robot interference force, and can meet the demand of active compliance control.
5. Conclusions
In this paper, we proposed an integrated method to compensate for the external forces applied to the six-axis force sensor. Considering the interference of drift, system error, and gravity, the proposed method used a deep learning model to predict the zero-point of the sensor and identified the tool load by using the least squares method. In the experiment, we designed and trained an NN model to predict the zero-point of the sensor, based on a 9966-item dataset obtained in the non-load condition. Moreover, we collected 15 groups of data when the robot arm was in different poses, including robot poses and sensor output. Such data were used to identify the tool load based on the least squares method. Finally, the force compensation values for the sensor were calculated by integrating the zero-point prediction and tool load identification obtained above. The experiment results show that the proposed method can perform more accurate and effective force compensation in different conditions compared with existing methods. This method can automatically complete force compensation of the six-axis force sensor in complex manufacturing scenarios, which can significantly improve the intelligence of the robot.
Author Contributions
Conceptualization, L.Y. and D.Z.; methodology, L.Y., Q.G., and W.Z.; software, L.Y. and Q.G.; validation, Y.C. and D.Z.; formal analysis, W.Z.; investigation, L.Y.; resources, Y.C. and D.Z.; data curation, L.Y., Y.C., and D.Z.; writing—original draft preparation, L.Y.; writing—review and editing, W.Z. and D.Z.; visualization, L.Y.; supervision, Y.C. and D.Z.; project administration, Y.C. and D.Z.; funding acquisition, Y.C. and D.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This work was funded by the Science and Technology Support Project of the National Science Foundation of China, grant number 51775215.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available on request from the corresponding author.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
The pseudo-code of the proposed method in this paper is shown in Algorithm A1.
| Algorithm A1: Integrated compensation method. |
| Input:: Current pose of the force sensor. : Set of robot poses to be moved. : Weight’s file generated by deep learning in Section 3.1. Output:: In the pose , the pure contact force applied to the end-effector. repeat if end-effector is replaced then for i to length () do Control the robot move to the pose in ; Save current pose information in ; Save current output of the force sensor in ; Save current zero-point in ; end for ; ; Calculate from ; ; end if while p is changed do Current output of the force sensor Get the rotation matrix at this pose ; Zero-point of the force sensor at this pose ; Effect of the end-effector: Calculate from end while until end of manufacturing task. |
Appendix B
Results for training data of different sizes. NN stands for Neural Network method and DM stands for Dynamic Model method.
Table A1.
Error results obtained from training sets of different sizes.
Table A1.
Error results obtained from training sets of different sizes.
| Data Size | Type | NN | DM | ||||||
|---|---|---|---|---|---|---|---|---|---|
| 1000 | Fx | 0.006 | 0.224 | 0.841 | 0.678 | −0.010 | 0.418 | 1.027 | 1.244 |
| Fy | −0.001 | 0.255 | 0.808 | 0.764 | 0.051 | 0.618 | 1.158 | 1.905 | |
| Fz | −0.019 | 0.205 | 0.531 | 0.596 | −0.006 | 0.368 | 0.951 | 1.098 | |
| Mx | 0.038 | 0.144 | 1.493 | 0.470 | −0.040 | 0.540 | 1.509 | 1.580 | |
| My | 0.028 | 0.571 | 1.783 | 1.741 | 0.001 | 0.983 | 1.841 | 2.950 | |
| 3000 | Fx | −0.008 | 0.208 | 0.682 | 0.616 | −0.024 | 0.418 | 1.043 | 1.230 |
| Fy | −0.018 | 0.273 | 0.925 | 0.801 | 0.038 | 0.619 | 1.146 | 1.895 | |
| Fz | −0.006 | 0.211 | 0.585 | 0.627 | 0.316 | 0.367 | 0.953 | 1.417 | |
| Mx | 0.062 | 0.449 | 1.526 | 1.409 | 0.452 | 0.536 | 1.488 | 2.060 | |
| My | −0.003 | 0.541 | 1.493 | 1.620 | 0.837 | 0.985 | 1.812 | 3.792 | |
| 5000 | Fx | −0.006 | 0.197 | 0.602 | 0.585 | −0.027 | 0.418 | 1.041 | 1.227 |
| Fy | −0.019 | 0.256 | 0.892 | 0.749 | 0.031 | 0.619 | 1.139 | 1.888 | |
| Fz | −0.139 | 0.198 | 0.592 | 0.455 | 0.004 | 0.367 | 0.960 | 1.105 | |
| Mx | 0.048 | 0.435 | 1.331 | 1.353 | −0.045 | 0.535 | 1.483 | 1.560 | |
| My | −0.005 | 0.537 | 1.363 | 1.606 | −0.041 | 0.984 | 1.810 | 2.911 | |
| 7000 | Fx | 0.019 | 0.182 | 0.616 | 0.565 | −0.031 | 0.418 | 1.045 | 1.223 |
| Fy | 0.004 | 0.245 | 0.858 | 0.739 | 0.028 | 0.619 | 1.147 | 1.885 | |
| Fz | −0.019 | 0.191 | 0.550 | 0.554 | 0.002 | 0.367 | 0.956 | 1.103 | |
| Mx | 0.020 | 0.413 | 1.159 | 1.259 | −0.044 | 0.535 | 1.487 | 1.561 | |
| My | 0.045 | 0.487 | 1.675 | 1.506 | −0.051 | 0.984 | 1.819 | 2.901 | |
where .
Appendix C
The robot postures data required in the experiment of tool load identification based on least squares method.
Table A2.
The robot postures data required in Section 4.2.
Table A2.
The robot postures data required in Section 4.2.
| Number | RX | RY | RZ |
|---|---|---|---|
| 1 | 2.0592 | 0.6381 | −0.0558 |
| 2 | 0.6526 | −2.7875 | 0.2766 |
| 3 | 1.8874 | −2.0127 | −0.2661 |
| 4 | 0.4880 | 2.1373 | 0.2597 |
| 5 | 0.4232 | −1.8506 | −0.6033 |
| 6 | −1.9615 | 1.9499 | 0.2685 |
| 7 | 1.5291 | −1.1463 | −1.6160 |
| 8 | −2.6872 | −0.6012 | −0.4095 |
| 9 | −1.2868 | 1.6152 | 0.1995 |
| 10 | −0.7769 | 2.4489 | 1.6684 |
| 11 | −1.9852 | 0.5888 | 0.2570 |
| 12 | −0.5084 | −2.1222 | −1.6644 |
| 13 | −0.4566 | −1.9525 | −1.8284 |
| 14 | −0.7513 | −0.2524 | 0.6352 |
| 15 | −0.8525 | −2.4366 | 0.7001 |
RX, RY, and RZ are the three elements of the rotation vector.
References
- Brogardh, T. Present and future robot control development—An industrial perspective. Annu. Rev. Control 2007, 31, 69–79. [Google Scholar] [CrossRef]
- Hua, J.; Zeng, L.; Li, G.; Ju, Z. Learning for a Robot: Deep Reinforcement Learning, Imitation Learning, Transfer Learning. Sensors 2021, 21, 1278. [Google Scholar] [CrossRef] [PubMed]
- Kemp, C.; Edsinger, A.; Torres-Jara, E. Challenges for robot manipulation in human environments [grand challenges of robotics]. IEEE Robot. Autom. Mag. 2007, 14, 20–29. [Google Scholar] [CrossRef]
- Villani, V.; Pini, F.; Leali, F.; Secchi, C. Survey on human–robot collaboration in industrial settings: Safety, intuitive interfaces and applications. Mechatronics 2018, 55, 248–266. [Google Scholar] [CrossRef]
- Bonci, A.; Cen Cheng, P.D.; Indri, M.; Nabissi, G.; Sibona, F. Human-Robot Perception in Industrial Environments: A Survey. Sensors 2021, 21, 1571. [Google Scholar] [CrossRef]
- Semini, C.; Barasuol, V.; Goldsmith, J.; Frigerio, M.; Focchi, M.; Gao, Y.; Caldwell, D. Design of the hydraulically actuated, torque-controlled quadruped robot HyQ2Max. IEEE/ASME Trans. Mechatron. 2016, 22, 635–646. [Google Scholar] [CrossRef]
- Campeau-Lecours, A.; Otis, J.; Gosselin, C. Modeling of physical human-robot interaction: Admittance controllers applied to intelligent assist devices with large payload. Int. J. Adv. Robot. Syst. 2016, 13, 1–12. [Google Scholar] [CrossRef] [Green Version]
- Li, Y.; Yang, C.; Wang, G.; Zhang, H.; Cui, H.; Zhang, Y. Research on the parallel load sharing principle of a novel self-decoupled piezoelectric six-dimensional force sensor. ISA Trans. 2017, 70, 447–457. [Google Scholar] [CrossRef]
- He, X.; Cai, P. Design and analysis of a low-range 6-D force sensor. Transducer Microsyst. Technol. 2012, 1, 20–22. [Google Scholar]
- Ma, J.; Song, A. Fast Estimation of Strains for Cross-Beams Six-Axis Force/Torque Sensors by Mechanical Modeling. Sensors 2013, 13, 6669–6686. [Google Scholar] [CrossRef] [Green Version]
- Oh, H.; Kim, U.; Kang, G.; Seo, J.; Choi, H. Multi-Axial Force/Torque Sensor Calibration Method Based on Deep-Learning. IEEE Sens. J. 2018, 18, 5485–5496. [Google Scholar] [CrossRef]
- Lin, C.-Y.; Ahmad, A.R.; Kebede, G.A. Novel Mechanically Fully Decoupled Six-Axis Force-Moment Sensor. Sensors 2020, 20, 395. [Google Scholar] [CrossRef] [Green Version]
- Raibert, M.; Craig, J. Hybrid position/force control of manipulators. ASME J. Dyn. Syst. Meas. Control 1981, 102, 126–133. [Google Scholar] [CrossRef]
- Luo, J.; Solowjow, E.; Wen, C.; Ojea, J.; Agogino, A.; Tamar, A.; Abbeel, P. Reinforcement learning on variable impedance controller for high-precision robotic assembly. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 3080–3087. [Google Scholar]
- Shetty, B.R.; Ang, M.H. Active compliance control of a PUMA 560 robot. In Proceedings of the IEEE International Conference on Robotics and Automation, Minneapolis, MN, USA, 22–28 April 1996; pp. 3720–3725. [Google Scholar]
- Tian, F.; Li, Z.; Lv, C.; Liu, G. Polishing pressure investigations of robot automatic polishing on curved surfaces. Int. J. Adv. Manuf. Technol. 2016, 87, 639–646. [Google Scholar] [CrossRef]
- Xu, X.; Zhu, D.; Zhang, H.; Yan, S.; Ding, H. Application of novel force control strategies to enhance robotic abrasive belt grinding quality of aero-engine blades. Chin. J. Aeronaut. 2019, 32, 2368–2382. [Google Scholar] [CrossRef]
- Loske, J.; Biesenbach, R. Force-torque sensor integration in industrial robot control. In Proceedings of the 15th International Workshop on Research and Education in Mechatronics (REM), El Gouna, Egypt, 9–11 September 2014; pp. 1–5. [Google Scholar]
- Lin, J. Research on Active Compliant Assembly System of Industrial Robot Based on Force Sensor. Ph.D. Thesis, South China University of technology, Guangzhou, China, 2013. [Google Scholar]
- Cai, M. Research on Homogeneous Hand Controller Based on Force Fusion Control for Telerobot. Master’s Thesis, Jilin University, Changchun, China, 2015. [Google Scholar]
- Li, S.; Zhang, Y.; Cao, Y.; Dai, X. Experiment research of gravity compensation for haptic device. Mech. Des. Res. 2007, 1, 95–97. [Google Scholar]
- Vougioukas, S. Bias estimation and gravity compensation for force-torque sensors. In Proceedings of the 3rd WSEAS Symposium on Mathematical Methods and Computational Techniques in Electrical Engineering, Athens, Greece, 17–23 December 2001; pp. 82–85. [Google Scholar]
- Zhang, L.; Hu, R.; Yi, W. Research on force sensing for the end-load of industrial robot based on a 6-axis force/torque sensor. Acta Autom. Sin. 2017, 43, 439–447. [Google Scholar]
- Zhang, Z.; Chen, Y.; Zhang, D.; Tong, Q. Research on numerical compensation method of tandem force sensor installed at the end of industrial robot. In Proceedings of the 5th International Conference on Automation, Control and Robotics Engineering, Dalian, China, 19–20 September 2020; pp. 725–731. [Google Scholar]
- El, D.; Kamal, M.; Sanchez, J.; Corrales, J.A.; Mezouar, Y. FaForce-torque sensor disturbance observer using deep learning. In International Symposium on Experimental Robotics; Springer: Cham, Switzerland, 2018. [Google Scholar]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Banerjee, I.; Ling, Y.; Chen, M.; Hasan, S.; Langlotz, C.; Moradzadeh, N.; Chapman, B.; Amrhein, T.; Mong, D.; Rubin, D.; et al. Comparative effectiveness of convolutional neural network (CNN) and recurrent neural network (RNN) architectures for radiology text report classification. Artif. Intell. Med. 2019, 97, 79–88. [Google Scholar] [CrossRef]
- Ciregan, D.; Meier, U.; Schmidhuber, J. Multi-column deep neural networks for image classification. In Proceedings of the 2012 IEEE Conference on Computer vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 13642–13649. [Google Scholar]
- Litjens, G.; Kooi, T.; Bejnordi, B.; Setio, A.; Ciompi, F.; Ghafoorian, M.; Laak, J.; Ginneken, B.; Sanchez, C. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
- Jiang, C.; Chen, S.; Chen, Y.; Bo, Y.; Han, L.; Guo, J.; Feng, Z.; Zhou, H. Performance Analysis of a Deep Simple Recurrent Unit Recurrent Neural Network (SRU-RNN) in MEMS Gyroscope De-Noising. Sensors 2018, 18, 4471. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Hu, J.; Hua, F.; Tian, W. Robot Positioning Error Compensation Method Based on Deep Neural Network. In Proceedings of the 2020 4th International Conference on Control Engineering and Artificial Intelligence (CCEAI 2020), Singapore, 17–19 January 2020; Volume 1487, p. 12045. [Google Scholar]
- Yegnanarayana, B. Artificial Neural Networks, 1st ed.; PHI Learning Pvt. Ltd.: New Delhi, India, 2009. [Google Scholar]
- Zhang, Z.; Chen, Y.; Zhang, D. Development and Application of a Tandem Force Sensor. Sensors 2020, 20, 6042. [Google Scholar] [CrossRef] [PubMed]
- Xiong, Y. Robotics: Modeling, Control and Vision, 1st ed.; Huazhong University of Science and Technology Press: Wuhan, China, 2018. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).