Next Article in Journal
Fabrication of Micro-Cantilever Sensor Based on Clay Minerals for Humidity Detection
Previous Article in Journal
Breathing Chest Wall Kinematics Assessment through a Single Digital Camera: A Feasibility Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Bucket Pose Estimation Based on Deep Neural Network and Registration Using Onboard 3D Sensor

1
School of Resources And Safety Engineering, Central South University, Changsha 410083, China
2
Digital Mine Research Center, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6958; https://doi.org/10.3390/s23156958
Submission received: 11 July 2023 / Revised: 29 July 2023 / Accepted: 2 August 2023 / Published: 5 August 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Real-time and accurate bucket pose estimation plays a vital role in improving the intelligence level of mining excavators, as the bucket is a crucial component of the excavator. Existing methods for bucket pose estimation are realized by installing multiple non-visual sensors. However, these sensors suffer from cumulative errors caused by loose connections and short service lives caused by strong vibrations. In this paper, we propose a method for bucket pose estimation based on deep neural network and registration to solve the large registration error problem caused by occlusion. Specifically, we optimize the Point Transformer network for bucket point cloud semantic segmentation, significantly improving the segmentation accuracy. We employ point cloud preprocessing and continuous frame registration to reduce the registration distance and accelerate the Fast Iterative Closest Point algorithm, enabling real-time pose estimation. By achieving precise semantic segmentation and faster registration, we effectively address the problem of intermittent pose estimation caused by occlusion. We collected our own dataset for training and testing, and the experimental results are compared with other relevant studies, validating the accuracy and effectiveness of the proposed method.

1. Introduction

In recent years, with the growing demand for resources and the increasing emphasis on the safety of mining operations, the development of mining is moving towards intelligence, with the ultimate goal of achieving autonomous mining. The technical conditions and environmental characteristics of open-pit mine are relatively unique compared to underground mining, and open-pit mines have the advantages of an open transport space, fixed driving routes, low speed and high-quality communication. For these reasons, open pit mines have become typical scenery for intelligent mining construction [1]. A traditional open pit mine site is shown in Figure 1, where the mining operations are carried out by a combination of bucket operators and site supervisors. This mode of excavation is low in efficiency and high in labor costs. Automatic mining [2] and safety assessments [3] can be realized by installing sensors such as LiDAR, camera, and GPS on the mining excavator. This is beneficial to improve the productivity of mining operations [4] and enhance the intelligence of the mining excavator [5,6]. As a key component and end-effector of the mining excavator, bucket pose estimation is an important way to improve the intelligence of the mining excavator [7]. On the one hand, the bucket pose estimation can assist the bucket driver in excavation and loading [8], improving the efficiency of mining operations. Real-time bucket pose estimation also provides an important basis for the realization of automatic mining and loading [9], which greatly contributes to promoting the intelligence of mining [10].
The traditional method for estimating bucket pose is installing multiple non-visual sensors, such as a potentiometer, rotary encoder, and resolver [11,12,13]. To accurately obtain the changing poses of the bucket when mining, several different types of non-visual sensors need to be installed. However, these suffer from cumulative errors caused by loose connections and a short service life caused by strong vibrations.
To solve these problems, we propose a visual-sensor-based bucket pose estimation method. The key contributions of this work are as follows:
To adapt the point transformer model for the bucket point cloud semantic segmentation task, we adjust the parameters of the original model. This adjustment significantly enhances the segmentation accuracy and overall model performance.
By employing pose canonicalization, we avoid the issue of limited overlap between consecutive frames’ bucket point clouds, which can negatively impact registration accuracy. As a result, we effectively address the challenge of achieving accurate bucket pose estimation when the bucket is occluded.
The proposed method does not rely on prior models for registration, making it highly generalizable. It can be applied to estimatations of the pose of end-effectors of other pieces of loading equipment in an open-pit mine.

2. Related Work

In this section, we will comprehensively review the relevant studies and technological advancements pertaining to pose estimation.
Considering the problems of non-visual sensor measurement methods, some researchers adopted visual sensor methods to estimate the bucket pose. Wen [14] present a monocular vision-based method for 3D excavator pose estimation, which can provide the groundwork for monitoring contact-driven accidents near excavators. Mahmood [15] proposed the use of synthetic images for training a CNN-based 3D pose estimation model, which was tested and applied to actual images from job sites. However, image-based pose estimation methods suffer from significant errors, and they are unable to estimate the bucket pose when it is occluded during mining operations.
To enhance the accuracy of pose estimation, Tang [16] proposed a multiple-onboard-sensors data fusion approach for full-body pose estimation of an excavator. By combining the advantages of IMU and camera sensors, they improved the accuracy of pose estimation. This method outperformed methods relying solely on IMU or camera for state estimation. However, it still exhibited a relatively large average error of 9.066 cm. Cui [17] proposed a novel memory augmented registration network for bucket pose estimation. This kind of method can effectively alleviate the forgetting problem and increase the estimation accuracy of the bucket poses with small samples. This method demonstrated relatively few errors (a rotation error of 2.141 and translation error of 4.8 cm). However, it still faces the challenge of requiring a large number of samples for training. Additionally, the experiments were conducted on empty buckets without any occlusion, and the accuracy of pose estimation for occluded buckets was not validated.
Therefore, in practical applications of mining operations, there is a need to explore a pose estimation method that offers high accuracy and is not affected by occlusion.
Recently, the combination of registration and deep neural networks has achieved a remarkable performance in pose estimation [18,19,20,21]. Wang et al. [22] utilized the modified U-SegNet segmentation network to segment the target and employed iterative closest point (ICP) [23] to register the segmented results with the pre-scanned model candidates for pose estimation. Zeng et al. [24] applied a fully convolutional network (FCN) [25] to segment and label multiple views of the scene. They used ICP to align the segmented point clouds with the pre-scanned model to estimate the 6D pose of objects. Similarly, V. Tran et al. [26] optimized the registration between target objects and CAD models based on R-CNN and structured light technology. Yang et al. [27] improved the pose estimation performance in partially occluded scenarios by combining object detection networks and ICP registration in a life-support robot operating in real-world environments. Luan et al. [28] proposed a multi-path convolutional neural network BiLuNet, which is based on the U-Net framework. The network extracts the object mask and merges it with the depth information to perform 6D pose estimation using the Iterative Closest Point (ICP) algorithm. Compared to the U-Net with limited training data, their method achieved better results.
These methods are all based on the use of pre-scanned models for registration, which leads to slower registration speeds. For example, Luan’s method took 2.76 s to process [28], and Wang’s method had a total processing time of from 2 to 3 s from sensor data capture to pose estimation [22]. These processing times do not meet the real-time requirements of loading scenarios. Moreover, these methods have not been proven to work in environments with significant occlusion. Therefore, before applying these methods in open-pit mine scenarios with a high level of interference such as dust, noise, and pile collapse, it is necessary to address the challenges of severe occlusion and real-time computation.
We apply a method that combines registration and a deep neural network in the mining field. By employing point cloud Pose Canonicalization, we effectively address the challenge of difficult registration when the bucket is occluded. Additionally, we accelerate the registration speed of the FICP algorithm by removing outlier points and downsampling the point cloud, ensuring real-time performance. This allows for continuous pose estimation of the bucket with occlusion in complex open-pit mine scenarios.

3. Methods

Our method is summarized in Figure 2, with four main steps: model training, point cloud segmentation, point cloud processing, and pose estimation. After receiving a sequence of point clouds as input, we first utilize a model for semantic segmentation to classify different objects in the point cloud. The segmentation results are subjected to outlier removal and downsampling. Next, the point cloud is pose-canonicalized and registered with the first frame’s point cloud using the FICP registration algorithm. Finally, the current frame’s pose estimation is computed based on the registration result.

3.1. Semantic Segmentation

The main types of point-cloud-processing task include object detection and semantic segmentation. In the source data of this research, the bucket and pile are involved in a dynamic interaction process. It is challenging to accurately separate the bucket and pile based solely on candidate bounding boxes generated by object detection. Therefore, semantic segmentation is employed, treating individual points as the unit of judgment. Points that do not belong to the bucket class are directly removed from the point clouds based on point labels, enabling the most precise bucket segmentation and ensuring accurate pose estimation.
Engel [29] proposed a Point Transformer (PT) network that applies self-attention to 3D point clouds. The PT network exhibits strong scalability, superior performance, and high segmentation accuracy, making it suitable for bucket segmentation tasks. It can achieve the desired segmentation performance even with relatively few training data and in the presence of occlusions.
To make the PT network more suitable for the bucket segmentation task and improve model performance, we adjusted parameters based on the original model. Considering the sparse density of point clouds in bucket loading scenes, we moderately reduced the number of groups used for attention calculation and set the number of nearest neighbors in the neighborhood for local attention computation to 8. This could enhance the model’s ability to capture spatial relationships and semantic information in bucket loading point clouds. We also accelerated the training and convergence speed by reducing the number of network layers and feature vector dimensions while ensuring model performance.
Since the FICP algorithm is used for registration, the registration accuracy can be affected by outliers. Therefore, when applying the model to our bucket point cloud segmentation task, we focused more on segmentation accuracy. We set the dropout rate parameter to 0.5 to enhance the model’s robustness and achieve high-precision segmentation. The model’s performance evaluation metric for bucket point cloud semantic segmentation is the mean Intersection over Union (mIoU). The optimal parameter combination for the model in the bucket segmentation task is presented in Table 1. Detailed experiments related to parameters adjustments are described in Chapter 4.

3.2. Pose Estimation

Achieving accurate bucket pose estimation is crucial for practical applications such as autonomous mining and loading. In addition to ensuring the accuracy of semantic segmentation results, the accuracy of the registration algorithm is also crucial. However, the bucket is often occluded by pile collapse and dust during the mining operations, which makes registration extremely challenging. Therefore, in this paper, we used the Fast Iterative Closest Point (FICP) [30] algorithm for registration. Compared to the classical ICP algorithm, FICP uses a robust error metric based on the Welsch function, which reduces misalignment in the presence of outliers and partial overlap.
Although the FICP method has made significant improvements compared to the ICP algorithm, there are still some issues. Despite the use of Anderson acceleration [31] to improve convergence speed, FICP is still sensitive to the initial relative pose, making it prone to local optima and meaning that it has a slower computation speed when the distance between the two frames is not close enough for registration. To address the sensitivity of the FICP algorithm to the initial relative pose, this paper adopts continuous frame registration for tracking. Due to the temporal continuity, the bucket’s pose changes minimally between consecutive frames, and the distance between the bucket point clouds is extremely close. This approach ensures fast registration using the FICP algorithm while avoiding getting trapped in local optima.
We can represent a 6DoF transformation T = { R , t } . Since more of the bucket point cloud is lost when there is occlusion, there is little similarity between the bucket point clouds of consecutive frames, which align poorly across different frames. Inspired by [32], we turned the estimation of T t + 1 into T t + 1 , which is widely used in instance-level 6D pose estimation and tracking works [33,34].
The input point cloud X t + 1 was pose-canonicalized for each frame except the first before alignment. The pose-canonicalized point cloud Z t + 1 is defined as the product of the inverse transformation of X t + 1 and T t .
Due to the continuity of the pose between frames, the pose-canonicalized point cloud is very close to the first frame point cloud, and this step could be considered the initial alignment. Note that T t + 1 can be expressed using T t + 1 and T t , namely, t t + 1 = R t t t + 1 + t t , R t + 1 = R t R t + 1 .
In this way, we can reduce the effect of bucket point cloud occlusion on pose estimation. By doing so, we significantly simplify the regression task and ensure accurate pose estimation even when part of the bucket point clouds is missing.

4. Experiment

A series of experiments were conducted in both a CARLA simulation environment and a metal mine to evaluate the accuracy and effectiveness of the proposed method. The experiments consisted of two main components: semantic segmentation and pose estimation. This section also introduces a new dataset that was used to train the network and validate the performance of the proposed method.

4.1. Dataset

The existing large-scale point cloud datasets, such as S3DIS, do not include data based on mining environments. Therefore, based on a simulation environment and a metal mine, we collected a new dataset specifically for bucket pose estimation.
We built a simulated mining scene based on a specific mine in the simulation platform CARLA. CARLA is an open-source research simulator for autonomous driving, based on the Unreal Engine 4 (UE4), which provides good simulation effects for real-world scenarios and physical activities.
In the simulated scene, the bucket is raised to a certain height and tilted backward to deposit the material into the transport vehicle after excavation from the pile. Then, the bucket returns to the excavation position to repeat the loading process. A LiDAR sensor installed on the driver’s cabin of the excavator synchronously collects environmental point clouds at a frequency of 10 Hz. The data collection duration is 3 min (one cycle of loading operation). We selected 332 frames containing the point cloud of the bucket to form the simulated dataset (Figure 3).
In the real metal mine, real mining and loading data were collected using a fixed-point acquisition scheme. The LiDAR sensor was fixed on the excavator’s cabin to capture the environmental point clouds during the excavator’s mining and loading operations. The total data collection duration was 2 min, from which 1158 frames containing the bucket point cloud were extracted. Due to the repetitive nature of the mining and loading process, a final selection of 100 frames was established as the real-world dataset.

4.2. Semantic Segmentation

To improve segmentation accuracy, we explored the effects of four parameters on model performance. The goal was to identify the best combination of PT parameters for bucket point cloud segmentation. For comparison purposes, an original model with unmodified parameters was set up. Figure 4 provides insights into how adjusting various parameters of the PT model affects its performance, using mIoU as the evaluation metric. The model parameter configurations are shown in Table 2.
Observing graph (a), it can be seen that reducing the number of neighboring points leads to better model training results compared to the original model. Since open-pit mine environments are more spacious, the number of point clouds collected by LiDAR is significantly lower than that in indoor and other densely populated areas. Therefore, a reduced number of neighboring points is sufficient to capture the spatial structure of most objects. Models with fewer neighboring points achieve a better performance as there is less interference and noise.
From graph (b), it is evident that when the number of groups is set too large, the final convergence results are lower compared to both the original model and the model with a reduced number of groups. Increasing the number of groups potentially improves the model’s ability to adjust its features while retaining error information during the computation process. For the bucket segmentation task, individual points may not require fine-grained feature representations but are more sensitive to error information. The improved performance of the model with reduced group numbers validates this analysis.
Graph © illustrates the impact of dropout rates of 0.25 and 0.5 on model performance. The experiment reveals that the highest dropout rate also achieves the highest segmentation accuracy, indicating that the model’s resistance to interference is indeed enhanced by an increase in the dropout rate.
Analyzing the results in graph (d), it can be observed that decreasing the number of network layers improves initial performance and speeds up convergence. On the contrary, increasing the number of network layers leads to more significant oscillation. Ultimately, the performance of models with increased and decreased network layers does not significantly differ from the original model, indicating that adjustments to the number of network layers are unnecessary.
Based on a comprehensive comparison, the best-performing parameter combination is selected for the PT model, focusing on its segmentation performance on the simulated dataset. The selected parameter combination is then used for model retraining.
We trained several different models using the same dataset, including Sparse-Unet, KP-FCNN, and the original PT model. The mean Intersection over Union (mIoU) curves of the training process are shown in Figure 5, with the mIoU curve of the parameter-optimized PT model represented by a bold red line. From the graph, it can be observed that both the original PT model and the parameter-optimized PT model significantly outperform the other two models. Furthermore, the mIoU curve of the parameter-optimized PT model exhibits reduced fluctuations and faster convergence compared to the original PT model.
We tested several different segmentation models, including PT, original PT, and Sparse-Unet, with the same test set including 55 frames of the simulated dataset and 55 frames of the real-world dataset. The purpose was to validate the effectiveness and accuracy of the segmentation models. In this experiment, the evaluation metrics include Overall Accuracy (OA), mean Intersection over Union (mIoU), mean Accuracy (mA), and Intersection over Union (IoU) for each class. The results for the simulated dataset and real-world dataset are presented in Table 3 and Table 4.
Our optimized PT model demonstrated an excellent performance on the simulated dataset, achieving an average IoU of 98.16% and an IoU of 96.05% for the bucket class. It outperformed the other two models across all evaluation metrics. On the real-world dataset, where the models were exposed to disturbances such as dust and noise, their performance declined. However, our method still achieved accurate segmentation results. The average IoU for the real-world dataset was 88.53%, with an IoU of 87.03% for the bucket class. While Sparse-Unet had a higher OA of 96.45% and a higher mA of 94.33%, its IoU for the bucket class was only 82.90%, with an accuracy of 86.02% for the bucket class, which was significantly lower than the PT model’s IoU of 87.03% and accuracy of 90.36% for the same class. These results demonstrate that the parameter-optimized PT model outperforms other architectures in terms of accuracy and IoU.
Unlike the simulated environment, real scenes present with more interference factors. In addition to bucket occlusion caused by the blast pile, dust near the bucket also introduces disturbances to semantic segmentation. As a result, the segmentation accuracy of the real scene dataset is slightly lower compared to the simulated environment dataset. However, it still achieves satisfactory segmentation results.
A visualized comparison of the output segmented results is shown in Figure 6. This compares the segmentation results of different models with the ground truth point clouds. From the comparison, it can be observed that the parameter-optimized PT model exhibits segmentation results that closely resemble the ground truth. The boundaries between the red, blue, and green point cloud classes are clearly visible, and the segmented bucket points closely match the ground-truth annotations.
In contrast, in the © group of images depicting three different bucket poses, the segmentation results of the red, blue, and green point cloud classes are poorer. Both the bucket and boom class segmentation results include outliers from the surrounding environment, which are included in the segmented point cloud and disrupt the registration process. In the third pose, due to the influence of surrounding dust, only a partial segment of the bucket is obtained.
In summary, as the degree of occlusion increases, the accuracy of the Sparse-Unet model’s segmentation decreases. It fails to accurately segment the surrounding environmental elements, such as dust and blast piles, along with the bucket. On the contrary, the PT model demonstrates remarkable accuracy, precisely segmenting objects of all three categories from cluttered point clouds. It effectively segments as many points as possible within the bucket class, preserving the shape characteristics of the bucket. This is crucial for the precise registration in subsequent frames.
The accuracy of semantic segmentation affects pose estimation. In our approach, we tested the semantic segmentation model on our own dataset, and the results demonstrate that our segmentation model achieves precise segmentation.

4.3. Pose Estimation

In this experiment, due to the difficulties of calibration in real-world scenarios, it is challenging to obtain accurate bucket pose estimation for quantitative analysis. Therefore, we selected a continuous sequence of 20 frames of point clouds from the simulated dataset for experimentation. To minimize the potential impact of outliers on registration accuracy and improve registration speed, the segmented point clouds are preprocessed before registration. Points located 25 m from the lidar are removed based on the actual range of bucket motion. After removing outliers and downsampling, we reduce the point cloud size from 3000 points to 1500 points. This significantly accelerates the registration speed of the FICP algorithm.
Table 5 presents the rotation error, translation error, and registration time for bucket pose estimation using various classic registration algorithms, including ICP, SAC-IA, RANSAC, Sift + RANSAC, GICP, and FICP. ICP performs the worst due to its susceptibility to local optima. SAC-IA and RANSAC, two feature-based matching algorithms, demonstrate improved accuracy compared to ICP, with RANSAC showing a slower computation speed. After incorporating the SIFT keypoint extraction step, the RANSAC algorithm significantly improves both registration time and accuracy. Variants in ICP, GICP, and FICP show substantial improvements in terms of both accuracy and time, especially FICP. FICP achieves exceptional registration accuracy with an average rotation error of 1.21° and translation error of 2.51 cm, meeting the high registration precision requirements of the application. While GICP exhibits a faster registration time of 20 ms compared to FICP’s 75 ms, FICP provides a higher registration accuracy. Moreover, considering the experimental use of a lidar frequency of 10 Hz, with one frame of point cloud data being transmitted every 100 ms, the real-time requirements of the application are satisfied by the FICP algorithm.
Figure 7 compares the registration results of different registration algorithms. It clearly demonstrates that FICP outperforms other registration algorithms in terms of registration accuracy. It can also be observed that ICP performs the worst among these registration algorithms, as indicated by the red point cloud, which becomes increasingly distant from the blue point cloud as the occlusion of the bucket increases. SAC-IA and RANSAC algorithms have a relatively better performance, with the red and blue point clouds appearing closer in poses 1, 2, and 3. However, in pose 4, the distance between the red and blue point clouds becomes larger. In contrast, FICP demonstrates a consistently good performance across all poses, with the red and blue point clouds almost overlapping.
Overall, the ICP algorithm exhibits significant registration errors between the transformed point cloud and the target registration point cloud. Moreover, as the degree of the occlusion of the bucket increases, the registration error also increases. The SAC-IA and RANSAC algorithms have an unstable registration performance. The two are prone to being trapped in local optima, which leads to poor registration results. However, the FICP algorithm is not affected by the degree of bucket occlusion and exhibits an excellent performance across the four different poses presented in the figure.
We conducted experiments using 55 frames from the real-world dataset, and Figure 8 shows the registration results between the target point cloud and the real-time point cloud with three different degrees of occlusion. From the figure, it is evident that the red registered point cloud aligns closely with the blue current frame point cloud. The registration performance remains unaffected by variations in bucket occlusion, demonstrating robust registration even under high levels of occlusion. The experimental results show that our method not only has higher accuracy in terms of bucket pose estimation on the simulated dataset but also performs well on the real-world dataset with more interference, such as dust and noise.
Figure 9 shows the registration results when the bucket is heavily occluded in both the simulated dataset and the real-world dataset. In the simulation environment, the bucket is heavily occluded during the excavation process, resulting in significant point cloud loss. In the real-world environment, in addition to the occlusion caused by ores during excavation, the bucket is also occluded due to dust. The results in Figure 9 demonstrate that our proposed method performs well in both the real-world environment and simulation environment, even when the bucket is heavily occluded. Moreover, this method effectively handles occlusions caused by both dust and ores.
The experiments were conducted on a computer with an NVIDIA RTX 4090 GPU and an Intel Core i5-10400F CPU. The average time of bucket pose estimation for our method is 0.095 s. This can effectively meet the practical requirements of bucket pose estimation in terms of both accuracy and computational efficiency.

5. Conclusions

In this paper, we propose a method for bucket pose estimation based on 3D vision to solve the large registration error problem caused by occlusion. We collected our own dataset for training and adjusted parameters based on the original PT model, effectively improving the model’s performance and segmentation accuracy (achieving an mIoU metric of 0.9816). We used point cloud Pose Canonicalization to transform the registration between consecutive frames into the registration between the current frame and the first frame’s point cloud. This approach helps to avoid the issue of the low overlap between bucket point clouds in consecutive frames, which can negatively impact registration accuracy. By doing this, we effectively address the challenge of difficult registration when the bucket is occluded. Based on the experiments and analysis presented above, the proposed method achieves the state-of-the-art accuracy (1.2058, 2.51 cm) in bucket pose estimation, even if the bucket suffers from variable degrees of occlusion, which plays an important role in enhancing the level of intelligence in mining excavators.

Author Contributions

Methodology, Z.X. and L.B.; software, Z.X.; validation, Z.X.; data curation, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Technology Project of China (2022YFC2904105).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, S.; Lu, C.; Jiang, S.; Shan, L.; Xiong, N.N. An Unmanned Intelligent Transportation Scheduling System for Open-Pit Mine Vehicles Based on 5G and Big Data. IEEE Access 2020, 8, 135524–135539. [Google Scholar] [CrossRef]
  2. Zhao, J.; Zhang, L. TaskNet: A Neural Task Planner for Autonomous Excavator. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021. [Google Scholar]
  3. Rasul, A.; Seo, J.; Oh, K.; Khajepour, A.; Reginald, N. Predicted Safety Algorithms for Autonomous Excavators Using a 3D LiDAR Sensor. In Proceedings of the 2020 IEEE International Systems Conference (SysCon), Montreal, QC, Canada, 24 August–20 September 2020. [Google Scholar]
  4. Shariati, H.; Yeraliyev, A.; Terai, B.; Tafazoli, S.; Ramezani, M. Towards autonomous mining via intelligent excavators. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Long Beach, CA, USA, 15–20 June 2019; pp. 26–32. [Google Scholar]
  5. Stentz, A.; Bares, J.; Singh, S.; Rowe, P. A robotic excavator for autonomous truck loading. In Proceedings of the 1998 IEEE/RSJ International Conference on Intelligent Robots and Systems. Innovations in Theory, Practice and Applications (Cat. No.98CH36190), Victoria, BC, Canada, 17 October 1998. [Google Scholar]
  6. Son, B.; Kim, C.; Kim, C.; Lee, D. Expert-Emulating Excavation Trajectory Planning for Autonomous Robotic Industrial Excavator. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020. [Google Scholar]
  7. Bewley, A.; Upcroft, B.; Lever, P.; Leonard, S. Automatic In-Bucket Volume Estimation for Dragline Operations; The Australian Institute of Mining and Metallurgy: Carlton, VIC, Australia, 2009. [Google Scholar]
  8. Li, Y.; Niu, T.; Qin, T.; Yang, L. Machine Vision Based Autonomous Loading Perception for Super-huge Mining Excavator. In Proceedings of the 2021 IEEE 16th Conference on Industrial Electronics and Applications (ICIEA), Chengdu, China, 1–4 August 2021; IEEE: Piscataway, NJ, USA, 2021. [Google Scholar]
  9. Sandzimier, R.J.; Asada, H.H. A Data-Driven Approach to Prediction and Optimal Bucket-Filling Control for Autonomous Excavators. IEEE Robot. Autom. Lett. 2020, 5, 2682–2689. [Google Scholar] [CrossRef]
  10. Zhao, J.; Long, P.; Wang, L.; Qian, L.; Lu, F.; Song, X.; Manocha, D.; Zhang, L. AES: Autonomous Excavator System for Real-World and Hazardous Environments. arXiv 2020, arXiv:2011.04848. [Google Scholar]
  11. Wang, F.; Lever, P.J.A. On-line trajectory planning for autonomous robotic excavation based on force/torque sensor measurements. In Proceedings of the 1994 IEEE International Conference on MFI’94. Multisensor Fusion and Integration for Intelligent Systems, Las Vegas, NV, USA, 2–5 October 1994; IEEE: Piscataway, NJ, USA, 1994. [Google Scholar]
  12. Haga, M.; Hiroshi, W.; Fujishima, K. Digging control system for hydraulic excavator. Mechatronics 2001, 11, 665–676. [Google Scholar] [CrossRef]
  13. Sun, D.; Ji, C.; Jang, S.; Lee, S.; No, J.; Han, C.; Han, J.; Kang, M. Analysis of the Position Recognition of the Bucket Tip According to the Motion Measurement Method of Excavator Boom, Stick and Bucket. Sensors 2020, 20, 2881. [Google Scholar] [CrossRef] [PubMed]
  14. Wen, L.; Kim, D.; Liu, M.; Lee, S. 3D Excavator Pose Estimation Using Projection-Based Pose Optimization for Contact-Driven Hazard Monitoring. J. Comput. Civ. Eng. 2023, 37, 04022048. [Google Scholar] [CrossRef]
  15. Mahmood, B.; Han, S.; Seo, J. Implementation experiments on convolutional neural network training using synthetic images for 3D pose estimation of an excavator on real images. Autom. Constr. 2022, 133, 103996. [Google Scholar] [CrossRef]
  16. Tang, J.; Wang, M.; Luo, H.; Wong, P.K.Y.; Zhang, X.; Chen, W.; Cheng, J.C. Full-body pose estimation for excavators based on data fusion of multiple onboard sensors. Autom. Constr. 2023, 147, 104694. [Google Scholar] [CrossRef]
  17. Cui, Y.; An, Y.; Sun, W.; Hu, H.; Song, X. Memory-Augmented Point Cloud Registration Network for Bucket Pose Estimation of the Intelligent Mining Excavator. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  18. Wang, X.X.; Gao, J.; Feng, L. Recognition and 3D Pose Estimation for Underwater Objects Using Deep Convolutional Neural Network and Point Cloud Registration. In Proceedings of the 2020 International Conference on System Science and Engineering (ICSSE), Kagawa, Japan, 31 August–3 September 2020. [Google Scholar]
  19. Lin, C.M.; Tsai, C.Y.; Lai, Y.C.; Li, S.A.; Wong, C.C. Visual Object Recognition and Pose Estimation Based on a Deep Semantic Segmentation Network. IEEE Sens. J. 2018, 18, 9370–9381. [Google Scholar] [CrossRef]
  20. Wong, J.M.; Kee, V.; Le, T.; Wagner, S.; Mariottini, G.L.; Schneider, A.; Hamilton, L.; Chipalkatty, R.; Hebert, M.; Johnson, D.M.; et al. SegICP: Integrated deep semantic segmentation and pose estimation. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; IEEE: Ithaca, NY, USA, 2017. [Google Scholar]
  21. Pramatarov, G.; De Martini, D.; Gadd, M.; Newman, P. BoxGraph: Semantic Place Recognition and Pose Estimation from 3D LiDAR. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022. [Google Scholar]
  22. Wang, Z.; Fan, J.; Jing, F.; Liu, Z.; Tan, M. A pose estimation system based on deep neural network and ICP registration for robotic spray painting application. Int. J. Adv. Manuf. Technol. 2019, 104, 285–299. [Google Scholar] [CrossRef]
  23. Besl, P.J.; McKay, N.D. A method for registration of 3-D shapes. In SPIE 1611, Proceedings of the Sensor Fusion IV: Control Paradigms and Data Structures, Boston, MA, USA, 12–15 November 1991; SPIE: Bellingham, WA, USA, 1992. [Google Scholar]
  24. Zeng, A.; Yu, K.T.; Song, S.; Suo, D.; Walker, E.; Rodriguez, A.; Xiao, J. Multi-view self-supervised deep learning for 6D pose estimation in the Amazon Picking Challenge. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017. [Google Scholar]
  25. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar]
  26. Lin, H.Y. 3D Object Detection and 6D Pose Estimation Using RGB-D Images and Mask R-CNN. In Proceedings of the 2020 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Glasgow, UK, 19–24 July 2020. [Google Scholar]
  27. Yang, G.; Wang, S.; Yang, J.; Shen, B. Active Pose Estimation of Daily Objects. In Proceedings of the 2018 IEEE International Conference on Mechatronics and Automation (ICMA), Changchun, China, 5–8 August 2018. [Google Scholar]
  28. Van Tran, L.; Lin, H.Y. BiLuNetICP: A Deep Neural Network for Object Semantic Segmentation and 6D Pose Recognition. IEEE Sens. J. 2021, 21, 11748–11757. [Google Scholar] [CrossRef]
  29. Engel, N.; Belagiannis, V.; Dietmayer, K. Point Transformer. IEEE Access 2021, 9, 134826–134840. [Google Scholar] [CrossRef]
  30. Zhang, J.; Yao, Y.; Deng, B. Fast and Robust Iterative Closest Point. IEEE Trans Pattern Anal Mach Intell. 2022, 44, 3450–3466. [Google Scholar] [CrossRef] [PubMed]
  31. Walker, H.F.; Ni, P. Anderson acceleration for fixed-point iterations. SIAM J. Numer. Anal. 2011, 49, 1715–1735. [Google Scholar] [CrossRef] [Green Version]
  32. Weng, Y.; Wang, H.; Zhou, Q.; Qin, Y.; Duan, Y.; Fan, Q.; Chen, B.; Su, H.; Guibas, L.J. CAPTRA: CAtegory-level Pose Tracking for Rigid and Articulated Objects from Point Clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  33. Wen, B.; Mitash, C.; Ren, B.; Bekris, K.E. se(3)-TrackNet: Data-driven 6D Pose Tracking by Calibrating Image Residuals in Synthetic Domains. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020. [Google Scholar]
  34. Li, Y.; Wang, G.; Ji, X.; Xiang, Y.; Fox, D. DeepIM: Deep Iterative Matching for 6D Pose Estimation. Int. J. Comput. Vis. 2020, 128, 683–698. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The traditional open-pit mine site.
Figure 1. The traditional open-pit mine site.
Sensors 23 06958 g001
Figure 2. The proposed bucket pose estimation framework.
Figure 2. The proposed bucket pose estimation framework.
Sensors 23 06958 g002
Figure 3. The simulation environment built in CARLA; (a) simulate open-pit mine scenarios; (b) working scenario of excavator.
Figure 3. The simulation environment built in CARLA; (a) simulate open-pit mine scenarios; (b) working scenario of excavator.
Sensors 23 06958 g003
Figure 4. The performance of the PT model when adjusting different parameters. (a) Neighboring points; (b) the number of groups; (c) dropout rate; (d) network layers.
Figure 4. The performance of the PT model when adjusting different parameters. (a) Neighboring points; (b) the number of groups; (c) dropout rate; (d) network layers.
Sensors 23 06958 g004aSensors 23 06958 g004b
Figure 5. The mIoU curves of different models.
Figure 5. The mIoU curves of different models.
Sensors 23 06958 g005
Figure 6. The visualized comparison of the output segmented point clouds. (a) Ground truth; (b) PT; (c) Sparse-Unet.
Figure 6. The visualized comparison of the output segmented point clouds. (a) Ground truth; (b) PT; (c) Sparse-Unet.
Sensors 23 06958 g006
Figure 7. The registration results of the bucket at poses (ad) with four registration methods. The target point clouds are marked in blue. The real-time point clouds are marked in green. The transformed point clouds are marked in red.
Figure 7. The registration results of the bucket at poses (ad) with four registration methods. The target point clouds are marked in blue. The real-time point clouds are marked in green. The transformed point clouds are marked in red.
Sensors 23 06958 g007
Figure 8. Registration results of bucket at poses (ac) for the real-world dataset. The target point clouds are marked in blue. The real-time point clouds are marked in green. The transformed point clouds are marked in red.
Figure 8. Registration results of bucket at poses (ac) for the real-world dataset. The target point clouds are marked in blue. The real-time point clouds are marked in green. The transformed point clouds are marked in red.
Sensors 23 06958 g008
Figure 9. Registration results when the bucket is heavily occluded. (a) Original point clouds; (b) segmented bucket point clouds; (c) registration results.
Figure 9. Registration results when the bucket is heavily occluded. (a) Original point clouds; (b) segmented bucket point clouds; (c) registration results.
Sensors 23 06958 g009aSensors 23 06958 g009b
Table 1. Best-performing parameter combination.
Table 1. Best-performing parameter combination.
CategoryParametersOriginal PTOptimized PT
Model PerformancemIoU0.97560.9813
Optimal Training Epoch9490
Self-Attention HyperparametersNeighboring Points168
Weight-EncodingGVAGVA
Groups(12, 24, 48)(6, 12, 24)
Network HyperparametersNetwork Layers(2, 6, 2)(2, 6, 2)
Feature Output Dimensions(96, 192, 384)(96, 192, 384)
Parameter Scale3.9 million3.8 million
Training HyperparametersBatch Size66
Dropout Rate00.5
Table 2. The PT model parameter adjustments.
Table 2. The PT model parameter adjustments.
ParametersBaseNeighboring PointsGroupsDropout RateNetwork Layers
mIoU0.97560.96770.98260.97360.97960.97910.98130.96910.9705
Neighboring Points16328161616161616
Groups(12, 24, 48)(12, 24, 48)(12, 24, 48)(24, 48, 96)(6, 12, 24)(12, 24, 48)(12, 24, 48)(12, 24, 48)(12, 24, 48)
Network Layers(2, 6, 2)(2, 6, 2)(2, 6, 2)(2, 6, 2)(2, 6, 2)(2, 6, 2)(2, 6, 2)(3, 8, 3)(1, 4, 1)
Parameter Scale3.9 million3.9 million3.9 million4 million3.8 million3.9 million3.9 million5.34 million2.47 million
Batch size666666666
Dropout Rate000000.250.500
Table 3. The results for the simulated dataset.
Table 3. The results for the simulated dataset.
MethodOAmAmIoUBackgroundBoomBucket
Sparse-Unet0.99320.97810.96930.99240.98450.9309
OPT0.99470.98630.97770.99540.98490.9529
PT0.99560.98730.98160.99600.98820.9605
Table 4. The results for the real-world dataset.
Table 4. The results for the real-world dataset.
MethodOAmAmIoUBackgroundBoomBucket
Sparse-Unet0.96450.94330.89600.96370.89530.8290
OPT0.92690.85680.80120.91610.85990.6276
PT0.95950.91440.88530.95330.83210.8703
Table 5. Rotation errors (◦) and translation errors (m) of bucket pose estimation.
Table 5. Rotation errors (◦) and translation errors (m) of bucket pose estimation.
MethodRE (◦)TE (m)Time (ms)
SAC-IA12.01960.2195725
ICP20.27330.40191140
RANSAC8.91540.68534386
SIFT+RANSAC4.06210.05591011
GICP2.06210.031220
FICP1.20580.025175
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Z.; Bi, L.; Zhao, Z. Real-Time Bucket Pose Estimation Based on Deep Neural Network and Registration Using Onboard 3D Sensor. Sensors 2023, 23, 6958. https://doi.org/10.3390/s23156958

AMA Style

Xu Z, Bi L, Zhao Z. Real-Time Bucket Pose Estimation Based on Deep Neural Network and Registration Using Onboard 3D Sensor. Sensors. 2023; 23(15):6958. https://doi.org/10.3390/s23156958

Chicago/Turabian Style

Xu, Zijing, Lin Bi, and Ziyu Zhao. 2023. "Real-Time Bucket Pose Estimation Based on Deep Neural Network and Registration Using Onboard 3D Sensor" Sensors 23, no. 15: 6958. https://doi.org/10.3390/s23156958

APA Style

Xu, Z., Bi, L., & Zhao, Z. (2023). Real-Time Bucket Pose Estimation Based on Deep Neural Network and Registration Using Onboard 3D Sensor. Sensors, 23(15), 6958. https://doi.org/10.3390/s23156958

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop