Next Article in Journal
Machine Learning Modeling for Spatial-Temporal Prediction of Geohazard
Previous Article in Journal
Highly Stable Spatio-Temporal Prediction Network of Wavefront Sensor Slopes in Adaptive Optics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Underground Tunnel Place Recognition Algorithm Based on Farthest Point Subsampling and Dual-Attention Transformer

1
54th Research Institute of China Electronics Technology Group Corporation, Shijiazhuang 050081, China
2
School of Mechanical Electronic and Information Engineering, China University of Mining and Technology, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(22), 9261; https://doi.org/10.3390/s23229261
Submission received: 10 October 2023 / Revised: 2 November 2023 / Accepted: 15 November 2023 / Published: 18 November 2023
(This article belongs to the Section Navigation and Positioning)

Abstract

:
An autonomous place recognition system is essential for scenarios where GPS is useless, such as underground tunnels. However, it is difficult to use existing algorithms to fully utilize the small number of effective features in underground tunnel data, and recognition accuracy is difficult to guarantee. In order to solve this challenge, an efficient point cloud position recognition algorithm, named Dual-Attention Transformer Network (DAT-Net), is proposed in this paper. The algorithm firstly adopts the farthest point downsampling module to eliminate the invalid redundant points in the point cloud data and retain the basic shape of the point cloud, which reduces the size of the point cloud and, at the same time, reduces the influence of the invalid point cloud on the data analysis. After that, this paper proposes the dual-attention Transformer module to facilitate local information exchange by utilizing the multi-head self-attention mechanism. It extracts local descriptors and integrates highly discriminative global descriptors based on global context with the help of a feature fusion layer to obtain a more accurate and robust global feature representation. Experimental results show that the method proposed in this paper achieves an average F 1 score of 0.841 on the SubT-Tunnel dataset and outperforms many existing state-of-the-art algorithms in recognition accuracy and robustness tests.

1. Introduction

Place recognition is a fundamental capability for robot state estimation and is widely applied in robotic systems such as autonomous cars. In restricted environments like underground tunnels where GPS [1,2] is not available, place recognition can provide precise global positioning information for vehicles or robots, ensuring their stable operation in underground tunnels. Specifically, in stagewise pose estimation, place recognition adds loopback constraints to a vehicle’s repeated arrival intervals, reducing the cumulative error in globally consistent localization. Secondly, place recognition also helps initialize the positioning system for continuous posture tracking and is also responsible for acting as a backup positioning scheme in case of odometer drift. Therefore, the study of an efficient position identification method is of great significance in realizing global localization estimation [3] in underground tunnels. Existing 3D LiDAR-based position recognition methods have primarily been developed to address position recognition problems on urban roads. To the best of our knowledge, current 3D LiDAR-based position recognition methods have not been developed for underground tunnel environments and have not been tested on underground tunnel datasets.
Underground tunnels exhibit significant large-scale geometric repetition due to their unique scene configuration. Sensor data acquired in such environments inevitably lack effective geometric features that can be easily distinguished. Existing algorithms, which are mainly experimented with in usual environments such as school campuses and highways, generally rely on recognizable descriptive features in the data for place recognition. However, when faced with underground tunnel scenes with a large number of repeating 3D structures and similar textures, it is difficult to guarantee the original recognition accuracy. It is even more difficult to deal with the tricky situations that often occur in underground tunnels, such as obstacles blocking the sensors and changes in sensor viewpoints.
A few researchers have designed place recognition algorithms based on semantic information, focusing more on the influence of point cloud categories on the recognition effect. SG-PR [4], on the other hand, transforms point clouds into graph-structured data with the help of existing semantic segmentation methods. It describes the scene at the semantic level and focuses on the encoding relationships between semantic objects. However, the recognition accuracy of this method is heavily dependent on the performance of the segmentation algorithm, and the algorithm runs with high complexity. RINet [5] designs a rotationally invariant global descriptor and utilizes semantic information and geometric features to improve its discriminative ability. It has low computational complexity, but relies on highly accurate semantic labeling and spends extra time converting data formats. In addition, descriptor-based place recognition methods have also received extensive attention. M2DP [6] parses the point cloud from a multi-view perspective and obtains the density distribution of each point cloud based on the spatial density distribution characteristics of the points on the plane. The method is novel and robust, but it is difficult to guarantee the recognition accuracy due to the loss of some data features when projecting the point cloud data. SegMatch [7] segments the point cloud and drastically reduces the number of matches by clustering. It first encodes the features using a 3D convolutional neural network (CNN), after which the corresponding candidate values are identified through k-nearest neighbors (KNN), and finally, a geometric validation step is used to convert the candidate values into location-aware candidates. This approach combines the advantages of local and global descriptors, but the connection between objects is not fully considered, and the recognition accuracy needs to be improved. PointNetVLAD [8] innovatively adopts PointNet [9] as the backbone network to extract point cloud features, with targeted consideration of point cloud substitution invariance and affine invariance. PCAN [10] improves on the above method. It first extracts multi-scale local contextual information to generate point attention maps via the group sampling method of PointNet++ [11], and then uses NetVLAD [12] to aggregate the attention-weighted local features into a global descriptor. DAGC [13] again improves the above algorithm by combining the dynamic graph architecture and the dual-attention mechanism to aggregate local context information and then using the NetVLAD layer to aggregate the global descriptor. The above algorithm has been continuously improved and has gained some improvement in accuracy. However, the generalized PointNet-based point cloud analysis architecture makes it difficult to improve the recognition accuracy of complex underground scenes in a targeted manner. SeqLPD [14] and LPD-Net [15] first fuse the neighborhood features of each point in feature space and Cartesian space, and then use NetVLAD to generate global descriptors. However, some of the effective information is lost when 3D data are projected to 2D. SOE-Net [16] accomplishes the recognition task end-to-end. The network first extracts local descriptors point-by-point using the PointOE module, which is a combination of Pointnet and an Orientation-encoding Unit, and then aggregates differentiated global descriptors using the Self-Attention module and NetVLAD. However, the loss function threshold of this method needs to be set in advance.
In summary, some advanced algorithms rely on the accuracy of semantic segmentation algorithms, and it is difficult to independently and stably realize the autonomous recognition of underground tunnels. Meanwhile, the advanced place recognition methods based on descriptors make it difficult to ensure recognition accuracy due to the difficulty of fully utilizing the small number of effective features in the underground tunnel data. To address these issues, this paper introduces an efficient point cloud place recognition algorithm, DAT-Net. This method first employs a farthest point subsampling module to remove invalid redundant points from the point cloud data, reducing the point cloud scale while preserving its basic shape and geometric features to improve algorithm efficiency and convergence speed. Next, this paper proposes a dual-attention Transformer module, which utilizes dot-product attention and sinc attention to adaptively extract important features to enhance distortion sensitivity. Specifically, by using the simple and efficient parallelizable dot-product attention, attention is focused on the K key V value pairs related to the Q query vector, capturing local features in the data and adaptively enhancing local information in the point cloud. Additionally, the proposed sinc attention is used for local feature filtering, improving fine-grained quality prediction performance in underground tunnel scenes. This dual-attention strategy effectively enhances the extraction of local descriptors and integrates highly discriminative global descriptors based on the global context using the feature fusion layer, resulting in more accurate and robust global feature representation. Here are the contributions of this article:
  • The use of the farthest point subsampling module significantly reduces the point cloud size, decreasing the computational complexity of the model while preserving point cloud features.
  • A point cloud analysis module based on the dual-attention layer Transformer has been developed, enhancing the accuracy and robustness of place recognition.
  • Experiments in place recognition and robustness testing on an underground track dataset demonstrate that our approach performs exceptionally well, achieving an average dispersion of 0.841.

2. Model Design

2.1. Overall Network Design

In order to improve the accuracy of the place recognition algorithm in underground tunnels, this paper proposes an efficient point cloud place recognition algorithm (DAT-Net), as shown in Figure 1. The algorithm first uses the farthest point downsampling module to eliminate the invalid redundant points in the point cloud data and retain the basic shape of the point cloud to reduce the size of the point cloud and at the same time reduce the impact of the invalid point cloud on the data parsing. After that, the transformed feature space solver module is used to dimensionally diffuse the point cloud and map the spatial flip suitable for feature extraction. Finally, the Transformer’s dual-attention layer is used to facilitate the exchange of multiple pieces of local information and integrate highly discriminative global descriptors according to the feature fusion module to obtain a more accurate and robust global feature representation.

2.2. Farthest Point Downsampling

Compared to images, a single frame of 3D LiDAR captures a significantly larger amount of data, and using this directly as input for a Transformer would lead to a substantial increase in model size. Refs [6,8,15] have used random downsampling methods or projected 3D data into 2D images to reduce the point cloud data volume. However, they lose too many point cloud features during downsampling, making it challenging to guarantee algorithm accuracy. Inspired by the research on PointNet [9], we believe that the outer points of an object hold more analytical value than the inner points because, in an ideal scenario, a sufficient number of outer points can adequately represent an object. Therefore, this paper adopts the farthest point downsampling algorithm, which reduces the scale of the point cloud while preserving environmental shape features as much as possible. The process is shown below:
  • Input: initial point cloud data { p 1 , p 2 , , p i }.
  • Select start point: randomly select a point as p o from the data or according to the density.
  • Initialize the subsampled point set: Begin by creating an empty subsampled point set S, and then add P 0 to the subsampled point set, resulting in S = { P 0 }.
  • Loop until the termination condition is met:
    • For each point p i in the initial point cloud, compute the Euclidean distance d i from point p o .
      d i = [ p i x p o x ] 2 + [ p i y p o y ] 2 + [ p i z p o z ] 2
    • Sort the calculated Euclidean distance values and select the point furthest from point p o in the downsampled point set S.
    • Update the selected point to p o .
  • Termination condition: the preset number of downsampled points N is reached.
  • Output: downsampled point cloud data X R N × 3 .
As shown in Figure 2, a frame of point cloud data from the SubT-Tunnel dataset typically contains around 120,000 points. Over-representation of the environment in some point-concentrated regions appears to be redundant. When the number of downsampling points N is set to 10,240 or 5120, it can be seen that the points in some areas are still very concentrated. After a series of attempts, it is found that when N is set to 1024, the downsampled point cloud can still express the shape of the original point cloud well, and will not be as sparse as the point cloud when N is set to 256. Therefore, this paper finally sets the number of downsampled points N to 1024. N = 1024 is the optimal value for this method. The specifications of the sensors and the environmental conditions during data collection influence the shape, density, and features of the original point cloud. The value of N for farthest point subsampling should be adjusted based on the specific environment and sensors.

2.3. Solution Module for Transformed Eigenspaces

Setting the step size of convolutional neural networks (CNNs) to 1 is capable of achieving translational isotropy in image analysis. However, in point clouds, 0 degrees and 359 degrees are adjacent, which means limiting the step size of CNNs to 1 does not fully realize rotational isotropy for point cloud analysis. Therefore, a convolutional neural network suitable for point cloud analysis is required to better advance the performance of place recognition. An intuitive solution is to use a recurrent convolutional neural network with a step size of 1. Its greatest property is rotational invariance, which is well-suited for point cloud analysis. First, the downsampled point cloud data X R L × 3 are subjected to three rotationally invariant convolution blocks, and then feature extraction is performed level by level to obtain different levels of features F 1 , F 2 , and F 3 . The rotation invariant convolution can be expressed in the following form:
V ¯ ( i ) = [ V × K ] ( i ) = c = 0 N 1 m = M M V [ ( i m ) mod N , c ] × K ( M + m , c )
where i [ 0 , N 1 ] and K R ( 2 M + 1 ) × N denotes the convolution kernel.
As shown in Figure 3, the shallower feature F 1 is passed into the coordinate transformation module for feature dimension diffusion using 1D convolution. After that, the transformation matrix R is gathered and computed with the help of a linear connection layer. After flipping the feature F 1 , it is fused with the deeper feature F 3 to enhance the adaptability of the algorithm to the scene perspective change.

2.4. Transformer

Transformer and self-attention mechanisms have made a number of contributions in the field of machine translation and NLP. This has inspired the development of self-attention models for images, the most influential of which is ViT [17]. Refs. [18,19] tried to analyze the whole point cloud with self-attention and found that its learning mechanism is very effective for point clouds. This paper introduces a dual-attention Transformer module, which utilizes both dot-product attention and sinc attention to adaptively extract important features, thereby enhancing sensitivity to distortions. Specifically, it focuses attention on K key–value pairs related to the Q query vector using the simple, efficient, and parallelizable dot-product attention. This helps capture local features in the data, adaptively reinforcing local information within the point cloud. Additionally, the proposed sinc attention is employed to perform local feature filtering and establish long-term dependencies, enhancing the algorithm’s fine-grained quality prediction performance in underground tunnel scenes. This dual-attention strategy effectively improves the extraction of local descriptors. Subsequently, with the help of the feature fusion layer, it integrates highly distinguishable global descriptors based on the global context to obtain a more accurate and robust global feature representation. More specifically, the feature fusion layer autonomously learns and assigns larger weight values to high-quality local descriptors, ensuring that the aggregated global descriptors can better describe the entire point cloud.
The multi-head self-attention layer is shown in Figure 4. The feature F is first equated into local descriptors Q R N × D / h , K R N × D / h , and V R N × D / h after a sigmoid layer, a rotationally invariant convolutional layer, and a linearly connected layer. Subsequently, the first local feature extraction is performed in the first attention layer. The computational process is represented by Equation (2) as follows:
M u l t i h e a d ( Q , K , V ) = C o n t a c t ( h e a d 1 , , h e a d h ) B O h e a d i = s o f t max ( Q B i Q ( K B i K ) D / h ) V B i V
where B is the matrix of learnable parameters. B i Q , B i K , and B i V R N × D / h and B O R D × D .
The second attention layer is proposed to parse the point cloud features in more depth and is named as the Sine Attention Mechanism. It first substitutes features Q, K, and V into the Sinc function to obtain a smoother numerical representation. After that, the parameters are exchanged and shared through Multilayer Perceptron (MLP). Based on this, feature multiplication by Q and K is performed to calculate the assigned weights of local features. The formula for the sinusoidal attention mechanism is expressed as follows:
M u l t i h e a d ( Q , K , V ) = C o n t a c t ( h e a d 1 , , h e a d h ) h e a d i = S A [ M L P ( Q i K i T ) ] i = 1 N S A [ M L P ( Q i K i T ) ] S A [ M L P ( V i ) ]
The size of the features remains constant after multiple attention layers and the quality of the local descriptors can be improved by stacking the network layers L times. The final output is a splice of N D-dimensional local descriptors F R N × D . The h in the multi-head attention is set to 12 and the number of loops L is also set to 12.

2.5. Feature Fusion Module

Inspired by SimGNN [20] and SG-PR [4], this paper assigns weights to local descriptors with the help of the attention mechanism. It learns autonomously and assigns larger weight values to high-quality local descriptors so that the aggregated global descriptors can better describe the whole point cloud.
As shown in Figure 5, the learnable parameter matrix J R D × D is set. First, J is multiplied with the average of all local descriptors to construct a global context to provide global structure and feature information. After that, a sigmoid function is used to ensure that the weights are in [0, 1] to assign a weight value to each local descriptor. Finally, the final global descriptor e R D is obtained through summation, which is expressed by the following:
e = i = 1 N a i F i a i = s i g m o i d F i tanh c o n t a c t 1 N m = 1 N U m 1 J 1 , 1 N m = 1 N U m 2 J 2 T

2.6. Loss Function

In this paper, two linear connected layers are used to reduce the dimensionality of the matching vectors, and a sigmoid function is used to obtain the similarity p ( y i ) [ 0 , 1 ] of the scenes. The loss function used for training is the binary cross-entropy loss function with the following equation:
L oss = 1 N i = 1 N y i log [ p ( y i ) ] + ( 1 y i ) log [ 1 p ( y i ) ]
where y i is the binary label 0 or 1, and p ( y i ) is the probability that the output belongs to label y i .

3. Test Results and Performance Analysis

In order to evaluate the effectiveness of the method proposed in this paper, the experiments are conducted using the SubT-Tunnel [21] underground tunnel dataset, which consists of three different tunnel scenarios, S1, S2, and E1. The server for training uses an Intel® CoreTM i7-6950X CPU processor with four NVIDIA GeForce GTX 1080Ti, each with 11G of memory. The experiments are implemented on a Pytorch [22] deep learning framework based on the Ubuntu 18.04 operating system with GPU (GTX1080Ti) for training and testing. In this paper, the model parameters are optimized using the Adam [23] optimizer, where the initial learning rate is set to 0.0001, learning decreases with the number of iterations, weight decay is 0.0005, batch size is 1024, and the maximum number of iterations is 500 rounds.

3.1. Experimental Setup

In the experiments, point cloud pairs with timestamps differing by more than 30 s and Euclidean distances less than 3 m will be defined as positive samples, and point cloud pairs with Euclidean distances more than 20 m will be defined as negative samples. During training, sample pairs with timestamps less than 30 s are also included in the training set to increase the interference term. During the evaluation process, positive and negative samples with a time difference of less than 30 s are eliminated. This means that the performance of the algorithm is not evaluated based on simple pairs of positive samples (neighboring scenarios), allowing for a more accurate assessment of its capabilities in real-world environments. In this paper, DAT-Net is compared with state-of-the-art algorithms such as M2DP [6], LPD-NET [15], DISCO [24], PointNetVLAD [8], SeqOT [25], and RINET [5]. For a fair comparison, the F 1 score [26] value is used uniformly to measure the algorithm performance. F 1 is defined as follows, where P denotes precision and R denotes recall.
F 1 = 2 × P × R P + R

3.2. Analysis of Experimental Results

Qualitative Analysis: Figure 6 shows the recall–precision plot, and the P-R curve of DAT-Net shows that it has higher recall and precision compared to the other algorithms in each scene, which indicates that the method proposed in this paper is able to maintain better recognition in different underground tunnel scenes. In addition, it can be seen that unlike the S1 and S2 scenarios, the performance of each algorithm decreases somewhat in the mega-scene E1. However, the proposed algorithm DAT-Net maintains a stable performance in the E1 scenario compared to other state-of-the-art algorithms, which indicates that the accuracy and stability of DAT-Net are excellent.
Quantitative Analysis: According to the results in Table 1, DAT-Net achieved excellent experimental results on the underground tunnel dataset, i.e., an average maximum score of 0.841. In this experiment, DAT-Net scored on average 0.134 points higher than the second-place RINet in the S1, S2, and E1 scenarios. The experimental results of M2DP and LPD-Net were unsatisfactory, mainly because they converted the data to top view, resulting in the loss of some features of the underground tunnel data. SeqOT also lost some valid information in the data conversion. PointNetVLAD only considers the features of a single point and ignores their connection with neighboring features. Underground tunnel scenes have more sparse effective features compared to normal scenes, so it is difficult to use the above algorithms to show higher accuracy in underground tunnels due to the lack of effective utilization of features. It is worth noting that the E1 series used in the experiments were recorded in an ultra-large-scale complex underground tunnel environment, and all the algorithms decreased in accuracy for this series. DAT-Net still obtained a high score as high as 0.797 in the E1 scene, indicating that the method proposed in this paper still has excellent recognition capability and stability in large underground tunnel scenes.

3.3. Robustness Testing

Obstacle occlusion and sensor view angle changes are issues that need to be addressed in the daily tasks of autonomous driving.
As shown in Figure 7, obstacle occlusion directly leads to incomplete data information, and the change of view angle is not visible from the human perspective, but it is a more drastic change to the original data for the electronic processor. In order to verify the impact of the above problems on the place recognition algorithm and to further test the stability and robustness of the algorithm, this paper simulates the underground tunnel scene with obstacle occlusion and view angle change to compare the experiment with other state-of-the-art algorithms.
Occlusion test: Sensor occlusion is a common phenomenon in automobile and robot operation that can easily lead to incomplete data acquired by sensors. For place recognition, using data with missing important information features can significantly affect the algorithm performance. In order to evaluate the processing ability of different methods for defective data, in this paper, points in a region are randomly removed from the point cloud. The results shown in Table 2 indicate that the method proposed in this paper significantly outperforms the other algorithms in this experiment, exceeding the second-place algorithm, RINet, by 0.340 on average. It exceeds by 0.234 in the S1 scenario, 0.244 in the S2 scenario, and 0.244 in the E1 scenario. This shows that the missing information has a great impact on RINet, which prevents it from maintaining its performance in normal situations. The DAT-Net method proposed in this paper reduces the amount of point cloud data in the data preprocessing stage and maintains the data features as much as possible, which effectively mitigates the performance degradation of the algorithm due to the missing information. This demonstrates that the method proposed in this paper also exhibits strong robustness in environments with obstacle occlusion.
Rotation test: Sensor viewpoint changes are also a common phenomenon in autonomous driving. For the human eye, it is not difficult to recognize a scene that is at a different view angle when passing by the same place. However, for position recognition, the change in view angle brings a completely different form of data. In order to evaluate the ability of the different methods to handle changes in sensor view angle, this paper randomly rotates each point in the point cloud. The results shown in Table 3 indicate that DAT-Net still significantly outperforms the other algorithms in the experiments, exceeding the second-place algorithm RINet by 0.246 on average. It exceeds by 0.256 in the S1 scene, 0.241 in the S2 scene, and 0.246 in the E1 scene. RINet produces a more severe accuracy degradation under the influence of perspective changes. The reason for this is that the missing information only affects part of the data, while the change in view angle changes all the data. The proposed method, DAT-Net, uses the transformed feature space solver module to autonomously learn favorable transform matrices to enhance the parsing effect on the data, which effectively mitigates the algorithm’s performance degradation caused by changes in the sensor’s view angle.
The obstacle occlusion and viewpoint change tests demonstrate the stability and robustness of the method proposed in this paper for underground tunnel scenarios.

3.4. Algorithm Efficiency

As shown in Table 4, our method has 3.18 M parameters, 1.12 GFLOPs, and a recognition time of 36.0 ms for a single frame of data. Taking the TX2 embedded device used in practical scenarios as an example, with its hardware performance reaching 1.3 TFLOPS, it is more than sufficient to meet the operational requirements of our method. Additionally, our algorithm also fulfills the fundamental real-time processing needs.

4. Conclusions

Using the existing algorithms, it is difficult to fully utilize the small number of effective features in the underground tunnel data, and the recognition accuracy is difficult to guarantee for other problems. This paper proposes an efficient underground tunnel location identification algorithm (DAT-Net) that combines point cloud preprocessing and a dual-attention Transformer. The algorithm firstly adopts the farthest point downsampling module to eliminate the invalid redundant points in the point cloud data and retain the basic shape of the point cloud so as to reduce the size of the point cloud and at the same time reduce the influence of the invalid point cloud on the data parsing. After that, the dual-attention Transformer module is proposed to utilize the multi-head self-attention mechanism to promote local information exchange and extract local descriptors. With the help of the feature fusion layer, the highly recognizable global descriptors are integrated according to the global context, obtaining a more accurate and robust global feature representation. The experimental results show that the method proposed in this paper achieves an average F 1 score of 0.841 on the NVIDIA GTX 1080Ti experimental platform for the underground tunnel dataset, and it performs well in the recognition accuracy and robustness tests compared to many existing state-of-the-art algorithms. Therefore, it is able to meet the demand for high accuracy and robust position recognition for complex underground tunnel scenes.

Author Contributions

Methodology, X.C.; Formal analysis, T.Y.; Investigation, C.D.; Data curation, C.D.; Writing—original draft, X.Y.; Writing—review & editing, X.C.; Visualization, J.Y.; Funding acquisition, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62101517.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The SubT-Tunnel dataset can be obtained from the following website: https://ieeexplore.ieee.org/abstract/document/9197156, accessed on 15 September 2020.

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

References

  1. Carvalho, G.S.; Silva, F.O.; Pacheco, M.V.O.; Campos, G.A.O. Performance Analysis of Relative GPS Positioning for Low-Cost Receiver-Equipped Agricultural Rovers. Sensors 2023, 23, 8835. [Google Scholar] [CrossRef]
  2. Cheng, C.; Li, X.; Xie, L.; Li, L. A Unmanned Aerial Vehicle (UAV)/Unmanned Ground Vehicle (UGV) Dynamic Autonomous Docking Scheme in GPS-Denied Environments. Drones 2023, 7, 613. [Google Scholar] [CrossRef]
  3. Lai, J.; Liu, S.; Xiang, X.; Li, C.; Tang, D.; Zhou, H. Performance Analysis of Visual–Inertial–Range Cooperative Localization for Unmanned Autonomous Vehicle Swarm. Drones 2023, 7, 651. [Google Scholar] [CrossRef]
  4. Xin, K.; Yang, X.M.; Zhai, G.Y.; Zhao, X.G.; Zeng, X.F. Semantic graph based place recognition for 3D point clouds. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25–29 October 2020; pp. 8216–8223. [Google Scholar]
  5. Li, L.; Kong, X.; Zhao, X.R.; Huang, T.X.; Li, W.L. RINet: Efficient 3D lidar-based place recognition using rotation invariant neural network. IEEE Robot. Autom. Lett. 2022, 7, 4321–4328. [Google Scholar] [CrossRef]
  6. He, L.; Wang, X.; Zhang, H. M2DP: A novel 3D point cloud descriptor and its application in loop closure detection. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 231–237. [Google Scholar] [CrossRef]
  7. Dubé, R.; Dugas, D.; Stumm, E.; Nieto, J.; Siegwart, R.; Cadena, C. SegMatch: Segment based place recognition in 3D point clouds. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5266–5272. [Google Scholar] [CrossRef]
  8. Uy, M.A.; Lee, G.H. PointNetVLAD: Deep point cloud based retrieval for large-scale place recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4470–4479. [Google Scholar]
  9. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  10. Zhang, W.X.; Xiao, C.X. PCAN: 3D attention map learning using contextual information for point cloud based retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 12436–12445. [Google Scholar]
  11. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5099–5108. [Google Scholar]
  12. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN architecture for weakly supervised place recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 27–30 June 2016; pp. 5297–5307. [Google Scholar]
  13. Sun, Q.; Liu, H.Y.; He, J.; Fan, Z.X.; Du, X.Y. DAGC: Employing dual attention and graph convolution for point cloud based place recognition. In Proceedings of the International Conference on Multimedia Retrieval, Dublin, Ireland, 8–11 June 2020; pp. 224–232. [Google Scholar]
  14. Liu, Z.; Suo, C.Z.; Zhou, S.B.; Xu, F.; Wei, H.S. SeqLPD: Sequence mtching enhanced loop-closure detection based on large-scale point cloud description for self-driving vehicles. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Macau, China, 4–8 November 2019; pp. 1218–1223. [Google Scholar]
  15. Liu, Z.; Zhou, S.B.; Suo, C.Z.; Liu, Y.T.; Yin, P. LPD-Net: 3D point cloud learning for large-scale place recognition and environment analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2831–2840. [Google Scholar]
  16. Xia, Y.; Xu, Y.S.; Li, S.; Wang, R.; Du, J. SOE-Net: A self-attention and orientation encoding network for point cloud based place recognition. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11348–11357. [Google Scholar]
  17. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.H. An image is worth 16x16 words:Transformers for image recognition at scale. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  18. Guo, M.H.; Cai, J.X.; Liu, Z.N.; Mu, T.J.; Martin, R.R. PCT: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  19. Zhao, H.S.; Jiang, L.; Jia, J.Y.; Torr, P.H.S.; Koltun, V. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  20. Bai, Y.; Ding, H.; Bian, S.; Chen, T.; Sun, Y.; Wang, W. Simgnn:A neural network approach to fast graph similarity computation. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, VIC, Australia, 11–15 February 2019; pp. 384–392. [Google Scholar]
  21. Rogers, J.G.; Gregory, J.M.; Fink, J.; Stump, E. Test your SLAM! The SubT-Tunnel dataset and metric for mapping. In Proceedings of the IEEE International Conference on Robotics and Automation, Paris, France, 31 May–31 August 2020; pp. 955–961. [Google Scholar]
  22. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J. Pytorch: An imperative style, high-performance deep learning library. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 8024–8035. [Google Scholar]
  23. Socher, R.; Chen, D.; Manning, C.D.; Ng, A. Reasoning with neural tensor networks for knowledge base completion. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 926–934. [Google Scholar]
  24. Xu, X.; Yin, H.; Chen, Z.; Li, Y.; Wang, Y. Disco: Differentiable scan context with orientation. IEEE Robot. Autom. Lett. 2021, 6, 2791–2798. [Google Scholar] [CrossRef]
  25. Ma, J.; Chen, X.; Xu, J.; Xiong, G. SeqOT: A Spatial–Temporal Transformer Network for Place Recognition Using Sequential LiDAR Data. IEEE Trans. Ind. Electron. 2023, 70, 8225–8234. [Google Scholar] [CrossRef]
  26. Chinchor, N. MUC-4 Evaluation Metrics. In Proceedings of the Fourth Message Understanding Conference, McLean, Virginia, 16–18 June 1992; pp. 22–29. [Google Scholar]
Figure 1. Place recognition method for efficient 3D point cloud (DAT-Net).
Figure 1. Place recognition method for efficient 3D point cloud (DAT-Net).
Sensors 23 09261 g001
Figure 2. Effect of farthest point downsampling.
Figure 2. Effect of farthest point downsampling.
Sensors 23 09261 g002
Figure 3. Coordinate transformation module.
Figure 3. Coordinate transformation module.
Sensors 23 09261 g003
Figure 4. Multi-head self-attention layers.
Figure 4. Multi-head self-attention layers.
Sensors 23 09261 g004
Figure 5. Feature fusion module.
Figure 5. Feature fusion module.
Sensors 23 09261 g005
Figure 6. Accuracy–recall for place recognition.
Figure 6. Accuracy–recall for place recognition.
Sensors 23 09261 g006
Figure 7. Schematic diagram of obstacle occlusion and view angle change.
Figure 7. Schematic diagram of obstacle occlusion and view angle change.
Sensors 23 09261 g007
Table 1. Average F 1 scores of each advanced algorithm on SubT-Tunnel dataset.
Table 1. Average F 1 scores of each advanced algorithm on SubT-Tunnel dataset.
MethodsS1S2E1Mean
M2DP0.4560.5500.4280.478
LPD-Net0.4720.5400.4260.479
SeqOT0.3500.4240.3050.360
DISCO0.4720.5470.4240.481
PointNetVLAD0.4660.5450.4150.475
RINet0.7160.7060.7000.707
DAT-Net (Our)0.8370.8880.7970.841
Table 2. Average F 1 scores of each advanced algorithm for the occlusion test.
Table 2. Average F 1 scores of each advanced algorithm for the occlusion test.
MethodsS1S2E1Mean
M2DP0.4890.5530.4100.484
LPD-Net0.4810.5430.4220.482
SeqOT0.3500.4240.3050.360
DISCO0.4750.5010.4210.466
PointNetVLAD0.4750.5530.4170.476
RINet0.5670.6100.5200.566
DAT-Net (Our)0.8010.8540.7640.806
Table 3. Average F 1 scores for each advanced algorithm for rotational testing.
Table 3. Average F 1 scores for each advanced algorithm for rotational testing.
MethodsS1S2E1Mean
M2DP0.4560.5490.4280.478
LPD-Net0.4610.5440.4180.474
SeqOT0.3500.4240.3050.360
DISCO0.4760.5400.4210.479
PointNetVLAD0.4640.5410.4170.474
RINet0.5010.5850.4760.521
DAT-Net (Our)0.7570.8230.7220.767
Table 4. Efficiency demonstration of various advanced algorithms.
Table 4. Efficiency demonstration of various advanced algorithms.
MethodsParameters (M)FLOPs (G)Runtime (ms)
LPD-Net2.495.5142.3
SeqOT1.990.6727.0
DISCO7.791.3338.3
PointNetVLAD2.810.9028.4
RINet1.820.05810.7
DAT-Net (Our)3.180.11236.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chai, X.; Yang, J.; Yan, X.; Di, C.; Ye, T. Efficient Underground Tunnel Place Recognition Algorithm Based on Farthest Point Subsampling and Dual-Attention Transformer. Sensors 2023, 23, 9261. https://doi.org/10.3390/s23229261

AMA Style

Chai X, Yang J, Yan X, Di C, Ye T. Efficient Underground Tunnel Place Recognition Algorithm Based on Farthest Point Subsampling and Dual-Attention Transformer. Sensors. 2023; 23(22):9261. https://doi.org/10.3390/s23229261

Chicago/Turabian Style

Chai, Xinghua, Jianyong Yang, Xiangming Yan, Chengliang Di, and Tao Ye. 2023. "Efficient Underground Tunnel Place Recognition Algorithm Based on Farthest Point Subsampling and Dual-Attention Transformer" Sensors 23, no. 22: 9261. https://doi.org/10.3390/s23229261

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop