Next Article in Journal
Research on LEACH Protocol Based on Dynamic Clustering and Routing Optimization
Previous Article in Journal
EHFOA-ID: An Enhanced HawkFish Optimization-Driven Hybrid Ensemble for IoT Intrusion Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DTD: Density Triangle Descriptor for 3D LiDAR Loop Closure Detection

1
School of Instrument Science and Engineering, Southeast University, Nanjing 210096, China
2
School of Electrical Engineering and Automation, Suzhou University of Technology, Suzhou 215500, China
*
Authors to whom correspondence should be addressed.
Sensors 2026, 26(1), 201; https://doi.org/10.3390/s26010201 (registering DOI)
Submission received: 25 November 2025 / Revised: 17 December 2025 / Accepted: 25 December 2025 / Published: 27 December 2025

Abstract

Loop closure detection is essential for improving the long-term consistency and robustness of simultaneous localization and mapping (SLAM) systems. Existing LiDAR-based loop closure approaches often rely on limited or partial geometric features, restricting their performance in complex environments. To address these limitations, this paper introduces a Density Triangle Descriptor (DTD). The proposed method first extracts keypoints from density images generated from LiDAR point clouds, and then constructs a triangle-based global descriptor that is invariant to rotation and translation, enabling robust structural representation. Furthermore, to enhance local discriminative ability, the neighborhood around each keypoint is modeled as a Gaussian distribution, and a local descriptor is derived from the entropy of its probability distribution. During loop closure detection, candidate matches are first retrieved via hash indexing of triangle edge lengths, followed by entropy-based local verification, and are finally refined by singular value decomposition for accurate pose estimation. Extensive experiments on multiple public datasets demonstrate that compared to STD, the proposed DTD improves the average F1 max score and EP by 18.30% and 20.08%, respectively, while achieving a 50.57% improvement in computational efficiency. Moreover, DTD generalizes well to solid-state LiDAR with non-repetitive scanning patterns, validating its robustness and applicability in complex environments.

1. Introduction

Loop closure detection (LCD) is a fundamental component of simultaneous localization and mapping (SLAM) systems, aiming to determine whether the robot’s current location corresponds to a previously visited place. By establishing data associations between the current frame and historical loop candidates, LCD effectively corrects accumulated localization drift, thereby significantly improving the accuracy of pose estimation and the consistency of the global map. This mechanism has been widely integrated into SLAM systems such as LIO-SAM [1], ORB-SLAM [2], and LV-SLAM [3]. Based on the type of sensor employed, LCD methods can be broadly categorized into vision-based and LiDAR-based approaches. In recent years, vision-based place recognition methods have been extensively studied and applied [4,5,6]. However, these methods often suffer from significant limitations in scenarios with severe illumination changes or rapid motion. In contrast, LiDAR-based methods directly capture the 3D structural information of the environment through laser pulses, making them inherently more robust to changes in lighting and viewpoint. Most existing LiDAR-based loop closure detection methods adopt descriptor-based frameworks. Representative approaches such as M2DP [7] and Scan Context [8] rely on projection-based representations to encode global structural information in point clouds. More recently, several methods have further enhanced descriptor discriminability by incorporating additional cues, such as visual features during keypoint extraction [9] or reflectivity information in RE-TRIP [10], thereby improving robustness in complex environments. Inspired by the above research progress, this paper focuses on the loop closure detection problem based on LiDAR, aiming to design a descriptor that is both efficient and robust to adapt to diverse scenarios.
Compared to camera images, LiDAR point clouds are more unstructured and complex, which poses challenges for place recognition. (1) First, point clouds are sensitive to the sensor’s position and orientation. To handle viewpoint changes, LiDAR-based methods should achieve rotation and translation invariance, and several geometric descriptors such as triangles [11,12], Link3D [13], and rings [14] have been proposed. (2) Second, the descriptor must effectively capture the structural features of the environment in a compact form. (3) Finally, the descriptor should be robust enough to support relative pose estimation between the current frame and loop candidates, helping reduce drift in mapping and localization.
To address these challenges, we propose Density Triangle Descriptor (DTD), a spatial descriptor based on density images. DTD extracts keypoints from density BEV (Bird’s Eye View) images, restores them to 3D coordinates, and connects them using triangles. In addition, to enhance the representation of local structures around each keypoint, the probabilistic distribution of neighboring points is modeled as a local descriptor. Specifically, our contributions are as follows:
  • We propose a novel Density Triangle Descriptor (DTD) based on density image features, which enables fast keypoint extraction from point cloud density maps and can be robustly applied to loop closure detection across various scenarios.
  • We introduce an entropy-based local descriptor, in which the neighborhood of each keypoint is modeled as a Gaussian distribution, and entropy is employed to quantify local uncertainty, thereby enriching the expressiveness of DTD.
  • We develop a two-stage loop closure detection method for DTD, consisting of candidate search and geometric verification, where an entropy-based similarity function is further incorporated to improve the real-time performance and robustness of the method.
  • Extensive experiments across multiple scenarios demonstrate the effectiveness of our method, with results compared to state-of-the-art approaches.

2. Related Work

Current loop closure detection approaches can be broadly categorized into Vision-based and LiDAR-based methods. The method proposed in this work integrates both visual-inspired cues and LiDAR texture information. In the following, we provide a review of the related literature.

2.1. Vision-Based Loop Closure Detection

Vision-based LCD has been widely applied in various SLAM systems, with Bag-of-Words (BoW) models being among the most commonly used due to their efficiency and generalization capability. These approaches typically rely on retrieving 2D image features, where classical descriptors such as SIFT [15], SURF [16], and ORB [17] offer invariance to rotation and translation. For instance, DBoW2 [4] builds binary descriptors based on BRIEF [18], significantly improving the efficiency of image retrieval. FAB-MAP [19] further extends the BoW model by introducing a probabilistic generative framework.
However, these methods mainly exploit the appearance features of 2D images and are prone to false positives in the presence of local structural repetition. To mitigate this, our method not only utilizes the keypoints extracted from LiDAR-based density maps but also preserves their 3D spatial information, thereby reducing the impact of locally similar structures.

2.2. LiDAR-Based Loop Closure Detection

In recent years, LiDAR-based LCD has received increasing attention, and existing methods can be broadly classified into handcrafted feature-based and learning-based approaches.
Handcrafted feature-based methods often utilize projection strategies to generate low-dimensional descriptors. A representative example is M2DP [7], which projects point clouds onto multiple 2D planes to generate density signatures, and then applies singular value decomposition (SVD) to derive global descriptors. Another widely adopted method is Scan Context (SC) [8], which projects point clouds into BEV space under a polar coordinate system and encodes each bin using the maximum height, thereby achieving rotation invariance. However, SC lacks translation invariance. SC++ [20] improves upon SC by enhancing robustness to lateral shifts and rotational movements while also increasing search efficiency. Several SC variants have also been proposed to encode point clouds using alternative feature cues [21,22,23]. NDD [24] employs a probabilistic 2D representation to better capture the density distribution of point clouds. BoW3D [13] constructs a bag-of-words representation using Link3D features, enabling efficient place recognition and real-time 6-DoF loop pose correction. More recently, STD [11] has been introduced by Yuan et al., which extracts keypoints and constructs triangle-based descriptors with both rotation and translation invariance. However, since STD relies on planar structures for keypoint extraction, it suffers from reduced efficiency and limited discriminability in low-texture environments.
Learning-based methods have also emerged with the advancement of deep learning. PointNetVLAD [25] combines PointNet [26] and NetVLAD [27] to extract global point cloud features with high matching accuracy. OverLapNet [28] proposes a lightweight architecture to extract multi-scale features and estimates the relative yaw angle and overlap between scans. SSC [29] leverages semantic information to encode 3D scenes and incorporates a two-stage global ICP strategy to improve registration performance. BEVPlace++ [30] further introduces a rotation equivariant and invariant network (REIN), achieving robust place recognition under significant rotational variations. Although learning-based methods show promising performance, they typically require large-scale annotated datasets and substantial computational resources, making them less suitable for real-time deployment on general platforms. Therefore, this work focuses on handcrafted descriptors that are training-free and computationally efficient.
To balance feature expressiveness and environmental adaptability, we propose a loop closure detection framework that integrates both visual and LiDAR information. Our method generates 3D keypoints based on density map projections and constructs global descriptor through triangle topology, enhancing structural representation and matching stability. Additionally, entropy-based local descriptor is designed to characterize the local neighborhood of each keypoint using the probabilistic distribution of point density, thereby improving local discriminability.

3. Methods

In this section, we present the overall framework of the proposed Density Triangle Descriptor (DTD), as illustrated in Figure 1. The pipeline begins with the keypoint extraction module, where the input point cloud is first preprocessed and keypoints are detected from the density-based BEV image. This is followed by the descriptor construction module, in which a descriptor is formed from three selected keypoints. The descriptor integrates both the geometric side lengths of the triangle and the probabilistic distribution characteristics of the keypoint neighborhoods. Finally, a two-stage loop closure detection algorithm is introduced, which integrates global geometric representation from triangle edge lengths with local entropy-based descriptors, ultimately yielding a 6-DoF pose estimation.

3.1. Keypoint Extraction

Bird’s Eye View (BEV) projection is widely used in place recognition because it preserves local 2D geometric structure while reducing computational cost through dimensionality reduction. Traditional BEV images built from maximum height [8] or intensity [31] are often sensitive to pose variations or external factors such as reflectivity and sensor noise. In contrast, density-based BEV representations [32] rely on point distributions within grid cells rather than raw point values, offering better robustness to viewpoint changes. Motivated by this advantage, we adopt a density image as the BEV representation in this work.
Specifically, the BEV plane is uniformly partitioned into square grid cells of equal area, upon which the density distribution is computed. The raw LiDAR point cloud is defined as follows:
P = { p 1 , p 2 , , p n   |   p i 3 }
where  p i   denotes a point in the 3D space. In typical road environments, the x, y, and z axes point toward the right, forward, and upward directions of the LiDAR sensor, respectively, with the x−y plane corresponding to the ground. Centered on the LiDAR, we define a square projection region covering  L   m ,   L   m   in the x–y plane, as illustrated in Figure 2 (point cloud projection). A keyframe is formed by accumulating a fixed number of LiDAR scans; the aggregated point cloud is then discretized into an  M × M   grid. For each grid cell, the number of points falling within it is counted as  C ( u , v )   based on the x and y coordinates of the points. These counts are subsequently normalized to obtain the density image  I ( u , v ) , defined as follows:
I ( u , ν ) = C ( u , ν ) C min C max C min | α β | + min ( α , β )   if   I ( u , ν ) I max σ k     0         if   I ( u , ν ) I max < σ k
C max = max C ( u , v ) ; C min = min C ( u , v ) ; I max = max I ( u , v ) .
where  α   and  β   control the range of normalized values. Furthermore, a proportional threshold  σ k is applied to the density image in this study, where all pixels with density values below this threshold are set to zero. This design effectively suppresses the influence of ground points and noise in the density image generation process, thereby ensuring greater consistency and reliability in the subsequent keypoint extraction.
After generating the density image, we perform feature extraction to identify salient regions. Density image typically contains rich local textures and pronounced structural variations. Since the sensor may undergo significant changes in orientation during vehicle motion, it is essential for the extracted features to exhibit rotational invariance. To this end, we adopt the Good Features to Track (GFTT) algorithm [33] to detect corner features within the density image.
The GFTT algorithm effectively captures salient characteristics in the point cloud density distribution, especially in areas with strong local structural variations. The detected 2D feature points are then projected back into 3D space by computing the mean coordinates of the raw 3D points within their corresponding grid cells (as illustrated in Figure 2). This process yields the keypoints for each keyframe, effectively combining the local structural information of the density map with the spatial properties of the point cloud, and providing a robust foundation for subsequent descriptor generation.

3.2. Density Triangle Descriptor Construction

We construct triangle descriptors using the keypoints generated in Section 3.1. Owing to the inherent translation and rotation invariance of triangles, a set of triangles formed by keypoints can effectively represent the global geometric structure of each keyframe. To ensure the uniqueness and geometric validity of the triangles, we employ a kd-tree to search for two neighboring points for each keypoint so that each triplet forms only one triangle. Additionally, the side lengths of triangles are constrained within a predefined threshold range to avoid overly sparse or dense connections.
When describing local point cloud features, many existing methods rely on simple attributes such as height, intensity, or density, which often lack robustness in complex outdoor environments. Since each vertex of a triangle corresponds to a grid cell, we model the point distribution within each cell as a Gaussian distribution  x   ~   N ( μ , Σ ) , where  μ is the mean and  Σ is the full covariance matrix. We compute its uncertainty using information entropy, the entropy vector for the three vertices is defined as  H = ( H 1 , H 2 , H 3 ) , where
H ( x ) = i = 1 n P ( x i ) log P ( x i ) = 1 2 log ( ( 2 π e ) d   | Σ | )
where  d represents the dimension of the data, which is set to 3 for a 3D point cloud. This probabilistic modeling based on Gaussian statistics effectively reduces the influence of outliers and improves local feature stability.
The proposed Density Triangle Descriptor combines density and entropy information to jointly capture both the spatial density and uncertainty of local regions. Each descriptor  D consists of the vertex coordinates  p 1 , p 2 , p 3 , side lengths  l 12 , l 23 , l 13 , the corresponding entropy values  h 1 , h 2 , h 3 , and the frame index  k :
D = ( p 1 , p 2 , p 3 , l 12 , l 23 , l 13 , h 1 , h 2 , h 3 , k )
An example of DTD is shown in Figure 3. Finally, each descriptor is indexed into a database using a hash key derived from its side lengths.

3.3. Loop Detection

In this section, we propose a two-stage loop closure detection algorithm based on the DTD, as outlined in Algorithm 1. The method consists of the following two stages:
Algorithm 1: Loop Closure Detection
Sensors 26 00201 i001
(1) Candidate Search: We first extract global and local descriptors from the current frame using the method described in Section 3.2. For each descriptor, a unique hash key is generated based on the triangle’s edge lengths. These hash keys are then used to query a preconstructed hash table to retrieve candidate descriptors from past frames that exhibit similar triangular structures.
To further refine the candidate set and identify the most relevant loop candidates, we define a local descriptor similarity function, which quantitatively evaluates the similarity between the current local descriptor  H c and a candidate loop local descriptor  H l . The similarity is defined as:
sim ( H c , H l ) = H c H l T H c   H l
It can be observed that the similarity function corresponds to the cosine similarity of the entropy vectors. If the similarity between  H c and  H l exceeds a predefined threshold  σ s , a vote is cast for the frame  F l . This voting process is applied to all descriptors in the current frame, and the top-k frames with the highest number of votes are selected as input to the subsequent geometric verification stage. This voting-based strategy effectively narrows down the set of candidate frames, significantly improving both the efficiency and accuracy of the geometric verification process.
(2) Geometric verification: After obtaining the top-k loop candidate frames, we perform geometric verification to eliminate false positives and estimate the 6-DoF relative pose transformation between the current frame and the loop frame. Specifically, for the current keyframe  F c and a candidate loop frame  F l , we identify N matched descriptor pairs, each corresponding to two point sets  P c and  P l , respectively. We first compute the centroids  p c c and  p l c of the corresponding point sets, and then construct the cross-covariance matrix:
W = i = 1 3 * N p c i p c c p l i p l c T
We apply Singular Value Decomposition (SVD) to  W :
W = U Σ V T
If  W is of full rank, the optimal rotation  R c , l and translation  t c , l between the two frames can be recovered as follows:
R c , l = V U T     t c , l = s l R c , l s c
Using this transformation, we project each point  P l to its estimated location  P e , and compute its Euclidean distance to the corresponding point  P c . A match is considered valid if
p c i ( R c , l p l i + t c , l ) < σ d
If the distance is less than  σ d , the two points are successfully matched. We count the number of valid correspondences, and only if this number  N d exceeds a specified minimum threshold  σ n , the candidate frame is finally accepted as a valid loop closure.

4. Results

4.1. Dataset and Experimental Settings

To comprehensively evaluate the loop closure detection performance of the proposed method, experiments were conducted on four publicly available datasets: KITTI [34], NCLT [35], FusionPortable [36], and HeLiPR [37], with detailed information summarized in Table 1.
The KITTI dataset provides ground truth (GT) data corresponding to each point cloud file. In our experiments, each bin file is treated as an individual point cloud frame.
For the NCLT dataset, the timestamps of point cloud data are not strictly aligned with the official GT. Therefore, we employ the FAST-LIO [38] odometry to resample the point clouds at a fixed spatial interval, generating frames at every 1 m of traveled distance.
The FusionPortable dataset contains a variety of complex environments (as shown in Figure 4). Similarly, FAST-LIO is used to process the raw data and generate the corresponding point cloud sequences for evaluation.
The above three datasets are collected using spinning LiDAR sensors. To further investigate the adaptability of our method across different LiDAR modalities, we additionally evaluate our approach on the HeLiPR dataset, which includes data acquired by two solid state LiDAR sensors. The HeLiPR dataset provides GT annotations for each point cloud, and we apply the same 1 m spatial sampling strategy to generate keyframes for experimental validation.
To validate the effectiveness of our method, we compared it against several state-of-the-art approaches. All experiments were executed on a computing platform equipped with an AMD Ryzen 5 5600 @ 3.3 GHz processor and 16 GB RAM, running Ubuntu 20.04 LTS.

4.2. Loop Detection Evaluation

In this section, we compare the proposed method with several state-of-the-art (SOTA) approaches, including Scan Context [8] and STD [11]. In our DTD implementation, every 10 consecutive frames are aggregated into a keyframe. A loop closure is considered a true positive when the ground-truth (GT) distance between the target frame and a candidate frame is less than 15 m, with at least 50 keyframes in between to avoid trivial matches.
For parameter settings, when constructing the density image, the projection area range  L is 50, the size of the density image  M is 300, and the density threshold  σ k is 0.05. The parameters used for constructing the triangle descriptor are kept consistent with those of STD, while the similarity threshold  σ s   for the local descriptor is set to 0.95. To evaluate the contribution of the local descriptor, we also compare DTD with its entropy-free variant, DTD (w/o entropy). For Scan Context and STD, we adopt their open-source implementations and followed the recommended parameter settings provided in their respective papers.
By varying the number of coincident points  N d , we generate the precision–recall (PR) curves for different sequences (as shown in the Figure 5). Furthermore, we evaluate the F1 max score and the Extended Precision (EP) value of each method (as shown in the Table 1). The F1 score is defined as follows:
F 1 = 2 × P × R P + R
where  P and  R denote precision and recall, respectively. The F1 max score provides a balanced measure between accuracy and completeness. The EP value is defined as follows:
E P = 1 2 P R 0 + R P 100
where  P R 0 represents the precision at minimum recall and  R P 100 represents the max recall at 100% precision. EP is particularly suitable for evaluating loop closure detection performance.
As shown in Figure 5 and Table 2, DTD consistently achieves superior PR curves, higher F1 max scores, and larger EP values across most sequences. Compared to STD, DTD shows average improvements of 18.30% and 20.08% in F1 max score and EP, respectively; compared to Scan Context, the improvements are 15.51% and 33.40%, respectively.
For the KITTI dataset, DTD (w/o entropy) performs less favorably on KITTI 00 and KITTI 02, as it relies solely on density-based triangles for matching. Incorporating entropy significantly improves the robustness, leading to 12.51% and 11.13% increases in the F1 max score, respectively. In the reverse-loop-only KITTI 08 sequence, Scan Context exhibits degraded performance, whereas DTD remains stable, demonstrating strong rotation invariance.
For the NCLT dataset, which contains diverse indoor–outdoor transitions, both STD and Scan Context show inconsistent results. Even without entropy, DTD (w/o entropy) remains stable, and the integration of entropy further enhances loop detection performance.
For the FusionPortable dataset, we evaluate four representative sequences, with a focus on two challenging cases: building_day and vehicle_campus01. In the building_day complex indoor environment, the keypoint selection strategy of STD becomes unstable, leading to numerous mismatches or matching failures (as is shown in Figure 6). In contrast, DTD benefits from density-based keypoint extraction, leading to a two-fold improvement in F1 max score over STD. In vehicle_campus01, similar road structures and limited height variation make Scan Context prone to false detections; specifically, Figure 5l visually demonstrates that Scan Context maintains relatively low loop closure performance, and Table 2 shows that STD improves the F1 max score by 68.04% compared to Scan Context. DTD maintains high performance owing to its robust keypoint extraction and entropy-based similarity evaluation.

4.3. Ablation Study

In Section 3.2, we introduced the entropy-based similarity function. To evaluate the contribution of entropy to the local descriptor, we conduct an ablation study to analyze its impact during the loop detection phase. Specifically, DTD (w/o entropy) refers to the variant that removes the entropy validation and performs matching solely based on triangle side lengths. As shown in Figure 5 and Table 2, incorporating entropy significantly improves the recall rate of DTD and achieves better F1 max score and EP. This improvement is consistent across all sequences, demonstrating strong robustness and stability.
Furthermore, the proposed entropy-based local descriptor effectively filters out incorrect loop candidates, accelerating the overall matching process while improving loop detection performance. To evaluate its contribution, we select representative sequences from the KITTI, NCLT, and FusionPortable datasets (KITTI 02, NCLT 28 September 2012, and vehicle_campus01) and conduct comparative experiments. Two metrics are adopted: the F1 max score and the total processing time. These metrics are used to (1) verify the performance gains brought by the introduction of entropy and (2) analyze the influence of different similarity thresholds  σ s after incorporating entropy-based verification, as shown in Figure 7.
Figure 7 shows the variations in F1 max scores and computation time under DTD (w/o entropy) and DTD with various similarity thresholds  σ s . Experimental results indicate that, after incorporating entropy verification, DTD achieves higher F1 max scores and lower time consumption than DTD (w/o entropy) even when  σ s = 0. Specifically, when  σ s = 0.95, the F1 max scores of these three sequences increased by 11.13%, 4.07%, and 44.89%, respectively, while the computation time decreased by 38.13%, 13.19%, and 40.84%. This demonstrates that although entropy computation introduces an additional cost, it can rapidly eliminate false candidates with similar density-triangle structures, thereby reducing the overall matching time.
Moreover, the choice of  σ s affects the performance of DTD. Generally, a larger threshold leads to better detection results; however, when  σ s approaches 1, correct loop candidates may be mistakenly discarded. Therefore, in practical applications, σ s is typically set within the range of 0.90–0.95 to balance accuracy and robustness.

4.4. Performance on LiDAR-Based SLAM

In this experiment, we evaluate the effectiveness of the proposed loop closure detection module within a LiDAR SLAM system. DTD, STD, and Scan Context are, respectively, integrated into FAST-LIO (denoted as DTD-FAST-LIO, STD-FAST-LIO, and SC-FAST-LIO) and tested on the NCLT and FusionPortable datasets. To quantify the impact of loop closure on trajectory accuracy, we use the EVO toolkit to compute the Absolute Pose Error (APE) and adopt Root Mean Square Error (RMSE) as the evaluation metric, where a lower RMSE indicates better global consistency. The trajectory optimization results are presented in Figure 8 and Table 3.
As shown in Table 3, compared with the original FAST-LIO system, DTD-FAST-LIO consistently yields lower RMSE values across almost all tested sequences, indicating that the proposed loop closure module effectively suppresses accumulated drift. Taking the NCLT dataset as an example, the average RMSE decreased by 51.02% after incorporating DTD. Further comparison with STD-FAST-LIO and SC-FAST-LIO demonstrates that DTD achieves the lowest RMSE on all sequences, validating the superiority of the proposed method.
Moreover, the trajectory comparison in Figure 8 visually illustrates that DTD successfully identifies loop closures and significantly improves accumulated drift. Overall, the proposed loop closure detection module can be integrated into existing LiDAR SLAM frameworks and substantially improves global pose estimation accuracy.

4.5. Computational Complexity

In this section, we evaluate the average computational time of different loop detection methods, the results are presented in Figure 9. Table 4 presents the runtime comparison of each submodule on the KITTI 00 sequence. For Scan Context, we used the official MATLAB (R2022b) implementation with default parameters, while both STD and DTD were implemented in C++.
The results show that Scan Context exhibits the highest computational cost. In contrast, the total processing time of DTD is reduced by 50.57% and 73.85% compared to STD and Scan Context, respectively, which is mainly due to the keypoint selection strategy based on the density map. Although the introduction of local descriptors slightly increases the time required to insert descriptors into the database, this overhead is negligible considering the improved robustness achieved. As illustrated in Figure 9, the overall runtime of DTD remains stable as the number of frames increases, demonstrating its clear advantage in terms of real-time performance.

4.6. Applicability to Different LiDAR Types

This section mainly evaluates the applicability of the proposed DTD method to different types of LiDAR sensors. We use two solid-state LiDARs, Avia and Aeva, provided by the HeLiPR dataset. These sensors have different Fields of View (FOVs) and employ non-repetitive scanning patterns, as detailed in Table 1.
In the experimental setup, we accumulate 10 frames as a keyframe, and a detected loop is regarded as a true positive when the GT distance is less than 15 m. The remaining algorithmic parameters are consistent with those described in Section 4.2. The corresponding PR curves and loop closure performance metrics are shown in Figure 10 and Table 5, respectively.
As can be observed from Table 5, DTD achieves higher F1 max scores and EP values than STD and Scan Context on almost all sequences. Compared with STD, the F1 max scores are improved by 24%, 116.61%, 95.03%, and 86.61%, respectively, indicating that DTD maintains strong loop closure performance under solid-state LiDAR data. In addition, DTD consistently outperforms DTD (w/o entropy) across all sequences, with F1 max scores increasing by 3.33%, 5.02%, 3.07%, and 0.79%, further verifying the effectiveness of the entropy-based local descriptor in enhancing loop closure detection performance.
To intuitively analyze the impact of different LiDAR types on DTD, we compare the keypoint extraction results on the KITTI 00, Avia KAIST04, and Aeva DCC04 datasets, as shown in Figure 11. Since KITTI 00 is collected using a LiDAR sensor with a 360° FOV, it has the richest set of keypoints, while Avia and Aeva have smaller FOVs but higher point cloud densities. The experimental results show that the density-image-based keypoint extraction strategy remains stable under different sensor conditions and effectively reflects regions with higher point cloud density.
In summary, DTD is not only applicable to traditional mechanically rotating LiDARs but can also be applied to non-repetitive scanning LiDARs, demonstrating good generality and adaptability.

5. Discussion

As shown in Figure 5 and Table 2, DTD demonstrates superior overall performance in loop closure detection compared with STD and Scan Context, achieving higher F1 max score and EP across most sequences. Further analysis reveals that DTD exhibits strong robustness across diverse scenarios: on multiple sequences from the NCLT and FusionPortable datasets, the comparison methods suffer from varying degrees of performance degradation. Specifically, Scan Context only uses the maximum height of the point cloud for projection, which easily leads to mismatches in open or scenes with insignificant height variations; STD directly extracts corner points from the 3D point cloud as keypoints, resulting in keypoints being more concentrated near the lidar, as intuitively shown in Figure 6. In contrast, inspired by previous BEV image-based feature extraction methods (such as BEVPlace [32]), DTD extracts keypoints based on density images, effectively covering longer distances and areas with richer structures, demonstrating stronger adaptability to complex environments.
In addition, Figure 5 and Table 2 further verify the effectiveness of the proposed entropy-based local descriptor. Similar to probabilistic or information-theoretic modeling methods used in previous loop detection research (such as NDD [24]), entropy can effectively characterize the distribution characteristics of point clouds. By incorporating entropy into DTD, more reliable similarity evaluation for candidate loop closures can be achieved. The ablation study results presented in Figure 7 demonstrate that this local verification strategy not only improves the F1 and EP metrics, but also significantly reduces the computational cost of loop detection by filtering out incorrect candidate frames early.
Furthermore, DTD is integrated into a real-world LiDAR odometry system, FAST-LIO. As shown in Figure 8 and Table 3, the introduction of loop closure detection effectively reduces accumulated drift and improves global trajectory consistency. Although STD and Scan Context can also improve trajectory quality to some extent, the results in Table 3 show that DTD still has a more significant advantage in the RMSE metric, which is consistent with the conclusions drawn from the loop closure detection experiments.
Figure 9 and Table 4 highlight the real-time performance advantages of DTD. Statistical analysis of the runtime of the submodules shows that compared to STD, the efficiency improvement of DTD mainly comes from its descriptor extraction strategy and two-stage loop detection method. Keypoint extraction based on density images is more efficient than directly operating on 3D point clouds, while the entropy-based local descriptor reduces unnecessary matching computations, thereby improving overall system efficiency.
Finally, we evaluate the applicability of DTD under other types of LiDAR. The results in Figure 10 and Table 5 show that DTD consistently outperforms the comparison methods under different LiDAR types and scanning modes. This is mainly due to the keypoint extraction strategy employed: as shown in Figure 11, even when the sensor type and acquisition method change, the keypoints extracted by DTD can still reliably reflect the high-density structural regions in the point cloud.

6. Conclusions

This paper proposes a novel global descriptor, the Density Triangle Descriptor (DTD), for LiDAR loop closure detection in complex environments. DTD consists of keypoint extraction based on density images, descriptor construction, and a two-stage loop closure detection strategy. By extracting keypoints from density BEV images, this method outperforms direct manipulation of 3D point clouds in terms of efficiency and stability. Simultaneously, an entropy-based local descriptor is introduced to enhance robustness to local structural changes, and geometric constraints are combined to achieve reliable loop closure verification and pose estimation.
The method is systematically validated on four public datasets: KITTI, NCLT, FusionPortable, and HeLiPR, covering various complex scenarios and different types of LiDAR sensors. Compared to STD and Scan Context methods, DTD demonstrates significant advantages in both loop closure detection accuracy and computational efficiency. Experimental results show that compared to STD, DTD improves the average F1 max score and EP by 18.30% and 20.08%, respectively, while increasing computational efficiency by 50.57%. Furthermore, integrating DTD into the FAST-LIO odometry system effectively suppresses long-term accumulated errors, reducing the trajectory RMSE by an average of 51.02%. These results demonstrate that DTD possesses good stability, robustness, and environmental adaptability in complex environments, validating its potential for application in practical SLAM systems.
In future work, we will further expand the descriptor’s expressive power and application scope. We will consider introducing richer keypoint description information (such as semantic features) to improve the ability to characterize complex scene structures. In addition, we will explore multi-sensor fusion strategies, introducing visual information in LiDAR degradation scenarios to provide additional constraints to the system, further improving the system’s robustness and applicability.

Author Contributions

Conceptualization, Q.W.; methodology, K.T.; software, K.T.; validation, K.T.; formal analysis, K.T.; investigation, K.T.; resources, K.T.; data curation, K.T.; writing—original draft preparation, K.T.; writing—review and editing, C.Y., Y.S. and S.L.; visualization, K.T.; supervision, Q.W. and C.Y.; project administration, Q.W. and C.Y.; funding acquisition, Q.W. and C.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 42374029) and the Natural Science Foundation of the Jiangsu Higher Education Institutions of China (No. 24KJB420001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: KITTI dataset: https://www.cvlibs.net/datasets/kitti/eval_odometry.php (accessed on 1 November 2025); NCLT dataset: http://robots.engin.umich.edu/nclt/ (accessed on 1 November 2025). FusionPortable Dataset: https://fusionportable.github.io/ (accessed on 1 November 2025). HeLiPR Dataset: https://sites.google.com/view/heliprdataset/download (accessed on 1 November 2025). EVO tool: https://github.com/MichaelGrupp/evo (accessed on 1 November 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Ratti, C.; Rus, D. LIO-SAM: Tightly-coupled Lidar Inertial Odometry via Smoothing and Mapping. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 5135–5142. [Google Scholar]
  2. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  3. Tong, B.; Jiang, F.; Lu, B.; Gu, Z.; Li, Y.; Wang, S. Three-Dimensional Environment Mapping with a Rotary-Driven Lidar in Real Time. Sensors 2025, 25, 4870. [Google Scholar] [CrossRef] [PubMed]
  4. Galvez-López, D.; Tardos, J.D. Bags of Binary Words for Fast Place Recognition in Image Sequences. IEEE Trans. Robot. 2012, 28, 1188–1197. [Google Scholar] [CrossRef]
  5. Pivoňka, T.; Přeučil, L. On Model-Free Re-Ranking for Visual Place Recognition with Deep Learned Local Features. IEEE Trans. Intell. Veh. 2024, 9, 7900–7911. [Google Scholar] [CrossRef]
  6. Zhang, X.; Cui, Y.; Ren, Y.; Duan, G.; Zhang, H. Loop Detection Method Based on Neural Radiance Field BoW Model for Visual Inertial Navigation of UAVs. Remote Sens. 2024, 16, 3038. [Google Scholar] [CrossRef]
  7. He, L.; Wang, X.; Zhang, H. M2DP: A Novel 3D Point Cloud Descriptor and Its Application in Loop Closure Detection. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 231–237. [Google Scholar]
  8. Kim, G.; Kim, A. Scan Context: Egocentric Spatial Descriptor for Place Recognition Within 3D Point Cloud Map. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4802–4809. [Google Scholar]
  9. Gupta, S.; Guadagnino, T.; Mersch, B.; Vizzo, I.; Stachniss, C. Effectively Detecting Loop Closures using Point Cloud Density Maps. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 10260–10266. [Google Scholar]
  10. Park, Y.; Pak, G.; Kim, E. RE-TRIP: Reflectivity Instance Augmented Triangle Descriptor for 3D Place Recognition. In Proceedings of the 2025 IEEE International Conference on Robotics and Automation (ICRA), Atlanta, GA, USA, 19–23 May 2025; pp. 2226–2232. [Google Scholar]
  11. Yuan, C.; Lin, J.; Zou, Z.; Hong, X.; Zhang, F. STD: Stable Triangle Descriptor for 3D place recognition. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 1897–1903. [Google Scholar]
  12. Yuan, C.; Lin, J.; Liu, Z.; Wei, H.; Hong, X.; Zhang, F. BTC: A Binary and Triangle Combined Descriptor for 3-D Place Recognition. IEEE Trans. Robot. 2024, 40, 1580–1599. [Google Scholar] [CrossRef]
  13. Cui, Y.; Chen, X.; Zhang, Y.; Dong, J.; Wu, Q.; Zhu, F. BoW3D: Bag of Words for Real-Time Loop Closing in 3D LiDAR SLAM. IEEE Robot. Autom. Lett. 2023, 8, 2828–2835. [Google Scholar] [CrossRef]
  14. Xu, X.; Lu, S.; Wu, J.; Lu, H.; Zhu, Q.; Liao, Y.; Xiong, R.; Wang, Y. RING++: Roto-Translation Invariant Gram for Global Localization on a Sparse Scan Map. IEEE Trans. Robot. 2023, 39, 4616–4635. [Google Scholar] [CrossRef]
  15. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  16. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded up robust features. In Proceedings of the Computer Vision-ECCV 2006, Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
  17. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  18. Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P. BRIEF: Binary Robust Independent Elementary Features. In Proceedings of the Computer Vision-ECCV 2010, Heraklion, Greece, 5–11 September 2010; pp. 778–792. [Google Scholar]
  19. Paul, R.; Newman, P. FAB-MAP 3D: Topological mapping with spatial and visual appearance. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation (ICRA), Anchorage, AK, USA, 3–7 May 2010; pp. 2649–2656. [Google Scholar]
  20. Kim, G.; Choi, S.; Kim, A. Scan Context++: Structural Place Recognition Robust to Rotation and Lateral Variations in Urban Environments. IEEE Trans. Robot. 2022, 38, 1856–1874. [Google Scholar] [CrossRef]
  21. Fan, Y.; Du, X.; Luo, L.; Shen, J. FreSCo: Frequency-Domain Scan Context for LiDAR-based Place Recognition with Translation and Rotation Invariance. In Proceedings of the 2022 17th International Conference on Control, Automation, Robotics and Vision (ICARCV), Singapore, 11–13 December 2022; pp. 576–583. [Google Scholar]
  22. Wang, H.; Wang, C.; Xie, L. Intensity Scan Context: Coding Intensity and Geometry Relations for Loop Closure Detection. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 2095–2101. [Google Scholar]
  23. Tian, X.; Yi, P.; Zhang, F.; Lei, J.; Hong, Y. STV-SC: Segmentation and Temporal Verification Enhanced Scan Context for Place Recognition in Unstructured Environment. Sensors 2022, 22, 8604. [Google Scholar] [CrossRef] [PubMed]
  24. Zhou, R.; He, L.; Zhang, H.; Lin, X.; Guan, Y. NDD: A 3D Point Cloud Descriptor Based on Normal Distribution for Loop Closure Detection. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 1328–1335. [Google Scholar]
  25. Uy, M.A.; Lee, G.H. PointNetVLAD: Deep Point Cloud Based Retrieval for Large-Scale Place Recognition. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4470–4479. [Google Scholar]
  26. Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  27. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN Architecture for Weakly Supervised Place Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5297–5307. [Google Scholar]
  28. Chen, X.Y.L.; Läbe, T.; Milioto, A.; Röhling, T.; Vysotska, O.; Haag, A.; Behley, J.; Stachniss, C. OverlapNet: Loop Closing for LiDAR-based SLAM. arXiv 2021, arXiv:2105.11344. [Google Scholar]
  29. Li, L.; Kong, X.; Zhao, X.; Huang, T.; Li, W.; Wen, F.; Zhang, H.; Liu, Y. SSC: Semantic Scan Context for Large-Scale Place Recognition. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 2092–2099. [Google Scholar]
  30. Luo, L.; Cao, S.-Y.; Li, X.; Xu, J.; Ai, R.; Yu, Z.; Chen, X. BEVPlace++: Fast, Robust, and Lightweight LiDAR Global Localization for Autonomous Ground Vehicles. IEEE Trans. Robot. 2025, 41, 4479–4498. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Tian, Y.; Yang, G.; Li, Z.; Luo, M.; Li, E.; Jing, F. Intensity Triangle Descriptor Constructed from High-Resolution Spinning LiDAR Intensity Image for Loop Closure Detection. IEEE Robot. Autom. Lett. 2024, 9, 8937–8944. [Google Scholar] [CrossRef]
  32. Luo, L.; Zheng, S.; Li, Y.; Fan, Y.; Yu, B.; Cao, S.-Y.; Li, J.; Shen, H.-L. BEVPlace: Learning LiDAR-based Place Recognition using Bird’s Eye View Images. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 8666–8675. [Google Scholar]
  33. Jianbo, S.; Tomasi. Good features to track. In Proceedings of the 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
  34. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  35. Carlevaris-Bianco, N.; Ushani, A.K.; Eustice, R.M. University of Michigan North Campus long-term vision and lidar dataset. Int. J. Robot. Res. 2015, 35, 1023–1035. [Google Scholar] [CrossRef]
  36. Jiao, J.; Wei, H.; Hu, T.; Hu, X.; Zhu, Y.; He, Z.; Wu, J.; Yu, J.; Xie, X.; Huang, H.; et al. FusionPortable: A Multi-Sensor Campus-Scene Dataset for Evaluation of Localization and Mapping Accuracy on Diverse Platforms. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 3851–3856. [Google Scholar]
  37. Jung, M.; Yang, W.; Lee, D.; Gil, H.; Kim, G.; Kim, A. HeLiPR: Heterogeneous LiDAR Dataset for inter-LiDAR Place Recognition under Spatiotemporal Variations. arXiv 2023, arXiv:2309.14590. [Google Scholar] [CrossRef]
  38. Xu, W.; Cai, Y.; He, D.; Lin, J.; Zhang, F. FAST-LIO2: Fast Direct LiDAR-Inertial Odometry. IEEE Trans. Robot. 2022, 38, 2053–2073. [Google Scholar] [CrossRef]
Figure 1. The pipeline of the DTD framework. The approach extracts 2D GFTT keypoints from the density BEV image, constructs triangular structures in 3D space to encode global geometry, and employs entropy to describe local distributions. During loop closure detection, candidate frames are retrieved via triangle-edge hash search and entropy similarity, followed by geometric verification to confirm true loops and estimate relative poses.
Figure 1. The pipeline of the DTD framework. The approach extracts 2D GFTT keypoints from the density BEV image, constructs triangular structures in 3D space to encode global geometry, and employs entropy to describe local distributions. During loop closure detection, candidate frames are retrieved via triangle-edge hash search and entropy similarity, followed by geometric verification to confirm true loops and estimate relative poses.
Sensors 26 00201 g001
Figure 2. Workflow of keypoint extraction. The input point cloud is first preprocessed and projected into a density image, from which GFTT keypoints are detected. These keypoints are then back-projected to obtain their corresponding 3D positions.
Figure 2. Workflow of keypoint extraction. The input point cloud is first preprocessed and projected into a density image, from which GFTT keypoints are detected. These keypoints are then back-projected to obtain their corresponding 3D positions.
Sensors 26 00201 g002
Figure 3. Illustration of the proposed DTD. A triangle is formed by three geometrically valid keypoints to capture global structural features, while entropy of the local neighborhood encodes point-wise local characteristics.
Figure 3. Illustration of the proposed DTD. A triangle is formed by three geometrically valid keypoints to capture global structural features, while entropy of the local neighborhood encodes point-wise local characteristics.
Sensors 26 00201 g003
Figure 4. Representative sequences from FusionPortable: (a) ugv_transition00, (b) ugv_parking01, (c) building_day, (d) vehicle_campus01. The images above and below are captured by a camera and LiDAR, respectively. The dataset includes indoor–outdoor transitions, open outdoor areas, complex building interiors, and visually similar road segments, which are prone to mismatches.
Figure 4. Representative sequences from FusionPortable: (a) ugv_transition00, (b) ugv_parking01, (c) building_day, (d) vehicle_campus01. The images above and below are captured by a camera and LiDAR, respectively. The dataset includes indoor–outdoor transitions, open outdoor areas, complex building interiors, and visually similar road segments, which are prone to mismatches.
Sensors 26 00201 g004
Figure 5. Precision–recall curves for KITTI (ad), NCLT (eh), and FusionPortable (il) datasets.
Figure 5. Precision–recall curves for KITTI (ad), NCLT (eh), and FusionPortable (il) datasets.
Sensors 26 00201 g005aSensors 26 00201 g005b
Figure 6. A representative comparison of the matching performance between STD and DTD in the building_day sequence. (a) Failed matching with the STD method; (b) successful matching with the DTD method. The red dots denote the keypoints, while the white and green lines represent the successfully matched descriptors from the current frame and the loop frame, respectively.
Figure 6. A representative comparison of the matching performance between STD and DTD in the building_day sequence. (a) Failed matching with the STD method; (b) successful matching with the DTD method. The red dots denote the keypoints, while the white and green lines represent the successfully matched descriptors from the current frame and the loop frame, respectively.
Sensors 26 00201 g006
Figure 7. F1 max score (a) and average computation time (b) of DTD on the KITTI 02, NCLT 28 September 2012, and vehicle_campus01 sequences under different similarity thresholds  σ s .
Figure 7. F1 max score (a) and average computation time (b) of DTD on the KITTI 02, NCLT 28 September 2012, and vehicle_campus01 sequences under different similarity thresholds  σ s .
Sensors 26 00201 g007
Figure 8. Example on NCLT 5 April 2013: (a,b) comparison of estimated and ground-truth trajectories for FAST-LIO and DTD-FAST-LIO; (c,d) Absolute Pose Error (APE) profiles of FAST-LIO (c) and DTD-FAST-LIO (d).
Figure 8. Example on NCLT 5 April 2013: (a,b) comparison of estimated and ground-truth trajectories for FAST-LIO and DTD-FAST-LIO; (c,d) Absolute Pose Error (APE) profiles of FAST-LIO (c) and DTD-FAST-LIO (d).
Sensors 26 00201 g008aSensors 26 00201 g008b
Figure 9. Runtime Evaluation on KITTI 00.
Figure 9. Runtime Evaluation on KITTI 00.
Sensors 26 00201 g009
Figure 10. Precision–recall curves for HeLiPR dataset: (a,b) captured by Livox Avia LiDAR; (c,d) captured by Aeva Aeries II LiDAR.
Figure 10. Precision–recall curves for HeLiPR dataset: (a,b) captured by Livox Avia LiDAR; (c,d) captured by Aeva Aeries II LiDAR.
Sensors 26 00201 g010
Figure 11. Comparison of DTD keypoint extraction results under different LiDAR types, the red dots denote the keypoints.
Figure 11. Comparison of DTD keypoint extraction results under different LiDAR types, the red dots denote the keypoints.
Sensors 26 00201 g011
Table 1. The details of the datasets used.
Table 1. The details of the datasets used.
DatasetKITTINCLTFusionPortableHeLiPR
SensorSpinningSpinningSpinningSolid state
Scanning PatternRepetitiveRepetitiveRepetitiveNon-Repetitive
LiDARVelodyne HDL-64EVelodyne HDL-32EOuster OS1-128Livox AviaAeva Aeries II
Field of View360° × 26.8°360° × 41.3°360° × 45°70° × 77°120° × 19.2°
Sequence0026 May 2012ugv_transition00KAIST04
Bridge01
DCC04
Riverside04
0220 August 2012ugv_parking01
0528 September 2012building_day
085 April 2013vehicle_campus01
EnvironmentUrbanCampusUnstructured, IndoorCampus, Bridge
Table 2. The F1 max scores and extended precision on datasets.
Table 2. The F1 max scores and extended precision on datasets.
SequenceDTDDTD (w/o Entropy)STDScan Context
F1 MaxEPF1 MaxEPF1 MaxEPF1 MaxEP
KITTI 000.9620.9470.8550.8740.9850.9850.9620.963
KITTI 020.8890.8970.8000.8330.8150.8130.8100.450
KITTI 050.9760.9760.9580.9600.9640.9640.9480.939
KITTI 080.9890.9890.9780.9780.9390.8140.7610.563
NCLT 26 May 20120.8790.8500.8500.8500.8400.6120.7660.503
NCLT 20 August 20120.8360.8450.8160.8390.7290.5950.7330.501
NCLT 28 September 20120.8960.8840.8610.8630.6180.5620.7710.517
NCLT 5 April 20130.9120.9190.8360.8550.6280.5320.7090.536
ugv_transition000.9230.9290.9090.9160.8840.6430.7440.752
ugv_parking010.9080.8330.8790.8000.5690.5670.6490.689
building_day0.6740.6810.6330.6380.2220.5210.6640.611
vehicle_campus010.6520.6430.4500.5890.5260.6790.3880
Note: The best results are marked in bold and the second-best results are underlined.
Table 3. The APE (meters) of FAST-LIO with different loop closure configurations on various datasets.
Table 3. The APE (meters) of FAST-LIO with different loop closure configurations on various datasets.
SequenceFAST-LIODTD-FAST-LIOSTD-FAST-LIOSC-FAST-LIO
RMSEMeanRMSEMeanRMSEMeanRMSEMean
NCLT 26 May 20122.3822.0661.5741.3461.6081.3861.6351.402
NCLT 20 August 20122.2882.0641.5981.3991.6751.4501.6181.387
NCLT 28 September 20122.9302.2311.0570.9231.1881.0301.9311.408
NCLT 5 April 20136.4745.2721.5501.2732.9782.5293.3222.816
ugv_transition000.2010.1470.2010.1470.2030.1490.2010.147
ugv_parking010.3820.3320.3590.3120.3860.3310.3630.318
building_day1.5601.3331.5141.2411.5311.2581.5451.308
vehicle_campus015.0493.9715.0473.9685.0503.9715.0533.973
Note: The best results are marked in bold.
Table 4. Average computational time of different submodules on KITTI 00.
Table 4. Average computational time of different submodules on KITTI 00.
MethodTime Consumption (ms)
Descriptor ConstructionLoop DetectionAdd to DatabaseTotal
Scan Context47.06330.2890.01777.517
STD27.24913.6610.09941.009
DTD13.4986.5730.19820.269
Note: The best results are marked in bold.
Table 5. The F1 max scores and extended precision on HeLiPR dataset.
Table 5. The F1 max scores and extended precision on HeLiPR dataset.
SequenceDTDDTD (w/o Entropy)STDScan Context
F1 MaxEPF1 MaxEPF1 MaxEPF1 MaxEP
Avia KAIST040.6200.7240.6000.7140.5000.6090.4190.593
Avia Bridge010.6910.5080.6580.5060.3190.5360.6370.500
Aeva DCC040.7060.7510.6850.7430.3620.5080.5700.501
Aeva Riverside040.8920.9010.8850.8910.4780.5060.4520.500
Note: The best results are marked in bold and the second-best results are underlined.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, K.; Wang, Q.; Yan, C.; Sun, Y.; Liu, S. DTD: Density Triangle Descriptor for 3D LiDAR Loop Closure Detection. Sensors 2026, 26, 201. https://doi.org/10.3390/s26010201

AMA Style

Tang K, Wang Q, Yan C, Sun Y, Liu S. DTD: Density Triangle Descriptor for 3D LiDAR Loop Closure Detection. Sensors. 2026; 26(1):201. https://doi.org/10.3390/s26010201

Chicago/Turabian Style

Tang, Kaiwei, Qing Wang, Chao Yan, Yang Sun, and Shengyi Liu. 2026. "DTD: Density Triangle Descriptor for 3D LiDAR Loop Closure Detection" Sensors 26, no. 1: 201. https://doi.org/10.3390/s26010201

APA Style

Tang, K., Wang, Q., Yan, C., Sun, Y., & Liu, S. (2026). DTD: Density Triangle Descriptor for 3D LiDAR Loop Closure Detection. Sensors, 26(1), 201. https://doi.org/10.3390/s26010201

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop