You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

26 November 2025

Cylindrical Scan Context: A Multi-Channel Descriptor for Vertical-Structure-Aware LiDAR Localization

,
,
,
,
,
and
Korea Institute of Robotics & Technology Convergence, Pohang 37666, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Section Navigation and Positioning

Abstract

This study introduces Cylindrical Scan Context (CSC), a novel LiDAR descriptor designed to improve robustness and efficiency in GPS-denied or degraded outdoor environments. Unlike the conventional Scan Context (SC), which relies on azimuth–range projection, CSC employs an azimuth–height representation that preserves vertical structural information and incorporates multiple physical channels—range, point density, and reflectance intensity—to capture both geometric and radiometric characteristics of the environment. This multi-channel cylindrical formulation enhances descriptor distinctiveness and robustness against viewpoint, elevation, and trajectory variations. To validate the effectiveness of CSC, real-world experiments were conducted using both self-collected coastal–forest datasets and the public MulRan–KAIST dataset. Mapping was performed using LIO-SAM with LiDAR, IMU, and GPS measurements, after which LiDAR-only localization was evaluated independently. A total of approximately 700 query scenes (1 m ground-truth threshold) were used in the self-collected experiments, and about 1200 scenes (3 m threshold) were evaluated in the MulRan–KAIST experiments. Comparative analyses between SC and CSC were performed using Precision–Recall (PR) curves, Detection Recall (DR) curves, Root Mean Square Error (RMSE), and Top-K retrieval accuracy. The results show that CSC consistently yields lower RMSE—particularly in the vertical and lateral directions—and demonstrates faster recall growth and higher stability in global retrieval. Across datasets, CSC maintains superior DR performance in high-confidence regions and achieves up to 45% reduction in distance RMSE in large-scale campus environments. These findings confirm that the cylindrical multi-channel formulation of CSC significantly improves geometric consistency and localization reliability, offering a practical and robust LiDAR-only localization framework for challenging unstructured outdoor environments.

1. Introduction

Accurate and reliable localization is a critical component in autonomous driving, robotic navigation, and maritime or port mobility systems. Although satellite-based navigation systems such as the Global Positioning System (GPS) provide dependable global positioning in open-sky environments, their performance can degrade severely in unstructured or cluttered outdoor areas—such as coastal regions, mountainous terrain, or dense forests—due to multipath distortion, signal occlusion, or intentional interference such as jamming and spoofing. These limitations have motivated increased interest in LiDAR-based localization, which leverages the rich geometric structure of the environment and remains functional even under GPS-denied or degraded conditions.
LiDAR sensors provide dense three-dimensional measurements that enable robust pose estimation in challenging outdoor environments [,]. Traditional registration-based methods, including Iterative Closest Point (ICP) and Normal Distribution Transform (NDT), estimate relative motion by aligning consecutive point clouds. While such approaches can yield high accuracy, their computational demand and sensitivity to initialization make them unsuitable for large-scale, real-time applications. To address these challenges, global descriptor–based localization methods have been proposed, transforming raw point clouds into compact and distinctive representations for fast place recognition. One notable descriptor is Scan Context (SC) [], which discretizes a LiDAR scan into a ring–sector grid on an azimuth–range plane for efficient similarity computation.
Despite its advantages, SC encodes the vertical (z-axis) structure of each grid cell using only a single maximum height value ( z max ), inevitably discarding multi-layered height information. Therefore, in environments characterized by complex vertical geometry—such as forested areas with overlapping canopy layers, coastal cliffs with large elevation gradients, or hilly terrain—SC may fail to capture subtle geometric distinctions, reducing its discriminability and robustness to viewpoint changes.
To overcome these limitations, this study introduces the Cylindrical Scan Context (CSC), a novel LiDAR descriptor that represents each scan in an azimuth–height domain rather than the traditional azimuth–range plane. CSC discretizes the space along the vertical axis, preserving multi-level height variations and capturing richer geometric structures. In addition, CSC integrates multiple physical channels—including maximum range, point density, and reflectance intensity—to jointly encode geometric and radiometric information []. This multi-channel cylindrical formulation enhances distinctiveness, improves robustness against viewpoint and trajectory variations, and aligns with recent trends toward multi-modal and semantically aware LiDAR localization [,]. Beyond descriptor design, this work also addresses a key operational challenge in LiDAR-only localization: how to convert global descriptor retrieval into an accurate final pose estimate. CSC provides coarse global place candidates that are robust even under large viewpoint differences; however, descriptors alone cannot yield a full 6-DoF pose. Therefore, we adopt a two-stage localization framework in which (1) CSC performs global retrieval to propose high-confidence pose candidates, and (2) ICP refines these candidates locally by aligning the query scan to a submap. In this pipeline, CSC handles global search and viewpoint invariance, while ICP provides fine-grained geometric alignment, enabling the system to produce accurate and consistent localization results even in complex environments.
The main contributions of this work are summarized as follows:
  • We construct a real-world dataset in a mixed coastal–forest environment and empirically validate the responsiveness and reliability of LiDAR-only localization using a pre-built SLAM map.
  • We develop a two-stage localization framework consisting of (i) CSC-based global retrieval, which generates coarse pose hypotheses, and (ii) ICP-based local pose refinement, which accurately aligns the query scan with the local map.
  • We present comprehensive comparisons between SC and CSC under identical conditions and quantitatively evaluate their performance using Precision–Recall (PR), Detection Recall (DR), Root Mean Square Error (RMSE), and Top-K retrieval accuracy metrics.
It is important to clarify that the focus of this work is not on improving SLAM-based map construction, but rather on evaluating operational performance of LiDAR-only localization when revisiting a pre-built map under real-world conditions. To this end, mapping and localization were performed in separate sessions, demonstrating the robustness of CSC under temporal variations, viewpoint changes, and heterogeneous traversal trajectories.
The remainder of this paper is organized as follows: Section 2 reviews related works on LiDAR-based localization; Section 3 describes the proposed CSC methodology; Section 4 presents experimental results and quantitative analyses; and Section 5 concludes the paper with discussions on scalability and future research directions.

3. Methodology

In this section, the algorithmic architecture and operational procedure of the proposed Cylindrical Scan Context (CSC)-based LiDAR-only localization framework are presented. First, the conceptual design and descriptor generation process of CSC are defined, followed by a step-by-step description of the localization pipeline. Finally, computational efficiency and implementation considerations are briefly discussed.

3.1. Concept and Descriptor Generation

The conventional Scan Context (SC) projects a point cloud onto a polar coordinate system, representing it as a two-dimensional matrix on the azimuth–range plane, where each cell records the maximum height value ( z max ) within its region. This approach achieves high computational efficiency and partial invariance to yaw rotation, but it fails to sufficiently capture structural information along the vertical direction or physical properties such as reflection intensity.
To overcome these limitations, this study extends the concept of SC and proposes the Cylindrical Scan Context (CSC), which employs an azimuth–height representation instead of the traditional azimuth–range projection. CSC converts the point cloud into cylindrical coordinates ( r , θ , z ) and records multiple statistical channels—including maximum range, point density, and intensity—on a two-dimensional grid with azimuth θ and height z axes. Through this design, CSC encodes both the geometric and radiometric characteristics of the environment simultaneously. Moreover, by discretizing along the height (z) axis, each bin in CSC maintains a constant spatial area independent of radial distance (r), thus providing a uniform spatial resolution across all ranges. This alleviates the over-expansion of outer ring bins in SC, which often causes the loss of fine structural details in distant regions.
Each channel captures distinct physical attributes: the range channel (r) quantifies spatial distribution relative to the sensor origin; the density channel expresses the concentration of points within a given region, allowing for effective discrimination of vertical structures or obstacles; and the intensity channel reflects surface reflectivity, enabling distinction between objects with similar geometry but different material or roughness properties. By integrating these channels, CSC provides a richer and more discriminative structural representation compared to the single-height encoding of SC.
Since CSC constructs its descriptor on the azimuth–height plane, it inherently preserves partial rotational robustness to yaw variation. In other words, even when the sensor’s viewing direction changes, the cylindrical projection ensures that similar descriptors are generated for the same environment. This design allows CSC to retain vertical structural information (e.g., trees, slopes, cliffs) while maintaining intensity and density distributions, resulting in more stable place recognition performance in unstructured natural environments.
Figure 1. conceptually compares the structural difference between SC and CSC. While SC consists of a single-channel representation, CSC incorporates multiple channels on the azimuth–height plane, providing a richer and more expressive environmental description.
Figure 1. Comparison between SC and CSC. CSC preserves vertical structural information by encoding multiple channels—maximum range, point density, and intensity on an azimuth–height grid.

3.2. Localization Algorithm Architecture

The proposed CSC-based LiDAR-only localization algorithm consists of five main stages, as illustrated in Figure 2. (a) Coordinate transformation and quantization convert the input point cloud into cylindrical coordinates. (b) CSC generation constructs a multi-channel descriptor. (c) Global retrieval searches for candidate locations in the pre-built database. (d) Candidate ICP refinement performs precise pose alignment. Finally, (e) the best-matching pose is selected as the LiDAR-only global localization result.
Figure 2. CSC-based localization pipeline. After coordinate transformation and CSC generation, the system determines the final pose through global retrieval and ICP refinement.
(a)
Coordinate Transformation and Quantization
Each point p = ( x , y , z ) is transformed into the cylindrical coordinate system:
r = x 2 + y 2 , θ = atan 2 ( y , x ) , z = z
The azimuth θ [ 0 , 2 π ) and height z [ z min , z max ] are uniformly quantized into N θ and N z bins, respectively. Each point is mapped to its corresponding indices as follows:
b θ = θ 2 π N θ , b z = z z min z max z min N z
(b)
Multi-Channel Feature Computation
For each bin ( b θ , b z ) , the proposed CSC descriptor extracts three statistical features from the corresponding set of points P b θ , b z . Specifically, the feature vector is defined as
a b θ , b z = [ r max , dens , I ¯ ] R 3 ,
where r max is the maximum range within the bin, dens denotes the point density, and I ¯ represents the mean intensity. These three channels provide a compact yet effective representation of geometric and radiometric characteristics: the maximum range encodes the structural extent of objects, density reflects the occupancy level of local regions, and intensity provides additional radiometric contrast.
The complete CSC descriptor is stored as a 3D tensor:
D R N z × N θ × 3 .
For efficient global retrieval, the channel dimension and height bins are flattened into a matrix representation
D ^ R N h × N θ , N h = 3 · N z ,
which allows cosine-distance comparison with circular column shifts.
(c)
Global Candidate Retrieval
The query point cloud P q is encoded into a descriptor D ^ q , which is compared against all database entries DB = { ( D ^ k , T k ) } k = 1 N . The similarity is evaluated using cosine distance with circular column-wise shifts:
d ( D ^ q , D ^ k ) = min Δ θ { 0 , , N θ 1 } 1 D ^ q , . shift Δ θ ( D ^ k ) D ^ q · D ^ k
where shift Δ θ ( · ) denotes a circular shift along the azimuth direction. The optimal shift Δ θ * corresponding to the minimum distance d is converted to a yaw correction term:
Δ ψ = Δ θ * · 2 π N θ
Based on the computed distances, the Top-K candidate indices C = { k 1 , , k K } are selected. For each candidate, the initial pose is computed as
T k ( 0 ) = T k · R z ( Δ ψ k ) , k C
(d)
Candidate ICP Refinement
For the selected Top-K candidates, a local map M k is built around each candidate position, typically within a radius of r local = 20 –30 m. Starting from the initial pose T k ( 0 ) , point-to-point ICP registration is performed:
T k = arg min T S E ( 3 ) ( u , v ) T p u q v 2
The alignment iterates until convergence, defined by Δ T < ϵ icp or a maximum iteration limit N max . Each candidate’s alignment quality is measured by a fitness score J k , and the candidate with the smallest J k proceeds to the final selection stage.
(e)
Final Pose Determination
Among all candidates, the result with the minimum fitness score below a threshold τ icp is selected as the final pose:
( k * , T * ) = arg min k J k , s . t . J k < τ icp
The final transformation T * represents the LiDAR-only global localization output, demonstrating stable and repeatable perform

3.3. Parameter Settings

To ensure reproducibility and enable fair comparison between SC- and CSC-based localization, all experimental parameters were consistently applied across both the self-collected and public datasets.
For the cylindrical projection, the 360° azimuth range was uniformly divided into N θ = 60 bins, corresponding to an angular resolution of 6° per bin. The height axis was discretized over the range z [ 10 m , + 20 m ] , producing a total vertical span of 30 m. This configuration provides consistent spatial resolution regardless of radial distance, ensuring that distant structures receive the same vertical representation as nearby regions.
Each cell in the CSC descriptor stores a three-dimensional feature vector
a b θ , b z = [ r max , dens , I ¯ ] ,
as defined in Equation (3). A channel-wise weight vector
w = [ w r max , w dens , w I ]
was applied during descriptor flattening to balance the contribution of geometric ( r max , dens ) and radiometric ( I ¯ ) cues. These weights were empirically tuned and fixed across all experiments to ensure consistent evaluation.
These weights were empirically tuned to enhance geometric cues (e.g., density, range) while preventing the intensity channel from dominating descriptor similarity. The same weight configuration was used for all datasets to ensure consistent evaluation.
For global retrieval, cosine distance with circular column shifts (Equation (5)) was used, and the system selected up to Top- K = 30 candidates from the descriptor database. This large retrieval pool provides robustness against viewpoint variation, environmental noise, and partial occlusions.
Among the retrieved candidates, the top three were used for ICP-based pose refinement. For each candidate, a local submap with a radius of 20–30 m was extracted, and point-to-point ICP was executed until the convergence threshold or iteration limit was reached. The candidate producing the minimum ICP fitness score was chosen as the final localization result (Equation (9)).
All modules were implemented in C++ and executed on the target computing platform. With the complete pipeline—including cylindrical projection, descriptor generation, global retrieval, and ICP refinement—the system operated at approximately 2 Hz, providing sufficient responsiveness for low-speed autonomous navigation and static localization tasks.

4. Experimental

To comprehensively validate the proposed CSC-based LiDAR-only localization framework, experiments were conducted using two types of datasets: (1) a self-collected outdoor dataset acquired in a complex coastal–forest environment, and (2) the publicly available MulRan-KIAST [] dataset containing repeated traversals of the same campus area. Together, these datasets allow evaluation under diverse conditions, including unstructured natural terrain, urban-like campus structures, temporal variations, and trajectory-dependent viewpoint changes. The complete workflow of the proposed method is summarized in Algorithm 1, which outlines the full CSC-based LiDAR-only global localization pipeline.

4.1. Self-Collected Dataset (Pohang, South Korea)

The first dataset was collected in a coastal–forest mixed terrain located in Pohang, South Korea. This region features elevation variations, curved paths, sloped road segments, and dense vegetation, making it a challenging environment for LiDAR-only localization. A sensing platform equipped with an Ouster OS1-32 LiDAR, a MicroStrain 3DM-GX5 IMU, and a SparkFun RTK-GPS module was driven along an outdoor route of approximately hl1 km. The detailed sensor platform configuration is shown in Figure 3, and the driving environment and UTM-referenced trajectories are illustrated in Figure 4a.
Figure 3. Configuration of the driving platform for data collection.
Figure 4. (a) Experimental environment (satellite imagery), (b) comparison of UTM trajectories, and (c) visualization of the mapping results. Since all maps share similar spatial structures, only representative visualizations are presented for clarity.
During the mapping stage, LiDAR, IMU, and GPS data were temporally synchronized and fused within the LIO-SAM framework to generate globally consistent maps. Both SC-based and CSC-based SLAM pipelines were executed independently, producing two sets of maps for each driving session. Descriptors extracted from individual LiDAR frames were stored together with their estimated poses to build the retrieval database.
Algorithm 1 CSC-Based LiDAR-Only Global Localization.
Require: 
Pre-built global map M , descriptor database DB = { ( D ^ k , T k ) } k = 1 N
Ensure: 
Global LiDAR-only pose estimate T *
  1:
Set descriptor parameters: N θ = 60 , z [ 10 , 20 ] m
  2:
Set retrieval/ICP parameters: Top- K = 30 , M = 3 , r local = 20 30 m , threshold τ icp
  3:
for each incoming query scan P q  do
  4:
                          ▹ CSC descriptor generation
  5:
    Convert all points p = ( x , y , z ) in P q to cylindrical coordinates ( r , θ , z )
  6:
    Quantize ( θ , z ) into bins ( b θ , b z ) using N θ and [ z min , z max ]
  7:
    for each bin ( b θ , b z )  do
  8:
        Collect point set P b θ , b z
  9:
        Compute maximum range r max in the bin
10:
        Compute point density dens and mean intensity I ¯
11:
        Form feature vector a b θ , b z = [ r max , dens , I ¯ ]
12:
    end for
13:
    Assemble CSC tensor D q and flatten to matrix form D ^ q
14:
    Apply channel-wise weighting (if enabled)
15:
                   ▹ Global retrieval (Top-K candidates)
16:
    for each database entry ( D ^ k , T k ) in DB  do
17:
        Compute cosine distance with circular shifts d k = d ( D ^ q , D ^ k )
18:
        Store best shift Δ θ k * associated with d k
19:
    end for
20:
    Select Top-K candidates C = { k 1 , , k K } sorted by ascending d k
21:
                ▹ Initial pose hypotheses from yaw correction
22:
    for each k C  do
23:
        Compute yaw correction Δ ψ k = Δ θ k * · 2 π N θ
24:
        Set initial pose T k ( 0 ) = T k · R z ( Δ ψ k )
25:
    end for
26:
                   ▹ ICP refinement on Top-M candidates
27:
     J min , T * I
28:
    Select Top-M indices from C with the smallest d k
29:
    for each selected candidate index k do
30:
        Extract local submap M k from M within radius r local around T k ( 0 )
31:
        Run ICP between P q and M k initialized at T k ( 0 )
32:
        Obtain refined pose T k and fitness score J k
33:
        if  J k < J min  then
34:
            J min J k , T * T k
35:
        end if
36:
    end for
37:
                          ▹ Final pose validation
38:
    if  J min < τ icp  then
39:
        Accept T * as the final LiDAR-only global pose
40:
    else
41:
        Mark localization as invalid (e.g., keep previous pose or odometry-only estimate)
42:
    end if
43:
end for
To evaluate pure LiDAR-only localization, GPS measurements were intentionally removed during the localization stage. This design simulates realistic autonomous operation in GPS-denied outdoor environments, in which the mapping and localization stages occur at different times.
Two separate LiDAR–IMU–GPS sequences, denoted as data1 and data2, were collected by driving the same route on different occasions. Although collected in the same environment, these sequences exhibit natural differences in vegetation, lighting, and viewpoint. For each sequence, SC- and CSC-based SLAM were performed independently, resulting in four maps:
  • SC-based maps: map1 (from data1), map2 (from data2);
  • CSC-based maps: map1 (from data1), map2 (from data2).
Figure 4c shows a representative map. These maps were used to analyze the localization accuracy, robustness to temporal changes, and map reusability when the robot revisits the same environment under slightly different conditions.

4.2. Experiments Using Public Dataset (MulRan-KIAST)

To examine the generalizability and scalability of the proposed method, additional experiments were conducted using the publicly available MulRan-KIAST dataset. This dataset contains LiDAR–IMU recordings collected on the KAIST campus, where the same spatial region was traversed three times along different driving trajectories. Compared to the self-collected natural-terrain dataset, the KAIST environment includes a mix of structured buildings, pedestrian walkways, and vegetation, offering complementary evaluation conditions.
For the experiments:
  • The second traversal was used as the reference mapping session, where SC-based and CSC-based SLAM were applied to build two maps.
  • The first and third traversals were used exclusively for LiDAR-only localization testing against the pre-built maps.
Because the three sequences cover the same areas while following different motion patterns, the dataset enables systematic analysis of viewpoint variation, loop-based localization stability, and map reusability. Figure 5 is visualization of the KAIST1 driving trajectory in the MulRan dataset overlaid on satellite imagery. By combining results from both the self-collected dataset and the MulRan-KIAST dataset, the proposed CSC framework is evaluated across diverse environments, natural and temporal variations, and trajectory-dependent perception differences—providing a comprehensive validation of its robustness and practicality.
Figure 5. Visualization of the KAIST1 driving trajectory in the MulRan dataset overlaid on satellite imagery.

4.3. Evaluation Metrics

To quantitatively evaluate the performance of the proposed Cylindrical Scan Context (CSC)-based localization algorithm, a set of standard metrics was employed. All experiments were conducted under identical environmental conditions and parameter settings for both the SC-based and CSC-based methods, ensuring a fair comparison focused solely on the effect of descriptor structure.
Root Mean Square Error (RMSE): The overall localization accuracy was evaluated using the Root Mean Square Error (RMSE) between the estimated position and the ground-truth trajectory. This metric represents the average localization deviation along the entire path and was computed in the UTM-K coordinate frame. Minimum, mean, and maximum position errors, as well as axis-wise statistics ( x , y , z ), were also reported. RMSE is defined as
RMSE = 1 N i = 1 N p i est p i GT 2 2 ,
where p i est and p i GT denote the estimated and ground-truth positions, respectively.
Top-K Accuracy: This metric represents the probability that the correct match is included among the Top-K retrieved candidates for each query scan []. Top-K Accuracy was computed for k = { 1 , 5 , 10 } , providing a measure of the retrieval success rate and search reliability in descriptor-based localization.
Precision–Recall (PR) Curve: The discriminative capability of the descriptors during global candidate retrieval was evaluated using the PR curve. A retrieved candidate was considered correct if the distance error between the estimated and ground-truth positions was within a predefined threshold ( t D , in meters). Precision and Recall were plotted over varying distance thresholds, and the Average Precision (AP)—computed as the Area Under the Curve (AUC)—was used as an integrated measure of retrieval performance [].
Detection Recall (DR) Curve: The DR curve was used to assess the relationship between descriptor similarity scores and actual localization success. For each query, the highest similarity score among correct matches was recorded, and Recall was measured over different score thresholds. This analysis quantifies how consistently the descriptor’s similarity score correlates with true spatial correspondence, thereby reflecting the reliability of similarity-based localization.
Overall, these metrics—RMSE, Top-k Accuracy, PR, and DR—were employed to comprehensively evaluate the accuracy, retrieval reliability, and robustness of the proposed CSC-based localization framework.

4.4. Localization Performance Evaluation

This section evaluates the localization performance of the proposed Cylindrical Scan Context (CSC) in comparison with the conventional Scan Context (SC). The experiments include Precision–Recall (PR) curves, Detection Recall (DR) curves, RMSE-based pose error, and Top-K retrieval accuracy, using both the self-collected Pohang datasets (D1_M2, D2_M1) and the MulRan–KAIST datasets (KAIST1–KAISTMAP2, KAIST3–KAISTMAP2).
For the self-collected dataset, the ground-truth correctness threshold was set to 1 m, and approximately 700 query scenes were used for evaluation. For the MulRan–KAIST dataset, a wider correctness threshold of 3 m was applied due to the large-scale campus environment and inherent trajectory variability, and approximately 1200 query scenes were used in the experiments. These settings ensure a fair and environment-appropriate comparison between SC and CSC.
The PR curve results shown in Figure 6a–d exhibit distinct characteristics depending on the dataset. For the self-collected D1_M2 pair, SC provides slightly higher precision in the mid-recall range (0.2–0.6), yielding a marginally larger AP. However, CSC becomes comparable as recall approaches 1. In contrast, for the D2_M1 case, CSC maintains consistently higher precision across almost the entire recall range, indicating that CSC is more robust to trajectory differences and environmental variations. For the MulRan–KAIST datasets, CSC demonstrates a significant advantage: in KAIST1–KAISTMAP2, CSC maintains near-perfect precision up to the mid-recall region, while SC shows a steep degradation. In KAIST3–KAISTMAP2, SC performs better in the mid-recall range, but CSC surpasses SC at higher recall levels, demonstrating more stable retrieval as similarity thresholds become strict. Although SC shows higher precision in some mid-recall regions—particularly in short-range or nearly identical revisit conditions—CSC demonstrates more stable performance overall across datasets, especially under viewpoint variation, elevation differences, and heterogeneous traversal paths.
Figure 6. Precision–Recall (PR) curves: (a) D1_M2, (b) D2_M1, (c) KAIST1–KAISTMAP2, (d) KAIST3–KAISTMAP2. Detection Recall (DR) curves: (e) D1_M2, (f) D2_M1, (g) KAIST1–KAISTMAP2, (h) KAIST3–KAISTMAP2.
The DR curve results in Figure 6e–h reinforce this observation. In D1_M2, CSC maintains higher recall than SC for score thresholds above 0.85, showing that CSC’s similarity score more reliably correlates with true geometric alignment. In D2_M1, both methods display similar DR behavior, but CSC retains a slight advantage in high-confidence regions. For KAIST1–KAISTMAP2, CSC significantly outperforms SC, maintaining recall above 0.95 over a wide threshold interval, while SC declines sharply. In KAIST3–KAISTMAP2, CSC consistently achieves higher recall across most thresholds, although SC remains competitive in the mid-threshold region. These results confirm that CSC provides more stable and reliable detection performance, especially in large-scale or complex campus environments.
The RMSE-based pose accuracy analysis, summarized in Table 1, further supports the superiority of CSC. For D1_M2, CSC yields a 17% reduction in overall distance RMSE (0.888 m vs. 1.064 m), with substantial reductions in Y- and Z-axis errors by 41% and 42%, respectively. In D2_M1, CSC again achieves slightly lower overall RMSE (1.244 m vs. 1.321 m), with a significant improvement in Y-axis accuracy (0.281 m vs. 0.661 m). CSC also performs strongly on the MulRan–KAIST datasets: in KAIST1–KAISTMAP2, CSC reduces the distance RMSE by approximately 45% (8.4 m vs. 15.113 m), with improvements across all axes. Although KAIST3–KAISTMAP2 shows similar distance RMSE for both methods, CSC maintains more consistent performance on the Z-axis and shows improved lateral accuracy.
Table 1. RMSE comparison of CSC and SC on self-collected and MulRan–KAIST datasets.
Table 2 presents the retrieval accuracy evaluation. For D1_M2, both SC and CSC achieve high Top-K accuracy, with SC slightly outperforming CSC at Top-1. However, CSC reaches perfect accuracy (1.000) at Top-10. For D2_M1, SC yields higher Top-1 accuracy, but CSC shows stronger performance at Top-5 and Top-10, indicating that CSC captures more correct candidates when the retrieval window expands. A similar trend is evident in the MulRan–KAIST results: CSC consistently maintains competitive or superior accuracy as K increases, demonstrating improved recall and robustness under viewpoint and path variations.
Table 2. Top-K retrieval accuracy comparison of CSC and SC on self-collected and MulRan–KAIST datasets.
In summary, the experimental results across PR/DR curves, RMSE accuracy, and Top-K retrieval demonstrate that CSC provides more stable and reliable localization performance than SC. By preserving vertical structure and offering a more consistent geometric encoding in cylindrical coordinates, CSC enhances both retrieval stability and final pose accuracy in diverse and unstructured outdoor environments. Although SC remains competitive in short-range or nearly identical revisit scenarios, CSC proves to be the more robust choice for real-world applications involving temporal changes, elevation differences, and heterogeneous trajectories.

4.5. Performance Analysis

The overall experimental results demonstrate that CSC provides more consistent and reliable localization performance than SC across diverse conditions. In terms of metric-based evaluation, CSC achieves lower RMSE values—particularly along the Y- and Z-axes—indicating improved spatial alignment when elevation changes or vertical structural variations are present. CSC also shows stronger retrieval stability, with rapidly increasing Top-K accuracy as the candidate range expands, suggesting that its cylindrical encoding captures more valid matches even under viewpoint differences.
The PR and DR curves further confirm this trend: CSC maintains higher precision and recall in high-confidence regions, especially in datasets involving trajectory or environmental changes. This indicates that the similarity scores produced by CSC more faithfully reflect true geometric correspondence. Altogether, these observations show that preserving vertical structure and representing the environment in cylindrical coordinates enable CSC to deliver robust LiDAR-only localization performance, particularly in unstructured or dynamically varying outdoor environments.

5. Conclusions and Future Works

This work presented a LiDAR-only localization framework based on the Cylindrical Scan Context (CSC), which projects point clouds onto an azimuth–height domain and incorporates multiple physical channels to preserve vertical structural information more effectively than the conventional Scan Context (SC). By separating mapping and localization sessions, the proposed framework was evaluated under realistic GPS-denied conditions using pre-built maps.
Across all datasets, CSC demonstrated improved localization stability and geometric accuracy. In the self-collected experiments, CSC achieved up to 17% lower distance RMSE and a 57% reduction in Y-axis error, while in the MulRan–KAIST dataset, it reduced distance RMSE by approximately 45%. CSC also showed stronger recall growth in Top-K retrieval and more stable performance in the high-recall regions of the PR/DR curves, highlighting its robustness to viewpoint changes, elevation differences, and traversal variations. Moreover, the cylindrical representation yielded favorable scalability, offering consistent performance across structurally different datasets with lightweight computational cost.
Despite these improvements, the evaluation remains focused on ground-level outdoor environments, and further validation is required for dense urban, multi-level, or industrial settings. Long-term environmental variations were not explicitly considered, and the mapping stage relied on LIO-SAM with fixed parameters, limiting broader sensitivity analyses.
Future work will include channel ablation and accuracy–complexity analysis, development of adaptive and scalable descriptor variants, and evaluation across more diverse and long-term datasets. Statistical significance testing and confidence-interval reporting will also be incorporated to strengthen the reliability of performance comparisons.
In summary, CSC effectively mitigates structural limitations of SC by preserving vertical geometry and enhancing descriptor consistency under environmental and temporal variations. The results confirm CSC as a practical and scalable solution for LiDAR-only localization in GPS-denied or degraded environments, with strong potential for next-generation autonomous navigation systems.

Author Contributions

Conceptualization, C.B.; Methodology, C.B. and J.B.; Software, C.B., S.P. and S.K.; Validation, C.B., G.R.C. and M.L.; Formal Analysis, C.B. and J.H.P.; Writing—Original Draft Preparation, C.B. and J.B.; Writing—Review and Editing, G.R.C. and J.B.; Supervision, G.R.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Institute of Civil Military Technology Cooperation, funded by the Defense Acquisition Program Administration and Ministry of Trade, Industry, and Energy of Korean government under grant No. 22-CM-GU-08.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data supporting the findings of this study are not publicly available due to institutional restrictions and privacy/security considerations.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LiDARLight Detection And Ranging
SCScan Context
CSCCylindrical Scan Context
SLAMSimultaneous Localization And Mapping
ICPIterative Closest Point
IMUInertial Measurement Unit
GPSGlobal Positioning System
UTMUniversal Transverse Mercator
RMSERoot Mean Square Error
PR CurvePrecision–Recall Curve
DR CurveDetection Recall Curve

References

  1. Zhang, Y.; Shi, P.; Li, J. Lidar-based place recognition for autonomous driving: A survey. ACM Comput. Surv. 2024, 57, 1–36. [Google Scholar] [CrossRef]
  2. Du, Z.; Ji, S.; Khoshelham, K. 3-D LiDAR-based place recognition techniques: A review of the past ten years. IEEE Trans. Instrum. Meas. 2024, 73, 1–24. [Google Scholar] [CrossRef]
  3. Kim, G.; Kim, A. Scan context: Egocentric spatial descriptor for place recognition within 3d point cloud map. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4802–4809. [Google Scholar]
  4. Bae, C.; Lee, Y.C.; Yu, W.; Lee, S. Spherically stratified point projection: Feature image generation for object classification using 3d lidar data. Sensors 2021, 21, 7860. [Google Scholar] [CrossRef] [PubMed]
  5. Yang, B.; Li, Z.; Li, W.; Cai, Z.; Wen, C.; Zang, Y.; Muller, M.; Wang, C. Lisa: Lidar localization with semantic awareness. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 15271–15280. [Google Scholar]
  6. Li, W.; Yang, Y.; Yu, S.; Hu, G.; Wen, C.; Cheng, M.; Wang, C. Diffloc: Diffusion model for outdoor lidar localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 15045–15054. [Google Scholar]
  7. Charroud, A.; El Moutaouakil, K.; Palade, V.; Yahyaouy, A.; Onyekpe, U.; Eyo, E.U. Localization and mapping for self-driving vehicles: A survey. Machines 2024, 12, 118. [Google Scholar] [CrossRef]
  8. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. In Proceedings of the Sensor fusion IV: Control Paradigms and Data Structures, Boston, MA, USA, 14–15 November 1991; Volume 1611, pp. 586–606. [Google Scholar]
  9. Lim, H.; Hwang, S.; Shin, S.; Myung, H. Normal distributions transform is enough: Real-time 3D scan matching for pose correction of mobile robot under large odometry uncertainties. In Proceedings of the 2020 20th International Conference on Control, Automation and Systems (ICCAS), Busan, Republic of Korea, 13–16 October 2020; pp. 1155–1161. [Google Scholar]
  10. Ren, Y.; Shen, Z.; Liu, W.; Chen, X. An Enhanced LiDAR-Based SLAM Framework: Improving NDT Odometry with Efficient Feature Extraction and Loop Closure Detection. Processes 2025, 13, 272. [Google Scholar] [CrossRef]
  11. Chen, S.; Ma, H.; Jiang, C.; Zhou, B.; Xue, W.; Xiao, Z.; Li, Q. NDT-LOAM: A real-time LiDAR odometry and mapping with weighted NDT and LFA. IEEE Sens. J. 2021, 22, 3660–3671. [Google Scholar] [CrossRef]
  12. He, L.; Wang, X.; Zhang, H. M2DP: A novel 3D point cloud descriptor and its application in loop closure detection. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 231–237. [Google Scholar]
  13. Wang, H.; Wang, C.; Xie, L. Intensity scan context: Coding intensity and geometry relations for loop closure detection. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Virtually, 31 May–31 August 2020; pp. 2095–2101. [Google Scholar]
  14. Kim, G.; Choi, S.; Kim, A. Scan context++: Structural place recognition robust to rotation and lateral variations in urban environments. IEEE Trans. Robot. 2021, 38, 1856–1874. [Google Scholar] [CrossRef]
  15. Zhao, L.; Hu, Y.; Han, F.; Dou, Z.; Li, S.; Zhang, Y.; Wu, Q. Multi-sensor missile-borne LiDAR point cloud data augmentation based on Monte Carlo distortion simulation. CAAI Trans. Intell. Technol. 2025, 10, 300–316. [Google Scholar] [CrossRef]
  16. Yang, Y.; Li, H.; Li, X.; Zhao, Q.; Wu, J.; Lin, Z. Sognet: Scene overlap graph network for panoptic segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12637–12644. [Google Scholar]
  17. Wang, Y.; Sun, Z.; Xu, C.Z.; Sarma, S.E.; Yang, J.; Kong, H. Lidar iris for loop-closure detection. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 5769–5775. [Google Scholar]
  18. Si, Y.; Han, W.; Yu, D.; Bao, B.; Duan, J.; Zhan, X.; Shi, T. MixedSCNet: LiDAR-Based Place Recognition Using Multi-Channel Scan Context Neural Network. Electronics 2024, 13, 406. [Google Scholar] [CrossRef]
  19. Kim, G.; Park, Y.S.; Cho, Y.; Jeong, J.; Kim, A. Mulran: Multimodal range dataset for urban place recognition. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Virtually, 31 May–31 August 2020; pp. 6246–6253. [Google Scholar]
  20. Mekuria, R.; Laserre, S.; Tulvan, C. Performance assessment of point cloud compression. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.