Previous Article in Journal
Management of Severe COVID-19 Diagnosis Using Machine Learning
Previous Article in Special Issue
Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Energy-Conscious Lightweight LiDAR SLAM with 2D Range Projection and Multi-Stage Outlier Filtering for Intelligent Driving

1
College of Automotive and Transportation Engineering, Yancheng Polytechnic College, Yancheng 224005, China
2
College of Automotive Engineering, Yancheng Institute of Technology, Yancheng 224005, China
3
College of Mechanical Engineering, Universiti Sains Malaysia, George Town 11800, Malaysia
*
Author to whom correspondence should be addressed.
Computation 2025, 13(10), 239; https://doi.org/10.3390/computation13100239
Submission received: 3 September 2025 / Revised: 27 September 2025 / Accepted: 4 October 2025 / Published: 10 October 2025
(This article belongs to the Special Issue Object Detection Models for Transportation Systems)

Abstract

To meet the increasing demands of energy efficiency and real-time performance in autonomous driving systems, this paper presents a lightweight and robust LiDAR SLAM framework designed with power-aware considerations. The proposed system introduces three core innovations. First, it replaces traditional ordered point cloud indexing with a 2D range image projection, significantly reducing memory usage and enabling efficient feature extraction with curvature-based criteria. Second, a multi-stage outlier rejection mechanism is employed to enhance feature robustness by adaptively filtering occluded and noisy points. Third, we propose a dynamically filtered local mapping strategy that adjusts keyframe density in real time, ensuring geometric constraint sufficiency while minimizing redundant computation. These components collectively contribute to a SLAM system that achieves high localization accuracy with reduced computational load and energy consumption. Experimental results on representative autonomous driving datasets demonstrate that our method outperforms existing approaches in both efficiency and robustness, making it well-suited for deployment in low-power and real-time scenarios within intelligent transportation systems.

1. Introduction

With the rapid advancement of autonomous driving technology, environmental perception and self-localization systems have become critical components of intelligent transportation systems. Among them, LiDAR-based Simultaneous Localization and Mapping (SLAM) techniques have attracted significant attention due to their strong adaptability to complex environments and high-precision performance [1,2]. However, mainstream SLAM algorithms often struggle to meet the practical requirements of autonomous driving scenarios, which demand high-frequency sensing and low-power operation [3,4]. These methods [5,6,7] tend to suffer from excessive computational redundancy, high memory consumption, and strong dependence on hardware resources, thus limiting their deployment on edge devices or energy-constrained platforms. Therefore, there is an urgent need for a robust yet lightweight LiDAR-internal SLAM framework that can maintain localization accuracy while effectively reducing system energy consumption and computational overhead [8,9,10,11].
In most mainstream LiDAR SLAM systems, feature extraction relies on ordered scan-line structures or neighborhood construction from raw 3D point clouds. Representative methods like LOAM [12,13,14,15] use sequential indexing to compute curvature and extract corner and planar features [16]. Specifically, the curvature calculation in LiDAR point clouds differs from the strict mathematical definition of curvature. Instead, it is approximated by measuring the variation in range values between a given point and several of its neighboring points along the same scan line. This procedure provides a fast and approximate evaluation of the local surface smoothness, where points with large curvature values correspond to edge-like structures, while those with small curvature values correspond to planar regions. By applying a simple threshold, edge and surface features can be efficiently selected. Consequently, the extraction process strongly depends on the availability of ordered scan-line structures or on the construction of local neighborhoods in the raw 3D point cloud. However, this approach suffers from low efficiency when processing large-scale or unstructured environments, and it can result in indexing mismatches under occlusion, sparsity, or motion distortion. Furthermore, traditional feature extraction pipelines often depend on OpenCV-based sorting or custom heap operations, which incur high computational and memory overhead on embedded platforms. Meeting the stringent real-time and energy-efficient requirements of autonomous driving demands a more compact and indexable point cloud structure that enables faster and more reliable feature extraction [17,18].
In real-world autonomous driving environments, LiDAR data often contains numerous outliers and noisy points caused by sensor artifacts, occlusion, or surface discontinuity. If not properly filtered, these outliers can introduce false constraints during feature association and back-end optimization, leading to inaccurate or divergent pose estimates. Most SLAM systems rely on simple single-pass filters—such as fixed depth thresholds or curvature filters—which struggle to handle complex noise distributions [12,13,14,15,19]. Although some methods employ occlusion detection or curvature-normal filtering, they lack geometric consistency validation and often retain mismatched points. Moreover, outlier rejection is usually static and lacks a hierarchical structure. A multi-stage filtering strategy that incorporates spatial, depth, and structural cues is therefore crucial for improving system robustness and energy efficiency [20].
The construction of local maps plays a central role in scan-to-map matching and back-end pose optimization [21]. Most existing SLAM frameworks use fixed-radius (spatially driven) or fixed-window (temporally driven) strategies to select historical keyframes [22]. While such methods perform well under steady linear motion, they cannot adapt to dynamic or nonlinear driving behaviors such as turning, accelerating, or stopping. In degenerate cases (e.g., repeated in-place rotation), local maps may fail to update in time, leading to pose drift or registration failure. Moreover, overly dense local maps result in unnecessary optimization load and increased power consumption. Therefore, an adaptive keyframe selection strategy that adjusts spatial coverage and density in response to vehicle motion is essential for enabling energy-efficient, high-precision SLAM in real-time driving applications.
To address these challenges, this paper proposes a lightweight and energy-efficient LiDAR SLAM framework tailored for autonomous driving scenarios. The proposed system balances localization accuracy, computational load, and robustness under diverse motion conditions. The main contributions are as follows:
(1)
Range Image-Based Feature Extraction without Ordered Indexing:
A novel projection method is introduced that transforms unordered 3D LiDAR data into a structured 2D range image. This enables efficient neighborhood access and curvature computation while preserving spatial geometry, significantly improving feature extraction speed and reducing memory overhead. The approach supports parallel implementation and is well-suited for embedded deployment.
(2)
Multi-Stage Outlier Rejection with Geometric and Structural Constraints:
A hierarchical filtering process is designed to progressively remove invalid and unstable points based on spatial proximity, depth continuity, and curvature consistency. Points affected by occlusion or geometric instability are filtered early, enhancing robustness in noisy and cluttered environments.
(3)
(Dynamically Filtered Local Map Construction for Pose Optimization:
An adaptive local mapping strategy is proposed that selects keyframes using spatial radius queries and dynamically adjusts voxel filter resolution based on system state. A temporal compensation mechanism ensures sufficient constraints during initialization. This method reduces frame redundancy and supports stable real-time optimization under varying motion conditions.
These components are integrated into a complete SLAM pipeline and evaluated on benchmark autonomous driving datasets. Experimental results demonstrate that the proposed method maintains competitive accuracy while significantly reducing computational load, making it suitable for real-time deployment on low-power autonomous platforms.

2. Lightweight and Robust Feature Extraction

This section presents an efficient and lightweight feature extraction algorithm tailored for LiDAR SLAM systems, aiming to enhance the robustness and computational efficiency of feature selection in complex driving environments. The proposed framework is implemented on top of the open-source LIO-SAM system. The method begins by projecting the raw 3D LiDAR point cloud into a structured 2D range image, significantly improving neighborhood access and reducing spatial redundancy. A novel curvature computation strategy based on 2D image neighborhoods is introduced, along with a multi-criteria feature screening mechanism that incorporates depth continuity, geometric stability, and neighborhood support. To avoid feature clustering and redundancy, a non-maximum suppression scheme is employed, ensuring spatially uniform and independent feature selection. This approach effectively suppresses false detections in occluded regions and enhances geometric consistency in sparse or distorted scans, providing high-quality, sparse, and reliable constraints for back-end pose optimization. The resulting system achieves reduced computational overhead while supporting energy-efficient and high-performance SLAM localization. Some details are in Figure 1.

2.1. Feature Point Extraction

Given an ordered point cloud P = { p 1 , p 2 , , p k } that has been corrected for motion distortion, each point p k contains 3D coordinates x k , y k , z k . First, the Euclidean distance of each point (i.e., the depth of each laser return in the point cloud) is computed and projected onto a two-dimensional matrix. Suppose the LiDAR sensor has N s c a n scan lines and a horizontal resolution corresponding to N h o r i z samples. If point p k corresponds to the i -th scan line and the j -th horizontal sample, then its range r k is assigned to the corresponding element of the range image as R i , j .
R ( i , j ) = r k = x k 2 + y k 2 + z k 2
In this work, the raw 3D LiDAR points are first projected into a structured 2D range image, where each row corresponds to one laser scan line and each column corresponds to a horizontal sampling index. The value of each matrix element stores the depth (range) measurement of the corresponding point. This dimensionality reduction from 3D to 2D provides an efficient and ordered data structure for subsequent processing.
For invalid measurements, their values in the range image are set to a maximum constant r max to facilitate subsequent processing. This 2D structure provides a foundation for efficient neighborhood access. To evaluate the smoothness of a point’s local region, this paper adopts the curvature computation method used in LOAM [13,14,15] (Lidar Odometry and Mapping). For any point p i , j in the range image, its curvature is defined as the squared difference between its range and the average range of its valid neighboring points in set S i , j . It is worth noting that in our formulation the curvature is computed using the squared depth difference, whereas the original LOAM defines curvature based on the squared difference of 3D Cartesian coordinates. This modification simplifies the computation and is more consistent with the 2D range image representation. The specific computation is given by:
c ( p i , j ) = p u , v S i , j R ( u , v ) | S i , j | R ( i , j ) 2
Here, R ( i , j ) is the range value of the center point p i , j , and S i , j represents the 2D valid neighborhood set of point p i , j in the range image. | S i , j | denotes the number of valid neighboring points. The construction of the neighborhood set S i , j is both dynamic and constrained. Whether a neighboring point p u , v can be selected as a valid neighbor of p i , j depends on the depth continuity between the two LiDAR points. A point p u , v is included in S i , j if it satisfies the following condition:
| R ( u , v ) R ( i , j ) | α R ( i , j )
where α is a proportional threshold factor (set to 0.1 in this paper). This condition ensures that only points lying on the same continuous surface are considered for curvature calculation, thereby significantly improving robustness against occlusion and edge discontinuities. To maintain statistical stability in curvature estimation, the valid neighboring points p u , v are selected as the 6 points to the left and right of p i , j along the scan line. Importantly, this depth continuity check (Equation (3)) also serves as the first layer of the proposed multi-stage outlier rejection strategy—one of the key innovations in this work. By enforcing strict local surface consistency at the neighborhood construction stage, the method eliminates unreliable measurements at an early stage, thereby enhancing the geometric purity of candidate features and reducing the risk of including occlusion-affected or structurally inconsistent points in subsequent processing.
To ensure the accuracy and stability of curvature statistics, points identified as geometrically unstable—primarily those near occlusions—must be excluded from feature selection. A point p i , j is considered an occluded or unstable point if it satisfies two conditions. First, the spatial proximity constraint requires that the point and its adjacent neighbor p i , j + 1 on the same scan line be physically continuous. This is verified by checking the index difference Δ idx in the original 1D point cloud. If this difference exceeds a threshold δ col (set to 10 in this work), the spatial continuity is violated, expressed as:
| I ( p i , j + 1 ) I ( p i , j ) | < δ col
Second, the depth discontinuity constraint ensures that even if the spatial proximity is satisfied, the two points must also have similar range values. If the absolute difference in depth exceeds a fixed threshold δ depth (set to 0.3), the point is marked as invalid. The condition is written as:
| R ( i , j + 1 ) R ( i , j ) | > δ depth
Points that meet both criteria, along with several neighboring points, are excluded from further feature extraction to avoid introducing unreliable geometric information. In addition, to guarantee the spatial uniformity of selected features, each scan line is divided into six subregions, and feature selection is performed independently within each region.
A point p i , j is classified as a corner point if it satisfies a set of strict geometric and statistical conditions. First, the point must exhibit a sufficiently high curvature, specifically c ( p i , j ) > τ edge , where the curvature threshold τ edge is set to 0.1 in this work. Second, the point must be supported by a sufficient number of valid neighbors in its local region, requiring | S i , j | > N min _ corner , with N min _ corner = 10 as defined. Third, the point must not be labeled as geometrically unstable or occluded in the earlier filtering stage. Finally, to further eliminate potential noise-induced outliers, a dynamic upper bound is imposed on the curvature value. Specifically, a point is rejected as a corner candidate if its curvature exceeds c ( p i , j ) < β ( R ( i , j ) ) 2 | S i , j | , where β is a noise suppression coefficient set to 0.1. This composite set of criteria ensures that the extracted corner features are both geometrically distinctive and statistically reliable, even in the presence of sensor noise or environmental complexity. This adaptive curvature suppression mechanism serves as the second stage of the proposed multi-stage outlier rejection strategy. By introducing a dynamic upper bound conditioned on the local range and neighborhood density, the method effectively filters out noise-induced high-curvature points that would otherwise degrade the reliability of structural feature extraction. Together with the initial depth-based filtering, this layered design significantly enhances the robustness and selectivity of the feature screening process, which is central to the lightweight and energy-efficient SLAM framework proposed in this study.
A point p i , j is identified as a surface point if it meets a set of well-defined criteria. First, the point must exhibit low curvature, specifically c ( p i , j ) < τ s u r f , where the surface threshold τ s u r f is set to 0.1 in this paper. In practice, a curvature value smaller than 0.1 is classified as a surface point, while a value greater than 1.0 is classified as a corner point. For intermediate values between 0.1 and 1.0, the points are uniformly treated as surface points. This thresholding strategy ensures a clear and robust separation of geometric features while avoiding ambiguous classifications. Second, to ensure the reliability of curvature estimation, the point must have adequate neighborhood support—meaning the number of neighboring points that pass the depth consistency check must be greater than six. Third, the point must not have been previously marked as geometrically unstable or occluded in the pre-filtering stage. Once these conditions are satisfied, the selected surface points—together with the previously extracted corner points—form two sparse and reliable sets of geometric features. These feature sets are further refined using a voxel grid filter, which reduces redundancy while preserving structural information, and are then passed to the back-end module for pose optimization. The Pseudo Code is provided in Table 1.

2.2. Local Map Construction

In LiDAR-based SLAM systems, the quality of local map construction in the scan-to-map module directly influences the accuracy of pose estimation and the computational efficiency of the system. The local map is typically a spatially adjacent submap dynamically extracted from a global set of historical keyframes. Its construction involves transforming, merging, and downsampling the feature point clouds of selected keyframes [14,15]. This process is designed to ensure sufficient geometric constraints while minimizing the volume of data involved in optimization, thereby achieving a balanced trade-off between real-time performance and localization accuracy. Traditional strategies for constructing the local map generally fall into two categories, each with inherent drawbacks. Spatially driven methods select keyframes within a fixed-radius region (e.g., 30 m) around the current pose, offering rich geometric context and strong translational constraints, which are beneficial for large-scale localization. However, when the platform moves slowly or remains stationary, this strategy can lead to long periods without map updates, resulting in repeated matching against outdated observations and the accumulation of pose drift. In contrast, temporally driven strategies select the most recent N keyframes based on time, ensuring that fresh observations are always incorporated and enabling rapid response to orientation changes. Nonetheless, during high-speed linear motion, these keyframes tend to be spatially concentrated, which may exclude distant but critical structural information, thereby weakening translational estimation. These limitations highlight the need for a more adaptive and context-aware local mapping strategy.
The first step in constructing the local map is to extract a subset of historical keyframes that are spatially relevant to the current pose. Let K p 3 = { p i } i = 1 N denote the set of 3D position vectors of all N stored keyframes. The position of the most recent keyframe is denoted as p l a t e s t . To identify keyframes near the current location, the system employs a K-d tree structure to index K p 3 , enabling efficient spatial queries. A one-time radius search centered at p l a t e s t is performed to generate an initial candidate set of relevant keyframes, denoted as:
I c a n d = { i |     p i p l a t e s t R s e a r c h }
where the search radius R s e a r c h is empirically set to 50.0 m. This step ensures that only keyframes with spatial proximity to the current pose are selected for inclusion in the local map, thereby laying the foundation for a compact and context-aware local submap used in scan-to-map registration.
To prevent excessive computational overhead and redundant geometric information caused by the dense spatial clustering of selected keyframes, the system performs voxel grid downsampling on the point clouds of the candidate keyframes to regulate spatial density. The voxel size for downsampling, denoted as ρ k , is determined dynamically based on the system state. Its baseline value ρ b a s e is set to 0.5 m, and the adjustment strategy is as follows:
ρ k = 0.25   m N < 11 0.5   m other
Here, N represents the number of keyframes currently stored. When N < 11 , the system is considered to be in an initialization phase. In this case, a finer resolution (i.e., smaller voxel size of 0.25 m) is used to preserve more keyframes in the local map, enhancing geometric constraint diversity and improving pose robustness during the early stages. This dynamic strategy ensures better stability and localization accuracy in scenarios with limited prior observations. To further reinforce system robustness in the initialization phase, an additional temporal-based keyframe supplement mechanism is introduced. Specifically, when N < 11 , a temporal window is used to include the most recent keyframes within 5.0 s. Let I f i l t e r e d denote the index set of candidate keyframes selected through spatial search and downsampling; the final keyframe index set is defined as:
I f i n a l = I f i l t e r e d { i | t curr t i | < 5.0   s }
where t curr is the timestamp of the current LiDAR scan and t i is the timestamp of historical keyframe iii. The 5.0 s temporal window helps ensure that temporally recent but spatially excluded keyframes are retained, which improves the richness and timeliness of the local map during initialization. In practice, the local map is maintained as a sliding window of keyframes, where the point clouds from selected frames are transformed into a common reference frame through 3D pose transformations. Following the approaches in [8,12,13,16], the transformed keyframe point clouds are then aggregated to construct the local map, which provides consistent geometric constraints for scan-to-map registration.

2.3. Scan-to-Map Registration and Pose Optimization

A Scan-to-Map registration framework is employed to perform real-time pose refinement for each incoming LiDAR scan. The core objective is to align the current LiDAR frame with a local submap constructed from historical keyframes in the global coordinate system. This alignment process is formulated as a nonlinear least-squares optimization problem, which aims to minimize the residuals between the current scan’s feature points and their corresponding planar and edge features in the local map. By minimizing both point-to-plane and point-to-edge residuals, the system obtains a precise estimation of the current pose in the global reference frame. To accelerate the correspondence search during iterative optimization, two separate K-d tree structures are built for edge and planar features, denoted as the corner map M c and surface map M s , respectively. The optimization proceeds in an iterative manner until convergence is reached or a maximum number of iterations is met, which is set to 30 in this implementation.
To construct residual constraints for edge features, each extracted edge point P c o r in the current LiDAR scan first searches for its five nearest neighbors in the local corner map M c . These neighbors form a candidate set c = p i i = 1 5 , which is used to define a potential line structure L via Principal Component Analysis. The mean of the five points is computed as the centroid P ¯ = 1 5 i = 1 5 p i , and a covariance matrix H describing the spatial distribution of the set is constructed by [23]:
H = i = 1 5 ( p i p ¯ ) ( p ¯ i p ¯ ) T
Eigen decomposition is then performed on matrix H . The eigenvector u ^ corresponding to the largest eigenvalue λ 1 indicates the direction of greatest variance in the data and is taken as the direction of the candidate line. Thus, the fitted line L is uniquely defined by the centroid p ¯ and the unit vector u ^ . To ensure the reliability of this line fitting, the system evaluates the linearity of the neighbor set by comparing the dominant eigenvalue λ 1 with the second largest eigenvalue λ 2 . The candidate line is deemed valid only if the condition λ 1 > k c λ 2 is satisfied, where k c is a curvature ratio threshold empirically set to 5.0. If the line passes this check, the point-to-line residual d c is computed as the perpendicular distance from the original point p s u r f to the line L , given by:
d c = ( p c o r p ) × u
This residual is then incorporated into the pose optimization process. For each surface point p s u r f in the current LiDAR scan, the system searches its five nearest neighbors from the local surface point map M s , forming a point set c = p i i = 1 5 . A least-squares method is then employed to fit a local 3D plane π to this set. The general form of a plane equation is a x + b y + c z + d = 0 , which can be rewritten as a x + b y + c z = d [24]. Assuming d 0 , both sides can be divided by d , and the normalized plane normal vector is defined as v = [ a / ( d ) , b / ( d ) , c / ( d ) ] T . The equation becomes p T v = 1 , where p i = [ x i , y i , z i ] T . Given the five points p i s , the system constructs an overdetermined linear system A v = b , where
A = x 1 y 1 z 1 x 2 y 2 z 2 x 5 y 5 z 5 , b = 1 1 1
This formulation allows for a closed-form least-squares solution to estimate the plane parameters from the neighborhood points, which will later be used to compute the point-to-plane residuals during pose optimization. To solve the overdetermined linear system A v = b , the system employs QR decomposition to obtain the least-squares solution v 0 that minimizes the residual norm A v b 2 . From this solution vector v 0 , the normalized plane parameters can be recovered. Letting v o = p s , the unit normal vector of the plane is defined as n ^ = v o / p s , and the plane’s offset from the origin is given by the constant term d = 1 / p s . At this point, the plane π is uniquely determined by its unit normal n ^ and distance d . To ensure the quality of the plane fitting, the system further requires that all points p i in the neighborhood used for fitting lie within a bounded distance from the estimated plane. Specifically, for all i { 1 , , 5 } , the point-to-plane residual must satisfy the constraint: | n ^ T p i + d | < δ s , where δ s is a predefined threshold for plane fitting accuracy, and is set to 0.15 in this implementation.
After obtaining the point-to-plane residuals d s for each current scan point with respect to the corresponding fitted planes π , the system calculates each residual as the perpendicular distance from the point p s u r f to the plane, using the equation [25]:
d s = n ^ T p surf + d = a x surf + b y surf + c z surf + d
where n ^ and d are the unit normal vector and offset of the plane, respectively. Once all valid feature residuals and their corresponding Jacobian matrices are obtained, the system formulates a nonlinear least-squares optimization problem, which aims to find a pose increment Δ x 6 that minimizes the sum of squared residuals: min Δ x i d i ( x k Δ x ) 2 . To solve this nonlinear optimization problem, the system employs the Gauss–Newton method. A first-order Taylor expansion is applied to linearize the residuals around the current pose estimate, yielding the standard Gauss–Newton equation: ( J T J ) Δ x = J T r , where J is the Jacobian matrix and r is the residual vector [16]. The system uses QR decomposition to solve this linear system for the pose increment Δ x . The resulting pose update is then applied to the current pose estimate: x k + 1 = x k Δ x . The optimization is deemed converged when both the rotation and translation components of Δ x fall below predefined thresholds (0.05 degrees and 0.05 m, respectively), at which point the iterative process terminates.

3. Results

The dataset used in this study is UrbanNav-HK-Data2020314, which was collected in a low-urbanization area in Kowloon, Hong Kong [26]. The environment features typical urban road scenarios and is well-suited for algorithm verification and benchmarking. As was shown in Figure 2, the data collection platform is a modified Honda Fit vehicle equipped with a range of sensors, including a Velodyne HDL-32E 3D LiDAR (360° HFOV, ±10° to ±30° VFOV, 80 m range, 10 Hz, Velodyne Lidar, Inc., San Jose, CA, USA), an Xsens MTi-10 IMU (100 Hz, AHRS, Velodyne Lidar, Inc., San Jose, CA, USA), a monocular camera with a resolution of 1920 × 1200 and a field of view of 79.4° × 96.8° at 10 Hz, and GNSS receivers such as the u-blox M8T (Velodyne Lidar, Inc., San Jose, CA, USA) and NovAtel SPAN-CPT (Velodyne Lidar, Inc., San Jose, CA, USA). The SPAN-CPT system integrates RTK GNSS and an inertial navigation system, achieving a positioning RMSE of 5 cm at 1 Hz. All sensors were mounted on the vehicle roof, with extrinsic calibrations and time synchronization completed. The data was collected on 14 March 2020, with a total size of 27.0 GB and a trajectory length of 1.21 km, covering diverse representative urban scenes. The official dataset also provides detailed sensor extrinsics, intrinsic parameters of the camera, and noise models of the IMU, serving as a reliable foundation for Scan-to-Map optimization and performance evaluation of SLAM algorithms.
This study adopts the UrbanNav-HK-Data20200314 sequence from the UrbanNav dataset for field testing [26]. The data was collected in a low-urbanization area of Kowloon, Hong Kong, featuring a diverse set of urban elements such as narrow streets, residential buildings, dense vegetation, and roadside parked vehicles. This environment poses considerable challenges and provides generality for the evaluation of SLAM and multi-sensor fusion localization algorithms. As illustrated in Figure 3, the dataset contains a closed-loop trajectory starting from Point A and passing through Points B, C, D, E, F, G, and H before returning to the starting position. The total path length is approximately 1.21 km. Eight key waypoints (A~H) are selected along the route, covering various typical scenes including intersections, curves, and straight lanes. These are used for subsequent analysis of keyframe positioning and trajectory accuracy. Figure 3b–i present real-world visual observations at each corresponding waypoint (A~H). The images reflect diverse conditions such as changes in road geometry, lighting variations, and the presence of dynamic objects like pedestrians and parked vehicles. Such complexity increases the difficulty of environmental perception and map construction, thereby offering a realistic and challenging benchmark for testing the robustness and adaptability of the Scan-to-Map method under practical urban navigation scenarios.
The experimental platform utilized in this study is a desktop computer equipped with an Intel Core i5-13400 processor, running the Ubuntu 20.04 operating system. The Robot Operating System (ROS1 Noetic) was selected as the software framework. To assess the resource efficiency of the proposed method, Energy-LIO, we conducted a quantitative comparison with the open-source LIO-SAM method using the same dataset, focusing specifically on memory consumption. The memory usage of both algorithms was recorded over a continuous runtime of approximately 800 s, as shown in Figure 4c. The horizontal axis represents the elapsed runtime of the vehicle, while the vertical axis indicates the memory occupied by the point cloud data. Throughout the entire experiment, Energy-LIO consistently maintained a lower memory footprint, generally staying below 4 MB with relatively small fluctuations. In contrast, LIO-SAM exhibited significantly higher memory demands, with a clear upward trend over time. During the early phase of the experiment (approximately the first 400 s), both methods demonstrated similar memory consumption. This is attributable to the necessity of accumulating a sufficient number of edge features (see Figure 4a) and planar features (see Figure 4b) during the initialization of the LiDAR odometry, which is essential for ensuring localization accuracy.
However, in the later phase of the experiment (after 400 s), the memory consumption of LIO-SAM increased rapidly, exceeding 5 MB and even peaking above 6 MB. In contrast, the growth of Energy-LIO’s memory usage remained more gradual and stable, indicating a more scalable and robust memory management mechanism. This memory efficiency can be largely attributed to the denoising and filtering strategies introduced in our method, which significantly reduce the backend memory load and contribute to a lightweight system architecture. The lightweight design of Energy-LIO translates directly into energy-saving benefits. By leveraging streamlined data structures and more efficient data processing pipelines, the method alleviates CPU burden associated with memory allocation and data traversal. This advantage is particularly crucial in real-world scenarios such as urban driving or highway navigation, where reducing overall computational overhead and instantaneous power consumption can extend the operational duration of autonomous vehicles and enhance the stability of long-term navigation tasks.
To further demonstrate the lightweight advantage of our proposed method, a comparative analysis of the resulting point cloud maps is conducted between Energy-LIO and LIO-SAM. This includes both an overview and close-up examinations of structural details (see Figure 5a,b). From a global perspective, the point cloud map generated by Energy-LIO (Figure 5a) exhibits superior clarity and sharpness. Structural features such as building facades, wall edges, and contours are rendered more distinctly and precisely, reflecting a higher fidelity in environmental representation. In contrast, the map produced by LIO-SAM (Figure 5b) appears less refined, with structural lines being blurrier and more dispersed. Artifacts such as “ghosting” or “dragging” effects are also observable in certain areas. This high-quality mapping performance is primarily attributed to the denoising and filtering strategies integrated into Energy-LIO. As illustrated in Figure 4a,b, our method applies multiple filtering stages during feature extraction, effectively eliminating outliers. In the backend processing stage, the algorithm emphasizes the retention of high-value edge features with strong geometric constraints, while significantly reducing the number of planar points that offer limited contribution to localization accuracy but demand substantial memory and computation.
The result is a set of rich and structured corner features that provide stronger and more deterministic constraints for pose estimation, thereby enhancing registration accuracy between successive LiDAR scans. This leads to more precise alignment and a higher-fidelity map. As shown in Figure 4b, the number of planar points processed by Energy-LIO remains significantly lower than that of LIO-SAM throughout the entire runtime. Importantly, this reduction is not a simple downsampling of the dataset; instead, it is achieved through intelligent identification and removal of unreliable planar points—often caused by dynamic objects or sensor noise.
As illustrated in Figure 4d—where the horizontal axis represents the runtime and the vertical axis indicates the number of keyframes maintained in the local map—Energy-LIO consistently sustains a significantly smaller keyframe set compared to LIO-SAM throughout the entire experiment. While LIO-SAM maintains a relatively high number of keyframes, typically ranging from 80 to 140, Energy-LIO operates within a much smaller range of 20 to 80 keyframes. On average, this represents a reduction of approximately 40% to 50% in keyframe set size. This notable decrease reflects the adoption of a more rigorous and efficient keyframe selection strategy in Energy-LIO. By maintaining a smaller yet more representative keyframe set, the system substantially reduces the computational burden associated with backend modules such as pose graph optimization and loop closure detection. Consequently, the overall computational cost and energy consumption are significantly lowered, highlighting the energy-efficient nature of the proposed method.
This strategy of “fewer but higher-quality” planar features ensures cleaner data input to the optimization backend, improving odometry accuracy while simultaneously reducing unnecessary computational overhead. In essence, the energy-efficient principle is embedded in the algorithm’s core by reducing the processing of ineffective data. Although both Energy-LIO and LIO-SAM are capable of reconstructing the overall topological structure of the environment (see Figure 5a,b), Energy-LIO demonstrates superior geometric consistency (see Figure 5e area C and D, Figure 5f area C and D). For instance, in areas featuring overpasses or parked vehicles, Energy-LIO generates sharper and more compact point cloud structures (see Figure 5c area A and B). In contrast, LIO-SAM suffers from coarser lines or duplicated boundaries, indicating the presence of accumulated pose drift over long durations (see Figure 5d area A and B). These misalignments result in slightly displaced representations of the same physical structures, degrading the geometric consistency of the reconstructed map.
The ability of Energy-LIO to achieve stable tracking with significantly fewer keyframes—as demonstrated in the aforementioned map quality analysis—stems from its intelligent keyframe selection strategy. Unlike traditional methods that rely on fixed spatial or temporal thresholds, Energy-LIO evaluates multiple criteria, including pose variation, information gain, and feature matching quality. A new frame is designated as a keyframe only if it contributes sufficient novel information to the system. This selective approach ensures that each retained keyframe carries high informational value, effectively preventing the unnecessary growth of the pose graph and avoiding redundant computations. By minimizing keyframe redundancy without compromising estimation accuracy, Energy-LIO achieves a more compact and computationally efficient backend, aligning with its core objective of energy-aware system design.
This study further employs the KITTI dataset [27,28] to conduct a more comprehensive quantitative evaluation of the proposed method in terms of odometry accuracy. The KITTI dataset was collected using a Velodyne HDL-64E 3D LiDAR scanner operating at 10 Hz, with each frame capturing approximately 100,000 three-dimensional points. These point clouds, which store the coordinates (x, y, z) along with the reflectance intensity (r), serve as the primary input data for LiDAR SLAM algorithms. Meanwhile, to provide the “ground truth” for evaluation, the vehicle was also equipped with an OXTS RT3003 high-precision GPS/IMU integrated navigation system. By combining GPS measurements with an inertial measurement unit and applying differential positioning techniques (RTK) in post-processing, this system can generate centimeter-level ground-truth vehicle poses (position and orientation). The test sequences include sequences 01, 02, 04, 05, 06, and 07, which cover different driving scenarios such as repeated-path driving in road environments (sequences 02 and 05), short straight-path driving (sequences 01 and 04), and complex-path driving (sequence 06). The experimental equipment setup is illustrated in Figure 6.
We evaluated the two methods in terms of the standard deviation (STD) and root mean square error (RMSE) of the Absolute Pose Error (APE). The results are summarized in Table 2, and the corresponding trajectory comparisons are illustrated in Figure 7. Table 1 presents the APE results across six test sequences, including the metrics RMSE, MEAN, and STD. Overall, Energy-LIO achieved consistent improvements over LIO-SAM in most sequences. Specifically, the average RMSE of Energy-LIO was reduced by approximately 6.4%, and the average STD decreased by about 12.5% compared with LIO-SAM. The most significant improvements were observed in Seq02 and Seq05, where the RMSE decreased by about 14.4% and 15.7%, respectively. These results indicate that excluding noisy points during the feature extraction stage has a substantial impact on localization accuracy, and importantly, this improvement was achieved while using fewer corner and surface points during scan-to-map matching (as demonstrated in previous experiments). However, in Seq04, the RMSE showed a slight increase of about 5.2%, while the STD decreased significantly. This can be attributed to the fact that Seq04 involves short straight-line driving (see Figure 7c), where the travel distance is limited. In such simple and short trajectories, LIO-SAM exhibited an advantage. Nevertheless, for longer and more complex driving sequences, Energy-LIO consistently demonstrated superior performance.

4. Conclusions

This study presents Energy-LIO, a lightweight and energy-conscious LiDAR SLAM framework tailored for autonomous driving in complex urban environments. By integrating range image-based feature extraction, a multi-stage outlier filtering strategy, and an adaptive keyframe selection mechanism, the proposed system effectively balances localization accuracy, computational efficiency, and memory usage. Experimental results on the UrbanNav dataset demonstrate that Energy-LIO achieves high-fidelity mapping with significantly fewer feature points and keyframes compared to LIO-SAM, while maintaining or surpassing its localization performance. Notably, the system’s memory footprint remains below 4 MB throughout long-duration operation, and the average reduction in keyframe count reaches 40~50%, reflecting substantial computational savings. The design of Energy-LIO embeds energy efficiency at its core by minimizing data redundancy, selectively processing high-value information, and avoiding unnecessary backend optimization. These characteristics make it particularly well-suited for deployment on resource-constrained platforms, such as embedded systems or autonomous vehicles with limited onboard computing capacity. Overall, Energy-LIO contributes a practical and scalable solution to real-time, power-aware SLAM and offers valuable insights into sustainable algorithm design for intelligent transportation applications.

Author Contributions

Conceptualization, C.W.; methodology, T.L.; software, X.H.; validation, X.H.; formal analysis, T.L.; investigation, C.W.; resources, T.L.; data curation, X.H.; writing—original draft preparation, C.W.; writing—review and editing, T.L.; visualization, C.W.; supervision, T.L.; project administration, X.H.; funding acquisition, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by Outstanding young backbone teachers of the “Qing Lan Project” in Jiangsu universities (Su Teacher Letter (2022) No.2) and Technology Innovation Team of Yancheng Polytechnic College (YGKJ202503).

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to express their gratitude to all those who helped them during the writing of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, C.; Qin, J.; Wang, S.; Yu, L.; Wang, Y. Accurate Rgb-D Slam in Dynamic Environments Based on Dynamic Visual Feature Removal. Sci. China-Inf. Sci. 2022, 65, 202206. [Google Scholar] [CrossRef]
  2. Zhao, C.; Hu, K.; Xu, J.; Zhao, L.; Han, B.; Wu, K.; Tian, M.; Yuan, S. Adaptive-Lio: Enhancing Robustness and Precision through Environmental Adaptation in Lidar Inertial Odometry. IEEE Internet Things J. 2025, 12, 12123–12136. [Google Scholar] [CrossRef]
  3. Geneva, P.; Eckenhoff, K.; Yang, Y.; Huang, G. Lips: Lidar-Inertial 3d Plane Slam. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018. [Google Scholar]
  4. Hu, Y.; Zhou, Q.; Miao, Z.; Yuan, H.; Liu, S. Outdoor Lidar-Inertial Slam Using Ground Constraints. Robotica 2024, 42, 1246–1261. [Google Scholar] [CrossRef]
  5. Xu, W.; Chen, Y.; Liu, S.; Nie, A.; Chen, R. Multi-Robot Cooperative Simultaneous Localization and Mapping Algorithm Based on Sub-Graph Partitioning. Sensors 2025, 25, 2953. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, Z.; Xu, Y.; Yuan, S.; Xie, L. Ig-Lio: An Incremental Gicp-Based Tightly-Coupled Lidar-Inertial Odometry. IEEE Robot. Autom. Lett. 2024, 9, 1883–1890. [Google Scholar] [CrossRef]
  7. Shen, B.; Xie, W.; Peng, X.; Qiao, X.; Guo, Z. Lio-Sam + +: A Lidar-Inertial Semantic Slam with Association Optimization and Keyframe Selection. Sensors 2024, 24, 7546. [Google Scholar] [CrossRef] [PubMed]
  8. Zhao, Z.; Zhang, W.; Gu, J.; Yang, J.; Huang, K. Lidar Mapping Optimization Based on Lightweight Semantic Segmentation. IEEE Trans. Intell. Veh. 2019, 4, 353–362. [Google Scholar] [CrossRef]
  9. Cao, F.; Wang, S.; Chen, X.; Wang, T.; Liu, L. Bev-Lslam: A Novel and Compact Bev Lidar Slam for Outdoor Environment. IEEE Robot. Autom. Lett. 2025, 10, 2462–2469. [Google Scholar] [CrossRef]
  10. Wang, H.; Wang, C.; Xie, L. Lightweight 3-D Localization and Mapping for Solid-State Lidar. IEEE Robot. Autom. Lett. 2021, 6, 1801–1807. [Google Scholar] [CrossRef]
  11. Chen, J.; Wang, H.; Yang, S. Tightly Coupled Lidar-Inertial Odometry and Mapping for Underground Environments. Sensors 2023, 23, 6834. [Google Scholar] [CrossRef] [PubMed]
  12. Shan, T.; Englot, B. Lego-Loam: Lightweight and Ground-Optimized Lidar Odometry and Mapping on Variable Terrain. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018. [Google Scholar]
  13. Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Ratti, C.; Rus, D. Lio-Sam: Tightly-Coupled Lidar Inertial Odometry via Smoothing and Mapping. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020. [Google Scholar]
  14. Shan, T.; Englot, B.; Ratti, C.; Rus, D. Lvi-Sam: Tightly-Coupled Lidar-Visual-Inertial Odometry Via Smoothing and Mapping. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021. [Google Scholar]
  15. Zhang, J.; Singh, S. Low-Drift and Real-Time Lidar Odometry and Mapping. Auton. Robot. 2017, 41, 401–416. [Google Scholar] [CrossRef]
  16. Zhou, P.; Guo, X.; Pei, X.; Chen, C. T-Loam: Truncated Least Squares Lidar-Only Odometry and Mapping in Real Time. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  17. Xu, W.; Zhang, F. Fast-Lio: A Fast, Robust Lidar-Inertial Odometry Package by Tightly-Coupled Iterated Kalman Filter. IEEE Robot. Autom. Lett. 2021, 6, 3317–3324. [Google Scholar] [CrossRef]
  18. Xu, W.; Cai, Y.; He, D.; Lin, J.; Zhang, F. Fast-Lio2: Fast Direct Lidad-Inertial Odometry. IEEE Trans. Robot. 2022, 38, 2053–2073. [Google Scholar] [CrossRef]
  19. Lee, D.; Jung, M.; Yang, W.; Kim, A. Lidar Odometry Survey: Recent Advancements and Remaining Challenges. Intell. Serv. Robot. 2024, 17, 95–118. [Google Scholar] [CrossRef]
  20. Chen, X.; Milioto, A.; Palazzolo, E.; Giguere, P.; Behley, J.; Stachniss, C. Suma++: Efficient Lidar-Based Semantic Slam. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019. [Google Scholar]
  21. Fu, H.; Yu, R.; Ye, L.; Wu, T.; Xu, X. An Efficient Scan-to-Map Matching Approach Based on Multi-Channel Lidar. J. Intell. Robot. Syst. 2018, 91, 501–513. [Google Scholar] [CrossRef]
  22. Ma, X.; Song, C.; Ji, Y.; Zhong, S. Related Keyframe Optimization Gaussian–Simultaneous Localization and Mapping: A 3d Gaussian Splatting-Based Simultaneous Localization and Mapping with Related Keyframe Optimization. Appl. Sci. 2025, 15, 1320. [Google Scholar] [CrossRef]
  23. Rusu, R.B.; Bradski, G.; Thibaux, R.; Hsu, J. Fast 3d Recognition and Pose Using the Viewpoint Feature Histogram. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010. [Google Scholar]
  24. Qin, C.; Ye, H.; Pranata, C.E.; Han, J.; Zhang, S.; Liu, M. Lins: A Lidar-Inertial State Estimator for Robust and Efficient Navigation. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020. [Google Scholar]
  25. Xie, G.; Zong, Q.; Zhang, X.; Tian, B. Loosely-Coupled Lidar-Inertial Odometry and Mapping in Real Time. Int. J. Intell. Robot. Appl. 2021, 5, 119–129. [Google Scholar] [CrossRef]
  26. Hsu, L.-T.; Kubo, N.; Wen, W.; Chen, W.; Liu, Z.; Suzuki, T.; Meguro, J. Urbannav: An Open-Sourced Multisensory Dataset for Benchmarking Positioning Algorithms Designed for Urban Areas. In Proceedings of the 34th International Technical Meeting of the Satellite Division of the Institute of Navigation (ION GNSS+ 2021), Online, 20–24 September 2021. [Google Scholar]
  27. Geiger, A.; Lenz, P.; Urtasun, R. Are We Ready for Autonomous Driving? The Kitti Vision Benchmark Suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  28. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision Meets Robotics: The Kitti Dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
Figure 1. Workflow of the proposed lightweight feature extraction method.
Figure 1. Workflow of the proposed lightweight feature extraction method.
Computation 13 00239 g001
Figure 2. Experimental platform and sensor configuration for data collection [26].
Figure 2. Experimental platform and sensor configuration for data collection [26].
Computation 13 00239 g002
Figure 3. The scene and trajectory diagrams of the UrbanNav-HK-Data20200314 sequence in the UrbanNav dataset: (a) Total points, (b) Point A, (c) Point B, (d) Point C, (e) Point D, (f) Point E, (g) Point F, (h) Point G, (i) Point H.
Figure 3. The scene and trajectory diagrams of the UrbanNav-HK-Data20200314 sequence in the UrbanNav dataset: (a) Total points, (b) Point A, (c) Point B, (d) Point C, (e) Point D, (f) Point E, (g) Point F, (h) Point G, (i) Point H.
Computation 13 00239 g003aComputation 13 00239 g003b
Figure 4. The comparison results of (a) corner point, (b) surf point, (c) memory usage and (d) keyframe count.
Figure 4. The comparison results of (a) corner point, (b) surf point, (c) memory usage and (d) keyframe count.
Computation 13 00239 g004
Figure 5. The point cloud maps and some details of (a) Energy-LIO, (c) local enlarged views of Energy-LIO, (e) local enlarged views of Energy-LIO and (b) LIO-SAM, (d) local enlarged views of LIO-SAM, (f) local enlarged views of LIO-SAM.
Figure 5. The point cloud maps and some details of (a) Energy-LIO, (c) local enlarged views of Energy-LIO, (e) local enlarged views of Energy-LIO and (b) LIO-SAM, (d) local enlarged views of LIO-SAM, (f) local enlarged views of LIO-SAM.
Computation 13 00239 g005aComputation 13 00239 g005b
Figure 6. Experimental equipment used in KITTI dataset collection [27].
Figure 6. Experimental equipment used in KITTI dataset collection [27].
Computation 13 00239 g006
Figure 7. Trajectory comparisons between LIO-SAM, Energy-LIO, and ground truth across different KITTI sequences. Subfigures (af) correspond to the trajectory results of KITTI sequences 01, 02, 04, 05, 06, and 07, respectively.
Figure 7. Trajectory comparisons between LIO-SAM, Energy-LIO, and ground truth across different KITTI sequences. Subfigures (af) correspond to the trajectory results of KITTI sequences 01, 02, 04, 05, 06, and 07, respectively.
Computation 13 00239 g007
Table 1. Pseudo Code.
Table 1. Pseudo Code.
Algorithm 1: Occlusion-Aware Curvature for Robust Feature Extraction
Input: Deskewed Point Cloud P , with per-point range r k and column index.
1: Deskewed cloud P Project to 2D image M r a n g e ;
2: Initialize curvature array c for all points in P ;
3: for each point p i , j P do
4:  Initialize valid neighborhood;
5:  for each neighbor p u , v of p i , j in M r a n g e do
6:    if | R ( u , v ) R ( i , j ) | α R ( i , j ) ;
7:       S i , j S i , j p u , v ;
8:    else
9:      Marked as invalid point;
10:   end if
11: end for
12:   if S i , j is sufficient for a stable calculation;
13:       c p i , j Calculate Curvature S i , j , p i , j ;
14:   else
15:      Mark as unreliable if neighborhood is ill-defined;
16:   end if
17: end for
18: Mark unstable occluded points using M r a n g e ;
19: Standard procedure using LOAM method to calculate curvatures c p i , j ;
20: Extract corner points and surface points based on curvature;
21: Voxel filtering;
Output: Corner Feature Set F c , Surface Feature Set F s .
Table 2. Quantitative comparison between LIO-SAM and Energy-LIO on different KITTI sequences.
Table 2. Quantitative comparison between LIO-SAM and Energy-LIO on different KITTI sequences.
Seq010204
STDMEANRMSESTDMEANRMSESTDMEANRMSE
LIO-SAM9.58084120.1668622.3273.9512769.93337210.690390.1359320.2747230.306513
Energy-LIO8.9070918.9091520.901973.2185878.5630339.1479420.1123850.3020990.322326
Seq050607
STDMEANRMSESTDMEANRMSESTDMEANRMSE
LIO-SAM1.0121032.0876312.3200335.19737613.3998714.372520.4485810.7969570.91453
Energy-LIO0.791291.7896251.9567575.18363613.3970314.36490.4041930.7503340.852275
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, C.; Li, T.; Hu, X. Energy-Conscious Lightweight LiDAR SLAM with 2D Range Projection and Multi-Stage Outlier Filtering for Intelligent Driving. Computation 2025, 13, 239. https://doi.org/10.3390/computation13100239

AMA Style

Wei C, Li T, Hu X. Energy-Conscious Lightweight LiDAR SLAM with 2D Range Projection and Multi-Stage Outlier Filtering for Intelligent Driving. Computation. 2025; 13(10):239. https://doi.org/10.3390/computation13100239

Chicago/Turabian Style

Wei, Chun, Tianjing Li, and Xuemin Hu. 2025. "Energy-Conscious Lightweight LiDAR SLAM with 2D Range Projection and Multi-Stage Outlier Filtering for Intelligent Driving" Computation 13, no. 10: 239. https://doi.org/10.3390/computation13100239

APA Style

Wei, C., Li, T., & Hu, X. (2025). Energy-Conscious Lightweight LiDAR SLAM with 2D Range Projection and Multi-Stage Outlier Filtering for Intelligent Driving. Computation, 13(10), 239. https://doi.org/10.3390/computation13100239

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop