Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,547)

Search Parameters:
Keywords = 3D-LiDAR

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 7416 KB  
Article
LiDAR SLAM for Safety Inspection Robots in Large Scale Public Building Construction Sites
by Chunyong Feng, Junqi Yu, Jingdan Li, Yonghua Wu, Ben Wang and Kaiwen Wang
Buildings 2025, 15(19), 3602; https://doi.org/10.3390/buildings15193602 - 8 Oct 2025
Abstract
LiDAR-based Simultaneous Localization and Mapping (SLAM) plays a key role in enabling inspection robots to achieve autonomous navigation. However, at installation construction sites of large-scale public buildings, existing methods often suffer from point-cloud drift, large z-axis errors, and inefficient loop closure detection, [...] Read more.
LiDAR-based Simultaneous Localization and Mapping (SLAM) plays a key role in enabling inspection robots to achieve autonomous navigation. However, at installation construction sites of large-scale public buildings, existing methods often suffer from point-cloud drift, large z-axis errors, and inefficient loop closure detection, limiting their robustness and adaptability in complex environments. To address these issues, this paper proposes an improved algorithm, LeGO-LOAM-LPB (Large-scale Public Building), built upon the LeGO-LOAM framework. The method enhances feature quality through point-cloud preprocessing, stabilizes z-axis pose estimation by introducing ground-residual constraints, improves matching efficiency with an incremental k-d tree, and strengthens map consistency via a two-layer loop closure detection mechanism. Experiments conducted on a self-developed inspection robot platform in both simulated and real construction sites of large-scale public buildings demonstrate that LeGO-LOAM-LPB significantly improves positioning accuracy, reducing the root mean square error by 41.55% compared with the original algorithm. The results indicate that the proposed method offers a more precise and robust SLAM solution for safety inspection robots in construction environments and shows strong potential for engineering applications. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

24 pages, 8595 KB  
Article
Integrated Geomatic Approaches for the 3D Documentation and Analysis of the Church of Saint Andrew in Orani, Sardinia
by Giuseppina Vacca and Enrica Vecchi
Remote Sens. 2025, 17(19), 3376; https://doi.org/10.3390/rs17193376 - 7 Oct 2025
Viewed by 44
Abstract
Documenting cultural heritage sites through 3D reconstruction is crucial and can be accomplished using various geomatic techniques, such as Terrestrial Laser Scanners (TLS), Close-Range Photogrammetry (CRP), and UAV photogrammetry. Each method comes with different levels of complexity, accuracy, field times, post-processing requirements, and [...] Read more.
Documenting cultural heritage sites through 3D reconstruction is crucial and can be accomplished using various geomatic techniques, such as Terrestrial Laser Scanners (TLS), Close-Range Photogrammetry (CRP), and UAV photogrammetry. Each method comes with different levels of complexity, accuracy, field times, post-processing requirements, and costs, making them suitable for different types of restitutions. Recently, research has increasingly focused on user-friendly and faster techniques, while also considering the cost–benefit balance between accuracy, times, and costs. In this scenario, photogrammetry using images captured with 360-degree cameras and LiDAR sensors integrated into Apple devices have gained significant popularity. This study proposes the application of various techniques for the geometric reconstruction of a complex cultural heritage site, the Church of Saint Andrew in Orani, Sardinia. Datasets acquired from different geomatic techniques have been evaluated in terms of quality and usability for documenting various aspects of the site. The TLS provided an accurate model of both the interior and exterior of the church, serving as the ground truth for the validation process. UAV photogrammetry offered a broader view of the exterior, while panoramic photogrammetry from 360° camera was applied to survey the bell tower’s interior. Additionally, CRP and Apple LiDAR were compared in the context of a detailed survey. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

28 pages, 879 KB  
Article
Performance Bounds of Ranging Precision in SPAD-Based dToF LiDAR
by Hao Wu, Yingyu Wang, Shiyi Sun, Lijie Zhao, Limin Tong, Linjie Shen and Jiang Zhu
Sensors 2025, 25(19), 6184; https://doi.org/10.3390/s25196184 - 6 Oct 2025
Viewed by 168
Abstract
LiDAR with direct time-of-flight (dToF) technology based on single-photon avalanche diode detectors (SPADs) has been widely adopted in various applications. However, a comprehensive theoretical understanding of its fundamental ranging performance bounds—particularly the degradation caused by pile-up effects due to system dead time and [...] Read more.
LiDAR with direct time-of-flight (dToF) technology based on single-photon avalanche diode detectors (SPADs) has been widely adopted in various applications. However, a comprehensive theoretical understanding of its fundamental ranging performance bounds—particularly the degradation caused by pile-up effects due to system dead time and the potential benefits of photon-number-resolving detectors—remains incomplete and has not been systematically established in prior work. In this work, we present the first theoretical derivation of the Cramér–Rao lower bound (CRLB) for dToF systems explicitly accounting for dead time effects, generalize the analysis to SPADs with photon-number-resolving capabilities, and further validate the results through Monte Carlo simulations and maximum likelihood estimation. Our analysis reveals that pile-up not only reduces the information contained within individual ToF but also introduces a previously overlooked statistical coupling between distance and photon flux rate, further degrading ranging precision. The derived CRLB enables the determination of the optimal optical photon flux, laser pulse width (with FWHM of 0.56τ), and ToF quantization resolution that yield the best achievable ranging precision, showing that an optimal precision of approximately 0.53τ/N remains theoretically achievable, where τ is TDC resolution and N is the number of laser pulses. The analysis further quantifies the limited performance improvement enabled by increased photon-number resolution, which exhibits rapidly diminishing returns. Overall, these findings establish a unified theoretical framework for understanding the fundamental limits of SPAD-based dToF LiDAR, filling a gap left by earlier studies and providing concrete design guidelines for the selection of optimal operating points. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

23 pages, 5434 KB  
Article
Deep Reinforcement Learning for Sim-to-Real Robot Navigation with a Minimal Sensor Suite for Beach-Cleaning Applications
by Guillermo Cid Ampuero, Gabriel Hermosilla, Germán Varas and Matías Toribio Clark
Appl. Sci. 2025, 15(19), 10719; https://doi.org/10.3390/app151910719 - 5 Oct 2025
Viewed by 296
Abstract
Autonomous beach-cleaning robots require reliable, low-cost navigation on sand. We study Sim-to-Real transfer of deep reinforcement learning (DRL) policies using a minimal sensor suite—wheel-encoder odometry and a single 2-D LiDAR—on a 30 kg differential-drive platform (Raspberry Pi 4). Two policies, Proximal Policy Optimization [...] Read more.
Autonomous beach-cleaning robots require reliable, low-cost navigation on sand. We study Sim-to-Real transfer of deep reinforcement learning (DRL) policies using a minimal sensor suite—wheel-encoder odometry and a single 2-D LiDAR—on a 30 kg differential-drive platform (Raspberry Pi 4). Two policies, Proximal Policy Optimization (PPO) and a masked-action variant (PPO-Mask), were trained in Gazebo + Gymnasium and deployed on the physical robot without hyperparameter retuning. Field trials on firm sand and on a natural loose-sand beach show that PPO-Mask reduces tracking error versus PPO on firm ground (16.6% ISE reduction; 5.2% IAE reduction) and executes multi-waypoint paths faster (square path: 112.48 s vs. 103.46 s). On beach sand, all waypoints were reached within a 1 m tolerance, with mission times of 115.72 s (square) and 81.77 s (triangle). These results indicate that DRL-based navigation with minimal sensing and low-cost compute is feasible in beach settings. Full article
Show Figures

Figure 1

20 pages, 2812 KB  
Article
Seven Decades of River Change: Sediment Dynamics in the Diable River, Quebec
by Ali Faghfouri, Daniel Germain and Guillaume Fortin
Geosciences 2025, 15(10), 388; https://doi.org/10.3390/geosciences15100388 - 4 Oct 2025
Viewed by 139
Abstract
This study reconstructs seven decades (1949–2019) of morphodynamic changes and sediment dynamics in the Diable River (Québec, Canada) using nine series of aerial photographs, a high-resolution LiDAR Digital Elevation Model (2021), and grain-size analysis. The objectives were to document long-term river evolution, quantify [...] Read more.
This study reconstructs seven decades (1949–2019) of morphodynamic changes and sediment dynamics in the Diable River (Québec, Canada) using nine series of aerial photographs, a high-resolution LiDAR Digital Elevation Model (2021), and grain-size analysis. The objectives were to document long-term river evolution, quantify erosion and deposition, and evaluate sediment connectivity between eroding sandy bluffs and depositional zones. Planform analysis and sediment budgets derived from DEMs of Difference (DoD) reveal an oscillatory trajectory characterized by alternating phases of sediment export and temporary stabilization, rather than a simple trend of degradation or aggradation. The most dynamic interval (1980–2001) was marked by widespread meander migration and the largest net export (−142.5 m3/km/year), whereas the 2001–2007 interval showed net storage (+70.8 m3/km/year) and short-term geomorphic recovery. More recent floods (2017, 2019; 20–50-year return periods) induced localized but persistent sediment loss, underlining the structuring role of extreme events. Grain-size results indicate partial connectivity: coarse fractions tend to remain in local depositional features, while finer sediments are preferentially exported downstream. These findings emphasize the geomorphic value of temporary sediment sinks (bars, beaches) and highlight the need for adaptive river management strategies that integrate sediment budgets and local knowledge into floodplain governance. Full article
18 pages, 14342 KB  
Article
A Multi-LiDAR Self-Calibration System Based on Natural Environments and Motion Constraints
by Yuxuan Tang, Jie Hu, Zhiyong Yang, Wencai Xu, Shuaidi He and Bolun Hu
Mathematics 2025, 13(19), 3181; https://doi.org/10.3390/math13193181 - 4 Oct 2025
Viewed by 181
Abstract
Autonomous commercial vehicles often mount multiple LiDARs to enlarge their field of view, but conventional calibration is labor-intensive and prone to drift during long-term operation. We present an online self-calibration method that combines a ground plane motion constraint with a virtual RGB–D projection, [...] Read more.
Autonomous commercial vehicles often mount multiple LiDARs to enlarge their field of view, but conventional calibration is labor-intensive and prone to drift during long-term operation. We present an online self-calibration method that combines a ground plane motion constraint with a virtual RGB–D projection, mapping 3D point clouds to 2D feature/depth images to reduce feature extraction cost while preserving 3D structure. Motion consistency across consecutive frames enables a reduced-dimension hand–eye formulation. Within this formulation, the estimation integrates geometric constraints on SE(3) using Lagrange multiplier aggregation and quasi-Newton refinement. This approach highlights key aspects of identifiability, conditioning, and convergence. An online monitor evaluates plane alignment and LiDAR–INS odometry consistency to detect degradation and trigger recalibration. Tests on a commercial vehicle with six LiDARs and on nuScenes demonstrate accuracy comparable to offline, target-based methods while supporting practical online use. On the vehicle, maximum errors are 6.058 cm (translation) and 4.768° (rotation); on nuScenes, 2.916 cm and 5.386°. The approach streamlines calibration, enables online monitoring, and remains robust in real-world settings. Full article
(This article belongs to the Section A: Algebra and Logic)
Show Figures

Figure 1

17 pages, 1322 KB  
Article
Robust 3D Object Detection in Complex Traffic via Unified Feature Alignment in Bird’s Eye View
by Ajian Liu, Yandi Zhang, Huichao Shi and Juan Chen
World Electr. Veh. J. 2025, 16(10), 567; https://doi.org/10.3390/wevj16100567 - 2 Oct 2025
Viewed by 145
Abstract
Reliable three-dimensional (3D) object detection is critical for intelligent vehicles to ensure safety in complex traffic environments, and recent progress in multi-modal sensor fusion, particularly between LiDAR and camera, has advanced environment perception in urban driving. However, existing approaches remain vulnerable to occlusions [...] Read more.
Reliable three-dimensional (3D) object detection is critical for intelligent vehicles to ensure safety in complex traffic environments, and recent progress in multi-modal sensor fusion, particularly between LiDAR and camera, has advanced environment perception in urban driving. However, existing approaches remain vulnerable to occlusions and dense traffic, where depth estimation errors, calibration deviations, and cross-modal misalignment are often exacerbated. To overcome these limitations, we propose BEVAlign, a local–global feature alignment framework designed to generate unified BEV representations from heterogeneous sensor modalities. The framework incorporates a Local Alignment (LA) module that enhances camera-to-BEV view transformation through graph-based neighbor modeling and dual-depth encoding, mitigating local misalignment from depth estimation errors. To further address global misalignment in BEV representations, we present the Global Alignment (GA) module comprising a bidirectional deformable cross-attention (BDCA) mechanism and CBR blocks. BDCA employs dual queries from LiDAR and camera to jointly predict spatial sampling offsets and aggregate features, enabling bidirectional alignment within the BEV domain. The stacked CBR blocks then refine and integrate the aligned features into unified BEV representations. Experiment on the nuScenes benchmark highlights the effectiveness of BEVAlign, which achieves 71.7% mAP, outperforming BEVFusion by 1.5%. Notably, it achieves strong performance on small and occluded objects, particularly in dense traffic scenarios. These findings provide a basis for advancing cooperative environment perception in next-generation intelligent vehicle systems. Full article
(This article belongs to the Special Issue Recent Advances in Intelligent Vehicle)
18 pages, 11220 KB  
Article
LM3D: Lightweight Multimodal 3D Object Detection with an Efficient Fusion Module and Encoders
by Yuto Sakai, Tomoyasu Shimada, Xiangbo Kong and Hiroyuki Tomiyama
Appl. Sci. 2025, 15(19), 10676; https://doi.org/10.3390/app151910676 - 2 Oct 2025
Viewed by 317
Abstract
In recent years, the demand for both high accuracy and real-time performance in 3D object detection has increased alongside the advancement of autonomous driving technology. While multimodal methods that integrate LiDAR and camera data have demonstrated high accuracy, these methods often have high [...] Read more.
In recent years, the demand for both high accuracy and real-time performance in 3D object detection has increased alongside the advancement of autonomous driving technology. While multimodal methods that integrate LiDAR and camera data have demonstrated high accuracy, these methods often have high computational costs and latency. To address these issues, we propose an efficient 3D object detection network that integrates three key components: a DepthWise Lightweight Encoder (DWLE) module for efficient feature extraction, an Efficient LiDAR Image Fusion (ELIF) module that combines channel attention with cross-modal feature interaction, and a Mixture of CNN and Point Transformer (MCPT) module for capturing rich spatial contextual information. Experimental results on the KITTI dataset demonstrate that our proposed method outperforms existing approaches by achieving approximately 0.6% higher 3D mAP, 7.6% faster inference speed, and 17.0% fewer parameters. These results highlight the effectiveness of our approach in balancing accuracy, speed, and model size, making it a promising solution for real-time applications in autonomous driving. Full article
Show Figures

Figure 1

24 pages, 4942 KB  
Article
ConvNet-Generated Adversarial Perturbations for Evaluating 3D Object Detection Robustness
by Temesgen Mikael Abraha, John Brandon Graham-Knight, Patricia Lasserre, Homayoun Najjaran and Yves Lucet
Sensors 2025, 25(19), 6026; https://doi.org/10.3390/s25196026 - 1 Oct 2025
Viewed by 197
Abstract
This paper presents a novel adversarial Convolutional Neural Network (ConvNet) method for generating adversarial perturbations in 3D point clouds, enabling gradient-free robustness evaluation of object detection systems at inference time. Unlike existing iterative gradient methods, our approach embeds the ConvNet directly into the [...] Read more.
This paper presents a novel adversarial Convolutional Neural Network (ConvNet) method for generating adversarial perturbations in 3D point clouds, enabling gradient-free robustness evaluation of object detection systems at inference time. Unlike existing iterative gradient methods, our approach embeds the ConvNet directly into the detection pipeline at the voxel feature level. The ConvNet is trained to maximize detection loss while maintaining perturbations within sensor error bounds through multi-component loss constraints (intensity, bias, and imbalance terms). Evaluation on a Sparsely Embedded Convolutional Detection (SECOND) detector with the KITTI dataset shows 8% overall mean Average Precision (mAP) degradation, while CenterPoint on NuScenes exhibits 24% weighted mAP reduction across 10 object classes. Analysis reveals an inverse relationship between object size and adversarial vulnerability: smaller objects (pedestrians: 13%, cyclists: 14%) show higher vulnerability compared to larger vehicles (cars: 0.2%) on KITTI, with similar patterns on NuScenes, where barriers (68%) and pedestrians (32%) are most affected. Despite perturbations remaining within typical sensor error margins (mean L2 norm of 0.09% for KITTI, 0.05% for NuScenes, corresponding to 0.9–2.6 cm at typical urban distances), substantial detection failures occur. The key novelty is training a ConvNet to learn effective adversarial perturbations during a one-time training phase and then using the trained network for gradient-free robustness evaluation during inference, requiring only a forward pass through the ConvNet (1.2–2.0 ms overhead) instead of iterative gradient computation, making continuous vulnerability monitoring practical for autonomous driving safety assessment. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

14 pages, 15260 KB  
Article
High-Performance 3D Point Cloud Image Distortion Calibration Filter Based on Decision Tree
by Yao Duan
Photonics 2025, 12(10), 960; https://doi.org/10.3390/photonics12100960 - 28 Sep 2025
Viewed by 213
Abstract
Structured Light LiDAR is susceptible to lens scattering and temperature fluctuations, resulting in some level of distortion in the captured point cloud image. To address this problem, this paper proposes a high-performance 3D point cloud Least Mean Square filter based on Decision Tree, [...] Read more.
Structured Light LiDAR is susceptible to lens scattering and temperature fluctuations, resulting in some level of distortion in the captured point cloud image. To address this problem, this paper proposes a high-performance 3D point cloud Least Mean Square filter based on Decision Tree, which is called the D−LMS filter for short. The D−LMS filter is an adaptive filtering compensation algorithm based on decision tree, which can effectively distinguish the signal region from the distorted region, thus optimizing the distortion of the point cloud image and improving the accuracy of the point cloud image. The experimental results clearly demonstrate that our proposed D−LMS filtering algorithm significantly improves accuracy by optimizing distorted areas. Compared with the 3D point cloud least mean square filter based on SVM, the accuracy of the proposed D−LMS filtering algorithm is improved from 86.17% to 92.38%, the training time is reduced by 1317 times and the testing time is reduced by 1208 times. Full article
Show Figures

Figure 1

26 pages, 3429 KB  
Article
I-VoxICP: A Fast Point Cloud Registration Method for Unmanned Surface Vessels
by Qianfeng Jing, Mingwang Bai, Yong Yin and Dongdong Guo
J. Mar. Sci. Eng. 2025, 13(10), 1854; https://doi.org/10.3390/jmse13101854 - 25 Sep 2025
Viewed by 333
Abstract
The accurate positioning and state estimation of surface vessels are prerequisites to autonomous navigation. Recently, the rapid development of 3D LiDARs has promoted the autonomy of both land and aerial vehicles, which has attracted the interest of researchers in the maritime community. However, [...] Read more.
The accurate positioning and state estimation of surface vessels are prerequisites to autonomous navigation. Recently, the rapid development of 3D LiDARs has promoted the autonomy of both land and aerial vehicles, which has attracted the interest of researchers in the maritime community. However, in traditional maritime surface multi-scenario applications, LiDAR scan matching has low point cloud scanning and matching efficiency and insufficient positional accuracy when dealing with large-scale point clouds, so it has difficulty meeting the real-time demand of low-computing-power platforms. In this paper, we use ICP-SVD for point cloud alignment in the Stanford dataset and outdoor dock scenarios and propose an optimization scheme (iVox + ICP-SVD) that incorporates the voxel structure iVox. Experiments show that the average search time of iVox is 72.23% and 96.8% higher than that of ikd-tree and kd-tree, respectively. Executed on an NVIDIA Jetson Nano (four ARM Cortex-A57 cores @ 1.43 GHz) the algorithm processes 18 k downsampled points in 56 ms on average and 65 ms in the worst case—i.e., ≤15 Hz—so every scan is completed before the next 10–20 Hz LiDAR sweep arrives. During a 73 min continuous harbor trial the CPU temperature stabilized at 68 °C without thermal throttling, confirming that the reported latency is a sustainable, field-proven upper bound rather than a laboratory best case. This dramatically improves the retrieval efficiency while effectively maintaining the matching accuracy. As a result, the overall alignment process is significantly accelerated, providing an efficient and reliable solution for real-time point cloud processing. Full article
Show Figures

Figure 1

32 pages, 33744 KB  
Article
Attention-Based Enhancement of Airborne LiDAR Across Vegetated Landscapes Using SAR and Optical Imagery Fusion
by Michael Marks, Daniel Sousa and Janet Franklin
Remote Sens. 2025, 17(19), 3278; https://doi.org/10.3390/rs17193278 - 24 Sep 2025
Viewed by 425
Abstract
Accurate and timely 3D vegetation structure information is essential for ecological modeling and land management. However, these needs often cannot be met with existing airborne LiDAR surveys, whose broad-area coverage comes with trade-offs in point density and update frequency. To address these limitations, [...] Read more.
Accurate and timely 3D vegetation structure information is essential for ecological modeling and land management. However, these needs often cannot be met with existing airborne LiDAR surveys, whose broad-area coverage comes with trade-offs in point density and update frequency. To address these limitations, this study introduces a deep learning framework built on attention mechanisms, the fundamental building block of modern large language models. The framework upsamples sparse (<22 pt/m2) airborne LiDAR point clouds by fusing them with stacks of multi-temporal optical (NAIP) and L-band quad-polarized Synthetic Aperture Radar (UAVSAR) imagery. Utilizing a novel Local–Global Point Attention Block (LG-PAB), our model directly enhances 3D point-cloud density and accuracy in vegetated landscapes by learning structure directly from the point cloud itself. Results in fire-prone Southern California foothill and montane ecosystems demonstrate that fusing both optical and radar imagery reduces reconstruction error (measured by Chamfer distance) compared to using LiDAR alone or with a single image modality. Notably, the fused model substantially mitigates errors arising from vegetation changes over time, particularly in areas of canopy loss, thereby increasing the utility of historical LiDAR archives. This research presents a novel approach for direct 3D point-cloud enhancement, moving beyond traditional raster-based methods and offering a pathway to more accurate and up-to-date vegetation structure assessments. Full article
Show Figures

Graphical abstract

23 pages, 14935 KB  
Article
Robust Pedestrian Detection and Intrusion Judgment in Coal Yard Hazard Areas via 3D LiDAR-Based Deep Learning
by Anxin Zhao, Yekai Zhao and Qiuhong Zheng
Sensors 2025, 25(18), 5908; https://doi.org/10.3390/s25185908 - 21 Sep 2025
Viewed by 418
Abstract
Pedestrian intrusion in coal yard work areas is a major cause of accidents, posing challenges for the safe supervision of coal yards. Existing visual detection methods suffer under poor lighting and a lack of 3D data. To overcome these limitations, this study introduces [...] Read more.
Pedestrian intrusion in coal yard work areas is a major cause of accidents, posing challenges for the safe supervision of coal yards. Existing visual detection methods suffer under poor lighting and a lack of 3D data. To overcome these limitations, this study introduces a robust pedestrian intrusion detection method based on 3D LiDAR. Our approach consists of three main components. First, we propose a novel pedestrian detection network called EFT-RCNN. Based on Voxel-RCNN, this network introduces an EnhancedVFE module to improve spatial feature extraction, employs FocalConv to reconstruct the 3D backbone network for enhanced foreground–background distinction, and utilizes TeBEVPooling to optimize bird’s eye view (BEV) generation. Second, a precise 3D hazardous area is defined by combining a polygonal base surface, determined through on-site exploration, with height constraints. Finally, a point–region hierarchical judgment method is designed to calculate the spatial relationship between pedestrians and the hazardous area for graded warning. When evaluated on the public KITTI dataset, the EFT-RCNN network improved the average precision for pedestrian detection by 4.39% in 3D and 4.68% in BEV compared with the baseline, while maintaining a real-time processing speed of 28.56 FPS. In practical tests, the pedestrian detection accuracy reached 92.9%, with an average error in distance measurement of 0.054 m. The experimental results demonstrate that the proposed method effectively mitigates complex environmental interference, enables robust detection, and provides a reliable means for the proactive prevention of pedestrian intrusion accidents. Full article
Show Figures

Figure 1

20 pages, 5335 KB  
Article
LiGaussOcc: Fully Self-Supervised 3D Semantic Occupancy Prediction from LiDAR via Gaussian Splatting
by Zhiqiang Wei, Tao Huang and Fengdeng Zhang
Sensors 2025, 25(18), 5889; https://doi.org/10.3390/s25185889 - 20 Sep 2025
Viewed by 481
Abstract
Accurate 3D semantic occupancy perception is critical for autonomous driving, enabling robust navigation in unstructured environments. While vision-based methods suffer from depth inaccuracies and lighting sensitivity, LiDAR-based approaches face challenges due to sparse data and dependence on expensive manual annotations. This work proposes [...] Read more.
Accurate 3D semantic occupancy perception is critical for autonomous driving, enabling robust navigation in unstructured environments. While vision-based methods suffer from depth inaccuracies and lighting sensitivity, LiDAR-based approaches face challenges due to sparse data and dependence on expensive manual annotations. This work proposes LiGaussOcc, a novel self-supervised framework for dense LiDAR-based 3D semantic occupancy prediction. Our method first encodes LiDAR point clouds into voxel features and addresses sparsity via an Empty Voxel Inpainting (EVI) module, refined by an Adaptive Feature Fusion (AFF) module. During training, a Gaussian Primitive from Voxels (GPV) module generates parameters for 3D Gaussian Splatting, enabling efficient rendering of 2D depth and semantic maps. Supervision is achieved through photometric consistency across adjacent camera views and pseudo-labels from vision–language models, eliminating manual 3D annotations. Evaluated on the nuScenes-OpenOccupancy benchmark, LiGaussOcc achieved performance competitive with 30.4% Intersection over Union (IoU) and 14.1% mean Intersection over Union (mIoU). It reached 91.6% of the performance of the fully supervised LiDAR-based L-CONet, while completely eliminating the need for costly and labor-intensive manual 3D annotations. It excelled particularly in static environmental classes, such as drivable surfaces and man-made structures. This work presents a scalable, annotation-free solution for LiDAR-based 3D semantic occupancy perception. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

43 pages, 1954 KB  
Review
Review of Uneven Road Surface Information Perception Methods for Suspension Preview Control
by Yujie Shen, Kai Jing, Kecheng Sun, Changning Liu, Yi Yang and Yanling Liu
Sensors 2025, 25(18), 5884; https://doi.org/10.3390/s25185884 - 19 Sep 2025
Viewed by 604
Abstract
Accurate detection of road surface information is crucial for enhancing vehicle driving safety and ride comfort. To overcome the limitation that traditional suspension systems struggle to respond to road excitations in real time due to time delays in signal acquisition and control, suspension [...] Read more.
Accurate detection of road surface information is crucial for enhancing vehicle driving safety and ride comfort. To overcome the limitation that traditional suspension systems struggle to respond to road excitations in real time due to time delays in signal acquisition and control, suspension preview control technology has attracted significant attention for its proactive adjustment capability, with efficient road surface information perception being a critical prerequisite for its implementation. This paper systematically reviews road surface information detection technologies for suspension preview, focusing on the identification of potholes and speed bumps. Firstly, it summarizes relevant publicly available datasets. Secondly, it sorts out mainstream detection methods, including traditional dynamic methods, 2D image processing, 3D point cloud analysis, machine/deep learning methods, and multi-sensor fusion methods, while comparing their applicable scenarios and evaluation metrics. Furthermore, it emphasizes the core role of elevation information (e.g., pothole depth, speed bump height) in suspension preview control and summarizes elevation reconstruction technologies based on LiDAR, stereo vision, and multi-modal fusion. Finally, it prospects future research directions such as optimizing robustness, improving real-time performance, and reducing labeling costs. This review provides technical references for enhancing the accuracy of road surface information detection and the control efficiency of suspension preview systems, and it is of great significance for promoting the development of intelligent chassis. Full article
Show Figures

Figure 1

Back to TopTop