Next Article in Journal
Real-Time Fluorescence-Based COVID-19 Diagnosis Using a Lightweight Deep Learning System
Next Article in Special Issue
FGO-PMB: A Factor Graph Optimized Poisson Multi-Bernoulli Filter for Accurate Online 3D Multi-Object Tracking
Previous Article in Journal
HCNet: Multi-Exposure High-Dynamic-Range Reconstruction Network for Coded Aperture Snapshot Spectral Imaging
Previous Article in Special Issue
Evaluating LiDAR Perception Algorithms for All-Weather Autonomy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HFSA-Net: A 3D Object Detection Network with Structural Encoding and Attention Enhancement for LiDAR Point Clouds

1
School of Transportation and Vehicle Engineering, Shandong University of Technology, Zibo 255000, China
2
Shuntai Automobile Co., Ltd., Zibo 255400, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 338; https://doi.org/10.3390/s26010338
Submission received: 7 November 2025 / Revised: 1 December 2025 / Accepted: 4 December 2025 / Published: 5 January 2026
(This article belongs to the Special Issue Recent Advances in LiDAR Sensing Technology for Autonomous Vehicles)

Abstract

The inherent sparsity of LiDAR point cloud data presents a fundamental challenge for 3D object detection. During the feature encoding stage, especially in voxelization, existing methods find it difficult to effectively retain the critical geometric structural information contained in these sparse point clouds, resulting in decreased detection performance. To address this problem, this paper proposes an enhanced 3D object detection framework. It first designs a Structured Voxel Feature Encoder that significantly enhances the initial feature representation through intra-voxel feature refinement and multi-scale neighborhood context aggregation. Second, it constructs a Hybrid-Domain Attention-Guided Sparse Backbone, which introduces a decoupled hybrid attention mechanism and a hierarchical integration strategy to realize dynamic weighting and focusing on key semantic and geometric features. Finally, a Scale-Aggregation Head is proposed to improve the model’s perception and localization capabilities for different-sized objects via multi-level feature pyramid fusion and cross-layer information interaction. Experimental results on the KITTI dataset show that the proposed algorithm increases the mean Average Precision (mAP) by 3.34% compared to the baseline model. Moreover, experiments on a vehicle platform with a lower-resolution LiDAR verify the effectiveness of the proposed method in improving 3D detection accuracy and its generalization ability.

1. Introduction

As a core technology in the environmental perception systems of autonomous vehicles and intelligent robots, 3D object detection aims to determine the precise 3D spatial position, dimensions, and category of objects within a scene. This provides a critical basis for subsequent decision-making, planning, and control [1,2]. Among the various 3D sensors available, Light Detection and Ranging (LiDAR) has become the predominant modality for developing advanced 3D object detection algorithms, owing to its capacity to directly acquire precise 3D geometric information and its excellent robustness to changes in illumination [3].
Despite the significant progress achieved in LiDAR-based detection methods [4,5,6], their performance remains constrained by the intrinsic characteristics of point cloud data [7,8]. By nature, a LiDAR point cloud is a sparse and unordered sampling of scene surfaces, which gives rise to two fundamental challenges [9,10]. The first is data sparsity: due to the discrete scanning mechanism of LiDAR sensors, the collected data is inherently non-continuous and unevenly distributed, especially for distant or small objects, which are composed of only a minimal number of points, this sparsity leads to an inadequate feature representation. The second is the fragility of geometric structural information. To enhance computational efficiency, conventional methods widely adopt voxelization to convert the point cloud into a regular grid representation. However, this process is often accompanied by the degradation of fine-grained local geometric structures, causing an irreversible loss of information.
The information loss issue, stemming from the physical characteristics of sensors and data preprocessing pipelines, constitutes a significant performance bottleneck in current LiDAR-based 3D object detection, particularly for distant and small objects. The degradation of geometric details directly leads to a reduction in detection accuracy and an increase in the miss detection rate, posing a potential threat to the safety of autonomous driving systems. Therefore, developing an algorithmic framework that can effectively preserve and enhance the structural information of point clouds is of great scientific significance and practical value for advancing the field.
To address the aforementioned challenges, this paper proposes an enhanced 3D object detection network, namely HFSA-Net (Hierarchical Focus and Structural-Aware Network). This framework, an advancement based on the CenterPoint architecture, systematically resolves the information loss problem in sparse point clouds in a synergistic manner by incorporating three specially designed modules. The main contributions of this paper are summarized as follows:
(1)
A structured voxel feature encoder is proposed, which explicitly compensates for the loss of local geometric information during voxelization by employing intra-voxel feature refinement and multi-scale neighborhood context aggregation. This is designed to improve the model’s representation capability for fine-grained structures.
(2)
A hybrid-domain attention-guided sparse backbone network is constructed. This network introduces a decoupled hybrid-domain attention mechanism that enables the network to dynamically focus on salient feature regions within sparse point clouds, thereby enhancing the effectiveness of feature extraction.
(3)
A scale-aggregated detection head is designed to enhance the model’s perception and localization capabilities for objects at varying distances and sizes. It achieves this by fusing a multi-level feature pyramid to adapt to variations in point cloud density.
Comprehensive experimental evaluations of the proposed model were conducted on the public KITTI dataset. The results demonstrate that the proposed algorithm achieves significant performance improvements across multiple metrics. Furthermore, extensive ablation studies and on-vehicle tests have verified the effectiveness of each innovative module and the practical utility of the framework.

2. Related Work

Existing 3D object detection algorithms can be primarily categorized into three main streams based on their distinct approaches to processing LiDAR point cloud data: point-based, voxel-based, and point-voxel fusion methods [11,12]. This section will review these mainstream technical routes, analyzing their respective advantages and limitations in handling point cloud information.

2.1. Point-Based Methods

Point-based methods operate directly on the raw point cloud set, aiming to fully preserve the geometric information of the scene. PointNet [13] and its successor, PointNet++ [14], pioneered this direction by designing symmetric network architectures capable of directly processing unordered point sets. These methods enable end-to-end learning on point clouds and can capture hierarchical local features. Building upon this foundation, PointRCNN [15] learns per-point features through set abstraction and other operations to generate high-quality 3D proposals.
In principle, this category of methods can retain the original information to the maximum extent, thus often achieving higher accuracy on tasks that require fine-grained geometric features. However, due to the necessity of processing large-scale point sets and performing complex neighborhood query operations, their computational complexity and memory footprint are substantial. Consequently, their inference speed is often far from meeting the real-time requirements of applications such as autonomous driving.

2.2. Voxel-Based Methods

Voxel-based methods leverage well-established Convolutional Neural Networks (CNNs) by converting the point cloud into a regular 3D voxel grid, thereby achieving a significant efficiency advantage when processing large-scale, unordered point clouds. VoxelNet [16] was a pioneering work in this direction, first proposing to partition the point cloud into voxels and perform per-voxel feature encoding. However, its reliance on dense 3D convolutions resulted in excessively high computational complexity. To address this issue, SECOND [17] innovatively introduced sparse 3D convolutions, which perform computations only on non-empty voxels, thus substantially improving processing speed. PointPillars [18] further simplified the point cloud into “pillars” along the vertical dimension and applied efficient 2D CNNs for feature extraction, achieving real-time inference performance. Building on these advancements, anchor-free detectors, exemplified by CenterPoint [19], have simplified the detection pipeline by directly predicting object centers. This approach has struck an excellent balance between accuracy and speed, establishing it as a widely adopted baseline model today.
However, the high efficiency of these methods comes at the cost of information fidelity. The feature encoding process within voxelization is, in essence, a form of lossy compression, aggregating all the geometric information of points within a local region into a single feature vector. For small objects, which are inherently represented by a sparse set of points, this loss of information is particularly severe and constitutes a major bottleneck for their detection performance.

2.3. Point-Voxel Fusion Methods

Point-voxel fusion methods endeavor to combine the advantages of the preceding two approaches. A quintessential example is PV-RCNN, which utilizes a voxel-based backbone network to efficiently generate high-quality 3D proposals [20]. Subsequently, a point-based branch network encodes the fine-grained geometric features of keypoints to refine the locations of these proposals. While this class of methods typically achieves high detection accuracy, their two-stage, complex architecture and the cross-representation feature interaction result in a substantial overall computational overhead and considerable deployment challenges.
In recent years, attention mechanisms and the Transformer architecture have also been introduced into the domain of point cloud processing [21]. For instance, Point Transformer [22] models local geometric relationships through self-attention mechanisms among points, whereas Voxel Transformer [23] captures the global context at the voxel level. Although these methods have demonstrated significant potential in feature modeling, they are often computationally expensive. Furthermore, when integrated with a voxelization front-end, they still face the persistent challenge of compensating for the initial loss of geometric information.
Most recently, deep learning paradigms for point cloud processing have witnessed significant advancements. For instance, Rehman et al. [24] provided a systematic review on the fusion of hyperspectral and LiDAR imagery, elucidating the critical transition from traditional machine learning to advanced CNNs. In a similar vein, Coglan et al. [25] demonstrated the efficacy of combining geometric feature analysis with deep learning, underscoring the importance of structural awareness. Furthermore, emerging works have focused on refining feature representations through attention mechanisms and voxel encoding. Specifically, Wang et al. [26] proposed a channel-wise attention network to dynamically prioritize informative features for 3D detection, while Naich et al. [27] introduced an intensity-aware voxel encoder to enhance robustness against environmental noise. Building upon these cutting-edge methodologies, our HFSA-Net integrates structured voxel encoding with a hybrid-domain attention mechanism to explicitly address the challenges of sparsity and scale variation.
In summary, all existing LiDAR-based 3D object detection methods negotiate a trade-off between computational efficiency and information completeness to varying degrees. Voxel-based methods, in their pursuit of real-time performance, sacrifice critical local geometric information during the preprocessing stage. Conversely, while point-based methods ensure information integrity, their prohibitive computational overhead restricts their application in real-world scenarios. The objective of this research is therefore to operate within the efficient paradigm of voxel-based frameworks, yet proactively enhance the critical information that is lost during data processing. We achieve this by developing purpose-built structured encoding and attention mechanisms, ultimately culminating in the construction of an enhanced 3D object detection network.

3. Methods

In this section, we present the detailed framework of the proposed HFSA-Net. We begin with an overview of the overall network architecture. Subsequently, we describe the three key components designed to handle point cloud sparsity: the structural feature encoder, the attention-guided backbone, and the multi-scale detection head. Finally, we introduce the loss function used for model training.

3.1. Overall Architecture

To address the challenges of existing anchor-free detectors in leveraging LiDAR point cloud structural information, focusing on key features, and perceiving multi-scale objects, we propose an enhanced 3D object detection framework: HFSA-Net (Hierarchical Focus and Structural-Aware Network). Built upon the CenterPoint architecture, this framework systematically improves the model’s overall performance by incorporating three specially designed, innovative components.
As illustrated in Figure 1, the architecture of HFSA-Net is sequentially composed of a Structured Voxel Feature Encoder (S-VFE), a Hybrid-Domain Attention-guided sparse Backbone (HDA-Backbone), and a Scale-Aggregation Head (SA-Head), which replace their corresponding modules in the original CenterPoint. The entire data processing pipeline begins with the S-VFE performing structured encoding on the raw point cloud, followed by the HDA-Backbone conducting hierarchical focused extraction on the structured features. Finally, the SA-Head performs multi-scale decoding and prediction in the 2D Bird’s-Eye View (BEV) space. The design of each component will be elaborated below.

3.2. Structured Voxel Feature Encoder

When processing LiDAR point clouds, traditional Voxel Feature Encoder (VFE) methods simplify voxelization to the mean of features of the points within each voxel, leading to an irreversible loss of information. This loss is twofold: first, micro-geometric structures, such as the intra-voxel point distribution, are obliterated; second, macro-contextual information, which is jointly constituted by adjacent voxels, is completely disregarded. To this end, we design the Structured Voxel Feature Encoder (S-VFE), which aims to re-encode this rich structural information back into the voxel features through a two-stage process.
As depicted in Figure 2, the first stage of the S-VFE is intra-voxel feature refinement. For a given voxel v with an initial mean feature f v , we first compute a refined feature f v r e f by passing it through a feature transformation network T MLP and an attention network Attn MLP . This process utilizes the Hadamard Product (⊙) to perform feature weighting:
w v = σ ( Attn MLP ( T MLP ( f v ) ) )
f v ref = T MLP ( f v ) w v
In the expressions above, T MLP and Attn MLP are both small Multi-Layer Perceptrons (MLPs), and σ denotes the Sigmoid activation function.
The second stage is multi-scale neighborhood context aggregation. In this work, we treat all non-empty voxels as nodes of an implicit graph. For each voxel v i , information from its k m nearest neighbors, N k m ( v i ) , is aggregated in parallel across M different scales. The context feature c i m at each scale is computed by a scale-specific MLP processor, P m :
c i m = P m 1 k m v j N k m ( v i ) f v j ref
In the specific implementation of HFSA-Net, the voxel resolution is configured as [0.05 m, 0.05 m, 0.1 m] to ensure sufficient granularity for small objects such as pedestrians. For the neighborhood aggregation, we employ M = 4 parallel scales with neighbor counts set to k {4, 8, 16, 32}. This multi-scale design allows the encoder to simultaneously capture fine local geometry and broad contextual information to compensate for sparsity.
After the context features from all scales, { c i m } m = 1 M , are concatenated, they are fed into a fusion network, F fuse . The final output feature, f v i out , is formed via a residual connection [28], where λ is a learnable scaling parameter:
f v i out = f v i ref + λ F fuse C o n c a t m = 1 .. M [ c i m ]
Through the S-VFE, the output feature of each voxel not only contains its own refined information but also explicitly encodes its position and relationships within the multi-scale local geometric structure.

3.3. Hybrid-Domain Attention-Guided Sparse Backbone

The inherent sparsity of a LiDAR point cloud leads to the formation of a vast number of empty voxels after voxelization. Although sparse convolutional networks can enhance efficiency by skipping computations on these empty voxels, they, by default, assign equal importance to all activated voxels. However, in real-world LiDAR perception scenarios, critical foreground objects such as distant pedestrians or small obstacles may be represented by only a few isolated, activated voxels. Standard networks struggle to effectively focus on these faint yet crucial foreground signals amidst a multitude of background voxels. To address this, we propose the Hybrid-Domain Attention-guided sparse Backbone (HDA-Backbone). The core idea is to decouple the attention mechanism into two orthogonal feature domains: the spatial domain, to locate critical regions, and the channel domain, to select key semantics.
As illustrated in Figure 3, the core of the HDA-Backbone is the Decoupled Hybrid Attention (DHA) module. For an input sparse feature tensor X , this module concurrently computes a spatial attention map, M s , generated by the Fast Coordinate Attention (FastCA) [29] mechanism, and a channel attention map, M c , produced by the lightweight Gated Channel Transformation (GCT) [30] module. Here, FastCA is selected over standard global pooling methods, such as the SE-Block, to explicitly preserve the spatial structure along coordinate axes. Meanwhile, GCT is chosen for its efficiency, as it filters semantic features using a normalization-based gating mechanism without the heavy computational burden of complex self-attention layers. These two attention maps are adaptively fused via a learnable gating parameter, α , and are applied to the original features in a residual manner to obtain the enhanced features X :
X = X + α ( X M s ) + ( 1 α ) ( X M c )
where M s and M c are computed by FastCA and GCT, respectively, and the symbol denotes element-wise multiplication (Hadamard product). The parameter α , implemented via a Sigmoid activation function, serves as an adaptive gating coefficient. Its primary role is to dynamically govern the trade-off between geometric and semantic information. During training, the network automatically adjusts α via back-propagation; specifically, a higher value prioritizes spatial structural cues derived from FastCA, whereas a lower value emphasizes semantic channel features extracted by GCT, thereby optimizing the feature representation layer-by-layer.
Furthermore, considering that features extracted at different network depths possess varying semantic levels, we adopt a hierarchical integration strategy. The DHA module is strategically embedded at the entrance of the backbone network and after each downsampling stage. This ensures that features throughout the entire pipeline receive dynamic and appropriate focus and enhancement.

3.4. Scale-Aggregation Head

The final detection head confronts a core real-world challenge: the significant variation in object scales within autonomous driving scenes. A standard single-scale detection head, constrained by its fixed receptive field, struggles to efficiently and simultaneously handle both large, nearby vehicles and small, distant pedestrians.
To address the issue of non-uniform density distribution in LiDAR point clouds, we designed the Scale-Aggregation Head (SA-Head), the core of which lies in explicitly handling multi-scale information through its structural design. As shown in Figure 4, the SA-Head introduces a multi-scale feature aggregation mechanism [31]. It receives high-resolution features, F h , and low-resolution features, F l , from the 2D BEV backbone. A fusion module, akin to a Feature Pyramid Network (FPN) [32,33], integrates these two feature types to generate a scale-complete, unified feature plane, F agg . This is achieved through lateral connections, ϕ h and ϕ l , implemented by 1 × 1 convolutional layers, and an upsampling operator, U :
F agg = ϕ h ( F h ) + ϕ l ( U ( F l ) )
F agg is subsequently fed into shared convolutional layers and separate prediction heads for heatmap prediction and bounding box attribute regression, respectively. In this manner, the SA-Head can leverage information from different semantic levels within a single decoding stage. This provides a solid structural foundation for concurrently detecting objects of various sizes, thereby significantly enhancing the model’s overall adaptability to complex scenes.

3.5. Loss Function

We employ a multi-task loss function, L , to train the network. This loss function is composed of a classification loss for supervising the center point heatmap prediction and a regression loss for optimizing the 3D bounding box attributes.
For the classification task, we adopt the Focal Loss [34], denoted as L f l , which is widely used in dense object detection. In point cloud detection scenarios, there is a severe foreground-background imbalance, as the number of object centers is far less than the number of background locations. Focal Loss effectively addresses this issue by introducing a modulating factor that down-weights the contribution of a large number of easy negative samples to the total loss, thereby enabling the model to focus more on learning from hard positive samples. For the regression of the various 3D bounding box attributes, we utilize the Smooth L1 Loss [35], denoted as L r e g . This loss is applied only at the locations corresponding to the ground-truth object centers (derived from the high-precision KITTI benchmark annotations [2]) and computes the deviation between the predicted values and the true values.
The total loss of the network, L , is a weighted sum of the aforementioned classification and regression losses. We introduce a balancing hyperparameter, λ r e g , to adjust the weight of the regression loss. The total loss is calculated as follows:
L = L f l + λ r e g L r e g

4. Experiments and Result Analysis

To comprehensively validate the effectiveness and robustness of HFSA-Net, extensive experiments were conducted on the KITTI benchmark. This section first details the dataset characteristics, implementation parameters, and evaluation metrics. Subsequently, we present a quantitative comparison with mainstream 3D object detection methods, followed by an ablation study that dissects the specific contributions of each proposed module.

4.1. Dataset

All experiments in this study were conducted on the widely used KITTI dataset. The data acquisition platform for this dataset is equipped with grayscale cameras, color cameras, and a Velodyne HDL-64E LiDAR sensor. This Velodyne LiDAR features 64 scan lines and captures high-precision 3D point cloud data at a frequency of 10 Hz. Key technical specifications of this sensor are summarized in Table 1. The KITTI dataset comprises a total of 7481 samples, covering three core categories: ‘Car’, ‘Pedestrian’, and ‘Cyclist’. Following the widely adopted standard data split protocol proposed by Chen et al. [36], we split the official training set (7481 samples) into a training set of 3712 samples and a validation set of 3769 samples. This division helps to avoid overfitting and ensures a fair comparison with other methods using the same configuration. These sets were used for model training and ablation studies, respectively.

4.2. Experimental Setup and Parameters

Implementation Details. Our research was conducted based on the OpenPCDet framework, an open-source toolbox for 3D object detection. The experimental environment was configured as follows: the operating system was Ubuntu 20.04, a Python 3.8 virtual environment was created using Anaconda3, and the training framework was built upon PyTorch 1.10.0 with CUDA 11.1. The hardware platform consisted of an Intel Core i5-12490F CPU (Intel Corporation, Santa Clara, CA, USA) and an NVIDIA GeForce RTX3060 GPU (NVIDIA Corporation, Santa Clara, CA, USA).
Training Parameters. During the training process, the batch size was set to 4 due to the memory constraints of the single NVIDIA GeForce RTX 3060 GPU. The model was trained for 100 epochs. We used the Adam optimizer combined with the OneCycle scheduling policy to ensure stable convergence. The initial learning rate was set to 0.002, which was adjusted based on the Linear Scaling Rule to align with the reduced batch size. The weight decay was set to 0.01, and gradient clipping (with a norm of 10) was applied to prevent exploding gradients. For the enhanced attention modules, their learning rate was set to 1/10 of that of the backbone network to ensure stability in feature extraction. Data Processing. For data processing, the point cloud range was defined as [0, −40, −3] to [70.4, 40, 1] along the x, y, and z axes, respectively. The voxel size was set to (0.05, 0.05, 0.1) meters. The maximum number of voxels was configured to 16,000 during training and 40,000 during inference. To improve the model’s generalization ability, we employed several data augmentation strategies during training, including: (1) random horizontal flipping with a probability of 0.5; (2) random rotation around the vertical axis within a range of [−45°, +45°]; and (3) global scaling with a random factor between 0.95 and 1.05 to enrich sample diversity.

4.3. Evaluation Metrics

We adopt the official evaluation metrics of the KITTI benchmark, namely Average Precision (AP), which is calculated using 40 recall positions. The evaluation is conducted from three perspectives: 3D detection, Bird’s-Eye View (BEV) detection, and Average Orientation Similarity (AOS).
According to the official KITTI criteria, the Intersection over Union (IoU) threshold is set to 0.7 for the ‘Car’ category and 0.5 for the ‘Pedestrian’ and ‘Cyclist’ categories. For each category, the AP is reported for three difficulty levels: Easy, Moderate, and Hard. The AP is defined as:
A P R 40 = 1 | R 40 | r R 40 p interp ( r )
p interp ( r ) = max r ˜ : r ˜ r p ( r ˜ )
R 40 is a set containing 40 equally spaced recall thresholds, | R 40 | is the cardinality of this set, and p interp ( r ) is the interpolated precision at a recall level r . p ( r ˜ ) denotes the measured precision at an actual recall of ř. The AOS is also sampled at these 40 recall positions, and its formula is defined as:
A O S R 40 = 1 | R 40 | r R 40 s interp ( r )
s interp ( r ) = max r ˜ : r ˜ r s ( r ˜ )
s ( r ˜ ) = 1 | D ( r ˜ ) | i D ( r ˜ ) 1 + cos ( Δ θ ( i ) ) 2 δ i
where s interp ( r ) is the interpolated orientation similarity at recall level r , and s ( r ˜ ) is the average orientation similarity of all True Positive (TP) detections at a recall of r ˜ . D ( r ˜ ) is the set of all detections classified as TP at recall r ˜ , and | D ( r ˜ ) | is the number of elements in this set. Δ θ ( i ) represents the angle error between the estimated orientation and the ground-truth orientation for the i-th detection.

4.4. Experimental Result Analysis

4.4.1. Loss Curve

During the training process of this experiment, the trend of the total loss value over time is illustrated in Figure 5. Specifically, on the experimental platform (RTX 3060 GPU), the average time per iteration (batch size = 4) is approximately 0.48 s, and the total training time for 100 epochs is about 12.5 h. It is evident that the proposed algorithm maintains a lower loss value throughout the entire training process.

4.4.2. Quantitative Analysis

To evaluate the effectiveness of our proposed model, we conducted a comprehensive comparison with mainstream 3D object detection methods. Specifically, CenterPoint [19] was selected as the primary baseline to verify the architectural improvements. We also included foundational voxel-based methods (VoxelNet [16], SECOND [17]) and the widely used real-time detector PointPillars [18] to benchmark against classic standards. Additionally, F-PointNet represents the point-based paradigm, providing a cross-category comparison. The results are presented in Table 2, Table 3 and Table 4.
3D detection performance is a core metric for assessing a model’s comprehensive perception capability. As shown in Table 3, HFSA-Net achieves an mAP of 66.93%, outperforming all comparative methods and marking an improvement of 3.34% over the CenterPoint baseline. Most notably, on the highly challenging pedestrian detection task, our model attains an AP of 51.17%, representing a substantial 6.04% gain over the baseline. This significant uplift demonstrates the effectiveness of the proposed S-VFE module in encoding fine-grained geometric structures, as well as the advantage of the SA-Head in fusing high-resolution features. Furthermore, the robust improvements in the ‘Car’ and ‘Cyclist’ categories indicate the general applicability of our proposed algorithmic framework.
The BEV detection performance reflects the model’s capabilities in 2D planar localization and size estimation. As presented in Table 2, HFSA-Net also exhibits outstanding performance from this perspective, achieving an mAP of 72.64% and comprehensively surpassing methods like CenterPoint and PointPillars. This indicates that our proposed modules, particularly the attention-enhanced features provided by the HDA-Backbone and the multi-scale fusion from the SA-Head, have effectively improved the quality of feature representation in the bird’s-eye view.
The AOS performance provides a composite evaluation of 2D detection and 3D orientation estimation. Our model achieves an mAP of 80.76% in this metric, which is also superior to the CenterPoint baseline. This demonstrates that the rich structural and semantic features extracted by our network are also beneficial for accurately predicting the 3D orientation of objects.

4.4.3. Qualitative Analysis

To provide a more intuitive demonstration of HFSA-Net’s performance, we present a qualitative comparison of detection results on the KITTI dataset in Figure 6. The figure displays two challenging scenarios that include distant small objects and partially occluded vehicles. It can be clearly observed that in regions where the baseline model, CenterPoint, either misses objects or provides inaccurate localization, HFSA-Net is able to detect the targets successfully and precisely. For instance, in the first scenario, HFSA-Net successfully identifies a pedestrian in an occluded area, whereas the baseline model completely overlooks this target. These visual results provide strong corroborating evidence for our quantitative analysis.

4.5. Ablation Study

To thoroughly investigate the effectiveness of the three proposed modules and their interactions, we designed a series of detailed ablation studies. Starting from the CenterPoint baseline model (denoted as the first row in the tables), we progressively integrated our innovative components. In the tables, S-VFE represents the Structured Voxel Feature Encoder, HDA stands for the Hybrid-Domain Attention-guided sparse Backbone, and SA-H signifies the Scale-Aggregation Head. The experimental results were evaluated from the BEV, 3D, and AOS perspectives, as presented in Table 5, Table 6 and Table 7.
First, we evaluated the individual contribution of each module. When SA-H was introduced alone, the model achieved the largest individual gain of 2.66% in 3D mAP, primarily attributed to its significant improvements on the ‘Car’ and ‘Cyclist’ categories. This advantage was equally evident from the BEV and AOS perspectives, which strongly demonstrates that multi-scale fusion is crucial for enhancing object localization and size estimation. The individual addition of S-VFE led to a 0.51% increase in 3D mAP, with its main advantage being the refined perception of the ‘Pedestrian’ category. Although its overall gains in BEV and AOS were modest, it provided higher-quality initial features for subsequent modules. Introducing HDA alone yielded a 0.21% improvement in 3D mAP, indicating that its effect is limited without the cooperation of other modules. However, it laid the foundation for subsequent synergistic enhancements.
Second, we assessed the synergistic effects between the modules. The performance of module combinations surpassed the sum of their individual contributions. Among them, the combination of HDA and SA-H was particularly outstanding, achieving a 3.76% improvement in 3D mAP, along with gains of 1.06% in BEV mAP and 0.42% in AOS mAP. This reveals that a powerful attention mechanism (HDA) provides high-quality features for an advanced multi-scale head (SA-H), thereby maximizing its fusion capabilities. This demonstrates the complementary nature of the proposed modules: the S-VFE enriches the input geometric representation, the HDA-Backbone refines feature selection, and the SA-Head ensures robust multi-scale perception. Together, they optimize the entire detection pipeline from encoding to prediction.
Finally, we evaluated the complete model, HFSA-Net (S-VFE + HDA + SA-H). With the support of all modules, it delivered the most comprehensively powerful and robust solution. From the 3D perspective, it achieved a 3.34% mAP improvement over the baseline. From the BEV perspective, the mAP gain was 2.77%, and from the AOS perspective, it was 1.23%. The consistent and significant performance growth across these three core evaluation dimensions uniformly validates the comprehensiveness and effectiveness of our proposed enhanced framework. The final model also demonstrated stronger robustness on other metrics, such as the ‘Hard’ difficulty level, confirming the unique value of S-VFE in processing sparse and challenging samples.

4.6. Real-Vehicle Experiment

To validate the adaptability and robustness of our proposed object detection algorithm in a real-world traffic environment, we conducted a series of real-vehicle experiments. Our experimental platform, as shown in Figure 7, is an autonomous research vehicle modified from a Haval H7. The vehicle is equipped with a Velodyne HDL-32E LiDAR sensor. The key technical specifications of this sensor are detailed in Table 8.
Compared to the Velodyne HDL-64E used to create the KITTI dataset, the HDL-32E has fewer vertical scan lines and a lower point cloud output rate, meaning the point clouds it generates are inherently sparser. Additionally, the vehicle is outfitted with other core equipment, including cameras, a combined GPS-IMU navigation system, and an industrial control computer. The installation positions and layout of the sensors are also depicted in Figure 7.
We collected traffic scenes from a campus road environment with the vehicle moving at a safe speed of approximately 20–30 km/h, and processed the raw data packets into a standardized dataset compatible with the KITTI format. Although vehicle motion theoretically introduces point cloud distortion, at this low speed, such distortion is negligible and has a minimal impact on detection accuracy. Therefore, the primary challenge stems from the relatively low resolution of the vehicle-mounted LiDAR and environmental factors such as occlusions, which result in a point cloud density significantly lower than that of the KITTI dataset. As shown in Figure 8, despite these challenges, the improved algorithm demonstrates strong detection capabilities.
As depicted in Figure 8c displays the detection results in a campus scene, where the model is able to accurately identify multi-class traffic participants. In panel (d), despite occlusions between a cyclist and a vehicle and the resulting point cloud sparsity, our algorithm still successfully distinguishes between the different targets. Furthermore, the experiment validates the model’s robustness against dynamic targets. Despite the cyclist moving at speeds of approximately 10–15 km/h, the proposed S-VFE module effectively captures the structural integrity of the point cloud, which remains largely unaffected by motion distortion at these velocities. Moreover, the model accurately predicts the heading angle across varying orientations. This capability is quantitatively corroborated by the high Average Orientation Similarity (AOS) of 80.76% reported in Table 4, confirming the network’s effectiveness in perceiving moving objects with diverse velocities and trajectories.
The experiment demonstrates that the proposed method is not reliant on a specific high-density sensor configuration. It possesses excellent cross-sensor generalization ability and can effectively address the detection challenges in complex traffic scenarios.

5. Discussion

In this section, we provide a critical analysis of the experimental findings. We first interpret the underlying reasons for the performance improvements, particularly regarding small and sparse objects. We then examine the model’s generalization capabilities across different sensor configurations. Finally, we discuss the computational trade-offs and current limitations to suggest directions for future research.

5.1. Performance Analysis and Interpretations

The experimental results on the KITTI dataset demonstrate that HFSA-Net effectively addresses the challenges of point cloud sparsity. With a 3D mAP of 66.93%, our model outperforms the baseline CenterPoint by 3.34%. A critical finding is the significant improvement in the ‘Pedestrian’ category (+6.04% AP). Pedestrians occupy very few voxels in LiDAR scans, making them highly sensitive to the quantization artifacts caused by standard voxelization. The success of HFSA-Net in this category is attributed to the synergy between the high voxel resolution and the Structured Voxel Feature Encoder (S-VFE). The fine resolution ensures sufficient spatial extent for small objects, while the S-VFE employs neighbor aggregation to preserve fine-grained intra-voxel details that are typically discarded by mean-pooling operations.

5.2. Generalization and Practicality

The real-vehicle experiments provide crucial insight into the model’s robustness. While the model was trained on the high-resolution KITTI dataset (64-beam), it was tested on a vehicle platform equipped with a lower-resolution Velodyne HDL-32E sensor (32-beam). Despite the domain gap and reduced point density, HFSA-Net successfully detected cyclists and pedestrians without fine-tuning. This indicates that the Scale-Aggregation Head (SA-Head) effectively mitigates the impact of varying point densities by fusing multi-scale features, making the algorithm highly adaptable for deployment on autonomous vehicles with varying sensor configurations.

5.3. Limitations and Future Work

Despite the promising results, this study presents certain limitations. First, regarding computational cost, the proposed S-VFE module utilizes multi-scale neighborhood aggregation, which introduces additional latency compared to simple mean-pooling operations. However, experimental evaluation on an RTX 3060 GPU demonstrates an average inference latency of 42.8 ms, corresponding to a speed of approximately 23.4 FPS. This processing speed meets the real-time constraints of standard autonomous driving systems that typically operate at 10 Hz. Second, the current experiments were conducted primarily in clear weather scenarios. Future work will focus on extending the framework to multi-modal fusion by integrating LiDAR and camera data. Specifically, we aim to incorporate image-derived semantic information into the S-VFE module to enrich point representation. Additionally, the attention mechanisms within the HDA-Backbone will be adapted to fuse cross-modal features, thereby enhancing detection robustness in adverse weather conditions where point cloud noise increases significantly.

6. Conclusions

In response to the limitations imposed on 3D object detection by the inherent sparsity and geometric structural fragility of LiDAR point clouds, this paper has proposed an enhanced 3D object detection framework: HFSA-Net. First, our designed Structured Voxel Feature Encoder (S-VFE) effectively overcomes the information loss associated with traditional VFE methods through intra-voxel refinement and multi-scale neighborhood context aggregation. Second, the constructed Hybrid-Domain Attention-guided sparse Backbone (HDA-Backbone) utilizes a decoupled hybrid attention mechanism optimized for sparse data, enabling the network to adaptively focus on critical features and thereby enhancing both the efficiency and quality of feature extraction. Finally, the proposed Scale-Aggregation Head (SA-Head) significantly improves the model’s perception capabilities for objects of varying sizes by fusing multi-scale BEV features. Extensive experiments conducted on the public KITTI dataset have thoroughly validated the effectiveness of our proposed method. Compared to the CenterPoint baseline, HFSA-Net improves the 3D mean Average Precision (mAP) by 3.34%. It achieves a remarkable increase of up to 6.04% in average precision on the challenging ‘Moderate’ difficulty pedestrian detection task, which demonstrates HFSA-Net’s superior performance in handling sparse, small-sized targets. The general applicability and robustness of the framework were further confirmed through ablation studies and real-vehicle experiments. Future work will explore extending this framework to multi-modal data fusion scenarios to further enhance 3D perception capabilities in complex scenes.

Author Contributions

Conceptualization, X.Y. and J.S.; methodology, X.Y. and Z.X.; software, Z.Q. and L.W.; validation, X.Y. and J.S.; formal analysis, X.Y. and Z.Q.; resources, Z.X. and L.W.; writing—original draft preparation, X.Y. and Z.Q.; writing—review and editing, Z.X.; visualization, Z.Q. and L.W.; supervision, Z.X.; funding acquisition, Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC were funded by Shandong Province Major Science and Technology Innovation Project (Grant No. 2023CXGC010111) and Shandong Provincial Natural Science Foundation [grant numbers ZR2023QE105].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors declare that upon reasonable request, the data and the code are available from the corresponding author.

Conflicts of Interest

Author Lei Wang was employed by the company Shuntai Automobile Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HFSA-NetHierarchical Focus and Structural-Aware Network
S-VFEStructured Voxel Feature Encoder
HDA-BackboneHybrid-Domain Attention-guided sparse Backbone
SA-HeadScale-Aggregation Head
FastCAFast Coordinate Attention
GCTGated Channel Transformation
FPNFeature Pyramid Network
APAverage Precision
AOSAverage Orientation Similarity
TPTrue Positive

References

  1. Janai, J.; Güney, F.; Behl, A.; Geiger, A. Computer Vision for Autonomous Vehicles: Problems, Datasets and State of the Art. FNT Comput. Graph. Vis. 2020, 12, 1–308. [Google Scholar] [CrossRef]
  2. Geiger, A. Object Detection Evaluation 2012; The KITTI Vision Benchmark Suite: Tübingen, Germany, 2012. [Google Scholar]
  3. Linnhoff, C.; Hofrichter, K.; Elster, L.; Rosenberger, P.; Winner, H. Measuring the Influence of Environmental Conditions on Automotive Lidar Sensors. Sensors 2022, 22, 5266. [Google Scholar] [CrossRef] [PubMed]
  4. Li, Y.; Ibanez-Guzman, J. Lidar for Autonomous Driving: The Principles, Challenges, and Trends for Automotive Lidar and Perception Systems. IEEE Signal Process. Mag. 2020, 37, 50–61. [Google Scholar] [CrossRef]
  5. Shi, W.; Rajkumar, R. Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1711–1719. [Google Scholar]
  6. Fernandes, D.; Silva, A.; Névoa, R.; Simões, C.; Gonzalez, D.; Guevara, M.; Novais, P.; Monteiro, J.; Melo-Pinto, P. Point-Cloud Based 3D Object Detection and Classification Methods for Self-Driving Applications: A Survey and Taxonomy. Inf. Fusion 2021, 68, 161–191. [Google Scholar] [CrossRef]
  7. Wang, R.; Peethambaran, J.; Chen, D. LiDAR Point Clouds to 3-D Urban Models: A Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 606–627. [Google Scholar] [CrossRef]
  8. Li, Y.; Ma, L.; Zhong, Z.; Liu, F.; Chapman, M.A.; Cao, D.; Li, J. Deep Learning for LiDAR Point Clouds in Autonomous Driving: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3412–3432. [Google Scholar] [CrossRef] [PubMed]
  9. Huang, S.; Liu, L.; Fu, X.; Dong, J.; Huang, F.; Lang, P. Overview of LiDAR Point Cloud Target Detection Methods Based on Deep Learning. Sens. Rev. 2022, 42, 485–502. [Google Scholar] [CrossRef]
  10. Liu, M.; Ma, J.; Zheng, Q.; Liu, Y.; Shi, G. 3D object detection based on attention and multi-scale feature fusion. Sensors 2022, 22, 3935. [Google Scholar] [CrossRef] [PubMed]
  11. Li, B.; Zhang, T.; Xia, T. Vehicle Detection from 3D Lidar Using Fully Convolutional Network 2016. arXiv 2016, arXiv:1608.07916. [Google Scholar]
  12. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-View 3D Object Detection Network for Autonomous Driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  13. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  14. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  15. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 770–779. [Google Scholar]
  16. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  17. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  18. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection From Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  19. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-Based 3D Object Detection and Tracking. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  20. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10529–10538. [Google Scholar]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.U.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  22. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  23. Mao, J.; Xue, Y.; Niu, M.; Bai, H.; Feng, J.; Liang, X.; Xu, H.; Xu, C. Voxel Transformer for 3D Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 3164–3173. [Google Scholar]
  24. Rehman, M.Z.U.; Islam, S.M.S.; Blake, D.; Ulhaq, A.; Janjua, N. Deep Learning for Land Use Classification: A Systematic Review of HS-LiDAR Imagery. Artif. Intell. Rev. 2025, 58, 272. [Google Scholar] [CrossRef]
  25. Coglan, J.; Gharineiat, Z.; Tarsha Kurdi, F. Automatic Rooftop Solar Panel Recognition from UAV LiDAR Data Using Deep Learning and Geometric Feature Analysis. Remote Sens. 2025, 17, 3389. [Google Scholar] [CrossRef]
  26. Wang, Z.; Huang, X.; Hu, Z. Attention-Based LiDAR–Camera Fusion for 3D Object Detection in Autonomous Driving. World Electr. Veh. J. 2025, 16, 306. [Google Scholar] [CrossRef]
  27. Naich, A.Y.; Carrión, J.R. LiDAR-Based Intensity-Aware Outdoor 3D Object Detection. Sensors 2024, 24, 2942. [Google Scholar] [CrossRef] [PubMed]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  29. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Virtual, 17 October 2021; pp. 13713–13722. [Google Scholar]
  30. Yang, Z.; Zhu, L.; Wu, Y.; Yang, Y. Gated Channel Transformation for Visual Recognition. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11794–11803. [Google Scholar]
  31. Liang, T.; Xie, H.; Yu, K.; Xia, Z.; Lin, Z.; Wang, Y.; Tang, T.; Wang, B.; Tang, Z. BEVFusion: A Simple and Robust LiDAR-Camera Fusion Framework. Adv. Neural Inf. Process. Syst. 2022, 35, 10421–10434. [Google Scholar]
  32. Lin, T.-Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2117–2125. [Google Scholar]
  33. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  34. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  35. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  36. Chen, X.; Kundu, K.; Zhu, Y.; Berneshawi, A.G.; Ma, H.; Fidler, S.; Urtasun, R. 3D Object Proposals for Accurate Object Class Detection. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
Figure 1. Block diagram of the proposed network.
Figure 1. Block diagram of the proposed network.
Sensors 26 00338 g001
Figure 2. Block diagram of the structured voxel feature encoder.
Figure 2. Block diagram of the structured voxel feature encoder.
Sensors 26 00338 g002
Figure 3. Block diagram of the hybrid-domain attention-guided sparse backbone network.
Figure 3. Block diagram of the hybrid-domain attention-guided sparse backbone network.
Sensors 26 00338 g003
Figure 4. Block diagram of the scale-aggregated detection head.
Figure 4. Block diagram of the scale-aggregated detection head.
Sensors 26 00338 g004
Figure 5. Curve of total loss versus time.
Figure 5. Curve of total loss versus time.
Sensors 26 00338 g005
Figure 6. Detection results of the two algorithms in different scenes. (a) Original photo of Scene 1; (b) Detection result of the baseline method for Scene 1; (c) Detection result of the proposed method for Scene 1; (d) Original photo of Scene 2; (e) Detection result of the baseline method for Scene 2; (f) Detection result of the proposed method for Scene 2.
Figure 6. Detection results of the two algorithms in different scenes. (a) Original photo of Scene 1; (b) Detection result of the baseline method for Scene 1; (c) Detection result of the proposed method for Scene 1; (d) Original photo of Scene 2; (e) Detection result of the baseline method for Scene 2; (f) Detection result of the proposed method for Scene 2.
Sensors 26 00338 g006
Figure 7. Vehicles used in the experiment.
Figure 7. Vehicles used in the experiment.
Sensors 26 00338 g007
Figure 8. Detection results of the real-vehicle experiment. (a) Original photo of Scene 1; (b) Original photo of Scene 2; (c) Detection result of the proposed method for Scene 1; (d) Detection result of the proposed method for Scene 2.
Figure 8. Detection results of the real-vehicle experiment. (a) Original photo of Scene 1; (b) Original photo of Scene 2; (c) Detection result of the proposed method for Scene 1; (d) Detection result of the proposed method for Scene 2.
Sensors 26 00338 g008
Table 1. Key technical specifications of the LiDAR sensor used in the KITTI dataset.
Table 1. Key technical specifications of the LiDAR sensor used in the KITTI dataset.
ParametersValues
Laser Harness (Wire)64
Measuring range (m)120
Range Accuracy (cm)2
Horizontal FoV (°)
Vertical FoV (°)
360
26.8
Output (points per second)1,300,000
Table 2. The detection accuracy of different algorithms on the dataset from the perspective of BEV (%).
Table 2. The detection accuracy of different algorithms on the dataset from the perspective of BEV (%).
ModelCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
Second88.0784.0075.3358.0950.2247.2083.6666.1962.1368.24
VoxelNet89.3579.2677.3946.1340.7438.1166.754.7650.5560.33
F-PointNet88.7084.0075.3358.0951.0547.5475.3861.6954.6866.27
PointPillars92.0587.8085.1956.5350.8346.4381.3265.0760.7369.55
Centerpoint91.4185.6383.0456.0251.7747.9680.1068.1164.8069.87
Ours91.3187.6286.0662.1657.0652.5883.1968.9364.8372.64
Table 3. The detection accuracy of different algorithms in the 3D perspective of the dataset (%).
Table 3. The detection accuracy of different algorithms in the 3D perspective of the dataset (%).
ModelCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
Second86.4476.9773.3947.4740.4736.2681.2863.4959.2962.78
VoxelNet77.4765.1157.7339.4833.6931.5061.2248.3644.3750.99
F-PointNet81.2070.3962.1951.2144.8940.2371.9656.7750.3958.80
PointPillars85.0375.7672.7450.0844.1839.5377.1360.9456.9162.48
Centerpoint86.8675.9873.0949.7045.1341.1676.7363.3460.3463.59
Ours85.8077.6375.4656.1851.1746.8582.3965.0161.9066.93
Table 4. The detection accuracy of different algorithms from the perspective of AOS on the dataset (%).
Table 4. The detection accuracy of different algorithms from the perspective of AOS on the dataset (%).
ModelCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
Second94.8490.9490.1160.0153.9250.7789.4072.8268.8774.63
SubCNN90.6188.4378.6378.3366.2861.3771.3963.4146.3471.64
AVOD-FPN89.9587.1379.7453.3644.9243.7767.6157.5354.1664.24
PointPillars95.0291.2488.4647.3344.4041.3184.7571.3567.2470.12
Centerpoint95.5689.7388.9570.1065.3161.6190.5978.6875.2579.53
Ours96.0392.5290.3674.3569.5965.2190.0176.3872.3480.76
Table 5. The detection accuracy of the ablation experiment from the BEV perspective on the KITTI dataset (%).
Table 5. The detection accuracy of the ablation experiment from the BEV perspective on the KITTI dataset (%).
S-VFEHDASA-HCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
91.4185.6383.0456.0251.7747.9680.1068.1164.8069.87
89.7986.1685.5855.6251.8847.8381.1368.7164.9370.18
89.3486.0885.6156.6551.7847.7283.5367.8063.6070.23
89.8986.4086.0158.9853.2748.9084.4970.6466.1771.64
90.0786.6685.9858.8753.8149.3082.0470.2566.0771.45
91.9987.9486.3359.5354.0750.0282.4667.8064.1671.59
89.6187.6486.1660.9356.5451.5182.0770.7166.7372.44
91.3187.6286.0662.1657.0652.5883.1968.9364.8372.64
Table 6. The detection accuracy of the ablation experiment in the 3D perspective on the KITTI dataset (%).
Table 6. The detection accuracy of the ablation experiment in the 3D perspective on the KITTI dataset (%).
S-VFEHDASA-HCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
86.8675.9873.0949.7045.1341.1676.7363.3460.3463.59
84.0976.0274.0251.2046.9242.6178.1963.0259.1563.91
85.4975.9174.0750.8345.9741.5581.5463.1859.0464.18
84.4177.8375.8553.4447.6843.3581.7066.9262.5965.97
86.4976.4674.4754.0149.3144.3079.1766.6962.6965.96
85.6378.1574.7053.1948.0944.2279.5462.9859.3265.09
84.9577.2975.6357.1752.4047.4577.9066.0361.9566.75
85.8077.6375.4656.1851.1746.8582.3965.0161.9066.93
Table 7. The detection accuracy of the ablation experiment from the AOS perspective on the KITTI dataset (%).
Table 7. The detection accuracy of the ablation experiment from the AOS perspective on the KITTI dataset (%).
S-VFEHDASA-HCarPedestrianCyclistmAP
EasyModeHardEasyModeHardEasyModeHard
95.5689.7388.9570.1065.31 61.6190.5978.6875.2579.53
94.8090.8490.0668.6165.4562.3588.6377.8273.3979.11
94.3092.0690.2371.6867.1062.6389.2275.7272.2379.47
94.7190.7890.1070.8566.5462.0991.2077.4572.8179.61
94.8291.0290.34 72.3667.8662.9490.9078.0974.1980.28
94.7592.0290.2071.5067.3763.5788.8576.8273.5279.84
94.4892.4690.2471.2667.8464.4088.5177.3173.1179.95
96.0392.5290.3674.3569.5965.2190.0176.3872.3480.76
Table 8. Key Technical Specifications of Velodyne HDL-32E.
Table 8. Key Technical Specifications of Velodyne HDL-32E.
ParametersValues
Laser Harness (Wire)32
Measuring range (m)80–100
Range Accuracy (cm)±2
Dimensions (mm)85 × 144
Horizontal FoV (°)360
Vertical FoV (°)+10.67° to −30.67°
Supply Voltage (VDC)9–32
Laser ClassClass 1
Power (W)31.4
Output (points per second)700,000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, X.; Xiao, Z.; Shao, J.; Qiu, Z.; Wang, L. HFSA-Net: A 3D Object Detection Network with Structural Encoding and Attention Enhancement for LiDAR Point Clouds. Sensors 2026, 26, 338. https://doi.org/10.3390/s26010338

AMA Style

Yin X, Xiao Z, Shao J, Qiu Z, Wang L. HFSA-Net: A 3D Object Detection Network with Structural Encoding and Attention Enhancement for LiDAR Point Clouds. Sensors. 2026; 26(1):338. https://doi.org/10.3390/s26010338

Chicago/Turabian Style

Yin, Xuehao, Zhen Xiao, Jinju Shao, Zhimin Qiu, and Lei Wang. 2026. "HFSA-Net: A 3D Object Detection Network with Structural Encoding and Attention Enhancement for LiDAR Point Clouds" Sensors 26, no. 1: 338. https://doi.org/10.3390/s26010338

APA Style

Yin, X., Xiao, Z., Shao, J., Qiu, Z., & Wang, L. (2026). HFSA-Net: A 3D Object Detection Network with Structural Encoding and Attention Enhancement for LiDAR Point Clouds. Sensors, 26(1), 338. https://doi.org/10.3390/s26010338

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop