Next Article in Journal
When Action Speaks Louder than Words: Exploring Non-Verbal and Paraverbal Features in Dyadic Collaborative VR
Previous Article in Journal
Dual-Stream Attention-Enhanced Memory Networks for Video Anomaly Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fusing Residual and Cascade Attention Mechanisms in Voxel–RCNN for 3D Object Detection

1
School of Automation, Guangxi University of Science and Technology, Liuzhou 545000, China
2
Liugong Machinery Co., Ltd., Liuzhou 545000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(17), 5497; https://doi.org/10.3390/s25175497
Submission received: 6 June 2025 / Revised: 18 July 2025 / Accepted: 1 September 2025 / Published: 4 September 2025
(This article belongs to the Section Communications)

Abstract

In this paper, a high-precision 3D object detector—Voxel–RCNN—is adopted as the baseline detector, and an improved detector named RCAVoxel-RCNN is proposed. To address various issues present in current mainstream 3D point cloud voxelisation methods, such as the suboptimal performance of Region Proposal Networks (RPNs) in generating candidate regions and the inadequate detection of small-scale objects caused by overly deep convolutional layers in both 3D and 2D backbone networks, this paper proposes a Cascade Attention Network (CAN). The CAN is designed to progressively refine and enhance the proposed regions, thereby producing more accurate detection results. Furthermore, a 3D Residual Network is introduced, which improves the representation of small objects by reducing the number of convolutional layers while incorporating residual connections. In the Bird’s-Eye View (BEV) feature extraction network, a Residual Attention Network (RAN) is developed. This follows a similar approach to the aforementioned 3D backbone network, leveraging the spatial awareness capabilities of the BEV. Additionally, the Squeeze-and-Excitation (SE) attention mechanism is incorporated to assign dynamic weights to features, allowing the network to focus more effectively on informative features. Experimental results on the KITTI validation dataset demonstrate the effectiveness of the proposed method, with detection accuracy for cars, pedestrians, and bicycles improving by 3.34%, 10.75%, and 4.61%, respectively, under the KITTI hard level. The primary evaluation metric adopted is the 3D Average Precision (AP), computed over 40 recall positions (R40). The Intersection over IoU thresholds used are 0.7 for cars and 0.5 for both pedestrians and bicycles.

1. Introduction

In recent years, with the rapid advancement of autonomous driving technologies, LiDAR sensors have become an essential component in the development of self-driving vehicles. The point cloud data they produce contain rich spatial information and offer reliable input for environmental perception, thereby attracting significant academic and industrial interest. Object detection is generally classified into two categories: two-dimensional (2D) and three-dimensional (3D). 3D object detection offers a more intuitive representation of the real world, incorporating size, orientation, and distance information [1]. As LiDAR technology matures and becomes more cost-effective, 3D point cloud-based object detection has emerged as a mainstream approach in autonomous driving. However, LiDAR-generated point clouds are unordered, uneven in density, and highly data-intensive, rendering direct application of 2D object detectors unsuitable for 3D data [2].
Research on point cloud-based object detection can be categorised into four main approaches according to the method of point cloud processing: point-based, voxel-based, projection-based, and hybrid point–voxel approaches. Point-based methods directly operate on raw point cloud data for feature extraction, effectively preserving geometric structures and spatial details. However, they are less suited to large-scale point cloud data. Representative works include PointNet, PointNet++, PointRCNN, and 3DSSD [3,4,5,6]. Voxel-based methods divide the point cloud into regular 3D voxel grids and model the spatial structure using voxel representations, which facilitates the extraction of local features in 3D space. The choice of voxel size significantly affects detection accuracy. Notable methods in this category include SECOND, VoxelNet, and Voxel–RCNN [7,8,9]. Projection-based approaches convert 3D point clouds into 2D images or pseudo-images, enabling the use of conventional 2D convolutional neural networks. This transformation simplifies the data while preserving the overall outline of objects. Representative methods include PointPillars and CenterPoint [10,11]. Hybrid point–voxel methods combine point-wise and voxel-wise representations to leverage both the fine-grained features of point clouds and the global structural information offered by voxels. These methods demonstrate strong object recognition capabilities across a variety of scenarios. Representative works include PV-RCNN, SASSD, and CT3D [12,13,14].
The baseline adopted in this study, Voxel–RCNN, is a classic and effective two-stage detector that not only extracts voxel-based features but also integrates BEV features [9]. Furthermore, the spatial awareness advantage provided by the BEV perspective enables the detector to identify smaller objects more effectively. Region proposals are first generated through the RPN, followed by refinement and classification to produce the final detection results. However, considering the limitations of this approach—namely, the suboptimal performance of the RPN and the excessive depth of both the 3D and 2D backbone networks—this paper proposes a CAN to iteratively refine the predictions. Additionally, residual connections are incorporated into both the 3D and 2D backbone networks. The 2D backbone further integrates a Squeeze-and-Excitation (SE) self-attention mechanism to dynamically reweight features, thereby enhancing the detection capability for small-scale objects [15].
Extensive experiments and rigorous ablation studies conducted on the KITTI dataset demonstrate the effectiveness of the proposed method. Specifically, the contributions of this work can be summarised as follows:
(1)
To address the limited performance of the RPN, a Cascaded Attention Network is designed to iteratively refine region proposals and produce high-quality predictions.
(2)
To mitigate the degradation in feature representation caused by excessively deep convolutional layers in both 3D and 2D backbones, a 3D Residual Network is introduced to enhance feature transmission across layers. Furthermore, a 2D Residual Attention Network is proposed to improve the model’s sensitivity to small-scale objects by incorporating attention-based feature weighting.
(3)
The proposed method is validated on the KITTI dataset, with the results demonstrating its effectiveness. Under the hard setting of the 3D Average Precision (AP) metric using R40 evaluation, detection performance for cars, pedestrians, and cyclists improves by 3.34%, 10.75%, and 4.61%, respectively.

2. Related Work

Currently, 3D object detection methods can be broadly categorised into single-stage and two-stage detectors. The baseline adopted in this study is Voxel–RCNN, a two-stage detector proposed by Jiajun Deng et al. In the first stage, the point cloud is divided into a fixed-size voxel grid, and the point-wise information within each voxel is encoded to form voxel-wise feature representations. These features are then processed by a 3D convolutional backbone network to extract high-level voxel features, and the final-layer features are compressed along the height axis to generate a BEV representation [16]. Subsequently, a RPN applies a 2D backbone network to the BEV features for further extraction, and uses an anchor-based approach to generate candidate object regions [17]. In the second stage, in order to refine the proposals, Voxel–RCNN employs a Multilayer Perceptron (MLP) to directly extract fine-grained geometric features from the raw point cloud within each candidate region [9]. Finally, the Detection Head further refines these proposals and performs classification and 3D bounding box regression to produce the final detection results.
For single-stage detectors, object categories and bounding boxes are typically predicted directly from point cloud data. These methods feature a streamlined architecture that facilitates end-to-end modelling. PointNet and PointNet++, inspired by 2D object detection techniques, operate directly on raw point clouds, thereby effectively preserving geometric details effectively [3,4]. VoxelNet introduces a voxelisation approach, dividing the point cloud into fixed-size 3D grids and employing a Voxel Feature Encoding (VFE) module to extract voxel-level features, enabling an end-to-end learning framework from raw data to detection output [8]. Building upon voxelisation, SECOND addresses the sparse nature of point clouds by incorporating sparse 3D convolutions and applying max pooling to aggregate features, thereby enhancing representational capacity [7]. PointPillars partitions the point cloud into vertical columnar structures (pillars) and uses the Pillar Feature Net (PFN) to extract features from points within each pillar [10]. These features are then projected into a BEV representation and processed using 2D convolutional networks. This approach integrates the advantages of voxel partitioning and BEV representation in structural modelling. However, the construction of pillars may result in the loss of some fine-grained point cloud information, potentially affecting the accuracy of boundary detection. CenterPoint directly projects point clouds into the BEV space and extracts spatial features using 2D convolution, simplifying the processing pipeline while preserving the overall object shape [11]. Nevertheless, the flattening of the 3D structure inherent in BEV representations can lead to the weakening of certain spatial geometric cues. RangeDet projects the point cloud into a depth map based on a radar-view polar coordinate system, maintaining the original spatial structure of the data [18]. It leverages 2D convolutional networks for feature extraction while preserving spatial relationships. However, the geometric distortion and variation in object scales introduced by the range view projection can present challenges, particularly in detecting small or edge-located objects. DCGNN proposes a distinctive point set optimisation method based on density clustering and incorporates both local and global graph neural network modules, which are designed to model intra-set and inter-set relationships, respectively [19]. This approach facilitates a more comprehensive learning of spatial features and offers a novel perspective for single-stage detection networks.
Single-stage detectors lack a region proposal process, which limits their effectiveness in detecting smaller objects—especially when these objects are subject to geometric occlusion. Inspired by the Faster R-CNN architecture in 2D object detection, researchers have developed two-stage frameworks for 3D object detection [20]. In such frameworks, the first stage utilises an RPN to generate candidate regions from point cloud data, while the second stage performs refined feature extraction and bounding box regression on these proposals. This approach enhances detection accuracy, making it well-suited for high-precision 3D object detection tasks. PointRCNN directly extracts features from raw point clouds using PointNet++ and generates 3D object proposals. It then applies RoI pooling to further extract point-level features within each candidate region for object classification and bounding box refinement [5,21]. Although this method effectively leverages the raw geometric information of the point cloud and achieves high detection accuracy, its computational cost is relatively high, limiting its practicality in real-time applications. PV-RCNN processes point clouds through voxelisation and employs sparse voxel convolution to extract features and generate 3D object proposals [12]. It then further refines these proposals by fusing local voxel features with fine-grained raw point features, thereby improving detection accuracy [22]. However, the repeated computation of both voxel-based and point-based features leads to high computational overhead. To address this, PV-RCNN++ introduces improvements to the Voxel Set Abstraction (VSA) module [23], reducing unnecessary computation while maintaining high accuracy. Voxel Transformer combines the advantages of Transformer architectures and voxel-based representations [16], leveraging self-attention mechanisms for global feature extraction and thereby enhancing detection performance, particularly for distant objects. Point2Seq builds upon this by modelling object detection as a sequence generation problem [24]. It adopts an auto-regressive decoding scheme to progressively predict the bounding boxes, making it more effective for detecting small objects and handling sparse point clouds.
In summary, within the field of object detection, single-stage detectors are generally characterised by higher speed [25], whereas two-stage detectors often demonstrate superior accuracy, particularly for small objects. Direct processing of raw LiDAR point cloud data can yield high detection accuracy, but it inevitably poses significant computational challenges. Conversely, projecting point clouds into two-dimensional representations can reduce computational complexity, but often results in information loss and limits the ability to fully exploit the spatial structure of the data—especially when detecting small-scale objects [26]. The Voxel–RCNN algorithm combines voxel-based processing with well-established 2D convolutional techniques, effectively extracting both voxel-wise and BEV features [9]. However, as the depth of convolutional networks increases, the ability to detect small objects may deteriorate. To address this issue, the present study incorporates residual connections and the Squeeze-and-Excitation (SE) attention mechanism to enhance the network’s sensitivity to small objects. In addition, a CAN is introduced to enable multi-level refinement of detection results, thereby significantly improving overall detection accuracy. These improvements not only enhance the performance of the detection algorithm but also further unlock the potential of point cloud data in 3D object detection tasks.

3. Methods

In this study, Voxel–RCNN is adopted as the baseline, with targeted enhancements introduced to address its existing limitations [9]. The proposed improvements primarily focus on three components: the 3D backbone network, the 2D backbone network, and the cascaded attention network applied after the RPN in the first stage. The overall framework of the proposed approach is illustrated in Figure 1.

3.1. Residual Backbone

In Voxel–RCNN, the 3D backbone network is employed following voxel encoding [13], with the objective of extracting spatial features from non-empty voxels. This network primarily comprises Submanifold Convolution Blocks and Sparse Convolutions, which substantially reduce computational complexity while retaining rich feature representations for use in the subsequent RPN. As illustrated in Figure 2a, the 3D backbone network of Voxel–RCNN receives a four-dimensional voxel input containing the X, Y, and Z coordinates, as well as reflectance intensity. Initially, a Submanifold Convolution Block is applied to perform 1× downsampling [27], increasing the number of channels from 4 to 16. This is followed by an additional Submanifold Convolution Block for further feature extraction. Subsequently, sparse convolutions are employed to perform 2×, 4×, and 8× downsampling operations using a 3 × 3 × 3 kernel, with the number of channels gradually increased from 16 to 64 [28]. After each sparse convolution, two Submanifold Convolution Blocks are applied to extract features at multiple scales. The structure of a Submanifold Convolution Block is depicted in Figure 2b and consists of a submanifold convolution layer, normalisation layers, and activation functions. The original 3D backbone network adopts a linear structure, which means that feature extraction is carried out in a sequential manner. This approach results in insufficient multi-scale feature fusion, making it difficult for high-level abstract features to retain low-level detail information. Consequently, when features are subsequently compressed into the BEV representation, more low-level details are lost. To address this issue, residual connections have been integrated into the existing 3D backbone network. These connections facilitate the fusion and interaction of features across different levels, thereby enhancing the effectiveness of feature extraction in subsequent stages [29]. Specifically, the proposed method modifies the final two Submanifold Convolution Blocks, which are responsible for increasing the number of feature channels, by replacing them with residual Submanifold Convolution Blocks, as illustrated in Figure 2d. This design ensures that the features of small-scale objects are better preserved during the channel expansion process. Let x denote the original input, F ( x ) the output of the intermediate network layer, and y the output of the residual connection. The corresponding computation is defined as follows:
y = F ( x ) + x
Meanwhile, during the 8× downsampling process, the number of input feature channels is maintained at 64, while the number of output feature channels is increased to 128 in order to extract more expressive features [30]. This enhancement facilitates spatial transformations and aggregation operations on the 3D feature map, enabling compression along the height dimension. As a result, the voxelised three-dimensional representation can be effectively projected onto a two-dimensional BEV. The modified network structure is illustrated in Figure 2c.

3.2. Residual Attention Network

This study builds upon the final feature map extracted from the 3D backbone network, which is compressed along the height dimension (Z-axis) to generate a BEV feature map [31]. This transformation effectively simplifies the data structure, reduces computational complexity, and enhances the delineation of object boundaries, thereby benefiting subsequent feature extraction and region proposal generation. The existing 2D backbone network primarily comprises two stages—downsampling and upsampling—as illustrated in Figure 3a. Initially, during standard 2D CNN operations, the number of feature channels is reduced from 256 to 64, followed by five consecutive convolutional layers to extract deep features. Subsequently, a 2D convolutional layer is employed to perform 2× downsampling, while increasing the number of feature channels to 128. After downsampling, five additional convolutional layers are applied for further feature extraction. Finally, an upsampling convolution is used to restore the spatial resolution of the original feature map. To fully exploit multi-scale information, the number of feature channels is increased from 64 to 128 through a convolutional operation prior to downsampling, and the resulting features are fused with the upsampled feature map to produce a final fused feature map with 256 channels. However, during the feature extraction phase, the aforementioned framework suffers from a gradual weakening of feature representation for small-scale objects. This is primarily due to the sequential stacking of a large number of convolutional layers in the deeper stages of the network, which tends to dilute fine-grained information.
To address this issue, we propose an RAN. This network enhances the representation of key information by introducing residual connections and incorporating the SE attention mechanism [32]. The SE module suppresses redundant features and adaptively adjusts channel-wise feature weights, thereby improving the feature representation of small-scale objects. As a result, it provides higher-quality feature inputs to the RPN. As illustrated in Figure 3b. In this paper, a convolutional module based on residual learning is proposed, comprising two standard convolutional blocks arranged in series. Within this structure, an identity mapping of the input features is retained and combined with the output of the convolutional path via element-wise addition prior to the application of the activation function in the second convolutional block. This design, as illustrated in Figure 3c, facilitates direct information flow and enables efficient gradient propagation. The Squeeze-and-Excitation (SE) attention module employed in this work is capable of automatically learning the relative importance of feature channels, thereby enhancing the representational capacity of 3D point cloud features. The core idea is to increase the network’s sensitivity to salient features through a channel attention mechanism, while simultaneously suppressing redundant information. The squeeze phase compresses a feature map of dimensions W × H × C into a 1 × 1 × C feature vector by performing global average pooling across the spatial dimensions. This operation captures global contextual information and mitigates inter-channel dependency. Specifically, z R C is conducted by generating a statistic by reducing U c over the spatial dimensions H × W, with the following calculation formula:
z c = F sq ( u c ) = 1 W × H i = 1 W j = 1 H u c ( i , j )
The Excitation stage employs an adaptive channel recalibration mechanism comprising two fully connected (FC) layers. The first FC layer reduces the number of channels from C to C / r , thereby decreasing computational overhead and introducing a non-linear transformation to enhance the feature representation capability. The second FC layer then restores the number of channels back to C and applies a Sigmoid activation function to generate normalised channel-wise attention weights. W 1 R C r × C , W 2 R C r × C . The calculation formula is as follows:
s = F ex ( z , W ) = σ ( g ( z , W ) ) = σ ( W 2 δ ( W 1 z ) )
Finally, the resulting channel attention weights are multiplied element-wise with the input features along the channel dimension, thereby adaptively modulating the representational capacity of each channel. The corresponding flowchart is presented in Figure 2d.

3.3. Cascade Attention Network

Inspired by the remarkable performance of the 2D Cascade Attention Network in object detection tasks, this study extends its underlying principles to the domain of 3D object detection, and proposes a 3D Cascade Attention Network to improve detection accuracy and feature representation capability [33]. The overall process is illustrated in Figure 4. Initially, pooling operations are applied to the candidate regions generated by the RPN in order to extract features from within these regions. The primary aim of this step is to standardise Region of Interest (RoI) features of varying scales into a fixed size, thereby ensuring consistent input for subsequent processing via a Multi-Layer Perceptron (MLP). During the cascading phase, the features are iteratively refined and updated at each stage. To maintain the flow of global information and avoid redundant feature learning at individual stages, a Shared Feature Connection layer is introduced. Furthermore, to strengthen the network’s representational capacity, Cross Attention is incorporated into the cascade framework [34]. F ^ j denotes the encoded features from the previous stage, whereas F j represents the encoded features at the current stage. Here, Q j , K j and V j , corresponding to the query, key, and value, respectively. The computation is formulated as follows:
F ^ i j = softmax Q i j ( K i j ) T c V i j
By learning the interactions between these features, the network is able to effectively integrate cross-stage information, thereby enhancing the robustness of object recognition. Ultimately, the optimised features are passed to the classification branch and regression branch, which are responsible for object category prediction and 3D bounding box regression, respectively, to produce the final detection results. The detailed process is illustrated in Figure 4.

3.4. Loss Function

The loss function in this article combines the RPN loss and the CAN loss with equal weights, L = L R P N + L C A N . The specific calculation formula is as follows:
L RPN = 1 N p i L cls a i , a ^ i + I ( I o U i > u ) i L reg δ i , δ ^ i
L CAN = 1 N b i j L cls a i j , a ^ i j + I ( I o U i j > u j ) i j L reg δ i j , δ ^ i j
For L R P N , N P denotes the number of anchors. a i and δ i represent the outputs for classification and bounding box regression, respectively, while a ^ i and δ ^ i correspond to the classification labels and regression targets. L refers to the regions where the loss is computed, specifically those with an IoU greater than a predefined threshold. For the loss function of L C A N , a i j and δ i j represent the predicted results and ground truth targets, respectively, whereas a ^ i j and δ ^ i j denote the predicted offsets and target offsets.

4. Experiment

4.1. Dataset and Evaluation Metrics

All model training and validation were conducted on the KITTI dataset, which consists of 7481 LiDAR point cloud samples for training and 7518 samples for testing. Following recent related works, the training set was further divided into 3712 samples for training and 3769 samples for validation. All network models were trained using the training split and evaluated on the validation split. For data augmentation, a mixed strategy was adopted. Specifically, random rotations and flipping were applied to prevent overfitting and to improve the generalisation capability of the models. The primary evaluation metric adopted is the 3D Average Precision (AP), computed over 40 recall positions (R40). The Intersection over IoU thresholds used are 0.7 for cars and 0.5 for both pedestrians and bicycles [35].

4.2. Experiment Details

All experiments were conducted on an NVIDIA GeForce GTX 3060 GPU with 12 GB of memory. The system was running Ubuntu 20.04, and the experiments were implemented using the PyTorch 1.10 deep learning framework, with CUDA 11.3 employed for model acceleration. The programming language used was Python 3.6.9. The point cloud input was restricted to the following spatial range: X-axis [0, 70.4] m, Y-axis [−40, 40] m, and Z-axis [−3, 1] m [35]. During the training phase, the number of training epochs was set to 80, and the batch size was fixed at 2. Other parameters, such as voxel size and initial learning rate, remained consistent with those provided in the publicly available code of the original paper.

4.3. Comparison with Other Algorithms

To evaluate the effectiveness of the proposed method, we conducted comparative experiments with several state-of-the-art algorithms on the KITTI validation dataset. The experimental results demonstrate that our approach achieves superior overall detection accuracy, with particularly notable improvements in the detection of small-scale objects. The detailed results are presented in Table 1.

4.4. Ablation Experiment

To thoroughly validate the effectiveness of the proposed approach, a series of systematic ablation studies were conducted to evaluate the individual contributions of each module to the overall model performance. Starting from the baseline model, the CAN, RAN, and 3D Residual module were introduced in succession. The experimental results demonstrate that the incorporation of these modules led to significant improvements in detection performance. Specifically, the 3D bounding box detection accuracy for cars, pedestrians, and bicycles under the three difficulty levels was improved by 0.47%, 3.09%, 3.34%, 10.71%, 11.81%, 10.75%, 5.61%, 4.67%, and 4.61%, respectively. The performance metrics for Voxel–RCNN were obtained using publicly available open-source code. Detailed results are provided in Table 2.

4.5. Visualisation of the Results

To provide a more intuitive understanding of the experimental results, visualisations were carried out using the KITTI dataset. During this process, raw point cloud data from multiple randomly selected scenes was input into the proposed model for inference. As illustrated in Figure 5, the 3D bounding box detection results across various scenes can be clearly observed, offering a visual representation of the model’s performance in diverse environments.
The proposed model was compared with the baseline model across a range of scenarios. As illustrated in Figure 6, the model presented in this paper exhibits clear superiority in detecting small objects at long distances, thereby providing strong evidence of the effectiveness of the proposed modules.
Several scenarios were randomly selected for further comparison, and the results indicate that the model proposed in this study achieves a lower false detection rate compared to the baseline model. The detailed visual comparisons are presented in Figure 7.

5. Discussion

5.1. Limitations

Although the model proposed in this study has achieved notable improvements in overall detection accuracy, certain limitations remain. Firstly, the incorporation of the cascade attention network inevitably increases the computational complexity of the model. Secondly, in complex scenes, particularly where multiple small-scale objects are densely clustered, the model remains prone to missed detections.

5.2. Improvement Methods

In future research, we will focus on optimising the computational efficiency of the proposed algorithm. To address the issue of region redundancy in the three-stage region proposal mechanism of the CAN network, we plan to adopt a feature caching and reuse strategy [36]. Specifically, a feature memory bank will be constructed for overlapping regions, and during iterative processing, feature pooling operations will be performed only on newly generated regions. This approach is expected to significantly reduce computational complexity.
The issue of missed detections in scenarios involving dense clusters of small objects may be attributed to the use of overly large voxel sizes during the voxelisation process. In such cases, occluded small-object point clouds may become merged within a single voxel, making them difficult to distinguish. To address this, we intend to investigate the relationship between voxel size and detection accuracy in greater depth, with the aim of identifying an optimal voxelisation parameter configuration. Additionally, we plan to introduce a multi-scale feature fusion strategy to ensure that small objects across different scales receive adequate feature representation [37]. By enhancing the sensitivity and accuracy of the detection algorithm to small objects, this approach aims to improve the model’s capability in complex scenarios characterised by dense small-object distributions.

6. Summary

In this study, we designed three modules, namely, the CAN, RAN and 3D Residual modules, to enhance the detection performance of the model. The Cascade Attention Network (CAN) aggregates features from different stages using a cross-attention mechanism, which progressively improves and refines the region proposals generated by the Region Proposal Network (RPN). To address the problem of feature degradation caused by overly deep convolutional layers in the original 3D and 2D backbone networks, we reduced the number of convolutional layers and introduced residual connections. This design facilitates the transmission of information across different levels and improves the capability of the network to represent features effectively. The 3D Residual module successfully mitigates the limitations in the 3D backbone network. However, it is not suitable for the 2D backbone network used in the BEV representation, since the BEV is produced by compressing 3D feature maps along the height axis. This compression leads to weaker features for small objects. To resolve this issue, we developed the RAN, which incorporates the SE attention mechanism into the residual connection process. This mechanism adaptively adjusts the weights of different channels and enhances the feature representation of small-scale targets. Experimental results confirm that the algorithm proposed in this study achieves a clear improvement in overall detection accuracy, particularly in the detection of small objects.

Author Contributions

Y.L.: methodology, writing—reviewing and editing, software. R.G., Y.Z. and D.C.: writing—review and editing, conceptualisation. X.F.: investigation, writing—review and editing, project administration, supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Guangxi Science and Technology Major Project (AA2023062091) and Guangxi Outstanding Youth Science Foundation Project (2025GXNSFFA069014).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

Author Dengsheng Cai was employed by the company Liugong Machinery Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RCAVoxel-RCNNResidual and Cascade Attention Mechanisms in Voxel–RCNN
RPNRegion Proposal Network
CANCascade Attention Network
RANResidual Attention Network
SESqueeze-and-Excitation
BEVBird’s-Eye View
3DThree-Dimensional
2DTwo-Dimensional
VFEVoxel Feature Encoding
PFNPillar Feature Net PFN
VSAVoxel Set Abstraction
FCFully Connected

References

  1. Qian, R.; Lai, X.; Li, X. 3D object detection for autonomous driving: A survey. Pattern Recognit. 2022, 130, 108796. [Google Scholar] [CrossRef]
  2. Wu, Y.; Wang, Y.; Zhang, S.; Ogai, H. Deep 3D object detection networks using LiDAR data: A review. IEEE Sens. J. 2020, 21, 1152–1171. [Google Scholar] [CrossRef]
  3. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  4. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Advances in Neural Information Processing Systems; NIPS Foundation: La Jolla, CA, USA, 2017; Volume 30. [Google Scholar]
  5. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 770–779. [Google Scholar]
  6. Yang, Z.; Sun, Y.; Liu, S.; Jia, J. 3DSSD: Point-based 3D single stage object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11040–11048. [Google Scholar]
  7. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  8. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  9. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel r-cnn: Towards high performance voxel-based 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; Volume 35, pp. 1201–1209. [Google Scholar]
  10. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  11. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3D object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  12. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10529–10538. [Google Scholar]
  13. He, C.; Zeng, H.; Huang, J.; Hua, X.S.; Zhang, L. Structure aware single-stage 3d object detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11873–11882. [Google Scholar]
  14. Sheng, H.; Cai, S.; Liu, Y.; Deng, B.; Huang, J.; Hua, X.S.; Zhao, M.J. Improving 3d object detection with channel-wise transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 2743–2752. [Google Scholar]
  15. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  16. Mao, J.; Xue, Y.; Niu, M.; Bai, H.; Feng, J.; Liang, X.; Xu, H.; Xu, C. Voxel transformer for 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 3164–3173. [Google Scholar]
  17. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. Foveabox: Beyound anchor-based object detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  18. Fan, L.; Xiong, X.; Wang, F.; Wang, N.; Zhang, Z. RangeDet: In defense of range view for lidar-based 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2918–2927. [Google Scholar]
  19. Xiong, S.; Li, B.; Zhu, S. DCGNN: A single-stage 3D object detection network based on density clustering and graph neural network. Complex Intell. Syst. 2023, 9, 3399–3408. [Google Scholar] [CrossRef]
  20. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems; NIPS Foundation: La Jolla, CA, USA, 2015; Volume 28. [Google Scholar]
  21. Gholamalinezhad, H.; Khosravi, H. Pooling methods in deep neural networks, a review. arXiv 2020, arXiv:2009.07485. [Google Scholar] [CrossRef]
  22. Tang, H.; Liu, Z.; Zhao, S.; Lin, Y.; Lin, J.; Wang, H.; Han, S. Searching efficient 3d architectures with sparse point-voxel convolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 685–702. [Google Scholar]
  23. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-voxel feature set abstraction with local vector representation for 3D object detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  24. Xue, Y.; Mao, J.; Niu, M.; Xu, H.; Mi, M.B.; Zhang, W.; Wang, X.; Wang, X. Point2seq: Detecting 3d objects as sequences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8521–8530. [Google Scholar]
  25. Wang, H.; Yu, Y.; Cai, Y.; Chen, X.; Chen, L.; Li, Y. Soft-weighted-average ensemble vehicle detection method based on single-stage and two-stage deep learning models. IEEE Trans. Intell. Veh. 2020, 6, 100–109. [Google Scholar] [CrossRef]
  26. Wang, R.; Peethambaran, J.; Chen, D. Lidar point clouds to 3-D urban models: A review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 606–627. [Google Scholar] [CrossRef]
  27. Graham, B.; Van der Maaten, L. Submanifold sparse convolutional networks. arXiv 2017, arXiv:1706.01307. [Google Scholar] [PubMed]
  28. Wang, H.; Chen, Z.; Cai, Y.; Chen, L.; Li, Y.; Sotelo, M.A.; Li, Z. Voxel-RCNN-complex: An effective 3-D point cloud object detector for complex traffic conditions. IEEE Trans. Instrum. Meas. 2022, 71, 2507112. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Dollár, P.; Appel, R.; Belongie, S.; Perona, P. Fast feature pyramids for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1532–1545. [Google Scholar] [CrossRef] [PubMed]
  31. Liu, Q.; Li, X.; Zhang, X.; Tan, X.; Shi, B. Multi-view joint learning and bev feature-fusion network for 3d object detection. Appl. Sci. 2023, 13, 5274. [Google Scholar] [CrossRef]
  32. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  33. Wu, H.; Deng, J.; Wen, C.; Li, X.; Wang, C.; Li, J. CasA: A cascade attention network for 3-D object detection from LiDAR point clouds. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5704511. [Google Scholar] [CrossRef]
  34. Chen, C.F.R.; Fan, Q.; Panda, R. Crossvit: Cross-attention multi-scale vision transformer for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 357–366. [Google Scholar]
  35. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  36. Liao, Z.; Jin, Y.; Ma, H.; Alsumeri, A. Distance Awared: Adaptive Voxel Resolution to help 3D Object Detection Networks See Farther. In Proceedings of the 2023 42nd Chinese Control Conference (CCC), Tianjin, China, 24–26 July 2023; pp. 7995–8000. [Google Scholar]
  37. Wu, P.; Wang, Z.; Zheng, B.; Li, H.; Alsaadi, F.E.; Zeng, N. AGGN: Attention-based glioma grading network with multi-scale feature extraction and multi-modal information fusion. Comput. Biol. Med. 2023, 152, 106457. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall framework diagram.
Figure 1. The overall framework diagram.
Sensors 25 05497 g001
Figure 2. (a) Original network structure, (b) Subm block, (c) modified networkstructure, (d) Subm block with residual connections.
Figure 2. (a) Original network structure, (b) Subm block, (c) modified networkstructure, (d) Subm block with residual connections.
Sensors 25 05497 g002
Figure 3. (a) Original 2D backbone network, (b) modified 2D backbone network, (c) residual convolutional block, (d) SE attention mechanism.
Figure 3. (a) Original 2D backbone network, (b) modified 2D backbone network, (c) residual convolutional block, (d) SE attention mechanism.
Sensors 25 05497 g003
Figure 4. Cascade Attention Network structure.
Figure 4. Cascade Attention Network structure.
Sensors 25 05497 g004
Figure 5. An example of results on the KITT validation dataset, featuring four different scenes, where cyclists, pedestrians, and bicycles are displayed in pink, green, and purple, respectively.
Figure 5. An example of results on the KITT validation dataset, featuring four different scenes, where cyclists, pedestrians, and bicycles are displayed in pink, green, and purple, respectively.
Sensors 25 05497 g005
Figure 6. The detection results of distant targets in two different scenes, with the differences marked with circles.
Figure 6. The detection results of distant targets in two different scenes, with the differences marked with circles.
Sensors 25 05497 g006
Figure 7. The detection results of distant targets in two different scenes, with the differences marked with circles.
Figure 7. The detection results of distant targets in two different scenes, with the differences marked with circles.
Sensors 25 05497 g007
Table 1. In the comparison of algorithms, the best results are indicated in bold.
Table 1. In the comparison of algorithms, the best results are indicated in bold.
MethodsCar.3D (APR40)Ped.3D (APR40)Cyc.3D (APR40)
EasyModHardEasyModHardEasyModHard
PV-RCNN (CVPR)92.5784.8382.6964.2656.6751.9188.6571.9566.78
PV-RCNN++92.1185.5582.2767.7460.5555.9288.7373.5869.05
PointPillars (CVPR)86.4277.2975.6053.648.3645.2282.8364.2460.05
SECOND (SENSOR)90.5581.8678.6155.9451.1446.1782.9666.7465.34
CT3D (ICCV)92.8585.8282.8665.7358.5653.0491.9971.667.34
VoxelNet (CVPR)87.8875.5872.7756.4650.9745.6578.1861.7454.68
P a r t - A 2 (TPAMI)89.3680.2378.8865.761.3255.485.669.2465.63
PointRCNN (CVPR)89.878.6578.0562.6955.7752.6584.4866.3760.83
Ours92.8185.7883.5171.4664.7458.7292.9276.6171.62
−0.04−0.04+0.65+5.74+3.42+3.32+0.93+4.66+4.28
Table 2. Improvement in performance through different modules, the best results of this paper’s algorithm are indicated in bold.
Table 2. Improvement in performance through different modules, the best results of this paper’s algorithm are indicated in bold.
MethodModuleCar.3D (APR40)Ped (APR40)Cyc.3D (APR40)
CAN RAN 3D Residual Easy Mod Hard Easy Mod Hard Easy Mod Hard
RNRAN
Voxel–RCNN 92.3482.7180.1760.7552.9347.9787.3171.9467.43
A 92.4783.8582.1869.8061.5356.8890.8273.5368.88
B 92.5883.8682.3369.961.5456.9591.2273.4669.91
C 92.8084.9382.9970.1362.5257.3191.8574.0569.92
Ours 92.8185.7883.5171.4664.7458.7292.9276.6171.62
Improvement +0.47+3.09+3.34+10.71+11.81+10.75+5.61+4.67+4.61
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, Y.; Zhang, Y.; Fan, X.; Cai, D.; Gong, R. Fusing Residual and Cascade Attention Mechanisms in Voxel–RCNN for 3D Object Detection. Sensors 2025, 25, 5497. https://doi.org/10.3390/s25175497

AMA Style

Lu Y, Zhang Y, Fan X, Cai D, Gong R. Fusing Residual and Cascade Attention Mechanisms in Voxel–RCNN for 3D Object Detection. Sensors. 2025; 25(17):5497. https://doi.org/10.3390/s25175497

Chicago/Turabian Style

Lu, You, Yuwei Zhang, Xiangsuo Fan, Dengsheng Cai, and Rui Gong. 2025. "Fusing Residual and Cascade Attention Mechanisms in Voxel–RCNN for 3D Object Detection" Sensors 25, no. 17: 5497. https://doi.org/10.3390/s25175497

APA Style

Lu, Y., Zhang, Y., Fan, X., Cai, D., & Gong, R. (2025). Fusing Residual and Cascade Attention Mechanisms in Voxel–RCNN for 3D Object Detection. Sensors, 25(17), 5497. https://doi.org/10.3390/s25175497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop