Next Article in Journal
Evaluation of Positioning Accuracy Using Smartphone RGB and LiDAR Sensors with the viDoc RTK Rover
Previous Article in Journal
A New Sensitive Sensor Test for Capturing and Evaluating Bacteria and Viruses in Airborne Aerosols
Previous Article in Special Issue
FMCW Radar-Aided Navigation for Unmanned Aircraft Approach and Landing in AAM Scenarios: System Requirements and Processing Pipeline
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Fusion of LiDAR Features for 3D Object Detection in Autonomous Driving

1
Image Processing Group, TSC Department, Polytechnic University of Catalonia (UPC), 08034 Barcelona, Spain
2
Key Laboratory of Advanced Manufacturing and Intelligent Technology Ministry of Education, School of Mechanical and Power Engineering, Harbin University of Science and Technology, Harbin 150080, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 3865; https://doi.org/10.3390/s25133865 (registering DOI)
Submission received: 30 May 2025 / Revised: 16 June 2025 / Accepted: 18 June 2025 / Published: 21 June 2025
(This article belongs to the Special Issue Sensor Fusion in Positioning and Navigation)

Abstract

:
In the field of autonomous driving, cooperative perception through vehicle-to-vehicle communication significantly enhances environmental understanding by leveraging multi-sensor data, including LiDAR, cameras, and radar. However, traditional early or late fusion methods face challenges such as high bandwidth and computational resources, which make it difficult to balance data transmission efficiency with the accuracy of perception of the surrounding environment, especially for the detection of smaller objects such as pedestrians. To address these challenges, this paper proposes a novel cooperative perception framework based on two-stage intermediate-level sensor feature fusion specifically designed for complex traffic scenarios where pedestrians and vehicles coexist. In such scenarios, the model demonstrates superior performance in detecting small objects like pedestrians compared to mainstream perception methods while also improving the cooperative perception accuracy for medium and large objects such as vehicles. Furthermore, to thoroughly validate the reliability of the proposed model, we conducted both qualitative and quantitative experiments on mainstream simulated and real-world datasets. The experimental results demonstrate that our approach outperforms state-of-the-art perception models in terms of mAP, achieving up to a 4.1% improvement in vehicle detection accuracy and a remarkable 29.2% enhancement in pedestrian detection accuracy.

1. Introduction

With the rapid development of autonomous driving technologies, self-driving cars [1,2,3] are gradually entering everyday life and becoming an essential component of future intelligent transportation systems. Emerging paradigms such as intelligent connected vehicles, vehicle–road cooperation, vehicular networks, and smart mobility are reshaping the automotive industry and accelerating the advancement of modern transportation systems [4,5]. Among the core technologies enabling autonomous driving, LiDAR point clouds play a crucial role in environmental perception by supporting accurate 3D object detection and precise localization in real-world scenarios. However, conventional 3D object detection algorithms often struggle in complex environments, particularly with small or occluded objects, resulting in reduced detection accuracy.
In recent years, the reliability of vehicle-to-vehicle (V2V) collaborative perception algorithms [6,7,8] has significantly improved, largely driven by advancements in neural network architectures and the intelligent fusion of multi-modal sensor data, such as LiDAR, images, and radar. Compared to single-vehicle perception, V2V collaboration allows multiple connected autonomous vehicles (CAVs) to share and integrate complementary sensory information across different viewpoints. This collaborative approach addresses limitations caused by occlusion and field-of-view constraints, improving global perception performance in dynamic traffic environments. Furthermore, sophisticated feature fusion strategies have demonstrated strong robustness in recognizing objects, even under conditions of adverse weather and congested traffic [9].
Current cooperative perception fusion methods among vehicles are mainly categorized into three types: early fusion, late fusion, and intermediate fusion [10]. These fusion strategies differ significantly in terms of sensor data redundancy, total data volume, and the effectiveness of fusion results. In early fusion approaches, raw sensor data from different connected autonomous vehicles (CAVs) are aggregated to build a global driving environment perspective [11]. Although such methods have demonstrated remarkable performance in addressing occlusion and field-of-view limitations inherent in single-vehicle perception, they come at the cost of high communication resource demands. The heavy transmission load and large volume of shared data can lead to communication network congestion and latency, thereby affecting the usability and stability of the models in real-world applications. Under the premise of limited communication bandwidth, early fusion becomes increasingly impractical and inefficient in complex traffic scenarios with large data volumes, ultimately constraining the effectiveness of perception performance.
Late sensor data fusion methods (late fusion) [12,13] achieve global collaborative perception by merging the perception results independently generated by individual vehicles. Compared with early fusion methods, late fusion only requires the transmission of processed detection results, allowing each vehicle to independently process its own sensor data, then perform unified data fusion afterward. This approach facilitates system modularization and enables autonomous detection and decision-making by individual vehicles, thereby reducing dependence on real-time, high-bandwidth communication. However, current collaborative perception approaches based on late fusion rely heavily on the local perception results of individual vehicles rather than aggregated global data. If all participating vehicles were able to share sensor data, more statistically meaningful data processing would be possible, leading to more accurate detection and tracking of objects in the environment. Therefore, to achieve optimal overall performance, it is essential to consider the global nature of sensor data within the perception range and perform thorough and effective fusion accordingly.
Intermediate-level feature fusion [14,15,16] refers to the extraction of intermediate feature maps within each connected autonomous vehicle (CAV) using a predictive model, followed by the filtering and aggregation of these features in the intermediate feature space. Unlike early fusion methods that require the transmission of raw sensor data, intermediate fusion techniques only transmit these processed feature maps to other CAVs or to edge computing servers in the infrastructure. These intermediate features are then fused and decoded by each autonomous vehicle to generate final perception results. As a compromise in V2V cooperative perception strategies, intermediate-level fusion has the potential to significantly reduce inter-vehicle communication bandwidth requirements compared to early fusion while also demonstrating strong performance in enhancing perception accuracy [17,18]. Compared to late fusion methods, this approach avoids the limitations caused by reliance on local perception results from individual vehicles by efficiently compressing representative global information of the environment into intermediate features, thereby achieving a better trade-off between transmission efficiency and perceptual effectiveness.
Based on the aforementioned challenges in perception accuracy and bandwidth constraints, this paper introduces a novel collaborative perception framework through an efficient Two-Stage Intermediate-level Feature Fusion (TS-IFF) strategy. The proposed framework focuses on the effective aggregation of multi-scale features while maintaining low communication overhead. By integrating a dynamic fusion model, TS-IFF enables adaptive and robust feature combination, leading to enhanced 3D object detection performance in complex traffic scenarios. The key contributions of this work are summarized as follows:
  • We design a collaborative perception architecture based on a novel TS-IFF framework that hierarchically fuses intermediate features to balance perception accuracy and communication bandwidth.
  • To enhance the detection of small and occluded objects, we propose a dual-branch fusion strategy that combines high-resolution pseudo-image features with contextual intermediate-level features for richer semantic representation.
  • We introduce a Dynamic Weight Learning Mechanism (DWLM) to learn fusion weights for different feature types and develop an Adaptive Feature Selection Module(AFSM) to selectively aggregate the most informative components during the fusion process.

2. Related Work

2.1. 3D Object Detection

Accurate object perception is crucial for safety in autonomous driving. The current leading 3D object detection models primarily use deep learning on 3D point clouds, a key area in 3D object detection [19]. These models directly process raw point-cloud data to reduce information loss and utilize 3D geometry. PointNet [20] achieves end-to-end recognition through point-wise feature extraction and global pooling. To improve local geometric modeling, DGCNN [21] introduces dynamic compositional convolution via a graph convolutional network that enables point adjacency adjustments. Transformer-based models like Point Transformer [22] further improve accuracy by integrating global and local information with self-attention. These methods refine feature extraction and point-cloud representation by utilizing sparse structures to balance computational efficiency and information preservation. Techniques such as the use of anchor points [23] and center strategies [24,25] improve accuracy and real-time performance. In addition, BirdNet+ adopts a BEV-based approach using Faster R-CNN to directly predict 3D object boxes, achieving competitive accuracy and efficiency on the KITTI and nuScenes datasets [26]. These results highlight the effectiveness of compact BEV representations for real-time 3D detection across diverse environments.
While using point clouds preserves 3D information, data sparsity, especially at long distances or in complex environments, poses a challenge for feature extraction. Sparse distributions hinder the network’s ability to generate accurate feature representations, impacting detection accuracy and robustness. Solutions such as voxelization and bird’s-eye-view projection are used to improve the geometry of LiDAR point clouds. For example, VoxelNet [27] encodes voxel features with PointNet++ [28] and applies a region proposal network, while SECOND [29] boosts performance with sparse convolution. CenterPoint refines the backbone outputs into feature maps and predicts object centers from heat maps. However, in real urban driving environments (with obstacles such as buildings, trees, and traffic signs), individual vehicle perception from a single point of view is prone to occlusion, leading to information loss or misclassification [30,31]. Therefore, the integration of sensor data from multiple CAVs is a promising approach to improve 3D object recognition in real traffic conditions.

2.2. Cooperative Perception

To overcome the limitations of single-vehicle perception in complex environments, cooperative perception with multiple AVs has become widely adopted [32]. LiDAR and camera data from surrounding vehicles or roadside infrastructures are important sources for sharing observations in cooperative perception. Intermediate-level feature fusion provides a balance between performance and efficiency by effectively merging features from nearby vehicles. F-Cooper [33], the first intermediate collaborative perception system, uses feature-level fusion by taking the maximum value of overlapping regions. Based on this, CoFF [34] addresses F-Cooper’s disregard for low-confidence features. Attention mechanisms, including visual transformers such as V2X-ViT [35] and CoBEVT [36], further improve the relationships between features. In high-resolution detection, MSwin [37] captures spatial interactions over large distances, while AttFusion [38] applies self-attention to specific spatial locations. AdaFusion [39] introduces adaptive fusion models with trainable neural networks. CORE [12] reconstructs incomplete scenes perceived by a single vehicle into a comprehensive view using a compressor, an attention module, and a reconstruction module. However, most existing cooperative perception methods focus on merging a single type of intermediate features, overlooking the benefits of combining multiple feature types. Therefore, we propose a novel perceptual model that integrates intermediate features across different stages.

3. Overall Network Architecture

The overall structure of the network is shown in Figure 1, which can be divided into the following five parts:
  • Data Generation: Following the methodology of [38], a spatial graph is first constructed to effectively integrate the relative poses and geographic locations of each connected and autonomous vehicle (CAV), enabling reliable sharing of localization information across the network. Then, the LiDAR data from each CAV in the network is projected onto a unified reference self-coordinate plane for alignment. The aligned point cloud features are broadcast to all participating CAVs in the cooperative perception system, forming the initial stage of inter-vehicle feature interaction and preparing for the next phase of point-cloud encoding and extraction.
  • Feature Encoding and Extraction: Each CAV processes the received point-cloud features using a combination of a Voxel Feature Encoding (VFE) module and a PointPillar-based feature extraction network. The VFE module generates voxelized features with different resolutions, resulting in pseudo-images. These pseudo-images from different viewpoints are handled in two ways: (1) they are broadcast to a central dynamic fusion module to be integrated with the intermediate-level features from the ego CAV, and (2) they are retained locally to enable the extraction of intermediate features by the CAV itself. This stage enables distributed local encoding and centralized fusion interactions. (Section 3.1).
  • Feature Projection: A Feature Pyramid Network (FPN) [40] is used to extract intermediate features from the pseudo-images. The network follows a top-down structure, first extracting semantic features through downsampling blocks with 2D convolution, batch normalization, and ReLU activation, then processing them through upsampling and lateral connections to generate multi-scale intermediate-level features. The projected features are unified in channel dimension, concatenated, and transmitted to the feature fusion module. Through the Dynamic Weight Learning Mechanism (DWLM), local pseudo-image features are adaptively fused, enabling fine-grained feature interaction across multiple CAVs. (Section 3.2).
  • Feature Fusion: All pseudo-images and intermediate-level features from participating CAVs are aggregated via the proposed dynamic fusion strategy. The system performs cross-agent feature integration by assigning adaptive weights to each feature channel based on its contribution. The Adaptive Feature Selection Module (AFSM) refines the joint features further to ensure that the final representation maintains discriminative cues from both local and shared contexts. (Section 3.3).
  • Object Detector: Finally, a standard Single-Shot Detector (SSD) network [41] is applied to the fused intermediate features to classify 3D objects and regress their locations. The end-to-end detection result is enhanced by the preceding multi-agent collaborative encoding and fusion steps.

3.1. Feature Encoding and Extraction

We used the VFE module from [27] to project the original point cloud onto the bird’s-eye-view (BEV) plane. This process involves calculating the 3D ( X , Y , Z ) indices of each point and transforming point-level features into voxel-level features, represented as a four-dimensional tensor, i.e., V R C × H × W × Z . To further process these features, we integrated the PointPillars method [31], which reorganizes the tensor by collapsing the Z dimension through scatter operations and pooling, resulting in a columnar structure. Essentially, PointPillars treats vertical columns (pillars) on the BEV plane as spatial bins, aggregating and encoding features from all points within the same pillar to create a dense 2D pseudo-image ( F p R C × H × W ) that effectively represents the 3D point cloud.
The pseudo-image generated by PointPillars and the intermediate features extracted by an FPN differ fundamentally in structure and representation. PointPillars converts the raw point cloud into a 2D pseudo-image by dividing the space into vertical columns and applying PointNet to each pillar. This process compresses the 3D spatial information into a BEV feature map, emphasizing efficiency and regular grid alignment suitable for 2D convolution.
In contrast, the intermediate features extracted via an FPN operate on multi-scale hierarchical representations of the input, often preserving richer semantic and spatial context across resolutions. When applied to point-cloud data (e.g., using sparse convolution backbones), FPN features retain more local geometric details and cross-scale dependencies, which are essential for detecting objects of varying sizes and densities in 3D space. In summary, while PointPillars emphasizes structured efficiency via BEV pseudo-images, FPN-derived features focus on multi-level abstraction and geometric richness, often at a higher computational cost but with improved accuracy in complex scenes.
To optimize the input resolution of the pseudo-image, we adjusted the voxel size, experimenting with values ranging from 0.4 m down to 0.12 m, which controls the dimensions [ C , H , W ] of the pseudo-image. Our experiments indicate that higher pseudo-image resolution improves the performance of downstream feature fusion-based object detection tasks. However, when extracting intermediate-level features from the pseudo-image using the FPN [40], the downsampling modules produce intermediate features with a fixed output resolution. Thus, the spatial resolution of the intermediate FPN features remains unchanged, despite variations in the resolution of the input pseudo-image. A schematic diagram illustrating the point-cloud feature encoding process is provided in Figure 2.

3.2. Feature Fusion and Object Detection

Pseudo-images generated from raw point-cloud data effectively capture the spatial structure of the environment, preserving detailed geometric information, while intermediate-level features extracted from point clouds provide rich multi-scale contextual semantics. In this paper, we propose a novel collaborative perception fusion strategy that adaptively integrates these two types of features, fully exploiting their complementary strengths in feature representation. The fused feature maps significantly enhance the accuracy of 3D object detection, particularly in complex environments involving small and distant targets. By incorporating both pseudo-images and intermediate-level features, the proposed fusion strategy diversifies the feature representation and improves detection robustness, outperforming other methods that rely solely on intermediate features.
The fusion process is carried out in two stages. In the first stage, a set of fusion weights ( W ) is generated by a DWLM, which dynamically adjusts and optimizes the contributions of the different pseudo-images and intermediate features based on their relevance. In the second stage, inspired by the structure of the SENet module [42], we propose the AFSM to define feature mappings by selecting and fusing channel information. These fusion weights are used to effectively integrate and refine all pseudo-images and intermediate features from all cooperating CAVs. This two-stage approach ensures optimal spatial and semantic fusion of features and significantly improves the model’s ability to perform accurate object detection in diverse and challenging driving scenarios.

3.3. Dynamic Weight Learning Mechanism

The DWLM is shown in Figure 3. Before fusion, we concatenate pseudo-images ( F n P ) with intermediate-level features ( F n I ) (where n identifies the CAV). Before concatenation, intermediate-level features are upsampled to match the resolution of the pseudo-images. The final concatenated feature corresponds to a tensor ( F concat R N × ( C + C * ) × H × W , where N represents the total number of fused CAVs and C and C * denote the channel numbers of the pseudo-images and intermediate-level features, respectively). Subsequently, global average pooling is applied to F concat to reduce the dimensionality in the last two dimensions, resulting in the feature vector ( S R N × ( C + C * ) ).
Adaptive fusion weights are learned based on the channel-wise aggregated statistics, allowing the network to emphasize more informative modalities or feature levels. This vector ( S ) is passed through two fully connected layers of the same dimension to learn the importance of each channel, thereby producing the fusion weight vector ( W R 2 N ). Certain channels may focus more on edge structures, dense regions, or local geometric features, which are differently captured by pseudo-images and intermediate features. Thus, the network is trained to automatically determine the appropriate balance between them, depending on their semantic richness and discriminative capacity. The final weight vector ( W ) is divided into two N-dimensional sub-vectors, controlling the fusion ratio of pseudo-image and intermediate-level features, respectively. To ensure training stability of the weight learner, softmax normalization is applied to W , yielding 2 N adaptive fusion weights. These weights control the contribution ratios of the pseudo-image features and intermediate-level features in the final fused representation. Finally, the input features are linearly weighted and fused based on the learned fusion weights. This approach enables efficient and robust integration of features across modalities and CAVs.

3.4. Adaptive Feature Selection Module

During the feature fusion stage, the input is a 4D tensor ( F I R N × C * × H × W ). To extract the importance of each channel, the AFSM applies global average pooling and global max pooling to the input feature, thereby generating channel attention weights. The structure of the AFSM is illustrated in the upper part of Figure 4. The channel weights ( F h * R N × 1 × 1 × 1 ) are applied to the input tensor ( F I ) via channel-wise multiplication. The enhanced features are then linearly fused with DWLM-learned weights to form the fused feature ( F * R N × C × H × W ). Subsequently, a 2D convolutional neural network (2D CNN) with channel compression is applied to refine the spatial dimensions and generate the final fused feature map ( F IP R 1 × C × H × W ). This operation preserves global information while standardizing the output dimensions, thereby improving the adaptability and efficiency of the network. The overall framework for multi-scale feature fusion is shown in Figure 4, and the complete multimodal fusion procedure is described in Algorithm 1. Finally, the fused feature map ( F IP ) is fed into an SSD detection head [41] to perform 3D object detection, including bounding-box localization and confidence score classification.
Algorithm 1: Adaptive Spatial and Channel Feature Fusion
Input: Feature map F I R N × C * × H × W , F concat R N × ( C + C * ) × H × W
Output: Fused feature F I P R 1 × C × H × W
AFSM: Channel Attention Branch
      1. h avg GAP ( F I ) ;
      2. h max GMP ( F I ) ;
      3. F h Concat ( h avg , h max ) ;
      4. F h * σ ( ReLU ( F h ) ) ;
      5. F I * R N × C × H × W F I F h * ;
      6. F I * R 1 × C × H × W ReLU ( Conv 3 D ( F I * ) ) ;
DWLM: Spatial Attention Branch
      7. s GAP ( F concat ) ;
      8. x ReLU ( FC 1 ( s ) ) ;
      9. w Softmax ( FC 2 ( x ) ) ;
      10. w reshape ( w , N , C , 1 , 1 ) ;
      11. F concat * R N × C × H × W F concat w ;
Feature Fusion
      12. F * F I * F concat * ;
      13. F I P ReLU ( Conv 2 D ( F * ) ) ;
return  F I P ;

3.5. Loss Function

The TS-IFF network proposed in this paper employs the loss function introduced in [27]. The total loss ( L total ) is composed of a classification loss and a regression loss:
L total = α L cls pos + β L cls neg + L reg
where α and β are positive constants that balance the relative importance and L cls pos and L cls neg denote the classification losses for positive and negative samples. The terms L cls pos and L cls neg are defined as follows:
L cls pos = 1 N p i = 1 N p L cls ( p i pos , 1 )
L cls neg = 1 N n j = 1 N n L cls ( p j neg , 0 )
where p i pos and p j neg are the softmax output probabilities for positive samples and negative samples, respectively. N p and N n denote the counts of positive and negative samples. L r e g is the regression loss, which we define as follows:
L r e g = 1 N p i = 1 N p L 1 ( u i u ^ i )
where u i and u ^ i represent the regression ground truth and the predicted position, respectively, and L 1 denotes the smooth L 1 function:
L 1 ( x ) = x 2 2 if | x | < 1 | x | 1 2 if x < 1 x > 1

4. Experimental Results

To evaluate the proposed model, we conducted targeted experiments separately on both simulated and real-world datasets: OPV2V [38] and V2V4Real [43] were utilized to assess cooperative perception capabilities, while CODD [44] was specifically used to evaluate performance in detecting small objects such as pedestrians. Additionally, extensive ablation studies and benchmark comparisons were carried out to demonstrate the superiority and effectiveness of the proposed cooperative perception model compared to existing state-of-the-art methods.

4.1. Datasets

The OPV2V dataset [38] is a simulation dataset that contains two subsets: Default Towns (DT) and Culver City (CC). The DT subset consists of data from eight default towns provided by CARLA [45] and contains on average about three CAVs per frame, with a minimum of two and a maximum of seven vehicles. The data in this subset was formally divided into a training set (6.7 K frames), a validation set (2 K frames), and a test set (2.7 K frames). The CC subset includes an independent test set of 550 frames to evaluate the model’s ability to generalize to new scenarios. All scenes last approximately 16.4 s and were captured using 64-channel LiDAR, generating approximately 1.3 million point clouds per second. This dataset simulates diverse urban driving conditions, including dynamic traffic flow, occlusions, and varying vehicle densities, providing a comprehensive benchmark for evaluating cooperative perception algorithms.
The CODD dataset [44] was also created with the help of the CARLA simulation platform. It contains 108 scene clips from eight different CARLA towns. For comparison with other methods, we use the same methodology as in [39]. Each scene consists of 125 frames, of which the first 100 frames are used for model training and the remaining 25 frames are used for testing. A notable feature of this dataset is that it includes varying numbers of vehicles and pedestrians, with the number of vehicles ranging from 4 to 15 and the number of pedestrians ranging from 2 to 8. CODD is the only collaborative sensing dataset that currently includes a pedestrian population. This diversity in participant types introduces additional complexity to the perception task, making it well-suited for evaluating models’ ability to detect and distinguish between heterogeneous traffic agents. Moreover, CODD provides detailed annotations for both vehicles and pedestrians, enabling fine-grained performance analysis across object categories and contributing to more realistic assessments of cooperative perception systems.
V2V4Real [43] is the first large-scale, publicly available real-world dataset for V2V cooperative perception, collected in Columbus, Ohio, across highways and urban streets. It includes 19 h of driving data with 310 K frames, from which 67 representative scenarios (10–20 s each) were selected. LiDAR and RGB frames were sampled at 10 Hz, yielding 20 K LiDAR point clouds and 40 K images. The dataset features high-density LiDAR point clouds and 240 K precisely annotated 3D bounding boxes for five classes. Sensor asynchronization between vehicles was kept below 50 ms. This dataset presents real-world challenges such as sensor noise, occlusion, and asynchronous multi-vehicle coordination, making it a valuable benchmark for validating the robustness and adaptability of cooperative perception models. Its diverse driving environments and dense traffic scenarios further enhance its utility for evaluating performance under complex real-world conditions.

4.2. Implementation Details

Our model was implemented with the PyTorch v1.7.1 framework [46] and trained and tested on a GeForce RTX 3090 GPU (NVIDIA Corporation, Santa Clara, CA, USA). The GPU has 24 GB RAM and runs in a CUDA v11.1 environment combined with cuDNN v8.0 for acceleration, ensuring efficient computation during inference. During the training process, the model uses a learning rate scheduler and an early stopping mechanism, and the optimizer was chosen to be Adam, with parameters set to ε = 0.1 and a weight decay factor of 10 4 . We trained the model for 30 epochs, and the model parameters were updated with a batch size of 2 and a learning rate of 2 × 10 3 . The momentum was set to a value between 0.85 and 0.95. During the inference process, we filtered low-confidence bounding boxes with a threshold of 0.3 and used a non-maximum suppression strategy to remove overlapping candidates by setting the IoU threshold to 0.2.
The driving scenario was selected at any time during the following training process, and the number of CAVs was selected in the interval of [ 2 , 7 ] , where the center vehicle is included in the interval as the EGO car (the car that receives all collaborative features). The number of CAVs was fixed for all scenes to ensure the fairness of the experiment. For data generation, we used the same parameters from [38,43] and set the range of LiDAR point clouds to [ 3 , 1 ] × [ 140.8 , 140.8 ] × [ 40 , 40 ] meters as the range of z , x , y values for both OPV2V and V2V4Real. Similarly, for the CODD dataset, the range was set to [ 6 , 4 ] × [ 140.8 , 140.8 ] × [ 40 , 40 ] meters. All datasets use the same body-column resolution of 0.4 m, which corresponds to a tensor size of [ H × X ] = [ 704 × 200 ] meters. For the SSD detection module, we used a vehicle anchor length, width, and height of [ 3.9 , 1.6 , 1.56 ] meters, a pedestrian anchor length, width and height of [ 0.6 , 0.6 , 1.7 ] meters and an anchor-box rotation angle range of [ 0 , 90 ] degrees.
In autonomous driving 3D point-cloud feature fusion experiments, we utilized AP metrics to provide a comprehensive evaluation of detection performance. AP captures the balance between precision and recall across varying confidence thresholds. Specifically, AP0.5 and AP0.7 correspond to the average precision computed at IoU thresholds of 0.5 and 0.7, respectively, which are commonly used to assess the detection accuracy of larger objects such as vehicles. For smaller or more variable targets like pedestrians, we adopted a lower IoU threshold, AP0.1, to more appropriately evaluate the model’s detection capabilities. By incorporating AP metrics at different IoU thresholds, we achieved a more thorough and nuanced assessment of the model’s effectiveness across diverse object categories and scales, thereby offering deeper insights into its strengths and limitations.

4.3. Experimental Results

Table 1 shows the quantitative experimental results of our proposed model on the simulated OPV2V and CODD datasets. Using the detection of individual vehicles without collaborative sensing as a baseline, we benchmark our model against the SOTA methods. Of the state-of-the-art methods listed in the table, DiscoNet, CoBEVT, and HM-ViT are specifically designed for the features and scenarios of the DT sub-dataset and the V2X-Sim dataset. In contrast, the CC sub-dataset and CODD datasets contain more diverse transformations, complex scenarios, and small targets, which fall outside the optimal application conditions for these methods. In summary, the comparison methods selected in this work are representative and closely related to our task. They are evaluated on similar datasets and metrics to ensure a fair comparison. Their implementations and results are publicly available, supporting reproducibility and meaningful benchmarking.
The experimental results show that our model achieves an improvement in AP of up to 65 % over the baseline. In particular, for object detection of surrounding vehicles, our model shows an AP improvement of about 2 % to 4 % over the SOTA for both the OPV2V and CODD datasets. In particular, for the DT and CC subsets of the OPV2V dataset, our model achieves a detection accuracy of over 90 % for both the AP@0.5 and AP@0.7 thresholds. For small-object pedestrian detection within the CODD dataset, our model achieves an accuracy of over 60 % , which is a significant increase of 29.2 % AP compared to the state-of-the-art best method. Moreover, to better adapt to the complexity of the real world and enhance perception consistency and decision reliability, we conducted model testing on the real-world V2V4Real dataset. As shown in Table 2, our method achieved AP values of 68.2% and 40.1% at thresholds of 0.5 and 0.7, respectively, for vehicle detection, outperforming other methods. Compared to the second best approach, our model demonstrated a performance improvement of 2.5% to 8% across different thresholds, further demonstrating the model’s outstanding performance and clear advantages in detecting surrounding vehicles in cooperative perception tasks.
The results presented in Figure 5 show the evident trend in collaborative perception: as the number of CAVs in the collaborative perception network increases (up to seven CAVs in the OPV2V dataset and five CAVs in the CODD dataset), there is a significant improvement in detection performance. The vehicle detection accuracy (AP@0.5) improved by 26.5% and 31.9% in the two datasets, respectively, while the pedestrian detection accuracy (AP@0.1) increased by 61.9% in the CODD dataset. Meanwhile, we also conducted extensive experiments on the CODD dataset, which contains more pedestrians. The qualitative results are shown in Figure 6, demonstrating that in driving scenarios with blind spots, we can successfully detect pedestrians through collaborative perception. These results show that the detection of small objects benefits significantly from collaborative perception and that our proposed method significantly improves the detection of these small objects.
Figure 7 and Figure 8 present the qualitative visualization results on the simulated DT and CODD simulated, as well as the V2V4Real dataset, showcasing multi-vehicle collaborative perception in simulated driving scenarios. As observed in Figure 7, when relying solely on a single central vehicle (without fusion), certain objects in the scene may be misidentified as vehicles, while some distant targets may be entirely missed due to occlusion. As more collaborative autonomous vehicles (CAVs) participate in cooperative perception, the central vehicle gains an expanded field of view and richer sensor data, enabling more accurate and reliable detection of distant objects while reducing false positives and missed detections. Consequently, our model demonstrates outstanding robustness in both simulated and real-world scenarios.
Additionally, in the simulated OPV2V dataset (Figure 7a) includes typical qualitative examples where sparse LiDAR inputs lead to occasional false positives and a slight degradation in detection performance. While the model generally performs well in identifying vehicle targets, in scenarios with extreme sparsity or missing information, some non-vehicle objects may be mistakenly classified as vehicles, and the detection accuracy for distant or occluded targets is somewhat reduced. This highlights the importance of multi-sensor fusion and collaborative perception in enhancing the comprehensiveness of scene understanding. In the analysis of the real-world V2V4Real dataset, visualization results reveal that factors such as occlusion, sparse object distribution, and sensor noise in real environments can still affect detection outcomes. Occlusion causes partial loss of point-cloud information in certain areas, increasing the difficulty of accurate recognition, while sensor noise may lead to occasional false detections or uncertainties.

4.4. Ablation Study

To evaluate the impact of the proposed AFSM and DWLM on 3D object detection performance, we conducted a series of ablation experiments with seven CAVs for the OPV2V dataset and three CAVs for the CODD dataset. In the baseline setup, we excluded the AFSM and DWLM and directly fused the intermediate-level features via simple concatenation without generating pseudo-images. We then incrementally activated each module (DWLM and AFSM) to evaluate their individual contributions. All experiments were performed using an SSD detection head. The results, summarized in Table 3, indicate significant performance improvements when the DWLM and AFSM are integrated. In particular, for the OPV2V dataset, vehicle detection accuracy increased by 9% at AP@0.5, while pedestrian detection accuracy improved by 18.5% at AP@0.1 for the CODD dataset. We provide further statistics on the model’s inference time compared to baseline methods, demonstrating that the proposed approach achieves notable performance gains while keeping computational overhead within a practical and acceptable range, supporting its feasibility for real-world deployment.
In a subsequent ablation study, we investigated the impact of fusing pseudo-images with different resolutions on 3D object detection in the environment. Consistent with previous experiments, we used seven CAVs for the OPV2V dataset as the upper bound for the ablation study and three CAVs for the CODD dataset. The results presented in Table 4 show that in the OPV2V dataset, the fusion of intermediate features with pseudo-images achieves an average precision of 94.1% for vehicle detection at AP@0.5, representing an improvement of 17.4%. For the CODD dataset, which focuses specifically on pedestrian detection, the fusion of intermediate features with pseudo-images achieves an average precision of 69.1% for small-object detection (e.g., pedestrians) at AP@0.1—nearly a 40% improvement compared to the baseline experiment.
Overall, the results presented in Table 4 indicate that increasing the resolution of pseudo-images consistently improves detection accuracy, especially for small objects such as pedestrians. However, the performance gains tend to plateau after reaching a certain resolution, with diminishing returns and a limited impact on overall perception performance from further increases. Therefore, in practical applications, selecting an appropriate resolution is crucial to achieving optimal system performance.
In V2V collaborative perception networks, communication bandwidth serves as a vital factor that directly influences both the speed and efficiency of information transmission between connected autonomous vehicles. To thoroughly assess the performance of our proposed method in terms of network communication bandwidth consumption, we conducted a series of detailed ablation experiments using two representative simulated datasets. These experiments aimed to explore and analyze the relationship between detection performance and bandwidth requirements. The results are presented in Figure 9a. It is evident that our method achieves the highest target recognition accuracy, though this comes with relatively high bandwidth use. However, as the feature resolution decreases, an inevitable but acceptable decline in AP is observed. We speculate that more aggressive downsampling leads to greater loss of keypoint information, reducing recognition accuracy. From the perspective of collaborative perception, the trade-off between performance and bandwidth, as shown in Figure 9b, is reasonable.
Moreover, in the small-target pedestrian recognition experiment, when we reduced the number of feature channels and the resolution to half of the original values, the resulting detection accuracy and bandwidth overhead reached an optimal balance. This also suggests that even with simple downsampling for feature compression, our model can still maintain optimal recognition accuracy for small-target detection.

5. Conclusions

In this paper, we introduce a novel perception architecture, TS-IFF, which integrates multiple feature types to improve the effectiveness of collaborative perception. Specifically, we propose a two-stage intermediate feature fusion strategy that optimizes and integrates intermediate features across different levels to enhance perception performance. Additionally, we designed a feature weight learning mechanism to adaptively fuse high-resolution pseudo-images with intermediate features. Pseudo-images preserve the spatial structure and geometry of point clouds, while intermediate features capture multi-scale contextual semantics at multiple levels. Experimental results demonstrate that the TS-IFF model excels in detecting small 3D objects, such as pedestrians, while maintaining lightweight bandwidth requirements. This effectively addresses the limitations of traditional non-fusion methods under occlusion and bandwidth constraints.
While our method achieves a good balance between perception accuracy and communication efficiency, its performance in extremely complex urban scenarios still faces robustness challenges. In future work, we plan to further reduce bandwidth consumption and improve system robustness by developing a more efficient autoencoder-based encoding and decoding mechanism, enabling optimal compression of features while preserving critical perception performance.

Author Contributions

Conceptualization, M.W., J.R.-H. and J.R.C.; methodology, M.W., J.R.-H. and J.R.C.; software, M.W., J.R.-H. and J.R.C.; validation, M.W., J.R.-H., J.R.C. and D.L.; formal analysis, J.R.-H.; investigation, J.R.C.; resources, J.R.C.; data curation, M.W. and J.R.C.; writing—original draft preparation, M.W.; writing—review and editing, M.W., J.R.-H. and J.R.C.; visualization, M.W.; supervision, J.R.-H., J.R.C. and D.L.; project administration, J.R.-H. and J.R.C.; funding acquisition, J.R.-H. and J.R.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Spanish Research Agency (AEI) under project PID2020-117142GB-I00 of call MCIN/AEI/10.13039/501100011033 and by the 6G-EWOC project of the Smart Networks and Services Joint Undertaking (SNS JU) under the European Union’s Horizon Europe research and innovation programme under Grant Agreement No. 101139182.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to privacy restrictions.

Acknowledgments

The authors are grateful for the financial support provided by the Program of the China Scholarship Council (Grant No. 202308230300).

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Abbreviations

The following abbreviations are used in this manuscript:
V2VVehicle to vehicle
AVsAutonomous vehicles
CAVsConnected autonomous vehicles
TS-IFFTwo-stage intermediate-level feature fusion
DWLMDynamic weight learning mechanism
AFSMAdaptive feature selection module
SSDSingle-shot detector
FPNFeature pyramid network
VFEVoxel feature encoding
BEVBird’s-eye view
3D CNN3D convolutional neural network
2D CNN2D convolutional neural network
DTDefault Towns
CCCulver City
GTGround truth
PIPseudo-image

References

  1. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-voxel feature set abstraction with local vector representation for 3D object detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  2. Fang, J.; Zhou, D.; Yan, F.; Zhao, T.; Zhang, F.; Ma, Y.; Wang, L.; Yang, R. Augmented LiDAR simulator for autonomous driving. IEEE Robot. Autom. Lett. 2020, 5, 1931–1938. [Google Scholar] [CrossRef]
  3. Caillot, A.; Ouerghi, S.; Vasseur, P.; Boutteau, R.; Dupuis, Y. Survey on cooperative perception in an automotive context. IEEE Trans. Intell. Transp. Syst. 2022, 23, 14204–14223. [Google Scholar] [CrossRef]
  4. Hu, X.; Chen, T.; Zhang, W.; Ji, G.; Jia, H. MonoAMP: Adaptive multi-order perceptual aggregation for monocular 3D vehicle detection. Sensors 2025, 25, 787. [Google Scholar] [CrossRef] [PubMed]
  5. Coelho, D.; Oliveira, M.; Santos, V. RLAD: Reinforcement learning from pixels for autonomous driving in urban environments. IEEE Trans. Autom. Sci. Eng. 2023, 21, 7427–7435. [Google Scholar] [CrossRef]
  6. Xing, Y.; Lv, C.; Cao, D.; Hang, P. Toward human-vehicle collaboration: Review and perspectives on human-centered collaborative automated driving. Transp. Res. C Emerg. Technol. 2021, 128, 103199. [Google Scholar] [CrossRef]
  7. Li, J.; Xu, R.; Liu, X.; Ma, J.; Chi, Z.; Ma, J.; Yu, H. Learning for vehicle-to-vehicle cooperative perception under lossy communication. IEEE Trans. Intell. Veh. 2023, 8, 2650–2660. [Google Scholar] [CrossRef]
  8. Du, X.; Ang, M.H.; Rus, D. Car detection for autonomous vehicle: LIDAR and vision fusion approach through deep learning framework. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 749–754. [Google Scholar]
  9. Wang, K.; Pu, L.; Zhang, J.; Lu, J. Gated adversarial network based environmental enhancement method for driving safety under adverse weather conditions. IEEE Trans. Intell. Veh. 2022, 8, 1934–1943. [Google Scholar] [CrossRef]
  10. Lahat, D.; Adali, T.; Jutten, C. Multimodal data fusion: An overview of methods, challenges, and prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef]
  11. Su, S.; Li, Y.; He, S.; Han, S.; Feng, C.; Ding, C.; Miao, F. Uncertainty quantification of collaborative detection for self-driving. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5588–5594. [Google Scholar]
  12. Wang, B.; Zhang, L.; Wang, Z.; Zhao, Y.; Zhou, T. Core: Cooperative reconstruction for multi-agent perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 8–14 October 2023; pp. 8710–8720. [Google Scholar]
  13. Yu, H.; Luo, Y.; Shu, M.; Huo, Y.; Yang, Z.; Shi, Y.; Guo, Z.; Li, H.; Hu, X.; Yuan, J.; et al. Dair-v2x: A large-scale dataset for vehicle-infrastructure cooperative 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 21361–21370. [Google Scholar]
  14. Lin, C.; Tian, D.; Duan, X.; Zhou, J.; Zhao, D.; Cao, D. V2VFormer: Vehicle-to-vehicle cooperative perception with spatial-channel transformer. IEEE Trans. Intell. Veh. 2024, 9, 3384–3395. [Google Scholar] [CrossRef]
  15. Fang, Q.; Wang, Z. Cross-modality attentive feature fusion for object detection in multispectral remote sensing imagery. Pattern Recognit. 2022, 130, 108786. [Google Scholar]
  16. Yuan, Y.; Cheng, H.; Sester, M. Keypoints-based deep feature fusion for cooperative vehicle detection of autonomous driving. IEEE Robot. Autom. Lett. 2022, 7, 3054–3061. [Google Scholar] [CrossRef]
  17. Lei, Z.; Ren, S.; Hu, Y.; Zhang, W.; Chen, S. Latency-aware collaborative perception. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 316–332. [Google Scholar]
  18. Yang, Q.; Fu, S.; Wang, H.; Fang, H. Machine-learning-enabled cooperative perception for connected autonomous vehicles: Challenges and opportunities. IEEE Netw. 2021, 35, 96–101. [Google Scholar] [CrossRef]
  19. Chen, L.; Wang, H.; Kong, H.; Yang, W.; Ren, M. PTC-Net: Point-wise transformer with sparse convolution network for place recognition. IEEE Robot. Autom. Lett. 2023, 8, 3414–3421. [Google Scholar] [CrossRef]
  20. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  21. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 146. [Google Scholar] [CrossRef]
  22. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  23. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel r-cnn: Towards high performance voxel-based 3D object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 22 February–1 March 2021; Volume 35, pp. 1201–1209. [Google Scholar]
  24. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3D object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  25. Zhou, Z.; Zhao, X.; Wang, Y.; Wang, P.; Foroosh, H. Centerformer: Center-based transformer for 3D object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 496–513. [Google Scholar]
  26. Barrera, A.; Beltrán, J.; Guindel, C.; Iglesias, J.A.; García, F. BirdNet+: Two-Stage 3D Object Detection in LiDAR Through a Sparsity-Invariant Bird’s Eye View. IEEE Access 2021, 9, 160299–160316. [Google Scholar] [CrossRef]
  27. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3D object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
  28. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 5105–5114. [Google Scholar]
  29. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef]
  30. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  31. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  32. Cai, K.; Qu, T.; Gao, B.; Chen, H. Consensus-based distributed cooperative perception for connected and automated vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 24, 8188–8208. [Google Scholar] [CrossRef]
  33. Chen, Q.; Ma, X.; Tang, S.; Guo, J.; Yang, Q.; Fu, S. F-cooper: Feature based cooperative perception for autonomous vehicle edge computing system using 3D point clouds. In Proceedings of the 4th ACM/IEEE Symposium on Edge Computing, Bellevue, WA, USA, 28–30 October 2019; pp. 88–100. [Google Scholar]
  34. Guo, J.; Carrillo, D.; Tang, S.; Chen, Q.; Yang, Q.; Fu, S.; Wang, X.; Wang, N.; Palacharla, P. CoFF: Cooperative spatial feature fusion for 3-D object detection on autonomous vehicles. IEEE Internet Things J. 2021, 8, 11078–11087. [Google Scholar] [CrossRef]
  35. Xu, R.; Xiang, H.; Tu, Z.; Xia, X.; Yang, M.-H.; Ma, J. V2x-vit: Vehicle-to-everything cooperative perception with vision transformer. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 107–124. [Google Scholar]
  36. Xu, R.; Tu, Z.; Xiang, H.; Shao, W.; Zhou, B.; Ma, J. CoBEVT: Cooperative Bird’s Eye View Semantic Segmentation with Sparse Transformers. In Proceedings of the Conference on Robot Learning, Auckland, New Zealand, 11–14 December 2023; pp. 989–1000. [Google Scholar]
  37. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin transformer v2: Scaling up capacity and resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12009–12019. [Google Scholar]
  38. Xu, R.; Xiang, H.; Xia, X.; Han, X.; Li, J.; Ma, J. Opv2v: An open benchmark dataset and fusion pipeline for perception with vehicle-to-vehicle communication. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 2583–2589. [Google Scholar]
  39. Qiao, D.; Zulkernine, F. Adaptive feature fusion for cooperative perception using lidar point clouds. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2023; pp. 1186–1195. [Google Scholar]
  40. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  41. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  42. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  43. Xu, R.; Xia, X.; Li, J.; Li, H.; Zhang, S.; Tu, Z.; Meng, Z.; Xiang, H.; Dong, X.; Song, R.; et al. V2v4real: A real-world large-scale dataset for vehicle-to-vehicle cooperative perception. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 20–25 October 2023; pp. 13712–13722. [Google Scholar]
  44. Arnold, E.; Mozaffari, S.; Dianati, M. Fast and robust registration of partially overlapping point clouds. IEEE Robot. Autom. Lett. 2021, 7, 1502–1509. [Google Scholar] [CrossRef]
  45. Dosovitskiy, A.; Ros, G.; Codevilla, F.; Lopez, A.; Koltun, V. CARLA: An open urban driving simulator. In Proceedings of the Conference on Robot Learning, Mountain View, CA, USA, 13–15 November 2017; pp. 1–16. [Google Scholar]
  46. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  47. Wang, T.-H.; Manivasagam, S.; Liang, M.; Yang, B.; Zeng, W.; Urtasun, R. V2vnet: Vehicle-to-vehicle communication for joint perception and prediction. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 605–621. [Google Scholar]
  48. Li, Y.; Ren, S.; Wu, P.; Chen, S.; Feng, C.; Zhang, W. Learning distilled collaboration graph for multi-agent perception. Adv. Neural Inf. Process. Syst. 2021, 34, 29541–29552. [Google Scholar]
  49. Xiang, H.; Xu, R.; Ma, J. HM-ViT: Hetero-modal vehicle-to-vehicle cooperative perception with vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 20–25 October 2023; pp. 284–295. [Google Scholar]
Figure 1. Overview of a collaborative perception framework based on Two-Stage Intermediate-Level Feature Fusion (TS-IFF). The system fuses LiDAR data from multiple autonomous vehicles, as demonstrated here with three collaborating CAVs. Each point-cloud datum is voxelized to generate voxel-level pseudo-images. Pseudo-images are passed through a feature extraction layer (FPN) to extract corresponding intermediate features. Our proposed fusion module integrates the features from both stages and from all CAVs. A final single-shot detector (SSD) produces the detection results. Note that ⊕ represents concatenation. In the visual representation, differently colored arrows illustrate data flow, while bold red lines highlight connections related to the fusion of pseudo-image features.
Figure 1. Overview of a collaborative perception framework based on Two-Stage Intermediate-Level Feature Fusion (TS-IFF). The system fuses LiDAR data from multiple autonomous vehicles, as demonstrated here with three collaborating CAVs. Each point-cloud datum is voxelized to generate voxel-level pseudo-images. Pseudo-images are passed through a feature extraction layer (FPN) to extract corresponding intermediate features. Our proposed fusion module integrates the features from both stages and from all CAVs. A final single-shot detector (SSD) produces the detection results. Note that ⊕ represents concatenation. In the visual representation, differently colored arrows illustrate data flow, while bold red lines highlight connections related to the fusion of pseudo-image features.
Sensors 25 03865 g001
Figure 2. Schematic representation of point-cloud feature encoding. When using a given voxel size, the voxelization of the point cloud leads to voxel features of size [ C , H , W , Z ] . After pillarization, the Z dimension is collapsed. If a sample or pillar has too little data to populate the tensor, zero-padding is applied. Through this encoding process, high-resolution pseudo-images of the point cloud can be generated, serving as the input for subsequent feature extraction and fusion.
Figure 2. Schematic representation of point-cloud feature encoding. When using a given voxel size, the voxelization of the point cloud leads to voxel features of size [ C , H , W , Z ] . After pillarization, the Z dimension is collapsed. If a sample or pillar has too little data to populate the tensor, zero-padding is applied. Through this encoding process, high-resolution pseudo-images of the point cloud can be generated, serving as the input for subsequent feature extraction and fusion.
Sensors 25 03865 g002
Figure 3. Dynamic weight learning mechanism: the intermediate point-cloud features ( F 1 I , F 2 I , F 3 I ) of three autonomous vehicles (car1, car2, and the ego vehicle) and their pseudo-image features ( F 1 P , F 2 P , F 3 P ) are combined as input features. A cascade operation generates concatenated features ( F concat , where W and H are the feature width and height, C and C * are the channel numbers of different modality features, N is the number of fused CAVs, S is the feature vector, and W is the feature fusion weight).
Figure 3. Dynamic weight learning mechanism: the intermediate point-cloud features ( F 1 I , F 2 I , F 3 I ) of three autonomous vehicles (car1, car2, and the ego vehicle) and their pseudo-image features ( F 1 P , F 2 P , F 3 P ) are combined as input features. A cascade operation generates concatenated features ( F concat , where W and H are the feature width and height, C and C * are the channel numbers of different modality features, N is the number of fused CAVs, S is the feature vector, and W is the feature fusion weight).
Sensors 25 03865 g003
Figure 4. Adaptive feature fusion via channel and spatial weighting mechanisms. This framework uses global pooling to compress global information within the channel descriptors. We use global max pooling and global average pooling to extract the two channel descriptors ( h max R N × 1 × 1 × 1 and h avg R N × 1 × 1 × 1 ). After concatenation to obtain the channel weights ( F h R 2 N × 1 × 1 × 1 ), we obtain the input channel descriptor weights ( F h * R N × 1 × 1 × 1 ) through a linear layer with ReLU activation, where N is the maximum number of input CAVs (N is taken as 3). The learned channel feature weights are multiplied element-wise along the channel dimension with the features produced by the DWLM ( F concat ), resulting in a new feature representation ( F * R N × C × H × W ). Finally, a 2D CNN with ReLU activation is applied to obtain the fused new feature ( F IP R 1 × C × H × W ).
Figure 4. Adaptive feature fusion via channel and spatial weighting mechanisms. This framework uses global pooling to compress global information within the channel descriptors. We use global max pooling and global average pooling to extract the two channel descriptors ( h max R N × 1 × 1 × 1 and h avg R N × 1 × 1 × 1 ). After concatenation to obtain the channel weights ( F h R 2 N × 1 × 1 × 1 ), we obtain the input channel descriptor weights ( F h * R N × 1 × 1 × 1 ) through a linear layer with ReLU activation, where N is the maximum number of input CAVs (N is taken as 3). The learned channel feature weights are multiplied element-wise along the channel dimension with the features produced by the DWLM ( F concat ), resulting in a new feature representation ( F * R N × C × H × W ). Finally, a 2D CNN with ReLU activation is applied to obtain the fused new feature ( F IP R 1 × C × H × W ).
Sensors 25 03865 g004
Figure 5. Effect of CAV numbers on the accuracy of cooperative perception: curves at different thresholds in the DT and CODD datasets (viewing in color is recommended for clarity).
Figure 5. Effect of CAV numbers on the accuracy of cooperative perception: curves at different thresholds in the DT and CODD datasets (viewing in color is recommended for clarity).
Sensors 25 03865 g005
Figure 6. Visualization of the results of the TS-IFF model on the CODD dataset. The figure shows a driving scene with a blind spot where the C A R e g o is able to accurately detect pedestrians and other objects outside its field of view through collaborative perception with C A R 1 . Ground truth (GT) is denoted by green rectangles, while predictions are shown in red.
Figure 6. Visualization of the results of the TS-IFF model on the CODD dataset. The figure shows a driving scene with a blind spot where the C A R e g o is able to accurately detect pedestrians and other objects outside its field of view through collaborative perception with C A R 1 . Ground truth (GT) is denoted by green rectangles, while predictions are shown in red.
Sensors 25 03865 g006
Figure 7. TS-IFF visualization of the effects of collaborative perception as the number of CAVs increases. The figure shows the prediction results for the DT sub-dataset in subfigures (ae). Ground truth (GT) is denoted by green rectangles, while predictions are shown in red. These images are best viewed in color.
Figure 7. TS-IFF visualization of the effects of collaborative perception as the number of CAVs increases. The figure shows the prediction results for the DT sub-dataset in subfigures (ae). Ground truth (GT) is denoted by green rectangles, while predictions are shown in red. These images are best viewed in color.
Sensors 25 03865 g007
Figure 8. TS-IFF visualization results in two real-world scenarios. The ground truth (GT) is represented by green rectangles, while predictions are indicated by red rectangles. The correspondence between the GT and predictions is highlighted with yellow rectangles. These images are best viewed in color.
Figure 8. TS-IFF visualization results in two real-world scenarios. The ground truth (GT) is represented by green rectangles, while predictions are indicated by red rectangles. The correspondence between the GT and predictions is highlighted with yellow rectangles. These images are best viewed in color.
Sensors 25 03865 g008
Figure 9. Ablation results showing the relationship between the performance and bandwidth of the latest models on two datasets. (a) Results on OPV2V. (b) Results on CODD. Red and blue points represent TS-IFF performance under different resolutions. Best viewed in color.
Figure 9. Ablation results showing the relationship between the performance and bandwidth of the latest models on two datasets. (a) Results on OPV2V. (b) Results on CODD. Red and blue points represent TS-IFF performance under different resolutions. Best viewed in color.
Sensors 25 03865 g009
Table 1. Quantitative comparison of the TS-IFF model with state-of-the-art methods across two datasets. Bold highlights denote best performance, with blue values in parentheses indicating AP improvement over the second best method and red values indicating parameter increases compared to this method. Underlined values indicate second best results. ↑: Larger values are better. ↓: Smaller values are better.
Table 1. Quantitative comparison of the TS-IFF model with state-of-the-art methods across two datasets. Bold highlights denote best performance, with blue values in parentheses indicating AP improvement over the second best method and red values indicating parameter increases compared to this method. Underlined values indicate second best results. ↑: Larger values are better. ↓: Smaller values are better.
MethodOPV2VCODDPara (M)
Total (↓)
DTCCVehiclePedestrian
AP@0.5AP@0.7AP@0.5AP@0.7 (↑)AP@0.5AP@0.7AP@0.1 (↑)
Baseline69.262.455.347.660.354.424.36.58
F-Cooper [33]88.779.084.672.877.674.332.87.27
V2VNet [47]89.782.286.073.480.375.832.014.61
AttFuse [38]90.881.585.473.581.477.738.16.58
DiscoNet [48]74.159.09.66
CoBEVT [36]82.863.78.35
V2X-ViT [35]89.182.687.373.782.378.933.813.45
HM-ViT [49]85.376.317.64
AdaFusion [39]91.685.688.179.086.283.945.27.27
Ours94.1 (+2.7%)89.3 (+4.1%)90.3 (+2.4%)82.1 (+3.8%)88.6 (+2.7%)85.8 (+2.2%)63.8 (+29.2%)8.16 (+10.9%)
Table 2. Comparison of the TS-IFF model with SOTA methods for vehicle detection on the V2V4Real dataset. Bold highlights indicate the best performance, with blue values in parentheses representing the accuracy improvement over the second best method. Underlined values indicate the second best results.
Table 2. Comparison of the TS-IFF model with SOTA methods for vehicle detection on the V2V4Real dataset. Bold highlights indicate the best performance, with blue values in parentheses representing the accuracy improvement over the second best method. Underlined values indicate the second best results.
MethodV2V4Real
AP@0.5AP@0.7
Baseline39.822.0
F-Cooper60.731.8
V2Vnet64.534.3
AttFuse64.733.6
V2X-ViT64.936.9
CoBEVT66.536.0
Ours68.2 (+2.5%)40.1 (+8.0%)
Table 3. Ablation study to investigate the impact of the proposed modules (AFSM and DWLM) on network performance without the fusion of pseudo-images. Baseline represents collaborative results without these modules. Best results are highlighted in bold, ↑: Larger values are better. ↓: Smaller values are better.
Table 3. Ablation study to investigate the impact of the proposed modules (AFSM and DWLM) on network performance without the fusion of pseudo-images. Baseline represents collaborative results without these modules. Best results are highlighted in bold, ↑: Larger values are better. ↓: Smaller values are better.
ModelAFSMDWLMOPV2V (AP@0.5↑)CODD (AP@0.1↑)Para. (M)Ave Infer. Time (ms/Frame)
Baseline××69.824.76.5815.63
×70.226.67.0326.18
TS-IFF×72.127.87.2727.33
76.730.38.1629.60
Table 4. Ablation study exploring the impact of intermediate feature resolution. Baseline represents collaborative results without fusing pseudo-images. The PI column represents the use of pseudo-images and their resolution with respect to intermediate features. The best results in each setting are clearly emphasized in bold for comparison.
Table 4. Ablation study exploring the impact of intermediate feature resolution. Baseline represents collaborative results without fusing pseudo-images. The PI column represents the use of pseudo-images and their resolution with respect to intermediate features. The best results in each setting are clearly emphasized in bold for comparison.
ModelPI (C × H × W)OPV2V (AP@0.5↑)CODD (AP@0.1↑)
Baseline76.730.3
TS-IFF C × ( H / 2 ) × ( W / 2 ) 85.262.9
( C / 2 ) × ( H / 2 ) × ( W / 2 ) 73.250.3
C × H × W 94.165.9
( C / 2 ) × H × W 86.355.7
C × ( 2 H ) × ( 2 W ) 90.769.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, M.; Li, D.; Casas, J.R.; Ruiz-Hidalgo, J. Adaptive Fusion of LiDAR Features for 3D Object Detection in Autonomous Driving. Sensors 2025, 25, 3865. https://doi.org/10.3390/s25133865

AMA Style

Wang M, Li D, Casas JR, Ruiz-Hidalgo J. Adaptive Fusion of LiDAR Features for 3D Object Detection in Autonomous Driving. Sensors. 2025; 25(13):3865. https://doi.org/10.3390/s25133865

Chicago/Turabian Style

Wang, Mingrui, Dongjie Li, Josep R. Casas, and Javier Ruiz-Hidalgo. 2025. "Adaptive Fusion of LiDAR Features for 3D Object Detection in Autonomous Driving" Sensors 25, no. 13: 3865. https://doi.org/10.3390/s25133865

APA Style

Wang, M., Li, D., Casas, J. R., & Ruiz-Hidalgo, J. (2025). Adaptive Fusion of LiDAR Features for 3D Object Detection in Autonomous Driving. Sensors, 25(13), 3865. https://doi.org/10.3390/s25133865

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop