Next Article in Journal
Dynamic Algorithm for Mining Relevant Association Rules via Meta-Patterns and Refinement-Based Measures
Previous Article in Journal
The Gallery of Memories (GA-ME): A Novel Virtual Navigation Tool for the Study of Spatial Memory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BiDFNet: A Bidirectional Feature Fusion Network for 3D Object Detection Based on Pseudo-LiDAR

School of Computer Science, University of South China, Hengyang 421001, China
*
Author to whom correspondence should be addressed.
Information 2025, 16(6), 437; https://doi.org/10.3390/info16060437
Submission received: 4 May 2025 / Revised: 22 May 2025 / Accepted: 23 May 2025 / Published: 26 May 2025

Abstract

:
This paper presents a bidirectional feature fusion network (BiDFNet) for 3D object detection, leveraging pseudo-point clouds to achieve bidirectional fusion of point cloud and image features. The proposed model addresses key challenges in multimodal 3D detection by introducing three novel components: (1) the SAF-Conv module, which extends the receptive field through improved submanifold sparse convolution, enhancing feature extraction from pseudo-point clouds while effectively reducing edge noise; (2) the bidirectional cross-modal attention feature interaction module (BiCSAFIM), which employs a multi-head cross-attention mechanism to enable global information interaction between point cloud and image features; and (3) the attention-based feature fusion module (ADFM), which adaptively fuses dual-stream features to improve robustness. Extensive experiments on the KITTI dataset demonstrate that BiDFNet achieves state-of-the-art performance, with a 3D AP (R40) of 88.79% on the validation set and 85.27% on the test set for the Car category, significantly outperforming existing methods. These results highlight the effectiveness of BiDFNet in complex scenarios, showcasing its potential for real-world applications such as autonomous driving.

1. Introduction

In the advancement of industrial domains such as autonomous driving, robotics, and intelligent traffic monitoring, 3D object detection technology plays a pivotal role. It directly impacts the environmental perception capabilities and decision-making accuracy of emerging transportation tools, including smart vehicles and drones. For instance, in autonomous driving, accurate 3D object detection is crucial for obstacle avoidance, path planning, and ensuring passenger safety. Despite significant progress in recent years, existing methods still face challenges in complex and dynamic environments, such as varying lighting conditions, occlusions, and sparse data from distant objects.
Although significant progress has been achieved in recent years with single-sensor-based 3D object detection, including both grid-based LiDAR methods like VoxelNet [1], SECOND [2], PointPillars [3], Voxel R-CNN [4], and VoxelNeXt [5], as well as point-based approaches such as PointRCNN [6], Point-GNN [7], and hybrid methods like PV-RCNN [8], these methods still encounter limitations in complex and dynamic traffic environments. For example, LiDAR provides precise three-dimensional spatial information but is susceptible to extreme weather conditions, such as rain or snow, which can scatter laser signals and reduce data quality. On the other hand, cameras capture rich color and texture details but struggle with accurate depth estimation, especially in low-light conditions or when objects are far away. These limitations highlight the need for multimodal approaches that can combine the strengths of different sensors to improve detection performance.
Consequently, multimodal 3D object detection which integrates data from diverse sensors such as LiDAR and cameras has emerged as an effective strategy with which to enhance detection performance. Early approaches like MV3D [9] explored fusing multiple views, while methods like Frustum PointNets [10] leveraged 2D detections to constrain the 3D search space. More recent works focus on fusing features at various stages, such as combining raw point and image features [11], enhancing point features with image semantics [12], fusing features at the BEV level [13], or employing deep fusion networks [14]. By combining the precise geometric information from LiDAR with the rich semantic details from cameras, multimodal methods can achieve more robust and accurate detection in complex environments. For example, in urban driving scenarios, multimodal approaches can better handle occlusions and varying lighting conditions, leading to improved safety and reliability. Recent studies have demonstrated that multimodal fusion can significantly enhance detection accuracy, particularly for distant or partially occluded objects.
However, existing multimodal methods still face several challenges. For instance, traditional feature fusion techniques fail to fully exploit the complementary information between modalities, resulting in suboptimal detection accuracy. These issues are particularly pronounced in complex scenarios where objects are partially occluded or located at a distance. Furthermore, methods relying on derived representations like pseudo-point clouds (discussed in Section 2) introduce specific challenges related to noise and feature quality. To address these challenges, this paper proposes an SAF-Conv module and BiCSAFIM model. The SAF-Conv module uses more precise 3D regions of point clouds for cropping and employs improved submanifold sparse convolution [15] to expand the receptive field for extracting richer information, effectively reducing edge noise. The BiCSAFIM module uses a multi-head cross-attention mechanism [16] to significantly enhance the information interaction between point cloud features and image features from a global perspective, effectively improving the model’s utilization of point cloud and image feature information.
To evaluate the performance of the proposed model, the widely recognized KITTI dataset, a benchmark for multimodal 3D object detection, was employed for network performance assessment. Comprehensive experimental results indicate that the model achieves notable performance improvements compared to traditional feature fusion networks. Furthermore, ablation experiments validate the significance and necessity of each module introduced in this study. The primary contributions of this paper are as follows:
  • We propose a bidirectional cross-sensor attention feature interaction module (BiCSAFIM), which significantly enhances the interaction between point cloud features and image features.
  • A feature extraction module, termed SAF-Conv, is introduced for pseudo-point clouds. This module enlarges the receptive field, thereby improving the extraction of 2D features and concurrently reducing noise to a certain extent.
  • Our method surpasses previous approaches, achieving state-of-the-art performance on the KITTI 3D object detection benchmark.

2. Related Work

2.1. LiDAR-Based 3D Object Detection

LiDAR provides precise geometric information and is a focal point in single-modal 3D detection. Early methods project point clouds into 2D bird’s-eye view (BEV) or range images for 3D detection [3,17,18]. Recently, mainstream LiDAR-based 3D object detection approaches can be divided into two major categories: point-based methods and grid-based methods. Point-based methods, such as PointNet [19] and PointNet++ [20], directly process raw point clouds using deep learning frameworks. These methods extract features through point-based backbone networks and predict 3D bounding boxes based on downsampled points and features. However, the process of sampling and aggregating features from irregular point clouds is computationally expensive.
Grid-based methods, meanwhile, transform point clouds into structured grid formats, such as voxels, pillars, or bird’s-eye view (BEV) feature maps. For example, VoxelNet [1] utilizes sparse voxel grids to extract features from points within voxel cells. SECOND [2] introduced 3D sparse convolution for voxels, significantly accelerating the convolution process. PointPillars [3] further simplifies voxels into pillars to enhance processing speed. Despite their efficiency, grid-based methods often struggle with distant or sparse objects due to the limited resolution of LiDAR. Addressing computational efficiency, Ada3D [21] focuses on optimizing inference speed by exploiting spatial redundancy. It employs an adaptive mechanism to dynamically allocate computation, reducing processing in empty or simple regions of the 3D space, thereby improving overall efficiency.

2.2. Fusion-Based 3D Object Detection

To address the limitations of single-modal approaches, multimodal 3D object detection methods have been developed to leverage both LiDAR and camera data. These methods can be broadly categorized into frustum-based and multi-view-based approaches. Frustum-based methods represent a sequential, result-level fusion approach. Works like FrustumPointNet [10], FrustumConvNet [22], and CenterFusion [23] use the output of image detectors to lift 2D detections into 3D frustums with explicit depth estimation. While these methods optimize the search space for 3D detection and reduce computational costs, their performance is often limited by inaccurate depth estimation. For example, Frustum-PointNet generates 3D proposals by extruding 2D bounding boxes into 3D space, but the accuracy of these proposals heavily depends on the quality of depth estimation, which can be unreliable in complex scenes.
Multi-view-based approaches utilize feature-level fusion to create 3D region proposals from bird’s-eye views (BEV) and conduct regression for 3D bounding boxes. A notable early contribution in this field is MV3D [9], developed by Chen. MV3D converts point clouds into BEV representations, extracts relevant features, and generates 3D proposals. These proposals are subsequently projected into range views, where they are combined with high-dimensional image features during the ROI pooling phase, followed by 3D bounding box regression. Building on MV3D, Ku introduced AVOD [24], which integrates image feature maps from VGG-16 with point cloud data during the 3D proposal generation stage. AVOD also incorporates an autoencoder (AE) structure to reduce feature space dimensionality, thereby lowering computational demands. Despite these advancements, such fusion techniques often project 3D point clouds onto 2D planes, leading to substantial information loss. Additionally, the limited alignment between point cloud features and image features hinders the full utilization of semantic information.

2.3. Pseudo-Point Cloud Methods

A significant direction within multimodal fusion leverages pseudo-point clouds, which are 3D representations generated primarily from camera data using depth estimation, rather than direct LiDAR sensing. This approach has gained traction due to several potential advantages, detailed as follows: (1) cost-effectiveness, which enables capable 3D perception using lower-cost camera-centric sensor suites; (2) leveraging 2D representations, which allows the application of powerful, pre-trained image networks to extract rich semantic features for 3D tasks; and (3) unified geometric representation—by converting image-derived depth information into a 3D point cloud format, pseudo-LiDAR brings both camera and (potentially) sparse real LiDAR data into a common geometric domain (point clouds). This unification can simplify the subsequent feature interaction process compared to directly fusing disparate representations like dense 2D feature maps and sparse 3D points, potentially reducing the complexity of explicit cross-dimensional feature alignment mechanisms. Finally, (4) scene densification potentially fills gaps in sparse LiDAR scans when fused, offering a denser scene representation.
Building upon these advantages, recent research has explored various techniques. For instance, MVP [25] uses depth estimation to generate virtual points to enhance the point cloud. PointAugmenting [26] introduces cross-modal data augmentation techniques to generate diverse pseudo-point clouds, enhancing model generalization. SFD [27] employs semantic feature distillation to improve the expression of geometric details and semantic information in pseudo-point clouds. patch refinement proposes a local patch-based optimization method for pseudo-point clouds, reducing the impact of depth estimation errors. DepthGAN [28] leverages generative adversarial networks to produce high-quality pseudo-point clouds, improving the accuracy and robustness of depth estimation. However, existing pseudo-point cloud methods face two major challenges: (1) inaccurate depth estimation often introduces incorrectly placed virtual points around objects, misleading object predictions, and (2) the large number of redundant points generated in pseudo-point clouds do not fully align with sparse point cloud data, leading to incomplete utilization of information during fusion. Recent work has also explored how to better utilize pseudo-points within detector architectures. For instance, SQD [29] proposes enhancing features around sparse object queries using pseudo-point information, aiming to improve the interaction between sparse queries and the densified feature maps for more accurate detection. These limitations motivate the need for improved feature extraction and fusion techniques, as proposed in this work.
In essence, challenges remain in effectively handling sparse LiDAR data, mitigating noise in pseudo-LiDAR representations, and achieving truly synergistic fusion between modalities. This work introduces BiDFNet to address these issues, proposing dedicated modules for robust pseudo-LiDAR feature extraction (SAF-Conv), deep bidirectional cross-modal interaction (BiCSAFIM), and adaptive fusion (ADFM), as detailed next.

3. BiDFNet

This section offers an in-depth exploration of the architectural framework of BiDFNet, a 3D object detection approach leveraging pseudo-point clouds, which combines submanifold sparse convolution with multi-head attention mechanisms. Section 3.1 outlines the overarching design principles of the architecture. Following this, Section 3.2 delves into the design and implementation of the sparse convolutional submanifold network (SAF-Conv) [15]. Section 3.3 details the structure of the bidirectional feature interaction module (BiCSAFIM). Finally, in Section 3.4, the grid feature fusion (ADFM) module is discussed.

3.1. Overall Architecture

In the BiDFNet model, pseudo-point clouds are first generated through depth completion. The initial point cloud is used to predict 3D ROIs (Regions of Interest). These ROIs are then employed to crop both the original point cloud and the pseudo-point cloud. For the pseudo-point cloud, point cloud information is extracted and voxelized. The voxelized features are fed into SAF-Conv for further feature extraction and a certain degree of denoising. Finally, the features from the initial point cloud and the pseudo-point cloud are input into the dual-stream feature interaction module. The interaction-enhanced features are fused by the ADFM (attention-based feature fusion module) and then passed to the detection head (detect head) to produce the output. The overall architecture of the BiDFNet model is illustrated in Figure 1.

3.2. Pseudo-Point Cloud Feature Extraction

For the original point cloud R, we employ the benchmark Voxel-RCNN [4] method to generate 3D detection regions. Similarly, the region proposals are further divided into G×G×G regular sub-voxels, with the center point serving as the grid point of the corresponding sub-voxel. We observe that a significant amount of non-Gaussian noise is introduced during the generation of pseudo-point clouds due to depth estimation errors. Notably, this noise tends to occur prominently at object edges. While edge detection can roughly remove such noise, it also results in the loss of a substantial number of useful boundary points. To address this, we crop the pseudo-point cloud using 3D proposal regions generated from point clouds rather than pseudo-point clouds. To extract 2D semantic information from the pseudo-point cloud, we utilize CPConvs [27]. Following this, the extracted features are voxelized and then processed for feature extraction. When extracting features from the pseudo-point cloud, we develop an improved submanifold sparse convolution (SAF-Conv), inspired by [30]. This approach expands the receptive field in the pseudo-point cloud flow to the 2D image space, enabling the simultaneous extraction of both 3D and 2D features from the pseudo-point cloud.The structure of this method is illustrated in Figure 2. It is designed specifically to handle the characteristics of pseudo-LiDAR. It expands the receptive field within the sparse convolution framework, allowing the model to incorporate broader spatial context, which is crucial for interpreting potentially noisy points, especially near object boundaries. Furthermore, SAF-Conv enables the simultaneous extraction and adaptive fusion of both 3D geometric features and projected 2D semantic features derived from the pseudo-point cloud, aiming for a more robust representation.
Specifically, given N input voxels represented by 3D index vectors R N × 3 and feature vectors Z R N × C in geometric features are first encoded separately in both 3D and 2D image spaces. These two geometric features are then combined to yield Y R N × C out where the number of input and output feature channels are denoted as C in and C out .
As depicted in the workflow of Figure 2, SAF-Conv processes features through parallel 2D and 3D spatial encoding paths before fusion. For three-dimensional space encoding, the geometric features are directly encoded using the 3D submanifold convolution kernel K3D(·). Z ^ i are calculated from the non-empty voxels within 3 × 3 × 3 neighborhood based on the corresponding 3D indices, as is expressed as Equation (1) [30].
Z ^ i = R K 3 D Z i , Z i f 1 , , Z i f j
For two-dimensional space encoding, the points are first projected onto the 2D image plane, after which geometric features are encoded using the 2D submanifold convolution kernel K2D(·). The 3D indices are converted into a set of grid points (denoted as G(·)) based on voxelization parameters. Due to the inherent differences between point cloud and image data, data augmentation techniques such as rotation and scaling are commonly applied, as denoted T(·) [27,30]. The grid points are then mapped back to the original coordinate system using the data augmentation parameters. Finally, these points are projected onto the 2D image plane, denoted P(·), using LiDAR-Camera calibration parameters. This process is mathematically expressed as Equation (2) [30].
H ^ = P T 1 ( G ( H ) )
Here, H ^ represents the 2D index vector. For each voxel feature Z i , we then calculate the noise-aware features from the non-empty voxels within a 3 × 3 neighborhood based on the corresponding 2D indices, as is expressed in Equation (3) [30].
Z ˜ i = R K 2 D Z i , Z ˜ i f 1 , , Z ˜ i f k
Z ˜ i represents the adjacent voxel features generated, and K2D(·) denotes the 2D submanifold convolution kernel. Given that multiple features may exist within a single 2D adjacent voxel, we employ max pooling to retain the most prominent feature for performing 2D convolution, ensuring the emphasis on significant characteristics. After encoding both 3D and 2D features, an attention mechanism is applied to combine these two features, ultimately yielding the ROI (Region of Interest) features of the pseudo-point cloud, as expressed in Equations (4)–(6).
α = σ ( MLP ( [ Z ˜ i ; Z ^ i ] ) )
Z i ¯ = α · Z ˜ i + ( 1 α ) · Z ^ i
Y = Z 1 ¯ T , . . . , Z N ¯ T T
Multiple SAF-Conv modules are linked in sequence with standard SparseConv3D layers to broaden the 2D receptive field during sparse convolution, resulting in the final output feature vector.

3.3. Interaction of Bidirectional Features

Previous methods relied primarily on ROI feature strategies to utilize multimodal data or simply employed channel attention modules to fuse two ROIs. However, these approaches did not fully exploit the complementary information between modalities, resulting in insufficiently effective feature fusion.
To address these limitations, we propose a BiCSAFIM (bidirectional cross-modal selective attention feature interaction module), as illustrated in Figure 3. This module consists of dual streams, each taking LiDAR features and pseudo-point cloud features generated from RGB images as input, and performs feature interaction and enhancement separately.
Simple fusion methods like concatenation or element-wise operations, or even basic channel attention, often fail to capture the complex, non-linear correlations and spatial misalignments between sparse LiDAR features and denser pseudo-LiDAR/image features. We employ a bidirectional cross-attention mechanism (BiCSAFIM) based on the Transformer architecture [16] because cross-attention explicitly allows features from one modality (acting as queries) to attend to and selectively aggregate information from another modality (acting as keys and values) based on learned relevance. The bidirectional structure, with parallel streams where each modality can query the other, ensures comprehensive information exchange (e.g., pseudo-LiDAR providing denser context where LiDAR is sparse, and real LiDAR providing accurate geometric cues to refine pseudo-LiDAR features). Multi-head attention further enhances this capability by allowing the model to jointly attend to information from different representation subspaces in different positions.
Given the need for cross-modal interaction and the requirement for features from one modality (e.g., pseudo-point ROI features) to perceive the overall features of another modality (e.g., point ROI features) while selectively aggregating and enhancing complementary information, we incorporate both multi-head cross-attention and multi-head self-attention mechanisms [16]. This enables the dual streams to perceive intermodal feature relationships while emphasizing their own feature information.
More specifically, initially, for LiDAR ROI features, F l features serve as the query, while pseudo-LiDAR ROI features F p act as the key and value for intermodal fusion, resulting in feature F1.
Attention ( F l , F p ) = Softmax F l F p T d k F p
We begin by applying the multi-head cross-attention mechanism, where independent learnable linear transformations are applied to the query FL, key FP, and value FP. The i-th head is denoted in Equation (8), where dimensions are represented as follows.
Q i = F l W i Q , K i = F p W i K , V i = F p W i V
We then employ m heads to execute multi-head cross-attention, and the result is summed with FL, as shown in Equations (9) and (10).
head i = Attention ( Q i , K i , V i )
F 1 = Concat ( head 1 , head 2 , . . . , head h ) + F l
where F 1 is the sum of the output from the multi-head attention module and the LiDAR ROI features F L . F 1 is used as queries, keys, and values. Subsequently, similar to the multi-head cross-attention mechanism, we utilize m heads to perform multi-head self-attention, aiming to emphasize important information within a single modality, thereby obtaining the intra-modal fusion feature F 2 , as shown in Equations (11)–(14).
Attention ( F 1 , F 1 ) = Softmax F 1 F 1 T d k F 1
Q i = F 1 W i Q , K i = F 1 W i K , V i = F 1 W i V
head i = Attention ( Q i , K i , V i )
F 2 = MultiHead ( F 1 ) = Concat ( head 1 , head 2 , . . . , head h )
After applying the aforementioned attention mechanisms, we obtain the fused feature F 2 from the point cloud stream branch, which achieves the fusion of pseudo-point cloud features and the enhancement of its own features. The other branch adopts a similar method. After selective fusion of pseudo-point cloud features with point cloud features, it enhances itself to obtain F 4 .

3.4. Attention-Driven ROI Fusion

After processing the features from the original LiDAR stream (yielding F 2 ) and the pseudo-LiDAR stream (yielding F 4 ) through the BiCSAFIM, a final fusion step is needed before prediction. Since the reliability and informativeness of each feature stream ( F 2 vs. F 4 ) might vary spatially across the ROI grid (e.g., pseudo-LiDAR might be more reliable in dense areas far from the ego vehicle, while real LiDAR is better close-up or for fine geometric details), a simple averaging or concatenation might be suboptimal.
To address this, we employ the attention-driven fusion module (ADFM), inspired by attention mechanisms [16] and selective fusion approaches [31,32]. Since the features F 2 and F 4 represent the same ROI and share a consistent grid structure, ADFM can operate effectively at the grid level. As illustrated in Figure 4, for each corresponding pair of grid features from F 2 and F 4 , ADFM dynamically computes spatial attention weights. This allows the network to learn which feature stream to emphasize more at each specific location within the ROI. By adaptively weighting the grid features based on their learned reliability or importance before combining them, ADFM produces a more robust final fused representation for the detection head.
Specifically, let b represent a single 3D ROI. In the previous section, we obtained the original flow ROI features F 2 and the pseudo flow ROI features F 4 . Here, n denotes the total number of grids in the 3D ROI, and C represents the number of feature channels per grid. Given a pair of ROI grid features ( F 2 , F 4 ), they are concatenated. Subsequently, two parallel 3 × 3 convolutional layers are applied to generate higher-level features. These features are then passed through a sigmoid function to produce a pair of weights ( w 1 , w 2 ), where both w 1 and w 2 are scalars. Finally, the fused grid feature F is obtained by weighting ( F 2 , F 4 ) with ( w 1 , w 2 ). The implementation of (F) is as follows:
F G = Concat F 2 , F 4
w 1 = sigmoid Conv F 2 , w 2 = sigmoid Conv F 4
F = CONCAT F 2 × w 1 , F 4 × w 2
By fusing the ROI features of the original flow and the pseudo-point cloud flow using the aforementioned method, the final fused ROI features are then utilized to predict class confidence scores and bounding boxes.

4. Experiments

In this section, we present a comprehensive evaluation of the BiDFNet model on the KITTI dataset [33], a widely recognized benchmark for 3D object detection. The evaluation is structured as follows. Section 4.1 provides an overview of the KITTI dataset and its evaluation metrics. Section 4.2 details the experimental setup, including the network architecture, loss function, data augmentation techniques, and training procedures. Section 4.3 discusses the design and implementation of the proposed model. Section 4.4 presents ablation studies to analyze the contributions of individual modules. Finally, Section 4.5 summarizes the experimental results and highlights the key findings.

4.1. KITTI Dataset

The KITTI dataset [33] is utilized for our experiments. It provides real-world driving data including LiDAR point clouds and camera images, along with annotations for 3D object detection. We conduct our experiments on this dataset as it remains a widely adopted and challenging benchmark for 3D object detection in autonomous driving research. Its real-world driving scenarios and standardized evaluation protocol facilitate fair comparison with numerous state-of-the-art methods [4,12,27,29,30,34,35,36,37,38,39]. While larger datasets like nuScenes or the Waymo Open Dataset offer greater diversity in environments and sensor configurations, KITTI is still highly relevant for evaluating core aspects of multimodal fusion and performance on varying object distances and occlusion levels, which are central to our proposed method. For evaluation, we follow the standard practice of splitting the official training set of 7481 samples into a training set (3712 samples) and a validation set (3769 samples). Performance is reported using the official Average Precision (AP) metric with IoU thresholds of 0.7 for the ‘Car’ category, calculated across Easy, Moderate, and Hard difficulty levels. These levels categorize objects based on their size (height in pixels), occlusion severity, and truncation degree, as specified in the KITTI evaluation protocol.

4.2. Experimental Details

The development environment used for the experiments includes (1) an Ubuntu 18.04 operating system; (2) GPU: RTX 4090; (3) an open-source framework: PyTorch 1.6.9. The BiDFNet model restricts the point cloud range to [0, 70.4] m on the X-axis, [−40, 40] m on the Y-axis, and [−3, 1] m on the Z-axis. The input voxel size is set to (0.05 m, 0.05 m, 0.1 m), with an IoU threshold of 0.7 and a recall rate of 40. The code for this work is implemented based on an open-source 3D object detection framework, Voxel-RCNN.

4.3. Loss and Data Augmentation

In the BiDFNet model, we use the same training loss as the benchmark method [4]. Local and global data enhancement methods such as ground-truth sampling, rotation, translation, and flipping are followed in the model [27,30].

Training and Inference Details

The BiDFNet model undergoes training from the ground up on the KITTI dataset, utilizing an adaptive optimization algorithm across 80 epochs. Training is performed on a single NVIDIA 4090 GPU, with learning rates dynamically adjusted to 0.01, 0.01, and 0.005 in various phases of the training process. To ensure an equitable comparison with existing models, all experimental settings—excluding the learning rate—are aligned with the baseline framework outlined in Voxel-RCNN [4].

4.4. Main Results

We evaluate BiDFNet on the widely-used KITTI dataset [33] following the standard evaluation protocol, reporting 3D Average Precision (AP) with an IoU threshold of 0.7 for the ‘Car’ category across Easy, Moderate, and Hard difficulty levels, using both R40 and R11 recall sampling points. For comparison against existing methods in Table 1 and Table 2, results for competing approaches are primarily sourced from their original publications or the official KITTI leaderboard, unless otherwise noted, ensuring a fair comparison under identical evaluation settings.
We present the following results on the KITTI validation set. From the experimental results of car detection accuracy comparison in Table 1, it can be observed that the BiDFNet model, by integrating the SAF-Conv, BiCSAFIM, and ADFM modules, significantly enhances its feature representation capability in complex environments. Specifically, compared to the baseline model Voxel-RCNN, BiDFNet shows improvements across Easy, Moderate, and Hard detection levels; the 3D AP (40) increased by 1.09%, 3.53%, and 4.15%, respectively, while the 3D AP (11) increased by 0.30%, 2.29%, and 6.77%, respectively. Furthermore, compared to some mainstream 3D vehicle detection algorithms, the proposed method achieves the best results across all three detection difficulty levels.
The results on the KITTI test set demonstrate that our BiDFNet model achieved significant improvements in 3D AP (R40) across the Easy, Moderate, and Hard vehicle categories, with increases of 0.89%, 3.65%, and 3.66%, respectively, compared to the baseline model Voxel RCNN [4]. Moreover, BiDFNet has consistently outperformed numerous state-of-the-art methods in recent years. For a comprehensive comparison, the detailed results are summarized in Table 2. While improvements on the ‘Easy’ subset might appear incremental due to near-saturation by the top methods, the significant gains observed for the ‘Moderate’ and particularly the ‘Hard’ difficulty levels are noteworthy. These Hard cases often involve distant, small, or heavily occluded objects, where robust feature extraction and effective fusion are most critical. The consistent improvements in these challenging scenarios strongly suggest the effectiveness of our proposed modules; SAF-Conv likely contributes by enhancing pseudo-point cloud features and mitigating noise, BiCSAFIM improves cross-modal feature interaction, and ADFM enables more robust final fusion. This ability to better handle difficult cases highlights the practical value of BiDFNet for real-world autonomous driving applications where robustness is paramount.
To validate the generalization capability of our approach, we extended BiDFNet to simultaneously detect pedestrians and cyclists using a unified model. As shown in Table 3, our method demonstrates consistent improvements across all object categories compared to the multi-class Voxel-RCNN baseline. These results confirm that BiDFNet’s bidirectional fusion mechanism effectively preserves its advantages when adapted to multi-class detection scenarios, achieving significant performance gains while maintaining architectural simplicity.

4.5. Ablation Study

To assess the efficacy of the proposed approach, a comprehensive set of experiments were carried out. As depicted in Table 4, the first one serves as a tailored baseline model derived from the Voxel R-CNN framework, solely utilizing the original point cloud data as input. Conversely, the others integrate pseudo-point clouds to enable multimodal enhancement, facilitating a balanced comparison with the single-modal approach of the first one. These experiments aimed to evaluate the synergistic effects and overall performance of the SAF-Conv, BiSCAFIM, and ADFM modules.
The integration of SAF-Conv and BiCSAFIM into the baseline model led to a notable enhancement in detection accuracy, as evidenced by the results in Table 4. Specifically, SAF-Conv contributed to improvements in 3D AP (R40) by 0.40%, 1.24%, and 1.12% for the Easy, Moderate, and Hard difficulty levels, respectively. Similarly, BiCSAFIM achieved gains of 0.44%, 1.32%, and 1.16% across the same categories. When both modules were combined, the accuracy saw further enhancement, with the Hard class accuracy reaching 86.10%. The addition of the ADFM fusion attention module further elevated the Hard class detection accuracy to 86.41%, representing a 4.15% overall improvement compared to the baseline. These results underscore the significant contributions of each module to the model’s performance.
Finally, we analyze the computational overhead introduced by our modules using the inference times reported in Table 4 (measured in ms on an RTX 4090). The Voxel-RCNN baseline runs in 40 ms. Integrating the pseudo-LiDAR stream and our proposed components increases the runtime incrementally; SAF-Conv adds approximately 8 ms, BiCSAFIM adds about 10 ms, and ADFM adds 4 ms. Our full BiDFNet model results in a total inference time of 84 ms. While this is roughly 2.1 times slower than the baseline (corresponding to 11.9 FPS), this increased computational cost directly correlates with significant performance gains, particularly the substantial 4.15% AP (R40) improvement in the difficult ‘Hard’ category. This demonstrates a clear trade-off, where the enhanced robustness and accuracy in challenging scenarios justify the additional computation, though further optimization could be explored for applications demanding higher frame rates.
The experimental results reveal that the SAF-Conv module significantly improves the feature representation of pseudo-point clouds, whereas the BiCSAFIM and ADFM modules enhance the integration and effective use of data across modalities. These advancements lead to a marked increase in detection accuracy, especially in challenging scenarios, where substantial improvements are evident for the Hard category.

5. Conclusions

This paper introduces BiDFNet, a novel 3D object detection framework integrating pseudo-point clouds and bidirectional feature fusion between point cloud and image data streams. Our approach, leveraging the proposed SAF-Conv, BiCSAFIM, and ADFM modules, ensures comprehensive information exchange across modalities while enhancing feature robustness, particularly for pseudo-LiDAR data. Experimental results on the KITTI dataset demonstrate that BiDFNet achieves state-of-the-art performance in the highly competitive Car category, significantly surpassing existing methods, especially in challenging ‘Hard’ scenarios. Furthermore, we validated the generalizability of our architecture by extending it to multi-class detection, showing consistent performance improvements for ‘Pedestrian’ and ‘Cyclist’ categories over the baseline (Table 3).
We acknowledge several limitations in the current study which outline important directions for future work. Firstly, while we demonstrated multi-class capability, the primary focus remained on the ‘Car’ category, and further dedicated evaluation and optimization are needed for pedestrian and cyclist detection to fully assess performance across all crucial road users. Secondly, and critically for autonomous driving applications, this work lacks a dedicated robustness analysis under adverse weather conditions (e.g., rain, snow) or scenarios simulating sensor degradation, which is essential for evaluating real-world reliability. Thirdly, while inference times were reported (Section 4.4), a comprehensive computational complexity analysis—including hardware-independent metrics like detailed efficiency comparisons—was not performed, limiting a full assessment of the model’s computational demands. Future research will prioritize addressing these limitations. Specifically, we aim to conduct rigorous robustness testing, perform detailed computational benchmarking, and extend optimization efforts to pedestrian and cyclist detection. Despite these acknowledged limitations, this study offers an efficient and innovative solution built upon attention-driven bidirectional fusion, establishing the effectiveness of the proposed components and providing a strong foundation for future advancements in robust and accurate multimodal 3D object detection.

Author Contributions

Conceptualization, Q.Z. and Y.W.; methodology, Q.Z.; software, Q.Z.; validation, Q.Z. and Y.W.; formal analysis, Q.Z.; investigation, Q.Z.; resources, Y.W.; data curation, Q.Z.; writing—original draft preparation, Q.Z.; writing—review and editing, Q.Z. and Y.W.; visualization, Q.Z.; supervision, Y.W.; project administration, Y.W.; funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Project 2024JJ7428 of the Hunan Provincial Natural Science Foundation of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in the KITTI dataset at [http://www.cvlibs.net/datasets/kitti/, accessed on 1 May 2025].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  2. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  3. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection From Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  4. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel r-cnn: Towards high performance voxel-based 3D object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 1201–1209. [Google Scholar]
  5. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. Voxelnext: Fully sparse voxelnet for 3D object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 21674–21683. [Google Scholar]
  6. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  7. Shi, W.; Rajkumar, R. Point-gnn: Graph neural network for 3D object detection in a point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1711–1719. [Google Scholar]
  8. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10529–10538. [Google Scholar]
  9. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  10. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum pointnets for 3D object detection from rgb-d data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 918–927. [Google Scholar]
  11. Xu, D.; Anguelov, D.; Jain, A. Pointfusion: Deep sensor fusion for 3D bounding box estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 244–253. [Google Scholar]
  12. Huang, T.; Liu, Z.; Chen, X.; Bai, X. Epnet: Enhancing point features with image semantics for 3D object detection. In Proceedings, Part XV 16, Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 35–52. [Google Scholar]
  13. Liang, T.; Xie, H.; Yu, K.; Xia, Z.; Lin, Z.; Wang, Y.; Tang, T.; Wang, B.; Tang, Z. Bevfusion: A simple and robust lidar-camera fusion framework. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 10421–10434. [Google Scholar]
  14. Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Lu, Y.; Zhou, D.; Le, Q.V.; et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17182–17191. [Google Scholar]
  15. Graham, B.; Van der Maaten, L. Submanifold sparse convolutional networks. arXiv 2017, arXiv:1706.01307. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–7 December 2017; Volume 30.
  17. Beltrán, J.; Guindel, C.; Moreno, F.M.; Cruzado, D.; Garcia, F.; De La Escalera, A. Birdnet: A 3D object detection framework from lidar information. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 3517–3523. [Google Scholar]
  18. Huang, J.; Huang, G.; Zhu, Z.; Ye, Y.; Du, D. Bevdet: High-performance multi-camera 3D object detection in bird-eye-view. arXiv 2021, arXiv:2112.11790. [Google Scholar]
  19. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  20. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–7 December 2017; Volume 30. [Google Scholar]
  21. Zhao, T.; Ning, X.; Hong, K.; Qiu, Z.; Lu, P.; Zhao, Y.; Zhang, L.; Zhou, L.; Dai, G.; Yang, H.; et al. Ada3D: Exploiting the spatial redundancy with adaptive inference for efficient 3D object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 17–24 June 2023; pp. 17728–17738. [Google Scholar]
  22. Wang, Z.; Jia, K. Frustum convnet: Sliding frustums to aggregate local point-wise features for amodal 3D object detection. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 1742–1749. [Google Scholar]
  23. Nabati, R.; Qi, H. Centerfusion: Center-based radar and camera fusion for 3D object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 1–6 January 2021; pp. 1527–1536. [Google Scholar]
  24. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  25. Yin, T.; Zhou, X.; Krähenbühl, P. Multimodal virtual point 3D detection. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–14 December 2021; Volume 34, pp. 16494–16507. [Google Scholar]
  26. Wang, C.; Ma, C.; Zhu, M.; Yang, X. Pointaugmenting: Cross-modal augmentation for 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11794–11803. [Google Scholar]
  27. Wu, X.; Peng, L.; Yang, H.; Xie, L.; Huang, C.; Deng, C.; Liu, H.; Cai, D. Sparse fuse dense: Towards high quality 3D detection with depth completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5418–5427. [Google Scholar]
  28. Li, Y.; Wang, Y.; Lu, Z.; Xiao, J. DepthGAN: GAN-based depth generation of indoor scenes from semantic layouts. arXiv 2022, arXiv:2203.11453. [Google Scholar]
  29. Mo, Y.; Wu, Y.; Zhao, J.; Hou, Z.; Huang, W.; Hu, Y.; Wang, J.; Yan, J. Sparse Query Dense: Enhancing 3D Object Detection with Pseudo points. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, VIC, Australia, 28 October–November 2024; pp. 409–418. [Google Scholar]
  30. Wu, H.; Wen, C.; Shi, S.; Li, X.; Wang, C. Virtual sparse convolution for multimodal 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 21653–21662. [Google Scholar]
  31. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  32. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  33. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  34. Mao, J.; Niu, M.; Bai, H.; Liang, X.; Xu, H.; Xu, C. Pyramid r-cnn: Towards better performance and adaptability for 3D object detection. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 2723–2732. [Google Scholar]
  35. Zhao, X.; Su, L.; Zhang, X.; Yang, D.; Sun, M.; Wang, S.; Zhai, P.; Zhang, L. D-conformer: Deformable sparse transformer augmented convolution for voxel-based 3D object detection. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  36. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-voxel feature set abstraction with local vector representation for 3D object detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  37. Yoo, J.H.; Kim, Y.; Kim, J.; Choi, J.W. 3D-cvf: Generating joint camera and lidar features using cross-view spatial feature fusion for 3D object detection. In Proceedings, Part XXVII 16, Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 720–736. [Google Scholar]
  38. Wu, H.; Deng, J.; Wen, C.; Li, X.; Wang, C.; Li, J. CasA: A cascade attention network for 3-D object detection from LiDAR point clouds. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5704511. [Google Scholar] [CrossRef]
  39. Yang, H.; Liu, Z.; Wu, X.; Wang, W.; Qian, W.; He, X.; Cai, D. Graph r-cnn: Towards accurate 3D object detection with semantic-decorated local graph. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 662–679. [Google Scholar]
Figure 1. 1. The initial point cloud is used to predict 3D ROIs (Regions of Interest). These ROIs are then employed to crop both the original point cloud and the pseudo-point cloud. For the pseudo-point cloud, point cloud information is extracted and voxelized. 2. The voxelized features are fed into SAF-Conv for further feature extraction and a certain degree of denoising. 3. the features from the initial point cloud and the pseudo-point cloud are input into the dual-stream feature interaction module. 4. The interaction-enhanced features are fused by the ADFM (attention-based feature fusion module) and then passed to the detection head (detect head) to produce the output.
Figure 1. 1. The initial point cloud is used to predict 3D ROIs (Regions of Interest). These ROIs are then employed to crop both the original point cloud and the pseudo-point cloud. For the pseudo-point cloud, point cloud information is extracted and voxelized. 2. The voxelized features are fed into SAF-Conv for further feature extraction and a certain degree of denoising. 3. the features from the initial point cloud and the pseudo-point cloud are input into the dual-stream feature interaction module. 4. The interaction-enhanced features are fused by the ADFM (attention-based feature fusion module) and then passed to the detection head (detect head) to produce the output.
Information 16 00437 g001
Figure 2. The structure of SAF-Conv. Input voxel features are processed through two parallel streams: (Top) Three-dimensional submanifold sparse convolution captures the local geometric structure. (Bottom) Features are projected onto the 2D image plane, processed by 2D submanifold sparse convolution to capture semantic context, and then fused back with the 3D features using an attention-based mechanism. This allows adaptive integration of 3D geometry and 2D semantics from pseudo-LiDAR data.
Figure 2. The structure of SAF-Conv. Input voxel features are processed through two parallel streams: (Top) Three-dimensional submanifold sparse convolution captures the local geometric structure. (Bottom) Features are projected onto the 2D image plane, processed by 2D submanifold sparse convolution to capture semantic context, and then fused back with the 3D features using an attention-based mechanism. This allows adaptive integration of 3D geometry and 2D semantics from pseudo-LiDAR data.
Information 16 00437 g002
Figure 3. The structure of BiCSAFIM, structured with two branches, both of which take point cloud features and pseudo-point cloud features as inputs. After interaction through multi-head cross-attention, the features are further enhanced via multi-head self-attention.
Figure 3. The structure of BiCSAFIM, structured with two branches, both of which take point cloud features and pseudo-point cloud features as inputs. After interaction through multi-head cross-attention, the features are further enhanced via multi-head self-attention.
Information 16 00437 g003
Figure 4. The structure of ADFM. Parallel branches generate attention weights ( w 1 , w 2 ) for input features F 2 (LiDAR) and F 4 (pseudo-LiDAR) using Conv + Signem. Weighted features are concatenated (C) for adaptive fusion before detection.
Figure 4. The structure of ADFM. Parallel branches generate attention weights ( w 1 , w 2 ) for input features F 2 (LiDAR) and F 4 (pseudo-LiDAR) using Conv + Signem. Weighted features are concatenated (C) for adaptive fusion before detection.
Information 16 00437 g004
Table 1. Comparison of experimental results on the KITTI validation set.
Table 1. Comparison of experimental results on the KITTI validation set.
MethodReferenceCar AP (R40)Car AP (R11)
EasyMod.HardEasyMod.Hard
Voxel-RCNN [4]AAAI 202192.3885.2682.2689.4184.5278.93
Pyramid-PV [34]ICCV 202192.2685.2382.9489.3784.3878.84
D-Conformer-TSD [35]ICASSP 2023---89.6985.3679.53
3D-CVF [37]ECCV 202089.6779.8878.47---
EPNet [12]ECCV 202092.2882.5980.14---
SFD [27]CVPR 202293.2187.8584.5688.9686.3284.79
VirConv-L [30]CVPR 202393.3688.7185.83---
Ours-93.4788.7986.4189.7186.8185.70
Table 2. Comparison of experimental results on the KITTI test set.
Table 2. Comparison of experimental results on the KITTI test set.
MethodReferenceCar AP (R40)
EasyMod.Hard
Voxel-RCNN [4]AAAI 202190.9081.6277.06
Pyramid-PV [34]ICCV 202188.3982.0877.49
CasA [38]TGRS 202291.5883.0680.08
D-Conformer-TSD [35]ICASSP 202389.1382.1879.75
PV-RCNN++ [36]IJCV 202390.5381.6077.07
3D-CVF [37]ECCV 202089.2080.8573.11
Graph-Vol [39]ECCV 202291.8983.2777.78
SFD [27]CVPR 202291.7384.7677.92
VirConv-L [30]CVPR 202391.4185.0580.56
Ada3D [21]ICCV202387.4679.4175.63
SQD [29]ACM MM 202491.5881.8279.07
Ours-91.7985.2780.72
Table 3. Three-dimensional detection results (3D AP (R40)) of multi-class detection (KITTI validation set).
Table 3. Three-dimensional detection results (3D AP (R40)) of multi-class detection (KITTI validation set).
ClassMethod3D AP (R40)
EasyMod.Hard
PedestrianBaseline70.5562.9257.35
PedestrianOurs73.4766.8261.01
CyclistBaseline89.8671.1366.67
CyclistOurs90.1373.8769.15
Table 4. Performance impact of various components on 3D detection of cars using the KITTI validation set, as measured by Average Precision at 40 recall positions.
Table 4. Performance impact of various components on 3D detection of cars using the KITTI validation set, as measured by Average Precision at 40 recall positions.
PseudoSAF-ConvBiCSAFIMADFMCar AP (R40)Times (ms)
EasyModHard
92.3885.2682.2640
92.4185.6383.9762
92.8186.8785.0970
92.8586.9585.1372
92.4985.9183.0464
93.2088.3486.1080
92.9787.6385.4176
93.0287.5485.5174
93.4788.7986.4184
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, Q.; Wan, Y. BiDFNet: A Bidirectional Feature Fusion Network for 3D Object Detection Based on Pseudo-LiDAR. Information 2025, 16, 437. https://doi.org/10.3390/info16060437

AMA Style

Zhu Q, Wan Y. BiDFNet: A Bidirectional Feature Fusion Network for 3D Object Detection Based on Pseudo-LiDAR. Information. 2025; 16(6):437. https://doi.org/10.3390/info16060437

Chicago/Turabian Style

Zhu, Qiang, and Yaping Wan. 2025. "BiDFNet: A Bidirectional Feature Fusion Network for 3D Object Detection Based on Pseudo-LiDAR" Information 16, no. 6: 437. https://doi.org/10.3390/info16060437

APA Style

Zhu, Q., & Wan, Y. (2025). BiDFNet: A Bidirectional Feature Fusion Network for 3D Object Detection Based on Pseudo-LiDAR. Information, 16(6), 437. https://doi.org/10.3390/info16060437

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop