You are currently viewing a new version of our website. To view the old version click .
Smart Cities
  • Article
  • Open Access

14 October 2025

Feature-Level Vehicle-Infrastructure Cooperative Perception with Adaptive Fusion for 3D Object Detection

,
,
,
and
School of Transportation, Southeast University, Nanjing 211119, China
*
Author to whom correspondence should be addressed.
Smart Cities2025, 8(5), 171;https://doi.org/10.3390/smartcities8050171 
(registering DOI)
This article belongs to the Topic 3D Computer Vision and Smart Building and City, 3rd Edition

Abstract

Highlights

What are the main findings?
  • The proposed feature-level VICP framework consistently outperforms state-of-the-art baselines on the DAIR-V2X-C dataset, achieving higher A P 3 D and A P B E V .
  • Experiments show that RFR delivers the largest gain, UWF improves robustness via adaptive uncertainty weighting, and CDCA enhances feature calibration.
What is the implication of the main finding?
  • Cooperative perception effectively overcomes occlusion and blind-spot limitations of vehicle-centric systems.
  • The proposed model provides a reference for scalable and generalizable deployment of cooperative perception within smart city infrastructure.

Abstract

As vehicle-centric perception struggles with occlusion and dense traffic, vehicle-infrastructure cooperative perception (VICP) offers a viable route to extend sensing coverage and robustness. This study proposes a feature-level VICP framework that fuses vehicle- and roadside-derived visual features via V2X communication. The model integrates four components: regional feature reconstruction (RFR) for transferring region-specific roadside cues, context-driven channel attention (CDCA) for channel recalibration, uncertainty-weighted fusion (UWF) for confidence-guided weighting, and point sampling voxel fusion (PSVF) for efficient alignment. Evaluated on the DAIR-V2X-C benchmark, our method consistently outperforms state-of-the-art feature-level fusion baselines, achieving improved A P 3 D and A P B E V (reported settings: 16.31% and 21.49%, respectively). Ablations show RFR provides the largest single-module gain +3.27% A P 3 D and +3.85% A P B E V , UWF yields substantial robustness gains, and CDCA offers modest calibration benefits. The framework enhances occlusion handling and cross-view detection while reducing dependence on explicit camera calibration, supporting more generalizable cooperative perception.

1. Introduction

As autonomous driving advances toward higher levels of automation, monocular, vehicle-centric perception encounters inherent limitations in scenarios characterized by occlusions, dense traffic participants, and complex environments. Such limitations often result in incomplete environmental perception, which is crucial to address because undetected road users or obstacles can compromise the safety and reliability of autonomous driving systems. Vehicle-infrastructure cooperative perception (VICP) has thus emerged as a promising paradigm to overcome these challenges. The development of V2I communication provides the prerequisite infrastructure for high-level vehicle-infrastructure information fusion, enabling the timely and reliable exchange of data between vehicles and roadside units []. By integrating the macroscopic, global information acquired by roadside infrastructure sensors (e.g., fixed cameras and LiDAR) with the local, dynamic data provided by onboard sensors, VICP mitigates the blind spots inherent to single-vehicle systems []. This cooperative framework not only enhances the completeness of environmental perception but also fortifies the robustness and integrity of the overall sensing architecture, establishing VICP as a critical direction for transcending the constraints of standalone perception.
Fusion methodologies in vehicle-infrastructure collaborative perception are broadly classified into three categories: early, intermediate, and late fusion []. Early fusion operates at the data level by directly combining raw sensor inputs, such as images and point clouds, preserving maximal information content but imposing stringent demands on bandwidth and sensor synchronization []. Late fusion, in contrast, performs independent detection at each sensing endpoint and subsequently merges object-level outputs, which minimizes computational overhead but limits the exploitation of complementary sensor details [,]. Intermediate fusion occupies the middle ground: local feature extraction is performed onboard, followed by cross-domain feature alignment and fusion. By balancing representational richness with communication efficiency [,], intermediate fusion has become the prevailing focus of contemporary research.
However, implementing effective feature-level fusion in VICP faces several critical challenges, as illustrated in Figure 1. Precise alignment between heterogeneous sensors is often impeded by spatio-temporal asynchrony and viewpoint mismatch []. Meanwhile, the inherently multi-scale and complementary features from different sources can introduce semantic noise if fused improperly []. In addition, many existing approaches rely on external priors (e.g., geometric calibration or known camera poses), which compromises robustness and generalization in real-world deployments where strict sensor synchronization may not hold. Moreover, most static fusion schemes neglect source-dependent uncertainty, hindering truly reliable and adaptive information integration. Therefore, it is crucial to address these limitations to achieve a reliable and generalizable cooperative perception system.
Figure 1. A visual example of VICP.
Recent advances in visual fusion and attention modeling have provided effective paradigms for enhancing perception reliability in complex multi-source environments. Specifically, some research has demonstrated that incorporating spatiotemporal interactions and adaptive attention can substantially enhance perception robustness under occlusion and dynamic illumination []. Similarly, multi-scale transformer-based fusion models [,] highlight the importance of scale-aware representation and hierarchical feature refinement, while context-aware recalibration strategies [] have been shown to effectively suppress semantic noise in cross-domain or occluded scenarios. Collectively, these studies reveal three key principles: (1) adaptive feature alignment across heterogeneous domains, (2) context-driven attention to balance multi-scale semantics, and (3) confidence-aware fusion to mitigate uncertainty in dynamic conditions.
In contrast to prior cooperative perception frameworks, such as V2X-ViT [], CoBEVT [], and EMIFF [], which primarily rely on multi-view attention or BEV-based feature alignment, our proposed framework removes dependence on explicit geometric calibration and fixed fusion heuristics. While V2X-ViT and CoBEVT enhance inter-view interaction through cross-attention and transformer structures, they still assume synchronized, high-quality inputs and lack adaptive mechanisms to handle feature-level uncertainty or degraded sensor observations. EMIFF improves efficiency through intermediate fusion but uses static weighting strategies that limit adaptability in dynamic, heterogeneous traffic conditions. Thus, existing frameworks have yet to fully resolve the issues of misalignment and uncertainty in cooperative perception, underscoring the necessity for a more adaptive and robust solution.
To address these limitations, this work integrates deformable attention, contextual channel adaptation, and uncertainty-driven fusion into a unified cooperative perception framework, thereby enhancing the robustness, adaptability, and interpretability of vehicle-infrastructure feature fusion. The main contributions can be summarized as follows:
  • A feature-level vehicle-infrastructure perception framework is proposed and experimentally validated, demonstrating superior occlusion handling and adaptability.
  • This study proposes a confidence-map-based regional feature reconstruction (RFR) that decouples roadside features and uses deformable attention to reconstruct and augment onboard features, thereby improving onboard representation and cross-view detection.
  • A context-driven channel attention (CDCA) module is proposed to exploit global image-level context for adaptive, channel-wise recalibration of multi-scale features on both vehicle and roadside sensors, thereby eliminating reliance on external calibration parameters.
  • An uncertainty-weighted fusion (UWF) mechanism is designed to estimate voxel-level uncertainty across heterogeneous feature sources and to allocate fusion weights dynamically based on confidence, significantly enhancing robustness under noise, occlusion, and projection errors.
The remainder of this paper is organized as follows. Section 2 reviews recent work on vision-based 3D object detection, vehicle-infrastructure cooperative perception, channel recalibration, and uncertainty-driven fusion. Section 3 presents the proposed methodology, including feature representation and regional feature reconstruction, the context-driven channel attention module, and the uncertainty-weighted fusion mechanism. Section 4 describes the experimental setup and reports comparative results on public datasets. Finally, Section 5 concludes the study and discusses directions for future research.

3. Methods

This work is inspired by the pipelines of EMIFF [] and other state-of-the-art VICP studies. It aims to further improve detection accuracy while relaxing the reliance on burdensome roadside camera calibration and enhancing scene generalization. Leveraging V2X communication to fuse vehicle- and infrastructure-side features, the overall framework comprises five main modules: regional feature reconstruction (RFR), which spatially decouples roadside information and directionally augments onboard representations; context-driven channel attention (CDCA), which exploits global image-level context for adaptive channel-wise recalibration; uncertainty-weighted fusion (UWF), which estimates voxel-level uncertainty across heterogeneous feature sources and dynamically allocates fusion weights by confidence; and point sampling voxel fusion (PSVF), which efficiently aligns and merges sampled point features with voxel representations. The full architecture and module details are illustrated in Figure 2.
Figure 2. Framework of the proposed model.
To better illustrate the synergy of the proposed framework, we emphasize that the four modules are designed to operate in a complementary manner rather than as isolated components. The backbone and FPN establish a unified multi-scale representation that ensures consistent feature hierarchies across infrastructure and vehicle views. On this foundation, the RFR module performs spatial alignment and compensates for geometric discrepancies, providing spatially coherent intermediate features.
Building upon these aligned representations, the CDCA module adaptively recalibrates channel responses according to the global semantic context extracted from both streams. This process effectively suppresses redundant or misaligned feature activations and enhances semantic consistency prior to fusion. The refined contextual features are then passed to the UWF module, which estimates voxel-level confidence and assigns adaptive weights to each source. The synergy between CDCA’s semantic calibration and UWF’s uncertainty weighting allows the framework to jointly reduce fusion noise and mitigate spatial misalignment, yielding more stable and interpretable representations than the standalone attention or fusion modules used in prior studies.

3.1. Feature Representation and Regional Feature Reconstruction

Inspired by recent work such as SETR-Net [], which highlights the effectiveness of enhancing a single feature representation while leveraging temporal recurrence for robust context modeling, we design our feature extraction pipeline to favor semantically rich, context-aware recalibration over purely multi-scale fusion strategies. During multi-scale feature extraction, vehicle-mounted and roadside images are processed through a shared ResNet-101 [] backbone augmented with a Feature Pyramid Network (FPN), yielding three levels of pyramid features f v e h , s , f i n f , s   ( s = 3 ,   4 ,   5 ) . Low-level features emphasize texture and edge information, facilitating fine-grained object localization, whereas high-level features capture global semantics, aiding robust detection in complex scenes. This multi-scale architecture balances representational richness with computational efficiency and provides diverse resolution cues for subsequent alignment and fusion. To address spatiotemporal asynchrony between the two streams, deformable convolutions perform pixel-level geometric correction, effectively mitigating projection shifts induced by timing offsets.
To deeply integrate heterogeneous perceptions from vehicle-mounted and roadside sensors while accounting for their spatial distributions, we introduce a region-based feature reconstruction method grounded in probabilistic confidence maps. First, confidence maps are generated for both streams to spatially decouple roadside features. Then, a deformable attention mechanism reconstructs vehicle-side features within these decoupled regions, substantially enhancing feature representation and adaptability under real-world deployment conditions.
As shown in Figure 3, considering that some regions in roadside features do not provide significant gains for improving vehicle-side perception performance, the regional decoupling module generates a corresponding confidence map P R H × W × 1 for the input vehicle-side or roadside image feature F R H × W × C through 1 × 1 convolution and Sigmoid function mapping. Each value in the confidence map reflects the probability estimate of a target existing at the corresponding position in the scene:
P v e h / i n f = σ ( C o n v ( F v e h / i n f ) )
where F v e h and F i n f denote the vehicle-side and roadside feature maps, and F v e h and F i n f correspond to the vehicle-side and roadside probability confidence maps. C o n v ( · ) represents a 1 × 1 convolutional layer; and σ ( · ) is the Sigmoid activation function.
Figure 3. Regional decoupling structure.
In object detection tasks, regions containing potential targets typically carry higher information value and are critical for downstream perception and decision-making. The introduction of probabilistic confidence maps serves as a spatial attention mechanism, enabling the network to adaptively prioritize these high-confidence regions and enhance perception performance. Building upon this, the roadside features are spatially decoupled to generate a shared perception mask c i and an exclusive perception mask e i , both are defined with respect to the vehicle-centric coordinate system. These masks are designed to capture the common and distinctive components of vehicle-infrastructure perception, respectively. This process can be formally expressed as follows:
c i = f t h d ( P v e h P i n f )
e i = f t h d ( 1 P v e h ) P i n f
where the thresholding function f t h d ( · ) binarizes the probabilistic confidence maps: it outputs 1 if the input probability exceeds the corresponding threshold τ c or τ e . Using this binary mask, the roadside features can be spatially decoupled as follows:
F i , c = c i F i n f
F i , e = e i F i n f
where F i , c and F i , e denote the shared-region and exclusive-region feature representations, respectively. This region-based decoupling strategy enables downstream vehicle-side perception to learn consistent embeddings from homologous features while exploiting complementary features to significantly amplify perceptual gains.
The binary decoupling in Equations (2)–(5) uses thresholds τ c or τ e to separate shared and exclusive perception regions. Conceptually, a lower threshold increases included region coverage, while a higher threshold enforces strict sparsity, which may risk omission of weak but informative cues. We selected the default values τ c = τ e = 0.01 as a conservative operating point that retains weak target-supporting features in low-contrast roadside views and suppresses homogeneous background activations produced by the Sigmoid output of the 1 × 1 confidence predictor. To validate robustness to threshold choice, we performed a structured sensitivity analysis on the validation split: thresholds were swept over { 0.001 ,   0.005 ,   0.01 ,   0.02 ,   0.05 } . For each threshold pair, we measured A P B E V , A P 3 D and the proportion of the feature map area selected by the binary mask. The analysis shows that model performance exhibits a broad plateau at approximately τ [ 0.005,0.02 ] , indicating insensitivity to small threshold perturbations; the chosen τ = 0.01 represents a stable trade-off between recall of weak cues and background suppression.
To address feature misalignment caused by dynamic scene changes and object motion, a deformable attention-based feature reconstruction module is proposed in Figure 4. The module comprises two parallel branches, one aggregating exclusive-area features with vehicle-side features, and the other aggregating common-area features with vehicle-side features. Although both branches share an identical network topology, they learn independent parameter sets. The key components are referred to as exclusive-area feature aggregation (EFA) [] and common-area feature aggregation (CFA) [], respectively.
Figure 4. Regional reconstruction structure.
Taking the EFA module as an example, the vehicle-side feature map F v e h R H × W × C is first projected through an embedding layer to generate the query vectors Q R N q × C , where N q denotes the number of query tokens, determined by the positions in the probabilistic confidence map that exceed the confidence threshold.
A positional embedding is then generated for each query via a linear transformation. For every query position, the network predicts a set of sampling offsets through a linear layer to dynamically adjust the sampling locations. Based on the reference point positions and their learned offsets, the corresponding feature values V R N q × K × C are extracted from the roadside feature map F i n f R H × W × C , where K is the predefined number of sampling keys.
A cross-attention fusion is performed between the query vectors Q and the sampled feature values V to generate the enhanced feature representation F a t t . Specifically, within the cross-attention block, the feature values V are linearly transformed into keys and values, and the similarity between Q and the keys is computed to produce attention weights through an activation function, which are then applied to the values. The output F a t t of the cross-attention block is linearly transformed and integrated into the original vehicle-side feature F v e h to obtain the exclusive-area feature representation F v e h , e . The generation process of F v e h , e follows the workflow illustrated in Figure 5.
Figure 5. Exclusive-Area Feature Aggregation Structure.
Following EFA and CFA, their outputs F v , e and F v , c are each passed through a 3 × 3 convolution to produce weight maps A e and A c . These maps enable adaptive fusion of the two feature streams, effectively leveraging the complementary strengths of distinct perceptual regions to enhance vehicle-side feature representation:
F v e h = A e F v , e + A c F v , c

3.2. Context-Driven Channel Attention

In previous studies, spatial geometric priors were typically injected into the network through the intrinsic and extrinsic parameters of cameras to guide feature recalibration. However, this method is highly dependent on the accuracy of external calibration, often leading to performance fluctuations in practical deployment due to calibration errors or synchronization issues. To address these shortcomings, this study proposes a CDCA module whose workflow comprises the following:
CDCA takes as input the infrastructure features F i n f and the preliminary enhanced vehicle features F v e h . Global average pooling (GAP) and global max pooling (GMP) are applied on each feature tensor F { F i n f , F v e h } to obtain channel descriptors:
v a v g = G A P ( F )
v m a x = G M P ( F )
where [ v a v g ; v m a x ] are concatenated along the channel dimension, and then pass through a two-layer MLP, with a bottleneck ratio r , followed by ReLU and Sigmoid activations to produce the attention weights:
z = σ ( W 2 δ ( W 1 [ v a v g ; v m a x ] ) ) R C
Next, we proceed with the intelligent recalibration of the channels by performing element-wise multiplication on F ; multiplying F with z , we obtain the recalibrated features:
F ~ = z F
The above recalibration is independently applied to and F i n f and F v e h , producing F ~ i n f and F ~ v e h for subsequent alignment and UWF processing.
The channel logits are computed from compact global descriptors obtained by spatial pooling: for a feature map F { F i n f , F v e h } , we form descriptors by GAP and GMP and feed their concatenation to a small two-layer bottleneck MLP. Concretely, the following is obtained:
z = S i g m o i d ( M L P ( [ G A P F , G M P ( F ) ] ) )
The MLP uses a bottleneck ratio ( r = 16 ) with shared weights across streams; to mitigate overfitting we apply dropout (rate = 0.1) and weight decay. In practice the MLP can be implemented as 1 × 1 convolutions on the pooled descriptors for efficiency; CDCA is applied independently at each FPN level and yields per-stream attention vectors.
The CDCA module is completely driven by global contextual data, without the need for external parameters, and has stronger robustness and generalization ability under heterogeneous and asynchronous deployment conditions.

3.3. Uncertainty-Weighted Fusion Mechanism

To enable adaptive fusion of heterogeneous features in the voxel domain, we introduce an UWF mechanism. Multi-scale features recalibrated by CDCA are projected onto a common 3D voxel grid, yielding vehicle-side voxel features V v e h and infrastructure-side voxel features V i n f . In our implementation, selected pixels are inverse-projected via X = π 1 ( u , d ( u ) ) and assigned to voxel indices by i , j , k = ( x x m i n ) / v x , ( y y m i n ) / v y , ( z z m i n ) / v z before per-voxel aggregation. Voxel-wise uncertainty σ v is estimated either via variance across multiple frame projections or through a lightweight MLP regressor. This uncertainty is then mapped to a confidence weight:
w v = exp α σ v ,   α > 0
which governs the relative contribution of each stream. The fused voxel representation is obtained by the following:
V f u s e v = w v V v e h v + ( 1 w v ) V i n f v
By dynamically modulating fusion weights according to feature reliability, UWF effectively mitigates the impact of noise, occlusion, and projection errors, thereby enhancing the robustness and discriminative power of the fused representation. The resulting voxel tensor V f u s e is subsequently processed by a 3D convolutional network and a BEV detection head to produce accurate 3D object predictions.

3.4. Point Sampling Voxel Fusion

To ensure reproducibility and computational efficiency in lifting image features into the shared voxel representation, we implement PSVF that concentrates representational capacity on pixels identified as informative by the RFR confidence map, which maps sampled pixels into a 3D voxel grid and emits compact per-voxel descriptors with sampling-confidence summaries consumed by the UWF stage. PSVF samples a bounded set of pixel locations from each selected feature map using a mixture strategy: most samples are drawn by importance sampling from the RFR confidence distribution C ( u ) , which is the probabilistic map produced by Equation (1), while a small fraction is drawn uniformly to preserve coverage of low-confidence areas. Concretely, importance sampling probabilities are formed as p i m p ( u ) C ( u ) α , where the exponent α controls concentration on high-confidence pixels; a uniform fallback fraction β prevents complete neglect of low-confidence regions. Empirically, modest values of α and small β balance robustness and selectivity.
Each sampled pixel u is associated with a depth proxy d ( u ) , which is the predicted depth, stereo estimate, or scene prior, and is inverse-projected to a 3D point in the sensor frame via X s e n s o r = π 1 ( u , d ( u ) ) . The point is transformed to the common/world frame by the sensor extrinsic T s e n s o r w o r l d , yielding X s e n s o r w o r l d . X s e n s o r Voxel indices are obtained with the convention.
i = x x m i n v x , j = y y m i n v y , k = z z m i n v z
where x , y , z = X w o r l d , ( x m i n , y m i n , z m i n ) denote the voxel grid origin and ( v x , v y , v z ) the voxel sizes. Points outside the grid bounds are discarded. The voxelization and filling process are illustrated in Figure 6. To bound memory and latency, each voxel collects up to a fixed cap N v points. Within each voxel, collected point features { f n } are aggregated with a learnable per-point weighting. A shallow MLP produces scalar scores s n = M L P ( F n ) , which are normalized by a softmax to obtain attention weights a n . The voxel descriptor is the weighted sum:
F v o x e l = n = 1 N v a n f n
Figure 6. Voxel filling diagram.
PSVF also computes a concise confidence summary for each voxel, for example, c = [ m a x n a n , m e a n n a n ] , which quantifies sampling concentration and provides an evidential cue. The aggregated voxel descriptor and its confidence summary are concatenated and provided as inputs to the UWF uncertainty regressor, thereby enabling UWF to consider both feature content and sampling reliability when producing fusion weights. In practice, this design enables the fusion module to down-weight voxels with diffuse or noisy evidence even when scattered point activations exist. According to the experiment, the parameters are set to the following: per-frame sample budget N m a x = 2048 , per-voxel cap N v = 8 , confidence exponent α = 2 , and uniform fallback fraction β = 0.1 . To evaluate sensitivity, we recommend sweeping N m a x { 1024 ,   2048 ,   4096 } , N v { 4 ,   8 ,   16 } , and β { 0 ,   0.1 ,   0.2 } , and recording A P 3 D and A P B E V on a per-frame latency to guide deployment tradeoffs.

4. Dataset Description and Experimental Setup

4.1. Dataset Description

DAIR-V2X constitutes China’s inaugural public large-scale vehicle-infrastructure collaborative multi-modal dataset, acquired within Beijing’s advanced autonomous driving demonstration zone. It encompasses complex urban environments, including intersections and roundabouts, under diverse meteorological conditions. Through rigorous spatiotemporal synchronization with microsecond-level precision, the dataset comprises 72,890 valid frames sampled at 10 Hz from 100 continuous 20 s sequences, partitioned into three subsets: the vehicle-infrastructure collaborative set (DAIR-V2X-C, comprising 40481 image-point cloud pairs), infrastructure-only set (DAIR-V2X-I), and vehicle-only set (DAIR-V2X-V).
Annotations employ LiDAR-coordinated 2D–3D joint labeling across 15 obstacle categories (e.g., pedestrians, vehicles). Each frame integrates 3D bounding boxes (position, dimensions, orientation), 2D image annotations, occlusion levels, truncation states, and cross-frame tracking identifiers. This protocol yields over 1.2 million high-quality 3D annotations spanning 230,000 unique instances, establishing foundational benchmarks for collaborative perception research.

4.2. Experimental Settings

In this study, all experiments were conducted on a single server equipped with 8 NVIDIA GeForce RTX 3090 GPUs and an Intel Xeon Gold 5320 CPU to ensure reproducibility and fair comparison. The software environment comprised Ubuntu 20.04, CUDA 11.6 with CuDNN 8.3, and Python 3.7. The model was implemented in pyTorch 1.9.1, with data preprocessing and visualization facilitated by OpenCV 4.7 and Matplotlib 3.5.2.
The experiments were implemented based on the MMDetection3D framework (version 1.4.0, Multimedia Laboratory, The Chinese University of Hong Kong, Hong Kong, China) with its default training configurations. The proposed model was trained for 20 epochs with a batch size of 8, using the AdamW optimizer with an initial learning rate of 1 × 10−3 and a weight decay of 0.0001. The RepeatDataset strategy in MMDetection3D was employed to maintain data balance, where each scene was repeated three times per epoch. The detection head adopted Rotated Non-Maximum Suppression (NMS) to filter redundant bounding boxes and generate the final predictions. It is worth noting that only vehicles were considered as the detection category in this study.

4.3. Evaluation Metrics

For the 3D object detection task, evaluation follows standard object detection protocols, using Precision, Recall, and Average Precision (AP) to quantify detection accuracy. AP measures the area under the Precision, Recall curve and is typically evaluated at specific Intersection over Union (IoU) thresholds, such as AP@0.5 or AP@0.7, to ensure sufficient spatial overlap between predictions and ground truth.
In 3D detection, two complementary AP metrics are commonly reported: A P B E V , which assesses localization accuracy on the BEV plane, and A P 3 D , which further incorporates the predicted bounding box’s height and center position, providing a stricter evaluation of volumetric precision. The mean Average Precision (mAP) represents the overall performance averaged across all object categories, which is defined as the arithmetic mean of AP values over all classes and provides a holistic assessment of a detector’s performance. It is calculated as follows:
m A P = 1 N c l s i = 1 N c l s A P i
where N c l s represents the number of target categories, and A P i is the AP value of the target in the i t h category.
For transparency and to facilitate reproducibility, we provide here the mathematical definition and computational steps for the 3D IoU used in our experiments. A 3D bounding box B is parameterized as ( x c ,   y c ,   z c ,   l ,   w ,   h ,   θ ) , where ( x c ,   y c ,   z c ) is the box center in world coordinates, l ,   w ,   h denote length, width, and height, and θ is the yaw around the vertical axis. The 3D IoU between two boxes, ground truth B g and prediction B p , is defined as
I o U 3 D ( B g , B p ) = V o l ( B g B p ) V o l ( B g B p )
The intersection volume computation is implemented by decomposing each box into a vertical prism: (i) compute the 2D polygon intersection area A between the yaw-rotated rectangles on the ground plane, and (ii) compute the overlapped height interval as follows:
H = m a x ( 0 , min z t o p g , z t o p p m a x ( z t o p g , z t o p p ) )
where z t o p = z c + h / 2 , z b o t = z c h / 2 . Thus,
V o l ( B g B p ) = A · H
Because the 2D polygon intersection A depends on the yaw θ , small heading errors θ can reduce A nonlinearly for elongated objects. For small angular perturbations, a first-order approximation yields the following:
A κ ( l ,   w , θ ) · | θ |
where κ ( · ) is a scene and geometry dependent factor larger for high aspect ratio boxes and for near-edge alignments; this explains why IoU drops faster for orientation-biased cases. Consequently, A P 3 D is more sensitive to heading errors than A P B E V , which ignores height and, to a lesser extent, vertical misalignment.
The BEV projection discards height information by evaluating the IoU of 3D bounding boxes projected onto the ground plane. Mathematically, this renders the BEV IoU invariant to pure vertical displacements, tolerating height errors that do not alter the 2D footprint. This makes A P B E V a valid and efficient metric for ground-level path planning, where lateral and longitudinal extents are primary constraints. In contrast, the full 3D IoU is sensitive to orientation and height, making A P 3 D indispensable for tasks requiring volumetric precision, such as vertical clearance assessment. We, therefore, report both metrics: A P B E V to quantify planning-centric horizontal localization and A P 3 D to quantify full spatial fidelity.

5. Results and Analysis

5.1. Analysis of Comparative Experimental Results

To rigorously evaluate the proposed method, we performed comparative experiments on the DAIR-V2X-C dataset against leading vision-based cooperative 3D object detection frameworks. In the first set of experiments, we assessed single-agent perception by adapting our network to vehicle-only and infrastructure-only workflows: we retained the feature-extraction and BEV-encoding modules while removing the region-level reconstruction branch, yielding the configurations denoted Ours_Veh and Ours_Inf, which we compared to the corresponding EMIFF_Veh and EMIFF_Inf baselines.
Next, decision-level fusion was implemented using the official DAIR-V2X fusion protocol: EMIFF_Veh and EMIFF_Inf were merged to form the baseline decision-level fusion (DL) configuration, and Ours_Veh and Ours_Inf were fused in the same manner to produce our decision-level fusion variant. Finally, we evaluated the full feature-level fusion (FL) capability of our model against representative feature-level fusion methods, including EMIFF [] and QUEST []. All comparisons focus exclusively on the car category, with detection accuracy measured by mean Average Precision (mAP) in both 3D and BEV projections at IoU thresholds of 0.3 and 0.5. Communication overhead accompanying each setting is also reported. Quantitative results are summarized in Table 1.
Table 1. Detection results for different types of models.
Experimental results demonstrate that, for all evaluated models and IoU thresholds, the mean Average Precision in A P 3 D is systematically lower than in A P B E V . This disparity arises because 3D detection must recover object height information, substantially increasing model complexity and learning difficulty; in typical traffic scenarios, however, all objects lie on a common ground plane, so height contributes minimally to planning and decision-making.
Moreover, both the baseline EMIFF models and the proposed method exhibit inferior single-agent performance on the vehicle side compared to the infrastructure side. This degradation is primarily due to the vehicle-mounted camera’s limited field of view and susceptibility to occlusion, which reduces data quality. Consequently, supplementing vehicle perception with infrastructure imagery yields significant practical benefits, as cooperative perception consistently outperforms vehicle-only detection under any fusion strategy.
Regarding fusion strategies, FL incurs higher communication overhead—owing to the transmission of full feature maps—than DL, which exchanges only detection outputs. Nevertheless, FL achieves superior accuracy by more effectively leveraging multi-source contextual semantics. Under FL, the proposed method surpasses both EMIFF and QUEST across all IoU thresholds. At I o U = 0.5 , our model attains an A P 3 D of 15.76%, compared to 15.07% for EMIFF and 14.10% for QUEST; similarly, A P B E V reaches 21.02%, exceeding the baselines. These results validate that the introduced BEV-encoding strategy enhances height information utilization, thereby improving 3D detection performance.
To investigate the influence of backbone depth on object detection performance, ResNet-50 [], ResNet-101 [], and VGG-16 [] were adopted as the feature extraction networks for both the EMIFF baseline and the proposed method. The mAP is evaluated at IoU thresholds of 0.3 and 0.5. The results are summarized in Table 2.
Table 2. Detection performance across different feature extraction backbones.
In vehicle-infrastructure cooperative object detection tasks, the choice of backbone network exerts a profound influence on overall performance. Compared with the shallow VGG16, ResNet architectures leverage residual connections to deliver substantially richer feature representations. In particular, the deeper ResNet101 extracts more discriminative high-level semantic features, boosting the EMIFF baseline’s average precision by 0.89% and its average recall by 0.77%, while our proposed model achieves corresponding gains of 0.57% and 0.55%. However, these accuracy improvements incur a marked computational overhead: our model’s parameter count rises from 54.72 million to 92.66 million and its GFLOPs increase from 107.44 billion to 184.93 billion, whereas the EMIFF baseline grows from 49.33 million to 87.31 million parameters and from 123.76 to 201.46 GFLOPs. Such an almost two-fold expansion in model size and inference workload challenges the strict real-time and resource-constrained requirements of intelligent transportation systems. Accordingly, to strike the optimal balance between detection accuracy and computational efficiency, this study adopts ResNet-50 as the feature extraction backbone.
To further verify computational efficiency, local inference tests were conducted without deployment acceleration. All evaluations were performed on the workstation described in Section 4.2. Two input resolutions (1280 × 720 and 1920 × 1080) were tested, and the corresponding inference results are shown in Table 3. We report the FPS of the Backbone + FPN stage, the overall Total FPS, and the Total Latency (ms). Each measurement was averaged over 300 frames, after 100 warm-up frames, using CUDA events. The results show that Res50_Ours achieves near real-time performance on a single GTX 3090, while Res101_Ours maintains reasonable latency with improved accuracy.
Table 3. Inference results under different input resolutions.

5.2. Ablation Study

To verify the effectiveness of each module in the proposed model, we designed ablation experiments to evaluate the impact of the RFR module, the camera parameter embedding module, its alternative based on global image-level context for adaptive recalibration (i.e., CDCA), and the UWF module on both 3D object detection and BEV object detection performance. The results are reported in Table 4.
Table 4. Ablation results on model architecture, where symbol √ indicates that the corresponding module is enabled.
At an IoU threshold of 0.5, replacing the camera parameter embedding module with the CDCA module yields relative improvements of 0.15% in A P 3 D and 0.18% in A P B E V . Although the gains are modest, they are notable given the simplicity of the module and its ability to eliminate the need for rule-based camera parameter settings across different scenes, thereby facilitating model transferability. The UWF module, by estimating voxel-level uncertainty across heterogeneous feature sources and dynamically allocating fusion weights, provides a more substantial benefit, improving A P 3 D and A P B E V by 1.47% and 2.19%, respectively. Most importantly, the RFR module achieves the largest improvements, with A P 3 D and A P B E V increasing by 3.27% and 3.85%, respectively, demonstrating that region-wise reconstruction of onboard features using roadside information significantly enhances model performance.
To further assess the role of the PSVF module, we compared it against two alternative voxelization schemes: (1) uniform voxel sampling, which randomly samples pixels with equal probability, and (2) full-grid voxelization, which converts all projected points into voxels without sampling.
As summarized in Table 5, PSVF achieves the highest efficiency and accuracy trade-off. Specifically, compared with full-grid voxelization, PSVF reduces total latency by 26%, while also slightly outperforming uniform sampling in both A P 3 D and A P B E V metrics. The results confirm that PSVF provides a compact yet informative voxel representation, effectively concentrating computational resources on salient high-confidence regions identified by the RFR confidence maps. This design preserves the structural integrity of the fusion pipeline while delivering significant gains in efficiency and detection performance.
Table 5. Comparison of different voxel construction strategies in the fusion stage.

5.3. Visualization

To demonstrate the detection performance of the proposed model, we present its detection results across different scenarios from three perspectives: the onboard camera view, the roadside camera view, and the ego-centric BEV. We also provide a performance comparison under identical scene settings between a single-vehicle detector, baseline methods, and our proposed model; the results are shown in Figure 7, Figure 8 and Figure 9. In the visualizations, green bounding boxes denote ground truth, red bounding boxes denote predicted detections, and the blue point indicates the ego vehicle’s position.
Figure 7. Visualization of results across different scenarios, where green bounding boxes represent ground truth and red bounding boxes denote the predicted results.
Figure 8. Visual comparison of proposed model and baselines, where green bounding boxes represent ground truth and red bounding boxes denote the predicted results.
Figure 9. Visual comparison of DAIR-V2X-V and DAIR-V2X-C, where green bounding boxes represent ground truth and red bounding boxes denote the predicted results.
In sparse traffic-flow scenarios, the proposed model achieves accurate detection for the vast majority of vehicles. In dense traffic-flow scenarios, our model not only effectively identifies mutually occluded vehicles but also detects vehicles that lie outside the onboard camera’s field of view, as indicated by the circled region in Scene 3 of Figure 7. However, there remain some missed detections in areas roughly 30m ahead of the ego vehicle and about 25 m laterally beyond the ego sensing range, indicating that long-range perception remains an area for improvement.
Across identical traffic scenes, observable differences exist between the baseline and our proposed model. As shown in Figure 8, EMIFF effectively fuses roadside perception data and partially mitigates the beyond-ego-field-of-view detection problem, substantially improving the detection of occluded vehicles and enabling more comprehensive perception in complex environments such as intersections. Nevertheless, EMIFF still exhibits small but systematic biases in orientation estimation and localization when processing occluded targets, which is likely attributable to insufficient modeling of voxel-level uncertainty across heterogeneous feature sources. By fully leveraging roadside augmentation and complementary information, our proposed model achieves significant improvements in both detection accuracy and spatial coverage.
To further validate the generalizability of the proposed framework, we additionally compare the single-vehicle detection model trained on DAIR-V2X-V with the vehicle-infrastructure cooperative model trained on DAIR-V2X-C under identical driving scenes.
As shown in Figure 9, the single-vehicle model exhibits severe performance degradation when target vehicles are partially or fully occluded by leading traffic, resulting in clear missed detections within the blind area of the onboard camera. In contrast, the cooperative perception model effectively compensates for these occlusions by integrating roadside visual cues, enabling accurate long-range detection and complete spatial coverage. This comparison provides intuitive evidence that feature-level vehicle-infrastructure fusion significantly enhances robustness and perception completeness in complex traffic environments.

6. Conclusions

This study proposed a feature-level VICP framework that integrates vehicle- and roadside-derived features through V2X communication. The framework combines RFR, CDCA, UWF, and PSVF to jointly address occlusion, calibration dependency, and feature heterogeneity. Comprehensive experiments on the DAIR-V2X-C benchmark demonstrated consistent improvements over state-of-the-art baselines, while ablation studies confirmed the complementary contributions of each module. These findings establish VICP as an effective means to extend sensing coverage and enhance robustness, offering a practical reference for scalable deployment within smart city infrastructure.
While attention mechanisms and uncertainty weighting have been applied independently in prior VICP works (e.g., EMIFF uses multi-scale fusion and channel masking; V2X-ViT applies multi-agent attention), our contribution is twofold and integrative: (1) RFR performs explicit spatial decoupling between shared and exclusive perception regions, which prevents misleading roadside-exclusive semantics from contaminating vehicle-side embeddings during cross-view reconstruction; and (2) the combination of CDCA and UWF acts at complementary granularities, CDCA provides channel-wise semantic rebalancing conditioned on global context, whereas UWF controls voxel-level fusion strength according to estimated uncertainty. The synergy of these components yields robustness to viewpoint mismatch and noisy inputs that is not achieved by standalone attention modules or by post-hoc weighting alone.
Future research will extend this work in three directions: integrating additional sensing modalities such as radar and event cameras, improving real-time efficiency and communication adaptability for large-scale V2X networks, and exploring cross-domain generalization to ensure reliable operation under diverse environmental and traffic conditions. Despite the promising results, the current framework still relies on high-quality temporal synchronization in the DAIR-V2X-C dataset, which may constrain its scalability to heterogeneous infrastructures and complex traffic scenarios. Moreover, long-range object perception remains challenging due to reduced feature consistency and signal attenuation across sensing modalities. Furthermore, the robustness of the current framework could be affected under conditions of significant motion blur or temporal asynchrony, which were not the focus of this study. Enhancing temporal alignment and mitigating these effects represent key objectives for future work, along with improving generalization and deployment feasibility in real-world cooperative perception systems.

Author Contributions

Conceptualization, S.Y. and S.W.; methodology, S.Y. and S.W.; software, S.Y. and J.P.; validation, S.Y. and J.P.; formal analysis, S.Y.; investigation, S.Y. and; resources J.P. and C.M.; data curation, S.W. and D.W.; writing—original draft preparation, S.Y. and S.W.; writing—review and editing, S.Y., J.P. and S.W.; visualization, S.Y. and D.W.; supervision, C.M.; project administration, J.P.; funding acquisition, S.Y. and J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key R&D Program of Shandong Province, China, grant number 2023CXGC010111; and the SEU Innovation Capability Enhancement Plan for Doctoral Students of southeast university, grant number CXJH_SEU 24180.

Data Availability Statement

The data used in this study are publicly available from the DAIR-V2X dataset at https://air.tsinghua.edu.cn/DAIR-V2X/index.html (accessed on 5 August 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Clancy, J.; Molloy, D.; Hassett, S.; Leahy, J.; Ward, E.; Denny, P.; Jones, E.; Glavin, M.; Deegan, B. Evaluating the feasibility of intelligent blind road junction V2I deployments. Smart Cities 2024, 7, 973–990. [Google Scholar] [CrossRef]
  2. Yang, L.; Zhang, X.; Yu, J.; Li, J.; Zhao, T.; Wang, L.; Huang, Y.; Zhang, C.; Wang, H.; Li, Y. MonoGAE: Roadside monocular 3D object detection with ground-aware embeddings. IEEE Trans. Intell. Transp. Syst. 2024, 25, 17587–17601. [Google Scholar] [CrossRef]
  3. Han, Y.; Zhang, H.; Li, H.; Jin, Y.; Lang, C.; Li, Y. Collaborative perception in autonomous driving: Methods, datasets, and challenges. IEEE Intell. Transp. Syst. Mag. 2023, 15, 131–151. [Google Scholar] [CrossRef]
  4. Arnold, E.; Dianati, M.; De Temple, R.; Fallah, S. Cooperative perception for 3D object detection in driving scenarios using infrastructure sensors. IEEE Trans. Intell. Transp. Syst. 2020, 23, 1852–1864. [Google Scholar] [CrossRef]
  5. Wang, T.H.; Manivasagam, S.; Liang, M.; Yang, B.; Zeng, W.; Urtasun, R. V2vnet: Vehicle-to-vehicle communication for joint perception and prediction. In Proceedings of the Computer Vision—ECCV 2020, Virtually, 23–28 August 2020; pp. 605–621. [Google Scholar]
  6. Hu, Y.; Lu, Y.; Xu, R.; Xie, W.; Chen, S.; Wang, Y. Collaboration helps camera overtake lidar in 3d detection. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 9243–9252. [Google Scholar]
  7. Luo, G.; Zhang, H.; Yuan, Q.; Li, J. Complementarity-enhanced and redundancy-minimized collaboration network for multi-agent perception. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 3578–3586. [Google Scholar]
  8. Xu, R.; Xiang, H.; Tu, Z.; Xia, X.; Yang, M.H.; Ma, J. V2x-ViT: Vehicle-to-everything cooperative perception with vision transformer. In Proceedings of the ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; pp. 107–124. [Google Scholar]
  9. Cui, G.; Zhang, W.; Xiao, Y.; Yao, L.; Fang, Z. Cooperative perception technology of autonomous driving in the internet of vehicles environment: A review. Sensors 2022, 22, 5535. [Google Scholar] [CrossRef]
  10. Liu, J.; Wang, P.; Wu, X. A Vehicle-Infrastructure Cooperative Perception Network Based on Multi-Scale Dynamic Feature Fusion. Appl. Sci. 2025, 15, 3399. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Zhang, Y.; Xiao, Y.; Wang, T. Spatiotemporal Dual-Branch Feature-Guided Fusion Network for Driver Attention Prediction. Expert Syst. Appl. 2025, 292, 128564. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Wang, S.; Zhang, Y.; Yu, P. Asymmetric light-aware progressive decoding network for RGB-thermal salient object detection. J. Electron. Imaging 2025, 34, 013005. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Yu, P.; Xiao, Y.; Wang, S. Pyramid-structured multi-scale transformer for efficient semi-supervised video object segmentation with adaptive fusion. Pattern Recognit. Lett. 2025, 194, 48–54. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Liu, T.; Zhen, J.; Kang, Y.; Cheng, Y. Adaptive downsampling and scale enhanced detection head for tiny object detection in remote sensing image. IEEE Geosci. Remote Sens. Lett. 2025, 22, 6003605. [Google Scholar] [CrossRef]
  15. Xu, R.; Tu, Z.; Xiang, H.; Shao, W.; Zhou, B.; Ma, J. CoBEVT: Cooperative bird’s eye view semantic segmentation with sparse transformers. arXiv 2022, arXiv:2207.02202. [Google Scholar]
  16. Wang, Z.; Fan, S.; Huo, X.; Xu, T.; Wang, Y.; Liu, J.; Chen, Y.; Zhang, Y.Q. Emiff: Enhanced multi-scale image feature fusion for vehicle-infrastructure cooperative 3d object detection. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 16388–16394. [Google Scholar]
  17. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; p. 28. [Google Scholar]
  19. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  20. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  21. Brazil, G.; Liu, X. M3d-rpn: Monocular 3d region proposal network for object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9287–9296. [Google Scholar]
  22. Liu, Z.; Wu, Z.; Tóth, R. Smoke: Single-stage monocular 3d object detection via keypoint estimation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 996–997. [Google Scholar]
  23. Tang, X.; Wang, W.; Song, H.; Zhao, C. CenterLoc3D: Monocular 3D vehicle localization network for roadside surveillance cameras. Complex Intell Syst. 2023, 9, 4349–4368. [Google Scholar] [CrossRef]
  24. Ding, M.; Huo, Y.; Yi, H.; Wang, Z.; Shi, J.; Lu, Z.; Luo, P. Learning depth-guided convolutions for monocular 3d object detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1000–1001. [Google Scholar]
  25. Wang, Y.; Chao, W.L.; Garg, D.; Hariharan, B.; Campbell, M.; Weinberger, K.Q. Pseudo-lidar from visual depth estimation: Bridging the gap in 3d object detection for autonomous driving. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8445–8453. [Google Scholar]
  26. Ma, X.; Wang, Z.; Li, H.; Zhang, P.; Ouyang, W.; Fan, X. Accurate monocular 3d object detection via color-embedded 3d reconstruction for autonomous driving. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27–28 October 2019; pp. 6851–6860. [Google Scholar]
  27. Chabot, F.; Chaouch, M.; Rabarisoa, J.; Teuliere, C.; Chateau, T. Deep manta: A coarse-to-fine many-task network for joint 2d and 3d vehicle analysis from monocular image. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2040–2049. [Google Scholar]
  28. He, T.; Soatto, S. Mono3d++: Monocular 3d vehicle detection with two-scale 3d hypotheses and task priors. Proc. AAAI Conf. Artif. Intell. 2019, 33, 8409–8416. [Google Scholar] [CrossRef]
  29. Wang, Y.; Guizilini, V.C.; Zhang, T.; Wang, Y.; Zhao, H.; Solomon, J. Detr3d: 3d object detection from multi-view images via 3d-to-2d queries. In Proceedings of the 5th Conference on Robot Learning (CoRL 2021), London, UK, 8–11 November 2022; pp. 180–191. [Google Scholar]
  30. Philion, J.; Fidler, S. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 194–210. [Google Scholar]
  31. Li, Y.; Ge, Z.; Yu, G.; Yang, J.; Wang, Z.; Shi, Y.; Sun, J.; Li, Z. Bevdepth: Acquisition of reliable depth for multi-view 3d object detection. Proc. AAAI Conf. Artif. Intell. 2023, 37, 1477–1485. [Google Scholar] [CrossRef]
  32. Song, Z.; Yang, L.; Xu, S.; Liu, L.; Xu, D.; Jia, C.; Jia, F.; Wang, L. Graphbev: Towards robust bev feature alignment for multi-modal 3d object detection. In Proceedings of the 2024 European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 347–366. [Google Scholar]
  33. Cai, H.; Zhang, Z.; Zhou, Z.; Li, Z.; Ding, W.; Zhao, J. Bevfusion4d: Learning lidar-camera fusion under bird’s-eye-view via cross-modality guidance and temporal aggregation. arXiv 2023, arXiv:2303.17099. [Google Scholar]
  34. Fan, S.; Wang, Z.; Huo, X.; Wang, Y.; Liu, J. Calibration-free bev representation for infrastructure perception. In Proceedings of the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; pp. 9008–9013. [Google Scholar]
  35. Xu, R.; Xiang, H.; Xia, X.; Han, X.; Li, J.; Ma, J. Opv2v: An open benchmark dataset and fusion pipeline for perception with vehicle-to-vehicle communication. In Proceedings of the 2022 IEEE International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 2583–2589. [Google Scholar]
  36. Xiang, C.; Feng, C.; Xie, X.; Shi, B.; Lu, H.; Lv, Y.; Yang, M.; Niu, Z. Multi-sensor fusion and cooperative perception for autonomous driving: A review. IEEE Intell. Transp. Syst. Mag. 2023, 15, 36–58. [Google Scholar] [CrossRef]
  37. Xu, R.; Chen, W.; Xiang, H.; Liu, L.; Ma, J. Model-agnostic multi-agent perception framework. arXiv 2022, arXiv:2203.13168. [Google Scholar]
  38. Zhou, Y.; Yang, C.; Wang, P.; Wang, C.; Wang, X.; Van, N.N. ViT-FuseNet: MultiModal fusion of vision transformer for vehicle-infrastructure cooperative perception. IEEE Access 2024, 12, 31640–31651. [Google Scholar] [CrossRef]
  39. Liu, H.; Gu, Z.; Wang, C.; Wang, P.; Vukobratovic, D. A lidar semantic segmentation framework for the cooperative vehicle-infrastructure system. In Proceedings of the 2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall), Hong Kong, China, 10–13 October 2023; pp. 1–5. [Google Scholar]
  40. Yang, D.; Yang, K.; Wang, Y.; Liu, J.; Xu, Z.; Yin, R.; Zhai, P.; Zhang, L. How2comm: Communication-efficient and collaboration-pragmatic multi-agent perception. In Proceedings of the 37th Conference on Neural Information Processing Systems (NeurIPS 2023), Orleans, LA, USA, 10–16 December 2023; Volume 36, pp. 25151–25164. [Google Scholar]
  41. Zhou, L.; Gan, Z.; Fan, J. CenterCoop: Center-based feature aggregation for communication-efficient vehicle-infrastructure cooperative 3d object detection. IEEE Robot. Autom. Lett. 2023, 9, 3570–3577. [Google Scholar] [CrossRef]
  42. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  43. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the Computer Vision—ECCV 2018: 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  44. Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
  45. Kendall, A.; Gal, Y. What uncertainties do we need in bayesian deep learning for computer vision? In Proceedings of the 31st International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 4–9 December 2017; p. 30. [Google Scholar]
  46. Nabati, R.; Qi, H. Centerfusion: Center-based radar and camera fusion for 3d object detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 5–9 January 2021; pp. 1527–1536. [Google Scholar]
  47. Liang, T.; Xie, H.; Yu, K.; Xia, Z.; Lin, Z.; Wang, Y.; Tang, T.; Wang, B.; Tang, Z. Bevfusion: A simple and robust lidar-camera fusion framework. In Proceedings of the 36th International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 28 November–9 December 2022; Volume 35, pp. 10421–10434. [Google Scholar]
  48. Li, G.; Yang, L.; Lee, C.G.; Wang, X.; Rong, M. A Bayesian deep learning RUL framework integrating epistemic and aleatoric uncertainties. IEEE Trans. Ind. Electron. 2020, 68, 8829–8841. [Google Scholar] [CrossRef]
  49. Zhang, Y.; Xiao, Y.; Zhang, Y.; Zhang, T. Video saliency prediction via single feature enhancement and temporal recurrence. Eng. Appl. Artif. Intell. 2025, 160, 111840. [Google Scholar] [CrossRef]
  50. Demir, A.; Yilmaz, F.; Kose, O. Early detection of skin cancer using deep learning architectures: Resnet-101 and inception-v3. In Proceedings of the 2019 Medical Technologies Congress (TIPTEKNO), Izmir, Turkey, 3–5 October 2019; pp. 1–4. [Google Scholar]
  51. Vangimalla, R.R.; Sreevalsan-Nair, J. A multiscale consensus method using factor analysis to extract modular regions in the functional brain network. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) in Conjunction with the 43rd Annual Conference of the Canadian Medical and Biological Engineering Society, Virtual, 20–24 July 2020; pp. 2824–2828. [Google Scholar]
  52. Zhang, Y.; Wu, C.; Guo, W.; Zhang, T.; Li, W. CFANet: Efficient detection of UAV image based on cross-layer feature aggregation. IEEE Trans. Geosci. 2023, 61, 5608911. [Google Scholar] [CrossRef]
  53. Fan, S.; Yu, H.; Yang, W.; Yuan, J.; Nie, Z. Quest: Query stream for vehicle-infrastructure cooperative perception. arXiv 2023, arXiv:2308.01804. [Google Scholar]
  54. Wen, L.; Li, X.; Gao, L. A transfer convolutional neural network for fault diagnosis based on ResNet-50. Neural Comput. Appl. 2019, 32, 6111–6124. [Google Scholar] [CrossRef]
  55. Haque, M.F.; Lim, H.Y.; Kang, D.S. Object detection based on VGG with ResNet network. In Proceedings of the 2019 International Conference on Electronics, Information, and Communication (ICEIC), Auckland, New Zealand, 22–25 January 2019; pp. 1–3. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.