Next Article in Journal
Debiasing Session-Based Recommendation for the Digital Economy: Propensity-Aware Training and Temporal Contrast on Graph Transformers
Previous Article in Journal
CAXSS: A Cookie Authentication Scheme Against XSS Attacks for HTTPS
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Bidirectional Complementary Cross-Attention and Temporal Adaptive Fusion for 3D Object Detection in Intelligent Transportation Scenes

1
School of Mechanical Engineering, Xi’an Shiyou University, Xi’an 710065, China
2
School of Automobile, Chang’an University, Xi’an 710064, China
3
Lishui Key Laboratory of New Energy Vehicle Sensing and Control Technology, Zhejiang Yili Auto Mobile Air Condition Co., Ltd., Lishui 323700, China
4
School of Intelligent Manufacturing, Longquan Celadon Sword Technician College, Lishui 323700, China
*
Authors to whom correspondence should be addressed.
Electronics 2026, 15(1), 83; https://doi.org/10.3390/electronics15010083
Submission received: 20 November 2025 / Revised: 13 December 2025 / Accepted: 23 December 2025 / Published: 24 December 2025

Abstract

Multi-sensor fusion represents a primary approach for enhancing environmental perception in intelligent transportation scenes. Among diverse fusion strategies, Bird’s-Eye View (BEV) perspective-based fusion methods have emerged as a prominent research focus owing to advantages such as unified spatial representation. However, current BEV fusion methods still face challenges with insufficient robustness in cross-modal alignment and weak perception of dynamic objects. To address these challenges, this paper proposes a Bidirectional Complementary Cross-Attention Module (BCCA), which achieves deep fusion of image and point cloud features by adaptively learning cross-modal attention weights, thereby significantly improving cross-modal information interaction. Secondly, we propose a Temporal Adaptive Fusion Module (TAFusion). This module effectively incorporates temporal information within the BEV space and enables efficient fusion of multi-modal features across different frames through a two-stage alignment strategy, substantially enhancing the model’s ability to perceive dynamic objects. Based on the above, we integrate these two modules to propose the Dual Temporal and Transversal Attention Network (DTTANet), a novel camera and LiDAR fusion framework. Comprehensive experiments demonstrate that our proposed method achieves improvements of 1.42% in mAP and 1.26% in NDS on the nuScenes dataset compared to baseline networks, effectively advancing the development of 3D object detection technology for intelligent transportation scenes.
Keywords: 3D object detection; multimodal; adaptive cross-attention; temporal aggregation 3D object detection; multimodal; adaptive cross-attention; temporal aggregation

Share and Cite

MDPI and ACS Style

Tian, D.; Wang, J.; Li, J.; Gong, M.; Shi, J.; Huang, Z.; Fu, Z. Bidirectional Complementary Cross-Attention and Temporal Adaptive Fusion for 3D Object Detection in Intelligent Transportation Scenes. Electronics 2026, 15, 83. https://doi.org/10.3390/electronics15010083

AMA Style

Tian D, Wang J, Li J, Gong M, Shi J, Huang Z, Fu Z. Bidirectional Complementary Cross-Attention and Temporal Adaptive Fusion for 3D Object Detection in Intelligent Transportation Scenes. Electronics. 2026; 15(1):83. https://doi.org/10.3390/electronics15010083

Chicago/Turabian Style

Tian, Di, Jiawei Wang, Jiabo Li, Mingming Gong, Jiahang Shi, Zhongyi Huang, and Zhongliang Fu. 2026. "Bidirectional Complementary Cross-Attention and Temporal Adaptive Fusion for 3D Object Detection in Intelligent Transportation Scenes" Electronics 15, no. 1: 83. https://doi.org/10.3390/electronics15010083

APA Style

Tian, D., Wang, J., Li, J., Gong, M., Shi, J., Huang, Z., & Fu, Z. (2026). Bidirectional Complementary Cross-Attention and Temporal Adaptive Fusion for 3D Object Detection in Intelligent Transportation Scenes. Electronics, 15(1), 83. https://doi.org/10.3390/electronics15010083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop