Next Article in Journal
Radar-Based Damage Detection in a Wind Turbine Blade Using Convolutional Neural Networks: A Proof-of-Concept Under Fatigue Loading
Previous Article in Journal
Cross-Layer Security for 5G/6G Network Slices: An SDN, NFV, and AI-Based Hybrid Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BETAV: A Unified BEV-Transformer and Bézier Optimization Framework for Jointly Optimized End-to-End Autonomous Driving

by
Rui Zhao
1,
Ziguo Chen
1,
Yuze Fan
1,
Fei Gao
1,2,* and
Yuzhuo Men
3
1
College of Automotive Engineering, Jilin University, Changchun 130025, China
2
National Key Laboratory of Automotive Chassis Integration and Bionics, Jilin University, Changchun 130025, China
3
School of Mechanical and Electrical Engineering, Suqian University, Suqian 223800, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3336; https://doi.org/10.3390/s25113336
Submission received: 28 April 2025 / Revised: 23 May 2025 / Accepted: 24 May 2025 / Published: 26 May 2025
(This article belongs to the Section Vehicular Sensing)

Abstract

:
End-to-end autonomous driving demands precise perception, robust motion planning, and efficient trajectory generation to navigate complex and dynamic environments. This paper proposes BETAV, a novel framework that addresses the persistent challenges of low 3D perception accuracy and suboptimal trajectory smoothness in autonomous driving systems through unified BEV-Transformer encoding and Bézier-optimized planning. By leveraging Vision Transformers (ViTs), our approach encodes multi-view camera data into a Bird’s Eye View (BEV) representation using a transformer architecture, capturing both spatial and temporal features to enhance scene understanding comprehensively. For motion planning, a Bézier curve-based planning decoder is proposed, offering a compact, continuous, and parameterized trajectory representation that inherently ensures motion smoothness, kinematic feasibility, and computational efficiency. Additionally, this paper introduces a set of constraints tailored to address vehicle kinematics, obstacle avoidance, and directional alignment, further enhancing trajectory accuracy and safety. Experimental evaluations on Nuscences benchmark datasets and simulations demonstrate that our framework achieves state-of-the-art performance in trajectory prediction and planning tasks, exhibiting superior robustness and generalization across diverse and challenging Bench2Drive driving scenarios.

1. Introduction

Autonomous driving technology has recently become a critical area of research in intelligent transportation and future mobility, gradually transitioning from conceptualization to practical application. Over the past few decades, autonomous driving technology has achieved remarkable advancements, leading to the development of a well-established layered architecture comprising multiple modules, including perception, path planning, decision-making, and control [1]. Despite their success, traditional hierarchical control systems face inherent limitations [2]. These systems rely heavily on meticulously designed intermediate interfaces between modules and require detailed modeling of the environment. Consequently, their performance is highly dependent on the accuracy and seamless integration of each submodule. Any inaccuracies or inconsistencies in one module can propagate through the pipeline, potentially compromising the overall system’s reliability and robustness in real-world scenarios [3,4,5,6,7].
End-to-end autonomous driving [8,9,10,11] has recently emerged as a promising direction in the field, directly mapping raw sensor data (e.g., cameras and LiDAR) to decision outputs via deep learning models. This learning-based paradigm minimizes reliance on handcrafted rules and error accumulation across traditional modules while unifying perception, planning, and control. The recent advancements in deep learning and optimization have enhanced its efficiency, decision-making speed, and robustness in real-world autonomous driving. The recent research [3,11,12] in the field of end-to-end autonomous driving primarily focuses on two core modules: (1) feature encoding in the perception module to ensure that the model can accurately and efficiently understand complex driving environments, and (2) trajectory generation and control policy synthesis directly from latent representations, enabling the model to translate scene understanding into executable driving actions.
In terms of perception encoding, common approaches like Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) extract features from sensor data (e.g., camera images and LiDAR) for direct input into decision-making modules. This approach [8,13,14] bypasses intermediate tasks such as object tracking and traffic semantic understanding, directly mapping sensor inputs to control outputs, mimicking the “perception-to-decision” mechanism of human driving. While this simplifies the model architecture, it faces challenges such as high training complexity, dependence on large-scale datasets, and limited interpretability, reducing robustness in complex or unseen scenarios [15]. Another widely adopted approach [8,10,16] incorporates intermediate tasks in scene perception. This method first encodes sensor data into features and decodes them into high-level semantic information (e.g., positions of traffic agents and lane boundaries) for planning and control. Typical tasks include object detection, semantic segmentation, or multi-task combinations, offering advantages like greater interpretability and reduced learning complexity for control modules [17]. Leveraging models such as YOLO [18] or CenterNet [19], this type of method improves system performance. However, they face inherent limitations in reliably estimating precise real-world coordinates and the dynamic states of traffic agents [20]. These challenges stem from the difficulty in inferring accurate depth and motion information from 2D visual data, particularly in scenarios involving occlusions, overlapping objects, or complex interactions in the 3D environment.
In terms of planning modules, the mainstream approaches can be categorized into two types [21,22]: (1) directly outputting low-level control signals for the vehicle, such as throttle, brake, and steering angle, and (2) generating trajectory planning results. From a conceptual perspective, the first approach aligns more closely with the philosophy of end-to-end autonomous driving as it directly maps perception to control [15]. However, this method imposes strict real-time requirements on the model and is often limited by hardware differences across vehicles (e.g., variations in powertrain response and steering ratios), which reduces its adaptability and broad applicability [11]. The second method employs a hierarchical control architecture, where a neural network model serves as the upper layer to generate multi-time-step trajectory planning for future sequences, while the lower layer—typically composed of conventional control algorithms (e.g., PID and MPC) or enhanced approaches such as RMPC and VUF-PID—translates these trajectory plans into specific low-level control signals [12,23,24]. This separation of high-level decision-making from low-level control enhances the model’s generalizability and adaptability. However, a notable limitation of this approach is the sparsity of discrete trajectory points in the planning stage, which can lead to suboptimal trajectory smoothness and introduce challenges in accurately representing the desired path [25]. Additionally, the method heavily relies on the richness and diversity of the training dataset as a lack of variety in the data can result in poor generalization to unseen or complex driving scenarios [26].
To address the critical challenges in autonomous driving systems, including the perception module’s limitations in 3D inference accuracy, training complexity, and limited interpretability, as well as the motion planning module’s issues with real-time adaptability, trajectory smoothness, and data dependency, we propose an innovative framework that unifies advanced perception and planning techniques. For perception, we utilize Vision Transformers (ViTs) to encode visual information into a Bird’s Eye View (BEV) representation, offering interpretable and robust spatial understanding of traffic agents, including precise positions, orientations, and future trajectories. This improves 3D inference accuracy and enhances interpretability, boosting performance in complex scenarios. For planning, we introduce a novel Bézier curve-based trajectory representation, which ensures kinematic feasibility, smooth transitions, and reduced computational demand. This approach mitigates the limitations of the traditional discrete methods, enhances generalization to diverse scenarios, and improves real-time adaptability.
Our contributions can be summarized as follows:
  • Query-Based Interaction Mechanism: We introduce a novel query-based interaction mechanism that enables efficient fusion of multi-modal sensor signals, facilitating the extraction of explicit environmental features and implicit contextual cues for ViTs. This approach enhances perception accuracy while improving the safety and adaptability of autonomous navigation in complex real-world scenarios.
  • Bézier-Parameterized Trajectory Framework: We propose an innovative trajectory representation framework utilizing Bézier curves to parameterize high-dimensional outputs, transforming discrete trajectory points into a low-dimensional and continuous representation. This approach significantly reduces the network’s learning complexity while ensuring trajectory smoothness, kinematic feasibility, and adaptability across diverse driving scenarios.
  • Performance-Validated Evaluation Framework: We conducted tests on the nuScenes dataset to evaluate trajectory planning accuracy and safety while performing generalizability tests in the Carla simulator for system performance verification. The experimental results demonstrate that our method achieves superior reliability and stability in dynamic and complex scenarios compared to the state-of-the-art approaches.

2. Related Works

Perception and Prediction: Accurate perception of driving scenes forms the foundation and safety guarantee of autonomous driving. In end-to-end driving systems, while 2D detection provides fundamental visual cues through cross-task feature sharing, 3D detection establishes hierarchical latent representations for precise environment modeling and obstacle identification, enabling synergistic integration within a unified architecture [27]. Traditional 3D perception methods typically address 3D object detection or map segmentation tasks independently. For 3D object detection, the early methods followed the principles of 2D detection, where 3D bounding boxes were predicted based on 2D bounding boxes. DETR3D [28] introduced a novel approach by utilizing 3D queries to sample corresponding 2D image features, enabling detection without the need for non-maximum suppression (NMS). PETR [29] further advanced this by embedding 3D positional encoding into 2D image features and leveraging detection queries to extract object features through an attention mechanism [30].
More recently, BEV representations have gained significant attention and made substantial contributions to perception [31,32,33,34,35,36,37]. LSS [38] pioneered the use of depth prediction to project perspective features into the BEV space. OFT [39] and ImVoxelNet [40] utilized predefined voxels that are projected onto image features to construct voxel-based scene representations. M2BEV [41] extended BEV representations to simultaneously perform multiple perception tasks, highlighting their versatility. FIERY [32] and BEVerse [37] leveraged BEV feature maps to predict dense map segmentation, while LaneGAP [34] introduced a novel path-based representation for lane modeling, effectively preserving lane continuity and encoding traffic context for planning tasks. In this work, we use a set of BEV queries, agent queries, and map queries to complete scene perception, and further use these query features and perception results for the motion prediction and planning stages.
Planning: Some approaches [42,43] bypass intermediate stages such as perception and motion prediction, directly predicting planning trajectories or control signals. While this end-to-end paradigm is conceptually straightforward, it suffers from limited interpretability and presents challenges in optimization and reliability. To address these limitations, explicit dense cost maps have been widely adopted due to their high interpretability [8,9,44]. These cost maps are constructed from either perception and motion prediction outputs or directly learned through data-driven modules. Handcrafted rules are often employed to select the optimal trajectory with the minimum cost. However, constructing dense cost maps is computationally expensive, and the reliance on handcrafted rules introduces challenges in robustness and generalization.
Recent advancements have explored new directions for planning. UniAD [13], for instance, effectively integrates information from multiple upstream tasks to support goal-oriented planning, achieving state-of-the-art performance in perception, prediction, and planning. PlanT [45] takes the perception ground truth as the input and encodes the scene into object-level representations for trajectory planning. Despite these improvements, challenges such as trajectory smoothness and high computational complexity persist, particularly when relying on intermediate stages like perception and motion prediction. VAD [46] improves the perception and planning performance of autonomous driving by representing driving scenes in vectorized maps and agent actions. To overcome these challenges, alternative methods have been proposed. Zhichao Han et al. [47] utilized the property of differential flatness to simplify the trajectory planning problem, significantly improving computational efficiency and trajectory smoothness. Similarly, Weihuang Chen et al. [48] introduced the feasible and smooth tree (FAST) method, which ensures both trajectory continuity and kinematic feasibility by leveraging a tree-structured planning approach. VADv2 [49] leverages human-like driving policies acquired from extensive large-scale driving demonstrations, producing a probabilistic action distribution that facilitates superior planning and driving performance.
In this work, we adopt Bézier curves to assist trajectory planning, leveraging their compact and parameterized representation. This approach substantially reduces the trajectory prediction dimensionality while inherently ensuring motion smoothness, kinematic feasibility, and computational efficiency.

3. Methods

This work presents an end-to-end autonomous driving architecture (Figure 1) composed of two core components: (1) a perception and prediction module, which extracts high-level semantic information about the driving environment, and (2) a Bézier curve-based trajectory planning module, designed to generate smooth, kinematically feasible trajectories that enable the ego vehicle to navigate safely, legally, and compliantly with traffic regulations. The subsequent sections provide a detailed description of the problem formulation and the design of each module.

3.1. Problem Setting

This work aims to design an autonomous driving system that generates continuous trajectory outputs for steering and acceleration/deceleration control based on perception data obtained from onboard sensors. The sensor inputs consist of six RGB images captured from multiple perspectives and ego-vehicle states, including speed, acceleration, and GPS position. Building upon these multi-modal inputs, our methodology employs imitation learning (IL) as the foundational training paradigm, incorporating a composite loss function L formed through weighted combinations of multiple loss terms specifically designed to address distinct aspects of driving behavior. This integrated framework enables the system to generate kinematically feasible and smooth trajectories for autonomous navigation.
Imitation Learning in Autonomous Driving: IL, also known as learning from demonstrations, is a foundational paradigm in machine learning that allows an agent to learn desired behaviors by mimicking expert demonstrations. In the context of autonomous driving, IL provides a mechanism for vehicles to acquire complex driving behaviors by utilizing data collected from human drivers, thereby eliminating the need for explicitly programming every driving decision. The IL framework typically involves training a policy π , which maps observations s (e.g., raw sensor data or environment states) to actions a (e.g., trajectory). The objective is to minimize the discrepancy between the agent’s predicted actions and those performed by an expert using a supervised approach based on a dataset of state–action pairs ( s , a ) . In this work, the policy π is trained in imitation learning using the collected data D with the loss function L :
arg min π E ( s , a ) D L ( a , π ( s ) )

3.2. Input and Output Parameterization

Input Representation: The six RGB cameras are employed as input sensors, strategically arranged to ensure omnidirectional environmental coverage through their calibrated spatial configuration. The setup includes one forward-facing camera, two forward-facing cameras angled at ±55°, one rear-facing camera, and two rear-facing cameras angled at ±70°. All cameras, except for the rear-facing camera with a horizontal field of view (FOV) of 110°, have an FOV of 70° and a resolution of 1600 × 900 . In addition to visual inputs, the trajectory planning module incorporates sparse navigation information and the ego vehicle’s dynamic state. It is important to note that the navigation information is provided as a discrete set of guidance points, typically spaced hundreds of meters apart, rather than a dense continuous trajectory. This sparse guidance structure reflects realistic navigation data used in autonomous driving systems and poses additional challenges for trajectory planning.
Output Representation: Bézier curves are implemented as parametric descriptors for trajectory prediction of the ego vehicle’s future motion W within the BEV coordinate framework. The trajectory points are sampled from a cubic Bézier curve defined by its control points P 0 , P 1 , P 2 , P 3 . Mathematically, the trajectory W is expressed as
W = { ( x , y ) | ( x , y ) sample B ( t ) , t [ 0 , 1 ] }
here, t is the curve parameter that determines the position along the Bézier curve, ranging from 0 to 1. B(t) represents Bézier curves:
B ( t ) = ( 1 t ) 3 P 0 + 3 ( 1 t ) 2 t P 1 + 3 ( 1 t ) t 2 P 2 + t 3 P 3
where P 0 represents the initial position of the ego vehicle, while P 1 , P 2 , and P 3 are learnable control points that shape the curve. By sampling points along the curve, we generate a smooth and kinematically feasible trajectory for the ego vehicle in the BEV space.

3.3. Perception and Prediction Module

The image data captured by vehicle-mounted cameras are inherently sparse and low-level, posing significant challenges for direct use in planning and decision-making tasks. To address this limitation, a differentiated encoder–decoder architecture is implemented to encode raw image data into BEV-aligned high-level semantic representations.
Specifically, a set of grid-shaped learnable parameters Q t R H × W are instantiated as queries for BEVFormer, where H and W represent the spatial dimensions of the BEV plane. Each query Q p R at position p = ( x b q , y b q ) corresponds to a specific grid cell region in the BEV plane, with each grid cell covering a real-world area of s b e v meters. By default, the BEV feature is centered on the location of the ego vehicle, and the grid resolution H × W is set to 180 × 180 . Additionally, the default query sizes for the map ( Q m ) and agent queries ( Q a ) are 100 × 20 and 500, respectively.
During inference at time step t, multi-camera images ( I t = { I t , 1 , I t , 2 , , I t , 6 } , I t , i R H × W × C ) are processed through a visual feature extraction network (ResNet-50) to extract features F t corresponding to different camera views:
F t = R e s n e t ( I t )
Temporal consistency and scene understanding are enhanced through preservation of preceding BEV features B t 1 during inference, maintaining cross-frame feature continuity in the temporal domain. BEV queries Q t are used to decode both the extracted multi-camera features F t and the historical BEV features B t 1 . This decoding process generates updated BEV queries B t for the current time step, enabling a more comprehensive understanding of the environment. Specifically, the model focuses attention on queries Q s p at different positions in Q t by setting Pos s p through the deformable self-attention mechanism, and then performs attention operations between them and the historical BEV features B t 1 , allowing different parts of Q t to interact with the historical BEV features to capture the relevant temporal implicit dependencies, generating a time-aware BEV representation F b , t e m p . This mechanism ensures the model effectively integrates historical scene information, improving its perception of the current scene. For instance, objects occluded in the current frame but visible in previous frames can be more accurately detected by leveraging this temporal context:
F b , t e m p = Q s p Q t D S A ( Q s p , Pos s p , B t 1 )
where DSA(·) denotes the multi-head deformable self-attention mechanism. Following temporal feature extraction, spatial feature encoding is performed through multi-modal fusion of Q t and multi-camera scene features F t , yielding enhanced spatial–temporal representation F b , s t . This step enhances the model’s spatial understanding of the scene, improving its accuracy in predicting the positions and boundaries of surrounding objects. Notably, by projecting 3D perspectives into a 2D BEV scene through intrinsic and extrinsic camera parameters, we derive the BEV and the mapped positions Pos s t for each camera view. These positions Pos s t , combined with positional encoding from Q t , generate corresponding Q s t queries. This design ensures that each query Q s t Q t interacts exclusively with regions of interest in multi-camera views, thereby achieving focused and efficient feature aggregation. This query-based attention mechanism facilitates precise localization and boundary refinement, particularly in complex and cluttered environments:
F b , s t = Q s t Q t i = 1 6 C A ( Q s t , Pos s t , F t , i )
where CA(·) denotes the multi-head cross-attention mechanism. Subsequent task-specific decoding is performed through dedicated neural decoders, culminating in the derivation of the BEV feature representation B t that encodes the instantaneous environmental topology:
B t = C A ( S A ( Q t ) , M L P ( F b , s t ) )
where SA(·) denotes the multi-head self-attention mechanism. Finally, three lightweight detection heads, map, agent, and traffic, are used to extract high-level semantic information Info from ( B t ) , where each detection head consists of an MLP and has a multi-task output head to extract the corresponding high-level semantic information.
Info = c { map , agent , traffic } MLP c CA ( SA ( B t ) , { F b , s t , Q t } ) ) [ : , S c ]
where S c represents the allocated feature channel interval. This representation ( Info ) encapsulates three distinct categories of critical information: (1) Map Information ( Info m Info ): This category predicts three types of key map points: road boundaries, lane centers, and lane dividing lines. These predictions collectively form a high-level understanding of the road structure, including lane directions, lane boundary coordinates, and drivable regions. This information is crucial for constructing feasible and safe trajectories. (2) Agent Information ( Info a Info ): Agent-related data focus on the motion characteristics of surrounding traffic agents, such as vehicles and pedestrians. This includes their positions, orientations, 3D boundaries, speeds, and future trajectories. By predicting these dynamic attributes, the system enables proactive and adaptive planning in dynamic environments. (3) Traffic Element Information ( Info e Info ): Traffic-related elements are integral to safe and efficient driving decisions. This category encompasses signals and signs, such as traffic lights, speed limit signs, and stop signs. Accurate detection and interpretation of these elements guide the vehicle’s compliance with traffic regulations and ensure safe navigation. By combining these three categories of information, Info provides a comprehensive and high-level representation of the driving environment. This serves as the foundation for reliable trajectory planning and robust decision-making in complex and dynamic scenarios.

3.4. Bézier Curve-Based Trajectory Planning Module

To overcome the limitations of traditional regression methods for discrete trajectory points, we propose a planning decoder based on Bézier curves. Higher-order Bézier curves inherently ensure continuity in position, velocity, and acceleration. In this work, we adopt cubic Bézier curves, represented by Equation (3), where the control points are generated by our planning head.
In the planning head, a learnable parameter Q e g o is initialized to capture scene information essential for trajectory planning. Initially, Q e g o interacts with the positions and motion information of other traffic agents through a transformer decoder to extract implicit interaction features:
Q e g o = D e c o d e r ( Q e g o , Q a , Q a , Pos a )
here, Pos a = M L P ( F I n f o a ) provides the positional embedding for traffic agents, where F I n f o a represents the decoding feature of Info a and the ego vehicle serves as the BEV coordinate origin and requires no additional position embedding. This interaction enables the model to account for dynamic elements in the driving environment.
Subsequently, the updated Q e g o interacts with static traffic elements, such as road infrastructure and traffic signals, to obtain implicit static scene information:
Q e g o = D e c o d e r ( Q e g o , Q m , Q m , Pos m )
where Pos m = M L P ( F I n f o m , F I n f o e ) , F I n f o m and F I n f o e represents the decoding feature of Info m and Info e . This interaction captures a comprehensive understanding of the static environment, enhancing trajectory planning accuracy in complex driving scenarios.
Finally, the combined features Q e g o and Q e g o , along with discrete navigation information Info n and the ego-vehicle state s e g o , are input into an MLP-based planning head to compute the control points P = { P 0 , P 1 , P 2 , P 3 } of the Bézier curve:
P = P l a n H e a d ( Q e g o , Q e g o , Pos n , s e g o )
where ( Pos n = M L P ( Info n ) ) . Following the determination of control points P , a continuous Bézier curve is constructed using the Bézier formula, serving as the planned path.
T b z e = b = 0 n n b ( 1 t ) n b t b P b
For practical implementation and control, this continuous curve is then discretized by sampling at uniform time intervals, yielding a sequence of discrete trajectory points P t r j . These points are utilized by the vehicle’s low-level controller to generate appropriate control commands:
T b z e d = Sample ( B )
While Bézier curves provide a compact and continuous representation for trajectory planning, challenges remain in ensuring adherence to vehicle kinematics, addressing nonlinear constraints imposed by obstacles, and maintaining effective directional control. To address these issues and ensure that the planned trajectory satisfies kinematic feasibility and driving safety requirements, we introduce a set of constraints, as illustrated in Figure 2. These constraints strategically incorporate the following: (1) Minimum Turning Radius Constraint C k , (2) Curvature Continuity Constraint C k max , (3) Collision Avoidance Constraint C col , (4) Road Boundary Adherence Constraint C bod , and (5) Motion Direction Constraint C dir , collectively enforcing the generation of smooth, collision-free, and dynamically feasible trajectories for autonomous navigation systems.
Minimum Turning Radius Constraint: The curvature of the trajectory must not exceed a threshold determined by the vehicle’s minimum turning radius ( R min ) as vehicles have limited steering capabilities. The curvature constraint is defined as
| k | > 1 R min
where k = x y y x ( x 2 + y 2 ) 3 2 represents the curvature at each sampling point T b z e d on the Bézier curve trajectory T bze , where ( x , y ) , respectively, represent the coordinate positions of the sampling point in the vehicle’s own coordinate system.
Curvature Constraint: The shape of a Bézier curve is dictated by the distribution of its control points. To prevent excessive curvature and ensure smooth trajectories, endpoint curvature constraints and control point smoothness constraints are used to suppress curve curvature:
2 ( P 1 P 0 ) × ( P 0 2 P 1 + P 2 ) 3 P 1 P 0 3 + 2 ( P 3 P 2 ) × ( P 3 2 P 2 + P 1 ) 3 P 3 P 2 3 < k e p P 3 2 P 2 + P 1 2 < k m a x
where k e p is the predefined maximum endpoint curvature threshold and k m a x is the predefined maximum control point smoothness threshold. This constraint ensures that the control points are appropriately distributed to avoid sharp turns.
Collision Constraint: To ensure safety and avoid dangerous interactions with other traffic agents, the trajectory B ( t ) must maintain a safe distance from all other agents at any future time step t + i , where i n . The safety constraint is defined as
d col i > δ col
Here, d col i represents the distance between the ego vehicle’s predicted position b ( t + i ) and the closest traffic participant at time t + i , and δ col is a specified safety threshold.
Road Boundary Constraint: To prevent collisions with roadside objects and ensure that the trajectory remains within drivable regions, the distance between the planned trajectory points and road boundaries must not exceed a specified threshold:
d bod i < δ bod , i n
This constraint ensures the trajectory remains within feasible and safe driving areas.
Direction Constraint: The trajectory direction must align with the ego vehicle’s motion direction and comply with road traffic rules. We impose two directional constraints:
(1) At the starting point, the trajectory’s initial direction B ( 0 ) = 3 ( P 1 P 0 ) must align with the ego vehicle’s current direction r ego :
| B ( 0 ) r ego | < δ dir
(2) At the endpoint, the trajectory’s final direction B ( 1 ) = 3 ( P 3 P 2 ) must align with the road’s driving direction r road :
| B ( 1 ) r road | < δ dir

4. End-to-End Learning

The end-to-end framework is trained using supervised learning. The training process involves separate loss functions for the scene perception ( L SP ) module and the planning decoder ( L PD ) module.

4.1. Scene Perception Loss

The scene perception loss focuses on the accuracy of environment perception to ensure the precision of the high-level BEV information Info derived from sensor data. It comprises losses for three categories of predictions: map L map and agents L agent .
Map Loss: The map loss comprises two components: a regression loss and a classification loss. The regression loss evaluates the discrepancy between predicted map points w t and ground truth map points w t g t using an L1 loss. The classification loss assesses the accuracy of map point classifications, where the map categories include ‘Broken’, ‘Solid’, ‘Solid Solid’, and ‘Center’. Traffic elements associated with the map are strictly classified into four categories: ‘traffic sign’, ‘traffic cone’, ‘traffic light’, and ‘others’ using the same loss function. These losses are combined to form the total map loss.
Agent Loss: The agent loss supervises the prediction of traffic participant attributes and trajectories using a combination of regression and classification losses. The regression loss ensures the accuracy of predicted location and boundary coordinates of agents. Additionally, the regression loss for trajectory prediction ensures the consistency between predicted trajectory points and actual motion trajectory points. The classification loss supervises the accuracy of agent category predictions, with agents categorized as ’bicycle’, ’car’, ’van’, ’truck’, and ’pedestrian’. These losses are combined into the total agent loss.
The overall scene perception loss combines these components to ensure a comprehensive understanding of the driving environment:
L SP = L map + L agent

4.2. Planning Decoder Loss

The planning decoder loss ensures the accuracy and safety of ego vehicle trajectory predictions. In our approach, the predictions are based on a relative coordinate system centered on the ego vehicle rather than an absolute map-based coordinate system. The onboard sensors provide feedback on the vehicle’s velocity, acceleration, and GPS position.
The planning decoder loss consists of two parts: (1). Bézier curve constraints L b e z : supervises adherence to trajectory constraints. (2). Trajectory prediction loss ( L traj ): L1 loss between the predicted and ground truth trajectories.
Bézier Curve Constraints: The Bézier curve constraint architecture, analogous to those in Section 3.4, is systematically structured into five components to ensure physical feasibility and safety compliance. The Minimum Turning Radius Constraint ( C R m i n ) ensures the curvature of the trajectory does not exceed the physical minimum turning radius. The curvature constraint ( C c c ) penalizes trajectories with curvatures and smoothness exceeding the predefined maximum allowable threshold. Road boundary adherence is enforced through the boundary constraint ( C bod ). Collision avoidance is guaranteed by the collision constraint ( C col ). Directional consistency is achieved via the composite constraint C dir , which quantifies alignment between the initial trajectory derivative and the ego vehicle’s heading and evaluates terminal directional deviation from the road’s orientation. These constraints all set penalty losses that increase as thresholds are exceeded, and they are jointly optimized to generate kinematically acceptable trajectories.
Trajectory Prediction Loss: The L1 loss function quantifies the deviations between predicted Bézier trajectories and their ground-truth counterparts by computing the L1-norm distance across trajectory coordinates.
The overall planning decoder loss is formulated as
L PD = L b e z + L traj
The proposed loss functions ensure precise perception and safe trajectory planning. L SP guarantees high-level environmental understanding by supervising map, agent, and traffic element predictions. L PD enforces kinematic feasibility and compliance with safety constraints, achieving smooth, reliable, and accurate trajectory generation. This robust loss design underpins the effectiveness of our end-to-end autonomous driving framework.

5. Experiments

This section methodologically presents our experimental framework, visualizes key findings, and provides systematic analysis through three structured components: Implementation Details, Results, and Ablation Study.

5.1. Implementation Details

Comprehensive evaluations were conducted on the nuScenes [50] benchmark dataset, which contains 1000 unique driving scenarios, with each sequence spanning approximately 20 s. The dataset provides 1.4 million meticulously annotated 3D bounding boxes across 23 distinct object categories. The visual data were captured through a synchronized six-camera array (front, front-left, front-right, back, back-left, and back-right), achieving full 360° environmental coverage, with critical frames annotated at 2 Hz temporal resolution. Following established evaluation protocols in prior research, we employed two principal metrics for planning performance assessment: (1) displacement error, measuring the L2 distance between predicted and ground-truth trajectories at specified timestamps; and (2) collision rate, quantifying the percentage of trajectory predictions involving potential vehicle collisions. In addition, to validate the generalization capability of the proposed method, we conducted a set of simplified generalization tests using the curated scenario library from the Bench2Drive [51] benchmark.
The proposed framework processes a 2-s historical observation window to generate 3-second predictive trajectories. The architecture employs a pre-trained ResNet50 backbone for multi-scale feature extraction from six camera inputs. The perception system operates within a 60 m (longitudinal) × 30 m (lateral) operational domain, implementing simultaneous vector mapping and motion prediction through our novel multi-task learning framework. Input images are standardized to 640 × 360 resolution with photometric normalization applied during preprocessing. The complete system required 80 training epochs (168 GPU-hours) on an 8× NVIDIA RTX 3090 (24 GB VRAM per GPU) using distributed data-parallel strategy with synchronized batch normalization.

5.2. Results

Comprehensive benchmarking against outstanding baselines was conducted on the nuScenes dataset, with quantitative comparisons systematically documented in the results in Table 1, and the visualization images are displayed in Figure 3. The compared methods are as follows:
  • NMP [52]: a holistic neural motion planning framework processes raw LiDAR data and HD maps to generate interpretable 3D detections, trajectory predictions, and a multi-modal cost volume encoding navigational preferences. It selects optimal trajectories by sampling physically feasible candidates and minimizing the learned cost function that inherently captures diverse driving scenario dynamics.
  • FF [53]: a self-supervised learning framework for motion planning, bypassing object-centric scene representations by directly modeling occupancy dynamics through automatically generated freespace boundaries. The approach integrates two novel components: (1) collision-prone trajectory identification via freespace violation analysis in planning horizons, and (2) annotation-free supervision through differentiable freespace constraints in neural planner training pipelines.
  • EO [54]: a motion-invariant scene representation for self-supervised trajectory forecasting using differentiable raycasting to project predicted 3D occupancy into future LiDAR sweep predictions for training. By rendering unobserved regions (occlusions) through neural volumetric reasoning, occupancy inherently disentangles ego motion from dynamic environment changes, enabling direct integration with planners to avoid non-drivable regions.
  • ST-P3 [10]: an interpretable vision-based end-to-end framework (ST-P3) that jointly learns spatial–temporal features for perception, prediction, and planning through three core components: egocentric-aligned 3D geometry preservation before perception, dual-pathway motion history modeling for trajectory forecasting, and temporal refinement to enhance vision-centric planning cues. It unifies scene understanding, dynamics inference, and control reasoning within a single coherent pipeline by explicitly addressing geometry distortion, motion context aggregation, and visual feature adaptation across tasks.
  • UniAD [13]: a planning-oriented unified framework that integrates full-stack driving tasks (perception, prediction, and interaction modeling) into a single network, prioritizing feature abstraction and inter-task synergy through shared queries and complementary module designs. It establishes task communication via unified interfaces to jointly optimize scene understanding, behavior reasoning, and trajectory generation in a globally coordinated manner for end-to-end planning.
  • VAD [46]: an end-to-end vectorized autonomous driving framework that models driving scenes with fully vectorized representations, achieving improved safety through instance-level planning constraints and significantly faster computation compared to raster-based methods while also attaining strong baseline planning performance on the nuScenes dataset.
The experimental results demonstrate significant improvements in both trajectory accuracy and safety metrics. Our method achieves superior performance across all time horizons, with an average L2 displacement error of 0.49 m (38.3% lower than VAD-Base’s 0.72 m) and an average collision rate of 0.16% (27.3% reduction from VAD-Base’s 0.22%). Notably, the Bézier-parameterized trajectory framework shows remarkable consistency in long-term predictions, maintaining a minimal error growth from 0.45 m at 1 s to 0.54 m at 3 s, significantly outperforming conventional methods that exhibit error accumulation over time (e.g., FF increases from 0.55 m to 2.54 m). The collision rate progression (0.04% → 0.11% → 0.33%) reveals enhanced safety characteristics compared to the baseline approaches, validating our query-based interaction mechanism’s effectiveness in capturing critical environmental constraints. Particularly noteworthy is the method’s performance in 3 s predictions—the most challenging horizon—where we achieve 51.4% lower L2 error and 19.5% lower collision rate than the prior state-of-the-art approach (VAD-Base). These improvements confirm our framework’s ability to maintain kinematic feasibility through Bézier curve parameterization while ensuring robust perception through explicit feature fusion. The results validate our design objectives of achieving smooth trajectory transitions and improved interpretability in complex driving scenarios.
The generalization capability was evaluated through three composite scenarios constructed from the Bench2Drive benchmark: (1) static hazard scenarios: designed to assess vehicle performance when encountering stationary obstacles or immobile hazards, comprising ParkedObstacle, ConstructionObstacle, ConstructionObstacleTwoWays, Accident, AccidentTwoWays, and HazardAtSideLane; (2) dynamic interaction scenarios: testing trajectory conflict resolution with moving agents, including ParkingCutIn, HighwayCutIn, MergerIntoSlowTraffic, EnterActorFlow, ParkingExit, and VehicleTurningRoutePedestrian; and (3) regulated junction scenarios: evaluating intersection operations under regulatory constraints, with sub-scenarios such as SignalizedJunctionRightTurn, VehicleTurningRoute, VanillaSignalizedTurnEncounterRedLight, NonSignalizedJunctionLeftTurn, NonSignalizedJunctionRightTurn, and VanillaNonSignalizedTurn. The driving performance was evaluated using three key metrics: (1) driving score: calculated as the aggregated mean score across all the routes within each scenario, reflecting comprehensive driving quality; (2) route completion: measured by the averaged percentage of successfully navigated route segments per scenario; and (3) infraction score: initialized at 1.0 and subject to proportional discounting upon traffic rule violations, with the final values representing the normalized compliance levels across all the scenario routes. The experimental results are presented in Table 2, and the visualization is shown in Figure 4.
The generalization tests demonstrate our framework’s scenario-adaptive capabilities. The superior performance in dynamic interaction scenarios (95.46% route completion) validates the effectiveness of Bézier-parameterized trajectories in handling agent motion uncertainties, while the consistent infraction scores (average 0.64) across all the scenarios prove the safety benefits from the query-based multi-modal fusion. Notably, the 99% completion rate in regulated junctions confirms the BEV-Transformer’s interpretable spatial reasoning ability for structured environments. However, the relatively low driving score in static hazards (37.00) suggests that pure vision-based perception still faces challenges in severe occlusion scenarios, which could be mitigated by integrating complementary sensors in future work. These results collectively substantiate our core claim of achieving enhanced perception-planning synergy through unified BEV encoding and continuous trajectory optimization.

5.3. Ablation Study

Ablation studies were conducted to validate the individual contributions of methodological components by sequentially deactivating the following modules: map query, agent query (including its ego-specific submodule), BEV query, and Bézier curve constraints.
Ablation studies conducted on key components of our framework reveal critical insights into their individual contributions to system performance. As shown in Table 3, removing map queries increases the average L2 error by 10.2% (0.54 vs. 0.49) and collision rate by 43.8% (0.23 vs. 0.16), demonstrating the essential role of explicit environmental encoding in perception accuracy. The absence of agent queries degrades performance most severely, with L2 error doubling (1.04 vs. 0.49) and collision rate increasing by 62.5%, confirming the necessity of modeling traffic participant interactions. Notably, eliminating ego-vehicle queries causes disproportionate collision rate degradation (0.32 vs. 0.16), suggesting this component’s critical role in self-awareness for collision avoidance. The BEV query removal experiment shows comparable L2 degradation to map query removal (0.56 vs. 0.54) but yields better collision metrics, implying that BEV features primarily enhance spatial localization rather than interactive safety. The Bézier parameterization ablation increases L2 error by 63.3% (0.80 vs. 0.49) while maintaining collision performance comparable to map query removal, validating its effectiveness in trajectory smoothing and kinematic feasibility. The cumulative degradation when removing all the query mechanisms (L2: 0.99 and collision: 0.32) highlights the synergistic effects of our multi-modal interaction design. Our complete framework achieves optimal balance between trajectory accuracy (0.49 L2) and safety (0.16 collisions), proving the necessity of integrated perception-planning components for robust autonomous navigation.

6. Conclusions

End-to-end autonomous driving demands precise perception, robust motion planning, and efficient trajectory generation to navigate complex and dynamic environments. This paper proposes BETAV, a novel framework that addresses the persistent challenges of low 3D perception accuracy and suboptimal trajectory smoothness in autonomous driving systems. The proposed framework, integrating Vision-Transformer-based BEV perception and Bézier curve-based trajectory planning, demonstrates significant advancements in autonomous driving systems by achieving state-of-the-art trajectory accuracy and collision avoidance performance. The query-based interaction mechanism and compact Bézier representation effectively address the challenges of multi-modal fusion and motion smoothness, while explicit kinematic constraints ensure safety across diverse scenarios. However, the current method exhibits limitations in computational efficiency due to the reliance on transformer architectures for real-time BEV encoding, which may hinder deployment on resource-constrained platforms. Additionally, Bézier parameterization, while simplifying trajectory optimization, could face challenges in highly dynamic environments with abrupt behavioral changes as its continuity assumptions may limit adaptability to extreme maneuvers.
Future work will focus on optimizing the framework’s computational footprint through lightweight transformer variants or hybrid architectures combining CNNs and attention mechanisms. Enhancing the Bézier decoder with adaptive curvature constraints and integrating uncertainty-aware planning modules could further improve robustness in unpredictable scenarios. Extending the framework to incorporate real-world traffic rule reasoning and multi-agent interaction modeling will also be critical for achieving human-like decision-making. Finally, validation in open-world environments with long-tail edge cases and diverse sensor configurations remains essential to bridge the gap between simulation-based performance and practical deployment. Addressing these challenges will advance the development of safer, more generalizable autonomous driving systems.

Author Contributions

Conceptualization, Z.C.; methodology, R.Z.; software, Z.C.; validation, Y.F.; formal analysis, Z.C.; investigation, Y.M.; resources, F.G.; writing—original draft preparation, Z.C.; writing—review and editing, R.Z.; visualization, Z.C.; funding acquisition, R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Science Foundation of China under Grant 52202494 and Grant 52202495, Jilin Provincial Natural Science Foundation General Project under Grant 20250102130JC, the Science and Technology Development Project of Jilin province under Grant 20230301008ZD, and Suqian University Talent Introduction Research Initiation Fund Project (Suqian University No. 2023XRC020).

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

This study does not involve humans.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

On behalf of all the authors, the corresponding author states that there are no conflicts of interest.

References

  1. Parekh, D.; Poddar, N.; Rajpurkar, A.; Chahal, M.; Kumar, N.; Joshi, G.P.; Cho, W. A review on autonomous vehicles: Progress, methods and challenges. Electronics 2022, 11, 2162. [Google Scholar] [CrossRef]
  2. Garikapati, D.; Shetiya, S.S. Autonomous vehicles: Evolution of artificial intelligence and the current industry landscape. Big Data Cogn. Comput. 2024, 8, 42. [Google Scholar] [CrossRef]
  3. Padmaja, B.; Moorthy, C.V.; Venkateswarulu, N.; Bala, M.M. Exploration of issues, challenges and latest developments in autonomous cars. J. Big Data 2023, 10, 61. [Google Scholar] [CrossRef]
  4. Chen, L.; Platinsky, L.; Speichert, S.; Osiński, B.; Scheel, O.; Ye, Y.; Grimmett, H.; Del Pero, L.; Ondruska, P. What data do we need for training an av motion planner? In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May 2021–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1066–1072. [Google Scholar]
  5. González, D.; Pérez, J.; Milanés, V.; Nashashibi, F. A review of motion planning techniques for automated vehicles. IEEE Trans. Intell. Transp. Syst. 2015, 17, 1135–1145. [Google Scholar] [CrossRef]
  6. Kendall, A.; Hawke, J.; Janz, D.; Mazur, P.; Reda, D.; Allen, J.M.; Lam, V.D.; Bewley, A.; Shah, A. Learning to drive in a day. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 8248–8254. [Google Scholar]
  7. Xu, W.; Wang, Q.; Dolan, J.M. Autonomous vehicle motion planning via recurrent spline optimization. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May 2021–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 7730–7736. [Google Scholar]
  8. Casas, S.; Sadat, A.; Urtasun, R. Mp3: A unified model to map, perceive, predict and plan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14403–14412. [Google Scholar]
  9. Cui, A.; Casas, S.; Sadat, A.; Liao, R.; Urtasun, R. Lookout: Diverse multi-future prediction and planning for self-driving. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 16107–16116. [Google Scholar]
  10. Hu, S.; Chen, L.; Wu, P.; Li, H.; Yan, J.; Tao, D. St-p3: End-to-end vision-based autonomous driving via spatial-temporal feature learning. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 533–549. [Google Scholar]
  11. Chib, P.S.; Singh, P. Recent advancements in end-to-end autonomous driving using deep learning: A survey. IEEE Trans. Intell. Veh. 2023, 9, 103–118. [Google Scholar] [CrossRef]
  12. Chen, L.; Wu, P.; Chitta, K.; Jaeger, B.; Geiger, A.; Li, H. End-to-end autonomous driving: Challenges and frontiers. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10164–10183. [Google Scholar] [CrossRef]
  13. Hu, Y.; Yang, J.; Chen, L.; Li, K.; Sima, C.; Zhu, X.; Chai, S.; Du, S.; Lin, T.; Wang, W. Planning-oriented autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 17853–17862. [Google Scholar]
  14. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do vision transformers see like convolutional neural networks? Adv. Neural Inf. Process. Syst. 2021, 34, 12116–12128. [Google Scholar]
  15. Guo, J.; Han, K.; Wu, H.; Tang, Y.; Chen, X.; Wang, Y.; Xu, C. Cmt: Convolutional neural networks meet vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12175–12185. [Google Scholar]
  16. Yang, K.; Zhong, M.; Fan, K.; Tan, J.; Xiao, Z.; Deng, Z. Multi-Task Traffic Scene Perception Algorithm Based on Multi-Scale Prompter. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4965636 (accessed on 20 April 2025).
  17. Lei, Y.; Wang, Z.; Chen, F.; Wang, G.; Wang, P.; Yang, Y. Recent advances in multi-modal 3d scene understanding: A comprehensive survey and evaluation. arXiv 2023, arXiv:2310.15676. [Google Scholar]
  18. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  19. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  20. Backhaus, D.; Engbert, R.; Rothkegel, L.O.; Trukenbrod, H.A. Task-dependence in scene perception: Head unrestrained viewing using mobile eye-tracking. J. Vis. 2020, 20, 3. [Google Scholar] [CrossRef]
  21. Liu, W.; Hua, M.; Deng, Z.; Meng, Z.; Huang, Y.; Hu, C.; Song, S.; Gao, L.; Liu, C.; Shuai, B. A systematic survey of control techniques and applications in connected and automated vehicles. IEEE Internet Things J. 2023, 10, 21892–21916. [Google Scholar] [CrossRef]
  22. Liang, J.; Yang, K.; Tan, C.; Wang, J.; Yin, G. Enhancing High-Speed Cruising Performance of Autonomous Vehicles Through Integrated Deep Reinforcement Learning Framework. IEEE Trans. Intell. Transp. Syst. 2025, 26, 835–848. [Google Scholar] [CrossRef]
  23. Liang, J.; Tian, Q.; Feng, J.; Pi, D.; Yin, G. A Polytopic Model-Based Robust Predictive Control Scheme for Path Tracking of Autonomous Vehicles. IEEE Trans. Intell. Veh. 2024, 9, 3928–3939. [Google Scholar] [CrossRef]
  24. Zhang, H.; Zhao, Z.; Wei, Y.; Liu, Y.; Wu, W. A self-tuning variable universe fuzzy PID control framework with hybrid BAS-PSO-SA optimization for unmanned surface vehicles. J. Mar. Sci. Eng. 2025, 13, 558. [Google Scholar] [CrossRef]
  25. Van, N.D.; Sualeh, M.; Kim, D.; Kim, G.W. A hierarchical control system for autonomous driving towards urban challenges. Appl. Sci. 2020, 10, 3543. [Google Scholar] [CrossRef]
  26. Lan, G.; Hao, Q. End-to-end planning of autonomous driving in industry and academia: 2022-2023. arXiv 2023, arXiv:2401.08658. [Google Scholar]
  27. Ji, H.; Liang, P.; Cheng, E. Enhancing 3D object detection with 2D detection-guided query anchors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 21178–21187. [Google Scholar]
  28. Wang, Y.; Guizilini, V.C.; Zhang, T.; Wang, Y.; Zhao, H.; Solomon, J. Detr3d: 3d object detection from multi-view images via 3D-to-2D queries. In Proceedings of the Conference on Robot Learning, London, UK, 8–11 November 2021; pp. 180–191. [Google Scholar]
  29. Liu, Y.; Wang, T.; Zhang, X.; Sun, J. Petr: Position embedding transformation for multi-view 3d object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 531–548. [Google Scholar]
  30. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  31. Chen, S.; Cheng, T.; Wang, X.; Meng, W.; Zhang, Q.; Liu, W. Efficient and robust 2d-to-bev representation learning via geometry-guided kernel transformer. arXiv 2022, arXiv:2206.04584. [Google Scholar]
  32. Hu, A.; Murez, Z.; Mohan, N.; Dudas, S.; Hawke, J.; Badrinarayanan, V.; Cipolla, R.; Kendall, A. Fiery: Future instance prediction in bird’s-eye view from surround monocular cameras. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15273–15282. [Google Scholar]
  33. Li, Z.; Wang, W.; Li, H.; Xie, E.; Sima, C.; Lu, T.; Yu, Q.; Dai, J. BEVFormer: Learning Bird’s-Eye-View Representation from Multi-Camera Images via Spatiotemporal Transformers. arXiv, 2022; arXiv:2203.17270. [Google Scholar]
  34. Liao, B.; Chen, S.; Jiang, B.; Cheng, T.; Zhang, Q.; Liu, W.; Huang, C.; Wang, X. Lane graph as path: Continuity-preserving path-wise modeling for online lane graph construction. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer: Cham, Switzerland, 2024; pp. 334–351. [Google Scholar]
  35. Liao, B.; Chen, S.; Wang, X.; Cheng, T.; Zhang, Q.; Liu, W.; Huang, C. Maptr: Structured modeling and learning for online vectorized hd map construction. arXiv 2022, arXiv:2208.14437. [Google Scholar]
  36. Liu, Z.; Chen, S.; Guo, X.; Wang, X.; Cheng, T.; Zhu, H.; Zhang, Q.; Liu, W.; Zhang, Y. Vision-based uneven bev representation learning with polar rasterization and surface estimation. In Proceedings of the Conference on Robot Learning, PMLR, Auckland, New Zealand, 14–18 December 2022; pp. 437–446. [Google Scholar]
  37. Zhang, Y.; Zhu, Z.; Zheng, W.; Huang, J.; Huang, G.; Zhou, J.; Lu, J. Beverse: Unified perception and prediction in birds-eye-view for vision-centric autonomous driving. arXiv 2022, arXiv:2205.09743. [Google Scholar]
  38. Philion, J.; Fidler, S. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIV 16. Springer: Cham, Switzerland, 2020; pp. 194–210. [Google Scholar]
  39. Roddick, T.; Kendall, A.; Cipolla, R. Orthographic feature transform for monocular 3d object detection. arXiv 2018, arXiv:1811.08188. [Google Scholar]
  40. Rukhovich, D.; Vorontsova, A.; Konushin, A. Imvoxelnet: Image to voxels projection for monocular and multi-view general-purpose 3d object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2397–2406. [Google Scholar]
  41. Xie, E.; Yu, Z.; Zhou, D.; Philion, J.; Anandkumar, A.; Fidler, S.; Luo, P.; Alvarez, J.M. M2 BEV: Multi-camera joint 3D detection and segmentation with unified birds-eye view representation. arXiv 2022, arXiv:2204.05088. [Google Scholar]
  42. Codevilla, F.; Santana, E.; López, A.M.; Gaidon, A. Exploring the limitations of behavior cloning for autonomous driving. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9329–9338. [Google Scholar]
  43. Prakash, A.; Chitta, K.; Geiger, A. Multi-modal fusion transformer for end-to-end autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7077–7087. [Google Scholar]
  44. Sadat, A.; Casas, S.; Ren, M.; Wu, X.; Dhawan, P.; Urtasun, R. Perceive, predict, and plan: Safe motion planning through interpretable semantic representations. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXIII 16. Springer: Cham, Switzerland, 2020; pp. 414–430. [Google Scholar]
  45. Renz, K.; Chitta, K.; Mercea, O.B.; Koepke, A.; Akata, Z.; Geiger, A. Plant: Explainable planning transformers via object-level representations. arXiv 2022, arXiv:2210.14222. [Google Scholar]
  46. Jiang, B.; Chen, S.; Xu, Q.; Liao, B.; Chen, J.; Zhou, H.; Zhang, Q.; Liu, W.; Huang, C.; Wang, X. Vad: Vectorized scene representation for efficient autonomous driving. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 8340–8350. [Google Scholar]
  47. Han, Z.; Wu, Y.; Li, T.; Zhang, L.; Pei, L.; Xu, L.; Li, C.; Ma, C.; Xu, C.; Shen, S. An efficient spatial-temporal trajectory planner for autonomous vehicles in unstructured environments. IEEE Trans. Intell. Transp. Syst. 2023, 25, 1797–1814. [Google Scholar] [CrossRef]
  48. Chen, W.; Chen, Y.; Wang, S.; Kong, F.; Zhang, X.; Sun, H. Motion planning using feasible and smooth tree for autonomous driving. IEEE Trans. Veh. Technol. 2023, 73, 6270–6282. [Google Scholar] [CrossRef]
  49. Chen, S.; Jiang, B.; Gao, H.; Liao, B.; Xu, Q.; Zhang, Q.; Huang, C.; Liu, W.; Wang, X. Vadv2: End-to-end vectorized autonomous driving via probabilistic planning. arXiv 2024, arXiv:2402.13243. [Google Scholar]
  50. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  51. Jia, X.; Yang, Z.; Li, Q.; Zhang, Z.; Yan, J. Bench2drive: Towards multi-ability benchmarking of closed-loop end-to-end autonomous driving. arXiv 2024, arXiv:2406.03877. [Google Scholar]
  52. Zeng, W.; Luo, W.; Suo, S.; Sadat, A.; Yang, B.; Casas, S.; Urtasun, R. End-to-end interpretable neural motion planner. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8660–8669. [Google Scholar]
  53. Hu, P.; Huang, A.; Dolan, J.; Held, D.; Ramanan, D. Safe local motion planning with self-supervised freespace forecasting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12732–12741. [Google Scholar]
  54. Khurana, T.; Hu, P.; Dave, A.; Ziglar, J.; Held, D.; Ramanan, D. Differentiable raycasting for self-supervised occupancy forecasting. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 353–369. [Google Scholar]
Figure 1. BETAV architecture. Multi-view camera features extracted by ResNet are fused with spatiotemporal embeddings via self-attention (temporal) and cross-attention (spatial) in the transformer encoder. Task-specific decoders then generate traffic agent predictions and Bézier-parameterized trajectories for safe autonomous navigation.
Figure 1. BETAV architecture. Multi-view camera features extracted by ResNet are fused with spatiotemporal embeddings via self-attention (temporal) and cross-attention (spatial) in the transformer encoder. Task-specific decoders then generate traffic agent predictions and Bézier-parameterized trajectories for safe autonomous navigation.
Sensors 25 03336 g001
Figure 2. Illustration of planning constraints. (a) Visualizes the ego agent’s trajectory planning, constrained by road alignment, driving direction, curvature limits (defined by the minimum turning radius), and trajectory smoothness. (b) Depicts the ego agent’s collision avoidance measures, ensuring lateral and longitudinal safety relative to other agents, as well as adherence to road boundary constraints for safe and feasible navigation.
Figure 2. Illustration of planning constraints. (a) Visualizes the ego agent’s trajectory planning, constrained by road alignment, driving direction, curvature limits (defined by the minimum turning radius), and trajectory smoothness. (b) Depicts the ego agent’s collision avoidance measures, ensuring lateral and longitudinal safety relative to other agents, as well as adherence to road boundary constraints for safe and feasible navigation.
Sensors 25 03336 g002
Figure 3. Visualization tested on the nuScenes benchmark.
Figure 3. Visualization tested on the nuScenes benchmark.
Sensors 25 03336 g003
Figure 4. Visualization images of generalization testing.
Figure 4. Visualization images of generalization testing.
Sensors 25 03336 g004
Table 1. Performance comparison on motion planning metrics.
Table 1. Performance comparison on motion planning metrics.
MethodL2 (m)Collision (%)
1 s2 s3 sAvg.1 s2 s3 sAvg.
NMP [52]2.311.92
SA-NMP [52]2.051.59
FF [53]0.551.202.541.430.060.171.070.43
EO [54]0.671.362.781.600.040.090.880.33
ST-P3 [10]1.332.112.902.110.230.621.270.71
UniAD [13]0.480.961.651.030.050.170.710.31
VAD-Tiny [46]0.460.761.120.780.210.350.580.38
VAD-Base [46]0.410.701.050.720.070.170.410.22
Ours0.450.480.540.490.040.110.330.16
Table 2. Generalization testing across multiple scenarios.
Table 2. Generalization testing across multiple scenarios.
ScenariosDriving ScoreRoute CompletionInfraction Score
Static Hazard37.0052.630.67
Dynamic Interaction69.1695.460.72
Regulated Junction53.2899.000.53
Average51.4778.990.64
Table 3. Ablation study of different components.
Table 3. Ablation study of different components.
w/o Q m w/o Q a w/o Q ego w/o Q t w/o C b L2 (Avg.)Collision (Avg.)
----0.540.23
----1.040.26
----0.980.32
----0.560.20
----0.800.26
-0.990.32
-----0.490.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, R.; Chen, Z.; Fan, Y.; Gao, F.; Men, Y. BETAV: A Unified BEV-Transformer and Bézier Optimization Framework for Jointly Optimized End-to-End Autonomous Driving. Sensors 2025, 25, 3336. https://doi.org/10.3390/s25113336

AMA Style

Zhao R, Chen Z, Fan Y, Gao F, Men Y. BETAV: A Unified BEV-Transformer and Bézier Optimization Framework for Jointly Optimized End-to-End Autonomous Driving. Sensors. 2025; 25(11):3336. https://doi.org/10.3390/s25113336

Chicago/Turabian Style

Zhao, Rui, Ziguo Chen, Yuze Fan, Fei Gao, and Yuzhuo Men. 2025. "BETAV: A Unified BEV-Transformer and Bézier Optimization Framework for Jointly Optimized End-to-End Autonomous Driving" Sensors 25, no. 11: 3336. https://doi.org/10.3390/s25113336

APA Style

Zhao, R., Chen, Z., Fan, Y., Gao, F., & Men, Y. (2025). BETAV: A Unified BEV-Transformer and Bézier Optimization Framework for Jointly Optimized End-to-End Autonomous Driving. Sensors, 25(11), 3336. https://doi.org/10.3390/s25113336

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop