Next Article in Journal
Pressure and Temperature Dependence of Optical Resonance Wavelength (ORW) of Large-Element Surface-Micromachined Optical Ultrasound Transducers (SMOUTs)
Previous Article in Journal
Deep Learning-Assisted Porosity Assessment for Additive Manufacturing Components Using Ultrasonic Coda Waves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Atten-LTC-Enhanced MoE Model for Agent Trajectory Prediction in Autonomous Driving

1
School of Automotive and Traffic Engineering, Jiangsu University, Zhenjiang 212013, China
2
Automotive Engineering Research Institute, Jiangsu University, Zhenjiang 212013, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(2), 479; https://doi.org/10.3390/s26020479
Submission received: 7 November 2025 / Revised: 31 December 2025 / Accepted: 9 January 2026 / Published: 11 January 2026
(This article belongs to the Section Vehicular Sensing)

Abstract

The development of sensor technology and deep learning has significantly improved the reliability and practicality of automatic driving technology. In an autonomous driving system, agent trajectory prediction is a complex challenge, which includes the understanding of different and unpredictable behavior patterns of various entities, including vehicles, pedestrians, and other traffic participants, among the data collected by sensors. In this paper, we deeply study two kinds of problems: Single-Agent Trajectory Prediction (SATP) and Multi-Agent Trajectory Prediction (MATP). We propose an innovative model, which combines the attention mechanism and integrates the Liquid Time-Constant (LTC) network with spatio-temporal features and the Mixture of Experts (MoE) framework, termed the Atten-LTC-MoE model. The model is general and extensible to support SATP and MATP problems in different autonomous driving environments. In order to improve computational efficiency and prediction accuracy, lane and agent vectorization, spatio-temporal features, agent data fusion, and trajectory endpoint generation technologies are studied. The effectiveness of our method is verified by comprehensive experiments on Argoverse and Interaction datasets. Our proposed model has been superior to the state-of-the-art models in terms of minADE6 and minFDE6 metrics and has shown significant advantages in the accuracy of agent trajectory prediction and computational performance.

1. Introduction

Accurately predicting the driving trajectory of agents in dynamic environments can make agents much safer, efficient, and reliable in the rapidly developing field of autonomous driving [1]. Using advanced predictive analysis technologies and deep learning algorithms, the existing agent trajectory prediction (ATP) system can accurately predict the future trajectory for surrounding agents [2]. It is essential for the decision-making process in autonomous driving, such as path planning [3], collision avoidance [4], and speed regulation [5], which are highly dependent on trajectory prediction results. Taking autonomous vehicles as an example, ATP systems can predict the behavioral intentions of other road users (such as pedestrians, non-motorized vehicles, and other vehicles), dynamically adjust their own driving trajectory, and ultimately achieve higher safety and traffic efficiency navigation in complex traffic scenarios. In addition, ATP also plays a critical role in the fields of traffic management and urban planning, providing data support and decision-making basis for enhancing the overall efficiency of transportation networks by analyzing traffic flow patterns. In addition, ATP is of vital importance in the field of traffic management and urban planning. Through the analysis of traffic flow patterns, it improves the overall efficiency of the traffic network and provides data support and decision-making basis.
Researchers of trajectory prediction generally divide the research in the ATP field into two categories: Single-Agent Trajectory Prediction (SATP) and Multi-Agent Trajectory Prediction (MATP). SATP is generally utilized to predict the future trajectory of a single agent around the target vehicle. This approach is especially suitable for scenarios with sparse traffic flow or the need to handle isolated intelligent agents separately. However, in dense traffic environments, SATP is difficult to fully capture the complex interactive relationships between intelligent agents, thus having certain limitations.
However, MATP extends the prediction dimension to multi-agent collaborative analysis, fully considering the dynamic interactions among agents in the prediction process, and recognizing that the motion state of a single agent will have a significant impact on the driving trajectories of other agents. In complex traffic environments, the interaction between multiple intelligent agents directly determines the traffic flow status and driving safety. Therefore, MATP technology has become the core support for autonomous driving navigation in this scenario. Developing complex MATP models that can accurately capture the interaction mechanisms of intelligent agents is a major research challenge in the current field and a key breakthrough point for fully unleashing the potential of autonomous driving technology applications. The core of MATP lies in the deep integration of advanced computational models and real-time data analysis techniques, which work together to predict the potential future motion states of surrounding intelligent agents. This predictive capability is not only the core basis for autonomous vehicle maneuver control and path planning, but also a key technical support for enhancing system situational awareness in complex traffic environments. With the rapid development of machine learning algorithms, especially in the fields of deep learning and reinforcement learning, the performance of trajectory prediction models has reached unprecedented levels. Meanwhile, the continuous improvement of high-quality sensor data acquisition capabilities has laid a solid foundation for achieving significant technological breakthroughs in this field.
Autonomous driving systems need to process massive amounts of spatial and temporal data to achieve precise trajectory prediction. In this process, autonomous driving systems not only need to take the influence of environmental factors into account, but also need to fully consider the interaction between intelligent agents. The effectiveness of such applications largely depends on whether the underlying prediction algorithms can efficiently capture two key types of information: temporal dependencies (i.e., the historical motion patterns of agents) and spatial relationships (i.e., the interaction relationships among agents). In the research of trajectory prediction technology, the breakthrough studies of Biktairov et al. [6] were highly representative; they introduced a bird’s-eye view perspective and combined it with a CNN algorithm to construct a trajectory distribution model, and successfully achieved accurate characterization of spatial patterns in specific scenes. IntentNet [7], as an extension of this research direction, innovatively used raw sensor data to invert vehicle driving intentions, which can capture subtle behavioral clues in complex interactive scenarios and provide richer information input for the optimization of decision-making. To further overcome the limitations of single data modality, the multimodal framework [8] effectively improves the robustness and accuracy of trajectory prediction by integrating multidimensional features such as vision and motion. Considering the impact of uncertainty factors on prediction results in a dynamic traffic environment, methods like [9] integrate uncertainty estimation into the motion prediction model and provide a risk avoidance basis for the safe navigation of the auto drive system by outputting the probability distribution range of the prediction results. In addition, graph-based neural networks such as LaneRCNN [10], which rely on distributed spatial reasoning capabilities to model spatial correlations between intelligent agents, have shown outstanding performance in precise trajectory prediction tasks, fully verifying the potential application of graph-based models in the field of motion prediction.
Although many researchers have made breakthroughs in trajectory prediction, they still face the problem of computational efficiency in trajectory prediction. The existing methods based on LSTM and Transformer still rely on a large number of parameters to represent complex dependencies, which will incur great computational costs. Improving the parameter efficiency of the trajectory prediction algorithm is the key to realizing the prediction accuracy of the model and improving the calculation feasibility. It can realize real-time operation while ensuring maximum performance and avoiding unnecessary performance degradation.
In this work, we propose a model with low computational cost, which combines the Liquid Time-Constant (LTC) network [11] with the attention mechanism and the Mixture of Experts with Transformer model to provide a low-complexity framework model to improve computational efficiency and prediction accuracy. The main contributions are as follows:
  • A Spatio-Temporal Attention-enhanced encoder–decoder with Liquid Time-Constant is designed. Aiming at the long-term time dependencies of the agent, the dynamic behavior of the agent is extracted from the historical trajectory information, and the spatial interaction of surrounding agents is involved to further improve the prediction accuracy.
  • Improved computational efficiency. Compared with the existing state-of-the-art model, our model can obtain higher prediction accuracy in a smaller parameter scale. The effectiveness of the Atten-LTC-MoE model is verified by a large number of experiments on the Argoverse dataset [12] and in the Interaction dataset [13], which proves that it is suitable for the real-time deployment of single-agent trajectory prediction and multi-agent trajectory prediction.
  • The proposed method in this paper has better expressive and uncertainty control capabilities in modeling multimodal trajectory distribution and dynamic interaction relationships in complex urban intersection scenes by using vectorized data representation, feature fusion, and prediction methods based on multi-input encoders.
The rest of this paper is as follows: Section 2 reviews relevant research works in the field of agent trajectory prediction; Section 3 elaborates on the architecture and technical details of our proposed model; Section 4 presents the experimental settings and results analysis; Section 5 summarizes and prospects the research content of the entire paper.

2. Related Work

2.1. Agent Trajectory Prediction

Single-agent tracks prediction (SATP) is an essential issue in autonomous driving technology, and its research focus is to predict the future trajectory of a single agent in a complex environment. Although this task may not seem as complex as predicting motion in multi-agent contexts, it also presents its own unique challenges and subtle differences. The core of SATP is to understand and predict the behavior of a single agent, such as a single car, pedestrian, or other traffic participant, in dynamic, unpredictable environments. This prediction is of vital importance for the safety of autonomous vehicles and the trafficability of road traffic, because it directly affects the decision-making processes such as path planning [3], collision avoidance [4], and speed regulation [5]. The complexity of this problem is affected by the historical movement mode, environmental background, and the potential interaction with invisible factors. Researchers strive to improve the accuracy and robustness of these predictions by using advanced machine learning technologies, especially in the field of deep learning and probabilistic modeling.
In the field of single-agent prediction for autonomous driving, some groundbreaking research has made great contributions. Biktairov et al. [6] proposed a method for predicting the motion of a single agent using a bird’s-eye view and a convolutional neural network (CNN, emphasizing the trajectory distribution within unique scenes. As a supplement, IntentNet [7] shifted the focus to directly understand the intention of the agent from the original data of the sensor, and used the deep learning technology to explore the interpretability of complex behavior information. In order to further explore the importance of considering multiple data models, the paper [8] demonstrated the method of integrating different data forms to improve the trajectory prediction accuracy of the monomer. As for the key aspect of uncertainty, the paper [9] proposed a method of motion prediction using uncertainty estimation, which had a much better prediction effect for unpredictable traffic scenes. LaneRCNN [10] provided a graph-centered motion prediction method, which uses distributed representation to predict the trajectory of a single agent. This method shows that graph-based neural networks have great prediction potential for spatial motion reasoning. Based on the above research, many single-agent trajectory prediction technologies have broken through different challenges, and also helped to develop more complex and effective single-agent trajectory prediction models.
On the other hand, Multi-Agent Trajectory Prediction (MATP) emerges as a pivotal and inherently challenging task, as real-world traffic scenarios are characterized by intricate interactions among multiple road agents. The specific prediction method of MATP is shown in Figure 1. Accurately anticipating the future trajectories of all surrounding agents is indispensable for autonomous vehicles (AVs) to make safe, efficient, and context-aware decisions, thereby ensuring smooth navigation in complex traffic ecologies while mitigating collision risks. The success of MATP systems hinges on their capability to capture not only the individual motion patterns of each agent but also the dynamic spatio-temporal dependencies and interactive behaviors that arise from the mutual influences among agents. Recently, a wealth of groundbreaking research has advanced the state-of-the-art in MATP, each addressing distinct facets of the multi-agent interaction problem and collectively shaping a more robust foundation for autonomous navigation. One notable line of work in MATP leverages graph neural networks (GNNs) to explicitly model the interactions among agents. Paper [14] contributed to Multi-Agent Trajectory Prediction (MATP) by proposing a multi-modal Graph Attention Isomorphism Network (GAIN)-based framework, which effectively understood and aggregated long-term interactions across agents while taking model complexity and computational efficiency into account to adapt to real-time application scenarios. Boris Ivanovic et al. [15] proposed Trajectron model that integrated a conditional variational auto encoder (CVAE), LSTM, and a dynamic spatio-temporal graph framework, which can explicitly address the multi-modality, dynamics, and variability of agent behavior, enabling simultaneous prediction of the distribution of future trajectories of multiple agents.
These studies above emphasize the multifaceted nature of MATP in autonomous driving, with each study addressing the unique challenge of capturing the complexity of multi-agent traffic. As autonomous driving technology continues to evolve towards large-scale deployment, there are still some challenges to the insights and methods of these works. The deployment of large models leads to a huge amount of computation within the models, resulting in low computational efficiency. In addition, in dynamic environments, achieving accurate trajectory prediction still faces significant challenges because of complex spatio-temporal correlations between agents. Existing models are often limited by high computational costs and large parameter scales, so it is difficult to deploy them in complex traffic scenes in real time. Therefore, a more lightweight design is urgently needed to meet the real-time requirements of the actual scene. In order to solve the problems above, it is extremely necessary to develop innovative algorithms that can ensure prediction accuracy and have efficient computing capabilities.

2.2. Model-Based LSTM and Attention Mechanism in Trajectory Prediction

In the task of autonomous driving, when decisions need to be made, especially when the trajectory of the agent is predicted, spatio-temporal modeling is of great importance, because the agent is affected by the long-term past behavior of surrounding vehicles in time and space in complex traffic scenes. Accurately extracting these temporal and spatial dependencies from historical information can make the model understand the temporal evolution and spatial interaction of vehicle trajectories in complex traffic scenes.
Long Short-Term Memory (LSTM) networks [16] has been widely used in trajectory prediction to simulate long-term dependencies. The model based on LSTM [17,18,19] can extract the continuous motion in the agent trajectory using the encoder–decoder framework. For example, the LG-LSTM framework [20] combined LSTM temporal modeling, GCN local interaction, and attention global fusion to achieve local–global dual-layer interactive modeling, while retaining the temporal advantages of LSTM and the interactive advantages of graph models, significantly improving the prediction accuracy of specific operation trajectories. Similarly, paper [21] proposed the Scene-LSTM model, which extracted scene information through a two-level structure of “scene grid–subgrid”, jointly trained dual LSTM (agent motion LSTM and grid Scene LSTM), filtered effective nonlinear motion information through scene data filters (SDF), and took position offset as input to improve the prediction performance of nonlinear intelligent agent trajectories in static crowded scenes.
Attention mechanism is a powerful approach that focuses on temporal or spatial historical information features to extract spatio-temporal information. Combining these mechanisms can further improve the prediction accuracy of trajectory prediction. The spatio-temporal attention LSTM model [22] skillfully uses spatio-temporal attention to extract the dynamic relationship and historical motion state features between the target vehicle and the surrounding vehicles, so that the model can prioritize the key interactive relationships, thus making the model more interpretable. In addition, in order to enhance the prediction of static environment constrained scenes, the paper [23] adopts dynamic and static context-aware networks to identify the weight of surrounding dynamic vehicles combined with an attention mechanism. It can be seen that the attention-based method can narrow the gap between a single vehicle trajectory and its dependence on surrounding vehicles.

2.3. Liquid Time-Constant Networks

The core goal of the trajectory prediction decision scenario is to design an algorithm that can not only reflect the generalization ability by learning the coherent representation of its environment but also provide an interpretable explanation of its dynamic changes. It is found that a single algorithm with only 19 control neurons and 32 encapsulated input features connected to the output through 253 synapses can learn to map high-dimensional inputs to steering instructions [24]. Compared with the large-scale black box learning system, the system shows better generalization, interpretability, and robustness. The neural agent can provide autonomous control ability for the trajectory prediction task.
Surprisingly, even tiny organisms such as Caenorhabditis elegans [25] have mastered the abilities of trajectory prediction [26], motion control [27] and navigation [28] by virtue of their near-optimal neural system structure and coordinated neural information processing mechanism. In complex real-world scenes such as autonomous driving, this neural-inspired computing approach is expected to yield AI with more expressive ability. Its model can not only ensure accuracy but also provide interpretability. The structure, connectivity patterns, and information processing mechanisms of the nervous system of Caenorhabditis elegans directly guide the design of LTC [24].
The adaptive time constant τ of LTC can dynamically adjust the response speed of the hidden state: τ increases in the constant speed scene, and the long-term motion law is stably retained; τ decreases in the sudden change scenario, and track changes are quickly captured. LTC spreads the gradient through the analytical solution of a linear ordinary differential equation (ODE), without the gradient disappearance problem of traditional RNN, balancing the long-time-dependent modeling and gradient stability.
In this work, in order to build a more accurate and computationally efficient agent trajectory prediction model, this study will combine the attention mechanism and the advanced Liquid Time-Constant (LTC) network [11] to extract and capture the past long-term spatio-temporal historical information. LTC is an innovative type of time continuous recurrent neural network, which is utilized to replace the traditional recurrent network, LSTM, which depends on static parameters to control dynamic characteristics, and is used to improve the modeling ability of complex historical spatio-temporal data and make the predicted trajectory more accurate.

2.4. Mixture of Expert Methods

Mixture of experts (MoE) plays a wide role in machine learning, as it can improve the performance and scalability of models. A new sparse MoE architecture [29] has been proposed, which strategically allocates computing resources across different “expert” networks based on input data, providing a new breakthrough for scaled visual models. On the basis of improving parameter efficiency, this method enhances the significant characteristics of large-scale image recognition tasks. The sparse gating mechanism of this method dynamically selects expert subsets for each input, providing a scalable alternative to dense architectures and offering a new understanding of efficient resource allocation in deep learning models. Furthermore, Zhou Y et al. [30] proposed an expert selection routing mechanism to optimize the expert selection process during the learning process. This technology enhances traditional MoE frameworks by allowing for more complex routing algorithms based on input data to select the most relevant experts. Their method not only improves predictive performance but also has a certain interpretability of model decisions.

3. Proposed Method

To improve the accuracy of agent trajectory prediction, optimize computational efficiency, and reduce the complexity of the model, this paper constructs the Atten-LTC-enhanced Mixture of Experts (MoE) model (Atten-LTC-MoE) to accurately predict the future trajectory of all agents in complex traffic scenarios. The model description is divided into the following aspects.

3.1. Overall Framework

The overall framework of the proposed structure is presented in Figure 2, where the entire processing pipeline consists of three essential modules. First, the input data includes the agents’ historical trajectories extracted from high-precision (HD) maps in specific traffic scenarios and the geometric information of surrounding lanes. The input geometric data is usually captured in real-time by sensors installed on the vehicle, such as LiDAR. The feature fusion module first extracts the spatio-temporal features of the agent, as well as key features of motion and lane structure, and then integrates all hidden information into a unified feature representation. The trajectory prediction module based on Atten-LTC-enhanced MoE predicts the key endpoints of the vehicle’s future path based on these fusion features. Finally, the trajectory generation module generates smooth, continuous trajectory curves that conform to physical constraints based on the predicted endpoints. In our proposed model, the LTC unit has stronger temporal and spatial expression ability, which can significantly improve the performance of time series prediction tasks. This model can not only effectively model short-term and long-term dynamic relationships but also fully consider spatial correlation among agents. This innovative fusion enables the model to generate accurate trajectory predictions in dynamic multi-agent traffic scenarios.

3.2. Lane and Agent Vectorization Encoding

In the dynamic and complex traffic environment, the data processing efficiency of trajectory prediction is of high importance. The traditional raster representation of maps and agents will produce high-dimensional data, which will lead to the time delay of endpoint processing of the predicted trajectory. In order to improve the efficiency of processing geometric data, we use a vectorized representation [31] to encode geometric data, including time series, lanes, and vehicle tracks. The advantage of vectorization is that it only focuses on the relevant features, which further reduces the computational complexity and data redundancy.
Specifically, we represent the lane information as a set given by M = m i . Here, the i-th lane information is composed of consecutive l i polyline points are encoded into a vector expressed as m i R l i × 2 , where l i = v 1 , v 2 , , v n , each v i = d i s , d i e , a i , j is a vector encoding attributes such as position, heading, or speed, where d i s and d i e denote the coordinates of the starting points and end points of the vector, and a i corresponds to attributes, such as object type, time stamp of track, or road feature type, or speed limit of lane. For each polyline, this work first applies a local GNN [31] to encode the interaction between vectors within the polyline. The polyline is regarded as a fully connected graph, where each node corresponds to a vector v k . For each polyline, the local GNN is represented by the message passing update vector:
h k l + 1 = M L P h k l + j N k h j l
where h k l is the feature of vector k in the l layer. After the sum operation aggregates the information from adjacent vectors N k to obtain the local polyline coding and further uses the global interaction graph to model the interaction between the agents and the map elements. The nodes in the global graph correspond to the encoded polyline features, and the edges between nodes capture the interaction patterns of spatial and temporal dimensions. Global GNN captures the complex relationship between different polylines through the following formula:
h i l + 1 = M L P h i l + j i h j l · e i j
where e i j represents the learning interaction weight between the polyline l i and l j , and h i l is the feature vector of polyline l in the i-th layer. As shown in Figure 2, we construct two independent graph networks—the lane graph and the agent graph—which are, respectively, used for encoding lane and vehicle features. Through these two dedicated graph networks, vehicle trajectory and lane information are efficiently encoded into a vectorization format. The agent feature encoding involves converting the agent trajectories into a vectorized format. The history trajectories of an agent i are expressed as A = a i , where a i R T × 2 denotes the 2D history location information of the agent i for the previous T time steps in the HD map.
In this work, the lanes in the map are modeled in a vectorization format. The encoding process first converts the input time series or images into tokens into a vector form called embeddings. Consider an input sequence with a length of n, expressed as x 1 , x 2 , , x n , where each x i is a vector in R d m o d e l . These embeddings are designed to encapsulate the essence of each token and form the basis of the model’s computational processes. In addition, we also use a vectorized graph structure, which is suitable for modeling relationship and structure information. We use the polyline subgraph to encode each scene element.

3.3. Feature Fusion for Lane and Agent

Meanwhile, the other critical part is to find an effective approach to model the hidden interaction between different elements in the given input data. The first step in preparing the transaction data for the model involves embedding each transaction t i into a higher-dimensional space, resulting in a sequence of embedded vectors E = e 1 , e 2 , , e n , where each e i R d m o d e l . This embedding process captures the essential features of each transaction, transforming them into a form that is amenable to the subsequent encoding steps in the Transformer architecture. In our proposed method, we adapt a linear layer to encode the input credit card transaction records. This learnable linear layer is crucial for transforming the raw transaction data into a format that is more comprehensible for the model.
Transformer architecture includes two main components: encoder and decoder [32], as shown in Figure 3. In order to detect and classify the inputs in this study, the focus is still on the encoder. The encoder operates by obtaining a series of input tags and converting them into a compressed low-dimensional form. Subsequent models use this transformation representation for prediction or classification. A key aspect of the transformer encoder is the integration of the self-attention mechanism. This mechanism helps to capture and process long-distance relationships in input data, so as to generate a more accurate representation of complex inputs.
In the task of marking sequence, the Transformer uses a self-attention mechanism, but lacks the internal method to identify the sequence. Therefore, the position encodings, which refer to Equation (3), are integrated into the input embedding to provide additional information about the position of each token in the sequence:
P E ( p o s , 2 i ) = sin p o s 10,000 2 i / d m o d e l P E ( p o s , 2 i + 1 ) = cos p o s 10,000 2 i / d m o d e l ,
Following the operation above, the Transformer encoder applies the self-attention mechanism to these embeddings. This mechanism allows the model to identify and measure the long-term dependencies in the input text by evaluating the relevance of each embedding relative to other embeddings. After the self-attention phase, the encoder uses one or more feed-forward layers on the result representation.
In this context, the input token comprises queries ( Q ), keys ( K ), and values ( V ), each with the dimensionality d m o d e l . These components are derived by multiplying the input by three learnable matrices W q , W k , W v :
A t t e n t i o n Q , K , V = S o f t m a x Q K T d k V
Q = W q · x ,   K = W k · x , V = W v · x
Here, d k is the hidden dimension, typically equivalent to d m o d e l . This work utilizes scaled dot-product attention.
Multi-head attention mechanism: This mechanism divides the input embeddings into multiple ‘heads’, applying self-attention independently to each. It allows the model to discern diverse dependencies in the input tokens by weighing the relevance of embeddings within each head. The distinctions between scaled dot-product attention and multi-head attention are showcased in Figure 4, and the output of multi-head attention is formulated as follows:
M u l t i H e a d Q , K , V = C o n c a t h e a d 1 , h e a d 2 , , h e a d h W O w h e r e   h e a d i = A t t e n t i o n W i Q x , W i K x , W i V x ,
The projection matrices are denoted as follows:
W i Q R d m o d e l × d q , W i K R d m o d e l × d k , W i V R d m o d e l × d v , W O R d m o d e l × h d v
where h d v = d k and h is typically set to 12.
We do not use sinusoidal position coding [32] or absolute position embedding to inject position information. Instead, we add one-dimensional relative position deviation to each attention matrix to consider the relative positions between features in the window. Self-attention in each head with relative bias is calculated as follows:
A t t e n t i o n Q , K , V = S o f t m a x Q K T d k + B V
where B = b i , j R M × M is a relative position bias, and an element b i , j = b ^ j 1 is brought from a learnable bias table B ^ = b ^ n M + 1 n M 1 .

3.4. Trajectory Prediction and Generation

Our proposed trajectory prediction module consists of two parts: the endpoint prediction module based on Atten-LTC-MoE, and the trajectory generation module based on MLP. Specific methods are introduced in detail in the following sections.

3.4.1. Atten-LTC-MoE Based Endpoint Prediction

In fact, the output features of agents and lanes in MHA only contain static spatio-temporal relationships, and do not fully capture the dynamic time sequence dependence of agent movement, such as agent acceleration and deceleration trend, long-term following behavior, while dynamic spatial interactions, such as the impact of adjacent agents when changing lanes, are also not contained. On these bases, to enhance the capture of historical information of long-term dependence, this study innovatively utilizes the Atten-LTC unit to establish continuous time and space dynamic features, while making the original highly complex model much more sparse.
Figure 5 introduces the flow of the spatio-temporal attention-enhanced LTC encoder–decoder module in our proposed model. Firstly, in the previous data preprocessing and vectorization module, the model receives the vectorized historical trajectories of the agent in the past t time steps from the 3 × 13 grids. These features are connected in vector form after normalization. In this paper, these characteristics of the i-th agent at time t are presented as follows:
p t i = x t i ,   y t i ,   u t i ,   v t i
where x t i and y t i are x coordinate and y coordinate, and u t i defines the agents’ acceleration, while v t i defines the lane change for the i-th agent at time t. After transforming the motion trajectory of agents into a vectorized data representation, the raw input data is mapped to a high-dimensional space using a multi-layer perceptron (MLP). This potential spatial representation can capture more complex agent behaviors and spatio-temporal dependencies. This work uses an MLP represented by the function R e L U ( · ) to perform an embedded transformation, which is shown below:
r t i = R e L U W r x t i + b e
where W r R d m o d e l ×   d L T C represents the weight matrix for the fully connected layer. b e R d m o d e l represents the bias vector. d L T C is the size of the input vectors ( d L T C = 4 for x t i , y t i , u t i , v t i ), r t i is mapped from d m o d e l to the hidden dimension of the LTC network d L T C through the linear layer, which is for the i-th agent, representing its potential embedding at the time step t.
The second step, as shown in Figure 5, then transmits the embedding generated by the MLP layer to the LTC unit, which is to form the time dependency of agents and generate the hidden states h t T + 1 , h t T + 2 , , h t and h t R d m o d e l for the historical trajectory of the agents.
LTC model [11] proposed a network hidden state flow composed of linear ordinary differential equations based on the evolution process of CT-RNN [33] hidden state. The specific equation is as follows:
d x t d t = 1 τ + f x t , I t , t , θ x t + f x t , I t , t , θ A L T C
where x t represents a hidden state, I t represents the input, t is the specific time, and the function f · is controlled by a parameter θ . The equation introduces a stable time constant τ that can make the hidden state approach the equilibrium state of the autonomous system. Formula (11) can dynamically adjust the response of the LTC network to historical information according to the time features, so that the LTC model has superior performance in dealing with time-varying dependence.
In terms of specific deployment, the discrete-time approximation method is utilized to achieve a more efficient solution [19]. When the fused ordinary differential equation needs to be solved, the state can be expressed as the following equation:
x t + t = x t + t f x t , I t , t , θ A L T C t 1 τ + f x t , I t , t , θ + 1
As shown in Figure 6, the internal structure of each LTC unit is presented. The input neuron first processes the input time series signal, and then inputs the processed time series signal into the liquid layer. Neurons in the liquid layer are interconnected by dynamic pathways, as shown by the arrow. The LTC network has neurons with adaptive time constants, which control the response speed of each neuron to the change in input information. Finally, the output neuron summarizes information after processing.
After generating the hidden states h t T + 1 , h t T + 2 , , h t of the historical trajectories by LTC unit, the temporal attention module will optimize these hidden states, as shown in Figure 5. The module is composed of fully connected layers (FC), the tanh function, and the softmax function, which can assign attention weight to key time steps. The attention mechanism can give priority to the key time steps by calculating the impact of the historical trajectory on the future motion of the target agent or surrounding agents. Given the hidden state H t k = h t T + 1 k , h t T + 2 k , , h t k of agent k , the time attention weight A t k can be calculated by the following equation:
A t k = S o f t m a x W t 2 t a n h W t 1 H t k
where W t i is the learnable weight, and S o f t m a x · is defined as follows:
S o f t m a x z i = e z i j = 1 n e z i
where z i is the i-th input for the S o f t m a x · function.
In the following step, encoded contexts with temporal information will be output with spatial information for further decoding:
G t k = A t k H t k
As we can see in Figure 5, the Atten-LTC decoder module proposed in this study combines the spatial information about the agents with the encoded temporal context through the spatial attention mechanism. This module comprises two key components: the spatial attention module and the LTC decoder. A s k represents the spatial attention weights and is computed as follows:
A s k = S o f t m a x W s 2 t a n h W s 1 G t k
where W s i is the learnable weight that controls each agent for the prediction of future trajectories. And the spatio-temporal representation J t k is computed as the following equation:
J t k = A s k G t k
The generated J t k is then transferred to LTC units, which are endpoints of the rough future trajectories x t + 1 , y t + 1 , x t + 2 , y t + 2 , . This process, through step-by-step decoding, makes sure the seamless integration of spatial and temporal dependencies and generates rough predicted trajectories in a complex multi-agent environment.
To obtain much more accurate predicted trajectories, in this work, we introduce a Mixture of Experts (MoE) [29,30] module to adapt and select the previous rough trajectories. From the previous work [29,30], MoE can more effectively deal with complex, high-dimensional, and nonlinear problems, and obtain excellent predictions by using the analysis and selection of various experts’ work. The basic idea of MoE is to divide the complex problem space into multiple homogeneous regions based on multiple expert networks, so that the system has better adaptability and reflects the universality and robustness.
As shown in Figure 7, the Mixture of Experts (MoE) method in this study employs multiple generic Feedforward Neural Networks (FNNs) as experts and uses linear softmax as the gating network. Similarly to the previously proposed hard MoE, this method achieves sparsity by computing a weighted sum of the outputs from only the t o p k experts, rather than a weighted sum of the outputs from all experts. Specifically, within one MoE layer, there exist FNNs, f 1 , f 2 , , f n , and a gating network g · . The gating network is defined as follows:
g i x = S o f t m a x t o p k W · x + n o i s e
where t o p k is a t o p k function that only remains the k largest elements in the given vectors. We also add random noise to avoid significant load imbalances. The S o f t m a x · function is computed as follows:
S o f t m a x h i x = e x p h i x j = 1 k e x p h j x
where h i x is the t o p k output of the gating network corresponding to the experts.
The final trajectories output y ^ of the MoE model is a weighted sum of expert predictions, where the gating network assigns probabilities g i x to each expert:
y ^ = g i x f i ( x )
This modular design significantly improves prediction accuracy and system robustness, especially when dealing with complex and uncertain driving scenarios. In addition, the MoE architecture has the ability to handle heterogeneous data and dynamically adapt to environmental changes, making it particularly suitable for autonomous driving application scenarios. During the training process, the overall loss function of the MoE model combines the prediction results of all experts and is assigned weights by a gating network. The loss function is expressed as the following equation:
L = 1 T t = 1 T y ^ t y t 2
where T represents time steps’ number, y ^ t denotes the endpoint position at time step t , while y t is the actual endpoint position of the time step t . The model is trained by stochastic gradient descent, while optimizing the parameters of the gating network and various expert models.

3.4.2. Trajectory Generation

The endpoints of the trajectory prediction are essentially a set of discrete coordinate data. In order to generate continuous trajectories that are more consistent with physical constraints, we refer to the method [34] and use the predicted endpoints and their corresponding eigenvectors to convert them into smooth and continuous trajectories. Specifically, the method first uses an MLP layer to predict the position offset of the initial endpoint. Then, an additional MLP layer is used to receive the offset endpoint and its related features and output the final smooth trajectory prediction. At the same time, MLP also generates the corresponding confidence probability distribution for each smooth trajectory.

4. Experiments

In this section, we will test the evaluation criterion based on different datasets. The following content analyzes the indicators used and the experimental results.

4.1. Experimental Setup

4.1.1. Dataset Specifications

In this work, our evaluation was conducted on the two most popular datasets: the Argoverse dataset [12] for a single-agent prediction task and the Interaction dataset [13] for a multi-agent prediction task.
First of all, we use the Argoverse dataset [12], which has a numerous and diverse dataset, including 323,557 individual scenes, and is a well-known benchmark in the field of single-agent trajectories prediction. In addition, the Interaction dataset [13] is used as the evaluation dataset for multi-agent trajectory prediction. The Interaction dataset contains 62,022 scenarios, and each scenario contains up to 40 agents. It has the future movement of all agents in a given scene and can provide a more complex and diverse multi-agent prediction environment. The above two datasets are randomly scrambled and uniformly divided into the training set, verification set, and test set according to the ratio of 7:1:2.

4.1.2. Evaluation Metrics

Similarly to previous works [34,35], our work utilizes several standard metrics for evaluation, including the minimum Average Displacement Error ( m A D E k ), minimum Final Displacement Error ( m F D E k ), Miss Rate ( M R k ), and Brier minimum Final Displacement Error ( b r i e r m F D E k ). These error metrics are calculated based on the trajectory with the closest endpoint to the ground truth over k trajectory predictions. These metrics measure the prediction quality from different angles. The m A D E k metric measures the average L2 distance between the entire prediction trajectory and the corresponding ground-truth results:
m A D E k = min i 1 , , k 1 T t = 1 T y ^ i , t y t 2
The m F D E k metric measures the distance between the predicted endpoints and the ground-truth endpoints:
m F D E k = min i 1 , , k y ^ i , T y T 2
M R k is the ratio of scenes where m F D E k is larger than 2 m. The b r i e r m F D E k metric is calculated as follows:
b r i e r m F D E k = 1 p 2 + m F D E k
where p is the probability predicted for the trajectory.

4.1.3. Model Configurations and Training

In this study, a network structure with L = 3 layers of depth was adopted for polyline subgraphs and interactive modeling. The model training process consists of 36 complete epochs, with a batch size set to 256 to balance computational efficiency and memory usage. The Atten-LTC-MoE-based trajectory prediction module is configured with a total of eight expert networks and dynamically selects the t o p k = 2 most relevant expert combination prediction results during the inference process. For optimization, we employed the Adam optimizer and set different initial learning rates for different datasets: the Argoverse dataset uses a learning rate of 1 × 10−4, while the Interaction dataset uses a learning rate of 2 × 10−4 to adapt to more complex multi-vehicle interactions. All experiments were conducted on RTX 4090 GPU servers equipped with 64 GB of memory, ensuring efficient execution of large-scale training.
Referring to the method in this work [34], our proposed method reduced the learning rate to 0.15 times the original value at the 70th and 90th nodes of the training process to promote model convergence to a better solution. In terms of data preprocessing, we generated detailed lane vector representations for roads within 50 m of any agent. To enhance the generalization ability of the model, various data augmentation techniques were implemented, including random scaling within the range of [0.75, 1.25], as well as applying a 0.15 probability random agent dropout strategy for Argoverse and Interaction experiments, simulating sensor occlusion and detection uncertainty. For the reference coordinate system setting, this paper performs scene translation and rotation processing relative to the target agent in the reference frame centered on the agent; Conversely, in the scene-centric frames orientation, the coordinate system direction was determined based on the average position of all relevant agents, providing a more global perspective.

4.2. Ablation Study

In order to quantify the independent contributions of the attention mechanism, LTC unit, and MoE in the Atten-LTC-MoE framework, three sets of model variants were designed and compared in terms of their predictive performance on the Argoverse dataset. The results are shown in Figure 8.
As shown in Figure 8, the m A D E 6 of Atten-LSTM-MoE reached 0.73, significantly higher than other variants; the m A D E 6 of LTC-MoE (0.68) and Atten-LTC (0.63) decrease sequentially, while the m A D E 6 of our model is only 0.61. This result indicates that replacing LTC with LSTM significantly increases the displacement error of short-term trajectories, verifying that LTC’s adaptive time constant is more suitable for dynamic motion patterns in traffic scenarios. The LTC-MoE error of removing spatial attention is higher than Atten-LTC, indicating that spatial attention can effectively integrate the heterogeneous features of agents and lanes, and improve short-term prediction accuracy.
It can be seen that the m F D E 6 of Atten-LSTM-MoE is 1.25, far higher than other models, while the long-term errors of LTC-MoE (1.14) and Atten-LTC (1.05) gradually decrease, and the m F D E 6 of our model is only 1.01. This trend further highlights the advantages of LTC: the static gating mechanism of traditional LSTM is difficult to capture the dependency relationship of long-term trajectories, while LTC avoids the problem of vanishing gradients in long sequences through the analytical gradient propagation of linear ODE, greatly reducing the long-term endpoint error. At the same time, the lack of attention weakens the modeling ability of multi-agent interaction, resulting in long-term errors of LTC-MoE being higher than Atten LTC.
The M R 6 of Atten-LSTM-MoE (0.22) is the highest among all methods, while the failure rates of LTC-MoE (0.14) and Atten-LTC (0.12) decrease sequentially. The M R 6 of our model is only 0.11. The change in failure rate is consistent with the error index; the LTC units can significantly improve the stability of trajectory prediction, while the synergistic effect of attention and MoE further reduces the risk of prediction failure, making the robustness of the model optimal.
The Interaction dataset focuses on multi-agent intensive interaction scenarios, which can better validate the modeling ability of the model for complex dynamic interactions. As shown in Figure 9, the m A D E 6 of Atten-LSTM-MoE reached 0.27, significantly higher than LTC-MoE (0.21), Atten-LTC (0.17), and our model (0.16). In multi-agent scenarios, vehicle motion is more affected by interference from adjacent agents, while LTC’s adaptive time constant can quickly capture trajectory changes caused by interactions, and spatial attention can encode the relative positional relationships between agents. The combination of the two modules reduces the short-term interaction prediction error of the original model by 40.7% compared to Atten-LSTM-MoE, reflecting the adaptability of the core module to complex dynamic interactions.
The m F D E 6 of Atten-LSTM-MoE is 0.81, which is nearly twice that of our model (0.42). The long-term trajectory of multi-agent systems relies on the dynamic evolution of interaction strategies, such as collaborative lane changing and competition for lanes. The long-term gradient stability of LTC avoids the gradient vanishing problem of traditional LSTM, while the dynamic expert selection of MoE can adapt to different interaction modes. The synergy of the two improves the long-term prediction accuracy of the original model by 48.1% compared to Atten-LSTM-MoE, highlighting the advantages of the framework in long-term interaction modeling.
Extreme interactions in multi-agent scenarios can easily lead to trajectory prediction failure. Therefore, the M R 6 of Atten-LSTM-MoE is 0.16, while our model is only 0.06, reducing the failure rate by 62.5%. The synergistic effect of attention and LTC enhances the robustness of the model to abnormal interactions. Combined with MoE’s sparse inference, it further reduces the risk of failure and significantly improves the robustness of the original model in complex scenarios.

4.3. Results and Analysis

4.3.1. Comparison with State-of-the-Art Models

Parameter test: as shown in Figure 10 and Table 1, it is clearly presented that the impact of the number of active experts ( t o p k = 1~4) in the Atten-LTC-enhanced MoE-based module on the three key metrics on the Argoverse dataset. From the data trend, we can see that (1) m F D E 6 as the endpoint error index, it always maintains the highest value (1.01~1.03), indicating that it is difficult to predict the trajectory endpoint; (2) the overall error of m A D E 6 is low (0.61~0.65), and it reaches the optimal value when the t o p k = 2, indicating that increasing the number of experts moderately helps to improve the overall accuracy of the trajectory predictions; (3) M R 6 is the most stable (0.11~0.13) as the conflict rate metric, which verifies the robustness of the MoE module. It is worth noting that when t o p k increases from 1 to 2, all metrics are improved, but the continued increase in the number of experts will lead to fluctuations in some metrics, which provides an important basis for model parameter optimization: in order to achieve the best balance between calculation efficiency and prediction accuracy, t o p k = 2 is selected in this work.
Compared with the state-of-the-art model: We first evaluate the single-agent trajectory prediction task, and predict data in Argoverse [12] compared with several of the most state-of-the-art baseline models. Figure 11, Figure 12, Figure 13 and Figure 14 summarize the performance comparison of different models ( b r i e r m F D E 6 , m A D E 6 , m F D E 6 and M R 6 as evaluation metrics) for fair comparison. The compared models considered in this test include LaneGCN [36], mmTransformer [37], DenseTNT [38], TPCN [39], SceneTransformer [40], HiVT [41], MultiPath++ [42], GANet [43], PAGA [44], and Wayformer [35]. It should be noted that the results shown in this paper do not use the integration method, except for MultiPath++ (only the integration results are shown).
As can be seen from Figure 11, the method proposed in this paper achieves the optimal result (1.68) on the b r i e r m F D E 6 metric on the Argoverse dataset, which is superior to the existing popular trajectory prediction algorithms, such as LaneGCN (2.06), DenseTNT (1.98), and Wayformer (1.74). As the model evolves from the traditional Graph Convolution Network structure to the Transformer and Mixture of Experts (MoE) mechanism, the prediction error metrics present a downward trend gradually. This indicates that our method has better expression ability and uncertainty control ability in modeling multimodal trajectory distribution and dynamic interaction in a complex urban intersection scene.
As shown in Figure 12, the m A D E 6 of this method on the Argoverse dataset reaches 0.74, which is better than all the models compared, including LaneGCN (0.87), DenseTNT (0.88), HiVT (0.77), and Wayformer (0.77). The overall trend shows that with the model from the traditional Convolution Network to the introduction of the Transformer structure and the fusion of the Mixture of Experts (MoE) mechanism, the trajectory prediction has achieved continuous optimization in the overall average displacement error. The vectorized feature extraction, temporal and spatial attention fusion, and dynamic expert selection strategies used in this paper can better model the complex interaction relationship and multimodal trajectory distribution, thus effectively reducing the overall deviation between the predicted trajectory and the real trajectory.
Figure 13 shows the m F D E 6 error performance of different models on the Argoverse dataset. The method proposed in this paper achieves the minimum error value of 1.12, which is better than Wayformer (1.16), GANet (1.16), MultiPath++ (1.13), and other existing advanced methods. The overall trend shows that, based on the Atten-LTC-advanced MoE mechanism, the model can dynamically select the prediction module according to different driving intentions and effectively improve the prediction accuracy of the end position of the future trajectory, combined with high-precision map vectorized coding. Since m F D E 6 emphasizes the prediction accuracy of the final time, the results further verify the ability of this method to generate a reasonable and coherent trajectory in complex traffic scenes, reflecting the comprehensive advantages of long-term dependence modeling and terminal optimization.
Figure 14 indicates that the comparison of various methods on M R 6 metric. Our method achieves a low error rate of 0.12, which is equal to or slightly superior to the state-of-the-art methods such as GANet and Wayformer, and is better than early methods such as LaneGCN (0.16) and mmTransformer (0.15). It reflects whether there is a large bias between the predicted trajectories and the real trajectories, especially for the safety of autonomous driving decision-making. The multimodal trajectory generation and spatio-temporal interactive modeling mechanism proposed in this paper effectively reduces the frequency of extreme trajectory deviation and improves the robustness and security ability of the system in a complex dynamic environment.
In addition, to verify the balance between accuracy and efficiency of the Atten-LTC-MoE framework, this section tested the core efficiency indicators (parameter count, floating-point operation count, inference delay, batch speed) of mainstream trajectory prediction models based on RTX 4090 GPU (batch size = 256). The results are shown in Table 2.
As shown in Table 2, the parameter count (8.77 M), MFLOPs (12.36), per-sample delay (2.1 ms), and FPS (1190) of our method proposed in this article are all optimal among the benchmark models compared. The advantage comes from the lightweight design of the framework, where the LTC unit adopts a sparse linear ODE structure instead of redundant gate control weights. At the same time, a top-k (k = 2) expert selection strategy is adopted in the MoE module, activating only two experts and avoiding parameter redundancy caused by full expert activation. It means that our method has the ability to balance accuracy and efficiency and is suitable for trajectory prediction needs in large-scale traffic scenarios.
On the other hand, in the more challenging task of multi-agent trajectory prediction, our proposed method is also compared with the state-of-the-art baseline models. Table 3 presents the accuracy comparison of multi-agent trajectory prediction tasks using the Interaction dataset. The comparison still focuses on three key metrics: m A D E 6 , m F D E 6 and M R 6 . Our proposed method performs well in all available metrics. Specifically, it achieves the lowest m A D E 6 (0.16), indicating that when considering the first six predicted trajectories, the average displacement error of all time steps is the smallest. Compared with SceneTransformer [40] and Thomas [45], this is a significant improvement, and its m A D E 6 is 0.26, respectively. In m F D E 6 , the proposed method also performs well, and the final displacement error is 0.42, which is better than SceneTransformer (0.47), ITRA [46] (0.49) and DenseTNT (0.67), indicating higher accuracy in the final prediction step. For the miss-detection rate M R 6 , our proposed method provides a competitive miss detection rate of 0.06, which is slightly higher than that of SceneTransformer and THOMAS (0.05). The M R 6 of our proposed methods is better than that of GOHOME [47]. It shows that the method proposed in this work can minimize the number of missed detections while maintaining the superior accuracy of other metrics. Generally, the proposed method shows significant improvement in the accuracy of average and final displacement, while maintaining a low rate of missed detection, making it the most reliable multi-agent prediction method in the dataset.

4.3.2. Qualitative Analysis

To gain a deeper understanding of the advantages and characteristics of the model proposed in this article, multiple typical scenarios were randomly selected from experimental data for detailed visualization analysis. Figure 15a–c shows the visualization results of multi-agent trajectory prediction for three different complex scenarios on the Interaction dataset [13]. These scenes represent complex road layouts commonly seen in the real world, where multiple agents interact in different traffic environments. The experimental samples in this article cover various typical road environments, including roundabouts and standard intersections.
In these visualizations, the red dots represent the possible trajectories predicted by the model, while the green curves represent the actual observed trajectories. From the visualization results, it can be clearly observed that the model proposed in this paper can generate high-quality multi-modal predictions, fully considering various possible motion paths under the current state of the vehicle and environmental constraints. The predicted trajectory strictly follows the lane structure and traffic rules, indicating that the model effectively integrates map features, such as road layout and lane lines, when predicting future movements, which is crucial for safe and reliable trajectory prediction in the autonomous driving system.
The roundabout scene in Figure 15a particularly demonstrates the model’s ability to handle complex circular road structures, and the predicted trajectory accurately reflects the turning behavior and speed changes in agents in the roundabout. The intersection scenarios in Figure 15b,c demonstrate the reliability of the model in predicting various behavioral patterns such as straight driving and turning. It is worth noting that the trajectories generated by the model not only approach the real trajectory in spatial position but also exhibit high consistency in velocity and acceleration characteristics, which is particularly important for safety planning and risk assessment in autonomous driving.

5. Conclusions

In this work, we combine the attention mechanism with LTC and the MoE model to build an extensible framework for trajectory prediction in single-agent or multi-agent complex traffic environments. Firstly, the Transformer model with an attention mechanism is used to fuse the feature vectors extracted from the lane graph and agent graph. Then, the fused feature vectors are mapped to the grids and input into the trajectory prediction module based on Atten-LTC-MoE, and then input into the trajectory generation module to gain the predicted trajectory of the target agent. We have conducted benchmark tests on two popular datasets, Argoverse and the Interaction dataset. The experimental results show that the model based on Atten-LTC-MoE is effective in solving the SATP and MATP tasks for autonomous driving. The LTC unit ensures the sparsity of our proposed model and the fusion attribute of space-time features. The prediction based on the MoE method improves the accuracy and efficiency of trajectory prediction. Future work could explore how to further enhance scalability and real-time processing capability, and reduce the scale of the expert network combined with knowledge distillation technology, consolidating the role of MATP in the technological progress of autonomous vehicles.

Author Contributions

Conceptualization: S.J. and R.W.; methodology: S.J.; software: S.J.; validation: R.D. and Q.Y.; formal analysis: R.W.; investigation: S.J.; resources: R.W.; data curation: S.J. and W.L.; writing—original draft preparation: S.J.; writing—review and editing: R.W. and S.J.; visualization: S.J.; supervision: R.W.; funding acquisition: R.W., R.D. and W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2023YFB2504500; National Natural Science Foundation of China, grant number 52472410; National Natural Science Foundation of China Youth Program, grant number 52502526; Jiangsu Funding Program for Excellent Postdoctoral Talent, grant number 2025ZB599; and Basic research program youth fund project of Jiangsu province, grant number BK20250841.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were used in this study. The Argoverse v1.1 dataset can be found at https://s3.amazonaws.com/argoverse/datasets/av1.1/tars/hd_maps.tar.gz (accessed on 27 September 2025), and the Interaction dataset can be found at INTERACTION Dataset GitHub https://github.com/interaction-dataset (accessed on 27 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bahram, M.; Hubmann, C.; Lawitzky, A.; Aeberhard, M.; Wollherr, D. A combined model-and learning-based framework for interaction-aware maneuver prediction. IEEE Trans. Intell. Transp. Syst. 2016, 17, 1538–1550. [Google Scholar] [CrossRef]
  2. Li, J.; Yang, F.; Tomizuka, M.; Choi, C. Evolvegraph: Multi-agent trajectory prediction with dynamic relational reasoning. Adv. Neural Inf. Process. Syst. 2020, 33, 19783–19794. [Google Scholar]
  3. Liu, C.; Lee, S.; Varnhagen, S.; Tseng, H.E. Path planning for autonomous vehicles using model predictive control. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 174–179. [Google Scholar]
  4. Brudigam, T.; Olbrich, M.; Wollherr, D.; Leibold, M. Stochastic Model Predictive Control With a Safety Guarantee for Automated Driving. IEEE Trans. Intell. Veh. 2021, 8, 22–36. [Google Scholar] [CrossRef]
  5. Xu, S.; Peng, H. Design, analysis, and experiments of preview path tracking control for autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2019, 21, 48–58. [Google Scholar] [CrossRef]
  6. Biktairov, Y.; Stebelev, M.; Rudenko, I.; Shliazhko, O.; Yangel, B. Prank: Motion prediction based on ranking. Adv. Neural Inf. Process. Syst. 2020, 33, 2553–2563. [Google Scholar]
  7. Casas, S.; Luo, W.; Urtasun, R. Intentnet: Learning to predict intention from raw sensor data. In Proceedings of the Conference on Robot Learning, Zürich, Switzerland, 29–31 October 2018; pp. 947–956. [Google Scholar]
  8. Cui, H.; Radosavljevic, V.; Chou, F.-C.; Lin, T.-H.; Nguyen, T.; Huang, T.-K.; Schneider, J.; Djuric, N. Multimodal trajectory predictions for autonomous driving using deep convolutional networks. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 2090–2096. [Google Scholar]
  9. Djuric, N.; Radosavljevic, V.; Cui, H.; Nguyen, T.; Chou, F.-C.; Lin, T.-H.; Singh, N.; Schneider, J. Uncertainty-aware short-term motion prediction of traffic actors for autonomous driving. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 2095–2104. [Google Scholar]
  10. Zeng, W.; Liang, M.; Liao, R.; Urtasun, R. Lanercnn: Distributed representations for graph-centric motion forecasting. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 532–539. [Google Scholar]
  11. Lechner, M.; Hasani, R.; Amini, A.; Henzinger, T.A.; Rus, D.; Grosu, R. Neural circuit policies enabling auditable autonomy. Nat. Mach. Intell. 2020, 2, 642–652. [Google Scholar] [CrossRef]
  12. Chang, M.-F.; Lambert, J.; Sangkloy, P.; Singh, J.; Bak, S.; Hartnett, A.; Wang, D.; Carr, P.; Lucey, S.; Ramanan, D. Argoverse: 3d tracking and forecasting with rich maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 8748–8757. [Google Scholar]
  13. Zhan, W.; Sun, L.; Wang, D.; Shi, H.; Clausse, A.; Naumann, M.; Kummerle, J.; Konigshof, H.; Stiller, C.; de La Fortelle, A. Interaction dataset: An international, adversarial and cooperative motion dataset in interactive driving scenarios with semantic maps. arXiv 2019, arXiv:1910.03088. [Google Scholar] [CrossRef]
  14. Liu, Y.; Qi, X.; Sisbot, E.A.; Oguchi, K. Multi-agent trajectory prediction with graph attention isomorphism neural network. In Proceedings of the 2022 IEEE Intelligent Vehicles Symposium (IV), Aachen, Germany, 5–9 June 2022; pp. 273–279. [Google Scholar]
  15. Ivanovic, B.; Pavone, M. The trajectron: Probabilistic multi-agent trajectory modeling with dynamic spatiotemporal graphs. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2375–2384. [Google Scholar]
  16. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  17. Song, X.; Chen, K.; Li, X.; Sun, J.; Hou, B.; Cui, Y.; Zhang, B.; Xiong, G.; Wang, Z. Pedestrian trajectory prediction based on deep convolutional LSTM network. IEEE Trans. Intell. Transp. Syst. 2020, 22, 3285–3302. [Google Scholar] [CrossRef]
  18. Zhang, P.; Ouyang, W.; Zhang, P.; Xue, J.; Zheng, N. Sr-lstm: State refinement for lstm towards pedestrian trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 12085–12094. [Google Scholar]
  19. Hou, L.; Xin, L.; Li, S.E.; Cheng, B.; Wang, W. Interactive trajectory prediction of surrounding road users for autonomous driving using structural-LSTM network. IEEE Trans. Intell. Transp. Syst. 2019, 21, 4615–4625. [Google Scholar] [CrossRef]
  20. Sun, H.; Chen, R.; Liu, T.; Wang, H.; Sun, F. LG-LSTM: Modeling LSTM-based interactions for multi-agent trajectory prediction. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar]
  21. Manh, H.; Alaghband, G. Scene-lstm: A model for human trajectory prediction. arXiv 2018, arXiv:1808.04018. [Google Scholar]
  22. Jiang, R.; Xu, H.; Gong, G.; Kuang, Y.; Liu, Z. Spatial-temporal attentive LSTM for vehicle-trajectory prediction. ISPRS Int. J. Geo-Inf. 2022, 11, 354. [Google Scholar] [CrossRef]
  23. Yu, J.; Zhou, M.; Wang, X.; Pu, G.; Cheng, C.; Chen, B. A dynamic and static context-aware attention network for trajectory prediction. ISPRS Int. J. Geo-Inf. 2021, 10, 336. [Google Scholar] [CrossRef]
  24. Hasani, R.; Lechner, M.; Amini, A.; Rus, D.; Grosu, R. Liquid time-constant networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; pp. 7657–7666. [Google Scholar]
  25. Kaplan, H.S.; Thula, O.S.; Khoss, N.; Zimmer, M. Nested neuronal dynamics orchestrate a behavioral hierarchy across timescales. Neuron 2020, 105, 562–576.e9. [Google Scholar] [CrossRef]
  26. Lu, Y.; Wang, W.; Bai, R.; Zhou, S.; Garg, L.; Bashir, A.K.; Jiang, W.; Hu, X. Hyper-relational interaction modeling in multi-modal trajectory prediction for intelligent connected vehicles in smart cites. Inf. Fusion 2025, 114, 102682. [Google Scholar] [CrossRef]
  27. Wu, W.; Li, Z.; Gu, Y.; Zhao, R.; He, Y.; Zhang, D.J.; Shou, M.Z.; Li, Y.; Gao, T.; Zhang, D. Draganything: Motion control for anything using entity representation. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 331–348. [Google Scholar]
  28. Vijayabaskaran, S.; Zeng, X.; Ghazinouri, B.; Wiskott, L.; Cheng, B. A taxonomy of spatial navigation in mammals: Insights from computational modeling. Neurosci. Biobehav. Rev. 2025, 176, 106282. [Google Scholar] [CrossRef] [PubMed]
  29. Riquelme, C.; Puigcerver, J.; Mustafa, B.; Neumann, M.; Jenatton, R.; Susano Pinto, A.; Keysers, D.; Houlsby, N. Scaling vision with sparse mixture of experts. Adv. Neural Inf. Process. Syst. 2021, 34, 8583–8595. [Google Scholar]
  30. Zhou, Y.; Lei, T.; Liu, H.; Du, N.; Huang, Y.; Zhao, V.; Dai, A.M.; Le, Q.V.; Laudon, J. Mixture-of-experts with expert choice routing. Adv. Neural Inf. Process. Syst. 2022, 35, 7103–7114. [Google Scholar]
  31. Gao, J.; Sun, C.; Zhao, H.; Shen, Y.; Anguelov, D.; Li, C.; Schmid, C. Vectornet: Encoding hd maps and agent dynamics from vectorized representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11525–11533. [Google Scholar]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  33. Hasani, R.; Lechner, M.; Amini, A.; Liebenwein, L.; Ray, A.; Tschaikowski, M.; Teschl, G.; Rus, D. Closed-form continuous-time neural networks. Nat. Mach. Intell. 2022, 4, 992–1003. [Google Scholar] [CrossRef]
  34. Aydemir, G.; Akan, A.K.; Güney, F. Adapt: Efficient multi-agent trajectory prediction with adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 8295–8305. [Google Scholar]
  35. Nayakanti, N.; Al-Rfou, R.; Zhou, A.; Goel, K.; Refaat, K.S.; Sapp, B. Wayformer: Motion forecasting via simple & efficient attention networks. arXiv 2022, arXiv:2207.05844. [Google Scholar] [CrossRef]
  36. Liang, M.; Yang, B.; Hu, R.; Chen, Y.; Liao, R.; Feng, S.; Urtasun, R. Learning lane graph representations for motion forecasting. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 541–556. [Google Scholar]
  37. Liu, Y.; Zhang, J.; Fang, L.; Jiang, Q.; Zhou, B. Multimodal motion prediction with stacked transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 7577–7586. [Google Scholar]
  38. Gu, J.; Sun, C.; Zhao, H. Densetnt: End-to-end trajectory prediction from dense goal sets. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15303–15312. [Google Scholar]
  39. Ye, M.; Cao, T.; Chen, Q. Tpcn: Temporal point cloud networks for motion forecasting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 11318–11327. [Google Scholar]
  40. Ngiam, J.; Caine, B.; Vasudevan, V.; Zhang, Z.; Chiang, H.-T.L.; Ling, J.; Roelofs, R.; Bewley, A.; Liu, C.; Venugopal, A. Scene transformer: A unified architecture for predicting multiple agent trajectories. arXiv 2021, arXiv:2106.08417. [Google Scholar]
  41. Zhou, Z.; Ye, L.; Wang, J.; Wu, K.; Lu, K. Hivt: Hierarchical vector transformer for multi-agent motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8823–8833. [Google Scholar]
  42. Varadarajan, B.; Hefny, A.; Srivastava, A.; Refaat, K.S.; Nayakanti, N.; Cornman, A.; Chen, K.; Douillard, B.; Lam, C.P.; Anguelov, D. Multipath++: Efficient information fusion and trajectory aggregation for behavior prediction. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 7814–7821. [Google Scholar]
  43. Wang, M.; Zhu, X.; Yu, C.; Li, W.; Ma, Y.; Jin, R.; Ren, X.; Ren, D.; Wang, M.; Yang, W.J. Ganet: Goal area network for motion forecasting. arXiv 2022, arXiv:2209.09723. [Google Scholar]
  44. Da, F.; Zhang, Y. Path-aware graph attention for hd maps in motion prediction. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 6430–6436. [Google Scholar]
  45. Gilles, T.; Sabatini, S.; Tsishkou, D.; Stanciulescu, B.; Moutarde, F. Thomas: Trajectory heatmap output with learned multi-agent sampling. arXiv 2021, arXiv:2110.06607. [Google Scholar]
  46. Ścibior, A.; Lioutas, V.; Reda, D.; Bateni, P.; Wood, F. Imagining the road ahead: Multi-agent trajectory prediction via differentiable simulation. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 720–725. [Google Scholar]
  47. Gilles, T.; Sabatini, S.; Tsishkou, D.; Stanciulescu, B.; Moutarde, F. Gohome: Graph-oriented heatmap output for future motion estimation. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 9107–9114. [Google Scholar]
Figure 1. Example of multi-agent trajectory prediction.
Figure 1. Example of multi-agent trajectory prediction.
Sensors 26 00479 g001
Figure 2. The overall framework of the proposed Atten-LTC-advanced MoE model for trajectory prediction. The proposed method consists of Feature Fusion, Atten-LTC-MoE-based Trajectory Prediction, and Trajectory Generation modules.
Figure 2. The overall framework of the proposed Atten-LTC-advanced MoE model for trajectory prediction. The proposed method consists of Feature Fusion, Atten-LTC-MoE-based Trajectory Prediction, and Trajectory Generation modules.
Sensors 26 00479 g002
Figure 3. Encoder–decoder architecture and Multi-Head attention (MHA) are used in the feature fusion module.
Figure 3. Encoder–decoder architecture and Multi-Head attention (MHA) are used in the feature fusion module.
Sensors 26 00479 g003
Figure 4. Illustration of the scaled dot-product attention and multi-head attention (MHA) from the Transformer block [32] used in the Feature Fusion Module.
Figure 4. Illustration of the scaled dot-product attention and multi-head attention (MHA) from the Transformer block [32] used in the Feature Fusion Module.
Sensors 26 00479 g004
Figure 5. The flow of the spatio–temporal Atten-LTC encoder–decoder module.
Figure 5. The flow of the spatio–temporal Atten-LTC encoder–decoder module.
Sensors 26 00479 g005
Figure 6. Internal structure of the Liquid Time-Constant (LTC) network.
Figure 6. Internal structure of the Liquid Time-Constant (LTC) network.
Sensors 26 00479 g006
Figure 7. Trajectory prediction using the ensemble Mixture of Experts (MoE).
Figure 7. Trajectory prediction using the ensemble Mixture of Experts (MoE).
Sensors 26 00479 g007
Figure 8. Comparison of Atten-LTC-MoE framework core module ablation experiment on the Argoverse dataset.
Figure 8. Comparison of Atten-LTC-MoE framework core module ablation experiment on the Argoverse dataset.
Sensors 26 00479 g008
Figure 9. Comparison of Atten-LTC-MoE framework core module ablation experiment on the Interaction dataset.
Figure 9. Comparison of Atten-LTC-MoE framework core module ablation experiment on the Interaction dataset.
Sensors 26 00479 g009
Figure 10. Trajectory prediction using the ensemble Mixture of Experts (MoE). The influence of the number of t o p k expert selections on trajectory prediction metrics ( m A D E 6 , m F D E 6 , M R 6 ) in the Atten-LTC-MoE framework.
Figure 10. Trajectory prediction using the ensemble Mixture of Experts (MoE). The influence of the number of t o p k expert selections on trajectory prediction metrics ( m A D E 6 , m F D E 6 , M R 6 ) in the Atten-LTC-MoE framework.
Sensors 26 00479 g010
Figure 11. b r i e r m F D E 6 error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Figure 11. b r i e r m F D E 6 error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Sensors 26 00479 g011
Figure 12. m A D E 6   error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Figure 12. m A D E 6   error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Sensors 26 00479 g012
Figure 13. m F D E 6   error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Figure 13. m F D E 6   error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Sensors 26 00479 g013
Figure 14. M R 6 error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Figure 14. M R 6 error metric comparison for single-agent trajectory prediction on the Argoverse dataset [12].
Sensors 26 00479 g014
Figure 15. Visualization of multi-agent trajectory prediction results on the Interaction dataset [13]: (a) scene CHN roundabout; (b) scene USA intersection 1; and (c) scene USA intersection 2.
Figure 15. Visualization of multi-agent trajectory prediction results on the Interaction dataset [13]: (a) scene CHN roundabout; (b) scene USA intersection 1; and (c) scene USA intersection 2.
Sensors 26 00479 g015
Table 1. Performance analysis of the impact of the number of activated experts in the Atten-LTC-advanced MoE-based trajectory prediction module on the Argoverse dataset [12].
Table 1. Performance analysis of the impact of the number of activated experts in the Atten-LTC-advanced MoE-based trajectory prediction module on the Argoverse dataset [12].
Top-kminADE6minFDE6MR6
10.651.030.13
20.611.010.11
30.631.010.12
40.631.020.11
Table 2. Comparison of computational efficiency indicators with baseline prediction models at batch size = 256.
Table 2. Comparison of computational efficiency indicators with baseline prediction models at batch size = 256.
ModelsParameter (M)MFLOPsPer-Sample Latency (ms)FPS
DenseTNT [38]12.9321.445.1610
LaneGCN [36]10.2418.523.4735
Wayformer [35]15.6927.815.2481
Our method8.7712.363.11190
Table 3. Performance analysis of the impact of the number of activated experts in the Atten-LTC-advanced MoE-based trajectory prediction module on the Interaction dataset [13].
Table 3. Performance analysis of the impact of the number of activated experts in the Atten-LTC-advanced MoE-based trajectory prediction module on the Interaction dataset [13].
ModelsminADE6minFDE6MR6
DenseTNT [38]0.210.67-
SceneTransformer [40]0.260.470.05
ITRA [46]0.170.49-
GOHOME [47]-0.450.07
THOMAS [45]0.260.460.05
Our method0.160.420.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, S.; Wang, R.; Ding, R.; Ye, Q.; Liu, W. Atten-LTC-Enhanced MoE Model for Agent Trajectory Prediction in Autonomous Driving. Sensors 2026, 26, 479. https://doi.org/10.3390/s26020479

AMA Style

Jiang S, Wang R, Ding R, Ye Q, Liu W. Atten-LTC-Enhanced MoE Model for Agent Trajectory Prediction in Autonomous Driving. Sensors. 2026; 26(2):479. https://doi.org/10.3390/s26020479

Chicago/Turabian Style

Jiang, Shangwu, Ruochen Wang, Renkai Ding, Qing Ye, and Wei Liu. 2026. "Atten-LTC-Enhanced MoE Model for Agent Trajectory Prediction in Autonomous Driving" Sensors 26, no. 2: 479. https://doi.org/10.3390/s26020479

APA Style

Jiang, S., Wang, R., Ding, R., Ye, Q., & Liu, W. (2026). Atten-LTC-Enhanced MoE Model for Agent Trajectory Prediction in Autonomous Driving. Sensors, 26(2), 479. https://doi.org/10.3390/s26020479

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop