You are currently viewing a new version of our website. To view the old version click .
World Electric Vehicle Journal
  • Article
  • Open Access

15 January 2024

Vehicle Trajectory Prediction Based on Local Dynamic Graph Spatiotemporal–Long Short-Term Memory Model

,
and
1
SHU-UTS SILC Business School, Shanghai University, Shanghai 201899, China
2
Smart City Research Institute, Shanghai University, Shanghai 201899, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Development towards Vehicle Safety in Future Smart Traffic Systems

Abstract

Traffic congestion and frequent traffic accidents have become the main problems affecting urban traffic. The effective location prediction of vehicle trajectory can help alleviate traffic congestion, reduce the occurrence of traffic accidents, and optimize the urban traffic system. Vehicle trajectory is closely related to the surrounding Point of Interest (POI). POI can be considered as the spatial feature and can be fused with trajectory points to improve prediction accuracy. A Local Dynamic Graph Spatiotemporal–Long Short-Term Memory (LDGST-LSTM) was proposed in this paper to extract and fuse the POI knowledge and realize next location prediction. POI semantic information was learned by constructing the traffic knowledge graph, and spatial and temporal features were extracted by combining the Graph Attention Network (GAT) and temporal attention mechanism. The effectiveness of LDGST-LSTM was verified on two datasets, including Chengdu taxi trajectory data in August 2014 and October 2018. The accuracy and robustness of the proposed model were significantly improved compared with the benchmark models. The effects of major components in the proposed model were also evaluated through an ablation experiment. Moreover, the weights of POI that influence location prediction were visualized to improve the interpretability of the proposed model.

1. Introduction

Traffic congestion and frequent traffic accidents have become serious problems affecting urban traffic with the rapid development of cities [1]. The Intelligent Transportation System (ITS) has been proposed to alleviate these traffic problems. The location prediction of vehicle trajectory is one of the research topics relating to the ITS [2]. Through effective location prediction, traffic congestion can be alleviated, which is of great significance in terms of optimizing traffic management. The goal of vehicle location prediction is to predict the next location of the vehicle according to the observed trajectory data [3]. Deep learning methods have been widely used in vehicle trajectory location prediction. Vehicle trajectory is usually considered a time series since it is a sequence of sampled points [1]. Long Short-Term Memory (LSTM) was mainly used to deal with time sequences. Convolution Neural Network (CNN) was used by Guo [1] to extract the spatial features of vehicles, and a Deep Bidirectional Long Short-Term Memory (DBLSTM) model was used to extract the temporal features to realize the location prediction of vehicles. A LSTM was applied by Li et al. [4] to capture the temporal features between vehicles and design a convolutional social pooling network to capture spatial features to predict vehicle trajectory. A weighted support vector regression algorithm was proposed by Liao [5] to allocate weight according to the calculation and comparison of trajectory points to achieve vehicle location prediction, improving the accuracy and robustness of the results. Methods based on deep learning algorithms can effectively improve prediction accuracy and reduce error. Through feature mining and extraction, deep learning models can effectively improve the accuracy and robustness of the prediction results. However, complex road structures and the surrounding environment could become the factors affecting vehicle location prediction. The impact of surrounding buildings, or to say, Point of Interest (POI), on prediction under complex traffic environments was usually ignored in the existing research. Based on this situation, the performance of vehicle location prediction in the real traffic network cannot be satisfied. Therefore, how to integrate external features with vehicle trajectory to improve the accuracy and robustness of the model needs further research.
Knowledge graph (KG), Graph Attention Network (GAT), and time attention mechanism are popular methods used lately in the field of transportation and have shown better performance in improving the accuracy and robustness of the model [6]. The knowledge graph can help fuse features. GAT can be used as a spatial attention mechanism to effectively extract spatial features, and a temporal attention mechanism can effectively extract temporal features. The application of knowledge graph has developed rapidly in the field of transportation. For example, a multi-layer traffic knowledge graph model was proposed by Wang [7] to realize destination prediction by modeling the complex traffic network, weighted summing, and fusing different node features. The driving intention in the traffic network is affected by the functional area and the surrounding Point of Interest (POI) [8]. Therefore, it is important to predict the location of vehicle trajectory by designing a traffic knowledge graph. The accuracy can be improved by interacting the vehicle trajectory with the surrounding environments [8]. Point of interest (POI) generally refers to a location or a specific point on a map or in a geographical area. Those points are often marked on maps to help people navigate and discover places of interest, such as hotels, restaurants, shopping malls, or other noteworthy locations. In trajectory prediction, the driving intention of vehicles is closely related to the surrounding POI [8]. Therefore, POI can be considered as prior knowledge and spatial feature to help model understanding of the context. It can be fused with trajectory points to improve prediction accuracy [9]. The trajectory may change when there are traffic jams or traffic accidents, and the POI knowledge around trajectory points has an impact on the future trend. It is of significance to the optimization of traffic management by integrating POI knowledge and vehicle trajectory to realize position prediction and improve the accuracy of the prediction. Graph Attention Network (GAT) [10] and various temporal attention mechanisms [11] are also widely used for extracting spatial and temporal features. A new GAT is proposed by Su et al. [12] to allocate attention to explain and express the spatial correlation of road networks, as well as effectively capture the dynamic update of the adjacent transformation matrix. LSTM combined with the attention mechanism was proposed by Ali et al. [13] to capture temporal features. However, how to fuse the surrounding environment features such as POI and the synergy of knowledge graph, GAT, and temporal attention mechanism on location prediction tasks still lacks research.
A Local Dynamic Graph Spatiotemporal–Long Short-Term Memory (LDGST–LSTM) is proposed in this paper to predict the next location of vehicle trajectory. Through the proposed model, POI knowledge is extracted and fused by constructing a traffic knowledge graph. Additionally, GAT and temporal attention mechanisms are combined to improve the accuracy, robustness, and interpretability of the prediction. The main contributions of this paper are as follows.
  • Knowledge graph and spatiotemporal attention mechanisms were combined in this paper to predict the vehicle location at the next moment. POI was integrated with historical trajectory, and the POI weights that affect the prediction were visualized additionally. Regions that have a great impact on the prediction are explored, and the interpretability of the model was enhanced.
  • A global traffic knowledge graph was constructed to learn and represent POI semantic information. POI nodes were considered as the entity, and the connection between POI was considered as the relationship. The representation vector of each node was obtained by using the Translate Embedding (TransE) algorithm and was considered as the feature vector for vehicle location prediction.
  • A spatiotemporal attention mechanism was designed to allocate weights for spatial and temporal features, thus enhancing the interpretability and accuracy of the model. The weight distribution of spatial features was achieved through GAT to obtain the corresponding graph representation vector. LSTM combined with a multi-head attention mechanism was applied to allocate weights of trajectory points at different time steps to improve the prediction accuracy.
Section 2 introduces a thorough review of the research on vehicle trajectory prediction from three aspects: deep learning, knowledge graphs, and attention mechanisms. The basic concepts and studied problems are formally defined in Section 3. The research methodology and details of the LDGST-LSTM model are presented in Section 4. Then, Section 5 provides a description of the experiment settings. Experiments on robustness, accuracy, and ablation are conducted to confirm the effectiveness of the proposed model. In Section 6, the findings of this paper are outlined. Moreover, limitations and further directions are also described.

3. Problem Statement

In this section, some basic concepts and notations in this paper are first introduced, and then the studied problem is formally defined.
Definition 1
(Raw trajectory T i ). A raw vehicle trajectory is usually represented by a sequence of points continuously sampled by the Global Positioning System (GPS). Given a raw trajectory dataset T r a j = { T 1 , T 2 , T n } , the trajectory i , T i = P i 1 , P i 2 , , P i k , is defined as a sequence of sampled points P i j , where P i j R 2 , i 1,2 , , n , j 1,2 , , k . A sampled point P i j is defined as a tuple ( l n g i j , l a t i j ) which represents the longitude and latitude of the GPS point at timestamp j . Due to varying sampling rate, trajectory data may have irregular distribution characteristics. For example, the data may be dense in some areas of the map but sparse in other areas.
Definition 2
(Point of Interest Q u ). Point Of Interest (POI) can be denoted as a spatial representation of urban infrastructure such as schools, hospitals, restaurants and so on. It can reflect land use and urban functional characteristics and has potential semantic information. Its distribution influences the intentions of travel. POI can be classified into different types. Given a POI dataset with h types P O I = { Q 1 , Q 2 , , Q h } , the type u of POI Q u = { I u 1 , I u 2 , I u m } is defined as a set of points I u j , where u 1 , , h ,   j 1 , , m . A POI I u j = ( n a m e u j ,   a d d r e s s u j , l n g u j , l a t u j , a t t r i b u t e u j ) is a five-tuple, that represents the semantic information including the name, address, longitude, latitude, as well as the attribute of the j t h POI. In this paper, eleven different categories of POI are applied, including stores, entertainment, food, hotels, scenic spots, finance, government, company, hospitals, life services, and sports.
Definition 3
(Road network G ). A road network is defined as a graph G = ( V , E , A ) , where V is the set of nodes; each v i denotes a road segment. E is the set of edges connecting nodes, and each e i , j = ( v i , v j ) denotes the connectivity between road segment v i and v j . The adjacency matrix of the graph can be denoted as A R V × V , and semantic information of different POI categories are considered as features. Each element a i , j in this matrix is a binary value, which is 1 when road segment v i is adjacent to v j and 0 otherwise.
Definition 4
(Normalized trajectory N T i ). Raw trajectory T i will be processed through a data conversion layer in this paper and transformed into a trajectory in normalization form N T i . Due to the measurement error, it is not appropriate to directly input raw trajectory to the model. Given a raw trajectory dataset T r a j = { T 1 , T 2 , T n } , the trajectory T i = P i 1 , P i 2 , , P i k . Match each point P i j to the nearest node of the road network. A normalized point N P i j is defined as a tuple ( v l n g i j , v l a t i j ) , which represents the longitude and latitude of node j . A normalized trajectory N T i is then defined as a sequence of road segments after projection.
Definition 5
(Normalized POI N Q u ). Match each POI I u j to the nodes of road network that has the shortest projection distance. A normalized POI N I u j is defined as the corresponding road segment after projection, and POI semantic information is assigned to the matched nodes as node features. The normalized POI of type u is denoted as N Q u and the semantic information of the normalized POI j of type u is denoted as N I u j .
Definition 6
(POI Knowledge graph K G P O I ). Defining a knowledge graph K G P O I = ( E P O I , R P O I , T P O I ) , where E P O I = { e 1 , e 2 , e 3 , e m } is the POI entity set, and where m is the number of the POI entities. R P O I = { r 1 , r 2 , r 3 , r n } is the POI relationship set, where n is the number of the POI relationship. T P O I = { ( h P O I , l P O I , t P O I ) } is a triple set, which refers to head entity, relation, and tail entity, respectively.
Problem 
(Vehicle trajectory location prediction). Given a raw trajectory dataset T r a j = { T 1 , T 2 , T n } and a road network G , for current point P i j of the trajectory T i , the task aims to predict the next location P i j + 1 , which is a tuple ( l n g i j + 1 , l a t i j + 1 ) consisting of the longitude and latitude of a GPS point at timestamp j + 1 .

4. Methodology

Vehicle trajectory is a sequence of continuously sampled points, and its intention is closely related to the Point of Interest (POI). The semantic information of POI near each trajectory point may have an impact on the next location prediction. Therefore, a Local Dynamic Graph Spatiotemporal–Long Short-Term Memory model (LDGST-LSTM) was proposed in this paper to realize location prediction for vehicle trajectory and explore the impact weight of the nearby POI.
Raw vehicle trajectory and POI were first matched to the traffic network through the data conversion layer in the proposed model. The data of POI and road network were extracted from an open-source map website, OpenStreetMap (OSM). Then, a global POI Knowledge Graph (KG) was constructed to obtain the representation vectors of the POI entities in the global POI knowledge extraction layer. Based on the global knowledge graph, local graphs related to each trajectory point were generated, and the graph representation vector was captured through the Graph Attention Network (GAT) in the local dynamic graph generation module. Finally, trajectory points with the related graph representation vectors were input into the Long Short-Term Memory model (LSTM) with a multi-head temporal attention mechanism (T-Attn) in the trajectory prediction layer to predict the next location.
In this section, details of the proposed model LDGST-LSTM are provided, which consists of four major components, as follows: a data conversion layer, a global POI knowledge extraction layer, a local dynamic graph generation module, and a trajectory prediction layer. The overall framework is first described in Section 4.1. Then, each component in this model is specifically introduced in Section 4.2, Section 4.3, Section 4.4, Section 4.5.

4.1. Overall Framework

The overall framework of LDGST-LSTM is shown in Figure 1. There are four major components in the proposed model: (1) a data conversion layer, (2) a global POI knowledge extraction layer, (3) a local dynamic generation module, and (4) a trajectory prediction layer. Additionally, the local dynamic generation module is concluded in the trajectory prediction layer. The observed vehicle trajectory T i = { P i j t , P i j t + 1 , , P i j } was considered as the input of the model, and the predicted next location P i j + 1 of the trajectory at timestamp j + 1 was considered the output.
Figure 1. The overall framework of Local Dynamic Graph Spatiotemporal–Long Short-Term Memory (LDGST-LSTM) with four main components, as follows: (1) data conversion layer, (2) global POI knowledge extraction layer, (3) local dynamic generation module, and (4) trajectory prediction layer.
As shown in Figure 1, vehicle trajectory T i was the input of the model. Firstly, normalized trajectory N T i and normalized POI N Q u were obtained through map matching and proximity algorithm, respectively, in the data conversion layer. Then, the representation vector V P O I i j of every POI was trained through the usage of the Translating Embedding (TransE) algorithm based on the knowledge graph in the global POI knowledge extraction layer. It will be considered semantic features of the trajectory points. In the trajectory prediction layer, local graphs G i j t : j related to trajectory points P i j t : j were first generated, and the corresponding graph representation vectors N R V i j t : j were obtained through GAT in the local dynamic graph generation module. Normalized trajectory points N P i j t : j and corresponding graph representation vectors N P V i j t : j were then concatenated and input into LSTM with a multi-head attention mechanism. The model finally output the predicted next location P i j + 1 of the trajectory. The overall framework can be denoted through the use of the following Formulas (1)–(4), where P j t : j = ( P j t , P j t + 1 , , P j ) .
  • Data conversion layer:
T i m a p   m a t c h i n g N T i , Q u p r o x i m i t y   a l g . N Q u
  • Global POI knowledge extraction layer:
K G P O I , N Q u T r a n s E V P O I i j
  • Local dynamic graph generation module:
P i j t : j V P O I i j G i j t : j G A T N R V i j t : j
  • Trajectory prediction layer:
N P i j t : j ; N R V i j t : j L S T M P i j + 1

4.2. Data Conversion Layer

This section will introduce road network data, trajectory data conversion and POI data conversion specifically.

4.2.1. Road Network

The open-source map website Open Street Map (OSM) provided the urban road network data needed in this work. Taking Chengdu as an example, the road network is shown in Figure 2 and it contains three parts of information. Figure 2a represents the road vector map and Figure 2b is the node vector map. Figure 2c visualizes the partial satellite projection map of Chengdu along with the node vectors of road (the red dots). This paper only maintains the road network data within the Third Ring Road of Chengdu as the research area, which can mainly represent the urban district.
Figure 2. Road network of Chengdu: (a) road vector map; (b) node vector map; (c) satellite projection map.

4.2.2. Trajectory Data Conversion

In the data conversion layer, raw trajectory data were converted into a normalized trajectory. Every point was denoted by the latitude and longitude corresponding to the matched road network node. The original geographic coordinate system (GCJ-02) of the trajectory data was first converted to WGS84, the geographic coordinate system of the road network data in this paper. A predetermined threshold was taken into consideration as the radius to search for candidate points, and the candidate points with the shortest distance were chosen as the normalized trajectory points. This approach was based on the earlier map-matching work of Lou et al. [34]. Points that were not successfully matched will be deleted.

4.2.3. POI Data Conversion

POI data of Chengdu in 2018 were used in this paper. The proximity algorithm was used in ArcGIS, and the nearest nodes were selected in the road network for each POI point based on distance. The context information of POI points was matched to the relevant nodes as the normalized POIs. POIs of different types were matched to the nodes of the road network. Taking life services (black dots), food (orange dots), and hospitals (blue dots) as examples, the partially matched visualization of POI can be seen in Figure 3.
Figure 3. Partial visualization of normalized POI.

4.3. Global POI Knowledge Extraction Layer

A global graph G = ( V , E , A ) and a knowledge graph K G P O I = ( E P O I , R P O I , T P O I ) were constructed in the global POI knowledge extraction layer, as shown in Figure 4. The entity set, relationship set, and related triples were defined to study the representation vectors of each POI entity through TransE algorithm. In the defined global graph, each node in the global graph contained the related POI knowledge through the data conversion layer.
Figure 4. Global POI knowledge extraction layer.
As shown in Figure 4, the normalized POI was considered as the entity, and the link between the normalized POI was considered as the relationship, which in this paper denotes that there was a connection between two POI nodes. Moreover, the triplet T P O I = { ( h , l , t ) } was considered as the training set of TransE algorithm, where h represented the head entity, t represented the tail entity, and l represented the relationship; h and t belongs to the entity set E P O I , and l belongs to the relationship set R P O I . The target of the TransE algorithm was to consider the relationship as the translation from the head entity to the tail entity, or to say, to made h + l equal to t as much as possible. The potential energy of a triplet is defined by the L2 norm of the difference between h + l and t , as shown in Formula (5), where N is the number of the triplets, and i represents the triplet i .
f h i , l i , t i = h i + l i t i 2 = i = 1 N h i + l i t i 2
Wrong triplets were identified and considered as negative samples in TransE algorithm for uniform sampling. A negative sample was generated when any one of the factors in a positive sample was replaced by the other entities or relationship randomly. The potential energy of the positive samples was reduced, and the potential energy of the negative samples were increased in the TransE algorithm. The objective function is defined in Formula (6).
L = h , l , t h , l , t f r h , t + γ f r h , t +
where is the set of positive samples, and is the set of negative samples. γ is a constant, usually is set as 1, which represents the distance between positive and negative samples. ( h , l , t ) and ( h , l , t ) are the triplets of positive and negative samples, respectively. f ( · ) is the potential energy function and [ · ] + is m a x ( 0 , · ) .
The distributed representation vector of the current head and tail entities were considered as the representation vectors V P O I i j R d . The pseudocode of the global POI knowledge extraction layer is shown in Algorithm 1.
Algorithm 1: Global POI knowledge extraction layer
Input   entity ,   relation ,   and   training   sets   E P O I , R P O I , T P O I = { ( h , l , t ) } ,   embedding   dim   k
1:normalize  l   uniform   ( 6 k , 6 k )   for   each   l R P O I
2:      l / l   for   each   l R P O I
3:      e   uniform   ( 6 k , 6 k )   for   each   e E P O I
4:loop
5: e / e   for   each   e E P O I
6: S b a t c h P O I   sample   ( T P O I , b )   / / sample   by   size   b
7: T b a t c h P O I     //initialize the triplets
8:for ( h , l , t ) S b a t c h P O I  do
9:   ( h , l , t )   sample   S ( h , l , t ) P O I  //extract negative samples
10:   T b a t c h P O I T b a t c h P O I { h , l , t , h , l , t } //extract positive and negative samples randomly
11:end for
12: Update   embeddings   h , l , t h , l , t f r h , t + γ f r h , t +  //L
13:end loop
Output   representation   vector   V i j P O I of the current entity

4.4. Local Dynamic Graph Generation Module

Local graphs were generated for each trajectory point in the local dynamic graph generation module. The Graph Attention network (GAT) was used as a spatial attention mechanism to allocate weight and update the parameters of every trajectory point and its neighbors. Corresponding graph representation vector N R V i j can be obtained through GAT.
As shown in Figure 5, every node in the global graph was embedded with the POI representation vector V P O I i j based on the POI knowledge graph in Section 4.3. The feature matrix F R N × k was constructed by all the embeddings of nodes, F i j = V P O I i j , where N is the number of the graph nodes, k is the embedding dimension of the feature, F i j represents the feature matrix of the trajectory point j of the trajectory T i . The related local graphs G i j t : j = ( V i j t : j , E i j t : j , A i j t : j ) were generated for each normalized trajectory point N P i j t : j in the normalized trajectory N T i , where V i j is the set of the current point N P i j and its neighbor nodes, E i j is the set of edges among the current point and its neighbors, A i j is the local adjacency matrix of the current point, and it concludes the features of both the current point and its neighbor nodes.
Figure 5. Local graph generation module.
GAT was used to calculate the attention weight between the trajectory points and their neighbor nodes and fused the features for the local graphs. The adjacency matrix was used to check whether there is a connection among nodes, and the resource was allocated by calculating the weights of the neighbor nodes in GAT. It can be considered as a spatial attention mechanism to enhance the interpretability of the proposed model. The definition of attention mechanism is shown as Formula (7).
A t t e n t i o n q u e r y , s o u r c e = s i m i l a r i t y q u e r y , k e y m v a l u e m
where s o u r c e is the original information, and it is formed by the k e y v a l u e pairs. A t t e n t i o n q u e r y , s o u r c e represents the information extraction through weight allocation from the s o u r c e under the condition of q u e r y . The aim of the GAT is to learn the relevance among target nodes and their neighbor nodes through the parameter matrix. In this paper, q u e r y is set as the feature vector F i j of the current point N P i j , s o u r c e is set as the feature vectors of all the neighbor nodes of N P i j , k e y m and v a l u e m are, respectively, the m t h neighbor node and its feature vector. The relevance coefficients among every trajectory point and its neighbor nodes were calculated in GAT. The calculation is shown in Formula (8).
e j m = L e a k y R e L U w T [ W F i j | | W F j m ]
where | | represents concatenation, F i j is the feature vector of the point j of the trajectory T i , and F j m is the feature vector of the neighbor node m of the point j . W is a learnable weight parameter. w is a learnable weight of the linear layer and L e a k y R e L U is the activation function. Moreover, all the coefficients were normalized by the function s o f t m a x in GAT to obtain the attention weight, as shown in Formula (9). a j m is the attention weight between node j and its neighbor node m , where N i j denotes the neighborhood of node j in the trajectory T i .
a j m = s o f t m a x e j m = exp e j m n N i j exp e j n
The aggregated and updated feature vector was then calculated, as shown in Formula (10). A multi-head attention mechanism was used to enhance the feature fusion of the neighbor nodes.
N P V i j = | | h = 1 H 1 σ ( m N i j a j m ( h ) W ( h ) F j m )
where H 1 is the number of the heads, σ is the activation function and N P V i j is the graph representation vector. The pseudocode of the local dynamic graph generation module is shown in Algorithm 2.
Algorithm 2: Local dynamic graph generation module
Input   normalized   trajectory   points   N P i j t : j ,   POI   feature   vectors   F i j = V i j P O I
1: G i j t : j ( V i j t : j , E i j t : j , A i j t : j ) for each trajectory point   // generate local graphs
2: Target   node   vector   F i j V i j   and   neighbor   node   vector   F j m V j m // according to local graphs
3: e j m L e a k y R e L U w T [ W F i j | | W F j m ]
4: a j m s o f t m a x e j m
5:weighted sum:
6:for  h H 1  do
7:             N P V i j N P V i j σ ( m N i j a j m ( h ) W ( h ) F j m )
8:end for
Output   graph   representation   vector   N P V i j for each trajectory point

4.5. Trajectory Prediction Layer

In the trajectory prediction layer, trajectory points and corresponding graph representation vectors were input, and the coordinate of the next location were obtained after going through LSTM with a multi-head attention mechanism, as shown in Figure 6.
Figure 6. The research framework of the trajectory prediction layer.
As shown in Figure 6, the trajectory points N P i j t : j and the corresponding graph representation vectors N P V i j t : j were concatenated and input into LSTM, as shown in Formulas (11)–(17).
x j = N P i j , N P V i j
f j = σ W f · h i j 1 , x j + b f
i j = σ W i · h i j 1 , x j + b i
C ~ j = tanh W C · h i j 1 , x j + b C
o j = σ W o · h i j 1 , x j + b o
C j = f j C j 1 + i j C ~ j
h i j = o j tanh C j
where W f , W i , W C and W o are, respectively, the weight of the forget gate, the input gate, the cell state, and the output gate. b f , b i , b C and b o are the respective bias. The trajectory points and corresponding graph representation vectors were concatenated as the input x j , as shown in Formula (11). x j goes through the forget, input, and output gates, and generate the cell state C j , as shown in Formulas (12) and (13). Necessary information was processed by the input gate and the updated information was activated by the function t a n h to obtain the C ~ j , as shown in Formula (14). The current cell state C j is shown in Formula (16), where f j and i j are, respectively, the output of the forget gate and the input gate. C j 1 is the cell state of the previous vector and C ~ j is the activated state updated by the input gate. The hidden state h i j of the current vector is obtained by multiplying the activated cell state C j and the output of the output gate o j , as shown in Formula (17).
A multi-head attention mechanism was used based on LSTM to allocate weight and enhance the interpretability of the proposed model, as shown in Formulas (18) and (19).
A h Q h , K h , V h = s o f t m a x Q h · K h T d V h
M u l t i A t t Q , K , V = A 1 | | | | A H 2 W O
where Q h = W i Q h i j , K h = W i K , h i j t : j , V h = W i V h i j t : j . h i j is the hidden state of the current vector and is considered as the query, h i j t : j denotes the hidden states of the vectors x j t : j and is considered as the key and value, W i Q , W i K , W i V , W O are learnable weight parameters, d is the dimension, and H 2 is the number of the heads.
The predicted result was obtained by adding a multilayer perceptron composed of two full-connected layers, as shown in Formula (20).
Y ^ j = W F 2 R e L U W F 1 h j + b F 1 + b F 2
where Y ^ j is the predicted result. W F 1 , W F 2 , b F 1 , and b F 2 are, respectively, the weights and biases of the two fully connected layers. h j is the output calculated by using the multi-head attention mechanism. Mean Square Error (MSE) was considered as the loss function to calculate the difference between the predicted results and truth, where N is the number of trajectories and K is the length of the trajectory. The calculation is shown in Formula (21).
L = i = 1 N j = 1 K Y ^ j N P i j 2
The pseudocode of the trajectory prediction layer is shown in Algorithm 3.
Algorithm 3: Trajectory prediction layer
Input   trajectory   points   in   normalization   form   N P i j t : j ,   representation   vectors   of   graph   N P V i j t : j
1:LSTM Module:
2:loop
3:             x j { N P i j , N P V i j }
4:             f j , i j , C j ~ , o j calculated by forget gate, input gate and output gate
5:             C j , h i j   cell   state   and   hidden   state   are   calculated   by   f j , i j , C j ~ , o j
6:      Temporal attention mechanism:
7:           Q h W i Q   h i j ,   K h W i K ,   h i j t : j , V h W i V h i j t : j
8:     for  h H 2  do
9:                       A h s o f t m a x Q h · K h T d V h
10:                       h j h j A h W O
11:      end for
12:      MLP:
13:           Y ^ j W F 2 R e L U W F 1 h j + b F 1 + b F 2
14:end loop
Output   coordinate   Y ^ j of the next location

5. Experiments

This section will demonstrate the details of the experiments, including datasets, experiment settings and result analysis. The experiments were conducted to analyze the accuracy and robustness of the proposed model by comparing the performances with the benchmarks. Additionally, the ablation experiment was set to explore the effectiveness of the proposed model by filtering the spatial and temporal attention mechanisms. Moreover, the POI weights that influence the prediction results are visualized on the map to demonstrate the significance of urban functional regions to vehicle trajectories.

5.1. Datasets

Trajectory data: This study utilized Chengdu taxi trajectory data from Didi Gaiya, specifically from October 2018 and August 2014. The dataset includes attributes such as driver ID, order ID, timestamp, longitudes, and latitudes. Despite covering trajectory data for a single month, the dataset comprises over 380,000 trajectories generated by 14,000 vehicles. Notably, 270,000 trajectories are generated on holidays, while 110,000 pertain to working days. In terms of data distribution, the dataset predominantly covers the urban area of Chengdu, spanning from 30.65283 to 30.72649° N and 104.04210 to 104.12907° E. Each trajectory in the dataset has a sampling frequency of 4 s, resulting in a relatively high data density on the city road network. Given the dataset’s substantial volume, comprehensive distribution, and frequent sampling, it provides sufficient data to effectively train the proposed model presented in this paper.
Road Network data: The integration of Point of Interest (POI) and trajectory information forms the basis for constructing a comprehensive global traffic graph. The road network data for Chengdu were obtained from a map website OpenStreetMap (OSM). The node vector data include the road node ID along with its corresponding longitude and latitude. Each piece of road vector data comprises the road ID and a sequence of node coordinates defining its path.
POI data: 11 types of POI data concerning Chengdu from AutoNavi map were sourced through the Chinese Software Developer Network (CSDN) website. The original data format includes information such as POI name, longitude and latitude coordinates, address, district, and POI categories.

5.2. Experimental Settings

The experimental setup in this study is detailed as follows: The experiment employed AMD RYZEN 75,800 h CPU and Nvidia Geforce RTX 3060 GPU. The operating system was Windows 10, with Python as the coding language and PyTorch as the deep learning framework. Furthermore, the parameter configurations are outlined as follows: All models utilized a division of input data into training, validation, and test sets in a ratio of 7:2:1. A batch size of 16 was set, with an initial learning rate of 0.0001. The training process employed Adam as the optimizer. To ensure model convergence, the number of iterations was set to 17,500.

5.2.1. Benchmark Models

There are five baselines used in experiments for the comparation and evaluation of the proposed model, including LSTM [35], GRU [36], BiLSTM [37], AttnLSTM [38], and AttnBiLSTM [39].
LSTM: It is commonly employed in time series prediction. In contrast to traditional RNN models, LSTM addresses a deficiency where conventional models only factor in recent states and struggle with long-term memory. LSTM achieves this by utilizing an internal cell state and a forget gate, allowing it to decide which states to retain or forget. Additionally, it can circumvent certain gradient vanishing issues inherent in traditional RNN models.
GRU: It features only two gates and three fully connected layers in contrast to LSTM, leading to a reduction in computational requirements and lowering the overfitting risk. GRU incorporates an update and a reset gate, allowing it to regulate information output, selectively retaining historical information while discarding irrelevant data.
BiLSTM: Building upon LSTM, BiLSTM incorporates both forward and backward propagation in its input, enabling each timestamp in the input sequence to retain both future and past historical information simultaneously.
AttnLSTM: The attention mechanism is designed to allocate more weight to crucial tasks when computational resources are constrained in this model, acting as a resource allocation scheme to address information overload. Integrating LSTM with the attention mechanism enhances its performance by amplifying the importance of key features or filtering out irrelevant feature information.
AttnBiLSTM: BiLSTM combined with the attention mechanism can fuse information, which involves elevating the weight of significant features and employing bi-directional propagation. This integration enhances the computational efficiency and accuracy of the model by incorporating abundant semantic information.

5.2.2. Evaluation Metrics

Evaluation metrics used in this paper are Mean Absolute Error (MAE), Mean Square Error (MSE), Root Mean Square Error (RMSE), Haversine Distance (HSIN), and Accuracy. The definitions and equations of the five metrics are shown as follows. Y i denotes the ground truth and Y ^ i denotes the predicted location, and the number of trajectory data for evaluation is denoted as N .
Mean Absolute Error (MAE): It denotes the mean of the absolute error between the value that was predicted and the actual value, which can be calculated as in Formula (22).
MAE = 1 N i = 1 N Y i Y ^ i
Mean Square Error (MSE): This metric denotes the square of the variation between the ground truth and forecasted values, which can be calculated as in Formula (23).
MSE = 1 N i = 1 N Y i Y ^ i 2
Root Mean Square Error (RMSE): The standard deviation of the difference between the ground truth and predicted results is RMSE. The model fits better if the RMSE is less. When it comes to high disparities, RMSE has greater penalties than MAE. It can be calculated as Formula (24), where S S E = i = 1 N ( Y i Y ^ i ) 2 denotes the squares of the error of all the samples.
RMSE = M S E = S S E N = 1 N i = 1 N Y i Y ^ i 2
Haversine Distance (HSIN): This metric denotes the distance between the actual and predicted locations. The model with lower distance error is better. It can be calculated as shown in Formula (25), where Y i = l a t i , l o n i is the truth and Y ^ i = l a t ^ i , l o n ^ i is the predicted longitude and latitude.
HSIN = 2 r   a r g s i n s i n 2 l a t i l a t ^ i 2 + cos l a t i cos l a t ^ i s i n 2 l o n i l o n ^ i 2
Accuracy: It indicates the proportion of the accurately predicted output in the total output. The higher the accuracy rate, the better the model training effect. It can be calculated as shown in Formula (26), and T P denotes the correctly predicted results, N P denotes the incorrectly predicted results, and c o u n t ( · ) denotes the counting number.
Accuracy = c o u n t T P c o u n t T P + c o u n t N P  

5.3. Result Analysis

The accuracy experiment and the robustness experiment are discussed in this paper. The LDGST-LSTM model is compared with the benchmark models to explore the performance in terms of accuracy and robustness. The ablation experiment is discussed to verify the spatial and temporal attention mechanisms of the proposed model to enhance interpretability. Moreover, the POI weights calculated using the temporal attention mechanism are visualized on the map. The POI is considered as the feature of the vehicle trajectory, and the influence of the POI on vehicle position prediction is revealed.

5.3.1. Accuracy Experiment

All models were trained, and the accuracy comparison results on the training set are shown in Figure 7. It can be seen that on both holidays and working days, the accuracy of LDGST-LSTM was significantly higher than the benchmarks.
Figure 7. Accuracy comparison of different models on training set: (a) holidays in October 2018, (b) working days in October 2018, (c) weekends in August 2014, and (d) working days in August 2014.
As shown in Figure 7a,b, the accuracy of the proposed model on both holidays and working days was nearly 4.5 times higher than the benchmarks in the dataset of October 2018. As shown in Figure 7c,d, the accuracy of the proposed model on both weekends and working days was nearly 3.5–4.5 times higher than the benchmarks in the dataset of August 2014. Therefore, the accuracy of location prediction can be effectively improved by considering POI knowledge and combining GAT and temporal attention mechanisms.
The accuracy comparison results of different models are shown in Table 1.
Table 1. Accuracy comparison of different models on holidays in October 2018, working days in October 2018, weekends in August 2014, and working days in August 2014.
In the October 2018 dataset, the accuracy of LDGST-LSTM was 21% higher compared to LSTM during holidays and 61% higher compared to Attn-LSTM during weekdays. In the August 2014 dataset, LDGST-LSTM improved by 64% compared to Attn-BiLSTM on weekends. Therefore, compared to the benchmarks, the model proposed in this paper can improve performance by selecting appropriate POI categories on weekdays and holidays, respectively. More importantly, integrating POI knowledge and combining spatial and temporal attention mechanisms can greatly improve the accuracy of vehicle trajectory location prediction.

5.3.2. Robustness Experiment

The results of the robustness experiment are discussed in this section. The convergence speed and the metrics after convergence of the proposed model and benchmarks are compared to analyze the robustness. The results of convergence speed are shown in Figure 8.
Figure 8. Convergence speed comparison of different models: (a) holidays in October 2018, (b) working days in October 2018, (c) weekends in August 2014, and (d) working days in August 2014.
As shown in Figure 8a,b, the loss of LDGST-LSTM drops slowly compared with the benchmark models before the 200th iteration on the dataset of October 2018. The convergence speed of the proposed model becomes faster from the 250th to the 300th iterations, and it converges at around the 350th iteration. As shown in Figure 8c,d, the loss of LDGST-LSTM drops slowly compared with the benchmark models before the 200th iterations on the dataset of August 2014. Additionally, the convergence speed of the proposed model becomes faster from the 250th to the 350th iterations and it converges at around the 450th iteration.
The performance of MAE, MSE, RMSE and HISN of different models are shown in Table 2, Table 3, Table 4 and Table 5. As shown in Table 2 and Table 3, the performance of the proposed model on evaluation metrics was the best compared with benchmarks when using the dataset of October 2018. The MAE, MSE, RMSE, and HSIN of the proposed model were, respectively, 7.73%, 28.57%, 11.11%, and 15.72% lower than LSTM, which was the most robust among the baselines in the holidays. The MAE, MSE, RMSE, and HSIN of the proposed model were, respectively, 6.19%, 33.33%, 27.27%, and 41.79% lower than GRU, which was the most robust among the benchmarks in the working days. As shown in Table 4 and Table 5, the performance of the proposed model on evaluation metrics were still the best when using the dataset of Chengdu August 2014. The four metrics of the proposed model were, respectively, 6.09%, 33.33%, 15.64%, and 41.66% lower than GRU in the weekend. The MAE, MSE, RMSE, and HSIN of the proposed model were, respectively, 4.88%, 28.57%, 18.55%, and 15.99% lower than LSTM in the working days. Therefore, it can be seen that the robustness of the proposed model were the best compared with all the benchmark models.
Table 2. Performance comparison of different models on holidays in October 2018.
Table 3. Performance comparison of different models on working days in October 2018.
Table 4. Performance comparison of different models on weekends in August 2014.
Table 5. Performance comparison of different models on working days in August 2014.

5.3.3. Ablation Experiment

The ablation experiment is discussed to analyze the effect of major components in the proposed model by filtering GAT and the temporal attention mechanism. In addition to the proposed model, three ablation models are analyzed, including (1) Local Dynamic Graph Convolutional Network–Long Short-Term Memory (LDGCN-LSTM), (2) Local Dynamic Graph Convolutional Network–Temporal Attention Long Short-Term Memory (LDGCN-TAttnLSTM), and (3) Local Dynamic Graph Attention–Long Short-Term Memory (LDGAT-LSTM).
As shown in Figure 9, the accuracy of LDGST-LSTM is the highest among the other three ablation models in both datasets. The accuracy of LDGST-LSTM is almost the same as the accuracy of LDGAT-LSTM. It may be the reason that, during holidays, the intention of the taxis is more focused on some functional regions, and the importance of spatial overweighs the temporal features.
Figure 9. Accuracy comparison of different ablation modules: (a) holidays in October 2018, (b) working days in October 2018, (c) weekends in August 2014, and (d) working days in August 2014.
The ablation results are shown in Table 6, Table 7, Table 8 and Table 9. As shown in Table 6 and Table 7, the evaluation metrics of LDGST-LSTM were the best among the ablation models when using a dataset in 2018. The MAE, MSE, RMSE, and HSIN of the proposed model were, respectively, 9.91%, 28.57%, 12.97%, and 15.46% lower, and the accuracy was 4.17% higher than LDGAT-LSTM, which means it performed the best among ablation models in the holidays. The MAE, MSE, RMSE, and HSIN of the proposed model were, respectively, 3.9%, 14.29%, 22.18%, and 20.12% lower, and the accuracy was 40% higher than LDGAT-LSTM on working days. As shown in Table 8 and Table 9, the performances of LDGST-LSTM were also the best in the 2014 dataset. The four metrics of LDGST-LSTM were, respectively, 25.40%, 33.33%, 18.31% and 18.52% lower, and the accuracy was 7.69% higher than LDGAT-LSTM; therefore, it performed the best among ablation models in the weekends. The MAE, MSE, RMSE and HSIN of LDGST-LSTM were, respectively, 10.55%, 28.57%, 11.81% and 10.42% lower, and the accuracy was 3.64% higher than LDGAT-LSTM on working days. In conclusion, the proposed model performed the best compared with other ablation models. Therefore, the combination of GAT and temporal attention mechanism can enhance the interpretability and improve the accuracy and robustness of the model.
Table 6. Ablation results of the proposed model on holidays in October 2018.
Table 7. Ablation results of the proposed model on working days in October 2018.
Table 8. Ablation results of the proposed model on weekends in August 2014.
Table 9. Ablation results of the proposed model on working days in August 2014.

5.3.4. POI Weights Visualization

The predicted coordinates of the next location, along with their corresponding weight, can be calculated by using the proposed model. The visualization of POI weights is realized through the nuclear density analysis in the ArcGIS.
The visualization of POI weight has a positive significance in terms of vehicle trajectory planning, traffic optimization, and vehicle location prediction. As shown in Figure 10, there are some regions of POI information that affect vehicle location prediction. For example, this is shown in the western and southern regions in Figure 10a,c, and the right and the top sides in Figure 10b,d. It can be seen that on holidays and weekends, POI in the western and southern regions can be considered important information that influences vehicle trajectories. It may denote that these regions are close to the center of the city and have high traffic flow on holidays and weekends. Therefore, it is important to plan the driving path in these regions on holidays and weekends. Moreover, POI regions that influence the location prediction are more dispersed on working days, and the trajectory decision can be more flexible.
Figure 10. Visualization of the POI weights: (a) holidays in October 2018, (b) working days in October 2018, (c) weekends in August 2014, and (d) working days in August 2014.

6. Conclusions and Future Work

A Local Dynamic Graph Spatiotemporal–Long Short-Term Memory (LDGST-LSTM) model was proposed in this paper to predict the next location of vehicle trajectory. The data conversion layer, POI global knowledge extraction layer, local dynamic graph generation module, and trajectory prediction layer were major components of the proposed model. Raw taxi trajectory and POI semantic information were first matched to the road network through the use of a map-matching algorithm and proximity algorithm in the data conversion layer. Then, the representation vectors of POI were learned through the use of the TransE algorithm by constructing a knowledge graph in the POI global knowledge extraction layer. Based on the global knowledge graph, a local graph related to each trajectory point was generated, and the graph representation vector was captured through GAT in the local dynamic graph generation module. Finally, trajectory points with the related graph representation vectors were input into LSTM with a multi-head attention mechanism in the trajectory prediction layer to predict the next location.
However, there are limitations in this paper. Firstly, exploring accurate trajectory recovery can be a potential avenue for future research due to non-uniform GPS sampling and the presence of GPS signal shielding areas. Moreover, only POI knowledge is considered as the external feature in this paper. The current approach heavily relies on Points of Interest for contextual information. However, there may be other relevant contextual factors, such as real-time weather conditions and road construction, which could influence trajectory prediction. Exploring these factors can provide a more comprehensive understanding of the environment and enhance the accuracy and robustness of the model. Additionally, knowledge graphs are currently static and do not dynamically update in real time. This limitation may impact the model’s adaptability to changes in the environment, such as sudden road closures or new construction. Therefore, implementing a mechanism for online updating of knowledge graphs based on real-time data can be a further enhancement. This ensures that the model adapts dynamically to changes in the environment, improving its ability to handle dynamic scenarios. Moreover, scenarios in this paper were the macro traffic roads and specific traffic scenarios such as complex intersections and traffic light timing could be a challenge for trajectory prediction. The current model may struggle to accurately predict trajectories in scenarios where multiple roads converge or diverge. Therefore, further research could focus on developing specialized mechanisms for handling complex intersections. This may involve incorporating advanced spatial attention mechanisms or considering the historical behavior of vehicles at such intersections. In recent years, the rapid development of connected automated vehicles has significantly transformed the landscape of transportation. The emergence of shared data on the road presents a unique opportunity to enhance various aspects of vehicular technologies, including trajectory prediction. Leveraging data from connected vehicles, such as cooperative perception using 3D LiDAR [40], can indeed contribute valuable insights to trajectory dynamics. The potential of shared information from connected vehicles to revolutionize the accuracy and reliability of trajectory predictions can be another new direction for research and innovation. Furthermore, extending the analysis in terms of dataset duration and geographic coverage is another further research direction. Assessing the model’s performance over a more extended period should be considered, such as capturing potential variations in different seasons or temporal patterns. The effectiveness of the model across a broader geographic area should also be considered, incorporating diverse scenarios to test its adaptability in different environments.

Author Contributions

Funding acquisition, J.C.; methodology, J.C.; project administration, J.C.; software, Q.F. and D.F.; supervision, J.C.; validation, D.F.; visualization, Q.F. and D.F.; writing—original draft, J.C.; writing—review and editing, Q.F. and D.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61104166.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are applied for download from Didi Gaia and are not publicly available due to Didi’s requirements. The application page is https://outreach.didichuxing.com/app-outreach/TBRP, accessed on 20 November 2023.

Conflicts of Interest

The authors declare there are no conflicts of interest regarding the publication of this paper. The authors have no financial and personal relationships with other people or organizations that could inappropriately influence our work.

Abbreviations

LDGST-LSTMLocal Dynamic Graph Spatiotemporal–Long Short-Term Memory
latlatitude
lnglongitude
OSMOpenStreetMap
POIPoint of interest
TransETranslating Embedding
GATGraph Attention Network
KGKnowledge Graph
LSTMLong Short-Term Memory model
T-AttnTemporal Attention mechanism

References

  1. Guo, L. Research and Application of Location Prediction Algorithm Based on Deep Learning. Ph.D. Thesis, Lanzhou University, Lanzhou, China, 2018. [Google Scholar]
  2. Havyarimana, V.; Hanyurwimfura, D.; Nsengiyumva, P.; Xiao, Z. A novel hybrid approach based-SRG model for vehicle position prediction in multi-GPS outage conditions. Inf. Fusion 2018, 41, 1–8. [Google Scholar] [CrossRef]
  3. Wu, Y.; Hu, Q.; Wu, X. Motor vehicle trajectory prediction model in the context of the Internet of Vehicles. J. Southeast Univ. (Nat. Sci. Ed.) 2022, 52, 1199–1208. [Google Scholar]
  4. Li, L.; Xu, Z. Review of the research on the motion planning methods of intelligent networked vehicles. J. China Highw. Transp. 2019, 32, 20–33. [Google Scholar]
  5. Liao, J. Research and Application of Vehicle Position Prediction Algorithm Based on INS/GPS. Ph.D. Thesis, Hunan University, Hunan, China, 2016. [Google Scholar]
  6. Wang, K.; Wang, Y.; Deng, X. Review of the impact of uncertainty on vehicle trajectory prediction. Autom. Technol. 2022, 7, 1–14. [Google Scholar]
  7. Wang, L. Trajectory Destination Prediction Based on Traffic Knowledge Map. Ph.D. Thesis, Dalian University of Technology, Dalian, China, 2021. [Google Scholar]
  8. Guo, H.; Meng, Q.; Zhao, X. Map-enhanced generative adversarial trajectory prediction method for automated vehicles. Inf. Sci. 2023, 622, 1033–1049. [Google Scholar] [CrossRef]
  9. Xu, H.; Yu, J.; Yuan, S. Research on taxi parking location selection algorithm based on POI. High-Tech Commun. 2021, 31, 1154–1163. [Google Scholar]
  10. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio’, P.; Bengio, Y. Graph Attention Networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  11. Li, L.; Ping, Z.; Zhu, J. Space-time information fusion vehicle trajectory prediction for group driving scenarios. J. Transp. Eng. 2022, 22, 104–114. [Google Scholar]
  12. Su, J.; Jin, Z.; Ren, J.; Yang, J.; Liu, Y. GDFormer: A Graph Diffusing Attention based approach for Traffic Flow Prediction. Pattern Recognit. Lett. 2022, 156, 126–132. [Google Scholar] [CrossRef]
  13. Ali, A.; Zhu, Y.; Zakarya, M. Exploiting dynamic spatio-temporal correlations for citywide traffic flow prediction using attention based neural networks. Inf. Sci. 2021, 577, 852–870. [Google Scholar] [CrossRef]
  14. Fan, H. Research and Implementation of Vehicle Motion Tracking Technology Based on Internet of Vehicles. Ph.D. Thesis, Beijing University of Posts and Telecommunications, Beijing, China, 2017. [Google Scholar]
  15. Hui, F.; Wei, C.; Shangguan, W.; Ando, R.; Fang, S. Deep encoder-decoder-NN: A deep learning-based autonomous vehicle trajectory prediction and correction model. Phys. A Stat. Mech. Its Appl. 2022, 593, 126869. [Google Scholar] [CrossRef]
  16. Kalatian, A.; Farooq, B. A context-aware pedestrian trajectory prediction framework for automated vehicles. Transp. Res. C Emerg. Technol. 2022, 134, 103453. [Google Scholar] [CrossRef]
  17. An, J.; Liu, W.; Liu, Q.; Guo, L.; Ren, P.; Li, T. DGInet: Dynamic graph and interaction-aware convolutional network for vehicle trajectory prediction. Neural Netw. 2022, 151, 336–348. [Google Scholar] [CrossRef]
  18. Zambrano-Martinez, J.L.; Calafate, C.T.; Soler, D.; Cano, J.C.; Manzoni, P. Modeling and characterization of traffic flows in urban environments. Sensors 2018, 18, 2020. [Google Scholar] [CrossRef]
  19. Zhang, X.; Onieva, E.; Perallos, A.; Osaba, E.; Lee, V. Hierarchical fuzzy rule-based system optimized with genetic algorithms for short term traffic congestion prediction. Transp. Res. C Emerg. Technol. 2014, 43, 127–142. [Google Scholar] [CrossRef]
  20. Yang, D.; He, T.; Wang, H. Research progress in graph embedding learning for knowledge map. J. Softw. 2022, 33, 21. [Google Scholar]
  21. Xia, Y.; Lan, M.; Chen, X. Overview of interpretable knowledge map reasoning methods. J. Netw. Inf. Secur. 2022, 8, 1–25. [Google Scholar]
  22. Zhang, Z.; Qian, Y.; Xing, Y. Overview of TransE-based representation learning methods. J. Comput. Appl. Res. 2021, 3, 656–663. [Google Scholar]
  23. Chen, W.; Wen, Y.; Zhang, X. An improved TransE-based knowledge map representation method. Comput. Eng. 2020, 46, 8. [Google Scholar]
  24. Zheng, D.; Song, X.; Ma, C.; Tan, Z.; Ye, Z.; Dong, J.; Xiong, H.; Zhang, Z.; Karypis, G. DGL-KE: Training Knowledge Graph Embeddings at Scale. In Proceedings of the SIGIR ′20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020. [Google Scholar]
  25. Ji, Q.; Jin, J. Reasoning Traffic Pattern Knowledge Graph in Predicting Real-Time Traffic Congestion Propagation. IFAC-PapersOnLine 2020, 53, 578–581. [Google Scholar] [CrossRef]
  26. Wang, X.; Lyu, S.; Wang, X.; Wu, X.; Chen, H. Temporal knowledge graph embedding via sparse transfer matrix. Inf. Sci. 2022, 623, 56–69. [Google Scholar] [CrossRef]
  27. Wang, C.; Tian, R.; Hu, J.; Ma, Z. A trend graph attention network for traffic prediction. Inf. Sci. 2023, 623, 275–292. [Google Scholar] [CrossRef]
  28. Wang, B.; Wang, J. ST-MGAT:Spatio-temporal multi-head graph attention network for Traffic prediction. Phys. A-Stat. Mech. Its Appl. 2022, 603, 127762. [Google Scholar] [CrossRef]
  29. Wang, T.; Ni, S.; Qin, T.; Cao, D. TransGAT: A dynamic graph attention residual networks for traffic flow forecasting. Sustain. Comput. Inform. Syst. 2022, 36, 100779. [Google Scholar] [CrossRef]
  30. Cai, K.; Shen, Z.; Luo, X.; Li, Y. Temporal attention aware dual-graph convolution network for air traffic flow prediction. J. Air Transp. Manag. 2023, 106, 102301. [Google Scholar] [CrossRef]
  31. Yan, X.; Gan, X.; Wang, R.; Qin, T. Self-attention eidetic 3D-LSTM: Video prediction models for traffic flow forecasting. Neurocomputing 2022, 509, 167–176. [Google Scholar] [CrossRef]
  32. Chen, L.; Shi, P.; Li, G.; Qi, T. Traffic flow prediction using multi-view graph convolution and masked attention mechanism. Comput. Commun. 2022, 194, 446–457. [Google Scholar] [CrossRef]
  33. Wang, K.; Ma, C.; Qiao, Y.; Lu, X.; Hao, W.; Dong, S. A hybrid deep learning model with 1DCNN-LSTM-Attention networks for short-term traffic flow prediction. Phys. A-Stat. Mech. Its Appl. 2021, 583, 126293. [Google Scholar] [CrossRef]
  34. Lou, Y.; Zhang, C.; Zheng, Y.; Xie, X.; Wang, W.; Huang, Y. Map-matching for low-sampling-rate GPS trajectories. In Proceedings of the GIS ′09: Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 4–6 November 2009. [Google Scholar]
  35. Cai, Y. Research on Vehicle Trajectory Prediction Based on RNN-LSTM Network. Ph.D. Thesis, Jilin University, Jilin, China, 2021. [Google Scholar]
  36. Zhang, H.; Huang, C.; Xuan, Y. Real time prediction of air combat flight trajectory using gated cycle unit. Syst. Eng. Electron. Technol. 2020, 42, 7. [Google Scholar]
  37. Guo, Y.; Zhang, R.; Chen, Y. Vehicle trajectory prediction based on potential features of observation data and bidirectional long short-term memory network. Autom. Technol. 2022, 3, 21–27. [Google Scholar]
  38. Liu, C.; Liang, J. Vehicle trajectory prediction based on attention mechanism. J. Zhejiang Univ. (Eng. Ed.) 2020, 54, 8. [Google Scholar]
  39. Guan, D. Research on Modeling and Prediction of Vehicle Moving Trajectory in the Internet of Vehicles. Ph.D. Thesis, Beijing University of Posts and Telecommunications, Beijing, China, 2020. [Google Scholar]
  40. Meng, Z.; Xia, X.; Xu, R.; Liu, W.; Ma, J. HYDRO-3D: Hybrid Object Detection and Tracking for Cooperative Perception Using 3D LiDAR. IEEE Trans. Intell. Veh. 2023, 8, 4069–4080. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.