Next Article in Journal
Analysis of an Impact of Inertia Parameter in Active Disturbance Rejection Control Structures
Next Article in Special Issue
An Application of Reinforced Learning-Based Dynamic Pricing for Improvement of Ridesharing Platform Service in Seoul
Previous Article in Journal
Enhancing Security on IoT Devices via Machine Learning on Conditional Power Dissipation
Previous Article in Special Issue
Bottleneck Based Gridlock Prediction in an Urban Road Network Using Long Short-Term Memory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Context-Aware Link Embedding with Reachability and Flow Centrality Analysis for Accurate Speed Prediction for Large-Scale Traffic Networks

1
Department of Artificial Intelligence, Big Data, Hongik University, 94 Wausan-ro, Sangsu-dong, Mapo-gu, Seoul 04068, Korea
2
Department of Computer Engineering, Hongik University, 94 Wausan-ro, Sangsu-dong, Mapo-gu, Seoul 04068, Korea
3
Neouly Incorporated, 94 Wausan-ro, Sangsu-dong, Mapo-gu, Seoul 04068, Korea
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(11), 1800; https://doi.org/10.3390/electronics9111800
Submission received: 26 September 2020 / Revised: 26 October 2020 / Accepted: 28 October 2020 / Published: 29 October 2020
(This article belongs to the Special Issue AI-Based Transportation Planning and Operation)

Abstract

:
This paper presents a novel method for predicting the traffic speed of the links on large-scale traffic networks. We first analyze how traffic flows in and out of every link through the lowest cost reachable paths. We aggregate the traffic flow conditions of the links on every hop of the inbound and outbound reachable paths to represent the traffic flow dynamics. We compute a new measure called traffic flow centrality (i.e., the Z value) for every link to capture the inherently complex mechanism of the traffic links influencing each other in terms of traffic speed. We combine the features regarding the traffic flow centrality with the external conditions around the links, such as climate and time of day information. We model how these features change over time with recurrent neural networks and infer traffic speed at the subsequent time windows. Our feature representation of the traffic flow for every link remains invariant even when the traffic network changes. Furthermore, we can handle traffic networks with thousands of links. The experiments with the traffic networks in the Seoul metropolitan area in South Korea reveal that our unique ways of embedding the comprehensive spatio-temporal features of links outperform existing solutions.

1. Introduction

Accurate prediction of speed on traffic networks helps improve traffic management strategies and generate efficient routing plans. However, precisely estimating the traffic speed in advance has been a non-trivial task since various factors determine traffic flows in many ways.
Various approaches have been used for traffic speed prediction using statistical methods [1,2,3,4,5,6,7] and machine learning with neural networks with deep hidden layers [8,9,10,11,12,13,14,15,16,17,18]. However, the existing solutions are mostly limited to estimating traffic speed for either a single road link or a small-scale sub-network with only a handful of traffic links (e.g., a crossroad). In reality, a road system is a large-scale connected graph with the traffic links affecting each other’s traffic flow over time in a more complicated fashion that cannot be explained easily with a simple statistical model. The works without a broader view of the traffic links may not adequately unravel the hidden, but critical speed prediction factors.
In this paper, we adapt the node embedding techniques introduced by Hamilton et al. [19]. We represent the relationship between links with a feature vector whose size is invariant even when certain changes are made to the traffic networks’ structure, such as the addition or deletion of roads. We analyze how the traffic flows in and out of a link through reachable multi-hop paths computed with the Floyd–Warshall algorithm [20]. How the links impact each other given the traffic flow analysis is quantified through a novel metric we refer to as Traffic Flow Centrality (TFC). We combine the data related to TFC with the external conditions around each link, such as climate and time information (e.g., time of day, the indication of holidays). We use a recurrent neural network algorithm to correlate between the composite feature’s temporal transition and the traffic speed of each link. Our method captures the traffic flow dynamics through sub-networks around every link without embedding the entire adjacency matrix. Therefore, our method is much more space efficient, and it can handle large-scale traffic networks with thousands of links. We also do not face the problem of considering irrelevant information such as the hollow points around traffic networks when they are represented either with a sparse adjacency matrix or raster graphics [10,18]. Our solution assesses the impact of the remote links in addition to the adjacent neighbors on traffic flows. Hence, with the broader contextual view and the exclusion of unnecessary information, we expect to outperform the works that based their speed prediction only on the limited view of adjacent links.
This paper is structured as follows: In Section 2, we first review the related work. In Section 3, we introduce the method for inferring traffic speed given the temporal transitions of links’ composite contextual features that are modeled with a novel link embedding technique. In Section 4, we benchmark the performance of our approach against existing works, and we conclude in Section 5.

2. Related Works

In this section, we put our work in the context of various related research works on traffic speed prediction. Recently, artificial neural networks with deep hidden layers have gained popularity, as they are effective at modeling the non-linear traffic speed dynamics. Some of the notable works have employed FNNs (Fuzzy Neural Networks) [8,9], DNNs (Deep Neural Networks) [11,21], RNNs (Recurrent Neural Networks) [13,14], DBNs (Deep Belief Networks) [12,15], and the IBCM-DL (Improved Bayesian Combination Model with Deep Learning) [17] models. These works have shown more accurate speed prediction than the approaches that are based on classic statistical methods such as ARIMA (Auto-Regressive Integrated Moving Average) models [1,2,3], SVR (Support Vector Regression) [4,5], and K-NN (K-Nearest Neighbor) [6,7].
When the models are obtained by learning the pattern on a single specific link [1,2,3,13,14,15,17], the distinct features of other links may not be adequately accounted for. Thus, modeling the traffic speed pattern per individual link was discussed in [22]. Nonetheless, the work by Kim et al. [22] still did not reflect the substructure of the traffic network around each link.
The works presented in [10,16,23,24] extracted spatial features from a visual representation of traffic networks. In particular, Zheng et al. [23] used a two-dimensional traffic flow that is embedded in Convolutional Neural Networks (CNN). They also used a Long Short-term Memory (LSTM) [25] algorithm to model long-term historical data. Similarly, Du et al. [16] represented the passenger flows among different traffic lines in a transportation network into multi-channel matrices with deep irregular convolutional networks. Guo et al. [10] analyzed the congestion propagation patterns based on the traffic observations recorded at fixed intervals in time and fixed locations in space. They represented the traffic observation in raster data. Then, all the raster data were fed into a 3D convolutional neural network to model the spatial information. They used a 3 × 3 convolutional kernel that includes the hollow points where no road lies and no traffic flows. The hollow points in the convolutional kernel may accidentally reflect stale traffic flows. Instead, Du et al. [16] used an irregular convolution kernel that refers to the traffic flow values from the adjacent traffic lines to fill in the values of the hollow points. These methods commonly exploit the CNN architecture that effectively models visual imagery [26,27,28]. Furthermore, a family of RNN algorithms such as GRU [29] and LSTM [25] was used so that repetitive temporal patterns can be discovered. However, these works did not capture the relation between the flow points on the two-dimensional spaces such as junctions, crossroads, and overpasses. Therefore, these models are susceptible to errors by correlating between irrelevant traffic flows. For instance, they may confuse overpasses and overlapping roads as crossroads. Furthermore, these works did not address the impact of the external conditions on the traffic flow. Learning the correlation between traffic flows and weather parameters was useful for flow prediction, as presented in [12,30]. However, they overlooked the impact of the substructure around traffic links on traffic flows.
More recently, modeling the traffic flow based on the graph representation of the traffic networks has emerged. The works presented in [18,31,32,33] combined GNNs (Graph Neural Networks) [34,35] and RNNs to capture the temporal flow transition patterns given adjacency matrices that explicitly reflect the complex interconnections. These methods do not have to unnecessarily deal with the information irrelevant to the traffic flow, such as the hollow points in the visual traffic networks that Du et al. [16] had to consider forcefully.
ST-TrafficNet [33] used Caltrans PeMS (Performance Measurement System) data from around 20 links between intersection points and predicted traffic speed with stacked LSTM using a spatially-aware multi-diffusion convolution block. This PeMS data were from 350 loop detectors at 5 min intervals from 1 January 2017 to 31 May 2017. This model reflects spatial influence through multi-diffusion convolution with forward, backward, and attentive channels.
TGC-LSTM was used by Cui et al. [18] to predict the traffic speed on four connected freeways in the Greater Seattle Area. They used publicly available traffic state data from 323 sensor stations over the entirety of 2015 at 5 min intervals. With their model, traffic speed was predicted with the RMSE (Root Mean Squared Error) as low as 2.1. However, the GNN architecture has to be restructured whenever the traffic networks undergo some changes. This is because TGC-LSTM uses the entire adjacency matrix as an input to the GNN instead of embedding the features of individual traffic links. Upon any change to the traffic networks, we have to re-train from scratch with the newly updated GNN architecture. Furthermore, since TGC-LSTM uses a very large adjacency matrix, both the time and space complexity of modeling the network structure becomes high. However, more importantly, the larger the adjacency matrix is, the more sparse it becomes. Therefore, TGC-LSTM still faces the problem of incorporating unnecessary data such as the hollow points captured in a regular convolution kernel, as discussed in [10]. The shortcomings of these GNN-based approaches motivated us to devise a new method for embedding the characteristics of the traffic network.
We adapt the node embedding techniques introduced by Hamilton et al. [19]. We represent the relationship between links on the traffic network with a feature vector whose size is invariant even when any part of the network structure changes. We analyze how the traffic routes through a link via reachable lowest cost multi-hop paths that are computed with the Floyd–Warshall algorithm [20]. We compute every link’s relative impact on other links based on its inbound/outbound traffic flow patterns and its neighbors’ collective conditions. We refer to the relative cross-link impact value as Traffic Flow Centrality (TFC). We combine the features related to TFC with the external conditions around each link, such as climate and time information. We use a recurrent neural network algorithm to learn how such a composite feature change over time determines the traffic speed of each link.
Our method does not involve the process of embedding the entire adjacency matrix. Therefore, our solution is more space efficient and can easily handle large-scale traffic networks with thousands of links. Furthermore, it avoids incorporating irrelevant information such as the hollow points that can be present in traffic networks when they are represented with a sparse adjacency matrix or raster graphics [10,18]. Our solution considers the conditions of the remote links beyond the adjacent neighbors. By ruling out irrelevant information and having the broader contextual view, we expect to outperform the works that base their speed prediction myopically on the conditions of the adjacent links.
The advantage of our work, named TFC-LSTM, is summarized in Table 1, which shows the comparison between existing related works we have discussed so far. The “Traffic Network Structure” refers to the usage of the abstract representation of interconnections between links. The “Surrounding Conditions” refer to the consideration of external situations around links such as climate and time information. The “Traffic Flow Reachability Analysis” refers to the process of analyzing the pattern of traffic flowing in and out of links through reachable paths. The “Centrality Analysis” refers to the usage of the link’s relative influence on others. The “Chains of Neighbors” column indicates the consideration of remote neighbors besides the adjacent ones when capturing the substructure around a link.

3. Methodology

We outline the overall procedure for predicting the link’s speed on the traffic networks, as shown in Figure 1. Given raw data such as adjacency, distance, and speed in a data tensor, we compute reachability information such as hop count, distance, and cost of traffic flowing between a pair of source and destination links through a chain of neighboring links. Paired with other external features such as weather conditions, time of day, and day of the week, we expect that encoding the dynamics of the temporal traffic flows on the neighboring links would significantly improve the prediction of links’ speed. How we generate the embedding of complex context-aware spatio-temporal features is explained in greater detail in Section 3.1. We model the correlation between the input feature and each link speed with a Long Short-Term Memory (LSTM) algorithm. We use 512 perceptrons in the hidden layer, ReLU for the activation function [36,37], and Adam for the optimizer [38,39].

3.1. Link Embedding

Feature representation of a link is done in the following three steps: First, we abstract the traffic system as a link adjacency matrix A. A traffic network illustrated in Figure 2A is abstracted further as a directed graph structure, as shown in Figure 2B, with every directional link being uniquely identified. Figure 3A shows the link adjacency matrix for the sample traffic system in Figure 2A. A l i l j denotes whether traffic can flow from link l i to link l j . For instance, according to the example in Figure 2A, A l 2 l 3 = 1 since traffic can flow from l 2 to l 3 . Otherwise, A l 2 l 3 equals 0. We exclude the links through which no traffic can reach any other parts of the traffic system. For instance, l 1 and l 16 are disconnected from the rest of the traffic system. Figure 3B illustrates a matrix of the distance between every pair of adjacent links. Given A l i l j , Figure 3C shows the speed of the traffic flowing from l i to adjacent link l j .
Second, we conduct a traffic flow reachability analysis. For every link l, we extract all inbound multi-hop paths through which traffic traverses towards l. Oppositely, we compute all outbound multi-hop paths through which the traffic originated from l diffuses. Given A l i l j , we run the Floyd–Warshall algorithm [20] to obtain the minimum time it takes to travel from every source link l i to all other destination links l j , as shown in Figure 3F. We generate a matrix of hop counts and minimum distances from every source link l i to all other destination link l j , as shown in Figure 3D,E, respectively. For instance, the traffic on l 5 (at Hop 3) takes l 6 (at Hop 2) to reach l 9 at time t 2 in Figure 4B as opposed to taking l 4 (at Hop 1) at time t 1 in Figure 4A due to the heavy congestion on l 4 . After we obtain the paths, we discard the links through which traffic cannot reach the source link before the beginning of the next time window. For example, if the traffic on l 9 is too slow to reach the remote link l 15 by the end of the current time window, then we discard l 15 from the reachable outbound path from l 9 , as shown in Figure 5.
P i n = i R R C i n , P o u t = i R R C o u t
ρ F i n = i R R v i n i f i n i d i n i , ρ F o u t = i R R v o u t i f o u t i d o u t i
L x i n = ρ F i n i R R ρ F i n , L x o u t = ρ F o u t i R R ρ F o u t
Z i n = ln ( n N i n N i n L n i n e h o p ) , Z o u t = ln ( n N o u t N o u t L n o u t e h o p )
Z i n = Z n i n n N N Z n i n , Z o u t = Z n o u t n N N Z n o u t
| Z i n Z i n | < ϵ , | Z o u t Z o u t | < ϵ
In the final step, we compute the Z value for every link, which we refer to as the traffic flow centrality. Given the matrix of minimum time to travel from source link l i to destination link l j (Figure 6A), we count the number of inbound and outbound reachable paths (R), ( P i n and P o u t ) for every link using Equation (1), as illustrated in Figure 6B. We also compute the speed and distance between every pair of adjacent links on the reachable paths. Equation (2) defines ( ρ F ) as the weighted sum of the average traffic speed (v) on every link i on the multi-hop reachable paths. The weight of each intermediate link i is the inverse of the product between the fanout f and the distance d to i, where f specifically represents the number of alternate paths on a junction. The weight represents the impact on a given link the current traffic is either destined to or stemmed from. We expect the impact to be sensitive to the f and d values. For instance, we can capture the circumstance where the traffic on links with higher f and d values are less likely to move towards a target link l than the traffic on a link with lower f and d values. This is because traffic on the link with higher f and d values is more likely to veer away by taking different turns on the junctions or halt the transition at any point on the path. As an example, computing the inbound and the outbound ρ F values for the link l 9 is illustrated in Figure 6C,D. The aggregation steps above are to account for the traffic flow dynamics around every link.
Note that the links adjacent to each other are also inter-dependent on each other with regards to computing their Z values. Due to the inter-dependence, we have to compute Equations (4) and (5) iteratively until the condition on Equation (6) holds. We obtain a converged Z value when Equation (6) is satisfied. We re-scale the ρ F values to L x through normalization as specified in Equation (3). Suppose we have a set of adjacent inbound and outbound neighbors N i n and N o u t , respectively, for a given link. Then, we take the sums of L x / e h o p of every neighbor in N i n and N o u t . The division by e h o p reflects that the impact of a link’s L x value on its immediate neighbor is inversely proportional to the hop distance between them. We take the natural logarithmic function on the sums as shown in Equation (4) and normalize the value as defined in Equation (5). Whenever Equation (4) repeats, L n is substituted by Z obtained in the previous iteration. The ϵ value in Equation (6) is for determining the converged Z. We empirically set the ϵ value that leads to the most accurate link speed prediction. The iterative computation of the Z value for every link in a traffic network is illustrated in Figure 7.
The traffic flow centrality is to capture the inter-link relationship, and we expect it to be one of the key factors for accurately predicting traffic speed. Intuitively, it is highly probable that a link would experience traffic swarming in and causing congestion, especially when a large portion of its immediate neighbors also experience high inbound traffic through the reachable paths. Furthermore, a link with several surrounding links that spread out traffic quickly is likely to disseminate its traffic more easily.
We embed the 7 features, (V, P i n , P o u t , ρ F i n , ρ F o u t , Z i n , Z o u t ), into a vector for every link, where V is the traffic speed. We make this link embedding more context-aware by simply concatenating an additional vector of external conditions surrounding the link. The external conditions include temperature, precipitation, time of day, day of the week, and an indication of whether a given day is a public holiday. The steps for generating the final input matrix we feed into a neural network for speed prediction is illustrated in Figure 8.
No matter how many adjacent neighbors a link has, the length of the input vector remains invariant, thus making our solution resilient to changes such as the addition or deletion of neighboring links. We do not take the entire adjacency matrix as an input to the prediction engine based on recurrent neural networks. Therefore, our approach becomes space-efficient. Our input vectors contain only the essential information instead of the initial raw adjacency matrix that is sparse. With every link expressed with highly distinct features, we expect it to yield better prediction results. Note that the input vectors are computed at every time window. In the following section, we introduce the method for modeling the transition of the features over time.

3.2. Modeling the Temporal Patterns with Recurrent Neural Networks

Given the input matrix generated at a time window, we predict the traffic speed of every link in the next time window using recurrent neural networks, such as GRU and LSTM. With the hidden layers, we can model a complex non-linear relationship between the input and the output values. Specifically, the example of feeding in the input features of every link and predicting its speed through an LSTM network is illustrated in Figure 9. The states summarized in the previous time window are fed into the block of hidden layers at the subsequent time window. Besides the previous states, we feed in the updated input feature vectors of every link. By training this network, we can model the spatial information’s temporal transition, such as the transition of the traffic system’s structure, traffic flow dynamics, and external conditions. We also make the correlation between the traffic speed of each link and the temporal transition of the most comprehensive set of features to date.

4. Evaluation

In this section, we evaluate our approach using the data from TOPIS (Seoul Traffic Operation and Information service) https://topis.seoul.go.kr/, which contain hourly average speed information for the 4670 major traffic links in the Seoul metropolitan area. We obtained data at one hour intervals for 8760 h worth of data from 00:00 on 1 January 2018 to 23:00 on 31 December 2018. Temperature and precipitation information of each link was obtained from the nearest weather station. We retrieved the climate readings from KMA’s https://data.kma.go.kr/ AWS (Automatic Weather System). We took 60% of the data as a training set, 20% as a validation set, and the rest as the test set. The attributes of the dataset used in this paper are listed in Table 2 with basic statistics, units, measurement intervals, and data types.
We conducted our experiments on NVIDIA DGX-1 with an 80-Core (160 threads) CPU, 8 Tesla V100 GPUs with 32 GB of exclusive memory and 512 GB of RAM. NVIDIA DGX-1 was operated with the Ubuntu 16.04.5 LTS server, and the machine learning tasks were executed through the Docker containers. The machine learning algorithms were implemented with Python (v3.6.9), Tensorflow (v1.15.0), and Keras (v2.3.1) libraries. We used ReLU for the activation function [36,37] and Adam for the optimization function [38,39]. The learning rate was empirically set to 0.001. The early stoppage was configured by setting the patience value to 200 through Keras. To measure the prediction performance, we employed the metrics as follows: (1) MAE (Mean Absolute Error) (Equations (7)); (2) RMSE (Root Mean Squared Error) (Equation (8)); and (3) MAPE (Mean Absolute Percentage Error) (Equation (9)). y t and y ^ t are the predicted speed and the actual speed, respectively. n is the number of test cases. The ϵ value for checking the convergence of the Z value, as defined in Equation (6), was empirically set as shown in Table 3. With ϵ = 0.005, we were able to achieve the lowest MAPE.
M A E = 1 n t = 1 n y t y ^ t
R M S E = 1 n t = 1 n ( y t y ^ t ) 2
M A P E = 1 n t = 1 n y t y ^ t y t 100 ( % )

4.1. Measurement of Prediction Performance

We measured the prediction performance between various approaches denoted as DNN, TFC-DNN, GRU, TFC-GRU, LSTM, and TFC-LSTM. The prefix TFC- stands for Traffic Flow Centrality, and it represents the following seven novel features that we introduced in Section 3, i.e., V, P i n , P o u t , ρ F i n , ρ F o u t , Z i n , and Z o u t . We evaluated the effectiveness of the information about external conditions such as climate and date information separately. DNN, GRU, and LSTM are the artificial neural network architectures we employed for machine learning. Table 4 contains the prediction performance of each approach. For each model, we picked the empirically best hyper-parameter settings, such as the number of hidden layers, perceptrons per layer, and the number of time windows for the recurrent neural networks. For double hidden layers, we had 64 and eight perceptrons for the first and the second layer, respectively. For a single hidden layer, we used 512 perceptrons. We cited the representative existing works that fall under the prediction model categories that do not use our techniques. We did not compare the methods that could not deal with the traffic networks that scale to thousands of links. MSE values during the training and validation stage are plotted on the graphs in Figure 10. MSEs converged at epoch = 400.
The consideration of every link’s external conditions was effective only when used by LSTM and TFC-LSTM. On the other hand, we observed that using the features related to traffic flow centrality consistently led to improvement over the baseline approaches. However, when both the external conditions and the features related to traffic flow centrality were used with LSTM, i.e., TFC-LSTM, we achieved the lowest MAPE of 10.39.
The DNN-based approaches without the information about the external conditions performed poorly compared to the recurrent neural network models because it does not model the temporal transitions of the link features.

4.2. The Effect of Reachable Path Length Cutoff

We noticed that a link in the Seoul traffic system could be reached from any other links within an hour regardless of the distance, even during rush hours with heavy traffic, according to the Floyd–Warshall algorithm we used for retrieving the lowest cost inbound and outbound paths between any OD pairs. It turns out that the Floyd–Warshall algorithm retrieved several unrealistic reachable paths between any OD pairs by ignoring specific restrictions on some of the traffic links. For example, the algorithm generated some routes that included wild U-turns that are not allowed on some roads. As a result, path lengths measured as the number of hops tended to be excessively long. Thus, we were unnecessarily taking into account the state of the links for which it is practically impossible to influence the state of the remote links.
One possible solution to this problem is to have a shorter time window than an hour. However, TOPIS only makes the hourly data available to the public. Therefore, we revised the algorithm to limit the path length instead. When overall traffic flows at the lowest average speed, we limited the reachable inbound and outbound path lengths to 15 hops. On the other hand, for the period when overall traffic flows at the highest average speed, we relaxed the length limit to 45 hops. Furthermore, the link’s low speed may be attributed to the congestion on the neighboring links in the vicinity. Thus, reachability from other links to the low-speed link is also limited. Therefore, it is sufficient to consider only the links in proximity. Compared to the period of lowest average traffic speed (“Lowest Speed Period” in Table 5), the reachability of the traffic from one link to the others is higher during the highest average traffic speed period. Thus, for such a period (“Highest Speed Period” in Table 5), we considered the states of the other links within a wider range. Our simple revision of the algorithm is empirically proven to be effective, as shown in Table 5. It shows the prediction performance for different path length cutoff settings. We observed that our approach performed most effectively with an MAPE of 10.39 when the range of neighboring links to consider was proportional to the traffic’s average speed.

4.3. The Discussion on Scalability

One of the merits of our approach is the capability to predict the traffic speeds even for a very large-scale traffic network such as the road system of Seoul. The larger the traffic network is, the higher the opportunity is to predict speed accurately. This is because we can consider more conditions around the links that are often overlooked by the existing works. The works that only consider the conditions of adjacent neighbors [18] exhibited a decline in prediction accuracy compared to what was originally reported and performed worse than our approach. This is because their approaches were unableto capture the farther away links’ highly probable influence.
Naively feeding in the raw adjacency matrix was the most space inefficient approach [40,41]. We could not even compare the prediction performance with such approaches, as they quickly encountered out-of-memory errors when dealing with Seoul’s large-scale road system. Our approach is agnostic of the scale of the network and even the structure change. Regardless of the scale and any changes, we ran the aggregation functions to compute the fixed-length input feature vector concerning the traffic flow centrality. Therefore, we did not need to restructure the neural network architecture upon changes to the traffic system (i.e., addition/deletion of links, change of adjacent links).
However, we dealt with a fragment of the entire national traffic system in South Korea. The traffic networks in Seoul are connected to the systems in other districts such as Gyeonggi Province and Incheon metropolitan area. Due to the fragmented view, we accidentally identified the links that bridge between different traffic networks as dead-ends, as shown in Figure 11. We could not accurately reflect the traffic flow centrality for these dead-end links. The Z o u t value was zero for these dead-end links because there is no way out. The Z i n value is not credible as the inbound traffic from other regions was not accounted for. The inaccurate Z values on the dead-end may negatively impact the neighboring links’ Z values.
For this issue, we applied the average Z value of the whole system as the Z value of the dead-end links, which still cannot be viewed as an ideal solution. This motivated us to venture into applying our approach to predicting speed prediction for all the links in the entire national traffic system. By expanding the view of the traffic network, we expect the accuracy to improve further. This is planned to be done soon after we are given integrated traffic information from all Korean regions.

5. Conclusions

We presented a novel method for describing the dynamic circumstances around any given traffic links. Specifically, by using a new measure called traffic flow centrality, we were able to concisely express the dynamics of the traffic flow in and out of any given link. Through this measure, we can reflect on the inherently complex ways in the interconnected traffic links affecting each other. Combined with the information about the external conditions surrounding the links, e.g., climate and time of day, the new features and their temporal patterns are used for predicting traffic speed. According to the information available from TOPIS (Seoul Traffic Operation and Information service), training with the LSTM algorithm given our comprehensive spatio-temporal features yielded the lowest prediction error with an MAPE of 10.39. Our experiment also shows that our solution is easily applicable to large-scale traffic systems. We could predict the traffic speed on the road network with thousands of links, unlike the existing works without any efficient feature embedding approaches.
As future work, we plan to apply our solution to the nation-wide traffic system.

Author Contributions

Conceptualization, Y.Y.; methodology, C.L. and Y.Y.; software, C.L.; validation, C.L. and Y.Y.; formal analysis, C.L. and Y.Y.; investigation, C.L.; resources, C.L.; data curation, C.L.; writing, original draft preparation, C.L. and Y.Y.; writing, review and editing, C.L. and Y.Y.; visualization, C.L.; supervision, Y.Y.; project administration, Y.Y.; funding acquisition, Y.Y. All authors read and agreed to the published version of the manuscript.

Funding

This research was supported by the Korean Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure, and Transport (Grant 20TLRP-B148970-03), by the Basic Science Research Programs through the National Research Foundation of Korea (NRF) funded by the Korea government (MSIT) (2020R1F1A104826411), by the Ministry of Health & Welfare, Republic of Korea (Grant Number HI19C0542020020), by the Ministry of Trade, Industry & Energy (MOTIE) and the Korea Institute for Advancement of Technology (KIAT), under Grants P0004602 and P0014268 Smart HVAC demonstration support, and by the 2020 Hongik University Research Fund.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; nor in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ODOrigin and Destination
TFCTraffic Flow Centrality (Z value)

References

  1. Williams, B.M.; Hoel, L.A. Modeling and Forecasting Vehicular Traffic Flow as a Seasonal ARIMA Process: Theoretical Basis and Empirical Results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef] [Green Version]
  2. Kumar, S.V.; Vanajakshi, L. Short-Term Traffic Flow Prediction using Seasonal ARIMA Model with Limited Input Data. Eur. Transp. Res. Rev. 2015, 7, 21. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, C.; Hu, J.; Meng, Q.; Zhang, Y. Short-Time Traffic Flow Prediction with ARIMA-GARCH Model. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 607–612. [Google Scholar]
  4. Jeong, Y.S.; Byon, Y.J.; Castro-Neto, M.M.; Easa, S.M. Supervised Weighting-Online Learning Algorithm for Short-Term Traffic Flow Prediction. IEEE Trans. Intell. Transp. Syst. 2013, 14, 1700–1707. [Google Scholar] [CrossRef]
  5. Wu, C.H.; Ho, J.M.; Lee, D.T. Travel-Time Prediction with Support Vector Regression. IEEE Trans. Intell. Transp. Syst. 2004, 5, 276–281. [Google Scholar] [CrossRef] [Green Version]
  6. Cai, P.; Wang, Y.; Lu, G.; Chen, P.; Ding, C.; Sun, J. A Spatiotemporal Correlative K-Nearest Neighbor Model for Short-Term Traffic Multistep Forecasting. Transp. Res. Part C Emerg. Technol. 2016, 62, 21–34. [Google Scholar] [CrossRef]
  7. Kim, T.; Kim, H.; Lovell, D.J. Traffic Flow Forecasting: Overcoming Memoryless Property in Nearest Neighbor Non-Parametric Regression. In Proceedings of the 2005 IEEE Intelligent Transportation Systems, Vienna, Austria, 16 September 2005; pp. 965–969. [Google Scholar]
  8. Yin, H.; Wong, S.; Xu, J.; Wong, C. Urban Traffic Flow Prediction using a Fuzzy-Neural Approach. Transp. Res. Part C Emerg. Technol. 2002, 10, 85–98. [Google Scholar] [CrossRef]
  9. Quek, C.; Pasquier, M.; Lim, B.B.S. POP-TRAFFIC: A Novel Fuzzy Neural Approach to Road Traffic Analysis and Prediction. IEEE Trans. Intell. Transp. Syst. 2006, 7, 133–146. [Google Scholar] [CrossRef]
  10. Guo, S.; Lin, Y.; Li, S.; Chen, Z.; Wan, H. Deep Spatial-Temporal 3D Convolutional Neural Networks for Traffic Data Forecasting. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3913–3926. [Google Scholar] [CrossRef]
  11. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic Flow Prediction with Big Data: A Deep Learning Approach. IEEE Trans. Intell. Transp. Syst. 2014, 16, 865–873. [Google Scholar] [CrossRef]
  12. Koesdwiady, A.; Soua, R.; Karray, F. Improving Traffic Flow Prediction with Weather Information in Connected Cars: A Deep Learning Approach. IEEE Trans. Veh. Technol. 2016, 65, 9508–9517. [Google Scholar] [CrossRef]
  13. Tian, Y.; Pan, L. Predicting Short-Term Traffic Flow by Long Short-Term Memory Recurrent Neural Network. In Proceedings of the 2015 IEEE International Conference on Smart City/SocialCom/SustainCom (SmartCity), Chengdu, China, 19–21 December 2015; pp. 153–158. [Google Scholar]
  14. Fu, R.; Zhang, Z.; Li, L. Using LSTM and GRU Neural Network Methods for Traffic Flow Prediction. In Proceedings of the 2016 31st Youth Academic Annual Conference of Chinese Association of Automation (YAC), Wuhan, China, 11–13 November 2016; pp. 324–328. [Google Scholar]
  15. Huang, W.; Song, G.; Hong, H.; Xie, K. Deep Architecture for Traffic Flow Prediction: Deep Belief Networks with Multitask Learning. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2191–2201. [Google Scholar] [CrossRef]
  16. Du, B.; Peng, H.; Wang, S.; Bhuiyan, M.Z.A.; Wang, L.; Gong, Q.; Liu, L.; Li, J. Deep Irregular Convolutional Residual LSTM for Urban Traffic Passenger Flows Prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 972–985. [Google Scholar] [CrossRef]
  17. Gu, Y.; Lu, W.; Xu, X.; Qin, L.; Shao, Z.; Zhang, H. An Improved Bayesian Combination Model for Short-Term Traffic Prediction with Deep Learning. IEEE Trans. Intell. Transp. Syst. 2019, 21, 1332–1342. [Google Scholar] [CrossRef]
  18. Cui, Z.; Henrickson, K.; Ke, R.; Wang, Y. Traffic Graph Convolutional Recurrent Neural Network: A Deep Learning Framework for Network-Scale Traffic Learning and Forecasting. IEEE Trans. Intell. Transp. Syst. 2019. [Google Scholar] [CrossRef] [Green Version]
  19. Hamilton, W.L.; Ying, R.; Leskovec, J. Representation Learning on Graphs: Methods and Applications. arXiv 2017, arXiv:1709.05584. [Google Scholar]
  20. Floyd, R.W. Algorithm 97: Shortest Path. Commun. ACM 1962, 5, 345. [Google Scholar] [CrossRef]
  21. Zhang, J.; Zheng, Y.; Qi, D.; Li, R.; Yi, X. DNN-based Prediction Model for Spatio-temporal Data. In Proceedings of the 24th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Burlingame, CA, USA, 31 October–3 November 2016; pp. 1–4. [Google Scholar]
  22. Kim, D.H.; Hwang, K.Y.; Yoon, Y. Prediction of Traffic Congestion in Seoul by Deep Neural Network. J. Korea Inst. Intell. Transp. Syst. 2019, 18, 44–57. [Google Scholar] [CrossRef]
  23. Zheng, Z.; Yang, Y.; Liu, J.; Dai, H.N.; Zhang, Y. Deep and Embedded Learning Approach for Traffic Flow Prediction in Urban Informatics. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3927–3939. [Google Scholar] [CrossRef]
  24. Sun, S.; Wu, H.; Xiang, L. City-wide traffic flow forecasting using a deep convolutional neural network. Sensors 2020, 20, 421. [Google Scholar] [CrossRef] [Green Version]
  25. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  27. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Li, F.-F. Large-Scale Video Classification with Convolutional Neural Networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
  28. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D Convolutional Neural Networks for Human Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
  29. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  30. Jia, Y.; Wu, J.; Xu, M. Traffic Flow Prediction with Rainfall Impact using a Deep Learning Method. J. Adv. Transp. 2017, 2017. [Google Scholar] [CrossRef]
  31. Yu, B.; Lee, Y.; Sohn, K. Forecasting Road Traffic Speeds by Considering Area-Wide Spatio-temporal Dependencies based on a Graph Convolutional Neural Network (GCN). Transp. Res. Part C Emerg. Technol. 2020, 114, 189–204. [Google Scholar] [CrossRef]
  32. Ge, L.; Li, S.; Wang, Y.; Chang, F.; Wu, K. Global Spatial-Temporal Graph Convolutional Network for Urban Traffic Speed Prediction. Appl. Sci. 2020, 10, 1509. [Google Scholar] [CrossRef] [Green Version]
  33. Lu, H.; Huang, D.; Song, Y.; Jiang, D.; Zhou, T.; Qin, J. St-trafficnet: A spatial-temporal deep learning network for traffic forecasting. Electronics 2020, 9, 1474. [Google Scholar] [CrossRef]
  34. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The Graph Neural Network Model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [Green Version]
  35. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  36. Nair, V.; Hinton, G.E. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the 27th International Conference on Machine Learning (ICML 2010), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  37. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  38. Kingma, D.P.; Ba, J. Adam: A method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Ruder, S. An Overview of Gradient Descent Optimization Algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  40. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  41. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3848–3858. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The overall procedure for traffic link embedding and training for traffic speed prediction.
Figure 1. The overall procedure for traffic link embedding and training for traffic speed prediction.
Electronics 09 01800 g001
Figure 2. Relation and bound of the network.
Figure 2. Relation and bound of the network.
Electronics 09 01800 g002
Figure 3. Matrices representing adjacency, distance, and information about reachable paths between links.
Figure 3. Matrices representing adjacency, distance, and information about reachable paths between links.
Electronics 09 01800 g003
Figure 4. Extraction of inbound reachable paths and the computation of the inbound Z value for I 9 .
Figure 4. Extraction of inbound reachable paths and the computation of the inbound Z value for I 9 .
Electronics 09 01800 g004
Figure 5. Extraction of outbound reachable paths and the computation of the outbound Z value for I 9 .
Figure 5. Extraction of outbound reachable paths and the computation of the outbound Z value for I 9 .
Electronics 09 01800 g005
Figure 6. An example of aggregating the traffic conditions of the neighboring links on reachable paths.
Figure 6. An example of aggregating the traffic conditions of the neighboring links on reachable paths.
Electronics 09 01800 g006
Figure 7. Iterative computation of traffic flow centrality for every link in the traffic network.
Figure 7. Iterative computation of traffic flow centrality for every link in the traffic network.
Electronics 09 01800 g007
Figure 8. Generation of the input tensor reflecting the context-aware traffic flow centrality of every link.
Figure 8. Generation of the input tensor reflecting the context-aware traffic flow centrality of every link.
Electronics 09 01800 g008
Figure 9. Using LSTM for modeling the correlation between the temporal patterns and the speed prediction at every time window.
Figure 9. Using LSTM for modeling the correlation between the temporal patterns and the speed prediction at every time window.
Electronics 09 01800 g009
Figure 10. MSEs during training and validation.
Figure 10. MSEs during training and validation.
Electronics 09 01800 g010
Figure 11. Bridge links between traffic networks subject to analysis and out-of-range traffic network accidentally being recognized as dead-end links.
Figure 11. Bridge links between traffic networks subject to analysis and out-of-range traffic network accidentally being recognized as dead-end links.
Electronics 09 01800 g011
Table 1. Comparison with other models. TFC, Traffic Flow Centrality.
Table 1. Comparison with other models. TFC, Traffic Flow Centrality.
Prediction ModelsRaw Data UsageTraffic Network Representation
Traffic
Speed/Volume
Traffic
Network
Structure
Surrounding
Conditions
Invariant
Input Feature
Vector Size
Traffic Flow
Reachability
Analysis
Centrality
Analysis
Chains of
Neighbors
TFC-LSTMOOOOOOO
TGC-LSTM [18]OOXXXXX
ST-3DNET [10]OOXXXXX
DST-ICRL [16]OOOXXXX
IBCM-DL [17]OXXXXXX
DBN [15]OXXXXXX
LSTM [13]OXXXXXX
GRU [14]OXXXXXX
DNN [12,22]OXOXXXX
FNN [8,9]OOXXXOX
K-NN [6,7]OOXXOXX
SVR [4,5]OOXXXXX
ARIMA [1,2,3]OXXXXXX
“X” means no, and “O” means yes.
Table 2. Data attributes.
Table 2. Data attributes.
DataSpeedTemperaturePrecipitationDayHoliday
Mean28.1712.880.14--
Median25.6213.70--
Max110.0040.57611
Min0.74−27.1000
Unitkm/h°Cmm--
Time
Interval
1 h1 h1 h1 day1 day
Typefloatfloatfloatbinarybinary
Table 3. Prediction performance according to the convergence condition ( ϵ ) setting.
Table 3. Prediction performance according to the convergence condition ( ϵ ) setting.
EpsilonMAERMSEMAPE
0.12.604.4010.84
0.052.694.4811.16
0.012.634.4111.07
0.0052.504.2810.39
0.0012.694.4311.15
0.00052.594.3610.79
0.00012.544.3210.51
Table 4. Comparison of prediction performance.
Table 4. Comparison of prediction performance.
Prediction ModelsUsage of SurroundingHidden Layer Structure# of TimeMAERMSEMAPE
Condition Information(# of Perceptrons per Layer)Windows
DNN [12,22]NoDouble layers (64-8)-4.056.2515.58
YesDouble layers (64-8)-4.366.4217.36
TFC-DNNNoDouble layers (64-8)-3.845.9815.27
YesDouble layers (64-8)-3.966.0117.02
GRU [14]NoSingle layer (512)183.705.5114.42
YesSingle Layer (512)187.5210.2031.12
TFC-GRUNoSingle layer (512)182.854.6011.42
YesSingle Layer (512)186.208.1222.57
LSTM [13]NoSingle layer (512)182.884.9111.89
YesSingle Layer (512)182.644.4810.87
TFC-LSTMNoSingle layer (512)182.634.4711.38
YesSingle layer (512)182.504.2810.39
Table 5. The effect of reachability path length limit on speed prediction.
Table 5. The effect of reachability path length limit on speed prediction.
Lowest Speed PeriodHighest Speed PeriodMAERMSEMAPE
15 hops15 hops2.614.3810.83
30 hops30 hops2.514.3010.48
45 hops45 hops2.614.4010.96
15 hops45 hops2.504.2810.39
45 hops15 hops2.634.4410.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, C.; Yoon, Y. Context-Aware Link Embedding with Reachability and Flow Centrality Analysis for Accurate Speed Prediction for Large-Scale Traffic Networks. Electronics 2020, 9, 1800. https://doi.org/10.3390/electronics9111800

AMA Style

Lee C, Yoon Y. Context-Aware Link Embedding with Reachability and Flow Centrality Analysis for Accurate Speed Prediction for Large-Scale Traffic Networks. Electronics. 2020; 9(11):1800. https://doi.org/10.3390/electronics9111800

Chicago/Turabian Style

Lee, Chanjae, and Young Yoon. 2020. "Context-Aware Link Embedding with Reachability and Flow Centrality Analysis for Accurate Speed Prediction for Large-Scale Traffic Networks" Electronics 9, no. 11: 1800. https://doi.org/10.3390/electronics9111800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop