You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Hypothesis
  • Open Access

16 October 2024

Prediction of Traffic Volume Based on Deep Learning Model for AADT Correction

Department of Highway & Transportation Research, Korea Institute of Civil Engineering and Building Technology, Goyang 10223, Republic of Korea
This article belongs to the Special Issue Intelligent Transportation System Technologies and Applications

Abstract

Accurate traffic volume data are crucial for effective traffic management, infrastructure development, and demand forecasting. This study addresses the challenges associated with traffic volume data collection, including, notably, equipment malfunctions that often result in missing data and inadequate anomaly detection. We have developed a deep-learning-based model to improve the reliability of predictions for annual average daily traffic volume. Utilizing a decade of traffic survey data (2010–2020) from the Korea Institute of Civil Engineering and Building Technology, we constructed a univariate time series prediction model across three consecutive sections. This model incorporates both raw and adjusted traffic volume data from 2017 to 2019, employing long short-term memory (LSTM) techniques to manage data discontinuities. A power function was integrated to simulate various error correction scenarios, thus enhancing the model’s resilience to prediction inaccuracies. The performance of the model was evaluated using certain metrics, such as the mean absolute error, the root mean squared error, and the coefficient of determination, thus validating the effectiveness of the deep learning approach in refining traffic volume estimations.

1. Introduction

The annual average daily traffic volume (AADT) is a crucial metric in road traffic management extensively used for traffic planning, road infrastructure development, and forecasting traffic demand. It also provides a comparative basis for assessing factors that adversely affect economic growth or quality of life across different cities. The metric is typically used to evaluate the road traffic congestion coefficient, which can highlight instances where the traffic exceeds road capacities, leading to congestion.
Predicting long-term traffic volume requires substantial computational resources and depends heavily on hardware capabilities; thus, short-term forecasting of traffic conditions often takes precedence. Extensive research has been conducted on these short-duration predictions, focusing on approximately one-hour intervals. Despite the development and adoption of various short-term prediction models, their reliability can be compromised by incomplete datasets. Hence, the acquisition of precise and comprehensive raw data is crucial.
Road traffic volume is defined as the total number of vehicles traversing a particular point or segment of a road within a specific time unit. Traffic data collection, which includes differentiating the traffic volume by vehicle type and direction on an hourly basis, is known as a traffic volume survey. Current practices in traffic volume surveys are categorized into occasional and continuous types. Occasional surveys, which provide essential data for analyzing the overall road usage, are performed annually and target specific points and sections. These are further subdivided by the targeted roads—general national roads and, for the October survey, expressways and local roads. Techniques for occasional surveys range from mechanical methods using portable traffic equipment to video surveys with closed-circuit television (CCTV) and manual counts by personnel.
Continuous surveys deploy automatic vehicle classification (AVC) systems equipped with sensors that continuously record traffic data around the clock at designated points. These systems gather detailed information on the vehicle type, direction, and timing. Technologies used in continuous surveys include loop and piezoelectric sensors embedded in the roadway, artificial-intelligence-driven video analysis from CCTV footage, and vehicle license plate recognition systems that integrate CCTV with a vehicle specifications database.
Among the continuous traffic volume survey methods, those utilizing buried sensors encounter several challenges. Frequent sensor damage caused by road wear from heavy vehicles often results in gaps in the traffic volume data. Similarly, video survey methods face data losses arising from equipment malfunctions and adverse weather conditions. Furthermore, the traffic volume survey equipment designated for 365-day measurement suffers from inconsistent maintenance times due to communication disruptions, controller malfunctions, and sensor damage. These factors result in prolonged periods where there are missing or abnormal data, rendering it challenging to provide accurate observed traffic volume data. Consequently, the traffic volume information collected from such continuous survey equipment may not accurately reflect the actual traffic volumes owing to a combination of technical limitations, controller defects, sensor errors, external environmental factors, and communication issues.
This study aims to develop and verify a deep-learning-based traffic volume correction and prediction algorithm to enhance the reliability of high-quality traffic volume information and raw data for use in national infrastructure planning, as defined in Article 102 of the Road Act and under Article 88 of the National Transport System Efficiency Act. The spatial scope of this study encompasses 551 points where general national road traffic volume survey equipment (AVC) is installed to calculate the AADT. The temporal scope includes hourly and daily continuous traffic volume data from the past decade (2010 to 2020) collected in real-time from the AVC equipment installed on general national roads.
The main objective is to correct and predict traffic volume by applying a cyclic model-based algorithm to short-cycle complex seasonal time series information while specifically addressing the issue of frequent missing data during the traffic volume data collection process. Most importantly, this study aims to minimize the out-of-sample issues that commonly occur during model construction by comparing and verifying the actual observed traffic volume data against those obtained using the proposed methodology, thereby ensuring reliable traffic volume information.
Additionally, this study focuses accurately predicting traffic volume, which has various potential applications in the field of road traffic. Specifically, it aims to provide more reliable traffic volume information by addressing the issue of missing data owing to the failure of traffic information collection equipment, thereby contributing to improved travel reliability and enhanced transportation planning and management.
In this study, a power-function-based long short-term memory (LSTM) model algorithm is proposed to enhance the model’s ability to recognize uncertainty and long-term trend patterns that accurately reflect the characteristics of traffic volume.

3. Analysis and Procedure of the Development Model

3.1. Analysis of the Development Model

This study evaluates the traditional statistical models (ARIMA, SARIMA) alongside deep learning models (recurrent neural network (RNN), LSTM, power-function-based LSTM) to select suitable models and propose effective algorithms by addressing their limitations.
(1)
ARIMA
The ARIMA model is a well-established time series analysis method that incorporates trends and seasonality by performing seasonal and first differencing, rendering the series stationary, as expressed in Equation (1):
γ t = T t + S t + E t .
Here, T represents the trend, S represents the seasonality, and E represents the randomness.
The ARIMA model extends a typical ARMA model where, if the first difference is zero, it represents a stationary series. The formula for estimating the ARIMA model, which combines first-differenced ARMA, is expressed in Equation (2),
γ t = I + α 1 y t 1 + + α p y t p + e t + θ 1 e t 1 + + θ q e t q ,
where e denotes the error term. However, the ARIMA model is best suited to contexts with short seasonal cycles owing to its structure, which limits its effectiveness in analyzing long seasonal cycles and handling multiple seasonalities.
(2)
SARIMA
The SARIMA model integrates the seasonal autoregressive model and seasonal moving average model within the ARIMA framework. The model, as formulated in Equation (3), accommodates the probabilistic characteristics of seasonality (m),
S A R I M A ( p ,   d ,   q ) ( P ,   D ,   Q ) m ,
where p denotes the non-seasonal autoregressive order, d denotes the degree of differencing, q denotes the non-seasonal moving average order, P denotes the seasonal autoregressive order, D denotes the degree of seasonal differencing, Q denotes the seasonal moving average order, and m represents the number of observations per year (indicating annual seasonality).
The distinction between ARIMA and SARIMA lies in SARIMA’s capacity to encapsulate the seasonal cycle (m), as expressed in Equation (4),
Φ p ( B s ) Φ p ( B ) ( 1 B ) d ( 1 B s ) D Y t = 1 + θ Q ( B s ) θ q ( B ) ε t ,
where ε t denotes the error term (also known as white noise) and B represents the backward shift operator. If the traffic volume data cycle is recorded monthly, this equation can be modified to Equation (5):
( 1 Φ p ) ( 1 Φ 1 B 12 ) ( 1 B ) ( 1 B 12 ) y t = ( 1 + θ 1 B ) ( 1 + θ 1 B 12 ) ε t .
(3)
RNN
RNN is predominantly utilized for time series analysis and is particularly adept at processing sequential data. A defining feature of the RNN is its state vector (also known as the hidden unit), which forms a singular neural structure for the sequence and retains the memory of all preceding elements. The schematic recursive structure of an RNN incorporating the backpropagation through a time algorithm is presented in Figure 1 [21].
Figure 1. Basic recursive structure of recurrent neural network.
A major assumption for different traditional statistical learning models is the independence among data samples. However, this assumption does not hold for sequential data, where dependencies exist among individual elements over time. This characteristic is particularly relevant to the traffic volume data analyzed in this study, negating the typical neural network advantage of processing each data sample independently.
Within an RNN, the transfer of information through the network’s weight matrices is described by the following equations:
s t = σ ( U x t + W s t 1 + b s ) ,
h t = s o f t m a x ( V s t + b h ) .
In RNNs, the weights are designed to handle variable-length sequences. Each timestep receives a new input value, thus updating the hidden state, s t . The information is sequentially relayed throughout the network over time, with each prior input stored as memory, thus enhancing the processing of incoming data.
(4)
LSTM
LSTM networks, a type of traditional RNN, are designed to address the gradient vanishing problem that occurs during the training processes, particularly when there is a significant time gap between the relevant information and its point of use in long-term trend data. This issue causes the gradient to diminish progressively during backpropagation, substantially reducing the network’s learning capability. In contrast to that of the standard RNNs, the LSTM architecture incorporates a cell state along with the hidden state, which helps in preserving information over longer sequences.
In this study, the LSTM model was employed to compare its predictive power against that of the newly developed power-function-based LSTM model. The standard LSTM model is structured as a relatively simple multi-layered neural network optimized for predicting short-term time steps (e.g., one hour ahead) using a single window. The predictive accuracy of the LSTM model varies with the time steps. Structurally, a linear transformation layer is inserted between the data input and prediction layers owing to limitations of the algorithm. While enabling fast computation, this layer reduces the model’s predictive performance. The model relies primarily on the values and features of the immediately preceding time step (t = 0) to predict the next time step, often failing to capture long-term trends and characteristics effectively.
(5)
Proposed Power-Function-Based LSTM Model
This study introduces a power-function-based LSTM model to address the challenges presented by the forget gate in the conventional LSTM models. The employed LSTM model is an RNN that includes an LSTM hidden layer. When dealing with lengthy input streams without explicitly identified start and end points, the state of the LSTM can expand indefinitely, thus potentially destabilizing the network. To overcome this issue, we implemented a method of resetting the memory cell contents before initiating a new sequence. The use of an LSTM with a forget gate aims to prevent instability or interruptions during the learning process.

3.2. Development Model Analysis Procedure

Figure 2 depicts the procedure for developing the deep-learning-based model. Initially, the process involved establishing baseline values for traffic volume data, detecting anomalies in the raw data, and conducting data validation and preprocessing to correct and predict the traffic volume. Subsequently, datasets were organized for anomaly detection in traffic volume data, including a training set exclusively comprising normal values, a validation set with normal values, a secondary validation set containing both normal values and anomalies, and a test set comprising both normal values and anomalies. The third step involved defining the characteristics of the data to ensure appropriate data transfer suitable for each deep learning structure before the actual implementation of the deep learning model. The fourth step addressed the challenge of gradient vanishing during data analysis by setting input sequences for traffic volume prediction to construct the deep learning model. An architecture was then developed to evaluate and validate the predictive power of the devised model.
Figure 2. Development of the architecture for deep-learning-based models.

4. Statistical Learning Verification and Machine Learning Model Analysis and Evaluation

4.1. Statistical Learning Verification Design

Previous studies have commonly employed a preprocessing process to correct data anomalies before training the deep learning models. This approach, although necessary in the absence of actual values, tends to complicate the analysis process and significantly reduces the processing speed. Moreover, the use of simple interpolation methods during the correction phase can degrade the data quality. In this study, where actual values are available, the predictive power was verified without prior data correction, thereby simplifying the analysis process and enhancing efficiency. This study evaluated the model’s performance using certain metrics, such as the mean absolute percentage error (MAPE), the MAE, the root mean squared error (RMSE), and the coefficient of determination (R2), and formulated additional data-missing scenarios for analysis. Figure 3 depicts the design procedure for the time series prediction performance verification.
Figure 3. Design procedure for verifying time series prediction performance.
  • Scenario 1: Assumes 10% of the data is missing.
  • Scenario 2: Assumes 50% of the data is missing.
Bayesian optimization was conducted using the GPyOpt 1.2.6 library, thus necessitating the determination of candidate values for each hyperparameter. Values for certain parameters, such as the learning rate and the dropout rate, were established based on the existing literature. Additionally, this study explored three distinct and arbitrarily selected network architecture variations.
For the coding and implementation of experiments, the Python 3.12.4 programming language and Keras 2.3 were employed. Keras is an open-source project that offers a high-level API that supports the implementation of deep neural networks along with libraries, such as TensorFlow 2.10 and Theano 1.0.0. In this study, Keras was utilized in addition to TensorFlow, thus providing advantages in rapid prototyping and experimental flexibility through its simple API. This API supports modular neural network construction via the integration of various layers, activation functions, loss functions, and optimizers, and it includes provisions for most conventional deep learning components. Despite its broad utility, Keras may present limitations for constructing custom or novel solutions, for which TensorFlow may be a more appropriate choice. Notably, Keras incorporates an LSTM implementation featuring a forget gate, as described by Gers et al. (2000) [22].

4.2. Machine Learning Model Analysis and Evaluation

(1)
SARIMA Model Analysis Results
In evaluating the SARIMA model, the analysis of sample errors revealed no discernible patterns in the residuals between the estimated and actual values, suggesting the absence of significant heterogeneity issues; therefore, no additional log transformations were necessary. The Ljung–Box (LB) test was conducted to verify the accuracy of the model judgments. The LB test yielded a significance probability of 0.04, which is below the threshold of 0.1, indicating minimal autocorrelation. Consequently, the optimal parameters for the SARIMA model were determined through the auto grid searching method without the need for further log transformations or differencing.
The selection of optimal parameters was guided by Akaike’s information criterion (AIC), the Bayesian information criterion (BIC), and various significance tests. The results are summarized in Table 2.
Table 2. Optimal parameter estimation results for the SARIMA model.
The overall prediction results indicated excellent performance, averaging approximately 85%; however, there was a notable decline in the prediction accuracy due to long-term missing data from 2019, which adversely affected the model’s ability to reflect the periodicity and trends accurately. Table 3 presents the evaluation results of the SARIMA (4,1,3)(4,0,3) 12 model, revealing the discrepancy between the actual and predicted values, which ranged from a minimum of 3.28% to a maximum of 14.59%. Although the SARIMA model typically uses MAPE to estimate prediction errors, for consistency in comparisons across the statistical, machine learning, and deep learning models, MAPE was converted to MAE. The converted MAE was analyzed to be 0.158.
Table 3. Evaluation results of the SARIMA (4,1,3)(4,0,3) 12 model.
(2)
Prophet Model Analysis
A separate analysis was conducted using the min–max scaling method to standardize the range of data values. Evaluation metrics included RMSE, MAE, and R2. The analysis focused on two specific points, as detailed in Table 4. Given its nature as a univariate regression model, R2, which indicates the linear fit, serves as a measure of prediction accuracy.
Table 4. Evaluation results of the Prophet model.
Figure 4 depicts the results of predicting the hourly traffic volume at point 10001 using the Prophet model, where MAE is 0.117, RMSE is 0.147, and R2 is 0.914, indicating considerably high predictive performance.
Figure 4. Prophet model’s prediction results (point 10001).
Figure 5 presents the results of predicting the hourly traffic volume at point 10004 using the Prophet model. The blue line represents the actual traffic values, and the yellow line represents the predicted values. For point 10001, the regularity in the past time series patterns (seasonality, periodicity) was confirmed, leading to a high predictive performance, with an MAE of 0.117 and an RMSE of 0.147.
Figure 5. Prophet model’s prediction results (point 10004).
However, for point 10004, irregular past time series patterns presented lower predictive performance, with an MAE of 111.687, an RMSE of 158.050, and an R2 of 0.847, when compared to point 10001. This indicates the presence of uncertainties that are not recognized by the data. To address these limitations, analyses utilizing deep-learning-based models were conducted.
(3)
LSTM Model
The LSTM model was utilized to evaluate the predictive power of the proposed LSTM model, which incorporates a power function. The predictive accuracy of the LSTM model varies based on the time point, as shown in Figure 6. Owing to the structural limitations of the algorithm, a linear transformation layer is incorporated between the input values (inputs) and the predicted outputs (predictions). This architecture lacks densely nonlinear hidden layers between the input and output layers, which, while enhancing the computational speed, considerably diminishes the predictive performance.
Figure 6. LSTM model’s prediction results (24 h).
Furthermore, the model predicts the subsequent time step based solely on the values and characteristics of the immediately preceding time step (t = 0). This approach does not adequately capture the long-term trends and characteristics inherent in the data.
(4)
Proposed LSTM Model Based on Power Function
Previous efforts to modify the LSTM architectures have aimed at enhancing the inference capabilities for data characterized by long-term trends, typically yielding similar performance levels. However, our analysis identified inconsistencies in the LSTM performance. In particular, within the forget gate, the memory cell, c t , which stores past data, is described by Equation (8):
c t = f t c t 1 + i t c t ¯ .
Despite the superior performance of the LSTM models with forget gates over the standard LSTM or RNN models, they frequently suffer significant information loss during the decay process, where non-essential information is discarded using a Sigmoid function. Shivangi et al. [23] suggested an improved decay function using exponential functions to address this issue. However, exponential functions still led to excessive information loss in datasets with pronounced long-term trends. This effect was notably severe in data with substantial long-term dependencies, such as the hourly traffic volume data. Therefore, this study proposes an LSTM model that employs a power function as the decay function within the forget gate, designed to accommodate long-term patterns and reduce unnecessary information decay in the time series data with significant long-term dependencies.
The memory cell in the forget gate is redefined by the power function to reduce the rate of information decay:
c t = c 0 ( t t 0 + 1 ) p = ( t t 0 + 1 t t 0 ) c t 1 .
To evaluate the absolute predictive performance of the proposed LSTM model, it was compared against those of linear combination models, multi-step combination models, and convolution neural networks. Figure 7 depicts the comparative prediction performance of the power-function-based LSTM model against that of the traditional LSTM model, demonstrating enhanced prediction accuracy for the traffic volume.
Figure 7. Comparison of performances between LSTM and power-function-based LSTM.
To further substantiate these findings, Table 5 presents the prediction performance metrics, demonstrating that the LSTM model achieves an MAE of 0.1288, whereas the power-function-based LSTM model significantly improves this with an MAE of 0.0733. This confirms that the LSTM model developed with the power function exhibits superior predictive performance when compared to the standard LSTM model.
Table 5. Prediction performance results of LSTM and power-function-based LSTM.
(5)
Traffic Volume Prediction Using the Power-Function-Based LSTM
Figure 8 depicts the 24 h traffic volume predictions at various random points and times using the power-function-based LSTM model developed in this study.
Figure 8. Traffic volume prediction results of the power-function-based LSTM (24 h).
Table 6 presents the evaluation results of the LSTM model at point 10001 under two scenarios, each assuming different data missing rates. The analysis indicated that scenario 2, with a higher missing rate, outperformed scenario 1, which had a lower missing rate. This suggests that the power-function-based LSTM model can effectively capture long-term trends and uncertainties in the traffic volume data. However, a higher missing rate does not universally ensure better predictions. If the points or times with significant missing data are randomly selected, the prediction performance at a 50% missing rate might actually deteriorate. Figure 9 and Figure 10 present the results for each scenario.
Table 6. Evaluation results of the LSTM model at point 10001.
Figure 9. LSTM model’s results (point 10001, 10% data missing).
Figure 10. LSTM model’s prediction results (point 10001, 50% data missing).

5. Model Prediction Performance Verification Results

This study aimed to develop an optimal model to address the missing traffic volume data by comparing and analyzing statistical learning, machine learning, and deep learning models. The SARIMA model was employed to account for seasonality in the traffic volume data, the Prophet model was used for univariate predictions specializing in time series estimation, such as traffic volume, and the LSTM model was utilized to address long-term trends and uncertainties.
Table 7 presents the prediction results based on MAE for the analyzed statistical learning, machine learning, and deep learning models. The verification results indicated that the deep learning models performed the best in predicting traffic volume, followed by machine learning models and statistical learning models. Notably, the LSTM model in deep learning exhibited a higher MAE than the Prophet model in machine learning.
Table 7. Model prediction performance verification results.
To address this problem, this study proposed the power-function-based LSTM model. This modified LSTM model, utilizing the developed power function, achieved superior performance in correcting and predicting traffic volume compared to the existing models.
Table 8 compares the error rates among the statistical learning, machine learning, and deep learning models, revealing that the power-function-based LSTM model demonstrated lower errors when compared to SARIMA (84.18%), Prophet (78.63%), and LSTM (80.47%). The power-function-based LSTM model recorded an MAE of 0.025, indicating the best performance among all evaluated models.
Table 8. Comparative analysis of error rates among models.
To further verify the prediction accuracy of the proposed power-function-based LSTM model, the AADT at point 10001 was assumed to have a 50% missing rate when compared to the actual values. The verification results demonstrated a difference of approximately 24 vehicles/day using the existing methods, with an error rate of 0.164%. The power-function-based LSTM model exhibited a difference of approximately four vehicles/day, with an error rate of 0.025%.
The accuracy verification for the AADT indicated a difference of 20 vehicles/day and 8760 vehicles/year, demonstrating that the proposed model significantly outperformed existing methods in correcting traffic volume based on actual values. Table 9 presents the accuracy verification results of the power-function-based LSTM model.
Table 9. Accuracy verification results of the power-function-based LSTM model.

6. Conclusions

The challenge of uncertainty in traffic volume data persists as an unavoidable issue for next-generation traffic systems. In particular, sensor-based time series traffic volume data, which accumulate vast amounts of real-time information, consistently exhibit anomalies and missing values despite the sophistication of the collection equipment. Therefore, it is crucial to accurately assess the uncertainty of the collected data and execute appropriate correction tasks.
This study applied statistical learning, machine learning, and deep learning models—all primarily utilized for time series algorithms—to develop and evaluate methods that accurately reflect the characteristics of the traffic volume data and enhance the accuracy of correcting and predicting the missing traffic volume data. The key findings of this study are summarized as follows.
First, a statistical learning model, SARIMA (4,1,3)(4,0,3) 12 [9], effectively captured the periodicity, seasonality, trend, and residual characteristics of traffic volume data. Predictions were also obtained using the Prophet model, a machine-learning-based additive algorithm that performs iterative computations. The SARIMA model predicted traffic volume with an average accuracy of 85%; a comparison of actual values to predicted values exhibited a variation ranging from a minimum of 3.28% to a maximum of 14.59%. For the Prophet model, the performance was evaluated by deleting 10% of the data arbitrarily, revealing an MAE of 0.117, an RMSE of 0.147, and an R2 of 0.914 for point 10001 and an MAE of 111.687, an RMSE of 158.050, and an R2 of 0.847 for point 10004.
Second, to address the long-term trends and data uncertainties evident in SARIMA and Prophet models, the performance of the LSTM model—a deep learning approach—was assessed for traffic volume prediction. The results indicated that the LSTM model outperformed the SARIMA model but was less accurate than the Prophet model. To enhance the LSTM model’s performance, this study introduced an improved LSTM model using a power-function-based forget gate memory cell. This adaptation aims to reduce information loss over time and more accurately reflect long-term patterns in time series data with high dependencies.
Third, the traffic volume prediction performance of the proposed power-function-based LSTM model was compared with those of the statistical learning model SARIMA, the machine learning model Prophet, and the deep learning model LSTM. The results confirmed that the proposed model achieved the lowest MAE for correcting and predicting the missing traffic volume.
Therefore, the proposed power-function-based LSTM model is expected to significantly contribute to enhancing the efficiency of traffic surveys and statistics. By improving traffic volume corrections and predictions, it can help prepare accurate and reliable road traffic statistics, which are imperative for national planning and decision making. Moreover, its ability to handle time series data collected in various formats from diverse sources demonstrates its versality and potential applications beyond traffic volume analysis, which is another significant achievement of this study.

Funding

This study was prepared with funding from the 2024 Traffic Information Provision System Service (Grant No: 20240266-004). The author extends appreciation to the related organizations for their cooperation.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

This study was conducted using data provided by the Korea Institute of Civil Engineering and Building Technology (Project Title: 2024 Traffic Volume Information Provision System Service), and the author extends appreciation to the related organizations for their cooperation.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Satish, S.; Lingras, P.; Zhong, M. Effect of missing value imputations on traffic parameters estimations from permanent traffic counts. In Proceedings of the 80th Annual Meeting, Washington, DC, USA, 9–12 April 1986; Transportation Research Board: Washington, DC, USA, 2003; pp. 1–33. [Google Scholar]
  2. Tim, L.; Shawn, T.; Richard, M. Monitoring Urban Roadway in 2000: Using Archived Operations Data for Reliability and Mobility Measurement; Texas Transportation Institute: College Station, TX, USA, 2001. [Google Scholar]
  3. Mei, H.; Ma, A.; Poslad, S.; Oshin, T.O. Short-term traffic volume prediction for sustainable transportation in an urban area. J. Comput. Civ. Eng. 2015, 1, 11. [Google Scholar] [CrossRef]
  4. Do, M.-S.; Lee, H.-M.; NamKoong, S. Outlier filtering and missing data imputation algorithm using TCS data. Korean Soc. Transp. 2008, 26, 241–250. [Google Scholar]
  5. Lee, J.; Do, M.-S.; Kim, S.-H.; Ryu, S.-K. Real-time adjustment of traffic volume—Based on the national highway Route 3. Korean J. Appl. Stat. 2003, 16, 203–215. [Google Scholar] [CrossRef][Green Version]
  6. Kim, J.-Y.; Lee, Y.-I.; Baek, S.-G.; Nam, G.-S. A study on the imputation for missing data in dual-loop vehicle detector system. Korean Soc. Transp. 2006, 24, 27–40. [Google Scholar]
  7. Jung, S.-W.; Oh, J.-S. Traffic volume correction system using axle counts from piezo sensors. J. Korea Contents Assoc. 2021, 21, 277–283. [Google Scholar] [CrossRef]
  8. Ha, J.A.; Park, J.H.; Kim, S.H. Missing data imputation using permanent traffic counts on national highways. Korean Soc. Transp. 2007, 25, 121–132. [Google Scholar]
  9. Han, D.; Lee, D.W.; Jung, D. A study on the traffic volume correction and prediction using SARIMA algorithm. J. Korea Inst. Intelligent Transp. Syst. 2021, 20, 1–13. [Google Scholar] [CrossRef]
  10. Korea Institute for Health and Social Affairs. Study on Machine Learning-Based Anomaly Detection Techniques; Korea Institute for Health and Social Affairs: Sejong-si, Republic of Korea, 2018.
  11. Korea Institute of Civil Engineering and Building Technology. Study on Improving and Evaluating the Reliability of ITS Information; Korea Institute of Civil Engineering and Building Technology: Goyang-si, Republic of Korea, 2008.
  12. Steffen, M.; Bartz-Beielstein, T. ‘impute TS: Time Series Missing Value Imputation in R,’ Contributed research article. R J. 2017, 9, 207–218. [Google Scholar]
  13. Zhou, Y.; Meng, P. Diagnosis of causes for high railway traffic based on Bayesian network. Math Modell. Eng. Probl. 2019, 6, 136–140. [Google Scholar] [CrossRef]
  14. Li, L.; Su, X.; Zhang, Y.; Lin, Y.; Li, Z. Trend modeling for traffic time series analysis: An integrated study. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3430–3439. [Google Scholar] [CrossRef]
  15. Hamed, M.M.; Al-Masaeid, H.R.; Said, Z.M.B. Short-term prediction of traffic volume in urban arterials. J. Transp. Eng. 1995, 121, 249–254. [Google Scholar] [CrossRef]
  16. Castro-Neto, M.; Jeong, Y.S.; Jeong, M.K.; Han, L.D. Online-SVR for short-term traffic flow prediction under typical and atypical traffic conditions. Expert Syst. Appl. 2009, 36, 6164–6173. [Google Scholar] [CrossRef]
  17. Van der Voort, M.; Dougherty, M.; Watson, S. Combining Kohonen maps with Arima time series models to forecast traffic flow. Transp. Res. C 1996, 4, 307–318. [Google Scholar] [CrossRef]
  18. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic flow prediction with big data: A deep learning approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar] [CrossRef]
  19. Fusco, G.; Colombaroni, C.; Comelli, L.; Isaenko, N. Short-term traffic predictions on large urban traffic networks: Applications of network-based machine learning models and dynamic traffic assignment models. In Proceedings of the International Conference on Models and Technologies for Intelligent Transportation Systems (MT-ITS), Budapest, Hungary, 3–5 June 2015; pp. 93–101. [Google Scholar] [CrossRef]
  20. Cho, K.; Lee, S.S.; Nam, D. Forecasting of rental demand for public bicycles using a deep learning model. J. Korea Inst. Intelligent Transp. Syst. 2020, 19, 28–37. [Google Scholar] [CrossRef]
  21. Won-Jae, J.; Korea Transport Institute. Application of Deep Learning in the Transportation Sector; Korea Transport Institute: Sejong, Republic of Korea, 2019.
  22. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  23. Shivangi, M.; Vo Vy, A.; Turek Javier, S.; Huth Alexander, G. Multi-timescale representation learning in LSTM language models. ICLR 2021, 2021, 1–19. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.