Forecasting Thailand’s Transportation CO 2 Emissions: A Comparison among Artificial Intelligent Models

: Transportation significantly influences greenhouse gas emissions—particularly carbon dioxide (CO 2 )—thereby affecting climate, health, and various socioeconomic aspects. Therefore, in developing and implementing targeted and effective policies to mitigate the environmental impacts of transportation-related carbon dioxide emissions, governments and decision-makers have focused on identifying methods for the accurate and reliable forecasting of carbon emissions in the transportation sector. This study evaluates these policies’ impacts on CO 2 emissions using three forecasting models: ANN, SVR, and ARIMAX. Data spanning the years 1993–2022, including those on population, GDP, and vehicle kilometers, were analyzed. The results indicate the superior performance of the ANN model, which yielded the lowest mean absolute percentage error (MAPE = 6.395). Moreover, the results highlight the limitations of the ARIMAX model; particularly its susceptibility to disruptions, such as the COVID-19 pandemic, due to its reliance on historical data. Leveraging the ANN model, a scenario analysis of trends under the “30@30” policy revealed a reduction in CO 2 emissions from fuel combustion in the transportation sector to 14,996.888 kTons in 2030. These findings provide valuable insights for policymakers in the fields of strategic planning and sustainable transportation development.


Introduction
Greenhouse gases (GHGs) are a group of gases that trap heat in the Earth's atmosphere, leading to the so-called greenhouse effect and contributing to global warming and climate change.Carbon dioxide (CO 2 ) produced as a result of human activities is the primary cause of global warming among these gases.The past decade (2011-2020) has been the hottest in recorded history [1].Furthermore, CO 2 levels in the past decade have also increased at historically high rates-rising more than 2 ppm per year-indicating a continuous increase in CO 2 [2].The increasing amounts of carbon emissions have various environmental impacts, including heatwaves, droughts, floods, storms, and more frequent and severe weather events [3].These events significantly affect ecosystems, human health, agriculture, and the global economy; leading to concerns about mitigating global climate change and ensuring environmental sustainability [4].At the same time, countries around the world are striving to reduce GHG emissions and promote sustainable development.The Paris Agreement, which is subsumed under the United Nations Framework Convention on Climate Change, is a significant international accord that reflects global efforts to address climate change.The main goals of the Paris Agreement are: (1) to limit global warming to below 2 degrees Celsius above pre-industrial levels and (2) to pursue continued efforts to limit the temperature increase to 1.5 degrees Celsius.
As one of the parties to the agreement, Thailand is also engaged in these global efforts.From 2011 to 2021, the average temperature in the country increased by 0.09 degrees Celsius per year.In 2018, the majority of GHG emissions in the energy sector resulted from fuel combustion [5].As shown in Figure 1, the Thai transportation sector plays a significant role in producing carbon emissions, contributing to 29.50% of GHG emissions; making it second only to the energy industry, which contributes 39.63% [6].Therefore, given this background, governments and decision-makers must find accurate and reliable methods of forecasting carbon emissions in the transportation sector; such methods are essential in developing and implementing targeted and effective policies to mitigate the environmental impacts of transportation-related CO 2 emissions [7].
Forecasting 2024, 6, FOR PEER REVIEW 2 As one of the parties to the agreement, Thailand is also engaged in these global efforts.From 2011 to 2021, the average temperature in the country increased by 0.09 degrees Celsius per year.In 2018, the majority of GHG emissions in the energy sector resulted from fuel combustion [5].As shown in Figure 1, the Thai transportation sector plays a significant role in producing carbon emissions, contributing to 29.50% of GHG emissions; making it second only to the energy industry, which contributes 39.63% [6].Therefore, given this background, governments and decision-makers must find accurate and reliable methods of forecasting carbon emissions in the transportation sector; such methods are essential in developing and implementing targeted and effective policies to mitigate the environmental impacts of transportation-related CO2 emissions [7].Various techniques have been employed in the past decade to forecast carbon emissions, ranging from traditional statistical methods to machine learning (ML) algorithms.Previous research has compared CO2-emission-forecasting models in various contexts.For instance, Ağbulut [8] compared deep learning (DL), support vector machine (SVM), and artificial neural network (ANN) models for forecasting CO2 emissions in Turkey.To forecast energy demand in the Turkish transportation sector, Sahraei et al. [9] used the multivariate adaptive regression splines (MARS) technique.Tawiah et al. [10] compared autoregressive integrated moving average (ARIMA), nonlinear autoregressive (NAR), exponential smoothing (ETS), naïve approach, and ANN models to forecast CO2 emissions in Pakistan.Meanwhile, Ning et al. [11] forecasted CO2 emissions and analyzed future CO2 emission trends in China using ARIMA.Xu et al. [12] used nonlinear autoregressive exogenous (NARX) to examine the increasing trend of CO2 emissions in China and analyzed the forecast results using scenario analysis.Liu et al. [13] forecasted energy use in China by comparing multiple linear regression (MLR), a gated recurrent unit artificial neural network (GRU ANN), and support vector regression (SVR).Sun and Liu [14] developed three models-namely, an ANN, an SVM, and the Grey model (GM)-for forecasting CO2 emissions in China.Thabani and Bonga [15] attempted to model and forecast CO2 emissions in India using ARIMA.Meanwhile, Fatima et al. [16] studied the relationships of CO2 gas data in nine Asian countries-namely, Japan, Bangladesh, China, Pakistan, India, Sri Lanka, Iran, Singapore, and Nepal-by comparing the simple exponential smoothing (SES) and ARIMA models; with each country having different suitable models.
Amidst the current landscape of CO2 forecasting in Thailand, several studies have contributed valuable insights using different modeling techniques.For example, Ratanavaraha and Jomnonkwao [17] forecasted the CO2 amount released from transportation energy consumption in Thailand by conducting a comprehensive comparison of Various techniques have been employed in the past decade to forecast carbon emissions, ranging from traditional statistical methods to machine learning (ML) algorithms.Previous research has compared CO 2 -emission-forecasting models in various contexts.For instance, A gbulut [8] compared deep learning (DL), support vector machine (SVM), and artificial neural network (ANN) models for forecasting CO 2 emissions in Turkey.To forecast energy demand in the Turkish transportation sector, Sahraei et al. [9] used the multivariate adaptive regression splines (MARS) technique.Tawiah et al. [10] compared autoregressive integrated moving average (ARIMA), nonlinear autoregressive (NAR), exponential smoothing (ETS), naïve approach, and ANN models to forecast CO 2 emissions in Pakistan.Meanwhile, Ning et al. [11] forecasted CO 2 emissions and analyzed future CO 2 emission trends in China using ARIMA.Xu et al. [12] used nonlinear autoregressive exogenous (NARX) to examine the increasing trend of CO 2 emissions in China and analyzed the forecast results using scenario analysis.Liu et al. [13] forecasted energy use in China by comparing multiple linear regression (MLR), a gated recurrent unit artificial neural network (GRU ANN), and support vector regression (SVR).Sun and Liu [14] developed three models-namely, an ANN, an SVM, and the Grey model (GM)-for forecasting CO 2 emissions in China.Thabani and Bonga [15] attempted to model and forecast CO 2 emissions in India using ARIMA.Meanwhile, Fatima et al. [16] studied the relationships of CO 2 gas data in nine Asian countries-namely, Japan, Bangladesh, China, Pakistan, India, Sri Lanka, Iran, Singapore, and Nepal-by comparing the simple exponential smoothing (SES) and ARIMA models; with each country having different suitable models.
Amidst the current landscape of CO 2 forecasting in Thailand, several studies have contributed valuable insights using different modeling techniques.For example, Ratanavaraha and Jomnonkwao [17] forecasted the CO 2 amount released from transportation energy consumption in Thailand by conducting a comprehensive comparison of modeling techniques, including log-linear regression, path analysis, ARIMA, and curve estimation models; considering related factors, such as GDP, population, and the number of registered vehicles.They concluded that the ARIMA model outperformed the others in terms of predictive accuracy.Similarly, Sutthichaimethee and Ariyasajjakorn [18] attempted to forecast CO 2 emissions from industrial energy use in Thailand using the autoregressive integrated moving average with exogenous variables (ARIMAX) model, incorporating GDP and population data into their analysis.In 2022, Salangam [19] proposed an effective and suitable method for forecasting CO 2 levels in Thailand by comparing regression analysis and an ANN; this incorporated variables such as GDP, population, energy consumption, and the number of registered vehicles.Their results indicated that the ANN predictions of CO 2 levels were six times more efficient and accurate than those derived from regression analysis methods.By demonstrating improved efficiency and accuracy over traditional regression techniques, such a finding suggests that ANNs hold promise as a superior tool for CO 2 forecasting in the context of Thailand.To systematically present relevant research results and provide a comprehensive perspective on analytical methods, we categorized studies by author, methods used, input and output variables, and the region each study pertains to; as shown in Table 1.
CO 2 forecasting faces several limitations, including the complexity of the drivers of emissions (e.g., economic activity, technology, and policies) and the uncertainty and variability of emissions and external shocks, such as natural disasters, the COVID-19 pandemic, and geopolitical events, among others.Furthermore, limited data availability can lead to violations of the statistical assumptions required for some forecasting methods.The most common methods for forecasting carbon emissions are ANN, ARIMA, and SVR, which have different strengths and weaknesses when dealing with forecasting problems.ANNs, in particular, are highly flexible and can adapt to various types of data patterns, including those involving nonlinear relationships [20] or violations of traditional statistical assumptions, such as normality (to some extent) [21].SVR is also adept at capturing nonlinear relationships [22] and is less affected by outliers than traditional regression methods.ARIMAX-or ARIMA with exogenous variables-allows it to account for external factors or intervention events that may influence time series data, thereby improving forecasting accuracy [23][24][25].However, compared with simpler models such as ARIMAX, ANN and SVR models can be very complex and challenging to interpret.To date, no studies have compared the performance of these three methods in forecasting transportationrelated CO 2 emissions in Thailand.Therefore, the current study aims to fill this gap by providing a comprehensive comparison of ANN, ARIMAX, and SVR models in this context.
First, we review the theoretical foundations of each technique.Then, we assess the performance of these models on past carbon emission datasets using various evaluation metrics; including root mean square error (RMSE), mean absolute error (MAE), and mean absolute percentage error (MAPE).Finally, we provide insights into the appropriateness of using these models for forecasting carbon emissions and suggest future research directions to enhance their predictive capabilities.

Data Collection
The data acquired for this study were compiled annually, covering a 30-year period from 1993 to 2022.The dataset comprised CO 2 emissions data from the transportation sector, population, GDP, VK Passenger, VK Freight, and VK Motorcycle.These data were collected from secondary sources provided by various institutions.Detailed sources and information regarding the data are presented in Table 2.These secondary data sources ensured that the information they provided is comprehensive and reliable for analysis.Each agency provides specific datasets relevant to their domain, contributing to a robust and detailed dataset.The data cleaning and preprocessing phase is crucial for ensuring the accuracy and reliability of an analysis.In this study, initially, missing values were addressed through imputation methods, where incomplete records were either filled with appropriate estimates or removed if they were deemed insufficiently representative.The data were then standardized to ensure comparability across different variables.Standardization involved scaling the data such that each variable had a mean of zero and a standard deviation of one.This process helps in normalizing the range of the variables, particularly when they are measured on different scales.Following standardization, outliers were managed using the Z-score method.For each variable, the Z-score is given by the following [35,36]: where X is the data point, µ is the mean, and σ is the standard deviation.Data points with Z-scores exceeding the threshold of 3 or −3 were considered outliers [35,37].These outliers were removed from the dataset or transformed.Table 3 shows a strong correlation between input and output variables such as population, GDP, annual vehicle kilometers (VK), and historical CO 2 emissions [38].Therefore, these inputs were utilized in training models to predict CO 2 emissions related to transportation [8].

Data Analysis
ANNs are adept at capturing complex, nonlinear relationships; making them suitable models for predicting CO 2 levels in various settings [8,19,26].SVR, which is known for its efficiency in high-dimensional spaces, can handle nonlinear data effectively and is another widely used ML method for CO 2 level prediction [14,32,34].ARIMA, a popular statistical approach for forecasting CO 2 -as depicted in Table 1-is specifically designed for time series data; analyzing and making predictions based on historical trends.However, we believe that incorporating appropriate exogenous variables into this model can further improve accuracy.Therefore, we employed ARIMAX.Overall, these methods provide a comprehensive toolkit for predicting CO 2 levels; each offering unique strengths with which to address the complexities of data analysis.

Artificial Neural Network
Artificial Neural Networks (ANNs) are a type of machine learning method inspired by the structure and functioning of the human brain.They consist of numerous interconnected processing units called neurons, which collaborate to process information and recognize patterns.The learning process of an ANN involves adjusting weights through a method known as backpropagation.Initially, input data are fed into the network and pass through multiple layers, where each layer applies weights and activation functions to transform the data, ultimately producing the final output.The network's output is then compared to the actual target values using a loss function, which quantifies the difference between the predicted and actual values.Common loss functions include the mean squared error (MSE) for regression tasks and cross-entropy loss for classification tasks.The error from the loss function is propagated back through the network, and the weights are adjusted to minimize this error by using optimization algorithms such as Gradient Descent.This cycle of forward propagation, loss calculation, and backward propagation continues iteratively until the error is minimized to an acceptable level, allowing the ANN to learn the underlying patterns in the data under analysis.
ANN are highly adaptable to various data patterns, including nonlinear relationships [20], and can handle complex data structures that traditional statistical methods may struggle with; for example, the violation of the assumption of normality [21].However, they also come with inherent complexity and often lack interpretability [39], making it difficult to understand how specific predictions are made.Moreover, ANN typically require large quantities of data and substantial computational power for training.The architecture of an artificial Multi-Layer Perceptron (MLP) neural network is shown in Figure 2.Meanwhile, the mathematical formula for obtaining the forecasting output for the j th sample can be calculated using the following equation [40]: where ω ki,j denotes the weights from the k th hidden node to an output node in the i th iteration, β j indicates the bias of the j th output node sample, f (h k ) is the outcome of hidden node h k after the activation function has been applied, and y j refers to the output of the j th sample.

Support Vector Regression
Support vector regression (SVR) is a machine learning method in which techniques from a support vector machine (SVM) are used to predict continuous outcomes.Initially developed for classification tasks, SVMs have been adapted to handle regression problems [41].SVR is based on the concepts of margin and support vectors.In SVR, the margin refers to the epsilon-insensitive tube around the regression function, within which prediction errors are not penalized.Support vectors are data points that lie on the boundary of this margin or outside it, playing a crucial role in defining the regression function.The decision function of SVR is primarily determined by these support vectors, making SVR generally robust to outliers, depending on the choice of the epsilon parameter [42].The kernel function is used to map input data to higher-dimensional spaces, enabling SVR to handle nonlinear relationships [43].Figure 3 displays a flowchart of an SVR approach.An SVR model with a linear kernel is represented as follows [13,44,45]:

Support Vector Regression
Support vector regression (SVR) is a machine learning method in which techniques from a support vector machine (SVM) are used to predict continuous outcomes.Initially developed for classification tasks, SVMs have been adapted to handle regression problems [41].SVR is based on the concepts of margin and support vectors.In SVR, the margin refers to the epsilon-insensitive tube around the regression function, within which prediction errors are not penalized.Support vectors are data points that lie on the boundary of this margin or outside it, playing a crucial role in defining the regression function.The decision function of SVR is primarily determined by these support vectors, making SVR generally robust to outliers, depending on the choice of the epsilon parameter [42].The kernel function is used to map input data to higher-dimensional spaces, enabling SVR to handle nonlinear relationships [43].Figure 3 displays a flowchart of an SVR approach.An SVR model with a linear kernel is represented as follows [13,44,45]: where ŷsvr , (α i − α * i ), and b represent the output, the difference between the Lagrange multipliers, and the bias, respectively.The kernel function for a linear SVR is denoted by K(x i , x j ), for which the following holds [45]: where ˆsvr y , * ( ) , and b represent the output, the difference between the Lagrange multipliers, and the bias, respectively.The kernel function for a linear SVR is denoted by ( , ) i j K x x , for which the following holds [45]:

Bayesian Optimization
In this study, we employed Bayesian optimization as a strategic method for fine-tuning the hyperparameters of the ML model [46,47].Bayesian optimization is an effective method for optimizing the hyperparameters of complex models such as ANN and SVR.It efficiently searches the hyperparameter space by building a probabilistic model (surrogate model) of the objective function using past evaluations, making the process more efficient than a grid search or random search [48].Studies have shown that Bayesian optimization is particularly suitable for hyperparameter tuning-enabling the optimization of blackbox functions without requiring analytical expressions or gradients-making it ideal for scenarios where evaluations are expensive and noisy [49,50].
The process of optimizing hyperparameters using Bayesian optimization involves a series of meticulously planned steps, as illustrated in Figure 4. Initially, the hyperparameter space is defined by identifying the hyperparameters that require optimization, and setting their possible ranges or values.These parameters can include the number of layers, the number of neurons per layer, activation functions, learning rate, batch size, and regularization strength.

Bayesian Optimization
In this study, we employed Bayesian optimization as a strategic method for fine-tuning the hyperparameters of the ML model [46,47].Bayesian optimization is an effective method for optimizing the hyperparameters of complex models such as ANN and SVR.It efficiently searches the hyperparameter space by building a probabilistic model (surrogate model) of the objective function using past evaluations, making the process more efficient than a grid search or random search [48].Studies have shown that Bayesian optimization is particularly suitable for hyperparameter tuning-enabling the optimization of blackbox functions without requiring analytical expressions or gradients-making it ideal for scenarios where evaluations are expensive and noisy [49,50].
The process of optimizing hyperparameters using Bayesian optimization involves a series of meticulously planned steps, as illustrated in Figure 4. Initially, the hyperparameter space is defined by identifying the hyperparameters that require optimization, and setting their possible ranges or values.These parameters can include the number of layers, the number of neurons per layer, activation functions, learning rate, batch size, and regularization strength.
Afterward, Bayesian optimization is initialized by selecting an initial set of hyperparameters.These can be chosen randomly or based on prior knowledge.The model's performance with these initial settings in place is then evaluated using metrics such as mean squared error (MSE) or root mean squared error (RMSE).Subsequently, a surrogate model-typically a Gaussian process-is constructed to approximate the objective function and predict the performance of various hyperparameter settings.
The next step involves proposing a new set of hyperparameters using an acquisition function.This function strikes a balance between exploring new configurations and exploiting known promising ones.The model is then trained using these proposed hyperparameters, and performance metrics are calculated to assess its effectiveness.The results from these evaluations are integrated into the surrogate model, refining its accuracy and predictive capability.
The next step in the optimization process consists of checking convergence criteria, which may include reaching a maximum number of iterations, achieving convergence in the objective function, or obtaining satisfactory performance improvement.If the stopping criteria are not met, the process returns to proposing the next set of hyperparameters.Once the stopping criteria are satisfied, the optimal set of hyperparameters is identified, and the model is finalized with these optimal parameters for deployment or further testing.Afterward, Bayesian optimization is initialized by selecting an initial set of hyperparameters.These can be chosen randomly or based on prior knowledge.The model's performance with these initial settings in place is then evaluated using metrics such as mean squared error (MSE) or root mean squared error (RMSE).Subsequently, a surrogate model-typically a Gaussian process-is constructed to approximate the objective function and predict the performance of various hyperparameter settings.
The next step involves proposing a new set of hyperparameters using an acquisition function.This function strikes a balance between exploring new configurations and exploiting known promising ones.The model is then trained using these proposed hyperparameters, and performance metrics are calculated to assess its effectiveness.The results from these evaluations are integrated into the surrogate model, refining its accuracy and predictive capability.
The next step in the optimization process consists of checking convergence criteria, which may include reaching a maximum number of iterations, achieving convergence in the objective function, or obtaining satisfactory performance improvement.If the stopping criteria are not met, the process returns to proposing the next set of hyperparameters.Once the stopping criteria are satisfied, the optimal set of hyperparameters is identified, and the model is finalized with these optimal parameters for deployment or further testing.

Autoregressive Integrated Moving Average with Exogenous Variables
ARIMA is a statistical model that is used to analyze time series data to forecast future values based on past values [51].This model decomposes data into three processes: autoregressive (AR), which forecasts a variable based on its past values; integrated (I), which serves to stabilize data or make them stationary; and moving average (MA), which forecasts a variable considering errors at previous points [52].The components of the model consist of three parameters, represented by integers in the form (p, d, q); where p indicates the Lag order or autoregressive of order p, d indicates the number of times the data are differenced or integrated to make them stationary, and q indicates the Lag order or MA of order q.Time series data can often be influenced by special events such as legislative activities, policy changes, environmental regulations, and other similar events.These are referred to as intervention events.By incorporating appropriate exogenous variables that capture the effects of intervention events, an ARIMAX model can significantly improve forecasting performance [23][24][25].This approach is particularly useful in contexts wherein external factors are known to influence the time series data.Meanwhile, the mathematical expression for ARIMAX is depicted as Equation ( 5), which is derived from merging the mathematical expressions of four components: AR, represented by Equation ( 6); I, represented by Equation ( 7); MA, illustrated by Equation ( 8); and Exogenous Variables (X), illustrated by Equation ( 9) [24,51,53,54].
Here, Y t is the value of the series at the t th time, ϕ denotes the coefficients of the AR part, ∆ d denotes differencing d times, B is the backshift operator, θ is the coefficient of the MA part, β k is the coefficients of the exogenous variables X tk , and ε is the white noise error term.
The ARIMAX framework is shown in Figure 5.After loading and preprocessing the data, the next step consists of checking for multicollinearity among exogenous variables using VIF.If high multicollinearity is found, variables are removed or combined.The stationarity of the endogenous variable (Y) is then checked; if it is non-stationary, differencing is applied.Once stationarity is achieved, ACF and PACF plots of the differenced series are used to determine the AR (p) and MA (q) orders.The ARIMAX model is fitted with these orders and the refined exogenous variables.The model is evaluated for residuals and forecast accuracy.If the model is deemed satisfactory, it is used; otherwise, the specification is reevaluated.This process ends once a satisfactory model is obtained.

Scenario Analysis
Scenario analysis is a tool used to analyze and assess the impacts that may arise from different events or situations in the future.In particular, this tool is used to assist in decision-making and strategic planning in uncertain or risky conditions.It includes the best case (optimistic scenario), which explores the impacts in a case wherein everything proceeds as well as possible; the worst case (pessimistic scenario), which explores the impacts

Scenario Analysis
Scenario analysis is a tool used to analyze and assess the impacts that may arise from different events or situations in the future.In particular, this tool is used to assist in decisionmaking and strategic planning in uncertain or risky conditions.It includes the best case (optimistic scenario), which explores the impacts in a case wherein everything proceeds as well as possible; the worst case (pessimistic scenario), which explores the impacts in a case wherein the worst situation occurs; and the base-case scenario, which explores the impacts in a case where conditions are normal or as expected.The likelihood and possible impacts of these scenarios should be considered alongside strategic planning [55].

Evaluation Metrics and Statistical Tests
In this study, we utilized three significant statistical measures-RMSE, MAE, and MAPE-to evaluate the efficacy of our model's forecasting abilities.Each of these metrics offers unique insights into our model's accuracy and precision.Lower values across these measures indicate better model performance.
RMSE is metric that not only measures the difference between the predicted values and the actual values in a dataset but also evaluates the average magnitude of the prediction errors made by the model.It can be expressed as shown in Equation (7).MAE is a commonly used metric in regard to ML statistics for evaluating the performance of predictive models.MAE measures the average magnitude of errors between the predicted and actual values, as shown in Equation (8).MAPE is a metric that assesses the accuracy of a model's predictions by measuring the average percentage difference between the predicted and actual values in a dataset, as shown in Equation (9).
In Equations ( 10)- (12), n is the total number of observations or data points, y i represents the actual value of the i th observation, and ŷi represents the predicted value of the i th observation.In addition, in previous studies [8,[56][57][58], the evaluation of the MAPE metric has been categorized into four levels, as shown in Table 4.In addition to these metrics, the Harvey, Leybourne, and Newbold (HLN) test was employed to statistically compare the predictive accuracy of the models [59].The HLN test, an extension of the Diebold-Mariano (DM) test, is particularly useful for small sample sizes, providing a more robust test statistic [60,61].The null hypothesis (H 0 ) of the HLN test posits that there is no difference in the predictive accuracy between two models (E(d i ) = 0), where d i represents the difference in forecast errors between the models.The alternative hypothesis (H 1 ) suggests that there is a significant difference in predictive accuracy (E(d i ) ̸ = 0).Several studies have successfully applied the HLN test to evaluate forecasting models.For instance, Mizen and Tsoukas [61], Jiao et al. [62], and Song et al. [63] have applied the HLN test to evaluate forecasting models, highlighting the HLN test's applicability and effectiveness in various forecasting contexts.

Data Descriptive
The dataset for forecasting CO 2 emissions in Thailand covers the period from 1993 to 2022.In particular, the current study used data from 1993 to 2013 to train the models, while data from 2014 to 2022 were used to test the models' performance.The ANN and SVM models used population, GDP, VK Passenger, VK Freight, and VK Motorcycle as input variables, while the ARIMAX model used the previous values of CO 2 emissions as input variables.Figure 6a presents a significant rise in CO 2 emissions from the transportation sector, which peaked around 2022.The data indicate increased vehicular activity and possibly lax emission standards, although there seems to have been a decreasing trend in recent years.Thailand's demographic dynamics, shown in Figure 6b, illustrate a continuous upward trend, also indicating a stable economic and social environment.Despite such data, the decelerating growth rate in recent years signifies a structural shift toward an aging society.GDP, as depicted in Figure 6c, highlights Thailand's economic ascendancy, particularly after 2000; with a discernible decrease during the global financial crisis of 2008-2009.Comparable trends can be seen in Figure 6d-f, which illustrate vehicle-kilometers for passenger vehicles, freight, and motorcycles.Furthermore, a notable surge was observed post-2010, which may have been due to economic expansion or increased urbanization.However, the evident decline post-2019 is attributable to travel restrictions stemming from pandemic prevention measures relating to COVID-19.
Forecasting 2024, 6, FOR PEER REVIEW trend in recent years.Thailand's demographic dynamics, shown in Figure 6b, illustrate continuous upward trend, also indicating a stable economic and social environment.D spite such data, the decelerating growth rate in recent years signifies a structural shift t ward an aging society.GDP, as depicted in Figure 6c, highlights Thailand's economic a cendancy, particularly after 2000; with a discernible decrease during the global financi crisis of 2008-2009.Comparable trends can be seen in Figure 6d-f, which illustrate vehicl kilometers for passenger vehicles, freight, and motorcycles.Furthermore, a notable surg was observed post-2010, which may have been due to economic expansion or increase urbanization.However, the evident decline post-2019 is attributable to travel restriction stemming from pandemic prevention measures relating to COVID-19.During the data preprocessing phase, a thorough search for missing data was conducted across all variables, and it was determined that no missing data points were present.Subsequently, outlier detection was performed utilizing the Z-score method.For each variable, Z-scores were calculated; any data points exceeding the threshold of ±3 were considered potential outliers.The analysis revealed that no outliers were detected within the dataset.These results indicate that the dataset is complete and devoid of extreme values, thereby making it suitable for subsequent analysis.

ANN Results
The selection of hyperparameters for the ANN model was driven by an optimization process using Bayesian optimization, which is designed to minimize the Mean Squared Error (MSE).The optimal model configuration included a single hidden layer with seven neurons.This size was chosen to balance complexity and generalizability, avoiding both underfitting and overfitting.The ReLU activation function was used for its efficiency and ability to introduce non-linearity, enhancing the model's ability to learn complex patterns.A regularization coefficient of 826.227 was determined to be optimal, providing a penalty that helps prevent overfitting.
In this study, the structure of the ANN model underwent Bayesian optimization, using predefined hyperparameter search ranges, as outlined in Table 5.The optimization process yielded a minimum MSE value of 12.320 × 10 7 , accompanied by an RMSE of 3715.4.The optimized ANN architecture comprised a single, fully connected layer with a hidden layer size of seven neurons; employing the ReLU activation function and a regularization coefficient of 826.227.This configuration resulted in an enhanced predictive performance on the test set, with performance metrics such as MAPE, RMSE, and MAE yielding values of 6.395, 5054.005, and 4259.170,respectively.When performing hyperparameter tuning using Bayesian optimization, the algorithm selected the set of hyperparameter values that minimized the upper confidence interval of the MSE objective model, rather than the set that minimized the MSE.The optimization process depicted in Figure 7 highlighted the convergence toward the minimum observed and predicted MSE values across 30 iterations, with the 20th iteration representing the minimum-error hyperparameter; while the 16th iteration represents the best-point hyperparameter.

SVR Results
The hyperparameters for the SVR model were meticulously chosen to optimize model performance while balancing complexity and error tolerance.Specifically, a linear kernel, a box constraint (C) of 0.001, and an epsilon value of 1859.835 were chosen.A linear kernel simplifies a model and reduces computational complexity, making it ideal for assessing linear relationships between input features and the target variable.The small box constraint value of 0.001 imposes strong regularization, preventing overfitting by allowing some errors in the training data; thus maintaining a balance between bias and variance.The large epsilon value of 1859.835enhances the model's robustness with respect to outliers and noise by ignoring minor deviations from true values.
Bayesian optimization played a crucial role in this hyperparameter selection process, providing guidance through predefined search ranges to achieve optimal values, as detailed in Table 6.This iterative refinement process effectively balanced the model's complexity and the error of the training data while maintaining tolerance margins around the predicted values.Notably, Figure 8 depicts the 22nd iteration, corresponding to both the minimum-error hyperparameter and the best-point hyperparameter, representing the optimal configuration with the lowest observed error.

SVR Results
The hyperparameters for the SVR model were meticulously chosen to optimize model performance while balancing complexity and error tolerance.Specifically, a linear kernel, a box constraint (C) of 0.001, and an epsilon value of 1859.835 were chosen.A linear kernel simplifies a model and reduces computational complexity, making it ideal for assessing linear relationships between input features and the target variable.The small box constraint value of 0.001 imposes strong regularization, preventing overfitting by allowing some errors in the training data; thus maintaining a balance between bias and variance.The large epsilon value of 1859.835enhances the model's robustness with respect to outliers and noise by ignoring minor deviations from true values.
Bayesian optimization played a crucial role in this hyperparameter selection process, providing guidance through predefined search ranges to achieve optimal values, as detailed in Table 6.This iterative refinement process effectively balanced the model's complexity and the error of the training data while maintaining tolerance margins around the predicted values.Notably, Figure 8 depicts the 22nd iteration, corresponding to both the minimum-error hyperparameter and the best-point hyperparameter, representing the optimal configuration with the lowest observed error.providing guidance through predefined search ranges to achieve optimal values, as detailed in Table 6.This iterative refinement process effectively balanced the model's complexity and the error of the training data while maintaining tolerance margins around the predicted values.Notably, Figure 8 depicts the 22nd iteration, corresponding to both the minimum-error hyperparameter and the best-point hyperparameter, representing the optimal configuration with the lowest observed error.

Hyperparameter
Search Range

ARIMAX Results
All the exogenous variables in the training set have VIF values over 10, indicating a high degree of multicollinearity.To address this issue, stepwise regression was employed for variable selection and to reduce multicollinearity.The VK-Freight variable was selected as the exogenous variable.Then, the Augmented Dickey-Fuller (ADF) test for stationarity was conducted on the data, as presented in Table 7, showing that the data were found to be stationary at the first difference.During the identification stage, a researcher visually examines the autocorrelation function (ACF) and partial autocorrelation function (PACF) plots of the differenced series in Figure 9.The PACF plot shows a significant spike at lag 1, which drops sharply afterward, indicating an autoregressive (AR) component at lag 1.The ACF plot also displays significant spikes at lag 1 that confirm the presence of an autoregressive component.Based on these observations, the suggested model was ARIMAX (1, 1, 1)-incorporating an AR (1) component and an MA (1) component-with a differencing of order 1 since the data are stationary at the first difference.The model also includes the exogenous variable "VK Freight", which was selected after addressing multicollinearity through stepwise regression.Furthermore, in Figure 10, the autocorrelation function (ACF) plot demonstrates that most autocorrelations lie within the confidence bounds.This suggests that the residuals are essentially random and do not exhibit significant autocorrelation.It also implies that the model has effectively captured the time series data's autocorrelation structure.Concurrently, the Quantile-Quantile (Q-Q) plot's alignment with a straight line indicates that the residuals are normally distributed, thus affirming the assumption of normality, which is crucial for the validity of statistical inferences.Together, these diagnostic checks suggest that the model is reliable.Furthermore, the performance of the ARIMAX model was evaluated using several statistical measures; resulting in MAPE, RMSE, and MAE values of 9.286, 7916.483, and 6775.431,respectively.These figures serve as a testament to the model's predictive accuracy and effectiveness when applied to our specific dataset.

HLN Test Results
The HLN test results in Table 8 provide evidence to reject the null hypothesis ( 0 H ) of no difference in predictive accuracy between the compared models in all cases.For the comparison between ANN and SVR, the null hypothesis is rejected at the 5% significance level (HLN Statistic = 4.182 **), indicating a significant difference in predictive accuracy.
Similarly, for the comparison between ANN and ARIMAX, the null hypothesis is rejected at the 1% significance level (HLN Statistic = 12.221 ***); this further demonstrates a significant difference.Additionally, the comparison between SVR and ARIMAX also leads to the rejection of the null hypothesis at the 5% significance level (HLN Statistic = 3.692**).These results collectively indicate that the predictive accuracies of ANN, SVR, and ARI-MAX models are significantly different from each other.

Model Performance
The performance of the ARIMAX model, although not subpar, may have been significantly influenced by its limitations-such as assuming stationarity and being sensitive to

HLN Test Results
The HLN test results in Table 8 provide evidence to reject the null hypothesis ( 0 H ) of no difference in predictive accuracy between the compared models in all cases.For the comparison between ANN and SVR, the null hypothesis is rejected at the 5% significance level (HLN Statistic = 4.182 **), indicating a significant difference in predictive accuracy.
Similarly, for the comparison between ANN and ARIMAX, the null hypothesis is rejected at the 1% significance level (HLN Statistic = 12.221 ***); this further demonstrates a significant difference.Additionally, the comparison between SVR and ARIMAX also leads to the rejection of the null hypothesis at the 5% significance level (HLN Statistic = 3.692**).These results collectively indicate that the predictive accuracies of ANN, SVR, and ARI-MAX models are significantly different from each other.

Model Performance
The performance of the ARIMAX model, although not subpar, may have been significantly influenced by its limitations-such as assuming stationarity and being sensitive to

HLN Test Results
The HLN test results in Table 8 provide evidence to reject the null hypothesis (H 0 ) of no difference in predictive accuracy between the compared models in all cases.For the comparison between ANN and SVR, the null hypothesis is rejected at the 5% significance level (HLN Statistic = 4.182 **), indicating a significant difference in predictive accuracy.Similarly, for the comparison between ANN and ARIMAX, the null hypothesis is rejected at the 1% significance level (HLN Statistic = 12.221 ***); this further demonstrates a significant difference.Additionally, the comparison between SVR and ARIMAX also leads to the rejection of the null hypothesis at the 5% significance level (HLN Statistic = 3.692 **).These results collectively indicate that the predictive accuracies of ANN, SVR, and ARIMAX models are significantly different from each other.

Model Performance
The performance of the ARIMAX model, although not subpar, may have been significantly influenced by its limitations-such as assuming stationarity and being sensitive to multicollinearity among predictors-and might not capture all the complexities of the analyzed data, such as external factors or sudden changes due to variants or policy changes.Given that this model's forecasting was solely reliant on historical CO 2 emission and VKfreight data, with the test set corresponding to the year 2019, the ensuing global COVID-19 pandemic may have influenced the outcome.In particular, Thailand's governmental lockdown measures led to a dramatic reduction in road traffic.This was primarily due to a shift in commuting behaviors for work, school, and other routine activities, as a growing number of individuals transitioned to remote work or online study.Furthermore, the transport and logistics sectors experienced considerable disruptions due to border closures and labor shortages.Collectively, these unforeseen circumstances constitute a significant event that has had a profound impact on CO 2 emissions within the transport sector.
Conversely, the ANN and SVR models exhibited superior and comparable results; as presented in Table 9 and Figure 11, respectively.The inherent adaptability of these models, as evidenced by their ability to consider a multitude of input variables, allowed them to better account for the widespread effects of the global COVID-19 pandemic.The ANN model, with its strength in capturing nonlinear relationships and its ability to learn from and generalize based on the input data, could model complex patterns and anomalies introduced by the pandemic.However, ANN models can sometimes act as black boxes, making it quite challenging to interpret the relationships between, and importance of, different input variables, and they may require larger datasets and more computational resources for training.multicollinearity among predictors-and might not capture all the complexities of the analyzed data, such as external factors or sudden changes due to variants or policy changes.
Given that this model's forecasting was solely reliant on historical CO2 emission and VKfreight data, with the test set corresponding to the year 2019, the ensuing global COVID-19 pandemic may have influenced the outcome.In particular, Thailand's governmental lockdown measures led to a dramatic reduction in road traffic.This was primarily due to a shift in commuting behaviors for work, school, and other routine activities, as a growing number of individuals transitioned to remote work or online study.Furthermore, the transport and logistics sectors experienced considerable disruptions due to border closures and labor shortages.Collectively, these unforeseen circumstances constitute a significant event that has had a profound impact on CO2 emissions within the transport sector.
Conversely, the ANN and SVR models exhibited superior and comparable results; as presented in Table 9 and Figure 11, respectively.The inherent adaptability of these models, as evidenced by their ability to consider a multitude of input variables, allowed them to better account for the widespread effects of the global COVID-19 pandemic.The ANN model, with its strength in capturing nonlinear relationships and its ability to learn from and generalize based on the input data, could model complex patterns and anomalies introduced by the pandemic.However, ANN models can sometimes act as black boxes, making it quite challenging to interpret the relationships between, and importance of, different input variables, and they may require larger datasets and more computational resources for training.In comparison, SVR-with its foundation in statistical learning theory-provides a robust and accurate predictive model, especially in scenarios with smaller datasets and high-dimensional space.SVR models are also capable of managing non-linearities by employing different kernel functions.However, the selection of appropriate kernel functions and the tuning of parameters such as the box constraint and kernel coefficients can be computationally intensive and may require domain expertise to avoid overfitting or underfitting issues.
Therefore, the resilience demonstrated by the ANN and SVR models under these challenging conditions underscores their potential suitability for accurately predicting CO2 emissions in the face of future unforeseeable events; albeit with considerations for In comparison, SVR-with its foundation in statistical learning theory-provides a robust and accurate predictive model, especially in scenarios with smaller datasets and high-dimensional space.SVR models are also capable of managing non-linearities by employing different kernel functions.However, the selection of appropriate kernel functions and the tuning of parameters such as the box constraint and kernel coefficients can be computationally intensive and may require domain expertise to avoid overfitting or underfitting issues.
Therefore, the resilience demonstrated by the ANN and SVR models under these challenging conditions underscores their potential suitability for accurately predicting CO 2 emissions in the face of future unforeseeable events; albeit with considerations for their respective strengths and limitations in terms of model interpretability, parameter tuning, and computational requirements.
The HLN test results confirm significant differences in predictive accuracy among the compared models.Specifically, the null hypothesis of no difference in predictive accuracy is rejected for the comparisons between ANN and SVR (HLN Statistic  4 and 9.For an overview, refer to Figure 11, which presents a radar graph that displays the results for each statistical metric.In this graph, the statistical metrics are scaled from 0 to 10, with RMSE and MAE specifically measured in units of million tons.The ANN model showed the lowest errors across all the metrics; with an MAE of 4.259, an RMSE of 5.054, and an MAPE of 6.395.This indicates superior performance in capturing the complex impacts of COVID-19 on CO 2 emissions.The SVR model, while slightly less accurate-with an MAE of 4.865, an RMSE of 6.194, and an MAPE of 7.628-still performed significantly better than the ARIMAX model.The ARIMAX model had the highest errors-with an MAE of 6.775, an RMSE of 7.916, and an MAPE of 9.286-reflecting its limitations in adapting to the sudden changes induced by the pandemic.The forecasting performance on the test set for each model is illustrated in Figure 12.
Forecasting 2024, 6, FOR PEER REVIEW 19 their respective strengths and limitations in terms of model interpretability, parameter tuning, and computational requirements.
The HLN test results confirm significant differences in predictive accuracy among the compared models.Specifically, the null hypothesis of no difference in predictive accuracy is rejected for the comparisons between ANN and SVR (HLN Statistic = 4.182, **), ANN and ARIMAX (HLN Statistic = 12.221, ***); and SVR and ARIMAX (HLN Statistic = 3.692, **).This indicates that the predictive accuracies of ANN, SVR, and ARIMAX models are significantly different from each other.In terms of model evaluation, each algorithm demonstrated high accuracy in forecasting, as indicated by the metric MAPE in Tables 4  and 9.For an overview, refer to Figure 11, which presents a radar graph that displays the results for each statistical metric.In this graph, the statistical metrics are scaled from 0 to 10, with RMSE and MAE specifically measured in units of million tons.The ANN model showed the lowest errors across all the metrics; with an MAE of 4.259, an RMSE of 5.054, and an MAPE of 6.395.This indicates superior performance in capturing the complex impacts of COVID-19 on CO2 emissions.The SVR model, while slightly less accurate-with an MAE of 4.865, an RMSE of 6.194, and an MAPE of 7.628-still performed significantly better than the ARIMAX model.The ARIMAX model had the highest errors-with an MAE of 6.775, an RMSE of 7.916, and an MAPE of 9.286-reflecting its limitations in adapting to the sudden changes induced by the pandemic.The forecasting performance on the test set for each model is illustrated in Figure 12.

Forecasting and Scenarios
Thailand's "30@30" policy aspires to ensure that electric vehicles (EVs) constitute at least 30% of the nation's total vehicle production by 2030.This policy prioritizes the promotion of EV usage across various transportation modes, including passenger vehicles, freight transport, and motorcycles.This initiative aims to transition energy consumption in the transportation sector to green energy sources, thereby enhancing energy efficiency and reducing greenhouse gas (GHG) emissions.Researchers have hypothesized that, pursuant to this policy, 30% of vehicle-kilometers will correspond to electric vehicles by 2030.The scenario analysis conducted based on this policy compares the expected impact of the "30@30" policy against a benchmark scenario without the policy influence.The aim is to evaluate the potential reduction in CO2 emissions if the policy is implemented successfully.To evaluate the potential impact of the "30@30" policy, a comparison of two scenarios was conducted, specifically for the year 2030.
1.The Benchmark Scenario: This scenario assumes the continuation of current vehicle usage patterns and reliance on traditional energy sources.

Forecasting and Scenarios
Thailand's "30@30" policy aspires to ensure that electric vehicles (EVs) constitute at least 30% of the nation's total vehicle production by 2030.This policy the promotion of EV usage across various transportation modes, including passenger vehicles, freight transport, and motorcycles.This initiative aims to transition energy consumption in the transportation sector to green energy sources, thereby enhancing energy efficiency and reducing greenhouse gas (GHG) emissions.Researchers have hypothesized that, pursuant to this policy, 30% of vehicle-kilometers will correspond to electric vehicles by 2030.The scenario analysis conducted based on this policy compares the expected impact of the "30@30" policy against a benchmark scenario without the policy influence.The aim is to evaluate the potential reduction in CO 2 emissions if the policy is implemented successfully.To evaluate the potential impact of the "30@30" policy, a comparison of two scenarios was conducted, specifically for the year 2030.

1.
The Benchmark Scenario: This scenario assumes the continuation of current vehicle usage patterns and reliance on traditional energy sources.

2.
The Policy Scenario: This scenario incorporates the effects of the "30@30" policy, hypothesizing that 30% of vehicle-kilometers will shift to electric vehicles by 2030.The independent variables VK-passenger, VK-freight, and VK-motorcycle are adjusted to reflect this shift, while GDP and population remain the same as they are in the benchmark scenario.
In prior research, datasets spanning 24 years were employed to predict outcomes over a 15-year span [17]; 15-years-ahead forecasts, as in this study, aligned with the LT-LEDS and Thailand's National Strategy for the years 2023 to 2037 [5].A 15-year forecast (from 2023 to 2037) was made to assess the long-term trends in CO 2 emissions under the benchmark scenario.
An ANN model was employed to forecast CO 2 emissions, using independent variables predicted from available data via the ARIMA model.The dataset was partitioned, with 70% designated to be used for model training, while the remaining 30% was utilized for performance evaluation using MAPE.The utilized models are shown in Table 10.For the population variable the model was specified as the ARIMA (0,2,1); this achieved a MAPE of 0.376%, indicating very precise predictions.The models for GDP, VK-Passenger, VK-Freight, and VK-Motorcycle were specified as ARIMA (0,1,0), indicating that these series follow a random walk [64,65].A random walk is a stochastic process formed by the cumulative summation of independent, identically distributed random variables [66].In these models, future values of the series cannot be predicted from past values, except through differencing; with each value resulting from the previous value plus a random shock.These series do not exhibit significant lagged relationships or moving average processes beyond what is captured through differencing, emphasizing their random walk characteristics.The results in Table 10 demonstrate acceptable MAPE values for all models, as previously mentioned.These results indicate the suitability of these models for forecasting independent variables in future CO 2 predictions.Table 11 illustrates an increasing trend in CO 2 emissions, reflecting the continuation of current vehicle usage patterns and reliance on traditional energy sources.The results of the 15-year forecast indicate that, without intervention, CO 2 emissions will continue to rise, reaching 82,880.635kTons by 2037.By 2030, the Benchmark Scenario predicts that emissions will reach 78,514.470kTons.In contrast, the Policy Scenario, influenced by the "30@30" policy, forecasts a reduction in emissions to 63,517.583kTons.The scenario analysis comparing these two scenarios reveals a substantial decrease in CO 2 emissions from fuel combustion in the transportation sector.This reduction underscores this policy's effectiveness in mitigating emissions through enhancing energy efficiency and the increased utilization of renewable energy sources.Importantly, the anticipated transition to electric vehicles substantially contributes to the reduction in emissions from fuel combustion, highlighting this policy's potential in fostering environmental sustainability.

Conclusions
This study, in which we utilized ML data from 1993 to 2022 and employed models such as SVR and an ANN, demonstrated superior forecasting performance when compared to the traditional ARIMAX model.These ML models incorporate additional inputs, thereby exhibiting a reduced impact from the COVID-19 pandemic compared with that for ARIMAX, which relied on historical CO 2 emission and VK-freight data.Nevertheless, all three models displayed high predictive accuracy, as evidenced by the MAPE being less than or equal to 10% [8,56,57].Upon consideration of performance metrics such as RMSE, MAE, and MAPE, the ANN model emerged as the most fitting choice for forecasting CO 2 emissions in Thailand.However, while this study primarily delved into Thailand's transportation-related carbon emissions, its methodologies and findings can be adapted for application to other countries facing similar challenges pertaining to transportation-related CO 2 emissions.By examining factors such as population growth, GDP, and vehicle-kilometers traveled, this study sheds light on the underlying dynamics shaping emissions trends.Other countries can replicate this approach by conducting similar analyses that are tailored to their specific contexts.Furthermore, while the specific results may vary depending on factors such as data availability and quality, the overarching methodology can be adapted and applied by other countries seeking to improve their own emission forecasting capabilities.The scenario analysis demonstrated that Thailand's "30@30" policy has the potential to make a significant impact on reduction of CO 2 emissions from fuel combustion in the transportation sector by encouraging the widespread adoption of electric vehicles and improving energy efficiency.By comparing the Policy Scenario with the Benchmark Scenario, the analysis predicted a substantial decrease in CO 2 emissions, underscoring this policy's effectiveness.This policy not only promotes environmental sustainability by reducing greenhouse gas emissions but also stimulates economic growth and technological advancements, underscoring the crucial role of policy measures in advancing both environmental and economic sustainability.

Limitations and Future Research
Our study is not without its limitations.In particular, the relatively small dataset, combined with the significant impact of the COVID-19 pandemic, introduced a degree of uncertainty into our results.Thus, future research should focus on the potential impact of electric vehicles on CO 2 emissions, thereby contributing to a more comprehensive understanding of the role of sustainable technologies in reducing Thailand's overall carbon footprint.

ω
ki j denotes the weights from the th k hidden node to an output node in the th i iteration, j β indicates the bias of the th j output node sample, ( ) k f h is the outcome of hidden node k h after the activation function has been applied, and j y refers to the output of the th j sample.

Figure 4 .
Figure 4.The process of optimizing hyperparameters using Bayesian optimization.

Figure 4 .
Figure 4.The process of optimizing hyperparameters using Bayesian optimization.

Figure 8 .
Figure 8. MSE optimization plot for the SVR.

Figure 7 .
Figure 7. MSE optimization plot for the ANN.
, quadratic, cubic A rigorous evaluation via three-fold cross-validation resulted in a minimum MSE value of 8.216 × 10 6 and an impressive RMSE of 2866.3.Further assessment on an independent dataset revealed compelling performance metrics, including MAPE, RMSE, and MAE values of 7.628%, 6193.925, and 4865.085,respectively.

Figure 8 .
Figure 8. MSE optimization plot for the SVR.

Figure 8 .
Figure 8. MSE optimization plot for the SVR.

Figure 11 .
Figure 11.Radar graph of the performance evaluation metrics.

Figure 11 .
Figure 11.Radar graph of the performance evaluation metrics.
= 4.182, **), ANN and ARIMAX (HLN Statistic = 12.221, ***); and SVR and ARIMAX (HLN Statistic = 3.692, **).This indicates that the predictive accuracies of ANN, SVR, and ARIMAX models are significantly different from each other.In terms of model evaluation, each algorithm demonstrated high accuracy in forecasting, as indicated by the metric MAPE in Tables

Figure 12 .
Figure 12.A comparison of the actual and predicted values for CO2 emissions for each model for the test set.

Figure 12 .
Figure 12.A comparison of the actual and predicted values for CO 2 emissions for each model for the test set.

Table 1 .
Summary of studies on emission forecasting found in the literature review.

Table 2 .
Variable, data source, and description.

Table 3 .
Correlation matrices of the variables.

Table 4 .
Guidelines for interpreting the ability of MAPE to forecast accuracy.

Table 5 .
Parameter ranges for ANN optimization.

Table 6 .
Parameter ranges for SVR optimization.

Table 6 .
Parameter ranges for SVR optimization.
Null hypothesis: The time series contains a unit root, which means that the data are nonstationary.

Table 9 .
Results regarding the performance evaluation metric.

Table 9 .
Results regarding the performance evaluation metric.

Table 10 .
Assessment of the accuracy of the ARIMA models in forecasting independent variables.

Table 11 .
Future forecast of transportation-based CO 2 emissions in Thailand.