Next Article in Journal
Optimization of Process Parameters for Pellet Production from Corn Stalk Rinds Using Box–Behnken Design
Previous Article in Journal
Aerodynamic Study of MotoGP Motorcycle Flow Redirectors
Previous Article in Special Issue
Comprehensive Benefit Analysis of Port Shore Power Based on Carbon Trading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Dynamics of US Gasoline Demand and Its Prediction: An Extended Dynamic Model Averaging Approach

School of Mathematics and Statistics, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Energies 2023, 16(12), 4795; https://doi.org/10.3390/en16124795
Submission received: 10 March 2023 / Revised: 28 April 2023 / Accepted: 4 May 2023 / Published: 19 June 2023
(This article belongs to the Special Issue Modeling Energy–Environment–Economy Interrelations)

Abstract

:
This study contributes to the body of literature on modeling and predicting gasoline demand by using nonlinear econometric techniques. For this purpose, dynamic model averaging (DMA) and Bayesian model averaging (BMA) combined with Artificial Bee Colony (ABC) are used to forecast gasoline consumption in the United States. The article’s independent variables include demographic characteristics, economic activity, income, driving expenditures, automobile price, and road availability for annual data from 1960 to 2020. In the proposed model, not only may the coefficients and elasticity of a predictor of gasoline demand change over time, but other sets of predictors can also emerge at different periods. Moreover, this study aims to automate the process of picking two forgotten variables of the DMA model using the ABC model. Our findings indicate that dynamic model averaging significantly improves forecasting performance when compared to basic benchmark techniques and advanced approaches. Additionally, integrating it with an Artificial Bee Colony (ABC) may result in improved outcomes when time-varying forgetting variables are present. The findings of this research provide policymakers in the fields of energy economics and the environment with helpful tools and information.

1. Introduction

Gasoline demand in the United States has been steadily increasing since the 1990s. In 2019, Approximately 143 billion gallons of gasoline was used in the United States, with the transportation sector accounting for over 70% of the total consumption [1]. The demand increase can be attributed to factors, such as population growth, urbanization, and increased consumer spending on vehicles. Furthermore, the EIA [1] reported that gasoline demand is highly sensitive to changes in economic activity, fuel prices, and weather patterns. For example, during the COVID-19 pandemic in 2020, gasoline demand in the US fell significantly due to reduced economic activity and stay-at-home orders. However, as the economy recovers and restrictions are lifted, demand is expected to increase once again. In addition, the EIA [1] predicts that gasoline demand will continue to rise in the coming years, reaching approximately 151 billion gallons by 2050.
Numerous research has been conducted on the effectiveness of gasoline demand factors and their capacity to forecast. In this context, some previous studies have adopted a direct approach to estimation by examining the demand for car sales [2,3,4,5]. Apart from forecasting vehicle sales, research in the area of travel demand has also looked at gasoline use as a response variable when evaluating fuel price elasticities [6,7,8,9]. Huo and Wang [5] discovered that pricing and income elasticities in China are based on consumer vehicle stock and projected vehicle sales in China up to 2050 using the FEEI model. Bento et al. [10] conducted similar research for the United States, using a simultaneous equations model for US households and taking into account the new discarded vehicle markets, among other factors. Graham and Glaister [9] conducted a thorough literature review of 113 studies conducted in the United Kingdom. Goetzke and Vance [11] and Bento et al. [10] found comparable results in terms of fuel consumption’s reactivity to fuel prices and in contrast to vehicle mile travel’s response to fuel prices. Meanwhile, Oladosu [6] described individual family fuel consumption choices using a vehicle–fuel expenditure allocation model (or AIDS model) for multi-vehicle families in the United States.
In addition, there is a lack of consensus on the estimated coefficients in the studies that have been carried out in the process of modeling and predicting gasoline demand. According to Goetzke and Vance’s [11] review of the literature, the average gasoline price elasticity is roughly −0.18, with estimates ranging from −1.01 to 0.01. Thus, the majority of studies interpret these fuel price elasticities as evidence for the existence of a rebound effect, in which the cost savings associated with a reduction in the cost of driving or a gain in fuel efficiency eventually result in an unforeseen rise in fuel consumption. The rebound effect describes a situation in which drivers are presented with lower travel costs (such as falling gas prices) and/or increased fuel efficiency, which unintentionally results in increased fuel consumption and/or vehicle travel. Consequently, in terms of policy, such a phenomenon might result in erroneous calculations and, thus, incorrect interpretations for decision makers. Dimitropoulos et al. [12] conducted a meta-analysis of 74 studies that included 1120 estimates of reported rebound effects and discovered an average rebound effect of 10% to 12%. This unintended impact on driving and fuel consumption habits has significant consequences for the efficacy of policy planning and interventions aimed at reducing emissions and fuel consumption. Overall, studies in the transportation literature employ a variety of methods in terms of model selection, with the majority of scholars being aware of obvious causes for observed discrepancies in the findings. As a result, the effective modeling and forecasting of gasoline demand may offer a critical foundation for policymakers to consider the policy implications of their energy market activities, which is the goal of this paper.
At a minimum, typical forecasting models have two shortcomings: first, numerous studies have shown that predictors change over time, and factors, such as market cycles and macroeconomic policy changes, may result in structural breakdowns in the relationship between fundamental principles and dynamics. Additionally, the effect of each input on the dependent variable changes according to the period and market conditions [13,14]. A model with a static list of predictors may also lose accuracy and consistency over time. Extensive and precise analysis may be performed at any time to pick a model. In other words, if we have N predictors, we must evaluate and compare, 2 N models at each time point (the number of subsets of N variables that accurately represent all possible combinations and inclusions of N variables in the model) with T × 2 N as the total number of models should be tested throughout T. Therefore, while N and T are large, their analysis is impossible or, at least, difficult.
The accuracy of forecasts has been improved by using model averaging approaches, such as “forecasting combination”, in recent research. Both “Bayesian Model Averaging” (BMA) and “ forecasting combination” models are characterized by fixed weight values given to models throughout time; however, they do not offer sufficient flexibility to manage the time gap between the contributions of the modeling [15,16]. Therefore, dynamic model selection (DMS) and dynamic model averaging (DMA) were suggested by Raftery et al. [17] to overcome the limitations of the other models. Findings show that macroeconomic forecasting may benefit from this method [18,19]. The appropriateness of each model throughout time is shown in several studies on this subject. The time-varying parameter (TVP) model may employ DMA to compute the average likelihood of each variable being present in the best prediction model. As a more exact definition, one may argue that the average forecast across models is based on an average likelihood of the existence of a variable at time t based on prior knowledge [19,20,21]. Selecting the optimal prediction model is based on determining which variables have the greatest likelihood of being present in this model, and the model’s prediction will be based on this calculation [19]. Although DMS picks a model that comprises variables most likely to be included in forecast models among those estimated in each period, it does so in a more efficient manner. Inspired by the works of Koop and Korobilis [18] and Bork and Mller [22], Raftery et al. [17] found that the DMA model’s forecasting accuracy was 30 percent higher than that of other time-series approaches, such as AR and OLS regression. A DMA technique is presented by Wei and Cao [23] to predict a housing price increase in Chinese cities. Research shows that DMA is a better forecasting model than BMA, equal-weighted averaging (EW), and information-theoretic modeling. Dong and Yoon [24] employed a DMA approach to explore the global economic drivers that have a large impact on developing Asian stock market returns, notably during the financial crisis. Moreover, other applications for predicting are noteworthy: aggregate equity returns [25], commodity prices [26,27], exchange rates [28,29], Government bond yields’ term structure [27], and commodity price volatility and equity return [30].
Therefore, the following is the study’s primary contributions: (1) This study aims to estimate and forecast the gasoline demand in the USA using TVP techniques, particularly the DMA approach, which is much more accurate than prior methods. (2) In most investigations, Bayesian TVP is used to estimate the model’s parameters [31,32]. Although this approach approximates the generation of model parameters and switching probabilities using two forgetting elements, the inclusion of forgotten factors might be helpful since full Bayesian models may be quite large and time consuming in terms of computational volume. It also assumes that the two factors are constant over time, which is not the case for the single mechanism addressed in the study by Koop and Korobilis [18]. In addition, removing this constraint to reduce the computing cost of the model may lead to an improvement in model prediction accuracy. In this study, we attempt to execute a random process of forgetting factor selection using an algorithm called the ABC. Therefore, another key contribution in this work is to integrate ABC with DMA to improve the forecast accuracy.
The remaining parts of the article are organized as described below. In the second section, a research approach is presented. In Section 3, we provide a summary of both our data and the empirical findings of the forecasting. The conclusion is presented in Section 5.

2. Research Methodology

The DMA technique employed in the study at hand was introduced by Raftery et al. [17]. The following is the standard models for State-Space approaches, namely the Kalman filter:
y t = z t θ t + ε t
θ t = θ t 1 + μ t
where θ t = φ t 1 , β t 1 , γ t 1 , , γ t p denotes a vector of m × 1 coefficients, and μ t ~ N ( 0 , Q t ) and ε t ~ N ( 0 , H t ) with a mean of zero and variances of Q t and H t are normally distributed. y t denotes a dependent variable, and z t = 1 , x t 1 , y t 1 , , y t p denotes a 1 × m vector of variable interruption and intercept estimators depending on the model. As a consequence, the State-Space method is defined as follows, given a subset of K models at a given time:
y t = z t ( k ) θ t ( k ) + ε t ( k )
θ t + 1 ( k ) = θ t ( k ) + μ t ( k )
In this equation, ε t ( k ) ~ N ( 0 , H t ( k ) ) and μ t ( k ) ~ N ( 0 , Q t ( k ) ) with ϑ t = ( θ t 1 , , θ t k ) reveal which model of K subsets performs best during whatever period. Dynamic model averaging is a technique that permits a distinct model to be estimated at every given moment [19]. Raftery et al. [17] proposed a DMA approach that involves two parameters of α and λ , dubbed the forgetting factors. A recurrence estimate or forecast is feasible based on the information of conventional filtering when the constants H t and Q t are being considered. The following formula serves as the foundation for the Kalman filtering (KF) process:
θ t 1 y t 1 ~ N ( θ ^ t 1 , t 1 t 1 )
In Equation (5), the calculation of t 1 t 1 and θ ^ t 1 is performed using a conventional approach that is a function of H t and Q t , and then the KF process is performed using the following equation:
θ t y t 1 ~ N ( θ ^ t 1 , t t 1 )
Since t t 1 = t 1 t 1 + Q t , to simplify, Raftery et al. [17] substituted t t 1 = 1 λ t t 1 t 1 t 1 with t t 1 = t 1 t 1 + Q t , accordingly with 0 < λ 1 , Q t = ( 1 λ t t 1 1 ) t 1 t 1 . The value of λ t that is near to one suggests that the coefficients change more gradually. Raftery et al. [17] awarded it a value of 0.99 for the last five years’ quarterly statistical data; the preceding figure shows that the observations from the previous five years account for 80 percent of the most current observation. If it is 95%, it indicates that the most recent five years of data accounted for 35% of the weight of the earlier observation. As a result, it is critical to choose the forgetting factors, which are often believed to be between 95 and 99 percent. The estimate in the model will be completed by using updated estimators using the following functions:
λ t t = λ t 1 t 1
θ t y t ~ N ( θ ^ t , t t )
In which
θ ^ t = θ ^ t 1 + t t 1 z t H t + z t t t 1 z t 1 ( y t z t θ ^ t 1 )
t t = t t 1 t t 1 z t H t + z t t t 1 z t 1 z t t t 1
Recursive prediction operates based on the predictive distribution in the following manner:
y t y t 1 ~ N z t θ ^ t 1 , H t + z t t t 1 z t
Depending on the model, the above-mentioned functions for k may be expressed as follows, whereas the KF in the fixed estimators’ model can be represented as (5)–(7), using ϑ t as a vector of all parameters (3) and (4).
ϑ t 1 L t 1 = k , y t 1 ~ N ( θ ^ t 1 k , t 1 t 1 k )
ϑ t L t = k , y t 1 ~ N ( θ ^ t 1 ( k ) , t t 1 ( k ) )
ϑ t L t = k , y t ~ N ( θ ^ t ( k ) , t t ( k ) )
The value of θ ^ t ( k ) and ( t t k ) and ( t t 1 ( k ) ) was acquired with the use of KF and Equations (8) and (9) and t t 1 = 1 λ t t 1 t 1 t 1 . We employed the Raftery et al. [17] technique, which incorporates a forgetting factor termed α for state equations in various estimating models, and so the aforementioned components are analogous to the forgetting factor. Equation (4) is the starting point for the Kalman filter’s application. When DMA is utilized, similar effects are obtained:
P ( ϑ t 1 y t 1 ) = k = 1 K p ( θ t 1 k L t 1 = k , y t 1 ) Pr L t 1 = k y t 1
The model’s prediction function was replaced by the following equation introduced by Raftery et al. [17].
π t t 1 , k = π t 1 t 1 , k α t t 1 l = 1 K π t 1 t 1 , l α t t 1
If 0 α < 1 , the interpretation will be identical to that of λ , resulting in the following updated function:
π t t , k = π t t 1 , k α t t 1 p k ( y t y t 1 ) l = 1 K π t t 1 , l α t t 1 p l ( y t y t 1 )
α t t = α t 1 t 1
where p l ( y t y t 1 ) indicates the predictive density in terms of y . The weighted mean may be applied to the predictive outputs of each model by using π t t 1 , k to perform recursive prediction on those outputs. As a result, the DMA point prediction is as follows:
E ( y t y t 1 ) = k = 1 K π t t 1 , k z t k θ ^ t 1 k
DMS operates in such a manner that it picks the model with the greatest quantity of π t t 1 , k at any point in time. When α equals 0.99, the effectiveness of the previous 5 periods will account for 80% of the weighting for the current time. When α equals 0.99, 80 percent of the weighting for the current period will be determined by the performance of the preceding five periods. When α equals one, π t t 1 , k is precisely determined using the BMA model. Moreover, when λ equals one, BMA uses a traditional linear prediction model with constant coefficients.
Additionally, the suggested model’s recursive estimation will begin with past values for π 0 0 , k and θ 0 ( k ) :
E ( y t y t ) = k = 1 K π t t , k z t k θ ^ t 1 k
After calculating the equations, period t information is used to update the values. As previously stated, the purpose of including forgotten components is to minimize computational volume, as employing comprehensive Bayesian models may significantly increase computational volume. On the other hand, the sole process provided by Koop and Korobilis [18] is the manual selection of random values, which cannot result in plain values and also presupposes that the two parameters remain constant throughout time. In this work, we attempted to randomize the process for the selection of forgetting factors, α , and λ , using the ABC method. This approach is designed to decrease the sum of squared errors, which indicates the difference between computed and observed data. The mathematical expression is as follows:
M i n i m i z e   e t = ( y t E ( y t y t ) ) 2
The following is the pseudocode of the algorithm’s implementation procedure:
  • Step 1: Choose a curve fitting function. Equations (16) and (18) may be combined to create the following function:
    E ( y t y t ) = k = 1 K π t t 1 , k α t t p k ( y t y t 1 ) l = 1 K π t t 1 , l α t t p l ( y t y t 1 ) z t k θ ^ t 1 k
    whereas recursive prediction operates using predictive distributions in the following manner:
    y t y t 1 ~ N z t θ ^ t 1 , H t + z t 1 λ t t t 1 t 1 z t
  • Step 2. Arrange the greatest quantity of repetitions (MNC), the total number of bees (N), and LIMIT.
  • Step 3. Create random numbers for all bees for whom the optimization procedure begins with a preliminary estimate of their food supply, s , source using Equation (21).
    s w j n e w s = w j l o w + γ ( w j u p w j l o w ) , w l o w w j w u p s = 1 , , S N
    where SN represents the food supply in total. w j u p and w j l o w represent the top and lower limits of the j t h design variable, while γ is a random real value between zero and one.
  • Step 4. Calculate the objected function for all bees using Equation (19).
  • Step 5. Select fifty percent of the finest feeding places and appoint the bee who frequented these areas as the engaged bee.
  • Step 6. Set c y c l e = 1 .
  • Step 7. Traverse each source of food ( i = 1 , , S N )
    (a)
    Create new options for an employed bee using the following equation, where a new candidate food source ( s w j n e w ) is identified using two prior food source locations remembered by an employed bee ( s w j o l d ) and a randomly chosen neighborhood of a food source ( s w k o l d ):
    s w j n e w = s w j o l d + φ ( s w j o l d s w k o l d )
    (b)
    The old superscript displays the value of the preceding iteration’s design variable, but the new superscript displays existing design variables, where φ is a random positive integer between −1 and 1. k is a number that is chosen at random and is not equal to s.
    (c)
    Select the ideal dietary intake for each food source. The new place becomes the food source if there are more food sources there than there were at the previous location; otherwise, the previous location remains the food source.
  • Step 8. Estimate probability ( p i ) using the following equation:
    p i = i i = 1 S N i
    where i represents a measure of the solution’s fitness i, as determined by the employed bee. This corresponds to the nectar content in the food supply at location i.
  • Step 9. Traverse each source of food ( i = 1 , , S N ) .
    (a)
    Employ unemployed bees.
    (b)
    Utilizing Equation (21), develop novel employment strategies for jobless bees.
    (c)
    Check to see whether the amount of food sources has improved. If there is a considerable change, the observer bee will be promoted to the hired bee position; if there is no change, the candidate food source that the observer bee visited will not be selected.
  • Step 10. When the best food spot has not improved after a certain number of cycles (LIMIT), the hired bee switches to scout mode and uses Equation (20) to look for a new food source.
  • Step 11. c y c l e = 1 + cycle.
  • Step 12. Stop the operation if the cycle is M N C ; otherwise, go on to Step.
Another objective of this study aimed to compare the effectiveness of various prediction methods. The Mean Absolute Forecast Error (MAFE) and the Mean Squared Forecast Error (MSFE) are employed as standard indices in this research.
M S F E = τ = τ 0 T y τ E ( y τ D a t a τ h ) 2 T τ 0 + 1
M A F E = τ = τ 0 + 1 T y τ E ( y τ D a t a τ h ) T τ 0 + 1
where D a t a τ h is the data that were obtained from the time τ h , h is the horizon for time prediction, and E ( y τ D a t a τ h ) is the forecast point of y τ . This study begins with the results of DMA and DMS, followed by the events that determine which variables are most suited for predicting the gasoline demand function. Then, the performance of DMS and DMA is contrasted. In addition, it assesses the sensitivity of models and prediction results concerning the selection of forgetting factors.

3. The Estimated Model and Data

Annual observations for the United States from 1960 to 2020 were utilized in this analysis. Exogenous variables include measurements of demographic traits, economic activity, income, driving expenses, car pricing, and road availability. These variables in Table 1 are chosen based on an extensive review of the available literature.
Table 2, below, provides a brief description of the variables included in our analysis, as well as a definition and reference to the source. Moreover, summary statistics for the variables that are used in the empirical analysis are presented in Table 3.

4. Results

By comparing DMA predictions, we examine forecast performance. We attempted to empirically test several configurations of the ABC model to increase the accuracy of the forecast while achieving the quickest feasible computation speed. In conclusion, the number of bees was fixed at five, the greatest quantity of repetitions at five, and the lower and upper bound between 0.9% and 1%. Finally, we demonstrate the sensitivity of our findings to the choice of forgetting factors, α   a n d   λ . We provide findings for prediction horizons of one year (h = 1) and four years (h = 4). A prediction horizon of 4 means that we used the values of the independent variables in the previous 4 periods to predict the dependent variable in the current period. Obviously, with an increase in the prediction horizon, the prediction accuracy of the independent variables decreases. Our models all incorporate an intercept and a single lag between the dependent and independent variables. Experiments with lag lengths up to two revealed that a single lag produces the highest prediction results. Using the ABC approach, we sought to randomize the forgetting components in this study. Thus, our methodology not only provides for the automated determination of the two forgetting elements but also for their evolution over time to minimize the prediction model’s inaccuracy. These computations are carried out at a low computational cost. Thus, rather than selecting manually, we use a more precise selection mechanism. Figure 1 illustrates the outcome of estimating the components across time and the prediction horizons one and four. After estimating the model using the combined DMA-ABC model, the chance of each of the model’s independent variables being present is supplied. The posterior inclusion probability is shown in Figure 2 and Figure 3. That is, they quantify the likelihood that a predictor will help predict at time t. They are equivalent to the weights applied using DMA to models that incorporate a predictor. These graphs illustrate which predictors are significant at any given moment in time. These graphs demonstrate that DMS nearly always selects sparse models. These results are compelling evidence of model evolution. DMA has a significant theoretical advantage over other forecasting methodologies in that it enables the forecasting model to evolve. Of course, this gain may be negligible in a given empirical application if the forecasting model does not vary much over time. While the same trend remains true to a lesser degree, it is apparent that there is a significant change over time. That is, the forecasting model’s collection of predictors evolves with time. After 1980, practically all surface variables enter the model with varying probability. Intermittent values, of course, provide various outcomes. Between 2000 and 2015, the likelihood of existence, or the initial lag, of the majority of model variables is questioned. In comparison to other variables, vehicle registration has the lowest likelihood of being present, while public road mileage at the level and first log values indicate a high possibility of being included in gasoline demand forecasting.
Figure 4 illustrates the actual and predicted value of gasoline along with the forecast horizons h = 1 and h = 4. The accuracy of the model in estimating gasoline demand is seen in Figure 3. Additionally, expanding the prediction horizon resulted in a decline in the estimated model’s accuracy. Our earlier DMA and DMS findings were for our benchmark example, in which we used the ABC technique to determine a random forgetting factor that changes over time. As previously stated, researchers in this area use predetermined values for α   a n d   λ . As a consequence, Raftery et al. [17] used λ = α = 0.99 and suggest that the findings will remain resilient to accepting modifications in these variables. To test these claims of resilience, the results of our forecasting experiment utilizing different combinations of forgetting components are shown in Table 4. MSFE and MAFE values for various models of DMA-ABC, DMS-ABC, DMS, DMA, BMA, TVP-BMA, and TVP are provided in Table 4 for prediction horizons 1 and 4. The results of the comparison of several models in Table 4 indicate that the combined model of DMS and ABC, with the option of automatically acquiring forgetting factors over time, obtains the greatest results in forecasting gasoline demand. According to the DMA-ABC model, the mean values of the forgetting components are equal to α = 0.9449 and λ = 0.9662. Even taking the constant mean values of computational forgetting factors into account produced satisfactory results. It is noteworthy that the value α = 0.9449 enables relatively fast model evolution over time. This is similar to a previous tale we mentioned: it seems that allowing models to evolve is more significant than allowing parameters to vary with λ = 0.9662 for increasing forecast performance. The BME model (with λ = α = 1) does not have any dynamic approach, which means that while the estimated coefficients are constant over time, the input variables to the model are also constant over time. To investigate the effects of adding dynamics to the model in increasing the forecasting accuracy, we added two more columns to Table 4. In these two columns, the ratio of MAFE and MSFE of different models is calculated with the MAFE and MSFE values of the BMA model (with B index). Based on the results, the prediction error values in model DMS-ABC are about 0.82 of the prediction error in model BMA with a prediction horizon of 1. In addition, this value is equal to 0.76 in the forecast horizon of 4. Therefore, by increasing the prediction horizon, moving towards dynamic models leads to a further increase in prediction accuracy.

5. Conclusions and Implications

Accurate modeling and forecasting of gasoline demand may provide a valuable framework for policymakers to consider the policy implications of their energy market activities, which is the goal of this study. The primary shortcoming in prior forecasting models was their inability to accurately predict over time. Policymakers, on the other hand, should disregard short-term and temporary variations in gasoline demand in favor of economic stability. Therefore, the objective of this study was to develop a nonlinear dynamic model DMA-ABS to forecast gasoline consumption in the United States using annual data from 1960 to 2020. These models may be used to determine changes in both the input variables and the parameters of variables through time. The inclusion of two forgetting variables in the DMA model may be used to control the speed of such dynamics in the model, which has been previously determined manually in earlier research. In this work, we sought to implement a random process of forgetting factor selection using the ABC method. Therefore, one of the primary objectives of this study is to merge ABC with DMA to increase prediction accuracy. The findings of the DMS estimation model indicated that the input variables fluctuate with time, emphasizing the need of employing dynamic models rather than constant input variables for estimating gasoline demand.
Gasoline demand prediction helps policymakers to make informed decisions on issues related to energy security, environmental regulations, and transportation infrastructure. For instance, it can assist in determining the number of gas stations required to meet demand in a particular area, the type of fuel to be used in different transportation modes, and the amount of investment needed to maintain or upgrade the transportation infrastructure. Moreover, gasoline demand prediction can aid in managing the price of gasoline. It helps in determining the price level that will meet the demand and supply equilibrium. Therefore, it is recommended that in future research, the DMA model will be integrated with other evolutionary algorithms, such as particle swarm optimization (PSO), genetic algorithm (GA), etc., to compare the results and provide a more accurate prediction of the gasoline market through the expansion of the model presented in this research.

Author Contributions

Conceptualization, S.H.H.; Methodology, S.H.H.; Software, S.H.H.; Formal analysis, S.H.H.; Resources, Q.L.; Writing—original draft, S.H.H. and Q.L.; Writing—review & editing, Q.L.; Visualization, S.H.H.; Supervision, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

No funding has been received for this research.

Data Availability Statement

Data is available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. EIA. Annual Energy Outlook (AEO) Retrospective Review: Evaluation of AEO 2020 and Previous Reference Case Projections. Anal. Proj. U.S. Energy Inf. Adm. 2020, 92010, 1–15. [Google Scholar]
  2. McCarthy, P.S. Market Price and Income Elasticities of New Vehicle Demands. Rev. Econ. Stat. 1996, 78, 543. [Google Scholar] [CrossRef]
  3. Shen, C.; Linn, J. The Effect of Income on Vehicle Demand: Evidence from China’S New Vehicle Market; Resources for the Future: Washington, DC, USA, 2021. [Google Scholar]
  4. Dargay, J.; Gately, D.; Sommer, M. Vehicle ownership and income growth, worldwide: 1960–2030. Energy J. 2007, 28, 143–170. [Google Scholar] [CrossRef] [Green Version]
  5. Huo, H.; Wang, M. Modeling Future Vehicle Sales and Stock in China. Energy Policy 2012, 43, 17–29. [Google Scholar] [CrossRef]
  6. Oladosu, G. An Almost Ideal Demand System Model of Household Vehicle Fuel Expenditure Allocation in the United States. Energy J. 2003, 24, 1–21. [Google Scholar] [CrossRef]
  7. Kim, Y.; Kim, S. Forecasting Charging Demand of Electric Vehicles Using Time-Series Models. Energies 2021, 14, 1487. [Google Scholar] [CrossRef]
  8. Fouquet, R. Trends in Income and Price Elasticities of Transport Demand (1850-2010). Energy Policy 2012, 50, 62–71. [Google Scholar] [CrossRef] [Green Version]
  9. Graham, D.J.; Glaister, S. Road Traffic Demand Elasticity Estimates: A Review. Transp. Rev. 2004, 24, 261–274. [Google Scholar] [CrossRef]
  10. Bento, A.M.; Goulder, L.H.; Jacobsen, M.R.; Von Haefen, R.H. Distributional and Efficiency Impacts of Increased US Gasoline Taxes. Am. Econ. Rev. 2009, 99, 667–699. [Google Scholar] [CrossRef] [Green Version]
  11. Goetzke, F.; Vance, C. An Increasing Gasoline Price Elasticity in the United States? Energy Econ. 2021, 95, 104982. [Google Scholar] [CrossRef]
  12. Dimitropoulos, A.; Oueslati, W.; Sintek, C. The Rebound Effect in Road Transport: A Meta-Analysis of Empirical Studies. Energy Econ. 2018, 75, 163–179. [Google Scholar] [CrossRef]
  13. Ghysels, E.; Plazzi, A.; Valkanov, R.; Torous, W. Forecasting Real Estate Prices. Handb. Econ. Forecast. 2013, 2, 509–580. [Google Scholar] [CrossRef]
  14. Plakandaras, V.; Gupta, R.; Gogas, P.; Papadimitriou, T. Forecasting the U.S. Real House Price Index. Econ. Model. 2015, 45, 259–267. [Google Scholar] [CrossRef] [Green Version]
  15. Próchniak, M.; Witkowski, B. Time Stability of the Beta Convergence among EU Countries: Bayesian Model Averaging Perspective. Econ. Model. 2013, 30, 322–333. [Google Scholar] [CrossRef]
  16. Man, G. Competition and the Growth of Nations: International Evidence from Bayesian Model Averaging. Econ. Model. 2015, 51, 491–501. [Google Scholar] [CrossRef]
  17. Raftery, A.E.; Kárný, M.; Ettler, P. Online Prediction under Model Uncertainty via Dynamic Model Averaging: Application to a Cold Rolling Mill. Technometrics 2010, 52, 52–66. [Google Scholar] [CrossRef] [Green Version]
  18. Koop, G.; Korobilis, D. Forecasting Inflation Using Dynamic Model Averaging. Int. Econ. Rev. 2012, 53, 867–886. [Google Scholar] [CrossRef] [Green Version]
  19. Koop, G.; Korobilis, D. UK Macroeconomic Forecasting with Many Predictors: Which Models Forecast Best and When Do They Do So? Econ. Model. 2011, 28, 2307–2318. [Google Scholar] [CrossRef] [Green Version]
  20. Del Negro, M.; Primiceri, G.E. Time Varying Structural Vector Autoregressions and Monetary Policy: A Corrigendum. Rev. Econ. Stud. 2010, 82, 1342–1345. [Google Scholar] [CrossRef] [Green Version]
  21. Pesaran, M.H.; Timmermann, A. A Recursive Modelling Approach to Predicting UK Stock Returns. Econ. J. 2000, 110, 159–191. [Google Scholar] [CrossRef]
  22. Bork, L.; Møller, S.V. Forecasting House Prices in the 50 States Using Dynamic Model Averaging and Dynamic Model Selection. Int. J. Forecast. 2015, 31, 63–78. [Google Scholar] [CrossRef]
  23. Wei, Y.; Cao, Y. Forecasting House Prices Using Dynamic Model Averaging Approach: Evidence from China. Econ. Model. 2017, 61, 147–155. [Google Scholar] [CrossRef]
  24. Dong, X.; Yoon, S.M. What Global Economic Factors Drive Emerging Asian Stock Market Returns? Evidence from a Dynamic Model Averaging Approach. Econ. Model. 2019, 77, 204–215. [Google Scholar] [CrossRef]
  25. Naser, H.; Alaali, F. Can Oil Prices Help Predict US Stock Market Returns? Evidence Using a Dynamic Model Averaging (DMA) Approach. Empir. Econ. 2018, 55, 1757–1777. [Google Scholar] [CrossRef]
  26. Drachal, K. Forecasting Spot Oil Price in a Dynamic Model Averaging Framework—Have the Determinants Changed over Time? Energy Econ. 2016, 60, 35–46. [Google Scholar] [CrossRef]
  27. Byrne, J.P.; Cao, S.; Korobilis, D. Forecasting the Term Structure of Government Bond Yields in Unstable Environments. J. Empir. Financ. 2017, 44, 209–225. [Google Scholar] [CrossRef] [Green Version]
  28. Beckmann, J.; Koop, G.; Korobilis, D.; Schüssler, R.A. Exchange Rate Predictability and Dynamic Bayesian Learning. J. Appl. Econom. 2020, 35, 410–421. [Google Scholar] [CrossRef] [Green Version]
  29. Byrne, J.P.; Korobilis, D.; Ribeiro, P.J. On the Sources of Uncertainty in Exchange Rate Predictability. Int. Econ. Rev. 2018, 59, 329–357. [Google Scholar] [CrossRef] [Green Version]
  30. Ma, F.; Liu, J.; Wahab, M.I.M.; Zhang, Y. Forecasting the Aggregate Oil Price Volatility in a Data-Rich Environment. Econ. Model. 2018, 72, 320–332. [Google Scholar] [CrossRef]
  31. Avramov, D. Stock Return Predictability and Model Uncertainty. J. Financ. Econ. 2002, 64, 423–458. [Google Scholar] [CrossRef]
  32. Koop, G.; Potter, S. Forecasting in Dynamic Factor Models Using Bayesian Model Averaging. Econom. J. 2004, 7, 550–565. [Google Scholar] [CrossRef] [Green Version]
  33. Hughes, J.E.; Knittel, C.R.; Sperling, D. Evidence of a Shift in the Short-Run Price Elasticity of Gasoline Demand. Energy J. 2008, 29, 113–134. [Google Scholar] [CrossRef] [Green Version]
  34. Wadud, Z.; Graham, D.J.; Noland, R.B. Gasoline Demand with Heterogeneity in Household Responses. Energy J. 2010, 31, 47–74. [Google Scholar] [CrossRef] [Green Version]
  35. Rentziou, A.; Gkritza, K.; Souleyrette, R.R. VMT, Energy Consumption, and GHG Emissions Forecasting for Passenger Transportation. Transp. Res. Part A Policy Pract. 2012, 46, 487–500. [Google Scholar] [CrossRef]
  36. Lin, C.Y.C.; Prince, L. Gasoline Price Volatility and the Elasticity of Demand for Gasoline. Energy Econ. 2013, 38, 111–117. [Google Scholar] [CrossRef]
  37. Wang, T.; Chen, C. Impact of Fuel Price on Vehicle Miles Traveled (VMT): Do the Poor Respond in the Same Way as the Rich? Transportation (Amst) 2014, 41, 91–105. [Google Scholar] [CrossRef]
  38. Dillon, H.S.; Saphores, J.D.; Boarnet, M.G. The Impact of Urban Form and Gasoline Prices on Vehicle Usage: Evidence from the 2009 National Household Travel Survey. Res. Transp. Econ. 2015, 52, 23–33. [Google Scholar] [CrossRef]
  39. Hymel, K.M.; Small, K.A. The Rebound Effect for Automobile Travel: Asymmetric Response to Price Changes and Novel Features of the 2000s. Energy Econ. 2015, 49, 93–103. [Google Scholar] [CrossRef] [Green Version]
  40. Levin, L.; Lewis, M.S.; Wolak, F.A. High Frequency Evidence on the Demand for Gasoline. Am. Econ. J. Econ. Policy 2017, 9, 314–347. [Google Scholar] [CrossRef] [Green Version]
  41. Taiebat, M.; Stolper, S.; Xu, M. Corrigendum to “Forecasting the Impact of Connected and Automated Vehicles on Energy Use: A Microeconomic Study of Induced Travel and Energy Rebound”. Appl. Energy 2019, 256, 297–308. [Google Scholar] [CrossRef] [Green Version]
  42. Gillingham, K.T. The rebound effect and the proposed rollback of U.S. fuel economy standards. Rev. Environ. Econ. Policy 2020, 14, 136–142. [Google Scholar] [CrossRef]
  43. Chakraborty, D.; Hardman, S.; Tal, G. Integrating plug-in electric vehicles (PEVs) into household fleets-factors influencing miles traveled by PEV owners in California. Travel Behav. Soc. 2022, 26, 67–83. [Google Scholar] [CrossRef]
Figure 1. α and λ over time (h = 1, h = 4).
Figure 1. α and λ over time (h = 1, h = 4).
Energies 16 04795 g001
Figure 2. Posterior probability (h = 1).
Figure 2. Posterior probability (h = 1).
Energies 16 04795 g002
Figure 3. Posterior probability (h = 4).
Figure 3. Posterior probability (h = 4).
Energies 16 04795 g003
Figure 4. The actual and predicted value of gasoline demand in the forecast horizon h = 1 and h = 4 α = 0.9449; λ.
Figure 4. The actual and predicted value of gasoline demand in the forecast horizon h = 1 and h = 4 α = 0.9449; λ.
Energies 16 04795 g004
Table 1. Research literature for estimating gasoline demand function to determine model variables.
Table 1. Research literature for estimating gasoline demand function to determine model variables.
Author TypeDep. VariableInd. Variable
Hughes et al. [33]Time series OLSFuel demand/capitaGas price
Wadud et al. [34]RE panel (quarterly)Fuel demandGas price
Rentziou et al. [35]SURE panel model (annual)State VMTGas price
Lin & Prince [36]Dynamic times seriesFuel demand/capitaGas price
Wang & Chen [37]SEM (daily)Household VMTGas price
Dillon et al. [38]SEM (daily)Household VMTGas price
Hymel & Small [39]Simultaneous equationsState VMTFuel cost/mile
Levin et al. [40]FE panel (daily/monthly)Fuel demand/capitaGas price
Dimitropoulos et al. [12]Lit. review/meta-analysisFuel demand & VMTGas price
Taiebat et al. [41]microeconomic model (daily)Household VMTGas price
Gillingham [42]Lit. review/Lit. surveyUS VMTgas price
Goetzke & Vance [11]pooled OLSHousehold VMTGas price
Chakraborty et al. [43]OLS regressionTOT hh VMTfuel cost (non-PEV)
Table 2. Variables and definitions.
Table 2. Variables and definitions.
VariableDefinitionSource
G U / P O P G U / P O P = l n ( G U p o p u l a t i o n )
G U = Motor gasoline total end-use consumption
EIA
D P I / P O P D P I / P O P = l n ( D P I p o p u l a t i o n )
D P I = Disposable personal income, Billions of Dollars
FRED
R P R R P R / P O P = l n ( R P R p o p u l a t i o n )
R P R = Unleaded Regular Gasoline, U.S. City Average Retail Price
EIA
E F E E F E = l n ( F E )
F E = All Motor Vehicles Fuel Efficiency (Miles per Gallon)
EIA
L D / P O P L D / P O P = l n ( L D p o p u l a t i o n )
L D = Total Licensed Drivers
FHWA
V R / P O P V R / P O P = l n ( V R p o p u l a t i o n )
V R = Total Motor Vehicle Registrations For All Motor Vehicles
FHWA
O R P / P O P O R P / P O P = l n ( P R p o p u l a t i o n )
P R = Public Road Mileage
FHWA
FRED: Federal Reserve Economic Data; https://fred.stlouisfed.org, accessed on 20 June 2022; EIA: Energy Information Administration; https://www.eia.gov/, accessed on 20 June 2022; FHWA: Federal Highway Administration; https://highways.dot.gov/, accessed on 18 June 2022.
Table 3. Descriptive statistics.
Table 3. Descriptive statistics.
MeanMedianMaximumMinimumStd. Dev.
G U / P O P 9.2529.2699.4088.9950.095
D P I / P O P 2.5872.8573.9700.7371.006
R P R 0.0480.1521.293−1.1910.783
E F E 2.7082.7972.9012.4770.158
L D / P O P −0.462−0.401−0.359−0.7270.112
V R / P O P −0.381−0.290−0.173−0.8890.208
O R P / P O P 9.6559.6559.8889.4530.142
Table 4. Comparison of models.
Table 4. Comparison of models.
Prediction MethodMAFEMSFE M A F E M A F E B M S F E M S F E B MAFEMSFE M A F E M A F E B M S F E M S F E B
h = 1 h = 4
DMA-ABC0.553.601.010.98DMA-ABC0.714.860.960.96
DMS-ABC0.473.020.860.82DMS-ABC0.583.870.780.76
DMA λ = 0.9662; α = 0.94490.533.600.980.98DMA λ = 0.9698; α = 0.95560.714.850.950.96
DMS λ = 0.9662; α = 0.94490.443.000.810.81DMS λ = 0.9698; α = 0.95560.583.870.780.76
DMA λ = α = 0.990.543.650.990.99DMA λ = α = 0.990.735.030.980.99
DMS λ = α = 0.990.453.040.820.83DMS λ = α = 0.990.603.980.800.79
DMA λ = α = 0.950.553.571.010.97DMA λ = α = 0.950.724.980.970.98
DMS λ = α = 0.950.462.990.850.81DMS λ = α = 0.950.603.980.800.79
DMA λ = 0.95; α = 0.990.553.561.010.97DMA λ = 0.95; α = 0.990.724.950.970.98
DMS λ = 0.95; α = 0.990.462.990.850.81DMS λ = 0.95; α = 0.990.603.930.800.78
DMA λ = 0.99; α = 0.950.543.660.991.00DMA λ = 0.99; α = 0.950.745.040.990.99
DMS λ = 0.99; α = 0.950.453.040.820.83DMS λ = 0.99; α = 0.950.603.980.800.79
TVP- BMA ( λ = 1)0.553.681.001.00TVP- BMA ( λ = 1)0.755.081.001.00
BMA ( λ = α = 1)0.543.681.001.00BMA ( λ = α = 1)0.755.071.001.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hamza, S.H.; Li, Q. The Dynamics of US Gasoline Demand and Its Prediction: An Extended Dynamic Model Averaging Approach. Energies 2023, 16, 4795. https://doi.org/10.3390/en16124795

AMA Style

Hamza SH, Li Q. The Dynamics of US Gasoline Demand and Its Prediction: An Extended Dynamic Model Averaging Approach. Energies. 2023; 16(12):4795. https://doi.org/10.3390/en16124795

Chicago/Turabian Style

Hamza, Sakar Hasan, and Qingna Li. 2023. "The Dynamics of US Gasoline Demand and Its Prediction: An Extended Dynamic Model Averaging Approach" Energies 16, no. 12: 4795. https://doi.org/10.3390/en16124795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop