Next Article in Journal
Beyond the Hodrick Prescott Filter: Wavelets and the Dynamics of U.S.–Mexico Trade
Previous Article in Journal
Inclusive Turnout for Equitable Policies: Using Time Series Forecasting to Combat Policy Polarization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

The Forecasting of Aluminum Prices: A True Challenge for Econometric Models †

by
Krzysztof Drachal
* and
Joanna Jędrzejewska
Faculty of Economic Sciences, University of Warsaw, 00-241 Warszawa, Poland
*
Author to whom correspondence should be addressed.
Presented at the 11th International Conference on Time Series and Forecasting, Canaria, Spain, 16–18 July 2025.
Comput. Sci. Math. Forum 2025, 11(1), 13; https://doi.org/10.3390/cmsf2025011013
Published: 31 July 2025

Abstract

This paper explores the forecasting of aluminum prices using various predictive models dealing with variable uncertainty. A diverse set of economic and market indicators is considered as potential price predictors. The performance of models including LASSO, RIDGE regression, time-varying parameter regressions, LARS, ARIMA, Dynamic Model Averaging, Bayesian Model Averaging, etc., is compared according to forecast accuracy. Despite the initial expectations that Bayesian dynamic mixture models would provide the best forecast accuracy, the results indicate that forecasting by futures prices and with Dynamic Model Averaging outperformed all other methods when monthly average prices are considered. Contrary, when monthly closing spot prices are considered, Bayesian dynamic mixture models happen to be very accurate compared to other methods, although beating the no-change method is still a hard challenge. Additionally, both revised and originally published macroeconomic time-series data are analyzed, ensuring consistency with the information available during real-time forecasting by mimicking the perspective of market players in the past.

1. Introduction

Forecasting commodity prices, particularly aluminum, is essential for producers, consumers, investors, and policymakers. Aluminum is a key material for industries such as construction, aerospace, electronics, and automotive manufacturing. Accurate price forecasting is crucial for making informed decisions regarding investments, production, and trade. However, this is a complex task due to the volatility and unpredictability inherent in commodity markets in general, especially because prices are influenced by various economic, geopolitical, and financial factors, which makes forecasting a very challenging task [1,2,3].
Commodity markets, including aluminum, are affected by supply disruptions, demand fluctuations, geopolitical tensions, and changes in macroeconomic conditions. For example, variations in global industrial production, particularly in major aluminum-consuming countries like China, can significantly affect aluminum demand and prices. Additionally, the claimed financialization of commodities has introduced further complexity, where futures markets and speculative pressures can influence price movements alongside fundamental economic factors [4,5,6]. These dynamics make forecasting commodity prices even more complex, as it requires consideration of both market fundamentals and numerous other factors, leading to model uncertainty or feature selection problems. Furthermore, the relationship between a commodity price and its predictors can vary over time. On the other hand, in many cases, it is difficult to beat the simple no-change (NAIVE) or ARIMA method by generating more accurate forecasts [7,8,9,10,11,12]. Moreover, it is real-time data, which is quickly available to decisionmakers. In contrast, revised data provides a clearer picture of the past, but it is available with a time delay. Constructed models are usually trained using revised data, rather than the one that was truly available to market players in the past. As a result, this can generate another issue in econometric forecasting [13].
A promising approach to the problems above can be obtained through the application of Bayesian dynamic mixture models (BDMMs). These models incorporate model and parameter uncertainty, allowing forecasting adaptation as new information becomes available. However, these methods have not been studied much in applications to commodity prices. Therefore, this study contributes to the literature by comparing the accuracy of forecasts derived from various methods dealing with time-varying parameters and feature selection, as well as originally employing BDMMs in various specifications [14,15]. Furthermore, both real-time and revised datasets are employed.

2. Materials and Methods

2.1. Data

Monthly data from 04/2008 and 10/2024 were analyzed. LME (London Metal Exchange) spot prices (in USD per metric ton; monthly averages) for aluminum (unalloyed primary ingots, high grade, minimum 99.7% purity) were taken from World Bank [16]. This variable was denoted by p_aluminum.
As explanatory variables, excess demand (d_aluminum) was adopted following, for instance, Buncic and Moretto [2]. It was constructed as the difference between global consumption and the sum of global production and inventories. Consumption was proxied from the 2006 global estimates and assumed to mimic the time path of the OECD index of industrial production for China (2015 = 100, calendar and seasonally adjusted), due to a lack of exact data. Indeed, industrial production can be an effective proxy for this variable, and the role of China is huge on the analyzed market [17,18]. Inventories were taken as monthly averages of stocks on LME from Westmetall [19] in thousand metric tonnes. They were also taken as another explanatory variable (stocks_aluminum), as they can proxy speculative pressures. Production was taken in thousand metric tonnes from the International Aluminium Institute [20]. Following Buncic and Moretto [2], a convenience yield (cy_aluminum) was also taken (in percentages). It is constructed as
r_short − (f_aluminum − p_aluminum)/p_aluminum,
where r_short is the 3-month risk-free rate (3-month treasury bill: secondary market rate, TB3MS, monthly averages) taken from ALFRED and FRED [21,22], and f_aluminum denotes the official 3-month aluminum prices from LME in USD (^AH_3.F) taken from Stooq [23].
The US Industrial Production Index (INDPRO) was also taken into account [21,22]. It was denoted by ind_prod. The Index of Global Real Economic Activity (IGREA) proposed by Kilian [24] was obtained from ALFRED and FRED [21,22]. This was denoted by ec_act. Both ind_prod and ec_act were taken as explanatory variables resembling economic conditions.
Market stress was measured by two variables. First, the term spread, denoted by term_spread (i.e., U.S. 10-year treasury constant maturity minus U.S. 2-year treasury constant maturity, T10Y2YM, monthly averages) was used [21,22]. The second, monthly averages of the VIX index of implied volatility [25] were denoted by VIX.
Stock market behavior was measured by the S&P 500 index (^SPX) and the Shanghai Composite Index (^SHC). They were denoted by SP500 and SSE, respectively. This allows for including both the U.S. and Chinese economy—to capture global changes in economies and commodity markets [23]. Additionally, two leading aluminum sector companies’ stock prices were included, i.e., Rio Tinto Plc (RIO.UK, in GBX) and Alcoa Corp (AA.US, in USD), denoted by Rio_Tinto and Alcoa [26,27].
Moreover, two exchange rate currencies to USD of the largest primary aluminum producers were included, i.e., China (USDCNY) and Russia (USDRUB). They were denoted by fx_CNY and fx_RUB, respectively [23,28,29].
Finally, to capture the particular market past or autoregressive behaviors, the World Bank Base Metals Price Index (2010 = 100, USD, including aluminum, copper, lead, nickel, tin, and zinc), denoted by p_metals, was used [16].
Unless otherwise specified, for the above variables, the last observations were always taken into account (and closing and adjusted, if suitable).

2.2. Data Transformations

Similar to, for example, Buncic and Moretto [2], ordinary returns were assumed for most variables. In other words, time-series yt was transformed into (yt − yt−1)/yt where t denotes time. However, cy_aluminum, ec_act, term_spread, and VIX were not differentiated.
Next, following, for example, Koop and Korobilis [30], all variables were standardized into approximately stationary time-series. In other words, yt is a time-series, μ is its mean of the first T observations, and σ is its standard deviation of the first T observations. Then standardization is understood as the transformation of yt = (yt − μ)/σ. In particular, T = 100 was taken (and it was kept throughout the entire study as the in-sample period, for computing forecast accuracy, etc.). A whole sample was not used in order to omit the forward-looking bias. Furthermore, such a transformation allows all the time-series to have similar magnitudes, which improves computational issues in the applied models [31].
However, all forecast evaluation was carried out for raw (levels) aluminum prices; therefore, the predictions of returns obtained directly from the models (except ARIMA, NAÏVE, futures-based, and historical averages, which were computed directly for levels, i.e., raw data) were transformed back into levels. In other words, the estimations of models and their predictions were made over transformed variables, but for evaluation, the predictions of aluminum price were taken.
Explanatory variables were delayed one period back (i.e., first lags), except ec_act, where two periods back (i.e., second lags) were taken. This procedure resembles real-time information availability on the market [21].
Table 1 reports descriptive statistics, and Table 2 shows stationarity tests [32,33,34,35]. When assuming a 5% significance level for all variables, except for cy_aluminum, it can be assumed that they are stationary. Usual tests were applied: augmented Dickey–Fuller (ADF), Phillips–Perron (PP), and (a stationary-level version of) Kwiatkowski–Phillips–Schmidt–Shin (KPSS).

2.3. Models

As mentioned before, the first 100 observations were taken as in-samples. However, all models were estimated in a recursive manner (i.e., expanding window in-samples), meaning that a forecast for time t was generated from the model trained over all (monthly) data available up to time t − 1.
Firstly, various types of Bayesian dynamic mixture models (BDMMs) were estimated [14,15], both with state-space (SS) and normal regression (NR) components. Besides the original versions, weighted averaging schemes were also modified to include weighted averaged forecasting (A), forecasting from the component model with the highest posterior probability (H), and a median probability model (M) from Barbieri and Berger [36]. These were denoted as BDMM-SS, BDMM-SS-A, BDMM-SS-H, BDMM-SS-M, BDMM-NR, BDMM-NR-H, and BDMM-NR-M [37]. The details can be found in [14,15,37].
Next, Dynamic Model Averaging (DMA), Dynamic Model Selection (DMS), and DMS based on a median probability model (DMP) were estimated. For these models, the standard values of forgetting factors, i.e., 0.99, were taken [38]. Also, Bayesian Model Averaging (BMA), Bayesian Model Selection (BMS), and BMS based on the median probability model (BMP) were estimated. These are versions of the previous models, but with forgetting factors equal to 1. Besides these models, which consider averaging over all possible multilinear regression models (i.e., 2K, where K is the number of explanatory variables, herein K = 14), modified models—averaging over only one-variable simple linear regression models—were also estimated. This is because such models highly reduce computation time, as they consider K + 1 models only (due to the inclusion of a model with constant variable only as well). These were denoted as DMA-1VAR, DMS-1VAR, BMA-1VAR, and BMS-1VAR [39].
Also, LASSO, RIDGE, and Elastic Net (EL-NET) regressions were estimated [40]. Bayesian versions [41] were also estimated (B-LASSO and B-RIDGE). The penalty parameter was selected in a recursive manner (consistently with the spirit of this study) based on t-fold cross-validation (t being the time period) with respect to Mean Square Error (MSE). Mixing parameters {0.1, 0.2, …, 0.9} were applied.
Moreover, the least-angle regression (LARS) was estimated [42] with similar t-fold cross-validation based on MSE.
Additionally, time-varying regressions were estimated (which are equivalent to DMA with only one component model, i.e., the one with constant and all explanatory variables). This was denoted as TVP-FOR. Similarly, time-varying regression with no forgetting was estimated, i.e., the forgetting factor was set equal to 1 (TVP).
The recursively estimated ARIMA model [43] was also used in an automated way, as well as the no-change (NAÏVE) method. Historical-average forecasts were computed both as an expanding window, i.e., including all past historical observations (HA), and as a rolling window of the past 100 periods (HA-ROLL).
Finally, futures-based forecasts were computed [7,44], including the basic ones (FUT-SIMPLE), i.e., futures prices, which are taken directly as the subsequent forecast period, and those based on the spread (FUT-SPREAD).

2.4. Methods

Estimations were carried out in R [45]. The forecast accuracy was measured by Root Mean Square Error (RMSE) and its normalized version (N-RMSE). For N-RMSE, the value of RMSE is divided by the mean of the forecasted time-series. Also, Mean Absolute Error (MAE) and Mean Absolute Scaled Error (MASE) were computed [46]. Competing forecasts were compared pairwise with the Diebold–Mariano test [43,47] with Harvey et al.’s [48] correction method. The full set of forecasts was evaluated with the Model Confidence Set (MCS) from Hansen et al. [49]. Squared error loss functions were applied [50]. Additionally, the Giacomini and Rossi fluctuation test [51] was employed. The rolling parameter μ = 0.6 was used, which corresponds to approximately 5-year periods for the analyzed data [52].

3. Results

3.1. Main Results

Table 3 reports values of forecast accuracy measures. It can be seen that FUT-SPREAD generated the most accurate forecast. Also, FUT-SIMPLE is the second-most accurate forecast. The third best is DMA. This is according to each of the considered metrics. In the case of RMSE or N-RMSE, BDMMs do not perform very well, but if MAE or MASE is applied, BDMM-SS-A seems to be in top positions. Still, futures-based forecasts are the most accurate, DMA places next, and Bayesian versions of penalty regression models follow afterwards.
Table 4 reports outcomes from the Diebold–Mariano test. The forecast from the most accurate model, i.e., FUT-SPREAD, is compared with those from other methods. The same procedure is performed for the most accurate BDMM model, i.e., BDMM-SS-A, and for DMA. Assuming a 5% significance level, it can be seen that in the majority of cases, the competing method can be assumed to be significantly less accurate than FUT-SPREAD, except for FUT-SIMPLE, BDMM-SS-A, and DMA. In the case of the BDMM-SS-A method, only BDMM-SS and historical-average methods can be assumed to be less accurate. In the case of DMA, models such as BDMM-SS-M, BMS-1VAR, and TVP can additionally be assumed to be less accurate. If the absolute scaled error loss function, i.e., the one corresponding to the MASE metric, is considered, then the outcomes are even stronger. All models are less accurate than FUT-SPREAD. In the case of BDMM-SS-A, both time-varying parameter regressions (TVP and TVP-FOR) can additionally be assumed to be less accurate. But in the case of DMA, BDMM-SS-M cannot be assumed to be less accurate. Instead, BDMM-NR-H and BDMM-NR-M can be assumed as such, as well as both time-varying parameter regressions. (Due to the limited space herein, detailed outcomes are not included.)
Surprisingly, the MCS test did not reject any of the models (p-value 0.0836), meaning that no statistically significant difference in forecast accuracies can be detected by this test. This was achieved with a 5% significance level, a “Tmax” version of the test statistic, and 5000 bootstrapped samples, which was used to construct the test statistic. However, this test with respect to absolute scaled error loss function, i.e., the one corresponding to the MASE metric, rejected (p-value 0.1892) both models based on historical averages (HA and HA-ROLL).

3.2. Robustness Checks

The Diebold–Mariano test, when applied pairwise to all the considered models, but firstly estimated with data as they were released, and secondly with revised data (as in the last period of the time span of this research), does not detect any significant differences for any models.
In the case of the Giacomini and Rossi fluctuation test (BDMM-SS-A vs. FUT-SPREAD and DMA vs. FUT-SPREAD), assuming a 5% significance level, the null hypothesis that the models’ forecasting performance (i.e., BDMM-SS-A or DMA) is the same as that of FUT-SPREAD cannot be rejected (in favor of the alternative that a model performs worse than FUT-SPREAD). Details are presented in Figure 1.
The estimated models were also constructed by replacing the dependent variable with the aluminum LME cash official closing prices (AH_C.F) values of the month (instead of monthly average) [23]. In such a case, the two top minimizing RMSE models were NAÏVE and ARIMA. Futures-based forecasts were third and fourth. Interestingly, the next most accurate models were BDMM-SS and BDMM-SS-A. In the case of MAE metric, NAÏVE and ARIMA were still the most accurate, but BDMM-SS was the third most accurate model. Similarly, considering MASE, ARIMA was most accurate, and BDMM-SS was second best. Details are presented in Table 5.
However, in the case of MASE, none of the models obtained an MASE less than 1. Unfortunately, Diebold–Mariano tests (not reported herein, but available upon request) suggest that competing methods are less accurate than NAÏVE, and BDMM-SS and DMA, at least in most cases, are not significantly more accurate than other non-trivial models. However, the MCS test did not eliminate any model (p-value 0.0688).
Another modification was to replace the exchange rates by the exchange rates of countries that are biggest producers of ore (bauxite), i.e., Australia (AUDUSD) and the Republic of Guinea (GNF) [23]. For GNF, monthly averages were taken from FXTOP [53]. In such a case, futures-based forecasts were still the most accurate according to all metrics. However, for the next most accurate models, these included BMS, BMA, DMS, and BMP, i.e., versions of the general DMA scheme [38,39]. Interestingly, in the case of MAE and MASE, it was LARS, EL-NET, LASSO, and RIDGE that were the most accurate. Still, the MCS test did not reject any model (p-value 0.0996). Detailed outcomes are not reported herein, due to the limited space.

4. Discussion

This study compared various aluminum price forecasting models, which are able to deal with variable selection problems and time-varying regression coefficients. In other words, they can deal with problems where the importance of price predictor changes over time, as well as its relationship strength with the predicted price changes. It was found that futures-based forecasts (in particular, end-of-month closing prices) can be a very accurate method when the monthly average aluminum price is predicted. Also, Dynamic Model Averaging is a promising forecasting scheme, whereas Bayesian dynamic mixture models performed less accurately than they were expected to. This adds some new value towards research into the usability of futures-based forecasts on commodity markets. However, in the case of predicting closing monthly prices, futures-based forecasts (also end-of-month closing prices) are still performing well, but Bayesian mixture models also become an interesting method, especially the state-space component version, as well as the one with an averaging scheme incorporated, outperforming Dynamic Model Averaging. On the other hand, for closing price forecasting, beating the no-change or ARIMA method is still a hard challenge. All of these results are in line with previous studies [2,54,55,56,57].

Author Contributions

Conceptualization, K.D.; methodology, K.D.; software, K.D.; validation, K.D. and J.J.; formal analysis, K.D.; investigation, K.D. and J.J.; resources, K.D.; data curation, K.D. and J.J.; writing—original draft preparation, K.D. and J.J.; writing—review and editing, K.D. and J.J.; visualization, K.D.; supervision, K.D.; project administration, K.D.; funding acquisition, K.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in whole by National Science Centre, Poland, grant number 2022/45/B/HS4/00510.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Republishing the full data is unavailable due to copyright restrictions. The majority of the data applied in the study are openly available as cited in the text (suitable tickers provided) and in the References. The forecast data generated in the study are openly available in RepOD at https://doi.org/10.18150/ERBJKK.

Acknowledgments

This research was funded in whole by National Science Centre, Poland, grant number 2022/45/B/HS4/00510. For the purpose of Open Access, the author has applied a CC-BY public copyright licence to any Author Accepted Manuscript (AAM) version arising from this submission.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of the data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Alquist, R.; Kilian, L. What do we learn from the price of crude oil futures? J. Appl. Econom. 2010, 25, 539–573. [Google Scholar] [CrossRef]
  2. Buncic, D.; Moretto, C. Forecasting copper prices with dynamic averaging and selection models. N. Am. J. Econ. Financ. 2015, 33, 1–38. [Google Scholar] [CrossRef]
  3. Reynolds, D.B.; Umekwe, M.P. Shale-oil development prospects: The role of shale-gas in developing shale-oil. Energies 2019, 12, 3331. [Google Scholar] [CrossRef]
  4. Kahraman, E.; Akay, O. Comparison of exponential smoothing methods in forecasting global prices of main metals. Miner. Econ. 2023, 36, 427–435. [Google Scholar] [CrossRef]
  5. Kwas, M.; Paccagnini, A.; Rubaszek, M. Common factors and the dynamics of industrial metal prices: A forecasting perspective. Resour. Policy 2021, 74, 102319. [Google Scholar] [CrossRef]
  6. Liu, Q.; Liu, M.; Zhou, H.; Yan, F. A multi-model fusion based non-ferrous metal price forecasting. Resour. Policy 2022, 77, 102714. [Google Scholar] [CrossRef]
  7. Alquist, R.; Kilian, L.; Vigfusson, R.J. Forecasting the price of oil. In Handbook of Economic Forecasting; Elliott, G., Timmermann, A., Eds.; Elsevier: Amsterdam, Netherlands, 2013; Volume 2, Part A, pp. 427–507. [Google Scholar] [CrossRef]
  8. Gargano, A.; Timmermann, A. Forecasting commodity price indexes using macroeconomic and financial predictors. Int. J. Forecast. 2014, 30, 825–843. [Google Scholar] [CrossRef]
  9. He, Z.; Huang, J. A novel non-ferrous metal price hybrid forecasting model based on data preprocessing and error correction. Resour. Policy 2023, 86, 104189. [Google Scholar] [CrossRef]
  10. Ma, F.; Wahab, M.I.; Huang, D.; Xu, W. Forecasting the realized volatility of the oil futures market: A regime switching approach. Energy Econ. 2017, 67, 136–145. [Google Scholar] [CrossRef]
  11. Nguyen, D.K.; Walther, T. Modeling and forecasting commodity market volatility with long-term economic and financial variables. J. Forecast. 2020, 39, 126–142. [Google Scholar] [CrossRef]
  12. Pincheira, P.; Hardy, N. Forecasting aluminum prices with commodity currencies. Resour. Policy 2021, 73, 102066. [Google Scholar] [CrossRef]
  13. Croushore, D. Frontiers of real-time data analysis. J. Econ. Lit. 2011, 49, 72–100. [Google Scholar] [CrossRef]
  14. Nagy, I.; Suzdaleva, E. Mixture estimation with state-space components and Markov model of switching. Appl. Math. Model. 2013, 37, 9970–9984. [Google Scholar] [CrossRef]
  15. Nagy, I.; Suzdaleva, E.; Karny, M.; Mlynarova, T. Bayesian estimation of dynamic finite mixtures. Int. J. Adapt. Control Signal Process. 2011, 25, 765–787. [Google Scholar] [CrossRef]
  16. World Bank. Commodity Markets. 2025. Available online: https://www.worldbank.org/en/research/commodity-markets (accessed on 3 February 2025).
  17. Menzie, W.D.; Barry, J.J.; Bleiwas, D.I.; Bray, E.L.; Goonan, T.G.; Matos, G. The Global Flow of Aluminum from 2006 Through 2025: Open-File Report 2010–1256; U.S. Geological Survey: Reston, VA, USA, 2010. Available online: https://pubs.usgs.gov/of/2010/1256 (accessed on 3 February 2025).
  18. OECD. OECD Data Explorer. 2025. Available online: https://data-explorer.oecd.org (accessed on 3 February 2025).
  19. Westmetall. Market Data. 2025. Available online: https://www.westmetall.com/en/markdaten.php (accessed on 3 February 2025).
  20. International Aluminium Institute. Statistics. 2025. Available online: https://international-aluminium.org (accessed on 3 February 2025).
  21. ALFRED. Archival FRED. 2025. Available online: https://alfred.stlouisfed.org (accessed on 3 February 2025).
  22. FRED. Economic Data. 2025. Available online: https://fred.stlouisfed.org (accessed on 3 February 2025).
  23. Stooq. Stooq. 2025. Available online: https://stooq.pl/index.html (accessed on 3 February 2025).
  24. Kilian, L. Not all oil price shocks are alike: Disentangling demand and supply shocks in the crude oil market. Am. Econ. Rev. 2009, 99, 1053–1069. [Google Scholar] [CrossRef]
  25. Chicago Board Options Exchange. Historical Data. 2025. Available online: https://www.cboe.com/tradable_products/vix/vix_historical_data (accessed on 3 February 2025).
  26. Yahoo. Yahoo!finance. 2025. Available online: https://finance.yahoo.com (accessed on 3 February 2025).
  27. Shrivastava, P.; Vidhi, R. Pathway to sustainability in the mining industry: A case study of Alcoa and Rio Tinto. Resources 2020, 9, 70. [Google Scholar] [CrossRef]
  28. Barry, J.J.; Matos, G.R.; Menzie, W.D.U.S. Mineral Dependence—Statistical Compilation of U.S. and World Mineral Production, Consumption, and Trade, 1990–2010, Open-File Report 2013-1184; U.S. Geological Survey: Reston, VA, USA, 2013. Available online: https://pubs.usgs.gov/of/2013/1184 (accessed on 3 February 2025).
  29. Idoine, N.E.; Raycraft, E.R.; Hobbs, S.F.; Everett, P.; Evans, E.J.; Mills, A.J.; Currie, D.; Horn, S.; Shaw, R.A. World Mineral Production 2018–22; British Geological Survey: Keyworth, Nottingham, UK, 2024. Available online: https://nora.nerc.ac.uk/id/eprint/537241/1/World%20Mineral%20Production%202018%20to%202022.pdf (accessed on 3 February 2025).
  30. Koop, G.; Korobilis, D. Large time-varying parameter VARs. J. Econom. 2013, 177, 185–198. [Google Scholar] [CrossRef]
  31. Drachal, K. Determining time-varying drivers of spot oil price in a Dynamic Model Averaging framework. Energies 2018, 11, 1207. [Google Scholar] [CrossRef]
  32. Dickey, D.A.; Fuller, W.A. Distribution of the estimators for autoregressive time series with a unit root. J. Am. Stat. Assoc. 1979, 74, 427–431. [Google Scholar] [CrossRef]
  33. Phillips, P.C.B.; Perron, P. Testing for a unit root in time series regression. Biometrika 1988, 75, 335–346. [Google Scholar] [CrossRef]
  34. Kwiatkowski, D.; Phillips, P.C.B.; Schmidt, P.; Shin, Y. Testing the null hypothesis of stationarity against the alternative of a unit root. J. Econom. 1992, 54, 159–178. [Google Scholar] [CrossRef]
  35. Trapletti, A.; Hornik, K. Tseries: Time Series Analysis and Computational Finance. 2024. Available online: https://CRAN.R-project.org/package=tseries (accessed on 3 February 2025).
  36. Barbieri, M.M.; Berger, J.O. Optimal predictive model selection. Ann. Stat. 2004, 32, 870–897. [Google Scholar] [CrossRef]
  37. Drachal, K. “Dynmix”: An R package for the estimation of dynamic finite mixtures. SoftwareX 2023, 22, 101388. [Google Scholar] [CrossRef]
  38. Raftery, A.E.; Karny, M.; Ettler, P. Online prediction under model uncertainty via dynamic model averaging: Application to a cold rolling mill. Technometrics 2010, 52, 52–66. [Google Scholar] [CrossRef] [PubMed]
  39. Drachal, K. Dynamic Model Averaging in economics and finance with fDMA: A package for R. Signals 2020, 1, 47–99. [Google Scholar] [CrossRef]
  40. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef]
  41. Gramacy, R.B. Monomvn: Estimation for MVN and Student-t Data with Monotone Missingness. 2019. Available online: https://CRAN.R-project.org/package=monomvn (accessed on 3 February 2025).
  42. Hastie, T.; Efron, B. lars: Least Angle Regression, Lasso and Forward Stagewise. 2013. Available online: https://CRAN.R-project.org/package=lars (accessed on 3 February 2025).
  43. Hyndman, R.J.; Khandakar, Y. Automatic time series forecasting: The forecast package for R. J. Stat. Softw. 2008, 26, 1–22. [Google Scholar] [CrossRef]
  44. Ellwanger, R.; Snudden, S. Futures prices are useful predictors of the spot price of crude Oil. Energy J. 2023, 44, 65–82. [Google Scholar] [CrossRef]
  45. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2018; Available online: https://www.R-project.org (accessed on 3 February 2025).
  46. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef]
  47. Diebold, F.X.; Mariano, R.S. Comparing predictive accuracy. J. Bus. Econ. Stat. 1995, 13, 253–263. Available online: https://www.jstor.org/stable/1392155 (accessed on 3 February 2025). [CrossRef]
  48. Harvey, D.; Leybourne, S.; Newbold, P. Testing the equality of prediction mean squared errors. Int. J. Forecast. 1997, 13, 281–291. [Google Scholar] [CrossRef]
  49. Hansen, P.R.; Lunde, A.; Nason, J. The model confidence set. Econometrica 2011, 79, 453–497. [Google Scholar] [CrossRef]
  50. Bernardi, M.; Catania, L. The model confidence set package for R. Int. J. Comput. Econ. Econom. 2018, 8, 144–158. [Google Scholar] [CrossRef]
  51. Giacomini, R.; Rossi, B. Forecast comparisons in unstable environments. J. Appl. Econom. 2010, 25, 595–620. [Google Scholar] [CrossRef]
  52. Jordan, A.; Krueger, F. Murphydiagram: Murphy Diagrams for Forecast Comparisons. 2019. Available online: https://CRAN.R-project.org/package=murphydiagram (accessed on 3 February 2025).
  53. FXTOP. Historical Rates. 2025. Available online: https://fxtop.com/en/historical-exchange-rates.php (accessed on 3 February 2025).
  54. Drachal, K. Forecasting selected energy commodities prices with Bayesian dynamic finite mixtures. Energy Econ. 2021, 99, 105283. [Google Scholar] [CrossRef]
  55. Drachal, K. Forecasting prices of selected metals with Bayesian data-rich models. Resour. Policy 2019, 64, 101528. [Google Scholar] [CrossRef]
  56. Farag, M.; Snudden, S.; Upton, G. Can futures prices predict the real price of primary commodities? USAEE Work. Pap. 2024, 24, 629. [Google Scholar] [CrossRef]
  57. Kwas, M.; Rubaszek, M. Forecasting commodity prices: Looking for a benchmark. Forecasting 2021, 3, 447–459. [Google Scholar] [CrossRef]
Figure 1. Giacomini and Rossi fluctuation test. Horizontal black line represents critical value.
Figure 1. Giacomini and Rossi fluctuation test. Horizontal black line represents critical value.
Csmf 11 00013 g001
Table 1. Descriptive statistics.
Table 1. Descriptive statistics.
MinMaxMeanMedianStandard
Deviation
Coefficient of VariationSkewness
p_aluminum−3.30492.78340.06610.07780.887613.4292−0.3039
d_aluminum−2.81463.3251−0.03700.00800.9809−26.49730.0032
stocks_aluminum−2.08973.6537−0.1230−0.27540.9149−7.43621.2928
cy_aluminum−0.579811.97772.0681−0.09193.78961.83241.5975
ind_prod−7.56063.45810.03370.16681.086232.2483−3.8626
ec_act−2.04172.6619−0.0249−0.08660.8256−33.11130.6761
term_spread−5.69161.7419−1.5855−1.40151.9364−1.2213−0.1627
VIX−1.15564.1777−0.1539−0.39180.8850−5.74912.2702
SP500−3.74032.60160.06290.20840.985515.6561−0.6114
SSE−2.88432.41560.01530.02400.794551.8841−0.3745
p_metals−3.86202.87320.07650.09270.861811.2699−0.6471
fx_CNY−2.84974.02570.0650−0.04211.187118.25630.5568
fx_RUB−2.64253.8429−0.0584−0.14980.9966−17.07470.9778
Rio_Tinto−3.95672.74230.07540.04610.848711.2503−0.4827
Alcoa−4.54784.48370.10530.04721.229711.6781−0.0903
Table 2. Stationarity tests.
Table 2. Stationarity tests.
ADF StatisticADF p-ValuePP StatisticPP p-ValueKPSS StatisticKPSS p-Value
p_aluminum−5.5373<0.01−137.4591<0.010.1050>0.1
d_aluminum−6.5077<0.01−238.4936<0.010.1053>0.1
stocks_aluminum−3.86860.0168−66.5018<0.010.64010.0190
cy_aluminum−2.95440.1764−5.44270.80371.7829<0.01
ind_prod−5.9235<0.01−158.5210<0.010.1406>0.1
ec_act−4.0650<0.01−23.28350.03200.64120.0189
term_spread−3.43390.0505−14.16650.30813.2267<0.01
VIX−3.44980.0488−36.3286<0.010.64460.0186
SP500−6.1128<0.01−201.2408<0.010.1750>0.1
SSE−5.4294<0.01−191.9523<0.010.0866>0.1
p_metals−6.1435<0.01−118.8156<0.010.0518>0.1
fx_CNY−6.1865<0.01−182.8249<0.010.1750>0.1
fx_RUB−5.9000<0.01−151.7995<0.010.0399>0.1
Rio_Tinto−6.5383<0.01−185.9790<0.010.0488>0.1
Alcoa−5.1368<0.01−206.0026<0.010.1573>0.1
Table 3. Forecast accuracy measures.
Table 3. Forecast accuracy measures.
RMSEMAEN-RMSEMASE
BDMM-SS146.050698.39880.06811.2168
BDMM-SS-A110.654574.83060.05160.9254
BDMM-SS-H120.579677.90830.05620.9634
BDMM-SS-M122.436679.92030.05710.9883
BDMM-NR113.851681.65850.05311.0098
BDMM-NR-H115.326883.63390.05381.0343
BDMM-NR-M115.558883.03970.05391.0269
DMA103.051473.96860.04800.9147
DMS105.172875.80290.04900.9374
DMP103.659276.02640.04830.9402
DMA-1VAR104.597976.19000.04880.9422
DMS-1VAR105.615077.71380.04920.9610
BMA103.366975.39220.04820.9323
BMS103.162675.39310.04810.9323
BMP104.696777.20490.04880.9548
BMA-1VAR110.110178.88180.05130.9755
BMS-1VAR112.830580.74670.05260.9986
LASSO106.533675.32490.04970.9315
RIDGE105.142375.14960.04900.9293
EL-NET106.687475.81480.04970.9376
B-LASSO103.273774.57420.04810.9222
B-RIDGE103.276974.70000.04810.9238
LARS105.940975.13290.04940.9291
TVP116.533089.43480.05431.1060
TVP-FOR112.968984.29380.05271.0424
ARIMA109.141176.92690.05090.9513
NAIVE113.476580.86390.0529
FUT-SIMPLE89.369862.71080.04170.7755
FUT-SPREAD88.934062.10250.04150.7680
HA412.0597306.53650.19213.7908
HA-ROLL431.3150306.87700.20113.7950
Table 4. Diebold–Mariano tests.
Table 4. Diebold–Mariano tests.
FUT-SPREAD StatisticFUT-SPREAD p-ValueBDMM-SS-A StatisticBDMM-SS-A p-ValueDMA
Statistic
DMA
p-Value
BDMM-SS2.83520.00282.37650.00972.68240.0043
BDMM-SS-A1.65320.0507 1.45480.0745
BDMM-SS-H1.99600.02441.09900.13721.61070.0552
BDMM-SS-M2.16890.01631.26860.10381.78900.0384
BDMM-NR2.51720.00670.32450.37311.37400.0863
BDMM-NR-H2.65890.00460.48230.31531.59150.0574
BDMM-NR-M2.61540.00520.48200.31541.52040.0658
DMA1.53520.0640−1.45480.9255
DMS2.04530.0218−0.76720.77760.75920.2248
DMP1.76630.0402−1.00540.84140.22220.4123
DMA-1VAR2.04900.0216−0.82660.79480.50790.3063
DMS-1VAR2.19260.0154−0.67780.75030.78930.2159
BMA1.97970.0253−0.95060.82790.10680.4576
BMS2.02540.0228−0.92090.82030.03060.4878
BMP2.22050.0143−0.73550.76810.45190.3262
BMA-1VAR2.57830.0057−0.06470.52571.31090.0965
BMS-1VAR2.85620.00260.25910.39811.73490.0430
LASSO1.90720.0297−0.72620.76531.46560.0730
RIDGE1.73790.0427−0.95090.82800.85290.1979
EL-NET1.92790.0284−0.70050.75741.54230.0631
B-LASSO1.94960.0270−0.95540.82910.07060.4719
B-RIDGE1.89730.0304−0.98170.83570.07640.4696
LARS1.76620.0402−0.83080.79590.94690.1730
TVP2.99890.00170.63640.26301.98200.0251
TVP-FOR2.43480.00840.27010.39381.51000.0671
ARIMA1.86770.0324−0.16540.56550.88700.1886
NAIVE2.43600.00830.29450.38451.33960.0917
FUT-SIMPLE1.25510.1062−1.63250.9471−1.49540.9310
FUT-SPREAD −1.65320.9493−1.53520.9360
HA4.67420.00004.64470.00004.67050.0000
HA-ROLL4.50220.00004.47550.00004.49900.0000
Table 5. Forecast accuracy measures (forecasted monthly closing prices).
Table 5. Forecast accuracy measures (forecasted monthly closing prices).
RMSEMAEN-RMSEMASE
BDMM-SS206.8595142.57730.09571.4725
BDMM-SS-A209.1845149.70940.09681.5461
BDMM-SS-H219.1199158.57950.10141.6377
BDMM-SS-M217.7943157.76490.10081.6293
BDMM-NR220.4677161.50140.10201.6679
BDMM-NR-H219.0531160.61520.10141.6588
BDMM-NR-M220.8676161.81590.10221.6712
DMA218.9025157.98560.10131.6316
DMS218.0156159.38620.10091.6461
DMP218.8284160.40630.10131.6566
DMA-1VAR219.4383160.46480.10161.6572
DMS-1VAR218.5520159.80070.10121.6504
BMA218.6148159.37020.10121.6459
BMS219.7133161.25180.10171.6653
BMP219.6996161.33290.10171.6662
BMA-1VAR219.7482161.04090.10171.6632
BMS-1VAR219.3107160.69290.10151.6596
LASSO217.7136158.59580.10081.6379
RIDGE217.8733157.85990.10081.6303
EL-NET218.0318158.82630.10091.6403
B-LASSO219.2751160.00910.10151.6525
B-RIDGE218.5653159.39730.10121.6462
LARS217.9196158.27460.10091.6346
TVP226.3519159.97960.10481.6522
TVP-FOR223.6009158.24340.10351.6343
ARIMA134.716799.35320.06241.0261
NAIVE132.167496.82830.0612
FUT-SIMPLE205.0920148.40400.09491.5327
FUT-SPREAD204.3735148.30780.09461.5317
HA420.2411311.48230.19453.2169
HA-ROLL436.6985311.46610.20213.2167
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Drachal, K.; Jędrzejewska, J. The Forecasting of Aluminum Prices: A True Challenge for Econometric Models. Comput. Sci. Math. Forum 2025, 11, 13. https://doi.org/10.3390/cmsf2025011013

AMA Style

Drachal K, Jędrzejewska J. The Forecasting of Aluminum Prices: A True Challenge for Econometric Models. Computer Sciences & Mathematics Forum. 2025; 11(1):13. https://doi.org/10.3390/cmsf2025011013

Chicago/Turabian Style

Drachal, Krzysztof, and Joanna Jędrzejewska. 2025. "The Forecasting of Aluminum Prices: A True Challenge for Econometric Models" Computer Sciences & Mathematics Forum 11, no. 1: 13. https://doi.org/10.3390/cmsf2025011013

APA Style

Drachal, K., & Jędrzejewska, J. (2025). The Forecasting of Aluminum Prices: A True Challenge for Econometric Models. Computer Sciences & Mathematics Forum, 11(1), 13. https://doi.org/10.3390/cmsf2025011013

Article Metrics

Back to TopTop