Next Article in Journal
Bi-Objective Scheduling Optimization for Discrete Time/Cost Trade-Off in Projects
Previous Article in Journal
A Framework for Spatiotemporal Analysis of Regional Economic Agglomeration Patterns
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Some Novel Bayesian Model Combination Schemes: An Application to Commodities Prices

by
Krzysztof Drachal
Faculty of Economic Sciences, University of Warsaw, 00-241 Warsaw, Poland
Sustainability 2018, 10(8), 2801; https://doi.org/10.3390/su10082801
Submission received: 10 July 2018 / Revised: 2 August 2018 / Accepted: 3 August 2018 / Published: 7 August 2018

Abstract

:
Forecasting commodities prices on vividly changing markets is a hard problem to tackle. However, being able to determine important price predictors in a time-varying setting is crucial for sustainability initiatives. For example, the 2000s commodities boom gave rise to questioning whether commodities markets become over-financialized. In case of agricultural commodities, it was questioned if the speculative pressures increase food prices. Recently, some newly proposed Bayesian model combination scheme has been proposed, i.e., Dynamic Model Averaging (DMA). This method has already been applied with success in certain markets. It joins together uncertainty about the model and explanatory variables and a time-varying parameters approach. It can also capture structural breaks and respond to market disturbances. Secondly, it can deal with numerous explanatory variables in a data-rich environment. Similarly, like Bayesian Model Averaging (BMA), Dynamic Model Averaging (DMA), Dynamic Model Selection (DMS) and Median Probability Model (MED) start from Time-Varying Parameters’ (TVP) regressions. All of these methods were applied to 69 spot commodities prices. The period between Dec 1983 and Oct 2017 was analysed. In approximately 80% of cases, according to the Diebold–Mariano test, DMA produced statistically significant more accurate forecast than benchmark forecasts (like the naive method or ARIMA). Moreover, amongst all the considered model types, DMA was in 22% of cases the most accurate one (significantly). MED was most often minimising the forecast errors (28%). However, in the text, it is clarified that this was due to some specific initial parameters setting. The second “best” model type was MED, meaning that, in the case of model selection, relying on the highest posterior probability is not always preferable.

1. Introduction

During the 2000s, a commodities price boom was observed. Naturally, this led to the question of what was the purpose of such price spikes. Some of the hypotheses were based on the growing financialization of commodities markets. In other words, it was observed since the 1990s that the links between commodities prices and equity markets become more and more tight. For the previous periods, more research emphasis was put on the fundamental factors like supply and demand, etc.
The problem of sustainability is clearly seen in agricultural commodities. Especially in recent years, this has been noticed not only by the final goods’ consumers, but also by the producers. Nevertheless, such initiatives spread over whole commodities’ markets. Sustainable products might generate more stable revenues to producers, stakeholders can reduce their risk exposure, consumers become more aware of natural resources. Within this context, it seems interesting to construct a tool which would be able to model commodities prices, and which would be able to deal with variable uncertainty, as well as having its initial feature of time-varying coefficients. These adaptive abilities might make it a superior method in a highly volatile market background.
Forecasting commodity prices is a very difficult task. The reason for this is that the structure and the behaviour of commodities markets can be quite complex. For example, certain nonlinear effects can emerge, which, furthermore, can lead to chaotic behaviours. However, even without incorporating the developed conceptual machinery, many previous studies indicated that, in different time periods, different factors are the leading determinants of the prices of various commodities [1,2,3]. Therefore, in some periods, certain models work quite well, but, in the other periods, some other models work much better.
In other words, when a longer time horizon is analysed, the uncertainty about the true model arises (i.e., the set of predictors). However, except for the model uncertainty, there are strong arguments to also allow for time-varying parameters in the considered models (i.e., to allow regression coefficients, if the regression models are considered, to vary in time). This corresponds to possible changes in the strength (or even in the direction) of the impact of a predictor on the considered commodity price.
Naturally, this problem can be tackled with the Bayesian methods, and indeed, Bayesian Model Averaging (BMA) and its slight modification, i.e., Bayesian Model Selection (BMS), have been recently used in various branches of economics with a success. In order to extend this method, recently, a new framework was proposed [4]. Dynamic Model Averaging (DMA) allows for both time-varying: the state space of the model and its parameters (i.e., regression coefficients). Moreover, the final DMA forecast is computed as a weighted forecast from the predictions of all included models. However, the weights used are updated recursively with respect to the quality of forecast (i.e., predictive density) produced by the models in preceding periods.
A few years after presenting the theoretical foundations of this method, a rapid interest in it for the purpose of economics and finance can be observed. Of course, most of the published results provide arguments in favour of DMA, which can beat some alternative forecasts. Therefore, the first aim of this paper is to report a unified and consistent application of DMA to possibly large number of commodities. In addition, to provide some hints about the “calibration” of the initial parameters for this method. In particular, 69 time-series provided by The World Bank [5] were taken. The second aim of this paper is to compare DMA with the other, similar Bayesian methods. In other words, to verify whether DMA is superior to the previously proposed averaging method, and whether the model averaging itself improves the forecast accuracy in the case of commodity prices. Following Gargano and Timmermann [6], this study is based on macroeconomic and financial predictors. The second motivation behind using macroeconomic and financial predictors is to obtain a common set of predictors and use quite numerous time-series of commodity prices.
Indeed, this paper tries to fill some literature gaps. Contrary to previous DMA studies focusing on one approach to a relatively small number of independent variables (for example, [7,8,9], here, applying a similar pattern to possibly many commodities is attempted to see if some pattern will be recovered. However, this is not a somehow veiled data dredging. It is obvious that a more tailored data set of predictors can result in higher forecast accuracy. On the other hand, if certain research argues that some macroeconomic and financial variables (therefore common ones for different commodities) can serve as satisfactory commodities prices’ predictors [6], then it is worth checking if DMA can produce more accurate forecasts in comparison to the existing methods. In other words, the results of this paper might provide some arguments that, indeed, DMA forecast combination scheme is highly beneficial. However, they might also bring some shadow on DMA and provide rather the following argument: the already reported success of DMA (see, for example [10,11,12]) is mostly due to the suitable choice of predictors (and analysed period), not due to the method itself.
In the next section, a brief overview of the results already obtained with DMA is presented. In addition, the short review motivating the selection of possible commodity price predictors is given. It is also discussed why model averaging can lead to better forecast performance. Next, data are described in details. Furthermore, a short overview of DMA and other methods used in this research is sketched. Next, the results are presented and discussed. Moreover, some preliminary results, motivating the selection of the initial parameters for the main DMA models are discussed. This is believed to be of some help for future researches based on DMA methods. For the reader’s convenience, the list of abbreviations is given at the end of the paper.

2. Literature Review

This section is divided into two parts. First, which fields DMA has already been applied in is summarized. Next, a few recent research papers are presented with a stress put on the novelty of the methodology. Finally, which variables can be initially considered as potentially useful predictors in the construction of the econometric model of the selected commodities is presented.

2.1. Dynamic Model Averaging

DMA has already been applied to model various time-series and, in the case of commodities: copper, crude oil and gold. Aye et al. [13] used DMA to forecast gold price returns. Most of the predictors were chosen on the basis of a recursive principal component analysis. Both DMA and DMS were found to be superior to BMA and the naive forecast. In addition, Baur et al. [10] found similar conclusions. Risse and Ohl [14] compared the DMA improved by the Dynamic Occam’s Window of Onorante and Raftery [15] with the original DMA and some other machine learning techniques. They found that such a modification produces the best forecasts of gold prices.
Buncic and Moretto [7] found that DMA and DMS produces more accurate forecasts than the historical mean and the naive forecast in case of copper price. They used as predictors both macroeconomic and financial variables as well as fundamental factors specific for copper. For example, supply and demand, returns from specific stocks (firms heavy depending on commodity market), etc. Drachal [8] found that, in case of crude oil prices, it is quite hard to produce a statistically significant more accurate forecast than the naive one and autoregressive models with DMA. However, these benchmark forecasts were also not found to be superior to DMA and DMS. On the other hand, Naser [11] used DMA with a large set of predictors, and concluded that DMA outperforms alternative models as well as forecasts based on future prices.
Baxa et al. [16] successfully applied DMA to forecast inflation in G7 countries. Similarly, inflation was modeled with DMA by Del Negro et al. [17]. However, Di Filippo [18] concluded that, in the case of inflation, DMS produced better forecasts than DMA. In other words, model selection was more appropriate than model averaging. Ferreira and Palma [19] concluded that, in the case of modeling, the Brazilian inflation DMA behaves very well. These studies were usually done in the context of the generalized Phillips curve [20]. Indeed, such studies were usually made in a general macroeconomic context. For example, the economic growth was also modeled with DMA [21,22].
DMA was also used in real property markets. Bork and Moller [23] used it (and also DMS) to forecast house prices in the US. Risse and Kern [9] found DMA a superior method for forecasting house prices in the Euro area countries. Wei and Cao [12] applied DMA to model house prices in major Chinese cities. They also implemented a slight modification of the original DMA method. They concluded that this method is more accurate than BMA, information-theoretic averaging, and that, indeed, the time-varying weights’ scheme significantly improves the forecast quality.
De Bruyn et al. [24] applied DMA to forecast South African rand to the US dollar exchange rate and to the UK pound. The comparison was made with BMA and the random walk model. DMA was found to be superior to the benchmark models. Gupta et al. [25] examined with DMA the growth of China’s foreign exchange reserves. DMA and DMS were found superior to the random walk, recursive OLS-AR(1), recursive OLS and BMA.
DMA was also found useful in modeling carbon prices [26]. Liu et al. [27] used this method to forecast the realized range-based volatility. They concluded that time-varying weights indeed generate more accurate forecasts. Similar conclusions were derived by Wang et al. [28]. Finally, Naser and Alaali [29] applied a DMA framework to model the US stock prices’ returns.

2.2. Other Models

Indeed, the fact that forecast combination can lead to a more accurate forecast than the selection of one model is both an empirical and theoretically understood fact. The seminal paper in this context was published quite long ago [30]. In case of commodities, the recent findings support averaging over various models [31] as well as the necessity to consider a possibly wide set of predictors [32,33]. However, there are various methods of forecast combinations [34,35] and, yet, within the context of data-rich models, it is natural to focus on the Bayesian methodology [36,37]. This is, for example, due to the fact that, if the number of predictors exceeds the number of observations, then the conventional methods, like OLS are no longer meaningful. On the other hand, the Bayesian formulas are still reasonable in such a case. For example, recently, Lee and Huh [38] proposed some improvements in choosing informative priors for Bayesian models of crude oil price. Yin et al. [39] found that replacing the predictive densities with modified hit ratios in DMA weights’ construction can result in forecast accuracy improvement for oil prices.
Of course, the Bayesian approach is not the only one, which has been applied to commodity prices. However, except for purely classical and fundamental methods like multilinear regression models, etc., various improvements were applied. This was done in order to capture the specific features of the commodity market. For example, Kriechbaumer et al. [40] joined a well-known ARIMA modeling with a wavelet decomposition approach to forecast metals’ prices. The motivation behind using the wavelet analysis was to encompass in the model the cyclical behaviour of metals prices. The obtained framework allowed for increasing the accuracy of one-month ahead forecasts in comparison with the basic ARIMA models. Wang and Sun [41] studied the relationship between oil prices and economic activity and political tension with a structural equation model.
In addition, VAR/VECM methods were applied in the case of bidirectional relationships and the presence of many variables influencing each other. For example, Cross and Nguyen [42] focused on links between China’s economic growth and global oil market fluctuations. They modified the basic VAR methodology to include time-varying parameters’ VAR. Gangopadhyay et al. [43] used VECM methodology to model gold prices in the Indian market, investment decision and inflation.
Gil-Alana et al. [44] used the fractional integration modeling framework to identify structural breaks in metals prices. Kim and Jung [45] improved the predictive ability of standard forecasting models of crude oil prices by a functional partial least squares approach. This was done to deal with multicollinearity of predictors. The framework was compared with principal component regression and least absolute shrinkage and selection operator models.
In the case of a nonconventional approach, neural networks and various types of machine learning techniques are also popular. The review of such methods can be found, for example, in the paper by Hamdi and Aloui [46]. Liu et al. [47] proposed a machine learning algorithm with a decision tree. This method was able to significantly improve copper price forecasting. Zhao et al. [48] focused on the problems with uncertainty about the predictors of oil price. The study was based on almost 200 time-series. They proposed a certain deep learning ensemble approach. They joined an advanced deep neural network model to capture the nonlinearities and an ensemble method to generate multiple training data sets. The implemented algorithm was able to outperform the alternative models.

2.3. Commodities Prices Predictors

In the case of particular commodities prices’ predictors this research follows Gargano and Timmermann [6]. They selected various macroeconomic and financial predictors, which can be common for different commodities. In other words, they were using predictors like interest rates, term spreads, industrial production growth, etc., whereas no specific predictors suitable for just one particular commodity (like, for example, production quotas) was used. Their selection and construction of predictors was much influenced by the studies on stock prices and equity premium forecasting [49]. Of course, this is some kind of a limitation. Indeed, the fundamental factors, for example, inventory quotas, are known to have a significant impact on commodities prices [50]. Nevertheless, the narrowing to macroeconomic and financial predictors is treated as another modeling approach, rather than a sign of ignoring the importance of other predictors. Moreover, such an approach is quite often adapted [51].
It has already been tried to find such common predictors for various commodities in general. For example, Kagraoka [52] within a generalized dynamic factor model concluded that the U.S. inflation rate, the world industrial production, the world stock index and the price of crude oil impact commodities prices. Such methodology was also used by Lübbers and Posch [53] in the analysis of futures prices.
However, even based on a different methodological approach Arango et al. [54], as well as Byrne et al. [55] provided arguments that interest rates significantly impact numerous commodities prices. Alam and Gilbert [56] within a VAR framework concluded that the impact of monetary policy, global economic conditions and U.S. dollar exchange rates are significant for agricultural commodities. However, this study was based also on commodity-specific predictors like inventories quotas, etc. The motivation of this study was similar to the one presented herein: to analyse a big data set of commodities and find common predictors. In case of analysing the impact of exchange rates on commodities prices, a more detailed research was done by, for example, Chen et al. [57].
Much research was also devoted to links between commodities spot and futures prices [58]. However, Hong and Yogo [59] concluded that open interest can serve as an even better proxy than future prices itself. Chen [60] provided arguments that commodity-sensitive stock prices contain an important information for forecasting commodities prices. Indeed, such predictors were used in the already mentioned DMA study of copper prices by Buncic and Moretto [7].
Finally, it should be stressed that, except for various already existing methodologies and continual improvements and development of new ones, forecasting commodities prices remains a very difficult task [61].
Summarizing this considerations, it can be seen that there is a recent rapid increasing interest in the methods combining time-varying parameters and uncertainty about the predictors (or dealing with data-rich models). Economically it is motivated by the increasing complexity of the markets, and time-varying relationships between predictors and commodity prices: both in a sense of the strength and the direction, and even in a sense which predictors are important in particular moments of time. From the mentioned literature of these novel methods, a certain group can be classified, i.e., BMA, BMS, DMA and DMS. All of them emerge from time-varying parameters (TVP) regressions with the Kalman filter. The differences between these methods lie in whether from numerous models the final forecast emerge through the model averaging or model selection—in addition, in case of estimation of the model’s performance, how the weights are updated. It seems that currently there is no paper presenting a unified and consistent approach comparing all these methods for a relatively large set of independent variables (i.e., various spot commodities prices). The current paper is a try to fill this gap.
In particular, it was checked whether DMA indeed outperforms these other Bayesian methods. Secondly, it was checked whether these methods and time-varying regression with all predictors produce significantly more accurate forecasts than ARIMA models or the naive forecast. Once again, it is stressed that the aim was not to focus on some similar set of commodities, but to consider a possibly wide set of commodities and apply a unified methodology to them and see if this methodology leads to some improvement in forecasting. For certain conventional methods, such an approach was reported by Gargano and Timmermann [6].

3. Data

The monthly data between January 1984 and October 2017 were taken. The frequency was chosen as a kind of compromise. Indeed, many macroeconomic data are available even only in quarterly frequency. Therefore, more frequent than monthly data would result in problems with macroeconomic data. On the other hand, too high frequency would need a separate analysis of particular problems connected with very short periods. For example, it can be expected that, in monthly frequency, short-term speculative behaviour is flattened. The period before 1984 was not included due to two facts. First, many commodities prices before that time were not available or their volatilities were so small that the analysis would not be interesting. Secondly, as it is explained further, some variables are missing for certain periods before 1984. Resolving this problem—how to deal with missing observations—would need another discussion and may not seem worth the current research.
Commodities prices were obtained from The World Bank [5]. After trimming the time-series to the above-mentioned period, the remaining time-series which contained missing observations were excluded. As a result, 69 time-series remained. The full list of the analysed commodities is given in Table 1.
The predictors were mostly based on the already discussed papers with a great emphasis on the work of Gargano and Timmermann [6]. Dividend to price ratio (dpr) was taken as the difference between the logarithm of dividends and the logarithm of prices. These data were obtained from Schiller [62]. In particular, prices were taken as S&P Comp. Dividends were transformed to 12-month moving sums. Treasury bill rate (str) was taken as the 3-month treasury bill: secondary market rate (TB3MS) from FRED [63]. Long-term rate of bonds (ltr), because of the lack of exact recent data and problems with methodology, was taken as long-term government bond yields: 10-year, main (including benchmark) for the United States (IRLTLT01USM156N). Term spread (ts) was taken as the difference between the long-term rate of bonds and treasury bill rate. Default return spread (drs) was taken as the difference between long-term corporate bonds yield and treasury bill rate. Long-term corporate bond yield was taken as the index based on bonds with maturities 20-years and above (AAA). Inflation (cpi) was taken as the logarithmic difference of the CPI (Consumer Price Index CPIAUCSL). Industrial production growth (ip) was taken as the logarithmic difference of industrial production (INDPRO). Money stock (m) was taken as the logarithmic difference of M1 money stock (M1SL). In order to measure the economic growth, the classically used GDP was not appropriate because it is available in quarterly—not monthly—frequency. However, it can be found in literature that Kilian [64] global economic activity index can serve as a good proxy (kei). All of these data, as well as the unemployment rate (une, UNRATE), were obtained from FRED [63]. Following Chen et al. [57], the logarithmic differences of Australian dollar (aud) and Indian rupee to the US dollar (inr) exchange rates were taken (CCUSMA02INM618N and CCUSSP01AUM650N). These currencies were selected as exchange rate representatives because the respective countries are the largest exporters of industrial and agricultural commodities [6].
Open interest (op) was obtained from U.S. Commodity Futures Trading Commission [65] and directly from the website of Yogo [66]. During an 11-month period from January 1982 to November 1982, U.S. Commodity Futures Trading Commission did not collect data, due to budgetary reasons. Because the monthly growth rate of open interest is noisy, it was smoothed by taking a 12-month geometric mean in the time-series [59]. For the period before 2009, data by Yogo [66] were taken. Since 2009, data from U.S. Commodity Futures Trading Commission [65] were taken. As these data are provided in higher than monthly frequency, mean values from every month were computed.
Behaviour of commodities prices through some one, general measure (tr) was proxied by logarithmic differences of Thomson Reuters Equal Weight Commodity Index [67]. By equal weighting, this index, featuring 17 commodities, provides quite even distribution into the major commodity sectors, rather than overweighting in certain sectors such as energy or agriculture.
Prices of the commodities were transformed to logarithmic differences. The abbreviations of commodities are explained in Table 1.
It should be mentioned that these predictors were already used in a DMA-type scheme applied to modeling selected energy commodities prices [68]. However, in that study, these predictors were only a part of a much bigger collection of various other predictors. In addition, in that study, the collection of predictors was individually tailored for each commodity. For example, predictors indicating supply and demand forces were included, whereas, herein, they were taken as an optimal collection of predictors common for all analysed commodities. Finally, here the original DMA scheme is used [4], whereas the mentioned study (due to the high number of predictors inserted) used the Dynamic Occam’s Window of Onorante and Raftery [15]. Moreover, the mentioned study covered a different period (mostly due to the fact that it was also using Internet search queries available since 2004).
The descriptive statistics are reported in Table 2.
ADF, Phillips–Perron and KPSS stationarity tests are reported in Table 3. No important discrepancies were found. Only in the case of dpr and une, stationarity cannot be assumed. kei is stationary by the construction [64]. The reported opposition is due to analysing just a cut part of this time-series. However, it should be stressed that, in case of DMA, stationarity is not required, as this model is time-adapting. As it was already described, variables were taken in their differences mostly due to the convention, and to provide comparable outcomes with the previously mentioned literature. On the other hand, it is quite desirable due to computational issues in DMA that variables are of similar size. Therefore, kei was divided by 100.

4. Methodology

All the methods used herein are well explained in the original papers. Therefore, to keep both the clarity of the explanation and the reader’s convenience, only a short sketch is presented below. The computational details can be recovered from the cited sources, while the emphasis is put on the general idea of the methods. All computations were done in R Core Team [69] with the help of some additional packages [70,71].

4.1. Dynamic Model Averaging (DMA)

DMA was proposed by Raftery et al. [4]. Suppose that there are m potential predictors for the independent variable y t . In this research, y t denotes the logarithmic differences of the selected commodity price and m = 15 because the possible predictors are: lagged one period back dpr, str, ltr, ts, drs, cpi, ip, m, kei, une, aud, inr, op, tr and the first lag of y t . The lag of y t was taken because of the serious arguments in the literature and significant improvement observed in pre-simulations. On the other hand, more lags were decided not to be included because one of the alternative models is ARIMA. Thus, it was tried to check if there is some trade-off between the inclusion of other predictors and autoregressive lags.
Out of these m = 15 predictors, K = 2 15 = 32768 linear regression models can be constructed. Therefore, let x t ( k ) be the vector of predictors in k-th model, with k = 1 , , K . Then, the state space model of DMA is given by
y t = ( x t ( k ) ) θ t ( k ) + ϵ t ( k ) , θ t ( k ) = θ t 1 ( k ) + δ t ( k ) ,
where θ t is the vector of regression coefficients. It is assumed that errors follow the normal distribution, i.e., ϵ t ( k ) N ( 0 , V t ( k ) ) and δ t ( k ) N ( 0 , W t ( k ) ) . All of these K models are estimated as Time-Varying Parameters regressions (TVP) with the help of the Kalman filter [4]. Due to the computational issues, a certain forgetting procedure is used in updating W t ( k ) . This needs a forgetting factor λ ( 0 , 1 ] to be specified [4]. From the interpretative point of view, this factor is responsible for the allowed variability of regression coefficients, and it corresponds to the effective rolling window size of 1 1 λ . Lower values correspond to higher variability of regression coefficients.
Variance V t ( k ) can be updated by the recursive moment estimation [4]. However, in the case of the suspicion of ARCH effects in residuals, some authors suggest using the Exponentially Weighted Moving Average (EWMA) method [72]. Here, both approaches were used and compared.
Further in this paper, by TVP, it will be denoted the model for all possible m = 15 predictors estimated in this way. Equivalently, it can be seen as the above DMA model, but reduced to K = 1 and keeping only this k-th model which contains all m = 15 predictors.
Now, having estimated all these K time-varying parameters models, the DMA introduces a set of two weights, which are recursively updated. Therefore, this method to forecast at time t uses only the information available up to time t 1 .
The weights are given by
π t | t 1 , k = ( π t 1 | t 1 , k ) α l = 1 K ( π t 1 | t 1 , l ) α
and
π t | t , k = π t | t 1 , k f k ( y t | Y t 1 ) l = 1 K π t | t 1 , l f l ( y t | Y t 1 ) ,
where α ( 0 , 1 ] is a (second) forgetting factor, and f k ( y t | Y t 1 ) is the predictive density of k-th model at y t .
In order to start the above recursive computations, some initial values have to be set. Assuming that initially all K models are equally “good” (i.e., using the noninformative prior), it is set π 0 | 0 , k = 1 K and θ 0 ( k ) = 0 for all k = 1 , , K . V 0 ( k ) should be initialized with respect to the magnitude of the variables used. It should be big enough to allow for the changes in parameters. On the other hand, too big V 0 ( k ) and small λ can result in catching noise rather than signal. From standard deviations reported in Table 2, it seems reasonable to set V 0 ( k ) = 1 for all k = 1 , , K . Unifying these values for all models used in averaging also has some computational advantage:
Finally, the DMA forecast is given by
y ^ t DMA = k = 1 K π t | t 1 , k y ^ t ( k ) = k = 1 K π t | t 1 , k ( x t ( k ) ) θ ^ t 1 ( k ) .
Notice that, if α = 1 , equal weights are produced. Secondly, if α = 1 = λ is taken, then DMA reproduces in a (computationally efficient) way the Bayesian Model Averaging, further in the text denoted simply as BMA [4]. Usually, it is suggested to set α = 0.99 = λ in DMA [4,72]. However, in some cases, more accurate forecasts can be generated, if different pairs of these parameters are taken [8]. Therefore, in this research, all possible pairs α , λ = { 1 , 0.99 , 0.98 , 0.95 , 0.90 } were considered. More frequent search around α = 1 = λ , and less—around α = 0.90 = λ —was decided, because the previous simulations suggest that forecast accuracy can change much when switching between higher values of forgetting parameters, whereas, for lower values, the change is rather small [73].
At this point, it can be mentioned that DMA is an extension of BMA in the following sense. First, it allows for time-varying parameters. Secondly, it allows the weights in model averaging to vary in time. Equivalently, DMA can be seen as dynamic (i.e., with time-varying weights) averaging over TVP models.

4.2. Dynamic Model Selection (DMS)

Of course, it is easy to replace the above averaging scheme with model selection [72]. Simply, Equation (2) can be modified to
y ^ t DMS = y ^ t ( h t ) ,
where h t denotes that model (out of K ones), which corresponds to the highest π t | t 1 , k . Such a scheme is called the Dynamic Model Selection (DMS).
Similarly, as with DMA and BMA, if in the above scheme α = 1 = λ is taken, then the resulting model is called the Bayesian Model Selection (BMS).

4.3. Median Probability Model (MED)

However, Barbieri and Berger [74] noticed that the selection of the model with the highest posterior probability π t | t 1 , k is not always desirable. Indeed, it is under certain very general conditions, but it is still optimal only in the case of only two models competing. Secondly, this is so in the case of linear models having orthogonal design matrices. Therefore, they proposed Median Probability Model (MED).
First, relative variable importance needs to be computed for every predictor. This is defined as the sum of posterior probabilities π t | t 1 , k for those models out of all K, which contain a particular predictor. Next, the model which contains as the predictors exactly those whose relative variable importances are equal to or greater than 0.5 is selected. Of course, relative variable importance can vary in time, therefore the selected model also varies in time.
As previously, if in the above scheme α = 1 = λ is taken, then the resulting model is called the Bayesian Median Probability Model (BMED). It should be noticed that models pre-named “Bayesian” are the special cases of their more general encompassing combination schemes, just that their forgetting parameters are α = 1 = λ . This name-based emphasis is done to stress the particular forgetting factor values, corresponding to equal-weighting in Equation (1) and no variability of regression coefficients assumed.

4.4. Evaluation of Models

To summarize, for each of the commodities, DMA, DMS and MED models were estimated. These models were estimated with all possible pairs α , λ = { 1 , 0.99 , 0.98 , 0.95 , 0.90 } , i.e., in 25 different versions. Moreover, as described before, each of these models were estimated with AR(1) term and without it, and with the mentioned EWMA procedure or with the recursive moment estimation. For each of the commodities, TVP models with λ = { 1 , 0.99 , 0.98 , 0.95 , 0.90 } were also estimated. They were also estimated with and without AR(1) terms, and with EWMA and recursive moment estimation. This resulted in 320 models being estimated for each of the commodity.
As for the conventional alternative forecasting model, two choices were done: Auto ARIMA model by Hyndman and Khandakar [71] implemented in a recursive way, and the naive method, i.e., the observation from time t is taken as the forecast for time t + 1 (NAIVE). The model of Hyndman and Khandakar [71] is choosing AR and MA order dynamically in each step by starting from the one term and checking if more terms improve the selected criterion (in this case: Akaike Information Criterion). To guarantee stopping, no more than five AR terms and 5 MA terms are considered.
The evaluation of the above described Bayesian schemes was done in the following way. First of all, the forecast accuracy was measured by RMSE (Root Mean Squared Error). The estimated models were divided into two groups. The first was consisting of DMA, DMS and MED models. In other words, they are the models based on some kind of a model combination scheme. The second group was consisting of TVP, ARIMA and NAIVE models, i.e., models not based on any model combination scheme.
In order to compare the forecast accuracy generated by two competing models, the Diebold–Mariano test (DM) was used [75]. This test was chosen because of its relative little assumptions and popularity. Its null hypothesis is that the two forecasts have the same accuracy.
This test was used to compare forecasts from the most accurate models representing the two above described groups.
Secondly, for each commodity, the model minimising RMSE was chosen out of the various versions of the DMA model. It was compared with ARIMA and NAIVE models, i.e., with the very common benchmark models.
The aim was to statistically check if DMA can produce more accurate forecasts than the conventional methods and if the considered Bayesian model combination schemes in general have some forecast accuracy advantage over the conventional methods.
It was also analysed if a certain pattern in forecast errors depending on the forgetting factors can be identified.
Finally, it should be remembered that the used methods were modeling the logarithmic differences of commodities prices. However, RMSE and DM tests were applied to errors from the forecast of commodities prices (i.e., the direct outcomes from the considered models were transformed).

5. Results and Discussion

This section is devoted to, at first, presenting the preliminary results and observations derived from the summary of the estimation of the considered models. Indeed, as the selected Bayesian techniques rely on recursive computations and a large quantity of models, the computational issues result in certain obstacles. For example, the estimation of one DMA model for one particular commodity took on an average computer no more than few minutes. It should be noticed that adding one extra predictor doubles the time needed for computations; therefore, the computational issues can easily switch from negligible to insuperable ones. As mentioned in the previous section, for all 69 commodities, a total of 22,218 models were estimated. The estimation was speeded up by the use of both the cloud computing and the parallel computations. Nevertheless, the whole estimation took over a day, despite the use of Amazon Web Services (AWS). Therefore, due to the number of the estimated models, a certain degree of subjectivity in the summary outcomes presentation is a must (i.e., how to present them, and on which put the emphasis).
The second part (i.e., Section 5.2) is devoted to the main results. In other words, RMSEs (Root Mean Squared Errors) are reported. Additionally, the forecasts’ accuracies from different models are statistically tested. It is also described whether some type of a model combination scheme seem to play the dominant role in producing most accurate forecasts.
The last part (i.e., Section 5.3) is devoted to a short discussion of the analysis, which seemed interesting after looking at the main results. As just mentioned, a large number of models was estimated, therefore it is not possible to present quite extensive analysis. In the first two parts, the outcomes are presented and discussed within the context of the aims of this research, as they were stated in the beginning. However, after the estimation, some other comparisons become interesting. For example, forecast accuracy of DMA with α = 0.99 and λ = 1 was compared with forecast accuracy of BMA (i.e., DMA with α = 1 = λ ). The importance of these two models become evident after the knowledge of the outcomes reported in the first and the second part.
Another example of subjectivity is that, due to the large amount of models if, for example, a DMA model combination scheme was analysed and the model minimising RMSE was sought, then all variations (with all considered forgetting factors, both with EWMA and recursive moment estimation method used, with and without AR(1) term) of the models were put into one group. Some readers might be willing to, for example, analyse all of these models in separate groups. However, as several variations were considered, the number of such choices is quite big, and reporting all of that would obscure the outcomes.
The considered Bayesian models require some initial number of observations to “learn”. In other words, for the first observations, some chaotic outcomes are usually generated as the initial parameters tune up to the upcoming new observations. For example, as explained in the previous section, all models are initially set with zero regression coefficients. Therefore, the first 100 observations, i.e., approximately 1/4 of all observations were treated as in-sample. All reported outcomes (RMSE, p-values of the Diebold–Mariano test, etc.) are based on the out-of-sample consisting of the next 3/4 of observations. It should be stressed that the considered models are recursive ones, therefore their predictors coefficients are updated with each extra observation. Therefore, these models still evolve during the out-of-sample period. The in-sample period was chosen to cut off observations during which models try to catch the signal. This method is well known and is sometimes called evaluation on a rolling forecasting origin, walk forward validation, forward-chaining validation, or time-series cross-validation [76].
All estimated models were divided into the following groups: DMA, DMS, MED, TVP, ARIMA and NAIVE. Models in different versions due to the choice of forgetting factors, method of variance updating: EWMA or recursive moment estimation, with and without AR(1) term, if not stated otherwise, were not analysed separately. In particular, BMA was treated as a special case of DMA, BMS—a special case of DMS, etc.

5.1. Descriptive Analysis of the First Results

First of all, it was observed that, in 87% of cases, DMA produced smaller RMSE than the NAIVE model. Assuming a 5% significance level, according to the Diebold–Mariano test, these results were statistically significant in 83% of cases. (The null hypothesis was that both methods have the same forecast accuracy. The alternative is that DMA is more accurate than the NAIVE method.) If the benchmark model was taken as the ARIMA, then DMA produced smaller errors in 80% of cases, and they were statistically significant in 78% of cases. Such results can serve as an argument that, in general, despite its quite high complexity, DMA is an interesting alternative for simple forecasting models.
However, if the considerations are restricted only to DMA, DMS and MED models, then α < 1 allowed to minimise RMSE for 42% of DMA models, 54% of DMS models and 51% of MED models. If the considerations are restricted only to DMA, DMS, MED and TVP models, then λ < 1 allowed to minimise RMSE for 9% of DMA models, 12% of DMS models, 14% of MED models, and 6% of TVP models. These outcomes suggest that λ = 1 is very strongly preferred for minimising RMSE, whereas setting the value of α is not so obvious.
Indeed, narrowing the considerations only to DMA, DMS and MED models, the combinations of forgetting factors such that α , λ < 1 minimised RMSE only in 1% of DMA models, 3% of DMS models, and 6% of MED models. Therefore, even though the Bayesian model combination schemes seem to be beneficial as reducing the forecast error, this is mostly due to the time-varying weights construction. As λ = 1 is the most often preferred value, the regression coefficients’ variability is not so beneficial for forecast accuracy.
The DMA model with both forgetting factors being strictly less than 1 was the model for POTASH. Interestingly, this model generated a significantly more accurate forecast (at 5% significance level) than the NAIVE and ARIMA models. It was also the one which (for this commodity) minimised RMSE out of all the considered types of models. However, this model did not generate a statistically significant more accurate forecast than TVP model.
The preferences of forgetting factors can be visualised on a bubble plot in Figure 1. Clearly, it can be seen that, in the case of DMA, DMS and MED models, λ = 1 is preferred if minimising RMSE is the aim. In many cases, α = 1 = λ is also preferred, meaning that BMA, BMS or BMED is preferred. However, in many cases, α < 1 remains the preferred option. More discussion about this result is postponed for the end of this section.
The next observation is that the model minimising RMSE was the one with the AR(1) term added in 75% cases if the DMA type was considered. For DMS models, it was 62%, for MED—68%, and for TVP—71%. This means that, in the majority of cases, it is desirable to add an autoregressive term; however, in a reasonable number of cases, it is not.
EWMA (Exponentially Weighted Moving Average) variance updating method was present only in 20% of DMA models which had the smallest RMSE. In the case of DMS models, it was 35%, for MED models—38%, and for TVP models—35%. This suggests that, if minimising errors is the aim, then considering EWMA method is not desirable, and one can stick to the recursive moment estimation (which was proposed in the original paper introducing the DMA method [4]).
Anyway, the idea behind the EWMA method is to erase the possible ARCH effect in residuals. Performing LM–ARCH test with a 5% significance level indicated that the selected models do not have ARCH effects in residuals in 35% of cases of DMA type models, 32% of DMS type models, 33% of MED type models and 28% of TVP type models. Fortunately, as explained in the previous section, the considered Bayesian model combination schemes are recursively updating the error term in equations, so this is not an important obstacle in the case of forecasting. It should be stressed that models were chosen due to minimisation of RMSE, and the verification of ARCH effects was done as an additional check. In other words, for the purpose of this research, the model with ARCH effects, but with smaller RMSE was preferred, even if some model did not possess ARCH effects, but would have larger RMSE.
Of course, it is interesting to estimate how often the lack of ARCH effects could be assumed at a 5% significance level, and, simultaneously, the chosen model was the one with the EWMA estimation method. For DMA type, it was in 10% of cases, for DMS—12%, for MED—14%, and for TVP—10%. Therefore, in line with the previous observation, it can be stated that EWMA allowed for erasing ARCH effects efficiently only in 1/3—1/2 cases in which this method was applied.
To sum up the above results, the following is indicated: first, that DMA is an interesting alternative method (if compared with ARIMA or NAIVE models); secondly, that λ = 1 is usually preferred, but α < 1 , meaning that most gains in forecast accuracy is due to the certain weighting procedure—not time-varying regression coefficients; third, that the EWMA method is not necessarily “better” than the recursive moment estimation; fourth, that adding the autoregressive term usually (but not always) leads to smaller forecast errors.

5.2. Main Results

Table 4 presents the main outcomes for various model combination schemes. In other words, Table 4 reports the normalized RMSE (nRMSE) from all estimated model. As each type of model was estimated in various versions, in Table 4, the outcomes are present for the one which minimised RMSE for the given type of a model (i.e., out of DMA type models, out of DMS type models, etc.). The normalization of RMSE is done simply by dividing by the mean value of the forecasted time-series. It can be seen (Table 2) that the orders of magnitudes of forecasted commodity prices are quite different. It is easy to switch between RMSE and normalized RMSE having the information from Table 2 and Table 4. However, normalized values seem easier for interpreting the forecasting benefit of each of the models.
Generally, the normalized RMSE is around 0.07. This means that quite accurate forecasts can be produced. In a few cases, very small errors were produced (0.01) or very high errors were produced (0.19). However, for the given commodity, the errors from different models do not differ much.
More information can be derived from Table 5, which presents the results of the Diebold–Mariano tests. The first columns of this table are devoted to comparison of the accuracies of forecasts generated by the DMA method and the NAIVE and the ARIMA models. As a reminder, the null hypothesis of the test is that both forecasts have the same accuracy. For each of the commodities, whether DMA produced smaller RMSE than the NAIVE, and whether it produced smaller RMSE than the ARIMA, was checked. If so, then the alternative hypothesis for the test was taken that the DMA forecast is more accurate than the one from the NAIVE (or from the ARIMA). Otherwise, the NAIVE forecast (or the ARIMA forecast) is more accurate. Clearly, it can be seen that, in the majority of cases, there is a strong evidence to treat DMA as a significantly more accurate forecasting method.
The notation X < Y in Table 5 represents the alternative hypothesis for the Diebold–Mariano test, and it should be read: “The forecast generated by model X is more accurate than the forecast generated by model Y”. Similarly, the notation X > Y should be read: “The forecast generated by model Y is more accurate than the forecast generated by model X”.
Finally, the last two columns in Table 5 are devoted to choosing the “best” model out of all considered ones, for the given commodity price. The “best” model is understood to be the one that generated the smallest RMSE. Next, as already mentioned in the Methodology section, the estimated models were divided into two groups. The first contained DMA, DMS and MED models and the second the TVP, ARIMA and NAIVE. In each group, the representatives with smallest (in the given group) RMSE were chosen. Then, the Diebold–Mariano test was performed to check that, indeed, the forecast generated by one of these representative models is significantly more accurate than the forecasts generated by the second representative model (from another group). The summary of these considerations is presented in Table 6.
Indeed, from Table 6, it can be seen that, in 26% of cases, DMA was the model that generated the smallest RMSE. Moreover, in 22% of cases, it did so and, simultaneously, this result was significant at a 5% significance level according to the described procedure (dividing models into two groups). The DMA model generated the smallest RMSE in 23% of cases. In 13% of cases, it did so, and, simultaneously, this result was significant. The MED model generated the smallest RMSE in 28% of cases. In 17% of cases, it did so, and, simultaneously, this result was significant. It seems worth noticing that the scheme of selecting the model with the highest posterior probability is more popular over the Median Probability Model (MED). The obtained results seem therefore interesting, as they confirm that it is worth putting more attention in practice on the MED combination scheme.
The TVP model generated the smallest RMSE only in 1% of cases. In none of cases did it do so significantly.
BMA, BMS and BMED models were never the ones minimising RMSE out of all considered forecasting models. Together with the already discussed observation from Figure 1, it means that models with α = 1 = λ are very often minimising RMSE amongst the discussed Bayesian model combination schemes, but they cannot generate significantly more accurate forecasts than the ARIMA or the NAIVE models. On the other hand, if the Bayesian model combination scheme generated such a significantly more accurate forecast, then it was the model with α < 1 or λ < 1 . Again, looking at Figure 1, it can be expected that this was rather due to α < 1 than to λ < 1 .
Finally, the NAIVE model generated the smallest RMSE in only 6% of cases. In none of cases did it do so significantly. The ARIMA model generated the smallest RMSE in 16% of cases. In only 4% of cases did it do so, and, simultaneously, this result was significant.
It can be concluded that the considered Bayesian model combination schemes can result in significantly more accurate forecasts. Secondly, both model averaging (like in DMA), and model selection (like in DMS and MED) are efficient methods. It should be stressed that, due to Equation (1) and the way the weights are updated recursively, the forgetting factor α plays an important role also in model selection schemes. Its role is connected with the way that the past information is treated. Roughly speaking, the information (how the model behaved) from i periods back is given α i weight [7,10].
Those figures present normalized RMSE from DMA models with an AR(1) term and with recursive moment estimation used for variance updating. In Figure 2, various α = λ are considered. It can be seen that, under such a restriction, higher values of forgetting factors are preferred. A similar conclusion is valid if α = 1 is fixed and λ is changing, as it can be seen in Figure 3. However, in Figure 4, it can be seen that, if λ = 1 is fixed, in most cases, changing the value of α does not lead to important changes of normalized RMSE. Anyway, in a few cases, the smaller values of α are preferred.

5.3. Further Remarks

By looking at rows in Table 4, it could be seen that the average gain in lowering normalized RMSE by applying some of the Bayesian model combination scheme over TVP, ARIMA or NAIVE models is 0.002, which is very small. On the other hand, from Table 5, it was seen that Bayesian model combination schemes very often produce significantly more accurate forecasts, so this gain is important.
It is also interesting to come back to Figure 1 and discussion about the preferred forgetting factors. Usually, if considerations are narrowed just to DMA, DMS and MED models, then the combination of α = 1 = λ minimises errors. However, when considerations are narrowed to those of DMA, DMS and MED models, which generate significantly more accurate forecasts than TVP, ARIMA and NAIVE, then λ = 1 is preferred, but α < 1 . Therefore, it is interesting to analyse whether the difference between these two types of models is statistically significant. For this, DMA models with AR(1) term and based on recursive moment estimation for variance updating were compared—in other words, those models with α = 1 = λ (BMA) with α = 0.99 and λ = 1 (DMA). At the 5% significance level, according to the Diebold–Mariano test, DMA generated a more accurate forecast than BMA in 14% of cases, at a 10% significance level—in 23% of cases.
Actually, in Figure 1, some interesting patterns can also be seen for fixed α = 1 . From the interpretative point of view, the combination α = λ is also quite natural. Therefore, Figure 2, Figure 3 and Figure 4 present normalized RMSE for all analysed commodities prices with different combinations of forgetting factors α and λ .

6. Conclusions

It was found that, in general, the Bayesian model combination schemes like Dynamic Model Averaging (DMA), Dynamic Model Selection (DMS) and Median Probability Model (MED) in most of the cases produced significantly more accurate forecasts than the ARIMA, NAIVE or TVP (Time-Varying Parameters regression) models. Interestingly, the MED model in many cases generated smaller RMSE (Root Mean Squared Error) than DMS.
It was found that, indeed, the schemes based on recursive estimation like in DMA, allowing for time-varying regression coefficients and time-varying models’ weights, resulted in smaller RMSE and more accurate forecasts. However, the more thorough analysis suggested that the forecasting benefit from this approach is rather due to time-varying weights, not time-varying regression coefficients. From the economic point of view, this might be interpreted that, in different periods, it is rather the model itself (the collection of predictors) which really matters as opposed to the strength of the relationship between predictors and the given commodity price, which should be modeled more carefully.
Indeed, all considered modeling schemes used in this research were based on the recursive estimations. In other words, forecast at time t was based only on the information available up to the time t 1 . Additionally, all information available up to time t 1 was used, so, as the new information become available, it was used to update the model’s parameters.
On the other hand, in a numerical sense, the decline in RMSE due to the use of the applied schemes was not very high. Anyway, according to the Diebold–Mariano (DM) test, the applied schemes generated significantly more accurate forecasts.
This research was not based on some index as a proxy of commodities prices. Instead, 69 individual time-series were modeled. Generally, it was confirmed that adding the autoregressive first lag as a predictor improves the forecast accuracy. The advantage of Exponentially Weighted Moving Average (EWMA) in variance updating over the originally proposed recursive moment estimation was not confirmed.
Summarizing the outcomes, the considered Bayesian model combination schemes were found useful, i.e., generating more accurate forecasts in most of the cases.

Funding

The research was funded by the Polish National Science Centre grant under the contract number DEC-2015/19/N/HS4/00205.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ARIMAAuto ARIMA model described by Hyndman and Khandakar [71]
BMABayesian Model Averaging as a special case of DMA with forgetting factors α = 1 = λ
BMEDBayesian Median Probability Model as a special case of MED with forgetting factors α = 1 = λ
BMSBayesian Model Selection as a special case of DMS with forgetting factors α = 1 = λ
DMThe Diebold–Mariano test [75]
DMADynamic Model Averaging proposed by Raftery et al. [4]
DMSDynamic Model Selection, i.e., model averaging in DMA replaced by selecting the model with the highest posterior probability
MEDMedian Probability Model of Barbieri and Berger [74]
NAIVEthe naive forecast, i.e., the last observation is the one-ahead forecast
RMSERoot Mean Squared Error
nRMSENormalized RMSE, i.e., RMSE divided by the mean value of the forecasted time-series
TVPTime-Varying Parameters, i.e., DMA reduced to exactly one model, i.e., the one with all predictors

References

  1. Arouri, M.E.H.; Dinh, T.H.; Nguyen, D.K. Time-varying Predictability in Crude-oil Markets: The Case of GCC Countries. Energy Policy 2010, 38, 4371–4380. [Google Scholar] [CrossRef]
  2. Henkel, S.J.; Martin, J.S.; Nardari, F. Time-varying Short-horizon Predictability. J. Financ. Econ. 2011, 99, 560–580. [Google Scholar] [CrossRef]
  3. Rapach, D.; Strauss, J.; Zhou, G. International Stock Return Predictability: What Is the Role of the United States? J. Finance 2013, 68, 1633–1662. [Google Scholar] [CrossRef] [Green Version]
  4. Raftery, A.; Kárný, M.; Ettler, P. Online Prediction under Model Uncertainty via Dynamic Model Averaging: Application to a Cold Rolling Mill. Technometrics 2010, 52, 52–66. [Google Scholar] [CrossRef] [PubMed]
  5. The World Bank. Commodity Price Data; The World Bank: Washington, DC, USA, 2017. [Google Scholar]
  6. Gargano, A.; Timmermann, A. Forecasting Commodity Price Indexes Using Macroeconomic and Financial Predictors. Int. J. Forecast. 2014, 30, 825–843. [Google Scholar] [CrossRef]
  7. Buncic, D.; Moretto, C. Forecasting Copper Prices with Dynamic Averaging and Selection Models. N. Am. J. Econ. Finance 2015, 33, 1–38. [Google Scholar] [CrossRef]
  8. Drachal, K. Forecasting Spot Oil Price in a Dynamic Model Averaging Framework—Have the Determinants Changed over Time? Energy Econ. 2016, 60, 35–46. [Google Scholar] [CrossRef]
  9. Risse, M.; Kern, M. Forecasting House-price Growth in the Euro Area with Dynamic Model Averaging. N. Am. J. Econ. Finance 2016, 38, 70–85. [Google Scholar] [CrossRef]
  10. Baur, D.; Beckmann, J.; Czudaj, R. A Melting Pot-Gold Price Forecasts under Model and Parameter Uncertainty. Int. Rev. Financ. Anal. 2016, 48, 282–291. [Google Scholar] [CrossRef]
  11. Naser, H. Estimating and Forecasting the Real Prices of Crude Oil: A Data Rich Model Using a Dynamic Model Averaging (DMA) Approach. Energy Econ. 2016, 56, 75–87. [Google Scholar] [CrossRef]
  12. Wei, Y.; Cao, Y. Forecasting House Prices Using Dynamic Model Averaging Approach: Evidence from China. Econ. Model. 2017, 61, 147–155. [Google Scholar] [CrossRef]
  13. Aye, G.; Gupta, R.; Hammoudeh, S.; Kim, W. Forecasting the Price of Gold Using Dynamic Model Averaging. Int. Rev. Financ. Anal. 2015, 41, 257–266. [Google Scholar] [CrossRef]
  14. Risse, M.; Ohl, L. Using Dynamic Model Averaging in State Space Representation with Dynamic Occam’s Window and Applications to the Stock and Gold Market. J. Empir. Finance 2017, 44, 158–176. [Google Scholar] [CrossRef]
  15. Onorante, L.; Raftery, A. Dynamic Model Averaging in Large Model Spaces Using Dynamic Occam’s Window. Eur. Econ. Rev. 2016, 81, 2–14. [Google Scholar] [CrossRef] [PubMed]
  16. Baxa, J.; Plašil, M.; Vašiček, B. Inflation and the Steeplechase between Economic Activity Variables: Evidence for G7 Countries. B.E. J. Macroecon. 2017, 17, 42. [Google Scholar] [CrossRef]
  17. Del Negro, M.; Hasegawa, R.; Schorfheide, F. Dynamic Prediction Pools: An Investigation of Financial Frictions and Forecasting Performance. J. Econom. 2016, 192, 391–405. [Google Scholar] [CrossRef]
  18. Di Filippo, G. Dynamic Model Averaging and CPI Inflation Forecasts: A Comparison between the Euro Area and the United States. J. Forecast. 2015, 34, 619–648. [Google Scholar] [CrossRef]
  19. Ferreira, D.; Palma, A. Forecasting Inflation with the Phillips Curve: A Dynamic Model Averaging Approach for Brazil. Rev. Bras. Econ. 2015, 69, 451–465. [Google Scholar] [CrossRef]
  20. Koop, G.; Korobilis, D. Forecasting Inflation Using Dynamic Model Averaging. Int. Econ. Rev. 2012, 53, 867–886. [Google Scholar] [CrossRef]
  21. Koop, G.; Korobilis, D. UK Macroeconomic Forecasting with Many Predictors: Which Models Forecast Best and When Do They Do So? Econ. Model. 2011, 28, 2307–2318. [Google Scholar] [CrossRef]
  22. Koop, G.; Korobilis, D. A New Index of Financial Conditions. Eur. Econ. Rev. 2014, 71, 101–116. [Google Scholar] [CrossRef] [Green Version]
  23. Bork, L.; Moller, S. Forecasting House Prices in the 50 States Using Dynamic Model Averaging and Dynamic Model Selection. Int. J. Forecast. 2015, 31, 63–78. [Google Scholar] [CrossRef]
  24. De Bruyn, R.; Gupta, R.; Van Eyden, R. Can We Beat the Random-walk Model for the South African Rand—U.S. Dollar and South African Rand—UK Pound Exchange Rates? Evidence from Dynamic Model Averaging. Emerg. Mark. Finance Trade 2015, 51, 502–524. [Google Scholar] [CrossRef]
  25. Gupta, R.; Hammoudeh, S.; Kim, W.; Simo-Kengne, B. Forecasting China’s Foreign Exchange Reserves Using Dynamic Model Averaging: The Roles of Macroeconomic Fundamentals, Financial Stress and Economic Uncertainty. N. Am. J. Econ. Finance 2014, 28, 170–189. [Google Scholar] [CrossRef]
  26. Koop, G.; Tole, L. Forecasting the European Carbon Market. J. R. Stat. Soc. Ser. A Stat. Soc. 2013, 176, 723–741. [Google Scholar] [CrossRef]
  27. Liu, J.; Wei, Y.; Ma, F.; Wahab, M. Forecasting the Realized Range-based Volatility Using Dynamic Model Averaging Approach. Econ. Model. 2017, 61, 12–26. [Google Scholar] [CrossRef]
  28. Wang, Y.; Ma, F.; Wei, Y.; Wu, C. Forecasting Realized Volatility in a Changing World: A Dynamic Model Averaging Approach. J. Bank. Finance 2016, 64, 136–149. [Google Scholar] [CrossRef]
  29. Naser, H.; Alaali, F. Can Oil Prices Help Predict US Stock Market Returns? Evidence Using a Dynamic Model Averaging (DMA) Approach. Empir. Econ. 2017. [Google Scholar] [CrossRef]
  30. Bates, J.; Granger, C. The Combination of Forecasts. Oper. Res. Q. 1969, 20, 451–468. [Google Scholar] [CrossRef]
  31. Baumeister, C.; Kilian, L. Forecasting the Real Price of Oil in a Changing World: A Forecast Combination Approach. J. Bus. Econ. Stat. 2015, 33, 338–351. [Google Scholar] [CrossRef]
  32. Kaya, H. Forecasting the Price of Crude Oil with Multiple Predictors. Siyasal Bilgiler Fakültesi Dergisi (İSMUS) 2016, 1, 133–151. [Google Scholar]
  33. Wang, Y.; Liu, L.; Wu, C. Forecasting the Real Prices of Crude Oil Using Forecast Combinations over Time-varying Parameter Models. Energy Econ. 2017, 66, 337–348. [Google Scholar] [CrossRef]
  34. Moral-Benito, E. Model Averaging in Economics: An Overview. J. Econ. Surv. 2015, 29, 46–75. [Google Scholar] [CrossRef]
  35. Steel, M. Model Averaging and its Use in Economics. arXiv, 2017; arXiv:1709.08221. [Google Scholar]
  36. Greenberg, E. Introduction to Bayesian Econometrics; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  37. Koop, G. Bayesian Methods for Empirical Macroeconomics with Big Data. Rev. Econ. Anal. 2017, 9, 33–56. [Google Scholar]
  38. Lee, C.Y.; Huh, S.Y. Forecasting Long-term Crude Oil Prices Using a Bayesian Model with Informative Priors. Sustainability 2017, 9, 190. [Google Scholar] [CrossRef]
  39. Yin, X.; Peng, J.; Tang, T. Improving the Forecasting Accuracy of Crude Oil Prices. Sustainability 2018, 10, 454. [Google Scholar] [CrossRef]
  40. Kriechbaumer, T.; Angus, A.; Parsons, D.; Rivas Casado, M. An Improved Wavelet-ARIMA Approach for Forecasting Metal Prices. Resour. Policy 2014, 39, 32–41. [Google Scholar] [CrossRef] [Green Version]
  41. Wang, Q.; Sun, X. Crude Oil Price: Demand, Supply, Economic Activity, Economic Policy Uncertainty and Wars—From the Perspective of Structural Equation Modelling (SEM). Energy 2017, 133, 483–490. [Google Scholar] [CrossRef]
  42. Cross, J.; Nguyen, B. The Relationship between Global Oil Price Shocks and China’s Output: A Time-varying Analysis. Energy Econ. 2017, 62, 79–91. [Google Scholar] [CrossRef]
  43. Gangopadhyay, K.; Jangir, A.; Sensarma, R. Forecasting the Price of Gold: An Error Correction Approach. IIMB Manag. Rev. 2016, 28, 6–12. [Google Scholar] [CrossRef]
  44. Gil-Alana, L.; Chang, S.; Balcilar, M.; Aye, G.; Gupta, R. Persistence of Precious Metal Prices: A Fractional Integration Approach with Structural Breaks. Resour. Policy 2015, 44, 57–64. [Google Scholar] [CrossRef]
  45. Kim, J.M.; Jung, H. Time Series Forecasting Using Functional Partial Least Square Regression with Stochastic Volatility, GARCH, and Exponential Smoothing. J. Forecast. 2017, 37, 269–280. [Google Scholar] [CrossRef]
  46. Hamdi, M.; Aloui, C. Forecasting Crude Oil Price Using Artificial Neural Networks: A Literature Survey. Econ. Bull. 2015, 35, 1339–1359. [Google Scholar]
  47. Liu, C.; Hu, Z.; Li, Y.; Liu, S. Forecasting Copper Prices by Decision Tree Learning. Resour. Policy 2017, 52, 427–434. [Google Scholar] [CrossRef]
  48. Zhao, Y.; Li, J.; Yu, L. A Deep Learning Ensemble Approach for Crude Oil Price Forecasting. Energy Econ. 2017, 66, 9–16. [Google Scholar] [CrossRef]
  49. Welch, I.; Goyal, A. A Comprehensive Look at The Empirical Performance of Equity Premium Prediction. Rev. Financ. Stud. 2008, 21, 1455–1508. [Google Scholar] [CrossRef]
  50. Ghalayini, L. Modeling and Forecasting Spot Oil Price. Eurasian Bus. Rev. 2017, 7, 355–373. [Google Scholar] [CrossRef]
  51. Tan, X.; Ma, Y. The Impact of Macroeconomic Uncertainty on International Commodity Prices: Empirical Analysis Based on TVAR Model. China Finance Rev. Int. 2017, 7, 163–184. [Google Scholar] [CrossRef]
  52. Kagraoka, Y. Common Dynamic Factors in Driving Commodity Prices: Implications of a Generalized Dynamic Factor Model. Econ. Model. 2016, 52, 609–617. [Google Scholar] [CrossRef]
  53. Lübbers, J.; Posch, P. Commodities’ Common Factor: An Empirical Assessment of the Markets’ Drivers. J. Commod. Market. 2016, 4, 28–40. [Google Scholar] [CrossRef]
  54. Arango, L.; Arias, F.; Florez, A. Determinants of Commodity Prices. Appl. Econ. 2012, 44, 135–145. [Google Scholar] [CrossRef]
  55. Byrne, J.; Fazio, G.; Fiess, N. Primary Commodity Prices: Co-movements, Common Factors and Fundamentals. J. Dev. Econ. 2013, 101, 16–26. [Google Scholar] [CrossRef]
  56. Alam, M.; Gilbert, S. Monetary Policy Shocks and the Dynamics of Agricultural Commodity Prices: Evidence from Structural and Factor-Augmented VAR Analyses. Agric. Econ. 2017, 48, 15–27. [Google Scholar] [CrossRef]
  57. Chen, Y.; Rogoff, K.; Rossi, B. Can Exchange Rates Forecast Commodity Prices? Q. J. Econ. 2010, 125, 1145–1194. [Google Scholar] [CrossRef]
  58. Arslan-Ayaydin, O.; Khagleeva, I. Chapter: The Dynamics of Crude Oil Spot and Futures Markets. In Energy Economics and Financial Markets; Springer: Berlin, Germany, 2013; pp. 159–173. [Google Scholar]
  59. Hong, H.; Yogo, M. What does Futures Market Interest Tell Us about the Macroeconomy and Asset Prices? J. Financ. Econ. 2012, 105, 473–490. [Google Scholar] [CrossRef]
  60. Chen, S.S. Commodity Prices and Related Equity Prices. Can. J. Econ. 2016, 49, 949–967. [Google Scholar] [CrossRef]
  61. Wang, Y.; Liu, L.; Diao, X.; Wu, C. Forecasting the Real Prices of Crude Oil under Economic and Statistical Constraints. Energy Econ. 2015, 51, 599–608. [Google Scholar] [CrossRef]
  62. Schiller, R. Online Data Robert Shiller. 2017. Available online: http://www.econ.yale.edu/~shiller/data.htm (accessed on 12 March 2018).
  63. FRED. Economic Data. Federal Reserve Bank of St. Louis Web Site, 2017. Available online: https://fred.stlouisfed.org (accessed on 12 March 2018).
  64. Kilian, L. Updated Version of the Index of Global Real Economic Activity in Industrial Commodity Markets. The University of Michigan Web Site. Available online: http://www-personal.umich.edu/~lkilian/reaupdate.txt (accessed on 12 March 2018).
  65. U.S. Commodity Futures Trading Commission. Historical Compressed. 2017; Commodity Futures Trading Commission Web Site. Available online: http://www.cftc.gov/MarketReports/CommitmentsofTraders/HistoricalCompressed/index.htm (accessed on 12 March 2018).
  66. Yogo, M. Research. Motohiro Yogo Web Site, 2017. Available online: http://sites.google.com/site/motohiroyogo (accessed on 12 March 2018).
  67. Thomson Reuters. Commodity Indices. Financial Thomson Reuters Web Site, 2017. Available online: http://financial.thomsonreuters.com/en/products/data-analytics/market-data/indices/commodity-index.html (accessed on 12 March 2018).
  68. Drachal, K. Comparison between Bayesian and Information-theoretic Model Averaging: Fossil Fuels Prices Example. Energy Econ. 2018, 74, 208–251. [Google Scholar] [CrossRef]
  69. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2013. [Google Scholar]
  70. Drachal, K. fDMA: Dynamic Model Averaging and Dynamic Model Selection for Continuous Outcomes. R Package Documentation Web Site, 2017. Available online: https://CRAN.R-project.org/package=fDMA (accessed on 12 March 2018).
  71. Hyndman, R.; Khandakar, Y. Automatic Time Series Forecasting: The forecast Package for R. J. Stat. Softw. 2008, 26, 1–22. [Google Scholar]
  72. Belmonte, M.; Koop, G. Model Switching and Model Averaging in Time-varying Parameter Regression Models. Adv. Econ. 2014, 34, 45–69. [Google Scholar]
  73. Drachal, K. Determining Time-varying Drivers of Spot Oil Price in a Dynamic Model Averaging Framework. Energies 2018, 11, 1207. [Google Scholar] [CrossRef]
  74. Barbieri, M.; Berger, J. Optimal Predictive Model Selection. Ann. Stat. 2004, 32, 870–897. [Google Scholar]
  75. Diebold, F.; Mariano, R. Comparing Predictive Accuracy. J. Bus. Econ. Stat. 1995, 13, 253–263. [Google Scholar] [Green Version]
  76. Bergmeir, C.; Hyndman, R.; Koo, B. A Note on the Validity of Cross-validation for Evaluating Autoregressive Time Series Prediction. Comput. Stat. Data Anal. 2018, 120, 70–83. [Google Scholar] [CrossRef]
Figure 1. Combination of forgetting factors for DMA, DMS and MED models, which minimise RMSE. Circles are proportional to the number of models.
Figure 1. Combination of forgetting factors for DMA, DMS and MED models, which minimise RMSE. Circles are proportional to the number of models.
Sustainability 10 02801 g001
Figure 2. Normalized RMSE for all commodities vs. different values of forgetting factors.
Figure 2. Normalized RMSE for all commodities vs. different values of forgetting factors.
Sustainability 10 02801 g002
Figure 3. Normalized RMSE for all commodities vs. different values of forgetting factors.
Figure 3. Normalized RMSE for all commodities vs. different values of forgetting factors.
Sustainability 10 02801 g003
Figure 4. Normalized RMSE for all commodities vs. different values of forgetting factors.
Figure 4. Normalized RMSE for all commodities vs. different values of forgetting factors.
Sustainability 10 02801 g004
Table 1. Commodities abbreviations. Details can be found in the original source: The World Bank [5].
Table 1. Commodities abbreviations. Details can be found in the original source: The World Bank [5].
IBEVERAGESBeverages index includes cocoa, coffee and tea.
IFOODFood index includes fats and oils, grains and other food items.
IFATS_OILSFats and oils index includes coconut oil, groundnut oil, palm oil, soybeans, soybean oil and soybean meal.
IGRAINSGrains index includes barley, maize, rice and wheat.
IOTHERFOODOther food index includes bananas, beef, chicken meat, oranges and sugar.
IRAW_MATERIALAgricultural raw materials index includes timber and other raw materials.
ITIMBERTimber index includes tropical hard logs and sawn wood.
IOTHERRAWMATOther raw materials index includes cotton, natural rubber and tobacco.
IAGRICULTUREAgriculture index includes beverages, food and agricultural raw materials.
ALUMINUMAluminium (LME) London Metal Exchange, unalloyed primary ingots, high grade, minimum 99.7% purity, settlement price beginning 2005; previously cash price
BANANA_USBananas (Central & South America), major brands, US import price, free on truck US (f.o.t.) Gulf ports
BARLEYBarley (Canada), feed, Western No. 1, Winnipeg Commodity Exchange, spot, wholesale farmers’ price
COAL_AUSCoal (Australia), thermal, f.o.b. piers, Newcastle/Port Kembla, 6300 kcal/kg (11,340 btu/lb), less than 0.8%, sulfur 13% ash beginning January 2002; previously 6667 kcal/kg (12,000 btu/lb), less than 1.0% sulfur, 14% ash
COCOACocoa (ICCO), International Cocoa Organization daily price, average of the first three positions on the terminal markets of New York and London, nearest three future trading months.
COCONUT_OILCoconut oil (Philippines/Indonesia), bulk, c.i.f. Rotterdam
COFFEE_ARABICCoffee (ICO), International Coffee Organization indicator price, other mild Arabicas, average New York and Bremen/Hamburg markets, ex-dock
COFFEE_ROBUSCoffee (ICO), International Coffee Organization indicator price, Robustas, average New York and Le Havre/Marseilles markets, ex-dock
COPPERCopper (LME), grade A, minimum 99.9935% purity, cathodes and wire bar shapes, settlement price
COPRACopra (Philippines/Indonesia), bulk, c.i.f. N.W. Europe
COTTON_A_INDXCotton (Cotton Outlook “CotlookA index”), middling 1-3/32 inch, traded in Far East, C/F beginning 2006; previously Northern Europe, c.i.f.
CRUDE_PETROCrude oil, average spot price of Brent, Dubai and West Texas Intermediate, equally weighed
CRUDE_BRENTCrude oil, U.K. Brent 38’ API, f.o.b. U.K ports, spot price
CRUDE_DUBAICrude oil, Dubai Fateh 32’ API, f.o.b. Dubai, spot price
CRUDE_WTICrude oil, West Texas Intermediate (WTI) 40’ API, f.o.b. Midland Texas, spot price
DAPDAP (diammonium phosphate), standard size, bulk, spot, f.o.b. US Gulf
IENERGYEnergy index, a Laspeyres Index with fixed weights based on 2002–2004 average developing countries export values, for coal, crude oil and natural gas.
IFERTILIZERSFertilizers index includes natural phosphate rock, phosphate, potassium and nitrogenous products.
FISH_MEALFishmeal (any origin), 64–65%, c&f Bremen, estimates based on wholesale price, beginning 2004; previously c&f Hamburg
GOLDGold (UK), 99.5% fine, London afternoon fixing, average of daily rates
GRNUT_OILGroundnut oil (any origin), c.i.f. Rotterdam
LEADLead (LME), refined, 99.97% purity, settlement price
LOGS_CMRLogs (West Africa), sapele, high quality (loyal and marchand), 80 centimeter or more, f.o.b. Douala, Cameroon beginning January 1996; previously of unspecified dimension
LOGS_MYSLogs (Malaysia), meranti, Sarawak, sale price charged by importers, Tokyo beginning February 1993; previously average of Sabah and Sarawak weighted by Japanese import volumes
MAIZEMaize (US), no. 2, yellow, f.o.b. US Gulf ports
BEEFMeat, beef (Australia/New Zealand), chucks and cow forequarters, frozen boneless, 85% chemical lean, c.i.f. U.S. port (East Coast), ex-dock, beginning November 2002; previously cow forequarters
CHICKENMeat, sheep (New Zealand), frozen whole carcasses Prime Medium (PM) wholesale, Smithfield, London beginning January 2006; previously Prime Light (PL)
IMETMINMetals and minerals index includes aluminum, copper, iron ore, lead, nickle, tin and zinc.
NGAS_USNatural Gas (U.S.), spot price at Henry Hub, Louisiana
NICKELNickel (LME), cathodes, minimum 99.8% purity, settlement price beginning 2005; previously cash price
INONFUELNon-energy index, a Laspeyres Index with fixed weights based on 2002–2004 average developing countries export values, for 34 commodities contain in the agriculture, fertilizer, and metals and minerals indices.
ORANGEOranges (Mediterranean exporters) navel, EEC indicative import price, c.i.f. Paris
PALM_OILPalm oil (Malaysia), 5% bulk, c.i.f. N. W. Europe
PLATINUMPlatinum (UK), 99.9% refined, London afternoon fixing
PLYWOODPlywood (Africa and Southeast Asia), Lauan, 3-ply, extra, 91 cm × 182 cm × 4 mm, wholesale price, spot Tokyo
POTASHPotassium chloride (muriate of potash), standard grade, spot, f.o.b. Vancouver
RICE_05Rice (Thailand), 5% broken, white rice (WR), milled, indicative price based on weekly surveys of export transactions, government standard, f.o.b. Bangkok
RUBBER1_MYSGRubber (Asia), RSS3 grade, Singapore Commodity Exchange Ltd. (SICOM) nearby contract beginning 2004; during 2000 to 2003, Singapore RSS1; previously Malaysia RSS1
SAWNWD_MYSSawnwood (Malaysia), dark red seraya/meranti, select and better quality, average 7 to 8 inches; length average 12 to 14 inches; thickness 1 to 2 inch(es); kiln dry, c. & f. UK ports, with 5% agents commission including premium for products of certified sustainable forest beginning January 2005; previously excluding the premium
SHRIMP_MEXShrimp, (Mexico), west coast, frozen, white, No. 1, shell-on, headless, 26 to 30 count per pound, wholesale price at New York
SILVERSilver (Handy & Harman), 99.9% grade refined, New York
SORGHUMSorghum (US), no. 2 milo yellow, f.o.b. Gulf ports
SOYBEAN_MEALSoybean meal (any origin), Argentine 45/46% extraction, c.i.f. Rotterdam beginning 1990; previously US 44%
SOYBEAN_OILSoybean oil (Any origin), crude, f.o.b. ex-mill Netherlands
SOYBEANSSoybeans (US), c.i.f. Rotterdam
SUGAR_EUSugar (EU), European Union negotiated import price for raw unpackaged sugar from African, Caribbean and Pacific (ACP) under Lome Conventions, c.I.f. European ports
SUGAR_USSugar (US), nearby futures contract, c.i.f.
SUGAR_WLDSugar (world), International Sugar Agreement (ISA) daily price, raw, f.o.b. and stowed at greater Caribbean ports
TEA_AVGTea, average three auctions, arithmetic average of quotations at Kolkata, Colombo and Mombasa/Nairobi.
TEA_COLOMBOTea (Colombo auctions), Sri Lankan origin, all tea, arithmetic average of weekly quotes.
TEA_KOLKATATea (Kolkata auctions), leaf, include excise duty, arithmetic average of weekly quotes.
TEA_MOMBASATea (Mombasa/Nairobi auctions), African origin, all tea, arithmetic average of weekly quotes.
TINTin (LME), refined, 99.85% purity, settlement price
TOBAC_USTobacco (any origin), unmanufactured, general import, cif, US
TSPTSP (triple superphosphate), up to September 2006 bulk, spot, f.o.b. US Gulf; from October 2006 onwards Tunisian, granular, f.o.b.
UREA_EE_BULKUrea, (Black Sea), bulk, spot, f.o.b. Black Sea (primarily Yuzhnyy) beginning July 1991; for 1985–1991 (June) f.o.b. Eastern Europe
WHEAT_US_HRWWheat (US), no. 1, hard red winter, ordinary protein, export price delivered at the US Gulf port for prompt or 30 days shipment
WHEAT_US_SRWWheat (US), no. 2, soft red winter, export price delivered at the US Gulf port for prompt or 30 days shipment
WOODPULPWoodpulp (Sweden), softwood, sulphate, bleached, air-dry weight, c.i.f. North Sea ports
ZINCZinc (LME), high grade, minimum 99.95% purity, settlement price beginning April 1990; previously special high grade, minimum 99.995%, cash prices
Table 2. Descriptive statistics.
Table 2. Descriptive statistics.
VariableMeanStandard DeviationMinMaxSkewnessKurtosis
dpr−0.01340.0035−0.0200−0.00570.2356−0.8496
str0.03580.02770.00010.10470.2098−1.0013
ltr0.05530.02630.01500.13560.5970−0.0447
ts0.01950.0107−0.00530.0376−0.2611−0.9187
drs0.03270.01260.00360.0593−0.2889−0.8796
cpi0.00220.0026−0.01790.0137−1.401711.2535
ip0.00170.0061−0.04400.0203−1.56429.4380
une0.06090.01460.03800.10000.7915−0.0500
aud−0.00040.0332−0.17980.0885−0.77222.7502
inr−0.00450.0186−0.19560.0594−3.142228.1909
kei−0.01670.2709−1.33240.66610.05051.5030
op0.01360.0224−0.07270.0937−0.11243.4690
tr0.00100.0338−0.20100.1168−0.71194.1604
m0.00470.0083−0.03370.05741.52179.2340
IBEVERAGES−0.00020.0462−0.15920.28700.85294.7751
IFOOD0.00090.0305−0.18770.1500−0.19214.6650
IFATS_OILS0.00060.0459−0.25400.1998−0.32303.6843
IGRAINS0.00050.0415−0.19620.18370.13312.2778
IOTHERFOOD0.00190.0336−0.10160.10550.11530.3257
IRAW_MATERIAL0.00120.0256−0.09010.11030.32252.7758
ITIMBER0.00160.0338−0.13170.20111.06827.1127
IOTHERRAWMAT0.00070.0347−0.19670.1147−0.50563.8493
IAGRICULTURE0.00080.0241−0.15910.0941−0.48365.1574
ALUMINUM0.00080.0561−0.32620.1801−0.56253.6371
BANANA_US0.00280.1619−0.45560.59340.39111.5758
BARLEY0.00150.0680−0.27880.2832−0.04792.5754
COAL_AUS0.00300.0558−0.32850.36370.26098.5692
COCOA−0.00050.0569−0.19480.23090.08230.8005
COCONUT_OIL0.00100.0782−0.25980.35120.33551.2619
COFFEE_ARABIC−0.00010.0756−0.35250.42270.69714.0604
COFFEE_ROBUS−0.00070.0673−0.24980.37790.63953.4206
COPPER0.00390.0635−0.35010.2492−0.34964.1479
COPRA0.00100.0760−0.21280.32580.36441.3311
COTTON_A_INDX−0.00030.0560−0.26900.2006−0.04792.7694
CRUDE_PETRO0.00160.0874−0.43880.4304−0.41093.7925
CRUDE_BRENT0.00170.0910−0.38340.4326−0.26482.7104
CRUDE_DUBAI0.00170.0928−0.54010.4910−0.52906.1827
CRUDE_WTI0.00140.0851−0.39680.3755−0.41182.7034
DAP0.00120.0619−0.45970.2551−0.968112.9591
IENERGY0.00150.0720−0.33370.3449−0.40102.8708
IFERTILIZERS0.00200.0461−0.27490.2425−0.38165.6836
FISH_MEAL0.00260.0433−0.15420.20020.22021.3521
GOLD0.00290.0356−0.12480.16010.34941.3229
GRNUT_OIL0.00120.0515−0.21020.26070.36733.9632
LEAD0.00450.0702−0.29330.3091−0.30712.6723
LOGS_CMR0.00210.0345−0.17430.1440−0.24173.5366
LOGS_MYS0.00130.0500−0.19310.29410.81655.0984
MAIZE0.00010.0586−0.24480.2975−0.13473.5803
BEEF0.00150.0395−0.17800.1432−0.16292.4215
CHICKEN0.00260.0205−0.05580.10450.80583.0832
IMETMIN0.00210.0496−0.30450.1585−0.80805.0908
NGAS_US0.00020.1241−0.40550.47790.06911.6753
NICKEL0.00220.0856−0.38240.58110.72315.7015
INONFUEL0.00120.0265−0.20320.1002−1.13119.2270
ORANGE0.00220.1233−0.41990.4127−0.21081.0353
PALM_OIL0.00010.0739−0.34690.2731−0.37352.4278
PLATINUM0.00210.0527−0.29290.2335−0.44844.7605
PLYWOOD0.00190.0393−0.15830.17810.83624.3207
POTASH0.00250.0403−0.41580.3438−1.037541.9405
RICE_050.00110.0580−0.24240.42331.28968.8429
RUBBER1_MYSG0.00090.0708−0.39040.2121−0.43353.1845
SAWNWD_MYS0.00180.0346−0.14010.23821.133510.5032
SHRIMP_MEX−0.00010.0424−0.24320.1839−0.63457.0185
SILVER0.00150.0621−0.21430.26040.11651.6884
SORGHUM0.00060.0634−0.27770.3419−0.04463.8254
SOYBEAN_MEAL0.00080.0567−0.18580.21760.21461.6060
SOYBEAN_OIL0.00070.0557−0.27850.2611−0.04533.2895
SOYBEANS0.00060.0503−0.25610.2140−0.24552.9699
SUGAR_EU0.00010.0364−0.30520.2568−1.470124.9057
SUGAR_US0.00060.0314−0.13660.16700.07694.7483
SUGAR_WLD0.00150.0832−0.30800.31280.18170.7727
TEA_AVG0.00070.0524−0.19600.23780.21141.7555
TEA_COLOMBO0.00140.0699−0.46810.4046−0.24968.3187
TEA_KOLKATA−0.00040.1224−0.31000.47181.13492.1171
TEA_MOMBASA0.00110.0704−0.46130.45500.067410.5021
TIN0.00120.0567−0.25170.1615−0.39402.5582
TOBAC_US0.00150.0185−0.04860.09970.46622.0420
TSP0.00160.0577−0.46520.3021−1.483919.0871
UREA_EE_BULK0.00200.0857−0.55480.2863−1.08117.0126
WHEAT_US_HRW0.00030.0594−0.21920.22910.35102.0921
WHEAT_US_SRW0.00050.0669−0.26040.2582−0.06201.7340
WOODPULP0.00170.0407−0.24300.1195−0.90214.2056
ZINC0.00330.0635−0.28730.2440−0.37771.6376
Table 3. Stationarity tests.
Table 3. Stationarity tests.
VariableADF Stat.ADF p-ValuePP Stat.PP p-ValueKPSS Stat.KPSS p-Value
dpr−2.10520.5329−5.75840.78803.87420.0100
str−4.24310.0100−10.89350.50096.27360.0100
ltr−5.36440.0100−26.34740.01877.27370.0100
ts−3.79990.0192−16.89270.16540.25840.1000
drs−3.64110.0291−13.16740.37370.64840.0182
cpi−7.23340.0100−213.85120.01001.26060.0100
ip−5.38680.0100−423.84740.01000.51080.0392
une−2.75540.2583−4.60990.85230.63160.0198
aud−7.15200.0100−378.71650.01000.09130.1000
inr−5.72440.0100−299.02630.01000.56560.0269
kei−2.56810.3374−24.25510.02800.98750.0100
op−6.12130.0100−48.53970.01000.65580.0176
tr−6.89770.0100−438.08740.01000.16130.1000
m−4.26910.0100−440.12500.01000.67700.0156
IBEVERAGES−7.00480.0100−308.14860.01000.12710.1000
IFOOD−7.32670.0100−268.45240.01000.10370.1000
IFATS_OILS−7.28730.0100−254.97480.01000.10660.1000
IGRAINS−7.42730.0100−265.49550.01000.07460.1000
IOTHERFOOD−8.29800.0100−324.53340.01000.04770.1000
IRAW_MATERIAL−5.94600.0100−257.00400.01000.07940.1000
ITIMBER−6.25040.0100−252.23720.01000.09520.1000
IOTHERRAWMAT−6.49090.0100−220.25850.01000.07100.1000
IAGRICULTURE−7.29530.0100−262.93620.01000.12470.1000
ALUMINUM−7.19610.0100−365.21500.01000.03410.1000
BANANA_US−11.96150.0100−386.47370.01000.01020.1000
BARLEY−6.61440.0100−295.18780.01000.03980.1000
COAL_AUS−7.25260.0100−300.59870.01000.05890.1000
COCOA−6.96160.0100−328.76480.01000.13770.1000
COCONUT_OIL−6.08530.0100−310.37570.01000.12010.1000
COFFEE_ARABIC−6.92420.0100−303.89050.01000.05040.1000
COFFEE_ROBUS−6.44210.0100−286.35180.01000.13320.1000
COPPER−8.30530.0100−229.56960.01000.05590.1000
COPRA−6.21230.0100−298.21830.01000.12910.1000
COTTON_A_INDX−7.62770.0100−186.27060.01000.04120.1000
CRUDE_PETRO−8.00390.0100−276.58550.01000.07380.1000
CRUDE_BRENT−8.07580.0100−302.91280.01000.07440.1000
CRUDE_DUBAI−8.14820.0100−272.92480.01000.07650.1000
CRUDE_WTI−7.89430.0100−267.01520.01000.06690.1000
DAP−7.89950.0100−154.94840.01000.05550.1000
IENERGY−7.74450.0100−269.36150.01000.08340.1000
IFERTILIZERS−6.57340.0100−208.01590.01000.07690.1000
FISH_MEAL−7.28240.0100−219.04140.01000.08760.1000
GOLD−6.28780.0100−334.65870.01000.39270.0803
GRNUT_OIL−5.64870.0100−203.14130.01000.04500.1000
LEAD−6.50340.0100−323.01230.01000.06100.1000
LOGS_CMR−7.60970.0100−326.35370.01000.09150.1000
LOGS_MYS−7.18260.0100−221.09010.01000.03560.1000
MAIZE−7.67580.0100−293.51800.01000.06670.1000
BEEF−7.87110.0100−259.08610.01000.08540.1000
CHICKEN−9.17320.0100−200.72750.01000.01440.1000
IMETMIN−6.99040.0100−293.49940.01000.05560.1000
NGAS_US−8.73170.0100−322.09520.01000.04270.1000
NICKEL−6.18450.0100−270.50220.01000.06120.1000
INONFUEL−7.52800.0100−256.67030.01000.10510.1000
ORANGE−12.87750.0100−249.76390.01000.01190.1000
PALM_OIL−7.23180.0100−271.77180.01000.08250.1000
PLATINUM−7.78530.0100−301.48660.01000.08890.1000
PLYWOOD−6.81550.0100−244.47420.01000.13930.1000
POTASH−4.57560.0100−268.40680.01000.12280.1000
RICE_05−8.81390.0100−224.98170.01000.04450.1000
RUBBER1_MYSG−6.11890.0100−302.86960.01000.07370.1000
SAWNWD_MYS−6.21020.0100−304.18560.01000.11670.1000
SHRIMP_MEX−6.90830.0100−313.09720.01000.02900.1000
SILVER−7.17890.0100−320.03270.01000.23180.1000
SORGHUM−8.67410.0100−306.92720.01000.05510.1000
SOYBEAN_MEAL−7.55510.0100−281.16730.01000.06790.1000
SOYBEAN_OIL−6.66350.0100−261.98710.01000.08840.1000
SOYBEANS−7.70680.0100−305.99130.01000.07510.1000
SUGAR_EU−7.82970.0100−392.68560.01000.29770.1000
SUGAR_US−6.60060.0100−263.90910.01000.03600.1000
SUGAR_WLD−7.51190.0100−264.40070.01000.03520.1000
TEA_AVG−7.99400.0100−407.11720.01000.15100.1000
TEA_COLOMBO−7.75830.0100−340.44240.01000.14500.1000
TEA_KOLKATA−12.46790.0100−318.43010.01000.03450.1000
TEA_MOMBASA−7.13340.0100−331.52620.01000.04680.1000
TIN−7.44420.0100−309.77540.01000.23390.1000
TOBAC_US−5.66080.0100−297.58820.01000.12410.1000
TSP−7.24200.0100−173.85120.01000.04680.1000
UREA_EE_BULK−7.25600.0100−221.07770.01000.04280.1000
WHEAT_US_HRW−7.68250.0100−291.12230.01000.04790.1000
WHEAT_US_SRW−8.59030.0100−304.06120.01000.03820.1000
WOODPULP−6.61870.0100−230.43870.01000.03320.1000
ZINC−6.35000.0100−291.22580.01000.05780.1000
Table 4. Normalized RMSEs.
Table 4. Normalized RMSEs.
COMMODITYDMADMSMEDTVPARIMANAIVE
IBEVERAGES0.04640.04830.04630.04770.04730.0475
IFOOD0.03160.03140.03150.03180.03340.0353
IFATS_OILS0.04610.04610.04700.04650.04870.0510
IGRAINS0.04790.04830.04810.04810.04930.0514
IOTHERFOOD0.03060.03050.03140.03200.03180.0313
IRAW_MATERIAL0.02550.02580.02580.02590.02550.0272
ITIMBER0.02790.02850.02860.02870.02710.0286
IOTHERRAWMAT0.04620.04610.04560.04650.04760.0507
IAGRICULTURE0.02490.02480.02470.02500.02730.0283
ALUMINUM0.04790.04920.04780.04970.05310.0526
BANANA_US0.12350.12310.12250.13240.11760.1223
BARLEY0.06260.06310.06310.06330.06320.0654
COAL_AUS0.09200.09280.09260.09370.11880.0981
COCOA0.05730.05710.05710.05970.05800.0578
COCONUT_OIL0.08540.08410.08390.08720.09020.0884
COFFEE_ARABIC0.08500.08560.08570.08780.08810.0858
COFFEE_ROBUS0.07820.07890.07790.07860.08010.0788
COPPER0.07380.07340.07380.07360.07930.0786
COPRA0.08950.08770.08770.09180.09840.0910
COTTON_A_INDX0.06940.06770.06900.06960.06890.0808
CRUDE_PETRO0.08110.08130.08120.08790.08670.0919
CRUDE_BRENT0.08410.08410.08340.09140.09080.0942
CRUDE_DUBAI0.07890.07950.07910.08580.08460.0924
CRUDE_WTI0.08660.08640.08580.09260.08920.0947
DAP0.07740.07430.07630.08100.08640.1104
IENERGY0.07100.07060.07100.07750.07930.0814
IFERTILIZERS0.06810.06470.06560.06910.07790.0904
FISH_MEAL0.04500.04480.04480.04610.04470.0475
GOLD0.04320.04240.04490.04470.04730.0454
GRNUT_OIL0.04370.04420.04500.04620.04670.0540
LEAD0.09490.09690.09640.09790.10150.0960
LOGS_CMR0.03370.03410.03410.03450.03390.0351
LOGS_MYS0.04990.04920.04940.04990.05110.0542
MAIZE0.06960.07010.07030.07210.07140.0717
BEEF0.04460.04380.04410.04550.04510.0474
CHICKEN0.01350.01320.01310.01340.01280.0159
IMETMIN0.05410.05430.05410.05600.05920.0602
NGAS_US0.18040.18220.18100.18170.19200.1794
NICKEL0.11160.10970.11000.11280.11770.1182
INONFUEL0.02750.02740.02730.02770.03060.0331
ORANGE0.12040.11960.11970.12720.11410.1210
PALM_OIL0.06490.06460.06420.06740.06820.0707
PLATINUM0.06280.06240.06350.06350.06810.0683
PLYWOOD0.03200.03240.03280.03290.03100.0336
POTASH0.08290.08360.08430.08470.10100.0891
RICE_050.08360.08280.08280.08240.08580.0805
RUBBER1_MYSG0.10070.10000.09980.10310.10800.1046
SAWNWD_MYS0.02920.03490.03370.03040.02830.0291
SHRIMP_MEX0.03890.03940.03920.03990.03890.0399
SILVER0.09100.09130.08950.09910.10390.0982
SORGHUM0.07000.07050.07140.07190.07090.0714
SOYBEAN_MEAL0.06310.06200.06250.06650.06390.0658
SOYBEAN_OIL0.05130.05040.05110.05310.05500.0565
SOYBEANS0.05880.05890.05890.06000.06000.0605
SUGAR_EU0.03230.03250.03210.03320.03220.0322
SUGAR_US0.03930.03900.03880.04000.03980.0406
SUGAR_WLD0.07990.07950.07940.08240.07840.0822
TEA_AVG0.04860.04860.04860.05090.04900.0483
TEA_COLOMBO0.05350.05300.05300.05840.05270.0529
TEA_KOLKATA0.12220.12240.12150.12610.11470.1207
TEA_MOMBASA0.06950.06890.06910.07290.06870.0678
TIN0.07480.07590.07580.07540.08200.0789
TOBAC_US0.01890.02240.01920.01960.01920.0192
TSP0.09180.09240.09150.09410.10350.1251
UREA_EE_BULK0.13270.12970.13010.12420.13930.1406
WHEAT_US_HRW0.07480.07490.07470.07690.07720.0768
WHEAT_US_SRW0.08530.08530.08520.08730.08800.0860
WOODPULP0.03300.03300.03310.03450.03310.0378
ZINC0.07780.07800.07870.08110.07890.0815
Table 5. p-Values of the Diebold–Mariano tests for forecast accuracy. Details are in the text.
Table 5. p-Values of the Diebold–Mariano tests for forecast accuracy. Details are in the text.
COMMODITYDMA < NAIVEDMA < ARIMADMA > NAIVEDMA > ARIMABESTDM p-Value
IBEVERAGES0.000.00 MED0.00
IFOOD0.000.00 DMS0.10
IFATS_OILS0.000.00 DMA0.02
IGRAINS0.000.00 DMA0.17
IOTHERFOOD0.000.00 DMS0.00
IRAW_MATERIAL0.00 1.00ARIMA1.00
ITIMBER0.00 1.00ARIMA1.00
IOTHERRAWMAT0.000.00 MED0.00
IAGRICULTURE0.000.00 MED0.03
ALUMINUM0.000.00 MED0.08
BANANA_US 0.000.00ARIMA0.00
BARLEY0.000.00 DMA0.00
COAL_AUS0.000.00 DMA0.00
COCOA0.000.00 MED0.00
COCONUT_OIL0.000.00 MED0.01
COFFEE_ARABIC0.000.00 DMA0.00
COFFEE_ROBUS0.000.00 MED0.36
COPPER0.000.00 DMS0.07
COPRA0.000.00 DMS0.00
COTTON_A_INDX0.00 1.00DMS0.00
CRUDE_PETRO0.000.00 DMA0.00
CRUDE_BRENT0.000.00 MED0.00
CRUDE_DUBAI0.000.00 DMA0.00
CRUDE_WTI0.000.00 MED0.00
DAP0.000.00 DMS0.08
IENERGY0.000.00 DMS0.00
IFERTILIZERS0.000.00 DMS0.38
FISH_MEAL0.00 1.00ARIMA1.00
GOLD0.000.00 DMS0.18
GRNUT_OIL0.000.00 DMA0.00
LEAD0.000.00 DMA0.00
LOGS_CMR0.000.00 DMA0.00
LOGS_MYS0.000.00 DMS0.17
MAIZE0.000.00 DMA0.00
BEEF0.000.00 DMS0.00
CHICKEN0.00 1.00ARIMA1.00
IMETMIN0.000.00 MED0.02
NGAS_US 0.001.00 NAIVE1.00
NICKEL0.000.00 DMS0.03
INONFUEL0.000.00 MED0.10
ORANGE1.00 0.00ARIMA0.00
PALM_OIL0.000.00 MED0.02
PLATINUM0.000.00 DMS0.11
PLYWOOD0.00 1.00ARIMA1.00
POTASH0.010.00 DMA0.18
RICE_05 0.000.99 NAIVE0.99
RUBBER1_MYSG0.000.00 MED0.76
SAWNWD_MYS 1.001.00ARIMA1.00
SHRIMP_MEX0.000.00 DMA0.00
SILVER0.000.00 MED0.00
SORGHUM0.000.00 DMA0.00
SOYBEAN_MEAL0.000.00 DMS0.00
SOYBEAN_OIL0.000.00 DMS0.01
SOYBEANS0.000.00 DMA0.00
SUGAR_EU 0.000.00MED1.00
SUGAR_US1.001.00 MED1.00
SUGAR_WLD1.00 0.00ARIMA0.00
TEA_AVG 0.001.00 NAIVE1.00
TEA_COLOMBO 1.001.00ARIMA1.00
TEA_KOLKATA 1.001.00ARIMA1.00
TEA_MOMBASA 1.001.00NAIVE1.00
TIN0.000.00 DMA0.08
TOBAC_US0.000.00 DMA0.00
TSP0.010.00 MED0.08
UREA_EE_BULK0.000.00 TVP0.89
WHEAT_US_HRW0.000.00 MED0.00
WHEAT_US_SRW0.000.00 MED0.00
WOODPULP0.000.00 DMS0.00
ZINC0.000.00 DMA0.00
Table 6. Summary of results from Table 5.
Table 6. Summary of results from Table 5.
MODELFREQ. BESTFREQ. SIGN. BEST
DMA26%22%
DMS23%13%
MED28%17%
TVP1%0%
ARIMA16%4%
NAIVE6%0%
BMA0%0%
BMS0%0%
BMED0%0%

Share and Cite

MDPI and ACS Style

Drachal, K. Some Novel Bayesian Model Combination Schemes: An Application to Commodities Prices. Sustainability 2018, 10, 2801. https://doi.org/10.3390/su10082801

AMA Style

Drachal K. Some Novel Bayesian Model Combination Schemes: An Application to Commodities Prices. Sustainability. 2018; 10(8):2801. https://doi.org/10.3390/su10082801

Chicago/Turabian Style

Drachal, Krzysztof. 2018. "Some Novel Bayesian Model Combination Schemes: An Application to Commodities Prices" Sustainability 10, no. 8: 2801. https://doi.org/10.3390/su10082801

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop