Next Article in Journal
Bridging the Gap between Automated Manufacturing of Fuel Cell Components and Robotic Assembly of Fuel Cell Stacks
Next Article in Special Issue
Review and Comparison of Intelligent Optimization Modelling Techniques for Energy Forecasting and Condition-Based Maintenance in PV Plants
Previous Article in Journal
Optimal Control of Microgrid Operation Based on Fuzzy Sliding Mode Droop Control
Previous Article in Special Issue
Research and Application of a Novel Hybrid Model Based on a Deep Neural Network Combined with Fuzzy Time Series for Energy Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Daily Crude Oil Prices Using Improved CEEMDAN and Ridge Regression-Based Predictors

1
School of Economic Information Engineering, Southwestern University of Finance and Economics, Chengdu 611130, China
2
Sichuan Province Key Laboratory of Financial Intelligence and Financial Engineering, Southwestern University of Finance and Economics, Chengdu 611130, China
3
College of Computer Science, Sichuan University, Chengdu 610064, China
4
Architectural Design Institute, Nuclear Power Institute of China, Chengdu 610213, China
*
Authors to whom correspondence should be addressed.
Energies 2019, 12(19), 3603; https://doi.org/10.3390/en12193603
Submission received: 8 August 2019 / Revised: 9 September 2019 / Accepted: 18 September 2019 / Published: 20 September 2019
(This article belongs to the Special Issue Intelligent Optimization Modelling in Energy Forecasting)

Abstract

:
As one of the leading types of energy, crude oil plays a crucial role in the global economy. Understanding the movement of crude oil prices is very attractive for producers, consumers and even researchers. However, due to its complex features of nonlinearity and nonstationarity, it is a very challenging task to accurately forecasting crude oil prices. Inspired by the well-known framework “decomposition and ensemble” in signal processing and/or time series forecasting, we propose a new approach that integrates the improved complete ensemble empirical mode decomposition with adaptive noise (ICEEMDAN), differential evolution (DE) and several types of ridge regression (RR), namely, ICEEMDAN-DE-RR, for more accurate crude oil price forecasting in this paper. The proposed approach consists of three steps. First, we use the ICEEMDAN to decompose the complex daily crude oil price series into several relatively simple components. Second, ridge regression or kernel ridge regression is employed to forecast each decomposed component. To enhance the accuracy of ridge regression, DE is used to jointly optimize the regularization item, the weights and parameters of each single kernel for each component. Finally, the predicted results of all components are aggregated as the final predicted results. The publicly available West Texas Intermediate (WTI) daily crude oil spot prices are used to validate the performance of the proposed approach. The experimental results indicate that the proposed approach can achieve better performance than some state-of-the-art approaches in terms of several evaluation criteria, demonstrating that the proposed ICEEMDAN-DE-RR is very promising for daily crude oil price forecasting.

Graphical Abstract

1. Introduction

Crude oil is one of the leading types of energy that has great impacts on the global economy. Trying to accurately expect changes in crude oil prices benefits the producers and consumers of crude oil. However, the prices are affected by many factors, such as climate, exchange rate, supply and demand, speculation activities, geopolitics and so on, and they have fluctuated drastically in the last decades [1,2]. For example, the prices of West Texas Intermediate (WTI) reached over 145 USD/barrel in July 2008 and then quickly reduced to about 30 USD/barrel in the next five months. Crude oil prices have shown significant nonlinearity and nonstationarity in the last three decades. The complex fluctuation of crude oil prices makes it a very challenging task to accurately predict crude oil prices. Despite this, many researchers have contributed to building automatic models to forecast crude oil prices accurately.
The task of forecasting crude oil prices is to expect future prices using existing data. From the perspective of the input of the forecasting task, it can be divided into two groups: multivariate forecasting and univariate forecasting. The former usually feeds the data associated with types of variables, such as macroeconomic variables, exchange rates, sentiment analysis, inventory variables, previous crude oil prices, and so on, to the predictors [1,2,3,4,5,6,7], while the latter uses the previous prices only [8,9,10,11,12]. These are two different perspectives for studying crude oil price forecasting. In practical applications, the former is usually used to forecast long-term crude oil prices, for example, monthly prices or weekly prices, while the latter is for daily prices in most cases. In the task of forecasting crude oil prices, the predictors can be mainly categorized into two classes: statistical models (econometric models) and artificial intelligence (AI) models. Mirmirani and Li employed vector auto-regression (VAR) to forecast the movements of U.S. oil prices [13]. Murat and Tokat found that a vector error correction model (VECM) with crack spread futures outperformed the traditional random walk (RW) model [14]. Moshiri and Foroutan applied auto-regressive integrated moving average (ARIMA) and another statistical model, generalized autoregressive conditional heteroskedasticity (GARCH), to forecasting daily crude oil futures prices [15]. Some extensions of GARCH have also been employed in recent years [16,17,18]. In Lyocsa and Molnar’s study, the authors investigated whether the heterogeneous autoregressive (HAR) model can improve the results of forecasting the volatility of crude oil prices by using information from related energy commodity, and the experimental results demonstrated that such information can not improve the volatility forecasting [19]. Lv studied the performance of the HAR model of realized volatility (HAR-RV) for forecasting crude oil futures price volatility [20]. Naser found that the dynamic model averaging (DMA) model showed better performance in forecasting crude oil prices than all the other alternative models, and it could also achieve better results of forecasting spot prices than futures prices [21]. Azevedo and Campos combined ARIMA, exponential smoothing, and dynamic regression to forecast WTI and Brent crude oil spot prices, and the experimental results indicated that the combined model was promising for crude oil price forecasting [22].
The statistical models are usually built on the assumption of linearity and stationarity of the predicted time series. However, most research has shown that crude oil prices are highly nonlinear and nonstationary [8,12,23], so such characteristics limit the accuracy of statistical models for forecasting crude oil prices. To cope with the complex characteristics of crude oil prices, more and more scholars use AI models to forecast crude oil prices. The most popular AI models include artificial neural network (ANN) [12,24,25,26,27], support vector regression (SVR) [28,29] and least squares SVR (LSSVR) [2,10,30], sparse Bayesian learning (SBL) [31,32], extreme learning machine (ELM) [23,33,34], extreme gradient boosting (XGBoost) [8], random vector functional link (RVFL) network [11], long short-term memory (LSTM) [35], and so on. Yu et al. used a feed-forward neural network (FNN) to forecast each decomposed component from the raw series of crude oil prices, and then integrated the results of all components as the final forecasting result by an adaptive linear neural network (ALNN) [12]. Xiong et al. also used FNN to conduct multi-step-ahead weekly crude oil spot price forecasting [25]. Barunik and Malinska found that a focused time-delay NN could achieve higher accuracy than the compared models when forecasting monthly crude oil prices [26]. Extensive research has demonstrated that the kernel-based methods, such as SVR, LSSVR and relevance vector machine, are promising for forecasting crude oil prices [2,10,28,29,30,36]. Very recently, the LSTM, an artificial recurrent NN architecture widely used in deep learning, has been applied to forecasting crude oil prices. Owing to its power in processing sequences of data, the LSTM-based approach has yielded very promising forecasting results [35]. Chiroma et al. presented an extensive review of the research that applied AI-based models to crude oil price forecasting [37].
Due to the nonlinearity and nonstationarity of crude oil price series, statistical models and AI models usually cannot achieve satisfactory results by conducting forecasting with the raw crude oil prices directly. A simple but effective way is to adopt the “divide and conquer” strategy, that is, to decompose the complex signal into several relatively simple components and then extract relevant features or handle each component for further work. Typical applications of such strategy include fault diagnosis [38,39], biosignal analysis [40,41], time series forecasting [42,43], and so on [44,45,46]. Following this strategy, a “decomposition and ensemble” framework has become very popular in the field of energy forecasting, such as wind speed forecasting [47,48], load forecasting [49,50,51] and price forecasting [8,11,52,53] in recent years. Ren et al. integrated empirical mode decomposition (EMD) and SVR to forecast wind speed. Specifically, EMD was used to decompose the raw wind speed series into a couple of components (a residue and a few intrinsic mode functions (IMFs)). After that, SVR was used to build an individual forecasting model for each component. At last, the predicted values of all the components were aggregated as the final forecasting result [47]. Similarly, Li et al. utilized an extended EMD, namely, ensemble EMD (EEMD), and random forests (RF) for electricity consumption forecasting [31], and Yang and Wang applied complementary EEMD (CEEMD) and back propagation NN (BPNN) to forecast wind speed [54]. All the research indicates that the approaches following the framework of “decomposition and ensemble” is capable of significantly improving the accuracy of energy forecasting.
Regarding the decomposition methods, although EMD, EEMD and CEEMD have the ability of improving accuracy, they may introduce new noise into the recovered signal and still suffer from the “mode mixing” problem [55,56]. To solve this problem, a complete EEMD with adaptive noise (CEEMDAN) and an improved CEEMDAN (ICEEMDAN) were proposed [57,58]. The existing study has demonstrated the power of CEEMDAN/ICEEMDAN in energy forecasting [8,59]. As far as the forecasting method for each component, so-called individual forecasting, any regression methods can be selected for this purpose in theory. Besides the above-mentioned methods such as SVR, ANN, ELM, LSTM and so on, ridge regression (RR) is a simple but powerful regression for forecasting. Moreover, the accuracy of regression can be further improved by introducing kernels into RR (KRR), and the KRR has been successfully applied in wind speed forecasting [60,61], object tracking [62], and preheat temperature prediction [63]. The basic idea of the kernel trick is to map the features of the low-dimensional space to the high-dimensional space to obtain more representative features. Naik et al. used wavelet kernel RR and low rank multi-kernel RR to forecast the components of wind speed and wind power decomposed by EMD and variational mode decomposition (VMD), respectively [61,64]. The low rank multi-kernel RR in their approach was a simple linear combination by a polynomial kernel and a wavelet kernel, that is to say, the multi-kernel was actually a combination of two simple kernels [64]. The number and type of kernels may limit the performance of the proposed approach. Qian et al. used multi-kernel RR for object tracking, where the final kernel included one linear kernel, five polynomial kernels and five Gaussian kernels, and the parameter of each kernel was specified in advance and the optimization process was only to optimize the weight of each kernel [62]. Regarding multi-kernel learning, an ideal way is to optimize the weights and the parameters of each kernel together. The nature-inspired algorithms, such as particle swarm optimization (PSO) [65,66,67], differential evolution (DE) [68,69], ant colony optimization [70,71] and so on, can be applied to optimizing both the weights and the parameters of the multi-kernel learning. Among the algorithms, DE has proven to be very powerful for numerical optimization.
Motivated by the potential of ICEEMDAN in signal decomposition, RR in regression and DE in numerical optimization, we proposed a novel approach integrating ICEEMDAN, DE and RR, namely, ICEEMDAN-DE-RR, for crude oil price forecasting in this paper. Specifically, the ICEEMDAN-DE-RR consists of three steps. First, ICEEMDAN is employed to decompose the raw daily crude oil price series into several relatively simple components. Second, we use RR or KRR optimized by DE to forecast each component individually. Finally, the predicted results from each component are aggregated as the final forecasting result.
The main contributions of this paper lie in the following:
(1)
We propose a new framework of multiple kernel learning, which simultaneously optimizes the weights and parameters of kernels using nature-inspired optimization.
(2)
We forecast crude oil prices by integrating ICEEMDAN, DE and RR, following the “decomposition and ensemble” framework. To the best of our knowledge, it is the first time that this combination is used for forecasting tasks.
(3)
The experimental results indicate that the proposed approach is effective for crude oil price forecasting.
It is worth pointing out that although there have existed lots of models following the “decomposition and ensemble” framework for energy forecasting, the proposed models are different from them in several aspects. First, it is an attempt to use RR to forecast crude oil prices for the first time. Existing research focuses on using RR to forecast wind speed, hydrologic time series, real estate appraisal and so on [60,72,73], not including crude oil prices. Second, a new multiple kernel learning framework is proposed using DE to optimize the parameters and/or weight of each base kernel, as well as the regularization item simultaneously. The experimental results demonstrate the effectiveness of the proposed approach. Third, the integration of ICEEMDAN, DE and RR is used to forecast time series for the first time.
The novelty of this paper is three-fold: (1) Based on the power of ICEEMDAN, DE, and RR in signal decomposition, numerical optimization, and regression, respectively, a novel combination of these three methods is proposed for time series forecasting; (2) To improve the representability of kernels, a novel multiple kernel learning framework using DE to simultaneously optimize the weights and parameters of every single kernel is proposed, which can be applied to both classification and regression; (3) The proposed ICEEMDAN-DE-RR approaches are firstly applied to forecasting crude oil prices and the results demonstrate the effectiveness of the approaches.
The remainder of this paper is structured as follows. In Section 2, we briefly introduce ICEEMDAN, DE and RR. Section 3 formulates the proposed ICEEMDAN-DE-RR method in detail. To evaluate the proposed ICEEMDAN-DE-RR, we report and analyze the experimental results in Section 4. Finally, we conclude the paper in Section 5.

2. Methods

2.1. Improved Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN)

ICEEMDAN was originated from EMD proposed by Huang et al. [55,56,57,58]. EMD is an adaptive method for time-space analysis, which decomposes a raw sequence that is non-linear and non-stationary into several IMFs and one residue. The main steps of EMD are described as follows:
Step 1:
Find out all local extrema of the raw data x ( t ) , t = 1 , 2 , 3 , , T ;
Step 2:
Link all local minima and local maxima to construct the lower envelopes x l o w ( t ) and upper envelopes x u p ( t ) , respectively;
Step 3:
Compute the local mean, i.e., m ( t ) = x u p ( t ) + x l o w ( t ) 2 ;
Step 4:
Extract the first IMF and residue by I M F 1 ( t ) = x ( t ) m ( t ) and R 1 ( t ) = m ( t ) , respectively;
Step 5:
For i = 1 , 2 , 3 , , n , if find out more than two local extrema of R i ( t ) , go back to step 2 and get I M F i + 1 ( t ) and R i + 1 ( t ) .
In EMD, it was found that there were similar parts of signals existing at the same corresponding position in different IMFs, which was called mode mixing. Because of it, IMFs have lost their physical meanings [55]. To cope with this issue, Wu and Huang proposed Ensemble EMD (EEMD) by performing EMD many times on the time series with added white noises [56]. The new time series with white noises x i ( t ) can be formulated as Equation (1):
x i ( t ) = x ( t ) + w i ( t ) ,
where x ( t ) represents the raw data series and w i ( t ) is the i-th white noise, i = 1 , 2 , 3 , , n .
Then, when every x i ( t ) is decomposed, we can get the corresponding I M F k i ( t ) . To compute the real k-th IMF, I M F k ¯ , EEMD calculates the average of the I M F k i ( t ) , which can remove the effect of the white noises. However, in practice, one of the limitations of EEMD is that the recovered signal will include residual noise. To solve this problem, Torres et al. proposed a new extension of EEMD, termed as CEEMDAN for signal decomposition [57]. For k = 2 , 3 , , K , the k-th IMF and residue can be computed as Equations (2) and (3):
I M F k ¯ = 1 / I i = 1 I ( E 1 ( r k 1 [ t ] + ε k 1 E 1 ( w i [ t ] ) ) ,
r k [ t ] = r k 1 [ t ] I M F k ¯ ,
where E 1 ( . ) represents the first IMF decomposed from the series, and ε i is used to set the signal-to-noise (SNR) at each stage.
In 2014, Colominas et al. found that the IMFs of CEEMDAN contained some residual noise and some “spurious” modes. Thus, they further proposed a method to improve CEEMDAN (ICEEMDAN) [57,58], whose main steps can be described as follows:
Step 1:
Add the first IMF of the given white noises to the original series x ( t ) , as shown in following:
x i ( t ) = x ( t ) + β 0 E 1 ( w i ( t ) ) ,
where β 0 is the level of noise.
Step 2:
Find out the local means M ( . ) of x i ( t ) and calculate the average of local means to get the following residue:
r 1 ( t ) = 1 N i = 1 N M ( x i ( t ) ) .
Step 3:
Then, we can get the first IMF, as shown in Equation (6):
I M F 1 = x ( t ) r 1 ( t ) .
Step 4:
For k = 2 , 3 , , K , the residue and the k-th IMF can be computed by Equation (7) and Equation (8):
r k ( t ) = 1 N i = 1 N M ( r k 1 ( t ) + ε k 1 E k ( w i ( t ) ) ) ,
I M F k = r k 1 [ t ] r k [ t ] ,
In this paper, the ICEEMDAN is used to decompose the original data series into several IMFs and one residue, standing for local physical features of original signals. The difficult task of forecasting the original signals is now becoming several relatively simple sub-tasks.

2.2. Kernel Ridge Regression (KRR)

Ridge regression (RR) is a typical linear regression that uses a sum-of-squares error function and regularization technique to control the bias variance trade-off, whose purpose is to discover the linear structures hidden in the data [74]. Kernel ridge regression (KRR) is an extension of RR by introducing a kernel function that maps the input data in a low dimensional space to a high one. The kernel function k is defined on an input space X R d and is with the formula: k : X × X R . The kernel function is actually a feature map from d dimensional space into a high-dimensional Hilbert Space H k , Ψ : X H k such that k ( x i , x j ) = Ψ ( x i ) , Ψ ( x j ) H k [75,76,77]. The most popular kernel functions include:
  • Linear kernel: k L ( x i , x j ) = x i T x j .
  • Polynomial kernel: k P ( x i , x j ) = ( a ( x i T x j ) + b ) c , where a, b, and c are the coefficient, constant and degree of k P , respectively.
  • Sigmoid kernel: k S ( x i , x j ) = tanh ( d ( x i T x j ) + e ) , where d and e are the coefficient and constant, respectively.
  • Radial basis function (RBF) kernel: k R ( x i , x j ) = e f x i x j 2 , where f is related to the width of the kernel.
With the kernel functions and n data samples ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) X × Y ( y i is the target value of corresponding x i , i = 1 , 2 , , n ), we can construct the kernel matrix as Equation (9):
K = k x 1 , x 1 k x 1 , x 2 k x 1 , x n k x 2 , x 1 k x 2 , x 2 k x 2 , x n k x n , x 1 k x n , x 2 k x n , x n .
Then the KRR problem can be formulated as Equation (10):
min w Y K w 2 + λ w 2 ,
where Y is the target vector of all the n data samples, w is the unknown vector to be found, and λ 0 is a regularization item to avoid a large range of w. The solution in terms of w can be easily given in a closed-from manner as Equation (11):
w = K + λ I n 1 Y ,
where I n is an n × n identity matrix with ones on the main diagonal and zeros elsewhere.
Kernel types and parameters are two important factors for KRR. Some existing research used only a single kernel with specified parameters or simple combinations of several kernels with a fixed weight of every kernel in practical problems, limiting the forecasting performance of KRR. To improve the performance, an ideal solution is to adaptively optimize the weight and/or the parameters of each single kernel using some nature-inspired optimization algorithms. In this paper, we use differential evolution (DE) to optimize such weights and/or parameters for kernels.

2.3. Differential Evolution (DE)

Differential evolution (DE) is a member of the family of nature-inspired algorithms, and it has been demonstrated that DE is very powerful in solving various science and engineering problems [68]. The main idea of DE is to optimize a problem by using a few operations to iteratively improve a set of candidate solutions with evaluation criteria. Basically, the evolutionary process of DE consists of four stages/operations: initialization, mutation, crossover, and selection.

2.3.1. Initialization

For a D-dimensional optimization problem, given the population size P, evolutionary generation G, and the lower and upper bounds of each decision variable X m i n = [ x 1 , m i n , x 2 , m i n , , x D , m i n ] and X m a x = [ x 1 , m a x , x 2 , m a x , , x D , m a x ] respectively, the d -th decision variable of the i -th individual ( i = 1 , 2 , , P ) can be initialized as Equation (12):
I i , d 1 = x d , m i n + r a n d ( 0 , 1 ) × ( x d , m a x x d , m i n ) ,
where the “1” at the right-up corner of I represents the current evolutionary generation, and r a n d ( 0 , 1 ) generates a random real number between 0 and 1.
The initialization produces a population with P individuals/solutions/vectors, and the d -th decision variable in each individual is in the range of [ x d , m i n , x d , m a x ] .

2.3.2. Mutation

The purpose of mutation is to generate a new vector V i g , so-called mutant vector, from several existing vectors/individuals with respect to each vector I i g , so-called target vector, in the population of the g -th generation. Some popular mutation strategies are shown as follows:
V i g = I r 1 g + F × ( I r 2 g I r 3 g ) ,
V i g = I b g + F × ( I r 1 g I r 2 g ) ,
V i g = I b g + F × ( I r 1 g I r 2 g ) + F × ( I r 3 g I r 4 g ) ,
V i g = I i g + F × ( I b g I i g ) + F × ( I r 1 g I r 2 g ) ,
where r 1 r 4 are mutually random indices of the individuals, F is a preset parameter for scaling the difference vector, and I b g is the best individual at the g -th generation.

2.3.3. Crossover

The purpose of crossover is to generate a trial vector U i g = { U i , 1 g , U i , 2 g , , U i , D g } from the target vector I i g and its corresponding mutant vector V i g with the following strategy:
U i , j g = V i , j g , if r a n d ( 0 , 1 ) C r or j = j rand I i , j g , otherwise ,
where C r is a user-defined crossover rate that satisfies C r [ 0 , 1 ] , and j r a n d is a random integer in [ 1 , D ] to ensure that at least one decision variable in V i g can be passed to U i g directly.

2.3.4. Selection

The selection operation is to select the better vector from the target vector I i g and its corresponding trial vector U i g for the next generation with evaluation by fitness function f. The selection strategy is mathematically shown in Equation (18):
I i g = U i g , if f U i g f I i g I i g , otherwise .

3. The Proposed ICEEMDAN-DE-RR Approach

3.1. Ridge Regression by DE

For any types of RR, the regularization item is an important parameter to be optimized. Kernel is a very powerful trick in machine learning, which maps the data that are linearly inseparable in the input space into a higher dimensional space where the mapped data are linearly separable using kernel functions. Regarding kernel Ridge regression (KRR), besides the regularization item in Equation (10), the parameters a f for the single kernels k P , k S and k R need to be optimized. To further explore the ability of multiple kernel RR (MKRR) for crude oil price forecasting, we build a multiple kernel that consists of one k L , one k P , one k S and n k R s, as shown in Equation (19),
k M ( x i , x j ) = w 1 k L ( x i , x j ) + w 2 k P ( x i , x j ) + w 3 k S ( x i , x j ) + l = 4 n + 3 w l k R ( x i , x j ) = w 1 x i T x j + w 2 ( a ( x i T x j ) + b ) c + w 3 tanh ( d ( x i T x j ) + e ) + l = 4 n + 3 w l e f l 3 x i x j 2 ,
where w 1 , w 2 , , w n + 3 are the weights of each single kernel, and a e and f 1 f n are parameters for each single kernel.
In our approach, we use DE to optimize the regularization item λ in RR, λ and a f for each single kernel, and λ , weights and a e , f 1 f n for k M , respectively. The weights for k M , w i ( i = 1 , 2 , , n 3 ) , are generated from n + 3 real values x j ( j = 1 , 2 , , n 3 ) optimized by DE to meet i = 1 n + 3 = 1 by Equation (20):
w i = x i j = 1 n + 3 x j .

3.2. The Proposed ICEEMDAN-DE-RR Approach

The proposed ICEEMDAN-DE-RR approach is a typical form of “decomposition and ensemble”, which consists of three stages, i.e., decomposition, individual forecasting, and ensemble forecasting, as shown in Figure 1.
The details of each stage are as follows:
Stage 1:
Decomposition. The daily raw crude oil price series is decomposed into two groups of components: several IMFs and one residue.
Stage 2:
Individual forecasting. The data samples in each component are divided into training set, validation set, and test set. The training set and validation set are used to build RR models, and then the test set is applied to evaluate the models. For each model, we use DE to optimize the regularization item, corresponding kernel parameters, and possible weights.
Stage 3:
Ensemble forecasting. The individual forecasting results of all the components in Stage 2 are aggregated as the final forecasting results by addition.
The proposed ICEEMDAN-DE-RR adopts the typical “divide-and-conquer” strategy, which divides the complexly non-linear and non-stationary crude oil prices into several relatively simple components and then handles each component with a relatively simple DE-based RR predictor. With the strategy, the tough task of forecasting raw crude oil price series becomes several relatively simple sub-tasks of forecasting each component.
It is worth pointing out that although there is a lot of work on energy forecasting following the framework of “decomposition and ensemble”, our proposed work is very different from the work in several aspects: (1) RR and KRR are first applied to crude oil forecasting; (2) a novel multiple kernel RR (MKRR) optimized by DE is proposed and it can be applied in other fields; and (3) the ICEEMDAN, DE and RR are integrated to forecast daily crude oil prices for the first time.

4. Experimental Results and Comparative Analysis

4.1. Data Description

To validate the effectiveness of the proposed approach, we select the West Texas Intermediate (WTI) daily crude oil spot closing prices from 2 January 1986 to 4 February 2019 as an experimental dataset. There are 8342 samples in total, and the daily crude oil prices from 2 January 1986 to 14 June 2012, with 6673 samples accounting for 80% of the total samples, are chosen as the training set, while the rest are for testing. Within the training set, 5338 (80%) and 1335 (20%) samples are used for training and validation, respectively.
The WTI daily crude oil spot prices and corresponding decomposed components by ICEEMDAN are shown in Figure 2. We can see that the complex raw crude oil prices are decomposed into two groups: high-frequency group (IMF1-IMF5) and low-frequency group (IMF6-IMF11 and residue). The high-frequency components fluctuate within a narrow range of amplitude while the low-frequency ones fluctuate within a wide range of amplitude, making forecasting crude oil prices with the components easier than with the raw crude oil prices.
We conduct several types of multi-step-ahead forecasting with a lag of L in the experiments. A type of m-step-ahead prediction means forecasting the crude oil prices on the ( t + m ) -th day with the L price samples before the t-th day but including the t-th day.
For a fair comparison, each decomposed component is scaled to [ 0 , 1 ] using the min-max normalization, as formulated by Equation (21).
x t = x t x m i n x m a x x m i n ,
where x t and x t are crude oil price series before and after normalization respectively, and x m a x and x m i n are the maximum value and the minimum value of the time series, respectively.

4.2. Evaluation Criteria

We use a set of criteria to evaluate the proposed approach as well as the compared models. Specifically, the selected criteria include the mean absolute percent error (MAPE), the root-mean-square error (RMSE), and the directional statistic (Dstat). The MAPE, RMSE and Dstat are defined as Equations (22)–(24):
M A P E = 1 N i = 1 N | y t y t ^ y t | ,
R M S E = 1 N i = 1 N ( y t y t ^ ) 2 ,
D s t a t = 1 N i = 1 N d i × 100 % ,
where y t and y t ^ are the actual and predicted values at time t respectively, N is the size of the prediction, and d i = 1 if ( y ^ t + 1 y t ) ( y t + 1 y t ) 0 ; otherwise d i = 0 . The smaller the values of MAPE and RMSE, the better the model. In contrast, a higher Dstat means a better forecasting model.
Besides, we take the Diebold–Mariano (DM) test to compare the statistic difference of the accuracy of prediction between pairs of models. At first, we compute the difference of the prediction of pairs of models at time t as Equation (25):
d t = t = 1 N ( y t y t , a ^ ) 2 t = 1 N ( y t y t , b ^ ) 2 ,
where y t , a ^ is the prediction of model a at time t and y t , b ^ is the prediction of model b at time t. Then, the DM statistic can be defined as Equation (26):
D M = d ¯ V d / N ,
where
d ¯ = 1 N t = 1 N d t ,
V d = c o v ( d t , d t ) + 2 l = 1 c o v ( d t , d t l ) ,
where c o v is a covariance matrix.
If the value of the DM test is negative and statistically significant (e.g., p-value 0.05 ), it is proven that there is a significant difference between the predictive accuracy of pairs of models [78].

4.3. Experimental Settings

We forecast daily crude oil prices from the raw price series and decomposed components, so-called single models and ensemble models, respectively. As for single models, we compare the RR-based methods (RR, LinRR, PolyRR, SigRR, RbfRR and MKRR in Table 1) with two state-of-the-art AI models: LSSVR and BPNN, as well as two classical statistical methods: ARIMA and RW. Regarding ensemble models, we compare ICEEMDAN with EEMD to show the power of decomposition, and all the forecasting methods with single models except for ARIMA are applied to ensemble models. The parameters in the experiments are shown in detail in Table 1. Note that for the MKRR, we use 23 single kernels, i.e., one linear kernel, one polynomial kernel, one Sigmoid kernel and 20 RBF kernels, to build the multiple kernel, and both the weight and parameters of each single kernel are optimized by DE. The values or the ranges of some parameters are from existing literature [8,31]. In addition, we use RMSE as the fitness function to evaluate the individuals in DE.
All the experiments were performed by Matlab R2016b on a PC with 64-bit Windows 10, a 3.6 GHz i7 CPU and 32 GB RAM.

4.4. Results and Analysis

4.4.1. Single Models

The single models are performed on the raw daily crude oil prices directly. We compared the RR-based methods with LSSVR, BPNN and ARIMA. The experimental results are reported in Table 2, with the best and the worst results being shown in bold and underline, respectively, in terms of each criterion with each horizon.
From the table, we can find that the AI models outperform the statistical models in 6 out of 9 cases. Among the AI models, BPNN obtains two worst results: the MAPE value of 0.0161 as well as the RMSE value of 1.3050 with Horizon 1, while LSSVR and PolyRR obtain the worst values once: the Dstat value of 0.4592 and 0.4862 with Horizon 3 and 6, respectively. Overall, the RR-based single models outperform other models in most cases. In particular, RR achieves the best results in 6 out of 9 cases, showing that it is superior to other single models. Regarding the statistical models, RW is very close to but slightly better than ARIMA. In terms of MAPE and RMSE, for each model, the results become worse and worse when the horizon increases.
Regarding directional statistics, RbfRR, MKRR and BPNN achieve the highest values of 0.5186, 0.5090 and 0.4976 with Horizon 1, 3, and 6, respectively. In contrast, ARIMA and PolyRR obtain the worst values of 0.4868 and 0.4862 with Horizon 1 and 6, respectively. For Horizon 3, both LSSVR and RW obtain the worst Dstat values of 0.4952. The intervals of the best values and the corresponding worst values are so narrow that all the results of all cases are around 0.5, just like the result of random guessing, showing that it is a tough task to forecast the direction using single models.
To further compare the single models, we report the DM test results in Table 3, with the statistics and the corresponding p-values (in brackets). From this table, we have some findings. First, compared with the statistical model ARIMA and RW, the statistics of all the AI models in all cases except for BPNN with Horizon 1 are far below −2.000 and the corresponding p-values are much less than 0.01, indicating that the AI models significantly outperform ARIMA and RW. ARIMA and RW have very similar results. Second, LSSVM and BPNN underperform RR-based approaches in most cases, showing that the RR-based predictors are more effective than the statistical AI models (LSSVM and BPNN) for forecasting raw crude oil prices, to some extent. Third, as for RR-based models, the RR, PolyRR and MKRR have very close performance, which are superior to LinRR, SigRR and RbfRR. All the findings confirm the analysis on MAPE, RMSE and Dstat.
Due to the nonlinearity and nonstationarity, the performance of directly forecasting on raw crude oil price series needs to be improved. To cope with this issue, we use ICEEMDAN to decompose the raw series into several components each of which shows relatively simple characteristics when compared with the raw series, and then each component is predicted using an AI model individually. At last, the predicted results of all components are aggregated as the final result.

4.4.2. Ensemble Models

To demonstrate the effectiveness of ICEEMDAN, we also use EEMD as one of the decomposition methods for comparison. For the forecasting methods, we compare RR-base predictors with two state-of-the-art AI methods: LSSVR and BPNN. The values of MAPE, RMSE and Dstat with ensemble models are shown in Table 4.
As far as MAPE and RMSE with EEMD are concerned, the results of different forecasting models except for BPNN and RW are significantly superior to those of the corresponding single models. For example, the best (lowest) MAPE, and RMSE with Horizon 1 are improved from 0.0154 to 0.0084, and from 1.2454 to 0.6399, respectively. Regarding Dstat, the best/worst value except those by RW is 0.8231/0.7344, which is far greater than the values of single models. Among the forecasting methods, RR-based predictors achieve all the best values while BPNN and RW obtain all the worst results. Specifically, RR, SigRR and PolyRR achieve the best values 4, 4 and 3 out of 9 times, respectively. Except for the values of MAPE and RMSE with Horizon 1, the ensemble models by BPNN are advantageous over single BPNN. Another finding is that RW obtains the worst values 8 out of 9 times. In particular, the Dstat values by RW are always around 0.5, and the possible reason is that RW performs poorly in forecasting high-frequency components. The experimental results indicate that the ensemble models except for RW can significantly improve the forecasting effectiveness when compared with the single models.
When we look at the results with ICEEMDAN and AI models in Table 4, we can see a significant improvement in the forecasting ability. As for MAPE, the value of each model with ICEEMDAN is superior to that of its competitor with EEMD. Specifically, RR achieves the best (lowest) MAPE for all the horizons, while BPNN obtains the worst values for the same horizons. For Horizon 1, PolyRR, SigRR and MKRR also achieve the best MAPE (0.0043) as RR does. The results of RMSE show similar characteristics that all the models with ICEEMDAN exceed those with EEMD. The best values of RMSE with Horizon 1, 3, and 6 are achieved by SigRR, RR, and RR, respectively. In contrast, LinRR, LinRR and BPNN obtain the worst RMSEs with Horizon 1, 3 and 6, respectively. It is worth pointing out that the directional statistics is significantly improved by ICEEMDAN. The best Dstat (0.9113) is achieved by SigRR with Horizon 1, and the other models achieve very close Dstat, indicating that this metric is very stable with ICEEMDAN. The worst (lowest) value of Dstat is 0.6661, which is much higher than the best one by single models (0.5186). Therefore, the models with ICEEMDAN and AI models are able to effectively improve the results of directional statistics. Regarding the results with ICEEMDAN and RW models, they still perform poorly and obtain the nine worst values, although RW models with ICEEMDAN perform slightly better than those with EEMD for Horizon 1 and 2.
For the models with both EEMD and ICEEMDAN, most results of MAPE, RMSE and Dstat will become worse when the horizon increases, showing that it is more difficult to forecast crude oil prices with a long horizon than with a short one.
We still apply the DM test to compare the ensemble models, and the results are shown in Table 5. From this table, we can see that when the forecasting methods with ICEEMDAN are compared with those with EEMD, the statistical values are far below zero and the p-value is very close to zero (usually less than 0.0001) with Horizon 1 and 3, indicating that the former significantly outperforms the latter with these two horizons. Regarding Horizon 6, the forecasting methods, except for BPNN with ICEEMDAN, still outperform the corresponding methods with EEMD. For each decomposition, the RR-based methods are usually superior to LSSVR and BPNN. Among the RR-based predictors, RR and SigRR have the best forecasting ability, followed by MKRR and PolyRR, while RbfRR and LinRR have a slightly worse predictive power. In addition, the models with AI are all superior to the corresponding models with RW, showing that the forecasting effectiveness does not stem from luck but by the forecasting superiority of the proposed approaches. All the DM test results confirm that ICEEMDAN and RR-based predictors are very effective for forecasting daily crude oil prices. The proposed approach that integrates ICEEMDAN and RR is capable of improving the results of crude oil price forecasting.

4.5. Discussion

In this subsection, we will discuss the impact of the parameter settings of the ICEEMDAN, the impact of the lag orders and the result of each individual component. Since the above results and analysis have shown that RR and SigRR usually outperform the other models, we will take both RR and SigRR as examples to discuss the following.

4.5.1. The Impact of the Parameter Settings of the ICEEMDAN

When we use the ICEEMDAN to decompose the daily crude oil price series, we need to add noises to the series and decompose the series many times. Therefore, the noise standard deviation n s d and the number of realizations n r are two important parameters. To study the impact of n s d on forecasting, we run the proposed approach with a variable n s d in the range of {0.01, 0.03, 0.05, 0.08, 0.1, 0.15, 0.2, 0.3, 0.4} while fixing other parameters. The experimental results are shown in Figure 3. Likewise, we use a variable n r in the range of {20, 50, 100, 200, 300, 500, 800, 1000, 1500, 2000} and fixed other parameters to study the impact of the number of realizations, as shown in Figure 4.
It can be seen from Figure 3 that the results in terms of RMSE, MAPE and Dstat are gradually improved when n s d increases from 0.01 to 0.08. In contrast, after that, all the evaluation indicators are getting worse and worse with the increase of noise strength when n s d is greater than 0.1. Both RR and SigRR have similar trends, and one of the two models is alternatively better than the other. The results show that the white noise strength has great impact on the forecasting performance and an ideal white noise strength is between 0.05 and 0.1.
When we look at Figure 4, we can find that when the number of the realization is 20, the results of RMSE, MAPE and Dstat are all rather bad. When the number of realization increases from 20 to 500, all the results become better and better. Specifically, the Dstat reaches the best values for both RR and SigRR when the number of realization equals 500, while the results of RMSE and MAPE are very close to the best values. After that, the values of the three indicators are very stable when the number of realization varies from 500 to 2000. The results indicate that 500 is very ideal for the number of realizations.

4.5.2. The Impact of the Lag Orders

Lag orders refer to the length of recent data points treated as explanatory variables to build time series models. We further investigate the impact of variable lag orders from 1 to 20 with horizon 1, and the results are shown in Figure 5. When the lag order is equal to 1, the results of the evaluation indicators are the worst. However, when it varies from 1 to 6, the corresponding results are all becoming better and better. After that, the results have remained almost unchanged for the lag order from 6 to 20. Therefore, the best lag order is 6 because it can provide satisfactory results with less input, which confirms the previous study [12,31].

4.5.3. The Result of Each Individual Component

Each decomposed component by the ICEEMDAN shows either high-frequency or low-frequency characteristics. In general, it is harder to forecasting a high-frequency component than a low-frequency one. We plot the predicted values and raw data of each component and the raw crude oil prices by RR and sigRR in Figure 6 and Figure 7, respectively.
It can be seen from these figures that both RR and SigRR are able to forecast the high-frequency components (IMF6-IMF11 as well as the residue) very well, and the predicted errors mainly come from the high-frequency components (IMF1-IMF5), especially from IMF1. Since the volatility of the hight-frequency components is relatively narrow, the forecasting errors from such components might be restricted. This is one of the possible reasons why the framework of “decomposition and ensemble” is effective for time series forecasting.

5. Summary and Conclusions

Forecasting daily crude oil prices is an important but challenging task. To improve the forecasting performance, a series of approaches using ICEEMDAN, DE and RR, termed as ICEEMDAN-DE-RR, are proposed in this paper. The proposed approaches firstly use ICEEMDAN to decompose the complex original crude oil prices into several components, and then each component is forecasted individually by DE-based RR predictors. In the end, the sum of the predicted results of all the components is taken as the final result. The extensive experiments demonstrated the proposed approaches can outperform some state-of-the-art methods.
Especially from the experimental results, we have the following interesting findings: (1) It is a very difficult task to accurately forecast daily crude oil prices with the raw price series because of its nonlinearity and nonstationarity; (2) AI-models usually outperform statistical methods when forecasting crude oil prices; (3) RR-based predictors with DE optimizing the parameters have good forecasting ability; (4) The framework of “decomposition and ensemble” can significantly improve the performance of forecasting daily crude oil prices; ICEEMDAN is advantageous over EEMD for the forecasting tasks; (5) The proposed ICEEMDAN-DE-RR approach outperforms the competitive methods in terms of several evaluation metrics, indicating that it is promising for daily crude oil price forecasting. (6) Regarding RR-based predictors, RR and SigRR with DE optimizing parameters can achieve very promising forecasting results in terms of several criteria.
In the future, we will apply the proposed approaches to forecasting other types of energy time series, such as natural gas prices, wind speed, wind power and electricity load.

Author Contributions

Formal analysis, T.L. and X.L.; Funding acquisition, X.L.; Investigation, J.W. and T.H.; Methodology, T.L.; Resources, T.H.; Software, T.L. and Y.Z.; Supervision, T.L. and X.L.; Writing—original draft, T.L. and Y.Z.; Writing—review & editing, T.L. and Y.Z.

Funding

This work was supported by the Fundamental Research Funds for the Central Universities (Grant No. JBK1902029), the Ministry of Education of Humanities and Social Science Project (Grant No. 19YJAZH047), Sichuan Science and Technology Program (Grant No. 2019YFG0117) and the Scientific Research Fund of Sichuan Provincial Education Department (Grant No. 17ZB0433).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Miao, H.; Ramchander, S.; Wang, T.; Yang, D. Influential factors in crude oil price forecasting. Energy Econ. 2017, 68, 77–88. [Google Scholar] [CrossRef]
  2. Naderi, M.; Khamehchi, E.; Karimi, B. Novel statistical forecasting models for crude oil price, gas price, and interest rate based on meta-heuristic bat algorithm. J. Pet. Sci. Eng. 2019, 172, 13–22. [Google Scholar] [CrossRef]
  3. Ye, M.; Zyren, J.; Shore, J. A monthly crude oil spot price forecasting model using relative inventories. Int. J. Forecast. 2005, 21, 491–501. [Google Scholar] [CrossRef]
  4. Movagharnejad, K.; Mehdizadeh, B.; Banihashemi, M.; Kordkheili, M.S. Forecasting the differences between various commercial oil prices in the Persian Gulf region by neural network. Energy 2011, 36, 3979–3984. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Ma, F.; Shi, B.; Huang, D. Forecasting the prices of crude oil: An iterated combination approach. Energy Econ. 2018, 70, 472–483. [Google Scholar] [CrossRef]
  6. Wen, F.; Gong, X.; Cai, S. Forecasting the volatility of crude oil futures using HAR-type models with structural breaks. Energy Econ. 2016, 59, 400–413. [Google Scholar] [CrossRef]
  7. Li, J.; Zhu, S.; Wu, Q. Monthly crude oil spot price forecasting using variational mode decomposition. Energy Econ. 2019, 83, 240–253. [Google Scholar] [CrossRef]
  8. Zhou, Y.; Li, T.; Shi, J.; Qian, Z. A CEEMDAN and XGBOOST-Based Approach to Forecast Crude Oil Prices. Complexity 2019, 2019, 4392785. [Google Scholar] [CrossRef]
  9. He, K.; Yu, L.; Lai, K.K. Crude oil price analysis and forecasting using wavelet decomposed ensemble model. Energy 2012, 46, 564–574. [Google Scholar] [CrossRef]
  10. Yu, L.; Dai, W.; Tang, L.; Wu, J. A hybrid grid-GA-based LSSVR learning paradigm for crude oil price forecasting. Neural Comput. Appl. 2016, 27, 2193–2215. [Google Scholar] [CrossRef]
  11. Tang, L.; Wu, Y.; Yu, L. A non-iterative decomposition-ensemble learning paradigm using RVFL network for crude oil price forecasting. Appl. Soft. Comput. 2018, 70, 1097–1108. [Google Scholar] [CrossRef]
  12. Yu, L.; Wang, S.; Lai, K.K. Forecasting crude oil price with an EMD-based neural network ensemble learning paradigm. Energy Econ. 2008, 30, 2623–2635. [Google Scholar] [CrossRef]
  13. Mirmirani, S.; Li, H.C. A comparison of VAR and neural networks with genetic algorithm in forecasting price of oil. Appl. Artif. Intell. Financ. Econ. 2004, 19, 203–223. [Google Scholar]
  14. Murat, A.; Tokat, E. Forecasting oil price movements with crack spread futures. Energy Econ. 2009, 31, 85–90. [Google Scholar] [CrossRef]
  15. Moshiri, S.; Foroutan, F. Forecasting nonlinear crude oil futures prices. Energy J. 2006, 27, 81–95. [Google Scholar] [CrossRef]
  16. Herrera, A.M.; Hu, L.; Pastor, D. Forecasting crude oil price volatility. Int. J. Forecast. 2018, 34, 622–635. [Google Scholar] [CrossRef]
  17. Nademi, A.; Nademi, Y. Forecasting crude oil prices by a semiparametric Markov switching model: OPEC, WTI, and Brent cases. Energy Econ. 2018, 74, 757–766. [Google Scholar] [CrossRef]
  18. Zhang, Y.J.; Yao, T.; He, L.Y.; Ripple, R. Volatility forecasting of crude oil market: Can the regime switching GARCH model beat the single-regime GARCH models? Int. Rev. Econ. Financ. 2019, 59, 302–317. [Google Scholar] [CrossRef]
  19. Lyocsa, S.; Molnar, P. Exploiting dependence: Day-ahead volatility forecasting for crude oil and natural gas exchange-traded funds. Energy 2018, 155, 462–473. [Google Scholar] [CrossRef]
  20. Lv, W. Does the OVX matter for volatility forecasting? Evidence from the crude oil market. Phys. A 2018, 492, 916–922. [Google Scholar] [CrossRef]
  21. Naser, H. Estimating and forecasting the real prices of crude oil: A data rich model using a dynamic model averaging (DMA) approach. Energy Econ. 2016, 56, 75–87. [Google Scholar] [CrossRef]
  22. Azevedo, V.G.; Campos, L.M.S. Combination of forecasts for the price of crude oil on the spot market. Int. J. Prod. Res. 2016, 54, 5219–5235. [Google Scholar] [CrossRef]
  23. Tang, L.; Dai, W.; Yu, L. A Novel CEEMD-Based EELM Ensemble Learning Paradigm for Crude Oil Price Forecasting. Int. J. Inf. Technol. Decis. Mak. 2015, 14, 141–169. [Google Scholar] [CrossRef]
  24. Tehrani, R.; Khodayar, F. A hybrid optimized artificial intelligent model to forecast crude oil using genetic algorithm. Afr. J. Bus. Manag. 2011, 5, 13130–13135. [Google Scholar] [CrossRef]
  25. Xiong, T.; Bao, Y.; Hu, Z. Beyond one-step-ahead forecasting: Evaluation of alternative multi-step-ahead forecasting models for crude oil prices. Energy Econ. 2013, 40, 405–415. [Google Scholar] [CrossRef] [Green Version]
  26. Barunik, J.; Malinska, B. Forecasting the term structure of crude oil futures prices with neural networks. Appl. Energy 2016, 164, 366–379. [Google Scholar] [CrossRef] [Green Version]
  27. Ding, Y. A novel decompose-ensemble methodology with AIC-ANN approach for crude oil forecasting. Energy 2018, 154, 328–336. [Google Scholar] [CrossRef]
  28. Fan, L.; Pan, S.; Li, Z.; Li, H. An ICA-based support vector regression scheme for forecasting crude oil prices. Technol. Forecast. Soc. Chang. 2016, 112, 245–253. [Google Scholar] [CrossRef]
  29. Yu, L.; Zhang, X.; Wang, S. Assessing Potentiality of Support Vector Machine Method in Crude Oil Price Forecasting. Eurasia J. Math. Sci. Technol. Educ. 2017, 13, 7893–7904. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Zhang, J. Volatility forecasting of crude oil market: A new hybrid method. J. Forecast. 2018, 37, 781–789. [Google Scholar] [CrossRef]
  31. Li, T.; Hu, Z.; Jia, Y.; Wu, J.; Zhou, Y. Forecasting Crude Oil Prices Using Ensemble Empirical Mode Decomposition and Sparse Bayesian Learning. Energies 2018, 11, 1882. [Google Scholar] [CrossRef]
  32. Wu, J.; Chen, Y.; Zhou, T.; Li, T. An Adaptive Hybrid Learning Paradigm Integrating CEEMD, ARIMA and SBL for Crude Oil Price Forecasting. Energies 2019, 12, 1239. [Google Scholar] [CrossRef]
  33. Yu, L.; Dai, W.; Tang, L. A novel decomposition ensemble model with extended extreme learning machine for crude oil price forecasting. Eng. Appl. Artif. Intell. 2016, 47, 110–121. [Google Scholar] [CrossRef]
  34. Wang, J.; Athanasopoulos, G.; Hyndman, R.J.; Wang, S. Crude oil price forecasting based on internet concern using an extreme learning machine. Int. J. Forecast. 2018, 34, 665–677. [Google Scholar] [CrossRef]
  35. Wu, Y.X.; Wu, Q.B.; Zhu, J.Q. Improved EEMD-based crude oil price forecasting using LSTM networks. Phys. A 2019, 516, 114–124. [Google Scholar] [CrossRef]
  36. Li, T.; Zhou, M.; Guo, C.; Luo, M.; Wu, J.; Pan, F.; Tao, Q.; He, T. Forecasting Crude Oil Price Using EEMD and RVM with Adaptive PSO-Based Kernels. Energies 2016, 9, 1014. [Google Scholar] [CrossRef]
  37. Chiroma, H.; Abdul-kareem, S.; Noor, A.S.M.; Abubakar, A.I.; Safa, N.S.; Shuib, L.; Hamza, M.F.; Gital, A.Y.; Herawan, T. A Review on Artificial Intelligence Methodologies for the Forecasting of Crude Oil Price. Intell. Autom. Soft Comput. 2016, 22, 449–462. [Google Scholar] [CrossRef]
  38. Deng, W.; Zhang, S.; Zhao, H.; Yang, X. A Novel Fault Diagnosis Method Based on Integrating Empirical Wavelet Transform and Fuzzy Entropy for Motor Bearing. IEEE Access 2018, 6, 35042–35056. [Google Scholar] [CrossRef]
  39. Zhao, H.; Yao, R.; Xu, L.; Yuan, Y.; Li, G.; Deng, W. Study on a Novel Fault Damage Degree Identification Method Using High-Order Differential Mathematical Morphology Gradient Spectrum Entropy. Entropy 2018, 20, 682. [Google Scholar] [CrossRef]
  40. Bajaj, V.; Pachori, R.B. Classification of Seizure and Nonseizure EEG Signals Using Empirical Mode Decomposition. IEEE Trans. Inf. Technol. Biomed. 2012, 16, 1135–1142. [Google Scholar] [CrossRef] [PubMed]
  41. Li, T.; Zhou, M. ECG Classification Using Wavelet Packet Entropy and Random Forests. Entropy 2016, 18, 285. [Google Scholar] [CrossRef]
  42. Zhang, H.; Wang, X.; Cao, J.; Tang, M.; Guo, Y. A multivariate short-term traffic flow forecasting method based on wavelet analysis and seasonal time series. Appl. Intell. 2018, 48, 3827–3838. [Google Scholar] [CrossRef]
  43. Pannakkong, W.; Sriboonchitta, S.; Huynh, V.N. An Ensemble Model of Arima and Ann with Restricted Boltzmann Machine Based on Decomposition of Discrete Wavelet Transform for Time Series Forecasting. J. Syst. Sci. Syst. Eng. 2018, 27, 690–708. [Google Scholar] [CrossRef]
  44. Li, T.; Yang, M.; Wu, J.; Jing, X. A novel image encryption algorithm based on a fractional-order hyperchaotic system and DNA computing. Complexity 2017, 2017, 9010251. [Google Scholar] [CrossRef]
  45. Li, T.; Shi, J.; Li, X.; Wu, J.; Pan, F. Image encryption based on pixel-level diffusion with dynamic filtering and DNA-level permutation with 3D Latin cubes. Entropy 2019, 21, 319. [Google Scholar] [CrossRef]
  46. Li, X.; Xie, Z.; Wu, J.; Li, T. Image encryption based on dynamic filtering and bit cuboid operations. Complexity 2019, 2019, 7485621. [Google Scholar] [CrossRef]
  47. Ren, Y.; Suganthan, P.N.; Srikanth, N. A Novel Empirical Mode Decomposition with Support Vector Regression for Wind Speed Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1793–1798. [Google Scholar] [CrossRef] [PubMed]
  48. Yang, Z.S.; Wang, J. A combination forecasting approach applied in multistep wind speed forecasting based on a data processing strategy and an optimized artificial intelligence algorithm. Appl. Energy 2018, 230, 1108–1125. [Google Scholar] [CrossRef]
  49. Abdoos, A.; Hemmati, M.; Abdoos, A.A. Short term load forecasting using a hybrid intelligent method. Knowl.-Based Syst. 2015, 76, 139–147. [Google Scholar] [CrossRef]
  50. Fan, G.F.; Peng, L.L.; Hong, W.C.; Sun, F. Electric load forecasting by the SVR model with differential empirical mode decomposition and auto regression. Neurocomputing 2016, 173, 958–970. [Google Scholar] [CrossRef]
  51. Qiu, X.H.; Ren, Y.; Suganthan, P.N.; Amaratunga, G.A.J. Empirical Mode Decomposition based ensemble deep learning for load demand time series forecasting. Appl. Soft. Comput. 2017, 54, 246–255. [Google Scholar] [CrossRef]
  52. Sun, W.; Zhang, C.C. Analysis and forecasting of the carbon price using multi resolution singular value decomposition and extreme learning machine optimized by adaptive whale optimization algorithm. Appl. Energy 2018, 231, 1354–1371. [Google Scholar] [CrossRef]
  53. Wang, D.Y.; Luo, H.Y.; Grunder, O.; Lin, Y.B.; Guo, H.X. Multi-step ahead electricity price forecasting using a hybrid model based on two-layer decomposition technique and BP neural network optimized by firefly algorithm. Appl. Energy 2017, 190, 390–407. [Google Scholar] [CrossRef]
  54. Yang, Z.S.; Wang, J. A hybrid forecasting approach applied in wind speed forecasting based on a data processing strategy and an optimized artificial intelligence algorithm. Energy 2018, 160, 87–100. [Google Scholar] [CrossRef]
  55. Huang, N.E.; Shen, Z.; Long, S.R. A new view of nonlinear water waves: the Hilbert Spectrum 1. Annu. Rev. Fluid Mech. 1999, 31, 417–457. [Google Scholar] [CrossRef]
  56. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal 2009, 1, 1–41. [Google Scholar] [CrossRef]
  57. Torres, M.E.; Colominas, M.A.; Schlotthauer, G.; Flandrin, P. A complete ensemble empirical mode decomposition with adaptive noise. In Proceedings of the 2011 IEEE international conference on acoustics, speech and signal processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 4144–4147. [Google Scholar]
  58. Colominas, M.A.; Schlotthauer, G.; Torres, M.E. Improved complete ensemble EMD: A suitable tool for biomedical signal processing. Biomed. Signal Process. Control 2014, 14, 19–29. [Google Scholar] [CrossRef]
  59. Dai, S.; Niu, D.; Li, Y. Daily peak load forecasting based on complete ensemble empirical mode decomposition with adaptive noise and support vector machine optimized by modified grey Wolf optimization algorithm. Energies 2018, 11, 163. [Google Scholar] [CrossRef]
  60. Douak, F.; Melgani, F.; Benoudjit, N. Kernel ridge regression with active learning for wind speed prediction. Appl. Energy 2013, 103, 328–340. [Google Scholar] [CrossRef]
  61. Naik, J.; Satapathy, P.; Dash, P.K. Short-term wind speed and wind power prediction using hybrid empirical mode decomposition and kernel ridge regression. Appl. Soft. Comput. 2018, 70, 1167–1188. [Google Scholar] [CrossRef]
  62. Qian, C.; Breckon, T.P.; Li, H. Robust visual tracking via speedup multiple kernel ridge regression. J. Electron. Imaging 2015, 24, 053016. [Google Scholar] [CrossRef] [Green Version]
  63. Maalouf, M.; Homouz, D.; Abutayeh, M. Accurate Prediction of Preheat Temperature in Solar Flash Desalination Systems Using Kernel Ridge Regression. J. Energy Eng. 2016, 142, E4015017. [Google Scholar] [CrossRef]
  64. Naik, J.; Bisoi, R.; Dash, P.K. Prediction interval forecasting of wind speed and wind power using modes decomposition based low rank multi-kernel ridge regression. Renew. Energy 2018, 129, 357–383. [Google Scholar] [CrossRef]
  65. Kennedy, J. Particle swarm optimization. In Encyclopedia of Machine Learning; Springer: Berlin, Germany, 2010; pp. 760–766. [Google Scholar]
  66. Deng, W.; Zhao, H.; Yang, X.; Xiong, J.; Sun, M.; Li, B. Study on an improved adaptive PSO algorithm for solving multi-objective gate assignment. Appl. Soft. Comput. 2017, 59, 288–302. [Google Scholar] [CrossRef]
  67. Deng, W.; Yao, R.; Zhao, H.; Yang, X.; Li, G. A novel intelligent diagnosis method using optimal LS-SVM with improved PSO algorithm. Soft Comput. 2019, 23, 2445–2462. [Google Scholar] [CrossRef]
  68. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  69. Das, S.; Suganthan, P.N. Differential Evolution: A Survey of the State-of-the-Art. IEEE Trans. Evol. Comput. 2011, 15, 4–31. [Google Scholar] [CrossRef]
  70. Dorigo, M.; Blum, C. Ant colony optimization theory: A survey. Theor. Comput. Sci. 2005, 344, 243–278. [Google Scholar] [CrossRef]
  71. Deng, W.; Xu, J.; Zhao, H. An Improved Ant Colony Optimization Algorithm Based on Hybrid Strategies for Scheduling problem. IEEE Access 2019, 20281–20292. [Google Scholar] [CrossRef]
  72. Yu, X.; Liong, S.Y. Forecasting of hydrologic time series with ridge regression in feature space. J. Hydrol. 2007, 332, 290–302. [Google Scholar] [CrossRef]
  73. Ahn, J.J.; Byun, H.W.; Oh, K.J.; Kim, T.Y. Using ridge regression with genetic algorithm to enhance real estate appraisal forecasting. Expert Syst. Appl. 2012, 39, 8369–8379. [Google Scholar] [CrossRef]
  74. Zhang, S.; Hu, Q.; Xie, Z.; Mi, J. Kernel ridge regression for general noise model with its application. Neurocomputing 2015, 149, 836–846. [Google Scholar] [CrossRef]
  75. Saunders, C.; Gammerman, A.; Vovk, V. Ridge Regression Learning Algorithm in Dual Variables. In Proceedings of the Fifteenth International Conference on Machine Learning, Madison, WI, USA, 24–27 July 1998; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1998; pp. 515–521. [Google Scholar]
  76. Maalouf, M.; Barsoum, Z. Failure strength prediction of aluminum spot-welded joints using kernel ridge regression. Int. J. Adv. Manuf. Technol. 2017, 91, 3717–3725. [Google Scholar] [CrossRef]
  77. Avron, H.; Clarkson, K.L.; Woodruff, D.P. Faster kernel ridge regression using sketching and preconditioning. SIAM J. Matrix Anal. Appl. 2017, 38, 1116–1138. [Google Scholar] [CrossRef]
  78. Diebold, F.; Mariano, R. Comparing predictive accuracy. J. Bus. Econ. Stat. 2002, 20, 134–144. [Google Scholar] [CrossRef]
  79. Sakamoto, Y.; Ishiguro, M.; Kitagawa, G. Akaike Information Criterion Statistics; Reidel, D., Ed.; Springer: Dordrecht, The Netherlands, 1986; Volume 81. [Google Scholar]
Figure 1. The flowchart of the proposed ICEEMDAN-DE-KRR.
Figure 1. The flowchart of the proposed ICEEMDAN-DE-KRR.
Energies 12 03603 g001
Figure 2. The daily WTI crude oil prices and the corresponding decomposed components by ICEEMDAN.
Figure 2. The daily WTI crude oil prices and the corresponding decomposed components by ICEEMDAN.
Energies 12 03603 g002
Figure 3. The impact of the white noise strength in the ICEEMDAN by RR and SigRR.
Figure 3. The impact of the white noise strength in the ICEEMDAN by RR and SigRR.
Energies 12 03603 g003
Figure 4. The impact of the number of the realizations in the ICEEMDAN by RR and SigRR.
Figure 4. The impact of the number of the realizations in the ICEEMDAN by RR and SigRR.
Energies 12 03603 g004
Figure 5. The impact of the lag orders by RR and SigRR.
Figure 5. The impact of the lag orders by RR and SigRR.
Energies 12 03603 g005
Figure 6. The individual and final forecasting results by RR.
Figure 6. The individual and final forecasting results by RR.
Energies 12 03603 g006
Figure 7. The individual and final forecasting results by SigRR.
Figure 7. The individual and final forecasting results by SigRR.
Energies 12 03603 g007
Table 1. The settings for the parameters.
Table 1. The settings for the parameters.
MethodDescriptionParameters
EEMDEnsemble empirical mode decompositionNoise standard deviation: 0.2;
Number of realizations: 100.
ICEEMDANImproved complete EEMD with adaptive noiseNoise standard deviation: 0.05;
Number of realizations: 500;
Maximum number of sifting iterations allowed: 5000.
RRRidge Regression λ : [0.001, 0.2].
LinRRRR with a linear kernel λ : [0.001, 0.2].
PolyRRRR with a polynomial kernel λ : [0.001, 0.2]; a: [0, 2]; b: [0, 10]; c: {1,2,3,4}.
SigRRRR with a Sigmoid kernel λ : [0.001, 0.2]; d: [0, 4]; e: [0, 8].
RbfRRRR with a radial basis functional kernel λ : [0.001, 0.2]; f: [ 2 10 , 2 12 ] .
MKRRRR with multiple kernels as formulated in Equation (19) λ : [0.001, 0.2];
a e : the same as the above single kernel;
n: 20, number of the RBF kernels;
f 1 f 20 : [ 2 10 , 2 12 ] ; w 1 w 23 : [ 0 , 1 ] s . t . i = 1 23 w i = 1 .
LSSVRLeast square support vector regression with a RBF kernelRegularization parameter: 2 { 10 , 9 , , 11 , 12 } ;
Width of the RBF kernel: 2 { 10 , 9 , , 11 , 12 } .
BPNNBack propagation neural networkSize of the hidden layer: {10, 20, 50, 100};
Maximum training epochs: {100, 1000, 10000};
Learning rate: {0.001, 0.01, 0.05, 0.1}.
ARIMAAutoregressive integrated moving averageAkaike information criterion (AIC) to determine parameters (p-d-q) [79].
DEDifferential EvolutionPopulation size: 20; Number of iterations: 40;
Crossover probability: 0.2.
Table 2. Results of single models.
Table 2. Results of single models.
HorizonCriterionRRLinRRPolyRRSigRRRbfRRMKRRLSSVRBPNNARIMARW
1MAPE0.01540.01540.01540.01540.01560.01540.01540.01610.01570.0156
RMSE1.24541.24621.24731.24831.25671.24721.24811.30501.27011.2700
Dstat0.50000.49400.51320.51560.51860.51620.51020.51320.48680.5054
3MAPE0.02620.02630.02640.02640.02650.02630.02660.02640.02740.0272
RMSE2.06272.06892.07082.07012.07542.07672.08012.07972.17132.1645
Dstat0.49880.49880.49640.49580.50000.50900.49520.49940.49820.4952
6MAPE0.03770.03790.03810.03810.03810.03800.03790.03940.04080.0401
RMSE2.89772.91012.92092.92082.92392.91492.91282.99433.18243.1195
Dstat0.49520.49580.48620.48980.49220.49640.49100.49760.49160.4928
Table 3. The Diebold-Mariano (DM) test results of single models.
Table 3. The Diebold-Mariano (DM) test results of single models.
HorizonTested ModelLinRRPolyRRSigRRRbfRRMKRRLSSVRBPNNARIMARW
1RR−1.1025
(0.2704)
−0.7203
(0.4714)
−1.5290
(0.1265)
−2.0533
(0.0402)
−0.6405
(0.5219)
−1.3788
(0.1681)
−5.5601
(0.0000)
−3.4678
(0.0005)
−3.4585
(0.0004)
LinRR −0.4701
(0.6383)
−1.3435
(0.1793)
−1.6807
(0.0930)
−0.3852
(0.7001)
−1.3221
(0.1863)
−5.6817
(0.0000)
−3.3577
(0.0008)
−3.2647
(0.0007)
PolyRR −0.4864
(0.6268)
−1.3044
(0.1923)
0.2103
(0.8334)
−0.4205
(0.6742)
−5.7019
(0.0000)
−2.8226
(0.0048)
−2.7642
(0.0000)
SigRR −1.2372
(0.2162)
0.5057
(0.6132)
0.1453
(0.8845)
−5.7208
(0.0000)
−2.8937
(0.0039)
−2.3072
(0.0002)
RbfRR 1.3294
(0.1839)
1.2083
(0.2271)
−3.4277
(0.0006)
−1.3067
(0.1915)
−1.2796
(0.0142)
MKRR −0.4335
(0.6647)
−5.7099
(0.0000)
−2.8012
(0.0052)
−2.7326
(0.0312)
LSSVR −5.8110
(0.0000)
−2.8584
(0.0043)
−2.6057
(0.0147)
BPNN 2.6579
(0.0079)
2.1439
(0.0001)
ARIMA 0.0996
(0.1206)
3RR−2.4997
(0.0125)
−1.6877
(0.0916)
−2.0321
(0.0423)
−2.8803
(0.0040)
−2.3015
(0.0215)
−3.2386
(0.0012)
−2.7039
(0.0069)
−5.6427
(0.0000)
−4.9664
(0.0000)
LinRR −0.2930
(0.7696)
−0.2368
(0.8129)
−2.0401
(0.0415)
−1.0401
(0.2984)
−2.9172
(0.0036)
−1.3601
(0.1740)
−5.1351
(0.0000)
−5.1652
(0.0000)
PolyRR 0.1966
(0.8442)
−0.6415
(0.5213)
−1.6590
(0.0973)
−1.1121
(0.2662)
−2.1367
(0.0328)
−4.6340
(0.0000)
−4.9189
(0.0000)
SigRR −1.0601
(0.2893)
−1.6399
(0.1012)
−1.6464
(0.0999)
−1.8857
(0.0595)
−4.8619
(0.0000)
−4.6375
(0.0000)
RbfRR −0.1690
(0.8658)
−3.6421
(0.0003)
−0.5004
(0.6169)
−4.4865
(0.0000)
−4.4237
(0.0000)
MKRR −0.4004
(0.6889)
−0.5933
(0.5530)
−4.3286
(0.0000)
−3.3925
(0.0000)
LSSVR 0.0403
(0.9679)
−4.2520
(0.0000)
−3.8976
(0.0000)
BPNN −4.4861
(0.0000)
−4.3547
(0.0001)
ARIMA 0.7708
(0.1300)
6RR−2.6901
(0.0072)
−3.0491
(0.0023)
−3.1495
(0.0017)
−3.2282
(0.0013)
−1.4575
(0.1452)
−2.4926
(0.0128)
−5.1039
(0.0000)
−7.6069
(0.0000)
−7.9403
(0.0000)
LinRR −1.3345
(0.1822)
−1.2552
(0.2096)
−2.1339
(0.0330)
−0.3964
(0.6919)
−0.3807
(0.7035)
−4.2189
(0.0000)
−7.0961
(0.0000)
−6.8125
(0.0000)
PolyRR 0.0718
(0.9428)
−0.6199
(0.5354)
0.6465
(0.5180)
2.4994
(0.0125)
−4.9939
(0.0000)
−6.4072
(0.0000)
−6.0013
(0.0000)
SigRR −0.5182
(0.6044)
0.6242
(0.5326)
2.2295
(0.0259)
−5.2852
(0.0000)
−6.3341
(0.0000)
−6.1752
(0.0000)
RbfRR 0.8653
(0.3870)
2.6871
(0.0073)
−4.0728
(0.0000)
−6.3042
(0.0000)
−6.4841
(0.0000)
MKRR 0.2139
(0.8307)
−4.9976
(0.0000)
−6.4398
(0.0000)
−5.8925
(0.0000)
LSSVR −5.0705
(0.0000)
−6.6890
(0.0000)
−6.5482
(0.0001)
BPNN −4.0592
(0.0001)
−3.7692
(0.0002)
ARIMA 0.7134
(0.2304)
Table 4. Results of ensemble models.
Table 4. Results of ensemble models.
DecompositionHorizonCriterionRRLinRRPolyRRSigRRRbfRRMKRRLSSVRBPNNRW
EEMD1MAPE0.00840.00890.00840.00840.00880.00850.00900.02000.0186
RMSE0.64010.68270.63990.63990.67990.64670.68051.60441.7455
Dstat0.82130.81120.82310.81890.79800.81350.80760.73440.5084
3MAPE0.00960.01110.00970.00970.01070.01000.01180.01950.0296
RMSE0.75690.87020.75830.75600.84100.78030.94061.55992.5344
Dstat0.77460.73140.77280.77940.75000.77100.72720.68470.5000
6MAPE0.01200.01460.01210.01220.01470.01220.01260.02100.0396
RMSE0.94401.16020.95470.97041.15600.96660.98961.62973.1068
Dstat0.71460.66070.71400.70020.66250.72900.69240.62650.4976
ICEEMDAN1MAPE0.00430.00500.00430.00430.00480.00430.00440.00510.0175
RMSE0.34580.40390.34690.34410.39010.35050.35280.39641.6209
Dstat0.91010.89390.91010.91130.89750.90830.90710.89450.5228
3MAPE0.00730.00890.00740.00760.00870.00740.00750.00920.0286
RMSE0.59260.71700.59530.60670.70010.59840.60440.70222.4296
Dstat0.84530.80400.83990.84170.81240.83930.83330.81000.4862
6MAPE0.01020.01380.01020.01070.01300.01030.01040.01870.0400
RMSE0.80271.09770.81000.85131.02760.81370.82361.35313.1926
Dstat0.75900.66610.75840.75300.68470.76260.75780.68650.4982
Table 5. The Diebold-Mariano (DM) test results of ensemble models.
Table 5. The Diebold-Mariano (DM) test results of ensemble models.
HorizonDecompositionTested ModelICEEMDAN EEMD
LinRRPolyRRSigRRRbfRRMKRRLSSVRBPNNRW RRLinRRPolyRRSigRRRbfRRMKRRLSSVRBPNNRW
1ICEEMDANRR−7.6331
(0.0000)
−0.6142
(0.5392)
0.9695
(0.3324)
−7.5172
(0.0000)
−1.9834
(0.0475)
−3.5554
(0.0004)
−12.8611
(0.0000)
−5.5386
(0.0000)
−21.3654
(0.0000)
−22.2857
(0.0000)
−21.2054
(0.0000)
−21.2416
(0.0000)
−21.3583
(0.0000)
−20.9534
(0.0000)
−22.0337
(0.0000)
−16.7261
(0.0000)
−4.0125
(0.0001)
LinRR 8.0753
(0.0000)
7.8183
(0.0000)
1.8874
(0.0593)
6.9073
(0.0000)
6.8284
(0.0000)
0.9153
(0.3602)
−5.4460
(0.0000)
−17.1007
(0.0000)
−19.9938
(0.0000)
−17.0228
(0.0000)
−17.1490
(0.0000)
−18.1073
(0.0000)
−17.0138
(0.0000)
−18.4270
(0.0000)
−16.4052
(0.0000)
−3.9545
(0.0001)
PolyRR 2.1682
(0.0303)
−7.7155
(0.0000)
−1.7738
(0.0763)
−3.6728
(0.0002)
−12.0394
(0.0000)
−5.5375
(0.0000)
−21.3854
(0.0000)
−22.4548
(0.0000)
−21.3348
(0.0000)
−21.3959
(0.0000)
−21.4127
(0.0000)
−21.0541
(0.0000)
−22.1490
(0.0000)
−16.7245
(0.0000)
−4.0116
(0.0001)
SigRR −8.3841
(0.0000)
−2.7328
(0.0063)
−5.2931
(0.0000)
−12.1439
(0.0000)
−5.5416
(0.0000)
−21.3483
(0.0000)
−22.3840
(0.0000)
−21.2987
(0.0000)
−21.3724
(0.0000)
−21.5018
(0.0000)
−21.0097
(0.0000)
−22.1334
(0.0000)
−16.7370
(0.0000)
−4.0142
(0.0001)
RbfRR 6.1499
(0.0000)
6.7789
(0.0000)
−0.8816
(0.3781)
−5.4706
(0.0000)
−18.2902
(0.0000)
−20.4238
(0.0000)
−18.2305
(0.0000)
−18.4002
(0.0000)
−20.0042
(0.0000)
−18.1252
(0.0000)
−19.5005
(0.0000)
−16.5070
(0.0000)
−3.9687
(0.0001)
MKRR −0.7986
(0.4246)
−11.7816
(0.0000)
−5.5313
(0.0000)
−21.1164
(0.0000)
−22.0611
(0.0000)
−21.0784
(0.0000)
−21.0921
(0.0000)
−20.9704
(0.0000)
−20.7933
(0.0000)
−21.8257
(0.0000)
−16.7113
(0.0000)
−4.0080
(0.0001)
LSSVR −10.4989
(0.0000)
−5.5281
(0.0000)
−20.9187
(0.0000)
−22.0006
(0.0000)
−20.8537
(0.0000)
−20.9486
(0.0000)
−21.1186
(0.0000)
−20.6389
(0.0000)
−21.8183
(0.0000)
−16.6977
(0.0000)
−4.0058
(0.0001)
BPNN −5.4562
(0.0000)
−17.9970
(0.0000)
−19.3554
(0.0000)
−17.9211
(0.0000)
−17.9391
(0.0000)
−18.2960
(0.0000)
−17.8094
(0.0000)
−19.1642
(0.0000)
−16.4971
(0.0000)
−3.9608
(0.0001)
RW 4.8947
(0.0000)
4.7746
(0.0000)
4.8956
(0.0000)
4.8959
(0.0000)
4.7796
(0.0000)
4.8761
(0.0000)
4.7762
(0.0000)
0.1120
(0.9109)
−0.4916
(0.6231)
EEMDRR −5.9878
(0.0000)
0.0981
(0.9218)
0.0921
(0.9266)
−5.8039
(0.0000)
−1.9935
(0.0464)
−7.2864
(0.0000)
−14.8109
(0.0000)
−3.6137
(0.0003)
LinRR 5.8354
(0.0000)
6.0493
(0.0000)
0.3207
(0.7485)
4.6149
(0.0000)
0.2467
(0.8052)
−14.3910
(0.0000)
−3.5387
(0.0004)
PolyRR 0.0195
(0.9844)
−5.7741
(0.0000)
−2.2603
(0.0239)
−7.7598
(0.0000)
−14.8346
(0.0000)
−3.6141
(0.0003)
SigRR −5.8048
(0.0000)
−2.4071
(0.0162)
−8.0538
(0.0000)
−14.8512
(0.0000)
−3.6144
(0.0003)
RbfRR 4.6221
(0.0000)
−0.1080
(0.9140)
−14.4850
(0.0000)
−3.5431
(0.0004)
MKRR −7.2584
(0.0000)
−14.8197
(0.0000)
−3.6027
(0.0003)
LSSVR −14.5725
(0.0000)
−3.5415
(0.0004)
BPNN −0.6384
(0.5233)
3ICEEMDANRR−9.9325
(0.0000)
−1.4417
(0.1496)
−3.7347
(0.0002)
−9.4380
(0.0000)
−2.4333
(0.0151)
−4.2254
(0.0000)
−13.4240
(0.0000)
−8.2617
(0.0000)
−12.9322
(0.0000)
−15.5357
(0.0000)
−13.0108
(0.0000)
−12.8337
(0.0000)
−14.9738
(0.0000)
−13.9333
(0.0000)
−16.8289
(0.0000)
−14.1737
(0.0000)
−8.1352
(0.0000)
LinRR 10.0119
(0.0000)
9.7840
(0.0000)
1.8443
(0.0653)
9.5078
(0.0000)
9.0895
(0.0000)
1.0852
(0.2780)
−8.0440
(0.0000)
−2.6337
(0.0085)
−12.5877
(0.0000)
−2.7454
(0.0061)
−2.6451
(0.0082)
−8.2937
(0.0000)
−4.0327
(0.0001)
−10.8639
(0.0000)
−13.0047
(0.0000)
−7.9386
(0.0000)
PolyRR −4.4570
(0.0000)
−9.4540
(0.0000)
−1.1594
(0.2465)
−3.0217
(0.0026)
−13.1060
(0.0000)
−8.2580
(0.0000)
−12.5752
(0.0000)
−15.5953
(0.0000)
−12.7567
(0.0000)
−12.6621
(0.0000)
−14.8334
(0.0000)
−13.6041
(0.0000)
−16.7324
(0.0000)
−14.1551
(0.0000)
−8.1320
(0.0000)
SigRR −9.3379
(0.0000)
1.9688
(0.0491)
0.5195
(0.6035)
−10.9649
(0.0000)
−8.2399
(0.0000)
−11.3756
(0.0000)
−15.2838
(0.0000)
−11.5825
(0.0000)
−11.6158
(0.0000)
−14.3842
(0.0000)
−12.4692
(0.0000)
−16.2315
(0.0000)
−14.0525
(0.0000)
−8.1157
(0.0000)
RbfRR 8.7712
(0.0000)
8.8140
(0.0000)
−0.1757
(0.8605)
−8.0736
(0.0000)
−3.7902
(0.0002)
−11.6173
(0.0000)
−3.9565
(0.0001)
−3.9195
(0.0001)
−11.1280
(0.0000)
−5.2578
(0.0000)
−11.7123
(0.0000)
−13.2845
(0.0000)
−7.9644
(0.0000)
MKRR −1.6406
(0.1011)
−12.7449
(0.0000)
−8.2505
(0.0000)
−12.6415
(0.0000)
−15.4153
(0.0000)
−12.7391
(0.0000)
−12.5443
(0.0000)
−14.6474
(0.0000)
−13.6362
(0.0000)
−16.5479
(0.0000)
−14.1136
(0.0000)
−8.1265
(0.0000)
LSSVR −13.4650
(0.0000)
−8.2427
(0.0000)
−11.9772
(0.0000)
−14.9897
(0.0000)
−12.1661
(0.0000)
−12.0859
(0.0000)
−14.6889
(0.0000)
−13.4358
(0.0000)
−16.4904
(0.0000)
−14.0859
(0.0000)
−8.1157
(0.0000)
BPNN −8.0500
(0.0000)
−3.9068
(0.0001)
−9.2782
(0.0000)
−4.1170
(0.0000)
−3.9602
(0.0001)
−8.0803
(0.0000)
−5.6840
(0.0000)
−11.3597
(0.0000)
−13.3170
(0.0000)
−7.9423
(0.0000)
RW 7.9299
(0.0000)
7.7031
(0.0000)
7.9259
(0.0000)
7.9322
(0.0000)
7.7496
(0.0000)
7.8725
(0.0000)
7.4621
(0.0000)
5.0472
(0.0000)
−0.5382
(0.5905)
EEMDRR −8.0803
(0.0000)
−0.6800
(0.4966)
0.2481
(0.8041)
−7.6514
(0.0000)
−4.8069
(0.0000)
−12.0032
(0.0000)
−12.7901
(0.0000)
−7.8385
(0.0000)
LinRR 7.9869
(0.0000)
8.2376
(0.0000)
2.2903
(0.0221)
5.9929
(0.0000)
−3.7962
(0.0002)
−11.4198
(0.0000)
−7.6264
(0.0000)
PolyRR 1.1239
(0.2612)
−7.6591
(0.0000)
−4.7817
(0.0000)
−11.8134
(0.0000)
−12.8335
(0.0000)
−7.8359
(0.0000)
SigRR −8.2382
(0.0000)
−4.8751
(0.0000)
−12.0577
(0.0000)
−12.8268
(0.0000)
−7.8411
(0.0000)
RbfRR 5.3608
(0.0000)
−7.2419
(0.0000)
−11.9950
(0.0000)
−7.6714
(0.0000)
MKRR −10.3296
(0.0000)
−12.6336
(0.0000)
−7.7891
(0.0000)
LSSVR −10.7944
(0.0000)
−7.4217
(0.0000)
BPNN −5.2626
(0.0000)
6ICEEMDANRR−13.9660
(0.0000)
−2.1767
(0.0296)
−6.0937
(0.0000)
−12.9968
(0.0000)
−2.4287
(0.0153)
−4.9536
(0.0000)
−22.6250
(0.0000)
−14.1579
(0.0000)
−12.3204
(0.0000)
−15.3905
(0.0000)
−12.0345
(0.0000)
−12.7060
(0.0000)
−15.1929
(0.0000)
−12.4624
(0.0000)
−13.2884
(0.0000)
−19.3882
(0.0000)
−17.6734
(0.0000)
LinRR 14.3951
(0.0000)
14.3875
(0.0000)
4.9217
(0.0000)
13.1664
(0.0000)
13.1583
(0.0000)
−9.0778
(0.0000)
−13.5116
(0.0000)
7.7778
(0.0000)
−4.3984
(0.0000)
7.4146
(0.0000)
6.9775
(0.0000)
−3.0020
(0.0027)
6.0234
(0.0000)
5.4957
(0.0000)
−13.2753
(0.0000)
−16.9797
(0.0000)
PolyRR −7.2407
(0.0000)
−13.3457
(0.0000)
−0.7001
(0.4840)
−2.9388
(0.0033)
−22.1291
(0.0000)
−14.1522
(0.0000)
−11.6702
(0.0000)
−15.6854
(0.0000)
−11.7628
(0.0000)
−12.7550
(0.0000)
−15.3002
(0.0000)
−11.7487
(0.0000)
−13.1322
(0.0000)
−19.3071
(0.0000)
−17.6775
(0.0000)
SigRR −12.8356
(0.0000)
4.1760
(0.0000)
3.6412
(0.0003)
−20.1380
(0.0000)
−14.0771
(0.0000)
−7.3939
(0.0000)
−14.8116
(0.0000)
−8.1197
(0.0000)
−9.6893
(0.0000)
−14.2892
(0.0000)
−7.9142
(0.0000)
−10.1005
(0.0000)
−18.4915
(0.0000)
−17.5994
(0.0000)
RbfRR 11.8646
(0.0000)
12.5228
(0.0000)
−12.1621
(0.0000)
−13.6712
(0.0000)
5.1460
(0.0000)
−6.9781
(0.0000)
4.6379
(0.0000)
3.9114
(0.0001)
−9.3717
(0.0000)
3.4358
(0.0006)
2.4569
(0.0141)
−14.7236
(0.0000)
−17.1397
(0.0000)
MKRR −1.5720
(0.1161)
−22.1601
(0.0000)
−14.1246
(0.0000)
−10.9045
(0.0000)
−14.7753
(0.0000)
−10.8633
(0.0000)
−11.5082
(0.0000)
−14.4164
(0.0000)
−11.3219
(0.0000)
−12.0115
(0.0000)
−19.2344
(0.0000)
−17.6322
(0.0000)
LSSVR −22.0285
(0.0000)
−14.1061
(0.0000)
−10.1588
(0.0000)
−14.5608
(0.0000)
−10.2971
(0.0000)
−11.2871
(0.0000)
−14.9082
(0.0000)
−10.9314
(0.0000)
−12.6866
(0.0000)
−18.7911
(0.0000)
−17.6113
(0.0000)
BPNN −12.2861
(0.0000)
16.3390
(0.0000)
6.6381
(0.0000)
16.0747
(0.0000)
15.2332
(0.0000)
6.7072
(0.0000)
16.0034
(0.0000)
14.0536
(0.0000)
−7.3185
(0.0000)
−15.2268
(0.0000)
RW 13.8085
(0.0000)
13.2857
(0.0000)
13.7910
(0.0000)
13.7611
(0.0000)
13.2775
(0.0000)
13.7266
(0.0000)
13.6873
(0.0000)
11.0380
(0.0000)
0.7513
(0.4526)
EEMDRR −11.9133
(0.0000)
−2.8917
(0.0039)
−5.0101
(0.0000)
−11.8544
(0.0000)
−3.3420
(0.0009)
−7.1119
(0.0000)
−17.4323
(0.0000)
−17.2382
(0.0000)
LinRR 11.6342
(0.0000)
11.3981
(0.0000)
0.2374
(0.8124)
9.5491
(0.0000)
9.4853
(0.0000)
−12.3880
(0.0000)
−16.6841
(0.0000)
PolyRR −4.2271
(0.0000)
−11.7039
(0.0000)
−1.5927
(0.1114)
−5.6236
(0.0000)
−17.1205
(0.0000)
−17.2187
(0.0000)
SigRR −11.6910
(0.0000)
0.4412
(0.6591)
−3.2085
(0.0014)
−16.7302
(0.0000)
−17.1929
(0.0000)
RbfRR 9.9300
(0.0000)
11.0223
(0.0000)
−12.7276
(0.0000)
−16.6240
(0.0000)
MKRR −2.6530
(0.0081)
−16.3767
(0.0000)
−17.1081
(0.0000)
LSSVR −16.6247
(0.0000)
−17.0843
(0.0000)
BPNN −13.4903
(0.0000)

Share and Cite

MDPI and ACS Style

Li, T.; Zhou, Y.; Li, X.; Wu, J.; He, T. Forecasting Daily Crude Oil Prices Using Improved CEEMDAN and Ridge Regression-Based Predictors. Energies 2019, 12, 3603. https://doi.org/10.3390/en12193603

AMA Style

Li T, Zhou Y, Li X, Wu J, He T. Forecasting Daily Crude Oil Prices Using Improved CEEMDAN and Ridge Regression-Based Predictors. Energies. 2019; 12(19):3603. https://doi.org/10.3390/en12193603

Chicago/Turabian Style

Li, Taiyong, Yingrui Zhou, Xinsheng Li, Jiang Wu, and Ting He. 2019. "Forecasting Daily Crude Oil Prices Using Improved CEEMDAN and Ridge Regression-Based Predictors" Energies 12, no. 19: 3603. https://doi.org/10.3390/en12193603

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop