Next Article in Journal
Synergy Effects in Electromagnetic Properties of Phosphate Ceramics with Silicon Carbide Whiskers and Carbon Nanotubes
Previous Article in Journal
A Cloud Information Monitoring and Recommendation Multi-Agent System with Friendly Interfaces for Tourism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Period-Sequential Index Forecasting Algorithm for Time Series Data

1
School of Mechanical Engineering, Tongji University, Shanghai 201804, China
2
School of Mechatronic and Power Engineering, Jiangsu University of Science and Technology, Zhenjiang 212003, China
3
Company of Sino-German Institute for Intelligent Technologies and PhD Study, Qingdao 266000, China
4
School of Mechanical Engineering, Jiangnan University, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(20), 4386; https://doi.org/10.3390/app9204386
Submission received: 9 October 2019 / Accepted: 15 October 2019 / Published: 17 October 2019
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
A period-sequential index algorithm with sigma-pi neural network technology, which is called the (SPNN-PSI) method, is proposed for the prediction of time series datasets. Using the SPNN-PSI method, the cumulative electricity output (CEO) dataset, Volkswagen sales (VS) dataset, and electric motors exports (EME) dataset are tested. The results show that, in contrast to the moving average (MA), exponential smoothing (ES), and autoregressive integrated moving average (ARIMA) methods, the proposed SPNN-PSI method shows satisfactory forecasting quality due to lower error, and is more suitable for the prediction of time series datasets. It is also concluded that: There is a trend that the higher the correlation coefficient value of the reference historical datasets, the higher the prediction quality of SPNN-PSI method, and a higher value (>0.4) of correlation coefficient for SPNN-PSI method can help to improve occurrence probability of higher forecasting accuracy, and produce more accurate forecasts for the big datasets.

1. Introduction

In the big data era, a large number of time series data are continuously generated in the network systems, such as stock price, sales volume, production capacity, weather data, ocean engineering, engineering control, and largely in any system of applied science and engineering which involves investigations of time-varying parameters [1,2,3]. In general, the distribution of time series data changes over time, and is non-stationary [4,5], while some data shows potential periodicity characteristics. Since the 1950s, time series forecasting has received much interest in prediction science.
Continuously growing numbers of new algorithms have been proposed and studied on time series forecasting. Firstly, the exponential smoothing (ES) method [5,6] and the moving average (MA) method [7] are simple and widely used, and performed well in forecasting competitions against more sophisticated approaches. Secondly, the autoregressive integrated moving average (ARIMA) model integrates autoregressive models (AR) and moving average models (MA), and is widely used as a linear time series forecasting method [8,9]. The ARIMA model gives good accuracy in forecasting relatively stationary time series data, but needs a strong assumption that the values of future data are linearly dependent on the values of historical data [10]. Thirdly, the artificial neural network (ANN) [11,12,13] and adaptive models [14] have also been used to forecast nonlinear time series data, and improve forecasting accuracy in different time scales. It is also possible to hybrid different methods to improve overall forecasting accuracy [15].
However, no traditional forecasting methods can meet all the targets [16,17,18], while applying heuristic methods are also worth researching [19]. Here, a period-sequential index algorithm with a sigma-pi neural network (SPNN-PSI) is proposed and dedicated to the prediction of time series data. Here a period-sequential index algorithm (PSI) by identifying structures from transformed data, where there are four indexes carrying implicitly structure information usable for forecasting, the period index, sequential index, small period index, and super sequential index, is combined with a sigma-pi neural network algorithm (SPNN) improving the accuracy and robustness of forecasting algorithm. The SPNN-PSI method has a universal application, and a satisfactory prediction quality improved as the correlation coefficient value of the reference historical datasets increased.

2. Theoretical Model

2.1. Period-Sequential Index (PSI) Algorithm

Finding index-values implicitly carrying structure information, a period-sequential index (PSI) algorithm, is proposed to predict the time series data. The index-values cover the complete period, while the period index and sequential index, as well as small period index and super sequential index, are introduced to describe the dataset structure information in vertical and horizontal dimensions, respectively. In this way, for time series data, the following year’s dataset can be predicted using only two consecutive years of reference historical datasets. Figure 1 shows the schematic diagram of the PSI algorithm. H−2, H−1 denote reference historical periods, i.e., the year before last and last year. H0 represents the forecasting period. The period for H−2, H−1, and H0 is uniform, and defined as T in this paper. At historical time of t, PI(t), SI(t), pi(t), and si(t) describe the period index, sequential index, small period index, and super sequential index, respectively.
We assume that the forecasting value follows the measurement Equation (1),
F ( t i ) = G ( PI ( t i 2 T ) , PI ( t i T ) , SI ( t i 2 T ) , SI ( t i T ) , pi ( t i 2 T ) , pi ( t i T ) , si ( t i 2 T ) , si ( t i T ) ) ,
where F(ti) is the forecasting value at time of ti (i = 1, 2 … N) during H0 period. N represents the number of model forecasting samples. PI(ti − 2T), PI(tiT), SI(ti − 2T), SI(tiT), pi(ti − 2T), pi(tiT), si(ti − 2T), and si(tiT), as the eight variables in Equation (1), are described as follows:
(1) Period index
The period index indicates the relationship between the reference historical data and the reference value which can be explained through Equations (2) and (3),
P I ( t i 2 T ) = y ( t i 2 T ) K 2 ,
P I ( t i T ) = y ( t i T ) K 1 ,
where y(ti − 2T) and y(tiT) describe the reference historical data at time of ti − 2T and tiT, respectively. K−2 and K−1 are reference functions of period index. A standard period average is originally set to be a reference function of period index, and where it is defined as a constant.
(2) Sequential Index
The sequential index indicates the relationship between two adjacent reference historical data with the defined time steps. It is calculated through Equations (4) and (5).
S I ( t i 2 T ) = y ( t i + 1 2 T ) y ( t i 2 T )
S I ( t i T ) = y ( t i + 1 T ) y ( t i T )
(3) Small Period Index
The small period index indicates the relationship between the reference historical data and the reference value, which can be explained through Equations (6) and (7),
p i ( t i 2 T ) = y ( t i 2 T ) k 2 ,
p i ( t i T ) = y ( t i T ) k 2 ,
where k−1 and k−2 are reference functions of a small period index. Here a small period (such as three months, because of seasonal factors) average, is originally set to be a reference function of small period index.
(4) Super Sequential Index
The super sequential index indicates the relationship between two interval reference historical data with the defined steps. It is calculated through Equations (8) and (9).
s i ( t i 2 T ) = y ( t i + 2 2 T ) y ( t i 2 T )
s i ( t i T ) = y ( t i + 2 T ) y ( t i T )
Therefore, the forecasting value is given by:
F ( t i ) = c · P I ( t i ) · K 0 + d · S I ( t i 1 ) · y ( t i 1 ) + e · p i ( t i ) · k 0 + f · s i ( t i 2 ) · y ( t i 2 ) ,
where,
P I ( t i ) = 0.5 · P I ( t i 2 T ) + 0.5 · P I ( t i T )
S I ( t i 1 ) = 0.5 · S I ( t i 1 2 T ) + 0.5 · S I ( t i 1 T )
p i ( t i ) = 0.5 · p i ( t i 2 T ) + 0.5 · p i ( t i T )
s i ( t i 2 ) = 0.5 · s i ( t i 2 2 T ) + 0.5 · s i ( t i 2 T )
K 0 = γ 0 · K 1
k 0 = γ 0 · k 1
γ 0 = m e d i a n { y ( t 1 T ) , y ( t 2 T ) , y ( t N T ) } m e d i a n { { y ( t 1 2 T ) , y ( t 2 2 T ) , y ( t N 2 T ) }
where, c, d, e, and f are the weighing factors of forecasting Equation (10). K0 and k0 are the correction coefficients for period index and small period index, respectively. γ0 is planning factor, which is defined by median method in Equation (17).

2.2. SPNN-PSI Method

SPNN has been proposed by Lyutikova [20]. The products of the different linear combinations of the inputs in SPNN are the output of the network. SPNN has simpler structure, less variance, and faster convergence speed. A high SPNN degree determines the function that defines the relationship between output and input to depend on more parameters and to have a more complex structure. This can contribute to better prediction results but can also cause overfitting, which requires more computation time for the training algorithm. In the study, the architecture of SPNN-PSI with degree 4 and 8 inputs is shown in Figure 2, which can reproduce modeling function of Equation (10). As shown in Figure 2, PI(ti), K0, SI(ti−1), y(ti−1), pi(ti), k0, si(ti−2), and y(ti−2) are the input parameter of SPNN-PSI; F(ti) is the predicted output value at time of ti; c, d, e, and f are the weight values of network connection.
The input vector zi is given by
z i = [ P I ( t i ) K 0 S I ( t i 1 ) y ( t i 1 ) p i ( t i ) k 0 s i ( t i 2 ) y ( t i 2 ) ] .
The weight matrix w is defined as follows
w = [ c d e f ] .
In order to obtain the optimized weight values of c, d, e, and f, the output vector is defined as the observation data y(tiT) during H−1 period.
y i = y ( t i T ) = w z i
In this study, multiple inputs and outputs are combined into the following equation:
Y = w Z
where Y and Z are a matrix of multiple outputs and matrix of multiple inputs, respectively.
Y = [ y ( t 1 ) y ( t 2 ) y ( t i 1 ) y ( t i T ) y ( t N T ) ]
Z = [ P I ( t 1 ) K 0 P I ( t 2 ) K 0 P I ( t 3 ) K 0 P I ( t i ) K 0 P I ( t i + 1 ) K 0 P I ( t N ) K 0 S I ( t N T ) y ( t N T ) S I ( t 1 ) y ( t 1 ) S I ( t 2 ) y ( t 2 ) S I ( t i 1 ) y ( t i 1 ) S I ( t i ) ( γ 0 y ( t i T ) ) S I ( t N 1 ) ( γ 0 y ( t N 1 T ) ) p i ( t 1 ) k 0 p i ( t 2 ) k 0 p i ( t 3 ) k 0 p i ( t i ) k 0 p i ( t i + 1 ) k 0 p i ( t N ) k 0 s i ( t N 1 T ) y ( t N 1 T ) s i ( t N T ) y ( t N T ) s i ( t 1 ) y ( t 1 ) s i ( t i 2 ) y ( t i 2 ) s i ( t i 1 ) y ( t i 1 ) s i ( t N 2 ) ( γ 0 y ( t N 2 T ) ) ]
Then, based on the measured value Y and input Z, the learning control method of the initial training neural network is used to get the optimal weight matrix w, which is given:
w T = ( Z Z T ) 1 Z Y T .

2.3. Error Evaluation

In order to evaluate the obtained results, the forecasting accuracy was measured with three error indicators, that are the mean absolute percentage error (MAPE), the root mean squared error (RMSE) and the mean absolute error (MAE) [20,21]. Calculation equations of error indicators are given in the below Equations (25)–(27). In addition, the Pearson correlation coefficient (r) in Equation (28) is also used to quantify the strength and direction of the linear relationship between two sets of data during the reference historical period [22].
M A P E = 1 N i = 1 N | y ( t i ) F ( t i ) y ( t i ) | × 100 %
R M S E = i = 1 N ( y ( t i ) F ( t i ) ) 2 N
M A E = i = 1 N | y ( t i ) F ( t i ) | N
r = i = 1 N ( y ( t i 2 T ) K 2 ) ( y ( t i T ) K 1 ) i = 1 N ( y ( t i 2 T ) K 2 ) 2 i = 1 N ( y ( t i T ) K 1 ) 2
where y(ti) is the measured value at time of ti during H0; F(ti) is the forecasting value at time of ti during H0.

3. Steps of Computation

The flow chart of the optimized network model is shown in Figure 3. The specific steps are as follows:
Step 1:
Initialize the reference historical datasets.
Step 2:
Training the reference historical two-cycle-years datasets, calculate the parameters Y and Z by solving Equations (22) and (23).
Step 3:
At the time step of ti, take Y and Z into the Equation (24), and solve it, and get the optimal weight matrix w.
Step 4:
Update the optimal values of c, d, e, and f in Equation (10), solve Equation (10), and get the forecasting value F(ti).
Step 5:
If the time steps of the stop condition (ti+1 > tN) are satisfied, the search stops, and output parameters of MAPE, RMSE, MAE, and r; Otherwise, the time step is added, and the procedure returns Step 2.

4. Results and Discussion

Three groups of actual time series datasets [23,24,25] are shown in Table 1. In these three samples, two-cycle-years datasets from January 2016 to December 2017 are used as reference historical data to forecast the data of 2018. In order to evaluate the confidence of the period-sequential index algorithm with sigma-pi neural network (SPNN-PSI), the correlation coefficients of datasets for 2016 and 2017 are further computed, and listed in Table 1. Then, a comparison analysis of the prediction value and the real value was implemented by using the SPNN-PSI, MA, ES, and ARIMA methods, so as to show a more direct observation of the prediction.

4.1. Periodic Recognition and Prediction on Electric Motors Exports (EME) Dataset

For the EME datasets, the electric motors exports in China between January 1995 and June 2019 is used as an example [23] to show the correlation coefficient detection and prediction results. Using the reference historical two-cycle-years dataset from January 2016 to December 2017, the monthly electric motors exports from January to December of 2018 are predicted by using the SPNN-PSI, MA, ES, and ARIMA methods, as shown in Figure 4. It can be seen from Figure 4 that, compared with MA, ES, and ARIMA methods, the SPNN-PSI method demonstrates a better prediction trend with good volatility and following quality.
Table 2 presents a more visual view of prediction errors of each model. According to the results obtained in Table 2, the corresponding error of MAPE is 5.34%, 6.79%, 8.11%, and 6.97% for the PSI, MA, ES, and ARIMA methods, respectively. It also can be noticed that the four errors are within a reasonable range, but the developed SPNN-PSI algorithm is more suitable for the prediction of the EME dataset used in this paper due to lower MAPE, RMSE, and MAE, as compared to the other three prediction methods. In addition, the correlation coefficient of its historical reference data is shown in Table 1. The example shows that the proposed SPNN-PSI algorithm achieves satisfactory accuracy in time series prediction on the EME dataset whose historical reference data has a relatively higher correlation coefficient value of 0.9388.

4.2. Periodic Recognition and Prediction on Volkswagen Sales (VS) Dataset

For the VS dataset, the Volkswagen sales in China between January 2007 and June 2019 is used as another example [24] to show the correlation coefficient detection and prediction results. Figure 5 shows the prediction results in 2018 by using the PSI, MA, ES, and ARIMA methods. As shown in Figure 5, between January 2018 and May 2018, the predicted VS using the PSI method can be almost identical with actual VS, comparing to the MA, ES, and ARIMA methods, and a lower MAPE of 4.48% is also achieved by Equation (18). Meanwhile, it can be seen that, the trend of the predicted values between June 2018 and December 2018 is similar to that of the actual values when using PSI method, but the difference between them at each time point is larger than that of using the MA, ES, and ARIMA methods, achieving a value of MAPE at 19.36% for the period between June 2018 and December 2018.
To clearly show the correlation coefficient value and the prediction results, we show correlation coefficient detection results from January 2016 to December 2017 (shown in Table 1), and Table 3 presents the further prediction errors of each model. Compared with the EME dataset, the time series dataset of VS has a lower correlation coefficient value (=0.8392) of reference historical dataset. However, a higher MAPE value of 9.94% for the VS dataset in 2018 can be achieved in Table 3. Therefore, the example shows that the prediction accuracy and quality of the proposed SPNN-PSI algorithm can be decreased due to the decreasing correlation coefficient value of reference historical dataset (shown in Table 1), as compared to that of the EME dataset.
To sum up, the MAPE, RMSE, and MAE values by SPNN-PSI method in Table 3 are all lower than that by MA, ES, and ARIMA methods. Thus, the forecasting quality of SPNN-PSI method is better, and the developed SPNN-PSI algorithm is still suitable for the prediction of VS dataset used in this paper.

4.3. Periodic Recognition and Prediction on Cumulative Electricity Output (CEO) Dataset

For the CEO dataset, a time series subset of cumulative electricity output in China [25] is used as an example to show the correlation coefficient detection and prediction results. For the large-scale reference historical time series dataset from January 2016 to December 2017, there is an obvious similarity periodicity in the CEO dataset with a length of 12 months (one year), and shows an upward trend from January to December. Based on the detected periodic model from January 2016 to December 2017, we predict the CEO for the next periodic (the whole year of 2018), as shown on the right side of Figure 6. We can see that the predicted values are very close to the actual values when using PSI method, while there are large deviations when using other methods, especially in January 2018.
Then, we evaluate the forecasting accuracy of the SPNN-PSI method by comparing the actual values and the predicted values, shown in Table 4. Because the correlation coefficient for reference historical dataset achieves a very high value (equal to 1.0, shown in Table 1), the MAPE, RMSE, and MAE by SPNN-PSI method are lower and equal 5.23%, 2912.23 × 108, and 2374.38 KWH, which indicates a smaller difference between the prediction value and actual value. By contrast, the higher MAPE, RMSE, and MAE by other forecasting methods are also given in Table 4. Their MAPE values are all more than 50%, which indicates a relatively large fluctuation of prediction error.
In a sum, based on the above error indicators, the proposed SPNN-PSI algorithm achieves very high accuracy and quality in time series prediction on the CEO dataset with its very high correlation coefficient value of 1.0.

4.4. Accuracy Analysis of SPNN-PSI Algorithm

After receiving the forecasting results of the above three groups of time series datasets, the forecasting accuracy (FA) was also calculated by using Equation (29). If the FA is close to 100% and the MAPE is close to 0, the model is considered to have excellent forecasting accuracy. Then, the FA of the VS, EME, and CEO datasets (in Table 1) were further analyzed and shown in Table 5. It can be seen that the FA can increase when the correlation coefficient (r) of the reference historical data increases. Thus, we think that there may be a positive correlation between r and FA.
F A = 1 M A P E
To further illustrate this correlation between the two, it was applied to the first case of the monthly export volume of specifically chosen 31 different kinds of products from January 1995 to December 2018 [23]. The scatter diagram of FA vs. r is shown in Figure 7a. It can be seen in Figure 7a that, dense point clouds are located in one area surrounded by red lines, while some individual points scatter in the other areas. It is evident that, when using SPNN-PSI method in this paper, the higher value of r (r > 0.4) can help to improve occurrence probability of higher value of FA (FA > 70%), and produce more accurate forecasts for the datasets of monthly export volume in the first case. Then, it is further examined empirically by using another case of big data, referring to vibration signal in hydraulic test rig with a sampling frequency of 1 Hz during 132,300 s [26]. This test rig cyclically repeats constant load cycles (duration 60 s) and measures process values. Figure 7b gives the scatter diagram of FA vs. r. It can be seen that Figure 7a,b shows a similar characteristic of scatter distribution, but whether r is positive or not in Figure 7b, the proposed PSI model can predict high values of FA (between 92% and 100%) and produces accurate forecasts of vibration. If we consider r is one dominant factor for forecasting accuracy, in the condition of r > 0.4, higher values of FA > 97% are easier to be applied and predicted using the proposed PSI algorithm.

5. Conclusions

  • The SPNN-PSI method with four indexes—the period index, sequential index, small period index, and super sequential index, by finding index-values implicitly carrying usable structure information, combined with a neural network, is initially proposed for predict the time series datasets.
  • In contrast to the MA, ES, and ARIMA methods, the proposed SPNN-PSI method shows satisfactory forecasting quality due to lower MAPE, RMSE, and MAE, and is more suitable for the prediction of time series datasets.
  • There is a trend that the higher the correlation coefficient value of the reference historical datasets, the higher the prediction quality of SPNN-PSI method; a higher value (>0.4) of the correlation coefficient for the SPNN-PSI method can help to improve the occurrence probability of higher forecasting accuracy, and produce more accurate forecasts for the big datasets.

Author Contributions

Conceptualization, H.J. and D.F., and K.S.; formal analysis, H.J. and D.F.; investigation, H.J. and D.F.; writing—original draft preparation, H.J. and F.C.; writing—review and editing, H.J., D.F., and B.L.

Funding

The authors gratefully acknowledge the support of key projects for international cooperation in scientific and technological innovation between governments, through grant No. 2017YFE0101600.

Acknowledgments

The authors would like to express their sincere thanks to Klaus Spicher from Sino–German Institute for Intelligent Technologies and PhD Study for his assistance in heuristic prediction method.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

c, d, e, fweighing factors of forecasting equation
FAforecasting accuracy
F(ti)forecasting value at time of ti during H0 period
G()measurement equation for forecasting value
H−2, H−1reference historical period
H0forecasting period
−2, −1, 0subscripts, and represent during period of H−2, H−1, H0, respectively
k0correction coefficients for small period index
k−1, k−2average of measured values during small periods of H−1, H−2
K0correction coefficients for period index
K−1, K−2average of measured values during periods of H−1, H−2
MAEmean absolute error
MAPEmean absolute percentage error
Nnumber of forecasting samples
rcorrelation coefficient
RMSEroot mean squared error
pi(t)small period index at time of t
PI(t)period index at time of t
si(t)super sequential index at time of t
SI(t)sequential index at time of t
ttime (t = t1 − 2T, t2 − 2T, … tN − 2T, t1T, t2T, … tNT, t1, t2, … tN)
Tperiod for H−2, H−1, and H0
y(t)measured value at time of t
yioutput vector
Ymatrix of multiple outputs
wweight matrix
ziinput vector
Zmatrix of multiple inputs
γ0planning factor

References

  1. Cao, L.J.; Tay, F.E.H. Support vector machine with adaptive parameters in financial time series forecasting. IEEE Trans. Neural Netw. 2003, 14, 1506–1518. [Google Scholar] [CrossRef] [PubMed]
  2. Shi, G.; Guo, J.; Huang, W.; Williams, B.M. Modeling seasonal heteroscedasticity in vehicular traffic condition series using a seasonal adjustment approach. J. Transp. Eng. 2014, 140, 4014012. [Google Scholar] [CrossRef]
  3. Stefanakos, C. Fuzzy time series forecasting of nonstationary wind and wave data. Ocean Eng. 2016, 121, 1–12. [Google Scholar] [CrossRef] [Green Version]
  4. Yaser, S.A.M.; Atiya, A.F. Introduction to financial forecasting. Appl. Intell. 1996, 6, 205–213. [Google Scholar]
  5. Dong, Z.; Yang, D.; Reindl, T.; Walsh, W.M. Short-term solar irradiance forecasting using exponential smoothing state space model. Energy 2013, 55, 1104–1113. [Google Scholar] [CrossRef]
  6. Billah, B.; King, M.L.; Snyder, R.D.; Koehler, A.B. Exponential smoothing model selection for forecasting. Int. J. Forecast. 2006, 22, 239–247. [Google Scholar] [CrossRef] [Green Version]
  7. Barrow, D.K. Forecasting intraday call arrivals using the seasonal moving average method. J. Bus. Res. 2016, 69, 6088–6096. [Google Scholar] [CrossRef]
  8. Büyüksahin, U.C.; Ertekin, S. Improving forecasting accuracy of time series data using a new ARIMA-ANN hybrid method and empirical mode decomposition. Neurocomputing 2019, 361, 151–163. [Google Scholar] [CrossRef] [Green Version]
  9. Wen, X.H.; Feng, Q.; Deo, R.; Wu, M. Two-phase extreme learning machines integrated with the complete ensemble empirical mode decomposition with adaptive noise algorithm for multi-scale runoff prediction problems. J. Hydrol. 2019, 570, 167–184. [Google Scholar] [CrossRef]
  10. Xie, T.; Zhang, G.; Hou, J.; Xie, J.; Lv, M.; Liu, F. Hybrid forecasting model for non-stationary daily runoff series: A case study in the Han River Basin, China. J. Hydrol. 2019, 577, 123915. [Google Scholar] [CrossRef]
  11. Voyant, C.; Muselli, M.; Paoli, C.; Nivet, M.L. Optimization of an artificial neural network dedicated to the multivariate forecasting of daily global radiation. Energy 2011, 36, 348–359. [Google Scholar] [CrossRef] [Green Version]
  12. Mellit, A.; Kalogirou, S.A. Artificial intelligence techniques for photovoltaic applications: A review. Prog. Energy Combust. Sci. 2008, 34, 574–632. [Google Scholar] [CrossRef]
  13. Linares-Rodríguez, A.; Ruiz-Arias, J.A.; Pozo-Vázquez, D.; Tovar-Pescador, J. Generation of synthetic daily global solar radiation data based on ERA-Interim reanalysis and artificial neural networks. Energy 2011, 36, 5356–5365. [Google Scholar] [CrossRef]
  14. Mellit, A.; Eleuch, H.; Benghanem, M.; Elaoun, C.; Pavan, A.M. An adaptive model for predicting of global, direct and diffuse hourly solar irradiance. Energy Convers. Manag. 2010, 51, 771–782. [Google Scholar] [CrossRef]
  15. Mostafavi, E.S.; Ramiyani, S.S.; Sarvar, R.; Moud, H.I.; Mousavi, S.M. A hybrid computational approach to estimate solar global radiation: An empirical evidence from Iran. Energy 2013, 49, 204–210. [Google Scholar] [CrossRef]
  16. Fang, D.J.; Weng, W.B. Sales Forecasting System for Chinese Tobacco Wholesalers. In Proceedings of the 2nd International Conference on Innovative Computing, Communication, Information Technology, and Ocean Engineering CICC-ITOE201, Macao, China, 5–6 March 2011; Volume 11, pp. 380–386. [Google Scholar]
  17. Yolcu, O.C.; Lam, H.K. A combined robust fuzzy time series method for prediction of time series. Neurocomputing 2017, 247, 87–101. [Google Scholar] [CrossRef] [Green Version]
  18. Bas, E.; Grosan, C.; Egrioglu, E. High order fuzzy time series method based on pi-sigma neural network. Eng. Appl. Artif. Intell. 2018, 72, 350–356. [Google Scholar] [CrossRef]
  19. Fang, D.; Zhang, Y.; Spicher, K. Forecasting Accuracy Analysis based on two new heuristical methods and Holt-Winters-Method. In Proceedings of the 2016 IEEE International Conference on Big Data Analysis, Hangzhou, China, 12–14 March 2016; pp. 152–157. [Google Scholar]
  20. Lyutikova, L.A. Sigma-Pi neural networks: Error correction methods. Procedia Comput. Sci. 2018, 145, 312–318. [Google Scholar] [CrossRef]
  21. Hannan, M.A.; Lipu, M.S.H.; Hussain, A. Neural network approach for estimating state of charge of lithium-ion battery using backtracking search algorithm. IEEE Access 2018, 6, 10069–10079. [Google Scholar] [CrossRef]
  22. Joshi, B.; Kay, M.; Copper, J.K.; Sproul, A.B. Evaluation of solar irradiance forecasting skills of the Australian Bureau of Meteorology’s ACCESS models. Sol. Energy 2019, 188, 386–402. [Google Scholar] [CrossRef]
  23. HeXun-Macro Data—National Accounts, China Electric Motors Exports Dataset. Available online: http://mac.hexun.com/Default.shtml?id=A511M (accessed on 9 October 2019).
  24. Owner’s Home, Volkswagen Sales (VS) Dataset. Available online: http://xl.16888.com/b/57411/ (accessed on 9 October 2019).
  25. HeXun-Macro Data—Three Major Industries, China Cumulative Electricity Output Dataset. Available online: http://mac.hexun.com/Default.shtml?id=E210M (accessed on 9 October 2019).
  26. Kaggle-Dataset-Condition Monitoring of Hydraulic Systems, Condition Assessment of a Hydraulic Test Rig Based on Multi Sensor Data. Available online: https://www.kaggle.com/jjacostupa/condition-monitoring-of-hydraulic-systems (accessed on 1 April 2018).
Figure 1. The schematic diagram of the period-sequential index algorithm (PSI) algorithm.
Figure 1. The schematic diagram of the period-sequential index algorithm (PSI) algorithm.
Applsci 09 04386 g001
Figure 2. The architecture of the period-sequential index algorithm with sigma-pi neural network (SPNN-PSI).
Figure 2. The architecture of the period-sequential index algorithm with sigma-pi neural network (SPNN-PSI).
Applsci 09 04386 g002
Figure 3. The flow chart of the optimized network model.
Figure 3. The flow chart of the optimized network model.
Applsci 09 04386 g003
Figure 4. Time series pattern and prediction on the electric motors exports (EME) dataset.
Figure 4. Time series pattern and prediction on the electric motors exports (EME) dataset.
Applsci 09 04386 g004
Figure 5. Time series pattern and prediction on the Volkswagen sales (VS) dataset.
Figure 5. Time series pattern and prediction on the Volkswagen sales (VS) dataset.
Applsci 09 04386 g005
Figure 6. Time series pattern and prediction on cumulative electricity output (CEO) dataset.
Figure 6. Time series pattern and prediction on cumulative electricity output (CEO) dataset.
Applsci 09 04386 g006
Figure 7. Scatter diagram of forecasting accuracy (FA) vs. correlation coefficient (r) r. (a) The monthly export volume of specifically chosen 31 different kinds of products from January 1995 to December 2018; (b) the vibration data of hydraulic test rig with a sampling frequency of 1 Hz during 132,300 s.
Figure 7. Scatter diagram of forecasting accuracy (FA) vs. correlation coefficient (r) r. (a) The monthly export volume of specifically chosen 31 different kinds of products from January 1995 to December 2018; (b) the vibration data of hydraulic test rig with a sampling frequency of 1 Hz during 132,300 s.
Applsci 09 04386 g007
Table 1. Time series datasets used in the experiments.
Table 1. Time series datasets used in the experiments.
Dataset NamesReference Historical DataForecasting Data
Month/YearCorrelation CoefficientMonth/Year
Electric motors exports (EME) [23] datasetJanuary 2016–December 20170.9388January 2018–December 2018
Volkswagen sales (VS) dataset [24] datasetJanuary 2016–December 20170.8392January 2018–December 2018
Cumulative electricity output (CEO) [25] datasetJanuary 2016–December 20171.0000January 2018–December 2018
Table 2. Error indicators of each prediction model on the EME dataset.
Table 2. Error indicators of each prediction model on the EME dataset.
ModelError Indicators
MAPE (%)RMSE (104 Sets)MAE (104 Sets)
SPNN-PSI5.341480.061179.87
MA6.792062.601393.56
ES8.112406.331667.28
ARIMA6.971873.411535.20
Table 3. Error indicators of each prediction model on the VS dataset.
Table 3. Error indicators of each prediction model on the VS dataset.
Prediction ModelError Indicators
MAPE (%)RMSE (Sets)MAE (Sets)
SPNN-PSI9.9432,14824,970
MA12.3341,61029,998
ES14.3846,81033,500
ARIMA15.1348,21135,702
Table 4. Error indicators of each prediction model on the CEO dataset.
Table 4. Error indicators of each prediction model on the CEO dataset.
Prediction ModelError Indicators
MAPE (%)RMSE (108 KWH)MAE (108 KWH)
SPNN-PSI5.232912.232374.38
MA79.0821,870.0517,608.59
ES61.1416,757.2610,136.89
ARIMA86.6323,782.3513,251.02
Table 5. Comparison of forecasting accuracy for the VS, EME, and CEO datasets.
Table 5. Comparison of forecasting accuracy for the VS, EME, and CEO datasets.
Dataset NamesrMAPE (%)FA (%)
VS0.83929.9490.06
EME0.93885.3494.66
CEO1.00005.2394.77

Share and Cite

MDPI and ACS Style

Jiang, H.; Fang, D.; Spicher, K.; Cheng, F.; Li, B. A New Period-Sequential Index Forecasting Algorithm for Time Series Data. Appl. Sci. 2019, 9, 4386. https://doi.org/10.3390/app9204386

AMA Style

Jiang H, Fang D, Spicher K, Cheng F, Li B. A New Period-Sequential Index Forecasting Algorithm for Time Series Data. Applied Sciences. 2019; 9(20):4386. https://doi.org/10.3390/app9204386

Chicago/Turabian Style

Jiang, Hongyan, Dianjun Fang, Klaus Spicher, Feng Cheng, and Boxing Li. 2019. "A New Period-Sequential Index Forecasting Algorithm for Time Series Data" Applied Sciences 9, no. 20: 4386. https://doi.org/10.3390/app9204386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop