Next Article in Journal
Simulation of One-Dimensional Solute Transport with Equilibrium-Controlled Non-Linear Sorption Using Modular Three-Dimensional Multispecies Transport Model
Previous Article in Journal
Preparation and Catalytic Study of Mn-NaX, Cu-NaX and Ag-/AgNPs-NaX Zeolites
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Recurrent Forecasting in Singular Spectrum Decomposition †

by
Maryam Movahedifar
1,*,‡,
Hossein Hassani
2,‡ and
Mahdi Kalantari
3,‡
1
Institute for Statistics, University of Bremen, 28359 Bremen, Germany
2
Research Institute of Energy Management and Planning, University of Tehran, Tehran 1417466191, Iran
3
Department of Statistics, Payame Noor University, Tehran 19395-4697, Iran
*
Author to whom correspondence should be addressed.
Presented at the 9th International Conference on Time Series and Forecasting, Gran Canaria, Spain, 12–14 July 2023.
These authors contributed equally to this work.
Eng. Proc. 2023, 39(1), 68; https://doi.org/10.3390/engproc2023039068
Published: 7 July 2023
(This article belongs to the Proceedings of The 9th International Conference on Time Series and Forecasting)

Abstract

:
In this paper, the Recurrent Singular Spectrum Decomposition (R-SSD) algorithm is proposed as an improvement over the Recurrent Singular Spectrum Analysis (R-SSA) algorithm for forecasting non-linear and non-stationary narrowband time series. R-SSD modifies the embedding step of the basic SSA method to reduce energy residuals. This paper conducts simulations and real-case studies to investigate the properties of the R-SSD method and compare its performance with R-SSA. The results show that R-SSD yields more accurate forecasts in terms of ratio root mean squared errors (RRMSEs) and ratio mean absolute errors (RMAEs) criteria. Additionally, the Kolmogorov–Smirnov Predictive Accuracy (KSPA) test indicates significant accuracy gains with R-SSD over R-SSA, as it measures the maximum distance between the empirical cumulative distribution functions of recurrent prediction errors and determines whether a lower error leads to stochastically less error. Finally, the non-parametric Wilcoxon test confirms that R-SSD outperforms R-SSA in filtering and forecasting new data points.

1. Introduction

Singular Spectrum Analysis (SSA) is a widely used tool for time series analysis and signal processing, first introduced by Broomhead and King [1] in 1986. Over the years, several studies, including [2,3,4,5,6,7,8,9], have attempted to improve the decomposition, reconstruction, and forecasting capabilities of SSA in various fields. The method breaks down a time series into a few principal components that are used to reconstruct the original series, making it an efficient analysis tool that focuses on the most relevant features of the data. Moreover, SSA does not rely on statistical assumptions such as linearity or stationarity, which are often unrealistic in real-world scenarios. Both univariate and multivariate time series data can be analyzed using SSA, with the former examining a single time series variable and the latter studying multiple time series variables simultaneously, for more details see [9,10,11,12,13,14,15,16,17]. Singular Spectrum Analysis (SSA) can be utilized for forecasting future trends. The first step in applying SSA to forecasting is to decompose the time series into its trend, seasonal, and noise components. Once these components have been identified, they can be extrapolated into the future using various methods, such as Vector SSA (V-SSA) and Recurrent SSA (R-SSA) [17]; while V-SSA has proven effective in many instances, there is still room for improvement in the R-SSA forecasting approach. This paper proposes an innovative recurrent forecasting algorithm called R-SSD, which is expected to generate more accurate results. The R-SSD method generates its coefficients from a modified trajectory matrix based on the new Singular Spectrum Decomposition (SSD) method over time-frequency datasets, see [18]. SSD is an iterative approach that is based on the SSA decomposition method and chooses the embedding dimension and principal components for the reconstruction and forecasting of a specific component series in a fully data-driven manner. In the Singular Spectrum Analysis (SSA) method, the number of observations needed to construct the trajectory matrix is not fixed and can vary. On the other hand, the Singular Spectrum Decomposition (SSD) method requires a fixed number of repetitions of observations to construct the trajectory matrix. The window length, denoted as L, determines the number of rows in the trajectory matrix in both methods. A larger window length is preferred if the goal is to retain more information, while a smaller window length is better for achieving statistical confidence, for more details see [19,20]. When addressing time series that exhibit different frequency domains, such as those with harmonic patterns where, for example, the first half of the signal has low-frequency and the second half has high-frequency oscillations, extracting the oscillatory components using the SSA method can be challenging as it requires setting an appropriate window length at each step. However, the SSD method overcomes this limitation by setting the embedding dimension or window length ( L N / 2 ) as a linear function of the inverse of the dominant frequency of the data, denoted as 1 / f m a x . This adaptive approach ensures that SSD is a flexible decomposition method that can increase its ability to capture oscillatory components while reducing residual energy, as detailed in Appendix A.1 of [21]. As a result, it can be expected that SSD, being an improved version of SSA, can provide more accurate predictions for new data points in time series with different frequency domains.
The structure of this paper is as follows. In Section 2, we provide an introduction to the methodology of the basic SSA method and the recurrent forecasting algorithm. In Section 3, we present the methodology of the novel R-SSD forecasting approach. The results of a simulation study, evaluating the properties and performance of the proposed R-SSD method and comparing it to the established R-SSA approach, as well as the analysis of real data, are reported in Section 4. All calculations were performed using R software, specifically the Rssa package. Finally, in Section 5, we provide concluding remarks and highlight the key findings of our study.

2. Singular Spectrum Analysis (SSA)

Singular Spectrum Analysis (SSA) is an effective nonparametric technique for analyzing data. It can decompose a series into multiple components and make predictions based on them. The method comprises two distinct stages: decomposition and reconstruction, each of which involves two separate steps. To perform the SSA method, Algorithm 1 outlines the general process, and we primarily rely on the guidelines presented in [22,23].
Algorithm 1: Singular Spectrum Analysis (SSA).
    Input
Time series Y = y 1 , , y N , N > 2 , embedding window length L, and number of eigentriples r
    Output
Underlying components of the time series
    1.
Embedding: Construct the trajectory matrix X by taking time-lagged vectors of length L from the time series.
    2.
Singular value decomposition (SVD): Compute the SVD of the trajectory matrix X to obtain the singular vectors and singular values.
    3.
Grouping: Select the first r eigentriples based on the characteristics of the singular values.
    4.
Reconstruction: Reconstruct the underlying components of the time series by multiplying the retained eigentriples with the appropriate columns of the trajectory matrix, and summing across these products.

R-SSA Forecasting Algorithm

Forecasting with SSA is applicable to time series that approximately satisfy a linear recurrent relation (LRR). The general process for forecasting using the SSA method is outlined by Algorithm 2, also described by Golyandina et al. [24].
Algorithm 2: Recurrent Forecasting in Singular Spectrum Analysis (SSA).
    Input:
Time series Y = y 1 , , y N , N > 2 , Window Length L , 1 < L < N , Linear space L r R L of dimension r < L . {It is assumed that e L L R where e L = ( 0 , 0 , , 0 , 1 ) T R L , in other terms, L r is not a ‘vertical’ space.}
    Output:
Forecasts for the next h time steps
    1.
Construct the trajectory matrix X = X 1 , , X K of the time series Y = y 1 , , y N .
    2.
Compute the singular value decomposition (SVD) of X to obtain the orthonormal basis vectors U i ( i = 1 , , r ) for the subspace L r .
    3.
Perform the orthogonal projection step by computing the matrix X ^ = X ^ 1 , , X ^ K = i = 1 r U i U i T X . The vector X ^ i is the orthogonal projection of X i onto the subspace L r .
    4.
Construct the matrix X ˜ = H X ^ = X ˜ 1 : : X ˜ K , which is referred to as the Hankelization step.
    5.
Set ν 2 = π 1 2 + + π r 2 , where π i is the last component of the vector U i ( i = 1 , , r ) . Moreover, assume that e L L r . This implies that L r is not a vertical space. Therefore, ν 2 < 1 .
    6.
Determine the vector R = ( α L 1 , , α 1 ) T : R = 1 ( 1 ν 2 ) i = 1 r π i U i * , where U * R L 1 is the vector consisting of the first L 1 components of the vector U. Note that this does not depend on the choice of a basis U 1 , , U r in the linear space L r .
    7.
Define the time series Y N + h = ( y 1 , , y N + h ) using the following formula:
y i = { y i ˜ for i = 1 , , N j = 1 L 1 α j y i j for i = N + 1 , , N + h ,
where y i ˜ ( i = 1 , , N ) are the reconstructed series. The values y N + 1 , , y N + h are the h step-ahead recurrent forecasts.

3. Singular Spectrum Decomposition (SSD)

In this section, we will introduce the Singular Spectrum Decomposition (SSD) method and the related recurrent forecasting technique. The SSD method consists of a two-stage approach with two steps in each stage as follows:
  • Stage 1. Decomposition (Modified Embedding and SVD)
The proposed approach enhances the basic SSA method by using a modified trajectory matrix for a given time series Y = y 1 , , y N . The trajectory matrix is of size ( L × N ) , where L is the embedding dimension, and is denoted as X S S D . It can be expressed as
X S S D = y 1 y 2 y K y 2 y 3 y K + 1 y L y L + 1 y N y K + 1 y N y K + 2 y 1 y 1 y L 1 = X A .
Compared to the basic SSA method, the trajectory matrix in the SSD method includes an additional block A , which leads to the incorporation of different permutations of the total time series vector in each row of the modified trajectory matrix denoted as X S S D . Further details can be found in references [18,21].
  • Stage 2. Reconstruction (Grouping and Diagonal Averaging)
Similar to the grouping step in basic SSA (Section 2), a group of l eigentriples is selected in the SSD method. In the diagonal averaging step, a matrix denoted as X ˜ S S D is computed as an approximation of X S S D . This is achieved by computing the sum of l matrices, each obtained by taking the outer product of the corresponding eigenvectors. Mathematically, X ˜ S S D = k = 1 l U ˜ i k U ˜ i k T X S S D , where U ˜ i s are the corresponding eigenvectors. The transition to a one-dimensional time series can be achieved as follows:
i + j = { k + 1 and k + 1 + N when i + j < N k + 1 when i + j N .

3.1. Choice of the Embedding Dimension

The choice of the embedding dimension in Singular Spectrum Analysis (SSA) is crucial for accurately capturing the underlying structure of a time series. The embedding dimension determines the number of time-lagged vectors used to construct the trajectory matrix, affecting the amount of information retained in the decomposition. The embedding dimension should be chosen large enough to capture all relevant information, but not too large so as to include noise or irrelevant information, which can lead to an inaccurate decomposition and overfitting. A common rule of thumb for choosing the embedding dimension L in SSA is L N / 2 , see [25]. Furthermore, Vautard [26] proposed a criterion for determining the appropriate window length in Singular Spectrum Analysis (SSA) when analyzing time series with intermittent oscillations. According to this criterion, SSA can isolate intermittent oscillations correctly if the inverse of the maximum spectral density of the time series, denoted as f m a x , is less than or equal to the window length L. In other words, L should be chosen such that 1 / f m a x L . However, for time series with varying frequency domains, extracting oscillatory components using the SSA method can be challenging due to the need to set an appropriate window length at each step, while in the SSD method, the window length L is selected as a linear function of 1 / f m a x and should be less than N / 2 , where N is the length of the time series. This approach captures local structures in the time series while minimizing noise inclusion.

3.2. R-SSD Forecasting Algorithm

Let λ ˜ 1 , , λ ˜ L be the eigenvalues of X S S D X S S D T , and U ˜ 1 , , U ˜ L be the corresponding eigenvectors for the trajectory matrix X S S D . Then, the new R-SSD coefficients can be computed as R ˜ = ( α ˜ L 1 , , α ˜ 1 ) = 1 / ( 1 ν ˜ 2 ) i = 1 r π ˜ i U ˜ i , where U ˜ i is the vector consisting of the first L 1 components of the vector U ˜ i , π ˜ i is the last component of the vector U ˜ i and ν ˜ 2 = i = 1 r π ˜ i 2 . Finally, to obtain the forecasting algorithm of R-SSD, we replace the α j values in Equation (1) with α ˜ j values, where y i ˜ s ( i = 1 , , N ) are the reconstructed series obtained using the SSD method.

4. Empirical Results

We assess the performance of the R-SSA and R-SSD forecasting methods on real and simulated time series in this section. A portion of the data is used for training, while the remaining data are reserved for testing. We evaluate the accuracy of forecasting using the root mean squared error (RMSE) and mean absolute error (MAE) criteria and compare the results using the ratios defined in Equations (4) and (5).
RRMSE h = t = m m + n h ( y t + h y ^ t + h | t ) 2 t = m m + n h ( y t + h y ^ ^ t + h | t ) 2 ,
RMAE h = t = m m + n h | y t + h y ^ t + h | t | t = m m + n h | y t + h y ^ ^ t + h | t | ,
where the lengths of the training sample, test sample, and forecast horizon are denoted by m, n and h, respectively. On the other hand, y ^ t + h | t denote the h-step ahead forecast obtained via the new R-SSD forecasting method and y ^ ^ t + h | t denote the h-step ahead forecast obtained via the R-SSA forecasting method. If the ratio of the average RMSE values obtained by R-SSD and R-SSA, denoted as R R M S E , is less than 1 at a given forecasting horizon h, denoted as R R M S E h < 1 , then the R-SSD procedure is more accurate than R-SSA at horizon h. Alternatively, when R R M S E h > 1 , it can be inferred that the accuracy of the R-SSD procedure is less than R-SSA. The same inference can be made using the ratio of the average MAE values obtained by R-SSD and R-SSA, denoted as R M A E . Additionally, to compare the accuracy of two sets of forecasts, the Kolmogorov–Smirnov Predictive Accuracy (KSPA) test is considered, as proposed in [27]. The KSPA test has two objectives: firstly, to determine if there is a significant statistical difference between the distribution of predictive errors by testing if the empirical cumulative distribution functions F S S D and F S S A for the forecast errors of the two methods are significantly different. The two-sided KSPA test evaluates this difference with H 0 : F S S D ( z ) = F S S A ( z ) and H 1 : F S S D ( z ) F S S A ( z ) . The second objective of the KSPA test is to determine if the method with the lowest error based on a given loss function also exhibits a statistically significantly smaller error than the corresponding method. The one-sided KSPA test is formulated as H 0 : F S S D ( z ) F S S A ( z ) and H 1 : F S S D ( z ) > F S S A ( z ) . Rejecting the null hypothesis indicates that the cumulative distribution function (c.d.f.) of forecast errors obtained from the SSD model is shifted toward the left and above the c.d.f. of forecast errors obtained from the SSA model, suggesting that the SSD method has a smaller stochastic error compared to the SSA method, for more details see [27].
In the following, two simulated time series with a length of 200 are generated, with the first 140 observations being designated as the training sample ( m = 140 ) and the remaining data as the test sample ( n = 60 ). The number of leading eigenvalues (r) for reconstructing and forecasting the time series is selected based on the rank of the corresponding trajectory matrix. This simulation is repeated 1000 times, and the mean of RRMSEs and RMAEs are calculated.

4.1. Simulated Examples

Example 1. 
In the first example, we examine a sine series that encompasses two distinct frequencies, as illustrated below:
y t = s i n ( 2 π t ) + ε t , 1 t 100 s i n ( 5 π t ) + ε t , 101 t 200
where the noise term ε t is generated from a normal distribution at varying levels of signal-to-noise ratio (SNR). In this example, both basic SSA and SSD methods are compared using a rank value of r = 5 for forecasting horizons of h = 1 , 3 , 6 , 12 , and 24 steps ahead. The R-SSD method outperforms the basic R-SSA method in terms of forecasting accuracy across all window lengths (L) and SNR levels tested, as shown in Figure 1 and Figure 2. For nearly all forecast horizons h, the values of RRMSE and RMAE are less than 1, indicating that the R-SSD method provides more accurate predictions than the R-SSA method. The accuracy is consistent across different metrics, with the lowest RRMSE and RMAE values occurring at the lowest window length level ( L = 6 ) for all SNR levels when h = 12 and 24. Overall, the results suggest that the R-SSD method is superior to the R-SSA method in providing accurate predictions.
Example 2. 
Example 2 involves an exponential series with two different frequencies as follows:
y t = e x p ( α 0 + α 1 t ) + c o s ( 2 π t / 6 ) + ε t , 1 t 100 e x p ( α 0 + α 1 t ) + c o s ( 5 π t / 6 ) + ε t , 101 t 200
where the term ε t represents the noise generated from a normal distribution at various levels of signal-to-noise ratio (SNR). In this study, both basic SSA and SSD methods use a rank of 25 for the trajectory matrix of the time series, with α 0 = 0 and α 1 = 0.01 . RRMSE and RMAE are computed for various forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) and SNR levels. As shown in Figure 3 and Figure 4, RRMSE and RMAE increase as the forecast horizon decreases, but decrease significantly when h = 24 . The results indicate that the R-SSD method performs better as the value of L decreases. However, for higher SNR levels, the R-SSA method outperforms the R-SSD method for larger values of L. The lowest RRMSE and RMAE values are achieved when the window length and SNR are at their lowest values.

4.2. Real Data Analysis

In this section, we compare the forecasting performance of the proposed R-SSD method with the basic R-SSA method using real data from fruit fly (Drosophila melanogaster) embryos. The caudal protein in fruit fly embryos plays a crucial role in tail formation, acting as a transcription factor that regulates the expression of other genes by binding to specific DNA sequences. The caudal protein is expressed in the cells of the “tail bud”, which gives rise to the tail, and its activation triggers a gene expression cascade that controls cell division, differentiation, and migration, ultimately leading to tail formation. Mutations in the caudal gene can result in a loss of function of the protein, leading to defects in tail formation such as a short or absent tail, as well as other developmental defects related to segmentation. However, it is important to note that causality detection techniques, as demonstrated by previous studies, can be sensitive to noise [23,28,29,30]. Here, we analyze four gene expression profiles with varying lengths and dominant frequencies to demonstrate the importance of utilizing an accurate noise filtering method, such as SSD, for conducting reliable causality studies. To compare the forecasting performance of the R-SSA and R-SSD approaches, we provide Table 1, Table 2, Table 3 and Table 4 to summarize the obtained results for four different time classes: ab2, ab18, be11, ad14. These tables display the respective forecasting metrics, including RRMSE and RMAE, for each time class, enabling a comprehensive comparison between the two methods. For each dataset, we considered the first 80% of observations as the training sample and the remaining 20% as the test sample. The number of leading eigenvalues ( r ) for reconstructing and forecasting the time series was selected based on the rank of the corresponding trajectory matrix. Additionally, the dominant frequency of the data ( 1 / f m a x ) was calculated for each dataset, and the window length was chosen as a multiple of 1 / f m a x and less than N / 2 . After selecting the appropriate L and r, we utilized the observations from the training set to forecast the test sample data and calculate the RRMSE and RMAE criteria for different h step-ahead recurrent forecasts, using Equations (4) and (5).
Based on the results presented in Table 1, it is evident that there is a discernible difference in the RRMSE and RMAE values obtained using the R-SSA and R-SSD methods for L = 20 , 30 , and 50. The performance metrics show contrasting outcomes for these window lengths, indicating that the choice of method can significantly impact the forecasting accuracy.
Table 2 shows the RRMSE and RMAE values obtained by each model for the cad profile ab18. As indicated in the table, the R-SSD method achieves a significant reduction in both RRMSE and RMAE values for L = 20 , which suggests that it generally provides better signal extraction and forecast results compared to the R-SSA model for this window length. Additionally, for L = 50 and 80, the accuracy of the two methods is similar, indicating that the R-SSD method can be preferable for smaller window lengths.
Table 3 summarizes the results of RRMSE and RMAE for the cad profile be11. The findings indicate that the R-SSD method outperforms the R-SSA method, particularly for L = 18 and 30 and horizons h = 12 and 24. Furthermore, a closer examination of the table reveals that the highest accuracy is obtained when L = 18 and h = 24 , as evidenced by the greatest reduction in both RRMSE and RMAE values.
Additionally, the forecasting methods R-SSD and R-SSA were evaluated for statistical significance using the non-parametric two-sample Wilcoxon test and the Kolmogorov–Smirnov Predictive Accuracy (KSPA) test. The results show a statistically significant difference between the two methods, with R-SSD forecasts having smaller errors than R-SSA forecasts with 95% confidence based on the one-sided KSPA test. The two-sided KSPA test further supports the significant differences between the two methods with 95% confidence, and these findings are consistent across different embryos and L values, especially for h = 24. These results demonstrate the superior accuracy of the R-SSD method and highlight the importance of utilizing an accurate noise filtering method such as SSD for precise causality studies. The Wilcoxon test also confirms the significant differences between the two methods for all tested embryos and L values, with p-values less than 0.05.

5. Discussion

In this paper, we introduced a new forecasting method called Recurrent Singular Spectrum Decomposition (R-SSD), which improves upon the standard R-SSA method by enhancing the identification of fluctuation content and enabling a fully data-driven selection of window length and principal components for reconstructing component series based on dominant frequency periods. The results were evaluated using the non-parametric two-sample Wilcoxon test and RRMSE/RMAE criteria, which demonstrated the superiority of R-SSD over basic R-SSA in the majority of cases for various window lengths and forecasting horizons. KSPA tests confirmed the ability of R-SSD to obtain significant components for accurate forecasting of new data points. In summary, the proposed R-SSD method with its improved trajectory matrix definition and window length selection shows promising results in time series forecasting. Overall, the R-SSD method offers a viable alternative to the standard R-SSA method and could lead to improved forecasting accuracy in a wide range of applications. Further research and investigation into the R-SSD method’s performance under different scenarios and datasets may be valuable for its continued development and potential adoption in practical settings.

Author Contributions

All authors contributed equally to the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available, upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Broomhead, D.S.; King, G. Extracting qualitative dynamics from experimental data. Phys. D Nonlinear Phenom. 1989, 20, 217–236. [Google Scholar] [CrossRef]
  2. Hassani, H.; Ghodsi, Z.; Silva, E.; Heravi, S. From nature to maths: Improving forecasting performance in subspace-based methods using genetics Colonial Theory. Digit. Signal Process. 2016, 51, 101–109. [Google Scholar] [CrossRef] [Green Version]
  3. Gong, Y.; Song, Z.; He, M.; Gong, W.; Ren, F. Precursory waves and eigenfrequencies identified from acoustic emission data based on Singular Spectrum Analysis and laboratory rock-burst experiments. Int. J. Rock Mech. Min. Sci. 2017, 91, 155–169. [Google Scholar] [CrossRef]
  4. Yu, C.; Li, Y.; Zhang, M. An improved Wavelet Transform using Singular Spectrum Analysis for wind speed forecasting based on Elman Neural Network. Energy Convers. Manag. 2017, 148, 895–904. [Google Scholar] [CrossRef]
  5. Rahman Khan, M.R.; Poskitt, D.S. Forecasting stochastic processes using singular spectrum analysis: Aspects of the theory and application. Int. J. Forecast. 2017, 33, 199–213. [Google Scholar] [CrossRef]
  6. Heravi, S.; Osborn, D.R.; Birchenhall, C.R. Linear versus neural network forecasts for European industrial production series. Int. J. Forecast. 2004, 20, 435–446. [Google Scholar] [CrossRef]
  7. Lai, L.; Guo, K. The performance of one belt and one road exchange rate: Based on improved singular spectrum analysis. Phys. A Stat. Mech. Its Appl. 2017, 483, 299–308. [Google Scholar] [CrossRef]
  8. Hassani, H.; Yeganegi, M.; Khan, A.; Silva, E. The Effect of Data Transformation on Singular Spectrum Analysis for Forecasting. Signals 2020, 1, 4–25. [Google Scholar] [CrossRef]
  9. Hassani, H.; Silva, E.S.; Antonakakis, N.; Filis, G.; Gupta, R. Forecasting accuracy evaluation of tourist arrivals. Ann. Tour. Res. 2017, 63, 112–127. [Google Scholar] [CrossRef]
  10. Movahedifar, M.; Hassani, H.; Yarmohammadi, M.; Kalantari, M.; Gupta, R. A robust approach for outlier imputation: Singular spectrum decomposition. Commun. Stat. Case Stud. Data Anal. Appl. 2021, 8, 234–250. [Google Scholar] [CrossRef]
  11. Chao, S.; Loh, C. Application of singular spectrum analysis to structural monitoring and damage diagnosis of bridges. Struct. Infrastruct. Eng. 2014, 10, 708–727. [Google Scholar] [CrossRef]
  12. Chen, Q.; van Dam, T.; Sneeuw, N.; Collilieux, X.; Weigelt, M.; Rebischung, P. Singular spectrum analysis for modeling seasonal signals from GPS time series. J. Geodyn. 2013, 72, 25–35. [Google Scholar] [CrossRef]
  13. Hassani, H.; Webster, A.; Silva, E.; Heravi, S. Forecasting U.S. Tourist arrivals using optimal Singular Spectrum Analysis. Tour. Manag. 2015, 46, 322–335. [Google Scholar] [CrossRef]
  14. Hutny, A.; Warzecha, M.; Derda, W.; Wieczorek, P. Segregation of Elements in Continuous Cast Carbon Steel Billets Designated for Long Products. Arch. Metall. Mater. 2016, 61, 2037–2042. [Google Scholar] [CrossRef] [Green Version]
  15. Liu, K.; Law, S.; Xia, Y.; Zhu, X.Q. Singular spectrum analysis for enhancing the sensitivity in structural damage detection. J. Sound Vib. 2014, 333, 392–417. [Google Scholar] [CrossRef]
  16. Muruganatham, B.; Sanjith, M.A.; Kumar, B.; Murty, S.A.V.; Swaminathan, P. Roller element bearing fault diagnosis using singular spectrum analysis. Mech. Syst. Signal Process. 2013, 35, 150–166. [Google Scholar] [CrossRef]
  17. Sanei, S.; Hassani, H. Singular Spectrum Analysis of Biomedical Signals; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  18. Movahedifar, M.; Yarmohammadi, M.; Hassani, H. Bicoid signal extraction: Another powerful approach. Math. Biosci. 2018, 303, 52–61. [Google Scholar] [CrossRef]
  19. Hiemstra, C.; Jones, J.D. Testing for Linear and Nonlinear Granger Causality in the Stock Price- Volume Relation. J. Financ. 1994, 49, 1639–1664. [Google Scholar]
  20. Ancona, N.; Marinazzo, D.; Stramaglia, S. Radial basis function approach to nonlinear Granger causality of time series. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 2004, 70, 056221. [Google Scholar] [CrossRef] [Green Version]
  21. Bonizzi, P.; Karel, J.; Meste, O.; Peeters, R. Singular Spectrum Decomposition: A new method for time series decomposition. Adv. Adapt. Data Anal. 2014, 6, 1450011. [Google Scholar] [CrossRef]
  22. Hassani, H. Singular Spectrum Analysis: Methodology and Comparison. J. Data Sci. 2007, 5, 239–257. [Google Scholar] [CrossRef] [PubMed]
  23. Golyandina, N.; Korobeynikov, A.; Zhigljavsky, A. Singular Spectrum Analysis with R; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  24. Golyandina, N.; Nekrutkin, V.; Zhigljavsky, A.A. Analysis of Time Series Structure: SSA and Related Techniques, 1st ed.; Chapman and Hall/CRC: Boca Raton, FL, USA, 2001. [Google Scholar] [CrossRef]
  25. Golyandina, N.; Zhigljavsky, A. Singular Spectrum Analysis for Time Series; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  26. Vautard, R.; Yiou, P.; Ghil, M. Singular-spectrum analysis: A toolkit for short, noisy chaotic signals. Phys. D Nonlinear Phenom. 1992, 58, 95–126. [Google Scholar] [CrossRef]
  27. Hassani, H.; Silva, E. A Kolmogorov–Smirnov Based Test for Comparing the Predictive Accuracy of Two Sets of Forecasts. Econometrics 2015, 3, 590–609. [Google Scholar] [CrossRef] [Green Version]
  28. Vautard, R.; Ghil, M. Singular spectrum analysis in nonlinear dynamics, with applications to paleoclimatic time series. Phys. D Nonlinear Phenom. 1989, 35, 395–424. [Google Scholar] [CrossRef]
  29. Zou, C.; Feng, J. Granger causality vs. dynamic Bayesian network inference: A comparative study. BMC Bioinform. 2009, 10, 122. [Google Scholar]
  30. Golyandina, N.E.; Holloway, D.M.; Lopes, F.J.; Spirov, A.V.; Spirova, E.N.; Usevich, K.D. Measuring gene expression noise in early Drosophila embryos: Nucleus-to-nucleus variability. Procedia Comput. Sci. 2012, 9, 373–382. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The RRMSE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 1.
Figure 1. The RRMSE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 1.
Engproc 39 00068 g001
Figure 2. The RMAE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 1.
Figure 2. The RMAE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 1.
Engproc 39 00068 g002
Figure 3. The RRMSE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 2.
Figure 3. The RRMSE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 2.
Engproc 39 00068 g003
Figure 4. The RMAE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 2.
Figure 4. The RMAE results for different forecast horizons ( h = 1 , 3 , 6 , 12 , and 24) in Example 2.
Engproc 39 00068 g004
Table 1. RRMSE and RMAE analysis of Cad Profile ab2, 1 / f m a x = 2 with r = 16 .
Table 1. RRMSE and RMAE analysis of Cad Profile ab2, 1 / f m a x = 2 with r = 16 .
HorizonL = 20L = 30L = 50
RRMSERMAERRMSERMAERRMSERMAE
10.960.961.021.021.121.12
30.860.911.031.051.131.16
60.780.800.961.021.091.19
120.590.620.860.921.111.22
240.170.210.590.671.001.09
Table 2. RRMSE and RMAE analysis for Cad profile ab18, 1 / f m a x = 2 , with r = 14 .
Table 2. RRMSE and RMAE analysis for Cad profile ab18, 1 / f m a x = 2 , with r = 14 .
HorizonL = 20L = 50L = 80
RRMSERMAERRMSERMAERRMSERMAE
10.820.821.271.271.141.14
30.880.861.151.211.011.07
60.910.921.171.261.031.10
120.900.911.231.311.081.17
240.840.861.251.361.161.31
Table 3. RRMSE and RMAE analysis of Cad Profile be11, 1 / f m a x = 3 with 3.
Table 3. RRMSE and RMAE analysis of Cad Profile be11, 1 / f m a x = 3 with 3.
HorizonL = 6L = 18L = 30
RRMSERMAERRMSERMAERRMSERMAE
11.161.161.211.211.151.15
31.161.171.071.111.131.17
61.221.201.051.091.111.16
121.351.380.800.850.950.98
241.892.030.210.250.430.44
Table 4. RRMSE and RMAE analysis of Cad Profile ad14, 1 / f m a x = 11 with 5.
Table 4. RRMSE and RMAE analysis of Cad Profile ad14, 1 / f m a x = 11 with 5.
HorizonL = 11L = 88L = 110
RRMSERMAERRMSERMAERRMSERMAE
11.261.261.281.281.151.15
31.221.301.231.301.171.17
61.121.251.271.331.161.20
120.961.031.371.431.311.32
240.730.781.501.591.471.53
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Movahedifar, M.; Hassani, H.; Kalantari, M. Recurrent Forecasting in Singular Spectrum Decomposition. Eng. Proc. 2023, 39, 68. https://doi.org/10.3390/engproc2023039068

AMA Style

Movahedifar M, Hassani H, Kalantari M. Recurrent Forecasting in Singular Spectrum Decomposition. Engineering Proceedings. 2023; 39(1):68. https://doi.org/10.3390/engproc2023039068

Chicago/Turabian Style

Movahedifar, Maryam, Hossein Hassani, and Mahdi Kalantari. 2023. "Recurrent Forecasting in Singular Spectrum Decomposition" Engineering Proceedings 39, no. 1: 68. https://doi.org/10.3390/engproc2023039068

Article Metrics

Back to TopTop