Next Article in Journal
Entropy-Aware Time-Varying Graph Neural Networks with Generalized Temporal Hawkes Process: Dynamic Link Prediction in the Presence of Node Addition and Deletion
Previous Article in Journal
Optimal Topology of Vision Transformer for Real-Time Video Action Recognition in an End-To-End Cloud Solution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks

by
Cristian Ubal
1,
Gustavo Di-Giorgi
2,
Javier E. Contreras-Reyes
1 and
Rodrigo Salas
3,4,*
1
Instituto de Estadística, Facultad de Ciencias, Universidad de Valparaíso, Valparaíso 2360102, Chile
2
Escuela de Administración Pública, Facultad de Ciencias Económicas y Administrativas, Universidad de Valparaíso, Valparaíso 2362797, Chile
3
Escuela de Ingeniería C. Biomédica, Facultad de Ingeniería, Universidad de Valparaíso, Valparaíso 2362905, Chile
4
Millennium Institute for Intelligent Healthcare Engineering (iHealth), Santiago 7820436, Chile
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2023, 5(4), 1340-1358; https://doi.org/10.3390/make5040068
Submission received: 24 July 2023 / Revised: 15 September 2023 / Accepted: 21 September 2023 / Published: 2 October 2023

Abstract

:
Long-term dependence is an essential feature for the predictability of time series. Estimating the parameter that describes long memory is essential to describing the behavior of time series models. However, most long memory estimation methods assume that this parameter has a constant value throughout the time series, and do not consider that the parameter may change over time. In this work, we propose an automated methodology that combines the estimation methodologies of the fractional differentiation parameter (and/or Hurst parameter) with its application to Recurrent Neural Networks (RNNs) in order for said networks to learn and predict long memory dependencies from information obtained in nonlinear time series. The proposal combines three methods that allow for better approximation in the prediction of the values of the parameters for each one of the windows obtained, using Recurrent Neural Networks as an adaptive method to learn and predict the dependencies of long memory in Time Series. For the RNNs, we have evaluated four different architectures: the Simple RNN, LSTM, the BiLSTM, and the GRU. These models are built from blocks with gates controlling the cell state and memory. We have evaluated the proposed approach using both synthetic and real-world data sets. We have simulated ARFIMA models for the synthetic data to generate several time series by varying the fractional differentiation parameter. We have evaluated the proposed approach using synthetic and real datasets using Whittle’s estimates of the Hurst parameter classically obtained in each window. We have simulated ARFIMA models in such a way that the synthetic data generate several time series by varying the fractional differentiation parameter. The real-world IPSA stock option index and Tree Ringtime series datasets were evaluated. All of the results show that the proposed approach can predict the Hurst exponent with good performance by selecting the optimal window size and overlap change.

1. Introduction

Time series analysis and forecasting are essential in many areas of application, such as finance and marketing [1], air pollution [2], electricity consumption [3], and weather forecasting [4,5], among others. However, selecting the appropriate model strongly depends on the degree of predictability of the time series [6].
Long-term dependencies play an essential role in time series forecasting because they are an inherent property of the degree of predictability of the observable time series [7,8]. It is necessary to infer the predictability level of the process based on memory and fractal characteristics in order to select the appropriate forecasting model. However, learning long-range dependencies embedded in time series is an obstacle for most algorithms [6]. Estimating the parameter that describes long memory is an essential part of describing the behavior of time series models. Moreover, most long memory estimation methods assume that this parameter has a constant value throughout the time series, and do not consider that the parameter may change over time.
The study of the relationship between Artificial Neural Networks and long memory time series was carried out by Siriopoulos et al. [9], where the authors studied the application of the multilayer perceptron in modeling the stock exchanges indexes. Lin et al. [10] used Recurrent Neural Networks to deal with the problem of learning long-term dependencies in Nonlinear Autoregressive models with eXogenous inputs (NARX models). Ledesma et al. [11] proposed a method for estimating the Hurst parameter using Artificial Neural Networks, where the experimental results show that this method outperforms traditional methods and can be used in applications such as traffic control in computer networks. Menezes et al. [12] proposed applying a feedforward time delay neural network (TDNN) as a NARX model for long-term prediction of univariate time series. Hua et al. [13] introduced a random connectivity LSTM model for predicting the dynamics of traffic and user locations through various temporal scales. Kovantsev et al. [6] proposed clustering the time series based on statistical indices such as entropy, correlation dimension, and the Hurst exponent in order to test their predictability. Li et al. [14] proposed a new time series classification model using long-term memory and convolutional neural networks (LCNN). The Hurst exponent was used to measure the long-term dependency of time series, and LCNN was found to improve classification performance and to be suitable for small datasets. Recently, Di-Giorgi et al. [1] proposed the application of deep recurrent neural networks for volatility forecasting as GARCH models. However, the works mentioned above have not addressed the need to create a method that can accurately and efficiently estimate and predict the time-varying long memory index of the time series.
In general, the sample autocorrelation function (ACF) is used in the literature to identify long memory processes. However, as was empirically demonstrated by Hassani et al. [15], it is not possible to determine long memory by summing the sample ACFs. They suggested alternative methods for detecting long-range dependence. In this sense, in this work we explore wavelet-based methods and fractional integration techniques in combination with neural networks containing both states and memory.
We propose an adaptive method that combines recurrent neural networks with statistical methods to learn the dependencies of long memory in time series and to predict the fractional differentiation parameter (and/or Hurst parameter) based on a moving window of the original time series. For the RNNs, we have evaluated four different architectures: the Simple RNN, LSTM, the BiLSTM, and the GRU. These models are built from blocks with gates controlling the cell state and memory. The rest of this work has the following structure: in Section 2, we briefly describe the basic theories of long memory processes and recurrent neural networks; in Section 3, the proposed approach is explained; and finally, Section 4 and Section 5 respectively present our results and concluding remarks.

2. Theoretical Framework

In a stationary time series, long-term dependence implies a non-negligible dependence between the current and all past points. The characteristics of long memory parameters are difficult to estimate, and even more so if the probability model evolves. Therefore, it is necessary to construct an adaptive method for their estimation. The Hurst exponent is an index of paramount importance in the analysis of the long-range dependence features of observable time series [16]. For instance, time series with a large Hurst exponent have a strong trend, making them more predictable than time series with a Hurst exponent closer to random noise.
Several statistical methods for long-term dependency estimation have been proposed in the literature. The oldest and most well known is the so-called re-scaled range analysis (R/S) described by Hurst [16] and popularized by Mandelbrot et al. [17], in which the Fractional Brownian Motion (FBM) and Fractional Gaussian Noise (FGN) are derived with their properties and representations. Alternative estimators include downward fluctuation analysis (DFA), proposed by Peng et al. [18], which was introduced in the study of the mosaic organization of DNA nucleotides. Geweke et al. [19] proposed a simple linear regression of the log-periodogram consisting of an ordinary least squares estimator of the parameter formed using only the lowest frequency ordinates of the logarithmic periodogram. The estimator proposed by Whittle [20] is based on the periodogram using the Fast Fourier Transform (FFT). Veitch et al. [21] proposed the wavelet estimation method based on the coefficients of discrete wavelet decomposition. Moreover, Taqqu et al. [22] studied several long-range dependency parameter estimators for Fractional Gaussian Noise.
In the following subsections, we introduce several fundamental concepts required to understand the basics of long dependency in stochastic processes. In addition, we review a number of the most widely used methods to estimate the fractional parameter.

2.1. ARFIMA Model for Long Memory Processes

Autoregressive Fractionally Integrated Moving Average (ARFIMA) models are used to model time series data that exhibit long memory or fractional integration, meaning that the autocorrelation of the series declines very slowly. These models extend the ARIMA models by incorporating a fractional differencing parameter, which allows them to capture the long memory effect. ARFIMA models are particularly useful in modeling and forecasting financial and economic time series with long memory, such as stock prices, exchange rates, and interest rates, and are used in option pricing and volatility forecasting as well. However, estimating these models can be computationally intensive, and interpreting their parameters can be challenging [23].
A stochastic process { Y t } follows an ARFIMA ( p , d , q ) process, where p and q are integers and d is a real number, if { Y t } can be represented as follows:
ϕ ( B ) ( 1 B ) d Y t = θ ( B ) ε t , ε t W N ( 0 , σ ϵ 2 ) ,
where ϕ ( B ) = 1 i = 1 p ϕ i B i and θ ( B ) = 1 + i = 1 q θ i B i are the polynomials of the autoregressive and moving average operators, respectively. These polynomials do not have roots in common.
The spectral density of the ARFIMA process is provided by
f ( λ ) = | 1 e i λ | 2 d σ ϵ 2 2 π | θ ( e i λ ) | 2 | ϕ ( e i λ ) | 2
where | 1 e i λ | = 2 sin ( λ 2 ) and i denotes the imaginary unit. Hosking [24] described the fractionally differentiated process (FN(d)) with polynomials ϕ ( B ) = θ ( B ) = 1 and with the spectral density provided by
f ( λ ) σ ε 2 2 π | 1 e i λ | 2 d .
Thus, the spectral density has a pole at 0 for d > 0 , leading to d = H 1 2 , thereby finding the relationship between the fractional differentiation parameter d and the Hurst exponent H.
As demonstrated in Hassani’s ½-theorem, it is important to note that the sum of the sample ACF is always negative one-half for any stationary time series with any length. For this reason, relying solely on the sample ACF to identify long memory processes can be misleading.

2.2. Long Memory Parameter Estimation Methods

2.2.1. Periodogram Regression Method

Under the assumption that the spectral density of a stationary process can be written as
f ( λ ) = f 0 ( λ ) ( 2 sin ( λ / 2 ) ) 2 d ,
where f 0 ( λ ) = 2 π σ 2 f y ( λ ) | λ | 2 d is a continuous function with f y as the strictly positive spectral density of { y t } , Geweke et al. [19] proposed a regression method for estimating the parameters; by defining y j = log ( I ( λ j ) ) , α = log ( f 0 ( 0 ) ) , β = d , x j = log ( [ 2 sin ( λ j / 2 ) ] 2 ) , and
ε j = log I ( λ j ) [ 2 sin ( λ / 2 ) ] 2 d f 0 ( 0 ) ,
the regression equation is obtained as
y j = α + β x j + ε j .
The least squares estimator of the long memory parameter d is provided by
d ^ m = j = 1 m ( x j x ¯ ) ( y j y ¯ ) j = 1 m ( x j x ¯ ) 2 ,
where x ¯ = 1 m j = 1 m x j and y ¯ = 1 m j = 1 m y j .

2.2.2. Whittle Estimator Method

Whittle’s estimator [20] is based on the periodogram. This involves the following equation:
Q ( η ) = π π I ( λ ) f ( λ , η ) d λ + π π log ( f ( λ , η ) ) d λ ,
where f ( λ , η ) is the spectral density at the frequency λ , η is the vectors of the unknown parameters, and I ( λ ) is the periodogram, defined here as
I ( λ ) = 1 2 π n j = 1 n Y j exp ( i j λ ) 2 .
The second term in Equation (8) can be set to equal to 0 by renormalizing f ( λ , η ) . The normalization only depends on a scale parameter, not on the rest of the components of η ; thus, we replace f with f * such that f * = β f and π π log ( f * ( λ , η ) ) d λ = 0. Because I ( λ ) is an estimator of the spectral density, a series with long-range dependence should have a periodogram which is proportional to | λ | 1 2 H at the origin. Whittle’s estimator is the value of η that minimizes the Q function. In actual application, instead of an integral, the corresponding sum over the Fourier frequencies λ j = 2 π j / n is computed, where j = 1 , 2 , ( n 1 ) / 2 and n is the length of the series. Thus, the actual function which the algorithm minimizes is
Q * ( η ) = j = 1 ( n 1 ) / 2 I ( λ ) f * ( λ j , η ) .
If { Y t } is fractional Gaussian noise, then η is the parameter H or d. If { Y t } follows an ARFIMA ( p , d , q ) process, η includes the unknown coefficients of the autoregressive and moving average parts of that model. This estimator assumes that the parametric form of the spectral density is known. For more details, see [25].

2.2.3. Detrended Fluctuation Analysis

Detrended fluctuation analysis (DFA) was introduced by Peng et al. [18], and proceeds as follows. Let { y 1 , y 2 , , y n } be a sample of a stationary process with a long memory and let x t = j = 1 t y j for t = 1 , , n . A sample { y 1 , y 2 , , y n } is divided into k blocks without overlap which contain m = n / k observations. A linear regression model of x t versus t within each block is fitted. Let σ k 2 be the estimated residual variance of the regression within the block, and let k be the dual variance of the regression within block k:
σ k 2 = 1 m t = 1 m ( x t α ^ k β ^ k t ) 2
where α ^ k and β ^ k are the least squares estimators of the intercept and slope of the regression line, respectively. Furthermore, let F 2 ( k ) be the average of these variances:
F 2 ( k ) = 1 k j = 1 k σ j 2 .
For a random walk, the last term behaves as F ( k ) c k 1 / 2 , while for a time series with long dependence we have F ( k ) c k d + 1 / 2 . Thus, an estimator of d can be obtained as d ^ = β ^ 1 / 2 by applying the least squares estimator to log ( F ( k ) ) = α + β log ( k ) + ε k .

2.2.4. Rescaled Range Method

Let { y 1 , y 2 , , y n } be a sample of a stationary long memory process, let x t = j = 1 t y j for t = 1 , , n , and let s n 2 = 1 ( n 1 ) t = 1 n ( y t y ¯ ) 2 be the sample variance, where y ¯ = x n / n . The rescaled range statistic introduced by Hurst [16] is defined by
R n = 1 s n max 1 t n x t t n x n min 1 t n x t t n x n .

2.2.5. Wavelet-Based Method

A real-value integrable function ψ ( t ) is defined as a wavelet if it satisfies ψ ( t ) d t = 0 . The family of dilations and translations of the wavelet function ψ is defined by
ψ j k ( t ) = 2 j / 2 ψ ( 2 j t k ) , j , k Z .
Here, the terms j and 2 j are called the octaves and scale, respectively. With this, we can define the discrete wavelet transform (DWT) of a process { y ( t ) } as
d j k = y ( t ) ψ j k ( t ) d t , j , k Z .
Moreover, the family { ψ j k ( t ) } forms an orthogonal basis, and the representation of the process { y ( t ) } is
y ( t ) = j = k = d j k ψ j k ( t ) .
Now, we define the statistic
μ ^ j = 1 n j k = 1 n j d ^ j k 2 ,
where n j is the number of coefficients in octave j available to be calculated. Veitch et al. [21] demonstrated that
μ ^ j z j n j χ n j ,
where z j = 2 2 d j c , c > 0 , and χ n j is a chi-square random variable with n j degrees of freedom.
The heteroscedastic regression model can be written as
y j = α + β x j + ε j ,
where ε j = log 2 ( log ( χ n j ) ) log 2 ( n j ) ψ ( n j / 2 ) + log ( n j / 2 ) with y j = log 2 ( μ ^ j ) ψ ( n j / 2 ) + log ( n j / 2 ) , α = log ( c ) , and β = 2 d . Therefore, when the estimator β ^ is obtained, an estimate for the long memory parameter d is provided by d ^ = β ^ / 2 . Furthermore, it follows that V a r ( d ^ ) = V a r ( β ^ ) / 4 .

2.3. Recurrent Artificial Neural Networks

Artificial Neural Network (ANN) models consist of layers of nonlinear processing units called neurons that are linked to each other by weighted connections. These models use the backpropagation algorithm to learn from data by fitting the weights of the connections between the neurons [26]. ANNs are highly parameterized nonlinear models capable of learning from data. Moreover, they are universal function approximations that learn from data (see Cybenko [27] and Hornik et al. [28]), and have been successfully applied in time series forecasting (for instance, see [2]). Specifically, ANN models outperform standard linear techniques when the time series is noisy and the underlying dynamical system is nonlinear [12].
Deep Recurrent Neural Networks (RNN) are a subclass of Artificial Neural Networks (ANN) in which the processing units, or neurons, may be grouped either in layers or blocks connected to the following units (feedforward connections) or to previous units (feedback or recurrent connections). These feedback connections introduce memory to the model structure. By using these recurrent connections, historical inputs can be “memorized” by the RNN and subsequently influence the network’s output. The “memory” that RNNs possess allows them to outperform feedforward neural networks (FNN) in many real-world applications.
Recurrent Neural Networks (RNN) are ANNs with at least one recurrent connection, and are capable of learning features and long-term dependencies from sequential and time series data [29]. Moreover, RNNs are universal approximations [30]. Hochreiter [31] introduced the Long Short-Term Memory Network (LSTM), an RNN consisting of memory cells and gate units. LSTMs address the vanishing gradient problem [32,33]. LSTMs can learn long-term dependencies, and are well known for working with sequential data. Later, Graves et al. [34] proposed the Bidirectional LSTM (BiLSTM) network, which consists of two LSTMs, the first taking the input in a forward direction and the second in a backward direction. Cho et al. [35] proposed a simplified version of the LSTM called Gated Recurrent Unit (GRU), which lacks an output gate.
Figure 1 shows the structure of a simple RNN [36], where x t is the mini-batch input of the t-th time step in the sequence and h t = f σ ( W i x t + W i h t 1 + b r ) is the hidden variable of the time step t, which is determined by both the input of the current time step and the hidden variable of the previous time step. The RNN stores the hidden variable h t 1 for the previous time step and introduces a new weight parameter W h to describe how the hidden variable of the previous time step is used for the current time step. An RNN can be understand as multiple replications of the same network; during each replication, a state is transferred to the next layer, and the hidden variables can be used to capture the historical information of the sequence up to the current time step. This means that the neural network is able to memorize information. The calculation formula for the output layer is as follows:
o t = f σ ( W o h t + W i + b o ) .
The parameters of the RNN include the hidden layer weights W i and W h , the hidden layer bias b r , the output layer weight W o , and the output layer bias b o .
A Long Short-Term Neural Network (LSTM) is a variant of the Recurrent Neural Network (RNN) proposed by Hochreiter et al. [31]. An LSTM has a similar basic structure to an RNN, except that memory blocks replace neurons. Each memory block contains three nonlinear units called gates. The input gate i t , the output gate o t , and the forget gate f t control the information in the network. The memory of the cell is controlled by the hidden state h t and the cell state c t . Figure 2 shows the diagram of the LSTM block.
At the time t, the input vector x t R d flows forward in the LSTM cell, where the operations formula are:
Input   Gate : i t = f σ ( W i x t + U i h t 1 + b i )
Forget   Gate : f t = f σ ( W f x t + U f h t 1 + b f )
Output   Gate : o t = f σ ( W o x t + U o h t 1 + b o )
Hidden   State : h t = o t tanh ( c t )
Cell   State : c t = f t c t 1 + i t tanh ( W c x t + U c h t 1 + b c )
where W i , W f , W o , W c U i , U f , U o , U c correspond to the weight matrices and b i , b f , b o , b c are the bias vectors. The initial values are c 0 = 0 and h 0 = 0 . The activation functions are the sigmoid function f σ ( z ) = 1 1 + e z and the hyperbolic tangent function tanh ( z ) = e z e z e z + e z . The operator ⊙ denotes the Hadamard product.
The Bidirectional LSTM Network (BiLSTM) proposed by Graves et al. [34] is a sequence processing model that consists of an LSTM with two hidden states that allows the information to flow both forward and backward. Figure 3 shows the architecture of the BiLSTM. After processing each time step t, the BiLSTM network generates two hidden states h t F and h t B .
A Gated Recurrent Unit (GRU) is a simpler version of the LSTM network proposed by Cho et al. [35]. Figure 4 shows the architecture of the GRU. The input vector x t is introduced to the network, passing through both the update gate z t and the reset gate z t . On the one hand, the update gate decides how the input x t and the previous output h t flow to the next cell. On the other hand, the reset gate determines how much past information can be forgotten. The equations that control the functionality of the GRU are:
Update   Gate : z t = f σ ( W z x t + U z h t 1 + b z )
Reset   Gate : r t = f σ ( W r x t + U r h t 1 + b r )
Hidden   State : h ˜ t = tanh ( W h x t + U h ( r t h t 1 ) + b h )
Output   state : h t = ( 1 z t ) h t 1 + z t h ˜ t
where W z , W r , W h , U z , U r , U h correspond to the weight matrices and b z , b r , b h , are the bias vectors.

3. Materials and Methods

3.1. Dataset Description

The synthetic data considered in this work for training the networks were artificially constructed and of size n = 10,000 , which can be seen in Figure 5. These data consisted of fractional noise (FN) obtained with different fractional differentiation parameters d.
The real data that were used to train the networks consisted of: (1) a percentage variation of the IPSA (used in the Santiago Stock Exchange as the primary utility index to measure the profitability of the leading forty stocks that are moving in the economy; these incorporate all the capital changes of each share, weighting the relative weight of each for the calculation) in the period 2000–2021, as can be seen in Figure 6a; and (2) tree-ring widths in dimensionless units recorded by Donald A. Graybill, 1980, from Gt Basin Bristlecone Pine 2805M, 3726-11810 in Methuselah Walk, California, as plotted in Figure 6b. In particular, this time series has been analyzed in several studies, for example in [37].

3.2. Methodology

The purpose of this work is to propose an automated methodology that combines the estimation methodologies of the fractional differentiation parameter (and/or Hurst parameter) with its application to recurrent neural networks such that the network learns and predicts long memory information dependencies obtained in nonlinear time series. The information that is entered in the RNN is obtained from the previously estimated parameter using the Whittle method (chosen from among other estimation methods from a previous study) given data windows of optimal size.
The proposal combines three methods that allow for better approximation in the prediction of the values of the parameters for each one of the windows obtained. The proposed approach makes predictions of the Hurst exponent values by learning the estimate obtained using conventional methods applied to a moving time window.
The scheme of this methodology is shown in Figure 7.
The proposed methodology consists of three main steps. The first step is to divide the sample { Y 1 , , Y T } into a number M of overlapping blocks of length N and with a shift of S such that T = S ( M 1 ) + N and the midpoint of the j-th block is obtained as t j = S ( j 1 ) + N / 2 with j = 1 , , M . When the data blocks are determined, the second step of the methodology consists of obtaining the Hurst exponent’s local estimates over each block using the Whittle estimation method. The idea of segmenting the time series is provided by Palma et al. [38], among others. The methodology for approximating the MLE is based on the calculation of the periodogram I ( λ ) by means of the fast Fourier transform (FFT), e.g., [39], and the use of the approximation of the Gaussian log-likelihood function follows Whittle [40] and Bisaglia [41]. Suppose that the sample vector Y = ( y 1 , y 2 , , y n ) is normally distributed with zero mean and with the autocovariance provided by
γ ( k j ) = π π f ( λ ) e i λ ( k j ) d λ ,
where f ( λ ) is defined as in (2) and is associated with the parameter set Θ of the ARFIMA model defined in (1). The log-likelihood function of the process Y is provided by
L ( Θ ) = 1 2 n [ log | Δ | Y T Δ 1 Y ] ,
where Δ = [ γ ( k j ) ] with k , j = 1 , , n . For calculating (31), two asymptotic approximations are made for the terms log ( | Δ | ) and Y T Δ 1 Y to obtain
L ( Θ ) 1 4 π π π log [ 2 π f ( λ ) ] d λ + π π I ( λ ) f ( λ ) d λ ,
as n where I ( λ ) = | j = 1 n y j e i λ j | 2 / ( 2 π n ) is the periodogram indicated before. Thus, a discrete version of (32) is the Riemann approximation of the integral, and is
L ( Θ ) 1 2 n j = 1 n log [ f ( λ j ) ] + j = 1 n I ( λ j ) f ( λ j ) ,
where λ j = 2 π j / n are the Fourier frequencies. Now, to find the estimator of the parameter vector Θ , we use the minimization of L T ( Θ ) produced by the relation
Θ ^ = arg min L ( Θ ) ,
where the minimization is over a parameter space Θ . This nonlinear minimization function carries out a minimization of L ( Θ ) using a Newton-type algorithm. Under regularity conditions, the Whittle estimator that maximizes the log-likelihood function provided in (33) is consistent and distributed normally (e.g., [42]). This estimation method has been used in studies on local seasonal series (see [43,44]), and the use of this estimator in this work is justified later on in the subsequent comparative study.
The time series is separated into two segments, where the first partition, corresponding to 90% of the samples, is used for training and the final segment, corresponding to 10% of the samples, is used for testing. The recurrent neural networks are fitted with the training set, then the models are compared and validated on the test set. The RNNs are explained in the following subsection.
In order to obtain the predictions and measure their performance, we use measures such as the Root Mean Square Error (RMSE) and Coefficient of Determination R 2 .
The RMSE is often preferred over the MSE, as it is on the same scale as the data. Historically, both the RMSE and MSE have seen widespread use due to their theoretical relevance in statistical modeling. However, they are more sensitive to outliers than the MAE, which has led a number of authors, e.g., [45], to recommend their use in assessing forecasting accuracy.
A two-sample two-sided Kolmogorov–Smirnov (KSPA) test, as proposed by Hassani et al. [15], was applied to determine the existence (or not) of statistical significant differences in the distribution of forecasts between the two models with the best performances.
The pseudo-code of the proposed methodology is provided in Algorithm 1.
Algorithm 1 Predicting the Hurst parameter
1:
Define the block length N and the shift S.
2:
Segment the blocks of size N from the Times Series.
3:
for each block j = 1 to M do
4:
    Apply the Whittle method given by Equation (33) to obtain the value of the Hurst Exponent at time t j = S ( j 1 ) + N / 2 .
5:
end for
6:
Separate the blocks of the original Time Series into training and test sets.
7:
Separate the Hurst’s Time Series into training and test sets.
8:
Fit the RNN (simpleRNN, LSTM, BiLSTM or GRU) using the training sets. The inputs are the blocks of the original Time series and the targets are the Hurst’s time series.
9:
for each block in the Test set do
10:
    Predict the value of the Hurst parameter for the next block using the RNN.
11:
end for
12:
Obtain the performance metrics for the test set: Root Mean Square Error (RMSE) and the Coefficient of Determination R 2 .

4. Results

4.1. Comparative Study of Estimation Methods

In this section, a comparative study of the following estimation methods is carried out: the periodogram regression method, Whittle estimator method, detrended fluctuation analysis, rescaled range method, and wavelet-based method. In addition, a Monte Carlo simulation was performed using simulated time series of sizes n = 10,000 with specific values of the fractional differentiation parameter and 1000 simulations.
Figure 8 and Figure 9 show the degree of fit of the different estimation methods with the Hurst parameter and the fractional differentiation parameter.
From Figure 8, it can be seen that with n = 10,000 , the estimation methods of the Hurst exponent concerning the differentiation parameter d stabilize around the relationship obtained in the ARFIMA(d), as provided by H = d + 0.5 . It can be observed that at around d = 0.5 ( H = 0 ) the Whittle method fits better than the rest of the methods, which either deviate from the true value or have unstable behavior, such as in the GPH method. Around d = 0 ( H = 0.5 ), all of the models fit the relationship between said parameters well except for the GPH method. Finally, around d = 0.5 ( H = 1 ) the behavior of the methods becomes unstable except in the case of the Whittle method. From this, it can be concluded that Whittle’s method is the one that best fits this relationship in the interval of a good definition of the fractional differentiation parameter. This behavior occurs in the Whittle and wavelet-based methods, as they are better able to fit the relationship, as shown by Figure 9 for all the values of the fractional differential parameter. These methods show less dispersion in the Hurst parameter estimates; however, the Whittle method shows more stability, while the GPH method shows a greater range of dispersion in its estimates. In particular, little dispersion is observed in almost all methods except the GPH method when d < 0 , although they are claimed from the objective value. When d approaches 0, the R/S and DFA methods increase their dispersion even though it approach the target value, as the R/S, DFA, and GPH methods are already unstable in their dispersion when d > 0 . We use the Whittle method in what follows based on the previous results, which are consistent with those obtained by Palma et al. [38].

4.2. Hurst Parameter Prediction Using Recurrent Neural Networks

4.2.1. Synthetic Data

Table 1, Table 2 and Table 3 show the performance results of these models. We ran ten simulations for each model, then the results were averaged, which are the results appearing in the tables below. It can be seen that the performance metrics for the test set are better on the BiLSTM network. For the best two models, we applied the Shapiro–Wilk test to check the normal distribution of the performances and the pooled variance t-test to verify statistical differences. Moreover, the KSPA test was applied to verify the statistical significance of the observed difference in performance.
For d = 0.3 and d = 0 , the best results of the coefficient of determination R 2 were obtained for small window sizes, specifically, for N = 20 and N = 18 , respectively. For d = 0.3 , the best result was obtained for N = 30 . In addition, it can be observed that the MAE and RMSE coefficients decrease as block size increases. As the value of the fractional differentiation parameter, and consequently the Hurst exponent, increases within the interval [ 0.5 ; 0.5 ] , the value of R 2 improves, which indicates that a better prediction is obtained for non-negative values of that parameter.
The same conclusion can be drawn concerning the MAE and RMSE performance indicators, which decrease as the interval mentioned above progresses. Finally, it can be observed that the values of all the indicators are significantly worse for S = 5 , and even yield results with little statistical meaning; thus, in this study, when using real data S = 1 is considered for the analysis. Regarding the training time of the neural networks, the results of which can be seen in the tables, it can be seen that the BiLSTM network takes the longest for all window sizes, which is clearly due to its architecture. However, this effectively increases the amount of information available to the network, improving its coefficient of determination.

4.2.2. Real Datasets

From the results of the fractional-noise synthetic data in the previous section, the BiLSTM neural network was used for the real dataset using different window sizes S = 1 . As in the case of the synthetic data, we ran ten simulations for each model and the results were averaged, which are the results appearing in the tables below. From Table 4 and Table 5, it can be observed that for the IPSA dataset, the best indices with N = 20 were obtained in the determination coefficient R 2 , while for the Tree Ring dataset the best results in the determination coefficient were obtained when N = 25 . It can observed that the MAE and RMSE coefficients both decrease as the size of the window increases. Figure 10 shows the prediction of the values of the Hurst exponent for the real data, reinforcing what was observed in the previous tables. Furthermore, in the last column of Table 4 and Table 5 it can be seen that the training time of the BiLSTM network depends on the size of the dataset and the assigned size of the window N. Table 6 indicates that, through the Kolmogorov-Smirnov test, the predictions obtained by the BILSTM network fit the test set of the real data time series. The idea of carrying out this test to check the goodness of fit of the predictions was based, among other studies, on [15].

5. Conclusions

In this work, we have presented a new approach for predicting the Hurst exponent using recurrent neural networks. By applying Whittle’s method using a sliding time window, a new time series corresponding to the estimation of long memory is constructed. Different recurrent neural network models were trained which received data blocks from the original time series as input and generated one-step-ahead predictions of the long memory parameter as output. Our results show that it is possible to have good predictions one step ahead of the long memory parameter; in particular, the BiLSTM network obtained the best results when using the proposed methodology. Additionally, these predictions can be made in real time due to the computational speed of the neural network models.
Further work could include a new procedure that incorporates more complex models with a long memory, and could even involve heteroscedastic behaviors. One of the limitations of our proposed method is that it relies on a fixed size block length, meaning that the RNN cannot successfully capture points located very distant from the signal. Further work is required to enhance prediction when the size of the overlapping blocks changes dynamically, together with a rehearsal mechanism for incremental learning. In our future work, we expect to correlate the temporal estimation of long-term memory in order to improve prediction of the volatility of GARCH models.

Author Contributions

Conceptualization, C.U., J.E.C.-R. and R.S.; Methodology, C.U., G.D.-G. and R.S.; Software, C.U., G.D.-G. and R.S.; Validation, C.U. and R.S.; Investigation, C.U., G.D.-G., J.E.C.-R. and R.S.; Resources, R.S.; Writing—original draft, C.U. and R.S.; Writing—review and editing, C.U., J.E.C.-R. and R.S.; Supervision, R.S. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the support provided by the ANID-Millennium Science Initiative Program ICN2021_004, ANID FONDECYT research grant number 1221938, and FONDECYT initiation research grant number 11190116. The work of C. Ubal was supported by the Universidad de Valparaiso scholarship.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used in this article are publicly available.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Di Giorgi, G.; Salas, R.; Avaria, R.; Ubal, C.; Rosas, H.; Torres, R. Volatility Forecasting using Deep Recurrent Neural Networks as GARCH models. Comput. Stat. 2023, 1–27. [Google Scholar] [CrossRef]
  2. Cordova, C.H.; Portocarrero, M.N.L.; Salas, R.; Torres, R.; Rodrigues, P.C.; López-Gonzales, J.L. Air quality assessment and pollution forecasting using artificial neural networks in Metropolitan Lima-Peru. Sci. Rep. 2021, 11, 24232. [Google Scholar] [CrossRef]
  3. Leite Coelho da Silva, F.; da Costa, K.; Canas Rodrigues, P.; Salas, R.; López-Gonzales, J.L. Statistical and artificial neural networks models for electricity consumption forecasting in the Brazilian industrial sector. Energies 2022, 15, 588. [Google Scholar] [CrossRef]
  4. Vivas, E.; de Guenni, L.B.; Allende-Cid, H.; Salas, R. Deep Lagged-Wavelet for monthly rainfall forecasting in a tropical region. Stoch. Environ. Res. Risk Assess. 2023, 37, 831–848. [Google Scholar] [CrossRef]
  5. Querales, M.; Salas, R.; Morales, Y.; Allende-Cid, H.; Rosas, H. A stacking neuro-fuzzy framework to forecast runoff from distributed meteorological stations. Appl. Soft Comput. 2022, 118, 108535. [Google Scholar] [CrossRef]
  6. Kovantsev, A.; Gladilin, P. Analysis of multivariate time series predictability based on their features. In Proceedings of the 2020 International Conference on Data Mining Workshops (ICDMW), Sorrento, Italy, 17–20 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 348–355. [Google Scholar]
  7. Qian, B.; Rasheed, K. Hurst exponent and financial market predictability. In Proceedings of the IASTED Conference on Financial Engineering and Applications, IASTED International Conference, Cambridge, MA, USA, 9–11 November 2004; pp. 203–209. [Google Scholar]
  8. Siriopoulos, C.; Markellos, R. Neural Network Model Development and Optimization. J. Comput. Intell. Financ. (Former. Neurovest J.) 1996, 7–13. [Google Scholar]
  9. Siriopoulos, C.; Markellos, R.; Sirlantzis, K. Applications of Artificial Neural Networks in Emerging Financial Markets; World Scientific: Singapore, 1996; pp. 284–302. [Google Scholar]
  10. Lin, T.; Horne, B.G.; Tino, P.; Giles, C.L. Learning long-term dependencies in NARX recurrent neural networks. IEEE Trans. Neural Netw. 1996, 7, 1329–1338. [Google Scholar]
  11. Ledesma-Orozco, S.; Ruiz-Pinales, J.; García-Hernández, G.; Cerda-Villafaña, G.; Hernández-Fusilier, D. Hurst parameter estimation using artificial neural networks. J. Appl. Res. Technol. 2011, 9, 227–241. [Google Scholar] [CrossRef]
  12. Menezes Jr, J.M.P.; Barreto, G.A. Long-term time series prediction with the NARX network: An empirical evaluation. Neurocomputing 2008, 71, 3335–3343. [Google Scholar] [CrossRef]
  13. Hua, Y.; Zhao, Z.; Li, R.; Chen, X.; Liu, Z.; Zhang, H. Deep learning with long short-term memory for time series prediction. IEEE Commun. Mag. 2019, 57, 114–119. [Google Scholar] [CrossRef]
  14. Li, X.; Yu, J.; Xu, L.; Zhang, G. Time Series Classification with Deep Neural Networks Based on Hurst Exponent Analysis. In Proceedings of the ICONIP 2017: Neural Information Processing, Guangzhou, China, 14–18 November 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 194–204. [Google Scholar]
  15. Hassani, H.; Silva, E.S. A Kolmogorov-Smirnov Based Test for Comparing the Predictive Accuracy of Two Sets of Forecasts. Econometrics 2015, 3, 590–609. [Google Scholar] [CrossRef]
  16. Hurst, H.E. Long-term storage capacity of reservoirs. Trans. Am. Soc. Civ. Eng. 1951, 116, 770–799. [Google Scholar] [CrossRef]
  17. Mandelbrot, B.B.; Van Ness, J.W. Fractional Brownian motions, fractional noises and applications. SIAM Rev. 1968, 10, 422–437. [Google Scholar] [CrossRef]
  18. Peng, C.K.; Buldyrev, S.V.; Havlin, S.; Simons, M.; Stanley, H.E.; Goldberger, A.L. Mosaic organization of DNA nucleotides. Phys. Rev. E 1994, 49, 1685. [Google Scholar] [CrossRef]
  19. Geweke, J.; Porter-Hudak, S. The estimation and application of long memory time series models. J. Time Ser. Anal. 1983, 4, 221–238. [Google Scholar] [CrossRef]
  20. Whittle, P. Hypothesis Testing in Time Series Analysis; Almqvist & Wiksells: Upsala, Sweeden, 1951; Volume 4. [Google Scholar]
  21. Veitch, D.; Abry, P. A wavelet-based joint estimator of the parameters of long-range dependence. IEEE Trans. Inf. Theory 1999, 45, 878–897. [Google Scholar] [CrossRef]
  22. Taqqu, M.S.; Teverovsky, V.; Willinger, W. Estimators for long-range dependence: An empirical study. Fractals 1995, 3, 785–798. [Google Scholar] [CrossRef]
  23. Palma, W.; Chan, N.H. Estimation and forecasting of long-memory processes with missing values. J. Forecast. 1997, 16, 395–410. [Google Scholar] [CrossRef]
  24. Hosking, J.R.M. Fractional differencing. Biometrika 1981, 68, 165–176. [Google Scholar] [CrossRef]
  25. Fox, R.; Taqqu, M.S. Large-sample properties of parameter estimates for strongly dependent stationary Gaussian time series. Ann. Stat. 1986, 14, 517–532. [Google Scholar] [CrossRef]
  26. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  27. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control. Signals Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  28. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  29. Salehinejad, H.; Sankar, S.; Barfett, J.; Colak, E.; Valaee, S. Recent advances in recurrent neural networks. arXiv 2017, arXiv:1801.01078. [Google Scholar]
  30. Schäfer, A.M.; Zimmermann, H.G. Recurrent neural networks are universal approximators. Int. J. Neural Syst. 2007, 17, 253–263. [Google Scholar] [CrossRef]
  31. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  32. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef] [PubMed]
  33. Hochreiter, S.; Bengio, Y.; Frasconi, P.; Schmidhuber, J. Gradient Flow in Recurrent Nets: The Difficulty of Learning Long-Term Dependencies; IEEE Press: Hoboken, NJ, USA, 2001. [Google Scholar]
  34. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef]
  35. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  36. Chen, L. Deep Learning and Practice with MindSpore; Springer Nature: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  37. Contreras-Reyes, J.E.; Palma, W. Statistical analysis of autoregressive fractionally integrated moving average models in R. Comput. Stat. 2013, 28, 2309–2331. [Google Scholar] [CrossRef]
  38. Palma, W.; Olea, R. An efficient estimator for locally stationary Gaussian long-memory processes. Ann. Stat. 2010, 38, 2958–2997. [Google Scholar] [CrossRef]
  39. Singleton, R. Mixed Radix Fast Fourier Transform; Technical Report; Stanford Research Inst.: Menlo Park, CA, USA, 1972. [Google Scholar]
  40. Whittle, P. Estimation and information in stationary time series. Ark. Mat. 1953, 2, 423–434. [Google Scholar] [CrossRef]
  41. Bisaglia, L.; Guegan, D. A comparison of techniques of estimation in long-memory processes. Comput. Stat. Data Anal. 1998, 27, 61–81. [Google Scholar] [CrossRef]
  42. Dahlhaus, R. Efficient parameter estimation for self-similar processes. Ann. Stat. 1989, 1749–1766. [Google Scholar] [CrossRef]
  43. Ferreira, G.; Olea Ortega, R.A.; Palma, W. Statistical analysis of locally stationary processes. Chil. J. Stat. 2013, 4, 133–149. [Google Scholar]
  44. Beran, J.; Feng, Y.; Ghosh, S.; Kulik, R. Long-Memory Processes; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  45. Armstrong, J.S. Evaluating Forecasting Methods. Principles of Forecasting; Springer: Berlin/Heidelberg, Germany, 2001; pp. 443–472. [Google Scholar]
Figure 1. The left side shows the simple RNN architecture; on the right side, the RNN is unfolded into a full network.
Figure 1. The left side shows the simple RNN architecture; on the right side, the RNN is unfolded into a full network.
Make 05 00068 g001
Figure 2. Block diagram of an LSTM recurrent neural network cell unit.
Figure 2. Block diagram of an LSTM recurrent neural network cell unit.
Make 05 00068 g002
Figure 3. Architecture of the BiLSTM Network.
Figure 3. Architecture of the BiLSTM Network.
Make 05 00068 g003
Figure 4. Block diagram of the GRU recurrent neural network cell unit.
Figure 4. Block diagram of the GRU recurrent neural network cell unit.
Make 05 00068 g004
Figure 5. Simulated data FN(d) for d = { −0.3, 0, 0.3}.
Figure 5. Simulated data FN(d) for d = { −0.3, 0, 0.3}.
Make 05 00068 g005
Figure 6. Real world datasets: (a) IPSA dataset for the period 2000–2021; (b) tree ring dataset.
Figure 6. Real world datasets: (a) IPSA dataset for the period 2000–2021; (b) tree ring dataset.
Make 05 00068 g006
Figure 7. Scheme of the methodology: Step 0, original dataset; Step 1, block construction (in red, the data blocks of the series from which the estimates are obtained); Step 2, Whittle estimation in each block; Step 3, training of the Hurst estimation dataset using the RNN (Blue: Training Data Set; Orange: Test Data Set); Step 4, prediction using the RNN (Blue: Target data series; Orange: Prediction data series).
Figure 7. Scheme of the methodology: Step 0, original dataset; Step 1, block construction (in red, the data blocks of the series from which the estimates are obtained); Step 2, Whittle estimation in each block; Step 3, training of the Hurst estimation dataset using the RNN (Blue: Training Data Set; Orange: Test Data Set); Step 4, prediction using the RNN (Blue: Target data series; Orange: Prediction data series).
Make 05 00068 g007
Figure 8. Comparison H vs. d of the estimation methods.
Figure 8. Comparison H vs. d of the estimation methods.
Make 05 00068 g008
Figure 9. Comparison of the estimation methods with Monte Carlo simulation.
Figure 9. Comparison of the estimation methods with Monte Carlo simulation.
Make 05 00068 g009
Figure 10. Predictions obtained with the BiLSTM network for the real dataset.
Figure 10. Predictions obtained with the BiLSTM network for the real dataset.
Make 05 00068 g010aMake 05 00068 g010b
Table 1. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0.3 ) simulated data using recurrent neural networks.
Table 1. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0.3 ) simulated data using recurrent neural networks.
d = 0.3
R 2 RMSETraining Time (s)
sRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRU
S = 1180.7210.7420.7060.7290.1710.1640.1750.168130.9262.3220.9258.7
200.7090.7460.7160.7360.1610.1510.1600.154104.5233.4207.0255.1
230.6730.7240.6860.7060.1500.1380.1480.143127.2261.7235.8270.2
250.6570.7230.6760.6990.1460.1310.1420.137128.6266.3265.6265.4
280.6520.7310.6660.6990.1410.1250.1380.132115.1226.8215.7231.1
300.6410.7300.6460.6850.1410.1220.1400.132134.5284.0226.4271.5
330.6010.6900.6160.6500.1360.1200.1330.127124.0240.9209.8246.3
350.5340.6700.6050.6440.1380.1160.1270.121127.1252.0213.6260.2
S = 5200.298−0.1170.5190.5320.2480.3140.2060.20334.470.462.479.1
250.282−0.3960.4400.5070.2060.2900.1840.17336.970.269.169.5
300.049−0.2170.4360.4610.2340.2640.1800.17632.976.260.469.3
Table 2. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0 ) simulated data using recurrent neural networks.
Table 2. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0 ) simulated data using recurrent neural networks.
d = 0
R 2 RMSETraining Time (s)
sRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRU
S = 1180.7710.8330.7760.7820.1670.1420.1660.163119.0271.3228.0288.6
200.7670.8130.7610.7790.1580.1410.1590.154116.2220.5209.8241.7
230.7300.8070.7210.7540.1440.1220.1470.138122.8226.1190.7237.7
250.7130.8000.7000.7470.1420.1180.1450.133168.5329.7303.2344.3
280.7210.8150.6970.7370.1320.1080.1380.128144.3316.6264.3298.5
300.6860.8120.6720.7200.1240.1040.1370.127113.3214.8197.0232.3
330.6540.7810.6500.6890.1250.0990.1260.118119.9212.5188.4221.3
350.6700.7820.6110.6730.1140.0930.1240.114131.6213.0183.4216.8
S = 5200.3540.1000.5160.5330.2600.3060.2250.22126.564.043.449.2
250.321−0.0600.4810.4970.2150.2690.1880.18632.451.442.549.0
300.1280.1250.4990.5260.2320.2330.1760.17131.952.744.061.6
Table 3. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0.3 ) simulated data using recurrent neural networks.
Table 3. Performance results and training time for prediction of the Hurst parameter for the F N ( d = 0.3 ) simulated data using recurrent neural networks.
d = 0.3
R 2 RMSETraining Time (s)
sRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRUsRNNBiLSTMLSTMGRU
S = 1180.8140.8520.8010.8010.1550.1380.1600.160128.3293.8237.6278.0
200.7690.8600.7910.8040.1570.1230.1500.145100.8221.7173.6202.0
230.7480.8690.7590.7780.1420.1020.1390.133124.1239.5208.3261.8
250.7560.8770.7530.7730.1350.0960.1350.130125.2249.9200.1240.1
280.7480.8820.7440.7680.1290.0880.1310.124129.2281.7245.7273.4
300.7160.8840.7210.7510.1290.0830.1280.121120.7260.9197.2247.7
330.6850.8590.6760.7140.1210.0810.1220.115117.6229.1193.3230.4
350.6810.8450.6540.6800.1140.0790.1190.114124.0223.2203.8243.2
S = 5200.3750.1730.4850.5180.2540.2920.2300.22337.283.081.482.5
250.3680.2300.4750.4790.2120.2340.1930.19239.786.775.3104.2
300.3070.1110.4610.5040.2080.2350.1830.17638.984.183.585.9
Table 4. Performance results for prediction of the Hurst parameter for the IPSA time series dataset using the BiLSTM Network.
Table 4. Performance results for prediction of the Hurst parameter for the IPSA time series dataset using the BiLSTM Network.
IPSA Dataset (BiLSTM)
Train Test
N R 2 MAE RMSE R 2 MAE RMSE Training Time (s)
180.9120.0720.0970.6790.1330.184147.659
200.9220.0620.0830.7160.1150.153136.575
230.9360.0480.0640.6510.1060.145111.459
250.9380.0440.0580.6210.1050.140131.349
280.9440.0380.0500.4370.1060.148122.480
300.9480.0340.0450.3960.1030.141112.823
330.9520.0290.0390.3980.0990.133133.320
350.9550.0270.0360.3480.0980.132109.261
Table 5. Performance results for prediction of the Hurst parameter for the Tree Ring time series dataset using the BiLSTM Network.
Table 5. Performance results for prediction of the Hurst parameter for the Tree Ring time series dataset using the BiLSTM Network.
Tree Ring Dataset (BiLSTM)
TrainTest
N R 2 MAERMSE R 2 MAERMSETraining (s)
180.9120.0760.1020.7990.1090.150169.029
200.9250.0660.0870.8180.0970.133149.533
230.9370.0520.0680.8220.0840.115251.819
250.9430.0460.0610.8300.0800.106259.888
280.9490.0420.0540.8280.0740.098259.807
300.9490.0390.0510.8190.0730.095224.713
330.9530.0350.0450.7900.0710.094247.332
350.9560.0320.0420.7860.0690.090245.959
Table 6. KS test for prediction of the real time series using the BILSTM Network.
Table 6. KS test for prediction of the real time series using the BILSTM Network.
KS Test
DataN OptStatistic Valuep-Value
IPSA200.06630.2244
Treering250.02890.8937
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ubal, C.; Di-Giorgi, G.; Contreras-Reyes, J.E.; Salas, R. Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks. Mach. Learn. Knowl. Extr. 2023, 5, 1340-1358. https://doi.org/10.3390/make5040068

AMA Style

Ubal C, Di-Giorgi G, Contreras-Reyes JE, Salas R. Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks. Machine Learning and Knowledge Extraction. 2023; 5(4):1340-1358. https://doi.org/10.3390/make5040068

Chicago/Turabian Style

Ubal, Cristian, Gustavo Di-Giorgi, Javier E. Contreras-Reyes, and Rodrigo Salas. 2023. "Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks" Machine Learning and Knowledge Extraction 5, no. 4: 1340-1358. https://doi.org/10.3390/make5040068

APA Style

Ubal, C., Di-Giorgi, G., Contreras-Reyes, J. E., & Salas, R. (2023). Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks. Machine Learning and Knowledge Extraction, 5(4), 1340-1358. https://doi.org/10.3390/make5040068

Article Metrics

Back to TopTop