Next Article in Journal
On a System of Equations with General Fractional Derivatives Arising in Diffusion Theory
Next Article in Special Issue
A New Approach to Determining the Network Fractality with Application to Robot-Laser-Hardened Surfaces of Materials
Previous Article in Journal
A Unified Description of the Electrical Properties with Complex Dynamical Patterns in Metal Halide Perovskite Photovoltaics
Previous Article in Special Issue
Self-Similar Growth and Synergistic Link Prediction in Technology-Convergence Networks: The Case of Intelligent Transportation Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Probabilistic Machine Learning Methods for Fractional Brownian Motion Time Series Forecasting

by
Lyudmyla Kirichenko
1,* and
Roman Lavrynenko
2
1
Applied Mathematics Department, Kharkiv National University of Radio Electronics, 61166 Kharkiv, Ukraine
2
Artificial Intelligence Department, Kharkiv National University of Radio Electronics, 61166 Kharkiv, Ukraine
*
Author to whom correspondence should be addressed.
Fractal Fract. 2023, 7(7), 517; https://doi.org/10.3390/fractalfract7070517
Submission received: 16 May 2023 / Revised: 14 June 2023 / Accepted: 28 June 2023 / Published: 29 June 2023

Abstract

:
This paper explores the capabilities of machine learning for the probabilistic forecasting of fractional Brownian motion (fBm). The focus is on predicting the probability of the value of an fBm time series exceeding a certain threshold after a specific number of time steps, given only the knowledge of its Hurst exponent. The study aims to determine if the self-similarity property is preserved in a forecasting time series and which machine learning algorithms are the most effective. Two types of forecasting methods are investigated: methods with a predefined distribution shape and those without. The results show that the self-similar properties of the fBm time series can be reliably reproduced in the continuations of the time series predicted by machine learning methods. The study also provides an experimental comparison of various probabilistic forecasting methods and their potential applications in the analysis and modeling of fractal time series.

1. Introduction

Fractional Brownian motion (fBm) is a stochastic process characterized by long-range dependence, self-similarity, and non-stationarity. These properties make fBm a suitable framework for modeling time series data with complex structures and memory effects, such as financial market data, Internet traffic, realizations of biomedical signals, dynamic indicators of natural phenomena, etc. [1,2,3,4,5]. Predicting the future behavior of stochastic fractal time series can help make more informed decisions. For instance, in finance, forecasts enable risk assessment and decision making regarding an asset purchase or sale. Predicting Internet traffic allows for optimal scaling of network resources, and anticipation of peak loads, ensuring sufficient bandwidth for high-quality user service. Moreover, forecasting fractal time series of seismic activity can provide insights into potential trends and probabilities of earthquakes, among other applications. These predictions contribute to a wide range of fields and can lead to more effective decision-making processes [6,7,8,9,10]. Research into fractional processes has the potential to advance emerging directions in computer science, such as high-order fractional complex-valued bidirectional associative memory neural networks with multiple time delays [11,12,13].
Despite the widespread usage of point forecasting and interval forecasting methods in time series analysis, these approaches alone prove insufficient for stochastic time series such as fBm. The application of probabilistic forecasting confers a range of advantages. Probabilistic forecasting provides a more comprehensive representation of future outcomes by estimating the entire probability distribution of the forecast variable, rather than just a single value.
Probabilistic forecasting methods have become increasingly important in various fields, such as finance, weather, and energy. These methods provide valuable information on the distribution of future outcomes, which is crucial for decision making and risk assessment. This information can be used to assess the risks associated with different decisions and to implement risk management strategies.
There is a multitude of research related to the forecasting of self-similar time series but the majority of existing methods primarily focus on point and interval forecasting methods [14,15,16]. However, the challenge of probabilistic forecasting for fractal time series, particularly fBm, remains unresolved. This highlights the need for further research and development of techniques that address probabilistic forecasting in the context of fractal time series.
Probabilistic forecasting methods can be broadly categorized into two classes based on the underlying distribution shape assumptions: (1) methods with a predefined distribution shape and (2) methods without a predefined distribution shape. The key difference between these two categories is the degree of flexibility in modeling the underlying uncertainty:
  • Methods that assume parameters of a specific probability distribution that the forecast variable follows, such as the Gaussian, Gamma, or Weibull distribution. These methods are characterized by a finite set of parameters that are estimated from the data. Examples of parametric methods include Natural Gradient Boosting for Probabilistic Prediction (NGBoost) [17], XGBoostLSS and LightGBMLSS, which model all moments of a parametric distribution: mean, location, scale, and shape (LSS) [18,19], Catboost with Uncertainty (CBU), Probabilistic Gradient Boosting Machines (PGBM) [20], and Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees (IBUG) [21].
  • Methods that estimate the distribution directly from the data without relying on any predefined shape, often focusing on predicting quantiles. These methods provide increased flexibility when modeling complex distributions, making them especially valuable when the true distribution is unknown or significantly deviates from standard parametric forms. Quantile regression is a popular nonparametric technique [22,23]. The newest approach is Nonparametric Probabilistic Regression with Coarse Learners (PRESTO) [24].
The objective of this research comprises two stages. The first stage focuses on exploring the potential for long-term probabilistic forecasting of fBm realizations while maintaining the Hurst exponent. In this stage, fBm realizations with various Hurst exponent values will be modeled using the Hosking algorithm [25], and forecasted realizations (continuations) will be obtained using several primary probabilistic forecasting methods. These continuations will be examined for preserving the Hurst exponent value using the Whittle estimator and normal distribution using the Anderson–Darling test [26]. The preservation of the Hurst exponent, normal distribution, and root mean square increments of fBm realizations and their continuations allows us to conclude that the properties of fBm realizations are preserved when forecasting with machine learning methods.
The second stage is directed toward conducting an experimental comparison of various methods for predicting fBm realizations. In this part of the research, both parametric and nonparametric probabilistic forecasting methods employing machine learning will be considered. Computational experiments will be carried out, where each method will be applied to forecast fBm realizations with different Hurst exponent values. The accuracy and efficiency of each method will then be compared to determine the most suitable forecasting method for these types of data.
Our Contribution:
  • We demonstrate that the self-similar properties of the fBm time series can be reliably reproduced in the continuations of the series predicted by machine learning methods.
  • We propose an approach for evaluating probabilistic methods in forecasting the fBm time series. Our method involves efficiently computing multiple continuations of a single fBm time series using the Hosking algorithm, resulting in a dataset with ground truth quantiles of possible continuations. Using a one-period-ahead model, iterated forward for the desired number of periods, we can predict quantiles of continuations of fBm time series over hundreds or thousands of time steps and compare them with the ground truth quantiles (Figure 1).
  • We conducted an experimental comparison of various probabilistic forecasting methods using this approach, providing insights into their performance and potential applications in the analysis and modeling of fractal time series.

2. Materials and Methods

2.1. Problem Setting

To evaluate the performance of a probabilistic forecasting model, it is necessary to compare the predicted probability distribution with the actual probability distribution of the ground truth data. One common way to do this is by calculating the quantiles of the predicted probability distribution and comparing them with the quantiles of the actual distribution.
Predicting only one-time step ahead is insufficient for modeling the behavior of a time series with fractal properties in the future and evaluating the method’s performance. It is necessary to construct sufficiently long continuations of the time series to assess the presence of similar fractal properties in these continuations. To achieve this, we used iterated approaches that utilize one-step-ahead prediction models, with multi-step predictions obtained by recursively feeding predictions into future inputs. This allows us to generate longer continuations of the time series while preserving the fractal characteristics, which can then be compared to the ground truth continuations for method evaluation and performance analysis.
Additionally, predicting merely one continuation of a fractal time series is not enough to evaluate the method. It is crucial to generate a representative number of possible time series continuations and compare the quantiles of the obtained continuations for each time step with the quantiles of the true possible continuations.
Our research will focus on the probabilistic forecasting of trajectories for a self-similar process. A process {X(t), t ∈ R} is said to be self-similar of Hurst exponent H, if, for any a > 0:
Law{X(at)} = Law{a H{X(t)},
that is the equality of n-dimensional distribution laws of X(t) [27,28].
There are several approaches to defining fractional Brownian motion [29,30], and here we apply one of them [28] that emphasizes the properties of normality and self-similarity. Fractional Brownian motion X(t) is a type of self-similar stochastic process, the increments of which on the time interval τ ΔX(τ) = X(t + τ) − X(t) have a Gaussian distribution:
P ( Δ X < x ) = 1 2 π σ τ H x exp ( z 2 2 σ 2 τ 2 H ) dz ,
where σ is the diffusion coefficient.
Exponent H determines the degree of long-range dependence of the fBm increments. The corresponding normalized correlation function is expressed as:
Cor(t,s) = 0.5 ∗ (|t|2H + |s|2H − |t − s|2H),
where t and s are time indices [28].
In general, for probabilistic forecasting, when Y is a real-valued response variable and X is a predictor variable, the goal is to find the conditional distribution function [31], which can be represented as
F(y|X = x) = P(Y ≤ y|X = x).
Quantiles provide complete information about the distribution. For a continuous distribution function, the α-quantile Qα(x) is defined as Qα(x) = inf{y: F(y|X = x) ≥ α}.
If the predictor variable is a time series
X(t) = {X(1), X(2), …, X(N)},
the task of probabilistic forecasting involves finding an algorithm that can predict the probability distribution of future values of the time series X(t + 1), X(t + 2), …, X(t + K), based on available past values X(t).
Thus, in the case of a fractional Brownian motion fBm time series X(t) of length N, with a Hurst exponent equal to H, the task is to investigate the properties of continuations
Xc(t) = {X(N + 1), X(N + 2), …, X(N + K)},
obtained using machine learning methods, such that
F(X(N + 1), X(N + 2), …, X(N + K) | X(N), X(N − 1), …, X(N − m)),
where F is the conditional distribution function, K is the forecast horizon, and m is the horizon of known values of the time series.
In each generated continuation Xc(t), the Hurst exponent H of the original time series X(t) should be preserved, as well as the scale (standard deviations of increments S). The quantiles of the continuations Xc(t) will be compared to the ground truth quantiles of possible continuations, and a metric assessing their discrepancies will be determined.

2.2. Method of Creating the Evaluating Dataset

There are several exact methods to simulate fBm realizations [25]. The Hosking method is simple, popular, and implemented in Python. The Hosking method is intended for the simulation of a stationary Gaussian sequence, which is increments of discrete-time fBm. An important feature of our research is that this method generates Xn+1 given Xn, …, X1 recursively.
We propose a method for creating a dataset of fBm time series and their ground truth continuations using the Hosking algorithm. Our method involves the parallel generation of M continuations Xc(t) for one original time series X(t) and the formation of a matrix of quantiles Qα = {Q(αi, tj)} for each time step t. This dataset can then be used to evaluate probabilistic forecasting methods the for fBm time series.
Our method of creating a single example for the dataset for a specific Hurst exponent consists of the following steps:
  • We use the Hosking method to generate values of the first N time steps of the predictor variable, which is a time series X(t) given by expression (5);
  • Starting from the state of the internal variables of the Hosking algorithm after generating step N, we continue generating values for the next K time steps, obtaining a continuation Xc(t), defined by (6);
  • We repeat step 2 M times, obtaining M different continuations from the original common starting point:
Xcm(t) = {Xcm(1), Xcm (2), …, Xcm (K)}, m = 1, 2,…, M;
4.
For each time step k, k = 1, 2, …, K, we determine the quantiles of the values of M continuations, obtaining the desired quantile matrix;
5.
The dataset stores the target Hurst exponent, for example, X(t), the predictor variable X(t) of length N, and matrix Qα = {Q(αi,tj)} of size 101× K, which contains, for each time step tj = 1, 2, …, K, quantiles αi, i = 1, 2, …, 101 of the distribution of M continuations {Xcm(t)}. The 0th and 1.0 quantiles represent the minimum and maximum values among the M variants at each time step tj.
This algorithm for creating a single example for the database can be used to populate the database with many examples for each of the various Hurst exponent values.
The example created using this method is illustrated in Figure 2. The figure displays a common row for the first 1024 time steps, followed by four colored curves representing the quantiles of the numerous possible continuations of the common beginning.
The code for generating a file of the dataset for a specific Hurst exponent can be found at the following link: https://www.kaggle.com/code/unfriendlyai/fbm-dataset-generator (accessed on 8 May 2023). Our fBm dataset is available on the Kaggle platform at: https://www.kaggle.com/datasets/unfriendlyai/fractal-brownian-motion (accessed on 8 May 2023).

2.3. Predicting Quantiles of Time Series for Multiple Future Time Steps

To iteratively predict quantiles for multiple future time steps using a one-step-ahead machine learning method, we propose the following approach:
  • Train the machine learning model to output the distribution of the next time series value given the previous ones. The result can be in the form of parameters of a certain distribution, quantiles, or any other form that we can sample values from.
  • Given a test example with a time series length of N X(t), the model predicts the distribution F represented by (7) of the next N + 1 value X(N + 1). Sample one value from the distribution F to obtain one possible next value of the continuation, Xc(1).
  • Concatenate X(t) with Xc(t) and iteratively predict the next time step distribution, sampling one value for Xc(t + 1) for the next K time steps, obtaining a continuation Xc(t), defined by Equation (6).
  • Repeat the second and third steps many times, obtaining many different continuations from the original single test example.
  • Determine the quantiles of the values of the continuations, obtaining the desired quantiles for each time step.
This algorithm for estimating quantiles for future time steps of a single test example X(t) can be used to predict quantiles for many such examples (Figure 1). Then predicted quantiles can be compared with ground truth quantiles from the dataset to evaluate the efficiency of the probabilistic machine learning method. The same algorithm can be used to predict the probability of the value of an fBm time series exceeding a certain threshold after a specific number of time steps.
The advantage of using iterative one-step-ahead prediction over directly predicting multiple time steps is that one-step prediction is accessible to a wide array of machine learning methods, unlike predicting a sequence of future values. Furthermore, once the model is trained, it can be used to predict an arbitrary number of future time steps.

2.4. Assessment of Accuracy and Reliability of Continuations

In our research, it is essential to compare the properties of the obtained continuations Xc(t) with the properties of the original time series X(t).
There are several properties we need to consider for comparison: the Hurst exponent H, the normality of the distribution, and the magnitude of increments. If the values of H, the standard deviation of the increments, and the normality of the distribution coincide, we can conclude that the obtained continuations are realizations of the fractional Brownian motion with the same H as the original ones.
The aforementioned properties are examined and verified through the following procedures:
  • Calculating the Hurst exponent (H) using the Whittle estimator demonstrated the best accuracy for time series with lengths of 512 and above, which we plan to evaluate, as shown in a comparative study [32]. This allows us to assess whether the fractal properties of the original time series X(t) are preserved in the generated continuations Xc(t).
  • Determining if the increments are normally distributed using the Anderson–Darling test. This test is a statistical method used to check whether a given sample of data follows a specific distribution (in our case, normal distribution). If the increments are normally distributed, it indicates that the generated continuations Xc(t) maintain the same distribution law as the original time series X(t).
  • Comparing the standard deviations of increments (S) between the original time series X(t) and the generated continuations Xc(t). By ensuring that the standard deviations of increments are preserved, we can confirm that the scale of the generated continuations Xc(t) is consistent with the original time series X(t).
In addition to the requirement that the generated continuations Xc(t) must have the same fractal and statistical properties as the original series X(t), we need to assess the quality of the forecasting. Ideally, for each example, X(t) with its corresponding value of H, the matrix of quantiles of the continuations Qc(αi) should match the true matrix Q(αi) presented in the dataset. As a measure of deviation Qdev, we will use the absolute difference between the true and predicted quantile values at the given time step tj, normalized by the difference between the true value of the quantile Q(αi) and the adjacent Q(αi+1) for each quantile from 0.01 to 0.99 and each time step, averaging it (Figure 3):
Qdev = 1 99 K j = 1 K i = 1 99 | Q c ( α i , tj ) Q ( α i , tj ) | Q ( α i + 1 , tj ) Q ( α i , tj ) ,
where QC(αi,tj) is predicted αi-quantile values at the time step tj; Q(αi,tj) is true αi-quantile values at the time step tj; K is the length of the continuation.
This interpretable metric provides an intuitive understanding of the model’s performance by expressing the error in the context of the distance between adjacent quantiles. A lower value for this metric indicates that the predicted quantiles are closer to the true quantiles, making the model’s forecasts more accurate.
The metric Qdev characterizes the deviation of the time series continuation Xc(t) for a single example X(t) in the dataset for a specific value of H. Since the values of Qdev depend on random time realizations, it is necessary to average over all calculated examples with the same Hurst exponent to improve accuracy and reliability:
QDmean = 1 Nexample i = 1 Nexample Qdev ( i )
where Qdev is the metric for a single example in the dataset, and Nexample is the number of examples in the dataset.
By performing the above-described checks, we can validate the quality of the generated continuations and ensure that our method effectively captures the essential fractal and statistical properties of the original time series X(t).

2.5. Machine Learning Methods for Probabilistic Forecasting

According to reviews [31,33,34], there are two broad classes of probabilistic methods: those predicting the parameters of a predefined distribution shape, and those addressing the problem when the distribution shape is not specified. Although the first class of methods is perfectly suitable for predicting fractional Brownian motion, this research aims to explore methods that are not limited by a predefined probability distribution shape and are applicable to studying complex fractal time series.
For this study, we used implementations of the methods listed in Table 1.
By comparing these methods, we aim to find the most suitable technique for modeling and forecasting fractal time series data while preserving the essential fractal properties. This will help us to better understand the advantages and limitations of different probabilistic prediction methods in the context of complex, long-range dependent, and self-similar time series data, ultimately leading to more accurate and reliable forecasts for various practical applications.
The code and results of our experiments can be accessed at https://www.kaggle.com/datasets/unfriendlyai/fractal-brownian-motion/code (accessed on 8 May 2023).

3. Results

3.1. Dataset

3.1.1. Training Dataset

In a supervised learning setting, a machine learning model requires a predictor and a ground truth target. Probabilistic forecasting methods use the same training datasets as point-predicting methods. For time series, the training process for each algorithm involves learning to find the distribution of the (n + 1)-th element in the series based on the known n previous values, according to Formula (7).
To train models, there is no need to store the dataset, as it is possible to generate millions of fBm examples with specified Hurst exponent within a few minutes. We used the Hosking method for Hurst exponent values starting from 0.9 and the Davies–Harte method [25] for the Hurst exponent values below 0.9.

3.1.2. Evaluating Dataset

The dataset was created with the following parameters:
  • Number of files—99, each for the value of H from 0.01 to 0.99 with a step of 0.01;
  • Number of records in each file—50;
  • The length of the original time series N—1024;
  • The length of the continuations of the original series, for which quantiles are provided K—1024;
  • The number of continuations of the time series to calculate quantiles M—10,001;
  • The number of quantiles for each time step is 101 (the 0th and 1.0th quantiles are equal to the maximum and minimum values among M examples);
  • Time series increments are normalized (divided by STD of increments); the standard deviations of increments for all examples are one;
  • Original time series are presented as cumulative sums of increments. Quantiles are calculated on their cumulative continuations.
Each file, which consists of 50 examples of a certain Hurst exponent mentioned in the filename, takes nearly 2 h to create.
For experiments, we use dataset examples for a range of Hurst exponent values to represent anti-persistent, persistent, and near-independent fBm examples: 0.30, 0.35, 0.45, 0.53, 0.60, 0.65, 0.72, 0.85, 0.90, and 0.93.

3.2. Experiment Setting

The experiment was conducted in two stages:
  • Feeding the entire time series of length 1024 from the dataset to the machine learning model and extending it by 1024 steps, allowing for the evaluation of the Hurst exponent of the continuations and the model’s ability to reliably extend the series for a large number of time steps. Since the dataset was prepared using Hosking’s algorithm, and the series is included entirely starting from the zero value and the first increment obtained by the algorithm, the models receive complete data in this case. The series depends only on itself and does not depend on past increments unknown to the model.
  • Feeding only the last 512 values out of the 1024 available in the dataset to the machine learning model. In this case, the true values depend on some unknown past that the model is not aware of. We extend the series by only 64 steps, which significantly reduces the computation time and allows for obtaining performance estimates of the models when they are not constrained by strict time frames.
Since time series have fractal properties, providing all 1024 (or 512) values as input to the models may not be efficient in terms of memory and computational performance constraints. Given the importance of the nearest dependency, it is reasonable to input a certain number of the most recent values (e.g., 8–16) in their entirety, while selecting every 8th (or every 16th, 32nd, or 64th, depending on the model’s memory requirements) value from the remaining length of the series.

3.3. Investigating the Properties of Continuations

3.3.1. Calculating the Hurst Exponent of Continuations

All continuations of the time series were evaluated using the Whittle method to determine the Hurst exponent and averaged. The results are presented in Table 2. For continuations of length 1024, the estimated Hurst exponent values were almost identical to the true values of the Hurst exponent for the original time series. The only exception is LightGBMLSS with the Gaussian_AutoGrad distribution when extending the series with the highest Hurst exponent H = 0.93 (in this case, all other indicators were also unsatisfactory).
As can be seen from the table, the lowest accuracy between the Hurst exponent of the original series and the predicted continuation was observed at high H values (0.90 and 0.93). At high Hurst exponent values (H > 0.85), the Hurst exponent values of the continuations deviate more significantly from the original time series values. This suggests that the methods may have more difficulty in capturing the long-range dependence of time series with higher Hurst exponent values.

3.3.2. Determining If the Increments Are Normally Distributed

The continuations obtained by all methods successfully passed the Anderson–Darling test [26] across the entire range of Hurst exponent values, except for CBU and LightGBMLSS with the Gaussian_AutoGrad distribution on high H values greater than 0.85.

3.3.3. Checking the Standard Deviation of the Increments

Another important aspect of evaluating the performance of the considered methods is to determine if the increments in the generated time series continuations Xc have a standard deviation equal to that of the original time series X, which in our case is equal to 1. This is crucial because it helps ensure that the methods maintain the same level of variability in the generated time series as in the original data.
All methods successfully extend the original time series, in which the standard deviation (SD) of the increments is equal to one, with continuations having SD of increments very close to one for Hurst exponent values below 0.85. For Hurst exponent values starting from 0.85, all methods show a decrease in the SD of increments, which is most noticeable at H = 0.93. In this case, CBU and LightGBMLSS with the Gaussian_AutoGrad distribution unsatisfactorily lower the SD to 0.2–0.4, while the other methods have SD of increments above 0.75 (Table 3).
This demonstrates that the two aforementioned methods (CBU and LightGBMLSS with the Gaussian_AutoGrad distribution) are the least adapted for extending fractal time series with long-range dependencies, while the other methods have relative issues only at the extreme Hurst exponent value of 0.93.

3.4. Investigating Forecasting Accuracy When Extending a Time Series of Length 1024 by 1024 Time Steps

In this experiment, when using the full evaluation dataset, not all methods were able to show results, as NGBoost, PGBM (even when using GPU), IBUG, Quantile Regression QuantReg (statsmodels package), and the quantile-forest package for Quantile Regression Forests turned out to be too slow to train sufficiently for comparable results.
We determine the deviation of forecasted quantiles Qdev using the “inter-quantile” measure (9). Lower values indicate better performance. Some methods provide high results only for H = 0.53 (i.e., very close to 0.5). These methods are LightGBMLSS Gaussian_AutoGrad and CBU (Table 4). However, when long-range dependencies are present in the time series, their results are significantly worse compared to other methods. Therefore, it is concluded that such methods are not suitable for the analysis of fractal time series.
An important result of the study is that Nonparametric Probabilistic Regression with Coarse Learners (PRESTO) [24] demonstrated comparable performance with the most successful LSS method (XGBoostLSS and LightGBMLSS with Gaussian distribution). The author of PRESTO primarily considered the application of decision trees as a classifier in his method. As the core idea does not limit us in the choice of classifier, we proposed the use of logistic regression from the sklearn package. Interestingly, this simplest approach turned out to be the most accurate in this experiment.
All methods were generally more accurate for Hurst exponent values below 0.85 (Figure 4).

3.5. Investigating Forecasting Accuracy When Extending a Time Series of Length 512 by 64 Time Steps

In this study, when providing only the last 512 values out of the 1024 available in the dataset, NGBoost showed very high results only for H = 0.53 (i.e., very close to 0.5), significantly lagging behind the leaders in other Hurst exponent values. A possible reason is the lack of features due to time constraints for this relatively slow method. Thus, this method may have potential, but its slowness is a significant obstacle to its use compared to other methods.
As in the previous study, CBU showed unsatisfactory results in capturing the long-range dependence of time series with higher Hurst exponent values. In addition, in the entire range except for H = 0.53, it significantly lagged behind the LSS method, modeled Gaussian distribution only, and is not attractive for the study of fractal time series.
Among the parametric methods, our choice is for the LSS method (XGBoostLSS is better when H > 0.6, and LightGBMLSS with Gaussian distribution is better when H < 0.6). The most significant interest is in the results of nonparametric methods.
The results are listed in Table 5.

3.6. Comparing Forecasting Accuracy Given Time Series of Length 1024 and 512

In this study, we compared the performance of the most successful model, PRESTO with Logistic Regression, when extending time series of different lengths. In all cases, the last 16 values of the time series were used as input, while the rest of the input varied between the experiments. The first experiment used one out of every 32 values of the original time series with a length of 1024, while the second experiment used one out of every 32 values of the first 512 values of the time series. The third experiment fed one out of every 16 values of the first 512 values to align the number of features. The results are listed in Table 6.
The results of this study show that the length of the input time series plays an important role when there are long-term dependencies (Hurst exponent significantly different from 0.5). When only half of the time series was fed to the model, it significantly underperformed in generating continuations, even when values were sampled more densely (every 16th value compared to every 32nd value). However, when the Hurst exponent was close to 0.5 (indicating an almost independent time series), the situation was reversed. Extra features from a longer time series were irrelevant and only worsened the prediction.

4. Discussion

By employing iterated approaches with one-period-ahead models, we can effectively generate numerous time series continuations that preserve the fractal properties of the original time series. This enables us to evaluate the performance of our model in capturing the complex, long-range dependencies, and self-similarity exhibited by fractal time series data. At the same time, it is essential to assess the quality of the continuations—whether they have the same measurable fractal or statistical characteristics—to guarantee the model’s reliability. In turn, this allows for a more comprehensive assessment of the model’s ability to provide accurate and reliable forecasts for time series with fractal properties, which is essential for various practical applications in fields such as finance, meteorology, and environmental studies.
The results suggest that machine learning methods can learn the distribution of the next increment of the time series, taking into account not only long-range dependence but also preserving the numerical values of the measurable indicator of this long-range dependence, namely, the Hurst exponent (Table 2). This provides a perspective that machine learning methods can be used not only for predicting quantiles of the distribution but also for generating plausible time series while preserving fractal and statistical properties.
Leveraging the fractal properties of time series data in this way allows for a reduction in computational demands and an increase in the training set size by reducing the number of features. The last 16 values of the time series were fed to the investigated model in their entirety, and to account for long-term dependencies, only every 8th (16th, 32nd) value of the remaining series was sufficient to obtain accurate predictions (Table 6).
Fractional Brownian motion time series exhibit standard (Gaussian) distribution of increments, making them ideally suited for forecasting using methods that assume a conditional distribution for the target variable, such as a normal distribution where both mean (μ) and standard deviation (σ) depend on the explanatory variables (X). It was anticipated that these methods would outperform nonparametric methods in forecasting tasks. The performance of parametric methods was investigated in comparison to nonparametric methods, providing valuable insights into the potential limitations of nonparametric approaches. Furthermore, it will enable practitioners to assess the risks associated with utilizing nonparametric methods for analyzing time series data exhibiting fractal properties when forecasting the distribution shape is challenging or not feasible.
It was found that not all parametric probabilistic methods perform well across the entire range of Hurst exponent values. The nonparametric PRESTO method proved to be comparable in effectiveness to parametric methods, while also being applicable for predicting time series with fractal characteristics whose probability distribution shape cannot be predetermined.
The main inspiring question of the study was to what extent parametric methods outperform nonparametric ones. Unexpectedly, the latest nonparametric method PRESTO showed itself quite acceptable, having the potential to study time series with any complex probability distribution of the next increment, and especially the ability to conduct detailed research and modeling of heavy-tailed distributions. Not to mention that it surpassed most parametric methods in terms of speed and accuracy.
This result highlights the potential of nonparametric methods such as PRESTO in the analysis and modeling of fractal time series, especially when the underlying probability distribution is unknown or complex. The flexibility and adaptability of nonparametric methods make them suitable for a wider range of applications and could lead to more accurate predictions and insights in various fields where fractal time series are relevant.
While parametric methods such as LSS (XGBoostLSS and LightGBMLSS with Gaussian distribution) showed promising results, the nonparametric PRESTO demonstrated that it could compete with these methods and even outperform them in certain situations. This finding could encourage further development and research into nonparametric methods, leading to more advanced and versatile tools for analyzing and modeling fractal time series.
The best performance was demonstrated by the PRESTO method with Logistic Regression as a classifier. The Quantile Regression method, QuantReg from the statsmodels package, also showed strong performance. This may be due to the fact that the time series was synthesized, meaning that it did not contain any outliers. Increments are normally distributed, and the task turned out to be perfectly suited for linear methods. In the case of real-world time series, other methods capable of capturing non-linear dependencies may demonstrate better performance.
In our analysis, we observed a specific trend. When the Hurst exponent is close to 0.5, the majority of models predict well. However, when the exponent is 0.85 or higher, the performance of all models deteriorates, some simply fail. This tendency is an indicative marker of the models’ quality, separating those that are effective from those that are not. The higher the Hurst exponent, the more challenging the prediction becomes. This is primarily because certain models struggle to capture long-term dependencies. While this is a challenge for all models, some particularly struggle. Thus, further research is needed to develop models that are capable of capturing long-term dependencies, especially in time series with high Hurst exponent values.
Parametric LSS methods (XGBoostLSS and LightGBMLSS with Gaussian distribution), demonstrated good performance in capturing long-range dependence in time series and extending them while maintaining Hurst exponent values. These methods were generally more accurate for Hurst exponent values below 0.85 (H < 0.85).
The nonparametric method PRESTO showed promising results, competing with the best parametric methods in terms of accuracy and speed. Its flexibility in modeling complex and unknown probability distributions makes it a suitable choice for a wide range of applications and indicates potential for further development.
Some methods, such as CBU, LightGBMLSS with Gaussian_AutoGrad, NGBoost, and PGBM, showed limitations in capturing long-range dependence in fractal time series, especially for high Hurst exponent values (>0.85). These methods may not be suitable for analyzing fractal time series or may require further refinement.
The best performance was demonstrated by the simplest methods, namely, PRESTO with Logistic Regression and Quantile Regression QuantReg from the statsmodels package.

5. Conclusions

This study aimed to evaluate the performance of various parametric and nonparametric probabilistic forecasting machine learning methods in extending the fBm time series with different Hurst exponent values. Self-similar properties of fBm time series have been reliably reproduced in the continuations of the time series predicted by machine learning methods. The values of the Hurst exponent have been preserved in the generated continuations. The increments of the continuations have had a normal distribution and their standard deviation has coincided with that of the original ones.
Overall, this study demonstrates that probabilistic forecasting machine learning methods allow the prediction of the dynamics of fractal time series while preserving their self-similarity and statistical characteristics. Furthermore, we propose a distinctive and effective approach to assess the performance of various probabilistic forecasting methods in predicting the fBm time series. This methodology involving the use of the Hosking algorithm for generating ground truth quantiles of possible continuations is novel.
The best performance was demonstrated by the method that converts the probabilistic regression problem into one of multi-classification—PRESTO, with Logistic Regression proving to be the most successful classifier. Additionally, the Quantile Regression method, QuantReg from the statsmodels package, also showed strong performance.
Further research and development in nonparametric methods could lead to more advanced and versatile tools for modeling and predicting fractal time series in various fields. The proposed approach enables the consideration of long-term dependencies and scaling properties of the data, leading to more accurate forecasts and a deeper understanding of the dynamics and structure of the investigated processes. The presented machine learning methods can be applied to various domains, including financial market forecasting, disease dynamics, seismic activity, telecommunication traffic load, and others.

Author Contributions

Conceptualization, methodology, investigation, validation, writing—original draft preparation, writing—review and editing, L.K. and R.L.; software, resources, data curation, visualization, R.L.; formal analysis, supervision, project administration, L.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are openly available here: https://www.kaggle.com/datasets/unfriendlyai/fractal-brownian-motion (accessed on 8 May 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feder, J. Fractals; Springer: New York, NY, USA, 2013. [Google Scholar]
  2. Radivilova, T.; Kirichenko, L.; Ageyev, D.; Tawalbeh, M.; Bulakh, V.; Zinchenko, P. Intrusion Detection Based on Machine Learning Using Fractal Properties of Traffic Realizations. In Proceedings of the 2019 IEEE International Conference on Advanced Trends in Information Theory (ATIT), Kyiv, Ukraine, 18–20 December 2019. [Google Scholar] [CrossRef]
  3. Brambila, F. (Ed.) Fractal Analysis—Applications in Physics, Engineering and Technology; InTech: London, UK, 2017. [Google Scholar] [CrossRef]
  4. Kirichenko, L.; Saif, A.; Radivilova, T. Generalized Approach to Analysis of Multifractal Properties from Short Time Series. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 183–198. [Google Scholar] [CrossRef]
  5. Burnecki, K.; Sikora, G.; Weron, A.; Tamkun, M.M.; Krapf, D. Identifying Diffusive Motions in Single-Particle Trajectories on the Plasma Membrane via Fractional Time-Series Models. Phys. Rev. E 2019, 99, 012101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Peters, E.E. Fractal Market Analysis: Applying Chaos Theory to Investment and Economics; Wiley: New York, NY, USA, 2009. [Google Scholar]
  7. Kirichenko, L.; Pavlenko, K.; Khatsko, D. Wavelet-Based Estimation of Hurst Exponent Using Neural Network. In Proceedings of the 2022 IEEE 17th International Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 10–12 November 2022. [Google Scholar] [CrossRef]
  8. Sidhu, G.S.; Ibrahim Ali Metwaly, A.; Tiwari, A.; Bhattacharyya, R. Short Term Trading Models Using Hurst Exponent and Machine Learning. SSRN Electron. J. 2021. [Google Scholar] [CrossRef]
  9. Pilgrim, I.; Taylor, R.P. Fractal Analysis of Time-Series Data Sets: Methods and Challenges. In Fractal Analysis; InTech Open: London, UK, 2018. [Google Scholar] [CrossRef] [Green Version]
  10. Kirichenko, L.; Radivilova, T.; Zinkevich, I. Forecasting Weakly Correlated Time Series in Tasks of Electronic Commerce. In Proceedings of the 2017 12th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 5–8 September 2017. [Google Scholar] [CrossRef] [Green Version]
  11. Huang, C.; Wang, J.; Chen, X.; Cao, J. Bifurcations in a Fractional-Order BAM Neural Network with Four Different Delays. Neural Netw. 2021, 141, 344–354. [Google Scholar] [CrossRef] [PubMed]
  12. Ou, W.; Xu, C.; Cui, Q.; Liu, Z.; Pang, Y.; Farman, M.; Ahmad, S.; Zeb, A. Mathematical Study on Bifurcation Dynamics and Control Mechanism of Tri-Neuron Bidirectional Associative Memory Neural Networks Including Delay. Math. Methods Appl. Sci. 2023. [Google Scholar] [CrossRef]
  13. Hou, H.; Zhang, H. Stability and Hopf Bifurcation of Fractional Complex–Valued BAM Neural Networks with Multiple Time Delays. Appl. Math. Comput. 2023, 450, 127986. [Google Scholar] [CrossRef]
  14. Garcin, M. Forecasting with Fractional Brownian Motion: A Financial Perspective. Quant. Financ. 2022, 2, 1495–1512. [Google Scholar] [CrossRef]
  15. Song, W.; Li, M.; Li, Y.; Cattani, C.; Chi, C.-H. Fractional Brownian Motion: Difference Iterative Forecasting Models. Chaos Solitons Fractals 2019, 123, 347–355. [Google Scholar] [CrossRef]
  16. Wang, J.; Liu, Y.; Wu, H.; Lu, S.; Zhou, M. Ensemble FARIMA Prediction with Stable Infinite Variance Innovations for Supermarket Energy Consumption. Fractal Fract. 2022, 6, 276. [Google Scholar] [CrossRef]
  17. Duan, T.; Avati, A.; Ding, D.Y.; Thai, K.K.; Basu, S.; Ng, A.Y.; Schuler, A. NGBoost: Natural Gradient Boosting for Probabilistic Prediction. arXiv 2020, arXiv:1910.03225. [Google Scholar] [CrossRef]
  18. März, A. XGBoostLSS—An Extension of XGBoost to Probabilistic Forecasting. arXiv 2019, arXiv:1907.03178. [Google Scholar] [CrossRef]
  19. März, A.; Kneib, T. Distributional Gradient Boosting Machines. arXiv 2022, arXiv:2204.00778. [Google Scholar] [CrossRef]
  20. Sprangers, O.; Schelter, S.; de Rijke, M. Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic Regression. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 1510–1520. [Google Scholar] [CrossRef]
  21. Brophy, J.; Lowd, D. Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees. arXiv 2022, arXiv:2205.11412. [Google Scholar] [CrossRef]
  22. Meinshausen, N. Quantile Regression Forests. J. Mach. Learn. Res. 2006, 7, 983–999. [Google Scholar]
  23. Koenker, R.; Hallock, K.F. Quantile Regression. J. Econ. Perspect. 2001, 15, 143–156. [Google Scholar] [CrossRef]
  24. Lucena, B. Nonparametric Probabilistic Regression with Coarse Learners. arXiv 2022, arXiv:2210.16247. [Google Scholar] [CrossRef]
  25. Banna, O.; Mishura, Y.; Ralchenko, K.; Shklyar, S. Fractional Brownian Motion; John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar] [CrossRef]
  26. Anderson, T.W.; Darling, D.A. A Test of Goodness of Fit. J. Am. Stat. Assoc. 1954, 49, 765–769. [Google Scholar] [CrossRef]
  27. Sato, K. Self-Similar Processes with Independent Increments. Probab. Theory Relat. Fields 1991, 89, 285–300. [Google Scholar] [CrossRef]
  28. Rao, P. Self-Similar Processes, Fractional Brownian Motion and Statistical Inference. Lect. Notes-Monogr. Ser. 2004, 45, 98–125. [Google Scholar] [CrossRef]
  29. Mandelbrot, B.B.; Van Ness, J.W. Fractional Brownian Motions, Fractional Noises and Applications. SIAM Rev. 1968, 10, 422–437. [Google Scholar] [CrossRef]
  30. Feder, J. Random Walks and Fractals. In Fractals. Physics of Solids and Liquids; Springer: Boston, MA, USA, 1988; pp. 163–183. [Google Scholar] [CrossRef]
  31. Tyralis, H.; Papacharalampous, G. A Review of Probabilistic Forecasting and Prediction with Machine Learning. arXiv 2022, arXiv:2209.08307. [Google Scholar] [CrossRef]
  32. Shang, H.L. A Comparison of Hurst Exponent Estimators in Long-Range Dependent Curve Time Series. J. Time Ser. Econom. 2020, 12, 20190009. [Google Scholar] [CrossRef]
  33. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamzadeh, M.; Fieguth, P.; Cao, X.; Khosravi, A.; Acharya, U.R.; et al. A Review of Uncertainty Quantification in Deep Learning: Techniques, Applications and Challenges. Inf. Fusion 2021, 76, 243–297. [Google Scholar] [CrossRef]
  34. Papacharalampous, G.; Tyralis, H. A Review of Machine Learning Concepts and Methods for Addressing Challenges in Probabilistic Hydrological Post-Processing and Forecasting. Front. Water 2022, 4, 961954. [Google Scholar] [CrossRef]
Figure 1. Summary of our approach. We compute multiple continuations of a single fBm time series using the Hosking algorithm, resulting in a dataset with ground truth quantiles of possible continuations. Using a one-period-ahead model, iterated forward for the desired number of periods, we can predict quantiles of continuations for fBm time series over hundreds or thousands of time steps. By comparing the predicted quantiles with the ground truth quantiles, we can evaluate the performance of the chosen machine learning method.
Figure 1. Summary of our approach. We compute multiple continuations of a single fBm time series using the Hosking algorithm, resulting in a dataset with ground truth quantiles of possible continuations. Using a one-period-ahead model, iterated forward for the desired number of periods, we can predict quantiles of continuations for fBm time series over hundreds or thousands of time steps. By comparing the predicted quantiles with the ground truth quantiles, we can evaluate the performance of the chosen machine learning method.
Fractalfract 07 00517 g001
Figure 2. Visualization of an example of the proposed dataset. Common beginning of length 1024 and quantiles of its continuation length 1024 (quantiles 0.99, 0.95, 0.84, 0.50).
Figure 2. Visualization of an example of the proposed dataset. Common beginning of length 1024 and quantiles of its continuation length 1024 (quantiles 0.99, 0.95, 0.84, 0.50).
Fractalfract 07 00517 g002
Figure 3. The figure illustrates the varying distances between adjacent quantiles for different time steps and distances from the median. To interpret the deviation Qdev in the “inter-quantile” measure, normalization is performed by dividing absolute deviation by the difference between adjacent quantiles.
Figure 3. The figure illustrates the varying distances between adjacent quantiles for different time steps and distances from the median. To interpret the deviation Qdev in the “inter-quantile” measure, normalization is performed by dividing absolute deviation by the difference between adjacent quantiles.
Fractalfract 07 00517 g003
Figure 4. Quantile prediction was produced using PRESTO with the Logistic Regression method for Hurst exponent values 0.65 (a) and 0.90 (b). At high Hurst exponent values (b) the predicted quantiles of the continuations deviate more significantly from the ground truth quantiles.
Figure 4. Quantile prediction was produced using PRESTO with the Logistic Regression method for Hurst exponent values 0.65 (a) and 0.90 (b). At high Hurst exponent values (b) the predicted quantiles of the continuations deviate more significantly from the ground truth quantiles.
Fractalfract 07 00517 g004
Table 1. Probabilistic machine learning methods and corresponding implementations.
Table 1. Probabilistic machine learning methods and corresponding implementations.
MethodOutputImplementation
Natural gradient boosting NGBoost [17]Estimates of the mean and standard deviation of the normal distributionPython package ngboost
Catboost with Uncertainty (CBU) Estimates of the mean and standard deviation of the normal distributionPython package catboost
Mean, location, scale, and shape (LSS) XGBoostLSS and LightGBMLSS [18,19]Estimates of the mean and standard deviation of the normal distributionAlexander März’s
(StatMixedML) repository on github
Nonparametric Probabilistic Regression (PRESTO) [24]Distribution without predefined shapeOur implementation of the author’s idea using CatBoost, LightGBM, and Scikit-learn’s Logistic Regression
Training a separate model for each quantileQuantiles predictionsPython packages catboost, lightgbm
Quantile regression (QuantReg) [23]Quantiles predictionsPython package statsmodels
Random Forest Quantile Regression [22]Quantiles predictionsPython package sklearn_quantile
Table 2. Comparison of Hurst Exponent Accuracy at different H Values of original time series. * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis.
Table 2. Comparison of Hurst Exponent Accuracy at different H Values of original time series. * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis.
MethodH = 0.3H = 0.35H = 0.45H = 0.53H = 0.6H = 0.65H = 0.72H = 0.85H = 0.9H = 0.93
CBU 500 iter.0.2990.3490.4510.5290.60.6490.7180.8480.8950.935
CBU 1000 iter.0.2990.3510.4490.5290.5990.6490.7190.8480.8970.935
LightGBMLSS Gaussian 0.2980.3490.4530.5280.5980.6480.7110.8430.8950.926
LightGBMLSS Gaussian_AutoGrad 0.2970.3490.4510.5310.6020.6540.7130.8490.8950.83 *
XGBoostLSS Gaussian0.3010.3560.4470.5340.5970.6510.7190.8440.90.922
PRESTO with CatBoost 0.2980.3490.4490.5290.5980.6480.7180.8460.8940.925
PRESTO with LightGBM 0.3020.3520.4520.5270.5950.6440.7130.8410.8890.922
PRESTO with Logistic Regression 0.2990.350.4480.5290.5990.6490.7190.8480.8970.928
CatBoost for each quantile0.3060.3570.4540.5250.5930.6430.7130.840.8860.914
LightGBM for each quantile0.3060.3570.4540.5250.5930.6440.7130.8390.8860.916
Table 3. The standard deviation of increments of continuations (expected to be close to 1). * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis.
Table 3. The standard deviation of increments of continuations (expected to be close to 1). * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis.
MethodH = 0.3H = 0.35H = 0.45H = 0.53H = 0.6H = 0.65H = 0.72H = 0.85H = 0.9H = 0.93
CBU 500 iter.0.9560.9740.9960.9980.9860.9650.921 *0.75 *0.573 *0.418 *
CBU 1000 iter.0.9550.9730.9950.9970.9850.9650.919 *0.753 *0.578 *0.414 *
LightGBMLSS Gaussian 0.9960.9971.0030.9981.0020.9960.9970.9750.8920.778
LightGBMLSS Gaussian_AutoGrad 0.9950.9960.9961.0011.0020.9960.9980.9810.9020.274 *
XGBoostLSS Gaussian0.9960.9980.9980.9980.9980.9970.9950.9760.8990.776
PRESTO with CatBoost 11.0011.0011.0011.00210.9990.9860.8970.786
PRESTO with LightGBM 1.0011.0021.0021.0021.0021.0010.9990.9780.9030.795
PRESTO with Logistic Regression 1.0011.0011.0021.0021.0021.00110.9770.9060.789
CatBoost for each quantile1.0041.0041.0041.0041.0051.0031.0020.9830.9050.809
LightGBM for each quantile1.0041.0041.0041.0041.0041.0031.0020.9830.9070.814
Table 4. The deviation of forecasted quantiles Qdev when extending a time series of length 1024 by 1024 time steps. * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis. The boldface used to highlight the models with the best performance in the table.
Table 4. The deviation of forecasted quantiles Qdev when extending a time series of length 1024 by 1024 time steps. * The red mark signifies an outlier, a data point that deviates significantly from other observations, and the marking was done intentionally for emphasis. The boldface used to highlight the models with the best performance in the table.
MethodH = 0.3H = 0.35H = 0.45H = 0.53H = 0.6H = 0.65H = 0.72H = 0.85H = 0.9H = 0.93
CBU 500 iter.1.8541.3310.7410.5441.3391.6142.8056.659 *10.597 *14.594 *
CBU 1000 iter.1.8031.2710.710.9631.1271.6043.026.289 *10.706 *14.917 *
LightGBMLSS Gaussian 1.4051.1750.9141.2561.0772.1343.1422.744.4776.573
LightGBMLSS Gaussian_AutoGrad 1.441.241.1130.5941.2232.5141.9713.7414.59436.382
XGBoostLSS Gaussian1.4931.1740.9550.8611.2361.6761.97511.028 *4.496.806
PRESTO with CatBoost 1.6061.5571.3071.6682.0351.6692.3553.9764.6096.925
PRESTO with LightGBM 1.621.4080.9751.0581.5121.5791.993.7035.6518.709
PRESTO with Logistic Regression 1.3191.0240.6470.7291.851.1741.411.8563.3615.863
CatBoost for each quantile2.3381.9781.261.3341.6022.7472.9647.037 *5.9849.793
LightGBM for each quantile2.1011.8461.0590.9231.5363.0143.7674.7425.9619.713
Table 5. The deviation of forecasted quantiles Qdev when extending a time series of length 512 by 64 time steps. The boldface used to highlight the models with the best performance in the table.
Table 5. The deviation of forecasted quantiles Qdev when extending a time series of length 512 by 64 time steps. The boldface used to highlight the models with the best performance in the table.
MethodH = 0.3H = 0.35H = 0.45H = 0.53H = 0.6H = 0.65H = 0.72H = 0.85H = 0.9H = 0.93
NGBoost2.5761.7561.020.81.3861.7822.1684.1634.197.863
CBU2.1781.9661.9841.6422.1462.1772.0366.93810.68813.962
XGBoostLSS Gaussian1.1230.8380.7810.8311.2071.1761.0252.1152.7516.001
LightGBMLSS Gaussian 1.1111.0660.5890.5991.3881.0421.0684.6073.20332.828
PRESTO with CatBoost 1.4991.1741.5651.491.7941.2191.5761.9773.5566.048
PRESTO with LightGBM 0.9950.8040.7270.7670.7741.0341.0771.8953.0225.783
PRESTO with Logistic Regression 0.940.7470.5050.5020.6690.9051.0771.6262.7475.839
LightGBM for each quantile1.1651.0220.6180.7981.0451.0131.2621.9312.9715.592
CatBoost for each quantile1.3181.1780.7640.7951.021.2961.8312.7894.2568.73
QR statsmodels1.0960.740.6890.6110.7690.8781.1351.4672.9615.53
Random Forest QR1.6291.5021.551.6391.3071.5791.6665.6724.136.45
Table 6. Comparison of forecasting quantiles using different lengths of the predictor.
Table 6. Comparison of forecasting quantiles using different lengths of the predictor.
FeaturesH = 0.3H = 0.35H = 0.45H = 0.53H = 0.6H = 0.65H = 0.72H = 0.85H = 0.9H = 0.93
1/32 of 10240.6590.6480.5050.6560.660.560.5440.862.3945.204
1/16 of 5120.9860.7020.5270.5370.6780.8581.0311.6012.9155.521
1/32 of 5121.0360.7220.5510.4660.6330.9491.0211.793.0195.348
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kirichenko, L.; Lavrynenko, R. Probabilistic Machine Learning Methods for Fractional Brownian Motion Time Series Forecasting. Fractal Fract. 2023, 7, 517. https://doi.org/10.3390/fractalfract7070517

AMA Style

Kirichenko L, Lavrynenko R. Probabilistic Machine Learning Methods for Fractional Brownian Motion Time Series Forecasting. Fractal and Fractional. 2023; 7(7):517. https://doi.org/10.3390/fractalfract7070517

Chicago/Turabian Style

Kirichenko, Lyudmyla, and Roman Lavrynenko. 2023. "Probabilistic Machine Learning Methods for Fractional Brownian Motion Time Series Forecasting" Fractal and Fractional 7, no. 7: 517. https://doi.org/10.3390/fractalfract7070517

Article Metrics

Back to TopTop