Next Article in Journal
Scalable Structure for Chiral Quantum Routing
Previous Article in Journal
Thermodynamic Limit of Electroluminescent Refrigeration Devices
Previous Article in Special Issue
Does Excellence Correspond to Universal Inequality Level?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Iterative Forecasting of Financial Time Series: The Greek Stock Market from 2019 to 2024

by
Evangelos Bakalis
* and
Francesco Zerbetto
Dipartimento di Chimica “G. Ciamician”, Università di Bologna, V. P. Gobetti 85, 40129 Bologna, Italy
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(5), 497; https://doi.org/10.3390/e27050497
Submission received: 31 March 2025 / Revised: 28 April 2025 / Accepted: 2 May 2025 / Published: 4 May 2025
(This article belongs to the Special Issue Entropy-Based Applications in Sociophysics II)

Abstract

:
Predicting the evolution of financial data, if at all possible, would be very beneficial in revealing the ways in which different aspects of a global environment can impact local economies. We employ an iterative stochastic differential equation that accurately forecasts an economic time series’s next value by analysing its past. The input financial data are assumed to be consistent with an α -stable Lévy motion. The computation of the scaling exponent and the value of α , which characterises the type of the α -stable Lévy motion, are crucial for the iterative scheme. These two indices can be determined at each iteration from the form of the structure function, for the computation of which we use the method of generalised moments. Their values are used for the creation of the corresponding α -stable Lévy noise, which acts as a seed for the stochastic component. Furthermore, the drift and diffusion terms are calculated at each iteration. The proposed model is general, allowing the kind of stochastic process to vary from one iterative step to another, and its applicability is not restricted to financial data. As a case study, we consider Greece’s stock market general index over a period of five years, from September 2019 to September 2024, after the completion of bailout programmes. Greece’s economy changed from a restricted to a free market over the chosen era, and its stock market trading increments are likely to be describable by an α -stable L’evy motion. We find that α = 2 and the scaling exponent H varies over time for every iterative step we perform. The forecasting points follow the same trend, are in good agreement with the actual data, and for most of the forecasts, the percentage error is less than 2%.

1. Introduction

A range of companies, securities, and other economic indices are combined to form a stock market’s general composite index (GCI). It is a marker that can convey a lot of information in an intelligible manner, and market performance can be evaluated by observing its behaviour. Working with financial data, if at all possible, it would be ideal to forecast the values of this index and thus its rate of change [1].
A financial time series is made up of the daily values of a GCI, but it is not only limited to those with a time lag of one day. Additional financial time series can be constructed by adopting a different time lag between two consecutive values. Assume that the opening or closing price of GCI on day i is x i . The logarithmic difference of x i over two consecutive days is r i = l o g ( x i x i 1 ) , which is the definition of the daily price return. One of the most important factors for stock market index analysis is volatility, assuming that stock market indexes are random variables [2]. Volatility measures dispersion around a central tendency and is commonly defined as the deviation from the mean. It is calculated as the variance of the daily price return, and it is a second-order statistic. As such, given that many random processes have fat tails or no mean due to outliers, its applicability is called into question [3]. However, volatility remains one of the most used methods in researching financial time series in conjunction with entropic contributions [4,5].
Financial markets can be viewed as complex, dynamic systems for which a wealth of data are available. By applying ideas from physics, financial time series can be viewed from a different perspective, potentially producing new discoveries [6]. Finding scaling exponents is a fundamental physics task in stationary or non-stationary stochastic processes. In fact, the search for these exponents in financial markets is presently a very rich field [7,8,9,10,11,12], where the growing body of empirical data is constantly enhancing our comprehension of their behaviours.
The mathematical foundation for the rescaled analysis was established by Mandelbrot and Wallis using a heuristic graphical method [13]. They demonstrated that a scaling property for stationary processes permits a relationship between past and future occurrences. Let x ( t ) be a stationary time series and τ be the length of the window that correlates differences between past and future events. Then, if there is a unique scaling, the formula x ( τ ) = c τ H holds true, where c is a positive constant and H is the scaling exponent, also known as the Hurst exponent in honour of Harold Edwin Hurst [14]. The Hurst exponent, H ( 0 , 1 ) , is used to characterise a stochastic process. The process is anti-persistent for 0 < H < 1 2 , which means that each new value is likely to go in the opposite direction of the one before it. The process is referred to as persistent for 1 2 < H < 1 , and each new value probably follows the path of the preceding one. The process is memoryless for H = 1 2 , and each new value is completely independent of the preceding values. For financial time series, a value of H > 0.5 reflects an economy at a low stage of development, as opposed to H < 0.5 , indicating an economy at an advanced stage of development. A financial time series is deemed monofractal if it is characterised by a single value of the Hurst exponent, but usually financial time series are better described as multifractals since they are described by multiple scalings [9,12]. Over time, numerous models have been developed in an attempt to pinpoint traits of stochastic processes, and they have been applied to the analysis of data in a range of scientific fields, including finance [3,6,15,16,17,18,19,20].
The finding of the scaling exponent of a self-similar process (Hurst exponent) is a widely used approach in econophysics [6,8] to understand the dependence of financial time series data on one another and the volatility of the stock market [13]. Back in 1963, Mandelbrot proposed an α -Lévy stable distribution as a suitable model for analysing price differences or logarithmic returns [7]. An α -Lévy stable distribution serves as a general framework for other widely used distributions, such as the log-normal that Scholes and Black [2] suggested for stock price forecasting. The latter, also referred to as Geometrical Brownian Motion (GBM), is once again α -Lévy stable with values of α = 2 and H = 0.5 , and it is a memoryless process. A statistical analysis of the Milan, Italy, stock exchange indices revealed that the price indices are well described by α -Lévy stable distributions because the financial data’s properties are compatible with a Lévy random walk [21]. Moreover, the analysis of European options data revealed that their multifractality can be attributed to an α -stable Lévy distribution [10]. As a consequence of real data analyses, fractional Brownian motion (fBm) [22] and fractional Lévy motion (fLm) [23] have been used as row models in the analysis of economic data, showing that financial data are more appropriately described as multifractals than monofractals (fBm) [10,23,24,25,26,27,28].
Employing an α -Lévy stable distribution as a model from which the increments or the logarithmic returns draw values, it is crucial to calculate from the row data the α index and the scaling exponent H 1 2 + 1 α [26]. The Hurst exponent can be calculated using a variety of techniques, such as power spectral density (PSD) [29], rescaled range analysis (R/S) [13], detrended fluctuation analysis (DFA) [16], dispersional analysis (DA) [30], multifractal detrended fluctuation analysis (MF-DFA) [17,31], generalised Hurst exponent (GHE) [3,32], and the generalised moments method (GMM) [19,33,34], to name some of them. The accuracy of each approach in delivering the scaling exponent depends on the length of the data and the type of distribution the data satisfy (whether or not heavy tails exist). Since PSD, DFA, R/S, and DA generate a unique scaling exponent (monofractality), it is unclear if they can “capture” the features of a α -stable Lévy distribution. Furthermore, R/S has been shown to overestimate the genuine Hurst exponent [35]. Additionally, GHE performs the best, while MF-DFA is unsuitable for data with smaller sample sizes and heavier tails [18].
In this article, we outline a path that can result in financial time series forecasting using an iterative scheme. A stochastic differential equation with a drift and a diffusion term serves as the starting point [2,27,28,36]. The drift term can be calculated as the mean of the increments of the process weighted by the actual value of the process. The calculation of the diffusion term is more demanding since it requires the autocorrelation of the underlying noise originating from the discretisation of an α -Lévy stable motion. We use the GMM method [19,33,34] to classify the type of motion. It is robust in providing the scaling exponent and is similar to the generalised Hurst exponent method (GHE), which is commonly used for financial data analysis [3]. In addition, GMM is one of the most reliable techniques for analysing non-stationary time series [19], and it also performs well for short time series [18,37]. The structure function [20], whose form yields the values of the Hurst exponent and the index α , is the method’s strength, aside from the fact that it may be applied even to short data sets. For each forecast, let the trading day be the kth GMM analysis of the previous k 1 values. The computed α and H are then used to generate the α -Lévy stable noise sequence of equal length and zero mean, whose value, after numerous iterations, contributes to the actual value of the index for the day k. The values of α and H may vary at every prediction point. It is important to note that as part of the data preprocessing procedure, we calculate the historical data sample entropy [38,39], whose value can reveal whether the time series behaves as a scale-free process [40] or is predictable or regular.
The current work is organised as follows: The application of the generalised moments approach to time series and the retrieved data is covered in Section 2.1. Section 2.2 discusses fractional Lévy stable motion and self-similar processes. Stochastic differential equations are covered in Section 2.3. Section 3 presents the results and discussions, and Section 4 summarises.

2. Materials and Methods

2.1. Time Series and Generalised Moments Method

Let { x i } with i = 1 , 2 , , N be a process or the measure of a field we study. We are interested in understanding how the fluctuations of the field behave, and for this purpose we define the increments of the field. In order to reveal scaling properties, where scaling is defined as the invariance of a process’s properties across a range of scales, we create many intermediate scales, for example, by dividing the field into non-overlapping intervals of certain resolution or time lag Δ , and producing new processes defined as y n ( Δ ) = | x ( n + Δ ) x ( n ) | . Scaling can be described using both fractal and multifractal theories. A unique scaling, or monofractal, is the focus of fractal theory [41] and only a few parameters are required to describe complex phenomena. Nonetheless, multifractal theory addresses multiscaling, which enables the scaling characteristics of a process to be generalised [42,43]. The scaling properties of the field can be determined either by the scaling of its probability distribution, P r ( y Δ > Δ α ) Δ f ( α ) or equivalently by the scaling of its different moments scaling function < y Δ > q Δ z ( q ) [20].
The scaling exponent z ( q ) , also referred to as the structure function, determines whether a process is monofractal (has a linear dependence of z ( q ) on q) or multifractal (has a convex shape). Among multifractals, universal multifractals, which essentially are a log-Lévy multifractals, are ubiquitous [20], and the scaling exponent reads
z ( q ) = q h C α 1 ( q α q )
where α = 1 C d 2 z ( q ) d q 2 | q = 1 is the Lévy index or index of multifractality. It takes values in the interval 0 < α 2 . For α = 2 , z ( q ) = q h C ( q 2 q ) , the logarithm of the field is normally distributed [44], and Equation (5) depicts the characteristic function of a Gaussian distribution. For α = 1 , z ( q ) = q h C q l o g ( q ) , and the field is distributed according to log-Cauchy. For all the other values of α in the range (0, 2), the field is distributed according to log-Lévy. The term C is called co-dimension information and measures the mean intermittency and is defined as C = d z ( q ) d q | q = 1 H . It takes values in the range C [ 0 , d ] , with d being the dimension of the support, which is 1 for one-dimensional time series. For C = 0 , the field is homogeneous; only one scale exists. From lower C to higher values of C, the degree of intermittency increases, and some extreme outliers will occur. The parameter h ( 0 , 1 ) , when the field is multifractal, shows the degree of fractional integration (persistent for h > 0.5 and antipersistent for h < 0.5 ). The value of z ( q ) for q = 1 provides the Hurst exponent for a multifractal field. Since z ( q ) = q h , it follows that z ( q = 1 ) = z ( q = 2 ) / 2 when a field is monofractal [45]. Therefore, the analytical form of the scaling exponent or structure function, Equation (1), gives information about the nature of the stochastic mechanisms; see, for details, refs. [19,31,33,34,45,46,47].
Let us assume that { x n } is a financial time series with the index n = 1 , 2 , , N to stand for the day, and N is the total number of recordings. We consider here as a minimum time lag one day, but it can also be smaller or greater, depending on the type of recordings. If { x n } is a self-similar process, then it is fully described by the scaling exponent (Hurst) [22,23]. We expect that such processes, when zoomed in or zoomed out, will reveal the same patterns scaled by a certain amount, { x a n } = d { a H x n } , where { . . } = d { . . } stands for the equality of finite dimensional distributions, and H is the scaling exponent. Let t and s with ( t > s ) be two distinct time moments, then it can be pointed out that x ( t ) x ( s ) | t s | H ; see ref. [31] for the proof. Taking into account that the various moments can scale differently, we write x ( t ) x ( s ) q | t s | q H ( q ) , with H ( q ) being the generalised Hurst exponent through which the structure function is defined z ( q ) = q H ( q ) . For discrete data sets, the time difference t s describes the time lag whose maximum value should be small with respect to the total length of the time series; one tenth of the total length is a safe choice. For the application of the GMM, we follow the steps:
(i)
The construction of time series characterized by different lag times, Δ , which contain the absolute change of the values between two points of the initial time series y n ( Δ ) = x ( n + Δ ) x ( n ) for n = 1 , 2 , , ( N Δ ) and for Δ = 1 , 2 , , N 10 , where Δ = N 10 is the maximum lag time.
(ii)
Calculation of the moments of various orders q for q > 0 , including also fractional orders m ( q , Δ ) = 1 T Δ n = 1 T Δ y n ( Δ ) q . Notice that the small order moments, 0 < q < 2 , are responsible for the core of the probability density function (pdf), and higher moments contribute to tails of a pdf [48].
(iii)
Finding the scaling exponent for each moment. Assuming that m ( q , Δ ) Δ q H ( q ) , then the slope of the linear regression in log–log scale returns the value of H ( q ) for each q.
(iv)
Construction of the structure function. Having the values of the term q H ( q ) for the various moments, fit the data with Equation (1) and find the parameters H , α , C .

2.2. Self-Similar Stochastic Models and Fractional Lévy Stable Motion

Typically, random processes are not stationary, as opposed to those that are stationary, for which it holds true that ( x ( t + b ) x ( t ) ) = d ( x ( b ) x ( 0 ) ) b 0 with { x ( t ) } t 0 . Those with stationary increments are a significant class among them. Fractional Brownian motion B H ( t ) (fBm) is the term used to describe a non-stationary and self-similar process whose increments are stationary and also derived from a Gaussian distribution as defined by Mandelbrot and van Ness [22]:
B H ( t ) : = 1 Γ ( H + 1 2 ) 0 t ( t τ ) + H 1 2 d B ( τ ) + 0 ( ( t τ ) + H 1 2 ( τ ) + H 1 2 ) d B ( τ )
where ( x ) + = m a x ( x , 0 ) . Since input data are typically discrete rather than continuous, it is preferable to create a process that contains the increments of the initial process. The one-step increment for a continuous fBm is called fractional Gaussian noise b H ( t ) (fGn). It is a self- similar stationary process; its sum up to the nth event yields the fBm valid at that time, and it is defined as
b H ( t ) = 1 Γ ( H + 1 2 ) t ( t + 1 τ ) H 1 2 ( t τ ) H 1 2 w ( τ ) d τ
with w ( t ) being a collection of independent and identically distributed (i.i.d.) Gaussian random variables (white noise), and Γ ( ) being the Euler’s gamma function. The autocovariance function of fGn reads
< b H ( t ) b H ( t ) > = Γ ( 1 2 H ) c o s ( π H ) 2 π H { ( t t + 1 ) 2 H + ( t t 1 ) 2 H 2 ( t t ) 2 H }
The properties of fGn depend on the definition of fBm. There are three definitions, Kolmogorov, Lévy, and Mandelbrot van Ness (see Equation (2) for the latter), and they do not share the same properties [49,50]. If a process is characterised as fBm, then the structure function, Equation (1), takes the simple form z ( q ) = q H , which easily determines the Hurst exponent [19,33,45]. In addition, the increments of fBm are drawn from a Gaussian distribution which serves as a mother process to generate the various fGn or fBm under integration/differentiation [51]. Gaussian distributions are stable distributions that are maintained under linear transformations and have been demonstrated to be effective models for highly variable data [23].
The term “stable” refers to a significant class of probability distributions with intriguing characteristics, in which any linear combination of two i.i.d. random variables with a common distribution can be linearly transformed to have the same distribution [23,24]. Stable distributions are also known as α -stable Lévy distributions. An α -stable distribution, L α ( X ; μ , β , σ ) , requires four parameters for its complete description. α ( 0 , 2 ] is the index of stability or Lévy index. There is a greater chance of outliers and heavier tails when α takes low values. β [ 1 , 1 ] is a skewness parameter; a left-skewed distribution has a negative value of β , while a right- skewed distribution has a positive value. For β = 0 , the distribution is symmetric, and its characteristic function describes a stretched exponential centred around its mean or location parameter μ R ; see Equation (5) below. Finally, σ R + * is a scale parameter. The probability density function is merely scaled and offset by the scale and location parameters; refer to ref. [24] for additional information on stable distributions. For α < 2 , the distribution has undefined variance, while for α 1 , it has an undefined mean. For α = 2 and for α = 1 , the normal and the Cauchy distributions are returned.
An α -stable distribution is defined through its characteristic function Φ ( k ) , which is the Fourier transform of the probability distribution, Φ ( k ) = P ( k ; α , β , μ , σ ) = e i k X P ( X ; α , β , μ , σ ) and always exists for any real-valued random variable as opposed to the moments generating function. The Φ ( k ) of an α -stable distribution reads [23]
l n ( Φ ( k ) ) = σ α | k | α { 1 i β s i g n ( k ) t a n ( π α 2 } + i μ k α 1 σ | k | { 1 + i β s i g n ( k ) 2 π l n | k | } + i μ k α = 1
The symmetric α -stable Lévy motion, also known as Lévy stable motion (Lsm) L α = { L α ( t ) , t 0 , is a Markovian stochastic process that starts at 0. It has stationary increments, and it is self-similar L α ( k t ) = Δ k 1 / α L α ( t ) with scaling exponent H = 1 / α . For α = 2 , Lsm returns the ordinary Brownian motion L 2 ( t ) = Δ B ( t ) (actually L 2 ( t ) = 2 B ( t ) ). Similar to the generalisation of Brownian motion to fractional Brownian motion (fBm) by Mandelbrot and van Ness [41], we define the fractional Lévy stable motion (fLsm) as the following integral [23]:
L α H ( t ) = { c f α , + H ( t , τ ) + d f α , H ( t , τ ) } M ( d τ )
where f α , + H ( t , τ ) = ( t τ ) + H 1 / 2 ( τ ) + H 1 / 2 , f α , H ( t , τ ) = ( t τ ) H 1 / 2 ( τ ) H 1 / 2 , ( x ) + = m a x ( x , 0 ) , ( x ) = m i n ( x , 0 ) , M is an α -stable random measure in R. fLsm is a self-similar with stationary increments process (H-sssi) L α H ( k t ) = k γ L α H ( t ) with γ = H 1 α and H [ 1 α , 1 ) [36,52]. For H = 1 / α , fLsm becomes an ordinary α -stable Lévy motion (Lsm) L α ( t ) . The self-similarity exponent of Lsm lies in the range ( 1 2 , ) , and for 1 α < 2 , Lsm is not the only self-similar process with stationary increments, as opposed to 0 < α < 1 and α = 2 , where it is the only one [23]. The increments of fLsm, Equation (6), form the fractional Lévy noise l α H ( t ) that presents long-range dependence for γ > 0 , and negative dependence for γ < 0 . For γ = 0 , the increments are i.i.d. symmetric α -stable variables:
l α H ( t ) = L α H ( t + 1 ) L α H ( t )
A process driven by fLsn is anticipated to not have moments for q > α . However, for finite-size discrete data sets, all moments, independent of the order q, exist. As a result of that, Equation (1) can be used to deliver the detailed form of the structure function. It has been argued that when z ( q = 1 ) is different from z ( q = 2 ) / 2 , any additive model cannot be applied, and unavoidably, multifractal behaviour emerges [20]. However, such a result can be understood as being a bi-fractal behaviour result of a truncated Lévy motion, either standard or fractional [53].

2.3. Stochastic Differential Equation

The stochastic differential equation for Lévy stable motion is given by [23,54]:
d x α ( t ) = μ ( t , x α ( t ) ) d t + σ ( t , x α ( t ) ) d L α ( t )
where x α ( t = 0 ) = x α , 0 is a random variable that takes values from an α -Lsm, the measure μ ( t , x ( t ) ) describes the drift term, and σ ( t , x ( t ) ) stands for the diffusion term. Equation (8) can also be presented in integral form
x α ( t ) = x α , 0 + 0 t μ ( τ , x α ( τ ) ) d τ + 0 t σ ( τ , x α ( τ ) ) d L α ( τ ) , t 0
In line with the derivation of the stochastic differential equation for fractional Brownian motion [55], and its integral representation, Equation (9), which is based on fractional Ito’s formula [27], the integral form of a process driven by an α -fLsm is written as [28]
x α H ( t ) = x α , 0 H + 0 t μ ( τ , x α H ( τ ) ) d τ + 0 t σ ( τ , x α H ( τ ) ) d L α H ( τ ) , t 0
where x α , 0 H is a random variable drawing values from an α -fLsm. The differential form of Equation (10) is given by an equation like Equation (8), where we replace x α ( t ) with x α H ( t ) and d L α ( t ) with d L α H ( t ) . It should be noted that the integral of the stochastic term in both Equations (9) and (10) requires careful handling [28,56].
The drift and diffusion terms, in a first approximation, can be expressed as μ ( t , x ( t ) ) = μ x ( t ) and σ ( t , x ( t ) ) = σ x ( t ) , respectively. Consequently, we can express in discrete space a stochastic differential equation that can prospectively forecast future values using an iterative scheme, but we require an estimate of μ and σ to do so. Let ξ be a generic process that takes values from a probability distribution; it can represent fLsm, fbm, or multifractional Brownian motion (mfBm). For ξ , the discrete equivalent of Equation (8) is as follows:
Δ x ( t ) = μ x ( t ) Δ t + σ x ( t ) Δ ξ ( t )
where Δ t is the minimum time lag related to the resolution of the available data set. Moreover, we consider that ξ ( t ) has stationary increments with zero mean, < Δ ξ ( t ) > = 0 , and variance that remains to be specified and is related to the type of the process. Notice that we can also write Δ ξ ( t ) = w ξ ( t ) ( Δ t ) s , where w ξ ( t ) is a white noise that draws values from ξ ( t ) and s is the self-similarity exponent, H for fBm, H 1 2 + 1 α for fLsm, and H ( t ) for mfBm [27]. We divide each term of Equation (11) by x ( t ) and write
Δ x ( t ) x ( t ) = μ Δ t + σ Δ ξ ( t )
Taking the expectation value of Equation (12), and since < Δ ξ ( t ) > = 0 , we find
μ = < Δ x ( t ) x ( t ) > = 1 N n = 1 N Δ x ( t n ) x ( t n )
with N being the length of the data at time t. In addition, we calculate the expectation of the term < Δ x ( t ) x ( t ) Δ x ( t ) x ( t ) > and we find that that equals to μ 2 + σ 2 A C F Δ ξ ( | t t | ) , where A C F Δ ξ ( | t t | ) is the autocorrelation of Δ ξ ( t ) . Combining our findings, we end up with
σ = ( A C F Δ ξ ( 0 ) ) 1 / 2 < Δ x ( t ) x ( t ) 2 > < Δ x ( t ) x ( t ) > 2
with A C F Δ ξ ( 0 ) being the value of the autocorrelation function for t = t . Equation (11) for a generic stochastic motion ξ ( t ) in conjunction with Equations (13) and (14) is the main equation utilised in the iterative forecast.

3. Results and Discussion

As a case study, we consider Greece’s stock market general index over a period of five years, from September 2019 to September 2024, after the completion of bailout programmes. Greece’s economy changed from a restricted to a free market over the chosen era, and its stock market trading increments are likely to be describable by an α -stable L’evy motion. For the period between 4 September 2019, and 6 September 2024, we examine the Athens General Composite (ATG) index. The website https://tradingeconomics.com/greece/stock-market (accessed on 10 September 2024) is where the data were obtained.
The trend of ATG as well as its increments are displayed in Figure 1: the upper left panel for the opening prices, the upper right panel for the closing prices, the bottom left panel for the increments of the opening prices, and the bottom right panel for the increments of the closing prices. Opening and closing prices are extremely similar and correspond to non-stationary processes.
The characteristics of a non-stationary time series alter over time. Trends, cycles, and random events derived from one or more probability distributions in an additive or multiplicative mode can all be found in a non-stationary time series. The degree of randomness that defines a time series is what remains after trends are eliminated by subtracting either the mean or the moving average mean, and cycles, if any, can be handled similarly.
The sample entropy ( S a m p E n ) value for a time series indicates how predictable or regular the time series is; low values indicate a more predictable system, while high values indicate randomness or chaos. Originally developed for physiological applications [38], it finds applications in a variety of fields, including finance [39]. The algorithm of S a m p E n draws inspiration from the embedding dimensions formulated by Grassberger and Procaccia in the early eighties [57,58]. Details of the algorithm and a comparison with the similar but not same concept of approximate entropy can be found in a recent tutorial [59]. S a m p E n ( d a t a , m , r ) [60] is the function that runs in Matlab and takes three arguments. The first one ( d a t a ) stands for the normalised input, where for normalisation we consider the relation x n = x < x > σ x ; see upper panel of Figure 2. The second argument m stands for the embedding dimension, and it is always smaller than the total length of the inp ut data; usually the first values ( m = 1 , 2 ) are of interest. The third argument r is the tolerance and is expressed as a fraction of the standard deviation.
In Figure 2, the bottom panel shows the value of S a m p E n for two different values of m ( m = 1 in red and m = 2 in green) for values of tolerance between 0.1 and 1. All values of S a m p E n are small, indicating a predictable system, and in addition, its value becomes smaller, increasing the value of the tolerance. Especially for values of r greater than 0.5, the conditional probability of the two patterns to be compared is close to 1, underlying the existence of a genuine order. However, such an order, the existence of patterns, is not detected by any other method; see below. Given that tolerance is a portion of the standard deviation and that the latter is affected by outliers, the criterion of matching can be true for all compared patterns only because the dynamic scale has been altered by the outliers. Consequently, S a m p E n suggests that the ATG index from Sep 2019 to 2024 behaves as a scale-free system, and its scaling exponent remains to be estimated [40,61].
In the following, we focus only on the opening prices; the flowchart shown in Figure 3 serves as the basis for their investigation.
The input data are used iteratively to characterise a process of length k, with a minimum value of 0.8 N The value of 0.8 N is used in this work, but it can be even smaller. GMM requires trajectories of a minimum of 100 data points. and a maximum of N 1 , where N is the total length of the trajectory depicted in Figure 1. For each given value of k, we apply GMM and calculate the values of α , h, and C, thus succeeding in characterising the type of process and its self-similarity exponent.
The degree of intermittency (C) falls between 0.046 and 0.053, and the degree of fractional integration (h) is consistently higher than 0.5, suggesting a persistent process. Furthermore, the process draws values from a Gaussian distribution since the index of multifractality α is always 2. In the interval (0.55, 0.58), the scaling or Hurst exponent (H), which is the value of the structure function for q = 1 , fluctuates slightly from trading day to trading day. The findings depicted in Figure 4 rather suggest a process with a time-varying Hurst exponent, or mfBm [27].
To forecast the next value of the opening prices data shown in Figure 1, using Equation (11), and since the input data correspond to an mfBm as the values of α and H indicate and are shown in Figure 4, it is necessary to generate an fBm process every time with scaling exponent H ( t ) , whose increments correspond to an fGn process. It is evident that in each iteration, the length of the generated fBm process increases, and its scaling exponent changes. Sequences of fBm are created using Matlab’s built-in function wfBm(H,k) with zero mean, variance 1, Hurst exponent H, and length k [62]. Next, we use the increments of fBm to generate fractional Gaussian noise sequences. For every iteration, 200 of these sequences are produced, and they yield the noise value in Equation (11).
As anticipated, given that μ is weighted by the value of ATG by definition, the drift term μ of Equation (11) varies over the course of each trading day and has a form that is very similar to the dependence of ATG on time. Additionally, the diffusion term σ provided by Equation (14) exhibits a decreasing behaviour. Figure 5 displays the ATG value for each trading day (in black) along with its predictive value (in red), which is calculated using Equation (11). With a larger percentage error value of less than 6%, both curves show excellent agreement; for a higher resolution, see Figure 6.
Figure 6, in which short time windows are used to increase resolution, shows some snapshots of the ATG evolution along with the corresponding predictions. It is noteworthy that the trend of the forecasts closely resembles the actual trend of changes, and the forecast values are close to the actual values. The bulk of the forecasts have a percentage error of less than 2%, while their largest value of 6% corresponds to trading day 1220, where the recorded stock market change exceeds nearly 80 basis points. Even in this instance, the forecast demonstrates its predictive power by providing the right trend and a change of about 50 basis points.

4. Conclusions

In this work, we propose a model that can predict the next value of a stochastic process through iteration. There are both deterministic and stochastic terms in the fundamental equation that form the basis of the iterative action. The method of generalised moments, which enables the classification of the type of stochastic process—which we consider a priori to be an α -stable Lévy process—is essential to the model’s application. We calculate the drift term, which is an average of the process increments weighted by the process values, and the diffusion term, which includes the effect of the autocorrelation that corresponds to the original process noise, after determining the values of α and the scaling exponent H. We use the Athens General Index as an actual data field from September 2019 to September 2024 in order to evaluate the model. Predicted values are in line with actual data trends, predictions are satisfactory, and the extreme value of the percentage error does not exceed 6%.

Author Contributions

Conceptualisation, E.B.; methodology, E.B.; software, E.B.; validation, E.B.; formal analysis, E.B.; investigation, E.B.; writing—original draft preparation, E.B.; writing—review and editing, E.B. and F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

As mentioned explicitly in the text, the information analysed in this study was obtained from the website https://tradingeconomics.com/greece/stock-market (accessed on 10 September 2024). No additional data have been produced.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
fBmfractional Brownian motion
fGnfractional Gaussian noise
mfBmmultifractional fractional Brownian motion
fLsmfractional Lévy stable motion
fLsnfractional Lévy stable noise

References

  1. Beran, J. Statistics for Long-Memory Processes; Chapman & Hall: London, UK, 2017; pp. 32–58. [Google Scholar]
  2. Black, F.; Scholes, M. The pricing of options and corporate liabilities. J. Polit. Econ. 1973, 81, 637–654. [Google Scholar] [CrossRef]
  3. Di Matteo, T. Multi-scaling in finance. Quant. Financ. 2007, 7, 21–36. [Google Scholar] [CrossRef]
  4. Bentes, S.R.; Menezes, R.; Mendes, D.A. Long memory and volatility clustering: Is the empirical evidence consistent across stock markets? Physica A 2008, 387, 3826–3830. [Google Scholar] [CrossRef]
  5. Drzazga-Szczesniak, E.A.; Szczepanik, P.; Kaczmarek, A.Z.; Szczesniak, D. Entropy of Financial Time Series Due to the Shock of War. Entropy 2023, 25, 823. [Google Scholar] [CrossRef]
  6. Plerou, V.; Gopikrishnan, P.; Rosenow, B.; Amaral, L.A.N.; Stanley, H.E. Econophysics: Financial time series from a statistical physics point of view. Physica A 2000, 279, 443–456. [Google Scholar] [CrossRef]
  7. Mandelbrot, B. The Variation of Certain Speculative Prices. J. Bus. 1963, 36, 394–419. Available online: https://www.jstor.org/stable/2350970 (accessed on 26 March 2025). [CrossRef]
  8. Mantegna, R.; Stanley, H.E. Scaling behaviour in the dynamics of an economic index. Nature 1995, 376, 46–49. [Google Scholar] [CrossRef]
  9. Bouchaud, J.P.; Potters, M.; Meyer, M. Apparent multifractality in financial time series. Eur. Phys. J. B 2000, 13, 595–599. [Google Scholar] [CrossRef]
  10. Scalas, E. Scaling in the market of futures. Physica A 1998, 253, 394–402. [Google Scholar] [CrossRef]
  11. Calvet, L.; Fisher, A. Multifractality in Asset Returns: Theory and Evidence. Rev. Econ. Stat. 2002, 84, 381–406. Available online: https://www.jstor.org/stable/3211559 (accessed on 26 March 2025). [CrossRef]
  12. Ruipeng, L.; Di Matteo, T.; Lux, T. Multifractality and long-range dependence of asset returns: The scaling behaviour of the Markov-switching multifractal model with lognormal volatility components. Adv. Complex Syst. 2008, 11, 669–684. [Google Scholar] [CrossRef]
  13. Mandelbrot, B.B.; Wallis, J.R. Robustness of the rescaled range R/S in the measurement of noncyclic long run Statistical dependence. J. Wat. Res. 1969, 5, 967–988. [Google Scholar] [CrossRef]
  14. Hurst, H.E. Long term storage capacity of reservoirs. Trans. Am. Soc. Eng. 1951, 116, 770–799. [Google Scholar] [CrossRef]
  15. Yin, Y.; Shang, P. Modified multidimensional scaling approach to analyze financial markets. Chaos 2014, 24, 022102. [Google Scholar] [CrossRef]
  16. Peng, C.-K.; Buldyrev, S.V.; Havlin, S.; Simons, M.; Stanley, H.E.; Goldberger, A.L. Mosaic organization of DNA nucleotides. Phys. Rev. E 1994, 49, 685–1689. [Google Scholar] [CrossRef] [PubMed]
  17. Kantelhardt, J.W.; Zschiegner, S.A.; Koscielny-Bunde, E.; Havlin, S.; Bunde, A.; Stanley, H.E. Multifractal detrended fluctuation analysis of nonstationary time series. Physica A 2002, 316, 87–114. [Google Scholar] [CrossRef]
  18. Barunik, J.; Kristoufek, L. On Hurst exponent estimation under heavy-tailed distributions. Physica A 2010, 389, 3844–3855. [Google Scholar] [CrossRef]
  19. Bakalis, E.; Höfinger, S.; Venturini, A.; Zerbetto, F. Crossover of two power laws in the anomalous diffusion of a two lipid membrane. J. Chem. Phys. 2015, 142, 215102. [Google Scholar] [CrossRef] [PubMed]
  20. Schertzer, D.; Lovejoy, S. Physical modeling and Analysis of Rain and Clouds by Anisotropic Scaling of Multiplicative Processes. J. Geophys. Res. 1987, 92, 9693–9714. [Google Scholar] [CrossRef]
  21. Mantegna, R.N. Lévy walks and enhanced diffusion in Milan stock exchange. Physica A 1991, 179, 232–242. [Google Scholar] [CrossRef]
  22. Mandelbrot, B.B.; Van Ness, J.W. Fractional brownian motions, fractional noises and applications. SIAM Rev. 1968, 10, 422–437. [Google Scholar] [CrossRef]
  23. Samorodnitsky, G.; Taqqu, M. Stable Non-Gaussian Random Processes: Stochastic Models with Infinte Variance, 1st ed.; Chapman and Hall: New York, NY, USA, 1994. [Google Scholar]
  24. Fama, E.F.; Roll, R. Some properties of symmetric stable distributions. J. Am. Stat. Assoc. 1968, 63, 817–836. [Google Scholar] [CrossRef]
  25. Kogon, S.M.; Manolakis, G. Signal modeling with self- similar α/-stable processes: The fractional Levy stable motion model. IEEE Trans. Signal Process. 1996, 44, 1006–1010. [Google Scholar] [CrossRef]
  26. Samorodnitsky, G.; Taqqu, M. Linear models with long-range dependence and with finite and infinite variance. In The IMA Volumes in Mathematics and Its Applications, 1st ed.; New Directions in Time Series Analysis; Brillinger, D., Caines, P., Geweke, J., Parzen, E., Rosenblatt, M., Taqqu, M.S., Eds.; Springer: New York, NY, USA, 1993; pp. 325–340. [Google Scholar]
  27. Song, W.; Cattani, C.; Chi, C.-H. Multifractional Brownian motion and quantum-behaved particle swarm optimization for short term power load forecasting: An integrated approach. Energy 2020, 194, 116847. [Google Scholar] [CrossRef]
  28. Liu, H.; Song, W.; Li, M.; Kudreyko, A.; Zio, E. Fractional Lévy stable motion: Finite difference iterative forecasting model. Chaos 2020, 133, 109632. [Google Scholar] [CrossRef]
  29. Eke, A.; Herman, P.; Kocsis, L.; Kozak, L.R. Fractal characterization of complexity in temporal physiological signals. Physiol. Meas. 2002, 23, R1–R38. [Google Scholar] [CrossRef]
  30. Caccia, D.C.; Percival, D.; Cannon, M.J.; Raymond, G.; Bassingthwaighte, J.B. Analyzing exact fractal time series: Evaluating dispersional analysis and rescaled range methods. Physica A 1997, 246, 609–632. [Google Scholar] [CrossRef]
  31. Bakalis, E.; Ferraro, A.; Gavriil, V.; Pepe, F.; Kollia, Z.; Cefalas, A.-C.; Malapelle, U.; Sarantopoulou, E.; Troncone, G.; Zerbetto, F. Universal Markers Unveil Metastatic Cancerous Cross-Sections at Nanoscale. Cancers 2022, 14, 3728. [Google Scholar] [CrossRef]
  32. Barabási, A.-L.; Viscek, T. Multifractality of self-affine fractals. Phys. Rev. A 1991, 44, 2730. [Google Scholar] [CrossRef]
  33. Sändig, N.; Bakalis, E.; Zerbetto, F. Stochastic analysis of movements on surfaces: The case of C60 on Au(1 1 1). Chem. Phys. Lett. 2015, 633, 163–168. [Google Scholar] [CrossRef]
  34. Bakalis, E.; Mertzimekis, T.J.; Nomikou, P.; Zerbetto, F. Breathing modes of Kolumbo submarine volcano (Santorini, Greece). Sci. Rep. 2017, 7, 46515. [Google Scholar] [CrossRef]
  35. Taqqu, M.S.; Teverovsky, V.; Willinger, W. Estimators for long-range dependence: An empirical study. Fractals 1995, 3, 785–798. [Google Scholar] [CrossRef]
  36. Laskin, N.; Lambadaris, I.; Harmantzis, F.C.; Devetsikiotis, M. Fractional Lévy motion and its application to network traffic modeling. Comput. Netw. 2002, 40, 363–375. [Google Scholar] [CrossRef]
  37. Gómez-Águila, A.; Trinidad-Segovia, J.E.; Sánchez-Granero, M.A. Improvement in Hurst exponent estimation and its application to financial markets. Financ. Innov. 2022, 8, 86. [Google Scholar] [CrossRef]
  38. Richman, J.S.; Moorman, J.R. Physiological time- series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  39. Olbrys, J.; Ostrowski, K. An entropy-based approach to measurement of stock market depth. Entropy 2021, 23, 568. [Google Scholar] [CrossRef] [PubMed]
  40. Sokunbi, M.O.; Gradin, V.B.; Waiter, G.D.; Cameron, G.G.; Ahearn, T.F.; Murray, A.D.; Steele, D.J.; Staff, R.T. Nonlinear Complexity Analysis of Brain fMRI Signals in Schizophrenia. PLoS ONE 2014, 9, e105741. [Google Scholar] [CrossRef] [PubMed]
  41. Mandelbrot, B.B. The Fractal Geometry of Nature, 1st ed.; W. H. Freeman & Company: New York, NY, USA, 2008; 480p. [Google Scholar]
  42. Grassberger, P. Generalized dimensions of strange attractors. Phys. Lett. A 1983, 97, 227–230. [Google Scholar] [CrossRef]
  43. Hentschel, H.G.E.; Procaccia, I. The infinite number of generalized dimensions of fractals and strange attractors. Physica D 1983, 8, 435–444. [Google Scholar] [CrossRef]
  44. Kolmogorov, A.N. A refinement of previous hypotheses concerning the local structure of turbulence in a viscous incompressible fluid at high Reynolds number. J. Fluid. Mech. 1962, 13, 82–85. [Google Scholar] [CrossRef]
  45. Bakalis, E.; Gavriil, V.; Cefalas, A.-C.; Kollia, Z.; Zerbetto, F.; Sarantopoulou, E. Viscoelasticity and Noise Properties Reveal the Formation of Biomemory in Cells. J. Phys. Chem. B 2021, 125, 10883–10892. [Google Scholar] [CrossRef] [PubMed]
  46. Bakalis, E.; Fujie, H.; Zerbetto, F.; Tanaka, Y. Multifractal structure of microscopic eye–head coordination. Physica A 2018, 512, 945–953. [Google Scholar] [CrossRef]
  47. Bakalis, E.; Mertzimekis, T.J.; Nomikou, P.; Zerbetto, F. Temperature and Conductivity as Indicators of the Morphology and Activity of a Submarine Volcano: Avyssos (Nisyros) in the South Aegean Sea, Greece. Geosciences 2018, 8, 193. [Google Scholar] [CrossRef]
  48. Andersen, K.; Castiglione, P.; Mazzino, A.; Vulpiani, A. Simple stochastic models showing strong anomalous diffusion. Eur. Phys. J. B 2000, 18, 447–452. [Google Scholar] [CrossRef]
  49. Fuliński, A. Fractional Brownian motions: Memory, diffusion velocity, and correlation functions. J. Phys. A Math. Theor. 2017, 50, 054002. [Google Scholar] [CrossRef]
  50. Bénichou, O.; Oshanin, G. A unifying representation of path integrals for fractional Brownian motions. J. Phys. A Math. Theor. 2024, 57, 2250012. [Google Scholar] [CrossRef]
  51. Bakalis, E.; Lugli, F.; Zerbetto, F. Daughter Coloured Noises: The Legacy of Their Mother White Noises Drawn from Different Probability Distributions. Fractal Fract. 2023, 7, 600. [Google Scholar] [CrossRef]
  52. Mercik, S.; Weron, K.; Burnecki, K.; Weron, A. Enigma of Self-Similarity of Fractional Lévy Stable Motions. Acta Polonica B 2003, 34, 3773–3791. [Google Scholar]
  53. Nakao, H. Multi-scaling properties of truncated Lévy flights. Phys. Lett. A 2000, 266, 282–289. [Google Scholar] [CrossRef]
  54. Janicki, A.; Weron, A. Can One See α-Stable Variables and Processes? Stat. Sci. 1994, 9, 109–126. [Google Scholar] [CrossRef]
  55. Yan, L.; Shen, G.; He, K. Itô’s formula for a sub-fractional Brownian motion. Commun. Stoc. Anal. 2011, 5, 135–159. [Google Scholar] [CrossRef]
  56. Jumarie, G. On the representation of fractional Brownian motion as an integral with respect to (dt)α. Appl. Math. Lett. 2005, 18, 739–748. [Google Scholar] [CrossRef]
  57. Grassberger, P.; Procaccia, I. Estimation of the Kolmogorov entropy from a chaotic signal. Phys. Rev. A 1983, 28, 2591(R). [Google Scholar] [CrossRef]
  58. Grassberger, P.; Procaccia, I. Characterization of Strange Attractors. Phys. Rev. Lett. 1983, 50, 346. [Google Scholar] [CrossRef]
  59. Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef]
  60. Martínez-Cagigal, V. Sample Entropy. Mathworks, 2018. Available online: https://mathworks.com/matlabcentral/fileexchange/69381-sample-entropy (accessed on 26 March 2025).
  61. Omidvarnia, A.; Mesbah, M.; Pedersen, M.; Jackson, G. Range Entropy: A Bridge between Signal Complexity and Self-Similarity. Entropy 2018, 20, 962. [Google Scholar] [CrossRef]
  62. The MathWorks Inc. MATLAB, Version: 9.13.0 (R2022b); The MathWorks Inc.: Natick, MA, USA, 2022. [Google Scholar]
Figure 1. The upper panel displays the ATG’s (Athens General Index) opening and closing prices over a five-year period, from September 2019 to September 2024. Although these days are not always calendar consecutive, two consecutive events indicate two trading days. The bottom panel shows the increments of the processes shown in the upper panel. The website https://tradingeconomics.com/greece/stock-market (accessed on 10 September 2024) is where the data were obtained.
Figure 1. The upper panel displays the ATG’s (Athens General Index) opening and closing prices over a five-year period, from September 2019 to September 2024. Although these days are not always calendar consecutive, two consecutive events indicate two trading days. The bottom panel shows the increments of the processes shown in the upper panel. The website https://tradingeconomics.com/greece/stock-market (accessed on 10 September 2024) is where the data were obtained.
Entropy 27 00497 g001
Figure 2. The normalised opening prices of ATG, see also Figure 1 for comparison, are displayed in the upper panel. Its values serve as input for the sample entropy calculation (bottom panel). Two distinct values of m (red for m = 1 and green for m = 2) and a range of tolerance r values are used to estimate the sample entropy.
Figure 2. The normalised opening prices of ATG, see also Figure 1 for comparison, are displayed in the upper panel. Its values serve as input for the sample entropy calculation (bottom panel). Two distinct values of m (red for m = 1 and green for m = 2) and a range of tolerance r values are used to estimate the sample entropy.
Entropy 27 00497 g002
Figure 3. The flowchart shows the steps that need to be taken from data entry to output, which is the forecast for each subsequent trading day.
Figure 3. The flowchart shows the steps that need to be taken from data entry to output, which is the forecast for each subsequent trading day.
Entropy 27 00497 g003
Figure 4. The calculated values of h, C, α , and H for the recorded stock market prices that cover a time window of at least 990 days to a maximum of 1240 days. Each new trading day increases the length of the recorded data by one.
Figure 4. The calculated values of h, C, α , and H for the recorded stock market prices that cover a time window of at least 990 days to a maximum of 1240 days. Each new trading day increases the length of the recorded data by one.
Entropy 27 00497 g004
Figure 5. The upper panel shows how the diffusion term, Equation (14), and the mean, Equation (13), change over the course of the analysis. The ATG index’s evolution over the course of five years is shown in the bottom panel in black, and iterative predictions made using Equation (11) are shown in red. Additionally, the relative percentage error is illustrated.
Figure 5. The upper panel shows how the diffusion term, Equation (14), and the mean, Equation (13), change over the course of the analysis. The ATG index’s evolution over the course of five years is shown in the bottom panel in black, and iterative predictions made using Equation (11) are shown in red. Additionally, the relative percentage error is illustrated.
Entropy 27 00497 g005
Figure 6. The ATG evolution is depicted in black in these snapshots of the analysed trajectory, along with the prediction, in red, made using the iterative scheme suggested in this work.
Figure 6. The ATG evolution is depicted in black in these snapshots of the analysed trajectory, along with the prediction, in red, made using the iterative scheme suggested in this work.
Entropy 27 00497 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bakalis, E.; Zerbetto, F. Iterative Forecasting of Financial Time Series: The Greek Stock Market from 2019 to 2024. Entropy 2025, 27, 497. https://doi.org/10.3390/e27050497

AMA Style

Bakalis E, Zerbetto F. Iterative Forecasting of Financial Time Series: The Greek Stock Market from 2019 to 2024. Entropy. 2025; 27(5):497. https://doi.org/10.3390/e27050497

Chicago/Turabian Style

Bakalis, Evangelos, and Francesco Zerbetto. 2025. "Iterative Forecasting of Financial Time Series: The Greek Stock Market from 2019 to 2024" Entropy 27, no. 5: 497. https://doi.org/10.3390/e27050497

APA Style

Bakalis, E., & Zerbetto, F. (2025). Iterative Forecasting of Financial Time Series: The Greek Stock Market from 2019 to 2024. Entropy, 27(5), 497. https://doi.org/10.3390/e27050497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop