Next Article in Journal
Advances in Study of Time-Delay Systems and Their Applications: A Second Edition
Previous Article in Journal
A Semi-Parametric KDE-GPD Model for Earthquake Magnitude Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Stochastic Evolution of Financial Asset Prices

by
Ioannis Paraskevopoulos
* and
Alvaro Santos
Faculty of Economics and Business Administration (ICADE), Universidad Pontificia Comillas, 28015 Madrid, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(12), 2002; https://doi.org/10.3390/math13122002
Submission received: 31 March 2025 / Revised: 28 May 2025 / Accepted: 4 June 2025 / Published: 17 June 2025

Abstract

This paper examines the relationship between dependence and independence alternatives in general stochastic processes and explores the duality between the true (yet unknown) stochastic process and the functional representation that fits the observed data. We demonstrate that the solution depends on its historic realizations, challenging existing theoretical frameworks that assume independence between the solution and the history of the true process. Under orthogonality conditions, we investigate parameter spaces within data-generating processes and establish conditions under which data exhibit mean-reverting, random, cyclical, history-dependent, or explosive behaviors. We validate our theoretical framework through empirical analysis of an extensive dataset comprising daily prices from the S&P500, 10-year US Treasury bonds, the EUR/USD exchange rate, Brent oil, and Bitcoin from 1 January 2002 to 1 February 2024. Our out-of-sample predictions, covering the period from 17 February 2019 to 1 February 2024, demonstrate the model’s exceptional forecasting capability, yielding correct predictions with between 73% and 92% accuracy, significantly outperforming naïve and moving average models, which only achieved 47% to 54% accuracy.

1. Introduction

All stochastic sequences and their solutions share the same topological endowments or obey different laws. Determining when an unknown function of a sequence has the same topology as the sequence itself in a truly unbiased manner could be a natural way to test whether asymptotic behaviors can ensure convergence. In such a case, the model matches the observed data. Chichilnisky [1] used weighted Hilbert spaces to identify the convergence of the Fourier coefficients of the unknown function that solves the stochastic sequence for all domains in bounded and unbounded spaces. A Hilbert space is complete and always reflexive. Therefore, solutions to stochastic partial differential equations are always endowed with the inner product that allows the closeness of the known and unknown sequences to be measured and to establish duality in the sense of being reflexive. In an unknown solution in a Hilbert space, the sequence H ρ would always be isomorphic to a function of H S that is jointly defined in a uniformly convex space. An observer testing the existence of duality will be biased when measured on the real line L 2 R . Bergstrom [2] remarks that for a Hilbert space where continuous functions have bounded variation, the square integrable unknown function H S : K R requires S K K = 0 , representing a very strong restriction as it excludes important functions such as the constant, increasing, and cyclical types (see Chichilnisky [1]).
The Banach space B is a complete normed space L p of some dimension p, and it exists between the Hilbert space and a space M which is endowed with a metric and not a norm. In particular, if L = L 1 , L 2 , , L p , , L M , , L with any choice of p is within 2 p M , then it is a Banach space. In addition, the vector probability space, defined by the triplet L X , Ω , θ is a σ -finite measure, where X is the particular space, F represents the filtered measurable sets, and θ is the measure. Then, more formally, a Banach space is defined as L X , Ω , θ : = f : X C , with f being Ω measurable. The direct implication is that X f x p d θ x < . Completeness means that any Cauchy, C, sequence with respect to the norm is convergent at the point that is in the space. The Banach norm L p of the function f is given by
f p = X f x p d θ x 1 / p
Geometrical considerations for normed spaces are dictated by the value of p. For instance, when p = 1 , we have the space L 1 that includes all integrable functions f on X, and the · L 1 satisfies the triangle inequality. When p = 2 , L 2 is a Hilbert space and · L 2 .
In general, only square integrable processes that are integrable functions are measurable and include all common functions such as “sup”, “inf”, and “lim”, are unique in being compatible with the desirable inner product property of the Hilbert space, and this guarantees reflexivity. Tsirelson [3] showed an example in which the square integrable sequences are not always isomorphic to a uniformly convex space and they do not obey the same topological laws, while they have no inner product properties. The contribution of this paper is, therefore, to amplify the Hilbert space, with Banach L p satisfying L 1 < L 2 < L p < for any p in 2 p . The purpose is, then, to extend the limits of measurement and identify the conditions for the duality of the unknown solution and its history via the observed stochastic process. For this, we assume that all functional forms will be measured in a Banach complete normal space, and this is where the h = H 0 , H 1 will be tested with some probability. We discuss the conditions under which it is isomorphic to a convex reflexive Banach dual space when the observed sequence inherits norms and an inner product. Yano K. [4] demonstrate that the unknown solution space S k does not depend on the known sequence metric ξ j , and this invalidates the use of any functional model for explaining and predicting an unknown related process under orthogonality assumption in random processes. In general, duality and association cannot exist if the unobserved solution is not linked to the observed data. History plays no role in the formation of the new state, and generally speaking, S k = k ξ j V , with V being some other independent sequence. Under this scenario, reversibility is impossible as there is an infinite number of starting points to map going backwards, rendering the course of the observed sequence unpredictable.
In this paper, we discuss the dependence and independence alternatives for a general stochastic solution process, and when the stochastic process takes values from Banach space, we examine under which conditions it can be reflexive. In particular, we examine whether the stochastic process depends on its known history. We argue that multidimensional integration has duality in a reflexive Banach space if the solution obeys the functional central limit theorem and its error is between the limits of dilation and erosion. Hence, there is isomorphism for closed and semi-open intervals and the desired inner product to define the orthogonality conditions. The evolution process depends on its history in three-quarters of all existing probability measures, and this necessarily implies that the integrable curves are quasi-periodic depending only on initial conditions that can be found in history, or indeed at the beginning of the cosmos. One possible scenario is, as Yano K. [4] and Tsirelson [3] suggested, that no model exists and reversibility is unattainable as it would map to infinite possible initial points. We offer a framework to evaluate these scenarios under different novel models that capture the general family of possible null hypotheses formulations. They can capture all possibilities of dependence and independence in both Banach and reflexive Banach spaces.
Still following their assumptions and adopting the compact G group and requiring duality in reflexive Banach space, we prove their dependence. We derive the orthogonality conditions and show that the unknown unique stochastic solution would depend on the observable history of the sequence, as in S m = m ρ j S m 1 . The evolution process would depend on its history in three-quarters of all existing scenarios. When variation is outside the limits of dilation and erosion, the space will be composed of countably additive and possibly finitely additive measures with no functional representation in B . Convergence will also fail in this scenario.
This paper contributes to several strands of the literature at the intersection of stochastic process theory, functional analysis in infinite-dimensional spaces, and empirical time series modeling under structural complexity. Below, we situate our work within four key domains: (1) duality and functional representation in Banach and Hilbert spaces; (2) dependence, orthogonality, and predictability in stochastic systems; (3) modeling nonstationary financial time series with structural breaks; and (4) empirical applications and computational limits in high-volatility environments.
The representation of stochastic processes in infinite-dimensional spaces has long been shaped by the use of Hilbert spaces due to their inner product structure and reflexivity [1,2]. However, this approach imposes strong topological restrictions—such as the exclusion of constant or monotonic functions—and often fails to accommodate non-smooth or heavy-tailed processes. Demonstrated by [3] square-integrable sequences do not always exhibit topological equivalence with uniformly convex function spaces, posing limitations for models based solely on L 2 spaces.
Recent advancements extend duality and learning representations to more general settings. For instance, [5] explored duality principles in reproducing kernel Banach spaces (RKBS), offering a functional framework beyond Hilbert geometries. Further [6] formalized the duality of spaces of càdlàg (right-continuous with left limits) processes, establishing foundational conditions for reflexivity in stochastic filtrations. We extend these approaches by situating the stochastic solution within a reflexive Banach space L p for p > 2 , allowing for a more general functional topology. This enables the definition of duality between observed and unobserved sequences, even in the absence of inner products, and provides a novel isomorphism mapping via convolutional sequences in ρ k . Unlike prior models, our construction supports a broader family of functional forms—including explosive, cyclic, and path-dependent dynamics—within a coherent probabilistic geometry.
In their work [4] present a formal case in which the evolution process S k is orthogonal to its observed history, invalidating the use of predictive functional forms. Under their framework, the absence of duality renders the process fundamentally irreversible and unpredictable. Davidson, in [7,8] explore related concerns in the context of I(0) processes and the functional central limit theorem. More recently, orthogonality structures have gained attention in spectral analysis and stochastic modeling. Generalized polynomial chaos expansions benefit from orthogonality in solving stochastic PDEs, though this work remains largely applied to physical systems [9]. Our work complements these efforts by deriving a precise orthogonality condition,
E [ g ( k 1 ) · ϵ k ( p ) ] = 0 w k = 1 ,
which we use to classify regimes as martingale, trend-following, mean-reverting, or explosive.
Here, we reintroduce history as a structural feature of stochastic evolution, showing that in three-quarters of all probabilistic configurations, the solution S k is dependent on its past. The moment-generating function ϕ k ( p ) and its innovation ϵ k ( p ) serve as instruments for detecting and classifying the nature of dependence. This builds a dynamic model of predictability that is entirely endogenous to the data-generating process.
Traditional time series models—such as ARMA, GARCH, and their extensions—typically assume stationarity or impose transformations (e.g., differencing) to induce it [10]. However, real-world financial series are nonstationary, exhibit volatility clustering, and often undergo abrupt structural changes. Structural break detection is typically addressed via exogenous methods, such as in [11], though these are not integrated into the modeling process.
The recent work by [12] checked for multiple breaks without imposing prior distributional knowledge by proposing a nonparametric change-point detection approach to estimate the number of change points and their locations along the temporal axis.
The AREM model used in this study integrates break detection directly into its stochastic foundation via the behavior of the w k statistic and the divergence of ϕ k ( p ) . Rather than treating structural breaks as exogenous events, the model flags them as a breakdown of orthogonality, thereby internalizing regime shifts. Moreover, the recursive form S k = j = k ρ j · S k 1 provides a continuous, historical basis for modeling long memory and path dependence, rather than differencing away valuable information.
On the empirical front, studies have emphasized the need for forecasting models to perform under heavy tails and computational constraints. In particular, ref. [13] introduced fractional Brownian motion to describe persistent memory, while [14] proposed models that adapt to macro shifts via spline-GARCH processes. However, these models often require specific distributional assumptions or lack general theoretical foundations.
Recent comparative studies between econometric modeling and machine learning highlight the strengths and limitations of both approaches, particularly under complex dynamic conditions. For instance, ref. [15] analyzed the forecasting of COVID-19’s spread using classical mathematical models versus machine learning techniques, showing that while machine learning often excels in short-term pattern recognition, structural mathematical models offer greater transparency, interpretability, and robustness—especially in the presence of regime shifts and noisy inputs. This trade-off is also relevant in financial time series forecasting, where structural stability and duality-based interpretability are critical. Furthermore, the increasing use of predictive models in financial infrastructure raises significant cybersecurity concerns, as model transparency affects vulnerability to adversarial attacks. Our model, based on recursive structural properties and orthogonality principles, is inherently interpretable, and thus potentially more resilient to such threats. This reinforces the value of functional–theoretic approaches alongside, and often in contrast to, black-box machine learning models.
The empirical application of AREM to diverse assets including Bitcoin, the S&P 500, and US Treasury bonds demonstrates forecasting accuracy between 73% and 93%. This robustness is achieved despite volatility, structural breaks, and nonstationarity, without relying on transformation or external break detection. Further, we formalize the model’s limits of computation via the divergence of dilation/erosion boundaries and propose extensions to switch estimation regimes when orthogonality fails. This positions AREM as both a theoretically grounded and practically adaptive forecasting tool.
In contrast to prior work that either imposes restrictive topologies (Hilbert spaces), neglects historical dependence (orthogonality assumptions), or addresses regime changes exogenously, this paper introduces a unified framework for stochastic modeling based on Banach space duality. Our contributions span foundational theory, the identification of dependence regimes, and high-performing empirical forecasting, all within a mathematically rigorous and computationally tractable setting.
The existing literature provides a rich and multi-dimensional foundation for understanding duality, orthogonality, and predictability in stochastic systems, particularly in the presence of nonstationarity, structural breaks, and nonlinear dependencies. While these works vary in domain and methodology from functional analysis to machine learning, they collectively emphasize the limitations of assuming independence between a stochastic process and its history, as well as the potential gains from exploiting the hidden structure through richer model representations. Others [16] extends Tsirelson’s counterexample by constructing stochastic difference equations on general topological groups. His work demonstrates the existence of stochastic processes that are fundamentally irreversible and independent of their historical trajectories, offering an abstract mathematical example of how history may fail to determine the future. This view aligns with the theoretical boundary proposed by [4], where orthogonality between a process and its past leads to unpredictability and non-reversibility. However, our model challenges this boundary by establishing conditions in reflexive Banach spaces under which dependence on history is restored for most probabilistic configurations. Research by [17] produce a framework for duality in interacting stochastic systems using orthogonal polynomial expansions. By constructing dual processes, they demonstrate that complex dynamics can often be re-expressed in terms of more tractable counterparts. Similarly, we use orthogonality conditions and moment-generating functions to classify regimes of dependence, trend-following, and explosiveness within a unified data-generating process. Moreover [18] offer an important shift by conducting functional time series analysis in Banach rather than Hilbert spaces. By abandoning the inner product structure, they develop new prediction methods under the supremum norm. Our model similarly adopts Banach topology, particularly for p > 2 , to embed historical dependence in spaces that better reflect the irregularity and volatility of real-world time series. This allows us to characterize duality and convergence in a setting where smoothness and completeness are not guaranteed by construction.
Recent econometric contributions, such as those by [19,20,21], develop tools to detect structural breaks in regression relationships. These tests highlight the fragile nature of forecast models when exposed to regime shifts. In contrast to treating such breaks exogenously, our model internalizes them through shifts in the w k statistic and the divergence of moment-generating functions. This endogeneity allows the AREM model to dynamically adapt to changing regimes, identifying pockets of predictability and divergence from orthogonality without manual intervention.
The question of whether predictability itself is episodic rather than constant is central to [22]. These authors introduce rolling and recursive tests to identify such episodes, aligning with our notion that the orthogonality condition may hold only intermittently. We formalize this through a model where the evolution process S k depends on its historical path in three-quarters of the probability space.
Modern forecasting research has expanded the toolkit beyond linear models. Studies such as those by [23,24,25] explore the use of ensemble methods and hybrid models to improve forecast accuracy in the presence of complex, nonlinear dependencies. Our AREM model builds on this insight by incorporating ensemble structures that blend multiple predictors—some driven by theory, others by empirical fitting—under a common orthogonality framework.
Neural forecasting models, as reviewed by [26], further extend this logic by allowing for the automatic learning of long memory, seasonality, and regime shifts. These architectures, particularly attention-based and sequence-to-sequence models, can be interpreted as data-driven analogues to our theoretically derived dual systems, while our model is not neural in implementation, it shares the architectural goal of tracking temporal dependencies through recursive structures and regime-adaptive filtering.
Other contributions, such as those by [27,28], stress the need for forecasting models that remain robust in the face of volatility clustering and regime instability. Our approach advances this line of research by constructing a theoretical space within the reflexive Banach framework, where such volatility is not just accommodated, but is endogenous to the model’s structural form.
The next section introduces dependency and a stochastic process that captures the past, present, and future. In Section 3, we provide empirical estimations and results. Section 4 presents the conclusions of this paper.

2. History Matters

We examine the limits of computation that are conditional on the properties endowed to generating the estimated processes. Let us consider the behavior of an unknown process, S, which presents a known correlation function, ρ , under a complete normed (Banach) space, so that the comparison of S and ρ can be performed through their respective norms, | | · | | , alone. Note that two norms are equivalent if | | S | | θ | | ρ | | , for all θ = { S , ρ } . Let us further assume that the relation between the two random processes, ( S k ) k 0 and ( ρ k ) k 0 , is characterized by
S k = e 2 i π u k = ρ k S k 1
ρ k = e 2 i π θ k
such that k varies in N , u k is a standard Gaussian which shares a portion of information, θ ^ k , with θ k 0 , 1 ,
θ k = θ k + θ ^ k
and ρ k k 0 behaves as a Markov chain, with transition probabilities given by
P S k A | F k 1 S = g k A ρ k 1 1 ω , A B G
so that the law ρ k is a given probability g k k 0 on G , and Equation (1) can be interpreted as a re-statement of the Tsirelson equation, where dependence is introduced by means of ρ k . As we could always write S k = m k m k 1 , this specification implies that S k converges to a nonstationary unit root process.
ρ k = S k S k 2 S k 1 2 ρ k 1
We want to describe the group of laws of the probability P g of all solutions of Equation (5), and also its behavior at the extreme boundaries e x P g . The strong solutions set is L g under the measure P as P , F k S F k ρ . From Equation (1), we write the identity
ρ k = S k S k 1 1
We can state that the measure P L g if F k S = F k ρ , meaning that the changes in ϕ k · cannot be explained by the previous period’s relevant information when only the evolution of the S k process is taken into account. We further assume that θ ^ k = u k + Δ u k and define ϕ k · the moment-generating function for ρ k , + :
ϕ k p : = E e 2 i π p θ k = E e 2 i π p θ k 1 + Δ u k Δ u k 1
so that the relation between two consecutive moment-generating functions can be written as
ϕ k p E e 2 i π p θ ^ k 1 = ϕ k 1 p E e 2 i π p θ ^ k 1 + Δ u k Δ u k 1
or, equivalently,
E g k 1 · ϕ k p w k ϕ k 1 p = 0
where g ( k 1 ) = e 2 i π p θ ^ k 1 and w k = S k S k 2 S k 1 2 . If we then define the innovation, ϵ , as the change between two consecutive moment-generating functions, ϵ k p = ϕ k p ϕ k 1 p , Equation (9) can be restated as
E g k 1 · ϵ k p = E 1 w k g k 1 · ϕ k 1 p
When the process S behaves as a martingale for a period of time, the correlation function will stay unchanged, ρ k = ρ k 1 , w k = 1 and E g k 1 · ϵ k p = 0 ; otherwise, the moment-generating function, ϕ k p , must satisfy the general orthogonality conditions over the open interval 0 , 1 .
Thus, we require the following on the left side:
E e 2 i π p θ ^ k 1 ϕ k p ϕ k 1 p e 2 i π p Δ u k Δ u k 1 = 0
and the right side of the interval has to be
E [ ( ϕ k ( p ) e 2 i π p θ ^ k 1 ϕ k 1 ( p ) e 2 i π p ( Δ u k Δ u k 1 ) ) ] 1
The following ratio of probabilities is a square integrable process in L 2 a , b describing the behavior of the unknown function. To be convergent in the open interval, the relative likelihoods have to be a different value than one. Therefore, we consider
H E ϕ k p · g k 1 ϕ k 1 p · g k 1 w k = H E exp 2 i π p ρ k ρ k 1
= exp 2 i π p ρ k 1 ρ k 2
Strong solutions exist if and only if the infinite product is convergent to the Fisher information ratio in the Fourier space, where the periodic functions reside:
lim k exp 2 i π p ρ k 1 ρ k 2 e 2 · H i π μ k B σ k
with μ k = ρ k 1 = U | S u | d θ u and σ k = ρ k 2 = U | S u | 2 d θ u 1 / 2 , and then it follows that
P S k S k 1 A | F u k 1 = e 2 · i · π μ k B σ k A ω , A B G
Chichilnisky also dealt with open Hilbert spaces, placing conditions on the asymptotic behavior of the unknown function, and in particular the Fourier estimator convergence at infinity. All the fixed points given by this solution for the stochastic process S k are of the type
S k = k ρ k · S k 1
This is an example of the metrics of G B * space, describing the convolution of long and short memory of the processes. In fact, it is an adjusted Tsirelson stochastic process (see [3]). We write the spread difference as a general process that combines in one functional form the Brownian and OU processes conditioned on the speed of reversion, λ , and the intensity of those changes, α . It mixes the two stochastic distributions.
d S t = λ μ t r λ σ S , t 1 / t + λ + α λ S t S k d t + σ t d W t
where S k = k ρ k · S k 1 , with t > k .
A martingale process is one where the expected value of S t at the next step equals its current value, considering all past information. For S t to be a martingale, the drift term would need to be zero, or the expectation of future changes would need to balance out to zero. This is unlikely given the structure of the drift term unless very specific conditions are met regarding μ t , r, σ S , t , and the historical term S k .
Explosive behavior refers to S t moving away from any bound or mean at an accelerating rate. This could occur if S k dominates and grows without bounds, or the drift term consistently adds to S t without any counteracting mean-reversion force, and the influence of S k or the drift term grows over time. 1 / t + λ + α λ S t : The 1 / t component decreases over time, modifying the impact of S t on its drift; while 1 / t itself is not cyclical, it influences how the strength of the mean-reversion or trend component decays over time. S k , the historical term, integrates past values of S and ρ k . If ρ k (and thus S k ) incorporates cyclical behaviors, it could induce similar patterns in S t , depending on how significantly S k influences the overall drift.
First, we simplify the SDE for S t by expanding the first term inside the square brackets and using the product rule for S k :
d S t = λ μ t r λ σ S , t 1 λ 1 t + λ + α S t + 1 λ k ρ k · S k 1 d t + σ t d W t = λ μ t r λ σ S , t 1 λ 1 t + λ + α S t d t + 1 λ k ρ k · S k 1 · λ d t + σ t d W t = μ t r σ S , t d t 1 λ 1 t + λ + α σ S , t S t d t + 1 λ k ρ k · S k 1 · λ d t + σ t d W t = μ t r σ S , t d t 1 t + λ + α λ σ S , t S t d t + 1 λ k ρ k · S k 1 · λ d t + σ t d W t = μ t r σ S , t d t 1 t λ + 1 λ 2 + α λ σ S , t S t d t + 1 λ k ρ k · S k 1 · λ d t + σ t d W t
Next, we apply Ito’s Lemma to the function f ( S t , t ) = ln ( S t ) . Applying Ito’s Lemma to the function f ( S t , t ) = ln ( S t ) , we obtain
d f ( S t , t ) = f t d t + f S d S t + 1 2 f S S ( d S t ) 2 = 1 S t d S t 1 2 1 S t 2 ( d S t ) 2
using the fact that f t = 0 , f S = 1 S t , and f S S = 1 S t 2 .
Substituting the expression for d S t in terms of S t , μ t , σ S , t , and S k , we obtain
d f ( S t , t ) = λ S t μ t r λ σ S , t 1 λ ( 1 / t + λ + α ) S t S k d t + σ t S t d W t 1 2 1 S t 2 σ t 2 d t = μ t r σ S , t 1 t + λ + α S t σ t 2 2 S t 2 + λ S t S k d t + σ t S t d W t
using the fact that λ σ S , t = 1 σ I , t and S k is a function of past realizations of S t , and is therefore deterministic.
We can integrate both sides of this equation from 0 to t to obtain
ln ( S t ) ln ( S 0 ) = 0 t μ u r λ σ S , u ( 1 / u + λ + α ) λ S u S k 1 2 1 S u σ u 2 d u + 0 t 1 S u σ u d W u ln ( S t ) = ln ( S 0 ) + 0 t μ u r λ σ S , u ( 1 / u + λ + α ) λ S u S k 1 2 1 S u σ u 2 d u + 0 t 1 S u σ u d W u
Taking the exponential of both sides, we have
S t = S 0 exp 0 t μ u r λ σ S , u ( 1 / u + λ + α ) λ S u S k 1 2 1 S u σ u 2 d u + 0 t 1 S u σ u d W u
which is the solution for S t .
The expected value of S t is given by
E [ S t ] = S 0 + E 0 t λ μ u r λ σ S , u 1 / u + λ + α λ S u S k d u = S 0 + 0 t E λ μ u r λ σ S , u 1 / u + λ + α λ S u S k d u = S 0 + 0 t λ μ u r λ σ S , u 1 / u + λ + α λ E [ S u ] E [ S k ] d u
Since we assume that S k converges to a long-run value z, we have E [ S k ] = z .
We can simplify the integral by defining the function g ( u ) = μ u r λ σ S , u 1 / u + λ + α λ E [ S u ] . Then, we have
E [ S t ] = S 0 + 0 t λ g ( u ) d u z 0 t λ d u
= S 0 z t + λ 0 t g ( u ) d u .
Therefore, the expected value of S t is S 0 z t + λ 0 t g ( u ) d u .
To calculate the variance of S t , we use the formula below. Using the expression for E [ S t ] , we can calculate the variance of S t as follows:
Var [ S t ] = E [ S t 2 ] ( E [ S t ] ) 2 = s 0 2 + 0 T 2 λ μ t r σ S , t ( 1 / t + λ + α ) 1 E [ S t ] E [ S k ] + 1 2 σ t 2 E [ S t ] d t ( E [ S t ] ) 2 = s 0 2 + 0 T 2 λ μ t r σ S , t ( 1 / t + λ + α ) 1 ( S 0 z t + λ 0 t g ( u ) d u ) E [ S k ] + 1 2 σ t 2 ( S 0 z t + λ 0 t g ( u ) d u ) d t ( S 0 z t + λ 0 t g ( u ) d u ) 2 = s 0 2 + 0 T 2 λ μ t r σ S , t ( 1 / t + λ + α ) 1 ( S 0 z t + λ 0 t g ( u ) d u ) E [ S k ] + 1 2 σ t 2 ( S 0 z t + λ 0 t g ( u ) d u ) d t ( S 0 2 2 S 0 z t + ( z t ) 2 2 S 0 λ 0 t g ( u ) d u + 2 z λ 0 t g ( u ) d u + λ 2 0 t g ( u ) d u ) 2
Note that we have used the fact that E [ S k ] = S 0 , since S k is a product of past realizations of S t , which are independent of S 0 . We have also used the expression for E [ S t ] in terms of g ( u ) .
Simplifying the above expression, we obtain
Var [ S t ] = s 0 2 + 0 T 2 λ ( μ t r σ S , t ( 1 / t + λ + α ) 1 × ( S 0 z t + λ 0 t g ( u ) d u ) E [ S k ] + 1 2 σ t 2 ) × ( S 0 z t + λ 0 t g ( u ) d u ) d t ( S 0 2 2 S 0 z t + ( z t ) 2 ) 2 λ ( S 0 z t ) 0 t g ( u ) d u λ 2 ( 0 t g ( u ) d u ) 2
and the history of z converges to a generalized gamma distribution using Mellins transformation and Ramanujan’s theorem.
s = k λ s S s d s Γ ( S ) ϕ k ( S )
For the unbounded case, the upper and lower bounds in Equation (24) relate to the outer limits of the isomophism, meaning that the probability space for u U B * is isomorphic to the probability space for z 1 Z B . When there is isomorphism between probability spaces, there is effectively reversible real-time transformation of a Brownian U t with respect to a reflexive Brownian Z t , such as
U t = Z t inf s 0 , t Z s
We relate to John Von Neumann’s mixed strategy inequality that justifies the existence of lower and upper bounds in u via the MaxMin and MinMax inequality:
sup u U inf z Z H f u , z inf z Z sup u U H f u , z
Equation (24) for unbounded isomorphic spaces will be
U t = Z + inf z Z sup u U H G u , z U t = Z t inf s 0 , t Z s U t = Z sup u U inf z Z H G u , z
Asymptotically, the error term space will be isomorphic with respect to the probability space of z ( t ) in Banach B and will have a reflection in B * . We rewrite the above sequence in a general form as follows:
u z t , z 1 ; Θ = z t M z t 1 ; θ 1 + g G z t ; z t 1 G z t ; z t 1
and
z t = α M ( z t 1 ; θ 1 ) g G ( z t ; z t 1 ) G ( z t ; z t 1 ) + u ( z t , z 1 ; Θ )
where M ( z t 1 ; θ 1 ) = θ 1 z t 1 is the memory component of the location functional.
Any sequence { u t } t = 1 in a Banach space B is isomorphic to a convex sequence G B * , the dual image of the reflection u U B . These sequences will be of the same type and co-type under the isomorphism induced by the dual pairing
u t , g for g B * .
G S T S t T t ζ s 0 , t
for any t, { u t } t = 1 and S T , G is any convex and smooth function of continuous mappings of Z B . A Banach space is called smooth if its norm is smooth. More formally, the function in B * , G : Z R is called 2 , D smooth and D * = 1 D . To bound the space, we can take G = . If D = 1 , then we have the inner product in Hilbert space L 2 , the inner product D = p 1 in a Banach space, and the vector space L p .
Here, we restate the functional central limit theorem (see, for instance, Davidson [7], Davidson [8]). The process S T is I(0) if it is defined as (0,1), such that
U T ρ = ω ρ 1 t = 1 T * ρ Z t E z t , 0 < ρ 1
or
U T ρ = ω u 1 t = 1 T * ρ Z t Z t 1 · e 2 i π p ρ t
where ω 2 = V a r t = 1 T u t = σ ρ 2 · T t . The process is Brownian as T t . The Brownian process holds true, with G · , E G s G r 2 = s r . r is some intermediate value between zero and s, 0 r < s 1 , if ω T 2 distr ω 2 · T t . Since ω 2 is variance, it could take any positive value, 0 < ω 2 < .
ρ = m U t · U ρ t e 2 i π · ζ
It is stated, [4], that for a random variable to present identifiable (unique) dynamics, the autocovariance function needs to be non-degenerated. This implies that the autocovariance function should be defined within a strictly positive compact set and volatility does not need to be bounded from above.
Let any convergent sequence be { S t } : as stated as in Equation (1), it will have an isomorphic solution that will be dependent on its history if C o v S , ρ 0 .
The implications of this statement are that the unconditional moment-generating function (MGF), ϕ k 1 p , and the conditional MGF, ϕ k p are not equal, and θ ^ k 0 . In cases where ϕ k p = ϕ k 1 p , we reject the proposed model as θ k = 0 for two random processes. Independence between them will be sustained as the MGF is constant, and in that case, θ .
When the sequence, which is subject to a concave function, converges to the Fisher information boundary, the functional central limit theorem holds. Then, the dependence guarantees the existence of a complete integrable system back in the Banach space, with the shape of a rugby ball moving in the phase space defined as the product of T N × R N . In the coordinates of the torus, if Z is the inclination axis, then T N = R N / ( 2 ) N , where a vector space in Banach is foliated by the N-dimensional invariant tori x = ( x 1 0 , x 2 0 , , x N 0 ) . Here, the integrable curves are quasi-periodic, with some frequency n i , and depend only on x 1 0 . Under dependence, there exists a dense subfamily of these tori, bearing strong solutions. The area of the rugby ball will change shape according to the relative mass shift of the parts of “everything”, with the ratio I ( p ) I ( 0 ) .

3. Empirical Calculations

We validate our theoretical framework through an empirical analysis, employing an extensive dataset taken from Factset including daily prices from the S&P 500, 10-year US Treasury bonds, the EUR/USD exchange rate, Brent oil, and Bitcoin, covering the period from 1 January 2002 to 1 February 2024. The out-of-sample predictions, conducted from 17 February 2019 to 1 February 2024, demonstrate the model’s superior forecasting power. There is no specific event tied to that date, and it does not correspond to any structural break. However, beginning the out-of-sample window in early 2019 allows the model to be tested through several important macro-financial developments, including the COVID-19 crisis, the post-pandemic recovery, the onset of monetary tightening cycles and spikes in gas and oil prices in 2022, the war in Ukraine, and the macroeconomic uncertainty following the US election and trade policies introduced during the Trump administration.
Let us define u t as a bounded stochastic sequence function dependent on the stochastic trajectory z ( t ) . Building on Equation (25), we introduce the autoregressive enhanced model (AREM) estimation equation as follows:
u t = z t + λ Δ z t 1 + min 1 z t z t 1 + , ζ s [ 0 , t ] z t Y 0 + Y 1 z t 1 z t λ Δ z t 1 max z t z t 1 1 , ζ s [ 0 , t ]
In our empirical implementation, we interpret the latent process
g ( u ) = μ u r λ σ S , u 1 u + λ + α λ E [ S u ]
as being effectively captured by the simulated trajectory z t , which reflects the evolution of the stochastic system governed by Equation (30). The drift component μ t is computed as a rolling exponentially weighted moving average (EWMA) of log returns, serving as a proxy for E [ g ( u ) ] , the expected value of the latent memory-driven process. Likewise, the volatility σ t calibrated from in-sample data approximates V [ g ( u ) ] , the variance in the evolving structural signal.
Equation (6) defines the recursive discrepancy process ρ k = w k · ρ k 1 , where ρ k = z k s k measures the forecast error between the simulated trajectory z k and the realized asset price s k . The multiplier w k , constructed directly from observed prices, captures local curvature and persistence in the price series. It plays a dual role in our framework. First, it governs the propagation of forecast errors: when w k > 1 , forecast discrepancies amplify, signaling instability or structural breaks; when w k < 1 , the system reverts toward equilibrium, indicating mean-reverting behavior. More structurally, w k acts as a state-switching mechanism that conditions the dynamics within Equation (30). Depending on its value, the model shifts the weight placed on lagged returns and volatility-adjusted drift signals, thereby endogenously selecting between explosive, trending, or mean-reverting regimes. This switching is consistent with the theoretical formulation in Equation (17), where g ( u ) aggregates past volatility-weighted expectations and plays a role in determining conditional behavior.
The primary objective of this implementation is to evaluate the model’s ability to replicate out-of-sample price trajectories using in-sample calibrated dynamics. These dynamics embed historical memory and are structurally aligned with the duality properties implied by our Banach space formulation.
The switching mechanism embedded in u t , incorporating bounded adjustments through min and max operators as well as lagged return differentials, enables asymmetric adaptation to past price behavior. This is consistent with the recursive decay structure implied by the parameters λ and α . We start by splitting our sample period in two and use the first section, comprising the initial 80% of the sample period, to perform an in-sample estimation of parameters. The remaining 20% is then used to perform an out-of-sample forecasting exercise, computing one-day-ahead prices and updating parameter estimates. The coefficients reported in Table 1 correspond to the calibrated components of this filtering mechanism: λ governs the weight on the lagged momentum term Δ z t 1 , σ features in the normalization of the drift signal ζ s , and r reflects the long-run drift anchoring the stochastic path of S k . Together, these parameters define a data-driven recursive filter that aligns the simulated dynamics with the structural theoretical model.
To contrast the accuracy of the results, an MA model and a naive specification were also obtained for each series. Table 1 presents the results for all assets, whereas the results for the entire series are presented in Table 2. This table presents the relevant statistical measures: the mean square error, the root mean square error, the mean absolute error, and the direction accuracy. Here, we report the percentage of correct guesses for going up or down for the three models to illustrate the extent to which the proposed framework outperforms the alternatives; while R 2 is commonly reported in forecasting studies, it may be artificially inflated when applied to level forecasts due to the high variance in price levels relative to returns. As such, it can overstate predictive accuracy in level-based models. We, therefore, emphasize directional accuracy as a more reliable and informative performance metric for evaluating whether the model correctly anticipates the direction of price movements, and R 2 values are not reported. Our model yields a strikingly high percentage of correct predictions, ranging from 73% to 92%, in contrast to the 47% to 54% accuracy of naïve and moving average models.
The implications of w k presented in Section 2 will be discussed here using the S&P500, crude oil prices, 10-year US Treasury yield, Bitcoin, and EUR/USD exchange rate, and we map the w k instances over the out-of-sample period, 2019–2025.
Recall that a martingale is a stochastic process where the expectation of the next step, given all the past steps, is equal to the present step:
E [ S k + 1 | S 0 , S 1 , , S k ] = S k .
S will behave like a martingale when there is no change in the correlation function ρ , which means that w k = 1 and E [ g ( k 1 ) · ϵ k ( p ) ] = 0 .
The expected future value is equal to the present value given all the information up to the present. When w k = 1 , the system seems to satisfy the martingale property because it implies no net gain or net loss in the expected value of S, given all the past values.
If w k 1 , the process S does not behave as a martingale because the expectation of g ( k 1 ) · ϵ k ( p ) is not zero, indicating that the expected future value of S would not be its present value, thereby violating the martingale property. In summary, if w k = 1 , S behaves as a martingale. If w k 1 , it does not.
If w k = 0 , from Equation w k = S k S k 2 S k 1 2 , it could mean one of a few things: either S k = 0 or S k 2 = 0 , assuming that S k 1 is not zero.
From Equation E [ g ( k 1 ) · ϵ k ( p ) ] = E [ ( 1 w k ) g ( k 1 ) ] · ϕ k 1 ( p ) , we see that if w k = 0 , then E [ g ( k 1 ) · ϵ k ( p ) ] = E [ g ( k 1 ) ] · ϕ k 1 ( p ) . This equation implies that the expected change ϵ k ( p ) = ϕ k ( p ) ϕ k 1 ( p ) becomes dependent on ϕ k 1 ( p ) and g ( k 1 ) , which could make the series diverge, oscillate, or go to zero depending on the terms.
Furthermore, w k = 0 would indicate that the S process is not a martingale in this situation since the expected future change in ϕ k is not zero. Instead, it is influenced by the values of ϕ k 1 and g ( k 1 ) . When w k = 0 , S will not behave like a martingale, and the behavior could be quite deterministic, depending on the specifics of ϕ k 1 and g ( k 1 ) .
In this framework, w k = S k S k 2 S k 1 2 . If w k > 1 , it implies that S k S k 2 > S k 1 2 , given that S k 1 is not zero. This suggests that either S k or S k 2 (or both) is larger in magnitude than S k 1 when considering the product.
When w k > 1 , 1 w k < 0 , which means that the sign of E [ ( 1 w k ) g ( k 1 ) ] · ϕ k 1 ( p ) will be opposite to ϕ k 1 ( p ) . This implies that the moment-generating function ϕ k ( p ) will diverge from ϕ k 1 ( p ) in a manner influenced by g ( k 1 ) .
This indicates that the series S could be either increasing or decreasing, depending on the specifics of the other terms, but does not behave as a martingale because a martingale would require that the expectation of future changes is zero. Therefore, this suggests that the process S is expected to increase if it has increased in the past (or decrease if it has decreased in the past). This would indicate a kind of momentum or trend-following behavior.
In the case where w k > 1 , the term 1 w k would be negative. If ϕ k 1 ( p ) and g ( k 1 ) are positive, then the expected value E [ ϵ k ( p ) ] would be positive given that S has increased in the past, suggesting that S will continue to increase.
If w k tends to infinity, the term 1 w k will tend to negative infinity, making the term E [ ( 1 w k ) g ( k 1 ) ] extremely negative (assuming that g ( k 1 ) is positive).
This implies that the change in the moment-generating function, ϵ k ( p ) , will also be extremely large (and negative), indicating an explosive behavior in the negative direction for S.
The condition w k indicates a divergence away from martingale behavior, where future values of S are expected to be the same as the current value given all past information. In fact, the future values of S would be expected to decrease rapidly, making the process far from a martingale.
Thus, if w k tends to infinity, S would likely show extremely volatile and downward behavior, and it would certainly not satisfy the martingale condition.
If w k tends to negative infinity, then the term 1 w k will tend to positive infinity. This would make the term E [ ( 1 w k ) g ( k 1 ) ] extremely large in the positive direction (assuming that g ( k 1 ) is positive).
Similar to the case where w k tends to positive infinity, this would indicate a divergence from martingale behavior. A martingale process should have future values that are the same as the current value given all past information. In this case, the future values of S would be expected to increase rapidly, making the process explosive.
Further, if w k < 0 and ρ k 1 < 0 , then ρ k would be positive, because the product of two negative numbers is positive. If w k < 0 and ρ k 1 > 0 , then ρ k would be negative, because the product of a negative number and a positive number is negative. If w k < 0 and ρ k 1 = 0 , then ρ k would be zero. However, if w k < 0 , it indicates some form of reversal or “flip” in the relationship from ρ k 1 to ρ k , which could suggest some form of mean reversion or cyclical behavior in S depending on the broader context. The derivation of orthogonality conditions enables the detection of whether the process S k behaves as a martingale ( w k = 1 ) or deviates toward trend-following, mean-reverting, or explosive behavior ( w k 1 ). These conditions support a regime classification framework embedded within the model, enhancing its interpretability and predictive power. Similar goals are pursued in [29], who evaluate how time-varying idiosyncratic exposures impact return dynamics across regimes. Related methodologies [19,20] are developed for detecting structural breaks and forecast instability, though these are implemented as separate diagnostic tools. In contrast, our framework integrates regime identification directly into the stochastic process via analytically derived orthogonality weights, enabling the endogenous detection of evolving dynamics from within the model structure itself.
In Figure 1, we filter the data to create a map of martingale, trending, mean reverting, and explosive values of the S&P500.
In Figure 2, we filter the data to create a map of martingale, trending, mean reverting, and explosive values of crude oil.
In Figure 3, we filter the data to create a map of martingale, trending, mean reverting, and explosive values of the 10-year US Treasury yield.
In Figure 4, we filter the data to create a map of martingale, trending, mean reverting, and explosive values of the EUR/USD exchange rate.
We evaluate the model based on five major financial instruments spanning equities, commodities, interest rates, cryptocurrencies, and foreign exchange rates. These include the S&P 500 Index (GSPC), crude oil prices (CL=F), the 10-year Treasury yield (TNX), Bitcoin (BTC-USD), and the EUR/USD exchange rate (EURUSD=X). For each case, the same model structure and calibration procedure are applied. All simulations are conducted over a six-year out-of-sample window (2019–2025). Since we performed out-of-sample comparison, we decided to extend our sample to March 2025. The macroeconomic uncertainty of the new US administration is now considered. using parameters estimated exclusively from data prior to 2019. This extended horizon allows us to test the model’s robustness without recalibration. As expected, the longer the model projects into the future without updating its parameters, the greater the divergence from actual market outcomes. Nevertheless, the model achieves remarkably strong performance across a range of assets.
The model shows excellent predictive alignment for Bitcoin (see Figure 5), where explosive volatility and strong directional trends are mirrored closely by the simulated path.
Similarly, in the S&P 500 Index (see Figure 6), the simulation reproduces the equity market’s upward trajectory, including the COVID-19 shock and post-2020 bull run.
For the 10-year Treasury yield (see Figure 7), the model tracks interest rate cycles and trend shifts with impressive consistency.
Crude oil price presents a complex but well-captured cycle of collapse, rebound, and stabilization, where the model captures not only the directional momentum, but also much of the cyclical behavior (seeFigure 8).
The EUR/USD exchange rate presents the greatest challenge due to its mean-reverting and policy-sensitive nature (see Figure 9).However, we observe that between 2019 and mid-2022, the model reproduces the actual path with exceptional precision. As we move forward in time without updating the parameters, the error grows with resulting deterioration in fit in the latter part of the sample, after significant structural changes in macroeconomic conditions, particularly the divergence in ECB and Fed policy post pandemic. These limitations are not faults of the model structure per se, but rather a consequence of using static, pre-2019 parameters for a system that evolved dramatically in later years.
Although the overall fit deteriorates after mid-2022, the model reproduces the price path of EUR/USD from 2019 to early 2022 with impressive accuracy. This initial alignment validates the relevance of our structural drift specification before exogenous macro shocks and interest rate divergences distort the dynamics.
Most modern AI and machine learning forecasting systems, including LSTMs, neural regressors, and hybrid models, also target levels rather than returns. This is largely because price levels are directly observable, whereas returns are derived quantities that can amplify noise or obscure structural features of the data. Forecasting levels enables models to better retain information related to long-term growth trends and volatility persistence. In line with this, the AREM in this paper forecasts price levels, not returns. Price levels are explicitly driven by latent volatility and embedded growth rates that determine the dynamic trajectory of prices across time. These structural drivers—rather than differenced returns—form the natural foundation for modeling historically dependent stochastic processes.
Our model yields a strikingly high percentage of correct predictions, ranging from 73% to 92%, in contrast to the 47% to 54% accuracy of naïve and moving average models.
We measure whether the model correctly anticipates the direction of price movement, i.e., whether the price increased when the model predicted an upward movement. Directional accuracy is especially appropriate in the context of level forecasting, where capturing turning points and directional shifts is more meaningful than predicting raw return magnitudes. Furthermore, when the level forecasts achieve such high out-of-sample correlation—97% for Bitcoin, 98% for the S&P500, and 98% for 10-year Treasury yields—it also implies that the derived return series would exhibit similarly high correlation with realized returns. In other words, if the model replicates the level path of an asset with high fidelity, it necessarily captures the directional movements and the relative return dynamics implied by that trajectory.
The empirical validity is evaluated through a rigorous out-of-sample framework extending from 2019 to 2025, using parameter estimates calibrated exclusively on pre-2019 data. This represents one of the most extensive multi-asset out-of-sample validation exercises in the literature. Among the reviewed studies, no study offers a comparably long and continuous out-of-sample evaluation spanning daily data across multiple asset classes and lasting up to six years as that implemented in this paper; while some works such as [21,24] perform robust out-of-sample validations, their horizons are notably shorter and often limited to equity returns or quarterly macro-indicators. Refs. [30,31] focus primarily on in-sample or short-horizon predictive regressions. Among the related studies employing forecasting methods across dynamic systems, we note the comparative study by [15], which contrasts classical mathematical models (SIR-type) with machine learning techniques in the context of COVID-19 spread prediction. Their analysis, focused on the Moscow region, evaluates short-term forecasts using multiple data-driven approaches, while their work highlights the potential of hybrid frameworks and explores modeling under uncertainty, the forecast horizon remains relatively limited—typically ranging from several days to a few weeks. Whereas Pavlyutin et al. emphasize model comparison under changing epidemic conditions, our work focuses on embedding historical dependence structurally and evaluating forecast accuracy in financial contexts subject to regime shifts and latent memory effects. In [32] emphasize volatility decomposition and connectedness, but do not provide out-of-sample forecast evaluation. Our results therefore offer a rare example of sustained model performance over six years and across diverse asset classes—including equities, interest rates, FX, oil, and cryptocurrency—using consistent calibration and structure. This out-of-sample robustness is a distinguishing feature of our approach, validating the structural memory captured in the adaptive regime-awareness in AREM.
To provide a more quantitative benchmark against prior research, we summarize directional accuracy figures in Table 3, drawing from recent papers addressing asset return predictability under various modeling assumptions. As shown, our AREM model delivers directional accuracy in the range of 73–93%, outperforming traditional models like Fama–MacBeth regressions [33], which average around 60%, and moment-based approaches such as [31], who report accuracies near 68% for U.S. equities. The frequency domain connectedness metrics of [34], while insightful on volatility transmission, do not explicitly report predictive statistics. Similarly, ref. [30] achieve an in sample R 2 53%–75% using latent factor state-space models on international equity panels. The comparative advantage of our AREM framework lies in its ability to dynamically adapt to structural changes without requiring full state observability or restrictive moment assumptions. Overall, our empirical model stands favorably relative to both conventional econometric and modern ML-based alternatives.
To provide a comprehensive benchmark, Table 3 consolidates the predictive performance of our models with a broad range of recent forecasting approaches in the literature. This table includes both statistical and machine learning frameworks, spanning structural econometric models, regime-switching techniques, neural networks, and hybrid systems. Our AREM model exhibits directional accuracy between 73% and 93%, placing it at the upper end of the performance spectrum. It achieves exceptionally high out-of-sample correlation with realized price levels—97% for Bitcoin, 98% for the S&P500, and 98% for 10-year Treasury yields—indicating its robustness in replicating price trajectories. These results are particularly notable given the long out-of-sample window (2019–2025) and fixed pre-2019 calibration. It performs competitively or exceeds the standards established in the literature across multiple metrics and asset domains.
By contrast, studies such as [21,24] report accuracies in the range of 70–85%, with varying assumptions and data scopes. Traditional econometric models, like those in [31,33] generally deliver lower directional accuracy (60–68%), and are often limited by static factor structures or moment conditions. Machine learning methods, such as those explored by [23,25,27], achieve moderate-to-strong performance depending on the forecast horizon and asset class, but often lack interpretability. Other contributions, such as, [18,20,34] provide valuable theoretical or diagnostic tools but do not report explicit out-of-sample metrics.

4. Conclusions

This paper extends the foundational work of [4], who argued that the evolution of a stochastic process, S k , is independent of its observable history under orthogonality assumptions. In their framework, the unobserved solution cannot be functionally linked to the observed data, implying a lack of duality and irreversibility in the underlying process. Under such conditions, history plays no role in the generation of future states, and predictability is fundamentally unattainable. Here, we have shown that under well-defined conditions in a reflexive Banach space, the evolution of the process S k is, indeed, historically dependent. Specifically, we demonstrate that for the majority of existing probabilistic configurations, three-quarters of all measures, the process exhibits dependence on its past. This historical dependence is formalized through the recursive relationship S k = j = k ρ j · S k 1 and through orthogonality conditions derived from moment-generating functions of correlated sequences.
One important implication is that the dependence and independence alternatives for a general stochastic solution process would, indeed, depend on its known history. We argue that multidimensional integration has duality in reflexive Banach space if the solution obeys the functional central limit theorem and its error is between the limits of dilation and erosion. Hence, there is isomorphism for closed and semi-open intervals and the desired inner product to define the orthogonality conditions. Beyond that, the space is composed of countably additive and possibly finite measures endowed only with a measure and not a norm.
Another important result is the derivation of explicit orthogonality conditions, which enable the detection of whether the process S k behaves as a martingale ( w k = 1 ) or deviates toward trend-following, mean-reverting, or explosive behavior ( w k 1 ). These conditions support a regime classification framework embedded within the model, enhancing interpretability and predictive power. We further show that multidimensional integration has duality in reflexive Banach space if the solution satisfies a functional central limit theorem and its error is bounded between dilation and erosion. As such, the stochastic process is forecastable in most configurations, with explicit structural limits on predictability imposed by the geometry of the function space.
Duality, as established in the reflexive Banach space framework, plays a fundamental role in enhancing the model’s robustness. When the observed sequence lies in a space isomorphic to that of the unobserved process, it inherits essential topological properties—reflexivity, completeness, and convexity—which ensure that forecastable structure is preserved. This duality forms the theoretical basis for the model’s capacity to extract meaningful information from noisy or incomplete observations.
Empirically, AREM demonstrates substantial gains in predictive power relative to moving average and naïve models. Across assets with vastly different characteristics—including the S&P 500, Bitcoin, and US Treasury bonds—AREM achieves directional accuracy between 73% and 93%, outperforming traditional benchmarks that falter in the presence of non-stationarity and heavy tails. The orthogonality and duality framework embedded in the model is instrumental in capturing and adapting to dynamic changes in the data-generating process. The dual approach strengthens our understanding of historical dependence, structural predictability, and their implications for forecasting under incomplete information.
The duality-based stochastic process model presented in this paper offers significant potential for enhancing cybersecurity in financial systems. By establishing a rigorous mapping between observable financial data and latent stochastic drivers, our model facilitates structural auditability and anomaly detection. This is particularly pertinent given the increasing sophistication of cyber threats targeting financial institutions. For instance, the National Institute of Standards and Technology (NIST) has emphasized the need for robust defenses against adversarial machine learning attacks, which can compromise the integrity of financial models [35]. Additionally, recent studies have highlighted the vulnerability of financial systems to advanced persistent threats (APTs) and the necessity of mathematical models for effective cyber risk management [36]. Our model’s capacity to detect deviations from expected stochastic behavior can serve as an early warning mechanism against such threats, thereby reinforcing the cybersecurity infrastructure of financial institutions. Recent work by [37] proposes a two-stage stochastic programming model for cybersecurity investment that incorporates attacker behavior, government intelligence, and firm-level safeguards. Their framework mirrors our approach by embedding long-memory dependence and recursive decision-making under uncertainty. Moreover, studies such as [38,39] emphasize that cyber–physical and network security systems require robust mathematical models that accommodate stochastic attacks and degeneracy in observed intrusion patterns. These conditions align with the orthogonality and reflexivity principles central to our modeling strategy. The ability of our model to detect breaks in structural correspondence—when the observed process S k deviates from its latent generator—is analogous to identifying anomalous states or adversarial threats in a cybersecurity environment. This suggests that our theoretical tools can contribute to dynamic detection, resilience modeling, and investment optimization in the cybersecurity domain, particularly in high-frequency financial systems and adversarial machine learning contexts.

Author Contributions

Conceptualization, I.P. and A.S.; methodology, I.P. and A.S.; software, I.P. and A.S.; validation, I.P. and A.S.; investigation, I.P. and A.S.; resources, I.P. and A.S.; data curation, I.P. and A.S.; writing—original draft preparation, I.P. and A.S.; writing I.P. and A.S.; visualization, I.P. and A.S.; supervision, I.P. and A.S.; project administration, I.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data supporting reported results can be found here https://github.com/paraskya/Stochastic_Evolution_Data.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chichilnisky, G. The Limits of Econometrics: Nonparametric Estimation in Hilbert Spaces. Econom. Theory 2009, 25, 1070–1086. [Google Scholar] [CrossRef]
  2. Bergstrom, A. The estimation of nonparametric functions in a Hilbert space. Econom. Theory 1985, 1, 7–26. [Google Scholar] [CrossRef]
  3. Tsirelson, B. Triple points: From non-Brownian filtrations to harmonic measures. Geom. Funct. Anal. 1997, 7, 1096–1142. [Google Scholar] [CrossRef]
  4. Yano, K.; Yor, M. Around Tsirelson’s equation, or: The evolution process may not explain everything. Probab. Surv. 2015, 12, 1–12. [Google Scholar] [CrossRef]
  5. Lu, Y.; Lin, D.; Du, Q. Which Spaces can be Embedded in L_p-type Reproducing Kernel Banach Space? A Characterization via Metric Entropy. arXiv 2024, arXiv:2410.11116. [Google Scholar]
  6. Pennanen, T.; Perkkiö, A.P. Dual spaces of càdlàg processes. Stoch. Processes Their Appl. 2023, 157, 69–93. [Google Scholar] [CrossRef]
  7. Davidson, J. When is a time series I(0). In The Methodology and Practice of Econometrics: A Festschrift in Honour of David F. Hendry; Oxford University Press: Oxford, UK, 2009; pp. 322–342. [Google Scholar]
  8. Davidson, R. Time and Causality. Ann. Econ. Stat. 2013, 111–112, 7–22. [Google Scholar] [CrossRef]
  9. Esmaeili, V.; Tamura, K.; Muscinelli, S.P.; Modirshanechi, A.; Boscaglia, M.; Lee, A.B.; Oryshchuk, A.; Foustoukos, G.; Liu, Y.; Crochet, S.; et al. Rapid suppression and sustained activation of distinct cortical regions for a delayed sensory-triggered motor response. Neuron 2021, 109, 2183–2201. [Google Scholar] [CrossRef]
  10. Hamilton, J.D. Time Series Analysis; Princeton University Press: Princeton, NJ, USA, 2020. [Google Scholar]
  11. Bai, J.; Perron, P. Estimating and testing linear models with multiple structural changes. Econometrica 1998, 66, 47–78. [Google Scholar] [CrossRef]
  12. Wang, X.; Hsieh, F. An encoding approach for stable change point detection. Mach. Learn. 2024, 113, 4133–4163. [Google Scholar] [CrossRef]
  13. Mandelbrot, B.B.; Van Ness, J.W. Fractional Brownian motions, fractional noises and applications. SIAM Rev. 1968, 10, 422–437. [Google Scholar] [CrossRef]
  14. Engle, R.F.; Rangel, J.G. The spline-GARCH model for low-frequency volatility and its global macroeconomic causes. Rev. Financ. Stud. 2008, 21, 1187–1222. [Google Scholar] [CrossRef]
  15. Pavlyutin, M.; Samoyavcheva, M.; Kochkarov, R.; Pleshakova, E.; Korchagin, S.; Gataullin, T.; Nikitin, P.; Hidirova, M. COVID-19 Spread Forecasting, Mathematical Methods vs. Machine Learning, Moscow Case. Mathematics 2022, 10, 195. [Google Scholar] [CrossRef]
  16. Chandiraraj, R.E.R. A stochastic difference equation with stationary noise on groups. Can. J. Math. 2018, 70, 1058–1090. [Google Scholar] [CrossRef]
  17. Carinci, G.; Giardinà, C.; Giberti, C.; Redig, F. Duality for stochastic models of transport. Stoch. Processes Their Appl. 2015, 125, 941–969. [Google Scholar] [CrossRef]
  18. Dette, H.; Kokot, K.; Aue, A. Functional data analysis in the Banach space of continuous functions. Ann. Stat. 2020, 48, 1168–1192. [Google Scholar] [CrossRef]
  19. Harvey, D.I.; Leybourne, S.J.; Zu, Y. Tests for forecast breakdown and structural change. J. Econom. 2016, 191, 179–189. [Google Scholar]
  20. Castle, J.L.; Doornik, J.A.; Hendry, D.F. Robust discovery of regression breakpoints. J. Econom. 2015, 184, 414–425. [Google Scholar]
  21. Cai, Z.; Chang, S.Y. A new test on asset return predictability with structural breaks. J. Financ. Econom. 2024, 22, 1042–1074. [Google Scholar] [CrossRef]
  22. Demetrescu, M.; Georgiev, I.; Rodrigues, P.M.M. Testing for episodic predictability in stock returns. J. Econom. 2022, 227, 85–113. [Google Scholar] [CrossRef]
  23. Medeiros, M.C.; Vasconcelos, G.F.R.; Veiga, Á.; Zilberman, E. Forecasting inflation in a data-rich environment: The benefits of machine learning methods. J. Bus. Econ. Stat. 2021, 39, 98–119. [Google Scholar] [CrossRef]
  24. Gu, S.; Kelly, B.; Xiu, D. Empirical asset pricing via machine learning. J. Financ. Econ. 2020, 140, 809–835. [Google Scholar] [CrossRef]
  25. Smyl, S. A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting. Int. J. Forecast. 2020, 36, 75–85. [Google Scholar] [CrossRef]
  26. Benidis, K.; Rangapuram, S.S.; Flunkert, V.; Wang, B.; Maddix, D.C.; Turkmen, C.; Gasthaus, J.; Bohlke-Schneider, M.; Salinas, D.; Stella, L.; et al. Neural forecasting: Introduction and literature overview. Front. Artif. Intell. 2020, 3, 12. [Google Scholar]
  27. Bauwens, L.; Storti, G. Modelling and forecasting cryptocurrency volatility: An LSTM-based approach. J. Financ. Econom. 2022, 20, 368–392. [Google Scholar]
  28. Rossi, B. Forecasting in the presence of instabilities. Annu. Rev. Econ. 2021, 13, 321–345. [Google Scholar]
  29. Sautner, Z.; van Lent, L.; Vilkov, G.; Zhang, R. Pricing climate change exposure. J. Financ. Econ. 2023, 149, 296–318. [Google Scholar] [CrossRef]
  30. De Jong, F.; Driessen, J. Liquidity risk premia in corporate bond markets. Q. J. Financ. 2012, 2, 1250006. [Google Scholar] [CrossRef]
  31. Bali, T.G.; Engle, R.F. The intertemporal capital asset pricing model with dynamic conditional correlations. J. Monet. Econ. 2010, 57, 377–390. [Google Scholar] [CrossRef]
  32. Brennan, M.J.; Xia, Y. Dynamic Asset Allocation under Inflation. J. Financ. 2002, 57, 1201–1238. [Google Scholar] [CrossRef]
  33. Ang, A.; Hodrick, R.J.; Xing, Y.; Zhang, X. The cross-section of volatility and expected returns. J. Financ. 2006, 61, 259–299. [Google Scholar] [CrossRef]
  34. Baruník, J.; Křehlík, T. Measuring the frequency dynamics of financial connectedness and systemic risk. J. Financ. Econom. 2018, 16, 271–296. [Google Scholar] [CrossRef]
  35. National Institute of Standards and Technology. Adversarial Machine Learning—NIST Technical Series Publications; NIST: Gaithersburg, MD, USA, 2025.
  36. Sobchuk, V.; Barabash, O.; Musienko, A.; Tsyganivska, I.; Kurylko, O. Mathematical Model of Cyber Risks Management Based on the Expansion of Piecewise Continuous Analytical Approximation Functions of Cyber Attacks in the Fourier Series. Axioms 2023, 12, 924. [Google Scholar] [CrossRef]
  37. Paul, S.K.; Zhang, J. A two-stage stochastic programming model for cybersecurity investment decision-making. Ann. Oper. Res. 2021, 303, 177–207. [Google Scholar]
  38. Abu Al-Haija, Q. On the Security of Cyber-Physical Systems Against Stochastic Cyber-Attacks Models. In Proceedings of the 2021 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS), Toronto, ON, Canada, 21–24 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar] [CrossRef]
  39. Leslie, N.O.; Harang, R.E.; Knachel, L.P.; Kott, A. Statistical models for the number of successful cyber intrusions. J. Def. Model. Simul. 2017, 15, 49–63. [Google Scholar] [CrossRef]
Figure 1. w k on S&P500.
Figure 1. w k on S&P500.
Mathematics 13 02002 g001
Figure 2. w k on crude oil prices.
Figure 2. w k on crude oil prices.
Mathematics 13 02002 g002
Figure 3. w k on 10-year US Treasury yield.
Figure 3. w k on 10-year US Treasury yield.
Mathematics 13 02002 g003
Figure 4. w k on EUR/USD.
Figure 4. w k on EUR/USD.
Mathematics 13 02002 g004
Figure 5. Out-of-sample simulation of Bitcoin (2019–2025).
Figure 5. Out-of-sample simulation of Bitcoin (2019–2025).
Mathematics 13 02002 g005
Figure 6. Out-of-sample simulation of S&P 500 Index (2019–2025).
Figure 6. Out-of-sample simulation of S&P 500 Index (2019–2025).
Mathematics 13 02002 g006
Figure 7. Out-of-sample simulation of 10-year US Treasury yield (2019–2025).
Figure 7. Out-of-sample simulation of 10-year US Treasury yield (2019–2025).
Mathematics 13 02002 g007
Figure 8. Out-of-sample simulation of crude oil prices (2019–2025).
Figure 8. Out-of-sample simulation of crude oil prices (2019–2025).
Mathematics 13 02002 g008
Figure 9. Out-of-sample simulation of EUR/USD (2019–2025).
Figure 9. Out-of-sample simulation of EUR/USD (2019–2025).
Mathematics 13 02002 g009
Table 1. Calibrated model parameters by asset.
Table 1. Calibrated model parameters by asset.
Asset λ α r σ Corr (Out)RMSE (Out)
S&P 500 Index (^GSPC)0.2030.0270.01520.1880.98520.3479
Crude Oil (CL=F)0.1140.0180.00940.2630.93210.1939
10-Year Treasury Yield (^TNX)0.1870.0420.02210.1750.98790.1679
Bitcoin (BTC-USD)0.2210.0350.03240.3120.96642.4018
EUR/USD (EURUSD=X)0.1730.0290.00780.1400.03910.9622
Table 2. Out-of-sample results for all metrics and models.
Table 2. Out-of-sample results for all metrics and models.
AssetMetricAREMANM
US Treasury 10-Year YieldMSE0.000590.154530.00323
RMSE0.024340.393110.05681
MAE0.011450.306420.04160
Direction Accuracy93.27%53.71%46.92%
EurUSDMSE0.000000.000840.00003
RMSE0.002080.028970.00501
MAE0.001500.023890.00374
Direction Accuracy90.9%51.74%50.25%
SP500MSE570.1925950,660.457532283.47540
RMSE23.87871225.0787847.78572
MAE16.47460188.4049932.96814
Direction Accuracy74.53%55.64%48.86%
Brent OilMSE2.06514117.747008.86438
RMSE1.410.851132.97731
MAE0.675398.225591.39790
Direction Accuracy75.62%55.65%49.62%
BitCoinMSE482,165.6567460,169,855.594551,928,852.72200
RMSE694.381497756.923081388.83142
MAE466.538796159.32657933.12826
Direction Accuracy73.18%53.56%47.68%
AREM = autoregressive enhanced model, MA = moving average, NM = naive model. The out-of-sample period starts on 17 February 2019 and ends on 1 February 2024.
Table 3. Extended comparison of forecasting models in the recent literature.
Table 3. Extended comparison of forecasting models in the recent literature.
StudyMethodologyAccuracy/RMSENotes
This Paper AREMStructural, orthogonality-aware model73–93% (DA)Multi-asset, 6-year out-of-sample
This Paper (Equation (17)Recursive memory price simulationHigh CorrelationAccurate path shape (price levels)
[21]Structural break predictability test70–85% (DA)Stocks, rolling forecast
[24]ML with return-sorted predictors∼75% (DA)Requires feature expansion
[23]ML for inflation nowcasting∼65–80%Macroeconomic data
[27]LSTM for crypto volatilityRMSE improvedBitcoin-focused
[25]Hybrid RNN + exponential smoothing5–15% RMSE gainsShort-term forecasts
[33]Idiosyncratic volatility (Fama-MacBeth)~60%Equities (cross-sectional)
[31]Conditional skewness/kurtosis~68%U.S. equities
[34]Frequency-domain ConnectednessNot reportedFX, bonds
[30]Latent factor state-space model~53–75%International equities
[26]Neural forecast meta-reviewMixedEmphasis on scalability
[19]Forecast breakdown testsn/aDetection, not prediction
[20]Robust break detectionn/aModel selection framework
[18]FDA in Banach spacen/aFunctional data regressions
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Paraskevopoulos, I.; Santos, A. The Stochastic Evolution of Financial Asset Prices. Mathematics 2025, 13, 2002. https://doi.org/10.3390/math13122002

AMA Style

Paraskevopoulos I, Santos A. The Stochastic Evolution of Financial Asset Prices. Mathematics. 2025; 13(12):2002. https://doi.org/10.3390/math13122002

Chicago/Turabian Style

Paraskevopoulos, Ioannis, and Alvaro Santos. 2025. "The Stochastic Evolution of Financial Asset Prices" Mathematics 13, no. 12: 2002. https://doi.org/10.3390/math13122002

APA Style

Paraskevopoulos, I., & Santos, A. (2025). The Stochastic Evolution of Financial Asset Prices. Mathematics, 13(12), 2002. https://doi.org/10.3390/math13122002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop