Next Article in Journal
A Complexity-Based Approach to Quantum Observable Equilibration
Previous Article in Journal
Linking Optimization Success and Stability of Finite-Time Thermodynamics Heat Engines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling and Generating Extreme Fluctuations in Time Series with a Multilayer Linear Response Model

1
Graduate School of Management, Tokyo Metropolitan University, 18F Marunouchi Eiraku Building, 1-4-1 Marunouchi, Chiyoda-ku, Tokyo 100-0005, Japan
2
Department of Liberal Arts, Hiroshima University of Economics, 5-37-1 Gion, Asaminami-ku, Hiroshima 731-0192, Japan
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(8), 823; https://doi.org/10.3390/e27080823 (registering DOI)
Submission received: 29 June 2025 / Revised: 30 July 2025 / Accepted: 1 August 2025 / Published: 3 August 2025
(This article belongs to the Section Complexity)

Abstract

A multilayer linear response model (MLRM) is proposed to generate time-series data based on linear response theory. The proposed MLRM is designed to generate data for anomalous dynamics by extending the conventional single-layer linear response model (SLRM) into multiple layers. While the SLRM is a linear equation with respect to external forces, the MLRM introduces nonlinear interactions, enabling the generation of a wider range of dynamics. The MLRM is applicable to various fields, such as finance, as it does not rely on machine learning techniques and maintains interpretability. We investigated whether the MLRM could generate anomalous dynamics, such as those observed during the coronavirus disease 2019 (COVID-19) pandemic, using pre-pandemic data. Furthermore, an analysis of the log returns and realized volatility derived from the MLRM-generated data demonstrated that both exhibited heavy-tailed characteristics, consistent with empirical observations. These results indicate that the MLRM can effectively reproduce the extreme fluctuations and tail behavior seen during high-volatility periods.

1. Introduction

Recently, remarkable progress has been made in data generation technology, especially in the text and image generation fields. In addition, attempts have been made in the financial sector to generate financial time-series data, e.g., stock prices, with numerous studies utilizing generative adversarial networks (GANs) [1] for this purpose. For example, Ref. [2] proposed Quant GAN, which comprises generator and discriminator functions using temporal convolutional networks (TCNs) [3] to capture long-range dependencies, e.g., the presence of volatility clusters. Similarly, Ref. [4] proposed Stock GAN, based on the Wasserstein GAN (WGAN) [5,6], to address the optimal transport problem [7]. More recently, Ref. [8] proposed Sig-Wasserstein GAN (Sig-WGAN), which combines rough path theory [9] and the expected signature [10] to improve training stability and sample quality. Building on this, Ref. [11] extended the Sig-WGAN framework to the conditional setting and proposed a conditional Sig-WGAN for modeling the distribution of future time series given past observations. Moreover, Ref. [12] proposed a conditional GAN-based market simulator that reacts to experimental agent behavior, enabling realistic and responsive market simulations to evaluate trading strategies. In addition, Ref. [13] introduced Fin-GAN, a GAN-based model for financial time series forecasting, which outperforms traditional methods like LSTM [14] and ARIMA [15] in both predictive accuracy and classification performance.
Recently, diffusion models [16] have been developed rapidly and employed for the generation of time-series data. For example, Ref. [17] introduced a two-way variational autoencoder, D 3 VAE, with diffusion, denoising, and disentanglement, to address multivariate time-series forecasting problems. The D 3 VAE model improves denoising diffusion and score matching by treating discrete observations as continuous values in the function space. In the context of missing value imputation, which is critical for practical financial applications, Ref. [18] proposed a conditional score-based diffusion model (CSDI). Furthermore, Ref. [19] introduced a structured state space diffusion (SSSD) model, which is particularly effective at capturing long-term dependencies in time-series data. More recently, Ref. [20] proposed a conditional diffusion model as a denoiser for financial time-series data, demonstrating an improved signal-to-noise ratio, return predictability, and trading performance compared to existing methods such as denoising autoencoders and moving averages. In addition, Ref. [21] proposed FTS-Diffusion, a diffusion-based generative framework that specifically models the irregular and scale-invariant patterns of financial time series, achieving high fidelity in synthetic data generation and significant improvements in downstream forecasting tasks.
In parallel with these developments, transformer-based architectures have attracted significant attention in financial time-series forecasting and generation. Unlike traditional RNNs and LSTMs, transformers leverage self-attention to capture long-term dependencies and complex patterns, and they have been shown to outperform conventional deep learning approaches for stock prediction tasks [22,23,24]. These advances highlight the growing importance of and attention-based models in financial and commodity time-series modeling.
However, there are several problems in applying machine learning models, including GANs, diffusion models, and transformer-based architectures, to financial time-series data. First, the numerous parameters in these models make them “black boxes,” making it difficult to explain the causal relationship between inputs and outputs. Second, there is a significant risk of overfitting to the training data, where even subtle differences between the input and training data can lead to substantially different outputs. In addition, existing machine learning approaches, including RNNs, LSTMs, and even transformer-based models, face fundamental limitations in financial contexts: traditional linear and parametric models struggle to capture the nonlinear and non-stationary dynamics of financial time series, while deep learning models often face difficulty modeling long-term dependencies and effectively selecting relevant features from high-dimensional or noisy inputs. This is particularly true due to the lack of explicit feature-level attention mechanisms, which can be crucial for accurate multivariate time-series forecasting [25]. Recently, it has been reported that explainable machine learning models, such as LightGBM, can achieve high prediction accuracy in financial time-series tasks by leveraging their feature importance and transparency [26]. Linear response theory (LRT), a physics-based method grounded in the framework of nonequilibrium statistical mechanics, offers a potential solution to these problems [27,28]. In LRT, the input is treated as an external force (i.e., a cause), and the output is the generated data (i.e., the result); this framework enables the system to produce robust responses to external force, as the coupling of these causal relationships is interpreted as a second-order fluctuation in the system.
In a related study applying LRT to finance, Ref. [29] analyzed NASDAQ data from 862 stocks and identified log-return as the most appropriate conjugate variable, demonstrating strong agreement between theoretical predictions and empirical average responses. Their study also examined the semi-strong efficiency of various financial markets, including Forex (EUR/USD, USD/HKD), Bitcoin, commodities (oil and gold), and the S&P500. Despite being a physical model, LRT accurately captures average price reactions to market events, providing valuable insights into market dynamics [30].
Previously, we proposed a linear response model (LRM) [31] to generate time-series data based on LRT. The LRM is derived under the assumption that the distribution of the time-series data follows an energy-based model (EBM) [32], with the energy function approximated up to the second-order around the mean. A key advantage of the LRM is that it avoids overfitting because model training is not required, and it avoids the black box issue because the causal relationships are explicitly known. In addition, the LRM offers several other benefits, e.g., very low computational costs. The LRM can be applied to financial time-series to generate data with fluctuations of various magnitudes; however, it is difficult to reproduce dynamics involving typically large fluctuations, e.g., those caused by COVID-19. While the LRM is primarily designed as a data generation method, there has been increasing attention in recent years to the detection and modeling of anomalous dynamics, which are often associated with extreme events. In this context, machine learning models such as normalizing flows have attracted considerable interest, and a variety of approaches have been explored to better capture rare anomalies and extreme dynamics in time-series data [33].
In order to overcome this issue, we propose the multilayer LRM (MLRM) to generate anonymous dynamics. The proposed MLRM is constructed by extending the LRM to multiple layers, which means that the MLRM incorporates the output data of the LRM iteratively. As a result, the fluctuations in the generated time-series data are amplified, thereby producing time-series data with anomalous dynamics. However, the data generated via the proposed MLRM may diverge. To solve this problem, it is necessary to consider methods to suppress divergence. Note that this approach is analogous to renormalization in quantum field theory [34], which is employed to avoid divergence in self-energy calculations.
By applying the proposed MLRM to financial time-series data, we attempted to determine whether large fluctuation dynamics can be generated using the information of fluctuations prior to COVID-19. Furthermore, we calculated the return distribution and the distribution of realized volatility from the data generated via the proposed MLRM. The return distribution exhibited pronounced fat-tailed behavior, with the estimated tail index being significantly smaller than 3 and close to 2, thereby deviating from the so-called inverse cubic law [35,36], and resembling the inverse square law observed during the COVID-19 pandemic and other extreme market periods [37,38]. The realized volatility distribution was well described by an inverse gamma distribution [39,40].
Praetz [41] demonstrated that Student’s t distribution provides a more accurate model for financial returns than the normal distribution, as it can accommodate the heavy tails observed in empirical data. Separately, the nonextensive statistical mechanics framework, originally formulated by Tsallis and his colleagues, offers a theoretical foundation for understanding the emergence of power-law distributions in complex systems, including financial markets [42]. In particular, this approach explains the frequent observation of fat-tailed return distributions and the breakdown of conventional Boltzmann–Gibbs statistics in empirical financial data. Our results, which show extremely heavy tails in the return distributions generated via the MLRM, are consistent with the predictions of nonextensive statistics and further support the relevance of this framework for modeling extreme market events, such as those observed during the COVID-19 pandemic.

2. Theory

First, in Section 2.1, we explain the theoretical framework of our previously proposed LRM [31] as a generative model for time-series data under external forces. Specifically, we begin by introducing the probability distribution of time-series data using an EBM, and then we incorporate external forces into the energy function to define the probability distribution of time-series data subject to such external forces. By approximating the energy function up to the second order, we demonstrate that time-series generation under external forces can be characterized by the product of the second-order fluctuation (evaluated in the absence of external forces) and the applied external force. Subsequently, in Section 2.2, we extend the LRM described in Section 2.1 to a multilayer framework. In this extension, the external force is treated as a random variable, enabling the generation of time-series data under various external forces. Furthermore, we assume that the fluctuations in data generated under various external forces themselves follow the LRM, and by iteratively applying this procedure, we derive the MLRM. Finally, since the derived MLRM may exhibit divergence under certain conditions, we also discuss methods to suppress such divergence and ensure the stability of the model.

2.1. LRM

Here, we review the theory of a single-layer LRM (SLRM) or a simple LRM for the generation of time-series data [31]. Let T : = { 1 , 2 , 3 , , N } be the time domain for which we observe the real-valued M-dimensional time-series. Then, for each t T , the observing values
x t : = ( x t , 1 , x t , 2 , , x t , M ) T R M
are considered to be the realization of the random variables X t : Ω R M , where Ω is the underlying set of a filtered probability space ( Ω , F , { F t } t T , P ) .
Assume that, for the random variable
X : Ω T × R M R N × M ,
the following function exists:
E X : R N × M R + .
This is referred to as an energy function, where the probability density function, f X , of X is expressed as follows:
ρ X ( x ) : = e E X ( x ) Z 0 .
Here, x : = ( x 1 , x 2 , , x N ) T R N × M , and we have the following:
Z 0 : = R N × M d x e E X ( x ) .
In other words, our description is based on the EBM introduced by [32].
In addition, the notation 0 represents the ensemble average, i.e., for a function, A i : R N × M R , we obtain the following for i = 1 , , N × M :
A i 0 : = R N × M d x A i ( x ) ρ X ( x ) .
Next, we consider the Taylor expansion of (4) around a mean value:
μ t : = x t 0 .
As a result, if we assume that E X can be approximated up to the second-order term of the displacement random vector Δ x t : = x t μ t , we obtain
E X ( x ) E X ( μ ) + t = 1 N g t · Δ x t + 1 2 s , t = 1 N Δ x s T H s t Δ x t ,
where
g t : = E X x t x t = μ t R M = E X x t , 1 , , E X x t , M T x t = μ t
is the gradient vector, and
H s t : = 2 E X x s x t x s = μ s , x t = μ t R M × M = 2 E X x s , 1 x t , 1 2 E X x s , M x t , 1 2 E X x s , 1 x t , M 2 E X x s , M x t , M x s = μ s , x t = μ t
is the Hessian matrix. Under the assumption that the gradient is zero ( g t = 0 ) at the mean value, (8) becomes the following:
E X ( x ) E X ( μ ) + 1 2 s , t = 1 N Δ x s T H s t Δ x t .
In addition, the first term, E X ( μ ) , in (10) is a constant. Thus, the numerator and denominator in (4) cancel each other out, and finally, with the substitution of (10) with (4), (4) is rewritten as follows:
ρ X ( x ) = 1 Z 0 e 1 2 s , t = 1 N Δ x s T H s t Δ x t ,
where
Z 0 : = R N × M d x e 1 2 s , t = 1 N Δ x s T H s t Δ x t .
Note that (11) takes the form of a multidimensional Gaussian distribution; however, it is dependent on both s and t. In addition, the Hessian matrix is equivalent to the following inverse matrix (refer to Section 1.4 in the literature [27]):
H s t = C s t 1 ,
where
C s t : = Δ x s Δ x t T 0 R M × M .
Next, assume that an external force vector f t R M acts on the system, and the external force energy term t = 1 N Δ x t · f t is added to the original energy function E X as follows:
E X ( x ) t = 1 N Δ x t · f t .
Thus, the probability distribution under external forces is expressed as follows:
ρ X f ( x ) : = 1 Z f e 1 2 s , t = 1 N Δ x s T C s t 1 Δ x t + t = 1 N Δ x t · f t ,
where
Z f : = R N × M d x e 1 2 s , t = 1 N Δ x s T C s t 1 Δ x t + t = 1 N Δ x t · f t .
To obtain the response Δ x t f under external forces, it is sufficient to take the partial derivative of the logarithm of (17) with respect to f t , as follows:
Δ x t f = f t ln Z f = 1 Z f f t Z f = 1 Z f R N × M d x Δ x t e 1 2 s , t = 1 N Δ x s T C s t 1 Δ x t + t = 1 N Δ x t · f t = R N × M d x ρ X f ( x ) Δ x t .
The notation f represents the ensemble average under the external force for i = 1 , , N × M :
A i f : = R N × M d x A i ( x ) ρ X f ( x ) .
To calculate (18), the variable transformation Δ x t : = Δ x t k = 1 N C t k f k for (17) makes it possible to obtain the following:
Z f = R N × M d x e 1 2 s , t = 1 N Δ x s T + k = 1 N f k T C k s C s t 1 Δ x t + k = 1 N C t k f k T + t = 1 N Δ x t + s = 1 N f s T C s t · f t = R N × M d x e 1 2 s , t = 1 N Δ x s T C s t 1 Δ x t + 2 t = 1 N Δ x t · f t + s , t = 1 N f s T C s t f t + t = 1 N Δ x t + s = 1 N f s T C s t · f t = R N × M d x e 1 2 s , t = 1 N Δ x s T C s t 1 Δ x t + 1 2 s , t = 1 N f s T C s t f t = Z 0 e 1 2 s , t = 1 N f s T C s t f t .
In this derivation, we have used the relation C s t = C t s and s = 1 N C k s C s t 1 = I k t R M × M , where I k t is the M × M identity matrix for k = t , and the zero matrix otherwise.
By substituting (20) with (18), we obtain the following:
Δ x t f = x t f μ t = s = 1 N C t s f s .
(21) is the equation of the LRM for time-series data and discrete time. If the time moves continuously in the interval [ t 0 , t 1 ] , then (21) can be rewritten as follows:
Δ x ( t ) f = t 0 t 1 C ( t , s ) f ( s ) d s ,
where Δ x ( t ) f , C ( t , s ) , and f ( s ) are regarded as Δ x t f , C t , s , and f s functions of continuous time, respectively. (22) takes the form of the green function in physics. We refer to C s t or C ( s , t ) as a two-time covariance matrix or function. In addition, (22) can be interpreted as the function f that minimizes the following variational function:
Φ [ f ] : = t 0 t 1 d t Δ x ( t ) f · f ( t ) 1 2 t 0 t 1 d t t 0 t 1 d s f ( t ) T C ( t , s ) f ( s ) .
With a return to the discrete time version, (21) can be rewritten as follows:
Δ x f = x f μ = C f ,
where
x f : = x 1 f x 2 f x N f , μ : = μ 1 μ 2 μ N , f : = f 1 f 2 f N , C : = C 11 C 12 C 1 N C 21 C 22 C 2 N C N 1 C N 2 C N N .
Here, Δ x f R L , C R L × L , and f R L for L : = N × M are the vector of Δ x t f , the matrix of the expanded versions in the time direction of C s t , and the vector of f t , respectively. Figure 1 shows a schematic representation of the LRM. Here, the input is treated as an external force (i.e., a cause), and the output is the generated data (i.e., the result); furthermore, the coupling of these causal relationships is considered a second-order fluctuation of the system.

2.2. MLRM

In the following section, we extend the conventional LRM to multiple layer in order to construct the proposed MLRM.
First, let the external force f be the observing values of a random variable, F : Ω R L . The probability density function of F is represented as ρ F ( f ) . Then, Δ x f becomes a conditional expectation of Δ x , given F = f . Here, Δ x f , n denotes the data generated via the n-layer LRM. Then, (24) of the SLRM can be expressed as follows:
Δ x f , 1 : = x f , 1 x f , 0 = C 0 f ,
where x f , 0 : = μ , C 0 : = C .
The two-time covariance matrix C 1 of the data generated via the SLRM is given as follows:
C 1 : = Δ x f , 1 Δ x f , 1 T = C 0 f f T C 0 T = C 0 Σ C 0 ,
where the relationship C 0 T = C 0 was used, and
Σ : = f f T : = R N × M d f ( f f T ) ρ F ( f ) .
Under the assumption that the data generated via the first layer of the LRM follow the second layer of the LRM, the displacement of the second layer of the LRM at that time is denoted as Δ x f , 2 , which is expressed as follows:
Δ x f , 2 : = x f , 2 x f , 1 = C 1 f .
(28) represents the data generation of the two-layer LRM. The equation for the two-layer LRM is expressed as follows:
x f , 2 = x f , 1 + C 1 f = x f , 0 + C 0 f + C 0 Σ C 0 f = μ + C f + C Σ C f .
By repeating these operations n times, we can derive the data generation equation for an n-layer LRM (n-LRM). Here, the following series is obtained:
x f , n = μ + C 0 f + C 1 f + + C n 1 f = μ + C 0 + C 1 + + C n 1 f = μ + p = 0 n 1 C p f ,
where
C p : = C p 1 Σ C p 1
for p = 1 , , n 1 .
By substituting (31) with (30), we obtain the following:
x f , n = μ + Σ 1 B + B 2 + B 4 + + B 2 n 1 f = μ + C ( n ) f ,
where
C ( n ) : = p = 0 n 1 C p = Σ 1 B ( n ) ,
B ( n ) : = B + B 2 + B 4 + + B 2 n 1 ,
B : = Σ C .
Note that B is generally an asymmetric matrix.
In the following, we consider a case when the number of layers proceeds to infinity. Here, we refer to the resulting model as the -LRM. In other words, if n in (32) is set to infinity, we obtain the following:
x f , = μ + Σ 1 B + B 2 + B 4 + f = μ + C eff f ,
where
C eff : = C ( ) = Σ 1 B eff ,
B eff : = B ( ) = B + B 2 + B 4 + .
Here, C eff is the two-time covariance matrix when (32) converges at infinity and becomes the finite values.
The eigenvalue problem for the asymmetric matrix B is solved as follows:
B V R = V R Λ ¯ ,
V L T B = Λ ¯ V L T ,
V L V R T = V R T V L = I .
Here, V R : = ( v 1 R , , v L R ) R L × L is a matrix of a column of right eigenvectors v i R for i = 1 , , L , V L R L × L is a matrix of a column of left eigenvectors v i L for i = 1 , , L , and Λ ¯ R L × L is a diagonal matrix whose ith diagonal element λ ¯ i ( i = 1 , , L ) is an eigenvalue corresponding to the ith column vector v i L of V L and v i R of V R . With the results of the eigenvalue problem in use, (34) can be decomposed by the eigenvectors as follows:
B ( n ) = V R Λ ¯ V L T + V R Λ ¯ 2 V L T + + V R Λ ¯ 2 n 1 V L T = V R H ( Λ ¯ , n ) V L T ,
where
H ( a , n ) : = k = 0 n 1 a 2 k = a + a 2 + a 4 + + a 2 n 1 .
Note that, when n , H ( a , ) = a + a 2 + a 4 + is one of the types called a Lacunary function ([43]). In addition, (42) is rewritten as follows:
B ( n ) = i = 1 L v i L H ( λ ¯ i , n ) ( v i R ) T .
When x 1 , (43) diverges because n is increased. In other words, to prevent x f , n from diverging, the eigenvalues of B must satisfy λ ¯ i < 1 for i = 1 , , L . Generally, (30) or (32) diverges because some eigenvalues with λ ¯ i 1 exist. To control this divergence, we introduce the hyperparameter ϵ 0 and modify (30) as follows:
x ϵ f , n = μ + p = 0 n 1 e ϵ p C p f .
When ϵ = 0 , (45) becomes (30); thus, x 0 f , n = x f , n . A well-chosen ϵ > 0 can prevent (45) from diverging. In addition, we define the following:
H ϵ ( a , n ) : = n = 0 e ϵ n a 2 n = a + e ϵ a 2 + e 2 ϵ a 4 + e 3 ϵ a 8 + .
Thus, the modified n-LRM is expressed as follows:
x ϵ f , n : = μ + C ϵ ( n ) f ,
C ϵ ( n ) : = p = 0 n 1 e ϵ p C p = Σ 1 B ϵ ( n ) ,
B ϵ ( n ) : = i = 1 L v i L H ϵ ( λ ¯ i , n ) ( v i R ) T .
In this study, (47) is used to generate the data in the proposed MLRM. Figure 2 shows a schematic representation of the proposed MLRM. Here, the proposed MLRM is constructed by extending the LRM to multiple layers, which means that the MLRM incorporates the output data of the LRM iteratively.

3. Method

Section 3.1 describes the principal component analysis (PCA) used for estimating external forces, introducing a dimensionality reduction technique. Section 3.2 details the estimation of two time covariance matrices, which are used for PCA and for data generation in the proposed MLRM. Section 3.3 outlines the method used for estimating external force vectors. Section 3.4 presents the approach that uses the proposed MLRM for data generation. Section 3.5 introduces dynamic time warping (DTW), which is employed to evaluate the data generated by the proposed MLRM.

3.1. PCA of Two-Time Covariance Matrix

PCA is employed in Section 3.3 as part of the procedure for estimating external forces. In this section, we consider a real–symmetric eigenvalue problem of C with an orthonormal condition that is similar to those found in PCA.
C V = V Λ ,
V V T = V T V = I .
Here, we obtain V , Λ , and I , which are the eigenvector, eigenvalue, and unit matrices, respectively. Note that Λ is a diagonal matrix whose ith diagonal element λ i ( i = 1 , , L ) is the variance of the ith PC, and the ith column vector v i of V is the corresponding eigenvector.
We consider approximating C using only a few top components with the contribution ratio of the eigenvalues. Typically, λ i is sorted in descending order such that the first PC shows the greatest variance λ 1 , and the corresponding column vector v 1 of V = ( v 1 , v L ) indicates the eigenvector of λ 1 .

3.2. Estimation of Two-Time Covariance Matrix

Here, let x d : = ( x 1 d , x 2 d , , x N d ) T R L for d = 1 , 2 , , D be the vector of M-dimensional stock price vectors in the intraday on the date indexed by d, where d = 1 and d = D represent the start and end dates of the interval used in the estimation period, respectively. In addition, let k be a positive integer, and consider the interval starting from the k days prior to the current date, d through d. Then, the two-time covariance matrix C d : k , estimated by the sample data from d k + 1 through d, is defined as follows:
C d : k : = 1 k i = d k + 1 d Δ x i ( Δ x i ) T ,
Δ x i : = x i μ d : k ,
μ d : k : = 1 k i = d k + 1 d x i .

3.3. Estimation of External Force Vector

Here, we consider f d and Δ x d f for d = 1 , 2 , , D to be the external force vector and the observed displacement vector in the intraday on day d, respectively. Using the observed Δ x d f and the estimated C d 1 , k , we can estimate f d by solving the following system of linear equations using the LRM:
Δ x d f = C d 1 : k f d .
However, the number L = N × M is large; thus, C d 1 : k may generally fall in an ill-conditioned matrix. Therefore, it is difficult to solve (55) directly. To address this problem, we apply a dimension reduction technique to (55) via PCA (refer to Section 3.1). Let Λ d 1 and V d 1 be the eigenvalue and eigenvector matrix of C d 1 : k , respectively. In this case, we fix the parameter λ c d 1 to cut off the eigenvalue close to zero. Let L c d 1 be a number of elements in the set { λ i d 1 | λ i d 1 > λ c d 1 , i = 1 , , L } , i.e., λ 1 d 1 > > λ L c d 1 d 1 > λ c d 1 . We consider the projection with V ˜ d 1 : = ( v 1 d 1 , , v L c d 1 d 1 ) as follows:
Δ x ˜ d f : = ( V ˜ d 1 ) T Δ x d f R L c d .
Then, (55) for this projection space is expressed as follows:
Δ x ˜ d f = C ˜ d 1 : k f ˜ d ,
where
C ˜ d 1 : k : = ( V ˜ d 1 ) T C d 1 : k V ˜ d 1 R L c d × L c d ,
f ˜ d : = ( V ˜ d 1 ) T f d R L c d .
By setting λ c d 1 successfully and removing the components with small eigenvalues, C ˜ c d 1 : k is well conditioned.
Finally, to convert f ˜ d to f d approximately in the original dimension L, the following equation is applied for d = 1 , , D :
f d V ˜ d f ˜ d R L .
We then use the force vectors estimated using (60). In addition, we estimate Σ = f f T as follows:
Σ : = 1 D d = 1 D f d f d T R L × L .

3.4. Data Generation by MLRM

Here, we explain the data generation process using the proposed MLRM. Let d { 1 , , D } be a fixed day, and let f d for d = 1 , , D be the approximated external force vectors given by (60). Then, we obtain the following data generation equation using the MLRM:
x ϵ , d f , n : = μ d : k + C ϵ , d : k ( n ) f d ,
C ϵ , d : k ( n ) : = p = 0 n 1 e ϵ p C p d : k ,
where
C 0 d : k : = C d : k ,
C p d : k : = C p 1 d : k Σ C p 1 d : k ,
for p = 1 , , n 1 , and ϵ 0 is selected to prevent (62) from diverging. Note that C d : k and Σ are calculated using (52) and (61), respectively.

3.5. DTW

In this study, we use DTW [44] to compare the generated and real time-series data. DTW represents the distance between the generated time-series data x : = { x 1 , , x L } T R L and the real time-series data y : = { y 1 , , y L } T R L . Note that a small DTW value means that the dynamics of the two sets of time-series data are similar. In addition, the cost function associated with a warping path p : = ( ( p 1 , p 1 ) , , ( p L , p L ) ) is obtained by computing with respect to all pairwise distances { d ( x i , y j ) | i , j = 1 , , L } R L × L :
c p ( x , y ) : = i = 1 L c ( X p i , Y p i ) .
The warping path p that incurs the minimum cost associated with the alignment is referred to as the optimal warping path. DTW is defined as the cost function with the optimal warping path as follows:
D T W ( x , y ) : = c p ( x , y ) : = min p c p ( x , y ) .

4. Experimental Results

4.1. Experimental Data

To evaluate the proposed model, we selected the Tokyo Stock Exchange (TSE) as the target market and used time-series data for stock prices extracted from the FLEX Full data which is a dataset that Japan Exchange Group (JPX) constructs with the real-time stock market data from the TSE on a daily basis and provides it as historical information. The Tokyo Stock Price Index (TOPIX) is a capitalization-weighted index of all companies listed in the First Section of the TSE. The black line in Figure 3 shows the time evolution in the TOPIX from March 2019 to March 2020. The COVID-19 shock caused the TOPIX to decline sharply in both February 2020 and March 2020. Here, we divided this period into two smaller periods without and with the COVID-19 shock, respectively, and we refer to these periods as the in-sample and out-of-sample periods, respectively. The latter period is shown as a gray region in Figure 3. As target stocks, we selected M = 3 stocks, i.e., Takeda Pharmaceutical Co., Ltd. (4502), Sony Corporation (6758), and Toyota Motor Corporation (7203) from the group of stocks listed on the TSE between March 2019 and March 2020 and included in the TOPIX CORE 30, which is the stock price index comprising 30 stocks with considerably high market capitalization and liquidity among all stocks on the TSE. In addition, we selected 09:00 to 11:30 (morning session) and 12:30 to 15:00 (afternoon session) as the time zones and used snapshots of the stock price data every 5 min. This resulted in time-series data for a total of 62 time points (9:00 to 11:30 ( representing 31-time points) and 12:30 to 15:00 (representing 31-time points) per day. In addition, all stock prices are preprocessed to make their averages zero in the time direction, which are denoted x t , m for t = 1 , , 62 and m = 1 , , 3 . Thus, t = 1 62 x t d = 0 R 3 and the dimension of x d for d = 1 , , D was 3 × 62 = 186 , and the index of this dimension was referred to as a Time-Stock Index. The red and blue lines in Figure 4 show the time-series data for the three stock prices in the in-sample and out-of-sample periods, respectively.

4.2. Estimation of External Force Vectors

In this evaluation, we set D = 731 and k = 200 , and we estimated { C d 1 : k } d = 1 , D using (52). In addition, we determined the number D of the eigenvalues { L c d 1 } d = 1 , , D , such that the sum of eigenvalue ratios for { C d 1 : k } d = 1 , D just exceeds 0.95 . This value of D was employed to remove eigenvalues that are close to zero. Using the number of reduced eigenvalues { L c d } d = 1 , , D , we calculated the reduced matrix { C ˜ d 1 : k } d = 1 , , D with the projection vectors { V ˜ d 1 : k } d = 1 , , D (refer to (58)).
The set x d f d = 1 , , D is the real stock price data, which we considered to be under the influence of external forces. We calculated Δ x d f d = 1 , , D by subtracting the expected values of x d f d = 1 , , D . We then obtained Δ x ˜ d f d = 1 , , D using the projection vectors { V ˜ d 1 : k } d = 1 , , D (refer to (56)). Next, we estimated { f ˜ d } d = 1 , , D using (57).
Note that { f ˜ d } d = 1 , , D are vectors in a lower-dimensional space; thus, we approximated { f d } d = 1 , , D by transforming them back to the original dimension using (60). From these transformations, we obtained various patterns of external forces based on the historical data. The results for Takeda Pharmaceutical Co., Ltd. (Tokyo, Japan) (4502), Sony Corporation (Tokyo, Japan) (6758), and Toyota Motor Corporation (Toyota, Japan) (7203) are shown in Figure 5a–c, respectively. Using the calculated external forces, we computed Σ using (61), and the results are presented in Figure 6a. As can be seen, the matrix Σ is symmetric, and it exhibits a strong positive correlation near the diagonal.

4.3. Estimation of C and B

We selected the day d = D , i.e., 30 December 2019, which is the last date in the in-sample period. The corresponding two-time covariance matrix C d : k is shown in Figure 6b, which shows the existence of high values near the diagonal and a negative correlation between the morning and afternoon sessions.
Figure 6c shows the matrix B = Σ C d : k estimated by (35), which is an asymmetric matrix. Here, we obtained the eigenvalues by solving the eigenvalue problem for B using (39) to (41). Figure 6d shows the eigenvalues of B from 1 to 25. As can be seen, the eigenvalues of B include values that are greater than 1.0, which means that the MLRM diverges as the number of layers increases. To address this divergence problem, data generation using the MLRM was performed by introducing the parameter ϵ to realize the sufficient suppression of divergence.
Next, we examined the convergence speed, depending on ϵ , using C ϵ , d : k ( n ) in (63). Here, we evaluated how the norm of C ϵ , d : k ( n ) changed as we increased the number of layers, with ϵ values ranging from 0.8 to 100.0. Figure 6e shows the results of the norm values of C ϵ , d : k ( n ) . The results confirm that, as the number of layers increases, a large ϵ value typically approaches the SLRM, and a small ϵ value diverges. In this study, we selected ϵ = 1 among candidates for which the convergence of the C ϵ , d : k ( n ) norm was confirmed. In addition, we confirmed that 10 layers are sufficient for convergence; thus, the number of layers for the MLRM was set to n = 10 , thereby resulting in a 10-layer LRM (10-LRM).

4.4. Data Generation Using MLRM

In this evaluation, we set ϵ = 1.0 and n = 10 , and the data generation for each stock by the proposed MLRM was performed using (62). Figure 7 shows the data generation results for Takeda Pharmaceutical Co., Ltd., Sony Corporation, and Toyota Motor Corporation, respectively. In Figure 7, the upper (a,b,c) and lower (d,e,f) graphs show the size of the data generated via the SLRM and MLRM (10-LRM) (blue lines), respectively, and the red lines represent the real out-of-sample data, which include samples with very large fluctuations due to the COVID-19 pandemic. A comparison of the SLRM and MLRM methods shows that the MLRM generates time-series with larger fluctuations than the SLRM. The time-series data generated via the proposed MLRM indicate that the out-of-sample fluctuations are covered. These results suggest that the data generation results obtained using the proposed MLRM contain information about the dynamics that is similar to that of the out-of-sample data.
Figure 8 shows box plots for the percentile of the stock price’s standard deviation (SD) calculated using the in-sample data, out-of-sample data, SLRM, and proposed MLRM. As can be seen, the fluctuations in the out-of-sample data are much larger than those in the in-sample data, which only partially capture the fluctuations observed in the out-of-sample period. The data generated via the SLRM show an improvement compared to the in-sample data, except for Sony Corporation (6752); however, the fluctuations captured via the SLRM do not fully account for those observed in the out-of-sample data. In contrast, the proposed MLRM generated significantly larger fluctuations, thereby effectively covering the out-of-sample fluctuations. These results demonstrate that the data generated via the proposed MLRM are more similar to the out-of-sample data than to the other sample data.

4.5. Comparison of DTW

We employed DTW to evaluate the similarity between the out-of-sample stock price data and the generated time-series dynamics. We also compared the results with those obtained using the WGAN. The WGAN is a machine learning method that generates time-series data by learning the in-sample data distribution through adversarial training. These comparisons are illustrated in Figure 9. We define minDTW as the minimum DTW distance between the out-of-sample data and any of the generated data samples. The left axis of Figure 9 shows the minDTW value for each day.
Here, a smaller minDTW indicates higher similarity between the two time-series data. The right axis shows the SD of the stock price. Notably, there are areas that exhibit high SD, particularly between 10 March 2020, and 19 March 2020, which is likely due to market disruptions caused by the COVID-19 pandemic. The in-sample results exhibit very high minDTW during periods heavily influenced by COVID-19, which indicates that the in-sample data does not contain time-series data that are similar to those for the COVID-19 period, thereby contributing to the high minDTW. The SLRM obtained better results than the in-sample results because the extrapolation representation of the SLRM, which is a first-order approximation relative to external forces, works very effectively. However, the proposed MLRM realized the most significant improvement in the minDTW during the COVID-19 period, which suggests that the MLRM captured large fluctuation dynamics effectively. In contrast, the results obtained via the WGAN were similar to the in-sample results, with higher minDTW during the COVID-19 period. These findings indicate that it is difficult to generate time-series data that are similar to those during the COVID-19 period using machine learning models that are dependent on historical data.
Table 1, Table 2 and Table 3 show the minDTW for the top 10 highest-SD days for each stock in the out-of-sample data, sorted by the SD, where the bold numbers indicate the lowest values in each column. The results in these tables confirm that the MLRM is able to generate data closer to the out-of-sample data than other methods on high-SD days. As expected, the MLRM successfully generates data with large fluctuations.
Next, Table 4, Table 5 and Table 6 show the minDTW for the top 10 lowest-SD days for each stock in the out-of-sample data, also sorted by the SD. The results confirm that the SLRM is able to generate data closer to the out-of-sample data than other methods on low-SD days. This suggests that the SLRM is effective for generating data without large fluctuations.
In summary, the evaluation results demonstrate that the proposed MLRM is a more effective tool for capturing the dynamics of anomalous events, such as those we faced during the COVID-19 pandemic. Moreover, by utilizing data generated from both the SLRM and MLRM, we can obtain artificial data that reflects the dynamics of both small and large fluctuations.

4.6. Comparison of Real Dynamics and Data Generated by MLRM

Furthermore, we compared the stock price dynamics of the generated and real data. Here, we focused on 10 March, 13 March, and 17 March 2020, when stock prices experienced significant fluctuations due to the impact of the COVID-19 pandemic. Figure 10 compares the dynamics between the real and generated data for the in-sample, WGAN, SLRM, and proposed MLRM, where each was selected based on the minDTW.
On 10 March, 13 March, and 17 March 2020, the minDTW values of the MLRM were lower than those of the in-sample, WGAN, and SLRM in most cases from Table 1, Table 2 and Table 3. Specifically, as shown in Figure 10, the stock price dynamics for Sony Corporation (6758) on 10 March 2020, and Toyota Motor Corporation (7203) on 17 March 2020, generated via the proposed MLRM were highly similar to the real data. This indicates that the MLRM faces challenges in generating stock price dynamics under highly uncertain conditions. Therefore, we expect that the MLRM will be applicable to various problems in the financial domain, e.g., risk management and scenario testing.

4.7. Tail Behavior of Log-Return Distributions

Figure 11 presents the distributions of log returns for three representative stocks (4502, 6758, and 7203). The vertical axis denotes the probability density on a logarithmic scale, while the horizontal axis corresponds to the log return. The empirical distributions for the in-sample period, out-of-sample period, and a subset of high-volatility observations are compared with those generated by the SLRM and the MLRM. The realized volatility used in this analysis is computed on a daily basis from intraday log returns sampled at 5 min intervals. Specifically, for each trading day, d, the realized volatility, R V d , is defined as follows:
R V d = t = 1 T r d , t 2 ,
where r d , t denotes the t-th 5-min log return on day d, and T is the number of such intervals within a trading day. The high-volatility sample is constructed by selecting out-of-sample observations for which the realized volatility exceeds the upper bound of the interquartile range (IQR), thereby isolating periods characterized by elevated market volatility. It is important to note that the SLRM and MLRM generate stock price series as zero-mean intraday fluctuations, lacking absolute price level information. To ensure consistency with the empirical data, we restore the price level by adding the mean of the in-sample stock price series to the generated outputs before computing the log returns. This procedure enables a meaningful comparison of return distributions across empirical and model-generated data. The distribution generated via the MLRM exhibits heavier tails than that produced using the SLRM, with tail behavior that lies between the empirical distributions of the out-of-sample period and the high-volatility sample. This finding indicates that the MLRM is capable of more effectively capturing the non-stationary dynamics and intricate correlation structures that emerge in the market under high-volatility conditions. By contrast, the SLRM yields a distribution with comparatively thinner tails, consistent with return behavior observed during more typical market regimes.
Table 7 shows the estimated excess kurtosis ( κ ) of the log-return distributions for stocks 4502, 6758, and 7203. The excess kurtosis κ of a random variable, X, with mean μ and standard deviation σ , is defined as
κ : = E ( X μ ) 4 σ 4 3 .
Excess kurtosis characterizes the heaviness of the distribution tails relative to a normal distribution. A normal distribution has κ = 0 ; positive values of κ indicate heavier tails and a higher likelihood of extreme returns, while negative values indicate lighter tails. In the MLRM-generated data, increasing the number of model layers leads to higher kurtosis, enabling the model to capture extreme market fluctuations and the heavy-tailed behavior observed in real financial markets effectively.
The parameter γ , estimated from the log-return distributions and reported in Table 7, corresponds to the exponent of the power-law tail, characterizing the rate of decay in the distribution’s tails. Empirical studies of financial markets have shown that the tail behavior of log returns typically follows P ( | r | > x ) x γ , where the so-called cubic law suggests that γ is approximately 3 under ordinary market conditions—specifically in developed markets, characterized by high liquidity and institutional maturity. Values of γ lower than 3 indicate heavier tails, implying a higher frequency of extreme returns. In this study, we estimated the tail index using the Hill estimator with a truncation level of 7.5%, following the approach recommended in the literature for developed markets [36]. Our results show that γ is consistently below 3 across all stocks in both the out-of-sample data and the data generated via the MLRM, demonstrating that the model effectively captures the heavy-tailed nature observed in real financial markets.
Given that the normal distribution cannot adequately capture such heavy-tailed behavior, we follow the approach of Praetz [41], who demonstrated that heavy-tailed financial returns are well described by Student’s t (or q-Gaussian) distribution. Figure 12 presents the results of fitting Student’s t distributions to the log-return distributions of three representative stocks (4502, 6758, and 7203). The Student’s t distribution with ν degrees of freedom is defined by the probability density function
p ( x | ν ) : = Γ ν + 1 2 ν π Γ ν 2 1 + x 2 ν ν + 1 2 ,
where Γ ( · ) denotes the gamma function. This distribution is symmetric and bell-shaped like the normal distribution but exhibits heavier tails, particularly for small values of ν . The degree of freedom, ν , governs the tail heaviness: smaller values of ν indicate heavier tails and a higher frequency of extreme returns. The fits are shown for the in-sample period, the out-of-sample period, and for data generated via the SLRM and the MLRM, with the estimated degrees of freedom ν indicated in each panel. The results show that the estimated degrees of freedom ν remain consistently around 2 across all data sets, reflecting the presence of heavy-tailed behavior in both empirical and model-generated distributions. Both the SLRM and the MLRM successfully reproduce this characteristic, with the MLRM providing slightly better alignment with the empirical tail behavior in some cases. These findings suggest that the models effectively capture the high frequency of extreme returns observed in real financial markets.

4.8. Tail Behavior of Realized Volatility Distributions

Figure 13 presents the distributions of realized volatility for three representative stocks (4502, 6758, and 7203), along with their corresponding inverse gamma distribution fits. The inverse gamma distribution, commonly used to model heavy-tailed positive variables such as volatility, is defined by its probability density function:
p ( x α , β ) : = β α Γ ( α ) x α 1 exp β x , x > 0 ,
where α is the shape parameter and β is the scale parameter. For each stock, the distributions are shown for the in-sample period, out-of-sample period, and for data generated via the SLRM and the MLRM. The MLRM-generated distributions closely match the empirical distributions, particularly in the tails, across all three stocks. It should be noted that the out-of-sample distributions exhibit greater variability due to the limited number of data points available, which results in less stable fits to the inverse gamma distribution.
Furthermore, it is known that there exists a theoretical relationship between the power-law exponent α of the inverse gamma distribution fitted to the realized volatility and the degrees of freedom ν of the Student’s t distribution fitted to the log-return distribution, given by ν = 2 α [45]. This relationship arises from models that describe the log returns as a mixture distribution driven by stochastic volatility, implying consistency between the tail behaviors of returns and volatility. In our results, the estimated ratio ν / α is generally close to 2 (see Table 7), which is in good agreement with the theoretical expectation. This observation further supports the conclusion that the data generated via the MLRM successfully capture both the tail behavior of returns and their relationship to the underlying volatility structure observed in real financial markets.

5. Conclusions

This paper has proposed the MLRM to generate time-series data. The proposed MLRM focuses on generating data for extreme fluctuations and anomalous dynamics and is constructed by extending the conventional SLRM to multiple layers. In this study, we evaluated the effectiveness of the proposed model by generating stock price data for the COVID-19 period using pre-COVID-19 pandemic data. The results demonstrated that the stock price data generated via the proposed MLRM exhibited a higher degree of similarity to real data during the COVID-19 period compared to other methods. However, because the MLRM estimates external forces based on historical data, it may be challenging to apply the MLRM to phenomena, such as regime switching, where the external forces driving future stock prices differ substantially from those observed historically. Nonetheless, this limitation may be overcome in future research by devising methodologies that model external forces independently of historical data.
In addition to generating realistic price trajectories, we further evaluated the tail behavior of the log-return and realized volatility distributions. The MLRM-generated data exhibited heavier tails in the log-return distributions, closely matching the empirical behavior observed during COVID-19 periods. The estimated power-law exponent γ and excess kurtosis κ confirm that the MLRM successfully reproduces the heavy-tailed nature of financial returns, which is critical for modeling extreme market movements. Moreover, the degrees of freedom ν of the Student’s t distribution fitted to the generated returns remained consistently low (around 2), which is consistent with empirical data, further supporting the model’s ability to replicate extreme return behavior.
Similarly, for realized volatility, the MLRM accurately captured the heavy-tailed nature of the empirical distributions, as evidenced by the fit to the inverse gamma distribution. Importantly, the theoretical relationship ν = 2 α between the volatility and return tail parameters was approximately satisfied by the MLRM-generated data, suggesting that the model can coherently capture the joint dynamics of returns and volatility observed in real markets.
Although it is generally difficult for machine learning models to generate data that is not included in the historical dataset, the MLRM can pre-generate rare events with extreme fluctuations, such as the COVID-19 pandemic. We believe that the artificial data reflecting both small and large fluctuations generated by the SLRM and the MLRM can be applied to various domains. In particular, the ability of the MLRM to reproduce realistic tail behavior is especially valuable for applications in financial risk management, stress testing, and scenario generation involving extreme market events. The results of this study have several practical implications for financial modeling and risk management. The proposed MLRM provides a transparent and computationally efficient framework for generating realistic synthetic financial time-series data, which can be utilized for scenario analysis, stress testing, and the development of algorithmic trading strategies. Furthermore, because the MLRM is based on explicit causal relationships, it offers enhanced interpretability relative to black-box machine learning models, making it particularly suitable for applications where model transparency is required, such as regulatory compliance and model validation. The ability to generate data that replicates extreme market fluctuations also enables more robust evaluation of risk measures and financial products under rare but critical market conditions. We believe that further development of the MLRM and related methodologies will contribute to more robust and interpretable approaches for modeling extreme events in financial markets.

Author Contributions

All authors declare that they jointly contributed to this research article. Y.N. was responsible for the majority of the work, including conceptualization, methodology, software implementation, data analysis, and manuscript preparation. T.A. supervised the project and provided primary guidance, while T.T. contributed as a co-supervisor and provided secondary supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original FLEX Full dataset was purchased from the JPX and cannot be shared due to contractual restrictions. However, the synthetic data generated in this study using the MLRM, which was trained on the JPX dataset, is available from the corresponding author upon reasonable request.

Acknowledgments

This research was supported by the Research Center for Quantitative Finance, Tokyo Metropolitan University.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar] [CrossRef]
  2. Wiese, M.; Knobloch, R.; Korn, R.; Kretschmer, P. Quant GANs: Deep generation of financial time series. Quant. Financ. 2020, 20, 1419–1440. [Google Scholar] [CrossRef]
  3. Oord, A.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar] [CrossRef]
  4. Li, J.; Wang, X.; Lin, Y.; Sinha, A.; Wellman, M.P. Generating Realistic Stock Market Order Streams. arXiv 2020, arXiv:2006.04212. [Google Scholar] [CrossRef]
  5. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar] [CrossRef]
  6. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein GANs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 4–9 December 2017; NIPS’17. pp. 5769–5779. [Google Scholar]
  7. Villani, C. Topics in Optimal Transportation; American Mathematical Society: Providence, RI, USA, 2003. [Google Scholar] [CrossRef]
  8. Ni, H.; Szpruch, L.; Sabate-Vidales, M.; Xiao, B.; Wiese, M.; Liao, S. Sig-wasserstein GANs for time series generation. In Proceedings of the Second ACM International Conference on AI in Finance, New York, NY, USA, 3–5 November 2021. ICAIF ’21. [Google Scholar] [CrossRef]
  9. Chevyrev, B.I.; Lyons, T. Characteristic Functions of Measures on Geometric Rough Paths. Annu. Probab. 2016, 44, 4049–4082. [Google Scholar] [CrossRef]
  10. Morrill, J.; Fermanian, A.; Kidger, P.; Lyons, T. A Generalised Signature Method for Multivariate Time Series Feature Extraction. arXiv 2021, arXiv:2006.00873. [Google Scholar] [CrossRef]
  11. Liao, S.; Ni, H.; Sabate-Vidales, M.; Szpruch, L.; Wiese, M.; Xiao, B. Sig-Wasserstein GANs for conditional time series generation. Math. Financ. 2024, 34, 622–670. [Google Scholar] [CrossRef]
  12. Coletta, A.; Prata, M.; Conti, M.; Mercanti, E.; Bartolini, N.; Moulin, A.; Vyetrenko, S.; Balch, T. Towards realistic market simulations: A generative adversarial networks approach. In Proceedings of the Second ACM International Conference on AI in Finance, Virtual Event, 3–5 November 2021; Association for Computing Machinery: New York, NY, USA, 2022. ICAIF ’21. [Google Scholar] [CrossRef]
  13. Vuletić, M.; Prenzel, F.; Cucuringu, M. Fin-GAN: Forecasting and classifying financial time series via generative adversarial networks. Quant. Financ. 2024, 24, 175–199. [Google Scholar] [CrossRef]
  14. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  15. Tsay, R.S. Analysis of Financial Time Series, 3rd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  16. Sohl-Dickstein, J.; Weiss, E.A.; Maheswaranathan, N.; Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, JMLR.org, Lille, France, 6–11 July 2015; ICML’15. Volume 37, pp. 2256–2265. [Google Scholar]
  17. Li, Y.; Lu, X.; Wang, Y.; Dou, D. Generative time series forecasting with diffusion, denoise, and disentanglement. In Proceedings of the 36th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 28 November 2022–9 December 2022. NIPS ’22. [Google Scholar]
  18. Tashiro, Y.; Song, J.; Song, Y.; Ermon, S. CSDI: Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation. In Advances in Neural Information Processing Systems; Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2021; Volume 34, pp. 24804–24816. [Google Scholar]
  19. Alcaraz, J.M.L.; Strodthoff, N. Diffusion-Based Time Series Imputation and Forecasting with Structured State Space Models. arXiv 2022, arXiv:2208.09399. [Google Scholar] [CrossRef]
  20. Wang, Z.; Ventre, C. A Financial Time Series Denoiser Based on Diffusion Models. In Proceedings of the 5th ACM International Conference on AI in Finance, New York, NY, USA, 14–17 November 2024; ICAIF’24. pp. 72–80. [Google Scholar] [CrossRef]
  21. Huang, H.; Chen, M.; Qiao, X. Generative Learning for Financial Time Series with Irregular and Scale-Invariant Patterns. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  22. Zhang, Q.; Qin, C.; Zhang, Y.; Bao, F.; Zhang, C.; Liu, P. Transformer-based attention network for stock movement prediction. Expert Syst. Appl. 2022, 202, 117239. [Google Scholar] [CrossRef]
  23. Wang, C.; Chen, Y.; Zhang, S.; Zhang, Q. Stock market index prediction using deep Transformer model. Expert Syst. Appl. 2022, 208, 118128. [Google Scholar] [CrossRef]
  24. Mozaffari, L.; Zhang, J. Predictive Modeling of Stock Prices Using Transformer Model. In Proceedings of the 2024 9th International Conference on Machine Learning Technologies, Oslo, Norway, 24–26 May 2024; ICMLT’24. pp. 41–48. [Google Scholar] [CrossRef]
  25. Huang, S.C.; Wu, C.F.; Chen, K.C.; Lin, M.C.; Chiou, C.C. Intelligent dual-stage attention-based deep networks for energy market predictions. Alex. Eng. J. 2025, 122, 625–644. [Google Scholar] [CrossRef]
  26. Wu, C.F.; Lin, M.C.; Chao, T.W.; Chiou, C.C. Omnipresent AI and Big data for financial early warning: Integrating financial indicators and text sentiment analysis in Chinese real estate. Enterp. Inf. Syst. 2025, 19, 2434742. [Google Scholar] [CrossRef]
  27. Kubo, R.; Toda, M.; Hashitsume, N. Statistical Physics II: Nonequilibrium Statistical Mechanics, 2nd ed.; Solid-State Sciences; Springer: Berlin/Heidelberg, Germany, 1991. [Google Scholar]
  28. Zubarev, D.N. Nonequilibrium Statistical Thermodynamics; Springer: New York, NY, USA, 1973. [Google Scholar]
  29. Puertas, A.M.; Trinidad-Segovia, J.E.; Sánchez-Granero, M.A.; Clara-Rahora, J.; de las Nieves, F.J. Linear response theory in stock markets. Sci. Rep. 2021, 11, 23076. [Google Scholar] [CrossRef]
  30. Puertas, A.M.; Clara-Rahola, J.; Sánchez-Granero, M.A.; de las Nieves, F.J.; Trinidad-Segovia, J.E. A new look at financial markets efficiency from linear response theory. Financ. Res. Lett. 2023, 51, 103455. [Google Scholar] [CrossRef]
  31. Naritomi, Y.; Adachi, T. Time Series Data Generation by Linear Response Model. Trans. Jpn. Soc. Artif. Intell. 2024, 39, 1–9. [Google Scholar] [CrossRef]
  32. Teh, Y.W.; Osindero, S.; Hinton, G.E. Energy-Based Models for Sparse Overcomplete Representations. J. Mach. Learn. Res. 2003, 4, 1235–1260. [Google Scholar]
  33. Zhao, X.; Liu, P.; Mahmoudi, S.; Garg, S.; Kaddoum, G.; Hassan, M.M. DDANF: Deep denoising autoencoder normalizing flow for unsupervised multivariate time series anomaly detection. Alex. Eng. J. 2024, 108, 436–444. [Google Scholar] [CrossRef]
  34. Peskin, M.E.; Schroeder, D.V. An Introduction to Quantum Field Theory; Westview Press: Boulder, CO, USA, 1995. [Google Scholar]
  35. Gopikrishnan, P.; Meyer, M.; Amaral, L.A.N.; Stanley, H.E. Inverse cubic law for the distribution of stock price variations. Eur. Phys. J. B-Condens. Matter Complex Syst. 1998, 3, 139–140. [Google Scholar] [CrossRef]
  36. Gu, Z.; Ibragimov, R. The “Cubic Law of the Stock Returns” in emerging markets. J. Empir. Financ. 2018, 46, 182–190. [Google Scholar] [CrossRef]
  37. Drożdż, S.; Kwapień, J.; Osświęcimka, P.; Stanisz, T.; Waṫorek, M. Complexity in Economic and Social Systems: Cryptocurrency Market at around COVID-19. Entropy 2020, 22, 1043. [Google Scholar] [CrossRef]
  38. Takaishi, T. Recent scaling properties of Bitcoin price returns. J. Phys. Conf. Ser. 2021, 1730, 012124. [Google Scholar] [CrossRef]
  39. Takaishi, T. Analysis of Realized Volatility in Superstatistics. Evol. Institutional Econ. Rev. 2010, 7, 89–99. [Google Scholar] [CrossRef]
  40. Andersen, T.G.; Bollerslev, T.; Diebold, F.X.; Labys, P. The Distribution of Realized Exchange Rate Volatility. J. Am. Stat. Assoc. 2001, 96, 42–55. [Google Scholar] [CrossRef]
  41. Praetz, P.D. The Distribution of Share Price Changes. J. Bus. 1972, 45, 49–55. [Google Scholar] [CrossRef]
  42. Tsallis, C.; Anteneodo, C.; Borland, L.; Osorio, R. Nonextensive statistical mechanics and economics. Phys. A Stat. Mech. Its Appl. 2003, 324, 89–100. [Google Scholar] [CrossRef]
  43. Mandelbrojt, S. Lacunary series. Rice Inst. Pam.-Rice Univ. Stud. 1923, 14, 261–284. [Google Scholar]
  44. Myers, C.; Rabiner, L.; Rosenberg, A. Performance tradeoffs in dynamic time warping algorithms for isolated word recognition. IEEE Trans. Acoust. Speech Signal Process. 1980, 28, 623–635. [Google Scholar] [CrossRef]
  45. Gelman, A.; Carlin, J.B.; Stern, H.S.; Dunson, D.B.; Vehtari, A.; Rubin, D.B. Bayesian Data Analysis, 3rd ed.; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
Figure 1. Data generation using the LRM.
Figure 1. Data generation using the LRM.
Entropy 27 00823 g001
Figure 2. Data generation using the MLRM.
Figure 2. Data generation using the MLRM.
Entropy 27 00823 g002
Figure 3. TOPIX: the black line represents the time evolution of the TOPIX from March 2019 to March 2020, while the gray region denotes the out-of-sample period (January 2020 to March 2020).
Figure 3. TOPIX: the black line represents the time evolution of the TOPIX from March 2019 to March 2020, while the gray region denotes the out-of-sample period (January 2020 to March 2020).
Entropy 27 00823 g003
Figure 4. Time-series data of stock prices: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203). The red line represents the in-sample data; the blue line denotes the out-of-sample data.
Figure 4. Time-series data of stock prices: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203). The red line represents the in-sample data; the blue line denotes the out-of-sample data.
Entropy 27 00823 g004
Figure 5. External force vectors: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203) for 1 , , D .
Figure 5. External force vectors: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203) for 1 , , D .
Entropy 27 00823 g005
Figure 6. Matrices using the MLRM and its convergence: (a) Σ = f f T , (b) C d : k , (c) B d : k , (d) eigenvalues of B d : k and (e) norms of C d : k ( d = D and k = 200 ).
Figure 6. Matrices using the MLRM and its convergence: (a) Σ = f f T , (b) C d : k , (c) B d : k , (d) eigenvalues of B d : k and (e) norms of C d : k ( d = D and k = 200 ).
Entropy 27 00823 g006
Figure 7. Comparison of the stock price data generated via the SLRM and the MLRM. (ac) show the results of the SLRM, and (df) show the results of the MLRM for three representative stocks: (a,d) Takeda Pharmaceutical Co., Ltd. (4502); (b,e) Sony Corporation (6758); and (c,f) Toyota Motor Corporation (7203). The blue lines represent the generated stock price data via the SLRM and MLRM (10-layer MLRM), and the red lines represent the real out-of-sample data.
Figure 7. Comparison of the stock price data generated via the SLRM and the MLRM. (ac) show the results of the SLRM, and (df) show the results of the MLRM for three representative stocks: (a,d) Takeda Pharmaceutical Co., Ltd. (4502); (b,e) Sony Corporation (6758); and (c,f) Toyota Motor Corporation (7203). The blue lines represent the generated stock price data via the SLRM and MLRM (10-layer MLRM), and the red lines represent the real out-of-sample data.
Entropy 27 00823 g007
Figure 8. Box plots of the percentiles of the stock price’s SD: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203). The black horizontal lines represent the 0th, 25th, 75th, and 100th percentiles (from the bottom), and the orange line indicates the 50th percentile. The triangles represent the averages.
Figure 8. Box plots of the percentiles of the stock price’s SD: (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), and (c) Toyota Motor Corporation (7203). The black horizontal lines represent the 0th, 25th, 75th, and 100th percentiles (from the bottom), and the orange line indicates the 50th percentile. The triangles represent the averages.
Entropy 27 00823 g008
Figure 9. Comparison of daily minDTW for (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), (c) Toyota Motor Corporation (7203). The minDTW (left axis) is shown for the in-sample (blue line), SLRM (black dotted line), proposed MLRM (black line), and WGAN (gray line). The light gray areas represent the SD of the stock prices (right axis).
Figure 9. Comparison of daily minDTW for (a) Takeda Pharmaceutical Co., Ltd. (4502), (b) Sony Corporation (6758), (c) Toyota Motor Corporation (7203). The minDTW (left axis) is shown for the in-sample (blue line), SLRM (black dotted line), proposed MLRM (black line), and WGAN (gray line). The light gray areas represent the SD of the stock prices (right axis).
Entropy 27 00823 g009
Figure 10. Comparison of dynamics on 10 March 2020, 13 March 2020, and 17 March 2020. The time-series data for the out-of-sample (red line), in-sample (blue line), WGAN (grey line), SLRM (dotted black line), and MLRM (black line) are shown. Each was selected based on the minDTW.
Figure 10. Comparison of dynamics on 10 March 2020, 13 March 2020, and 17 March 2020. The time-series data for the out-of-sample (red line), in-sample (blue line), WGAN (grey line), SLRM (dotted black line), and MLRM (black line) are shown. Each was selected based on the minDTW.
Entropy 27 00823 g010
Figure 11. Distributions of log returns for three representative stocks: (a) 4502, (b) 6758, and (c) 7203. The vertical axis shows the probability density on a logarithmic scale; the horizontal axis indicates the log return. The empirical distributions are shown for the in-sample period (blue solid line), out-of-sample period (red solid line), and high-volatility period (green solid line). The distributions generated via the SLRM (black dotted line) and the MLRM (black solid line) are also displayed. The MLRM-generated distributions exhibit heavier tails and closely match the empirical distributions observed during high-volatility periods.
Figure 11. Distributions of log returns for three representative stocks: (a) 4502, (b) 6758, and (c) 7203. The vertical axis shows the probability density on a logarithmic scale; the horizontal axis indicates the log return. The empirical distributions are shown for the in-sample period (blue solid line), out-of-sample period (red solid line), and high-volatility period (green solid line). The distributions generated via the SLRM (black dotted line) and the MLRM (black solid line) are also displayed. The MLRM-generated distributions exhibit heavier tails and closely match the empirical distributions observed during high-volatility periods.
Entropy 27 00823 g011
Figure 12. Student’s t distribution fits for the log-return distributions of three representative stocks (4502, 6758, and 7203). For each stock, four types of data are shown: in-sample data (a,e,i), out-of-sample data (b,f,j), data generated via the SLRM (c,g,k), and data generated via the MLRM (d,h,l). Each distribution is fitted with a Student’s t distribution, and the estimated degrees of freedom ν are indicated in the legend. The probability density is shown on a logarithmic scale to highlight the heavy-tailed behavior. The results demonstrate that both the SLRM and the MLRM can reproduce the heavy-tailed nature of real financial returns, with the MLRM providing a better fit across all three stocks.
Figure 12. Student’s t distribution fits for the log-return distributions of three representative stocks (4502, 6758, and 7203). For each stock, four types of data are shown: in-sample data (a,e,i), out-of-sample data (b,f,j), data generated via the SLRM (c,g,k), and data generated via the MLRM (d,h,l). Each distribution is fitted with a Student’s t distribution, and the estimated degrees of freedom ν are indicated in the legend. The probability density is shown on a logarithmic scale to highlight the heavy-tailed behavior. The results demonstrate that both the SLRM and the MLRM can reproduce the heavy-tailed nature of real financial returns, with the MLRM providing a better fit across all three stocks.
Entropy 27 00823 g012
Figure 13. Distributions of realized volatility for three representative stocks (4502, 6758, and 7203), along with their corresponding inverse gamma distribution fits. For each stock, results are presented for the in-sample period (a,e,i), the out-of-sample period (b,f,j), and for data generated via the SLRM (c,g,k) and the MLRM (d,h,l). The estimated parameters of the inverse gamma distribution ( α , β ) are reported in each panel. These results demonstrate that, while the SLRM tends to underestimate the probability density in the tails, the MLRM more accurately captures the heavy-tailed characteristics of the empirical volatility distributions observed in real financial markets.
Figure 13. Distributions of realized volatility for three representative stocks (4502, 6758, and 7203), along with their corresponding inverse gamma distribution fits. For each stock, results are presented for the in-sample period (a,e,i), the out-of-sample period (b,f,j), and for data generated via the SLRM (c,g,k) and the MLRM (d,h,l). The estimated parameters of the inverse gamma distribution ( α , β ) are reported in each panel. These results demonstrate that, while the SLRM tends to underestimate the probability density in the tails, the MLRM more accurately captures the heavy-tailed characteristics of the empirical volatility distributions observed in real financial markets.
Entropy 27 00823 g013
Table 1. minDTW of the top 10 highest-SD days for 4502. Bold indicates the minimum value among the methods in each row.
Table 1. minDTW of the top 10 highest-SD days for 4502. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
113 Mar 202071.82070.31882.61457.91756.3
210 Mar 202065.3983.3930.0793.2914.1
312 Mar 202058.51014.0505.3472.4724.2
417 Mar 202050.71095.31131.1892.41032.5
525 Mar 202045.5506.8698.6488.3597.7
631 Mar 202044.7936.2742.0495.3577.2
728 Feb 202034.4461.9452.3504.1265.2
83 Mar 202033.0454.1366.9411.1465.6
918 Mar 202029.5657.4641.3505.0362.0
1027 Feb 202028.1426.1407.4357.1373.7
Table 2. minDTW of the top 10 highest-SD days for 6752. Bold indicates the minimum value among the methods in each row.
Table 2. minDTW of the top 10 highest-SD days for 6752. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
113 Mar 2020182.15306.82778.12479.94672.5
210 Mar 2020140.13463.71610.91440.93468.5
319 Mar 2020134.03449.51972.21596.63109.9
417 Mar 2020102.22425.01776.41894.92808.9
52 Mar 202081.01075.01009.3954.21243.7
69 Mar 202074.61134.3928.8935.21229.9
712 Mar 202071.61197.01251.11071.21242.7
818 Mar 202070.41823.51538.71249.51647.9
911 Mar 202066.91084.9680.6679.3892.4
105 Feb 202063.51076.0748.6722.51235.7
Table 3. minDTW of the top 10 highest-SD days for 7203. Bold indicates the minimum value among the methods in each row.
Table 3. minDTW of the top 10 highest-SD days for 7203. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
117 Mar 202035.71084.81086.8302.41049.7
213 Mar 202025.8711.7610.7511.5668.8
310 Mar 202019.9438.1455.1230.2471.0
431 Mar 202019.6506.9298.7147.4531.6
518 Mar 202018.8703.0519.9501.2611.0
611 Mar 202014.9343.9179.6163.8299.5
725 Mar 202014.7261.0253.2220.5275.8
83 Mar 202013.5279.9132.2142.0260.7
92 Mar 202013.5213.4203.4156.1225.9
1027 Mar 202012.7298.8257.9222.0294.6
Table 4. minDTW of the top 10 lowest-SD days for 4502. Bold indicates the minimum value among the methods in each row.
Table 4. minDTW of the top 10 lowest-SD days for 4502. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
127 Jan 20205.6158.8144.9145.3174.6
222 Jan 20205.8137.9123.2113.5166.1
323 Jan 20205.8141.0128.2136.9192.4
415 Jan 20206.4145.9137.7135.9198.1
510 Jan 20206.8150.6137.8164.2174.9
619 Feb 20206.9166.7147.2134.0180.4
720 Jan 20207.1182.2144.1163.2147.0
89 Jan 20207.7159.5121.8115.8134.7
928 Jan 20207.7176.5160.0178.2228.2
1030 Jan 20207.8206.9177.1209.0226.0
Table 5. minDTW of the top 10 lowest-SD days for 6552. Bold indicates the minimum value among the methods in each row.
Table 5. minDTW of the top 10 lowest-SD days for 6552. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
120 Jan 20207.1222.7172.7247.1308.4
223 Jan 202011.0255.2269.8319.3387.7
314 Feb 202012.0312.5285.1337.3383.3
428 Jan 202012.7301.5288.1301.6363.5
510 Jan 202015.8301.6344.6382.6404.3
610 Feb 202016.2336.5322.4459.9411.5
724 Jan 202019.1258.2268.5275.7321.2
821 Jan 202019.2293.1263.1366.8317.4
95 Mar 202019.6400.1358.3405.1383.9
1021 Feb 202019.7359.5382.5387.9377.7
Table 6. minDTW of the top 10 lowest-SD days for 7203. Bold indicates the minimum value among the methods in each row.
Table 6. minDTW of the top 10 lowest-SD days for 7203. Bold indicates the minimum value among the methods in each row.
NoDateSDIn-SampleSLRMMLRMWGAN
116 Jan 20200.933.824.429.634.9
217 Jan 20201.235.927.229.444.3
314 Jan 20201.739.143.148.745.0
45 Feb 20201.843.941.638.150.2
523 Jan 20201.838.831.530.640.0
69 Jan 20201.939.332.841.839.4
722 Jan 20201.946.446.749.649.2
813 Feb 20201.938.638.841.948.6
910 Jan 20202.142.524.329.142.9
1029 Jan 20202.135.835.241.643.6
Table 7. Parameter estimates for the in-sample data, the out-of-sample data, and data generated via the MLRM with varying numbers of layers.
Table 7. Parameter estimates for the in-sample data, the out-of-sample data, and data generated via the MLRM with varying numbers of layers.
CodeParameterIn-SampleOut-of-SampleLayer = 1Layer = 2Layer = 3Layer = 4Layer = 10
4502 κ 1.812.773.123.103.143.213.20
γ 2.482.152.042.052.042.052.05
ν 1.991.991.991.981.981.981.98
α 2.701.301.221.121.061.051.05
ν / α 0.741.531.631.771.871.891.89
6758 κ 1.742.673.193.523.633.653.65
γ 2.532.242.022.022.002.012.01
ν 2.541.991.991.991.981.981.98
α 2.631.121.221.020.920.900.90
ν / α 0.971.781.631.952.152.202.20
7203 κ 1.873.411.792.092.282.332.33
γ 2.462.182.422.312.252.232.23
ν 2.541.991.981.981.981.981.98
α 2.500.871.421.161.041.021.02
ν / α 1.022.291.401.711.901.951.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Naritomi, Y.; Takaishi, T.; Adachi, T. Modeling and Generating Extreme Fluctuations in Time Series with a Multilayer Linear Response Model. Entropy 2025, 27, 823. https://doi.org/10.3390/e27080823

AMA Style

Naritomi Y, Takaishi T, Adachi T. Modeling and Generating Extreme Fluctuations in Time Series with a Multilayer Linear Response Model. Entropy. 2025; 27(8):823. https://doi.org/10.3390/e27080823

Chicago/Turabian Style

Naritomi, Yusuke, Tetsuya Takaishi, and Takanori Adachi. 2025. "Modeling and Generating Extreme Fluctuations in Time Series with a Multilayer Linear Response Model" Entropy 27, no. 8: 823. https://doi.org/10.3390/e27080823

APA Style

Naritomi, Y., Takaishi, T., & Adachi, T. (2025). Modeling and Generating Extreme Fluctuations in Time Series with a Multilayer Linear Response Model. Entropy, 27(8), 823. https://doi.org/10.3390/e27080823

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop