1. Introduction
In modern financial risk management, accurate forecasting of asset return volatility is essential. Volatility serves as a fundamental risk measure that informs investment decisions and constitutes the primary input for risk assessment frameworks such as Value-at-Risk (VaR). As a widely adopted metric for market risk, VaR is extensively used by financial institutions for regulatory capital requirements and internal risk management [
1]. Consequently, producing precise and adaptive VaR forecasts is critical for effectively navigating the complexities and turbulence of financial markets.
Throughout the history of VaR estimation, various methods have been developed, with the parametric approach being one of the primary paradigms. This method evaluates risk by analyzing the return distribution of an asset over a look-back period and estimating its volatility and expected return. Within this framework, accurate modeling of volatility has become crucial, leading to the development of many sophisticated volatility models, among which the Generalized Autoregressive Conditional Heteroskedasticity (GARCH) family is the most prominent [
2]. Efforts to improve predictive accuracy have also explored the use of more complex probability density functions and the incorporation of time-varying higher-order conditional moments. While GARCH models are highly regarded for their statistical rigor and ability to capture ‘stylized facts’ such as volatility clustering, their foundation relies on strict statistical assumptions—for example, specific error distributions. This structural rigidity often results in suboptimal performance amid the profound non-linearity and dynamic complexity of financial markets, particularly during periods of extreme stress [
3].
To overcome the limitations of traditional models, both academia and industry have increasingly turned to deep learning (DL), which offers fewer modeling constraints and enhanced feature extraction capabilities [
4,
5]. Given the pervasive long-memory property in financial time series [
6], Recurrent Neural Networks (RNNs), capable of retaining historical information, are considered more appropriate for processing such data than Feed-forward Neural Networks (FNNs) [
7]. However, conventional RNNs encounter difficulties with vanishing or exploding gradients during the training of long sequences [
8]. The Long Short-Term Memory (LSTM) network, introduced by Hochreiter & Schmidhuber [
9], addresses these issues through its gating mechanisms, enabling the model to capture features over extended time horizons. Consequently, LSTMs are theoretically well-suited for volatility forecasting tasks [
5].
Building on this foundation, researchers have pursued various approaches to further enhance LSTM performance in financial forecasting. A prominent trend involves developing hybrid models that combine the statistical insights of GARCH models with the sequential learning strengths of LSTMs. Early studies, such as those by Kim & Won [
4] and Hu et al. [
10], demonstrated the effectiveness of this approach by feeding GARCH-derived predictions as external features into LSTM networks. More recent advancements have integrated these methodologies more deeply; for instance, Zhao et al. [
11] proposed re-engineering and embedding the mathematical structure of GARCH models directly within the LSTM cell, which not only improved the model’s interpretability in a financial context but also yielded superior predictive performance over traditional econometric models. Simultaneously, another avenue of research enhances model capabilities by incorporating external information, notably utilizing Natural Language Processing (NLP) techniques to extract market sentiment from textual data as an additional input [
12]. Since these hybrid and information-enhanced methods aim to improve volatility forecast accuracy, they have naturally been extended to the more complex task of VaR estimation, with promising results [
3].
Despite notable advances in fusion strategies, current state-of-the-art forecasting models still encounter two fundamental bottlenecks rooted in the intrinsic properties of financial data. The first pertains to input feature fidelity: financial time series are inherently highly non-stationary [
13], with their mean and variance evolving dynamically over time. Deeper still, this non-stationarity often manifests as intricate fractal structures, where the complex temporal correlations give rise to multifractal characteristics—beyond mere fat-tail distributions [
14]. Multifractal analysis, a powerful tool in nonlinear dynamics, can reveal the “bursty” and heterogeneous nature of market fluctuations at fine scales [
15]. However, a persistent technical challenge in standard multifractal analysis—the use of non-overlapping segmentation—can introduce spurious fluctuations that undermine the stability of fractal measurements [
16]. Nonetheless, its core metric, the multifractal spectrum width (
), remains an effective indicator of the degree of inhomogeneity and local complexity in market dynamics. Recent studies have demonstrated that incorporating such refined fractal features into machine learning models can substantially improve prediction accuracy [
17]. This underscores that, without a robust fractal perspective, even advanced features derived from GARCH models or sentiment analysis may only offer a partial picture, failing to fully capture the deep, high-fidelity structural state of the market.
Beyond this input-level challenge lies a more fundamental bottleneck: the responsiveness of the model architecture itself. Most existing deep learning models rely on static gating mechanisms and simple activation functions (e.g., ReLU or tanh), which reveal inherent limitations when processing highly dynamic and non-stationary financial signals [
8]. This mismatch stems from a disparity between the complexity of the signals—characterized by chaos, nonlinearity, and rapid changes—and the fixed logic embedded within such models. While the nature of financial chaos remains an active research area, its significance in describing nonlinear market behaviors is widely recognized [
18]. Static gates, for instance, cannot adaptively modify their memory management strategies in response to fluctuations’ intensity—which can be quantified by the multifractal spectrum width—leading to suboptimal information retention. Additionally, conventional activation functions like tanh are insufficient for capturing the rich, chaotic nonlinearities inherent in financial data, impeding the model’s ability to characterize intrinsic complexity and abrupt transitions. This mechanistic rigidity means that, even with perfect input features, the internal processing may fail to respond appropriately to market dynamics, resulting in information decay during transmission [
19]. Therefore, developing an adaptive architecture capable of addressing both feature fidelity and responsiveness—by dynamically adjusting to market complexity—is a key challenge for advancing financial forecasting models.
To systematically address the challenges outlined above, this paper introduces the Dynamic Fractal–Chaotic Long Short-Term Memory (DFC-LSTM) network. The central idea is to embed the intrinsic dynamical principles of financial markets—namely, fractal geometry and deterministic chaos—directly into the core computational units of the network, thereby simultaneously tackling both the responsiveness and fidelity bottlenecks. The proposed architecture incorporates two synergistic innovations. First, the traditional static forget gate is replaced by a multifractal-driven dynamic forget gate, which leverages the multifractal spectrum width () to adaptively regulate memory retention in real time, according to the evolving complexity of the market. Second, the conventional tanh activation function is substituted with a chaotic oscillator-based dynamic activation, which evolves along a complex nonlinear trajectory; its peak response is used as the activation output, addressing the complexity mismatch inherent in static nonlinearities. Additionally, by using the multifractal spectrum width—an effective high-fidelity measure of the market state—as the direct input to modulate the network’s gating mechanisms, the DFC-LSTM enhances its fidelity to the true internal market structure, ensuring that decision-making is grounded in a more accurate perception of market dynamics.
The main contributions of this paper can be summarized as follows:
We introduce the DFC-LSTM, an architecture that embeds the intrinsic dynamical features of financial markets—namely, fractal complexity and chaos—directly into the core mechanisms of the neural network cell.
We develop a dynamic forget gate modulated by a robust, real-time fractal indicator derived from an Overlapped Sliding Window Multifractal Detrended Fluctuation Analysis (OSW-MF-DFA), which enables an adaptive memory policy responsive to shifts in market regimes.
We are the first to incorporate the maximum trajectory response of a chaotic oscillator as a dynamic activation function within a VaR forecasting LSTM framework, effectively addressing the complexity mismatch challenge in deep learning models.
We empirically demonstrate that the DFC-LSTM delivers more accurate and reliable VaR forecasts across datasets with varying volatility characteristics, including a broad market index and a highly volatile individual stock.
The remainder of this paper is organized as follows:
Section 2 provides a detailed exposition of the DFC-LSTM methodology.
Section 3 describes the data and experimental design.
Section 4 presents the empirical results, while
Section 5 discusses their implications. Finally,
Section 6 offers concluding remarks and outlines potential avenues for future research.
5. Discussion
The empirical results presented in
Section 4 consistently highlight the superior statistical robustness of the proposed DFC-LSTM architecture. A deeper interpretation of these findings begins with the ablation study (
Section 4.1), which provides critical insights into the model’s inner workings. For the challenging S&P 500 series (
Table 3), both the Gate-Only and Chaos-Only variants failed the rigorous DQ test. The full DFC-LSTM, however, successfully passed this test. This strongly suggests that for a persistent, broad market index, the two innovations are not individually sufficient. Instead, they act synergistically: the fractal gate adapts the model’s memory to shifting volatility regimes, while the chaotic activation provides the necessary nonlinear capacity to model the dynamics within those regimes. Only the full model, combining both, was able to achieve the correct dynamic specification. This synergistic relationship is further clarified by the results on the volatile AAPL dataset (
Table 4), where the Standard LSTM baseline failed the L-B independence test. Here, both individual components (Gate-Only and Chaos-Only) were robust enough on their own to fix this issue and pass all tests. This indicates that for highly non-stationary series, either innovation is a significant improvement. However, the full DFC-LSTM achieved a perfect DQ
p-value of 1.0000, demonstrating a level of calibration that even the individual components could not reach. This confirms that the innovations are complementary, leading to the most robust model.
This superior statistical robustness is reinforced when evaluating the model against the full benchmark suite. The primary criterion for a VaR model’s viability is its statistical validity, and our expanded backtesting framework, incorporating the DQ and Ljung-Box tests, underscores this point. On the S&P 500 (
Table 5), we found that violation independence is an extremely difficult property to capture, with all models failing the L-B test. However, the DFC-LSTM was one of the few models (unlike the TCN and GRU) to pass the rigorous DQ test, proving its superior dynamic specification. On the AAPL dataset (
Table 6), the DFC-LSTM’s superiority is unambiguous. It resolves the failure of the Standard LSTM baseline and achieves a perfect DQ score (1.0000), a mark of robustness unmatched by any other model. This demonstrates that the DFC-LSTM’s theory-informed design translates into empirically superior calibration.
From a practical perspective, this statistical reliability is paramount. While the DFC-LSTM demonstrates competitive economic loss scores (RQL and FS), its true value lies in its balance of statistical reliability and economic performance. A model with a slightly lower RQL but a failing DQ p-value (like the TCN on S&P 500) or a failing independence test (like the Standard LSTM on AAPL) is a poor choice for any risk manager. The DFC-LSTM provides an exceptional balance: it is among the best in terms of economic loss while also being the most statistically robust and dynamically well-specified model in the entire study. This success stands in contrast to the failure of the tuned Transformer-based models, which, especially on the S&P 500, suggests that their general-purpose self-attention mechanism may be ill-suited for the high-noise, low-signal environment of single-step-ahead VaR forecasting. In contrast, our DFC-LSTM, which embeds domain knowledge (fractals, chaos) directly into a recurrent cell, proves to be a far more effective architecture.
A final practical consideration is the computational cost. In our experiments, the DFC-LSTM models completed training in approximately 700 s on average. This is not only a reasonable absolute time but is also highly competitive, proving notably faster than more complex benchmarks like the GARCH-LSTM and TCN. This efficiency stems from its design. The fractal analysis (OSW-MF-DFA) component is a one-time preprocessing step performed before training begins and thus adds no overhead to the model training loop. Furthermore, the chaotic activation function’s cost is well-managed; while not computationally identical to a standard tanh, its reliance on a pre-computed lookup table for forward propagation and a simple surrogate gradient () for backward propagation means its additional overhead is minimal. The primary marginal cost arises from the dynamic fractal gate (Equation (16)), which requires one additional set of “expert” weights compared to a standard LSTM’s forget gate. Therefore, the total training cost represents only a modest increase over a standard LSTM and is more efficient than several other benchmark models. We argue this minor computational cost is a highly justifiable price for the model’s demonstrably superior statistical reliability and dynamic calibration.
6. Conclusions
This paper addresses the persistent challenge of accurate Value-at-Risk (VaR) forecasting in the face of complex, non-stationary financial market dynamics. To overcome the limitations of conventional models, we propose the Dynamic Fractal–Chaotic LSTM (DFC-LSTM), a novel architecture that integrates principles from fractal analysis and chaos theory directly into the core computational units of the LSTM cell. The model introduces two synergistic innovations: a multifractal-driven dynamic forget gate that adapts memory retention to real-time market complexity, and a chaotic oscillator-based dynamic activation function that resolves the “complexity mismatch” inherent in static processors.
Our empirical results, based on backtesting on both the S&P 500 index and the volatile AAPL stock, demonstrate the superiority of the DFC-LSTM. The model consistently delivers statistically valid VaR forecasts, passing both unconditional and conditional coverage tests with significantly higher p-values than a comprehensive suite of state-of-the-art benchmarks, including GARCH-LSTM, TCN, and Transformer-based models. This statistical robustness is particularly salient on the challenging AAPL dataset, where several benchmarks fail. While maintaining highly competitive economic loss scores, the DFC-LSTM’s primary strength lies in its exceptional calibration and reliability.
The findings strongly suggest that directly embedding the intrinsic dynamical principles of financial markets into the neural architecture is an effective strategy for improving financial risk modeling. The success of the DFC-LSTM validates our hypothesis that a theory-informed, adaptive architecture leads to more accurate and dependable risk forecasts. Future research could extend this framework in several directions: applying it to other asset classes such as commodities or cryptocurrencies; testing its performance at more extreme risk levels (e.g., 99% VaR) and for complementary risk measures like Expected Shortfall (ES); validating its robustness under different market conditions using more rigorous protocols, such as rolling-window estimations; assessing its practical economic implications; investigating a broader range of chaotic systems for the activation mechanism; or expanding the architecture to a multivariate setting to model systemic risk.