1. Introduction
Time-series analysis has become a cornerstone of modern data science, providing essential tools for understanding, modeling, and forecasting data that evolves over time. Unlike cross-sectional data, which assume independent observations, a time series inherently preserves temporal ordering, making it crucial for capturing dependencies, trends, cycles, and persistence. This temporal structure is particularly important in finance, where accurately modeling and forecasting volatility, returns, and risk directly inform portfolio management, derivative pricing, and risk assessment strategies [
1,
2]. Capturing volatility dynamics is challenging due to clustering, long memory, and abrupt market shifts, which traditional autoregressive (AR) models struggle to represent, as they are limited to short-term dependencies with rapidly decaying autocorrelations, whereas many financial series exhibit slowly decaying correlations indicative of long-range dependence. Fractional time-series models, such as the fractional autoregressive (FAR) and fractional autoregressive moving average (FARMA) models, address this limitation by introducing fractional differencing,
, where
L is the lag operator and
d is a fractional parameter. This enables the process to exhibit intermediate behavior between stationarity and nonstationarity, effectively capturing long memory and making fractional approaches particularly valuable for improving volatility forecasting and supporting informed decision-making under uncertainty [
3,
4].
Unfortunately, like traditional AR models, FAR and FARMA models often face practical limitations because they assume a fixed underlying structure, which restricts their ability to capture evolving dynamics, time-varying dependencies, and abrupt shifts commonly observed in financial volatility. This rigidity can reduce forecasting accuracy and hinder their capacity to reflect market irregularities or regime changes. To address these shortcomings, we extend the classical FAR framework by incorporating time-varying coefficients, resulting in a novel fractional time-varying autoregressive (FTVAR) process, as introduced in Equation (
1), which flexibly adapts to both persistent long-memory effects and dynamic structural changes in financial time series.
where the fractional weights
are defined as Equation (
2).
and
denotes the Gamma function. To model financial volatility using the FTVAR framework, we extend the traditional generalized additive model (GAM) to incorporate the FTVAR process, allowing it to capture both time-varying dynamics and long-range dependencies induced by fractional differencing. By representing time-varying coefficients through flexible, smooth basis functions, GAMs offer interpretable and structured estimation of nonlinear temporal patterns while simultaneously capturing gradual shifts and persistent effects in financial volatility, balancing flexibility with robustness and making them well-suited for processes exhibiting both short-term fluctuations and long-memory behaviors. Additionally, we incorporate the FTVAR process as a structural constraint within Physics-Informed Neural Networks (PINNs), enabling the model to leverage prior knowledge of fractional, time-varying dynamics while retaining the flexibility and expressive power of deep neural networks. This integration enhances generalization and robustness to noise by directly embedding governing constraints into the learning process, guiding the model to respect the underlying dynamics and making PINNs particularly well-suited compared to conventional neural networks. Moreover, the flexible neural network backbone allows PINNs to capture nonlinear, time-varying relationships that are challenging to model with purely data-driven approaches, including standard deep learning architectures. These two distinct methodologies provide complementary perspectives: GAM offers interpretability and smooth functional estimation, while PINNs provide flexibility and data-driven modeling capacity for complex financial volatility patterns. The primary contributions of this research are summarized as follows:
Building on the classical FAR framework, we introduce time-varying parameters into the autoregressive component. This extension accommodates nonstationarity, enhancing flexibility in capturing the evolving dynamics of market volatility.
For modeling the FTVAR process, we adapt the classical GAM framework to incorporate fractional components, and apply a similar extension within the PINNs setting. We evaluate and discuss the strengths and limitations of both approaches in the context of financial volatility modeling.
We provide an open-source implementation of both GAM- and PINN-based frameworks, developed using TensorFlow 2.18 for neural network training and SciPy for numerical optimization, to facilitate reproducibility and future research.
The structure of the paper is organized as follows:
Section 3 introduces the proposed FTVAR process along with the foundational concepts of the GAM- and PINNs-based approaches.
Section 4 reports simulation results under two distinct scenarios, linear and periodic time-varying settings.
Section 5 validates the effectiveness of the proposed methods using real-world financial time-series data from the S&P 500 and the VIX index.
2. Related Work
Currently, dynamic modeling, particularly via autoregressive frameworks, is a fundamental tool for analyzing and forecasting financial market volatility, as it systematically captures temporal dependencies and serial correlations in financial time series while providing a structured framework to reveal complex patterns and evolving dynamics over time. Traditional approaches include Structural Vector Autoregressions [
5], which facilitate the identification of contemporaneous interactions among multiple financial variables; Autoregressive Conditional Heteroskedasticity models [
6], which are effective in modeling time-varying volatility; as well as Nonlinear Autoregressive Distributed Lag [
7,
8] and Quantile Autoregressive Distributed Lag models [
9], which capture asymmetric and distribution-specific dynamics in volatility.
Besides these classical approaches, advances in both theoretical and practical probabilistic modeling, including fractional Brownian motion [
10,
11] and probabilistic machine learning techniques such as autoencoding and diffusion models [
12,
13], have enhanced volatility analysis by enabling a more nuanced representation of uncertainty and stochastic dynamics. Moreover, the rapid advancement of artificial intelligence has driven the development of data-driven methods for volatility modeling. Machine learning techniques [
14] and deep learning architectures, such as Long Short-Term Memory (LSTM) networks [
15,
16,
17] and traditional feedforward neural networks [
18], have been increasingly applied to uncover complex, nonlinear patterns in financial time series. However, while these models often achieve high predictive accuracy, they frequently suffer from a lack of interpretability and limited ability to incorporate domain-specific constraints, which reduces their explanatory power in financial applications.
To address this limitation, Physics-Informed Neural Networks (PINNs) were introduced as a hybrid framework that integrates physical or mathematical constraints with data-driven learning. This framework, first proposed in 2019 [
19], provides a mesh-free and data-efficient methodology that integrates governing equations directly into the loss function through automatic differentiation. The design of PINNs enforces both data fidelity and physical consistency, making PINNs particularly effective for solving ordinary and partial differential equations (ODEs/PDEs), especially in scenarios with sparse or noisy observations. Since their introduction, PINNs have been successfully applied to diverse domains, including wave equations [
20,
21], the Schrödinger equation [
22], and the Huxley equation [
23]. More recently, beyond classical ODE/PDE problems, PINNs have also been extended to fields with well-defined mathematical models, such as cybersecurity (e.g., false data injection detection [
24]), wireless communication (e.g., WiFi signal propagation [
25]), and finance, including option pricing and volatility analysis [
26,
27,
28,
29].
Then, in the context of classical time-varying autoregressive (TVAR) processes, where no fractional component is present, estimation is typically carried out using either the Generalized Additive Model (GAM) framework or kernel-smoothing techniques. GAMs approximate time-varying coefficients via flexible basis functions, while kernel-smoothing methods employ localized weighting schemes to track gradual structural changes in the process. Both approaches have been widely applied across disciplines, including the influence analysis and evaluation of COVID-19 [
30,
31,
32], research in psychology and behavioral sciences [
33,
34], Electroencephalography signal processing [
35,
36,
37], marine science [
38], and the study of financial and commodity market volatility [
39,
40,
41]. These applications highlight the adaptability of TVAR models in capturing dynamic structural changes and their relevance across theoretical and applied domains. Finally, beyond the time-varying framework, several extensions of the fractional autoregressive (FAR) process have also been explored, including the seasonal FAR [
42], nonlinear FAR [
43], periodic FAR [
44], and FARMA models [
45,
46], reflecting the growing interest in fractional dynamics and long-memory structures in time-series analysis.
Despite these advances, a critical gap remains in extending time-varying autoregressive frameworks to fractional processes, which are essential for modeling long-memory and persistent behaviors commonly observed in financial volatility. Existing approaches struggle to balance data efficiency, interpretability, and flexibility when incorporating fractional dynamics, highlighting the need for novel methodologies to bridge this gap.
3. Proposed Methods for Fractional Time-Varying Process Modeling
Recall that the classical AR process is expressed as Equation (
3).
where
denotes the observed one-dimensional time series at time
t,
represents the fixed autoregressive coefficients that capture dynamic dependencies at lag
i, and
is a noise term, typically assumed to be white noise. Next, to incorporate long-memory effects, we follow the fractional differencing formulation proposed by [
47], assuming a zero-order moving average polynomial. This leads to the fractional autoregressive (FAR) model defined in Equation (
4):
where
denotes the fractional differencing parameter, and
L is the lag operator. The operator
, known as the fractional differencing operator, generalizes the conventional integer differencing used in ARMA and ARIMA models. This generalization allows the FAR model to capture long-range dependence that traditional short-memory processes cannot represent. The fractional differencing operator can be expanded using the generalized binomial theorem, as shown in Equation (
5), providing a numerically tractable representation that retains the long-memory property.
where
denote Gamma function. Substituting Equation (
5) back into Equation (
4) and rearranging
, the FAR model can be rewritten in infinite-sum form Equation (
6):
Next, define the variable
, so that Equation (
4) can be rewritten as:
By evaluating
and solving for
in Equation (
7), we obtain:
Substituting
back into Equation (
8) and isolating
yields:
Finally, by separating the infinite sum, the FAR process can be represented in a finite form with an appended tail:
where the first sum captures the short-term dynamics via fixed autoregressive coefficients, while the second summation represents the long-memory structure induced by fractional differencing. Additional theoretical properties and asymptotic behavior of this formulation are further discussed in [
47,
48,
49]. In [
47,
50], they discussed the stationarity of fractional models. When
, the series remains weakly stationary when all the roots of
lie outside the unit circle, whereas values of
produce nonstationary but mean-reverting dynamics.
To further accommodate nonstationarities and evolving dynamics, the time-varying coefficients were introduced into Equation (
10), which allows the autoregressive structure to evolve over time. This leads to the fractional time-varying autoregressive (FTVAR) process Equation (
11).
where
are time-varying coefficients capturing evolving dependencies weights, and
are the fractional weights. Next, based on Equation (
9), we can derive the explicit expression for
as:
From Equation (
5), we know that:
Define
, where
. Substituting this relation yields:
Then, we assume that the polynomial term
remains within the interval
for
. Under this assumption, the equation in Equation (
11) can be approximated as:
which serves as the working formulation in our modeling framework. Several representative FTVAR sample paths based on this approximation are shown in
Figure 1.
Next, we examine the stationarity of the FTVAR model. Following the proof of Theorem 2 in [
47], we first rewrite the process as:
where in our model, the first
R coefficients are time-dependent by assumption. The fractional differencing operator
admits a convergent power series for
when
, ensuring it is well-defined. However, due to the time-varying nature of the time-varying coefficients,
cannot be expressed as a simple power series, and the operator cannot be inverted analytically, leading to global nonstationarity. But locally, where the time-varying terms can be approximated as constants, with the assumption that all the roots of the equation
lie outside the unit circle,
admits a convergent power series representation. Under these conditions, the series expansion of
converges for all
, implying that
is locally stationary when
. For
, the locally spectral density of
satisfies
[
47], indicating locally nonstationarity.
In order to model and make predictions based on the FTVAR process, we consider two methodological approaches: First, we employ Generalized Additive Model (GAM)-based techniques, which have been widely used in the estimation of time-varying autoregressive processes and offer a flexible framework for modeling smooth temporal variations in the coefficient . Second, we introduce a novel Physics-Informed Neural Network (PINN)-based approach, which embeds the governing equations of the FTVAR model directly into the training process, thereby combining data-driven learning with physical constraints to more effectively capture both short-term dynamics and long-range dependencies.
3.1. Generalized Additive Based Approach
The Generalized Additive Model (GAM) is widely regarded as a flexible extension of classical linear models, designed to accommodate nonlinear patterns in data while retaining interpretability. Unlike linear models, which assume that predictors exert fixed linear effects on the response, GAMs replace these fixed coefficients with smooth, non-parametric functions known as smoothers. The key assumption of this approach is that each predictor’s contribution to the response can be represented by a smooth function, thereby relaxing the restrictive linearity constraint and enabling the model to capture complex, nonlinear relationships without the need to prespecify a functional form. In practice, the smooth functions in a GAM are constructed using basis functions, such as splines, radial basis functions, or other suitable families, where the general structure of a GAM with
n basis functions can be expressed as Equation (
14):
where
denotes a chosen smooth basis function, and
is the corresponding linear coefficient in the GAM framework. In this research, we estimate the time-varying coefficients of the FTV-AR process by assuming the existence of
r nonzero smooth functions,
, each of which captures the temporal evolution of a corresponding lagged effect. Within the GAM framework, these coefficients can be represented as a linear combination of smooth basis functions, expressed as Equation (
15):
Next, we consider the estimation of the fractional differencing parameter; estimating all parameters in the FTV-AR model can be formulated as the following optimization problem Equation (
16).
where
d is the fractional differencing parameter,
are the autoregressive coefficients, and
denotes the fractional weights. This formulation seeks the parameter values that minimize the error between the observed series and its reconstructed form.
A key consideration in implementing the GAM framework is the selection of basis functions, as overparameterization can facilitate benign overfitting that enhances model flexibility, whereas excessive flexibility may lead to unstable predictions. Common spline bases offer varying degrees of smoothness and adaptability but are sensitive to the number and placement of knots, potentially causing underfitting or overfitting, whereas alternative basis functions can be employed depending on the target function and the desired trade-offs between interpretability, smoothness, and computational efficiency. In this study, we adopt a combination of spline bases and other selected functions, including Gaussian and tanh, to capture both smooth and nonlinear patterns in the time-varying coefficients, with a subset of the basis functions illustrated in
Figure 2.
Finally, after selecting the appropriate basis functions, several strategies are available to solve the optimization problem in Equation (
16). Unfortunately, unlike classical GAMs used for TVAR simulations, this problem cannot be reformulated as a linear regression due to the presence of fractional weights dependent on
d, making gradient-based optimization a more suitable approach. Currently, there are several methods available, including naive gradient descent, conjugate gradient, and the Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm, which can handle both constrained and unconstrained optimization. Hence, to address the constraints in Equation (
16) while maintaining computational efficiency, we employ a Lagrangian gradient-based solver from the Python 3.9.19 SciPy package, specifically the trust-constr algorithm. The workflow of the GAM-based approach is presented in
Figure 3.
3.2. The Physics-Informed Neural Networks Based Approach
Physics-Informed Neural Networks (PINNs) extend conventional neural architectures by embedding domain-specific knowledge directly into the training procedure, thereby overcoming one of the central limitations of black-box learning: the lack of interpretability and adherence to known physical principles. Rather than relying solely on data-driven empirical fitting, PINNs augment the loss function with additional terms that enforce consistency with governing equations, such as partial differential equations (PDEs), conservation laws, or other structural invariants. In practice, the total loss function is constructed to reflect multiple sources of physical information, typically combining the residuals of the PDE, initial conditions, and boundary conditions, which are expressed as:
where each component quantifies the discrepancy between the neural network approximation and the corresponding physical constraint. Consequently, training a PINN can be formulated as a constrained optimization problem, in which the network parameters are iteratively updated to minimize the composite loss through gradient-based optimization methods. Upon convergence, the trained model provides a mesh-free and continuous surrogate of the underlying physical solution, enabling efficient evaluation at arbitrary points within the domain by directly supplying the independent variables.
In the case of the FTVAR model, the structural constraints cannot be directly formulated as physical laws, which limits their compatibility with the standard PINN framework. This challenge arises from the time-varying nature of the model coefficients, which prevents the system from being represented by a single, time-invariant differential or algebraic equation. By contrast, traditional deep learning approaches to time-series analysis directly estimate or forecast the observed trajectories
, but despite their forecasting ability, they often fail to disentangle intrinsic temporal dynamics from noise or exogenous effects and tend to overlook the autoregressive structure that underpins interpretable and stable representations of dynamical systems. Hence, for time-varying processes, a more principled modeling strategy is to focus on learning the evolution of time-dependent coefficients rather than directly fitting the observed trajectories
, in line with prior studies such as [
51], which emphasize classical PINNs-based TVAR modeling. From this perspective, under the assumption that
d is constant while the coefficients
evolve over time, the total loss functions are formulated as presented in Equation (
17).
where
denotes the mean estimate of
d across all time steps
t. Considering the long-range temporal dependencies inherent in the FTVAR process, the Long Short-Term Memory (LSTM) network represents a suitable modeling choice due to its strong capability to capture dynamic and time-dependent patterns. To further enhance the model’s capacity for representing static and nonlinear relationships, a hybrid architecture integrating both a sequential neural network (NN) component and an LSTM module is proposed. This design allows the NN component to learn deterministic and nonlinear structures, while the LSTM focuses on capturing the temporal dynamics and stochastic variations. Accordingly, the proposed PINN-based framework is organized around two core components:
A sequence of k layers, each equipped with an activation function , trainable input weight matrices , and intercept vectors ;
Hidden-state value and weight matrices for capturing recurrent dynamics in the LSTM component.
Formally, the operation of the LSTM, which models temporal dependencies across lagged inputs
, is expressed as follows:
In the output layer, the output value was separated into two parts: time-varying coefficients
and fractional parameter
d, where an extra layer was added into the fractional parameter due to its constraint from 0 to 1, as illustrated in Equation (
18) and Equation (
19).
where
, and
. Accordingly, the proposed LSTM-based TVAR-PINNs can be interpreted as a solution to the optimization problem defined in Equation (
20).
where
represents the loss function that integrates the time-varying autoregressive components into the neural network framework. The overall architecture of this approach is illustrated in
Figure 4.
Next, to solve the optimization problem defined in Equation (
20), various gradient-based algorithms can be employed, including the Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm, the conjugate gradient method, and Adaptive Moment Estimation (Adam). In this study, we adopt the Adam optimizer due to its complementary strengths in handling large-scale optimization tasks. Adam combines adaptive learning rates with momentum-based updates, enabling efficient and stable convergence even in high-dimensional parameter spaces, which makes it particularly well-suited for models with complex, smooth, and differentiable loss landscapes. Additionally, when paired with methods that support bound constraints, Adam facilitates accurate and reliable estimation of time-varying coefficients, while maintaining computational efficiency and reducing memory overhead—features that are critical for training PINN-based FTVAR models on large datasets.
Finally, the prediction mechanism in our proposed FTVAR framework fundamentally differs from conventional deep learning approaches. Instead of directly forecasting future time-series values, the model estimates the underlying time-varying parameters: the autoregressive coefficients , and the fractional differencing parameter d. Predicting the value at time in an FTVAR model with r lags involves the following steps. First, construct the lagged input vector from observed data spanning to to generate the TVAR component of the prediction. Second, use the estimated fractional parameter and the observed data from to to compute the fractional differenced component. Third, combine these two components to obtain the final prediction . This procedure is then iteratively repeated until reaching the forecast horizon .
3.3. Performance Evaluation
To evaluate the performance of the proposed methods, three complementary measurement criteria were employed to capture both prediction accuracy and the ability to reconstruct the underlying temporal dependence structure of the data.
Mean absolute error (MAE): To assess the accuracy of the predictions, we consider the Mean Absolute Error (MAE). MAE provides a straightforward and interpretable metric by averaging the absolute deviations between the predicted values
and the true values
. The MAE is computed over a forecast horizon of length
H as follows Equation (
21)
Autocorrelation Function (ACF): While MAE captures pointwise predictive accuracy, it does not measure how well the model preserves the temporal correlation structure of the data. To evaluate reconstruction performance, we employ the ACF, which measures the correlation of a time series with its own lagged values and serves as an essential tool for characterizing the dynamics of time-dependent processes. The ACF at lag
K is defined as Equation (
22).
In our evaluation, the ACF is first computed separately for both the original and reconstructed time series across a range of lags, and the MAE between the resulting ACF sequences is subsequently calculated. A lower ACF-based error indicates that the reconstructed series retains more of the temporal dependency present in the original data.
Shapiro–Wilk test: A statistical test used to evaluate whether a sample comes from a normally distributed population, with the null hypothesis
stating that the data is normally distributed. The test statistic
W is defined as Equation (
23):
where
are the ordered sample values from smallest to largest, and
are constants derived from the expected values of order statistics of a standard normal distribution. Values of
W close to 1 indicate approximate normality, while smaller values suggest deviations, with the associated
p value formally testing
, such that
implies that
is not rejected and data are consistent with a normal distribution.
Together, MAE, ACF, and the Shapiro–Wilk test form a comprehensive evaluation framework: MAE measures forecast accuracy in the time domain, while ACF and the Shapiro–Wilk test analyze residuals, with ACF assessing temporal dependencies and structural patterns and the Shapiro–Wilk test evaluating normality.
4. Simulation
In practical implementations, the infinite summation of fractional components must be truncated to a finite lag
, with
M selected to be sufficiently large to preserve the long-memory characteristics while maintaining computational feasibility. Accordingly, the process in Equation (
13) can be approximated by the following finite-lag representation Equation (
24).
while
denotes the time-varying coefficient for lag
k at time
t, and noise term
is assumed to follow a Gaussian distribution with zero mean and variance
. Here, the truncation at lag
M introduces an approximation error due to the neglected tail of the infinite fractional summation. To quantify this error, we define the truncation error term as Equation (
25).
Assume that there exists a finite constant
B such that
for all
t. Substituting this bound into Equation (
25) yields
From the generalized binomial expansion of the fractional differencing operator and the asymptotic approximation of the ratio of Gamma functions, for large
j, the coefficients
satisfy
Putting Equation (
26) into Equation (
25) gives the asymptotic upper bound of the truncation error:
Therefore, the truncation error decays polynomially with order
, implying that larger
M values improve approximation accuracy at a sub-exponential rate. Finally, to determine an appropriate truncation lag
M for a target tolerance
k, which requires
. Using the asymptotic form of
, we have
Hence, to ensure that the truncation error remains below the specified tolerance
k, the
M should satisfy:
From Equation (
28), we observe a critical dependence of the truncation lag
M on the fractional differencing parameter
d. As d approaches 1, the fractional coefficients decay faster, meaning the influence of distant past observations becomes negligible; hence, a smaller truncation lag is sufficient to achieve a desired approximation accuracy. In contrast, as
d is close to
, the term
increases sharply, implying that
M should be significantly larger to achieve the same approximation accuracy. Specifically, when
d becomes sufficiently close to 0, the fractional component effectively vanishes, and the process reduces to a standard TVAR process, while the asymptotic bound in Equation (
28) diverges, reflecting the progressively slower decay of fractional weights.
In summary, the GAM framework models time-varying coefficients via basis functions from historical data, whereas the PINNs approach learns them directly from temporal inputs for multi-step forecasting over horizon H. All analyses were implemented in Python 3.9.19 and executed on a workstation with an NVIDIA RTX 3050 Ti GPU and AMD Ryzen 7 5800H CPU, using GPU acceleration to enhance training efficiency.
4.1. Simulation Preparation
In the simulation study, we examine multiple configurations defined by the fractional differencing parameter , series length with normalized time , Gaussian noise with zero mean and variance , and truncation lags of for and M = 10,000 for . Each setting is characterized by the following time-varying coefficient functions:
Scenario 1: A univariate time series with lag order
and an initial value
. The coefficient is initialized by a constant scalar
and evolves linearly:
Scenario 2: The same setting for lag order
r and initial value
, but with a periodic coefficient. The coefficient is initialized by a constant scalar
and evolves as:
Scenario 3: The same settings are used for the lag order
r and initial value
, but with a discrete coefficient consisting of two constants,
and
, defined for a given time
m as follows:
For the GMA-based methods, optimization was performed using the trust-constr algorithm from SciPy, which implements a constrained optimization framework based on Lagrangian gradient descent, while the PINNs-based approach employed the Adam optimizer. Prediction performance was evaluated using the mean absolute error (MAE) for one-step and five-step forecasts, consistent with established practice in financial time-series analysis, whereas reconstruction performance was assessed using the ACF and the KS test.
4.2. Simulation Results
In this study, flexibility in the GAM-based methods was achieved by employing a diverse set of basis functions, including linear and quadratic terms for polynomial trends, hyperbolic tangent and exponential functions for saturation effects and exponential growth, Gaussian functions for localized smooth approximations, and cosine functions for periodic or oscillatory patterns. This combination enhances the expressive capacity of the models, enabling them to approximate a broad range of functional forms in the estimation of time-varying coefficients. For the PINNs-based approach, we employed a hybrid architecture that integrates a dense sequential network with LSTM components. The model consists of three hidden layers with 16, 32, and 16 neurons, respectively, uses the tanh and sigmoid activation functions, and is trained for up to 1200 iterations.
For each scenario, one-step-ahead predictions were made using 50 randomly generated time series of length 1000 with time t normalized from 0 to 1, and model performance was evaluated across these steps using MAE, ACF, and the Shapiro–Wilk test. To provide a comprehensive comparison, three benchmark models were included: (1) a traditional AR model; (2) a hybrid ARFIMA-GARCH model (denoted as “ARF-GAR”), where ARFIMA estimates the conditional mean and GARCH models the stochastic volatility; and (3) a pure LSTM network composed of three LSTM layers with 32 nodes each and a tanh activation function.
All models were estimated with lag orders
, and the results for Scenarios 1–2 with
and
are presented in
Table 1,
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6. To compute the ACF and Shapiro–Wilk test, we first obtain the one-step predictions (
), and calculate the residuals (
), then apply the ACF and the Shapiro–Wilk test to the residuals, assessing their conformity with standard Gaussian noise, while also reporting the MAE of the ACF as a quantitative measure of goodness-of-fit. After that, a sample of representative one-step prediction results for all methods with lag order
is shown in
Figure 5. Finally, to demonstrate the convergence behavior of the proposed PINNs-based approach, the error convergence rates for three scenarios with
and lag order
are presented in
Figure 6.
4.3. Discussion
Initially, when examining the sensitivity of the proposed models to the finite lag length M, we observed that when d is close to zero, both models exhibit minimal sensitivity as M increases. This outcome aligns with theoretical expectations, since when d approaches zero, the process behaves as a short-memory series, and additional long lags contribute little to predictive performance. For larger values of d, model performance improves slightly with increasing M, although the effect remains marginal. This behavior may arise partly because the series length is finite, larger M values reduce the effective training data, and partly due to the strength of modern open-source optimizers, which can efficiently capture dependencies even with relatively small lag structures. Hence, to balance this trade-off, we recommend setting M to approximately 5% of the total series length, which provides sufficient temporal depth for FTVAR modeling while retaining enough data for effective training. Next, we discuss the overall strengths and weaknesses of all considered methods in light of the results of simulations.
Stationary vs. non-stationary: Based on the MAE of one-step prediction, the FTVAR model generally outperforms the traditional AR model, though our experiments reveal that this advantage is not guaranteed, as in certain scenarios the AR model can perform better. A primary reason for this reversal lies in overfitting: both the GAM and the PINNs-based implementations of FTVAR introduce substantially more parameters than the comparatively simple AR model. When the data are dominated by high levels of noise, or when the underlying dynamics exhibit only limited variability, these additional degrees of freedom can lead to overfitting to noise. In such cases, the expected advantage of non-stationary modeling may be diminished or even eliminated, allowing the more parsimonious stationary AR model to achieve superior predictive accuracy. Moreover, although the AR model typically yields competitive MAE with residuals that are statistically acceptable, residual ACF values below 0.05, and relatively high mean p-values in the Shapiro–Wilk test, its one-step predictions are often smoother than those produced by GAM- or PINNs-based FTVAR models. This effect is especially evident in Scenario 1, which aligns well with the characteristics of the AR model, as its predictions can be interpreted as a form of conditional mean estimation. Next, compared with the ARFIMA-GARCH model, both the GAM- and PINNs-based approaches exhibit consistently superior performance across MAE and ACF. This improvement highlights their enhanced capacity to capture nonlinear and non-stationary behaviors inherent in the FTVAR process. In contrast, the ARFIMA-GARCH model, which assumes a relatively stable mean and volatility structure, exhibits greater randomness but limited adaptability to time-varying dynamics, resulting in reduced accuracy when modeling such complex processes.
PINNs-based vs. GAM-based: Our simulations suggest that neither PINNs nor GAM-based models can be regarded as universally superior. PINNs offer greater flexibility by bypassing predefined basis functions, but this comes at the cost of a substantially larger parameter space that makes them more susceptible to noise and highly sensitive to the choice of activation functions, introducing additional variability in performance. Conversely, the performance of GAM models depends heavily on the choice of basis functions, and inadequate specification can lead to underfitting and reduced adaptability in complex or rapidly changing environments. As a result, the trade-offs between flexibility and robustness make direct comparisons between PINNs- and GAM-based approaches nontrivial, with their relative effectiveness ultimately depending on the characteristics of the data and the modeling objectives.
Length and lag order: Our simulations indicate that increasing the order generally improves performance: as the lag length grows, both the MAE and ACF decrease, suggesting that higher-order specifications allow the PINNs- and GAM-based approaches to better capture the underlying temporal structure of the process. Regarding sample length, although we did not explicitly report it in the main results, additional experiments revealed that our proposed methods are largely insensitive to the length of the simulated series. Specifically, when evaluating the same process with lengths of 750, 1000, and 2000 observations, the resulting performance metrics remained consistent, indicating that the approaches maintain robustness across different data sizes. This robustness highlights the adaptability of both frameworks to varying empirical settings without requiring strict tuning of sample length.
Convergence and stability of the PINNs-based approach: As illustrated in
Figure 6, the training loss exhibits a consistent downward trend with minor oscillations, suggesting that the proposed PINNs loss formulation is well aligned with the optimization dynamics of the open-source ADAM minimizer. The adopted learning configuration effectively maintains a balance between numerical stability and fitting accuracy, as evidenced by the steady reduction in loss that reflects the network’s progressive capture of the underlying system dynamics without signs of divergence. Overall, these results confirm the reliability of the proposed framework in achieving stable and convergent training performance, even in the presence of non-stationary and time-varying components.
4.4. Limitations
The main limitation of both the GAM- and PINNs-based approaches lies in their large number of parameters. For instance, in the GAM-based framework, assuming the use of 6 basis functions with lag order (r), the total number of parameters to be estimated is
without an intercept and
with an intercept. Similarly, in the PINNs-based approach, if the network consists of
n sequential layers with dense sizes
, the total number of parameters to be estimated is
. Specifically, from a computational standpoint, the proposed framework exhibits moderate complexity arising from its large parameter space and the iterative nature of the ADAM and BFGS optimization algorithms. While these optimizers ensure stable and reliable convergence, the overall training cost increases substantially with the number of hidden nodes, revealing a trade-off between robustness and efficiency. This limitation highlights the need for future research to systematically analyze the computational characteristics of different network architectures and optimization schemes to achieve a more balanced compromise among robustness, accuracy, and computational efficiency. Furthermore, to clearly illustrate these computational costs, particularly for the PINN-based framework relative to the GAM-based approach, the detailed time consumption of each method is summarized in
Table 7.
Another key limitation concerns the risk of overfitting. Owing to the large number of parameters, both GAM- and PINNs-based approaches exhibit heightened sensitivity to noise in the data. This sensitivity increases the likelihood that the models capture spurious patterns rather than genuine temporal dependencies, leading to overfitting. As a result, the supposed advantage of the flexible FTVAR framework over traditional AR models may be reduced, or even eliminated, when the fitted models adapt excessively to noise. This issue highlights the importance of careful model regularization, hyperparameter tuning, and validation strategies to ensure that the enhanced flexibility of GAM- and PINNs-based FTVAR models in predictive performance rather than artifacts of overfitting.
5. Application to Financial Volatility Series
In this real-world application, we model financial time series using key market indicators that capture both historical dynamics and forward-looking investor expectations. Specifically, we consider the S&P 500 3-month volatility index, representing overall U.S. equity market performance, and the VIX index, which reflects market volatility and anticipates near-term market risk. The S&P 500 3-month volatility index effectively capturing the patterns of historical volatility clustering and persistence. In contrast, the VIX provides a forward-looking measure of implied volatility, reflecting market participants’ expectations of future risk. By integrating these complementary measures, our analysis accounts for both observed and expected volatility, providing a comprehensive framework for modeling and understanding financial market dynamics.
All financial data used in this study were obtained from publicly available sources, including the Federal Reserve Economic Data (FRED,
https://fred.stlouisfed.org/) and the Chicago Board Options Exchange (CBOE,
https://www.cboe.com/), as shown in
Figure 7. These indices are widely recognized in both academic research and industry practice for their ability to capture market behavior, investor sentiment, and macro-financial conditions. Their properties, including persistence, volatility clustering, and long-memory effects, make them particularly suitable for FTVAR modeling, allowing us to evaluate how these frameworks capture both the historical and forward-looking characteristics of financial volatility.
To evaluate the performance of our methods on real-world financial data, we use the dataset spanning 4 January 2016 to 31 December 2024 to train the TV-VAR model under both the GAM- and PINNs-based frameworks with lag orders of
and
, performing one-step-ahead forecasting and comparing the predicted values to the actual observations. In addition to MAE, we assess temporal dependence to evaluate model adequacy. Unlike simulations, where ACF can be applied to stationary residuals, real-world series are often non-stationary, making residuals less meaningful. Since both the VIX and the S&P 500 3-month volatility can be considered approximately stationary, we apply the ACF directly to the predicted sequence to preserve the temporal ordering and long-range dependence of the original data, allowing us to assess whether the model captures the underlying stationary or quasi-stationary structure of financial volatility. The complete results are presented in
Figure 8 and
Figure 9, and
Table 8. Next, to enable a deeper analysis during the COVID-19 pandemic, a specific result is presented in
Table 9.
The findings show that both the PINNs- and GAM-based approaches achieve strong predictive performance for the VIX, indicating that the proposed FTVAR framework effectively captures the dynamics of implied volatility. In contrast, the S&P 500 3-month volatility presents a greater challenge: predictive accuracy declines sharply around 2020, when the COVID-19 shock triggered unprecedented fluctuations. Compared with the traditional ARFIMA-GARCH and LSTM models, our two approaches demonstrate notably higher performance, particularly when larger lag structures are considered. This highlights the limitation of both methods in handling abrupt structural breaks, while also demonstrating the flexibility of the FTVAR approach through the estimated baseline curve for S&P 500 volatility. Unlike the original straight-line structure, the model captures a non-linear increase-then-decrease trend, particularly evident in the GAM-based estimates, reflecting its ability to adapt to gradual regime shifts. For the VIX, both methods again perform well, though the PINNs results display more variability than the smoother GAM outputs, indicating a trade-off between flexibility and stability in model choice. Overall, these results emphasize the strengths of our framework in capturing persistent volatility structures, while also underscoring the difficulty of modeling extreme market episodes.
Finally, as shown in
Table 9, during the COVID-19 period, both the PINN- and GAM-based approaches achieve lower prediction errors for S&P 500 3-month volatility and the VIX. These methods also produce lower ACF values, indicating that the reconstructed series better preserves the temporal dependence present in the original data. Comparing
Table 8 and
Table 9 shows that MAEs for both volatility measures increase during the stress period, reflecting heightened turbulence. Nevertheless, ACFs are lower in the COVID-19 period than over the entire sample. Taken together, these results indicate that, even under stress, our model outperforms the traditional benchmark in predictive accuracy and that the reconstructed series maintains stronger temporal correlation with the observed data during the crisis.