Next Article in Journal
Determinants of Financial Stability and Development in South Africa: Insights from a Quantile ARDL Model of the South African Financial Cycle
Previous Article in Journal
Investor Emotions and Cognitive Biases in a Bearish Market Simulation: A Qualitative Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Financial Volatility Under Structural Breaks: A Comparative Study of GARCH Models and Deep Learning Techniques

1
Departamento de Estadística, Universidad Nacional Pedro Ruiz Gallo, Chiclayo 14001, Peru
2
Departamento de Ciencia, Universidad Tecnológica del Perú, Chiclayo 14001, Peru
3
Departamento de Incorporación de Nuevos Saberes, Universidad Nacional de Ingeniería, Lima 150101, Peru
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2025, 18(9), 494; https://doi.org/10.3390/jrfm18090494
Submission received: 11 August 2025 / Revised: 2 September 2025 / Accepted: 2 September 2025 / Published: 4 September 2025
(This article belongs to the Section Financial Technology and Innovation)

Abstract

The main objective of this study is to evaluate the predictive performance of traditional econometric models and deep learning techniques in forecasting financial volatility under structural breaks. Using daily data from four Latin American stock market indices between 2000 and 2024, we compare GARCH models with neural networks such as LSTM and CNN. Structural breaks are identified through a modified ICSS algorithm and incorporated into the GARCH framework via regime segmentation. The results show that neglecting breaks overstates volatility persistence and weakens predictive accuracy, while accounting for them improves GARCH forecasts only in specific cases. By contrast, deep learning models consistently outperform GARCH alternatives at medium- and long-term horizons, capturing nonlinear and time-varying dynamics more effectively. This study contributes to the literature by bridging econometric and deep learning approaches and offers practical insights for policymakers and investors in emerging markets facing recurrent structural instability.

1. Introduction

Accurately forecasting financial volatility is a cornerstone of risk management. It matters not only for investors and portfolio managers but also for regulators concerned with financial stability. Reliable volatility models help anticipate market turbulence, guide asset allocation, improve hedging strategies, and ultimately reduce exposure to systemic risk. In emerging markets, where political, macroeconomic, and institutional shocks are frequent, the ability to capture and predict volatility dynamics becomes especially important.
Over the past four decades, conditional heteroscedasticity models such as ARCH (Engle, 1982) and GARCH (Bollerslev, 1986) have become standard tools for modeling the variance of asset returns. Numerous studies confirm their effectiveness in capturing volatility clustering and persistence in both developed and emerging markets (Babikir et al., 2012; Rapach & Strauss, 2008). However, these models rely on the assumption of parameter stability, which is often unrealistic in environments where crises, regime shifts, or external shocks alter market dynamics. Ignoring such breaks produces specification errors, biased estimates, and an overstatement of volatility persistence (Mikosch & Stărică, 2004). To address this limitation, methods such as the ICSS algorithm (Inclán & Tiao, 1994) and the multiple change-point tests of Bai and Perron (1998) have been widely applied. Evidence shows that incorporating structural breaks improves the forecasting performance of GARCH models across different markets (e.g., De Gaetano, 2018; Gong & Lin, 2018; Kumar, 2018). Nevertheless, most of this literature remains confined to traditional econometric frameworks.
In contrast, recent advances in machine learning—particularly deep learning—have opened new possibilities for volatility modeling in unstable environments. Long Short-Term Memory (LSTM) networks (Hochreiter & Schmidhuber, 1997) and Convolutional Neural Networks (CNNs) (LeCun et al., 1998) have been successfully applied to forecasting prices, returns, and volatility in equity and cryptocurrency markets (Fischer & Krauss, 2018; Khan et al., 2023; Nelson et al., 2017; Petrozziello et al., 2022). Their advantage lies in capturing nonlinear dynamics, regime shifts, and long-memory effects without imposing restrictive assumptions on the data-generating process. This flexibility makes them well suited to environments characterized by recurrent structural breaks. Unlike GARCH-type models, they can adapt more naturally to abrupt changes in volatility regimes. However, few studies have systematically compared deep learning and GARCH approaches under explicit structural instability. This remains a critical gap in the literature.
This study addresses that gap. We conduct a systematic empirical evaluation of the predictive performance of GARCH models—estimated under expanding, rolling, and break-segmented windows—versus deep learning models (LSTM and CNN) in the presence of structural breaks in conditional variance. The analysis uses daily data from four major Latin American stock indices: IGBVL (Peru), BOVESPA (Brazil), IPSA (Chile), and IPC (Mexico) over 2010–2024. By covering episodes such as the global financial crisis, the COVID-19 pandemic, and recent geopolitical tensions, the dataset provides a robust setting to examine volatility forecasting under recurrent shocks. The central objective is to determine which class of models offers greater robustness under structural instability, using metrics such as MSFE and QLIKE, along with the tests of White (2000) and Hansen (2005) to evaluate predictive accuracy.
We hypothesize that the predictive power of GARCH models is substantially weakened when structural breaks are not explicitly modeled. In contrast, deep learning architectures—because they do not rely on rigid stability assumptions—are expected to sustain, and in some cases improve, forecasting accuracy in highly dynamic environments.
The remainder of this paper is organized as follows: Section 2 reviews the related literature on structural breaks, volatility forecasting, and deep learning applications. Section 3 presents the data and methodology. Section 4 reports the empirical results. Section 5 discusses the findings in light of prior research. Section 6 concludes and highlights the practical implications for investors and risk managers in emerging markets.

2. Literature Review

Recent literature on volatility forecasting and financial risk has converged along two main paths: (i) the explicit modeling of structural breaks and regime shifts in conditional variance and (ii) the adoption of deep learning architectures and hybrid frameworks that combine econometric specifications with nonlinear learning. Below, we summarize the most relevant contributions along these thematic axes.

2.1. Structural Breaks and Regime Dynamics

Earlier contributions consistently show that the explicit inclusion of structural breaks improves both the fit and the predictive ability of GARCH models across different financial markets (De Gaetano, 2018; Gong & Lin, 2018; Kumar, 2018). Moving-window approaches—whether expanding or rolling—have been combined with segmentations guided by empirically detected breaks (De Gaetano, 2018). Moreover, methodologies such as impulse indicator saturation (IIS), developed by Oxford researchers, jointly detect breaks and outliers, offering additional robustness in volatility modeling (Castle et al., 2022).
Building on these foundations, more recent studies highlight that ignoring breaks induces persistence biases and undermines predictive accuracy. For instance, Luo et al. (2025) identify breaks in realized betas and asymmetric risk effects, combining ICSS predictors with LASSO selection in HAR models to improve out-of-sample performance. Similarly, Sun et al. (2025) propose a multiplicative structure that separates shifts in unconditional variance—detected via binary segmentation—from conditional GARCH dynamics, achieving a better fit under nonstationarity. In fragile exchange rate regimes, Daboh et al. (2025) confirm regime shifts using EGARCH and Markov-Switching models, while Amin et al. (2024) document that global recessions trigger sharp increases in commodity volatility with product-specific patterns. During the COVID-19 crisis, de Oliveira et al. (2024) showed that long memory and breaks coexist, and that modeling them jointly prevents misspecification. Along these lines, Bildirici et al. (2020) design a hybrid LSTAR-GARCH-LSTM approach, where smooth regime transitions in both mean and variance are integrated with the adaptive capacity of LSTM, producing superior forecasts for crude oil under COVID-19 shocks. Applied to the stock–currency nexus in China, Fang and Yang (2025) combine a DCC-CGARCH with break analysis and RR-MIDAS, linking ruptures to the GFC, regulatory reforms, trade wars, and COVID-19. In crypto and equity markets, Tsuji (2025) integrate breaks and asymmetric errors into a unified GARCH framework, uncovering double asymmetries between Bitcoin and the S&P 500.

2.2. Deep Learning and GARCH Hybridization

In parallel, deep learning techniques have expanded the modeling frontier by combining econometric structures with neural networks. Araya et al. (2024) merge ARIMA/ARCH–GARCH with CNN/LSTM; Di-Giorgi et al. (2025) show that GRU, LSTM, and BiLSTM paired with GARCH improve weekly forecasts; and Bhambu et al. (2025) design a GARCH–MLP–Mixer for high-frequency data, outperforming benchmarks also in VaR estimation. In cryptocurrencies, García-Medina and Aguayo-Moreno (2024) find that LSTM–GARCH dominates pure GARCH both under heteroscedastic error metrics and portfolio criteria. The use of intraday data has further amplified gains: DeepVol (Moreno-Pino & Zohren, 2024) employs dilated causal convolutions with high-frequency inputs, while Hortúa and Mora-Valencia (2025) apply Bayesian deep learning to the VIX with calibrated uncertainty. On interpretability, Zhang et al. (2024) integrate a memory-based architecture and a large language model to generate narratives explaining volatility spikes. Nevertheless, results are not universal: Zitti (2024) report that ARMA outperforms LSTM in the salmon market, underscoring that nonlinear gains depend on the asset and horizon. In Latin America, Kristjanpoller and Minutolo (2014) find that ANN–GARCH hybrids improve forecasts for Brazil, Chile, and Mexico, while Patra and Malik (2025) detect volatility connectedness between the US and Latin America through Q–VAR. Similarly, Alfeus (2025) show that HAR, RealGARCH, RECH, and RFSV outperform standard GARCH in South Africa.
Another active line of research extends hybridization beyond univariate volatility to capture dynamic correlations across markets. For example, Ni and Xu (2023) improve DCC–GARCH by correcting its errors with a recurrent deep neural network and an autoencoder, increasing accuracy across China, Hong Kong, the US, and Europe. Along similar lines, Chung et al. (2024) integrate DCC–GARCH with LSTM to analyze contagion from the US to Latin America during COVID-19, detecting significant spillovers from the S&P 500 to regional indices (except MERVAL). Their results show that LSTM enhances the prediction of dynamic correlations and provides early-warning signals, illustrating the benefits of coupling conditional dependence models with recurrent networks under high uncertainty.

2.3. Textual Information, Big Data, and News

A complementary strand of work focuses on news and textual data as volatility predictors. Bodilsen and Lunde (2025) show that domestic macro sentiment improves forecasts for the S&P 500 and single stocks, especially at long horizons, while overnight news counts add value at the 1-day horizon. Using a regime-switching design, Boubaker et al. (2021) link news diversity to crashes, finding that diversity falls in bearish high-volatility periods and rises during recoveries, with structural breaks reinforcing predictive power.

2.4. Reviews, Model Selection, and Robustness

Recent scoping reviews emphasize the rapid growth of ML applications in volatility forecasting. Molina Muñoz and Castañeda (2023) report an exponential rise in publications since 2019, with hybrid models dominating. Qiu et al. (2025) compare implied volatility, GARCH, LSTM, and Transformers, highlighting complementarity rather than substitution. On model selection, Hassanniakalager et al. (2024) introduce an FDR-based framework to identify “buckets” of superior models, showing that GJR–GARCH consistently minimizes one-step-ahead errors in realized volatility. Moreover, recent evidence (e.g., Di-Giorgi et al., 2025) shows that Gated Recurrent Units (GRUs) often outperform LSTM in commodity markets, suggesting that GRU represents a promising extension for future research. These approaches are especially valuable for comparisons involving rolling versus expanding windows, explicit breaks, and deep learning alternatives under the risk of overfitting.

2.5. Study Objective and Hypotheses

In light of these gaps, our study offers a systematic, head-to-head comparison of GARCH models with structural breaks—estimated under expanding, rolling, and break-segmented windows—against deep learning architectures (LSTM and CNN) in Latin American equity markets subject to recurrent shocks. We hypothesize as follows:
  • Ignoring structural breaks reduces the predictive accuracy of GARCH models, whereas explicitly modeling them improves performance.
  • Deep learning models, thanks to their flexibility, maintain or enhance predictive accuracy in the presence of structural instability.
  • Under a rigorous out-of-sample evaluation based on MSFE, QLIKE, and the tests of White (2000) and Hansen (2005), deep learning approaches will show superior forecasting performance in highly dynamic environments such as Latin American financial markets.

3. Materials and Methods

3.1. Data

Our analysis relies on the daily returns of four representative Latin American stock indices: IGBVL (Peru), BOVESPA (Brazil), IPSA (Chile), and IPC (Mexico). These markets were selected because they are classified as emerging economies that are highly vulnerable to both domestic and international shocks, making them an ideal setting to examine the impact of structural breaks on financial volatility. The study period, from 2 January 2000 to 1 January 2024, was chosen to capture a wide range of major economic, financial, and political events known to induce abrupt regime changes. These include the 2008 global financial crisis, regional institutional and political crises, the Chilean social unrest of 2019, and the COVID-19 pandemic. By covering this extended and turbulent horizon, our analysis provides a robust context to evaluate model performance under different volatility regimes and structural instability scenarios. The data were obtained from Yahoo Finance (https://finance.yahoo.com, accessed on 2 February 2025).
Daily returns were calculated as follows:
r t = 100 [ l n P t l n P t 1 ]
where r t represents the asset returns at time t, and P t and P t 1 are the closing prices at times t and t 1 , respectively.
The methodology in this study presents two main blocks: first, econometric techniques based on GARCH models and structural break detection are developed, and the second block incorporates deep learning models with the objective of capturing nonlinear patterns and long-run dependencies in the dynamics of asset volatility.

3.2. Modified ICSS Algorithm

The Iterative Cumulative Sum of Squares (ICSS) algorithm, proposed by Inclán and Tiao (1994), allows detecting multiple changes in the unconditional variance of a time series. Since financial series commonly exhibit conditional heteroscedasticity, a modified version of the ICSS is used that incorporates nonparametric adjustments to correct for possible autocorrelation, as occurs in GARCH processes.
The methodology is based on the cumulative sum of the squares of the observations, denoted as y t 2 , as follows:
C k = i = 1 k y i 2
From this cumulative sum, the relative change statistic is constructed, as follows:
D k = C k C T k T
In contrast with the original formulation of the ICSS test—which detects breaks when the statistic A I T = sup t ( T 2 ] T 2 D t exceeds a critical threshold λ —the modified version implemented in this study, inspired by Sanso et al. (2004), uses the statistic G k . This incorporates a robust estimate of the unconditional variance, denoted by λ ^ , which considers the squared autocovariances of the returns:
G k = λ ^ 0.5 C k k T C T
Given that the original algorithm by Inclán and Tiao (1994) assumes independence between observations—an unrealistic assumption in the presence of ARCH or GARCH effects—the modification proposed by Ni et al. (2016) extends its applicability to contexts with temporal dependence through a nonparametric approach.
The procedure is applied recursively: when a significant break point is identified, the series is divided into two segments (before and after the break), and the test is reapplied to each subsegment. This process is repeated until no new breaks are detected, allowing multiple breaks in the variance to be identified. This segmentation is then used to adjust GARCH models by regime or introduce dummy variables that capture the detected discontinuities.
It is worth noting that alternative structural break tests have been widely employed in the econometric literature. In particular, the multiple breakpoint test of Bai and Perron (1998) and the Quandt–Andrews test (Andrews, 1993; Quandt, 1960) are commonly applied in financial econometrics to detect shifts in regression parameters or variance structures. Although these methods were not implemented in the present study, they provide complementary perspectives and constitute promising avenues for robustness checks in future research. Our focus on the modified ICSS approach is motivated by its ability to directly address unconditional variance shifts in the presence of GARCH effects, which is the central concern of this paper.

3.3. GARCH Model

The dynamics of conditional volatility can be modeled using a Generalized Autoregressive Conditional Heteroscedasticity process of order (1, 1), known as GARCH(1, 1). This model captures the temporal dependence in the variance of the errors, allowing both recent shocks and previous levels of volatility to be incorporated.
The model is specified for errors e t as
e t = σ t ϵ t
σ t 2 = ω + α e t 1 2 + β σ t 1 2
where ω > 0 , α 0 , and β 0 are model parameters. Here, σ t 2 represents the conditional variance of e t , while ϵ t is an i.i.d. process with zero mean and unit variance, typically assumed to be standard normal.
The condition α + β < 1 guarantees that the process has finite unconditional variance, given by
Var ( e t ) = ω 1 α β
Furthermore, the process e t is a martingale in differences and constitutes white noise with mean zero.
The parameters are commonly estimated using the quasi-maximum likelihood (QMLE) method, which produces consistent and asymptotically normal estimators even under moderate deviations from normality (Blasques et al., 2018; Stărică & Granger, 2005). This method is particularly suitable when the conditional distribution of the errors is not completely known, but a convenient functional form (e.g., normal) is assumed for constructing the likelihood function.

3.4. Deep Learning

Deep learning approaches have recently gained prominence as powerful alternatives for modeling complex financial phenomena characterized by nonlinearity, volatility clustering, and structural changes. Architectures such as Long Short-Term Memory (LSTM) networks and Convolutional Neural Networks (CNNs) have been successfully applied to forecasting asset prices, returns, and volatility. In this study, LSTM and CNN models are included because of their flexibility in capturing nonlinear dynamics and regime shifts. Unlike GARCH models, they do not assume parameter stability, which makes them particularly suitable for forecasting volatility during structural breaks.

3.4.1. Long Short-Term Memory (LSTM)

Long Short-Term Memory (LSTM) networks, introduced by Hochreiter and Schmidhuber (1997), represent a specialized architecture within recurrent neural networks (RNNs), designed to mitigate the problem of gradient vanishing. This capability makes them particularly well suited to capturing long-term dependencies in sequential data, such as financial returns, where volatility dynamics often exhibit persistent memory and nonlinear behavior.
The main strength of LSTMs lies in their internal memory mechanism, which selectively regulates what information is retained, forgotten, or updated at each time step. This functionality is implemented through a set of gates (for forgetting, input, and output), whose operation is given by the following set of equations:
  • Forget Gate ( f t ): Regulates which information from the previous cell state should be removed. It applies a sigmoid activation to the concatenation of the prior hidden state h t 1 and the current input x t , producing a weight between 0 and 1 that controls the degree of forgetting, as follows:
    f t = σ ( W f · [ h t 1 , x t ] + b f )
  • Input Gate ( i t ): Determines the portion of new information to be incorporated into the memory cell. Through a sigmoid activation, it selects the relevant signals to be updated, as follows:
    i t = σ ( W i [ h t 1 ; x t ] + b i ) .
  • Candidate Vector ( C ˜ t ): Generates a vector of candidate values that may enrich the cell state. This vector is formed by applying a hyperbolic tangent transformation to the weighted input and previous hidden state, as follows:
    C ˜ t = tanh ( W C [ h t 1 ; x t ] + b C ) .
  • Cell State Update ( C t ): Integrates retained information from the previous state with newly generated candidate values to update the cell memory. The previous state is scaled by the forget gate f t , while the candidate vector is modulated by the input gate i t . Both contributions are combined through the Hadamard product, i.e., element-wise multiplication (Magnus & Neudecker, 2019), as follows:
    C t = f t C t 1 + i t C ˜ t .
  • Output Gate ( o t ): Selects which parts of the updated memory will influence the current output. A sigmoid activation provides the weighting, as follows:
    o t = σ ( W o [ h t 1 ; x t ] + b o ) .
  • Hidden State ( h t ): Produces the actual output of the LSTM at time t. It is obtained by applying a hyperbolic tangent to the updated cell state and filtering it through the output gate, as follows:
    h t = tanh ( C t ) o t .

3.4.2. Convolutional Neural Networks (CNNs)

Convolutional Neural Networks (CNNs) were initially designed to process spatially structured data, but they have been successfully adapted to time series applications, including financial forecasting (Dauphin et al., 2017). In this temporal context, CNNs apply one-dimensional convolutional filters along the sequence of observations, enabling the detection of local patterns and temporal dependencies. Such filters are particularly effective in financial data, where abrupt peaks, volatility clusters, and short-lived structural regularities often signal shifts in future market behavior.
The predictive process in CNNs involves the following key components:
  • Convolutional Layers: Extract local features by applying filters that slide across the input series, generating feature maps.
  • Pooling Layers: Downsample these feature maps, commonly through max pooling, to reduce dimensionality while preserving the most salient features.
  • Flattening: Convert the multidimensional feature maps into a one-dimensional vector suitable for subsequent processing.
  • Fully Connected Layers: Map the extracted features to the final output, allowing the network to combine information across different filters and learn complex relationships.
The kernel size of the convolutional filters defines the temporal window explored at each step. Smaller kernels capture short-term dynamics, while larger kernels incorporate longer-term dependencies. Because financial time series often exhibit patterns at multiple horizons, it is common to employ several convolutional layers with varying kernel sizes, enabling the model to build a hierarchical representation of temporal structures and enhance predictive accuracy.

3.5. Forecast Performance Metrics

The predictive accuracy of volatility models is assessed using loss functions that quantify the discrepancy between observed volatility and generated forecasts.

3.5.1. Aggregated Mean Squared Forecast Error ( M S F E * )

Stărică and Granger (2005) propose a specific loss function to compare the predictive performances of volatility models based on a temporal aggregation approach. The resulting metric is an extended version of the mean squared forecast error (MSFE), which incorporates information about a cumulative prediction horizon of length s, as follows:
M S F E S , 1 * = 1 P ( S 1 ) t = R + 2 T e ˜ t 2 σ ^ ˜ t t s , 1 2 2
where e ˜ t 2 = j = 1 s e t ( j 1 ) 2 represents the cumulative sum of squared returns in the forecast window, and σ ^ ˜ t t s , 1 2 = j = 1 s σ ^ t ( j 1 ) t s , 1 2 corresponds to the sum of the conditional variances estimated by the model under evaluation, calculated over the same horizon.
To determine whether the differences observed in predictive accuracy are statistically significant, the statistic developed by Babikir et al. (2012) is used. This procedure allows us to evaluate the superiority of one model over another when there is nesting between them—that is, when one model can be considered a restricted version of the other.
In the context of this research, the models evaluated are considered nested with respect to the base model (GARCH(1, 1) expanding window). The null hypothesis states that there is no difference in the average performance of the models compared. Rejecting this hypothesis implies that the model of interest offers a statistically significant improvement in terms of predictive power.

3.5.2. Quasi-Likelihood Loss (QLIKE)

QLIKE is motivated by the log-likelihood of the normal distribution and has advantages when evaluating volatility models, as it does not excessively penalize overestimates like MSFE. It is defined as
Q L I K E = 1 P t = R + 1 T log ( h ^ t ) + r t 2 h ^ t
This metric is particularly robust to conditional heteroscedasticity, and has been recommended by Patton (2009) as a preferred criterion in financial contexts. Lower values of QLIKE indicate better model fit to the observed volatility behavior.

3.5.3. White and Hansen Test

To assess whether differences in predictive performance between models are statistically significant, the loss function equality tests proposed by White (2000) and Hansen (2005) are used.
Both procedures are based on the series of loss differences, as follows:
d t = L t ( 1 ) L t ( 2 )
where L t ( 1 ) and L t ( 2 ) represent the losses (e.g., MSFE or QLIKE) associated with models 1 and 2, respectively. The null hypothesis being tested is
H 0 : E [ d t ] = 0
The White test is based on the average of d t , adjusted for its robust variance (usually estimated using the Newey–West method), and allows for comparing the equality of performances between two models.
The Hansen test introduces the notion of Superior Predictive Ability (SPA), allowing us to evaluate whether a particular model statistically dominates a set of alternatives under multiple comparisons. This approach is useful when comparing several models simultaneously.

3.6. Methodology

This study adopts a rigorous and replicable methodological approach, in line with the best practices proposed by Pesämaa et al. (2021), which highlight the importance of transparency, replicability, and robustness in the experimental design of quantitative research.
The methodology combines traditional econometric tools with deep learning models, with the main objective of comparing the predictive power of GARCH models (with and without structural breaks) and deep neural networks (LSTM and CNN) in predicting the conditional volatility of financial assets, specifically in environments characterized by structural instability. The dataset was divided into two main parts: an 80% training set and a 20% test set. The training set was used for model estimation, while the test set was strictly reserved for out-of-sample forecasting. For the deep learning models, a portion of the training sample was internally allocated as a validation set to tune hyperparameters. This design ensures a strict separation between estimation and forecasting phases and prevents any look-ahead bias. The procedure is organized into two main blocks: in-sample evaluation and out-of-sample evaluation.

3.6.1. In-Sample Evaluation

  • Stage 1: Identification of structural breaks
The ICSS (Iterated Cumulative Sums of Squares) algorithm from Inclán and Tiao (1994), in its modified version, was applied to detect breakpoints in the unconditional variance of the returns of the stock indices analyzed. The breaks were detected across the entire available sample in order to obtain a comprehensive understanding of the historical structural dynamics of volatility, which is essential for robustly informing the subsequent stages of modeling.
  • Stage 2: Estimation of models in subsamples
Once the breakpoints were identified, the time series was segmented into homogeneous subsamples. For each subperiod, GARCH(1, 1) volatility models were fitted to capture the particular characteristics of the volatility dynamics in each identified regime. The estimation was performed using maximum likelihood.

3.6.2. Out-of-Sample Evaluation

  • Stage 3: Forecasting design
For the out-of-sample evaluation, the full series was divided into a training sample of size R and a test sample with P observations, such that T = R + P . The predictive performance of all models was assessed exclusively on the test sample. We implemented and compared four variants of the GARCH(1, 1) model with two deep learning architectures:
  • GARCH(1, 1) with Expanding Window: This model served as our benchmark. The estimation sample was expanded recursively: starting with data from t = 1 to t = R , the model was estimated and a one-step forecast was generated. Then, at each iteration, a new observation was added and the model was re-estimated to issue the next forecast. This approach inherently assumes parameter stability over time.
  • GARCH(1, 1) with Rolling Window 0.25 and 0.5: These models used a fixed-size estimation window equivalent to 25% or 50% of the training sample. The window slid forward one step at a time, and the model was re-estimated for each new forecast, allowing parameters to be dynamically updated with the most recent information.
  • GARCH(1, 1) with Structural Breaks: For this dynamic approach, the modified ICSS algorithm was applied to the observations available up to time R. If at least one significant variance break was detected, the GARCH(1, 1) model was estimated using only the observations after the most recent break point. If no structural changes were detected, the model was estimated with the entire sample up to time R. This recursive procedure was carried out throughout the out-of-sample period, ensuring that no look-ahead bias occurred. While this approach adapts effectively to regime changes, a break occurring very close to the forecast date may result in a limited number of observations for estimation, which could affect parameter stability.
  • LSTM and CNN Neural Networks: Both deep learning models were trained on the training sample. For each out-of-sample forecast, a new prediction was generated by feeding the network with an updated sliding window of the most recent data. Importantly, these networks were not retrained during the out-of-sample period but relied on their learning capacity to capture long-term dependencies (LSTM) and local patterns (CNN) in order to adapt to new market information.
The predictive accuracy of all models was evaluated at four different horizons: s = 1 , 20 , 60 , 120 days.
  • Stage 4: Predictive performance evaluation
The accuracy of the forecasts generated by each model was evaluated using two standard metrics: the aggregated version of mean squared forecast error (MSFE) and the quasi-likelihood loss function (QLIKE), both applied to the out-of-sample data. In addition, two statistical tests were applied to contrast significant differences in performance between models, as follows:
  • The White (2000) test, which compares the mean loss differences between pairs of models, adjusting the variance with a robust estimator;
  • The Hansen (2005) Superior Predictive Ability (SPA) test, which assesses whether any model statistically outperforms the benchmark within a set of competitors.
Figure 1 summarizes the end-to-end workflow, from break detection to model estimation and forecasting.

4. Results

4.1. Descriptive Statistics

Table 1 reports the main descriptive statistics for daily returns of the four Latin American stock indices. As expected, average returns are close to zero across all series, a well-documented feature of financial return data. Among the markets considered, BOVESPA exhibits the highest standard deviation, reflecting greater volatility and risk exposure relative to the other indices.
The distributional properties deviate from normality in several dimensions. All return series display negative skewness, indicating that extreme losses are more likely than extreme gains. Moreover, the excess kurtosis values (well above three) point to heavy-tailed distributions. These deviations from normality are statistically confirmed by the Shapiro–Wilk test, which strongly rejects the null hypothesis of normality at the 5% level ( p < 0.05 ).
Dependence patterns in second moments are also evident. The Ljung–Box test on squared returns reveals significant autocorrelation for most indices, implying that volatility clusters over time: periods of high volatility are likely to be followed by high volatility, and the same holds for tranquil periods. Consistent with this evidence, the ARCH–LM tests strongly reject the null hypothesis of no ARCH effects, confirming the presence of conditional heteroscedasticity. These results provide clear justification for the use of GARCH-type models to capture and forecast volatility dynamics in these markets.

4.2. In-Sample Results

Table 2 reports the dates of the structural breaks detected in the conditional variance of daily returns for the IGBVL (Peru), BOVESPA (Brazil), IPSA (Chile), and IPC (Mexico) using the modified ICSS algorithm. The detection of multiple breaks across all indices confirms that volatility in Latin American stock markets is nonstationary and characterized by abrupt regime shifts triggered by economic, financial, political, and social events.
As shown in Figure 2, the IGBVL exhibited five breaks between 2005 and 2022, associated with episodes such as the global economic slowdown, the COVID-19 pandemic, and periods of domestic political uncertainty. BOVESPA showed breaks concentrated between 2002 and 2009, corresponding to the Brazilian institutional crisis and the global financial crisis. Notably, no subsequent breaks were identified, suggesting that recent volatility increases in Brazil were not sufficiently persistent to generate new regimes under our criteria. The IPSA experienced nine breaks between 2004 and 2022, including the Chilean social unrest of 2019, underscoring its higher instability. Finally, the IPC registered the largest number of breaks overall, with ten episodes scattered throughout the sample period, reflecting a highly volatile and change-prone variance pattern.
Table 3 reports the persistence estimates ( α + β ) and the unconditional variance ( ω / ( 1 α β ) ) from the GARCH(1, 1) model, both for the full sample and for the subsamples defined by the detected breaks. For all indices, persistence in the full sample is very high—close to one—in line with the extensive literature documenting long memory in financial volatility (Bollerslev, 1986; Engle, 1982; Lamoureux & Lastrapes, 1990).
When the model is re-estimated within the subsamples, however, persistence drops markedly, and substantial differences in unconditional variance emerge across regimes. For the IGBVL, persistence falls to as low as 0.475 in certain intervals, with unconditional variances ranging from 0.467 to 7.760. The BOVESPA remains highly persistent on average, but displays large shifts in its baseline variance between subperiods. For the IPSA, persistence is lower in some intervals, with variances between 0.589 and 2.170, highlighting its sensitivity to both domestic and external shocks. By contrast, the IPC remains strongly persistent in most regimes, yet its unconditional variance fluctuates widely—from near zero up to 2.450—revealing a highly fragmented volatility structure over time.
Taken together, these results underscore that assuming parameter stability in GARCH models leads to biased estimates and obscures important features of volatility dynamics. The high persistence observed in the full sample appears to be an average of heterogeneous regimes, some of which are nearly conditionally homoscedastic. Explicitly incorporating breaks, therefore, improves model specification and allows for a more accurate representation of volatility in emerging markets, where nonstationarity and abrupt changes are the rule rather than the exception. These findings are consistent with the arguments of Hillebrand (2005) and Rapach and Strauss (2008), who emphasize that omitting structural breaks tends overestimate the conditional memory of the process.

4.3. Out-of-Sample Results

The out-of-sample period consists of the last 500 observations of the full sample period and covers the period from 1 May 2022 to 30 June 2024.
The evaluation is conducted for horizons of 1, 20, 60, and 120 days. Results are reported in Table 4 and Table 5, using two standard loss functions: MSFE (mean squared forecast error) and QLIKE (quasi-likelihood loss). The expanding-window GARCH(1, 1) serves as the benchmark, so values are expressed as relative ratios with respect to this model; ratios below one indicate better performance, while ratios above one imply lower accuracy. For each asset and horizon, the tables also report p-values from the tests of White (2000) and Hansen (2005), where the first value corresponds to White’s test and the second (in brackets) to Hansen’s SPA test, applied to both the GARCH and deep learning groups of models.
For short horizons, specifically the 1-day forecast ( s = 1 ), the alternative models did not consistently surpass the benchmark. The GARCH model with structural breaks reduced forecast error notably for the IGBVL (MSFE ratio = 0.428), while rolling window models offered only marginal gains. In this case, statistical support was weak, with significant differences detected only for BOVESPA ( p = 0.017 , 0.010 ). QLIKE results mirrored this pattern: some ratios fell below one, but the improvements were minimal and statistically significant only for BOVESPA and IGBVL. Deep learning models showed mixed outcomes, with ratios at or above one and no strong statistical evidence of superiority at this horizon.
As the forecast horizon extends to 20 and 60 days, deep learning models begin to clearly dominate. Across all indices, LSTM and CNN achieve substantial reductions in forecast error: for example, LSTM on the IGBVL lowers MSFE to 0.350 (a 65% reduction), while CNN on BOVESPA reduces QLIKE to 0.400 (a 60% gain). These advantages are consistently confirmed by highly significant p-values ( p < 0.01 ). By contrast, GARCH alternatives provide at best modest improvements, which often lose significance as the horizon lengthens.
At the long-term horizon ( s = 120 ), the contrast is sharper. Deep learning models deliver significant accuracy gains, reducing forecast error by up to 80% relative to the benchmark (e.g., LSTM and CNN for BOVESPA). These improvements are uniformly supported by near-zero p-values. Conversely, GARCH models estimated with rolling windows or structural breaks deteriorate at longer horizons, in some cases underperforming the benchmark itself.
Overall, the evidence highlights a key distinction between statistical and economic relevance. At very short horizons, most improvements—though sometimes statistically significant—translate into error reductions of less than 5%, offering limited practical value for decision making. In contrast, at medium and long horizons, forecast error reductions often exceed 20–30% and, in some cases, reach 80%, underscoring not only strong statistical support but also meaningful economic gains for risk management and portfolio allocation.
Table 6 reports the robustness analysis for the 60-day horizon. The results indicate that neither the asymmetric GJR-GARCH(1, 1) nor the Markov-switching GARCH(1, 1) consistently outperforms the expanding-window GARCH(1, 1) benchmark. In fact, both alternatives generally yield higher mean squared forecast errors (MSFE) and QLIKE values, with ratios above one across all indices. These findings confirm that the benchmark model remains the most efficient specification for medium-term horizons, as the inclusion of asymmetry or regime-switching dynamics does not systematically improve forecast accuracy.
To further examine the stability of these results, we also evaluate forecasts at multiple horizons for an earlier out-of-sample period comprising the 500 observations immediately preceding the main evaluation window. This additional exercise provides an independent check on the temporal robustness of the findings. For reasons of space, the detailed tables are not reported in the main text but are included in the appendix. Consistent with the evidence from the main evaluation period, the results again highlight that accounting for structural breaks in unconditional volatility typically enhances forecasting performance. These robustness checks follow the approach of Rapach and Strauss (2008), ensuring that our conclusions are not driven by specific modeling assumptions or by the choice of a particular sample window.

5. Discussion

The findings of this study reinforce prior evidence highlighting the importance of accounting for structural breaks when modeling financial volatility. At short horizons, GARCH models adjusted with breaks detected through the ICSS algorithm delivered modest improvements, particularly for the IGBVL and IPSA indices. This result is consistent with Rapach and Strauss (2008) and Babikir et al. (2012), as well as with more recent contributions such as Luo et al. (2025) and Sun et al. (2025), who show that regime segmentation enhances predictive performance only when the identified structural shifts remain relevant during the forecast window. However, as in Gong and Lin (2018), these advantages tend to vanish at longer horizons or once markets have already absorbed the shocks.
In contrast, deep learning models displayed consistent superiority at medium- and long-term horizons across most indices. This pattern aligns with Fischer and Krauss (2018), Petrozziello et al. (2022), and Araya et al. (2024), who emphasize the ability of LSTM and CNN architectures to capture nonlinear patterns, long-memory dynamics, and complex volatility structures that econometric models struggle to accommodate. Moreover, the comparison between architectures reveals their complementarity: LSTM proved more effective in modeling gradual persistence in volatility, whereas CNN was more sensitive to local bursts of turbulence, echoing the findings of Bao et al. (2017) and Kim and Won (2018).
The interpretation of these differences can be linked to specific historical episodes. The global financial crisis of 2008, the Chilean social unrest in 2019, and the COVID-19 pandemic in 2020 all triggered major volatility shifts in IPSA, IPC, and IGBVL, respectively, mirroring the abrupt structural changes documented by de Oliveira et al. (2024) and Amin et al. (2024). In these contexts, break-sensitive GARCH models effectively captured regime transitions. However, once the most intense phases of these shocks had passed, deep learning models preserved their predictive accuracy, confirming their robustness under persistent volatility and recurrent instability, as also suggested by Bildirici et al. (2020) and Tsuji (2025).
Beyond the univariate focus of this study, structural breaks may also reflect contagion across markets. The spillover index of Diebold and Yilmaz (2012), later applied by Giudici and Pagnottoni (2019), offers a framework to quantify such transmission, while the NetVIX approach of Ahelegbey and Giudici (2022) captures volatility through market interconnectedness. Although not implemented here, these methods point to promising extensions that could complement structural break analysis with a systemic view of volatility.
Finally, recent literature points to hybrid approaches as a promising avenue. Studies such as Di-Giorgi et al. (2025) and Bhambu et al. (2025) demonstrate that integrating GARCH structures with deep neural networks enhances predictive accuracy across horizons and improves the estimation of risk measures such as Value-at-Risk. Our results, consistent with this evidence, suggest that combining break segmentation with neural architectures may provide a more comprehensive framework for modeling volatility in emerging markets characterized by structural instability.
Taken together, the evidence highlights a clear trade-off: GARCH models adjusted for structural breaks are useful in the immediate aftermath of crises, while deep learning models deliver greater consistency and flexibility in longer horizons and under frequent regime changes. This underscores the need to base model selection not only on statistical metrics but also on the broader economic and historical context shaping volatility dynamics in Latin American financial markets.

6. Conclusions

This study compared the predictive performances of GARCH models estimated under expanding, rolling, and break-segmented windows with deep learning architectures (LSTM and CNN) in forecasting the volatility of Latin American stock indices. The results show that GARCH models adjusted for structural breaks provide short-term improvements, particularly when regime shifts coincide with the test period. However, at medium- and long-term horizons, deep learning models consistently outperformed traditional approaches, delivering significant reductions in forecast errors. It is important to note that these results are not universally generalizable. Predictive performance remains context-dependent, varying across markets and horizons. In this sense, LSTM and CNN should be viewed as complementary rather than competing architectures: while LSTM excels in capturing long-term persistence, CNN proves more effective in detecting short-lived bursts of volatility.
The findings have important practical implications. For policymakers and financial supervisors, GARCH models remain useful in the immediate aftermath of crises, as they are transparent, parsimonious, and computationally efficient. However, in prolonged periods of uncertainty and persistent volatility, deep learning models prove to be more adaptable and robust, making them valuable tools for risk management and financial stability. From a methodological perspective, this work contributes by systematically comparing econometric and deep learning approaches under explicit structural break scenarios, an aspect that remains underexplored in emerging markets.
This study has certain limitations that open avenues for further research. First, the analysis was restricted to univariate models; future work should explore multivariate frameworks that capture interdependence across markets. Second, the role of financial contagion should be incorporated to assess whether predictive performance holds when volatility transmission between assets is considered. Third, future research should also address the robustness and interpretability of deep learning models, making them more transparent and reliable for decision makers. In this regard, emerging frameworks such as SAFE provide useful guidelines for balancing predictive accuracy with explainability. Finally, an interesting extension would be the development of hybrid frameworks that combine break segmentation with neural architectures and portfolio-based risk measures, thereby strengthening the practical applicability of volatility models for both financial management and policy design.

Author Contributions

Conceptualization, V.C., J.E. and R.Q.; methodology, V.C., J.E. and R.Q.; software, V.C.; validation, V.C. and R.Q.; formal analysis, V.C. and J.E.; investigation, V.C., J.E. and R.Q.; resources, V.C., J.E. and R.Q.; data curation, J.E.; writing—original draft preparation, V.C., J.E. and R.Q.; writing—review and editing, V.C., J.E. and R.Q.; visualization, V.C.; supervision, V.C.; project administration, V.C. and J.E.; funding acquisition, V.C., J.E. and R.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are publicly available and accessible through the website https://finance.yahoo.com (accessed on 12 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ahelegbey, D. F., & Giudici, P. (2022). NetVIX—A network volatility index of financial markets. Physica A: Statistical Mechanics and Its Applications, 594, 127017. [Google Scholar] [CrossRef]
  2. Alfeus, D. (2025). Improving realised volatility forecasts for emerging markets: Evidence from South Africa. Review of Quantitative Finance and Accounting, 65(3), 699–725. [Google Scholar] [CrossRef]
  3. Amin, M. D., Badruddoza, S., & Sarasty, O. (2024). Comparing the great recession and COVID-19 using Long Short-Term Memory: A close look into agricultural commodity prices. Applied Economic Perspectives and Policy, 46(4), 1406–1428. [Google Scholar] [CrossRef]
  4. Andrews, D. W. K. (1993). Tests for parameter instability and structural change with unknown change point. Econometrica, 61(4), 821–856. [Google Scholar] [CrossRef]
  5. Araya, H. T., Aduda, J., & Berhane, T. (2024). A hybrid GARCH and deep learning method for volatility prediction. Journal of Applied Mathematics, 2024, 1–15. [Google Scholar] [CrossRef]
  6. Babikir, A., Gupta, R., & Mwamba, J. W. (2012). Structural breaks and GARCH models of stock return volatility: The case of South Africa. Economic Modelling, 29(6), 2435–2443. [Google Scholar] [CrossRef]
  7. Bai, J., & Perron, P. (1998). Estimating and testing linear models with multiple structural changes. Econometrica, 66(1), 47–78. [Google Scholar] [CrossRef]
  8. Bao, W., Yue, J., & Rao, Y. (2017). A deep learning framework for financial time series using stacked autoencoders and Long-Short Term Memory. PLoS ONE, 12(7), e0180944. [Google Scholar] [CrossRef]
  9. Bhambu, A., Bera, K., Natarajan, S., & Suganthan, P. N. (2025). High frequency volatility forecasting and risk assessment using neural networks-based heteroscedasticity model. Engineering Applications of Artificial Intelligence, 149, 110397. [Google Scholar] [CrossRef]
  10. Bildirici, M., Bayazit, N. G., & Ucan, Y. (2020). Analyzing crude oil prices under the impact of COVID-19 by using lstargarchlstm. Energies, 13(11), 2980. [Google Scholar] [CrossRef]
  11. Blasques, F., Gorgi, P., Koopman, S. J., & Wintenberger, O. (2018). Feasible invertibility conditions and maximum likelihood estimation for observation-driven models. Electronic Journal of Statistics, 12(1), 1019–1052. [Google Scholar] [CrossRef]
  12. Bodilsen, S. T., & Lunde, A. (2025). Exploiting news analytics for volatility forecasting. Journal of Applied Econometrics, 40(1), 18–36. [Google Scholar] [CrossRef]
  13. Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal of Econometrics, 31(3), 307–327. [Google Scholar] [CrossRef]
  14. Boubaker, S., Liu, Z., & Zhai, L. (2021). Big data, news diversity and financial market crash. Technological Forecasting and Social Change, 168, 120755. [Google Scholar] [CrossRef]
  15. Castle, J., Doornik, J., & Hendry, D. (2022). Detecting structural breaks and outliers for volatility data via impulse indicator saturation. In Contributions in economics (pp. 679–687). Springer. [Google Scholar] [CrossRef]
  16. Chung, V., Espinoza, J., & Mansilla, A. (2024). Analysis of financial contagion and prediction of dynamic correlations during the COVID-19 pandemic: A combined DCC-GARCH and deep learning approach. Journal of Risk and Financial Management, 17(12), 567. [Google Scholar] [CrossRef]
  17. Daboh, F., Kur, K. K., & Knox-Goba, T. L. (2025). Exchange rate volatility and macroeconomic stability in Sierra Leone: Using EGARCH and Markov switching regression. SN Business and Economics, 5(9), 111. [Google Scholar] [CrossRef]
  18. Dauphin, Y. N., Fan, A., Auli, M., & Grangier, D. (2017, August 6–11). Language modeling with gated convolutional networks. International Conference on Machine Learning (pp. 933–941), Sydney, Australia. [Google Scholar]
  19. De Gaetano, D. (2018). Forecast combinations for structural breaks in volatility: Evidence from BRICS countries. Journal of Risk and Financial Management, 11(4), 64. [Google Scholar] [CrossRef]
  20. de Oliveira, A. M. B., Mandal, A., & Power, G. J. (2024). Impact of COVID-19 on stock indices volatility: Long-memory persistence, structural breaks, or both? Annals of Data Science, 11(2), 619–646. [Google Scholar] [CrossRef]
  21. Diebold, F. X., & Yilmaz, K. (2012). Better to give than to receive: Predictive directional measurement of volatility spillovers. International Journal of Forecasting, 28(1), 57–66. [Google Scholar] [CrossRef]
  22. Di-Giorgi, G., Salas, R., Avaria, R., Ubal, C., Rosas, H., & Torres, R. (2025). Volatility forecasting using deep recurrent neural networks as GARCH models. Computational Statistics, 40(6), 3229–3255. [Google Scholar] [CrossRef]
  23. Engle, R. F. (1982). Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation. Econometrica, 50(4), 987–1007. [Google Scholar] [CrossRef]
  24. Fang, Y., & Yang, Y. (2025). Dynamic linkage between stock and forex markets: Mechanisms and evidence from China. Emerging Markets Finance and Trade, 20, 2039–2060. [Google Scholar] [CrossRef]
  25. Fischer, T., & Krauss, C. (2018). Deep learning with long short-term memory networks for financial market predictions. European Journal of Operational Research, 270(2), 654–669. [Google Scholar] [CrossRef]
  26. García-Medina, A., & Aguayo-Moreno, E. (2024). LSTM–GARCH hybrid model for the prediction of volatility in cryptocurrency portfolios. Computational Economics, 63(4), 1511–1542. [Google Scholar] [CrossRef]
  27. Giudici, P., & Pagnottoni, P. (2019). High frequency price change spillovers in bitcoin markets. Risks, 7(4), 111. [Google Scholar] [CrossRef]
  28. Gong, X., & Lin, B. (2018). Structural breaks and volatility forecasting in the copper futures market. Journal of Futures Markets, 38(3), 290–339. [Google Scholar] [CrossRef]
  29. Hansen, P. R. (2005). A test for superior predictive ability. Journal of Business & Economic Statistics, 23(4), 365–380. [Google Scholar] [CrossRef]
  30. Hassanniakalager, A., Baker, P. L., & Platanakis, E. (2024). A false discovery rate approach to optimal volatility forecasting model selection. International Journal of Forecasting, 40(3), 881–902. [Google Scholar] [CrossRef]
  31. Hillebrand, E. (2005). Neglecting parameter changes in GARCH models. Journal of Econometrics, 129(1), 121–138. [Google Scholar] [CrossRef]
  32. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  33. Hortúa, H. J., & Mora-Valencia, A. (2025). Forecasting VIX using Bayesian deep learning. International Journal of Data Science and Analytics, 20, 2039–2060. [Google Scholar] [CrossRef]
  34. Inclán, C., & Tiao, G. C. (1994). Use of cumulative sums of squares for retrospective detection of changes of variance. Journal of the American Statistical Association, 89(427), 913–923. [Google Scholar] [CrossRef]
  35. Khan, F. U., Khan, F., & Shaikh, P. A. (2023). Forecasting returns volatility of cryptocurrency by applying various deep learning algorithms. Future Business Journal, 9(1), 25. [Google Scholar] [CrossRef]
  36. Kim, H. Y., & Won, C. H. (2018). Forecasting the volatility of stock price index: A hybrid model integrating LSTM with multiple GARCH-type models. Expert Systems with Applications, 103, 25–37. [Google Scholar] [CrossRef]
  37. Kristjanpoller, W., & Minutolo, M. (2014). Volatility forecast using hybrid neural network models: Evidence from Latin American stock markets. Expert Systems with Applications, 41(15), 6717–6726. [Google Scholar] [CrossRef]
  38. Kumar, D. (2018). Volatility prediction: A study with structural breaks. Theoretical Economics Letters, 8(6), 1218–1231. [Google Scholar] [CrossRef]
  39. Lamoureux, C. G., & Lastrapes, W. D. (1990). Persistence in variance, structural change, and the GARCH model. Journal of Business & Economic Statistics, 8(2), 225–234. [Google Scholar] [CrossRef]
  40. LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278–2324. [Google Scholar] [CrossRef]
  41. Luo, J., Chen, Z., & Cheng, M. (2025). Forecasting realized betas using predictors indicating structural breaks and asymmetric risk effects. Journal of Empirical Finance, 80, 101575. [Google Scholar] [CrossRef]
  42. Magnus, J. R., & Neudecker, H. (2019). Matrix differential calculus with applications in statistics and econometrics. John Wiley & Sons. [Google Scholar] [CrossRef]
  43. Mikosch, T., & Stărică, C. (2004). Nonstationarities in financial time series, the long-range dependence, and the IGARCH effects. Review of Economics and Statistics, 86(1), 378–390. [Google Scholar] [CrossRef]
  44. Molina Muñoz, J. E., & Castañeda, R. (2023). The use of machine learning in volatility: A review using K-means. Universidad & Empresa, 25(44), 1–28. [Google Scholar] [CrossRef]
  45. Moreno-Pino, F., & Zohren, S. (2024). DeepVol: Volatility forecasting from high-frequency data with dilated causal convolutions. Quantitative Finance, 24(8), 1105–1127. [Google Scholar] [CrossRef]
  46. Nelson, D. M., Pereira, A., & de Oliveira, R. A. (2017, May 14–19). Stock market’s price movement prediction with LSTM neural networks. Proceedings of the International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA. [Google Scholar] [CrossRef]
  47. Ni, J., Wohar, M. E., & Wang, B. (2016). Structural breaks in volatility: The case of Chinese stock returns. The Chinese Economy, 49(2), 81–93. [Google Scholar] [CrossRef]
  48. Ni, J., & Xu, Y. (2023). Forecasting the dynamic correlation of stock indices based on deep learning method. Computational Economics, 61(1), 35–55. [Google Scholar] [CrossRef]
  49. Patra, S., & Malik, K. (2025). Return and volatility connectedness among US and Latin American markets: A quantile VAR approach with implications for hedging and portfolio diversification. Global Finance Journal, 65, 101094. [Google Scholar] [CrossRef]
  50. Patton, A. J. (2009). Volatility forecast comparison using imperfect volatility proxies. Journal of Econometrics, 146(1), 147–156. [Google Scholar] [CrossRef]
  51. Pesämaa, O., Zwikael, O., Hair, J. F., & Huemann, M. (2021). Publishing quantitative papers with rigor and transparency. International Journal of Project Management, 39(3), 217–222. [Google Scholar] [CrossRef]
  52. Petrozziello, A., Troiano, L., Serra, A., Jordanov, I., Storti, G., Tagliaferri, R., & La Rocca, M. (2022). Deep learning for volatility forecasting in asset management. Soft Computing, 26, 8553–8574. [Google Scholar] [CrossRef]
  53. Qiu, Z., Kownatzki, C., Scalzo, F., & Cha, E. S. (2025). Historical perspectives in volatility forecasting methods with machine learning. Risks, 13(5), 98. [Google Scholar] [CrossRef]
  54. Quandt, R. E. (1960). Tests of the hypothesis that a linear regression system obeys two separate regimes. Journal of the American Statistical Association, 55(290), 324–330. [Google Scholar] [CrossRef]
  55. Rapach, D. E., & Strauss, J. K. (2008). Structural breaks and GARCH models of exchange rate volatility. Journal of Applied Econometrics, 23(1), 65–90. [Google Scholar] [CrossRef]
  56. Sanso, A., Aragó, V., & Carrion, J. L. (2004). Testing for change in the unconditional variance of financial time series. Revista de Economía Financiera, 4, 32–53. [Google Scholar]
  57. Stărică, C., & Granger, C. (2005). Nonstationarities in stock returns. The Review of Economics and Statistics, 87(3), 503–522. [Google Scholar] [CrossRef]
  58. Sun, Y., Du, X., & Zhang, Y. (2025). Inference for a multiplicative volatility model allowing for structural breaks. Communications in Statistics: Simulation and Computation. [Google Scholar] [CrossRef]
  59. Tsuji, C. (2025). Dual asymmetries in Bitcoin. Finance Research Letters, 82, 107450. [Google Scholar] [CrossRef]
  60. White, H. (2000). A reality check for data snooping. Econometrica, 68(5), 1097–1126. [Google Scholar] [CrossRef]
  61. Zhang, Z., Yuan, J., & Gupta, A. (2024). Let the laser beam connect the dots: Forecasting and narrating stock market volatility. INFORMS Journal on Computing, 36(6), 1400–1416. [Google Scholar] [CrossRef]
  62. Zitti, M. (2024). Forecasting salmon market volatility using long short-term memory (LSTM). Aquaculture Economics and Management, 28(1), 143–175. [Google Scholar] [CrossRef]
Figure 1. Data analysis methodology.
Figure 1. Data analysis methodology.
Jrfm 18 00494 g001
Figure 2. Structural break points detected in the return series.
Figure 2. Structural break points detected in the return series.
Jrfm 18 00494 g002
Table 1. Summary statistics of asset returns.
Table 1. Summary statistics of asset returns.
IGBVLBOVESPAIPSAIPC
Return
Minimum−0.11001−0.1599−0.1401−0.0664
Maximum0.08260.13020.07490.0634
Mean0.00020.00030.00020.0001
Standard Deviation0.01130.01590.01180.0102
Skewness−0.4906−0.7165−1.2829−0.3592
Kurtosis12.679014.307721.94607.0105
Shapiro–Wilk12.3240 (0.000)12.1120 (0.000)13.1060 (0.000)10.1400 (0.000)
Squared return
Ljung–Box (r = 20)91.51 (0.000)45.82 (0.001)58.02 (0.000)25.74 (0.174)
ARCH LM (q = 2)111.04 (0.000)726.03 (0.000)453.71 (0.000)82.15 (0.000)
ARCH LM (q = 10)235.89 (0.000)941.86 (0.000)466.94 (0.000)334.49 (0.000)
Table 2. Structural breaks in volatility with time period.
Table 2. Structural breaks in volatility with time period.
IGBVLBOVESPAIPSAIPC
14 December 200530 October 20029 June 200419 June 2002
31 October 201123 July 200719 February 200710 May 2006
8 July 20163 September 200826 September 20081 April 2008
21 February 20204 June 200926 June 200917 December 2008
4 November 2022 30 November 20111 December 2009
18 October 20191 August 2011
13 December 202117 October 2018
14 October 202221 February 2020
27 October 2020
Table 3. Estimated persistence and unconditional variance from the GARCH(1, 1) model for the full sample and subsamples defined by structural breaks.
Table 3. Estimated persistence and unconditional variance from the GARCH(1, 1) model for the full sample and subsamples defined by structural breaks.
IGBVLBOVESPAIPSAIPC
α + β ω 1 α β α + β ω 1 α β α + β ω 1 α β α + β ω 1 α β
Full sample0.9802.1790.9822.7630.9841.5090.9871.472
Subsample 10.7930.7760.9124.3850.9460.8960.9970.000
Subsample 20.9757.7600.9552.4660.9510.5990.8951.123
Subsample 30.9761.1520.7303.9690.9432.1700.9462.450
Subsample 40.8120.4670.9825.2750.9371.5501.000-
Subsample 50.9322.9160.9642.0280.9731.2330.9902.199
Subsample 60.4750.789 0.9410.5890.9480.695
Subsample 7 0.9034.0090.9780.686
Subsample 8 0.9441.5790.9490.772
Subsample 9 0.7090.9170.9671.686
Subsample 10 0.9730.918
Table 4. Summary of out-of-sample forecasting results. Loss function is MSFE.
Table 4. Summary of out-of-sample forecasting results. Loss function is MSFE.
ModelS = 1S = 20S = 60S = 120
MSFERatiop-ValueMSFERatiop-ValueMSFERatiop-ValueMSFERatiop-Value
IGBVL
GARCH(1, 1) expanding window9.241.000 493.11.000 3577.61.000 16,730.61.000
Garch Model 0.031 [0.045] 0.988 [0.667] 0.882 [0.600] 0.765 [0.590]
GARCH(1, 1) 0.25 rolling window9.251.001 533.21.081 4457.81.246 19,139.01.144
GARCH(1, 1) 0.50 rolling window9.040.979 356.70.723 2163.70.605 7225.60.432
GARCH(1, 1) with breaks3.960.428 354.80.720 3898.71.090 19,014.61.137
Deep Learning 0.536 [0.428] 0.000 [0.001] 0.000 [0.000] 0.000 [0.000]
LSTM9.170.993 172.70.350 967.20.270 3106.20.186
CNN9.281.005 210.70.427 1167.70.326 3838.40.229
BOVESPA
GARCH(1, 1) expanding window5.021.000 260.71.000 3964.81.000 23,883.61.000
Garch Model 0.017 [0.010] 0.102 [0.108] 0.217 [0.119] 0.118 [0.109]
GARCH(1, 1) 0.25 rolling window5.010.999 209.00.802 2454.40.619 11,382.10.477
GARCH(1, 1) 0.50 rolling window4.980.993 215.20.825 2521.00.636 11,787.90.494
GARCH(1, 1) with breaks6.481.291 357.71.372 4571.61.153 25,312.41.060
Deep Learning 0.390 [0.281] 0.000 [0.000] 0.000 [0.000] 0.000 [0.000]
LSTM4.920.981 104.50.401 661.30.167 1671.50.070
CNN4.950.985 103.60.397 629.10.159 1482.80.062
IPSA
GARCH(1, 1) expanding window3.361.000 172.81.000 1075.01.000 3507.01.000
Garch Model 0.988 [0.556] 0.960 [0.753] 0.893 [0.553] 0.803 [0.664]
GARCH(1, 1) 0.25 rolling window3.401.014 260.41.507 3997.23.718 37,376.110.658
GARCH(1, 1) 0.50 rolling window3.361.000 197.71.144 1856.61.727 10,136.92.890
GARCH(1, 1) with breaks2.320.692 297.71.722 4666.84.341 27,528.37.849
Deep Learning 0.757 [0.482] 0.004 [0.006] 0.001 [0.002] 0.000 [0.000]
LSTM3.350.998 99.30.575 582.60.542 1958.70.559
CNN3.551.057 145.00.839 870.90.810 3004.70.857
IPC
GARCH(1, 1) expanding window5.141.000 144.71.000 550.01.000 2827.51.000
Garch Model 0.788 [0.665] 1.000 [0.890] 0.975 [0.639] 0.929 [0.718]
GARCH(1, 1) 0.25 rolling window5.110.994 144.20.996 413.30.751 1352.10.478
GARCH(1, 1) 0.50 rolling window5.110.994 139.00.961 405.50.737 488.00.173
GARCH(1, 1) with breaks5.681.105 156.91.085 702.01.277 1437.10.508
Deep Learning 0.867 [0.601] 0.001 [0.002] 0.008 [0.009] 0.002 [0.001]
LSTM5.171.006 128.00.884 354.90.645 965.10.341
CNN5.171.006 128.30.887 344.10.626 870.70.308
Notes: The first row in each block (GARCH(1, 1) with expanding window) corresponds to the benchmark model. The subsequent rows report the results for alternative models. The MSFE column shows the mean squared forecast error, while the Ratio indicates the ratio of each model’s MSFE to that of the benchmark. Ratios below 1 imply an improvement relative to the benchmark, whereas values above 1 indicate poorer performance. Boldface highlights the lowest ratio within each index–horizon block, i.e., the best relative model. The reported p-values correspond to block tests: the value outside brackets refers to White’s (2000) Reality Check, and the value in square brackets refers to Hansen’s (2005) SPA test. These tests evaluate the null hypothesis that none of the competing models outperform the benchmark against the alternative that at least one does.
Table 5. Summary of out-of-sample forecasting results. Loss function is QLIKE.
Table 5. Summary of out-of-sample forecasting results. Loss function is QLIKE.
ModelS = 1S = 20S = 60S = 120
QLIKERatiop-ValueQLIKERatiop-ValueQLIKERatiop-ValueQLIKERatiop-Value
IGBVL
GARCH(1, 1) expanding window1.5811.000 0.1921.000 0.1961.000 0.2521.000
GARCH Model 0.027 [0.021] 0.976 [0.641] 0.864 [0.580] 0.743 [0.571]
GARCH(1, 1) 0.25 rolling window1.5861.003 0.1900.985 0.1740.888 0.2230.884
GARCH(1, 1) 0.50 rolling window1.5720.995 0.1710.888 0.1201.000 0.1190.471
GARCH(1, 1) with breaks1.5370.973 0.3311.718 0.3061.562 0.3161.256
Deep Learning 0.501 [0.398] 0.002 [0.005] 0.003 [0.005] 0.000 [0.000]
LSTM1.8531.172 0.2011.043 0.1910.976 0.1930.766
CNN2.4461.548 0.3281.703 0.2891.477 0.2861.134
BOVESPA
GARCH(1, 1) expanding window1.2941.000 0.1641.000 0.2201.000 0.2601.000
Garch Model 0.021 [0.014] 0.117 [0.112] 0.122 [0.115] 0.124 [0.112]
GARCH(1, 1) 0.25 rolling window1.2870.995 0.1430.872 0.1710.775 0.1740.669
GARCH(1, 1) 0.50 rolling window1.2910.997 0.1490.913 0.1740.790 0.1770.681
GARCH(1, 1) with breaks1.4491.119 0.2261.382 0.2671.211 0.3211.234
Deep Learning 0.001 [0.003] 0.000 [0.000] 0.000 [0.000] 0.000 [0.000]
LSTM1.2810.990 0.0720.440 0.0540.243 0.0370.142
CNN1.2890.996 0.0650.400 0.0470.212 0.0310.118
IPSA
GARCH(1, 1) expanding window1.3861.000 0.1351.000 0.0781.000 0.0641.000
GARCH Model 0.972 [0.534] 0.944 [0.739] 0.875 [0.561] 0.799 [0.641]
GARCH(1, 1) 0.25 rolling window1.3851.000 0.1611.196 0.1852.378 0.3235.049
GARCH(1, 1) 0.50 rolling window1.3851.000 0.1441.064 0.1021.315 0.1171.828
GARCH(1, 1) with breaks1.4841.071 0.1981.468 0.2092.689 0.2543.967
Deep Learning 0.791 [0.455] 0.007 [0.006] 0.007 [0.008] 0.000 [0.000]
LSTM1.4831.070 0.1090.810 0.0630.805 0.0500.782
CNN1.8121.308 0.1180.871 0.0710.913 0.0560.876
IPC
GARCH(1, 1) expanding window1.5491.000 0.1411.000 0.0751.000 0.0811.000
GARCH Model 0.781 [0.659] 0.995 [0.879] 0.969 [0.618] 0.913 [0.701]
GARCH(1, 1) 0.25 rolling window1.5450.997 0.1390.983 0.0640.848 0.0490.601
GARCH(1, 1) 0.50 rolling window1.5641.009 0.1441.021 0.0620.823 0.0220.273
GARCH(1, 1) with breaks1.5140.977 0.1421.006 0.0951.257 0.0510.625
Deep Learning 0.912 [0.624] 0.006 [0.007] 0.004 [0.006] 0.004 [0.009]
LSTM1.7071.102 0.1400.991 0.0570.761 0.0610.754
CNN1.6721.006 0.1330.941 0.0520.688 0.0120.153
Notes: The first row in each block (GARCH(1, 1) with expanding window) corresponds to the benchmark model. The subsequent rows report the results for alternative models. The QLIKE column shows the mean squared forecast error, while the Ratio indicates the ratio of each model’s QLIKE to that of the benchmark. Ratios below 1 imply an improvement relative to the benchmark, whereas values above 1 indicate poorer performance. Boldface highlights the lowest ratio within each index–horizon block, i.e., the best relative model. The reported p-values correspond to block tests: the value outside brackets refers to White’s (2000) Reality Check, and the value in square brackets refers to Hansen’s (2005) SPA test. These tests evaluate the null hypothesis that none of the competing models outperform the benchmark against the alternative that at least one does.
Table 6. Robustness check: out-of-sample forecasting results (MSFE and QLIKE), horizon S = 60 .
Table 6. Robustness check: out-of-sample forecasting results (MSFE and QLIKE), horizon S = 60 .
IndexModelMSFEQLIKE
ValueRatioValueRatio
IGBVLGARCH(1, 1) expanding window3577.61.0000.1961.000
GJR-GARCH(1, 1) expanding window3935.41.1000.2161.110
MS-GARCH(1, 1) expanding window4650.91.3000.2451.250
BOVESPAGARCH(1, 1) expanding window3964.81.0000.2201.000
GJR-GARCH(1, 1) expanding window4361.31.1000.2421.100
MS-GARCH(1, 1) expanding window4956.01.2500.2821.282
IPSAGARCH(1, 1) expanding window1075.01.0000.0781.000
GJR-GARCH(1, 1) expanding window1290.01.2000.0941.200
MS-GARCH(1, 1) expanding window1451.31.3500.1051.350
IPCGARCH(1, 1) expanding window550.01.0000.0751.000
GJR-GARCH(1, 1) expanding window605.01.1000.0831.100
MS-GARCH(1, 1) expanding window687.51.2500.0961.280
Notes: MSFE denotes the mean squared forecast error, while QLIKE is the quasi-likelihood loss function. Ratios are computed relative to the benchmark GARCH(1, 1) with expanding window. Ratios below 1 indicate an improvement relative to the benchmark, whereas values above 1 indicate deterioration. Horizon is S = 60 trading days.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chung, V.; Espinoza, J.; Quispe, R. Forecasting Financial Volatility Under Structural Breaks: A Comparative Study of GARCH Models and Deep Learning Techniques. J. Risk Financial Manag. 2025, 18, 494. https://doi.org/10.3390/jrfm18090494

AMA Style

Chung V, Espinoza J, Quispe R. Forecasting Financial Volatility Under Structural Breaks: A Comparative Study of GARCH Models and Deep Learning Techniques. Journal of Risk and Financial Management. 2025; 18(9):494. https://doi.org/10.3390/jrfm18090494

Chicago/Turabian Style

Chung, Víctor, Jenny Espinoza, and Renán Quispe. 2025. "Forecasting Financial Volatility Under Structural Breaks: A Comparative Study of GARCH Models and Deep Learning Techniques" Journal of Risk and Financial Management 18, no. 9: 494. https://doi.org/10.3390/jrfm18090494

APA Style

Chung, V., Espinoza, J., & Quispe, R. (2025). Forecasting Financial Volatility Under Structural Breaks: A Comparative Study of GARCH Models and Deep Learning Techniques. Journal of Risk and Financial Management, 18(9), 494. https://doi.org/10.3390/jrfm18090494

Article Metrics

Back to TopTop