Next Article in Journal
Prime Strictly Concentric Magic Squares of Odd Order
Previous Article in Journal
Ninth-Order Two-Step Methods with Varying Step Lengths
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transformer-Based Downside Risk Forecasting: A Data-Driven Approach with Realized Downward Semi-Variance

1
School of Finance and Business, Shanghai Normal University, Shanghai 200234, China
2
School of Mathematics, Shandong University, Jinan 250100, China
3
Securities Institute for Financial Studies, Shandong University, Jinan 250100, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(8), 1260; https://doi.org/10.3390/math13081260
Submission received: 12 March 2025 / Revised: 6 April 2025 / Accepted: 9 April 2025 / Published: 11 April 2025

Abstract

:
Realized downward semi-variance (RDS) has been realized as a key indicator to measure the downside risk of asset prices, and the accurate prediction of RDS can effectively guide traders’ investment behavior and avoid the impact of market fluctuations caused by price declines. In this paper, the RDS rolling prediction performance of the traditional econometric model, machine learning model, and deep learning model is discussed in combination with various relevant influencing factors, and the sensitivity analysis is further carried out with the rolling window length, prediction length, and a variety of evaluation methods. In addition, due to the characteristics of RDS, such as aggregation and jumping, this paper further discusses the robustness of the model under the impact of external events, the influence of emotional factors on the prediction accuracy of the model, and the results and analysis of the hybrid model. The empirical results show that (1) when the rolling window is set to 20, the overall prediction effect of the model in this paper is the best. Taking the Transformer model under SSE as an example, compared with the prediction results under the rolling window length of 5, 10, and 30, the RMSE improvement ratio reaches 24.69%, 15.90%, and 43.60%, respectively. (2) The multivariable Transformer model shows a better forecasting effect. Compared with traditional econometric, machine learning, and deep learning models, the average increase percentage of RMSE, MAE, MAPE, SMAPE, MBE, and SD indicators is 52.23%, 20.03%, 62.33%, 60.33%, 37.57%, and 18.70%, respectively. (3) In multi-step prediction scenarios, the DM test statistic of the Transformer model is significantly positive, and the prediction accuracy of the Transformer model remains stable as the number of prediction steps increases. (4) Under the impact of external events of COVID-19, the Transformer model has stability, and the addition of emotional factors can effectively improve the prediction accuracy. In addition, the model’s prediction performance and generalization ability can be further improved by stacked prediction models. An in-depth study of RDS forecasting is of great value to capture the characteristics of downside risks, enrich the financial risk measurement system, and better evaluate potential losses.

1. Introduction

Research on financial market risk plays an important role in understanding the operating mechanism of the financial market, improving the financial theoretical system, evaluating and managing investment risks, etc. Volatility is an indicator that measures the range and frequency of market price changes over a period of time, reflecting the uncertainty of financial asset prices and the risk of asset investment [1]. Although volatility is a common risk measure, it measures both the upward and downward movements of asset prices. However, investors generally pay more attention to the potential losses caused by falling asset prices, known as downside risk. Therefore, the research on downside risk, especially the prediction accuracy of the downside risk measurement index, also known as realized downward semi-variance (RDS), is of great significance to help investors reduce extreme losses and assist financial institutions in effectively evaluating risk management strategies.
According to whether various factors of the market are included, volatility prediction models are divided into autoregressive models and multiple regression models. The autoregressive model realizes reasonable prediction by capturing the features of volatility itself. Early scholars estimated and predicted volatility under low-frequency data through traditional econometric models, such as the autoregressive conditional heteroskedasticity (ARCH) model [2], the generalized autoregressive conditional heteroskedasticity (GARCH) model [3], and stochastic volatility (SV) model [4]. These models can extract the linear and deterministic information and save the uncertain information in the residual sequence. However, it usually needs to satisfy the assumptions of a linear relationship, stationarity, a normal distribution, etc., so it has obvious shortcomings when dealing with complex nonlinear high-dimensional data. With the continuous improvement of the accessibility of high-frequency data, its rich intraday feature information is gradually mined and utilized, and many scholars consider increasing the modeling frequency down to the day and minute. Andersen and Bollerslev [5] proposed realized volatility (RV) and provided a method to estimate volatility directly from intraday high-frequency data by calculating the cumulative square daily return under high-frequency data. Thus, the potential unobservability of volatility and the estimation of conditional variance can be effectively solved. Classical RV models include the autoregressive fractionally integrated moving average-RV (ARFIMA-RV) model proposed by Andersen et al. [6] and the heterogeneous autoregressive-RV (HAR-RV) model proposed by Corsi [7], which characterized the autocorrelation and long memory characteristics of RV to achieve effective prediction. Izzeldin et al. [8], Liang et al. [9], and Korkusuz et al. [10] forecast multi-market and multi-stock RV based on the HAR model, the HAR-RV-Average model, and the HAR-RV-X framework, respectively.
Although the traditional econometric model can describe the volatility aggregation characteristics and has the advantages of clear statistical principle, strong explanation, clear structure, and stable parameters, the model has strict requirements on assumptions such as data distribution, which makes it difficult to fully capture the nonlinear relationship in the time series, and has high processing cost and limited prediction accuracy. Therefore, machine learning models, which are good at capturing complex nonlinear relationships and extracting high-dimensional data features, are widely used to predict RV in financial markets with non-stationary, nonlinear, aggregated, and jumping characteristics. For example, Fałdziński et al. [11] applied support vector regression (SVR) to predict the volatility of energy commodity futures contracts such as crude oil and natural gas. Compared with the single machine learning model with outlier sensitivity and weak generalization ability, the ensemble machine learning model achieves a more robust prediction model by combining multiple weak learners. For example, Luong and Dokuchaev [12] and Demirer et al. [13] used RF models to forecast the RV of the stock index and crude oil, respectively. However, the ensemble learning model still needs a lot of feature engineering work, and the ability to process high-dimensional complex structure data needs to be improved. Compared to the ensemble learning model, the deep learning model approximates complex nonlinear features infinitely using multi-layer nonlinear transformation and has strong automatic feature learning and generalization ability. Traditional deep learning models, such as long short-term memory (LSTM) and the gated recurrent unit (GRU), have been widely used to predict financial stock market volatility [14,15,16].
The trend and fluctuation of the financial stock market are easily affected by company fundamentals, macroeconomic factors, market sentiment, and other aspects. Multiple regression methods can consider multiple influencing factors at the same time, conduct a comprehensive analysis of the financial market, extract relevant features, and accurately understand market dynamics. Some scholars use multivariate regression models to forecast RV in the stock market. For example, Todorova and Souček [17] incorporated overnight information flow into the RV prediction of ASX 200 and seven highly liquid Australian stocks. Audrino et al. [18] made use of a large amount of emotion and search information on the website, combined with emotion classification technology, an attention mechanism, and the high-dimensional predictive regression model, to predict volatility. Kim and Baek [19] proposed the FAHAR-LSTM model, which incorporated multiple relevant transnational stock market volatility information into the RV prediction of the Asian stock market. Kambouroudis et al. [20] extended implied volatility (IV), the leverage effect, and overnight returns into the HAR and predicted the RV of 10 international stock indexes. Lei et al. [21] used the LSTM model to forecast stock RV by taking trading information, on-floor trading sentiment, and unstructured forum text information as input variables. Jiao et al. [22] took trading information and news text information in the crude oil market as input variables and used the LSTM to forecast the RV of crude oil. Ma et al. [23] proposed a parallel neural network model of CNN-LSTM and LSTM based on mixing big data, which considered a variety of micro and macro variables and processed mixed frequency data through the RR-MIDAS model. Zhang et al. [24] introduced transaction information and sentiment indicators into the LSTM model, combined with the jump decomposition method, to predict RV.
There are two shortcomings in the existing literature. First, most of the target variables for prediction focus on RV, which is used as an indicator to measure risk. RV is calculated by the sum of the squares of the intraday high-frequency logarithmic returns, so falls and rises in stock prices have the same effect on RV, but it should be noted that there is significant leverage in the market, and the distribution of returns is asymmetric. Granger [25] points out that the risk of an investment is mainly related to low or even negative returns. In reality, investors tend to be more sensitive to investment losses, and the left tail of the return distribution constitutes a downside risk, which can better reflect investors’ concern about trading losses. However, RV ignores the effect of leverage on stock market transactions in the calculation process, making it difficult to reflect investors’ concerns about downside risks. Barndorff-Nielsen et al. [26] proposed the RDS based on the high-frequency negative return rate. Therefore, this paper takes high-frequency RDS, which is more concerned with the market, as the forecast object and fully studies the amplitude and frequency of downward fluctuations of asset prices.
In addition, from the perspective of volatility prediction models, most studies use traditional deep learning models (RNN, LSTM, and GRU) to predict volatility. According to the descriptive statistical analysis below, it can be found that RDS has a strong long-term autocorrelation, but the traditional deep learning model cannot capture global information and deal with long-distance dependencies in long sequences. Therefore, this paper adopts a new deep learning model called Transformer, which abandons the cyclic structure of recurrent neural networks (RNNs) and uses self-attention mechanisms and parallel computation to process sequential data. The model is not only good at capturing global information and long-term dependencies but is also able to capture ultra-short-term feature changes, which is suitable for RDS data with long-term autocorrelations and short-term aggregation jumps.
To sum up, compared with the existing literature, this study is more in line with the downside risk indicator RDS of investors’ risk preference. Multiple factor variables related to RDS, including market trading information, floor trading sentiment, the volatility of trading indicators, etc., are selected to study the correlation between various influencing factors and RDS. Secondly, the downside risk measure RDS is predicted and analyzed using a multivariable Transformer prediction model that is good at capturing global information and short-term characteristics. Finally, various evaluation criteria and the DM test are used to analyze the sensitivity of scroll window size, prediction step size, and prediction model performance. The robustness of the model under structural fracture, the influence of the emotion index on model prediction, and the RDS prediction results under the stacked-ML-DL framework are discussed.
The marginal contribution of this paper mainly includes the following three points:
(1)
Predict objects: Compared with the existing literature, the research objective of this paper focuses on downside risk measurement, also known as RDS prediction based on high-frequency data. The downside risk is consistent with investors’ risk preference, which describes the potential loss of the portfolio. The study of downside risk measurement RDS is conducive to further distinguishing the impact of different price changes on market risk and traders’ behavior and more accurately capturing the negative volatility characteristics of assets, thus helping traders avoid possible major losses when the market falls and optimize the investment portfolio to achieve more robust investment returns.
(2)
Prediction model: In this paper, combined with multiple market influencing factors such as trading volume, trading volume volatility, overnight information, turnover rate, and moving average, Transformer is used to forecast RDS. Compared to traditional deep learning LSTM and GRU models, the Transformer model uses a self-attention mechanism to simultaneously process global data, which is better able to capture long-distance dependencies when processing the long series data. In addition, the model abandons serialization processing and makes full use of the advantages of parallel computing to effectively improve the model’s training and operation efficiency. By inputting a variety of relevant variables into the Transformer model, its high computation and strong long-range capture capability can be used to improve the prediction accuracy and efficiency of RDS. In addition, the predictive effects of the ML and DL stack models are discussed and compared with the single Transformer model.
(3)
Sensitivity analysis: In this paper, a variety of evaluation criteria and improvement ratios are used to explore the impact of the size of each model’s rolling window on the prediction accuracy under one-step prediction. In addition, by studying the evaluation criteria results of each model under different prediction lengths, the robustness of each model is analyzed and tested. Combined with evaluation criteria and the DM test, the results of one-step prediction and multi-step prediction among different models are compared, and the accuracy and robustness of the Transformer model in RDS prediction are comprehensively tested through event shock and emotion factor analysis.
The research content of this paper is to construct a Transformer multivariable model to predict the RDS and analyze the empirical results. The remainder of this article is structured as follows: Section 2 defines RDS and introduces the Transformer model and prediction evaluation methods. In Section 3, multivariate rolling prediction is carried out by combining nine forecasting variables, demonstrating the empirical results of various multivariate forecasting models, and using a variety of evaluation criteria, the DM test, and other methods to analyze the sensitivity of the rolling window length, prediction length, and prediction model to give the optimal model. Section 4 discusses the robustness of the model under event shock and the influence and effect of adding an emotion index on the RDS prediction models. Section 5 explores the forecasting performance of the stacked-ML-DL hybrid forecasting framework and provides empirical analysis. Section 6 summarizes the content of the paper and provides its limitations and future extension.

2. Theoretical Basis

2.1. Realized Downward Semi-Variance (RDS)

Barndorff-Nielsen et al. [26] proposed RDS based on the high-frequency negative rate of return, which has become an indicator in the financial field to measure the downside risk of financial assets or investment portfolios. Unlike the traditional RV, RDS only focuses on the volatility of the asset price with a negative yield, and the semi-variance measure achieved using high-frequency intraday data provides a deeper understanding of the risk characteristics of the asset, especially during periods of market stress. Therefore, RDS is considered to be a more accurate risk measurement for stocks and other assets in the case of market declines, especially for those investors who are only concerned about the risk of asset values falling.
Under the risk-free arbitrage hypothesis, the logarithmic asset price process follows the semi-martingale process. Suppose that logarithmic asset price processes X t t 0 defined on the filtered probability space Ω , F , F t t 0 , P are Itô semi-martingale processes, which have the following form:
X t = X 0 + 0 t a s d s + 0 t σ s d W s + 0 t x < 1 κ δ s , x μ ν d s , d x + 0 t x 1 κ δ s , x μ d s , d x ,
where W is a standard Brownian motion, a s s 0 is a locally bounded predictable drift process, and σ s s 0 is a volatility process. δ ( ω , t , x ) is a predictable càglàd function satisfying that s u p x E | δ ( ω , t , x ) | γ ( x ) is locally bounded, where γ is a nonnegative function ( E is a subset of R ). μ is a Poisson random measure on R + × E with the intensity measure ν ( d s , d x ) = d s λ ( d x ) . In addition, κ is a continuous truncation function and is bounded on the compact support set.
In practice, only discrete data such as transaction prices can be observed in financial markets. Therefore, based on the high-frequency discrete observation point data of intraday asset price X t at a given time T , RDS can be defined as
R D S t = i = 1 1 Δ r t , i 2 I r t , i 0 , t = 0 , , T ,
where Δ is the sample interval, 1 / Δ is the number of samples, r t , i = l n X t 1 + Δ i l n X t 1 + Δ i 1 ) , and I x is the indicative function, whereby the value is 1 when {∙} is true and 0 when {∙} is false.

2.2. Transformer Model

Transformer is a self-attentional deep learning model originally proposed by Vaswani et al. [27] to solve sequence modeling tasks in natural language processing (NLP). However, due to its excellent performance in processing time series data, Transformer has also been widely used in financial time series analysis in recent years, such as predicting realized volatility. The Transformer model is mainly composed of two parts: the encoder and the decoder, and in time series prediction tasks, we usually only need the encoder part. Encoders capture complex dependencies between data through multiple stacked hierarchies.
The overall structure of the Transformer model depicted in Figure 1 can be summarized as the alternating stacking of self-attention modules and feedforward neural network modules. Each layer contains two key components: the multi-head self-attention mechanism and the feedforward neural network.
In Figure 1, the input is entered into the encoder after the positional encoding and embedding layer. The encoder consists of several stacked layers, each consisting of a self-attention module and a feedforward neural network module.
(1)
Self-Attention Mechanism
The core of the self-attention mechanism is to assign different weights to each element in the sequence to capture global context information. For the input sequence X R n × d (where n is the number of time steps and d is the feature dimension), the steps for calculating self-attention are as follows:
Step 1. Calculate the query vector, key vector, and value vector:
Q = X W Q , K = X W K , V = X W V ,
where W Q , W K , and W V are weight matrices of the query, key, and value, respectively.
Step 2. Calculate the attention score:
A = softmax Q K T d k ,
where d k is the dimension of the key vector used for scaling so that the dot product remains in the appropriate range of values.
Step 3. Generate the output:
Z = A V .
Multi-head self-attention mechanisms are extensions of self-attention mechanisms that enhance the learning ability by executing multiple attention mechanisms in parallel and then stitching together the respective outputs. The formula for the self-attention of bulls is
MultiHead Q , K , V = Concat head 1 , , head h W O ,
where h indicates the use of h parallel attention heads. The output of each attentional head is
head i = Attention X W i Q , X W i K , X W i V .
Since the Transformer model lacks timing awareness, timing information needs to be injected through location coding. This paper adopts sinusoidal coding:
P E p o s , 2 i = sin   pos   10000 2 i d ,
P E p o s , 2 i + 1 = cos   pos   10000 2 i d ,
where p o s is the time step position and i is the dimension index. The encoding results are added to the input features and fed into the model.
(2)
Feedforward Neural Network
The part of the feedforward neural network consists of two fully connected layers and usually uses the activation function R e L U in the middle. Feedforward networks can be represented as
FFN x = ReLU x W 1 + b 1 W 2 + b 2 .
(3)
Residual Connection and Layer Normalization
To mitigate the gradient loss problem, each sublayer in the Transformer model (i.e., the self-attention layer and the feedforward layer) introduces residual connectivity and layer normalization:
LayerNorm x + Sublayer x .
This structure contributes to the effective transmission of information and improves the training efficiency of the model.
In the downward volatility prediction task achieved in this paper, the input sequence X R n × d contains d -dimensional features of n historical time steps, d -dimensional embedding is obtained via linear projection, and multi-scale timing features are extracted by stacking N encoder layers. Finally, the predicted values of T future time steps are output through the fully connected layer:
Y ^ = W h H N + b h ,
where H ( N ) is the top-level encoder output and W h R T × d is the prediction weight matrix.

2.3. Evaluation Index

2.3.1. Loss Evaluation Criteria

In this paper, the mean absolute percentage error (MAPE), root mean square error (RMSE), mean absolute error (MAE), symmetric mean absolute percentage error (SMAPE), mean bias error (MBE), and standard deviation (SD) are used to evaluate and compare the error sequences of each model [28,29]. The calculation method is as follows:
M A P E = 1 N n = 1 N y ^ n y n y n ,
R M S E = 1 N n = 1 N y ^ n y n 2 ,
MAE = 1 N n = 1 N y ^ n y n ,
S M A P E = 1 N n = 1 N y ^ n y n ( y ^ n + y n ) / 2 ,
M BE = 1 N n = 1 N ( y ^ n y n ) ,
SD = n = 1 N y ^ n A r i t h m e t i c   m e a n 2 t o t a l   n u m b e r ,
where y n indicates the true value and y ^ n indicates the predicted value.

2.3.2. DM Test

The Diebold–Mariano test (DM test) [30] is a method used to compare the two time series prediction models to check whether there are significant differences in the prediction errors of different models to determine which model has better prediction results. The DM test is a statistical significance test based on the average difference in prediction errors. Assuming two models make predictions throughout T , the error sequence of model α and model β can be expressed as
E α = f y 1 α , y 1 α ^ , f y 2 α , y 2 α ^ , , f y T α , y T α ^ ,
E β = f y 1 β , y 1 β ^ , f y 2 β , y 2 β ^ , , f y T β , y T β ^ ,
where f · represents the error function (MAPE is adopted in this paper); y and y ^ represent the real value and predicted value, respectively; the superscript represents the model; and the subscript represents the sequence length.
DM statistics can be calculated as follows:
D M = D ¯ i = 1 T d i D ¯ 2 T 1 N 0,1 ,
where D = E α E β = d 1 , d 2 , , d T , D ¯ = i = 1 T d i T , and DM test statistics follows the standard normal distribution N ( 0,1 ) . The null hypothesis and alternative hypothesis related to DM in this paper are as follows:
H 0 : when | D M | = 0 , it indicates that the predictive performance of the two models is the same;
H 1 : when | D M | 0 , it indicates that there is a difference in the performance of these two models.
If D M   < 0, it indicates that model A has a better forecasting ability than model B; otherwise, it indicates that model B has a better forecasting ability than model A. The more significant the difference between the models, the greater the value of | D M | .

3. Empirical Results

3.1. Data Sources and Variables

This paper selects the Shanghai Stock Exchange (SSE) stock index from 17 February 2011 to 31 October 2022 (2846 days in total) as the analysis object and obtains 48 high-frequency trading data every day with 5 min as the sampling frequency. The daily RDS is calculated using the negative return rate portion of the adjacent 5 min. To discuss the RDS volatility levels of the SSE, Figure 2 shows the RDS chart, which is calculated using the SSE over the study year, and shows the external event shocks corresponding to the time when the RDS produces sharp fluctuations. Overall, the sharp fluctuations of RDS can usually correspond to some external event shocks, including macroeconomic situations, international trade relations, government policy adjustments, geopolitical conflicts, world epidemic impacts, trading market sentiment, etc. Observing the SSE, it is found that the downside risk of the stock index has increased sharply around 2015 and 2020, and the RDS has fluctuated significantly. The specific reasons for the fluctuations are as follows: in 2015, the global market fluctuated sharply due to events and factors such as the expected interest rate hike of the Federal Reserve, the liquidation of leveraged funds, geopolitical tensions, and market panic; the year 2020 was hit hard by the COVID-19 pandemic.
Therefore, studying the changes in the level of RDS plays an important role in studying the downside risk of the market. The accurate prediction of RDS can help investors quickly adjust their trading strategies in the face of global shocks, thereby reducing investment risks. In the fourth chapter, this paper selects COVID-19 to deeply study the robustness of the forecast model. At the time of the outbreak, market panic spread, and investors were pessimistic about future economic growth expectations, leading to rapid declines in global stock markets. By analyzing the influence of market panic and other emotional factors, the impact of the change in investor sentiment on the stock market and downside risk indicators is further discussed. The study of external shocks and emotional factors helps to evaluate the performance of the model under extreme market conditions and also provides a theoretical basis for investors to adjust their strategies in similar crises.
In addition, by using the access and computing of nine associated variables like trading information and investors’ emotions as the predictor variables of RDS prediction, all variable data are taken from the JoinQuant database (URL: https://www.joinquant.com (accessed on 31 December 2024)). The meanings of specific variables are shown in Table 1. For variable selection and calculation formulas, refer to Lei et al. [21].
Figure 3 shows the Spearman correlation coefficient results among various related influencing factors under SSE. As can be seen from Figure 3, the correlation coefficient between volume and RDS is the highest, reaching a positive value of 0.41, which means that volume and RDS have a significant positive correlation, indicating that the transmission of new information will affect the market price fluctuation and trading volume at the same time, and traders often choose to trade when the market is active. On the contrary, bias, CR, and DMA are negatively correlated with RDS. The correlation between the covariance tests of each influence index is generally low, indicating that introducing these variables into the model at the same time will not cause serious multicollinearity problems.
In the case of SSE, Table 2 shows descriptive statistics for RDS and various predictors. From the perspective of the skewness value, most variables are high, indicating that the data distribution is not completely symmetric, with obvious left or right skewness. For example, the skewness of OI is −2.236, which is much less than 1, meaning that the distribution may be skewed to the left. The skewness values of AR, BR, and RV-V are all greater than 10, indicating that the distribution may be significantly skewed to the right. From the perspective of the kurtosis value, while the volume and CDP are less than three, the other variables are much higher than the critical value of three; that is, the distribution of data is sharper than the normal distribution, and there are more extreme values. The Jarque–Bera test results show that at the significance level of 5%, the null hypothesis that each variable sequence is normally distributed is rejected. The results of the ADF stationarity test show that when the significance level is 1%, except for the CDP index, the t statistics of all other sequences are smaller than the critical ADF test value, failing the unit root test, and the data have non-stationarity characteristics. At the same time, for Q statistics with a lag of 5, 10, and 20, the null hypothesis that the corresponding order has no autocorrelation is rejected at the significance level of 5%, which reflects the characteristics of the long memory of the sequence.

3.2. Data Preprocessing and Rolling Prediction

Due to the different scales between variables, the direct use of raw data can lead to certain large-scale features dominating the model, affecting the stability and convergence rate of machine learning and deep learning models. Therefore, this paper uses the min–max normalization algorithm to preprocess data and scale the data to the interval [ 0,1 ] through linear transformation. The calculation formula is as follows:
x i = x i m i n ( x ) m a x ( x ) m i n ( x ) , i = 1,2 , , n ,
where x i is the original data point; m i n ( x ) and m a x ( x ) are the minimum and maximum values in dataset x , respectively; and x i is the normalized data point.
To capture new trends and patterns in the data in time, this paper selects the rolling prediction method to predict the RDS value of the next day by rolling the historical data of the past L days, as shown in Figure 4. The rolling forecasting approach is more closely aligned with the latest market trends and conditions, allowing the model to use the latest data at each forecast step to maintain accuracy and relevance. In this paper, L is selected as 5, 10, 20, and 30, respectively. By comparing the prediction states and capabilities under different rolling windows, the optimal rolling window can be determined and applied to all prediction models.

3.3. One-Step Prediction Result

In this study, all models and tests are run on Python3.10 on a computer with a 1.60 GHz CPU and 16 GB of memory. In this paper, seven different predictive models are used, including the traditional econometric model (Ridge regression), the single machine learning model (SVR), the ensemble machine learning model (RF, XGBoost 3.0.0), the traditional deep learning model (LSTM, GRU), and the novel deep learning model (Transformer). To investigate the influence of historical data length on the roll prediction model and compare the performance of various models on the SSE dataset, this section will use four different rolling window sizes (5, 10, 20, and 30) to make a one-step prediction of RDS. Under different rolling windows, the results of the one-step prediction evaluation criteria of RDS of SSE are shown in Table 3. To more intuitively present the specific impact of various lengths of historical information on the prediction accuracy of each model, as well as the changing trend of evaluation criteria with the increase in the length of historical information, Figure 5 shows the changing trend of the RDS one-step prediction error of each prediction model of SSE with historical information under different evaluation criteria.
According to the empirical results in Table 3 and Figure 5, from the perspective of the prediction model, first of all, the six evaluation criteria of the traditional econometric model perform poorly under different rolling windows. This is because the Ridge regression model, as a basic regularized linear model, is mainly used to deal with multicollinearity problems and overfitting problems in linear relations. Faced with nonlinear RDS sequences, the Ridge regression model makes it difficult to handle nonlinear relations between input features and output variables, resulting in large prediction errors and low sensitivity to historical data. Secondly, the machine learning model is superior to the traditional measurement model overall, as the numerical results of MAE, MAPE, SMAPE, and SD indexes are all lower than those of the corresponding Ridge model under SSE. Compared with the Ridge model, the machine learning model can capture nonlinear features in the RDS sequence. In general, the prediction performance of the deep learning model is more advantageous than that of the machine learning model. Among them, the traditional deep learning model GRU is particularly sensitive to the setting of the rolling window size, and the prediction accuracy is mostly similar to that of the machine learning model. This may be due to the simple structure and few parameters of the GRU model, which makes it more sensitive to changes in model input features such as window size during the training process. Inappropriate window size settings will directly affect the gradient calculation and backpropagation process of the model, thus affecting the training effect of the model. The prediction effect of the LSTM model is slightly better than that of the GRU model, showing that the prediction results fluctuate with the amount of historical information used, and the error criteria are usually lower than that of the traditional machine learning model. This shows that LSTM, despite having more parameters, also provides a higher model capacity for the model, enabling it to capture more complex patterns and trends, thereby slightly reducing the instability caused by improper window size settings. Overall, the Transformer model ranks first in the comprehensive results of evaluation criteria in one-step prediction. It can be seen from Table 3 that the Transformer model has the lowest error evaluation criteria when the rolling window size is 10 and 20.
Quantitative analysis is carried out according to the advantages of the Transformer model compared with enhancement. The formula is as follows:
I m p r o v e m e n t   r a t i o = x 2 x 1 x 2 ,
where x 2 is the value of the evaluation criteria before promotion (large) and x 1 is the value after promotion (small). For SSE, taking a rolling window size of 20 as an example, the average RMSE, MAE, MAPE, SMAPE, MBE, and SD improvement ratio of the Transformer model and other models are 52.23%, 20.03%, 62.33%, 60.33%, 37.57%, and 18.70%, respectively. In addition, compared with the LSTM model, the improvement ratios of the six evaluation criteria for SSE are 38.30%, 56.25%, 77.01%, 34.09%, 36.66%, and 6.98%, respectively, indicating that the Transformer model has significantly improved the prediction accuracy compared to LSTM and other models.
From the perspective of the rolling window size, the results in Figure 4 show that the evaluation criteria of all models are basically at the lowest point when the historical data length is set to 20 days, that is, the model with the best prediction effect when the rolling window is 20. For further analysis of the data in Table 3, taking the RMSE of the Transformer model as an example, compared with the historical data of 5, 10, and 30 lengths, the model performance for SSE can be improved by 24.69%, 15.90%, and 43.60%, respectively. There are two main reasons for choosing the optimal rolling window size (20 days): the learning length of the historical information of the models and the economic implications. If the length of the historical data learned by the model is too short (such as 5 days and 10 days), the model may not be able to capture enough long-term memory information and may ignore the cyclical fluctuations of the market, resulting in bias in prediction judgment. Conversely, if the historical data are too long (such as 30 days), sensitivity to immediate market reactions may be lost due to the over-fitting of noise and the disturbance of information in the historical data, and the ability to generalize may be reduced. Setting the historical data length to 20 days, including fluctuations in several market cycles, ensures that the model has enough information to capture trends and patterns in the data, remain sensitive and adaptable to new data, and effectively avoid the risk of overfitting. The 20 trading days cover about a month, which is consistent with some of the cyclical characteristics of financial markets. Specifically, monthly cycles often correspond to monthly reports, corporate earnings reports, and macroeconomic data release cycles, which can have an impact on price volatility and market sentiment. By analyzing 20 days of historical data, the model is able to identify cyclical changes and better predict short-term fluctuations in the market.
Therefore, this paper sets the roll window size to 20 and further verifies the predictive advantages of the Transformer model through the DM test. The DM test results of one-step prediction under SSE are shown in Table 4, from which it can be found that DM statistics are significant at the 0.1% confidence level. By focusing on the Transformer column, it can be found that the DM statistics between the Transformer and other comparison models are all positive, indicating that the Transformer model has the best prediction performance. According to the principle that the greater the difference in predictive performance between models, the greater the absolute value of DM statistics, all DM statistical values are greater than 10, highlighting the significant advantages of the Transformer model. In summary, the Transformer model has the best prediction effect in one-step prediction, and the input of 20-length history information is more secure. The Transformer model offers the best predictive performance for two reasons. First of all, since volatility is often affected by a variety of factors, including macroeconomic indicators, market sentiment, etc., it is not only necessary to mine local characteristics and patterns but also to grasp the overall market dynamic information. Taking SSE as an example, according to Q statistics with lags of 5, 10, and 15, the autocorrelation of the RDS series is enhanced with the increase in the number of lag terms, reaching 10³ or higher, and the SSE time series has ultra-long-term autocorrelations. To this end, the Transformer model combines location coding and self-attention mechanisms to effectively capture global information about the input data and capture long distance dependencies in the sequence. While the traditional RNN, LSTM, and GRU models can capture long-term dependencies in the sequence, there may be gradient disappearance or explosion when processing large-scale long sequence data. Therefore, the Transformer model is better equipped to analyze long-term data trends, cyclical changes, and other features in a longer time window. In addition, the analysis of SSE high-frequency data in Figure 2 shows that market volatility often changes dramatically in a short period, such as the impact of major events or the promulgation of government policies, which requires a rapid response and processing mechanism. Due to the cyclic structure of LSTM and GRU, the computation of each time step depends on the output of the previous time step, and this limitation is particularly evident when dealing with ultra-short-term features that require fast responses and processing. The parallelization capability of the Transformer model can significantly improve model training speed and effectively capture ultra-short-term features.

3.4. Multi-Step Prediction Result

To explore the long-term forecasting ability and robustness of the model, this paper selects 20-day historical data for multi-step forecasting based on one-step forecasting results, including 5-step (one week), 10-step (two weeks), 15-step (three weeks), and 20-step (one month) forecasting. Table 5 describes the evaluation criteria results of different prediction models under different prediction steps of SSE, and Figure 6 shows the change trend chart of the evaluation criteria of different prediction models under different prediction steps of SSE. Figure 6 displays that with the increase in the number of prediction steps, the prediction effect of each model generally presents a downward trend, especially when the long-term prediction is made on the 20th day, where the data of each error evaluation criteria have a significant increase. This indicates that with the extension of forecasting time, the difficulty of the model to capture future trends increases, and the cumulative effect of future market uncertainty and forecasting errors leads to a decline in long-term forecasting accuracy.
On this basis, this paper explores the advantages and disadvantages of various models in multi-step prediction. Consistent with the effect of one-step prediction, the prediction accuracy of the deep learning model is generally better than that of the machine learning model. According to the multi-step prediction results, it can be found that the GRU model shows a certain advantage in predicting the RDS value of the next week, with RMSE, MAE, MAPE, SMAPE, MBE, and SD decreasing to 0.3354, 0.2064, 0.6869, 0.7957, 0.0963, and 0.2938, respectively. In the case of two and three weeks ahead, the GRU model and the Transformer model’s comprehensive evaluation criteria are comparable. However, when considering the RDS data of the forecast for the next month, it can be seen from the results of Figure 6 that the black dotted line representing the Transformer model can maintain a stable situation and rises slightly with the increase in the number of prediction steps. Moreover, it can be found from the bolded data in Table 5 that the Transformer model has the lowest error evaluation criteria. It is even better than the performance of traditional econometric models and machine learning models in one-step prediction. Next, the prediction effect of the Transformer and GRU models decreases with the increase in steps. Taking MAE and MAPE indicators as examples, the average decline ratio of different prediction steps in the Transformer model is 0.3026 and 0.3340, respectively, which is lower than the GRU model’s average decline ratios of 0.3792 and 0.6212. The result shows that the Transformer model can resist the problems caused by the accumulation and superposition of errors in long-term forecasting and has flexibility and stability in capturing the long-term dependence of overall data and market trends, which reflects some advantages and stability in long-term forecasting. This result also confirms the self-attention framework and parallelization of the Transformer model, which enables the model to grasp the short-term and global characteristics of the input sequence.
To explore the prediction ability of each model in the multi-step prediction scenario, the DM test is used for quantitative analysis. The confidence level is set at 5%, and the DM test results predicted by multiple steps of each model under SSE are shown in Table 6. Table 6 shows that most DM statistical results are significant at the 5% significance level, and the statistics of the Transformer model are basically positive at 5-, 10-, 15-, and 20-step prediction lengths. Taking the RDS data forecast for the next month as an example, the Transformer model and the other six models are, respectively, tested using DM, and the numerical results are 19.372, 8.865, 10.480, 13.527, 18.276, and 7.238, which verifies the high accuracy and stability of Transformer model prediction in long-term forecasting scenarios. In summary, the Transformer model is selected as the best model for multi-step RDS prediction.

4. Future Study

4.1. External Event Shock

Based on the analysis of RDS fluctuations in Figure 2, it can be found that the sharp fluctuations in the stock market are most likely caused by the impact of external events. To study the robustness of the model under the impact of external events, the moment that fluctuates sharply in Figure 2 is analyzed, and the COVID-19 epidemic is selected as an external impact event to verify whether the optimal Transformer model can maintain its inherent advantages in the case of structural market fractures and its ability to capture and adapt to major market downside risks. This paper compares the Transformer model with six other models to test the relative cumulative error of the Transformer model over the 2020 COVID-19 cycle. Figure 7 describes the cumulative relative prediction error of the SSE index RDS prediction of the Transformer model under different historical lengths, with the dashed blue line highlighting the COVID-19 outbreak. Considering that the occurrence of external shock events will often trigger a significant price fluctuation, the results in Figure 7 are consistent with expectations. For a while after the outbreak of COVID-19, the relative cumulative errors of the Transformer model and other models have significantly increased under different historical information lengths. Empirical results show that the RDS downside risk index fluctuates sharply when market shocks occur. Traditional machine learning and deep learning models are unable to capture features quickly and respond promptly, resulting in reduced prediction accuracy, while the Transformer model can maintain stable prediction accuracy and effectively respond to market changes, as it has a certain robustness.

4.2. Sentiment Factor

As major participants in the financial market, individual investors are highly emotional in their behavior, and external events may have an impact on investors’ emotions. Since traders may focus more on downside risk, market panic can lead to irrational movements in the stock market, resulting in a sharp increase in downside risk. Section 4.1 discusses the robustness of the model under the impact of external events. In this section, the sentiment index is constructed and incorporated into the RDS prediction impact index system. Taking SSE data as an example, the title and body text content of posts published on the Eastmoney stock forum between 1 March 2013 and 31 March 2022 (9 years) are captured, and the title, publication time, and post text are finally taken as a group. In this article, we first clean up the data by de-weighting and removing spaces. Then, considering that the extracted surface content of the title and text is somewhat misleading and incomplete and the combination of the two can express a complete and clear definition without losing any information, we take the title and text of each post (separated by commas) as a class and put the class and the release time of the text as a group. Finally, a total of 1700,535 sets of final text data are obtained.
Since negative words and positive words in financial markets may be different from the general corpora, the original BERT model based on Chinese corpora published by Google [31] is not chosen, but the FinBERT model published by Value Simplex in 2021 is used. This model is a Chinese BERT model pre-trained on a large-scale corpus in the financial field. The BERT model is fine-tuned for the classification of financial sentiment. The FinBERT model is used to classify the comments captured by online stock forums, and the number of daily positive and negative comments is counted. Furthermore, Antweiler and Frank’s [32] sentiment index construction model is further referred to construct and quantify the investor sentiment index S e n t t . The calculation formula is as follows:
S e n t t = M t p o s M t n e g M t p o s + M t n e g l n 1 + M t ,
where M t = M t p o s + M t n e g , M t p o s , a n d   M t n e g are the weighted sum of the number of positive and negative posts, i.e., weighted by the number of reads, in day t, respectively.
In this paper, descriptive statistical analysis is made on the sentiment index (from 8 January 2013 to 19 May 2022), and it is observed that the mean is −1.7285, which is less than 0, indicating that stockholders are more pessimistic about the stock market during the whole sample period. From the skewness and kurtosis value, the skewness of the sent sequence is −0.212, and the kurtosis is 2.555, indicating that the data distribution is slightly skewed to the right and relatively smooth compared with the normal distribution. The Jarque–Bera test also shows that the sent sequence has a non-normal distribution when the significance level is 1%. In addition, the results of the ADF statistical test show that the series has a certain smoothness at the significance level of 1%. Figure 8 shows the trend of RDS and sentiment indicators over the full sample period. It can be seen that during the bull and bear markets of 2015, stock prices fluctuated wildly, and with the increase in RDS, the number of forum comments increased compared to the previous period. Especially in the second half of 2015, when the stock market was in a bear market, the negative sentiment indicator in forum comments exceeded −6, reaching an extreme value.

4.3. Empirical Results of the Transformer Model Under the Sentiment Factor

The paper discusses the one-step prediction results of the Transformer model. Figure 9 and Table 7 show in detail the evaluation criteria of the Transformer model for the one-step prediction of RDS under different history lengths with and without the sentiment factor. First, by comparing the two situations, it can be found that the information reflecting investor sentiment in stock forum posts has a significant positive impact on RDS forecasts. Figure 8 shows this phenomenon more directly. It is found that after adding the sentiment factor, the six error evaluation criteria all decrease significantly, which effectively reduces the sensitivity of the Transformer model to history length in the prediction process. Especially under the rolling window length of 5 and 30, that is, under short and long historical information, the addition of the sentiment factor can greatly reduce the RDS prediction error and improve the overall robustness of the model. Taking the rolling window length of 30 as an example, the improvement ratio of RMSE, MAE, MAPE, SMAPE, MBE, and SD reaches 0.4259, 0.6402, 0.7451, 0.4313, 0.4616, and 0.1692, respectively. Second, the RDS prediction error decreases with the increase in the history length, and the prediction effect is better when the rolling window size of 20 or 30 is selected. Considering the calculation amount and prediction effect comprehensively, it is suggested to choose 20 as the length of the rolling window, which is consistent with the previous content. To sum up, under the forecasting framework of the Transformer model, adding the sentiment indicator of the stock forum can significantly improve the prediction accuracy of RDS, especially in the case of a long rolling window.
In addition, to verify whether there is an improvement in the multi-step prediction of forum sentiment participation, this study uses historical 20-day data to make 5-day, 10-day, 20-day, and 30-day multi-step predictions. The experimental results are shown in Figure 10 and Table 8, and it can be seen that the addition of stock forum sentiment significantly improves the accuracy of multi-step prediction. Specifically, Figure 10 shows that in the range of 1 to 15 prediction steps, the improvement effect of the sentiment factor on prediction accuracy boosts with the rise in prediction steps. This may be because in the short-term forecast cycle of 2–3 weeks, due to the lag of the transmission of forum information and the cumulative effect of sentiment, the impact of forum sentiment on the volatility of the financial market is relatively large and has a certain persistence. In the longer forecast cycle, other factors such as market fundamentals gradually become the dominant factor, and the impact of forum sentiment is relatively small. The results in Table 8 also show that under each forecasting step size, the six error evaluation criteria of the Transformer model with the sentiment factor added have improved, and the fluctuation of prediction accuracy has weakened. To sum up, the addition of relevant investor sentiment factors is conducive to improving the RDS prediction accuracy and stability of the Transformer model in the short term.

4.4. Empirical Results of the S&P500

To ensure the universality of the model prediction results and to avoid the individual bias caused by sample limitations, we add the Standard and Poor’s 500 (S&P500) up to 15 August 2024 (4539 days in total) as the analysis object for the prediction test. The results of the one-step prediction evaluation criteria of the S&P500 are shown in Table 9 below. As can be seen from Table 9, similar to SSE, the Transformer model ranks first overall under different historical information lengths. For the S&P500, when the rolling window size is 20, the improvement ratio of the six evaluation criteria is 3.89%, 3.34%, −1.20%, 2.21%, 9.48%, and 3.66%, respectively. Except for the slight increase in the MAPE index, the other indexes have improved. For further analysis of the data in Table 4, taking the RMSE of the Transformer model as an example, compared with the historical data with lengths of 5, 10, and 30, the model performance for the S&P500 can be improved by 3.49%, 12.70%, and 6.62%, respectively.

5. Discussion

Based on the above analysis of the forecasting effect of the single model, this section discusses the forecasting effect of an advanced stacked hybrid model and its advantages [29]. The main idea of the stacked model is as follows: The first layer of the model will train multiple base models (including machine learning, deep learning, etc.) to initially predict the target variables. The second layer of the model uses the prediction results of the base model as input features to train simple models (such as linear regression). This layer in the model learns the assigned weight of each base model, integrates the prediction results of each base model, and uses the model output as the final prediction result to complete the hybrid model’s prediction. According to the previous analysis of prediction results, the Transformer model with the best overall prediction effect of a single model is compared with two stacked hybrid models (stacked-ML and stacked-ML-DL). Stacked-ML includes submodels for SVR, RF, XGBoost, and Ridge, while stacked-ML-DL includes submodels for SVR, RF, XGBoost, Ridge, LSTM, GRU, and Transformer. Using linear regression as a meta model, the predicted value of each model is taken as the feature; the real value is taken as the target, and the linear regression model is trained to learn the weight. Taking SSE as an example, the one-step prediction evaluation results of RDS under different rolling windows are shown in Table 10.
As shown in Table 10, stacked-ML-DL has the best prediction effect on the whole. From the point of view of the setting of the rolling window, it is consistent with the conclusion of the single prediction model. When the window size is 20, the overall prediction error of each index is the lowest. From the perspective of the prediction model, the stacked model can effectively adapt to changes in historical data distribution, control model complexity, reduce overfitting, and improve model prediction accuracy through the complementarity between multiple base models and the adjustment of base model weights by the second-layer model according to data changes. Taking SSE as an example, when selecting the rolling window as 20, the error evaluation criteria RMSE, MAE, MAPE, SMAPE, MBE, and SD of the stacked-ML-DL model are 0.2810, 0.1795, 0.8997, 0.5657, −0.0087, and 0.2808, respectively. The advantage of the stacked model is that by combining the prediction results of multiple models, using the sensitivity of different models to different aspects of the data, reducing the possible bias and variance of the single model, and capturing the complex pattern of the data to improve the overall prediction accuracy. Stacked models can be adapted to various types and lengths of data without strong dependence on a particular model or algorithm. Moreover, through regularization of the second layer model, the stacked model can effectively reduce the overfitting risk when the base model is relatively complex, and the second layer model can provide the contribution degree or weight of each base model, improving the explainability of the model.

6. Conclusions and Extension

The downside risk refers to the risk that the asset value is lower than the expected target, which describes the extent of losses that investors may face. Its importance lies in that it is consistent with investors’ sensitivity to losses, which helps investors better evaluate potential losses and achieve more stable investment returns. In this paper, RDS is used to measure the downside risk, and multiple predictive indicators, as well as traditional econometric models, machine learning models, traditional deep learning models, and Transformer, are combined to predict RDS. Furthermore, the sensitivity analysis of the rolling window length, prediction length, and prediction model is carried out using various evaluation criteria, the DM test, etc.
This paper mainly draws the following three conclusions. First, the degree of use of historical information will affect the prediction effect of various models. In this paper, the rolling window sizes of 5, 10, 20, and 30 are used for empirical analysis to verify the prediction advantage when the window size is set to 20, which provides strong data support for selecting the optimal historical data length. Second, the Transformer model has the best prediction accuracy in one-step prediction. RMSE, MAE, MAPE, SMAPE, MBE, and SD are used to evaluate the prediction effect of the seven models. Compared with each model, the average improvement ratios of the Transformer model under SSE are 0.5223, 0.2003, 0.6233, and 0.6033, 37.57%, and 18.70%, respectively. In particular, the Transformer model is compared with the traditional deep learning model, and the improvement ratio of each evaluation criterion is more than 30%. The Transformer model can maintain a certain degree of accuracy in prediction due to its self-attention mechanism, location coding positioning, and parallel processing capabilities, which make the model efficient in capturing long-distance dependencies in time series prediction. Third, according to the prediction results of various models in the multi-step prediction scenario, it can be found that the Transformer model not only has advantages in prediction accuracy but also shows significant stability in long-term RDS prediction. The DM test is used to compare the advantages and disadvantages of the model, and it is found that all DM test statistics of the Transformer model are positive, which further confirms the superiority and stability of the Transformer model in long-term prediction, and the potential of grasping the long-term dependence relationship of nonlinear and non-stationary complex data. The robustness of the Transformer model and the advantages of the stacked hybrid model are further confirmed in the discussion of robustness under structural fracture, the influence of the emotion index on model prediction, and RDS prediction results under the stacked ML-DL framework.
The research in this paper provides empirical support for the deep learning prediction of RDS. In future volatility research, the architecture and parameters of the Transformer model can be further improved and optimized, such as introducing new attention mechanisms and adjusting the number of attention heads, learning rate, and batch size, to find a better model structure and configuration. In addition, this paper focuses on the model construction and prediction result analysis of the downside risk indicator RDS. In terms of investment decision-making and risk management, future studies may explore the economic impact of prediction models, such as using model prediction results to assist investors in formulating specific trading strategies and helping risk managers optimize investment strategies, to further enhance the application value of prediction models.

Author Contributions

Conceptualization, Y.S.; methodology, Y.S.; software, P.N.; validation, J.P. and C.K.; formal analysis, Y.Z.; investigation, Y.Z.; resources, Y.S.; data curation, P.N.; writing—original draft preparation, Y.Z.; writing—review and editing, L.H.; visualization, J.P. and C.K.; supervision, Y.S.; project administration, Y.S.; funding acquisition, L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Shanghai Planning Project of Philosophy and Social Science (2023BJB009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset for the empirical analysis can be derived from the following resource available in the public domain: https://www.joinquant.com/data/ (accessed on 31 December 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, Y.; Lin, Y.; Hou, X.; Zhang, X. Novel optimization approach for realized volatility forecast of stock price index based on deep reinforcement learning model. Expert. Syst. Appl. 2023, 233, 120880. [Google Scholar] [CrossRef]
  2. Engle, R.F. Autoregressive Conditional Heteroscedasticity with Estimates of the Variance of United Kingdom Inflation. Econometrica 1982, 50, 987–1007. [Google Scholar] [CrossRef]
  3. Bollerslev, T. Generalized autoregressive conditional heteroskedasticity. J. Econom. 1986, 31, 307–327. [Google Scholar] [CrossRef]
  4. Taylor, S.J. Modeling Stochastic Volatility: A Review and Comparative Study. Math. Financ. 1994, 4, 183–204. [Google Scholar] [CrossRef]
  5. Andersen, T.G.; Bollerslev, T. Answering the skeptics: Yes, standard volatility models do provide accurate forecasts. Int. Econ. Rev. 1998, 39, 885–905. [Google Scholar] [CrossRef]
  6. Andersen, T.G.; Bollerslev, T.; Diebold, F.X.; Labys, P. Modeling and forecasting realized volatility. Econometrica 2003, 71, 579–625. [Google Scholar] [CrossRef]
  7. Corsi, F. A simple approximate long-memory model of realized volatility. J. Financ. Econom. 2009, 7, 174–196. [Google Scholar] [CrossRef]
  8. Izzeldin, M.; Hassan, M.K.; Pappas, V.; Tsionas, M. Forecasting realised volatility using ARFIMA and HAR models. Quant. Financ. 2019, 19, 1627–1638. [Google Scholar] [CrossRef]
  9. Liang, C.; Wei, Y.; Lei, L.; Ma, F. Global equity market volatility forecasting: New evidence. Int. J. Financ. Econ. 2022, 27, 594–609. [Google Scholar] [CrossRef]
  10. Korkusuz, B.; Kambouroudis, D.; McMillan, D.G. Do extreme range estimators improve realized volatility forecasts? Evidence from G7 Stock Markets. Financ. Res. Lett. 2023, 55, 103992. [Google Scholar] [CrossRef]
  11. Fałdziński, M.; Fiszeder, P.; Orzeszko, W. Forecasting volatility of energy commodities: Comparison of garch models with support vector regression. Energies 2021, 14, 6. [Google Scholar] [CrossRef]
  12. Luong, C.; Dokuchaev, N. Forecasting of Realised Volatility with the Random Forests Algorithm. J. Risk Financ. Manag. 2018, 11, 61. [Google Scholar] [CrossRef]
  13. Demirer, R.; Gkillas, K.; Gupta, R.; Pierdzioch, C. Risk aversion and the predictability of crude oil market volatility: A forecasting experiment with random forests. J. Oper. Res. Soc. 2022, 73, 1755–1767. [Google Scholar] [CrossRef]
  14. Jia, F.; Yang, B. Forecasting volatility of stock index: Deep learning model with likelihood-based loss function. Complexity 2021, 2021, 1–13. [Google Scholar] [CrossRef]
  15. Zhang, C.-X.; Li, J.; Huang, X.-F.; Zhang, J.-S.; Huang, H.-C. Forecasting stock volatility and value-at-risk based on temporal convolutional networks. Expert. Syst. Appl. 2022, 207, 117951. [Google Scholar] [CrossRef]
  16. Koo, E.; Kim, G. A new neural network approach for predicting the volatility of stock market. Comput. Econ. 2023, 61, 1665–1679. [Google Scholar] [CrossRef]
  17. Todorova, N.; Souček, M. Overnight information flow and realized volatility forecasting. Financ. Res. Lett. 2014, 11, 420–428. [Google Scholar] [CrossRef]
  18. Audrino, F.; Sigrist, F.; Ballinari, D. The impact of sentiment and attention measures on stock market volatility. Int. J. Forecast. 2020, 36, 334–357. [Google Scholar] [CrossRef]
  19. Kim, D.; Baek, C. Factor-augmented HAR model improves realized volatility forecasting. Appl. Econ. Lett. 2020, 27, 1002–1009. [Google Scholar] [CrossRef]
  20. Kambouroudis, D.S.; McMillan, D.G.; Tsakou, K. Forecasting realized volatility: The role of implied volatility, leverage effect, overnight returns, and volatility of realized volatility. J. Futures Mark. 2021, 41, 1618–1639. [Google Scholar] [CrossRef]
  21. Lei, B.; Liu, Z.; Song, Y. On stock volatility forecasting based on text mining and deep learning under high-frequency data. J. Forecast. 2021, 40, 1596–1610. [Google Scholar] [CrossRef]
  22. Jiao, X.R.; Song, Y.P.; Kong, Y.; Tang, X.L. Volatility forecasting for crude oil based on text information and deep learning PSO-LSTM model. J. Forecast. 2022, 41, 933–944. [Google Scholar]
  23. Ma, W.; Hong, Y.; Song, Y. On stock volatility forecasting under mixed-frequency data based on hybrid RR-MIDAS and CNN-LSTM models. Mathematics 2024, 12, 1538. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Song, Y.; Peng, Y.; Wang, H. Volatility forecasting incorporating intraday positive and negative jumps based on deep learning model. J. Forecast. 2024, 43, 2749–2765. [Google Scholar]
  25. Granger, C.W.J. In Praise of Pragmatic Econometrics. In The Methodology and Practice of Econometrics: A Festschrift in Honour of David F. Hendry, 1st ed.; Castle, J.L., Shephard, N., Eds.; Oxford University Press: Oxford, UK, 2008; pp. 105–116. [Google Scholar]
  26. Barndorff-Nielsen, O.; Kinnebrock, S.; Shephard, N. Measuring downside risk-realised semivariance. In Volatility and Time Series Econometrics: Essays in Honor of Robert Engle, 1st ed.; Bollerslev, T., Jeffrey, R., Mark, W., Eds.; Oxford Academic: Oxford, UK, 2010; Volume 7, pp. 117–136. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  28. Asgarkhani, N.; Kazemi, F.; Jakubczyk-Gałczyńska, A.; Mohebi, B.; Jankowski, R. Seismic response and performance prediction of steel buckling-restrained braced frames using machine-learning methods. Eng. App. Artif. Intell. 2024, 128, 107388. [Google Scholar] [CrossRef]
  29. Kazemi, F.; Asgarkhani, N.; Jankowski, R. Optimization-based stacked machine-learning method for seismic probability and risk assessment of reinforced concrete shear walls. Expert. Syst. Appl. 2024, 255, 124897. [Google Scholar]
  30. Diebold, F.X.; Mariano, R.S. Comparing predictive accuracy. J. Bus. Econ. Stat. 2022, 20, 134–144. [Google Scholar]
  31. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  32. Antweiler, W.; Frank, M.Z. Is all that talk just noise? The information content of internet stock message boards. J. Financ. 2004, 59, 1259–1294. [Google Scholar] [CrossRef]
Figure 1. Transformer structure.
Figure 1. Transformer structure.
Mathematics 13 01260 g001
Figure 2. RDS chart of the SSE stock index.
Figure 2. RDS chart of the SSE stock index.
Mathematics 13 01260 g002
Figure 3. Spearman correlation coefficient heatmap of predictor and predicted variables of SSE.
Figure 3. Spearman correlation coefficient heatmap of predictor and predicted variables of SSE.
Mathematics 13 01260 g003
Figure 4. Schematic diagram of rolling prediction.
Figure 4. Schematic diagram of rolling prediction.
Mathematics 13 01260 g004
Figure 5. Under SSE, the change trend chart of each evaluation criteria is predicted using different models under different forecasting history lengths.
Figure 5. Under SSE, the change trend chart of each evaluation criteria is predicted using different models under different forecasting history lengths.
Mathematics 13 01260 g005
Figure 6. Under SSE, the change trend chart of each evaluation criteria based on different prediction models and various forecasting horizons is shown.
Figure 6. Under SSE, the change trend chart of each evaluation criteria based on different prediction models and various forecasting horizons is shown.
Mathematics 13 01260 g006
Figure 7. Cumulative relative errors of the Transformer model versus other models at various historical lengths under SSE.
Figure 7. Cumulative relative errors of the Transformer model versus other models at various historical lengths under SSE.
Mathematics 13 01260 g007
Figure 8. Dual coordinate main view of RDS and the sentiment indicator under SSE.
Figure 8. Dual coordinate main view of RDS and the sentiment indicator under SSE.
Mathematics 13 01260 g008
Figure 9. Evaluation criteria change trend chart of the Transformer model with or without the sentiment factor under different history lengths of SSE.
Figure 9. Evaluation criteria change trend chart of the Transformer model with or without the sentiment factor under different history lengths of SSE.
Mathematics 13 01260 g009
Figure 10. Change trend chart of each evaluation criteria of the Transformer model with or without the emotion factor under different forecasting horizons of SSE.
Figure 10. Change trend chart of each evaluation criteria of the Transformer model with or without the emotion factor under different forecasting horizons of SSE.
Mathematics 13 01260 g010
Table 1. Specific explanations of predictor variables.
Table 1. Specific explanations of predictor variables.
VariablesExplanation
VolumeThe number of transactions between the buyers and sellers of stocks in a given period. It is usually measured by the number of shares traded and the amount of transactions, which can reflect the market activity.
BiasA measure of the gap between stock prices and moving averages. It is used to assess whether a stock price is too high or too low and may be similar to the moving average trend.
DMAAn indicator that assesses the power of buyers and sellers and future price movements, and it is derived from the difference between the moving averages of two different cycles.
CDPContrarian operating system. It is mainly used for short-term transactions, reflecting the relationship between the highest price, the lowest price, and the closing price in the market.
ARIndicators of market sentiment. AR measures the strength of buying and selling in the market; a high AR value indicates strong buying, while a low AR value indicates strong selling.
BRA buyer’s ratio. As an indicator of the market’s desire to buy and sell, an increase in BR indicates the buyer’s strength, while a decrease in BR indicates the seller’s strength.
OIOvernight information, that is, market information accumulated during non-trading hours. It can often affect the market opening and trading behavior of the next trading day.
RV-VVolatility of trading volume. It reflects the degree of change in market trading volume.
CRChange rate. It is a measure of how much a price has changed over a given period.
Table 2. Descriptive statistics of RDS and predictor variables of SSE.
Table 2. Descriptive statistics of RDS and predictor variables of SSE.
SSERDSVolumeBiasDMACDPARBROIRV-VCR
Mean0.5442.137 × 10101.410 × 1041.422 × 10−12.919 × 1036.7888.843−2.9633.3348.100 × 105
Standard deviation1.2541.329 × 10101.956 × 1024.010 × 10−15.454 × 1026.686 × 1018.832 × 1011.918 × 1013.477 × 1011.298 × 102
Min1.808 × 10−23.963 × 109−1.729 × 101−3.472 × 1021.933 × 103−1.0000.000−2.598 × 1026.684 × 10−1−8.490 × 102
Max1.868 × 10−18.571 × 10109.520 × 1022.019 × 1025.154 × 1032.164 × 1033.7222.883 × 1021.362 × 1035.760 × 102
Skewness7.9271.253−9.379 × 10−1−1.5142.299 × 10−12.477 × 1013.092 × 101−2.2363.772 × 1017.932 × 10−1
Kurtosis8.166 × 1011.6696.7041.197 × 10−13.189 × 10−17.024 × 1021.170 × 1035.538 × 1011.426 × 1036.212
Jarque–Bera8.177 × 105 **1.072 × 103 **5.724 × 103 **1.800 × 104 **3.690 × 101 **5.859 × 107 **1.623 × 108 **3.647 × 105 **2.408 × 108 **4.854 × 103 **
ADF−5.098 ***−3.707 ***−9.665 ***−8.393 ***−2.242−5.330 × 101 ***−5.355 × 101−7.046 × 101 ***−9.130 ***−9.612 ***
Q (5)3.973 × 103 **1.307 × 104 **2.459 × 103 **6.379 × 103 **1.402 × 104 **2.350 × 10−1 **9.082 × 10−1 **4.109 × 101 **7.114 × 102 **8.463 **
Q (10)5.839 × 103 **2.273 × 104 **2.469 × 103 **6.489 × 103 **2.760 × 104 **4.334 × 10−1 **9.731 × 10−1 **7.970 × 101 **7.114 × 102 **3.091 × 101 **
Q (20)9.534 × 103 **4.135 × 104 **2.508 × 103 **6.515 × 103 **5.342 × 104 **8.882 × 10−1 **1.374 **1.483 × 102 **7.117 × 102 **6.240 × 101 **
Note: The symbols ** and *** signify the rejection of the null hypothesis at significance levels of 5% and 1%, respectively. The Jarque–Bera test assesses whether the dataset conforms to a normal distribution. The stationarity of the time series is evaluated using the augmented Dickey–Fuller (ADF) test. Q (n) denotes the Ljung–Box Q statistic, with n indicating the lag order, which tests if the sequence data behave as white noise.
Table 3. Six evaluation criteria for the one-step prediction of different prediction models under different rolling windows of SSE.
Table 3. Six evaluation criteria for the one-step prediction of different prediction models under different rolling windows of SSE.
ModelHistory_Length = 5History_Length = 10
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge0.85140.79905.07271.19490.77500.35820.86530.80925.12771.19780.79440.3442
SVR1.19070.73734.11671.08520.48321.09280.75760.48122.59721.12630.04950.7481
RF0.94730.74564.77501.03090.70630.63610.93920.75344.67801.06290.72320.6010
XGBoost0.70770.45772.59460.87850.31610.63590.89790.55292.75551.01210.34330.8280
LSTM0.54710.35771.54600.66430.17860.51710.59950.45782.71160.91530.36240.4775
GRU0.63020.59013.95891.07690.52610.34690.70100.65844.51201.12210.59590.3693
Transformer0.41310.34242.29780.8372−0.30800.35510.36990.20030.62330.60330.32180.3394
ModelHistory_Length = 20History_Length = 30
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge0.84110.78454.96551.18310.76970.33430.81540.75754.80681.16760.73520.3789
SVR0.68450.48162.75481.04740.23360.59830.68950.53763.35031.10150.32640.5982
RF0.74500.62014.10441.01610.56840.47020.78740.62844.19530.98760.58640.5481
XGBoost0.68090.56473.61781.02670.49610.33140.70400.41872.15380.82590.26000.6638
LSTM0.43920.34722.16100.81570.27140.34530.62390.53593.38231.00710.46050.4393
GRU0.60180.46231.72101.81650.40570.29090.57580.49652.12672.0000−0.52300.3618
Transformer0.31110.19300.98880.6056−0.17190.32120.55160.51693.64971.0424−0.29310.3616
Note: The number underlined in bold indicates the optimal value under the column criteria.
Table 4. DM test of one-step prediction error of different models under SSE.
Table 4. DM test of one-step prediction error of different models under SSE.
ModelForecasting Horizon = 1 Days
SVRRFXGBoostLSTMGRUTransformer
Ridge13.834 ***10.694 ***21.120 ***21.752 ***17.025 ***19.795 ***
SVR −7.567 ***−6.016 ***4.635 ***6.469 ***11.102 ***
RF 5.622 ***13.211 ***11.717 ***14.627 ***
XGBoost 15.556 ***12.662 ***16.868 ***
LSTM −5.080 ***12.423 ***
GRU 16.658 ***
Note: *** means p < 0.001.
Table 5. Six evaluation criteria of different prediction models under different prediction steps of SSE.
Table 5. Six evaluation criteria of different prediction models under different prediction steps of SSE.
ModelForecasting Horizon = 5 DaysForecasting Horizon = 10 Days
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge0.89240.83270.89240.83270.81260.36600.92590.86205.55431.22040.83860.3821
SVR0.78860.59570.78860.59570.31860.68750.86000.65733.62911.28800.04460.8358
RF0.65510.54210.65510.54210.48580.41470.89040.75535.16841.10610.72080.5126
XGBoost0.67970.58620.67970.58620.52320.37670.74040.62724.16631.06370.56010.4093
LSTM0.44800.32460.44800.3246−0.16190.43760.74190.61994.21251.04430.29590.5737
GRU0.33540.20640.33540.20640.09630.29380.34060.26331.74690.75470.22910.3201
Transformer0.40590.36000.40590.3600−0.13630.29260.32560.26091.79740.75120.11640.2945
ModelForecasting Horizon = 15 DaysForecasting Horizon = 20 Days
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge0.95460.88690.95460.88690.86270.39191.00560.93826.20191.25230.92380.3944
SVR1.14370.85611.14370.85610.73640.85301.49961.14037.40491.22621.09191.0472
RF0.95120.77820.95120.77820.73920.59181.00970.81455.87591.10670.77880.6397
XGBoost0.74690.63740.74690.63740.56840.39900.91330.80295.60361.15700.79340.4662
LSTM0.44580.34730.44580.34730.21330.39140.80650.74345.10191.18050.73520.3435
GRU0.43630.32170.43630.32170.72620.30760.54470.50493.54381.03420.45270.3030
Transformer0.54240.48940.54240.48940.50540.5300.50350.46723.33431.00510.41060.2937
Note: The number underlined in bold indicates the optimal value under the column criteria.
Table 6. DM test of multi-step prediction error of different models under SSE.
Table 6. DM test of multi-step prediction error of different models under SSE.
ModelForecasting Horizon = 5 DaysForecasting Horizon = 10 Days
SVRRFXGBoostLSTMGRUTransformerSVRRFXGBoostLSTMGRUTransformer
Ridge10.500 ***17.128 ***19.987 ***19.290 ***20.980 ***21.771 ***9.374 ***3.677 ***18.705 ***15.830 ***22.466 ***22.042 ***
SVR −0.482−2.399 *10.546 ***14.675 ***15.280 *** −6.391 ***−2.582 *4.789 ***11.942 ***6.796 ***
RF −3.025 **11.632 ***16.316 ***16.892 *** 9.237 ***11.873 ***16.801 ***14.633 ***
XGBoost 13.442 ***16.721 ***17.506 *** 9.633 ***16.385 ***13.222 ***
LSTM 11.020 ***11.793 *** 6.427 ***0.534
GRU 4.265 *** −17.575 ***
ModelForecasting Horizon = 15 DaysForecasting Horizon = 20 Days
SVRRFXGBoostLSTMGRUTransformerSVRRFXGBoostLSTMGRUTransformer
Ridge2.207 *1.80519.174 ***24.740 ***11.674 ***14.263 ***−3.337 ***1.7624.416 ***17.258 ***19.208 ***19.372 ***
SVR −1.0504.321 ***11.670 ***1.1485.932 *** 3.854 ***4.218 ***5.914 ***8.690 ***8.865 ***
RF 8.652 ***14.211 ***2.931 **9.234 *** 1.8354.223 ***9.674 ***10.480 ***
XGBoost 16.356 ***−9.619 ***1.615 4.582 ***12.233 ***13.527 ***
LSTM −23.398 ***−11.819 *** 17.565 ***18.276 ***
GRU 6.471 *** 7.238 ***
Note: *** means p < 0.001, ** means p < 0.01, and * means p < 0.05.
Table 7. Six evaluation criteria for one-step prediction of the Transformer model with or without the sentiment factor under different history lengths.
Table 7. Six evaluation criteria for one-step prediction of the Transformer model with or without the sentiment factor under different history lengths.
IndicatorHistory_Length = 5History_Length = 10
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Excluding Sent0.41310.34242.29780.8372−0.30800.35510.36990.20030.62330.6033−0.32180.3394
Include Sent0.36380.21770.87010.6824−0.19400.35050.35670.20590.94380.5981−0.11870.3361
ModelHistory_Length = 20History_Length = 30
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Excluding Sent0.31110.19300.98880.6056−0.17190.32120.55160.51693.64971.0424−0.29310.3616
Include Sent0.37160.19210.74770.5822−0.16210.30220.31670.18600.93040.5928−0.15780.3004
Note: The bold numbers represent the optimal values under the column criteria.
Table 8. Six evaluation criteria for the Transformer model with or without the emotion factor under different prediction steps of SSE.
Table 8. Six evaluation criteria for the Transformer model with or without the emotion factor under different prediction steps of SSE.
IndicatorForecasting Horizon = 5 DaysForecasting Horizon = 10 Days
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Excluding Sent0.40590.36002.55530.8845−0.13630.29260.32560.26091.79740.75120.11640.2945
Include Sent0.31250.18580.98780.5955−0.16070.31220.30220.17620.90180.5855−0.11570.2312
ModelForecasting Horizon = 15 DaysForecasting Horizon = 20 Days
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Excluding Sent0.54240.48943.54121.00510.50540.53060.50350.46723.33431.00510.41060.2937
Include Sent0.30110.16870.81200.5677−0.17570.29570.45570.24471.49450.6510−0.15240.3539
Note: The bold numbers represent the optimal values under the column criteria.
Table 9. Six evaluation criteria for the one-step prediction of different prediction models under different rolling windows of the S&P500.
Table 9. Six evaluation criteria for the one-step prediction of different prediction models under different rolling windows of the S&P500.
ModelHistory_Length = 5History_Length = 10
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge1.93080.737712.36761.6058−0.68901.80372.57711.590987.89441.27981.07722.3411
SVR1.66690.861542.49901.14000.39231.62011.96690.573317.36140.9351−0.30101.9437
RF1.89790.545714.80200.8753−0.23181.88371.86930.533414.39860.8627−0.18721.8599
XGBoost1.96710.574317.27080.9352−0.30401.94341.96690.572317.36140.9251−0.31981.9437
LSTM1.77850.556110.49121.1663−0.34741.74421.90360.528113.85280.8624−0.31751.8770
GRU1.82620.581318.64850.9426−0.20941.81421.78650.54898.30701.1972−0.39301.7427
Transformer1.57140.524420.40760.8489−0.06301.57011.73720.48208.15300.8397−0.27271.7156
ModelHistory_Length = 20History_Length = 30
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Ridge2.14241.262767.24471.27510.53422.07481.97140.574317.35730.9353−0.30331.9479
SVR1.96910.574017.37650.9354−0.30201.94581.77000.825239.60981.11210.24911.7524
RF1.98460.561613.06510.9261−0.39061.94581.93260.527712.01030.8531−0.28881.9109
XGBoost1.96910.574017.37650.9354−0.30201.94581.97140.574317.35730.9353−0.30331.9479
LSTM1.84270.52905.41010.9236−0.35711.80781.85470.538118.17480.8850−0.21291.8424
GRU1.96820.55498.53470.9707−0.45311.91531.77040.613912.84141.4208−0.38631.7277
Transformer1.51650.540424.08070.86260.09131.51381.62400.508710.47130.8628−0.12641.6191
Note: The number underlined in bold indicates the optimal value under the column criteria.
Table 10. Six evaluation criteria for the one-step prediction of different types of prediction models under different rolling windows of SSE.
Table 10. Six evaluation criteria for the one-step prediction of different types of prediction models under different rolling windows of SSE.
ModelHistory_Length = 5History_Length = 10
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Transformer0.41310.34242.29780.8372−0.30800.35510.36990.20030.62330.6033−0.32180.3394
Stacked-ML0.32620.20391.02570.61190.00230.32620.30120.19500.98710.60810.00150.3012
Stacked-ML-DL0.30840.19420.98020.57380.01410.30810.29960.19540.98950.60690.00290.2996
ModelHistory_Length = 20History_Length = 30
RMSEMAEMAPESMAPEMBESDRMSEMAEMAPESMAPEMBESD
Transformer0.31110.19300.98880.6056−0.17190.32120.55160.51693.64971.0424−0.29310.3616
Stacked-ML0.28260.18300.94670.5766−0.00220.28260.33270.19981.03070.5782−0.00050.3327
Stacked-ML-DL0.28100.17950.89970.5657−0.00870.28080.32800.20181.02760.59300.00630.3280
Note: The bold numbers represent the optimal values under the column criteria.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, Y.; Zhang, Y.; Ning, P.; Peng, J.; Kao, C.; Hao, L. Transformer-Based Downside Risk Forecasting: A Data-Driven Approach with Realized Downward Semi-Variance. Mathematics 2025, 13, 1260. https://doi.org/10.3390/math13081260

AMA Style

Song Y, Zhang Y, Ning P, Peng J, Kao C, Hao L. Transformer-Based Downside Risk Forecasting: A Data-Driven Approach with Realized Downward Semi-Variance. Mathematics. 2025; 13(8):1260. https://doi.org/10.3390/math13081260

Chicago/Turabian Style

Song, Yuping, Yuetong Zhang, Po Ning, Jiayi Peng, Chunyu Kao, and Liang Hao. 2025. "Transformer-Based Downside Risk Forecasting: A Data-Driven Approach with Realized Downward Semi-Variance" Mathematics 13, no. 8: 1260. https://doi.org/10.3390/math13081260

APA Style

Song, Y., Zhang, Y., Ning, P., Peng, J., Kao, C., & Hao, L. (2025). Transformer-Based Downside Risk Forecasting: A Data-Driven Approach with Realized Downward Semi-Variance. Mathematics, 13(8), 1260. https://doi.org/10.3390/math13081260

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop