Next Article in Journal
Research on Dynamic Decision-Making of Shared Supply Chain Considering Multi-Agent Altruistic Preferences
Previous Article in Journal
Incorporating Prior Information in Latent Structures Identification for Panel Data Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect

School of Statistics and Data Science, Nanjing Audit University, No. 86 Yushan Western Road, Nanjing 211815, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(9), 1506; https://doi.org/10.3390/math13091506
Submission received: 1 April 2025 / Revised: 29 April 2025 / Accepted: 30 April 2025 / Published: 2 May 2025
(This article belongs to the Section E5: Financial Mathematics)

Abstract

:
To describe the stylized features of volatility comprehensively, this paper embeds the time-varying leverage effect of volatility into the Realized Generalized AutoRegressive Conditional Heteroskedasticity (RG) model and proposes a new volatility model with a time-varying leverage effect. The Quasi-Maximum Likelihood-Kernel (QML-K) method is proposed to approximate the density function of returns and to estimate the parameters in the new model. Under some mild regularity conditions, the asymptotic properties of the resulting estimators are achieved. Simulation studies demonstrate that the proposed model yields better performances than traditional RG models under different situations. Finally, the empirical analysis shows better finite sample performance of the estimation method and the new model on real data compared with existing methods.

1. Introduction

The financial market is constantly changing with both returns and risks due to the existence of various unstable factors. In this context, volatility has become one of the most important tools to evaluate the risks caused by these factors. Therefore, volatility plays a crucial role in the research of financial markets and attracts widespread attention from many scholars. Thus, much literature has developed various time series models to measure volatility. One of the popular existing research areas in volatility modeling is the GARCH model proposed by Bollerslev [1]. However, the traditional GARCH family models were based on only low-frequency data, which means the models are not well suited for situations where volatility rapidly shifts to a new level. As Andersen et al. [2] pointed out, this is because a GARCH model is slow at ’catching up,’ and it may require more periods for the conditional variance (implied by the GARCH model) to attain its new level.
It is worth mentioning the Realized GARCH (RG) model proposed by Hansen et al. [3], which can more effectively capture market changes by utilizing high-frequency data. Numerous studies have extensively investigated the RG model and yielded significant results. Gerlach and Wang [4] proposed the realized range, which is observed at a 5 min frequency, as the realized measure, thus improving the out-of-sample predictive likelihood and the forecasting performance of Value at Risk (VaR) and Expected Shortfall (ES). Wang et al. [5] considered the two-sided Weibull distribution in tail risk forecasting for index and asset returns and confirmed the proposed model had better performance than the existing models. A recent study considers the use of the Itô process in the GARCH model to deal with high-frequency financial data, such as Song et al. [6], Fu et al. [7] and Fu et al. [8].
In addition, the RG model takes the existence of the leverage effect in the financial market into account. Black [9] pointed out that based on financial market information, the same degrees of positive and negative news lead to varying degrees of ups and downs in a company’s assets, which may affect stock prices and cause asymmetric fluctuations. This phenomenon is called the leverage effect. Cheung and NG [10], Poon and Taylor [11], Koutmos [12], etc., have conducted sufficient research on this effect and confirmed that it widely presents in mature financial markets. The applications of the leverage effect in high-frequency data have also been made. Wang and Mykland [13] explored the estimation method of the leverage effect in high-frequency data; Aït-Sahalia, Fan, Laeven et al. [14] focused on the leverage effect and proposed a kind of nonparametric estimation for stochastic leverage measures using high-frequency data; Chong and Todorov [15] proposed model-free (nonparametric) estimators of the volatility of volatility and leverage effect using high-frequency observations of short-dated options. The above literuare shows that the leverage effect plays an important role in volatility research. Considering this effect, Hansen et al. [3] introduced a leverage function in the RG model to better capture the leverage effect.
However, since the above studies only considered the constant leverage effect, few of them have discussed the time-varying leverage effect. Existing research showed that the time-varying leverage effect actually exists. Yu [16] developed a nonparametric approach to estimating leverage, which depends on jumps in stochastic volatility models, and found economically significant time-varying leverage that is more negative at higher variance levels. Wu and Wang [17] proposed the Realized Stochastic Volatility model with a Time-Varying Leverage effect (RSV-TVL) and demonstrated the effect of their model with empirical applications. Catania and Proietti [18] proposed a bivariate model that captures key characteristics of realized volatility and asset returns, including long-term memory, heavy-tailedness, and negative dependence, and demonstrated its out-of-sample forecasting performance outperforms the benchmark HAR-GARCH model.
Although GARCH-family models are widely used, the stochastic volatility (SV) framework provides distinct advantages, such as modeling latent volatility dynamics. The SV model conceptualizes volatility as an unobservable latent process driven by stochastic shocks, enabling it to capture complex market dynamics more effectively than purely deterministic specifications like GARCH. Unlike GARCH models that condition volatility solely on historical information, SV models incorporate exogenous stochastic innovations, enhancing their flexibility in describing financial time series. However, the latent nature of volatility in SV specifications introduces significant estimation challenges, particularly regarding state variable identification and parameter inference. Despite these computational complexities, SV models remain indispensable in modern financial econometrics due to their theoretical coherence and empirical accuracy. The SV framework has inspired extensive methodological developments: Tauchen and Pitts [19] and Taylor [20] pioneered the application of stochastic principles to financial volatility modeling; Chib, Nardari, and Shephard [21] advanced Bayesian estimation techniques for high-dimensional multivariate SV models with time-varying correlations; Jensen and Maheu [22] introduced a semiparametric Bayesian approach incorporating Markov chain Monte Carlo methods to address distributional uncertainty; Fernández-Villaverde, Guerrón-Quintana, and Rubio-Ramírez [23] developed computationally efficient particle filtering algorithms tailored for large-scale SV models. Recent innovations continue to expand the SV paradigm, as evidenced by contributions from Rømer [24], Yazdani, Hadizadeh, and Fakoor [25], Bolko, Christensen, Pakkanen et al. [26], and Chan [27], among others. Notwithstanding these advancements, SV models remain computationally intensive, particularly for parameter estimation and short-term forecasting. This computational burden renders them better suited for long-term dynamic volatility analysis than high-frequency applications. Given our focus on high-frequency trading risk management, we prioritize GARCH-family models for their tractability, closed-form estimation, and proven efficacy in operational settings.
Unlike SV models that require computationally intensive latent variable estimation, our proposed GARCH-family model is more easily calculated, making it more suitable for high-frequency risk management. In this paper, we introduce a time-varying leverage function with a linear spline representation to the RG model, thus constructing the Realized GARCH model with the Time-Varying Leverage effect (RG-TVL) model. Different from the leverage function in the RG model, the new function is sufficient to reflect the time-varying leverage effect. Besides, due to the complexity of the market data, the error terms usually do not follow a normal distribution but a kind of heavy tail distribution. Thus, the traditional QML method is not effective enough. A QML estimation method based on the Kernel density estimator (QML-K) is used to address this issue. The consistency and asymptotic normality of QML-K estimators are proved in this paper. Simulation studies and applications indicate that (1) The RG-TVL model has better performance in capturing the time-varying leverage effect than the RG model; (2) The QML-K method is more robust than the QML for the finite sample estimation. Thus, our paper has important theoretical significance and application value for financial risk prediction and macroeconomic market regulation.
The rest of this paper is organized as follows. In Section 2, we introduce the structure of the RG model with the time-varying effect. In Section 3, a semiparametric method is described. The convergence properties of the method and the related optimization algorithm are also proposed. In Section 4, we perform some simulation studies. In Section 5, empirical data are analyzed. In Section 6, we provide a conclusion and some comments.

2. Realized GARCH with Time-Varying Leverage Effect

2.1. The Model Construction

Hansen et al. [3] introduced the RG model, a novel framework that integrates a GARCH structure for returns with a joint modeling approach for realized volatility measures. By effectively incorporating high-frequency data, the model demonstrates superior performance compared with conventional low-frequency time series models. Moreover, the leverage function embedded in the model effectively captures the asymmetric volatility response observed in financial markets, thereby enhancing its ability to model the leverage effect.
The mathematical form of the RG( p , q ) model is as follows:
r t = μ + h t 1 / 2 z t ,
log h t = α + i = 1 p β i log h t i + j = 1 q γ j log x t j ,
log x t = ξ + ϕ log h t + τ ( z t ) + u t ,
where r t denotes the logarithmic return at time t, and  μ and h t represent the conditional expectation and variance of r t , respectively. The leverage effect is captured by the function τ ( z t ) , where z t is the error term of the logarithmic return. Additionally, x t stands for the realized measure. In the RG( p , q ) model, the leverage function takes the form τ ( z t ) = τ 1 z t + τ 2 ( z t 2 1 ) . Finally, u t corresponds to the error term of the realized measure. Furthermore, z t i . i . d . ( 0 , 1 ) and u t i . i . d . N ( 0 , σ u 2 ) . In order to better consider the property, we assume a mild condition that z t follows a symmetric distribution. In addition, we require 1 < i = 1 p β i + ϕ j = 1 q γ j < 1 to ensure the stability of the model.
However, a global linear relationship between τ ( z t ) and z t in the RG model may be too strict to describe the time variety of the leverage effect. To solve this problem, inspired by Yu [16], we insert a linear spline leverage function into the measure equation. The expression of the linear spline leverage function is as follows:
τ ( z t ) = τ 1 z t + i = 1 n τ 2 i ( z t 2 1 ) I ( b i 1 z t < b i ) ,
where ( b 0 , b 1 , , b n ) is a monotonically increasing sequence, n represents the number of intervals and there are n + 1 nodes. I is the indicative function. Therefore, the leverage function discussed in this article is a generalization of Hansen et al. [3]. The form of RG-TVL( p , q ) model consists of Equations (1)–(4).
Remark 1.
In the proposed model, we need to determine n and b i s. Yu [16] suggested that n needs to increase with the growth of the sample size; however, he also pointed out that the larger n indicates more parameters in the model and hence leads to higher computational cost. Essentially, increasing n results in a tradeoff where a smaller bias is achieved at the cost of a larger variance. The reason is that fewer effective observations are utilized to estimate τ 2 i . To control the computational costs, we assume n = 2 for the model construction and asymptotic properties. The form of the leverage function is as follows:
τ ( z t ) = τ 1 z t + τ 21 ( z t 2 1 ) I ( z t 0 ) + τ 22 ( z t 2 1 ) I ( z t > 0 ) .
The situation of n = 3 has also been considered in the simulation study, which has no significant advantage compared. Since this does not affect the analysis of the estimators’ properties, we take n = 2 as an example in the asymptotic properties and Appendix A sections.
According to Equation (1), since h t 1 / 2 > 0 , if  r t μ 0 , then z t 0 , the leverage function in Equation (5) is τ ( z t ) = τ 1 z t + τ 12 ( z t 2 1 ) ; otherwise, z t > 0 , the leverage function is τ ( z t ) = τ 1 z t + τ 22 ( z t 2 1 ) . When τ 21 τ 22 , the values of the leverage function will be different, which reflects the time variety of the leverage effect. Under the assumptions of Hansen et al. [3], E [ τ ( z t ) ] = 0 holds, which also holds in our leverage function to make sure the equation is stable. Compared with the RG model, the RG-TVL model fully considers the different values that may exist in the coefficients of the leverage function under two situations. In addition, if  τ 21 = τ 22 , the model degenerates into an RG model, where the leverage effect parameters are the same across different periods and cannot reflect the time-variety nature of the leverage effect. When the realized measure is replaced by the lag volatility term, the model degenerates into a GARCH model, which ignores high-frequency data and the leverage effect.

2.2. Strict Stationarity

After proposing the model, we then discuss the sufficient and necessary conditions for the existence of a strict stationary solution of the RG-TVL ( p , q ) model. Considering an ARMA representation as follows:
log h t = α 1 + i = 1 p β i log h t i + ϕ j = 1 q γ j log h t j + j = 1 q γ j η t j ,
where α 1 = α + ξ j = 1 q γ j and η t = τ ( z t ) + u t . In Equation (6), the second and the third terms represent the AR terms, and the fourth term represents the MA term. Furthermore, when applying the ARMA representation, the absolute value of unit roots of AR terms should be less than 1 to ensure the stationarity of the model.
Denote θ as the parameter space of θ = ( θ 1 , θ 2 , σ u 2 ) , where θ 1 = ( α , β 1 , , β p , γ 1 , , γ q ) and θ 2 = ( ξ , ϕ , τ 1 , τ 21 , τ 22 ) . Without loss of generality, we assume Θ = Θ 1 × Θ 2 × Θ 3 , where Θ 1 , Θ 2 , Θ 3 represent the space of θ 1 , θ 2 , σ u 2 , respectively. For the convenience of analysis, let m = p q , A θ 1 ( s ) = j = 1 q γ j s j and B θ ( s ) = 1 i = 1 m ( β i + ϕ γ i ) s i . Then Equation (6) can be written with backshift operator notation as follows:
B θ ( B ) log h t = α 1 + A θ 1 ( B ) η t ,
where B is the backshift operator. Thus, we have the following propositions to ensure the strict stationarity of RG-TVL ( p , q ) model, which follows from the discussion of the classical ARMA ( p , q ) time series Brockwell and Davis [28].
Proposition 1.
If A θ 1 ( s ) and B θ ( s ) have no common roots, then the stationary solution log h t exists if and only if B θ ( s ) 0 for all s C such that | s | 1 .
Then, let C θ 1 ( s ) = 1 i = 1 p β i s i . Thus, Equation (2) can be rewritten as
C θ 1 ( B ) log h t = α + A θ 1 ( B ) log ( x t )
Proposition 2.
Denote log + ( s ) as maximum ( log ( s ) , 0 ) . Suppose log ( x t ) is a stationary and ergodic process with E [ log + ( | log ( x t ) | ) ] < . Meanwhile, assume A θ 1 ( x ) and C θ 1 ( x ) have no common roots. If  C θ 1 ( x ) 0 for all x C such that | x | 1 , then the solution log h t of the RG-TVL ( p , q ) model has a linear representation in terms of log ( x t ) .
Remark 2.
The condition E [ log + ( | log x t | ) ] < is a very weak condition. By Lemma 2.5.3 in Straumann [29], E | log x t | d < for some d > 0 indicates that E [ log + ( | log x t | ) ] < .
With the argument above, we can make sure that log h t is stationary and ergodic. Since η t are i.i.d. random variables, { ( r t , log x t ) } are stationary and ergodic. This conclusion will be important in the estimation of the RG-TVL ( p , q ) model.

3. Semiparametric Estimation

Quasi-Maximum Likelihood (QML) estimation has been widely applied in fields of spatial econometrics, dynamic panel data models, nonlinear models, etc. The main idea is to construct an approximate likelihood function, which can still obtain a consistent and asymptotically normal estimator in the case of mild conditions. Pfaff [30] considered the quick consistency of the QML estimator and proposed its exponential convergence rate under general conditions; Lee and Hansen [31] applied the QML method to the GARCH(1,1) model, thus constructing the consistency and asymptotic normality; Since then, the QML method has been utilized in various econometric models; see Komunjer I [32], Lee L. F. [33], Fan J., Qi L., and Xiu D. [34], etc.
However, an important issue for estimation is that the distribution of z t is unknown in the RG-TVL model. The QML method pretends a standard normal distribution for the solution, which might lead to a certain degree of deviation. Fang and Han [35] and Wang et al. [36] performed a series of experiments that confirmed that inserting the Kernel density estimator in the QML method can effectively improve estimation performance. Referring to Wang et al. [36]’s method, we apply the QML-K method in our model.

3.1. QML-K Estimation

With the RG-TVL model, the likelihood function is as follows:
L T ( θ | r , x ) = 1 T t = 1 T l t ( θ ) ,
where T is the sample size of the data, r and x represent the sequences of r t and x t , respectively; l t ( θ ) is as follows:
l t ( θ ) = 1 2 2 log f z ( z t ) + log h t + log ( σ u 2 ) + u t 2 σ u 2 ,
where f z ( z ) is the probability density function of z t .
We apply the Quasi-Maximum Likelihood estimation to gain the estimators. Denote the true parameter as θ 0 = ( θ 1 , 0 , θ 2 , 0 , σ u , 0 2 ) . Rewrtie the variance equation, replace log ( h t ) by log h ˜ t ( θ 1 ) and we have
log h ˜ t ( θ 1 ) = α + i = 1 p β i log h ˜ t i ( θ 1 ) + j = 1 q γ j log x t j .
Here the function log h ˜ t ( θ 1 ) = log h t when θ 1 = θ 1 , 0 . Since the recursive way may not well define log h t ( θ 1 ) , the following is assumed:
Assumption 1.
A θ 1 ( s ) and C θ 1 ( s ) do not have any common roots, and  | C θ 1 ( s ) | = 0 as | s | > 1 , for any θ 1 Θ 1 .
If log x t is stationary and ergodic with E [ log + ( | log ( x t ) | ) ] < and Assumption 1 holds, then we can define
log h ˜ t ( θ 1 ) : = C θ 1 1 ( B ) α + A θ 1 log x t , θ 1 Θ 1 .
Similarly, we can define u ˜ t ( θ 1 , θ 2 ) as follows:
u ˜ t ( θ 1 , θ 2 ) = log x t ξ ϕ log h ˜ t ( θ 1 ) τ r t μ h ˜ t 1 / 2 ( θ 1 ) ,
and u ˜ t ( θ 1 , 0 , θ 2 , 0 ) = u t . log h ˜ t ( θ 1 ) and u ˜ ( θ 1 , θ 2 ) will be used in the likelihood function. Replacing h t and u t by h ˜ t ( θ 1 ) and u ˜ t ( θ 1 , θ 2 ) , we can obtain
L ˜ T ( θ | r , x ) = 1 T t = 1 T l ˜ t ( θ ) ,
where
l ˜ t ( θ ) = 1 2 2 log f z ( z t ) + log h ˜ t ( θ 1 ) + log ( σ u 2 ) + u ˜ t ( θ 1 , θ 2 ) 2 σ u 2 .
Note that h ˜ t ( θ 1 ) and u t will be used in the likelihood function. We apply an alternating iterative approach to estimate the two sequences. log h ^ t can be recursively defined as follows:
log h ^ t ( θ 1 ) = log h ^ 0 , t = 0 α + i = 1 q β i log h ^ t i ( θ 1 ) + j = 1 q γ log x t j , t > 0 ,
Thus, u ^ t can be defined as follows:
u ^ t ( θ 1 , θ 2 ) = log x t ξ ϕ log h ^ t ( θ 1 ) τ r t μ h ^ t 1 / 2 ( θ 1 ) .
Based on the conditional distribution of r t and x t , the logarithmic likelihood function of the model can be described as
L ˇ T ( θ | r , x ) = 1 T t = 1 T l ˇ t ( θ ) ,
where
l ˇ t ( θ ) = 1 2 2 log f z ( z t ) + log h ^ t ( θ 1 ) + log ( σ u 2 ) + u ^ t ( θ 1 , θ 2 ) 2 σ u 2 .
Since the distribution of z t is unknown, the probability density function of z t needs to be estimated. We apply the Kernel density method to estimate the probability density function of z t . The form of the estimator is as follows:
f ^ z , T ( z t ) = 1 T b T i = 1 T K ( z t z ^ i b T ) ,
where f ^ z , T ( z t ) is the estimator of f z ( z t ) , b T is the bandwidth of Kernel estimation, T is the sample size, z ^ t = ( r t μ ) / h ^ t 1 / 2 and K ( · ) is a Gaussian Kernel function:
K ( s ) = 1 2 π e s 2 2 .
The size of the bandwidth affects the accuracy of Kernel density estimation. In this section, we follow Silverman’s (1978) idea and select a fixed bandwidth suitable for Gaussian Kernel functions:
b T = ( 4 3 T ) 1 5 σ ^ z 1.06 σ ^ z T 1 5 ,
where σ ^ z 2 is the variance of z ˜ t and the estimator of V a r ( z t ) . By substituting f ^ z ( z t ) for f z ( z t ) in L ˇ T , we have
L ^ T ( θ | r , x ) = 1 T t = 1 T l ^ t ( θ ) ,
where
l ^ t ( θ ) = 1 2 2 log f ^ z ( z t ) + log h ^ t ( θ 1 ) + log ( σ u 2 ) + u ^ t ( θ 1 , θ 2 ) 2 σ u 2 .
We will refer to this method as the QML Kernel method, abbreviated as QML-K, in the following text. The estimator of QML-K is defined as follows:
θ ^ T = arg max θ L ^ T .

3.2. Asymptotic Properties

In this subsection, we focus on the asymptotic properties of the QML-K estimator.
Assumption 2.
For any real number ( τ ˜ 1 , τ ˜ 21 , τ ˜ 22 ) ( 0 , 0 , 0 ) , it is obvious that τ ˜ ( z t ) = τ ˜ 1 z t + τ ˜ 21 ( z t 2 1 ) I ( z t 0 ) + τ ˜ 22 ( z t 2 1 ) I ( z t > 0 ) does not degenerate, for any t Z .
Assumption 3.
A θ 1 ( s ) and B θ ( s ) have no common zeros, and  B θ ( s ) 0 for all s C such that | s | 1 .
Assumption 4.
There exists s > 0 such that E [ exp ( s | log x t | ) ] < .
Remark 3.
Assumption 2 makes sure the distribution of z t is not concentrated at two points, which means the RG-TVL( p , q ) model is identifiable. Assumption 3 shows the future-independence of the model. Assumption 4 on | log x t | indicates that the density function of x t around 0 does not explode too fast.
Theorem 1.
Suppose that log h t is a stationary RG-TVL ( p , q ) process defined by Equations (1)–(4) with parameters θ 0 . If  θ 0 Θ , where Θ is compact, and Assumptions 1–3 hold, then we have
θ ^ T a . s . θ 0 as T .
Then, we consider the asymptotic normality. For ease of explanation, we use the following notation:
E 0 [ · ] = E [ · | θ = θ 0 ] , 1 = / θ 1 , 2 = / θ 2 , and 12 = / ( θ 1 , θ 2 ) .
And define
A = E 0 [ 1 ( log h ˜ t ) 1 ( log h ˜ t ) ] , F = E [ z t 4 3 ] A , B 11 = 2 σ u , 0 2 E 0 [ 1 u ˜ t 1 u ˜ t ] , B 21 = 2 σ u , 0 2 E 0 [ 2 u ˜ t 1 u ˜ t ] , and B 22 = 2 σ u , 0 2 E 0 [ 2 u ˜ t 2 u ˜ t ] .
Theorem 2.
Suppose that the assumptions in Theorem 1 and Assumption 4 are satisfied. If  θ 0 is an interior point of Θ , E | z t | 4 < and E | u t | 4 < , then we have
T b T ( θ ^ T θ 0 ) d N ( 0 , V ) ,
where
V = V 11 0 0 0 0 0 0 0 0 + b T H 1 J H 1 ,
V 11 is a bounded unknown variable that is discussed in Appendix A.
J = F + 2 A + 2 B 11 2 B 21 2 E [ u t 3 ] σ u , 0 6 E 0 [ 1 u ˜ t ] 2 B 21 2 B 22 2 E [ u t 3 ] σ u , 0 6 E 0 [ 2 u ˜ t ] 2 E [ u t 3 ] σ u , 0 6 E 0 [ 1 u ˜ t ] 2 E [ u t 3 ] σ u , 0 6 E 0 [ 2 u ˜ t ] E [ u t 4 ] ( σ u , 0 4 ) σ u , 0 8 ,
and
H = A + B 11 B 21 0 B 21 B 22 0 0 0 1 σ u , 0 4 .
Remark 4.
The asymptotic covariance matrix V is singular if and only if u t 2 σ u , 0 2 , for all t.
The first term of V represents the variance of the Kernel estimator, and the second represents that of the QML estimator. Since the Kernel method only focuses on z t , which only affects θ 1 in Equation (2), other elements of the matrix are equal to 0 except V 11 .

3.3. Optimization Algorithm

In this subsection, our aim is to search for a θ ^ T = arg max L ^ T ( θ ) . The penalty function method is selected to solve constrained optimization problems. The accuracy of this method is limited by the parameter selection of the augmented objective function; thus, improper parameter selection may lead to lower accuracy of the optimization method.
Considering that the model contains inequality constraints mentioned in Section 2, the interior point method is selected to solve an optimization problem with inequality constraints. Based on the idea of the method, we first need to construct an augmented objective function as follows:
F ( θ | r , x , δ ) = L ^ T ( θ ) + δ Ψ ( θ ) ,
where δ > 0 represents obstacle factors to control the impact degree of Ψ ( θ ) , Ψ ( θ ) > 0 , and as θ approches the feasible domain boundary, Ψ ( θ ) tends towards infinity. A common construction of Ψ ( θ ) is as follows:
Ψ ( θ ) = i = 1 n [ p i ( θ ) ] 1 ,
where p i ( θ ) represents n restrictive conditions. In this paper we have p 1 ( θ ) = i = 1 p β i ϕ j = 1 q γ j 1 and p 2 ( θ ) = i = 1 p β i + ϕ j = 1 q γ j 1 . The parameter estimation results of the model can be obtained by solving the problem:
θ ^ = arg min θ W F ( θ r , x ) ,
where W = { w p i ( w ) < 0 , i = 1 , 2 , 3 , , m } .
Denote θ ^ T , k as the estimator of parameters in the k-th iteration.
With the optimization algorithm (Algorithm 1), we can calculate the estimator of our model. There may be a more effective algorithm than we used in this paper; however, we do not focus on the optimization algorithm. More research and exploration can be conducted.
Algorithm 1: The interior point method
Mathematics 13 01506 i001

4. Simulation Study

In this section, we aim to rigorously verify and assess the fitting effectiveness of the RG-TVL model utilizing QML and QML-K. The simulation is conducted with different control variables: estimation methods, sample sizes, and distributions of z t . By analyzing the simulations, we hope to gain deeper insights into the robustness and adaptability of the RG-TVL model across various scenarios and conditions, thus contributing to its better understanding and potential applications in various fields.
The evaluation criteria in the simulation study are bias and mean square error (MSE). In the simulation study and application, we take the RG-TVL(1,1) model as a sample. Additionally, since μ in Equation (6) has a negligible effect on parameter estimation, we assume it to be zero. A smaller absolute value of bias shows the higher the accuracy of parameter estimation, and a smaller MSE indicates stronger stability of parameter estimation. In this section, we mainly pay attention to MSE.
Under the restrictive condition in Section 2, some true parameters are set as follows:
α , β , γ , ξ , ϕ , τ 1 , τ 21 , τ 22 , σ u , h 0 = 0.2 , 0.7 , 0.25 , 0.1 , 0.9 , 0.5 , 0.5 , 0.2 , 0.1 , 1.0 .
To better describe the degree of leverage effect, we use the correlation between τ ( z t ) and z t as the indicator, which is called the leverage coefficient and noted as ρ . The value of ρ is between −1 and 1, and it can reflect a leverage effect if it is negative. The value of ρ close to −1 means a strong leverage effect. The formula for ρ is as follows:
ρ = C o r r [ τ ( z t ) , z t ] = C o v [ τ ( z t ) , z t ] V a r [ τ ( z t ) ] V a r ( z t ) = τ 1 + ( τ 21 τ 22 ) c 1 τ 1 2 + 1 2 ( τ 21 2 + τ 22 2 ) c 2 + 2 τ 1 ( τ 21 τ 22 ) c 1 ,
where c 1 = 0 ( z 3 z ) f ( z ) d z , c 2 = ( z 2 1 ) 2 f ( z ) d z and f ( z ) represents the probability distribution function of z t .
By Equation (9), if z t follows a standard normal distribution and a standardized t ( 4 ) distribution, we have ρ = 0.864 and ρ = 0.757 , respectively. Both values show a relatively strong leverage effect. The results are shown as follows:
In Table 1, when the sample size is equal to 100, the MSEs of α and σ u using QML-K are lower than those using QML, while the other parameters are the opposite. When the sample size grows to 500, the MSEs of β , γ , ϕ , and σ u using QML-K are less than those using QML, while the other parameters are on the contrary; When the sample size reaches 1000, the MSEs of β , γ , ϕ , and σ u are less than those using QML, while the other parameters are in contrast.
In Table 2, when the sample size is equal to 100, the MSEs of β , ξ , τ 1 , τ 21 , and τ 22 using QML-K are less than those using QML, while the other parameters are contrary to the former; when the sample size reaches 500 or 1000, the MSEs of all parameters using QML-K are less than those using QML.
Finally, we are interested in the impact of n in Equation (6) on the estimation. In Section 2, we mentioned that the increasing n may lead to more accuracy but more cost. Thus, we select the Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC) to balance cost and accuracy. In this part, we perform an experiment on which model to choose. Denote the situations n = 2 and n = 3 as Spline1 RG-TVL and Spline2 RG-TVL models, respectively. Yu [16] previously examined alternative node specifications, such as partitioning the interval into two or three segments with nodes at 0 or (−0.4, 0.4), motivated by the empirical properties of the Gaussian distribution. Thus, we set the leverage function in the Spline2 RG-TVL model as
τ ( z t ) = τ 1 z t + ( z t 2 1 ) [ τ 21 I ( z t 0.4 ) + τ 22 I ( 0.4 < z t 0.4 ) + τ 23 I ( z t > 0.4 ) ] .
Compared with Spline1 RG-TVL, this model adds τ 23 in parameters. Then, we generate data from Spline2 RG-TVL and Spline3 RG-TVL models and apply the two models to those data to compare their effects. In this experiment, we set τ 23 = 0.7 , z t N ( 0 , 1 ) , and other parameters the same as above. We conducted 100 estimations and recorded the mean and standard deviation as follows:
It can be seen from the table that the AICs and BICs of Spline1 do not show a significant difference between Spline1 and Spline2 in general.
According to Table 1, Table 2 and Table 3, we can draw conclusions:
(a) As the sample size rises, the MSEs of parameters tend to decrease using both methods and both distributions of z t , which conforms to general statistical theory;
(b) MSEs of parameters for both QML and QML-K are relatively small;
(c) Under the standard normal distribution assumption, the MSEs using QML and QML-K do not show a sigificant difference, while under the other assumption, the MSEs using QML-K are remarkably less than those using QML, especially when the sample size is larger than 500;
(d) There is no evidence to suggest that the Spline2 RG-TVL models have better finite sample performance than the Spline1 model.
The conclusions drawn from (a) and (b) demonstrate that the RG-TVL model exhibits good performance when applied to finite sample scenarios, highlighting its efficacy and reliability within such constrained datasets. Meanwhile, the observations made in (c) imply that, in comparison to the QML method, the QML-K method emerges as a more robust one for estimation purposes. This suggests that QML-K may offer advantages in terms of stability and accuracy, particularly in scenarios where the assumptions underlying QML may be violated. (d) demonstrates that increasing the number of nodes may not significantly improve the estimation performance of the model.

5. Applications

5.1. Data Acquisition and Processing

In this section, we study the applications of the RG-TVL model in the Hang Seng Index (HSI) and Standard & Poor’s 500 Index (S&P 500). The former is a market capitalization-weighted stock index in Hong Kong, tracking the performance of the 50 largest and most liquid companies listed on the Hong Kong Stock Exchange. It serves as an important benchmark for the Hong Kong stock market. The latter is one of the most representative stock indices in the US stock market, consisting of approximately 500 large publicly traded companies and covering about 80% of the total market value of the US stock market. It covers multiple industries such as technology, finance, healthcare, and consumer goods. Due to the quality and diversity of its constituent stocks, it has become a “barometer” of global investors’ attention for the US stock market. The required data were collected from the Wind Financial Terminal.
The date ranges selected are from 16 May 2022 to 17 September 2024, with a total of 568 working days of data (HSI), and from 16 January 2024 to 21 March 2025, with a total of 297 working days of data (S&P 500). The daily logarithmic return rate formula and 5-min logarithmic return rate formulas are
r t = log y t log y t 1 , r t , i = log y t , i log y t , i 1 ,
where y t represents the transaction price at day t and y t , i represents the ith transaction price at day t.
According to Andersen and Bollerslev [37], we use realized volatility (RV) as the realized measure inserted in the RG-TVL model. The form of RV is as follows:
RV t = i = 1 T r i , t 2 .
Note that RV is always positive, which ensures the stability of the model. Corsi et al. [38] have proved that the variable has unbiased and stable characteristics.
Firstly, we conduct descriptive statistics on the data and draw a daily logarithmic return rate of HSI and S&P 500 as shown in Figure 1.
Figure 1 shows that the daily logarithmic return rates of HSI and S&P 500 both have a small degree of change over time. The two sets of data do not show a noteworthy trend of growth or decrease. However, the HSI indicates significant fluctuations in October 2022 and February 2024, while the S&P 500’s fluctuations are in April 2024, September 2024, and January 2025. The RV changes in HSI and S&P 500 are plotted as shown in Figure 2.
From Figure 2, we can obtain that the values of RV are relatively small and had significant fluctuations, which correspond to the changes in daily logarithmic rates of return, indicating that RV can effectively reflect the changes in returns. The descriptive statistics of the daily logarithmic return rate and RV of HSI and S&P500 are shown in Table 4.
Table 4 indicates that, for the HSI, the maximum daily logarithmic return is 4.550 ×   10 2 , while the minimum is −4.600 ×   10 2 , indicating substantial variability in returns. The mean is close to zero, with a positive skewness, suggesting a right-skewed distribution. The kurtosis exceeds 3, reflecting leptokurtic behavior. As a proxy for integrated volatility, the realized volatility (RV) is strictly positive and exhibits a right-skewed distribution with significantly higher kurtosis (>3), consistent with a pronounced leptokurtic and heavy-tailed distribution. For the S&P 500, the daily logarithmic returns range from a maximum of 1.655 ×   10 2 to a minimum of −2.945 × 10−2, demonstrating considerable dispersion. The near-zero mean and negative skewness indicate a left-skewed distribution, while the kurtosis above 3 confirms leptokurtosis. The RV is strictly positive and displays a right-skewed, highly leptokurtic distribution (kurtosis >   3 ), characteristic of a heavy-tailed process. Then, perform normality tests on the samples separately and create Q–Q diagrams as follows:
From Figure 3, it is evident that the daily logarithmic return rate of both indices does not exhibit a straight line, suggesting that it does not follow a normal distribution. To further validate this assumption, the Jarque–Bera test was conducted on the return rate. The results indicate a p-value of 0.001, which further confirms that the return rate does not conform to normal distributions.

5.2. In-Sample Estimation

Next, we compare the performance of two estimation methods in two models (RG and RG-TVL) within the sample. The comparison indicators of the fitting effect are the log-Likelihood function (LL) and the AIC criterion. The estimated results are shown in Table 5.
From Table 5, we can obtain that the estimated results of ρ are all negative and further away from 0, indicating a significant leverage effect in the data. There is a significant difference between τ 21 and τ 22 , reflecting the time variety of the leverage effect. According to the numerical values of AIC and BIC, the order from low to high is RG-TVL-QML-K, RG-TVL-QML RG-QML-K, and RG-QML.

5.3. Out-of-Sample Prediction

The selected period for out-of-sample prediction is a total of 15 working days. We choose MSE, mean absolute error (MAE), and root mean square error (RMSE) to be the standards to measure the effectiveness of out-of-sample prediction. According to the form of the model, the predictive model is as follows:
log h ^ T + i = α ^ + β ^ log h ^ T + i 1 + γ ^ log x ^ T + i 1 ,
log x ^ T + i = ξ ^ + ϕ ^ log h ^ T + i + τ ^ ( z T + i ) + u ^ T + i ,
where θ ^ = ( α ^ , β ^ , γ ^ , ξ ^ , ϕ ^ , τ ^ 1 , τ ^ 21 , τ ^ 22 , σ ^ u ) can be obtained by the result of in-sample estimation; z i can be obtained from standard normal random numbers; u i can be obtained from normal random numbers with an expectation of 0 and a variance of σ u . The multi-step prediction results can be obtained by repeatedly calculating Equations (10) and (11), thus obtaining the predictor of x i .
The results of the prediction are shown in Table 6.
With the results above, the differences between predicted values are relatively small, which demonstrates that both RG and RG-TVL models can well predict the volatility by QML and QML-K.
Overall, the results indicate that the RG-TVL model consistently outperforms the standard RG specification across all error metrics (MSE, MAE, RMSE), highlighting its superior volatility forecasting capability. Furthermore, the QML-K estimator demonstrates enhanced precision compared with QML, reinforcing the importance of incorporating Kernel-based adjustments for robust parameter estimation in volatility modeling.

6. Conclusions and Discussion

We start from the perspective of the time-varying leverage effect and consider the asymmetric characteristics of volatility in the current capital market. Referring to Yu [16], a linear spline function is introduced into the RG model to reflect the time variety of the leverage effect, thus constructing the RG-TVL model. The QML-K method is applied to the RG-TVL model to estimate the parameters. Some asymptotic properties of the estimator have been proved. An experiment with different variables is conducted to explore the fitting effect of the RG-TVL model. Finally, we validate the predictive performance of the model with representative stock index data from Hong Kong.
The complexity and dynamism inherent in the time-varying leverage effect pose significant challenges for accurate measurement and prediction, making it a notoriously elusive phenomenon to capture. Our method offers a reliable and innovative approach to tackling this intricate problem. It is designed specifically to account for the temporal variations in the leverage effect, leveraging advanced statistical and econometric techniques that enable it to adapt dynamically to changes in market conditions. By offering an adaptable solution to the challenging task of capturing the time-varying leverage effect, our method represents a significant step forward in the field of financial risk management.

Author Contributions

Conceptualization, H.H.; Methodology, J.L. and H.H.; Software, Y.M.; Validation, G.L.; Formal analysis, Y.M.; Resources, H.H.; Writing—original draft, Y.M.; Writing—review & editing, G.L.; Supervision, J.L.; Project administration, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Natonal Natural Science Foundaton of China (No. 12371267) and the Postgraduate Research & Practice Innovation Program of Jiangsu Province (No. KYCX24_2342).

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of Theorems

Appendix A.1. Proof of Theorem 1

Under standard arguments, the strong consistency is obtained by showing the following intermediate results:
(i) L ^ T L T Θ a . s . 0 ;
(ii) if h ˜ 1 ( θ 1 ) = h ˜ 1 ( θ 1 , 0 ) , and u ˜ 1 ( θ 1 , θ 2 ) = u ˜ 1 ( θ 1 , 0 , θ 2 , 0 ) a.s., then θ = θ 0 ;
(iii) θ 0 is the unique maximum of L in Θ ;
(iv) any θ θ 0 has a neighborhood V ( θ ) such that lim inf T inf θ * V ( θ ) L ^ T ( θ * ) L ( θ 0 ) , a.s.
Assumption A1.
(a) f z ( z t ) is three times continuously differentiable; (b) 0 < f 1 inf f z ( z t ) sup f z ( z t ) f 2 , where f 1 and f 2 are two positive constants.
Assumption A2.
K : [ 1 , 1 ] R + is bounded, symmetric about 0, Lipschitz continuous, has bounded variation with 1 1 K ( u ) d u = 1 , C m = 1 1 u m K ( u ) d u , lim u K ( u ) = 0 , and u log ( u ) 1 / 2 d K ( u ) < .
Assumption A3.
b T satisfies that b T 0 and T b T log T .
Lemma A1.
With Assumptions A1–A3, the uniformly strong consistency of the Kernel density estimator f ^ z , T ( z t ) can be obtained:
sup z t f ^ z , T ( z t ) f z ( z t ) a . s . 0 , as T .
This property also holds for the derivative function of f ^ z , T ( z t ) and f z ( z t ) .
Remark A1.
Assumption A1 makes sure the function f z ( z ) is smooth and bounded. Assumption A2 has been often used in Kernel density estimation, where the bounded limit on K ( · ) simplifies the analysis. Assumption A3 is stricter conditions b T to ensure the uniform strong consistency of the Kernel estimator.
The proof of Assumption A1 can be found in Silverman [39].
Definition A1.
A sequence { v n } n Z of random elements with values in a normed vector space ( B , · ) is said to converge to zero exponentially fast a.s. when n , denoted by v n e . a . s . 0 , if there exists c > 0 so that c n v n a . s . 0 .
Lemma A2
(Straumann [29], Proposition 2.5.1). Let { ζ n } n Z be a sequence of random variables with ζ n e . a . s . 0 and { v n } n Z be a sequence of identically distributed random elements with values in a normed vector space ( B , · ) . If E [ log + v 0 ] < , then n = 0 ζ n v n converges a.s. and ζ n v n e . a . s . 0 as n .
Lemma A3
(Existence of log-moments). Let { ( r t , log x t ) } be a stationary RG-TVL( p , q ) process with the parameter vector θ 0 . If E | η t , 0 | d < and d 1 , then
(i) E | log h t | d < ;
(ii) E | log x t | d < ;
(iii) Furthermore, if Assumption 1 holds, we have E log h ˜ t ( θ 1 ) Θ 1 d < .
Lemma A4.
log h ^ t and log h ˜ t are random elements defined by Equations (7) and (8). Then, log h ^ t and log h ˜ t are, t Z + , almost surely twice continuously differentiable in the interior of Θ 1 . If E | η t , 0 | d < for some d 1 , then log h ^ t log h ˜ t e . a . s . 0 .
The proof of Lemmas A3 and A4 can be obtained by Lemmas A.2 and A.3 in Li et al. [40].
For intermediate result (i), we have log h ^ t log h ˜ t Θ 1 e . a . s . 0 by Lemma A4.
Lemma A5.
With the conditions in Lemma A4, we have 1 h ^ t 1 h ˜ t Θ 1 e . a . s . 0 and
u ^ t u ˜ t Θ 1 × Θ 2 e . a . s . 0 .
Proof. 
By Lemma A3, we have log h ^ t log h ˜ t e . a . s . 0 . Thus, log h ^ t log h ˜ t Θ 1 1 . For any fixed ( θ 1 ) , we can obtain
1 h ^ t 1 h ˜ t Θ 1 = exp [ log h ^ t ( θ 1 ) ] exp [ log h ˜ t ( θ 1 ) ] = exp ( log h ˜ t ( θ 1 ) ) 1 [ exp ( log h ˜ t ( θ 1 ) log h ^ t ( θ 1 ) ) ] exp | log h ˜ t ( θ 1 ) | + 1 log h ^ t ( θ 1 ) log h ˜ t ( θ 1 ) .
Since log h ^ t log h ˜ t e . a . s . 0 together with
E log + ( exp | log h ˜ t | + 1 ) Θ 1 = E log + ( exp | log h ˜ t | + 1 ) Θ 1 = 1 + E log h ˜ t Θ 1 < .
We have 1 h ^ t 1 h ˜ t Θ 1 e . a . s . 0 by the above results and Lemma A1. Then we turn to the second statement. By calculation, we have
u ^ t u ˜ t = ϕ ( log h ^ t log h ˜ t ) τ 1 ( r t μ ) ( h ^ t 1 / 2 h ˜ t 1 / 2 ) τ 12 r t 2 ( h ^ t h ˜ t ) I ( r t μ ) 0 τ 22 ( r t μ ) 2 ( h ^ t h ˜ t ) I ( r t μ ) > 0 .
If E | η t , 0 | < , we can conclude E [ log h t ] < by Lemma A2, thus, E [ log + h t 1 / 2 ] < . Together with E [ log | z t | ] < , we have E [ log + | ( r t μ ) | ] E [ log + h t 1 / 2 ] + E [ log | z t | ] < . By Lemma A1, since the sign of r t has no effect on h ^ t and h ˜ t , we have u ^ t u ˜ t e . a . s . 0 . □
The proof of intermediate result(i) then follows from Lemmas A1 and A5.
Lemma A6.
If h ˜ 1 ( θ 1 ) = h ˜ 1 ( θ 1 , 0 ) and u ˜ 1 ( θ 1 , θ 2 ) = u ˜ 1 ( θ 1 , 0 , θ 2 , 0 ) a.s. and Assumptions 1 and 2 hold, then θ 1 = θ 1 , 0 .
Proof. 
First, we can prove that for any θ 1 Θ 1 , if h ˜ 1 ( θ 1 ) = σ 1 2 a.s., then θ 1 = θ 1 , 0 . Note that
C θ 1 ( B ) log h ˜ t = α + A θ 1 ( B ) log x t .
If h ˜ 1 ( θ 1 ) = h ˜ 1 ( θ 1 , 0 ) a.s., by stationarity we have log h ˜ t ( θ 1 ) = log h ˜ t ( θ 1 , 0 ) for all t, thus, we have a.s.
A θ 1 ( B ) C θ 1 ( B ) A θ 1 , 0 ( B ) C θ 1 , 0 ( B ) log x t = α C θ 1 ( B ) α C θ 1 , 0 ( B ) .
Since log x t is a nontrivial random variable, we have A θ 1 ( B ) C θ 1 ( B ) = A θ 1 , 0 ( B ) C θ 1 , 0 ( B ) , thus, by Assumption 1, we have θ 1 = θ 1 , 0 .
Second, if u ˜ 1 ( θ 1 , θ 2 ) = u t a.s., which shows
ξ ξ 0 + ( ϕ ϕ 0 ) log ( σ 1 2 ) + [ τ ( z 1 ) τ 0 ( z 1 ) ] = 0 ,
and by Assumption 2, we have ( ξ , ϕ , τ 1 , τ 21 , τ 22 ) = ( ξ 0 , ϕ 0 , τ 1 , 0 , τ 21 , 0 , τ 22 , 0 ) .
Finally, since σ u , 0 2 = V a r [ u ˜ 1 ( θ 1 , 0 , θ 2 , 0 ) ] , the lemma has been proved. □
Step(ii) is proved in Lemma A6.
Lemma A7.
If Assumptions 1–3 hold, θ 0 is the unique minimum of E ( L T ) on Θ.
The proof of Lemma A7 is similar to Lemma A.6 in Li et al. [40]. For intermediate result (iii), Lemma A7 has proved that θ 0 is the unique minimum of L on Θ .
Refering to intermediate result(iv), we prove L T ( θ 0 ) < .
By Equation (7), we can see that l t is stationary and ergodic. Lemma A3 shows that E | log h ˜ t ( θ 1 , 0 ) | < . Thus, we have E [ l t ( θ 0 ) ] < . The rest of (iv) is the same as the proof of Theorem 2.1(iv) of Francq and Zakoian [41].

Appendix A.1.1. Proof of Theorem 2

By Taylor expansion, due to the existance of first and second derivatives of L T ( θ ) , we have
L T ( θ ^ T ) L T ( θ 0 ) = L T ( θ ˇ T ) ( θ ^ T θ 0 ) ,
where θ ˇ T satisfies θ ˇ T θ 0 θ ^ T θ 0 , almost surely. The thought of proof is similar to the proof of Theorem 7.1 in Straumann andd Mikosch [42]. The asymptotic normality can be obtained by proving the following intermediate results:
(i) T L T ( θ 0 ) d N ( 0 , J ) , as n ;
(ii) L T ( θ ˇ T ) H converges a.s. to 0 for any sequence ( θ ˇ T ) converges a.s. to θ 0 and H is invertible;
(iii) T L ˇ T ( θ ^ T ) L T ( θ ^ T ) converges a.s. to 0;
(iv) T b T [ L ^ T ( θ ^ T ) L ˇ T ( θ ^ T ) ] a . s . 0 , as T b T .
For intermediate result(i), we need to show that { l t ( θ 0 ) } t Z is a martingale difference sequence with respect to the filtration { F t } t Z . Note that 12 u ˜ t ( θ 1 , 0 , θ 2 , 0 ) is conditionally independent of u t , then we have E 0 [ 12 u ˜ t u t | F t 1 ] = 0 . In addition, 1 h ˜ t h t is F t 1 measurable and E [ 1 h ˜ t z t 2 h t | F t 1 ] = 1 h ˜ t h t E [ z t 2 | F t 1 ] = 1 h ˜ t h t . Since σ u 2 = E ( u t 2 ) , we can conclude that E [ l t ( θ 0 ) ] = 0 . Thus, { l t ( θ 0 ) } is a stationary and ergodic zero-mean martingale difference sequence with respect to the filtration { F t } t Z . By the cental limit theorem for square-integrable stationary and ergodic martingale difference sequences, we have
L T ( θ 0 ) d N ( 0 , E [ l t ( θ 0 ) l t ( θ 0 ) ] ) as T ,
where E [ l t ( θ 0 ) l t ( θ 0 ) ] exists.
Lemma A8.
Suppose that the assumptions in Theroem 2 are satisfied. Then for any d 1 , there exists a compact set Θ ˇ 1 ( d ) , and θ 1 , 0 Θ ˇ 1 ( d ) , such that
(i) E h t / h ˜ t ( θ 1 ) Θ ˇ 1 ( d ) d < ;
(ii) E log ( h t / h ˜ t ( θ 1 ) ) Θ ˇ 1 ( d ) d < ;
(iii) E u ˜ t 2 ( θ 1 , θ 2 ) θ ˇ 1 ( d ) × θ ˇ 2 d < .
Proof of Lemma A8 is identiacal to the proof of Lemma A.7 in Li et al. [40].
Lemma A9.
Suppose that the assumptions in Lemma A8 are satisfied such that Θ ˇ 1 ( d ) exists. Set Θ ˇ ( d ) = Θ ˇ 1 ( d ) × Θ ˇ 2 . Then, we have
(i) E 1 ( log h ˜ t ) Θ ˇ 1 ( d ) d < ;
(ii) E 1 u ˜ t Θ ˇ 1 ( d ) d < , provided that E | z t | 2 d < ;
(iii) E 2 u ˜ t Θ ˇ 1 ( d ) d < , provided that E | z t | 2 d < .
Proof. 
(i) Random element 1 ( log h ˜ t ) is the stationary solution of the equation
1 ( log h ˜ t ) = i = 1 p β i 1 ( log h ˜ t i ) + ( 1 , log h ˜ t 1 , , log h ˜ t p , log x t 1 , , log x t q ) T .
Assumption 1 argues that 1 ( log h ˜ t ) is linearly correlated with log x t 1 . By Lemma A8 (ii), we have E 1 ( log h ˜ t ) Θ ˇ 1 ( d ) d < .
For (ii) and (iii), we can calculate that
1 u ˜ t = ϕ + 1 2 τ 1 ( r t μ ) h ˜ t 1 / 2 + τ 21 ( r t μ ) h ˜ t I z t 0 + τ 22 ( r t μ ) h ˜ t I z t > 0 1 ( log h ˜ t )
and
2 u ˜ t = 1 , log h ˜ t , ( r t μ ) h ˜ t 1 / 2 , 1 ( r t μ ) 2 h ˜ t 1 I ( z t 0 ) , 1 ( r t μ ) 2 h ˜ t 1 I ( z t > 0 ) T .
Since ( r t μ ) = h t 1 / 2 z t , and z t is independent of ( h t h ˜ t ) 1 / 2 and 1 log h ˜ t , by Assumption 3. The results can be obtained by E | z t | 2 d < , Lemma A7, and Cauchy–Schwarz inequality. □
The square-integrability of ( θ 0 ) has been shown by Lemma A9. Then, we consider intermediate result (ii). Letting
Δ 11 = 2 / θ 1 T θ 1 , Δ 12 = 2 / θ 1 T θ 2 a n d Δ 22 = 2 / θ 2 T θ 2 ,
we have the following lemma:
Lemma A10.
Suppose that the assumptions in Lemma A7 are satisfied that θ ˇ 1 ( d ) exists. If E | z t | 2 d < , restricted on the vector space with norm · θ ˇ 1 ( d ) or · θ ˇ 1 ( d ) × θ ˇ 2 , we have
(i) ( 1 ( r t μ ) 2 h ^ t ) 1 ( log h ^ t ) ( 1 ( r t μ ) 2 h ˜ t ) 1 ( log h ˜ t ) e . a . s . 0 ;
(ii) u ^ t 2 u ˜ t 2 e . a . s . 0 ;
(iii) 1 u ^ t 2 1 u ˜ t 2 e . a . s . 0 and 2 u ^ t 2 2 u ˜ t 2 e . a . s . 0 .
A representation for the second derivative of the log-likelihood function is as follows:
l t = ( 1 ( r t μ ) 2 h ˜ t ) Δ 11 ( log h ˜ t ) + ( r t μ ) 2 h ˜ t 1 ( log h ˜ t ) 1 ( log h ˜ t ) + 1 σ u 2 Δ 11 ( u ˜ t 2 ) 1 σ u 2 Δ 12 ( u ˜ t 2 ) 1 σ u 4 1 ( u ˜ t 2 ) 1 σ u 2 Δ 12 ( u ˜ t 2 ) 1 σ u 2 Δ 22 ( u ˜ t 2 ) 1 σ u 4 2 ( u ˜ t 2 ) 1 σ u 4 1 ( u ˜ t 2 ) 1 σ u 4 2 ( u ˜ t 2 ) 2 u ˜ t 2 σ u 2 σ u 6 .
Set Θ ˇ = Θ ˇ 1 ( 2 ) × Θ ˇ 2 in Lemma A10 such that sup θ 0 Θ ˇ E [ l t ( θ ) ] exists.
Lemma A11.
Suppose that the assumptions in Lemma A7 are satisfied that θ ˇ 1 ( d ) exists. If E | z t | 2 d < . If E | z t | 4 < , then the following moment conditions hold:
(i) E Δ 11 ( log h ˜ t ) Θ ˇ 1 ( d ) d < , for any d > 0 ;
(ii) E Δ i j ( u ˜ t ) Θ ˇ 1 ( d ) × Θ 2 2 < , for i , j = 1 , 2 ;
(iii) E Δ i j ( u ˜ t 2 ) Θ ˇ 1 ( d ) × Θ 2 < , for i , j = 1 , 2 .
By consistency, we have θ ^ T Θ , for sufficiently large T. Then, we have inequality
L T ( θ ˇ T ) H L T ( θ ) E [ l t ( θ ) ] + E [ l t ( θ ^ T ) ] H ,
where the first term converges a.s. to 0 by Lemma A11, and the second term converges a.s. to 0 by the twice continuous differentiability of E [ l t ( θ ) ] . With the above argument, we can conclude that L T ( θ ˇ T ) H converges a.s. to 0 as θ ^ T converges a.s. to θ 0 .
Lemma A12.
Suppose that the assumptions in Lemma A10 are satisfied that θ ˇ 1 ( 2 ) exists. Restricted on the vector space with norm · θ ˇ 1 ( d ) or · θ ˇ 1 ( d ) × θ ˇ 2 , we have
(i) ( 1 ( r t μ ) 2 h ^ t ) Δ 11 ( log h ^ t ) ( 1 ( r t μ ) 2 h ˜ t ) Δ 11 ( log h ˜ t ) e . a . s . 0 ;
(ii) ( r t μ ) 2 h ^ t 1 ( log h ^ t ) 1 ( log h ^ t ) T ( r t μ ) 2 h ˜ t 1 ( log h ˜ t ) 1 ( log h ˜ t ) T e . a . s . 0 ;
(iii) Δ i j u ^ t 2 Δ i j u ˜ t 2 e . a . s . 0 , for i , j = 1 , 2 .
Lemma A13.
With the assumptions in Theorem 2, we have H = E ( l t θ 0 ) is positive definite.
Remark A2.
Since the spline structure has no effect on the exponential almost sure convergence, the proof of Lemmas A9–A13 is similar to Lemmas A.8–A.12 in Li et al. [40].
For (iii), similar to l t , denote l ˇ t ( θ ) = 1 2 2 log f z ( z t ) + log h ^ t + log ( σ u 2 ) + u ^ t 2 σ u 2 . Then we only need to prove that l ˇ t l t e . a . s . 0 as T . By calculation, we have
l t = ( 1 ( r t μ ) 2 h ˜ t ) 1 ( log h ˜ t ) + 1 σ u 2 1 ( u ˜ t 2 ) , 1 σ u 2 2 ( u ˜ t 2 ) , 1 σ u 2 u ˜ t 2 σ u 4 ,
and
l ˇ t = ( 1 ( r t μ ) 2 h ˜ t ) 1 ( log h ˜ t ) + 1 σ u 2 1 ( u ^ t 2 ) , 1 σ u 2 2 ( u ^ t 2 ) , 1 σ u 2 u ^ t 2 σ u 4 .
By Lemma A.10, the above proportion holds; thus, intermediate result (iii) holds. In addition, we need to prove that J is a singular matrix. This property has been proved in Li et al. [40].
For intermediate result (iv), first write l ^ t as
l ^ t = f ^ z , T ( z t ) 2 f ^ z , T ( z t ) z t 1 ( log h ˜ t ) + 1 σ u 2 1 ( u ^ t 2 ) , 1 σ u 2 2 ( u ^ t 2 ) , 1 σ u 2 u ^ t 2 σ u 4 ,
and rewrite l ˇ t as
l ˇ t = f z ( z t ) 2 f z , T ( z t ) z t 1 ( log h ˜ t ) + 1 σ u 2 1 ( u ^ t 2 ) , 1 σ u 2 2 ( u ^ t 2 ) , 1 σ u 2 u ^ t 2 σ u 4 .
Thus,
L ^ T L ˇ T = E ( l ^ t l ˇ t ) = I 11 + I 12 , I 2 , I 3 ,
where
I 11 = 1 2 f ^ z , T ( z t ) f ^ z , T ( z t ) f z ( z t ) f z , T ( z t ) z t 1 ( log h ˜ t ) , I 12 = I 2 = I 3 = 0 .
For I 11 , we have
I 11 = 1 2 f ^ z , T ( z t ) f ^ z , T ( z t ) f z ( z t ) f z , T ( z t ) z t 1 ( log h ˜ t ) = 1 2 f ^ z , T ( z t ) f ^ z , T ( z t ) f ^ z , T ( z t ) f z , T ( z t ) + f ^ z , T ( z t ) f z , T ( z t ) f z , T ( z t ) f z , T ( z t ) z t 1 ( log h ˜ t ) = 1 2 f ^ z , T ( z t ) 1 f ^ z , T ( z t ) 1 f z , T ( z t ) + f ^ z , T ( z t ) f z , T ( z t ) 1 f z , T ( z t ) z t 1 ( log h ˜ t ) .
Since f ^ z , T ( z t ) , f z , T ( z t ) and 1 ( log h ˜ t ) are bounded, by Lemma A1, we have I 11 a . s . 0 .
Then, we consider the variance of I 11 . Note 1 2 1 ( log h ˜ t ) as C and f ^ z , T ( z t ) f ^ z , T ( z t ) f z ( z t ) f z , T ( z t ) as F ( z t ) , where c is bounded with Lemma A9 (i) and assumption of z t .
V a r ( I 11 ) = C 2 V a r [ F ( z t ) z t ] C 2 V a r [ F ( z t ) ] V a r ( z t ) + V a r [ F ( z t ) ] E ( z t ) 2 + V a r ( z t ) E [ F ( z t ) ] 2 + 2 C o v [ F ( z t ) , z t ] E [ F ( z t ) ] E ( z t ) = C 2 V a r [ F ( z t ) ] + E [ F ( z t ) ] 2 = C 2 E [ F 2 ( z t ) ] .
By Härdle [43], we have
E ( f ^ z , T ( z t ) ) f z ( z t ) = 1 2 b T 2 f z ( z t ) u 2 K ( u ) d u + o ( b T 2 ) ,
thus,
E ( f ^ z , T ( z t ) ) f z ( z t ) = 1 4 b T 2 u 2 K ( u ) d u f z ( z t ) z t 1 log h t = O p ( b T 2 ) ,
along with
F ( z t ) = f z ( z t ) [ f ^ z , T ( z t ) f z ( z t ) ] + f z ( z t ) [ f z ( z t ) f ^ z , T ( z t ) ] f ^ z , T ( z t ) f z ( z t ) = O p ( b T 2 ) ,
therefore, with b T = O ( T 1 / 5 ) ,
V a r ( I 11 ) = O p ( b T 4 ) = O p ( T b T ) .
Define V 11 : = ( T b T ) 1 V a r ( I 11 ) . Based on central limit theorem, we have
T b T I 11 d N ( 0 , V 11 ) .
Thus, by Slutsky’s theorem, the proof of Theorem 2 completes. □

References

  1. Bollerslev, T. Generalized autoregressive conditional heteroskedasticity. J. Econom. 1986, 31, 307–327. [Google Scholar] [CrossRef]
  2. Andersen, T.G.; Bollerslev, T.; Diebold, F.X.; Labys, P. Modeling and forecasting realized volatility. Econometrica 2003, 71, 579–625. [Google Scholar] [CrossRef]
  3. Hansen, P.R.; Huang, Z.; Shek, H.H. Realized GARCH: A joint model for returns and realized measures of volatility. J. Appl. Econom. 2012, 27, 877–906. [Google Scholar] [CrossRef]
  4. Gerlach, R.; Wang, C. Forecasting risk via realized GARCH, incorporating the realized range. Quant. Financ. 2016, 16, 501–511. [Google Scholar] [CrossRef]
  5. Wang, C.; Chen, Q.; Gerlach, R. Bayesian realized-GARCH models for financial tail risk forecasting incorporating the two-sided Weibull distribution. Quant. Financ. 2019, 19, 1017–1042. [Google Scholar] [CrossRef]
  6. Song, X.; Kim, D.; Yuan, H.; Cui, X.; Lu, Z.; Zhou, Y.; Wang, Y. Volatility analysis with realized GARCH-Itô models. J. Econom. 2021, 222, 393–410. [Google Scholar] [CrossRef]
  7. Fu, J.Y.; Lin, J.G.; Hao, H.X. Volatility analysis for the GARCH–Itô–Jumps model based on high-frequency and low-frequency financial data. Int. J. Forecast. 2023, 39, 1698–1712. [Google Scholar] [CrossRef]
  8. Fu, J.Y.; Lin, J.G.; Liu, G.; Hao, H.X. Statistical inference for GQARCH-Itô-jumps model based on the realized range volatility. J. Time Ser. Anal. 2024, 45, 613–638. [Google Scholar] [CrossRef]
  9. Black, F. Studies of stock market volatility changes. In Proceedings of the American Statistical Association Bisiness and Economic Statistics Section; American Statistical Association: Washington, DC, USA, 1976. [Google Scholar]
  10. Cheung, Y.W.; Ng, L.K. Stock price dynamics and firm size: An empirical investigation. J. Financ. 1992, 47, 1985–1997. [Google Scholar] [CrossRef]
  11. Poon, S.H.; Taylor, S.J. Stock returns and volatility: An empirical study of the UK stock market. J. Bank. Financ. 1992, 16, 37–59. [Google Scholar] [CrossRef]
  12. Koutmos, G. Asymmetric volatility and risk return tradeoff in foreign stock markets. J. Multinatl. Financ. Manag. 1993, 2, 27–43. [Google Scholar]
  13. Wang, C.D.; Mykl, P.A. The estimation of leverage effect with high-frequency data. J. Am. Stat. Assoc. 2014, 109, 197–215. [Google Scholar] [CrossRef]
  14. Aït-Sahalia, Y.; Fan, J.; Laeven, R.J.; Wang, C.D.; Yang, X. Estimation of the Continuous and Discontinuous Leverage Effects. J. Am. Stat. Assoc. 2017, 112, 1744–1758. [Google Scholar] [CrossRef] [PubMed]
  15. Chong, C.H.; Todorov, V. Volatility of volatility and leverage effect from options. J. Econom. 2024, 240, 105669. [Google Scholar] [CrossRef]
  16. Yu, J. A semiparametric stochastic volatility model. J. Econom. 2012, 167, 473–482. [Google Scholar] [CrossRef]
  17. Wu, X.; Wang, X. Forecasting volatility using realized stochastic volatility model with time-varying leverage effect. Financ. Res. Lett. 2020, 34, 101271. [Google Scholar] [CrossRef]
  18. Catania, L.; Proietti, T. Forecasting volatility with time-varying leverage and volatility of volatility effects. Int. J. Forecast. 2020, 36, 1301–1317. [Google Scholar] [CrossRef]
  19. Tauchen, G.E.; Pitts, M. The Price Variability-Volume Relationship on Speculative Markets. Econometrica 1983, 51, 485. [Google Scholar] [CrossRef]
  20. Taylor, S.J. Modelling Financial Time Series; John Wiley & Sons: Hoboken, NJ, USA, 1986. [Google Scholar]
  21. Chib, S.; Nardari, F.; Shephard, N. Analysis of high dimensional multivariate stochastic volatility models. J. Econom. 2006, 134, 341–371. [Google Scholar] [CrossRef]
  22. Jensen, M.J.; Maheu, J.M. Bayesian semiparametric stochastic volatility modeling. J. Econom. 2010, 157, 306–316. [Google Scholar] [CrossRef]
  23. Fernández-Villaverde, J.; Guerrón-Quintana, P.; Rubio-Ramírez, J.F. Estimating dynamic equilibrium models with stochastic volatility. J. Econom. 2015, 185, 216–229. [Google Scholar]
  24. Rømer, S.E. Empirical analysis of rough and classical stochastic volatility models to the SPX and VIX markets. Quant. Financ. 2022, 22, 1805–1838. [Google Scholar] [CrossRef]
  25. Yazdani, S.; Hadizadeh, M.; Fakoor, V. Computational analysis of the behavior of stochastic volatility models with financial applications. J. Comput. Appl. Math. 2022, 411, 114–258. [Google Scholar] [CrossRef]
  26. Bolko, A.E.; Christensen, K.; Pakkanen, M.S.; Veliyev, B. A GMM approach to estimate the roughness of stochastic volatility. J. Econom. 2023, 235, 745–778. [Google Scholar] [CrossRef]
  27. Chan, J.C.C. Comparing stochastic volatility specifications for large Bayesian VARs. J. Econom. 2023, 235, 1419–1446. [Google Scholar] [CrossRef]
  28. Brockwell, P.J.; Davis, R.A. Time Series: Theory and Methods; Springer: Berlin/Heidelberg, Germany, 1991. [Google Scholar]
  29. Straumann, D. Estimation in Conditionally Heteroscedastic Time Series Models; Springer Science & Business Media: Dordrecht, The Netherlands, 2005. [Google Scholar]
  30. Pfaff, T. Quick Consistency of Quasi Maximum Likelihood Estimators. Ann. Stat. 1982, 10, 990–1005. [Google Scholar] [CrossRef]
  31. Lee, S.W.; Hansen, B.E. Asymptotic theory for the GARCH (1, 1) quasi-maximum likelihood estimator. Econom. Theory 1994, 10, 29–52. [Google Scholar] [CrossRef]
  32. Komunjer, I. Quasi-maximum likelihood estimation for conditional quantiles. J. Econom. 2005, 128, 137–164. [Google Scholar] [CrossRef]
  33. Lee, L.F. Asymptotic distributions of quasi-maximum likelihood estimators for spatial autoregressive models. Econometrica 2004, 72, 1899–1925. [Google Scholar] [CrossRef]
  34. Fan, J.; Qi, L.; Xiu, D. Quasi-maximum likelihood estimation of GARCH models with heavy-tailed likelihoods. J. Bus. Econ. Stat. 2014, 32, 178–191. [Google Scholar] [CrossRef]
  35. Fang, S.H.Y. Semi-parametric GARCH model based on transformed kernel density estimation. Syst. Eng. Theory Pract. 2014, 34, 1963–1970. [Google Scholar]
  36. Wang, J.; Jiang, Y.; Zhu, Y.; Yu, J. Prediction of volatility based on realized-GARCH-kernel-type models: Evidence from China and the US. Econ. Model. 2020, 91, 428–444. [Google Scholar] [CrossRef]
  37. Andersen, T.G.; Bollerslev, T. Answering the skeptics: Yes, standard volatility models do provide accurate forecasts. Int. Econ. Rev. 1998, 39, 885–905. [Google Scholar] [CrossRef]
  38. Corsi, F.; Mittnik, S.; Pigorsch, C.; Pigorsch, U. The volatility of realized volatility. Econom. Rev. 2008, 27, 46–78. [Google Scholar] [CrossRef]
  39. Silverman, B.W. Weak and strong uniform consistency of the kernel estimate of a density and its derivatives. Ann. Stat. 1978, 6, 177–184. [Google Scholar] [CrossRef]
  40. Li, Y.N.; Zhang, Y.; Zhang, C. Statistical inference for measurement equation selection in the log-realgarch model. Econom. Theory 2019, 35, 943–977. [Google Scholar] [CrossRef]
  41. Francq, C.; Zakoian, J.M. Maximum likelihood estimation of pure GARCH and ARMA-GARCH processes. Bernoulli 2004, 10, 605–637. [Google Scholar] [CrossRef]
  42. Straumann, D.; Mikosch, T. Quasi-maximum-likelihood estimation in conditionally heteroscedastic time series: A stochastic recurrence equations approach. Ann. Stat. 2006, 34, 2449–2495. [Google Scholar] [CrossRef]
  43. Härdle, W. Applied Nonparametric Regression; Cambridge University Press: Cambridge, UK, 1990. [Google Scholar]
Figure 1. Daily logarithmic return rate of HSI and S&P 500.
Figure 1. Daily logarithmic return rate of HSI and S&P 500.
Mathematics 13 01506 g001aMathematics 13 01506 g001b
Figure 2. Daily RV of HSI and S&P 500.
Figure 2. Daily RV of HSI and S&P 500.
Mathematics 13 01506 g002aMathematics 13 01506 g002b
Figure 3. Q–Q chart of daily logarithmic returns of rate of HSI and S&P500.
Figure 3. Q–Q chart of daily logarithmic returns of rate of HSI and S&P500.
Mathematics 13 01506 g003
Table 1. Results of simulation study with z t following standard normal distribution ( × 10 3 ).
Table 1. Results of simulation study with z t following standard normal distribution ( × 10 3 ).
Sample Size1005001000
MethodQMLQML-KQMLQML-KQMLQML-K
α Bias−71.218−43.4096.84438.06516.36434.121
MSE522.667410.0000.7052.2720.5341.634
β Bias24.16033.3630.311−0.1650.5110.450
MSE31.38639.9800.0920.0710.0380.030
γ Bias−5.476−2.191−0.7290.042−0.316−0.057
MSE0.8211.2010.0730.0470.0290.021
ξ Bias−35.051−75.455−22.681−80.928−31.988−65.200
MSE39.22162.1443.87210.4342.5326.419
ϕ Bias8.243−6.1663.6892.5620.281−0.503
MSE3.5497.4900.3070.2670.1560.109
τ 1 Bias0.017−10.706−1.416−9.984−23.309−47.293
MSE0.2510.5100.0540.1700.8302.834
τ 21 Bias4.60536.4247.76150.22722.23645.592
MSE1.4842.9580.6463.4820.8272.671
τ 22 Bias4.60657.9314.31830.54114.48128.862
MSE3.4077.2800.2731.2980.3331.056
σ u Bias1.5871.2700.9100.8640.0280.037
MSE0.5060.2970.0120.0110.0080.007
Table 2. Results of simulation study with z t following standardized t ( 4 ) distribution ( × 10 3 ) .
Table 2. Results of simulation study with z t following standardized t ( 4 ) distribution ( × 10 3 ) .
Sample Size1005001000
MethodQMLQML-KQMLQML-KQMLQML-K
α Bias42.755−2.49638.52433.58225.83127.543
MSE17.13625.83224.8292.35123.0761.582
β Bias7.7997.5227.809−0.2667.5310.486
MSE6.0425.4116.4660.1195.8820.045
γ Bias−0.619−0.616−1.4361.715−2.993−0.577
MSE0.9640.9730.9670.1010.9720.040
ξ Bias−111.174−30.199−113.242−64.578−85.109−54.789
MSE19.15212.11024.3239.70114.0436.384
ϕ Bias−3.181−2.4370.802−2.8942.8320.488
MSE0.8081.2591.5990.4321.0790.140
τ 1 Bias−80.346−24.331−15.001−9.918−54.902−37.313
MSE7.2692.7620.2700.1613.4552.703
τ 21 Bias74.68222.19374.02649.58652.45735.748
MSE6.3352.4436.1213.5553.1512.451
τ 22 Bias44.69713.08945.46230.40231.23121.332
MSE2.2610.9302.3251.3761.1330.880
σ u Bias94.78495.310−1.5790.168−1.1400.368
MSE230.514230.6900.2880.0130.2410.007
Table 3. Comparison results of Spline1 RG-TVL and Spline2 RG-TVL under 100 samples.
Table 3. Comparison results of Spline1 RG-TVL and Spline2 RG-TVL under 100 samples.
ModelSpline1Spline2
DataSpline1Spline2Spline1Spline2
LL−191.970−215.087−212.258−191.480
AIC401.939448.174444.515402.960
BIC425.386471.620470.567429.012
ModelSpline1Spline2
DataSpline1Spline2Spline1Spline2
LL−391.596−439.119−432.489−396.312
AIC801.192896.238884.977812.625
BIC830.877925.923917.960845.608
Table 4. Descriptive statistics of HSI and S&P 500.
Table 4. Descriptive statistics of HSI and S&P 500.
HSIS&P500
rRVrRV
Mean−3.437 ×   10 4 1.258 ×   10 4 −4.821 ×   10 5 4.169 ×   10 5
Max4.550 ×   10 2 7.939 ×   10 4 1.655 ×   10 2 2.894 ×   10 4
Min−4.600 ×   10 2 2.279 ×   10 5 −2.945 ×   10 2 6.007 ×   10 6
Std1.170 ×   10 2 8.424 ×   10 5 6.600 ×   10 3 3.977 ×   10 5
Skewness0.3193.500−0.8502.607
Kurtosis4.08821.0984.54812.238
25% Quantile−8.400 ×   10 3 7.831 ×   10 5 −3.028 ×   10 3 1.680 ×   10 5
Median−1.100 ×   10 3 1.044 ×   10 4 5.051 ×   10 4 2.787 ×   10 5
75% Quantile6.900 ×   10 3 1.443 ×   10 4 4.401 ×   10 3 5.188 ×   10 5
Table 5. Results of in-sample estimation.
Table 5. Results of in-sample estimation.
RGRG-TVL
QMLQML-KQMLQML-K
α 0.3000.2180.2060.114
β 0.7300.7290.7390.737
γ 0.3010.2960.2810.277
ξ −2.120−1.902−1.827−1.573
ϕ 0.7760.7880.8080.824
τ 1 −0.046−0.043−0.009−0.001
τ 2 0.0220.019
τ 21 0.0770.070
τ 22 −0.023−0.023
σ u 0.3730.3730.3680.368
ρ −0.890−0.902−0.850−0.821
LL−1506.503−1510.676−1512.973−1517.231
AIC−2995.007−3003.352−3005.946−3014.462
BIC−2949.586−2957.931−2962.525−2971.041
α −3.570−2.897−2.521−2.780
β 0.3000.3580.2790.278
γ 0.3390.3500.4610.438
ξ −11.367−5.3061.7252.381
ϕ −0.0930.5021.1951.258
τ 1 −0.082−0.167−0.024−0.025
τ 2 0.014−0.012
τ 21 0.0310.031
τ 22 −0.016−0.016
σ u 0.8680.7390.5660.565
ρ −0.971−0.995−0.964−0.963
LL721.701785.719828.333836.232
AIC−1427.403−1555.437−1636.665−1652.463
BIC−1397.853−1529.419−1525.887−1615.526
Table 6. Results of out-of-sample prediction.
Table 6. Results of out-of-sample prediction.
MeasureRGRG-TVL
QMLQML-KQMLQML-K
MSE ( × 10 9 ) 5.7215.1134.0613.949
MAE ( × 10 5 ) 4.6184.2764.0824.063
RMSE ( × 10 4 ) 2.9292.7692.4682.434
RGRG-TVL
QMLQML-KQMLQML-K
MSE ( × 10 6 )1.2631.2541.2521.250
MAE ( × 10 4 )6.5206.4766.4556.442
RMSE ( × 10 3 )1.1241.1201.1191.118
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, J.; Mao, Y.; Hao, H.; Liu, G. Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect. Mathematics 2025, 13, 1506. https://doi.org/10.3390/math13091506

AMA Style

Lin J, Mao Y, Hao H, Liu G. Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect. Mathematics. 2025; 13(9):1506. https://doi.org/10.3390/math13091506

Chicago/Turabian Style

Lin, Jinguan, Yizhi Mao, Hongxia Hao, and Guangying Liu. 2025. "Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect" Mathematics 13, no. 9: 1506. https://doi.org/10.3390/math13091506

APA Style

Lin, J., Mao, Y., Hao, H., & Liu, G. (2025). Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect. Mathematics, 13(9), 1506. https://doi.org/10.3390/math13091506

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop