Next Article in Journal
Using Small Area Estimation to Produce Official Statistics
Next Article in Special Issue
A New Predictive Algorithm for Time Series Forecasting Based on Machine Learning Techniques: Evidence for Decision Making in Agriculture and Tourism Sectors
Previous Article in Journal
A New Benford Test for Clustered Data with Applications to American Elections
Previous Article in Special Issue
Bootstrap Prediction Intervals of Temporal Disaggregation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling Realized Variance with Realized Quarticity

by
Hiroyuki Kawakatsu
Business School, Dublin City University, Dublin 9, D09 Dublin, Ireland
Stats 2022, 5(3), 856-880; https://doi.org/10.3390/stats5030050
Submission received: 11 August 2022 / Revised: 2 September 2022 / Accepted: 5 September 2022 / Published: 7 September 2022
(This article belongs to the Special Issue Modern Time Series Analysis)

Abstract

:
This paper proposes a model for realized variance that exploits information in realized quarticity. The realized variance and quarticity measures are both highly persistent and highly correlated with each other. The proposed model incorporates information from the observed realized quarticity process via autoregressive conditional variance dynamics. It exploits conditional dependence in higher order (fourth) moments in analogy to the class of GARCH models exploit conditional dependence in second moments.

1. Introduction

The commonly used measure of financial risk, the conditional variance, of an asset return is not directly observable. The large class of observation-driven GARCH type or parameter-driven stochastic volatility models use asset return data to estimate the conditional variance process. A more recent alternative approach exploits the availability of intraday high-frequency data and treats the realized variance ( R V ) process as observable data. This approach was popularized by the HAR (heterogeneous autoregressive) specification of Corsi [1].
Understanding the time variation in financial risk and the ability to predict its future course is of primary interest for financial practitioners and regulators who need to manage financial risk. In the extensive literature that examines the conditional dynamics of asset return second moments, some have considered utilizing information contained in higher order moments such as skewness and kurtosis [2,3,4]. This paper contributes to this literature by examining the information in the observed fourth-moment process to predict the realized variance R V .
The proposed model exploits two empirical features of the R V and the realized quarticity ( R Q ) processes. First, both processes are highly persistent with slowly decaying autocorrelations. The early literature modeled the slow decay in volatility as a long memory process with Hurst exponent H > 1 / 2 [5,6,7]. Recently, Gatheral et al. [8] have suggested a rough volatility model with H < 1 / 2 . As they note, though a rough volatility model is not strictly a long memory process, their autocorrelation is statistically difficult to distinguish from that of a long memory process. For the fourth-moment process Huang et al. [9] find positive but weak dependence in the volatility of volatility measure from options on VIX. Da Fonseca and Zhang [10] document roughness in volatility of volatility with H < 1 / 2 .
The second empirical feature the proposed model exploits is the high correlation between R V and R Q processes. A positive correlation between volatility and volatility of volatility is also documented in Da Fonseca and Zhang [10]. This correlation motivates the use of information in R Q to model R V . As the variance of variance is the fourth moment and this fourth moment is persistent (first empirical feature), the proposal is to model the conditional variance of the realized variance with GARCH-type dynamics that can capture the persistence. The proposed model is a higher moment extension of the realized GARCH model of Hansen et al. [11].
The common approach to model the persistence in R V is to use a restricted distributed lag model of Corsi [1]. Bollerslev et al. [12] extend the HAR specification to account for possible measurement error in R V as an estimator of the integrated variance I V . Their HARQ specification allows the parameters of the HAR model to be time-varying and depend on R Q . Buccheri and Corsi [13] consider a HAR model with time-varying parameters in a parameter-driven state-space model. The variance of the observation equation for ln R V is set to R Q and the state equation parameters are driven by the scores as in Creal et al. [14]. To capture long memory type persistence in R V , the state vector has dimension 22. To economize on the number of parameters to estimate the system is sparse, with score dynamics restricted to follow a diagonal random walk. Ding [15] extends the GARCH model to capture dynamics in the variance of variance.
Section 3 formally describes the proposed model specifications and their properties. Unlike the HAR type models, the dependence in R V is indirectly modeled through dependence in R Q . More specifically, R V is modeled as a function of the (unobservable) conditional variance of R V . The proposed model, therefore, extends Hansen et al. [11] to a higher order moment with a R Q -in-mean type model. The GARCH-in-mean models with asset returns on the left-hand side have a persistence mismatch as returns have little persistence while conditional variances are highly persistent [16]. For the higher-order specification in this paper, the left-hand side R V is persistent as with the right-hand side conditional variance of R V .
Section 3 also discusses the somewhat neglected issue of the use of nonlinear transformations. The three most commonly used transformations in the literature are R V (no transformation), R V [1], and ln R V [17]. The first two forms are the variable of interest in financial applications. However, in the majority of the models that are estimated by least squares, the parameters are not restricted to ensure R V 0 and may result in negative out-of-sample forecasts. The log transformation ensures non-negative forecasts without parameter restrictions. Another important advantage of the log transformation, as documented in the empirical analysis Section 4, is the removal of the excess kurtosis in R V and R Q . The proposed model with the log transformation can, therefore, be estimated by maximizing the Gaussian likelihood.
The empirical performance of the proposed model is considered in Section 4 using data for the 27 stocks analyzed in Bollerslev et al. [12]. This section first documents the claimed two empirical features of R V and R Q for the sample of 27 stocks. The proposed model is then fitted using a rolling estimation window for each stock. The rolling pseudo-out-of-sample predictions from the proposed model are then compared to those from the baseline HARQ model of Bollerslev et al. [12]. The HARQ model is chosen as the baseline for comparison as it also exploits information from R Q to predict R V .
The rest of the paper is organized as follows. Section 2 provides a summary of related literature, Section 3 formally describes the proposed model specification, Section 4 examines the empirical performance of the proposed model specification, and Section 5 provides some concluding comments.

2. Related Literature

An early parametric model with a long memory for the R V process was suggested in Andersen et al. [18]. They specify a fractionally integrated VAR where the fractional difference parameter d is estimated by the log-periodogram regression of Geweke and Porter-Hudak [19]. The long memory in R V is modeled by a restricted AR(22) specification in Corsi [1]. This long AR specification, known as the HAR (Heterogeneous AutoRegressive) model, has become the standard model to capture long memory like persistence in R V since it can be estimated by simple least squares.
Corsi et al. [17] extends the HAR specification by modeling conditional dependence in the second moment of the residuals. As the outcome variable in a HAR model is the second moment R V , the residual second moment is the variance of variance. This model of conditional fourth-moment dynamics is the higher-order extension of the GARCH class of conditional second-moment dynamic models. Like the GARCH model, the conditional fourth moment is a latent variable that is not directly observable in the model of Corsi et al. [17].
The proposal in this paper is to use the observed fourth-moment variable, the realized quarticity R Q . R Q was used in the HAR extension of Bollerslev et al. [12], which motivates its use as an instrument for potential measurement error in the lagged R V on the right-hand side of the regression. Their HARQ model can be estimated by simple least squares. Buccheri and Corsi [13] deal with the R V measurement error in a state-space model where the observed R V is modeled as the latent signal I V plus a measurement error noise. The unobserved latent signal I V is the state variable that needs to be filtered. To capture long memory type persistence, I V has a HAR type long AR specification resulting in a latent state vector of dimension 22. Their HARK model uses the observed R Q process as the time-varying variance of the observation equation for R V . Although the state-space model can be estimated by maximum likelihood with the use of Kalman filter, the likelihood is nonlinear in the parameters and estimation is computationally more expensive than the least squares based HARQ model of Bollerslev et al. [12].
Bollerslev et al. [12] and Buccheri and Corsi [13] both use the observed fourth moment variable R Q in the HAR class model to capture long memory type persistence in R V . Rather than use long lags of R V to capture persistence, the proposal in this paper is to exploit similar long memory type persistence in R Q to model the R V dynamics. To exploit potential dependence in the residual second moments as in Corsi et al. [17], the proposed model is a higher moment extension of the realized GARCH model of Hansen et al. [11] and Hansen and Huang [20]. Compared to their realized GARCH specification, the proposed model is an R Q -in-mean specification as described below. The likelihood function of the proposed model is nonlinear in the parameters and requires numerical optimization for maximum likelihood estimation.

3. Realized Variance Model

Online Appendix A.1 provides a brief summary of asymptotic distributions of realized variances.

3.1. Model Specification

The proposed model extends the realized GARCH model of Hansen et al. [11] in two ways. First, instead of the asset return (first moment), the conditional mean equation is for the realized variance (second moment) of asset returns. Second, as the realized variance is persistent, the conditional mean equation has a conditional variance of variance term analogous to the GARCH-in-mean specification to capture the persistence in realized variance.
y t = c 0 + c 1 ln κ t + κ t ϵ t , ϵ t i i d ( 0 , 1 )
ln κ t = ω + α x t 1 + β ln κ t 1
x t = ξ + ϕ ln κ t + τ ( ϵ t ) + σ u u t , u t i i d ( 0 , 1 )
where ϵ t and u t are assumed independent.
The observed outcome variable y t is some function of the realized variance R V t . The choice of transformation for y t is discussed below in Section 3.2. κ t is the (unobserved) conditional variance of y t . As κ t is the variance of variance, the parameter c 1 is the variance-in-mean parameter analogous to the GARCH-in-mean specification [21]. This term is included in the conditional mean Equation (1a) to capture the dependence in y t .
The conditional dynamics for κ t is specified in log form in (1b) to ensure κ t is non-negative without restricting the parameters in (1b). This specification is analogous to that of exponential GARCH of Hansen and Huang [20] and Nelson [22]. Corsi et al. [17] considered a GARCH analogue for κ t where x t 1 = κ t 1 ϵ t 1 2 in Equation (1b) with κ t on the left-hand side. For the proposed model x t is an observed variable analogous to Hansen et al. [11] and Hansen and Huang [20]. Hansen et al. [11] ‘augmented’ the GARCH model by using information from realized variance by setting x t = R V t . For this application, κ t is the conditional fourth moment and x t is set to some transformation of the realized quarticity R Q t .
In Equation (1c) for the observed variable x t , the term τ ( ϵ t ) captures potential asymmetric response to negative and positive shocks ϵ t . As in Hansen et al. [11], the quadratic specification τ ( z ) = τ 1 z + τ 2 ( z 2 1 ) is used in the empirical application. τ ( ϵ t ) is then a zero mean process where the parameter τ 1 captures the asymmetric response. (1b) can include additional lags of x t and ln κ t but we use only one lag for parsimony.
Model (Section 3.1) implies that ln κ t follows an AR(1) process and x t an ARMA(1,1) process [11]
ln κ t = ω + α ξ + ( β + α ϕ ) ln κ t 1 + α e t 1 x t = ω ϕ + ξ ( 1 β ) + ( β + α ϕ ) x t 1 + e t β e t 1
where e t τ ( ϵ t ) + σ u u t is a zero mean i i d process. The stationarity condition for ln κ t and x t is β + α ϕ < 1 with unconditional means
E [ ln κ t ] = ω + α ξ 1 β α ϕ , E [ x t ] = ω ϕ + ξ ( 1 β ) 1 β α ϕ
ρ β + α ϕ is a measure of persistence of the model implied series ln κ t and x t as their autocorrelations decay with powers of ρ .

3.2. Variable Transformations

The proposed model (Section 3.1) can be considered a class of models depending on the choice of transformed variables y t and x t . The literature that models realized variance R V t have used alternative transformations of R V t for y t . The HAR model of Corsi [1] uses y t = R V t , Corsi et al. [17] consider y t = R V t , y t = ln R V t and Bollerslev et al. [12] consider y t = R V t . Strictly speaking the non-negativity of y t = R V t , y t = R V t requires restrictions on the mean Equation (1a) parameters to ensure out of sample forecasts remain non-negative. Such restrictions, however, do not appear to be imposed for least squares based estimators in Corsi et al. [17] and Bollerslev et al. [12]. As a consequence out-of-sample predictions may produce negative values of y t = R V t or y t = R V t .
For this reason, the empirical application below uses the log transformation y t = ln R V t which ensures non-negative R V t predictions of R V t without restrictions on the mean equation parameters (1a). A problem with the use of log transformation is that the variable of interest in financial applications is often the variance itself R V t or the volatility R V t , not the log transformation. If we use the log transformation in (Section 3.1), the predicted values need to go through a nonlinear transformation to obtain predictions of the variable of interest. The nonlinear transformation results in approximate inference for the variable of interest even under the rather strong assumption of a correctly specified distribution for the transformed variable y t .
The variable x t is an observed measure of the conditional variance of y t . As y t is a second moment variable, it is natural to use a realized fourth moment or quarticity R Q t for x t . For transformations y t = R V t or y t = ln R V t , the asymptotic variance expression in online Appendix A.1 have I V t in the denominator. This may cause problems for the values of R V t close to zero, the so-called inlier problem. As the model (Section 3.1) restricts x t to follow an ARMA(1,1) process, an alternative approach is to choose a transformation of R Q t in the sample that has similar dependence as an ARMA(1,1) process. A further consideration is that the model implied process for x t does not ensure positive x t without restrictions on the parameters. The log transformation x t = ln R Q t ensures positive R Q t without imposing additional restrictions on the parameters.
For these reasons, the empirical application below uses the log transformation x t = ln R Q t . Alternative transformations of the conditional variance have been considered for the (G)ARCH-in-mean term. The R Q -in-mean term with coefficient c 1 in (1a) can also use κ t itself or transformations such as κ t . However, (1a) uses the log transformation as the dynamics of κ t in logs in (1b) ensures non-negative κ t without restricting the parameters. Furthermore, the log transformation of κ t matches the log transformation of R V t for y t and of R Q t for x t .

3.3. Maximum Likelihood Estimation

The parameters of the model can be estimated by maximum likelihood assuming the error terms have a Gaussian distribution. While one can interpret the Gaussian assumption as a quasi-likelihood, for finite sample performance, it is desirable that the transformations discussed in the previous section are chosen so that the distribution of ϵ t and u t are not ‘too’ different from the Gaussian. The levels of R V t and R Q t are well known to have very high kurtosis and an important reason to prefer the log transformation for y t and x t is to remove the excess kurtosis. Corsi et al. [17] and Barndorff-Nielsen and Shephard [23] find that the finite sample distribution of ln R V t is closer to the Gaussian than R V t .
For model (Section 3.1), the contribution to the Gaussian log-likelihood from the t-th observation is
t ( y t , x t ) = t ( ϵ t ) 1 2 ln κ t + t ( u t ) 1 2 ln σ u 2 = ln ( 2 π ) 1 2 ( ϵ t 2 + ln κ t + u t 2 + ln σ u 2 )
For the quadratic leverage function τ ( z ) = τ 1 z + τ 2 ( z 2 1 ) used in Hansen et al. [11], the parameter vector is θ = ( θ 1 , θ 2 , θ 3 ) , θ 1 = ( c 0 , c 1 ) , θ 2 = ( ω , α , β ) , θ 3 = ( ξ , ϕ , τ 1 , τ 2 , σ u ) . The parameter restrictions are σ u > 0 and for stationarity of κ t and x t , β + α ϕ < 1 .
To start the recursion in (1b), we need to specify presample values of x t , κ t . If stationarity of x t , κ t are imposed, we can set these to the unconditional means x 0 = E [ x t ] , κ 0 = E [ κ t ] given in (3). A simple alternative used in the empirical application below is to use the sample variance of y t .
For statistical inference, the ‘sandwich’ QML (quasi-maximum likelihood) parameter covariance matrix can be used by evaluating the first and second derivatives of the contributions to the Gaussian log-likelihood. Analytical expressions to recursively evaluate the first derivatives are given in the online Appendix A.2. The second derivatives can be evaluated by numerically differentiating the analytical first derivatives.

3.4. Model Evaluation by Pseudo Out-of-Sample Forecasting

A standard approach to evaluating model performance is to compare the accuracy of pseudo-out-of-sample forecasts. For alternative models with the same outcome variable y t and provided the actual outcome y t is observable, the comparison can be made by specifying a loss, or scoring, function.
There are two additional issues to consider in this application where the outcome of interest is (some function of) the daily realized variance R V t . First, if the statistic of interest is the integrated variance I V t , the outcome y t based on R V t is likely measured with error. Patton [24] suggested scoring functions robust to additive error in the outcome variable.
The second issue is the comparison of models where the outcome variable of interest is some transformation of the modeled variable y t . For example, in finance applications rather than y t = ln R V t we are often more interested in R V t = exp ( y t / 2 ) . To compare models with outcomes y 1 t = R V t , y 2 t = R V t , y 3 t = ln R V t , we need to specify a common outcome variable of interest. In this case setting R V t as the variable of interest is the least problematic as obtaining forecasts of R V t from R V t under additive error is mathematically intractable. Furthermore R V t is the observed data variable constructed from intraday data.
A commonly used approach to obtain forecasts for R V t is to assume Gaussianity. For y t = ln R V t and ϵ t N ( 0 , 1 ) , the h-step forecast of y is Gaussian with
E t [ y t + h ] N ( y ^ t + h | t , κ t ) y ^ t + h | t = c 0 + c 1 ln κ t
and the forecasts for the transformed variables are
E t [ R V t + h ] = E t [ exp ( y t + h ) ] = exp ( y ^ t + h | t + 1 2 κ t ) E t [ R V t + h ] = E t [ exp ( y t + h / 2 ) ] = exp ( 1 2 y ^ t + h | t + 1 8 κ t )
An alternative approach that does not assume Gaussianity is the ‘smearing’ estimate of Duan [25] and Wooldridge ([26], 6-4). Let e ^ t = y t y ^ t , t = 1 , , T denote the in-sample residuals. The smearing forecasts are
R V ^ T + h = exp ( y ^ T + h ) 1 T t = 1 T exp ( e ^ t ) R V ^ T + h = exp ( 1 2 y ^ T + h ) 1 T t = 1 T exp ( 1 2 e ^ t )

4. Empirical Application

This section uses the 5-minute return data from Bollerslev et al. [12] to evaluate the performance of the proposed specification (Section 3.1) for the realized variance. The data, made publicly available by the authors, is a daily sample of 27 Dow Jones constituent stocks from 22 April 1997 to 31 December 2013 (4200 trading days). A list of the 27 stocks and their ticker symbols are provided in online Appendix A.3.

4.1. Preliminary Analysis

Before we fit the model to the data, we examine the log transformations chosen for the observed variables y t = ln R V t and x t = ln R Q t . As explained above, the log transformations ensure non-negative R V t and R Q t without restricting the model parameters. Another important reason for using the log transformation for y t is to approximate the Gaussian distribution assumed for the likelihood function closely. Table A2 in the online Appendix A.3 shows the sample third (skewness) and fourth (kurtosis) moments of R V t , R V t , and ln R V t for the 27 stocks in the sample.
Table A2 shows that all three transformations are positively skewed. The untransformed R V has the largest positive skewness followed by R V , which have skewness all above one. The log transformation ln R V has the smallest skewness all below one, but they are all significantly different (at size 0.05) from the Gaussian value of zero. The untransformed R V also has the largest kurtosis, all well above the Gaussian value of three. R V has somewhat smaller kurtosis, but the smallest value, 7.42 (for INTC), is still well above three. The log transformation ln R V has kurtosis much closer to Gaussian with the largest value at 4.15 (for CVX).
As mentioned in the introduction, the proposed specification (Section 3.1) is based on two empirical features of R V and R Q : their common persistence and their correlation. Figure A1 in the online Appendix A.3 shows the sample autocorrelations of ln R V and ln R Q for the 27 stocks in the sample. For all stocks, ln R V is somewhat more persistent than ln R Q . The autocorrelations for ln R V slowly decay from about 0.8 while those for ln R Q slowly decay from about 0.6. Both autocorrelations die out slowly with the lag, a feature of long memory series. The two series are highly correlated with each other with all pairwise sample correlations above 0.95.
The model (Section 3.1) implies that the observed realized process x t should follow the restricted ARMA(1,1) process (2). The slowly decaying autocorrelations for x t = ln R Q t in Figure A1 is not inconsistent with an ARMA(1,1) process with a large AR(1) coefficient. As a further check, Table A5 in the online Appendix A.3 reports estimates of an unrestricted ARMA(1,1) model fitted to x t for the 27 stocks in the sample. Both the AR and MA coefficients are statistically significant, with a large positive AR(1) coefficient and a large negative MA(1) coefficient. The estimated AR(1) coefficient ranges from 0.976 (XOM) to 0.992 (MCD) and the MA(1) coefficient from 0.709 (XOM) to 0.844 (NKE). Although the portmanteau test for residual correlation up to lag 5 rejects the white noise residual null (at size 0.05), the residual serial correlations are small in size. The first order residual serial correlation ranges from 0.103 (NKE) to 0.036 (CVX).

4.2. In-Sample Estimates

To evaluate the performance of the model out of the sample, pseudo-out-of-sample rolling forecasts were obtained. As parameter estimation by maximum likelihood is computationally expensive compared to least squares, a rolling forecast window of one (calendar) month was moved forward each month starting from January 2006. The first set of parameter estimates were obtained for the estimation sample from the beginning of the data sample April 1997 to December 2005 (104 months). Using these parameter estimates, h-step forecasts for trading days in the month of January 2006 are obtained. These forecasts use the same parameter estimates but the conditioning information set, i.e., the lagged variables on the right-hand side are updated as we move forward within the forecast window.
The next set of estimates were obtained for the sample from May 1997 to January 2006 and forecasts for trading days in the month of February 2006 were obtained. This resulted in 96 sets of parameter estimates with forecasts from the beginning of January 2006 to the end of December 2013. This forecast sample included the financial crisis period 2008–2009 when there was a spike in R V .
Numerical maximization of the Gaussian likelihood can be sensitive to the choice of starting values. To guard against getting stuck in local maxima, a few alternative random starting values are tried for each estimation window. To start the recursion, the presample values for x t and ln κ t were set to the estimation sample variance of y t . (The alternative of setting these presample values to the model unconditional means (3) often resulted in the nonconvergence of the numerical optimizer and was sensitive to the choice of starting parameter values).
There are a large number of estimated parameters (10 parameters for each of the 96 estimation windows for each of the 27 stocks). Figure 1 shows a summary of the estimated parameters for stepsizes h = 1 ,   5 ,   22 days. Following Bollerslev et al. [12], for the h-step forecast the outcome variable on the left-hand side of (1a) is
y t ( h ) 1 h j = 1 h y t + j 1 , y t = R V t , R V t ln 1 h j = 1 h Y t + j 1 , y t = ln ( R V t ) , Y t = R V t , R V t
The h-step ahead variable for the log transformation is defined so that the forecasts from y t ( h ) can be compared to those from the other transformations. For h = 1 , taking the log of R V or R V just results in a different scaling. For h > 1 , the log of R V averages and the log of R V averages are considered since one cannot be recovered from the other just by rescaling. Strictly speaking, the model parameters change with the forecast stepsize h and should be written as a function of h.
Each panel in Figure 1 corresponds to a parameter and the shaded area is the interquartile range (from the 0.25 to 0.75 quantile) across the 27 stocks. The thick solid line is the median estimate across the 27 stocks. The intercept c 0 and the R Q -in-mean parameter c 1 of Equation (1a) are both positive. A positive R Q -in-mean parameter c 1 is to be expected given the similarity of the ln R V and ln R Q dynamics documented above. c 1 declines with stepsize h and for h > 1 , c 1 is smaller for the log average of R V than for the log average of R V .
The parameters α , β , ϕ that determine the persistence ρ β + α ϕ of the ln κ t and x t processes are all positive and imply ρ close to but below the stationarity boundary of one (Figure 2). As a function of stepsize h, α does not change much for the log average of R V but somewhat increases for the log average of R V . β increases with h from about 0.7 for h = 1 to above 0.8 for h = 22 and ϕ decreases with h from about 6 for h = 1 to less than 4 for h = 22 . Figure 2 shows the rolling estimates of the persistence parameter ρ increase with the stepsize h. This is to be expected as the h-step outcome variable y t ( h ) gets smoother and hence more persistent with h.
The asymmetric response parameter τ 1 is positive and decreases with stepsize h from about 1.3 for h = 1 to below 0.5 for h = 22 . The coefficient on the quadratic term τ 2 switches sign from positive for h = 1 to negative for h = 5 , 22 . The volatility parameter σ u for the x t = ln R Q t Equation (1c) increases with stepsize h.

4.3. Pseudo Out-of-Sample Forecasts

This section evaluates the performance of the proposed model in terms of (pseudo) out-of-sample rolling forecasts described above in Section 4.2. If the model captures the time series dependence of the outcome variable, the forecast errors should not be serially correlated. A formal such test needs to account for sampling error in generating the forecasts based on the estimated parameters. Figure A2, Figure A3 and Figure A4 in the online Appendix A.4 provide an informal check by plotting the autocorrelation functions of the rolling forecast errors. In addition to forecasts error from the proposed model (Section 3.1), these figures also compare the forecast error correlations from the HAR model of Corsi [1] with outcome y = R V and the HARQ model of Bollerslev et al. [12] with outcome y = R V .
For stepsize h = 1 , the forecast error serial correlation is small in magnitude, but most of them fall outside the asymptotic interval for a white noise series. (A portmanteau test for serial correlation up to lag 5 all reject the null hypothesis of a white noise at the conventional size of 0.05. However, these are asymptotic p-values that ignore parameter estimation sampling error.) For stepsizes h > 1 , the forecast errors show a large positive correlation up to lag h for all models. This dependence is due to the overlapping sample used in generating the multistep outcome variable (4).
To evaluate the relative forecast performance of the proposed model, its forecast accuracy is compared against a baseline model. The baseline model is the HARQ model of Bollerslev et al. [12], which extends the HAR model of Corsi et al. [17] with an additional interaction term involving R Q . The model proposed in this paper also uses the realized quarticity R Q but without the lagged R V terms of HAR(Q). Following Bollerslev et al. [12], the outcome variable of the baseline HARQ model is the untransformed y = R V . The model parameters are estimated by least squares without restricting them to ensure the predicted values are non-negative. Bollerslev et al. ([12], footnote 17) apply the ‘insanity’ filter and replace predicted values that are outside the in-sample range with the in-sample mean value. The baseline predictions in this paper do not apply this somewhat ad hoc insanity filter.
An alternative natural baseline that ensures non-negative predicted R V values is the HARQ model with y = ln ( R V ) and an interaction term with ln ( R Q ) . This log version of the HARQ specification, however, does not perform as well as the untransformed specification of Bollerslev et al. [12]. Table A3 in the online Appendix A.3 reports estimated coefficients on the interaction term for the full sample for stepsizes h = 1 ,   5 ,   22 . For the level (untransformed) specification, these coefficients are all negative with t-ratios above two in absolute value (with two exceptions for h > 1 ) as reported in Bollerslev et al. [12]. For the log specification, many coefficients are positive and insignificant. For h > 1 all t-ratios (except one) are below two in absolute value. Bollerslev et al. [12] motivate the use of R Q to correct for possible measurement error in R V . The negative coefficient of the interaction term may be correcting for this error mainly in the tails of the high excess kurtosis of R V . This may explain the insignificant coefficient in the interaction for the log specification as the excess kurtosis largely disappears for ln ( R V ) . The issue of appropriate HARQ specification is not the focus of this study and is left for further research.
To evaluate the relative accuracy of predictions from alternative models, we need to specify a loss or scoring function. For this study, two members from the Bregman family of consistent scoring functions that are robust to additive noise in R V [24] are used.
S m s ( R V , F ) = ( R V F ) 2 S q l ( R V , F ) = R V F ln ( R V F ) 1
where R V is the actual realized value and F its forecast. S m s is the mean squared and S q l is the QLIKE scoring function. These two scoring functions were also used in Bollerslev et al. [12]. S m s is defined for all values of F while S q l is undefined for F < 0 . As the baseline HARQ predictions are not filtered, the small cases of non-positive predictions are removed when evaluating S q l .
The difference in scores between the baseline and comparison model is tested with the t-ratio of equal forecast accuracy of Diebold and Mariano [27]. As mentioned above, the forecast errors are correlated for h > 1 . To account for this correlation, the denominator of the t-ratio, the standard error of the difference in average scores, is computed using the Bartlett kernel with bandwidth set to the forecast stepsize h. Table 1 reports these t-ratios for the full forecast sample. A positive τ value indicates a better forecast (smaller score) from the comparison model than the baseline HARQ.
Table 1 shows that in about half of the 27 stocks, a negative R V forecast is produced for the baseline HARQ model. As mentioned above, τ Q L is computed for the forecast sample, excluding these negative predictions. The sign switch between τ M S and τ Q L may be due to the difference in the forecast evaluation sample for the two score functions. With this caveat, the test result is somewhat sensitive to the choice of scoring function and forecast stepsize h. For h = 1 , both τ M S and τ Q L indicate lower average score, i.e., better forecast accuracy, for the baseline HARQ model than the proposed model (Section 3.1). The forecast accuracy of the proposed model relative to the baseline generally improves with the forecast stepsize h. For h = 22 , τ M S is inconclusive in the sense that all their values are less than two in absolute value. None of the τ Q L values are less than 2 , but about a quarter of the stocks have values above +2 indicating more accurate forecasts from model (Section 3.1) than from the baseline.
The tests in Table 1 are based on the full forecast sample. The test results could be specific to the choice of the forecast sample and a particular sample could be cherry-picked to obtain certain results. As a guard against such potential cherry picking, Figure A5, Figure A6 and Figure A7 in the online Appendix A.5 show running t-ratios of the Diebold and Mariano [27] equal forecast accuracy tests for all possible end-of-forecast samples. These are the values of t-ratios when the test is applied to the forecast sample from the beginning of the forecast sample (3 January 2007) up to each date of an expanding evaluation sample. (The three figures use alternative (un)transformations to obtain forecasts for R V from ln R V ).
One common feature of the running t-ratio results is the change in performance shortly after the financial crisis period 2008–2009. As expected the test performance up to the financial crisis period is quite noisy, but the majority of the t-ratios for both MS and QL are positive for h > 1 , indicating better forecast accuracy from model (Section 3.1) compared to the baseline. The problem with the HARQ model in levels producing negative forecasts mostly occurs during the financial crisis period. Table A4 in the online Appendix A.3 lists all dates with negative predictions from the HARQ model, the majority of which occur during 2008. The t-ratios remain quite stable after the financial crisis period indicating robustness to the choice of forecast sample post-financial crisis. When judging the significance of these running t-ratios, one should be aware of the multiple comparisons problem and that the usual critical values are likely to be too small.

5. Concluding Remarks

This paper proposed a model for R V dynamics that exploits the information in the higher order moment of R Q . In contrast to the HAR(Q) models that exploit dependence in the R V variable itself, the idea is to exploit dependence in the R Q series. The empirical analysis using 27 stocks suggests that the proposed model may perform better than the HAR(Q) type specifications for multistep predictions and during periods of market turmoil such as the financial crisis of 2008.
As the empirical analysis was based on large cap (blue chip) US stocks, it remains to be seen how the proposed model performs for other markets or asset classes. These include small-cap stocks, which are known to be volatile and fat-tailed, or cryptocurrencies. Performance during other market turmoil periods, such as the recent pandemic, or for commodities during the recent conflict in Ukraine could also be analyzed.
For policymakers, the performance of the proposed model for risk management purposes could be of interest. The fundamental review of the trading book for the latest Basel regulation has introduced the use of expected shortfall for market risk capital requirements. The forecast performance analysis in this paper could be extended to evaluation of the accuracy of risk measures such as expected shortfalls. Additional tools or models for managing tail risks for regulatory purposes could prove beneficial for large financial institutions.
A somewhat puzzling result is the better accuracy for multistep forecasts from the proposed model compared to the baseline HARQ model. The HARQ specification uses R V t 22 while the proposed specification only uses one lag (though, in principle, additional lags could be considered). Whether the implied ARMA(1,1) process for x t = ln R Q can better capture long-term dependence than the AR(22) term could be further investigated.
Several extensions of the proposed model could be considered. Bollerslev et al. [12] consider the use of alternative estimators of R V such as realized kernel of Barndorff-Nielsen et al. [28] and of R Q such as Q P and T P mentioned in Appendix A.1. Alternatively, option-based fourth-moment measures from VVIX as used in Huang et al. [9] could also be considered. Rather than select one of these alternative measures, one can try to incorporate information from all of these alternative measures as in Hansen and Huang [20].
The challenge in further generalizing the model is the increase in the number of parameters to estimate. With ten parameters to estimate, the model is already somewhat over-parameterized as a model of a single outcome variable. To reduce the number of parameters to estimate, one can either impose a priori ‘reasonable’ restrictions or sparsity consistent with the data. Alternatively, a penalty term can be added to the likelihood function for regularization [29,30].

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Appendix A.1. Summary of Asymptotic Distributions of Realized Variances

Assume the log asset price p t follows a semimartingale process
d p t = μ ( t ) d t + σ ( t ) d W ( t )
with drift μ ( t ) and volatility σ ( t ) . W ( t ) is the standard Wiener process. The realized variance over day t is R V t = j = 1 m r t , j 2 where r t , j is the log return over the j-th subinterval in day t and m is the number of subintervals over one trading day. In the diffusion limit m [31,32],
R V t M N ( I V t , 2 m I Q t )
where I V t = t 1 t σ 2 ( s ) d s is the integrated variance and I Q t = t 1 t σ 4 ( s ) d s is the integrated quarticity. The (approximate) asymptotic distribution of nonlinear transformations of R V t can be obtained by the delta method [17,23]
R V t M N ( I V t , 1 2 m I Q t I V t ) ln R V t M N ( ln I V t , 2 m I Q t I V t 2 )
To estimate the (asymptotic) variance of R V t , we need an estimate of integrated quarticity I Q t . Commonly used consistent estimators (under no microstructure noise) are
R Q t = m 3 j = 1 m r t , j 4 Q P t = m π 2 4 j = 1 m | r t , j | | r t , j 1 | | r t , j 2 | | r t , j 3 | T P t = m Γ ( 1 / 2 ) 3 4 Γ ( 7 / 6 ) 3 j = 1 m | r t , j | 4 / 3 | r t , j 1 | 4 / 3 | r t , j 2 | 4 / 3
R Q t is the realized quarticity, Q P t is the realized quad-power quarticity [33,34], and T P t is the realized tri-power quarticity [35].

Appendix A.2. Derivatives of Log-Likelihood

For (Section 3.1) the contribution to the Gaussian log-likelihood from the t-th observation is
t ( y t , x t ) = t ( ϵ t ) 1 2 ln κ t + t ( u t ) 1 2 ln σ u 2 = ln ( 2 π ) 1 2 ( ϵ t 2 + ln κ t + u t 2 + ln σ u 2 )
Contribution to score is
t θ = ϵ t ϵ t θ u t u t θ 1 σ u σ u θ 1 2 ln κ t θ
where
ln κ t θ 1 = 0 , ln κ t θ 2 = 1 x t 1 ln κ t 1 + β ln κ t 1 θ 2 , ln κ t θ 3 = β ln κ t 1 θ 3 ϵ t θ 1 = 1 κ t 1 ln κ t ( ϵ t 2 + c 1 κ t ) ln κ t θ 1 , ϵ t θ j = ( ϵ t 2 + c 1 κ t ) ln κ t θ j , j = 2 , 3 u t θ j = ϕ σ u ln κ t θ j 1 σ u ( τ 1 + 2 τ 2 ϵ t ) ϵ t θ j , j = 1 , 2 , u t θ 3 = 1 σ u 1 ln κ t ϵ t ϵ t 2 1 u t ϕ σ u ln κ t θ 3 1 σ u ( τ 1 + 2 τ 2 ϵ t ) ϵ t θ 3
For t = 1 , if x 0 , ln κ 0 are fixed, e.g., set to estimation sample variance of y t ,
ln κ 1 θ 2 = 1 x 0 ln κ 0 , ln κ 1 θ 3 = 0
and if x 0 , ln κ 0 are set to their unconditional means
ln κ 1 θ 2 = 1 1 β α ϕ 1 x 0 ln κ 0 , ln κ 1 θ 3 = α 1 β α ϕ 1 ln κ 0 0 3

Appendix A.3. Ticker Symbols

Twenty-seven constituents from the Dow Jones (Table 2, [12]).
Table A1. Ticker symbols from Dow Jones constituents. Reproduced from (Table 2, [12]).
Table A1. Ticker symbols from Dow Jones constituents. Reproduced from (Table 2, [12]).
SymbolExchangeCompany
AXPNYSEAmerican Express Company
BANYSEBoeing Company
CATNYSECaterpillar Inc.
CSCONASDAQCisco Systems, Inc.
CVXNYSEChevron Corporation
DDNYSEDuPont de Nemours, Inc.
DISNYSEWalt Disney Company
GENYSEGeneral Electric Company
HDNYSEHome Depot
IBMNYSEInternational Business Machine Corporation
INTCNASDAQIntel Corporation
JNJNYSEJohnson & Johnson
JPMNYSEJPMorgan Chase & Co.
KONYSECoca-Cola Company
MCDNYSEMcDonald’s Corporation
MMMNYSE3M Company
MRKNYSEMerck & Co., Inc.
MSFTNASDAQMicrosoft Corporation
NKENYSENike, Inc.
PFENYSEPfizer Inc.
PGNYSEProcter & Gamble Company
TRVNYSETravelers Companies, Inc.
UNHNYSEUnitedHealth Group Incorporated
UTXNYSEUnited Technologies Corporation
VZNYSEVerizon Communications Inc.
WMTNYSEWalmart Inc.
XOMNYSEExxonMobil Corporation
Table A2. Sample third ( S k e w ) and fourth ( K u r t ) moments of transformations of realized variance R V . p-values in square brackets are for K u r t = 3 for the log transformation. All other p-values for S k e w = 0 and K u r t = 3 are less than 0.05 and are not reported to avoid cluttering the table. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Table A2. Sample third ( S k e w ) and fourth ( K u r t ) moments of transformations of realized variance R V . p-values in square brackets are for K u r t = 3 for the log transformation. All other p-values for S k e w = 0 and K u r t = 3 are less than 0.05 and are not reported to avoid cluttering the table. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
RV RV ln RV
Skew Kurt Skew Kurt Skew Kurt
AXP11.59262.022.7417.680.342.95[0.53]
BA6.5776.692.2112.350.343.19[0.01]
CAT7.42110.652.3313.470.393.32[0.00]
CSCO4.6336.341.958.610.432.90[0.17]
CVX15.71400.323.8735.700.494.15[0.00]
DD5.8264.821.9010.010.222.88[0.12]
DIS7.95130.392.1312.200.372.78[0.00]
GE9.79153.913.2621.520.533.52[0.00]
HD7.83121.542.3213.240.443.07[0.39]
IBM6.6378.682.2211.720.452.91[0.25]
INTC4.1331.951.757.420.412.82[0.02]
JNJ8.23123.382.2513.720.272.82[0.02]
JPM9.63150.452.9918.910.443.20[0.01]
KO6.5583.302.0711.000.342.90[0.18]
MCD12.52283.702.4618.660.112.86[0.06]
MMM13.84348.472.9523.130.453.38[0.00]
MRK21.02774.033.7036.260.513.89[0.00]
MSFT4.6940.031.828.400.372.82[0.02]
NKE5.2855.881.828.770.352.71[0.00]
PFE5.3955.001.979.890.383.01[0.91]
PG9.63171.202.5516.200.463.01[0.94]
TRV15.45401.553.3526.920.543.16[0.03]
UNH8.32126.422.7816.560.613.55[0.00]
UTX7.65106.792.4013.990.433.21[0.00]
VZ7.49114.792.2312.430.413.00[0.98]
WMT8.83176.111.9911.480.332.59[0.00]
XOM13.52322.343.2626.450.393.63[0.00]
Table A3. Estimates of interaction term in HARQ models. ‘level’ is HARQ with y = R V , q = R Q and ‘log’ is with y = ln ( R V ) , q = ln ( R Q ) . κ is the sample kurtosis of y where κ = 3 for a Gaussian. τ h are the HAR t-ratios for the interaction term q t 1 y t 1 for forecast step size h days. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Table A3. Estimates of interaction term in HARQ models. ‘level’ is HARQ with y = R V , q = R Q and ‘log’ is with y = ln ( R V ) , q = ln ( R Q ) . κ is the sample kurtosis of y where κ = 3 for a Gaussian. τ h are the HAR t-ratios for the interaction term q t 1 y t 1 for forecast step size h days. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Levellog
κ τ 1 τ 5 τ 22 κ τ 1 τ 5 τ 22
AXP262.02−11.84−6.52−4.042.951.231.12−0.39
BA76.69−4.65−3.67−3.333.190.48−0.69−0.76
CAT110.65−6.40−2.85−2.703.320.690.250.31
CSCO36.34−4.15−3.54−4.332.902.651.710.61
CVX400.32−5.57−1.81−2.124.151.760.910.96
DD64.82−5.78−3.01−3.542.880.970.26−0.02
DIS130.39−3.90−5.44−4.602.78−0.08−0.33−0.74
GE153.91−5.61−4.79−2.423.522.290.77−0.27
HD121.54−6.96−6.33−4.833.070.05−0.22−0.34
IBM78.68−2.68−5.52−5.312.913.420.680.58
INTC31.95−8.25−3.40−3.972.821.730.911.08
JNJ123.38−4.02−5.29−4.052.822.340.540.37
JPM150.45−5.92−7.57−5.833.202.471.260.11
KO83.30−9.07−6.21−1.322.900.160.450.24
MCD283.70−3.78−3.10−3.352.86−2.05−1.78−2.16
MMM348.47−9.81−6.43−5.933.381.330.831.31
MRK774.03−4.70−5.44−5.443.89−0.17−0.42−0.73
MSFT40.03−4.28−5.22−2.212.822.681.490.70
NKE55.88−5.78−4.20−4.112.71−0.53−0.80−0.11
PFE55.00−6.08−5.86−5.533.01−2.11−1.20−1.13
PG171.20−4.68−5.68−4.263.010.760.090.31
TRV401.55−4.15−3.98−3.383.160.980.840.54
UNH126.42−3.38−3.32−2.213.550.370.431.23
UTX106.79−3.06−4.61−3.573.211.730.661.25
VZ114.79−4.63−4.75−3.993.00−0.79−0.190.73
WMT176.11−4.98−7.64−7.702.591.690.92−0.27
XOM322.34−6.14−2.92−2.883.632.070.860.64
Table A4. Dates when HARQ predictions for R V are negative. The one month rolling forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Table A4. Dates when HARQ predictions for R V are negative. The one month rolling forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
h = 1 h = 5 h = 22
AXP13 October 200813 October 200830 September 2008
CAT8 October 200813 October 2008
13 October 2008
CVX13 October 200813 October 200816 July 2008
DD 25 July 2007
GE17 September 200817 September 200817 September 2008
22 September 200822 September 200819 September 2008
22 September 2008
INTC13 October 200813 October 200813 October 2008
JNJ7 May 20107 May 201013 October 2008
7 May 2010
JPM31 December 201313 October 200813 October 2008
KO22 September 200822 September 200822 September 2008
MMM13 October 200813 October 20087 May 2010
7 May 20107 May 2010
MRK28 January 200828 January 200828 January 2008
NKE7 May 20107 May 20107 May 2010
TRV19 September 200823 April 200723 April 2007
22 September 200819 September 200819 September 2008
22 September 200822 September 2008
13 October 200826 September 2008
UNH22 September 200822 September 2008
WMT9 October 20089 October 20089 October 2008
13 October 200813 October 200813 October 2008
XOM13 October 200813 October 2008
Figure A1. Sample autocorrelations of log realized variance (solid line) and log realized quarticity (dashed line). Numbers in parentheses next to the ticker symbols are the sample correlation between to the two log realized series. The shaded area is the two standard error band for a white noise process. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Figure A1. Sample autocorrelations of log realized variance (solid line) and log realized quarticity (dashed line). Numbers in parentheses next to the ticker symbols are the sample correlation between to the two log realized series. The shaded area is the two standard error band for a white noise process. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Stats 05 00050 g0a1
Table A5. ARMA(1,1) estimates for daily log realized quarticity. The ARMA(1,1) model is parametrized as x t = ϕ x t 1 + e t + θ e t 1 where x t = ln ( R Q t ) μ , μ is the mean parameter. t are the t-ratios of the estimated parameters, ρ 1 is the first order autocorrelation of the residuals e ^ t , p 5 is the p-value from the Ljung–Box portmanteau test of residual correlation up to lag 5. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Table A5. ARMA(1,1) estimates for daily log realized quarticity. The ARMA(1,1) model is parametrized as x t = ϕ x t 1 + e t + θ e t 1 where x t = ln ( R Q t ) μ , μ is the mean parameter. t are the t-ratios of the estimated parameters, ρ 1 is the first order autocorrelation of the residuals e ^ t , p 5 is the p-value from the Ljung–Box portmanteau test of residual correlation up to lag 5. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days).
Ticker ϕ ( t ϕ ) θ ( t θ ) ρ 1 [ p 5 ]
AXP 0.989 ( 379.8 ) 0.769 ( 56.3 ) 0.079 [0.000]
BA 0.982 ( 264.4 ) 0.798 ( 59.1 ) 0.055 [0.000]
CAT 0.985 ( 304.1 ) 0.796 ( 61.5 ) 0.074 [0.000]
CSCO 0.983 ( 296.6 ) 0.717 ( 46.2 ) 0.084 [0.000]
CVX 0.977 ( 241.8 ) 0.742 ( 55.0 ) 0.036 [0.035]
DD 0.985 ( 305.3 ) 0.793 ( 58.4 ) 0.063 [0.000]
DIS 0.985 ( 302.6 ) 0.792 ( 57.6 ) 0.060 [0.000]
GE 0.981 ( 278.2 ) 0.713 ( 45.8 ) 0.067 [0.000]
HD 0.988 ( 349.7 ) 0.802 ( 64.7 ) 0.068 [0.000]
IBM 0.983 ( 292.6 ) 0.743 ( 53.9 ) 0.052 [0.000]
INTC 0.984 ( 299.0 ) 0.735 ( 47.5 ) 0.084 [0.000]
JNJ 0.986 ( 326.6 ) 0.799 ( 62.8 ) 0.075 [0.000]
JPM 0.985 ( 320.0 ) 0.736 ( 49.9 ) 0.072 [0.000]
KO 0.987 ( 334.2 ) 0.805 ( 63.9 ) 0.064 [0.000]
MCD 0.992 ( 422.7 ) 0.842 ( 72.0 ) 0.083 [0.000]
MMM 0.982 ( 275.0 ) 0.788 ( 59.2 ) 0.071 [0.000]
MRK 0.980 ( 241.5 ) 0.806 ( 57.3 ) 0.075 [0.000]
MSFT 0.980 ( 266.5 ) 0.724 ( 48.5 ) 0.063 [0.000]
NKE 0.991 ( 390.8 ) 0.844 ( 64.2 ) 0.103 [0.000]
PFE 0.982 ( 267.6 ) 0.793 ( 56.2 ) 0.084 [0.000]
PG 0.984 ( 302.9 ) 0.771 ( 56.4 ) 0.063 [0.000]
TRV 0.986 ( 322.2 ) 0.786 ( 61.8 ) 0.039 [0.009]
UNH 0.981 ( 253.0 ) 0.794 ( 54.8 ) 0.078 [0.000]
UTX 0.982 ( 272.9 ) 0.791 ( 57.6 ) 0.041 [0.007]
VZ 0.983 ( 282.4 ) 0.788 ( 56.0 ) 0.074 [0.000]
WMT 0.989 ( 373.0 ) 0.798 ( 61.1 ) 0.068 [0.000]
XOM 0.976 ( 239.6 ) 0.709 ( 48.6 ) 0.051 [0.000]

Appendix A.4. Forecast Error Diagnostics

Figure A2. Autocorrelations of forecast errors e t = y t ( h ) y ^ t ( h ) for stepsize h = 1 day. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with y = ln ( R V ) (red line) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A2. Autocorrelations of forecast errors e t = y t ( h ) y ^ t ( h ) for stepsize h = 1 day. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with y = ln ( R V ) (red line) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a2
Figure A3. Autocorrelations of forecast errors e t = y t h y ^ t ( h ) for stepsize h = 5 days. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with log of average R V (red) and with log of average R V (blue) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A3. Autocorrelations of forecast errors e t = y t h y ^ t ( h ) for stepsize h = 5 days. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with log of average R V (red) and with log of average R V (blue) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a3
Figure A4. Autocorrelations of forecast errors e t = y t h y ^ t ( h ) for stepsize h = 22 days. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with log of average R V (red) and with log of average R V (blue) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A4. Autocorrelations of forecast errors e t = y t h y ^ t ( h ) for stepsize h = 22 days. The autocorrelations for HAR with y = R V (gray), for HARQ with y = R V (dashed gray), and for R Q -in-mean with log of average R V (red) and with log of average R V (blue) should all be white noise if e t is i i d . The gray shaded area is the two standard error band for a white noise series. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a4

Appendix A.5. Running t-Ratios

Figure A5. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is the exponential transform of the log forecast from model (Section 3.1) assuming log-normality. A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A5. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is the exponential transform of the log forecast from model (Section 3.1) assuming log-normality. A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a5
Figure A6. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is a simple exponential transform of the log forecast from model (Section 3.1). A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A6. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is a simple exponential transform of the log forecast from model (Section 3.1). A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a6
Figure A7. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is the exponential transform of the log forecast from model (Section 3.1) adjusted with a smearing factor [25]. A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Figure A7. Running t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. Each panel has 27 lines for the 27 stocks in the sample. The baseline forecast is from the HARQ model (no transformation) and the comparison forecast is the exponential transform of the log forecast from model (Section 3.1) adjusted with a smearing factor [25]. A positive value indicates better forecast from the comparison model against the baseline HARQ forecasts. M S for equality of mean squared error loss and Q L for QLIKE loss of Patton [24]. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Stats 05 00050 g0a7

References

  1. Corsi, F. A Simple Approximate Long-Memory Model of Realized Volatility. J. Financ. Econom. 2009, 7, 174–196. [Google Scholar] [CrossRef]
  2. Rockinger, M.; Jondeau, E. Entropy densities with an application to autoregressive conditional skewness and kurtosis. J. Econom. 2002, 106, 119–142. [Google Scholar] [CrossRef]
  3. Jondeau, E.; Rockinger, M. Conditional volatility, skewness, and kurtosis: Existence, persistence, and comovements. J. Econ. Dyn. Control 2003, 27, 1699–1737. [Google Scholar] [CrossRef]
  4. Bali, T.G.; Mo, H.; Tang, Y. The role of autoregressive conditional skewness and kurtosis in the estimation of conditional VaR. J. Bank. Financ. 2008, 32, 269–282. [Google Scholar] [CrossRef]
  5. Ding, Z.; Granger, C.W.; Engle, R.F. A long memory property of stock market returns and a new model. J. Empir. Financ. 1993, 1, 83–106. [Google Scholar] [CrossRef]
  6. Comte, F.; Renault, E. Long memory in continuous-time stochastic volatility models. Math. Financ. 1998, 8, 291–323. [Google Scholar] [CrossRef]
  7. Comte, F.; Coutin, L.; Renault, E. Affine fractional stochastic volatility models. Ann. Financ. 2012, 8, 337–378. [Google Scholar] [CrossRef]
  8. Gatheral, J.; Jaisson, T.; Rosenbaum, M. Volatility is rough. Quant. Financ. 2018, 18, 933–939. [Google Scholar] [CrossRef]
  9. Huang, D.; Schlag, C.; Shaliastovich, I.; Thimme, J. Volatility-of-Volatility Risk. J. Financ. Quant. Anal. 2019, 54, 2423–2452. [Google Scholar] [CrossRef]
  10. Da Fonseca, J.; Zhang, W. Volatility of volatility is (also) rough. J. Futur. Mark. 2019, 39, 600–611. [Google Scholar] [CrossRef]
  11. Hansen, P.R.; Huang, Z.; Shek, H.H. Realized GARCH: A joint model for returns and realized measures of volatility. J. Appl. Econom. 2012, 27, 877–906. [Google Scholar] [CrossRef]
  12. Bollerslev, T.; Patton, A.J.; Quaedvlieg, R. Exploiting the errors: A simple approach for improved volatility forecasting. J. Econom. 2016, 192, 1–18. [Google Scholar] [CrossRef]
  13. Buccheri, G.; Corsi, F. HARK the SHARK: Realized Volatility Modeling with Measurement Errors and Nonlinear Dependencies. J. Financ. Econom. 2021, 19, 614–649. [Google Scholar] [CrossRef]
  14. Creal, D.; Koopman, S.J.; Lucas, A. Generalized Autoregressive Score Models with Applications. J. Appl. Econom. 2013, 28, 777–795. [Google Scholar] [CrossRef]
  15. Ding, Y.D. A simple joint model for returns, volatility and volatility of volatility. J. Econom. 2021; in press. [Google Scholar] [CrossRef]
  16. Ghysels, E.; Santa-Clara, P.; Valkanov, R. There is Risk-Return Trade-off After all. J. Financ. Econ. 2005, 76, 509–548. [Google Scholar] [CrossRef]
  17. Corsi, F.; Mittnik, S.; Pigorsch, C.; Pigorsch, U. The Volatility of Realized Volatility. Econom. Rev. 2008, 27, 46–78. [Google Scholar] [CrossRef]
  18. Andersen, T.G.; Bollerslev, T.; Diebold, F.X.; Labys, P. Modeling and Forecasting Realized Volatility. Econometrica 2003, 71, 529–626. [Google Scholar] [CrossRef]
  19. Geweke, J.; Porter-Hudak, S. The Estimation and Application of Long Memory Time Series Model. J. Time Ser. Anal. 1983, 4, 221–238. [Google Scholar] [CrossRef]
  20. Hansen, P.R.; Huang, Z. Exponential GARCH Modeling With Realized Measures of Volatility. J. Bus. Econ. Stat. 2016, 34, 269–287. [Google Scholar] [CrossRef]
  21. Engle, R.F.; Lilien, D.M.; Robins, R.P. Estimating Time Varying Risk Premia in the Term Structure: The ARCH-M Model. Econometrica 1987, 55, 391–407. [Google Scholar] [CrossRef]
  22. Nelson, D.B. Conditional Heteroskedasticity in Asset Returns: A New Approach. Econometrica 1991, 59, 347–370. [Google Scholar] [CrossRef]
  23. Barndorff-Nielsen, O.E.; Shephard, N. How Accurate is the Asymptotic Approximation to the Distribution of Realised Variance? In Identification and Inference for Econometric Models; Andrews, D.W.K., Stock, J.H., Eds.; Cambridge University Press: Cambridge, UK, 2005; Chapter 13; pp. 306–331. [Google Scholar]
  24. Patton, A.J. Volatility forecast comparison using imperfect volatility proxies. J. Econom. 2011, 160, 246–256. [Google Scholar] [CrossRef]
  25. Duan, N. Smearing Estimate: A Nonparametric Retransformation Method. J. Am. Stat. Assoc. 1983, 78, 605–610. [Google Scholar] [CrossRef]
  26. Wooldridge, J.M. Introductory Econometrics: A Modern Approach, 7th ed.; Cengage Learning: Boston, MA, USA, 2020. [Google Scholar]
  27. Diebold, F.X.; Mariano, R.S. Comparing Predictive Accuracy. J. Bus. Econ. Stat. 1995, 13, 253–263. [Google Scholar]
  28. Barndorff-Nielsen, O.E.; Hansen, P.R.; Lunde, A.; Shephard, N. Designing Realized Kernels to Measure the ex post Variation of Equity Prices in the Presence of Noise. Econometrica 2008, 76, 1481–1536. [Google Scholar]
  29. Bickel, P.J.; Li, B.; Tsybakov, A.B.; van de Geer, S.A.; Yu, B.; Valdés, T.; Rivero, C.; Fan, J.; van der Vaart, A. Regularization in statistics. Test 2006, 15, 271–344. [Google Scholar] [CrossRef]
  30. Hastie, T. Ridge Regularization: An Essential Concept in Data Science. Technometrics 2020, 62, 426–433. [Google Scholar] [CrossRef]
  31. Barndorff-Nielsen, O.E.; Shephard, N. Econometric analysis of realized volatility and its use in estimating stochastic volatility models. J. R. Stat. Soc. Ser. B 2002, 64, 253–280. [Google Scholar] [CrossRef]
  32. Barndorff-Nielsen, O.E.; Shephard, N. Estimating Quadratic Variation Using Realised Variance. J. Appl. Econom. 2002, 17, 457–477. [Google Scholar] [CrossRef]
  33. Barndorff-Nielsen, O.E.; Shephard, N. Power and Bipower Variation with Stochastic Volatility and Jumps. J. Financ. Econom. 2004, 2, 1–37. [Google Scholar] [CrossRef]
  34. Barndorff-Nielsen, O.E.; Shephard, N.; Winkel, M. Limit theorems for multipower variation in the presence of jumps. Stoch. Process. Appl. 2006, 116, 796–806. [Google Scholar] [CrossRef] [Green Version]
  35. Andersen, T.G.; Bollerslev, T.; Diebold, F.X. Roughing it up: Including jump components in the measurement, modeling and forecasting of return volatility. Rev. Econ. Stat. 2007, 89, 701–720. [Google Scholar] [CrossRef]
Figure 1. Rolling parameter estimates for log-in-mean specification (Section 3.1) for stepsizes h = 1 ,   5 ,   22 . For h > 1 , two models are estimated, one using the log of the average of R V (red) and one with the log of the average of R V (blue). The shaded area is the interquartile range across the 27 stocks and the solid line is the median. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days). The rolling estimation window is 8.67 years (104 months) followed by a one-month prediction window shifted each month over the sample for 96 estimation windows.
Figure 1. Rolling parameter estimates for log-in-mean specification (Section 3.1) for stepsizes h = 1 ,   5 ,   22 . For h > 1 , two models are estimated, one using the log of the average of R V (red) and one with the log of the average of R V (blue). The shaded area is the interquartile range across the 27 stocks and the solid line is the median. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days). The rolling estimation window is 8.67 years (104 months) followed by a one-month prediction window shifted each month over the sample for 96 estimation windows.
Stats 05 00050 g001
Figure 2. Rolling estimates of the implied persistence parameter ρ for different stepsize h. For h > 1 , two models are estimated: one using the log of the average of R V (red) and one with the log of the average of R V (blue). The shaded area is the interquartile range across the 27 stocks and the solid line is the median. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days). The rolling estimation window is 8.67 years (104 months) followed by a one-month prediction window shifted each month over the sample for 96 windows.
Figure 2. Rolling estimates of the implied persistence parameter ρ for different stepsize h. For h > 1 , two models are estimated: one using the log of the average of R V (red) and one with the log of the average of R V (blue). The shaded area is the interquartile range across the 27 stocks and the solid line is the median. The daily sample is from 22 April 1997 to 31 December 2013 (4200 trading days). The rolling estimation window is 8.67 years (104 months) followed by a one-month prediction window shifted each month over the sample for 96 windows.
Stats 05 00050 g002
Table 1. t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. The baseline forecast from the HARQ model (no transformation) is compared against the forecast from the exponential transform of the log forecast from model (Section 3.1) assuming log-normality. A positive τ value indicates a better forecast from the comparison model against the baseline HARQ forecasts. τ M S for equality of mean squared error loss and τ Q L for QLIKE loss of Patton [24]. R V ^ < 0 is the number of negative forecasts (in the forecast sample) produced by the HARQ model. The bottom four rows are the fraction out of 27 stocks that satisfy the inequalities; for R V ^ < 0 , it is the fraction of 27 stocks that produced at least one negative forecast. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one-month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
Table 1. t-ratios of Diebold and Mariano [27] test of equal predictive accuracy of h-step R V rolling forecasts. The baseline forecast from the HARQ model (no transformation) is compared against the forecast from the exponential transform of the log forecast from model (Section 3.1) assuming log-normality. A positive τ value indicates a better forecast from the comparison model against the baseline HARQ forecasts. τ M S for equality of mean squared error loss and τ Q L for QLIKE loss of Patton [24]. R V ^ < 0 is the number of negative forecasts (in the forecast sample) produced by the HARQ model. The bottom four rows are the fraction out of 27 stocks that satisfy the inequalities; for R V ^ < 0 , it is the fraction of 27 stocks that produced at least one negative forecast. The rolling forecasts are generated with a rolling estimation window of 8.67 years (104 months) followed by a one-month prediction window shifted each month over 96 windows. The forecast sample is from 3 January 2006 to 31 December 2013 (2013 trading days).
h = 1 h = 5 h = 22
τ MS τ QL RV ^ < 0 τ MS τ QL RV ^ < 0 τ MS τ QL RV ^ < 0
AXP 0.94 2.65 1 0.92 0.96 1 0.14 1.75 1
BA 2.12 0.55 0 1.44 0.44 0 0.31 0.43 0
CAT 1.10 3.33 2 1.04 0.58 1 1.05 1.25 0
CSCO 2.48 6.23 0 1.69 2.46 0 1.80 0.92 0
CVX 0.89 1.40 1 0.96 0.56 1 1.16 1.28 1
DD 2.49 0.71 0 0.04 0.32 0 0.99 2.16 1
DIS 1.47 2.47 0 0.79 2.09 0 1.47 0.73 0
GE 1.05 0.86 2 0.66 1.28 2 0.61 2.57 3
HD 1.96 2.62 0 1.24 1.96 0 0.01 0.89 0
IBM 0.28 2.17 0 0.02 2.43 0 0.95 0.21 0
INTC 1.60 6.43 1 0.49 5.41 1 0.20 1.08 1
JNJ 1.64 3.24 1 1.53 1.70 1 0.89 0.13 2
JPM 1.65 4.18 1 1.64 0.12 1 1.65 2.42 1
KO 1.49 4.84 1 0.10 0.53 1 0.39 2.39 1
MCD 0.01 1.65 0 0.49 6.11 0 0.41 5.44 0
MMM 1.03 5.68 2 1.20 0.91 2 1.11 0.59 1
MRK 0.98 1.35 1 1.01 1.57 1 1.03 4.62 1
MSFT 1.96 2.62 0 1.44 0.07 0 1.37 0.34 0
NKE 2.53 5.08 1 1.38 2.04 1 1.24 0.85 1
PFE 2.14 3.35 0 1.68 0.82 0 1.14 0.55 0
PG 0.51 0.41 0 0.87 1.27 0 1.71 2.22 0
TRV 1.15 4.57 2 1.01 2.84 4 1.14 1.90 4
UNH 0.94 4.75 1 0.32 2.72 1 0.85 0.43 0
UTX 0.57 0.38 0 0.95 3.51 0 0.21 0.53 0
VZ 1.31 3.06 0 0.73 3.78 0 0.12 0.26 0
WMT 0.54 1.16 2 0.81 1.52 2 0.67 1.45 2
XOM 0.59 0.94 1 1.01 2.12 1 1.11 0.20 0
τ < 0 [0.63][0.85][0.56][0.52][0.67][0.56][0.56][0.22][0.48]
τ < 2 [0.19][0.59] [0.00][0.37] [0.00][0.00]
τ > 0 [0.37][0.15] [0.48][0.33] [0.44][0.78]
τ > 2 [0.00][0.04] [0.00][0.04] [0.00][0.26]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kawakatsu, H. Modeling Realized Variance with Realized Quarticity. Stats 2022, 5, 856-880. https://doi.org/10.3390/stats5030050

AMA Style

Kawakatsu H. Modeling Realized Variance with Realized Quarticity. Stats. 2022; 5(3):856-880. https://doi.org/10.3390/stats5030050

Chicago/Turabian Style

Kawakatsu, Hiroyuki. 2022. "Modeling Realized Variance with Realized Quarticity" Stats 5, no. 3: 856-880. https://doi.org/10.3390/stats5030050

APA Style

Kawakatsu, H. (2022). Modeling Realized Variance with Realized Quarticity. Stats, 5(3), 856-880. https://doi.org/10.3390/stats5030050

Article Metrics

Back to TopTop