3.2. Model Estimation and Fit
This sections explains how GCLEs can be fit to the empirical data described above. First of all, we modeled the three mentioned returns with a trivariate spherical law. More precisely, we considered a trivariate Gaussian, logistic, and hyperbolic-secant density.
According to Equations (
24), (
26) and (
28) in
Section 2, the spherical Gaussian, logistic, and hyperbolic-secant laws, denoted by
,
, and
, are given by:
Taking into account Equations (
32)–(
34) in Corollary 2, the GCLE versions of these distributions, denoted by GCN, GCL, and GCHS respectively, are:
Here,
is the excess-kurtosis of the empirical trivariate series with respect to the (Mardia) kurtosis index (see Equation (12) in
Appendix A) of the spherical Gaussian, logistic, and hyperbolic-secant laws, given in Equations (
35)–(
37).
By denoting with
the kurtosis of the empirical trivariate series and with
K the (Mardia) kurtosis index of the spherical Gaussian, logistic, and hyperbolic-secant law that have been fit, in turn, to the empirical series, the excess kurtosis
can be estimated via the Method of Moments (MM) as follows:
Alternatively, Maximum Likelihood (ML) estimation can be employed to estimate
,
, and
from the data, numerically maximizing (via the
optim function available in the R software, (
R Core Team 2020)) the log-likelihood:
where
is one of the trivariate densities in (
38)–(
40) and
, thus obtaining the related
estimate.
Table 2 reports the estimates of
, the AIC information criterion values for the GCLEs, together with those of the multivariate Student
t (M
t;
Kotz and Nadarajah 2004) and the power exponential (MPE;
Gómez et al. 1998) distributions, which have been fit to the series for comparison reasons. Furthermore, for the GCLEs, the levels of kurtosis of each parent distribution (
) and of the associated expansions (
) are also reported. For the MM distribution, the latter clearly corresponds to the empirical Mardia’s kurtosis index of the data.
Looking at the AIC results, when these criteria are applicable, we conclude that GCLEs fit better the empirical series than the other distributions, and this is particularly true for the GCHS.
In order to better investigate the goodness of fit of these polynomially adjusted densities in comparison with the popular M
t and MPE distributions, the multivariate version of the Wilcoxon rank test (
Liu and Singh 1993) was also employed. To implement the test, the following procedure was devised:
An artificial population of one million data points was generated from each distribution, using the
estimates in
Table 2. For the GCLE distributions, since no standard routine is available, the Hamiltonian Monte Carlo (HMC;
Neal 2011) was employed;
samples were extracted from each population. The MW test was performed via the R package
DepthProc (
Kosiorowski and Zawadzki 2014) on each of the samples coming from each population. This yields a
B-dimensional vector containing the binary outcome of the tests (acceptance = 0/rejection = 1);
A non-parametric bootstrap procedure was then performed on the above vector, to obtain a confidence interval on the proportion of rejections. The higher the proportion of rejections is, the worse the performance of the underlying model.
Table 3 shows the average
p-value of the MW test across the
B replicates, along with the proportion of rejections
and its 95% confidence interval
, for each model. According to the test, a
p-value higher than the significance threshold provides evidence that the two datasets come from the same underlying population. In our situation, one dataset represents a simulation from a given model (thus, it can be regarded as a proxy of the model itself), while the other represents the empirical data. This entails that the models with an average
p-value above the significance threshold have a sufficiently adequate fit to the data. The results provide evidence that GCLEs perform well compared to M
t and MPE, with the only exception of the GCN, since they are the only distributions with an average
p-value above the 10% level. Looking at the results shown in the other columns of the table, we can see that GCLEs show also a much lower rate of rejection of the null hypothesis of the MW test across the
B replications, except for the GCN. Furthermore, looking at both the average MW
p-value and the rejection rate, we conclude that the GCHS distributions obtained via the method of moments estimation also adequately fit the data.
3.3. Evaluation of VaR and ES via GCLEs
So far, we have evaluated the capability of the GCLEs to fit financial series. In the following, the validity of these densities in computing risk measures such as the value at risk and the expected shortfall is also investigated (see
Chen 2018;
McNeil et al. 2005), both in- and out-of-sample, for a linear portfolio:
To this end, the entire observation period, which ranges from 1 January 2009 to 31 December 2012 (1903 observations), was split into two sub-periods having approximately the same length: , going from 1 January 2009 to 31 December 2012 (952 observations), and , going from 1 January 2013 to 31 December 2016 (951 observations).
The observations of the period were used to obtain parameter estimates of the GCLEs, the Mt, and the MPE law. Then, the latter were employed to estimate risk measures, such as the value at risk and the expected shortfall.
Given the spherical nature of the mentioned distributions,
and
were estimated by using the approach proposed by
Kamdem (
2005). The value at risk at a given level
,
, of a linear portfolio described by an
n-dimensional elliptical random variable
with vector mean
and covariance matrix
, was evaluated as:
Here,
is a vector of weights,
with
as defined in (
17), and
is the unique positive solution of the following equation:
In (
45),
is the density of the vector
, and
denotes the gamma function. In the case of GCLEs of spherical variables and by setting equal weights
, Equation (
44) simply becomes:
where
is the solution of (
45) obtained by replacing
with
defined as in (
15).
Following
Kamdem (
2005) and
Nesmith et al. (
2017), the expected shortfall at level
,
, is given by:
where the symbols are defined as in (
44)–(
46). In the case of GCLEs of spherical variables, by setting equal weights in
and by replacing
in (
47) with
as defined in (
15), the above equation becomes:
Equation (
48) was used to compute
by using
estimated as in (
44). In particular, GCLEs as defined in (
38)–(
40) were employed to compute
as given in (
48). The empirical counterparts of VaR and ES,
and
, respectively, of the linear portfolio
are given by:
where
is the empirical cumulative distribution function of the portfolio defined as in Equation (
43).
Table 4 compares the estimates of
, computed as in (
46) by using GCN, GCL, and GCHS, with the corresponding
in the first sub-sample
. In this table, the percentile bootstrap intervals,
, for
at the 95% confidence level are also reported, along with the absolute difference between
and
. These intervals were obtained by extracting 1000 bootstrap samples from the whitened data of the first period. To ensure preservation of the structure of the data, the maximum entropy bootstrap of the method of (
Vinod 2006) was employed, via the R package
meboot (
Vinod and López-de Lacalle 2009). A distribution whose
falls inside the 95% confidence interval of
can be regarded as adequate for this risk measure.
Looking at the results shown in the table, we see that GCLEs manage to fall inside the confidence intervals across all ’s. In particular, the GCN is adequate for the two highest risk levels, while the GCL and the GCHS are adequate for the two lowest risk levels. Neither the Mt nor the MPE follow a particular pattern and tend to underestimate the risk for and . The only distribution that is able to estimate satisfactorily across all levels of is the GCHS estimated via the method of moments. Finally, looking at the results reported in the last column of the table, we can conclude that the GCHS distributions provide the closest fit to for and , while the MPE the closest fit for and the Mt the closest fit for , immediately followed by the GCHS.
A similar analysis was carried out for the expected shortfall.
Table 5 shows
and
for the period
together with the percentile bootstrap intervals,
, at the 95% confidence level. Looking at the results reported in the table, we see that the ES is correctly estimated by GCLEs across all levels of
with the exception of the GCL, which slightly underestimates the expected VaR-exceeding losses for
; the same does not occur for both the M
t and the MPE densities, especially for
. Looking at the results provided in the last column of the table, we can conclude that GCHS distributions offer the best estimates of
for all levels of
, except for
, which is better estimated by GCN.
In particular, the last two tables highlight how the most leptokurtic distribution, the GCHS, has a great advantage in estimating losses associated with high risk levels, thanks to a tail heaviness that is more pronounced relatively to the other GCLEs
The out-of sample performance of the CGLE in estimating the value at risk was evaluated by using Kupiec’s Likelihood Ratio test (LR;
Kupiec 1995) and the Binomial Test (BT;
Lee and Su 2012). Both the LR and BT null hypotheses state that the percentage of portfolio losses that in the second part of the sample exceed
is equal to
. Since
is estimated via GCLEs in the first period of the sample, this analysis should provide evidence of the out-of-sample stability of the GCLEs in assessing this risk measure.
The
p-values of both the LR (
p-value LR) and binomial test (
p-value BT) are shown in
Table 6 together with two loss measures: the Average Binary Loss Function (ABLF) and the Average Quadratic Loss Function (AQLF). The former is the average of the Binary Loss (BL) function, which gives a penalty of one when each return
exceeds the VaR, without concern for the magnitude of the excess:
If a VaR model truly provides the level of coverage defined by its confidence level
, then the ABLF should be as close as possible to the latter.
The AQLF (
Lopez 1999) is the average of the Quadratic Loss (QL) function:
and pays more attention to the magnitude of the violations, as large violations are more penalized than the small ones.
Table 6 shows the results obtained by performing the LR and the BN test. They confirm the good performance offered by GCLEs, with the exception of the GCN. Looking at the loss functions, we can conclude that the GCHS outperforms all the other distributions for
. In particular, the GCL and the M
t are equivalent for
; the GCN (ML) and the GCHS (MM) provide the best fitting for
(the former for the ABLF, the latter for the AQLF); and the GCN is the only distribution not underestimating the ABLF for
, which is better approximated by the MPE according to the AQLF.
Finally, the out-of-sample GCLE performance in estimating the ES was evaluated by performing the
and
tests of
Acerbi and Szekely (
2014). The null hypothesis of both tests assumes that the estimated
tallies with the ES calculated from the data. This entails, in the present application, that
should provide a good estimate of the empirical expected shortfall computed in the second part of the sample. Under the alternative, this does not happen, and
systematically underestimates the effective loss mean,
, thus implying financial damage.
The first test is based on the following statistic:
where
is the number of losses
, which in the second part of the sample, break the threshold
, and
is an indicator variable that is equal to one if
and zero otherwise.
As observed by
Acerbi and Szekely (
2014), this test is completely insensitive to an excessive number of exceptions as it is simply an average taken over the exceptions themselves. Under the null, the realized value
is expected to be zero, and it signals a problem when it displays values significantly greater than zero.
The second test statistic is:
where
T denotes the sample size of the second period, and the other symbols are defined as above. As for the previous test, under the null, the realized value
is expected to be zero, and it signals a problem when it assumes values significantly greater than zero.
To perform the test and obtain the associated p-values, the following parametric bootstrap approach was used:
A population of one million observations was generated from each distribution under investigation (Mt, MPE, or one of the GCLEs considered in the paper), with parameters estimated by using the data of the first sample period. For the class of the GCLEs, HMC was employed to this aim;
B = 5000 samples, of a size equal to the dimension of the out-of-sample dataset, were drawn from each of these populations;
the bootstrap distributions of the statistics
and
were obtained and used to calculate the
p-values,
, of the statistics
and
computed by using the observed sample data:
The results are reported in
Table 7. They allow concluding that the GCLEs adequately estimate the ES, with the exception of the GCN. According to the
statistic, the M
t distribution underestimates the expected losses for
, while the MPE underestimates the ES according to the
statistic for
. As in the case of the in-sample assessment, distributions that are appreciably more leptokurtic than the GCN show a better tail sensitivity and are thus particularly adequate to evaluate risk measures at high levels of
.