3.2. Model Estimation and Fit
This sections explains how GCLEs can be fit to the empirical data described above. First of all, we modeled the three mentioned returns with a trivariate spherical law. More precisely, we considered a trivariate Gaussian, logistic, and hyperbolic-secant density.
According to Equations (
24), (
26) and (
28) in 
Section 2, the spherical Gaussian, logistic, and hyperbolic-secant laws, denoted by 
, 
, and 
, are given by:
Taking into account Equations (
32)–(
34) in Corollary 2, the GCLE versions of these distributions, denoted by GCN, GCL, and GCHS respectively, are: 
Here, 
 is the excess-kurtosis of the empirical trivariate series with respect to the (Mardia) kurtosis index (see Equation (12) in 
Appendix A) of the spherical Gaussian, logistic, and hyperbolic-secant laws, given in Equations (
35)–(
37).
By denoting with 
 the kurtosis of the empirical trivariate series and with 
K the (Mardia) kurtosis index of the spherical Gaussian, logistic, and hyperbolic-secant law that have been fit, in turn, to the empirical series, the excess kurtosis 
 can be estimated via the Method of Moments (MM) as follows:
Alternatively, Maximum Likelihood (ML) estimation can be employed to estimate 
, 
, and 
 from the data, numerically maximizing (via the 
optim function available in the R software, (
R Core Team 2020)) the log-likelihood:
        where 
 is one of the trivariate densities in (
38)–(
40) and 
, thus obtaining the related 
 estimate.
Table 2 reports the estimates of 
, the AIC information criterion values for the GCLEs, together with those of the multivariate Student 
t (M
t; 
Kotz and Nadarajah 2004) and the power exponential (MPE; 
Gómez et al. 1998) distributions, which have been fit to the series for comparison reasons. Furthermore, for the GCLEs, the levels of kurtosis of each parent distribution (
) and of the associated expansions (
) are also reported. For the MM distribution, the latter clearly corresponds to the empirical Mardia’s kurtosis index of the data.
 Looking at the AIC results, when these criteria are applicable, we conclude that GCLEs fit better the empirical series than the other distributions, and this is particularly true for the GCHS.
In order to better investigate the goodness of fit of these polynomially adjusted densities in comparison with the popular M
t and MPE distributions, the multivariate version of the Wilcoxon rank test (
Liu and Singh 1993) was also employed. To implement the test, the following procedure was devised:
An artificial population of one million data points was generated from each distribution, using the 
 estimates in 
Table 2. For the GCLE distributions, since no standard routine is available, the Hamiltonian Monte Carlo (HMC; 
Neal 2011) was employed;
 samples were extracted from each population. The MW test was performed via the R package 
DepthProc (
Kosiorowski and Zawadzki 2014) on each of the samples coming from each population. This yields a 
B-dimensional vector containing the binary outcome of the tests (acceptance = 0/rejection = 1);
A non-parametric bootstrap procedure was then performed on the above vector, to obtain a confidence interval on the proportion of rejections. The higher the proportion of rejections is, the worse the performance of the underlying model.
Table 3 shows the average 
p-value of the MW test across the 
B replicates, along with the proportion of rejections 
 and its 95% confidence interval 
, for each model. According to the test, a 
p-value higher than the significance threshold provides evidence that the two datasets come from the same underlying population. In our situation, one dataset represents a simulation from a given model (thus, it can be regarded as a proxy of the model itself), while the other represents the empirical data. This entails that the models with an average 
p-value above the significance threshold have a sufficiently adequate fit to the data. The results provide evidence that GCLEs perform well compared to M
t and MPE, with the only exception of the GCN, since they are the only distributions with an average 
p-value above the 10% level. Looking at the results shown in the other columns of the table, we can see that GCLEs show also a much lower rate of rejection of the null hypothesis of the MW test across the 
B replications, except for the GCN. Furthermore, looking at both the average MW 
p-value and the rejection rate, we conclude that the GCHS distributions obtained via the method of moments estimation also adequately fit the data.
   3.3. Evaluation of VaR and ES via GCLEs
So far, we have evaluated the capability of the GCLEs to fit financial series. In the following, the validity of these densities in computing risk measures such as the value at risk and the expected shortfall is also investigated (see 
Chen 2018; 
McNeil et al. 2005), both in- and out-of-sample, for a linear portfolio:
To this end, the entire observation period, which ranges from 1 January 2009 to 31 December 2012 (1903 observations), was split into two sub-periods having approximately the same length: , going from 1 January 2009 to 31 December 2012 (952 observations), and , going from 1 January 2013 to 31 December 2016 (951 observations).
The observations of the period  were used to obtain parameter estimates of the GCLEs, the Mt, and the MPE law. Then, the latter were employed to estimate risk measures, such as the value at risk and the expected shortfall.
Given the spherical nature of the mentioned distributions, 
 and 
 were estimated by using the approach proposed by 
Kamdem (
2005). The value at risk at a given level 
, 
, of a linear portfolio described by an 
n-dimensional elliptical random variable 
 with vector mean 
 and covariance matrix 
, was evaluated as:
Here, 
 is a vector of weights, 
 with 
 as defined in (
17), and 
 is the unique positive solution of the following equation:
In (
45), 
 is the density of the vector 
, and 
 denotes the gamma function. In the case of GCLEs of spherical variables and by setting equal weights 
, Equation (
44) simply becomes:
        where 
 is the solution of (
45) obtained by replacing 
 with 
 defined as in (
15).
Following 
Kamdem (
2005) and 
Nesmith et al. (
2017), the expected shortfall at level 
, 
, is given by:
        where the symbols are defined as in (
44)–(
46). In the case of GCLEs of spherical variables, by setting equal weights in 
 and by replacing 
 in (
47) with 
 as defined in (
15), the above equation becomes:
Equation (
48) was used to compute 
 by using 
 estimated as in (
44). In particular, GCLEs as defined in (
38)–(
40) were employed to compute 
 as given in (
48). The empirical counterparts of VaR and ES, 
 and 
, respectively, of the linear portfolio 
 are given by:
        where 
 is the empirical cumulative distribution function of the portfolio defined as in Equation (
43).
Table 4 compares the estimates of 
, computed as in (
46) by using GCN, GCL, and GCHS, with the corresponding 
 in the first sub-sample 
. In this table, the percentile bootstrap intervals, 
, for 
 at the 95% confidence level are also reported, along with the absolute difference between 
 and 
. These intervals were obtained by extracting 1000 bootstrap samples from the whitened data of the first period. To ensure preservation of the structure of the data, the maximum entropy bootstrap of the method of (
Vinod 2006) was employed, via the R package 
meboot (
Vinod and López-de Lacalle 2009). A distribution whose 
 falls inside the 95% confidence interval of 
 can be regarded as adequate for this risk measure.
 Looking at the results shown in the table, we see that GCLEs manage to fall inside the confidence intervals across all ’s. In particular, the GCN is adequate for the two highest risk levels, while the GCL and the GCHS are adequate for the two lowest risk levels. Neither the Mt nor the MPE follow a particular pattern and tend to underestimate the risk for  and . The only distribution that is able to estimate  satisfactorily across all levels of  is the GCHS estimated via the method of moments. Finally, looking at the results reported in the last column of the table, we can conclude that the GCHS distributions provide the closest fit to  for  and , while the MPE the closest fit for  and the Mt the closest fit for , immediately followed by the GCHS.
A similar analysis was carried out for the expected shortfall.
Table 5 shows 
 and 
 for the period 
 together with the percentile bootstrap intervals, 
, at the 95% confidence level. Looking at the results reported in the table, we see that the ES is correctly estimated by GCLEs across all levels of 
 with the exception of the GCL, which slightly underestimates the expected VaR-exceeding losses for 
; the same does not occur for both the M
t and the MPE densities, especially for 
. Looking at the results provided in the last column of the table, we can conclude that GCHS distributions offer the best estimates of 
 for all levels of 
, except for 
, which is better estimated by GCN.
 In particular, the last two tables highlight how the most leptokurtic distribution, the GCHS, has a great advantage in estimating losses associated with high risk levels, thanks to a tail heaviness that is more pronounced relatively to the other GCLEs
The out-of sample performance of the CGLE in estimating the value at risk was evaluated by using Kupiec’s Likelihood Ratio test (LR; 
Kupiec 1995) and the Binomial Test (BT; 
Lee and Su 2012). Both the LR and BT null hypotheses state that the percentage of portfolio losses that in the second part of the sample exceed 
 is equal to 
. Since 
 is estimated via GCLEs in the first period of the sample, this analysis should provide evidence of the out-of-sample stability of the GCLEs in assessing this risk measure.
The 
p-values of both the LR (
p-value LR) and binomial test (
p-value BT) are shown in 
Table 6 together with two loss measures: the Average Binary Loss Function (ABLF) and the Average Quadratic Loss Function (AQLF). The former is the average of the Binary Loss (BL) function, which gives a penalty of one when each return 
 exceeds the VaR, without concern for the magnitude of the excess:
		If a VaR model truly provides the level of coverage defined by its confidence level 
, then the ABLF should be as close as possible to the latter.
The AQLF (
Lopez 1999) is the average of the Quadratic Loss (QL) function:
        and pays more attention to the magnitude of the violations, as large violations are more penalized than the small ones.
Table 6 shows the results obtained by performing the LR and the BN test. They confirm the good performance offered by GCLEs, with the exception of the GCN. Looking at the loss functions, we can conclude that the GCHS outperforms all the other distributions for 
. In particular, the GCL and the M
t are equivalent for 
; the GCN (ML) and the GCHS (MM) provide the best fitting for 
 (the former for the ABLF, the latter for the AQLF); and the GCN is the only distribution not underestimating the ABLF for 
, which is better approximated by the MPE according to the AQLF.
 Finally, the out-of-sample GCLE performance in estimating the ES was evaluated by performing the 
 and 
 tests of 
Acerbi and Szekely (
2014). The null hypothesis of both tests assumes that the estimated 
 tallies with the ES calculated from the data. This entails, in the present application, that 
 should provide a good estimate of the empirical expected shortfall computed in the second part of the sample. Under the alternative, this does not happen, and 
 systematically underestimates the effective loss mean, 
, thus implying financial damage.
The first test is based on the following statistic:
        where 
 is the number of losses 
, which in the second part of the sample, break the threshold 
, and 
 is an indicator variable that is equal to one if 
 and zero otherwise.
As observed by 
Acerbi and Szekely (
2014), this test is completely insensitive to an excessive number of exceptions as it is simply an average taken over the exceptions themselves. Under the null, the realized value 
 is expected to be zero, and it signals a problem when it displays values significantly greater than zero.
The second test statistic is:
        where 
T denotes the sample size of the second period, and the other symbols are defined as above. As for the previous test, under the null, the realized value 
 is expected to be zero, and it signals a problem when it assumes values significantly greater than zero.
To perform the test and obtain the associated p-values, the following parametric bootstrap approach was used:
A population of one million observations was generated from each distribution under investigation (Mt, MPE, or one of the GCLEs considered in the paper), with parameters estimated by using the data of the first sample period. For the class of the GCLEs, HMC was employed to this aim;
B = 5000 samples, of a size equal to the dimension of the out-of-sample dataset, were drawn from each of these populations;
the bootstrap distributions of the statistics 
 and 
 were obtained and used to calculate the 
p-values, 
, of the statistics 
 and 
 computed by using the observed sample data:
            
The results are reported in 
Table 7. They allow concluding that the GCLEs adequately estimate the ES, with the exception of the GCN. According to the 
 statistic, the M
t distribution underestimates the expected losses for 
, while the MPE underestimates the ES according to the 
 statistic for 
. As in the case of the in-sample assessment, distributions that are appreciably more leptokurtic than the GCN show a better tail sensitivity and are thus particularly adequate to evaluate risk measures at high levels of 
.