1. Introduction
In the aftermath of the financial crisis, it was realized that the mathematical models for financial markets carry a substantial amount of model risk. Ref. [
1], with a discussion from a banking point of view, and [
2], who provide an overview with a focus on systemic risk, investigate various aspects of model risk, while [
3,
4] give some early examples. Accounting for model risk is particularly important for the forecast of risk inherent to portfolios consisting of financial instruments as this forecast is used to calculate economic and regulatory capital within financial institutions. Hence, it is no surprise that if a financial institution opts (e.g., under Solvency II or Basel III) for an application of an internal model, its stakeholders, such as regulators and rating agencies, require that such internal models are
reliable. For that reason, the EU directive for Solvency II imposes explicitly that internal models have to meet
statistical quality,
calibration, and
validation standards and have to pass the
use test (see Articles 120, 121, 122 and 124, respectively, of the Solvency II directive [
5]). By means of these quantitative and qualitative requirements, it is intended to achieve a high level of comfort with respect to the reliability of internal models. Furthermore, regulators require explicitly the
robustness of a model as an overarching
modelling principle (see [
6]). This is especially important because both internal and external model users, who include the results of the models in their decision-making process, are often non-experts in stochastic models. In this respect, robustness of models is a necessary condition to introduce a sustainable
risk culture.
Motivated by the above, we pose the problem of a consistent robust modeling approach by considering an axiomatic setting for monetary risk measures that ensures robustness on the one hand and emphasizes the
approximative nature of internal models on the other hand by making the need for formulating a suitable
distance explicit. Thus, we focus on the statistical robustness of the risk measurement procedure and in doing so complement recent approaches to evaluate model risk as in [
7] or to use statistical concepts to evaluate the performance of risk measures as suggested in e.g., [
8] or [
9].
Within our approach, we consider risk functionals, which treat risk measures on the level of distributions, and allow so to exploit established statistical theory. Risk functionals are typically not continuous with respect to the weak topology as this topology ignores the tails of the distribution. Ref. [
10] pointed out this fact and suggested to use a stronger topology. We will thus use robustness in the sense of continuity with respect to an appropriate metric, namely the Wasserstein metric, which was introduced in a general setting in [
11] and in the financial context in [
12] (p. 79), as the uniform metric between inverse distribution functions. Our approach is in line with [
13,
14], who provide extensions and introduce an index of qualitative robustness, and the approaches suggested by [
15], who prefer robust methods for estimating parameters (Value-at-Risk, expectations, etc.) of forecast distributions. Our framework allows for utilisation of several results that have been established recently in [
16].
In addition, we provide various real-life examples to support and to emphasise the applicability of our results. We thus extend and clarify results in [
17], who provide real-life examples that highlight the need for robust models for both valuation and calculation of risk in order to control model risk. In addition, in an insurance context [
18] recently, applied robust methods to the estimation of loss triangles.
2. A Suitable Distance for Risk Management
We consider monetary risk measures that are related to capital requirements. These risk measures,
ρ, are typically introduced for (random) financial positions
X defined on an appropriate probability space
by
where
denotes the space of real valued bounded functions.
We refer the reader for definitions and the various variants of monetary risk measures to [
19,
20]. Important for us is the contribution by [
21], who introduced
natural risk statistics. This concept bridges the gap between risk measures and statistics by defining natural risk statistics on a
sample space in contrast to the definition of risk measures on a probability space (
1).
In situations when we are dealing with possibly unbounded financial positions
X, the space
is not suitable. Following the observation of [
22] that Orlicz spaces or Orlicz hearts (see Proposition 2 below) are appropriate in this case, [
16] show that for
is an Orlicz heart and thus can serve as a space of financial positions. In fact, based on [
23], Theorem 2.10 of [
16] shows that law-invariant convex risk measures on
has a unique extension to
(convex, monotone, and lower semicontinuous with respect to the
-norm), which inherits continuity properties.
In order to discuss the robustness (estimation) of risk measures, we need the following concept:
Definition 1. Statistical functional. Let denote the set of all cumulative distribution functions on and an appropriate subset. We consider maps Now, let be a sample from a population with distribution function and empirical distribution . If a statistic can be written as a functional T of , say , then we will call T a statistical functional.
This is particularly important as we assess the risk
of a financial position
X with distribution function
F. Typically, we will use a Monte Carlo simulation or a sequence of historical data. For law-invariant
ρ, a natural estimate of
is
, where
is the empirical distribution function and
Now, robustness of the risk measure ρ can be related to continuity of . Here, robustness refers to a certain insensitivity of ρ to variations in the approximating distribution for F. For this, the concept of continuity is crucial.
Note that any statistical functional can be approximated with , provided that is in a neighborhood of F and the functional T is sufficiently well behaved. This leads us to consider F as a “point” in and to investigate continuity, differentiability and other regularity properties for functionals defined on (subsets of) .
In order to define these concepts properly, we need to introduce:
Definition 2. Simple probability metric. Let be a convex class of distribution functions on containing all degenerate distributions. A mappingis called a simple probability metric (or distance) if- 1.
,
- 2.
,
- 3.
.
Unfortunately, there are plenty of possible metrics to choose. In order to select a specific metric as a preferable one from a practitioner’s point of view, two questions have to be answered:
Does the chosen metric fit well in the context of application, i.e., has the chosen metric a natural interpretation? Is it perhaps the canonical metric for a particular field of application?
Is the metric not too strong? This means—do we have a sufficiently rich set of continuous objects?
The following example lists very common risk measures that should fit into the risk measurement framework from a practitioner’s point of view.
Example 1. Examples of statistical functionals and risk measures: The central piece of many (internal) risk management models is risk measures such a Value-at-Risk (VaR) and Tail-VaR (TVaR). We give representations of these important risk measures as statistical functionals. Let be the empirical distribution function related to a given data set. In addition, recall that, for a cumulative distribution function F, the (upper) quantile of order is defined as .Value-at-Risk. For a fixed (usually ), we define the Value-at-Risk at level α (VaR) asThe statistical functional for estimating Value-at-Risk is thenwhere puts mass one to . Tail-VaR (TVaR).1 The statistical functional for estimating TVaR is
If we substitute
F for
in the examples above, we obtain the associated
risk measures. In particular, any law-invariant
2 risk measure can be written as a statistical functional (see [
19] for further discussion). For the VaR, this yields
. From (
4) and (
5), we conclude the important role of the empirical distribution for determining risk measures. Hence, a metric for risk management should be based on the whole cumulative distribution function.
We now turn to the question of robustness of risk measures in more detail. Given that we want to use the empirical distribution in praxis, we need to specify how to quantify the effect of a small deviation from the true distribution on the risk measure. We need the following definition:
Definition 3. Continuity of statistical functionals. Let be a convex class of distribution functions on containing all degenerate distributions. A functional T defined on is continuous at ifwhere denotes a distance of two distribution G and F. In the axiomatic approach of [
24], the continuity of a risk measure is not required. Consequently, the class of coherent risk measures includes discontinuous risk measures such as the TVaR. Given the approximative character of stochastic models, we think that continuity, and, hence, robustness, are, from a practitioner’s point of view, even more important than subadditivity.
Example 2. TVaR is not continuous w.r.t. the weak topology: Consider the sequencewhere denotes the uniform distribution on and puts its mass one to , . Obviously, converges weakly to : However, the associated TVaRs do not. For all α, we have:whereasis finite. The above example clearly highlights the fact that risk measures, which put emphasis on the tail of a distribution cannot be expected to be continuous with respect to topologies that ignore the tails of distributions. In contrast, it can be shown that the VaR at level
α is continuous w.r.t the weak topology (see Theorem 3.7 in [
25]) at
if
is continuous at
α. Given that VaR and TVaR are the risk measures that are of overwhelming importance in practice, we suggest using a metric that implies the continuity of the associated functionals.
We will now discuss properties of a metric that are particularly useful for the risk management process. In terms of practical applications, it is important to strike a balance between properties in the centre of a distribution and its tail behaviour. Thus, as outlined in Example 1, the following requirements are natural in order that the statistical functionals allow for estimating all of the risk measures given in Example 1 consistently:
for some
and
. Note that the linear functional
is now continous in
F w.r.t. a given metric. For applications in risk management, the cases
and
are of particular importance. Thus, we are looking for a metric that combines these two properties.
We will show that with the Wasserstein metric (especially, the case ), both types of risk measures are considered: one-tail measures (such as VaR) and two-tail measures (such as volatility). Hence, this approach comprises both concepts.
We will only need the Wasserstein metric for two distributions on the real line.
Definition 4. Wasserstein Metric. Let random variables with their corresponding distribution functions. The Wasserstein distance, for two distributions is given by: For a general definition of the Wasserstein metric in Banach spaces, the reader will find appropriate definitions in [
11]. Refs. [
26,
27] show the equivalence of the above definition with the general case. An example for the use of the Wasserstein metric in the context of investments is [
28].
Proposition 1. Properties of Wasserstein Metric. The Wasserstein metric has the following properties:- 1.
For random variables we have the following scaling properties of :The scaling property (8) is important for risk management because it allows the change of numéraire, e.g., change of home currency. - 2.
For in and If and , then we have the regularity property: - 3.
Let be independent and assume that the laws are in . Then,This property is key to proving the consistency for bootstrapping mean statistics (see [11]). - 4.
The Wasserstein metric is a one-ideal metric (see the Appendix A for a definition of one- ideal).
Example 3. Wasserstein and distance: For distribution functions, the -distance, is defines as: We compare this distance with the Wasserstein metric for a standard normal distribution Φ, which we perturbate such that there is a difference in the right tail. Thus, letwhere is an exponential distribution. Table 1 shows that the Wasserstein distance increases in p while the distance decreases. Thus, the weighting of differences in the tail increase for the Wasserstein metric in p, while it decreases for . The convexity properties are closely related to the independence axiom of the Von–Neumann–Morgenstern approach to utility theory. Therefore, the Wasserstein metric fits smoothly into this framework. As the results of the risk measuring exercise need to have an impact on decisions within the company (the use test in Solvency II) and as utility theory is widely used in this context, this is a very desirable property.
Recently, [
16] have investigated the robustness properties of convex risk measures, in particular regarding the Wasserstein metric.
Proposition 2. [
16] (Theorems 2.5 & 2.7).
Let be a stationary and ergodic sequence of random variables with the same law as 3 withThen, for a law-invariant convex risk measure ρ with corresponding statistical functional , the estimatoris consistent, that is, as a.s., and continuous with respect to the p-Wasserstein metric4.
The proposition implies, for instance, the continuity of the TVaR w.r.t. the Wasserstein metric.
Many statistical functionals are so-called
L - functionals (see [
21] for a discussion of the role of L-statistics in terms of natural risk measures). We can relate them to the Wasserstein-metric with the following:
Theorem 1. Continuity of L-statistical functional w.r.t the Wasserstein metric. Let T be an L-statistical functional of the formfor , , and J a weights generating function bounded on [0,1]. Then, T is continuous w.r.t to Wasserstein metric . Proof. The claim follows as a special case from [
16], Proposition 2.22. We provide an independent proof in the
Appendix B. □
Theorem 1 implies the continuity of mean, trimmed mean, and TVaR w.r.t. the Wasserstein metric. □
As we discussed above, we extended the classical concept of qualitative robustness, which is based on the weak topology, by using a more suitable topology, and therefore we made it useful for tail sensitive functionals such as most risk measures. Ref. [
16] propose the alternative notion of
ψ-robustness. Here,
is a left-continuous, nondecreasing convex (Young) function, which we call weight function in the following. We denote by
the set of all probability measures
5 on
and by
the set of all probability measures
μ on
with
. Additionally, we have to impose a product structure on the underlying probability space, setting
,
for
,
, and, for
,
.
Definition 5. Qualitative robustness and ψ-robustness. Let , and let and be metrics on and , respectively. The functional is called (qualitatively) -robust with respect to and , if for all and , there exist and such that Let ψ be a weight function. A risk functional is called ψ-robust on if is robust with respect to and on all subsets , which are uniformly integrating, that is The key inside (already pointed out in [
10]) is that, when insisting on distances that generate the usual weak topology of measures, one can have distributions that are rather close in the weak topology but will have completely different tail behaviour. Thus, the notion of robustness should allow an insensitivity with respect to tail behaviour. As [
16] point out, the Prohorov distance
and the Prohorov
metric
are appropriate choices. Ref. [
16] (Theorem. 2.14 rso Theorem 2.25) characterize robust risk measures rsp. show that it is sufficient to study law-invariant convex risk measures on
.
A special case for , the function that is also implicitly used in Proposition 2, allows for stating the result for the p- Wasserstein metric.
Proposition 3. Every law-invariant convex risk measure , is -robust with respect to the p-Wasserstein metric.
In particular, the proposition shows that the TVaR is
-robust for any
. The value at risk is qualitatively robust even in the classical sense, that is, with
and
equal to the Prohorov metric and
(see ([
14], Example 4.3 )).
Example 4. Wasserstein metric for p=1: for heavy tailed distributions as they appear in finance, the mean absolute deviation is preferable, due to its robustness properties. Hence,where denotes the median, is a suitable substitute for , and thus for volatility-sensitive risk measures, in (6) which is much more in line with our overall robustness approach. In requiring We are in the setting of the Wasserstein framework. Note that for the case, we have: This may be used with advantage, when we consider Fréchet differentiable functions. Following [29], we consider the following norms: For , this means that if a statistical functional is strongly differentiable w.r.t supremum norm, then it is automatically differentiable with respect to the sum of the Wasserstein and supremum norm.
This norm implies, for statistical functionals which satisfy a Fréchet differentiability condition, that jackknifing and bootstraps yield consistent estimators.