Next Article in Journal
Valuation of Large Variable Annuity Portfolios Using Linear Models with Interactions
Previous Article in Journal
Association Rules for Understanding Policyholder Lapses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Log-Normal or Over-Dispersed Poisson?

Department of Economics, University of Oxford & Oriel College, Oxford OX1 4EW, UK
Risks 2018, 6(3), 70; https://doi.org/10.3390/risks6030070
Submission received: 18 June 2018 / Revised: 5 July 2018 / Accepted: 6 July 2018 / Published: 9 July 2018

Abstract

:
Although both over-dispersed Poisson and log-normal chain-ladder models are popular in claim reserving, it is not obvious when to choose which model. Yet, the two models are obviously different. While the over-dispersed Poisson model imposes the variance to mean ratio to be common across the array, the log-normal model assumes the same for the standard deviation to mean ratio. Leveraging this insight, we propose a test that has the power to distinguish between the two models. The theory is asymptotic, but it does not build on a large size of the array and, instead, makes use of information accumulating within the cells. The test has a non-standard asymptotic distribution; however, saddle point approximations are available. We show in a simulation study that these approximations are accurate and that the test performs well in finite samples and has high power.

1. Introduction

Which is the better chain-ladder model for claim reserving: over-dispersed Poisson or log-normal? While the expert may have a go-to model, the answer should be informed by the data. Choosing the wrong model could substantially influence the quality of the reserve forecast. Yet, so far, no statistical theory is available that supports the actuary in his/her decision and that allows him/her to make a solid argument in favour of either model.
We develop a test that can distinguish between over-dispersed Poisson and log-normal data generating processes, both of which have a long history in claim reserving. The test exploits that the former model fixes the variance to mean ratio across the array, while the latter assumes a common standard deviation to mean ratio. Consequently, the test statistic is based on estimators for the variation in the respective models. The idea is drawn from the econometric literature on encompassing. Intuitively, the test asks whether the null-model can accurately predict the behaviour of the rival model’s variation estimator when the null-model is true.
The over-dispersed Poisson model is appealing since it naturally pairs with Poisson quasi-likelihood estimation, replicating the popular chain-ladder technique in run-off triangles (Kremer 1985, pp. 130). Furthermore, this model makes for an appealing story due to its relation to compound Poisson distributions. Such distributions give the aggregate incremental claims an interpretation as the sum over a Poisson number of claims with random individual claim amounts (Beard et al. 1984, Section 3.2). A popular method to generate distribution forecasts for the over-dispersed Poisson model is bootstrapping (England and Verrall 1999; England 2002). While in widespread use, there is so far no theory proving the validity of the bootstrap in this setting. Furthermore, in some settings, the method seems to produce unsatisfactory results.
Recently, Harnau and Nielsen (2017) developed a theory that gives the over-dispersed Poisson model a rigorous statistical footing. They propose an asymptotic framework based on infinitely-divisible distributions that keeps the dimension of the data array fixed and instead builds on large cell means. This resolves the incidental parameter problem (Lancaster 2000; Neyman and Scott 1948) that renders a standard asymptotic theory based on a large array invalid and arises since the number of parameters grows with the size of the array. The class of infinitely-divisible distributions includes compound Poisson distributions, which are appealing in an insurance context, as noted above. We can then interpret large cell means as the result of a large latent underlying number of claims. Other infinitely-divisible distributions that can be reconciled with the over-dispersed Poisson structure include Poisson, gamma and negative binomial.
The intuition for the theory by Harnau and Nielsen (2017) is that the array is roughly normally distributed for large cell means, so that the results remind us of a classical analysis of variance (ANOVA) setting. Harnau and Nielsen (2017) show that Poisson quasi-likelihood estimators are t-distributed, and F-tests based on Poisson likelihoods can be used to test for model reduction, such as for the absence of a calendar effect. Finally, chain-ladder forecast errors are t-distributed, giving rise to closed-form distribution forecasts including for aggregates, such as the reserve or cash-flow. In their simulations, Harnau and Nielsen (2017) find that while the bootstrap (England and Verrall 1999; England 2002) matches the true forecast error distribution better on average, the t-forecast produces fewer outliers and appears more robust.
Building on the asymptotic framework put forward by Harnau and Nielsen (2017), Harnau (2018a) proposes misspecification tests for two crucial assumptions of the over-dispersed Poisson model. First, the variance to mean ratio is assumed to be common across the array. Second, accident effects are not allowed to vary over development years and vice versa. To check for a violation of these assumptions, Harnau (2018a) suggests splitting the run-off triangle into sub-samples and then testing whether a reduction from individual models for each sub-sample to a single model for the full array can be justified. While the idea of splitting the sample is borrowed from time-series econometrics (Chow 1960), the theory for a reduction to a single model is again reminiscent of an ANOVA setting. A classical Bartlett test (Bartlett 1937) can be used to assess whether we can justify common variance to mean ratios. This is followed by an independent F-test for the absence of breaks in accident and development effects. Again, the asymptotics needed to arrive at these results keep the dimension of the array fixed, growing instead the cell means. Harnau (2018a) also shows that these misspecification tests can be used in a similar fashion in a finite sample log-normal model.
The log-normal model introduced by Kremer (1982), who relates it to the ANOVA literature, features a predictor structure that is reminiscent of the classical chain-ladder. Verrall (1994) refers to this as the chain-ladder linear model, while Kuang et al. (2015) use the term geometric chain-ladder. The latter authors show that the maximum likelihood estimators in the log-normal model can be interpreted as development factors of geometric averages, compared to an interpretation of arithmetic averages arising for the classical chain-ladder. An advantage of the log-normal model is that an exact Gaussian distribution theory applies to the maximum likelihood estimators. However, since these estimators are computed on the log scale, a bias is introduced on the original scale. Verrall (1991) tackles this issue and derives unbiased estimators for the mean and standard deviation on the original scale. One issue for full distribution forecasts in the log-normal model is that the insurer is usually not interested in forecasts for individual cells, but rather for cell sums such as the reserve or the cash-flow. However, the log-normal distribution is not closed under convolution, so that cell sums are not log-normally distributed.
Recently, Kuang and Nielsen (2018) proposed a theory that includes closed-form distribution forecasts for cell sums, such as the reserve, in the log-normal model, thus remedying one of its drawbacks. Kuang and Nielsen (2018) combined the insight by Thorin (1977) that the log-normal distribution is infinitely divisible and the asymptotic framework by Harnau and Nielsen (2017). Based on this, they propose a theory for generalized log-normal models, a class that nests the log-normal model, but is not limited to it. In particular, the distribution is not assumed to be exactly log-normal, but merely needs to be infinitely divisible with a moment structure close to that of the log-normal model. The asymptotics in this framework again leave the dimension of the array untouched to avoid an incidental parameter problem. In contrast to the theory for large cell means in the over-dispersed Poisson model, results are now for small standard deviation to mean ratios.
For the generalized log-normal model, Kuang and Nielsen (2018) show that least squares estimators computed on the log scale are asymptotically t-distributed, and simple F-tests based on the residual sum of squares can be used to test for model reduction. Reassuringly, these results match the exact results in a log-normal model. Beyond that, they also prove that forecast errors on the original scale are asymptotically t-distributed so that distribution forecasting for cell sums is straightforward. Further, they show that the misspecification tests by Harnau (2018a) are asymptotically valid for the generalized log-normal model, just as they were in finite samples for the log-normal model.
We remark that besides over-dispersed Poisson and log-normal models, there exist a number of reserving models that we do not consider further in this paper. England and Verrall (2002) give an excellent overview. Perhaps the most popular contender is the “distribution-free” model by Mack (1993). This model also replicates the classical chain-ladder, but differs from the over-dispersed Poisson model. Mack (1993) derives the expression for forecast standard errors. However, so far, no full distribution theory exists for this model.
With a range of theoretical results in place for over-dispersed Poisson and (generalized) log-normal models, discussed further in Section 3, a natural question is when we should employ which model. The misspecification tests by Harnau (2018a) seem like a natural starting point. For example, if we can reject the specification of the log-normal, but not the over-dispersed Poisson model, the latter seems preferable. However, the misspecification tests may not always have enough power to make this distinction, as we show in Section 2.
Since generalized log-normal and over-dispersed Poisson models are not nested, a direct test between them is not trivial. Cox (1961, 1962) introduced a theory for non-nested hypothesis testing with a null model. Vuong (1989) provided a theory for non-nested model selection without a null model; in selection, the goal is to choose the better, not necessarily the true, model. However, both procedures are likelihood based, so that the results are not applicable here, since we did not specify exact distributions and, thus, do not have likelihoods available.
Given the lack of likelihoods for the models, we look to the econometric encompassing literature for inspiration. The theory for encompassing allows for a more general way of non-nested testing. As Mizon and Richard (1986) put it, “Among other criteria, it seems natural to ask whether a specific model, say M 1 , can mimic the DGP [data generating process], in that statistics which are relevant within the context of another model, M 2 say, behave as they should were M 1 the DGP.” The encompassing literature originates from Hendry and Richard (1982) and Mizon and Richard (1986); for a less technical introduction, see Hendry and Nielsen (2007, Section 11.5). Ermini and Hendry (2008) applied the encompassing principle in a time-series application. They tested whether disposable income is better modelled on the original scale or in logs. Taking the log model as the null hypothesis, they evaluated whether the log model can predict the behaviour of estimators for the mean and variance of the model on the original scale.
Building on the encompassing literature, we find the distribution of the over-dispersed Poisson model estimators under a generalized log-normal data generating process and vice versa. It turns out that both Poisson quasi-likelihood and log data least squares estimators for accident and development effects are asymptotically normal, regardless of the data generating process. Differences arise in the second moments. This manifests in the limiting distributions of the variation estimators. While these are asymptotically χ 2 under the correct model, their distribution is a non-standard quadratic form of normals under the rival model. However, these distributions involve the unknown dispersion parameter, which needs to be estimated. Employing the variation estimator of the correct model for this purpose, we arrive at a test statistic with a non-standard asymptotic distribution: the ratio of dependent quadratic forms. Saddle point approximations to such distributions are available (Butler and Paolella 2008; Lieberman 1994). Further, we can show that the power of the tests originates from variation in the means across cells. This is intuitive given that the main difference between the models disappears when all means are identical; then, both standard deviation to the mean and variance to mean ratios are constant across the array. These findings are collected in Section 4.
With the theoretical results for encompassing tests between over-dispersed Poisson and generalized log-normal models in place, we show that they perform well in a simulation study. First, we demonstrate that saddle point approximations to the limiting distributions of the statistics work very well. Second, we tackle an issue that disappears in the limit: we have the choice between a number of asymptotically-identical estimators that generally differ in finite samples. Simulations reveal substantial heterogeneity in finite sample performance, but also show that some choices generally do well. Third, we show that the tests have high power for parameterizations we may realistically encounter in practice. We also find that power grows quickly with the variation in the means. The simulation study is in Section 5.
Having convinced ourselves that the tests do well in simulations, we demonstrate their application in a range of empirical applications in Section 6. First, we revisit the empirical illustration of the problem from the beginning of the paper. We show that the test has no problem rejecting one of the two rival models. Second, we consider an example that perhaps somewhat cautions against starting with a model that may be misspecified to begin with. In this application, dropping a clearly needed calendar effect turns the results of the encompassing tests upside down. Third, taking these insights into account, we implement a testing procedure that makes use of a whole range of recent results: deciding between the over-dispersed Poisson and generalized log-normal model, evaluating misspecification and testing for the need for a calendar effect.
We conclude the paper with a discussion of potential avenues for future research in Section 7. These include further misspecification tests, a theory for the bootstrap and empirical studies assessing the usefulness of the recent theoretical developments in applications.

2. Empirical Illustration of the Problem

We illustrate in an empirical example that the choice between the over-dispersed Poisson and (generalized) log-normal model is not always obvious. Table 1 shows a run-off triangle taken from Verrall et al. (2010, Table 1) with accident years i in the rows and development years j in the columns. Calendar years k = i + j 1 are on the diagonals.
While Kuang et al. (2015) and Harnau (2018a) model the data in Table 1 as log-normal, it is not obvious whether a log-normal or an over-dispersed Poisson model is more appropriate. In a log-normal model, the aggregate incremental claims Y i j are independent:
M L N : log ( Y i j ) = N ( α i + β j + δ , ω 2 )
where α and β are accident and development effects, respectively. On the original scale, this implies that:
M L N E ( Y i j ) = exp α i + β j + δ + ω 2 2 , sd ( Y i j ) E ( Y i j ) = exp ( ω 2 ) 1 .
Thus, the standard deviation to mean ratio, as well as the log data variance are common across cells. If we instead chose an over-dispersed Poisson model, we would maintain the independence assumption and specify the first two moments of the claims Y i j as:
M O D P : E ( Y i j ) = exp ( α i + β j + δ ) , var ( Y i j ) E ( Y i j ) = σ 2 .
Thus, the variance to mean ratio σ 2 is identical for all cells.
To choose between the two models, we could take the misspecification tests by Harnau (2018a) as a starting point. To implement the tests, the data are first split into sub-samples. For the Verrall et al. (2010) data, Harnau (2018a) considers a split into two sub-samples consisting of cells relating to the first and last five accident years, as illustrated in Table 1. The idea is then to test for common parameters across the sub-samples. In the log-normal model, we first perform a Bartlett test for common log data variances ω 2 across sub-samples and, if this is not rejected, an F-test for common accident and development effects. Similarly, in the over-dispersed Poisson model, we first test for common over-dispersion σ 2 and then again for common accident and development effects.
If one of the models is flagged as misspecified, but not the other, the choice becomes obvious. However, in this application, we cannot reject either model based on these tests. For the log-normal model, the Bartlett test for common log data variances yields a p-value of 0.09 and the F-test for common effects a p-value of 0.91 ; the p-values for the equivalent tests in the over-dispersed Poisson model are 0.78 and 0.64 , respectively. Therefore, the question remains: Which model should we choose?

3. Overview of the Rival Models

We first discuss two common elements of the rival models, namely the data structure and the chain-ladder predictor and its identification. Then, we in turn state assumptions, estimation and known theoretical results for the over-dispersed Poisson and the generalized log-normal chain-ladder model.

3.1. Data

We assume that we have data for a run-off triangle of aggregate incremental claims. We denote the claims for accident year i and development year j by Y i j . Further, we count calendar years with an offset, so the calendar year k = i + j 1 . Then, we can define the index set for a run-off triangle with I accident, development and calendar years by:
I = { ( i , j ) : 1 i , j , k I } .
We define the number of observations in I as n. We could also allow for data in a generalized trapezoid as defined by Kuang et al. (2008) without changing the results of the paper. Loosely, generalized trapezoids allow for an unbalanced number of accident and development years, as well as missing calendar years both in the past and the future.

3.2. Identification

We briefly discuss the identification problem of the chain-ladder predictor α i + β j + δ that is common to both over-dispersed and generalized log-normal models. Kremer (1985) showed that based on this predictor, Poisson quasi-likelihood estimation replicates the classical chain-ladder point forecasts in a run-off triangle.
The identification problem is that for any a and b,
μ i j = α i + β j + δ = ( α i + a ) + ( β j + b ) + ( δ a b )
where α i and β j are accident and development effects, respectively. Thus, no individual effect is identified. Several ad hoc identification methods are available; for example, we could set i α i = j β j = 0 . Kuang et al. (2008) suggest a parametrization that is canonical in a Poisson model and allows for easy counting of degrees of freedom. The idea is to re-write the linear predictor in terms of a level and deviations from said level as:
μ i j = μ 11 + s = 2 I 1 ( i s ) Δ α s + t = s I 1 ( j s ) Δ β s .
Thus, we can write:
μ i j = x i j ξ
where the design x i j and identified parameter ξ are given by:
x i j = ( 1 , 1 ( i 2 ) , , 1 ( i I ) , 1 ( j 2 ) , , 1 ( j I ) ) ξ = ( μ 11 , Δ α 2 , , Δ α I , Δ β 2 , , Δ β I ) R p .
For the asymptotic theory in the over-dispersed Poisson model, it turns out to be useful to explicitly decouple the level and its deviations by decomposing as:
ξ = ( μ 11 , ξ ( 2 ) ) and x i j = ( 1 , x i j ( 2 ) ) .
We can then define the aggregate predictor τ and the frequencies π i j as:
τ = i j I exp ( μ i j ) and π i j = exp ( μ i j ) τ = exp ( x i j ( 2 ) ξ ( 2 ) ) i j I exp ( x i j ( 2 ) ξ ( 2 ) ) .
Importantly, the frequencies π i j are invariant to the level μ 11 , the first component of ξ . Therefore, we can vary the aggregate predictor τ by varying μ 11 without affecting the frequencies π i j . The frequencies π i j are, in turn, functions of ξ ( 2 ) alone. Further, we note that given ξ ( 2 ) , there is a one-to-one mapping between μ 11 and τ through τ = exp ( μ 11 ) i j I exp ( x i j ( 2 ) ξ ( 2 ) ) .
While this choice of identification scheme is useful for derivation of the theory in this paper, any scheme may be used in applications of the results. This is because, as Kuang et al. (2008) point out, the linear predictor μ i j is identified, unlike the individual effects. Since the main results of the paper rely on estimates of the linear predictors μ i j alone, they are unaffected by the choice of a particular identification scheme.
Furthermore, the results in this paper are not limited to the chain-ladder predictor; we could, for example, include a calendar effect. Nielsen (2015) derives the form of the design vector for extended chain-ladder predictors in generalized trapezoids. The identification method is implemented in the R (R Core Team 2017) package apc (Nielsen 2015), as well as in the Python package of the same name (Harnau 2017).
We note that the identification method can introduce arbitrariness into the forecast for models that require parameter extrapolation, such as the extended chain-ladder model with calendar effects. In the standard chain-ladder model, we can forecast claim reserves without parameter extrapolation; in a continuous setting, Lee et al. (2015) refer to this as in-sample forecasting. In contrast, in the extended chain-ladder model, we cannot estimate parameters for future calendar years from the run-off triangle. For this case, Kuang et al. (2008) and Nielsen and Nielsen (2014) explain how forecasts can be influenced by ad hoc constraints and lay out conditions for the identification method that make forecasts invariant to these arbitrary and untestable constraints.

3.3. Over-Dispersed Poisson Model

We give the assumptions of the over-dispersed Poisson model and discuss its estimation by Poisson quasi-likelihood. We state the sampling scheme proposed by Harnau and Nielsen (2017) and the asymptotic distribution of the estimators.

3.3.1. Assumptions

The first assumption imposes the over-dispersed Poisson structure on the moments. We can write it as:
M O D P : E ( Y i j ) = exp ( μ i j ) , var ( Y i j ) E ( Y i j ) = σ 2 .
The second assumption is distributional and allows for the asymptotic theory later on. We assume that the independent aggregate claims Y i j have a non-degenerate, non-negative and infinitely-divisible distribution with at least three moments. As noted by Harnau and Nielsen (2017), an appealing example for claim reserving of such a distribution is compound Poisson. The interpretation is that the aggregate incremental claims Y i j can be written as Y i j = = 1 N i j X for a Poisson number of claims N i j = D Poisson { exp ( μ i j ) } independent of the independent and identically distributed random claim amounts X .

3.3.2. Estimation

We estimate the over-dispersed Poisson model by Poisson quasi-likelihood. The appeal is that, as noted in Section 3.2, Poisson quasi-likelihood estimation replicates the chain-ladder technique. We explicitly distinguish between the model, subscripted with O D P , and its standard estimators, sub- or super-scripted with q l , to avoid confusion later on when we evaluate the estimators under the rival model.
The fitted values for the linear predictors are given by:
μ ^ i j q l = x i j ξ ^ q l where ξ ^ q l = arg max ξ R p i j I { Y i j ( x i j ξ ) exp ( x i j ξ ) } .
The fitted value for the aggregate predictor τ is then given by:
τ ^ q l = i j I exp ( μ ^ i j q l ) = i j I Y i j ,
a result implied by the fact that the re-parametrization of the Poisson likelihood in terms of the mixed parameter ( τ , ξ ( 2 ) ) is linearly separable, so the parameters are variation independent; see, for example, Harnau and Nielsen (2017); Martínez Miranda et al. (2015) or, for a more formal treatment, Barndorff-Nielsen (1978, Theorem 8.4). This implies that the estimator for the aggregate predictor is unbiased for the aggregate mean.
As an estimator for the over-dispersion σ 2 , Harnau and Nielsen (2017) use the Poisson deviance D scaled by the degrees of freedom. The deviance is the log likelihood ratio statistic against a model with as many parameters as observations, giving a perfect fit. The estimator is given by:
σ ^ 2 = D n p where D = 2 i j I Y i j { log ( Y i j ) μ ^ i j q l } .

3.3.3. Sampling Scheme

For the asymptotic theory, we adopt the sampling scheme proposed by Harnau and Nielsen (2017). The idea is to grow the overall mean τ = i j I E ( Y i j ) while holding the frequencies π i j and thus ξ ( 2 ) fixed. We note that this also implies that μ 11 is O { log ( τ ) } . In this sampling scheme, information accumulates in the estimated frequencies. In this sense, it is reminiscent of multinomial sampling as used, for example, by Martínez Miranda et al. (2015) in a Poisson model conditional on the data sum. Furthermore, we assume that τ increases in such a way that the skewness vanishes. Harnau and Nielsen (2017) remark that this is implicit for distributions such as Poisson, negative binomial and many compound Poisson distributions. Importantly, the sampling scheme holds the number of cells in the run-off triangle fixed. If we instead grew the dimension of the array, the number of parameters would also increase, thus making an asymptotic theory difficult.

3.3.4. Asymptotic Theory

Based on the assumptions in Section 3.3.1 and the sampling scheme Section 3.3.3, Harnau and Nielsen (2017) derived the asymptotic distribution of the estimators.
The theory hinges on Harnau and Nielsen (2017, Theorems 1, 2), which for our purposes can be formulated as:
τ 1 / 2 { Y i j exp ( μ i j ) } = τ 1 / 2 ( Y i j / τ π i j ) D N ( 0 , σ 2 π i j ) and Y i j τ P π i j .
An implication of the sampling scheme is that we cannot consistently estimate μ 11 since the overall mean τ and thus the level μ 11 grow. However, the remaining parameters ξ ( 2 ) are fixed and can be estimated in a consistent way. To ease notation, we define the design matrix X and the diagonal matrix of frequencies Π so:
X = { x i j : ( i , j ) I } and Π = diag { π i j : ( i , j ) I } .
Harnau and Nielsen (2017, Lemma 1) derive the distribution of the estimator for the mean parameters in terms of the mixed parametrization ( τ , ξ ( 2 ) ) . The advantage is that the two components of the mixed parameter are variation independent, so the covariance matrix featured in the asymptotic distribution is block-diagonal. This property turns out to be useful for example in the derivation of distribution forecasts. However, we opt to state the results in terms of the original parameterization by ξ to ease the analogy with the generalized log-normal model below. For our purposes, this does not complicate the theory.
As a corollary to Harnau and Nielsen (2017, Lemma 1), we can then state the distribution of the quasi-likelihood estimator ξ ^ q l as follows. All proofs are in Appendix A.
Corollary 1.
In the over-dispersed Poisson model Section 3.3.1 and Section 3.3.3,
τ ( ξ ^ q l ξ ) = τ ( μ ^ 11 μ 11 ) ξ ^ q l ( 2 ) ξ ( 2 ) D N { 0 , σ 2 ( X Π X ) 1 } .
Thus, even though the level μ 11 , the difference between estimator and level ( μ ^ 11 μ 11 ) vanishes in probability. We note that τ X Π X corresponds to the Fisher information about ξ in a Poisson model.
Further, Harnau and Nielsen (2017, Lemma 1) find that the asymptotic distribution of the deviance is proportional to a χ 2 :
D D σ 2 χ n p 2 .
Thus, the estimator σ ^ 2 has an asymptotic distribution, which is unbiased for σ 2 .

3.4. Generalized Log-Normal Model

Following the same structure as for the over-dispersed Poisson model above, we set up the generalized log-normal model as introduced by Kuang and Nielsen (2018) and discuss its estimation and theoretical results. This model nests the log-normal model. While the log-normal model allows for an exact distribution theory for the estimators, Kuang and Nielsen (2018) provide an asymptotic theory that covers the generalized model. We are going to employ this asymptotic theory for the encompassing tests below.

3.4.1. Assumptions

The assumptions for the generalized log-normal model mirror those for the over-dispersed Poisson model closely. The assumption of independent Y i j with a non-negative, non-degenerate infinitely-divisible distribution and at least three moments is maintained. The difference lies in the moment assumptions, which are replaced with:
M G L N : E ( Y i j ) = exp μ i j + ω 2 2 and sd ( Y i j ) E ( Y i j ) = ω 2 { 1 + o ( 1 ) }
where o ( 1 ) vanishes as ω 2 goes to zero. Thus, in the generalized log-normal model, the standard deviation to mean ratio, also known as the coefficient of variation, is common across the data for small ω 2 . This is in contrast to the variance to mean ratio in the over-dispersed Poisson model. Kuang and Nielsen (2018, Theorem 3.2) point out that the log-normal model log ( Y i j ) = D N ( μ i j , ω 2 ) satisfies these assumptions. There, the standard deviation to mean ratio is exp ( ω 2 ) 1 as in (1).
Based on the infinite divisibility assumption, we can construct a story similar to the compound Poisson story for the over-dispersed Poisson model. By definition, Y is infinitely divisible if for any m > 0 , there exist independent and identically distributed random variables X 1 , , X m , so = 1 m X has the same distribution as Y. Thus, as pointed out by Kuang and Nielsen (2018), we can again think of m as the unknown number of claims and of X as the individual claim amounts.

3.4.2. Estimation

We estimate the generalized log-normal model on the log scale by least squares. We define:
Z = { log ( Y i j ) : ( i , j ) I } .
Then, least squares fitted values for the linear predictors μ i j are, with the design X as defined in (5), given by:
μ ^ i j l s = x i j ξ ^ l s where ξ ^ l s = ( X X ) 1 X Z .
We estimate the variation parameter ω 2 based on the residual sum of squares written as:
ω ^ l s 2 = R S S n p where R S S = i j I ( Z i j μ ^ i j l s ) 2 = Z M Z for M = I X ( X X ) 1 X .
The estimator for the aggregate predictor τ as defined in (2) is then:
τ ^ l s = i j I exp ( μ ^ i j l s ) .
Unlike in the over-dispersed Poisson model, this estimator is generally not unbiased. Instead, the sum of linear predictors is unbiased for the sum of logs since i j I μ ^ i j l s = i j I Z i j .

3.4.3. Sampling Scheme

We adopt the sampling scheme Kuang and Nielsen (2018) put forward for the generalized log-normal model. In this scheme, ω 2 vanishes in such a way that the skewness of Y i j goes to zero while ξ remains fixed. In a log-normal model, this corresponds to letting the log data variance ω 2 , thus the standard deviation to mean ratio exp ( ω 2 ) 1 , go to zero. Again, the dimension of the array I remains fixed.

3.4.4. Asymptotic Theory

The asymptotic theory Kuang and Nielsen (2018) introduced for the generalized log-normal model allows one to find parameter uncertainty, testing for nested model reduction and closed-form distribution forecasts.
Kuang and Nielsen (2018, Theorem 3.4) find that for small ω 2 ,
( ω 2 ) 1 / 2 { Y i j exp ( μ i j ) } D N { 0 , exp ( 2 μ i j ) } .
Thus, generalized log-normal random variables are asymptotically normal, but heteroskedastic on the original scale. Furthermore, Kuang and Nielsen (2018, Theorem 3.3) prove that:
( ω 2 ) 1 / 2 ( Z i j μ i j ) D N ( 0 , 1 ) .
Therefore, conversion to the log scale yields asymptotic normality, as well. The difference is that the variance is now homoskedastic. We recall that μ i j is fixed under the sampling scheme in the generalized log-normal model. Therefore, these results imply that Y i j P exp ( μ i j ) and Z i j P μ i j . This also means that the data sum i j I Y i j P τ .
The small ω 2 distribution of the estimators in the generalized log-normal model is given by Kuang and Nielsen (2018, Theorem 3.5) as:
( ω 2 ) 1 / 2 ( ξ ^ l s ξ ) D N { 0 , ( X X ) 1 } and R S S ω 2 D χ n p 2 .
In an exact log-normal model, the results in (8) hold for any ω 2 .
In contrast to the over-dispersed Poisson model, the full parameter vector ξ , including the level μ 11 , can now be consistently estimated since it is fixed under the sampling scheme. This comes at the cost that ω 2 and, thus, the standard deviation to mean ratio move towards zero.

4. Encompassing Tests

With the two rival models in place, we aim to test the over-dispersed Poisson against the generalized log-normal model and vice versa. Since the models are generally not nested, we cannot simply test for a reduction from one to the other. Instead, we investigate whether the null model can correctly predict the behaviour of the statistics of the rival model if the null model is true. We first consider identifiable differences between the two models. Then, we in turn look at scenarios where the null is the over-dispersed Poisson model versus where it is the generalized log-normal model.

4.1. Identifiable Differences

It is interesting to consider what key features let us differentiate between the generalized log-normal and the over-dispersed Poisson model. Looking first at the means, we find that differences between the two models are not identifiable. This is because for any ξ = ( μ 11 , ξ ( 2 ) ) and ω 2 in the generalized log-normal model, we can define ξ = ( μ 11 + ω 2 / 2 , ξ ( 2 ) ) for the over-dispersed Poisson model, so:
E G L N ( Y i j ; ξ , ω 2 ) = exp x i j ξ + σ 2 2 = exp ( x i j ξ ) = E O D P ( Y i j ; ξ ) .
Thus, we could not even tell the models apart based on the means if we knew their true values.
In contrast, differences in the second moments are identifiable. In the generalized log-normal model, the standard deviation to mean ratio is constant for small ω 2 , while the variance to mean ratio is constant in the over-dispersed Poisson model. Since:
var ( Y i j ) E ( Y i j ) = sd ( Y i j ) E ( Y i j ) 2 E ( Y i j ) ,
constancy in one ratio generally implies variation in the other, except when all means are identical. Thus, the standard deviation to mean ratio in an over-dispersed Poisson model varies by cell, and so does the variance to mean ratio in a generalized log-normal model. Thus, if nature presented us with the true ratios, we could tell the models apart. As noted, an exception arises when all cells have the same mean, a scenario that seems unlikely in claim reserving. If this were the case, the assumptions of the two models are identical: the over-dispersed Poisson model becomes a generalized log-normal model and vice versa. Thus, non-identifiable differences between the ratios imply that both models are congruent with the data generating process in this dimension. Loosely, the two models become more different as the variation in the means increases. We may thus conjecture that there is a relationship between the power of tests based on standard deviations and variance to mean ratios and the variation in the means.

4.2. Null Model: Over-Dispersed Poisson

We find the asymptotic distribution of the least squares estimators, motivated in the generalized log-normal model, when the data generating process is over-dispersed Poisson. We propose a test statistic based on these estimators and find its limiting distribution under an over-dispersed Poisson data generating process.
The estimators from the log-normal model are computed on the log scale. Thus, we first find the limiting distribution of over-dispersed Poisson Y i j on the log scale.
Lemma 1.
In the over-dispersed Poisson model Section 3.3.1 and Section 3.3.3, lim τ P ( Y i j = 0 ) = 0 . For positive Y i j , with Z i j = log ( Y i j ) ,
τ ( Z i j μ i j ) = τ { log ( Y i j / τ ) log ( π i j ) } D N ( 0 , σ 2 π i j 1 ) .
We stress again that μ i j is not fixed under the sampling scheme so that the result does not imply that Z i j converges to μ i j , rather it implies that their difference ( Z i j μ i j ) vanishes. We can relate this lemma to Harnau and Nielsen (2017, Theorem 2), which states that Y i j / exp ( μ i j ) P 1 . This implies that log { Y i j / exp ( μ i j ) } = log ( Y i j ) μ i j P 0 , matching what we find here.
Given the limiting distribution on the log scale, we can find the distribution of the estimators in the same way as we would in a Gaussian model. Since the asymptotic distribution of τ ( Z i j μ i j ) is now heteroskedastic, unlike in the generalized log-normal model as shown in (7), we can anticipate that the results will not match those found in the generalized log-normal model. This is confirmed by the following lemma, using the notation for the design matrix X and the diagonal matrix of frequencies Π introduced in (5).
Lemma 2.
Define Ω = ( X X ) 1 X Π 1 X ( X X ) 1 , and let U = N ( 0 , I ) . Then, in the over-dispersed Poisson model Section 3.3.1 and Section 3.3.3,
τ ( ξ ^ l s ξ ) = τ ( μ ^ 11 μ 11 ) ξ ^ l s ( 2 ) ξ ( 2 ) D N ( 0 , σ 2 Ω ) , τ ^ l s τ P 1 a n d τ R S S D σ 2 U Π 1 / 2 M Π 1 / 2 U .
As could be expected given Lemma 1, the results in Lemma 2 match finite sample results in a heteroskedastic independent Gaussian model. Notably, the residual sum of squares R S S is not asymptotically χ 2 . However, the over-dispersion σ 2 enters their distribution only multiplicatively. The frequency matrix Π enters as a nuisance parameter that we can, however, consistently estimate since it is a function of ξ ( 2 ) alone. For example, we could use plug-in estimators Π ^ q l = Π ( ξ ^ q l ( 2 ) ) or Π ^ l s = Π ( ξ ^ l s ( 2 ) ) . If we knew σ 2 , we could feasibly approximate the limiting distribution of R S S . Besides Monte Carlo simulation, numerical methods are available; see, for example, Johnson et al. (1995, Section 18.8). These methods exploit that the distribution of the quadratic form can be written as a weighted sum of χ 1 2 . Generally, for a real symmetric matrix A and independent χ 1 2 variables V i j ,
U A U = D i j I λ i j V i j
where λ i j are the eigenvalues of A; this follows directly by the eigendecomposition of A.
Unfortunately, the over-dispersion σ 2 is generally unknown, so that we cannot simply base an encompassing test on the residual sum of squares R S S . Therefore, we require an estimator for σ 2 . An obvious choice in the over-dispersed Poisson model is the estimator σ ^ 2 = D / ( n p ) . However, computed on the same data, D and R S S are not independent. We could tackle this issue in two ways. First, similar to Harnau (2018a), we could split the data I into disjoint and thus independent sub-samples. Then, we could compute R S S on one sub-sample and D on the other, making the two statistics independent. However, in doing so, we would incorporate less information into each estimate and likely lose power. Beyond that, it seems little would be gained by this approach since no closed-form for the distribution of R S S is available in the first place. The second way to tackle the issue is to find the asymptotic distribution of the ratio R S S / D with each component computed over the full sample. This is the way we are going to go.
Before we proceed, we derive an alternative estimator for the over-dispersion σ 2 that gives us more choice later on for the encompassing test. Lemma 1 is suggestive of a weighted least squares approach on the log scale since the form of the heteroskedasticity is known, taking Π as given. For:
X * = Π 1 / 2 X , Z * = Π 1 / 2 Z and M * = I X * ( X * X * ) 1 X * ,
the weighted least squares estimators on the log scale are given by:
ξ ^ * = ( X * X * ) 1 X * Z * and R S S * = Z * M * Z * .
Of course, Π is unknown, so these estimators are infeasible. However, we can consistently estimate Π . Thus, we can compute feasible weighted least squares estimators. For a first stage estimation of the weights by least squares, we write:
Π ^ l s = Π ( ξ ^ l s ( 2 ) ) , X l s * = Π ^ l s 1 / 2 X , Z l s * = Π ^ l s 1 / 2 Z and M l s * = I X l s * ( X l s * X l s * ) 1 X l s * ,
so the (least squares) feasible weighted least squares estimators are:
ξ ^ l s * = ( X l s * X l s * ) 1 X l s * Z l s * and R S S l s * = Z l s * M l s * Z l s * .
Similarly, using instead the quasi-likelihood-based plug-in estimator Π ^ q l = Π ( ξ ^ q l ( 2 ) ) for the weights, we write:
Π ^ q l = Π ( ξ ^ q l ( 2 ) ) , X q l * = Π ^ q l 1 / 2 X , Z q l * = Π ^ q l 1 / 2 Z and M q l * = I X q l * ( X q l * X q l * ) 1 X q l * ,
so the (quasi-likelihood) feasible weighted least squares estimators are:
ξ ^ q l * = ( X q l * X q l * ) 1 X q l * Z q l * and R S S q l * = Z q l * M q l * Z q l * .
While we would generally expect them to differ in finite samples, it turns out that the Poisson quasi-likelihood and the (feasible) weighted least squares estimators are asymptotically equivalent. We formulate this in a lemma.
Lemma 3.
In the over-dispersed Poisson model Section 3.3.1 and Section 3.3.3, τ ( ξ ^ * ξ ^ q l ) P 0 , and for the Poisson deviance D as in (3), τ R S S * D P 0 . These results still hold if ξ ^ * is replaced by ξ ^ l s * or ξ ^ q l * , R S S * is replaced by R S S l s * or R S S q l * , or τ is replaced by τ ^ q l or τ ^ l s .
We are now armed with four candidate statistics for an encompassing test:
R l s = τ ^ l s R S S D , R q l = τ ^ q l R S S D , R l s * = R S S R S S l s * , and R q l * = R S S R S S q l * .
To find their asymptotic distribution, we exploit that the distribution of each one is asymptotically equivalent to a quadratic form of the same random vector Y. This is reflected in the limiting distribution. which we formulate in a theorem.
Theorem 1.
In the over-dispersed Poisson model Section 3.3.1 and Section 3.3.3, R l s , R q l , R l s * and R q l * are asymptotically equivalent, so that the difference of any two vanishes in probability. For U = D N ( 0 , I ) , Π as in (5), M as in (6) and M * as in (9), each statistic is asymptotically distributed as:
R O D P = U Π 1 / 2 M Π 1 / 2 U U M * U .
Crucially, the asymptotic distribution R O D P is invariant to σ 2 . While it is again a function of the unknown, but consistently estimable frequencies π i j , for large τ , the plug-in version R ^ O D P = R O D P ( Π ^ ) has the same distribution as R O D P ( Π ) .
Theorem 1 allows us to test whether the over-dispersed Poisson model encompasses the generalized log-normal model. For a given critical value, if we reject that the R-statistic was drawn from R ^ O D P , then we reject that the over-dispersed Poisson model M O D P encompasses the generalized log-normal model. While this indicates that the over-dispersed Poisson model is likely wrong, it could mean that the generalized log-normal model is correct or that some other model is appropriate. Conversely, non-rejection means that we cannot reject that the over-dispersed Poisson model encompasses the generalized log-normal model.
The distribution R O D P does not have a closed-form, but precise saddle point approximations are available, as we show below. Furthermore, it is of interest to investigate the impact of the choice among the different test statistics and plug-in estimators for Π appearing in R O D P in finite samples. Above that, we may question the power properties of the test. We discuss these points below in Section 5.

4.3. Null Model: Generalized Log-Normal

We first derive the small- ω 2 asymptotic distribution of Poisson quasi-likelihood and weighted least squares estimators when the data generating process is generalized log-normal. Then, we find the asymptotic distribution of the R-statistic proposed for an encompassing test above.
First, given asymptotic standard-normality on the log scale as in (7), we can easily show asymptotic normality of the weighted least squares estimator. As it turns out, Poisson quasi-likelihood estimators are also asymptotically equivalent to the weighted least squares estimators when the data generating process is generalized log-normal. We formalize this result in a lemma.
Lemma 4.
Define Σ = ( X Π X ) 1 X Π 2 X ( X Π X ) 1 , and let U = D N ( 0 , I ) . Then, in the generalized log-normal model Section 3.4.1 and Section 3.4.3,
( ω 2 ) 1 / 2 ( ξ ^ * ξ ) D N ( 0 , Σ ) and ( ω 2 ) 1 R S S * D U Π 1 / 2 M * Π 1 / 2 U .
Further, ( ω 2 ) 1 / 2 ( ξ ^ * ξ ^ q l ) P 0 and ( ω 2 ) 1 ( R S S * D / τ ) P 0 . These results still hold if ξ ^ * is replaced by ξ ^ l s * or ξ ^ q l * , R S S * is replaced by R S S l s * or R S S q l * , or τ is replaced by τ ^ q l or τ ^ l s .
With these results in place, we can find the distribution of the R-statistics in the generalized log-normal model.
Theorem 2.
In the generalized log-normal model Section 3.4.1 and Section 3.4.3, R l s , R q l , R l s * and R q l * as in (10) are asymptotically equivalent so that the difference of any two vanishes in probability. For U = D N ( 0 , I ) , Π as in (5), M as in (6) and M * as in (9), each statistic is asymptotically distributed:
R G L N = U M U U Π 1 / 2 M * Π 1 / 2 U .
Thus, the test statistics are asymptotically distributed as the ratio of quadratic forms in both data generating processes. The difference arises in the sandwich-matrices. While the orthogonal projections M and M * feature in both distributions, the frequency matrix Π acts in different ways on R O D P and R G L N . Intuitively, R O D P is the ratio of “bad” least squares to “good” weighted least squares residuals computed in a heteroskedastic Gaussian model. In contrast, R G L N has the interpretation as the ratio of “good” least squares to “bad” weighted least squares residuals now computed in a homoskedastic model. Thus, we may expect draws from R G L N to likely be smaller than those from R O D P .

4.4. Distribution of Ratios of Quadratic Forms

We discuss the support of and numerical saddle point approximations to the limiting distributions of the encompassing tests under either data generating process.
The limiting distribution under the null hypothesis in both models is a ratio of dependent quadratic forms in normal random variables. This class of distributions is rather common. Besides standard F distributions, which are a special case, they appear for example in the Durbin–Watson test for serial correlation (Durbin and Watson 1950, 1951). While the distributions generally do not permit closed-form computations of the cdf, fast and precise numerical methods are available.
Butler and Paolella (2008) study a setting that includes ours, but is more general. They consider R = ϵ A ϵ / ϵ B ϵ where A and B are symmetric n × n matrices, B is positive semidefinite and ϵ = D N ( ν , I ) . In our scenario, both A and B are positive semidefinite, and ν = 0 .
Butler and Paolella (2008, Lemma 2) state that R is degenerate if and only if A = c B for some constant c. In our setting, this occurs if Π = n 1 I , so all cells have the same mean. This matches our observation from Section 4.1 that generalized log-normal and over-dispersed Poisson model are indistinguishable if all cells Y i j have the same mean. In that case, both the standard deviation to mean and the variance to mean ratio are constant across cells. This manifests in the collapse of both R O D P and R G L N to a point mass at n.
Further, Butler and Paolella (2008, Lemma 3) derive the support of R for a variety of cases depending on the properties of A and B. Building on their work, we can prove the following result.
Lemma 5.
The distributions R G L N and R O D P have the same support. In non-degenerate cases, the support is ( l , r ) for 0 < l < r < .
The cumulative distribution functions and densities of ratios of quadratic forms admit saddle point approximations. We adapt the discussion in Butler and Paolella (2008) to our scenario in which ν = 0 ; a setting that matches Lieberman (1994). We aim to approximate:
P ( R r ) = P ϵ A ϵ ϵ B ϵ r = P ( X r 0 ) where X r = ϵ ( A r B ) ϵ
First, we compute the eigenvalues of A r B denoted λ 1 , , λ n . We can write the cumulant generating function K ( s ) , the log of the moment generating function φ ( s ) = E [ exp { s ( A r B ) } ] , of X r and its -th derivative as:
K ( s ) = 1 2 t = 1 n log ( 1 2 s λ t ) , K ( ) ( s ) = s K ( s ) = ( 2 2 ) ! ! t = 1 n λ t 1 2 s λ t
where a ! ! = a ( a 2 ) ( a 4 ) is the double factorial with the usual definition that 0 ! ! = 1 . The saddle point is the root:
s ^ : K ( 1 ) ( s ^ ) = t = 1 n λ t 1 2 s ^ λ t = 0 .
Except for the special case when all eigenvalues λ t are zero, so K ( 1 ) ( s ) = 0 , s ^ is unique since K ( 1 ) ( s ) is strictly increasing. The former case occurs if and only if E ( X r ) = 0 , which is the case for r = trace ( A ) / trace ( B ) . This case is dealt with separately. For the other cases, we compute:
w ^ = sgn ( s ^ ) 2 K ( s ^ ) and u ^ = s ^ K ( 2 ) ( s ^ ) .
Then, denoting by Φ ( . ) and ϕ ( . ) the standard normal cdf and density, respectively, the first order approximation to the cdf of R is:
P ^ ( R r ) = Φ ( w ^ ) + ϕ ( w ^ ) ( w ^ 1 u ^ 1 ) , if E ( X r ) 0 1 2 + K ( 3 ) ( 0 ) 6 2 π K ( 2 ) ( 0 ) 3 / 2 , if E ( X r ) = 0 .
This saddle point approximation is a special case of the more general form in Lugannani and Rice (1980). This is what Lieberman (1994) built on. Lugannani and Rice (1980) analysed the error behaviour for a sum of independent and identically distributed random variables and showed uniformity of the errors for a large sample. Butler and Paolella (2008) instead considered a fixed sample size and show uniformity of errors in the tail of the distribution. This seems appealing for our scenario, since we would expect the rejection region of the test to correspond to the tail of the distribution.

4.5. Power

We show that the conjecture of a link between the power of the tests and variation in the means raised above in Section 4.1 is correct. To prove this, we consider a sequential asymptotic argument in which first, depending on the data generating process, τ becomes large or ω 2 becomes small and then the means become “more dispersed” in a sense made precise below. Based on this argument, we can justify a one-sided test where the rejection region corresponds to the upper tail when the null model is generalized log-normal and to the lower tail when it is over-dispersed Poisson.
The sequential asymptotics allows us to exclusively consider the impact of more dispersed means on R G L N and R O D P without worrying about the effect on the distribution of R l s , R q l , R l s * or R q l * . However, larger mean dispersion would be linked to changes in ξ ( 2 ) , a parameter that we keep fixed when deriving the asymptotic distribution of the test statistics in the first stage of the asymptotics. Therefore, we would expect the approximation quality achieved in the first stage to be affected by the second stage. The interpretation of the results is thus for a given first stage approximation quality, however large τ or small ω 2 may be needed to achieve this.
We model “more dispersed” means by increasing the variation in the frequencies π i j and specifically by letting some frequencies go to zero. In this way, we do not make a statement about the means in absolute terms, but merely say that some cell means become large relative to others.
For our analysis, we exclude cells for which estimation would yield a perfect fit; equivalently, we can impose that the frequencies do not exclusively vanish for perfectly-fitted cells. For example, in a chain-ladder model for the run-off triangle in Table 1, the corner cells ( 1 , 10 ) and ( 10 , 1 ) would be fit perfectly as they have their own parameters Δ β 10 and Δ α 10 .
To increase the variation in the frequencies π i j , we decide on n q cells of the run-off triangle for which we want the frequencies to vanish. We require that the remaining q cells with non-vanishing frequencies make up an array on which we can estimate a model with the same structure for the linear predictor μ i j as for the full data without obtaining a perfect fit. For example, for a chain-ladder model in which μ i j = α i + β j + δ , this would be the case for rectangular arrays with at least two columns and rows or for triangular arrays with at least three rows and columns.
For the ease of notation, we sort rows and columns of the frequency matrix Π defined in (5) such that the cells with vanishing frequencies are in the bottom right block of the matrix. Then, for a q × q matrix Π 1 and an ( n q ) × ( n q ) matrix Π 2 , we define a new frequency matrix:
Π ( t ) = s ( t ) Π 1 0 0 t Π 2 where s ( t ) = { trace ( Π 1 ) + t · trace ( Π 2 ) } 1
so s ( t ) takes care of the normalization such that the elements of Π ( t ) are still frequencies. The idea is to model the vanishing frequencies by letting t 0 . Clearly, Π ( 1 ) corresponds to Π , whereas Π ( 0 ) has all frequencies in the the bottom right block equal to zero. We assume that Π 1 q 1 I , so that the limiting case does not correspond to a scenario without variation in the frequencies.
Similarly, we sort rows and columns of the design matrix X to obtain a convenient partition. We sort the rows such that the q cells relating to non-vanishing frequencies are in the first q rows. Further, we sort the columns so the p 1 , say, parameters relevant for these q cells are in the first p 1 columns. Then, we can partition:
X = X 11 X 12 X 21 X 22
where X 11 is q × p 1 and X 22 is ( n q ) × ( p p 1 ) . Column sorting ensures that X 12 = 0 . Imposing that there is no perfect fit for the q cells without vanishing frequencies implies that p 1 < q , so there are fewer parameters than cells.
We are now interested in the properties of the large τ or small ω 2 limiting distributions of the R statistics’ when some frequencies are small. For fixed t, Theorems 1 and 2 apply. Thus, for fixed frequencies Π ( t ) , the large τ and small ω 2 distributions of the R statistics in an over-dispersed Poisson and generalized log-normal model, respectively, are:
R O D P ( t ) = U Π ( t ) 1 / 2 M Π ( t ) 1 / 2 U U M ( t ) * U and R G L N ( t ) = U M U U Π ( t ) 1 / 2 M ( t ) * Π ( t ) 1 / 2 U
where M ( t ) * is the weighted least squares orthogonal projection matrix I X ( t ) * ( X ( t ) * X ( t ) * ) 1 X ( t ) * for X ( t ) * = Π ( t ) 1 / 2 X . Thus, Π ( t ) enters not only directly, but also indirectly through M ( t ) * .
We study the tests’ power by looking at the limit of R O D P ( t ) and R G L N ( t ) as t 0 . We reiterate that the sequential asymptotics neglect interactions between first stage asymptotics for large τ or small ω 2 and the second stage small t asymptotics. A first intuition that neglects the potential influence of M ( t ) * may tell us that R G L N ( t ) should be well behaved while R O D P ( t ) blows up for small t. This turns out to be correct.
Theorem 3.
Let U = D N ( 0 , I ) and let U 1 contain the first q elements of U. Further define Π ˘ 1 = s ( 0 ) Π 1 , X ˘ 11 * = Π ˘ 1 1 / 2 X 11 and M ˘ 11 * = I X ˘ 11 * ( X ˘ 11 * X ˘ 11 * ) 1 X ˘ 11 * . Then, as t 0 ,
R G L N ( t ) a . s . U M U U 1 Π ˘ 1 1 / 2 M ˘ 11 * Π ˘ 1 1 / 2 U 1 : = R G L N ( 0 ) while R O D P ( t ) a . s . .
Further, for α ( 0 , 1 ) , let q G L N , α ( t ) be the α-quantile of R G L N ( t ) and similarly for q O D P , α ( t ) . Then, R O D P ( t ) > q G L N , α ( t ) and R G L N ( t ) q O D P , α ( t ) almost surely as t 0 .
Theorem 3 justifies one-sided tests and shows that the power of the tests under either data generating process goes to unity in the sequential asymptotic argument. Since the distribution of R O D P and R G L N coincides for equal means, the power of the tests to distinguish between the data generating processes comes entirely from the variation in means. As the mean variation becomes large, R O D P first order stochastic dominates R G L N . Thus, we can consider the lower tail of R O D P and the upper tail of R G L N as rejection regions. While still controlling the size of the test under the null, we gain power compared to two-sided tests as the mean variation increases.
The denominator of R G L N ( 0 ) can be interpreted as “bad” weighted least squares residuals in a homoskedastic Gaussian model computed on just the subset of q cells with non-vanishing frequencies. For a brief intuition as to why only cells and parameters relating to X 11 matter in the limit of the denominator, we consider weighted least squares estimation for Z = X ξ + Π ( t ) 1 / 2 ϵ , taking Π ( t ) as given. We solve this by minimizing | Π ( t ) 1 / 2 ( Z X ξ ) | 2 . For t > 0 , the minimum is given by Z Π ( t ) 1 / 2 M ( t ) * Π ( t ) 1 / 2 Z . When t = 0 , the last n q elements of Z and rows of X corresponding to the vanishing frequencies do not contribute to the norm. The same holds for the last p p 1 parameters in ξ that are then not identified. Thus, for t = 0 , letting Z 1 contain the first p 1 elements of Z, the minimum of the norm equals Z 1 Π ˘ 1 1 / 2 M ˘ 11 * Π ˘ 1 1 / 2 Z 1 .

5. Simulations

With the theoretical results for encompassing tests between over-dispersed Poisson and generalized log-normal models in place, we show that they perform well in a simulation study. First, we show that saddle point approximations to the limiting distributions R O D P and R G L N are very accurate. Second, we tackle an issue that disappears in the limit; namely, the choice between asymptotically identical estimators that generally differ in finite samples. We show that finite sample performance is indeed affected by this choice. However, we find that for some choices, finite sample and asymptotic distributions are very close. Third, we show that the tests have high power in finite samples and, considering the behaviour of the limiting distributions alone, that power increases quickly with the variation in means. For the simulations and empirical applications below, we use the Python packages quad_form_ratio (Harnau 2018b) and apc (Harnau 2017). The package was inspired by the R (R Core Team 2017) package apc (Nielsen 2015) with similar functionality.

5.1. Quality of Saddle Point Approximations

We show that saddle point approximations work well compared to large Monte Carlo simulations.
We consider three parameterizations. First, we let the design X correspond to that of a chain-ladder model for a ten-by-ten run-off triangle and set the frequency matrix Π to the least squares estimates Π ^ l s = Π ( ξ ^ l s ( 2 ) ) of the Verrall et al. (2010) data in Table 1 ( V N J ). Second, for the same design, we now set the frequency matrix to the least squares plug-in estimates based on a popular dataset by Taylor and Ashe (1983) ( T A ). We provide these data in the Appendix A in Table A2. Third, we consider a design X for an extended chain-ladder model in an eleven-by-eleven run-off triangle and set Π to the least squares plug-in estimates of the Barnett and Zehnwirth (2000) data ( B Z ), also shown in the Appendix in Table A1. We remark that in the computations, we drop the corner cells of the triangles that would be fit perfectly in any case; this helps to avoid numerical issues without affecting the results.
Given a data generating process R chosen from R O D P and R G L N , a design matrix X and a frequency matrix Π , we use a large Monte Carlo simulation as a benchmark for the saddle point approximation. First, we draw B = 10 7 realizations r b from R . For the Monte Carlo cdf P ^ M C ( R q ) = B 1 b = 1 B 1 ( r b q ) , we then find the quantiles q α , so P ^ M C ( R q α ) = α for α = 0 , 0.005 , 0.01 , , 1 . To compute the saddle point approximation P ^ S P ( R q ) , we use the implementation of the procedure described in Section 4.4 in the package quad_form_ratio. Then, for each Monte Carlo quantile q α , we compute the difference P ^ S P ( R q α ) α . Taking the Monte Carlo cdf as the truth, we refer to this as the saddle point approximation error.
Figure 1a shows the generalized log-normal saddle point approximation error P ^ S P ( R G L N q α ) α plotted against α . One and two (pointwise) Monte Carlo standard errors α ( 1 α ) / B are shaded in blue and green, respectively. While the approximation errors for T A are generally not significantly different from zero, the same cannot be said for the other two sets of parameters. For the parameterizations V N J and B Z , the errors start and end in zero and are negative in between. Despite statistically-significant differences, the approximation is very good with a maximum absolute approximation error of just over 0.006 . The errors in the tails are much smaller, as we might have expected given the results by Butler and Paolella (2008) discussed in Section 4.4.
Figure 1b shows the plot for the approximation error to R O D P produced in the same way as Figure 1a. The approximation error is positive and generally significantly different from zero across parameterizations. Yet, the largest error is about 0.005 with smaller errors in the tails.
We would argue that the saddle point approximation errors, while statistically significant, are negligible in applications. That is, using a saddle point approximation rather than a large Monte Carlo simulation is unlikely to affect the practitioner’s modelling decision.

5.2. Finite Sample Approximations under the Null

The asymptotic theory above left us without guidance on how to choose between test statistics R and estimators for the nuisance parameter Π that appears in the limiting distributions R . While the choice is irrelevant for large τ or small ω 2 , we show that it matters in finite samples and that some combinations perform much better than others when it comes to approximation under the null hypothesis.
In applications, we approximate the distribution of R by R ^ = R ( Π ^ ) . That is, defining the α quantile of R ^ as q α R ^ , we hope that P ( R q α R ^ ) α under the null hypothesis. To assess whether this is justified, we simulate the approximation quality across 16 asymptotically identical combinations of R-statistics and ratios of quadratic forms R ^ . We describe the simulation process in three stages. First, we explain how we set up the data generating processes for the generalized log-normal and over-dispersed Poisson model. Second, we lay out explicitly the combinations we consider. Third, we explain how we compute the approximation errors. As in Section 5.1, we point out that we drop the corner cells of the triangles in simulations. This aids numerical stability without affecting the results.
For the generalized log-normal model, we simulate independent log-normal variables Y i j , so log ( Y i j ) = D N ( x i j ξ , ω 2 ) . We consider three settings for the true parameters corresponding largely to the estimates from the same three datasets we used in Section 5.1, namely the Verrall et al. (2010) data ( V N J ), Taylor and Ashe (1983) data ( T A ) and Barnett and Zehnwirth (2000) data ( B Z ). Specifically, we consider pairs ( ξ , ω 2 ) set to the estimated counterparts ( ξ ^ l s , ω ^ 2 / s ) for s = 1 , 2 . The estimates ω ^ 2 are 0.39 for V N J , 0.12 for T A and 0.001 for B Z . Theory tells us that the approximation errors should decrease with ω 2 , thus as s increases.
For the over-dispersed Poisson model, we use a compound Poisson-gamma data generating process, largely following Harnau and Nielsen (2017) and Harnau (2018a). We simulate independent Y i j = = 1 N i j X s where N i j = D Poisson { exp ( x i j ξ ) } and X s are independent Gamma distributed with scale σ 2 1 and shape ( σ 2 1 ) 1 . This satisfies the assumptions for the over-dispersed Poisson model in Section 3.3.1 and Section 3.3.3. For the true parameters ( τ , ξ ( 2 ) , σ 2 ) , we consider three sets of estimates ( s τ ^ q l , ξ ^ l s ( 2 ) , σ ^ 2 ) from the same data as for the log-normal data generating process. We use least squares estimates ξ ^ l s ( 2 ) so that the frequency matrix Π is identical within parameterization between the two data generating processes. The estimates for σ ^ 2 are 10,393 for V N J , 52,862 for T A and 124 for B Z . Those for τ ^ q l are 14 , 633 , 814 for V N J , 34 , 358 , 090 for T A and 10 , 221 , 194 for B Z . Again, we consider s = 1 , 2 , but this time scaling the aggregate predictor. If this increases, so should the approximation quality. We recall that ξ ( 2 ) and τ pin down μ 11 through the one-to-one mapping τ = exp ( μ 11 ) i j I exp ( x i j ( 2 ) ξ ( 2 ) ) . Thus, multiplying τ by s corresponds to adding log ( s ) to μ 11 .
For a given data generating process, we independently draw B = 10 5 run-off triangles b = { Y i j , b : ( i , j ) I } and compute a battery of statistics for each draw. First, we compute the four test statistics R l s , R q l , R l s * and R q l * as defined in (10). Second, we compute the estimates for the frequency matrices Π based on least squares estimates, quasi-likelihood estimates and feasible weighted least squares estimates with least squares and with the quasi-likelihood first stage. This leads to four different approximations to the limiting distribution, which, dropping the subscript for the data generating process, we denote by:
R ^ l s = R { Π ( ξ ^ l s ( 2 ) ) } , R ^ q l = R { Π ( ξ ^ q l ( 2 ) ) } , R ^ l s * = R { Π ( ξ ^ l s * ( 2 ) ) } , and R ^ q l * = R { Π ( ξ ^ q l * ( 2 ) ) } .
Given a data generating process and a choice of test statistic and limiting distribution approximation ( R , R ^ ) , we approximate P ( R q α R ^ ) by Monte Carlo simulation. For each combination ( R , R ^ ) , we have B paired realizations; for example, R b and the distribution R ^ b are based on the triangle b . Denote the saddle point approximation to the cdf of R ^ b as G b ( q ) = P ^ S P ( R ^ b q ) . Neglecting the saddle point approximation error, we then compute P ( R q α R ^ ) as P ^ M C ( R q α R ^ ) = B 1 b = 1 B 1 { G b ( R b ) α } , exploiting that G b ( R b ) α whenever R b G b 1 ( α ) = q α , b R ^ . We do this for α A = { 0.005 , 0.01 , , 0.995 } .
To evaluate the performance, we consider three metrics: area under the curve of absolute errors (also roughly the mean absolute error), maximum absolute error and error at (one-sided) 5 % critical values. We compute the area under the curve as AUC = = 1 199 | P ^ M C ( R q α R ^ ) α | Δ α where α = 0.005 · , so Δ α = 0.005 ; we can also roughly interpret this as the mean absolute error MAE = 200 / 199 · AUC since α = 200 1 . The maximum absolute error is max α A | P ^ M C ( R q α R ^ ) α | . Finally, the error at 5 % critical values is P ^ M C ( R > q G L N , 0.95 R ^ ) 0.05 for the generalized log-normal and P ^ M C ( R q O D P , 0.05 R ^ ) 0.05 for the over-dispersed Poisson data generating process.
Figure 2 shows bar charts for the area under the curve for all 16 combinations of R and R ^ stacked across the three parameterizations for s = 1 . The chart is ordered by the sum of errors across parameterizations and data generating processes within combination, increasing from top to bottom. The maximum absolute error summed over parameterizations is indicated by “+”. Since a bar chart for the maximum absolute errors is qualitatively very similar to the plot for the area under the curve, we do not discuss it separately and instead provide it as Figure A1 in the Appendix A.
Looking first at the sum over parameterizations and data generating processes within combinations, we see large differences in approximation quality both for the area under the curve of absolute errors and the maximum absolute error. The former varies from about 5 pp (percentage-points) for ( R l s * , R ^ l s * ) to close to 30 pp for ( R q l , R ^ l s * ) , the latter from 8 pp to 45 pp. It is notable that the four combinations involving R q l are congregated at the bottom of the pack. In contrast, the three best performing combinations all involve R l s * . These three top-performers have a substantial head start compared to their competition. While their AUC varies from 4.8 pp to 6.0 pp, there is a jump to 13.9 pp for fourth place. Similarly, the maximum absolute errors of the top three contenders lie between 7.5 pp and 9.3 pp, while those for fourth place add up to 20.5 pp.
Considering next the contributions of the individual parameterizations to the area under the curve across data generating processes, the influence is by no means balanced. Instead, the average contribution over combinations of the V N J , T A and B Z parameterizations is about 35 % , 57 % and 8 % , respectively. This ordering is well aligned in magnitude and ordering with that of ω 2 and σ 2 / τ , loosely interpretable as a measure for the expected approximation quality. Still, considering the contributions of the parameterizations within combinations, we see substantial heterogeneity. For example, the T A parameterization contributes much less to ( R q l * , R ^ q l * ) than V N J , while the reverse is true for ( R l s , R ^ l s ) .
Finally, we see substantial variation between the two data generating processes. While the range of areas under the curve of absolute errors aggregated over parameterizations for the generalized log-normal is 0.7 pp to 10 pp, that for the over-dispersed Poisson is 3.2 pp to 19.4 pp. The best performer for the generalized log-normal is, perhaps unsurprisingly, ( R l s * , R ^ l s ) . Intuitively, since the data generating process is log-normal, the asymptotic results would be exact for this combination if we plugged the true parameters into the frequency matrices. Just shy of these, we plug in the least squares parameter estimates, which are maximum likelihood estimated. It is perhaps more surprising that using R q l is not generally a good idea for the over-dispersed Poisson data generating process even though the fact that these combinations take the bottom four slots is largely driven by the T A parametrization. Reassuringly, the top three performers across data generating processes also take the top three spots within data generating processes, albeit with a slightly changed ordering.
Figure 3a shows box plots for the size error at 5 % nominal size computed over the three parameterizations and two data generating processes within combinations ( R , R ^ ) for s = 1 . Positive errors indicate an over-sized and negative errors an under-sized test. In the plots, medians are indicated by blue lines inside the boxes. The boxes show the interquartile range. Whiskers represent the full range. The ordering is increasing in the sum of the absolute errors at 5 % critical values from top to bottom.
Looking at the medians, we can see that these are close to zero, ranging from 0.15 pp to 0.37 pp. However, there is substantial variation in the interquartile range, 0.1 pp for ( R l s , R ^ q l * ) to 0.7 pp for ( R q l * , R ^ l s ) and range, 0.4 pp for ( R l s * , R ^ l s * ) to 6.7 pp for ( R q l , R ^ l s ) ). The best and worst performers from the analysis for the area under the curve and maximum absolute errors are still found in the top and bottom positions. Particularly the performance of ( R l s * , R ^ l s * ) seems close to perfection with a range from 0.2 pp to 0.2 pp.
Figure 3b is constructed in the same way as Figure 3a, but for s = 2 , halving the variance for the generalized log-normal and doubling the aggregate predictor for the over-dispersed Poisson data generating process. Theory tells us that the approximation quality should improve, and this is indeed what we see. The medians move towards zero, now taking values between 0.05 pp and 0.14 pp; the largest interquartile range is now 1.1 pp and the largest range 2.9 pp.
Overall, the combination ( R l s * , R ^ l s * ) performs very well across the considered parameterizations and data generating processes. This is not to say that we could not marginally increase performance in certain cases, for example by picking ( R l s * , R ^ l s ) when the true data generating process is log-normal. However, even in this case in which we get the data generating process exactly right, not much seems to be gained in approximation quality where it matters most, namely in the tails relevant for testing. Thus, it seems reasonable to simply use ( R l s * , R ^ l s * ) regardless of the hypothesized model, at least for size control.

5.3. Power

Having convinced ourselves that we can control size across a number of parameterizations, we show that the tests have good power. First, we consider how the power in finite sample approximations compares to power in the limiting distributions. Second, we investigate how power changes as the means become more dispersed based on the impact on the limiting distributions R G L N and R O D P alone, as discussed in Section 4.5.

5.3.1. Finite Sample Approximations Under the Alternative

We show that combinations of R-statistics and approximate limiting distributions R ^ that do well for size control under the null hypothesis also do well when it comes to power at 5 % critical values. The data generating processes are identical to those in Section 5.2 and so are the three considered parameterizations V N J , B Z and T A . To avoid numerical issues, we again drop the perfectly-fitted corner cells of the triangles without affecting the results.
To avoid confusion, we stress that we do not consider the impact of more dispersed means in this section. Thus, if we mention asymptotic results, we refer to large τ when the true data generating process is over-dispersed Poisson and for small ω 2 when it is generalized log-normal, holding the frequency matrix Π fixed.
For a given parametrization, we first find the asymptotic power. When the generalized log-normal model is the null hypothesis, we find the 5 % critical values c G L N : P ( R G L N > c G L N ) = 0.05 , using the true parameter values for Π . Then, we compute the power P ( R O D P > c G L N ) . Conversely, when the over-dispersed Poisson is the null model, we find c O D P : P ( R O D P c O D P ) = 0.05 and compute the power P ( R G L N c O D P ) . Lacking closed-form solutions, we again use saddle point approximations, iteratively solving the equations for the critical values to a precision of 10 4 .
Next, we approximate the finite sample power of the top four combinations for size control in Section 5.2, ( R l s * , R ^ l s * ) , ( R l s * , R ^ l s ) , ( R l s , R ^ q l ) and ( R l s * , R ^ q l * ) , by the rejection frequencies under the alternative for s = 1 . For example, say the generalized log-normal model is the null hypothesis, and we want to compute the power for the combination ( R l s * , R ^ l s * ) . Then, we first draw B = 10 5 triangles b from the over-dispersed Poisson data generating process. For each draw b, we find 5 % critical values c G L N , l s , b R ^ : P ( R ^ G L N , l s , b * > c G L N , l s , b R ^ ) = 0.05 . We compute these based on saddle point approximations, solving iteratively up to a precision of 10 4 . Then, we approximate the power as B 1 b = 1 B 1 { R G L N , l s , b * > c G L N , l s , b R ^ } . For the over-dispersed Poisson null hypothesis, we proceed equivalently, using the left tail instead. In this way, we approximate power for all three parameterizations and all four combinations.
Before we proceed, we point out that we should be cautious to interpret power without taking into account the size error in finite samples. A test with larger than nominal size would generally have a power advantage purely due to the size error. One way to control for this is to consider size-adjusted power, which levels the playing field by using critical values not at the nominal, but at the true size. In our case, this would correspond to critical values from the true distribution of the test statistic R, rather than the approximated distribution R ^ . Therefore, the choice of R ^ would not play a role any more. To sidestep this issue, we take a different approach and compare how close the power of the finite sample approximations matches the asymptotic power.
Table 2 shows the asymptotic power and the gap between power in finite sample approximations and asymptotic power.
Looking at the asymptotic power first, we can see little variation between data generating processes within parameterizations. The power is highest for the V N J parameterization with 99 % , followed by B Z with 95 % and T A with 65 % . This ordering aligns with that of the standard deviations of the frequencies π i j under these parameterizations, which are given by 0.016 , 0.012 and 0.009 for V N J , B Z and T A , respectively.
When considering the finite sample approximations, we see that their power is relatively close to the asymptotic power. For V N J , absolute deviations range from 0.14 pp to 0.34 pp and for B Z from 0.94 pp and 1.15 pp. Compared to that, discrepancies for the T A parameterization are larger. The smallest discrepancy of 0.14 pp arises for ( R l s * , R ^ l s ) when the data generating process is generalized log-normal. As before, this is intuitive since it corresponds to plugging maximum likelihood estimated parameters ξ ^ ( 2 ) into Π . With 5.28 pp, the largest discrepancy arises for ( R l s , R ^ q l ) for an over-dispersed Poisson data generating process. Mean absolute errors across parameterizations and data generating processes are rather close, ranging from 1.01 pp for ( R l s * , R ^ l s ) to 2.1 pp for ( R l s , R ^ q l ) . Our proposed favourite from above ( R l s * , R ^ l s * ) comes in second with 1.27 pp. We would argue that we can still justify the use of ( R l s * , R ^ l s * ) regardless of the data generating process.

5.3.2. Increasing Mean Dispersion in Limiting Distributions

We consider the impact of more dispersed means on power based on the the test statistics’ limiting distributions R G L N and R O D P . We show that the power grows quickly as we move from identical means across cells to a scenario where a single frequency hits zero.
For a given diagonal frequency matrix Π with values π i j , we define the linear combination:
Π ( t ) = t Π + ( 1 t ) n 1 I n .
Thus, for t = 1 , we recover Π , while for t = 0 , we are in a setting where all cells have the same frequencies, so all means are identical. In the latter scenario, R G L N and R O D P collapse to a point-mass at n, as discussed in Section 4.4. We consider t ranging from just over zero to just under t m a x : t m a x min i j I ( π i j ) + ( 1 t m a x ) n 1 = 0 . The significance of t m a x is that Π ( t m a x ) corresponds to the matrix where the smallest frequency is exactly zero.
For each t, we approximate one-sided 5 % critical values of R G L N ( t ) = R G L N ( Π ( t ) ) and R O D P ( t ) = R O D P ( Π ( t ) ) through:
c G L N ( t ) : P ^ S P ( R G L N ( t ) > c G L N ( t ) ) = 0.05 and c O D P ( t ) : P ^ S P ( R O D P ( t ) c O D P ( t ) ) = 0.05 .
We iteratively solve the equations up to a precision of 10 4 . Theorem 3 tells us that the critical values should grow for both models, but that c G L N ( t ) converges as t approaches t m a x , while c O D P ( t ) goes to infinity.
Then, for given t and critical values, we find the power when the null model is generalized log-normal P ( R O D P ( t ) > c G L N ( t ) ) and when the null model is over-dispersed Poisson P ( R G L N ( t ) c O D P ( t ) ) . Again, we use saddle point approximation. Based on Theorem 3, we should see the power go to unity as t approaches t m a x .
We consider the same parameterizations V N J , T A and B Z of frequency matrices Π and design matrices X as above. The values for t m a x are 1.083 for V N J , 1.396 for T A and 1.103 for B Z . To avoid numerical issues, we again drop the perfectly-fitted corner cells from the triangles. In this case, while the power is not affected, the critical values are scaled down by the ratio of τ ^ l s computed over the smaller array without corner cells to that computed over the full triangle. Since this is merely proportional, the results are not affected qualitatively.
Figure 4a shows the power when the generalized log-normal model is the null hypothesis. For all considered parameterizations, this is close to 5 % for t close to zero, increasing monotonically with t and approaching unity as t approaches t m a x , as expected.
For t = 1 , where Π ( t ) corresponds to the least squares estimated frequencies from the data, the power matches what we found in Table 2.
Figure 4b shows the difference in power between the two models plotted over t. For the three settings we consider, these curves have a similar shape and start and end at zero. Generally, the power is very comparable, with differences between 2 pp and 1 pp, again matching our findings from Table 2 for t = 1 .
Figure 5a shows the one-sided 5 % critical values c G L N ( t ) plotted over t. As expected, these are increasing for all settings. Figure 5b shows the ratio of the critical values c O D P ( t ) to c G L N ( t ) . This starts at unity, initially decreases, then increases and, finally, explodes towards infinity as we approach t m a x .
Taking the plots together, we get the following interpretation. We recall that the two distributions are identical for t = 0 . Further, the rejection regions for the generalized log-normal null is the upper tail, while the lower tail is relevant for the over-dispersed Poisson model. However, for small t, the mass of both R G L N ( t ) and R O D P ( t ) is highly concentrated around n, and the distributions are quite similar. This explains why the power is initially close to 5 % for either. Further, due to the concentration, c G L N ( t ) and c O D P ( t ) are initially close. As t increases, both distributions become more spread out and move up the real line, with R O D P ( t ) moving faster than R G L N ( t ) . This is reflected in the increase in power. Initially, c G L N ( t ) increases faster than c O D P ( t ) , so their ratio decreases. Yet, for t large enough, c O D P ( t ) overtakes c G L N ( t ) , indicating the point at which power reaches 95 % for either model. The power differential is necessarily zero at this point. Finally, c O D P ( t ) explodes while c G L N ( t ) converges as t approaches t m a x , so the ratio diverges.

6. Empirical Applications

We consider a range of empirical examples. First, we revisit the empirical illustration of the problem from the beginning of the paper in Section 2. We show that the proposed test favours the over-dispersed Poisson model over the generalized log-normal model. Second, we consider an example that perhaps somewhat cautions against starting off with a model that may be misspecified to begin with: dropping a clearly needed calendar effect turns the results of the encompassing tests upside down. Third, taking these insights into account, we implement a testing procedure that makes use of a number of recent results: deciding between the over-dispersed Poisson and generalized log-normal model, evaluating misspecification and testing for the need for a calendar effect.

6.1. Empirical Illustration Revisited

We revisit the data in Table 1 discussed in Section 2 and show that we can reject that the (generalized) log-normal model encompasses the over-dispersed Poisson model, but cannot reject the alternative direction. Thus, the encompassing tests proposed in this paper have higher power to distinguish between the two models than the misspecification tests Harnau (2018a) applied to these data. We remark that the encompassing tests were designed explicitly to distinguish between the two models, in contrast to the more general misspecification tests.
Table 3 shows p-values for all 16 combinations of R-statistics and R ^ under both null hypotheses.
Computing the four R-statistics in (10) yields:
R l s = 104.87 , R q l = 105.61 , R l s * = 113.19 and R q l * = 108.39 .
Thus, while not identical, the test statistics appear quite similar.
First, we consider the generalized log-normal model as the null model so:
H 0 : generalized log-normal vs . H A : over-dispersed Poisson .
This is consistent with the applications in Kuang et al. (2015) and Harnau (2018a), who consider these data in a log-normal model. Looking at our preferred combination ( R l s * , R ^ l s * ) , we find a p-value of 0.001 , rejecting the model. Reassuringly, we reject the generalized log-normal model for any combination of R and R ^ . The most favourable impression to this null hypothesis is given by ( R l s , R ^ l s ) with a p-value of 0.004 .
If we instead take the over-dispersed Poisson model as the null, so:
H 0 : over-dispersed Poisson vs . H A : generalized log-normal ,
the model cannot be rejected with a p-value of 0.17 for ( R l s * , R ^ l s * ) . Again, this decision is quite robust to the choice of estimators with a least favourable p-value of 0.09 obtained based on ( R l s , R ^ q l * ) . If we accept the null, we can evaluate the power against the generalized log-normal model. For instance, the 5 % critical value under the over-dispersed Poisson model is 95.7 . The probability of drawing a value smaller than that from the generalized log-normal model is 0.99 . Thus, the power at the 5 % critical value is close to unity. We can also find the power at the value taken by R l s * , interpretable as the 17 % critical value if we like. This is simply one minus the p-value of the generalized log-normal model, thus equal to 1 0.001 = 0.999 .

6.2. Sensitivity to Invalid Model Reductions

The Barnett and Zehnwirth (2000) data are known to require a calendar effect for modelling. We show those data in Table A1 in the Appendix. Barnett and Zehnwirth (2000), Kuang et al. (2015) and Harnau (2018a) approached these datasets using log-normal models. Here, we find that an encompassing test instead heavily favours an over-dispersed Poisson model. Further, we show that dropping the needed calendar effect substantially affects the test results.
We again first consider a generalized log-normal model; however, we initially allow for a calendar effect. Adding the prefix “extended” to models with calendar effect, we test:
H 0 : extended generalized log-normal vs . H A : extended over-dispersed Poisson .
Our preferred test statistic R l s * = 114.40 . Paired with R ^ l s * , this yields a p-value of 0.02 . Thus, the generalized log-normal model is clearly rejected. For illustrative purposes, we continue anyway and test whether we can drop the calendar effect from the generalized log-normal model. Thus, the hypothesis is:
H 0 : generalized log-normal vs . H A : extended generalized log-normal .
Kuang and Nielsen (2018) show that for small ω 2 , we can use a standard F-test for this purpose. If we assumed that the data generating process is not generalized log-normal, but log-normal, the F-test would be exact. This test rejects the reduction with a p-value of 0.00 . If again we decide to continue anyway, we can now test the generalized log-normal against an over-dispersed Poisson model, both without the calendar effect. Thus, the hypothesis is:
H 0 : generalized log-normal vs . H A : over-dispersed Poisson .
Interestingly, the log-normal model does not look so bad any more now. For this model, R l s * = 87.54 , which yields a p-value of 0.10 . Of course, this should not encourage us to assume that the generalized log-normal model without the calendar effect is actually a good choice. Rather, it draws attention to the fact that tests computed on inappropriately-reduced models may yield misleading conclusions. The tests proposed in this paper assume that the null model is well specified, and the results are generally only valid if this is correct. In applications, we may relax this statement to “the tests only give useful indications if the null model describes the data well”. In this case, we did not only ignore the initial rejection of the generalized log-normal model, but also that calendar effects are clearly needed to model the data well.
We now start over, switching the role of the two models, thus starting with an extended over-dispersed Poisson model. The first hypothesis is the mirror image from above:
H 0 : extended over-dispersed Poisson vs . H A : extended generalized log-normal .
The test statistic is still R l s * = 114.40 , but now, we cannot reject the null hypothesis with a p-value of 0.14 . We may thus feel comfortable to model the data using an over-dispersed Poisson model with a calendar effect. Next, we investigate whether the calendar effect can be dropped, testing:
H 0 : over-dispersed Poisson vs . H A : extended over-dispersed Poisson .
Harnau and Nielsen (2017) showed that for large τ , this can be done with an F-test based on Poisson deviances. This reduction is clearly rejected, again with a p-value of 0.00 . We move on anyway, drop the calendar effect and test:
H 0 : over-dispersed Poisson vs . H A : generalized log-normal .
In this case, the p-value is 0.01 , and we reject the null, so we get the opposite result.
Comparing the outcomes of the tests, it seems clear that an over-dispersed Poisson model with the calendar effect is the most reasonable choice. However, if we had not started at this point, but rather never considered a calendar effect in the first place, we might have come to a very different conclusion. This indicates that the starting point can matter a great deal for the model choice and that it may be a good idea to start with a more general model and test for reductions, even if we were fairly certain that the reduced model is a good choice.

6.3. A General to Specific Testing Procedure

The Taylor and Ashe (1983) data have frequently been modelled as over-dispersed Poisson, for example by England and Verrall (1999), England (2002) and Harnau (2018a). We provide those data in Table A2 in the Appendix A. Based on the insight from the application to the Barnett and Zehnwirth (2000) data above, we start with a general model with the calendar effect and use a whole battery of tests to see if a generalized log-normal or over-dispersed Poisson chain-ladder model can be justified. We find that an over-dispersed Poisson chain-ladder model is reasonable for these data.
We first consider a generalized log-normal model with the calendar effect. We test:
H 0 : extended generalized log-normal vs . H A : extended over-dispersed Poisson .
The null hypothesis is clearly rejected with a test statistic of R l s * = 81.5 and a p-value of 0.001 . Thus, we do not proceed further with this model.
Instead, we now start with an over-dispersed Poisson model with the calendar effect. The hypothesis:
H 0 : extended over-dispersed Poisson vs . H A : extended generalized log-normal
cannot be rejected with a p-value of 0.92 . We point out that this indicates that the draw is in the right tail of R ^ O D P . While we would argue this is not the case here, we may worry about values that are too far out in the right tail of R ^ O D P , which would perhaps indicate that we should reject both models.
Next, we apply the misspecification tests by Harnau (2018a). We first split the run-off triangle into four sub-samples, as indicated in Figure 6.
We can now test whether the over-dispersion is common across sub-samples:
H 0 : σ 2 = σ 2 .
Harnau (2018a) showed that we can use a Bartlett test based on the Poisson deviance for this purpose. In the model with the calendar effect, this test yields a p-value of just above 0.05 , a rather close call. In light of the fact that the ultimate goal of the exercise is forecasting reserves and that forecasting often benefits from simpler models, we decide to accept the hypothesis. Next, we consider the hypothesis that there are no breaks in accident, development and calendar effects between sub-samples:
H 0 : α i , + β j , + γ k , + δ = α i + β j + γ k + δ .
As demonstrated by Harnau (2018a), this can be tested with a deviance based F-test that is independent of the Bartlett tests for large τ . This test yields a p-value of 0.07 . Based on the same argument as above, we accept the hypothesis.
Now that we are reasonably happy with the over-dispersed Poisson extended chain-ladder model, we test whether the calendar effects can be dropped.
H 0 : over-dispersed Poisson vs . H A : extended over-dispersed Poisson .
Based on an F-test, this hypothesis cannot be rejected with a p-value of 0.30 . Thus, we move on, retesting whether the over-dispersed Poisson model still encompasses the log-normal model.
H 0 : over-dispersed Poisson vs . H A : generalized log-normal .
Based on a test statistic R l s * = 73.5 , this cannot be rejected with a p-value of 0.73 . We can now go back and apply the misspecification tests by Harnau (2018a) once again, except this time for models without the calendar effect. Using the same sub-sample structure, a Bartlett test cannot reject the hypothesis of common over-dispersion H 0 : σ 2 = σ 2 with a p-value of 0.08 . Further, an F-test for the hypothesis of the absence of breaks in the mean parameters H 0 : α i , + β j , + δ = α i + β j + δ cannot be rejected with a p-value of 0.93 .
In conclusion, an over-dispersed Poisson chain-ladder model for the Taylor and Ashe (1983) data survived a whole battery of specification tests, and we may at least be more comfortable with this model choice, having found no strong evidence telling us otherwise. In contrast, the generalized log-normal model was clearly rejected.

7. Discussion

While there has been a range of recent advances for both over-dispersed Poisson and (generalized) log-normal models, there are still several areas left for further research. This spans from further misspecification tests and refinements thereof over a potential theory for the bootstrap to empirical studies evaluating the impact of the theoretical procedures in practice.
As pointed out by Harnau (2018a), the misspecification tests require a specific choice for the number of sub-samples and their shape. A generalization that is agnostic about these choices would be desirable. Harnau (2018a) also remarked that a misspecification test for independence would be useful. The assumption of independence across cells is common to both over-dispersed Poisson and generalized log-normal models. It seems likely that a test that is valid in one model would translate easily to the other.
The closed-form distribution forecasts proposed by Harnau and Nielsen (2017) for the over-dispersed Poisson model and by Kuang and Nielsen (2018) for the generalized log-normal model are both based on t-distributions and thus symmetric. These forecasts seem to perform rather well and, in some settings, appear more robust than the bootstrap by England and Verrall (1999) and England (2002). However, with an appealing asymptotic theory in place for both types of models, it may be worth considering whether a theory for the bootstrap could be developed to allow for potential asymmetry of the forecast distribution that we might expect in finite samples.
Finally, given the range of recent theoretical developments, an empirical study that evaluates the impact of the contributions in applications seems appropriate. Since the main concern in claim reserving is forecasting, such a study would likely require data not just for run-off triangles, but also for the realized values in the forecast array, that is the lower triangle. Such data are available, for example, from Casualty Actuarial Society (2011). For instance, it would be interesting to see how the forecast performance between rival models differs if one were rejected by the theory, but not the other.

Funding

The author was supported by the European Research Council, Grant AdG 694262.

Acknowledgments

The author thanks the three anonymous referees for their helpful comments. Discussions with Bent Nielsen (Department of Economics, University of Oxford & Nuffield College, U.K.) are gratefully acknowledged.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Appendix A.1. Proof of Corollary 1

The proof is similar to that for the distribution of the mixed parameter in Harnau and Nielsen (2017, Lemma 1). We define the Poisson quasi-likelihood estimator g ( Y ) = ξ ^ q l for Y = ( Y i j : i , j I ) . We make use of the fact that g ( Y ) = ( μ ^ 11 , ξ ^ q l ( 2 ) ) and g ( Y / τ ) = { μ ^ 11 log ( τ ) , ξ ^ q l ( 2 ) } for identical μ ^ 11 and ξ ^ q l ( 2 ) . This follows from the Poisson score equation i j I Y i j x i j = exp ( μ ^ 11 ) i j x i j exp ( x i j ( 2 ) ξ ^ q l ( 2 ) ) in which replacing Y i j by Y i j / τ goes hand in hand with replacing μ ^ 11 by μ ^ 11 log ( τ ) . Thus, only μ ^ 11 is affect by scaling the variables. By Johansen (1979, Theorem 7.1), g ( . ) is Fisher consistent so g { E ( Y ) } = ( μ 11 , ξ ( 2 ) ) and g { E ( Y ) / τ } = ( μ 11 log ( τ ) , ξ ( 2 ) ) . By Johansen (1979, Lemma 7.2), g / Y | Y = { E ( Y ) / τ } = ( X Π X ) 1 X .
By independence of Y i j , (4) generalizes to τ 1 / 2 { Y / τ E ( Y ) / τ } D N ( 0 , σ 2 Π ) . Applying g ( . ) to this result using the δ -method, similar to Johansen (1979, Theorem 7.3), and taking into account that g ( Y / τ ) g { E ( Y ) / τ } = ξ ^ q l ξ yields the desired result. For the δ -method see, for example, Casella and Berger (2002, Theorem 5.5.24); to avoid confusion, we point out that in our notation, the sequence is not over n but rather over τ (for proofs relating to the generalized log-normal model below the sequence is over ω 2 ).

Appendix A.2. Proof of Lemma 1

First, we show that lim τ P ( Y i j = 0 ) = 0 . Recall that Y i j / τ P π i j > 0 . Thus, lim τ P ( Y i j / τ = 0 ) = 0 . Since P ( Y i j = 0 ) = P ( Y i j / τ = 0 ) the results follows. Next, we know from (4) that τ ( Y i j / τ π i j ) D N ( 0 , σ 2 π i j ) . For Y i j > 0 we can employ the δ -method to apply log ( . ) to Y i j / τ . The result follows since log ( Y i j / τ ) log ( π i j ) = log ( Y i j ) log ( τ ) [ log { exp ( μ i j ) } log ( τ ) ] and log ( x ) / x | x = E ( Y i j / τ ) = π i j 1 .

Appendix A.3. Proof of Lemma 2

Define Z = ( Z i j : i , j I ) and μ = ( μ i j : i , j I ) . Taking into account the independence, the multivariate version of Lemma 1 is τ ( Z μ ) D N ( 0 , σ 2 Π 1 ) .
To obtain the distribution of the least-squares estimator, we pre-multiply by ( X X ) 1 X . This yields τ ( ξ ^ l s ξ ) D N ( 0 , σ 2 Ω ) with Ω as defined in the lemma. We find the distribution of the residual sum of squares using the continuous mapping theorem. With that, τ Z M Z = { τ ( Z μ ) } M { τ ( Z μ ) } D U Π 1 / 2 M Π 1 / 2 U for U = D N ( 0 , I ) .
Finally, we show that τ ^ l s / τ P 1 where τ ^ l s = i j I exp ( x i j ξ ^ l s ) . Define f ( ξ ) = i j I exp ( x i j ξ ) . For this map, with ξ = ( μ 11 , ξ ( 2 ) ) and defining ξ τ = ( μ 11 log ( τ ) , ξ ( 2 ) ) , we have f ( ξ τ ) = f ( ξ ) / τ . Further by the equivalent argument made in the Proof of Corollary 1, subtracting log ( τ ) from Z (element-wise) affects only the estimate for the intercept μ 11 . That is, for the least squares estimator g ( Z ) = ( μ ^ 11 l s , ξ ^ l s ( 2 ) ) and with Z τ = { Z i j log ( τ ) : i , j I } , we have g ( Z τ ) = ( μ ^ 11 l s log ( τ ) , ξ ^ l s ( 2 ) ) = : ξ ^ l s τ . Thus, ξ ^ l s τ ξ τ = ξ ^ l s ξ and τ ( ξ ^ l s τ ξ τ ) D N ( 0 , σ 2 Ω ) . Now, we apply f ( . ) by the δ -method to get that τ { f ( ξ ^ l s τ ) f ( ξ τ ) } = O p ( 1 ) . Since f ( ξ ^ l s τ ) f ( ξ τ ) = τ ^ l s / τ 1 it follows that τ ^ l s / τ = 1 + o p ( 1 ) .

Appendix A.4. Proof of Lemma 3

Define the vector Y = ( Y i j : i , j I ) and let exp ( μ ) = { exp ( μ i j ) : i , j I } . Then, using the independence, (4) generalizes to τ 1 / 2 { Y exp ( μ ) } D N ( 0 , σ 2 Π ) .
In Corollary 1, we followed the approach by Harnau and Nielsen (2017, Lemma 1) to derive the asymptotic distribution of the Poisson quasi-likelihood estimator ξ ^ q l through the δ -method. Using Johansen (1979, Theorems 7.1 and 7.3, Lemma 7.2), we showed that the mapping Y ξ ^ q l estimator is asymptotically equivalent to the linear mapping Y ( X Π X ) 1 X Y .
Meanwhile, the weighted least squares estimator maps Z ( X Π X ) 1 X Π 1 / 2 Π 1 / 2 Z = ξ ^ * . Thus, the only non-linear component of the mapping Y ξ ^ * is the transformation from Y to Z. However, while this mapping is non-linear in finite sample, for large τ it is equivalent to the linear map from Y to Π 1 Y as seen in the proof of Lemma 1. Asymptotically, this conforms to sequentially applying the transformations Π 1 followed by ( X Π X ) 1 X Π 1 / 2 Π 1 / 2 to Y. Taken together, the map reduces to ( X Π X ) 1 X . Thus, both the Poisson quasi-likelihood and weighted least squares mapping asymptotically apply the same transformation to τ 1 / 2 { Y exp ( μ ) } D N ( 0 , σ 2 Π ) . Thus, τ ( ξ ^ * ξ ^ q l ) P 0 .
The proof for τ R S S * D P 0 follows by the same argument. The main insight is that the asymptotic distribution of the Poisson deviance is asymptotically equivalent to that of the quadratic form τ 1 { Y exp ( μ ) } ( Π 1 X ( X Π X ) 1 X ) { Y exp ( μ ) } , as Harnau and Nielsen (2017, Proof of Lemma 1) show building on Johansen (1979, Theorems 7.7, 7.8). This is again asymptotically identical to the sequential mapping from Y to Z followed by the map from Z to the scaled residual sum of weighted least squares τ R S S * = { τ Π 1 / 2 ( Z μ ) } M * { τ Π 1 / 2 ( Z μ ) } .
To show that we can replace the weight matrix Π in the weighted least squares estimator by Π ^ l s or Π ^ q l we note that both matrices converge in probability to Π and then apply Slutsky’s theorem Casella and Berger 2002, Theorem 5.5.17). Combining this argument with the proof of the equivalence of D and R S S * in the last paragraph, it also follows that we can replace the weights in R S S * without affecting the result. Finally, both τ ^ l s / τ P 1 and τ ^ q l / τ P 1 so we can replace τ as well by Slutsky’s theorem.

Appendix A.5. Proof of Theorem 1

Taking into account the results from Lemma 3, it follows that ( τ R S S * ) / D P 1 and that the result still holds if we replace R S S * by R S S l s * or R S S q l * and τ by τ ^ l s or τ ^ q l . Thus, for example, R l s P R q l so their difference vanishes and similarly for any other of the six total combinations.
Both R S S = Z M Z and R S S * = Z * M * Z * = Z Π 1 / 2 M * Π 1 / 2 Z are quadratic forms in the same random vector Z. It follows from the proofs of Lemma 2 and Lemma 3 that τ R S S D U Π 1 / 2 M Π 1 / 2 U and τ R S S * D U M * U for the same U = D N ( 0 , I n ) . The distribution of R S S / R S S * follows by the continuous mapping theorem. Since τ R S S * D P 0 as in Lemma 3, τ R S S * / D P 1 so that R S S / R S S * τ R S S / D P 0 follows.
We can replace τ by τ ^ q l since Harnau and Nielsen (2017, Theorem 2) gives us that τ / τ ^ q l P 1 . From Lemma 2, τ ^ l s / τ P 1 . Further, both Π ( ξ ^ l s ( 2 ) ) and Π ( ξ ^ q l ( 2 ) ) converge to Π in probability. Then, by Slutsky’s theorem, we can replace the true parameters with their estimates in τ R S S / D and R S S / R S S * without affection the limiting distribution.

Appendix A.6. Proof of Lemma 4

The asymptotic distribution of the weighted least squares estimators follows by the same argument as in Lemma 2, except now taking ( ω 2 ) 1 / 2 ( Z μ ) D N ( 0 , I n ) , as shown by Kuang and Nielsen (2018, Theorem 3.3), as a starting point.
The asymptotic equivalence of weighted least squares and Poisson quasi likelihood estimation follows from the same argument as in Lemma 3, except now ( ω 2 ) 1 / 2 { Y exp ( μ ) } D N [ 0 , diag { exp ( μ ) } ] . The argument for replacing true parameters in the frequency matrix Π and the aggregate means τ by estimates is identical to that in Lemma 3 as well.

Appendix A.7. Proof of Theorem 2

This follows by the same argument as the proof for Theorem 1 above, except now combining the asymptotic distribution of the least squares estimator in the generalized log-normal model from Kuang and Nielsen (2018, Theorem 3.5) in (8) and Lemma 4.

Appendix A.8. Proof of Lemma 5

First, we show that R G L N and R O D P share a common support. If we recall that:
R G L N ( U ) = U M U U Π 1 / 2 M * Π 1 / 2 U and R O D P ( U ) = U Π 1 / 2 M Π 1 / 2 U U M * U ,
the main insight is that R G L N ( Π 1 / 2 U ) = R O D P ( U ) . Formally, both R G L N : R n R and R O D P : R n R are random variables on { R n , B ( R n ) , P } where P is the measure associated with N ( 0 , I n ) . We now show that P { R G L N S } = 1 implies that P { R O D P S } = 1 ; the opposite direction follows. Notationally, P { R G L N S } = P { u R n : R G L N ( u ) S } and, for some set A R we denote by R G L N 1 ( A ) the pre-image { u R n : R G L N ( u ) A } . Now, R G L N is measurable since it is continuous almost everywhere, the exception being the measure zero set where the denominator is zero. Thus,
P { R G L N S } = P { U R G L N 1 ( S ) } .
Since the support of N ( 0 , I n ) is R n , we must have that R G L N 1 ( S ) = R n and hence S = R G L N ( R n ) . Since Π is invertible, Π 1 / 2 R n = R n so that R G L N ( R n ) = R G L N ( Π 1 / 2 R n ) = R O D P ( R n ) . Thus, R O D P ( R n ) = S and so R G L N 1 ( S ) = R O D P 1 ( S ) . Taken together,
P { R G L N S } = P { U R G L N 1 ( S ) } = P { U R O D P 1 ( S ) } = P { R O D P S } = 1 .
Since this holds in both directions and for any such S , it holds for the support which is a special case of S .
Now that we showed that R G L N and R O D P have identical support, we show that the support is a bounded compact set ( l , r ) . To do so, we specify the support of R G L N . The key insight is that the real symmetric matrices A = M and B = Π 1 / 2 M * Π 1 / 2 commute so A B = B A . Thus, they are simultaneously diagonalizable (Newcomb 1961). Beyond that, both matrices are of rank n p . Thus, we can find an orthogonal matrix of Eigenvectors P such that:
P A P = Λ A 0 0 0 and P B P = Λ B 0 0 0
where Λ A and Λ B are diagonal n p × n p matrices of Eigenvalues. Since M is a projection matrix, Λ A = I . Then, making use of Butler and Paolella (2008, Lemma 3 Case 2(c)), the upper bound of the support of R G L N is given by the largest element of Λ B 1 Λ A = Λ B 1 . Thus, it is finite. To find the lower bound, we consider the upper bound of R G L N , thus swapping A for A . By the same argument, the lower bound is the largest element of Λ B 1 times 1 , thus the smallest element of Λ B 1 . Since B is positive semi-definite and because Λ B contains only the non-zero eigenvalues, this must be larger then zero. Denoting the diagonal elements of Λ B sorted in descending magnitude by λ B = ( λ B , ( 1 ) , λ B , ( n p ) ) , we can write the support as ( l , r ) = ( λ B , ( 1 ) 1 , λ B , ( n p ) 1 ) .

Appendix A.9. Proof of Theorem 3

We assume that rows and columns from the design X and the frequency matrix Π that relate to cells with perfect fit, such as the corners in a run-off triangle, have been removed. In this way, there is no need to keep track of the restriction that not only frequencies relating to such cells can go to zero. Further, we assume without loss of generality that rows and columns of X and Π have been sorted as described in Section 4.5.
First, we establish a useful equivalence of the denominators of both R G L N ( t ) and R O D P ( t ) . We can write these as a quadratic form of the type V ( t ) M ( t ) * V ( t ) where V ( t ) = Π ( t ) 1 / 2 U for R G L N ( t ) and V ( t ) = U for R O D P ( t ) . Now, for any t > 0 ,
V ( t ) M ( t ) * V ( t ) = | M ( t ) * V ( t ) | 2 = min ξ R p | V ( t ) X ( t ) * ξ | 2 .
To understand the last equality, we note that minimizing argument for the last expression is the least squares estimator ξ ^ = ( X ( t ) * X ( t ) * ) 1 X ( t ) * V ( t ) . The expression itself corresponds to the squared length of the least squares residuals. We compute these residuals as M ( t ) * V ( t ) and their squared length is given by | M ( t ) * V ( t ) | 2 .
Now, we derive the small t limit of R G L N ( t ) . We denote the first q and last n q elements of U by U 1 and U 2 , respectively, and similarly for ξ . Then, using the partition of Π ( t ) in (11) and X in (12) where X 12 = 0 , we can write the denominator as:
U Π ( t ) 1 / 2 M * Π ( t ) 1 / 2 U = s ( t ) min ξ R p Π 1 1 / 2 ( U 1 X 11 ξ 1 ) t Π 2 1 / 2 { U 2 ( X 21 ξ 1 + X 22 ξ 2 ) } 2
where s ( t ) = { trace ( Π 1 ) + t · trace ( Π 2 ) } 1 . We note that the norm can only every decrease as t becomes smaller. This is because for t 1 < t 2 , we could still choose the optimal ξ * under t 2 and make the norm smaller by replacing t 2 with t 1 . For t = 0 , the right hand side simplifies to:
min ξ 1 R p 1 | Π ˘ 1 1 / 2 ( U 1 X 11 ξ 1 ) | 2 = U 1 Π ˘ 1 1 / 2 M ˘ 11 * Π ˘ 1 1 / 2 U 1 .
Therefore,
lim t 0 U Π ( t ) 1 / 2 M ( t ) * Π ( t ) 1 / 2 U = U 1 Π ˘ 1 1 / 2 M ˘ 11 * Π ˘ 1 1 / 2 U 1 .
The numerator does not depend on t. Thus, as long as U 0 , which happens with probability one,
lim t 0 R G L N ( t ) = U M U lim t 0 U Π ( t ) 1 / 2 M ( t ) * Π ( t ) 1 / 2 U = U M U U 1 Π ˘ 1 1 / 2 M ˘ 11 * Π ˘ 1 1 / 2 U 1 = R G L N ( 0 ) .
It follows that R G L N ( t ) a . s . R G L N ( 0 ) .
Next, we consider R O D P ( t ) . For the denominator, we have:
U M ( t ) * U = min ξ R p | U X ( t ) * ξ | 2 | U | 2
where the inequality follows since we can always set ξ = 0 . Thus, the denominator is bounded from above. Looking at the numerator U Π ( t ) 1 / 2 M Π ( t ) 1 / 2 U , we see that the limit of Π ( t ) 1 / 2 does not exist. We look at this in more detail. Partition M so M 11 is the q × q top right element and so on. Then we can write the numerator as:
s ( t ) 1 ( U 1 Π 1 1 / 2 M 11 Π 1 1 / 2 U 1 + t 1 U 2 Π 2 1 / 2 M 22 Π 2 1 / 2 U 2 + 2 t 1 / 2 U 1 Π 1 1 / 2 M 12 Π 2 1 / 2 U 2 ) .
The normalization s ( t ) converges to { trace ( Π 1 ) } 1 as t 0 . The first term of the sum is non-negative and does not vary with t. The second term is non-negative and O p ( t 1 ) while the third term is O p ( t 1 / 2 ) with ambiguous sign. Further, dropping cells with perfect fit from X ensures M 22 0 . Thus, since U 2 0 with probability one, the second term is positive with probability one so overall the numerator U Π ( t ) 1 / 2 M Π ( t ) 1 / 2 U a . s . . Thus, since the denominator is bounded, R O D P ( t ) a . s . .
Finally, R O D P ( t ) > q G L N , α ( t ) almost surely since for α ( 0 , 1 ) , the quantile q G L N , α ( t ) a . s q G L N , α ( 0 ) < . Conversely, q O D P , α ( t ) a . s so R G L N ( t ) q O D P , α ( t ) almost surely.
Figure A1. Bar chart of maximum absolute errors for the considered combinations of R and R ^ . Ordered by the sum of errors within combination across data generating processes and parameterizations increasing from top to bottom. Sum of maximum absolute errors across parameterizations indicated by “+”. V N J , T A , and B Z is short for parameters set to their estimates from the Verrall et al. (2010), Taylor and Ashe (1983) and Barnett and Zehnwirth (2000) data, respectively. Based on 10 5 repetitions for each parametrization. s = 1 .
Figure A1. Bar chart of maximum absolute errors for the considered combinations of R and R ^ . Ordered by the sum of errors within combination across data generating processes and parameterizations increasing from top to bottom. Sum of maximum absolute errors across parameterizations indicated by “+”. V N J , T A , and B Z is short for parameters set to their estimates from the Verrall et al. (2010), Taylor and Ashe (1983) and Barnett and Zehnwirth (2000) data, respectively. Based on 10 5 repetitions for each parametrization. s = 1 .
Risks 06 00070 g0a1
Table A1. Insurance run-off triangle taken from Barnett and Zehnwirth (2000, Table 3.5) as used in the empirical application in Section 6.2 and the simulations in Section 5.
Table A1. Insurance run-off triangle taken from Barnett and Zehnwirth (2000, Table 3.5) as used in the empirical application in Section 6.2 and the simulations in Section 5.
i , j 1234567891011
01153,638188,412134,53487,45660,34842,40431,23821,25216,62214,44012,200
02178,536226,412158,894104,68671,44847,99035,57624,81822,66218,000-
03210,172259,168188,388123,07483,38056,08638,49633,76827,400--
04211,448253,482183,370131,04078,99460,23245,56838,000---
05219,810266,304194,650120,09887,58262,75051,000----
06205,654252,746177,506129,52296,78682,400-----
07197,716255,408194,648142,328105,600------
08239,784329,242264,802190,400-------
09326,304471,744375,400--------
10420,778590,400---------
11496,200----------
Table A2. Insurance run-off triangle taken from Taylor and Ashe (1983) as used in the empirical application in Section 6.3 and the simulations in Section 5.
Table A2. Insurance run-off triangle taken from Taylor and Ashe (1983) as used in the empirical application in Section 6.3 and the simulations in Section 5.
i , j 12345678910
01357,848766,940610,542482,940527,326574,398146,342139,950227,22967,948
02352,118884,021933,8941,183,289445,745320,996527,804266,172425,046-
03290,5071,001,799926,2191,016,654750,816146,923495,992280,405--
04310,6081,108,250776,1891,562,400272,482352,053206,286---
05443,160693,190991,983769,488504,851470,639----
06396,132937,085847,498805,037705,960-----
07440,832847,6311,131,3981,063,269------
08359,4801,061,6481,443,370-------
09376,686986,608--------
10344,014---------

References

  1. Barndorff-Nielsen, Ole E. 1978. Information and Exponential Families. Chichester: Wiley. [Google Scholar]
  2. Barnett, Glen, and Ben Zehnwirth. 2000. Best estimates for reserves. Proceedings of the Casualty Actuarial Society LXXXVII: 245–321. [Google Scholar]
  3. Bartlett, Maurice S. 1937. Properties of Sufficiency and Statistical Tests. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 160: 268–82. [Google Scholar] [CrossRef]
  4. Beard, Robert E., Teivo Pentikäinen, and Erkki Pesonen. 1984. Risk Theory: The Stochastic Basis of Insurance, 3rd ed. London and New York: Chapman and Hall. [Google Scholar]
  5. Butler, Ronald W., and Marc S. Paolella. 2008. Uniform saddle point approximations for ratios of quadratic forms. Bernoulli 14: 140–54. [Google Scholar] [CrossRef]
  6. Casella, George, and Berger Roger L. 2002. Statistical Inference, 2nd ed. Pacific Grove: Duxbury/Thomson Learning. [Google Scholar]
  7. Casualty Actuarial Society. 2011. Loss Reserving Data Pulled From Naic Schedule P. Available online: http://www.casact.org/research/index.cfm?fa=loss_reserves_data (accessed on 8 July 2018).
  8. Chow, Gregory C. 1960. Tests of Equality Between Sets of Coefficients in Two Linear Regressions. Econometrica 28: 591–605. [Google Scholar] [CrossRef]
  9. Cox, David R. 1961. Tests of separate families of hypothesis. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability. Berkeley: University of California Press, pp. 105–23. [Google Scholar]
  10. Cox, David R. 1962. Further Results on Tests of Separate Families of Hypotheses. Journal of the Royal Statistical Society, Series B 24: 406–24. [Google Scholar]
  11. Durbin, James, and Geoffrey S. Watson. 1950. Testing for serial correlation in least squares regression. I. Biometrika 37: 409–28. [Google Scholar] [PubMed]
  12. Durbin, James, and Geoffrey S. Watson. 1951. Testing for Serial Correlation in Least Squares Regression. II. Biometrika 38: 159–77. [Google Scholar] [CrossRef] [PubMed]
  13. England, Peter, and Richard Verrall. 1999. Analytic and bootstrap estimates of prediction errors in claims reserving. Insurance: Mathematics and Economics 25: 281–93. [Google Scholar] [CrossRef]
  14. England, Peter D., and Richard J. Verrall. 2002. Stochastic Claims Reserving in General Insurance. British Actuarial Journal 8: 443–518. [Google Scholar] [CrossRef] [Green Version]
  15. England, Peter. 2002. Addendum to “Analytic and bootstrap estimates of prediction errors in claims reserving”. Insurance: Mathematics and Economics 31: 461–66. [Google Scholar] [CrossRef]
  16. Ermini, Luigi, and David F. Hendry. 2008. Log income vs. linear income: An application of the encompassing principle. Oxford Bulletin of Economics and Statistics 70: 807–27. [Google Scholar] [CrossRef]
  17. Harnau, Jonas, and Bent Nielsen. 2017. Over-dispersed age-period-cohort models. Journal of the American Statistical Association. [Google Scholar] [CrossRef]
  18. Harnau, Jonas. 2017. apc. Available online: https://pypi.org/project/apc/ (accessed on 8 July 2018).
  19. Harnau, Jonas. 2018a. Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models. Risks 6: 25. [Google Scholar] [CrossRef]
  20. Harnau, Jonas. 2018b. quad_form_ratio. Available online: https://pypi.org/project/quad-form-ratio/ (accessed on 8 July 2018).
  21. Hendry, David F., and Bent Nielsen. 2007. Econometric Modeling: A Likelihood Approach. Princeton: Princeton University Press. [Google Scholar]
  22. Hendry, David F., and Jean-Francois Richard. 1982. On the formulation of empirical models in dynamic econometrics. Journal of Econometrics 20: 3–33. [Google Scholar] [CrossRef]
  23. Johansen, Søren. 1979. Introduction to the Theory of Regular Exponential Families. Copenhagen: Institute of Mathematical Statistics, University of Copenhagen. [Google Scholar]
  24. Johnson, Norman Lloyd, Samuel Kotz, and N. Balakrishnan. 1995. Continuous Univariate Distributions Volume 1, 2nd ed. Chichester: Wiley. [Google Scholar]
  25. Kremer, Erhard. 1982. IBNR-claims and the two-way model of ANOVA. Scandinavian Actuarial Journal 1982: 47–55. [Google Scholar] [CrossRef]
  26. Kremer, Erhard. 1985. Einführung in die Versicherungsmathematik, 7th ed. Gottingen: Vandenhoeck & Ruprecht. [Google Scholar]
  27. Kuang, Di, and Bent Nielsen. 2018. Generalized Log-Normal Chain-Ladder. arXiv, arXiv:1806.05939. [Google Scholar]
  28. Kuang, Di, Bent Nielsen, and Jens P. Nielsen. 2008. Identification of the age-period-cohort model and the extended chain-ladder model. Biometrika 95: 979–86. [Google Scholar] [CrossRef]
  29. Kuang, Di, Bent Nielsen, and Jens Perch Nielsen. 2008. Forecasting with the age-period-cohort model and the extended chain-ladder model. Biometrika 95: 987–91. [Google Scholar] [CrossRef]
  30. Kuang, Di, Bent Nielsen, and Jens P. Nielsen. 2015. The geometric chain-ladder. Scandinavian Actuarial Journal 2015: 278–300. [Google Scholar] [CrossRef]
  31. Lancaster, Tony. 2000. The incidental parameter problem since 1948. Journal of Econometrics 95: 391–413. [Google Scholar] [CrossRef] [Green Version]
  32. Lee, Young K., Enno Mammen, Jens P. Nielsen, and Byeong U. Park. 2015. Asymptotics for in-sample density forecasting. Annals of Statistics 43: 620–51. [Google Scholar] [CrossRef]
  33. Lieberman, Offer. 1994. Saddle point approximation for the distribution of a ratio of quadratic forms in normal variables. Journal of the American Statistical Association 89: 924–28. [Google Scholar] [CrossRef]
  34. Lugannani, Robert, and Stephen Rice. 1980. Saddle point approximation for the distribution of the sum of independent random variables. Advances in Applied Probability 12: 475–90. [Google Scholar] [CrossRef]
  35. Mack, Thomas. 1993. Distribution free calculation of the standard error of chain ladder reserve estimates. ASTIN Bulletin 23: 213–25. [Google Scholar] [CrossRef]
  36. Martínez Miranda, María Dolores, Bent Nielsen, and Jens Perch Nielsen. 2015. Inference and forecasting in the age-period-cohort model with unknown exposure with an application to mesothelioma mortality. Journal of the Royal Statistical Society: Series A (Statistics in Society) 178: 29–55. [Google Scholar] [CrossRef]
  37. Mizon, Grayham E., and Jean-Francois Richard. 1986. The Encompassing Principle and its Application to Testing Non-Nested Hypotheses. Econometrica 54: 657–78. [Google Scholar] [CrossRef]
  38. Newcomb, Robert W. 1961. On the simultaneous diagonalization of two semi-definite matrices. Quarterly of Applied Mathematics 19: 144–46. [Google Scholar] [CrossRef]
  39. Neyman, Jerzy, and Elizabeth L. Scott. 1948. Consistent estimates based on partially consistent observations. Econometrica 16: 1–32. [Google Scholar] [CrossRef]
  40. Nielsen, Bent, and Jens P. Nielsen. 2014. Identification and forecasting in mortality models. The Scientific World Journal 2014: 347043. [Google Scholar] [CrossRef] [PubMed]
  41. Nielsen, Bent. 2015. apc: An R Package for Age-Period-Cohort Analysis. The R Journal 7: 52–64. [Google Scholar]
  42. R Core Team. 2017. R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing. [Google Scholar]
  43. Taylor, Greg C., and Frank. R. Ashe. 1983. Second moments of estimates of outstanding claims. Journal of Econometrics 23: 37–61. [Google Scholar] [CrossRef]
  44. Thorin, Olof. 1977. On the infinite divisibility of the lognormal distribution. Scandinavian Actuarial Journal 1977: 121–48. [Google Scholar] [CrossRef]
  45. Verrall, Richard, Jens Perch Nielsen, and Anders Hedegaard Jessen. 2010. Prediction of RBNS and IBNR claims using claim amounts and claim counts. ASTIN Bulletin 40: 871–87. [Google Scholar]
  46. Verrall, Richard J. 1991. On the estimation of reserves from loglinear models. Insurance: Mathematics and Economics 10: 75–80. [Google Scholar] [CrossRef]
  47. Verrall, Richard J. 1994. Statistical methods for the chain ladder technique. Casualty Actuarial Society Forum 1: 393–446. [Google Scholar]
  48. Vuong, Quang H. 1989. Likelihood Ratio Tests for Model Selection and Non-Nested Hypotheses. Econometrica 57: 307–33. [Google Scholar] [CrossRef]
Figure 1. Approximation error of the first order saddle point approximation to R G L N , shown in (a), and R O D P , displayed in (b). Monte Carlo simulation with 10 7 draws taken as the truth. One and two Monte Carlo standard errors shaded in blue and green, respectively.
Figure 1. Approximation error of the first order saddle point approximation to R G L N , shown in (a), and R O D P , displayed in (b). Monte Carlo simulation with 10 7 draws taken as the truth. One and two Monte Carlo standard errors shaded in blue and green, respectively.
Risks 06 00070 g001
Figure 2. Bar chart of the area under the curve of absolute approximation errors (also roughly mean absolute error) for the considered combinations of R and R ^ . Ordered by the sum of errors within the combination across the data generating processes and parameterizations increasing from top to bottom. Sum of maximum absolute errors across parameterizations indicated by “+”. V N J , T A and B Z are short for parameters set to their estimates from the Verrall et al. (2010) data in Table 1, the Taylor and Ashe (1983) data in Table A2 and the Barnett and Zehnwirth (2000) data in Table A1, respectively. Based on 10 5 repetitions for each parametrization. s = 1 .
Figure 2. Bar chart of the area under the curve of absolute approximation errors (also roughly mean absolute error) for the considered combinations of R and R ^ . Ordered by the sum of errors within the combination across the data generating processes and parameterizations increasing from top to bottom. Sum of maximum absolute errors across parameterizations indicated by “+”. V N J , T A and B Z are short for parameters set to their estimates from the Verrall et al. (2010) data in Table 1, the Taylor and Ashe (1983) data in Table A2 and the Barnett and Zehnwirth (2000) data in Table A1, respectively. Based on 10 5 repetitions for each parametrization. s = 1 .
Risks 06 00070 g002
Figure 3. Box plots of size error at 5 % critical values over parameterizations ( V N J , T A and B Z ) and data generating processes (generalized log-normal and over-dispersed Poisson). Results for s = 1 shown in (a) and for s = 2 in (b). Medians indicated by blue lines inside the boxes. The boxes show the interquartile range. Whiskers represent the full range.
Figure 3. Box plots of size error at 5 % critical values over parameterizations ( V N J , T A and B Z ) and data generating processes (generalized log-normal and over-dispersed Poisson). Results for s = 1 shown in (a) and for s = 2 in (b). Medians indicated by blue lines inside the boxes. The boxes show the interquartile range. Whiskers represent the full range.
Risks 06 00070 g003
Figure 4. Power as t increases from zero to t m a x . Values for t m a x are 1 . 083 for V N J , 1 . 396 for T A and 1 . 103 for B Z . (a) shows power when the null model is generalized log-normal, and (b) shows the difference in power between the two models.
Figure 4. Power as t increases from zero to t m a x . Values for t m a x are 1 . 083 for V N J , 1 . 396 for T A and 1 . 103 for B Z . (a) shows power when the null model is generalized log-normal, and (b) shows the difference in power between the two models.
Risks 06 00070 g004
Figure 5. Critical values as t increases. (a) shows critical values of R G L N ( t ) and (b) the ratio of critical values of R O D P ( t ) to R G L N ( t ) .
Figure 5. Critical values as t increases. (a) shows critical values of R G L N ( t ) and (b) the ratio of critical values of R O D P ( t ) to R G L N ( t ) .
Risks 06 00070 g005
Figure 6. Split of a run-off triangle into four sub-samples as in Harnau (2018a).
Figure 6. Split of a run-off triangle into four sub-samples as in Harnau (2018a).
Risks 06 00070 g006
Table 1. Run-off triangle taken from Verrall et al. (2010) with an indication for splitting into sub-samples corresponding to the first and last five accident years. Accident years i in the rows; development years j in the columns.
Table 1. Run-off triangle taken from Verrall et al. (2010) with an indication for splitting into sub-samples corresponding to the first and last five accident years. Accident years i in the rows; development years j in the columns.
i , j 12345678910
01451,288339,519333,371144,988093,243045,51125,21720,40631,4821729
02448,627512,882168,467130,674056,044033,39756,07126,52214,346-
03693,574497,737202,272120,753125,046037,15427,60817,864--
04652,043546,406244,474200,896106,802106,75363,688---
05566,082503,970217,838145,181165,519091,313----
06606,606562,543227,374153,551132,743-----
07536,976472,525154,205150,564------
08554,833590,880300,964-------
09537,238701,111--------
10684,944---------
Table 2. Power in % at 5 % critical values for large τ (over-dispersed Poisson DGP) and small ω 2 (generalized log-normal DGP) along with the power gap in pp for the top four performers from Table 3. DGP is short for data generating process. Based on 10 5 repetitions. s = 1 .
Table 2. Power in % at 5 % critical values for large τ (over-dispersed Poisson DGP) and small ω 2 (generalized log-normal DGP) along with the power gap in pp for the top four performers from Table 3. DGP is short for data generating process. Based on 10 5 repetitions. s = 1 .
P ( R GLN c ODP ) P ( R c ODP R ^ ) P ( R GLN c ODP )
H 0 DGP Π ( R l s * , R ^ l s * ) ( R l s * , R ^ l s ) ( R l s , R ^ q l ) ( R l s * , R ^ q l * )
G L N O D P V N J 99.0200.2500.140.1800.22
B Z 94.61−0.94−0.96−0.97−0.94
T A 65.3904.1803.415.2803.15
P ( R ODP > c GLN ) P ( R > c ODP R ^ ) - P ( R ODP > c GLN )
( R l s * , R ^ l s * ) ( R l s * , R ^ l s ) ( R l s , R ^ q l ) ( R l s * , R ^ q l * )
O D P G L N V N J 99.23−0.30−0.25−0.34−0.20
B Z 94.67−1.14−1.15−1.12−1.12
T A 64.7300.8100.1504.6802.49
Table 3. p-Values in % for the Verrall et al. (2010) data.
Table 3. p-Values in % for the Verrall et al. (2010) data.
H 0 : Generalized Log-Normal H 0 : Over-Dispersed Poisson
R ls R ql R ls * R ql * R ls R ql R ls * R ql *
R ^ l s 0 0.43 0 0.39 0 0.14 0 0.27 0 8.53 0 9.00 14.59 10.89
R ^ q l 0 0.32 0 0.29 0 0.10 0 0.19 11.80 12.40 19.35 14.79
R ^ l s * 0 0.35 0 0.32 0 0.11 0 0.22 10.42 10.97 17.34 13.14
R ^ q l * 0 0.38 0 0.34 0 0.13 0 0.24 0 9.48 0 9.99 15.96 12.01

Share and Cite

MDPI and ACS Style

Harnau, J. Log-Normal or Over-Dispersed Poisson? Risks 2018, 6, 70. https://doi.org/10.3390/risks6030070

AMA Style

Harnau J. Log-Normal or Over-Dispersed Poisson? Risks. 2018; 6(3):70. https://doi.org/10.3390/risks6030070

Chicago/Turabian Style

Harnau, Jonas. 2018. "Log-Normal or Over-Dispersed Poisson?" Risks 6, no. 3: 70. https://doi.org/10.3390/risks6030070

APA Style

Harnau, J. (2018). Log-Normal or Over-Dispersed Poisson? Risks, 6(3), 70. https://doi.org/10.3390/risks6030070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop