Next Article in Journal
Use of Derivatives and Market Valuation of the Banking Sector: Evidence from the European Union
Next Article in Special Issue
Modelling of Loan Non-Payments with Count Distributions Arising from Non-Exponential Inter-Arrival Times
Previous Article in Journal
Exchange Rate Volatility Effect on Economic Growth under Different Exchange Rate Regimes: New Evidence from Emerging Countries Using Panel CS-ARDL Model
Previous Article in Special Issue
On Financial Distributions Modelling Methods: Application on Regression Models for Time Series
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Contaminated Weighted Exponential Distribution: Applications to Modeling Insurance Claim Data

by
Abbas Mahdavi
1,
Omid Kharazmi
1 and
Javier E. Contreras-Reyes
2,*
1
Department of Statistics, Vali-e-Asr University of Rafsanjan, Rafsanjan 7718897111, Iran
2
Instituto de Estadística, Facultad de Ciencias, Universidad de Valparaíso, Valparaíso 2360102, Chile
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2022, 15(11), 500; https://doi.org/10.3390/jrfm15110500
Submission received: 21 September 2022 / Revised: 16 October 2022 / Accepted: 25 October 2022 / Published: 27 October 2022
(This article belongs to the Special Issue Financial Data Analytics and Statistical Learning)

Abstract

:
Deriving loss distribution from insurance data is a challenging task, as loss distribution is strongly skewed with heavy tails with some levels of outliers. This paper extends the weighted exponential (WE) family to the contaminated WE (CWE) family, which offers many flexible features, including bimodality and a wide range of skewness and kurtosis. We adopt Expectation-Maximization (EM) and Bayesian approaches to estimate the model, providing the likelihood and the priors for all unknown parameters. Finally, two sets of claims data are analyzed to illustrate the efficiency of the proposed method in detecting outliers.

1. Introduction

In many applied areas, particularly in finance and actuarial sciences, data are usually positive, right-skewed, leptokurtic and multimodal (Cummins et al. 1990). To capture a wide range of population heterogeneity and tail behavior, one practical way is to conduct analyses over subsets of claims with distinct claim characteristics. But the approach falls short of providing a full picture of claim dynamics. Classical distributions are not flexible enough to cater to heavy-tailed datasets due to extreme values that are far from the other observed data points. These unusual observations are usually called outliers. The presence of outliers in the data may distort both the estimated model parameters and the model’s goodness-of-fit. Recently, many authors have focused on a finite mixture approach that shares the efficiency of parametric modeling and the flexibility of non-parametric density estimation techniques. The flexibility of finite mixtures is accommodating various shapes of insurance and economic data (Bernardi et al. 2012; Hennig and Liao 2013; Maruotti et al. 2016; Punzo et al. 2018).
Okhli and Nooghabi (2021) introduced the contaminated exponential (CE) distribution as an alternative platform for analyzing positive-valued insurance datasets with some level of outliers. The pdf of CE distribution with scale parameter λ and contamination factor θ is defined as follows:
f C E ( y ; λ , θ , ω ) = ( 1 ω ) λ e λ y + ω λ θ e λ θ y , y > 0 , λ > 0 ,
where ω ( 0 , 1 ) is the proportion of contaminated points. The Bayesian approach is developed for computing the parameter estimates. It is demonstrated that the effect of outliers is automatically reflected in the posterior distribution for any sample size. This way, an outlier observation has the highest posterior probability of outlying, but the main observations have a relatively small such probability, indicating that the CE model can detect outliers well.
Weighted distributions are used to adjust the probabilities of events as observed and recorded (Chung and Kim 2004; Gupta and Kirmani 1990; Larose and Dey 1996); (Navarro et al. 2006). Patil (1991) proceeded from applications involving statistical ecology to generate and review many useful general results concerning weighted distributions. Mild outliers, on which this paper focuses, can be dealt with by using heavy-tailed distributions for data. Weighted distributions offer the flexibility needed for achieving mild outlier robustness, while the usual distributions like exponential, gamma and Weibull models lack sufficient fit. For more information and applications of weighted distributions see Patil and Rao (1977).
A two-parameter weighted exponential (WE) distribution (Gupta and Kundu 2009) was developed as a lifetime model which has been widely used in engineering, medicine and insurance. The sensitive skewness parameter governs essentially the shape of the probability density function (pdf) of the WE distribution. A random variable Y is said to have a weighted exponential distribution with a shape parameter α > 0 , and scale parameter λ > 0 , denoted by W E ( α , λ ) , if its pdf is given by
f W E ( y ; α , λ ) = 1 + 1 α λ e λ y ( 1 e α λ y ) , y > 0 .
In this paper, we introduce a class of contaminated weighted exponential (CWE) distributions to account for all possible features of insurance and economic data. Crucially, the CWE model is a two-component mixture in which one component, with a large prior probability, represents the reference distribution, and another, with small prior probability and inflated variability, represents the degree of contamination. For Bayesian inference, we consider several asymmetric and symmetric loss functions like squared error loss, modified squared error, precautionary, weighted squared error, linear exponential, general entropy, and K-loss functions to estimate the parameters of the CWE model. Further, using the independent prior distributions, Bayesian 95 % credible and highest posterior density (HPD) intervals (see Chen et al. 1999) are provided for each parameter of the proposed model.
The paper is organized as follows. Section 2 presents the CWE model and some illustrations of the density, skewness and kurtosis. In Section 3 and Section 4, the EM algorithm and Bayesian inference are respectively developed for CWE parameters. Section 5 illustrates several simulations of proposed estimation methods of Section 3 and Section 4. Section 6 and Section 7 illustrates numerical examples for insurance data fitting using proposed estimation methods of Section 3 and Section 4, respectively. Finally, discussions and conclusions are presented in Section 8.

2. The CWE Model

The pdf of a CWE model with contamination factor θ can be written as
f C W E ( y ; α , λ , θ , ω ) = ( 1 ω ) f W E ( y ; α , λ ) + ω f W E ( y ; α , λ θ ) ,
where θ > 0 and ω [ 0 , 1 ] denotes the proportion of outliers or unusual points and Θ = ( ω , α , λ , θ ) contains all model parameters. The CE model given in (1) is obtained as a special case of (3) when α . The effect of varying each parameter when one varies, but keeping others fixed, is illustrated by a set of CWE densities shown in Figure 1. The plots show that the distribution is more likely to be bimodal as ω increases, whereas flatness parameter vector α controls tail behavior. This implies that the CWE model provides a component of the WE distribution to capture the vast majority of small losses, whereas the contaminated component accommodates clusters of larger losses with an enhanced tail to capture extreme losses. Furthermore, the skewness and kurtosis 3D plots of the CWE model for numerous values of α and θ with fixed λ = 1 are depicted in the Figure 2. The fitting of this four-parameter CWE model via the likelihood approach is difficult because of the log-likelihood function’s complexity. But the EM and Bayesian approaches can help.

3. Maximum Likelihood Estimation via EM Algorithm

The EM algorithm (Dempster et al. 1977) and some of its extraordinary variants such as the expectation conditional maximization (ECM) algorithm (Meng and Rubin 1993) and the expectation-conditional maximization either (ECME) algorithm (Liu and Rubin 1994) are broadly applicable methods to carry out ML estimation for mixture distributions and variety of incomplete-data problems (Aitkin and Wilson 1980; McLachlan and Krishnan 2007; Redner and Walker 1984). Mahdavi et al. (2021a, 2021b) and Cavieres et al. (2022) developed novel EM-based procedures designed under the selection mechanism to compute the ML estimates of scale-shape mixtures of flexible generalized skew-normal and multivariate flexible skew-symmetric-normal distributions. Here, we develop a novel EM-based procedure designed under the selection mechanism to compute the ML estimates of the proposed model.
A random variable Y W E ( α , λ ) is said to follow WE distribution with shape parameter α and scale parameter λ if it has the following stochastic selection representation:
Y = d X 0 | U < 1 ,
where U = X 1 / ( α X 0 ) and X 0 and X 1 are two independent exponential random variables with mean 1 λ . To perform an EM-type algorithm for fitting the CWE model, we introduce a latent variable τ = U | U < 1 based on (4). The joint pdf of ( Y , τ ) is given by
f Y , τ ( y , τ ) = 1 P ( U < 1 ) f X 0 , U ( y , τ ) = 1 + 1 α f X 0 ( y ) f U | X 0 ( τ ) = ( α + 1 ) λ 2 y e λ y e λ α τ y , y > 0 , 0 < τ < 1 .
Dividing (5) by (2) yields
f τ | Y ( τ ) = α λ y e α λ y τ 1 e α λ y , 0 < τ < 1 .
Using (6), it is clear that
τ | Y = y T E x p α λ y ; ( 0 , 1 ) ,
where T E x p λ ; ( 0 , b ) represents the truncated exponential distribution with mean 1 λ on interval ( 0 , b ) .
Let us introduce an n-dimensional binary random variable γ = ( γ 1 , , γ n ) where a particular element γ i is equal to 1 if Y i belongs to unusual observations and is equal to zero otherwise. Note that, γ i follows a Bernoulli random variable with success probability ω denoted by γ i B e r ( ω ) .
Now, consider n independent random variables Y 1 , , Y n , which are taken from a mixture model (3) and latent variable τ = ( τ 1 , , τ n ) , where Θ = ( ω , α , λ , θ ) denotes the unknown vector of parameters. Clearly,
Y i | ( γ i = 0 ) W E ( α , λ ) and Y i | ( γ i = 1 ) W E ( α , λ θ ) , τ i | ( Y i = y i , γ i = 0 ) T E x p ( α λ y i ; ( 0 , 1 ) ) , τ i | ( Y i = y i , γ i = 1 ) T E x p ( α λ θ y i ; ( 0 , 1 ) ) .
According to (3) and (5), it is clear that
f Y i , τ i | γ i ( y i , τ i ) = ( α + 1 ) λ 2 y i e λ y i e i λ α τ i y i 1 γ i ( α + 1 ) λ 2 θ 2 y i e λ θ y i e λ θ α τ i y i γ i .
The complete log-likelihood function of Θ given y c = ( y , τ , , γ ) is
c ( Θ | y c ) = ln f Y , γ , τ ( y , γ , τ ) = ln f γ ( γ ) f Y , τ | γ ( y , τ ) = i = 1 n { γ i ln ω + ( 1 γ i ) ln ( 1 ω ) + ln ( α + 1 ) + 2 ln λ + 2 γ i ln θ ( 1 γ i ) λ y i ( 1 γ i ) λ α τ i y i γ i λ θ y i γ i λ θ α τ i y i } .
To evaluate the Q-function, the necessary conditional expectations include
γ ^ i ( k ) = E γ i | Y i = y i , Θ ^ ( k ) = ω ^ ( k ) f W E ( y i ; α ^ ( k ) , λ ^ ( k ) Θ ^ ( k ) ) f C W E ( y i ; α ^ ( k ) , λ ^ ( k ) , Θ ^ ( k ) ) , τ ^ 1 i ( k ) = E ( 1 γ i ) τ i | Y i = y i , Θ ^ ( k ) = ( 1 γ ^ i ( k ) ) 1 α ^ ( k ) λ ^ ( k ) y i 1 e α ^ ( k ) λ ^ ( k ) y i 1 , τ ^ 2 i ( k ) = E γ i τ i | Y i = y i , Θ ^ ( k ) = γ ^ i ( k ) 1 α ^ ( k ) λ ^ ( k ) θ ^ ( k ) y i 1 e α ^ ( k ) λ ^ θ ^ ( k ) y i 1 .
Therefore, the Q-function is given by
Q ( Θ | Θ ^ ( k ) ) = i = 1 n { γ ^ i ( k ) ln ω + ( 1 γ ^ i ( k ) ) ln ( 1 ω ) + ln ( α ^ ( k ) + 1 ) + 2 ln λ ^ ( k ) + 2 γ ^ i ( k ) ln θ ^ ( k ) λ ^ ( k ) ( 1 γ ^ i ( k ) ) y i λ ^ ( k ) α ^ ( k ) τ ^ 1 i ( k ) y i λ ^ ( k ) θ ^ ( k ) γ ^ i ( k ) y i λ ^ ( k ) θ ^ ( k ) α ^ ( k ) τ ^ 2 i ( k ) y i } .
In summary, the implementation of the ECM algorithm proceeds as follows:
E-step: Given Θ = Θ ^ ( k ) , compute γ ^ i ( k ) , τ ^ 1 i ( k ) and τ ^ 2 i ( k ) for i = 1 , , n .
CM-step 1: Calculate
ω ^ ( k + 1 ) = 1 n i = 1 n γ ^ i ( k ) .
CM-step 2: Fix λ = λ ^ ( k ) , θ = θ ^ ( k ) and update α ^ ( k ) by maximizing (9) over α , which gives
α ^ ( k + 1 ) = n λ ^ ( k ) i = 1 n τ ^ 1 i ( k ) y i θ ^ ( k ) τ ^ 2 i ( k ) y i 1 .
CM-step 3: Fix α = α ^ ( k + 1 ) , θ = θ ^ ( k ) and update λ ^ ( k ) by
λ ^ ( k + 1 ) = 2 n i = 1 n ( 1 γ ^ ( k ) ) y i + α ^ ( k + 1 ) τ ^ 1 i ( k ) y i + θ ^ ( k ) γ ^ i ( k ) y i + α ^ ( k + 1 ) θ ^ ( k ) τ ^ 2 i y i .
CM-step 4: Fix α = α ^ ( k + 1 ) , λ = λ ^ ( k + 1 ) and update θ ^ ( k ) by
θ ^ ( k + 1 ) = 2 i = 1 n γ ^ i ( k ) λ ^ ( k + 1 ) i = 1 n { γ ^ i ( k ) y i + α ^ ( k + 1 ) τ ^ 2 i ( k ) y i } .
This process is repeated until a suitable convergence rule is satisfied. The convergence appears when the relative difference between two successive log-likelihood values is less than tolerance ( ϵ ). In our numerical experiments, ϵ = 10 6 is used. An R code about EM algorithm is available in Appendix A.

4. Bayesian Inference

In this section, we discuss the Bayesian estimation for the CWE distribution parameters in terms of several symmetric and asymmetric loss functions such as squared error loss function (SELF), weighted squared error loss function (WSELF), modified squared error loss function (MSELF), precautionary loss function (PLF) and K-loss function (KLF). The considered loss functions and their Bayesian estimators with corresponding posterior risks are reported in Table 1.
For pertinent details about these loss functions, refer to Kharazmi et al. (2021, 2022) and references therein.

4.1. Joint and Marginal Posterior Distributions

Assume that the parameters of the CWE distribution have independent prior distributions as follows: α G a m m a ( α 0 , α 1 ) , θ G a m m a ( θ 0 , θ 1 ) , λ G a m m a ( λ 0 , λ 1 ) , and ω B e t a ( ω 0 , ω 1 ) , where all hyper-parameters are positive. Consequently, the joint prior density is formulated as
π ( α , λ , θ , ω ) = ω ω 0 ( 1 ω ) ω 1 α 1 α 0 θ 1 θ 0 λ 1 λ 0 B e t a ( ω 0 , ω 1 ) Γ ( α 0 ) Γ ( θ 0 ) Γ ( λ 0 ) α α 0 1 θ θ 0 1 λ λ 0 e ( α 1 α + θ 1 θ + λ 1 λ ) .
For simplicity, we define function ζ as
ζ ( α , θ , λ , ω ) = α α 0 1 β β 0 1 λ λ 0 e ( α 1 α + β 1 β + λ 1 λ ) ω ω 0 ( 1 ω ) ω 1 .
From (10) and likelihood function L ( d a t a ) , the joint posterior distribution is
π ( α , θ , λ , ω | d a t a ) π ( α , θ , λ , ω ) L ( d a t a ) .
Therefore, the exact joint posterior pdf is given by
π ( α , θ , λ , ω | x ̲ ) = K ζ ( α , θ , λ , ω ) L ( x ̲ , Ψ ) ,
where
L ( x ̲ ; Ψ ) = λ 1 + 1 α n i = 1 n ( 1 ω ) e λ x i ( 1 e α λ x i ) + ω θ e θ λ x i ( 1 e α θ λ x i ) ,
Ψ = ( α , θ , λ , ω ) and K is a normalizing constant with form
K 1 = 0 1 0 0 0 ζ ( α , θ , λ , ω ) L ( x ̲ , ξ ) α β λ ω .
Moreover, the marginal posterior density of α , θ , λ and ω (assuming Ψ = ( Ψ 1 , Ψ 2 , Ψ 3 , Ψ 4 ) = ( α , θ , λ , ω ) ) can be expressed as
π ( Ψ i | x ̲ ) = 0 1 0 0 π ( Ψ | x ̲ ) Ψ j Ψ k Ψ 4 , i = 1 , 2 , 3 , 0 0 0 π ( Ψ | x ̲ ) Ψ 1 Ψ 2 Ψ 3 , i = 4 ,
where j , k = 1 , 2 , 3 , j k i and Ψ i is the ith member of vector Ψ .

4.2. Bayesian Point Estimation

From the marginal posterior pdf in (12) and under framework of the loss functions listed in Table 1, the Bayesian point estimation for parameter vector Ψ = ( Ψ 1 , Ψ 2 , Ψ 3 , Ψ 4 ) = ( α , θ , λ , ω ) is formulated via minimizing the expectation of loss function with respect to the marginal posterior pdf in (12) as follows:
argmin C δ 0 L ( Ψ i , δ ) π ( Ψ i | x ̲ ) Ψ i .
In practice, because of the intractable integral in (13), we can use the Gibbs sampler (Geman and Geman 1984) or Metropolis-Hastings algorithms (Hastings 1970; Metropolis et al. 1953) to generate posterior samples. We will argue this issue more precisely in Section 4.5.

4.3. Credibility Interval

In the Bayesian framework, interval estimation is done via credibility interval conception. Consider parameter vector Ψ = ( Ψ 1 , Ψ 2 , Ψ 3 , Ψ 4 ) = ( α , θ , λ , ω ) , which is associated with CWE distribution and π ( Ψ j | x ̲ ) the marginal posterior pdf of parameter Ψ j , j = 1 , 2 , 3 , 4 , as in (12). For a given value of η ( 0 , 1 ) , the ( 1 η ) 100 % credibility interval C I ( L Ψ j , U Ψ j ) is defined as
L Ψ j π ( Ψ j | x ̲ ) Ψ j = 1 η 2 ,
U Ψ j π ( Ψ j | x ̲ ) Ψ j = η 2 .
By considering relation (14) and (15), it is not feasible to obtain the explicit marginal pdf from the joint posterior distribution. To overcome this difficulty, we use the Gibbs sampler algorithm and generate posterior samples from the CWE distribution. Let Ψ 1 , , Ψ k (where Ψ i = ( Ψ 1 i , Ψ 2 i , Ψ 3 i , Ψ 4 i ) ) be a posterior random sample of size k which is extracted from the joint posterior pdf in (10). Using these samples, the marginal posterior pdf of Ψ j given x ̲ is defined by
1 K i = 1 K π ( Ψ j , Ψ j i | x ̲ ) , j = 1 , 2 , 3 , 4 ,
where Ψ j i represents the vector of posterior samples when the j th component is removed. Inserting (16) in (15), it is possible to compute the credibility intervals for Ψ j , j = 1 , 2 , 3 , 4 , as follows
1 K i = 1 K L Ψ j π Ψ j , Ψ j i | x ̲ Ψ j = 1 η 2 ,
1 K i = 1 K U Ψ j π Ψ j , Ψ j i | x ̲ Ψ j = η 2 .

4.4. Highest Posterior Density Interval

Highest posterior density (HPD) interval is a credibility interval under a specific restriction. A ( 1 η ) 100 % HPD interval for Ψ j , j = 1 , 2 , 3 , 4 is the simultaneous solution of integral equations
1 K i = 1 K L Ψ j U Ψ j π Ψ j , Ψ j i | x ̲ Ψ j = 1 η ,
i = 1 K π L Ψ j , Ψ j i | x ̲ = i = 1 K π U Ψ j , Ψ j i | x ̲ .

4.5. Generating Posterior Samples

It is clear from Equations (10) and (12) that there are no explicit expressions for the Bayesian point estimators under the loss functions in Table 1. Because of intractable integrals associated with joint posterior and marginal posterior distributions, we require numerical software to solve the integral equations numerically via MCMC methods such as the Metropolis-Hastings algorithm and Gibbs sampling (Contreras-Reyes et al. 2018). Assuming general model f ( x ̲ | ψ ) is associated with parameter vector ψ = ( ψ 1 , ψ 2 , , ψ p ) and observed data x ̲ , the joint posterior distribution is π ( ψ 1 , ψ 2 , , ψ p | x ̲ ) . We also assume that ψ 0 = ( ψ 1 ( 0 ) , ψ 2 ( 0 ) , , ψ p ( 0 ) ) is the initial vector to start the Gibbs sampler (Quintero et al. 2017). The steps for any iteration, say iteration k, are as follows:
  • Starting with an initial estimate ( ψ 1 ( 0 ) , ψ 2 ( 0 ) , , ψ p ( 0 ) ) ;
  • draw ψ 1 k from π ψ 1 | ψ 2 k 1 , ψ 3 k 1 , , ψ p k 1 , x ̲ ;
  • draw ψ 2 k from π ψ 2 | ψ 1 k , ψ 3 k 1 , , ψ p k 1 , x ̲ ; and so on down to
  • draw ψ p k from π ψ p | ψ 1 k , ψ 2 k , , ψ p 1 k , x ̲ .
In the case of the CWE distribution, by considering parameter vector Ψ = ( α , θ , λ , ω ) and initial parameter vector Ψ 0 = ( α 0 , θ 0 , λ 0 , ω 0 ) , the posterior samples are extracted based on Gibbs sampler where the full conditional distributions are
π α | θ k 1 , λ k 1 , ω k 1 , x ̲ α + 1 α n α α 0 e α 1 α i = 1 n Υ ( x i , Ψ ) ,
π θ | α k 1 , λ k 1 , ω k 1 , x ̲ β θ 0 e θ 1 θ i = 1 n Υ ( x i , Ψ ) ,
π λ | α k 1 , θ k 1 , ω k 1 , x ̲ λ λ 0 + n e λ 1 λ i = 1 n Υ ( x i , Ψ ) ,
and
π ω | α k 1 , θ k 1 , λ k 1 , x ̲ ω ω 0 ( 1 ω ) ω 1 i = 1 n Υ ( x i , Ψ ) ,
where Υ ( x i , Ψ ) = ( 1 ω ) e λ x i ( 1 e α λ x i ) + ω θ e θ λ x i ( 1 e α θ λ x i ) .
In practice, simulations related to Gibbs sampling can be done with special software WinBUGS. This software was developed in 1997 to simulate data of complex posterior distributions, where analytical or numerical integration techniques cannot be applied. Moreover, Gibbs sampling processes can be carried out via OpenBUGS software, which is an open source version of WinBUGS. Since there isn’t any prior information about hyper-parameters in (10), we follow Congdon (2001) and the hyper-parameter values are set as α i = θ i = λ i = ω i = 0.0001 , i = 0 , 1 , so we can use the MCMC procedure to extract posterior samples of (10) by means of Gibbs sampling process in OpenBUGS software.

5. Simulation Study: Recovery of the True Underlying Parameters

An experiment intends to investigate the ability of the proposed EM algorithm to recover the true underlying parameters. We generate 5000 synthetic Monte Carlo samples of different sample sizes n = 30 , 70, 100 and 200 from the CWE distribution and following three parameter scenarios (each scenario corresponding to density plotted as “dotdash” line in Figure 1):
Scenario 1:  α = 0.5 , λ = 1 , θ = 2 , ω = 0.2 .
Scenario 2: α = 2 , λ = 1 , θ = 0.3 , ω = 0.2 .
Scenario 3: α = 0.5 , λ = 1 , θ = 0.3 , ω = 0.6 .
The accuracies of the parameter estimates are measured by computing the mean absolute bias (MAB) and the root mean square error (RMSE), defined as
M A B = 1 5000 i = 1 5000 | θ ^ i θ A | and R M S E = 1 5000 i = 1 5000 ( θ ^ i θ A ) 2 ,
where θ ^ i denotes the prediction of a specific parameter at the i-th replication and θ A denotes the actual specific parameter value. Table 2 shows the simulation results for the CWE distribution. As expected, the MAB and RMSE tend toward zero when the sample size increases, showing empirically the consistency of the ML estimates obtained via the EM algorithm.

6. Numerical Examples for Insurance Data Fitting

In this section, we evaluate the performance and various aspects of the proposed model using insurance claims data. The proposed distribution is fitted to the data by implementing the ECM algorithm described in Section 3. For the sake of comparison, the reduced WE, CE and exponential (Exp) models are also fitted as sub-models of CWE distribution. To compare how well the models fit the data, we adopt the Akaike information criterion (AIC) (Akaike 1973) and the Bayesian information criterion (BIC) (Schwarz 1978), defined as AIC = 2 p 2 m a x and BIC = p log n 2 m a x , where p is the number of free parameters in the model and m a x the maximized log-likelihood value. For both AIC and BIC, a smaller value indicates a better model fit.
The first dataset (DS1) comprises Danish fire losses analyzed in McNeil (1997). This dataset is frequently used for comparison of methods; see Eling (2012) and references therein. These data represent Danish fire losses in million Danish Krones and were collected by a Danish reinsurance company. The dataset contains individual losses above 1 million Danish Krones, a total of 2167 individual losses, covering the period from 3 January 1980 to 31 December 1990. Data are adjusted for inflation to reflect 1985 values and are available in R packages evir and fExtremes.
The second dataset (DS2), analyzed by Cummins and Freifelder (1978), contains 80 fire losses from 500 buildings a large university owned from 1951 to 1973. Cummins et al. (1990) found that the log-normal and gamma distributions did not have sufficient heavy tails to model the data, so they considered the generalized beta of the second kind (GB2) distribution.
Figure 3 presents two histograms for the considered datasets. Both histograms reveal a typical feature of insurance claims data: a large number of small losses and a small number of very large losses. Table 3 reports parameter estimates, standard error and model fit criteria for all fitted models. Observing the Table 3, it is evident from the AIC and BIC values that the CWE model provides better fit than other fitted models. The posterior probability of each observation belonging to unusual observations is depicted in Figure 4 and Figure 5, those reveal that the unusual data have the highest posterior probability and the original data have small posterior probability, showing clearly the impact of outliers.

7. Bayesian Numerical Results

We used an MCMC procedure based on 10,000 replicates with 1000 samples discarded as burn-in to compute the Bayesian estimators. The corresponding Bayesian point estimation and posterior risk based on DS1 and DS2 datasets are provided in Table 4. It can be seen that for the both datasets, the resulting log-likelihood values ( m a x ) are close to the obtained ones by the EM-algorithm given in Table 3, indicating the efficiency of the Bayesian approach to estimate the model parameters. It is noteworthy to mention that the KLF and PLF loss functions yields the highest log-likelihood values for DS1 and DS2 datasets, respectively.
Table 5 provides 95% credible and HPD intervals for the parameters of the CWE distribution. The posterior samples are extracted using Gibbs sampling technique. Moreover, we provide the posterior summary plots in Figure 6, Figure 7 and Figure 8. These plots confirm that the convergence of the Gibbs sampling process occurred.
In order to avoid repetition in evaluation of the MCMC procedure in Bayesian analysis, we just reported the Gelman–Rubin and Geweke–Raftery–Lewis diagnostics measures for checking the convergence based on data set DS1 in Table 6. For more details on these indexes see Lee et al. (2014). The Gelman–Rubin diagnostic is equal to 1 for parameters λ , θ , α and ω . Hence, the chains could be accepted, and this indicates the estimates come from a state space of the parameter, as depicted in Figure 9.
From Table 6, Geweke–Raftery–Lewis test statistics for parameters λ , θ , α and ω are 0.588 , 0.320 , 0.761 and 0.368 , respectively. Therefore, also in this case, the chain is acceptable, as shown in Figure 10 and Figure 11. Moreover, the reported diagnostics statistics for parameters α , β and λ based on the Geweke–Raftery–Lewis measure don’t show significant correlations between estimates. Hence, the estimated values have good mixing.

8. Conclusions

This paper extended the WE distribution to a richer family, the CWE distribution, to deal with data displaying large and positive skewness as well as a wide right tail. This four-parameter model is a mixture of two WE distributions in which one has an enhanced scale and hence a thicker tail to capture extreme losses. EM and Bayesian computational techniques were used to estimate parameters. The effectiveness and efficiency of the EM algorithm were evaluated by conducting one simulation study. By analyzing two real insurance claims datasets, we found that the CWE distribution outperformed the CE distribution in terms of model fit. The result show that both EM and Bayesian approaches are appropriate tools to estimate the model parameters. In addition, it is possible to consider proposed distribution to fit lifetimes, and how the suggested algorithms will be adjusted in case of truncated or censored data. Another application could be done in actuarial science context; specifically, how CWE distribution could be employed to calculate the VaR and TVaR (Barges et al. 2009).

Author Contributions

Conceptualization, A.M. and O.K.; methodology, A.M. and O.K.; software, A.M. and O.K.; validation, A.M., O.K. and J.E.C.-R.; formal analysis, O.K.; investigation, A.M., O.K. and J.E.C.-R.; resources, J.E.C.-R.; data curation, A.M.; writing—original draft preparation, A.M., O.K. and J.E.C.-R.; writing–review and editing, A.M., O.K. and J.E.C.-R.; visualization, A.M. and O.K.; supervision, J.E.C.-R.; project administration, O.K.; funding acquisition, J.E.C.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was fully supported by FONDECYT (Chile) grant No. 11190116.

Data Availability Statement

The datasets analyzed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

The authors also thank the editor and two anonymous referees for their helpful comments and suggestions. All R and OpenBUGS codes used in this paper are available upon request from the corresponding author. The datasets analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare that there is no conflict of interest in the publication of this paper.

Appendix A. R Code to Fit the CWE Distribution Using EM-Algorithm

EM.CWE <- function(y, om, al, la, th, iter.max = 500, tol=10^-6){
f.CWE <- function(y,om,al,la,th)
(1-om)*(al+1)/al*la*exp(-la*y)*(1-exp(-la*al*y))+om*(al+1)/al*th*la
*exp(-th*la*y)*(1-exp(-th*la*al*y))
n <- length(y); LL <- 1 ; dif <- 1 ; count <- 1
while ((dif > tol) & (count <= iter.max)) {
# E steps
gam <-  om*(al+1)/al*th*la*exp(-th*la*y)*(1-exp(-th*la*al*y))/
f.CWE(y,om,al,la,th)
ta1 <- (1-gam)*(1/(la*al*y)-1/(exp(la*al*y)-1) )
ta2 <- gam*(1/(th*la*al*y)-1/(exp(th*la*al*y)-1) )
# M steps
om <- sum(gam)/n
al <- n/(la*sum(ta1*y+th*ta2*y))-1
la <- 2*n/sum((1-gam)*y+al*ta1*y+th*gam*y+al*th*ta2*y)
th <- (2*sum(gam))/(la*sum(gam*y+al*ta2*y))
LL.new <- sum(log(f.CWE(y,om,al,la,th)))
count <- count +1
dif <- abs(LL.new/LL-1)
}
print.foo <- function(x) print(x[1:8])
aic <- -2 * LL.new + 2 * 4
bic <- -2 * LL.new + log(n) * 4
Ret <-list(omega=om, alpha=al, lambda=la, theta=th, loglik=LL.new,
AIC=aic, BIC=bic, iter=count, out.prob=gam)
class(Ret) <- "foo"
return(Ret)
}

References

  1. Aitkin, Murray, and Granville Tunnicliffe Wilson. 1980. Mixture models, outliers, and the em algorithm. Technometrics 22: 325–31. [Google Scholar] [CrossRef]
  2. Akaike, Hirotogu. 1973. Information theory as an extension of the maximum likelihood principle. In Second International Symposium on Information Theory. Budapest: BNPBF Csaki Budapest, Academiai Kiado, Hungary. [Google Scholar]
  3. Bargès, Mathieu, Hélène Cossette, and Etienne Marceau. 2009. TVaR-based capital allocation with copulas. Insurance: Mathematics and Economics 45: 348–61. [Google Scholar] [CrossRef] [Green Version]
  4. Bernardi, Mauro, Antonello Maruotti, and Lea Petrella. 2012. Skew mixture models for loss distributions: A bayesian approach. Insurance: Mathematics and Economics 51: 617–23. [Google Scholar] [CrossRef] [Green Version]
  5. Cavieres, Joaquin, German Ibacache-Pulgar, and Javier E. Contreras-Reyes. 2022. Thin plate spline model under skew-normal random errors: Estimation and diagnostic analysis for spatial data. Journal of Statistical Computation and Simulation. in press. [Google Scholar] [CrossRef]
  6. Chen, Ming-Hui, Joseph G. Ibrahim, and Debajyoti Sinha. 1999. A new bayesian model for survival data with a surviving fraction. Journal of the American Statistical Association 94: 909–19. [Google Scholar] [CrossRef]
  7. Chung, Younshik, and Chansoo Kim. 2004. Measuring robustness for weighted distributions: Bayesian perspective. Statistical Papers 45: 15–31. [Google Scholar] [CrossRef]
  8. Congdon, Peter. 2001. Bayesian Statistical Modelling. New York: John Wiley & Sons. [Google Scholar]
  9. Contreras-Reyes, Javier E., Freddy O. López Quintero, and Rodrigo Wiff. 2018. Bayesian modeling of individual growth variability using back-calculation: Application to pink cusk-eel (Genypterus blacodes) off Chile. Ecological Modelling 385: 145–53. [Google Scholar] [CrossRef]
  10. Cummins, J. David, Georges Dionne, James B. McDonald, and B. Michael Pritchett. 1990. Applications of the GB2 family of distributions in modeling insurance loss processes. Insurance: Mathematics and Economics 9: 257–72. [Google Scholar] [CrossRef] [Green Version]
  11. Cummins, J. David, and Leonard R. Freifelder. 1978. A comparative analysis of alternative maximum probable yearly aggregate loss estimators. Journal of Risk and Insurance 45: 27–52. [Google Scholar] [CrossRef]
  12. Dempster, Arthur P., Nan M. Laird, and Donald B. Rubin. 1977. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society: Series B (Methodological) 39: 1–22. [Google Scholar]
  13. Eling, Martin. 2012. Fitting insurance claims to skewed distributions: Are the skew-normal and skew-student good models? Insurance: Mathematics and Economics 51: 239–48. [Google Scholar] [CrossRef]
  14. Geman, Stuart, and Donald Geman. 1984. Stochastic relaxation, gibbs distributions, and the bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence 6: 721–41. [Google Scholar] [CrossRef] [PubMed]
  15. Gupta, Ramesh C., and S. N. U. A. Kirmani. 1990. The role of weighted distributions in stochastic modeling. Communications in Statistics-Theory and Methods 19: 3147–62. [Google Scholar] [CrossRef]
  16. Gupta, Rameshwar D., and Debasis Kundu. 2009. A new class of weighted exponential distributions. Statistics 43: 621–34. [Google Scholar] [CrossRef]
  17. Hastings, W. Keith. 1970. Monte carlo sampling methods using markov chains and their applications. Biometrika 57: 97–109. [Google Scholar] [CrossRef]
  18. Hennig, Christian, and Tim F. Liao. 2013. How to find an appropriate clustering for mixed-type variables with application to socioeconomic stratification. Journal of the Royal Statistical Society: Series C (Applied Statistics) 62: 309–69. [Google Scholar]
  19. Kharazmi, Omid, Ali Saadati Nik, Behrang Chaboki, and Gauss M. Cordeiro. 2021. A novel method to generating two-sided class of probability distributions. Applied Mathematical Modelling 95: 106–24. [Google Scholar] [CrossRef]
  20. Kharazmi, Omid, G. G. Hamedani, and Gauss M. Cordeiro. 2022. Log-mean distribution: Applications to medical data, survival regression, bayesian and non-bayesian discussion with MCMC algorithm. Journal of Applied Statistics, 1–26, in press. [Google Scholar] [CrossRef]
  21. Larose, Daniel T., and Dipak K. Dey. 1996. Weighted distributions viewed in the context of model selection: a bayesian perspective. Test 5: 227–46. [Google Scholar] [CrossRef]
  22. Lee, Cheol-Eung, Sang Ug Kim, and Sangho Lee. 2014. Time-dependent reliability analysis using bayesian MCMC on the reduction of reservoir storage by sedimentation. Stochastic Environmental Research and Risk Assessment 28: 639–54. [Google Scholar] [CrossRef]
  23. Liu, Chuanhai, and Donald B. Rubin. 1994. The ECME algorithm: A simple extension of EM and ECM with faster monotone convergence. Biometrika 81: 633–48. [Google Scholar] [CrossRef]
  24. Mahdavi, Abbas, Vahid Amirzadeh, Ahad Jamalizadeh, and Tsung-I. Lin. 2021a. Maximum likelihood estimation for scale-shape mixtures of flexible generalized skew normal distributions via selection representation. Computational Statistics 36: 2201–30. [Google Scholar] [CrossRef]
  25. Mahdavi, Abbas, Vahid Amirzadeh, Ahad Jamalizadeh, and Tsung-I. Lin. 2021b. A multivariate flexible skew-symmetric-normal distribution: Scale-shape mixtures and parameter estimation via selection representation. Symmetry 13: 1343. [Google Scholar] [CrossRef]
  26. Maruotti, Antonello, Valentina Raponi, and Francesco Lagona. 2016. Handling endogeneity and nonnegativity in correlated random effects models: Evidence from ambulatory expenditure. Biometrical Journal 58: 280–302. [Google Scholar] [CrossRef] [PubMed]
  27. McLachlan, Geoffrey J., and Thriyambakam Krishnan. 2007. The EM Algorithm and Extensions. New York: John Wiley & Sons, vol. 382. [Google Scholar]
  28. McNeil, Alexander J. 1997. Estimating the tails of loss severity distributions using extreme value theory. ASTIN Bulletin: The Journal of the IAA 27: 117–37. [Google Scholar] [CrossRef] [Green Version]
  29. Meng, Xiao-Li, and Donald B. Rubin. 1993. Maximum likelihood estimation via the ECM algorithm: A general framework. Biometrika 80: 267–78. [Google Scholar] [CrossRef]
  30. Metropolis, Nicholas, Arianna W. Rosenbluth, Marshall N. Rosenbluth, Augusta H. Teller, and Edward Teller. 1953. Equation of state calculations by fast computing machines. The Journal of Chemical Physics 21: 1087–92. [Google Scholar] [CrossRef] [Green Version]
  31. Navarro, Jorge, Jose M. Ruiz, and Yolanda Del Aguila. 2006. Multivariate weighted distributions: A review and some extensions. Statistics 40: 51–64. [Google Scholar] [CrossRef]
  32. Okhli, Kheirolah, and Mehdi Jabbari Nooghabi. 2021. On the contaminated exponential distribution: A theoretical bayesian approach for modeling positive-valued insurance claim data with outliers. Applied Mathematics and Computation 392: 125712. [Google Scholar] [CrossRef]
  33. Patil, Ganapati P, and C. R. Rao. 1977. The Weighted Distributions: A Survey and Their Applications. Applications of Statistics. Amsterdam: North Holland, vol. 383, p. 405. [Google Scholar]
  34. Patil, Ganapati P. 1991. Encountered data, statistical ecology, environmental statistics, and weighted distribution methods. Environmetrics 2: 377–423. [Google Scholar] [CrossRef]
  35. Punzo, Antonio, Angelo Mazza, and Antonello Maruotti. 2018. Fitting insurance and economic data with outliers: A flexible approach based on finite mixtures of contaminated gamma distributions. Journal of Applied Statistics 45: 2563–84. [Google Scholar] [CrossRef]
  36. Lopez Quintero, Freddy Omar, Javier E. Contreras-Reyes, Rodrigo Wiff, and Reinaldo B. Arellano-Valle. 2017. Flexible bayesian analysis of the von Bertalanffy growth function with the use of a log-skew-t distribution. Fishery Bulletin 115: 13–26. [Google Scholar] [CrossRef]
  37. Redner, Richard A., and Homer F. Walker. 1984. Mixture densities, maximum likelihood and the EM algorithm. SIAM Review 26: 195–239. [Google Scholar] [CrossRef]
  38. Schwarz, Gideon. 1978. Estimating the dimension of a model. The Annals of Statistics 6: 461–64. [Google Scholar] [CrossRef]
Figure 1. Density plots for different CWE distributions.
Figure 1. Density plots for different CWE distributions.
Jrfm 15 00500 g001
Figure 2. 3D plots of skewness and kurtosis of CWE distribution for two fixed values of ω .
Figure 2. 3D plots of skewness and kurtosis of CWE distribution for two fixed values of ω .
Jrfm 15 00500 g002
Figure 3. Data histograms corresponding to DS1 and DS2 datasetes.
Figure 3. Data histograms corresponding to DS1 and DS2 datasetes.
Jrfm 15 00500 g003
Figure 4. Posterior probability that each observation is unusual, corresponding to DS1 dataset. (Left) panel is for the first 2060 observations and (right) panel for the 107 last observations.
Figure 4. Posterior probability that each observation is unusual, corresponding to DS1 dataset. (Left) panel is for the first 2060 observations and (right) panel for the 107 last observations.
Jrfm 15 00500 g004
Figure 5. Posterior probability that each observation is unusual, corresponding to the DS2 dataset. Left panel is for the first 74 observations and right panel for the six last observations.
Figure 5. Posterior probability that each observation is unusual, corresponding to the DS2 dataset. Left panel is for the first 74 observations and right panel for the six last observations.
Jrfm 15 00500 g005
Figure 6. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Trace plots of each CWE distribution parameter.
Figure 6. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Trace plots of each CWE distribution parameter.
Jrfm 15 00500 g006
Figure 7. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Autocorrelation plots of each CWE distribution parameter.
Figure 7. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Autocorrelation plots of each CWE distribution parameter.
Jrfm 15 00500 g007
Figure 8. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Histogram plots of each CWE distribution parameter.
Figure 8. Plots of Bayesian analysis and performance of Gibbs sampling for DS1 dataset. Histogram plots of each CWE distribution parameter.
Jrfm 15 00500 g008
Figure 9. Gelman plot diagnostic for each CWE distribution parameter based on DS1 dataset.
Figure 9. Gelman plot diagnostic for each CWE distribution parameter based on DS1 dataset.
Jrfm 15 00500 g009
Figure 10. Geweke plot diagnostic (chain1) for each CWE distribution parameter based on DS1 dataset.
Figure 10. Geweke plot diagnostic (chain1) for each CWE distribution parameter based on DS1 dataset.
Jrfm 15 00500 g010
Figure 11. Geweke plot diagnostic (chain2) for each CWE distribution parameter based on DS1 dataset.
Figure 11. Geweke plot diagnostic (chain2) for each CWE distribution parameter based on DS1 dataset.
Jrfm 15 00500 g011
Table 1. Bayes estimator and posterior risk under several loss functions.
Table 1. Bayes estimator and posterior risk under several loss functions.
Loss Function  L ( ψ , δ ) Bayes Estimator ψ B Posterior Risk ρ ψ
S E L F = ( ψ d ) 2 E ( ψ | x ) V a r ( ψ | x )
W S E L F = ( ψ d ) 2 ψ ( E ( ψ 1 | x ) ) 1 E ( ψ | x ) ( E ( ψ 1 | x ) ) 1
M S E L F = 1 d ψ 2 E ( ψ 1 | x ) E ( ψ 2 | x ) 1 E ( ψ 1 | x ) 2 E ( ψ 2 | x )
P L F = ( ψ d ) 2 d E ( ψ 2 | x ) 2 E ( ψ 2 | x ) E ( ψ | x )
K L F = d ψ ψ d E ( ψ | x ) E ( ψ 1 | x ) 2 E ( ψ | x ) E ( ψ 1 | x ) 1
Table 2. Simulation results, based on 5000 replications, to evaluate the EM algorithm under three scenarios.
Table 2. Simulation results, based on 5000 replications, to evaluate the EM algorithm under three scenarios.
Sample Size n = 30 n = 70 n = 100 n = 200
ParameterMABRMSEMABRMSEMABRMSEMABRMSE
Scenario 1 α 0.3570.4190.2700.3290.2320.2840.1710.213
λ 0.2040.2590.1390.1760.1180.1500.0830.104
θ 1.9067.1581.1762.9430.9551.8250.6451.003
ω 0.0940.1480.0730.1090.0650.0930.0500.069
Scenario 2 α 1.7492.8921.1651.6260.9591.2860.6970.898
λ 0.3080.4190.1970.2590.1630.2130.1150.148
θ 0.1890.2820.1020.1510.0800.1150.0520.069
ω 0.1180.1580.0980.1260.0880.1120.0690.087
Scenario 3 α 0.4490.6650.3590.5080.3060.4120.2360.310
λ 0.3660.5640.2290.3230.1870.2520.1320.172
θ 0.0920.1210.0590.0750.0500.0640.0360.045
ω 0.1690.2040.1260.1550.1090.1350.0820.102
Table 3. Summary results from fitting various models to the data. The bold entries highlight the smallest AIC and BIC values for each model.
Table 3. Summary results from fitting various models to the data. The bold entries highlight the smallest AIC and BIC values for each model.
DatasetModel λ ^ θ ^ α ^ ω ^ p max AICBIC
DS1Exp0.2951−4809.3969620.7929626.474
WE0.3504.4202−4576.3279160.6559183.379
CE0.4010.1070.0433−4556.6469119.2929136.335
CWE0.8180.1130.1940.0644−4119.4758246.9508269.675
DS2Exp 0.590 × 10 5 1−859.04141720.0831722.465
WE 0.596 × 10 4 103.6672−858.5481725.0961734.624
CE 0.223 × 10 3 0.0330.0963−796.8151599.6301606.776
CWE 0.258 × 10 3 0.0359.1120.1044−793.0871594.1751603.703
Table 4. Bayesian estimates and their posterior risks of the CWE distribution parameters under different loss functions based on DS1 and DS2 datasets. The bold entries highlight the highest m a x values for each model.
Table 4. Bayesian estimates and their posterior risks of the CWE distribution parameters under different loss functions based on DS1 and DS2 datasets. The bold entries highlight the highest m a x values for each model.
DataDS1
Bayesian Estimation
Loss Function λ ^ ( r λ ^ ) θ ^ ( r θ ^ ) α ^ ( r α ^ ) ω ^ ( r ω ^ ) max
SELF0.74215 (0.00119)0.10790 (0.00006)0.48874 (0.02041)0.06219 (0.00004)−4120.942
WSELF0.74054 (0.00160)0.10734 (0.00056)0.44265 (0.04609)0.06150 (0.00068)−4120.876
MSELF0.73894 (0.00216)0.10677 (0.00527)0.38906 (0.12106)0.06081 (0.01121)−4121.926
PLF0.74296 (0.00161)0.10818 (0.00056)0.50920 (0.04090)0.06253 (0.00067)−4121.245
KLF0.74135 (0.00217)0.10762 (0.00525)0.46513 (0.10154)0.06184 (0.01109)−4120.797
DataDS2
Bayesian Estimation
Loss Function λ ^ ( r λ ^ ) θ ^ ( r θ ^ ) α ^ ( r α ^ ) ω ^ ( r ω ^ ) max
SELF0.000275 ( 1.889 × 10 9 )0.0366 ( 3.6 × 10 5 )6.90240 (1.0208)0.1038 (0.0019)−793.234
WSELF0.000268 ( 6.411 × 10 6 )0.0356 (0.0010)6.75207 (0.1503)0.0824 (0.0214)−793.4759
MSELF0.000262 (0.022395)0.0345 (0.0309)6.60050 (0.0224)0.0605 (0.2651)−794.104
PLF0.000278 ( 6.821 × 10 6 )0.0371 (0.0009)6.97590 (0.1471)0.1128 (0.0179)−793.208
KLF0.000271 (0.023714)0.0361 (0.0285)6.82680 (0.0221)0.0925 (0.2448)−809.881
Table 5. Credible and HPD intervals of parameters λ , θ , α and ω for DS1 and DS2 datasets.
Table 5. Credible and HPD intervals of parameters λ , θ , α and ω for DS1 and DS2 datasets.
DataDS1
Credible IntervalHPD Interval
λ (0.7184, 0.7645)(0.6740, 0.8108)
θ (0.1025, 0.1132)(0.09299, 0.12300)
α (0.3901, 0.5777)(0.2336, 0.7923)
ω (0.05767, 0.06658)(0.04907, 0.07438)
DataDS2
Credible IntervalHPD Interval
λ (0.00024, 0.00030)( 0.00019, 0.00035)
θ (0.03269, 0.04018)(0.02322, 0.04958)
α (6.16500, 7.68300)(5.04200, 8.76500)
ω (0.07135, 0.12980)(0.03289,0.19750)
Table 6. Diagnostics using the Gelman-Rubin and Geweke-Raftery-Lewis methods for parameters α , β and λ based on DS1 dataset.
Table 6. Diagnostics using the Gelman-Rubin and Geweke-Raftery-Lewis methods for parameters α , β and λ based on DS1 dataset.
ParameterGelman-RubinGeweke ( Z 0.025 = ± 1.96 )Raftery-Lewis
λ 1−0.58805.1
θ 10.32054.8
α 10.76075.01
ω 10.36794.632
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mahdavi, A.; Kharazmi, O.; Contreras-Reyes, J.E. On the Contaminated Weighted Exponential Distribution: Applications to Modeling Insurance Claim Data. J. Risk Financial Manag. 2022, 15, 500. https://doi.org/10.3390/jrfm15110500

AMA Style

Mahdavi A, Kharazmi O, Contreras-Reyes JE. On the Contaminated Weighted Exponential Distribution: Applications to Modeling Insurance Claim Data. Journal of Risk and Financial Management. 2022; 15(11):500. https://doi.org/10.3390/jrfm15110500

Chicago/Turabian Style

Mahdavi, Abbas, Omid Kharazmi, and Javier E. Contreras-Reyes. 2022. "On the Contaminated Weighted Exponential Distribution: Applications to Modeling Insurance Claim Data" Journal of Risk and Financial Management 15, no. 11: 500. https://doi.org/10.3390/jrfm15110500

Article Metrics

Back to TopTop