You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

13 February 2023

A New Reciprocal Weibull Extension for Modeling Extreme Values with Risk Analysis under Insurance Data

,
,
,
and
1
Department of Statistics, Mathematics and Insurance, Benha University, Benha 13518, Egypt
2
Department of Statistics and Operations Research, Faculty of Science, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
3
Department of Mathematical Sciences, Ball State University, Muncie, IN 47306, USA
4
Department of Applied, Mathematical and Actuarial Statistics, Faculty of Commerce, Damietta University, Damietta 34517, Egypt
This article belongs to the Special Issue New Advances and Applications of Extreme Value Theory

Abstract

Probability-based distributions might be able to explain risk exposure well. Usually, one number, or at the very least, a limited number of numbers called the key risk indicators (KRIs), are used to describe the level of risk exposure. These risk exposure values, which are undeniably the outcome of a specific model, are frequently referred to as essential critical risk indicators. Five key risk indicators, including value-at-risk, tail variance, tail-value-at-risk, and tail mean-variance, were also used for describing the risk exposure under the reinsurance revenues data. These measurements were created for the proposed model; hence, this paper presents a novel distribution for this purpose. Relevant statistical properties are derived, including the generating function, ordinary moments, and incomplete moments. Special attention is devoted to the applicability of the new model under extreme data sets. Three applications to real data show the usefulness and adaptability of the proposed model. The new model proved its superiority against many well-known related models. Five key risk indicators are employed for analyzing the risk level under the reinsurance revenues dataset. An application is provided along with its relevant numerical analysis and panels. Some useful results are identified and highlighted.
MSC:
62N02; 62E10; 60E05; 62N01; 62G05; 62N05; 62P30

1. Introduction and Motivation

Actuarial risks refer to the financial risks associated with insurance and pension plans. Statistical analysis is an important tool for actuaries to understand and quantify these risks. Actuaries use probability distributions to model the likelihood of various events, such as claims, deaths, and policy cancellations. Common distributions used in actuarial science include the Poisson distribution, the exponential distribution, and the Weibull distribution. Survival analysis is used to model the time until a particular event occurs, such as death or policy cancellation. This technique is used to estimate the probability of survival for a given time period and to calculate life expectancy. Stochastic modeling is used to model random processes, such as claims and policy cancellations. This technique is used to estimate the expected value of future claims and to determine the variability of these estimates. Loss distributions are used to model the distribution of losses due to events, such as claims and policy cancellations. This technique is used to calculate the expected value of future losses and to determine the risk associated with these losses. Actuaries use statistical techniques used to assess and manage financial risks, including methods for hedging and portfolio optimization.
In conclusion, statistical analysis is a critical tool for actuaries to understand and manage financial risks in insurance and pension plans. Actuaries use a variety of statistical techniques, including probability distributions, survival analysis, stochastic modeling, loss distributions, and risk management, to estimate and manage financial risks and ensure the long-term financial stability of insurance and pension plans.
Because they depict situations where extreme occurrences or outliers occur far more frequently than expected by traditional or lighter-tailed distributions, such as the normal distribution, heavy-tailed probability distributions are significant. Numerous real-world systems, including financial markets, communication networks, and natural disasters, may be significantly impacted by these occurrences. To effectively reflect the risk and uncertainty associated with such severe events and to make decisions based on the probabilities of these events, it is crucial to comprehend and model heavy-tailed distributions.
In the actuarial and statistical literature, the issue of measuring the risk resulting from some accidents is considered a matter of great importance. This is because determining the level of risk and measuring its degree will entail a lot of strategic actions and decisions. One of the most important fields in which we, as statisticians, are interested in measuring risk is the field of insurance and reinsurance and the field of actuarial sciences in general. Insurance companies, for example, are very interested in anticipating the volume of claims that they may be bound by soon. Statistical prediction of the volume of claims will make the insurance company stable for a certain period of time. Thus, if the expectations are correct, the company will be in a safe position with the insured parties. Additionally, if it is not correct, then the company will be in a safer position in most cases.
Based on the importance of this issue, whether in terms of actuarial modeling or in terms of statistical forecasts, we are excited about this work as it presents a flexible probabilistic model to help insurance and reinsurance companies determine the volume of claims that may occur in the near future. The new probability model is a probability distribution that presents many characteristics that contributed to its nomination to this actuarial task.
The issue of measuring risk has attracted the attention of many researchers in recent decades. The researchers’ interests focused on presenting and generalizing some actuarial indicators that may contribute to measuring risk under certain conditions. These indicators varied in their forms according to the statistical basis on which they were built. Some depend on the quantile function; some depend on moments; some depend on the tail of the probability distribution, and so on.
Heavy-tailed probability-based distributions refer to probability distributions that have “heavier” tails than the normal distribution, meaning that they have a higher probability of observing extreme values (outliers). This is in contrast to light-tailed distributions, such as the normal distribution, where the probability of observing extreme values is low. Examples of heavy-tailed distributions include; the Pareto distribution, which is used to model the distribution of wealth and income and is characterized by a long tail on the right side of the distribution; the Student’s t-distribution, which is used in hypothesis testing and has heavier tails than the normal distribution, making it more suitable for modeling data with outliers; the Lévy distribution, which is used in financial mathematics and physics and has heavier tails than the normal distribution; and the Cauchy distribution, which has infinite variance and is used to model phenomena with long tails, such as response times in networks and earthquakes.
Heavy-tailed distributions are important in many fields, including finance, economics, physics, and engineering, as they provide a more accurate representation of real-world phenomena where extreme events are more common than expected under a normal distribution. Heavy tailed probability-based distributions might be capable of explaining risk exposure well. Normally, one number, or at the very least a limited number of numbers, is used to describe the level of risk exposure. The term “crucial KRIs” refers to these risk exposure figures, which are undeniably the output of a certain model (see Artzner [1] and Wirch [2]). Such KRIs give actuaries and risk managers knowledge about the level of a company’s exposure to particular risks. There are many KRIs that can be considered and studied, including value-at-risk (VAR), tail-value-at-risk (TVAR), conditional-value-at-risk (CVAR), tail variance (TV), and tail Mean-Variance (TMV), among others. The VAR is specifically included in the quantile distribution of aggregate losses. Actuaries and risk managers usually concentrate on calculating the chance of a bad outcome, which can be measured using the VAR indicator at a particular probability/confidence level. This indicator is typically used to estimate how much money will be required to handle such potentially unfavorable events. The ability of the insurance firm to handle such events is a worry for actuaries, policymakers, investors, and rating agencies.
For this purpose, a novel probability-based reciprocal Weibull distribution called the generalized geometric Rayleigh reciprocal Weibull (GGR-RW) is provided for an adequate explanation of risk exposure under the reinsurance revenues data set. The probability density function (PDF) and cumulative distribution function (CDF) of the reciprocal Weibull (RW) distribution are given by
g δ 1 , δ 2 ( 𝓎 ) = δ 1 δ 2 δ 2 𝓎 ( δ 2 + 1 ) e x p ( δ 1 δ 2 𝓎 δ 2 )   | 𝓎 0 ,
and
G δ 1 , δ 2 ( 𝓎 ) = e x p ( δ 1 δ 2 𝓎 δ 2 )   | 𝓎 0 ,
where δ 2 > 0 is a shape parameter and δ 1 > 0 is a scale parameter. For δ 2 = 2 , we obtain the reciprocal Rayleigh (RR) model. For δ 2 = 1 , we obtain the reciprocal exponential (RE) model, and for δ 1 = 1 , we obtain the one parameter reciprocal Weibull (1PRW) model. Let g δ 1 , δ 2 ( 𝓎 ) and G δ 1 , δ 2 ( 𝓎 ) denote the PDF and CDF of the RW model with parameters δ 1 and δ 2 and consider the CDF of the generalized Rayleigh (GR) family, then the CDF of the generalized Rayleigh reciprocal Weibull (GR-RW) model
H b , δ 1 , δ 2 ( 𝓎 ) = 1 e x p [ b , δ 1 , δ 2 2 ( 𝓎 ) ] | 𝓎 0 ; b > 0 ,
where
b , δ 1 , δ 2 ( 𝓎 ) = e x p ( b δ 1 𝓎 δ 2 ) 1 e x p ( b δ 1 𝓎 δ 2 ) | 𝓎 0 , b > 0 ,
and 𝓀 b , δ 1 , δ 2 ( 𝓎 ) = d H b , δ 1 , δ 2 ( 𝓎 ) / d 𝓎 is the PDF corresponding to (1). For any arbitrary baseline RV having CDF ,   and V _ representing the baseline parameters vector, then the CDF of the geometric family is defined by
F a , V _ ( 𝓎 ) = a H V _ ( 𝓎 ) 1 ( 1 a ) H V _ ( 𝓎 ) | 𝓎 R ; a > 0 .
Then, the new model is derived by combining (1) and (2). The CDF of the GGR-RW model can be defined by
F ϕ _ ( 𝓎 ) = a a e x p [ b , δ 1 , δ 2 2 ( 𝓎 ) ] 1 ( 1 a ) { 1 e x p [ b , δ 1 , δ 2 2 ( 𝓎 ) ] } | 𝓎 0 ; a , b , δ 1 , δ 2 > 0 ,
where ϕ _ = ( a , b , δ 1 , δ 2 ) refers to the parameter vector. Then, the PDF corresponding to (3) is
f ϕ _ ( 𝓎 ) = 2 a b δ 1 δ 2 δ 2 𝓎 ( δ 2 + 1 ) e x p ( 2 δ 1 δ 2 b 𝓎 δ 2 ) [ 1 e x p ( b δ 1 δ 2 𝓎 δ 2 ) ] 3 e x p [ b , δ 1 , δ 2 2 ( 𝓎 ) ] ( 1 ( 1 a ) { 1 e x p [ b , δ 1 , δ 2 2 ( 𝓎 ) ] } ) 2 | 𝓎 0 , a > 0 , b > 0 .
Using the generalized binomial expansion and the power series, the PDF in (4) can be expressed as
f ϕ _ ( 𝓎 ) = 2 a b δ 1 δ 2 δ 2 𝓎 ( δ 2 + 1 ) e x p ( δ 1 δ 2 𝓎 δ 2 ) 𝒾 , 𝒿 , = 0 ( 1 a ) 𝒾 ( 1 ) ! ( 1 + 𝒾 ) e x p { [ 2 b ( + 1 ) 1 ] δ 1 δ 2 𝓎 δ 2 } [ 1 e x p ( b δ 1 δ 2 𝓎 δ 2 ) ] 3 + 2 ( 2 𝒾 ) ( 𝒾 𝒿 )
Using Taylor expansion, we can write
f ϕ _ ( 𝓎 ) = , 𝓂 = 0 c , 𝓂 g b [ 2 ( + 1 ) + 𝓂 ] , δ 1 , δ 2 ( 𝓎 ) | b [ 2 ( + 1 ) + 𝓂 ] > 0 ,
where
c , 𝓂 = 2 a b 𝒾 = 0 𝒾 = 0 𝒿 ( 1 a ) 𝒾 ( 1 ) ( 1 + 𝒾 ) ! b [ 2 ( + 1 ) + 𝓂 ] ( 2 𝒾 ) ( 𝒾 𝒿 ) ( 3 2 𝓂 ) ,
and
G b [ 2 ( + 1 ) + 𝓂 ] , δ 1 , δ 2 ( 𝓎 ) = b [ 2 ( + 1 ) + 𝓂 ] δ 1 δ 2 δ 2 𝓎 ( δ 2 + 1 ) e x p ( b [ 2 ( + 1 ) + 𝓂 ] δ 1 δ 2 𝓎 δ 2 )
refers to the RW density with scale parameter δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) and shape parameter δ 2 . Thus, some mathematical properties of the GGR-RW model can be obtained simply from those properties of the RW density. The CDF of the GGR-RW model can also be expressed as a mixture of RW densities. By integrating (5), we obtain the same mixture representation
F ϕ _ ( 𝓎 ) = , 𝓂 = 0 c , 𝓂 G b [ 2 ( + 1 ) + 𝓂 ] , δ 1 , δ 2 ( 𝓎 ) | b [ 2 ( + 1 ) + 𝓂 ] > 0 ,
where G b [ 2 ( + 1 ) + 𝓂 ] , δ 1 , δ 2 ( 𝓎 ) = e x p ( b [ 2 ( + 1 ) + 𝓂 ] δ 1 δ 2 𝓎 δ 2 ) is the CDF of the RW model with scale parameter δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) and shape parameter δ 2 . In fact, the GGR-RW model is motivated by its important flexibility in applications. We demonstrate that the GGR-RW model offers better fits than many other competing models using three applications. The “asymmetric monotonically increasing hazard rate function (HRF)” shown in Figure 1 suggests that the GGR-RW model could be beneficial in real-life modeling data (first row, right panels).
The GGR-RW model can be recommended for modeling real data which have some extreme values. Also, the GGR-RW model could be useful for analyzing and modeling real data which has no extreme observations, as shown in Figure 7 (the second row of the right and the left panels). Moreover, the GGR-RW distribution can be considered a useful model for modeling real data for which its nonparametric Kernel density is symmetric and unimodal. The GGR-RW could be a good choice for dealing with the real data which its nonparametric Kernel density is the asymmetric bimodal and heavy tail, as illustrated in Figure 1, Figure 4 (the first row, left panels). Finally, the GGR-RW model may be useful in modeling the real data which cannot be fitted by the common theoretical distributions, such as normal, uniform, exponential, logistic, beta, lognormal, and Weibull distributions as illustrated in Figure 1.
The rest of the paper is organized as follows. We derive a few mathematical features for the new model in Section 2. Some risk indicators are discussed in Section 3. We introduce the maximum likelihood technique in Section 4. We present three applications to actual data in Section 5 to demonstrate the adaptability of the new family. Section 6 provides the risk analysis for the reinsurance revenues data. Finally, Section 6 addresses a few closing thoughts.

2. Properties

The 𝓇 t h moment of Y , say μ 𝓇 , Y , follows from (5) as μ 𝓇 , Y = E ( Y 𝓇 ) = , 𝓂 = 0 c , 𝓂 E ( Y { b [ 2 ( + 1 ) + 𝓂 ] } 𝓇 ) .
Henceforth, Y { b [ 2 ( + 1 ) + 𝓂 ] } denotes the RW distribution with scale parameter δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) and shape parameter δ 2 . Then, we have
E ( Y { b [ 2 ( + 1 ) + 𝓂 ] } 𝓇 ) = δ 1 δ 2 δ 2 { b [ 2 ( + 1 ) + 𝓂 ] } 0 𝓎 ( δ 2 + 1 ) + 𝓇 e x p ( { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 𝓎 δ 2 ) d 𝓎 ,
Then,
μ 𝓇 , Y = E ( Y 𝓇 ) = Γ ( 1 𝓇 δ 2 ) , 𝓂 = 0 c , 𝓂 ( δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) 𝓇 ,
the numerical value of the constant , 𝓂 = 0 c , 𝓂 ( δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) 𝓇 can be evaluated using any software such as R and MATHCAD among others. The moment generating function (MGF) M Y ( 𝓉 ) = E ( e x p ( 𝓉 Y ) ) of Y can be derived from Equation (5) as
M Y ( 𝓉 ) = , 𝓂 = 0 c , 𝓂 M { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ( 𝓉 ) ,
where M { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ( 𝓉 ) is the MGF of Y { b [ 2 ( + 1 ) + 𝓂 ] } . Hence,
M Y ( 𝓉 ) = Γ ( 1 𝓇 δ 2 ) , 𝓂 , 𝓇 = 0 c , 𝓂 𝓇 ! ( 𝓉 δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) 𝓇 .
The s t h incomplete moment, say ϕ s , Y ( 𝓉 ) , of Y can be expressed from (5) as
ϕ s , Y ( 𝓉 ) = 𝓉 𝓎 s f ϕ _ ( 𝓎 ) d 𝓎 = , 𝓂 = 0 c , 𝓂 I 𝓉 ( 𝓎 s ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) ,
where I 𝓉 ( 𝓎 s ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = 0 𝓉 𝓎 s g { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ( 𝓎 ) d 𝓎 . Then,
ϕ s , Y ( 𝓉 ) = , 𝓂 = 0 c , 𝓂 ( δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) 𝓇 γ ( 1 𝓇 δ 2 , { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 𝓉 δ 2 ) .
The n t h moment of the residual life of Y is given by
𝓂 n , Y ( 𝓉 ) = 1 1 F ϕ _ ( 𝓉 ) , 𝓂 = 0 𝓇 = 0 n c , 𝓂 ( n 𝓇 ) ( 𝓉 ) n 𝓇 I 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) ,
where
I 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = 𝓉 𝓎 n g { b [ 2 ( + 1 ) + 𝓂 ] } ( 𝓎 ) d 𝓎
then,
I 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = ( δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) n Γ ( 1 n δ 2 , { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 𝓉 δ 2 ) .
The n 𝓉 𝓀 moment of the reversed residual life, say M n , Y ( 𝓉 ) = E [ ( 𝓉 Y ) n | Y 𝓉 ] for 𝓉 > 0 and n = 1 ,   2 ,…, follows as M n , Y ( 𝓉 ) = 1 F ϕ _ ( 𝓉 ) 0 𝓉 ( 𝓉 𝓎 ) n d F ( 𝓎 ) . Therefore, the n th moment of the reversed residual life of Y becomes
M n , Y ( 𝓉 ) = 1 F ϕ _ ( 𝓉 ) , 𝓂 = 0 𝓇 = 0 n c , 𝓂 ( 1 ) 𝓇 ( n 𝓇 ) 𝓉 n 𝓇 I 0 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) ,
where
I 0 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = 0 𝓉 𝓎 n g b * ( 𝓎 ) d 𝓎
then,
I 0 𝓉 ( 𝓎 n ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = ( δ 1   { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ) n
× γ ( 1 n δ 2 , { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 𝓉 δ 2 ) .

3. KRIs

The KRIs are metrics that use statistical methods to measure and monitor an organization’s key risks. KRSIs are a type of KRIs that use statistical techniques, such as probability distributions, regression analysis, and hypothesis testing, to provide a quantitative and data-driven assessment of risk. These KRIs measure the frequency and magnitude of losses due to various risks, such as accidents, losses due to fraud, or losses due to natural disasters. An actuarial estimate of the possible loss that could arise in the future as a result of a particular action or set of circumstances is the risk exposure. Risks are often assessed according to their likelihood of happening in the future multiplied by the potential loss if they did as part of a review of the business’s risk exposure. By assessing the possibility of future losses, a company can distinguish between small and significant losses. Speculative risks frequently lead to losses, including failing to adhere to regulations, a decrease in brand value, security vulnerabilities, and liability issues.
The examination of historical insurance data using time series analysis or continuous distributions has, nevertheless, received a great deal of attention. Actuaries have recently used continuous distributions, especially ones with broad tails, to represent actual insurance data. Using continuous heavy-tailed probability distributions, real data has been modeled in a number of real-world applications, including engineering, risk management, dependability, and the actuarial sciences. The skewness of the insurance data sets can be left, right, or right with huge tails.
Risk exposure is an inevitable event for any insurance firm. Actuaries developed a variety of risk indicators to measure risk exposure as a result. The VAR indicator determines the risk of a potential loss for the insurance company with a specified probability and calculates the amount that a group of investments could lose. An increasingly popular benchmark risk metric for determining risk exposure is this indicator. The VAR often determines how much capital is required, given a specific likelihood, to ensure that the business will not officially go out of business. The chosen confidence level is arbitrary. As a result, a significant VAR amount may be considered for various confidence levels. For the entire company, it may be a high proportion, for example, 99.95% or greater. The inter-unit or inter-risk type of diversification that exists can be represented by these different percentages.
In this paper, these five insurance indicators were chosen due to their importance and prevalence in the statistical literature. These five indicators can express the size of the expected loss for insurance companies and, thus, help insurance companies avoid unexpected and sudden random losses. In insurance analysis, loss refers to the amount of money an insurance company must pay to an insured individual or business as a result of a covered event or claim. Losses can include things such as property damage, medical expenses, and liability judgments. The calculation of loss is a crucial aspect of the insurance industry, as it helps insurance companies determine the financial risks of providing coverage and determine the rates charged for insurance policies.

3.1. VAR Indicator

The VAR is a financial metric that measures the maximum loss that an investment or portfolio is expected to experience with a certain level of confidence over a specified time period. VAR is used to quantify market risk and is a widely used risk management tool for financial institutions. The VAR indicator provides a single value that summarizes the potential loss of an investment or portfolio. This method uses historical data to simulate the distribution of returns and estimate the VAR for a portfolio. This method uses statistical models, such as the normal distribution or the GARCH model, to estimate the VAR for a portfolio. The VAR is a useful risk management tool, as it provides a clear and concise measure of the potential loss of an investment or portfolio. However, it has limitations, as it only provides a point estimate of the potential loss and does not take into account tail risks or extreme events. To address these limitations, financial institutions often use other risk management tools in conjunction with VAR, such as stress testing and scenario analysis.
Definition 1.
Let Y denote a loss random variable, The VAR of Y at the 100 ε % level, say VAR ( Y ) or p ( ε ) , is the 100 ε % quantile (or the 100 ε % percentile ( Q Y )) of the distribution of Y .
Then, based on Definition 1 for the GGR-RW distribution, we can simply write
P r ( Y > Q Y ) = { 1 % | ε = 99 % 5 % | ε = 95 % ,
From Definition 1, for a one-year time when ε = 90 % , the interpretation is that there is only a very small chance ( 10 % ) that the insurance company will be bankrupted by an adverse outcome over the next year. The quantity VAR ( Y ; ε ) does not satisfy one of the five criteria for coherence (see Wirch [2]).

3.2. TVAR Risk Indicator

The VAR indicator is frequently used as a risk assessment tool in the management of financial risk over a defined relatively short time period. Gains and losses are commonly explained in these circumstances using the normal distribution.
The quantity VAR ( Y ; ε ) satisfies all coherence requirements if the distribution of gains (or losses) is restricted to the normal distribution.
Definition 2.
Let Y denote a random loss variable, then the TVAR of Y at the 100 ε % confidence level is the expected loss given that the loss exceeds the 100 ε % of the distribution of Y can be expressed as
TVAR ( Y ) = E ( Y | Y p ( ε ) ) = 1 1 F ϕ _ ( p ( ε ) ) p ( ε ) 𝓎 f ϕ _ ( 𝓎 ) d 𝓎 = 1 1 ε p ( ε ) 𝓎 f ϕ _ ( 𝓎 ) d 𝓎 .
Then,
TVAR ( Y ) = 1 1 ε   I p ( ε ) ( 𝓎 ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) ,
where I p ( ε ) ( 𝓎 ; { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ) = p ( ε ) 𝓎 g { b [ 2 ( + 1 ) + 𝓂 ] } , δ 1 , δ 2 ( 𝓎 ) d 𝓎 . Then,
TVAR ( Y ) =   δ 1 1 ε { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) , 𝓂 = 0 c , 𝓂 Γ ( 1 1 δ 2 , { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 ( p ( ε ) ) δ 2 ) .  
Thus, the quantity in (9) is an average of all VAR values above at the confidence level ε , which provides more information about the tail of the GGR-RW distribution. Further, it can also be expressed as
TVAR ( Y ; ε ) = VAR ( Y ; ε ) + l ( VAR ( Y ; ε ) ) ,
where l ( VAR ( Y ; ε ) ) is the mean excess loss function evaluated at the 100 ε % th quantile. So, TVAR ( Y ; ε ) is larger than its corresponding VAR ( Y ; ε ) by the amount of average excess of all losses that exceed the EL ( Y ; ε ) value of VAR ( Y ; ε ) . The VAR ( Y ; ε ) has been independently developed and is also known as the conditional tail expectation in the insurance literature (Wirch [2]). According to Tasche [3] and Acerbi and Tasche [4], it has also been referred to as the expected shortfall (ES) or the conditional tail expectation (TCE).

3.3. TV Risk Indicator

Furman and Landsman [5] established the TV risk indicator, which determines the loss’s departure from the mean along a tail. Furman and Landsman [6] have created explicit calculations for the TV risk indicator under the multivariate normal distribution.
Definition 3.
Let Y denote a random loss variable, then the TV risk indicator (TR( Y )) can be expressed as
TV ( Y ; ε ) = E ( Y 2 | Y p ( ε ) ) [ TVAR ( Y ; ε ) ] 2 .
Then,
TV ( Y ; ε ) = [ δ 1 { b [ 2 ( + 1 ) + 𝓂 ] } ( 1 / δ 2 ) ] 2 1 ε , 𝓂 = 0 c , 𝓂 { ω p ( ε ) ( 2 ; δ 1 , δ 2 )   1 1 ε [ ω p ( ε ) ( 1 ; δ 1 , δ 2 ) ] 2 } ,    
where
ω p ( ε ) ( ; δ 1 , δ 2 ) = Γ ( 1 δ 2 , { b [ 2 ( + 1 ) + 𝓂 ] } δ 1 δ 2 ( p ( ε ) ) δ 2 ) .
Thus, the quantities in (9) and (10) can be evaluated using any software such as R and MATHCAD, among others, and we will give a numerical example under the reinsurance data with all details and panels in Section 6.

3.4. TMV Risk Indicator

As a metric for the best portfolio choice, Landsman [5] developed the TMV risk indicator, which is based on the TCE risk indicator and the TV risk indicator.
Definition 4.
Let Y denote a random loss variable, then the TMV risk indicator can be expressed as
TMV ( Y ; ε ) = TVAR ( Y ; ε ) + p TV ( Y ; ε ) | 0 < p < 1 .
Then, for any LRV
  • the TMV ( Y ; ε ) > TV ( Y ; ε ) ,
  • for p = 0 , TMV ( Y ; ε ) = TVAR ( Y ; ε ) ,
  • for p = 1 , TMV ( Y ; ε ) = TVAR ( Y ; ε ) + TV ( Y ; ε ) .

4. Modeling

Both simulation and real data have their advantages and disadvantages when comparing models. Using simulation data allows for complete control over the underlying data-generating process and enables the researcher to systematically vary the parameters and compare the models under different conditions. This can provide insight into the robustness and generalizability of the models. On the other hand, using real data provides a more realistic evaluation of the models as it reflects real-world scenarios and allows for the evaluation of the model’s ability to handle real-world complexities and uncertainties. Ultimately, the choice between simulation and real data will depend on the research question, the purpose of the comparison, and the availability of data. In many cases, a combination of both simulation and real data can provide a more comprehensive evaluation of the models being compared.
To illustrate the wide flexibility of the GGR-RW model, we considered three real-life data sets, and the new model is compared with many other relevant and competitive models. Table 1 reports some of the competitive models (see Fréchet [7], Nadarajah and Kotz [8], and Krishna et al. [9]). Real-life data can be examined quantitatively, visually, or using both methods. In order to examine the first fit to theoretical distributions, such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull distributions, we will take into consideration a variety of graphical tools, including the skewness-kurtosis panel (or the Cullen and Frey panel). For more accuracy, bootstrapping is applied. The Cullen and Frey panel summarizes the characteristics of distribution by comparing distributions in the space of squared skewness and kurtosis. The “nonparametric Kernel density estimation (NKDE)” approach for examining initial density shape, the “Quantile-Quantile (Q-Q)” panel for examining the “normality” of the data, the “total time in test (TTT)” panel for examining the initial shape of the empirical HRFs, the “box panel” for examining the extremes, and the scattergrams are also paneled, which are all taken into consideration.
Table 1. Some competitive models.

4.1. Comparing the Competitive Extensions under the Stress Data

A total of 100 observations on the breaking stress of carbon fibers make up the first uncensored data set (see Nichols and Padgett [10]). Figure 1 shows the NKDE panel (the first row, left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel). The breaking stress of carbon fibers is shown in Figure 1 (first-row left panel) to be the asymmetric bimodal and heavy right tail. It is evident from Figure 1’s panel in the first row to the right that the HRF for the current data is monotonically rising. It can be seen from Figure 1’s second-row left panel and second-row right panel that these data contain some extreme values. The current data cannot be described by theoretical distributions, such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull distributions, as seen in Figure 1’s third row right panel. The Shapiro test (ST) was performed to examine the extent to which the first data depended on a normal distribution, and the results were as ST = 0.7165, p-value = 1.33 × 10−12 which means that this data set does not follow the normal distribution.
Figure 1. The NKDE panel (the first row, left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel) for the breaking stress of carbon fibers data.
The statistics Cramér–von Mises criterion (CVM), Anderson–Darling test (AD), Kolmogorov–Smirnov test (KS), and the corresponding p-values ( P υ ) for all fitted models are presented in Table 2. The MLEs and corresponding standard errors (SEs) are reported in Table 3. From Table 2, the GGR-RW model gives the lowest values CVM = 0.0612, AD = 0.4467, KS = 0.05887, and P υ = 0.8789 as compared to the other models. As a result, the GGR-RW is the model that should be selected. In Figure 2, the estimated PDF and CDF are shown. For the current data, Figure 3 shows the Probability-Probability (P-P) panel and estimated HRF. We can see from Figure 2 and Figure 3 that the new GGR-RW model offers good fits for the empirical functions.
Table 2. CVM, AD, KS, and P υ for the breaking stress of carbon fibers data.
Table 3. MLEs and SEs for the breaking stress of carbon fibers data.
Figure 2. Estimated PDF (left right) and estimated CDF (right) for the breaking stress of carbon fibers data.
Figure 3. P-P panel (left) and estimated HRF (right) for the breaking stress of carbon fiber data.

4.2. Comparing the Competitive Extensions under the Glass Fiber Data

The second data set is a list of glass fiber strengths provided by Smith and Naylor (1987) [11]. Figure 4 gives the NKDE panel (the first-row left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), the scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel). Figure 4 (the first row, left panel) indicates that the glass fiber data is asymmetric bimodal and heavy right tail. Figure 4 (the first row, right panel) indicates that the HRF of the glass fiber data is monotonically increasing. The glass fiber data in Figure 4 (second-row left panel and second-row right panel) include some extreme values. Glass fiber data cannot be described by theoretical distributions, such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull distributions, according to Figure 4’s third-row right panel. The ST was performed to examine the extent to which the second data depended on a normal distribution. The results were as ST = 0.72409, p-value = 1.443 × 10−9, which means that this data set does not follow the normal distribution.
Figure 4. The NKDE panel (the first row, left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel) for the glass fibers data.
Table 4 lists the statistics of CVM, AD, KS, and P υ for each fitted model. Table 5 lists the MLEs and related SEs. In comparison to other models, the GGR-RW model provides the lowest values, CVM = 0.11304, AD = 0.89752, KS = 0.12348, and P υ = 0.2691 (Table 4). The GGR-RW distribution can be selected as the best model as a result. In Figure 5, the estimated PDF and CDF are shown. The P-P panel and estimated HRF for the data on glass fiber are shown in Figure 6. It is evident from Figure 5 and Figure 6 that the GGR-RW model adequately fits the empirical distribution functions.
Table 4. CVM, AD, KS and P υ for the glass fiber data.
Table 5. MLEs and SEs for the glass fibers data.
Figure 5. Estimated PDF (left) and CDF (right) for the glass fibers data.
Figure 6. P-P panel (left) and estimated HRF (right) for the glass fibers data.

4.3. Comparing the Competitive Extensions under the Relief Times Data

The third data collection, dubbed “Wingo data”, contains all of the results from a clinical trial that measured the number of hours of pain alleviation for 50 individuals with arthritis. Figure 7 gives the NKDE panel (the first row, left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), the scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel). Figure 7 (the first row, left panel) indicates that the relief time data can be considered as symmetric data. Based on Figure 7 (the first row, right panel), it is noted that the HRF of these data is monotonically increasing. Based on Figure 1 (the second row, left panel, and the second row, right panel), the relief times do not include any extreme values. It can be seen from Figure 7’s third row right panel those theoretical distributions, such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull distributions cannot explain the relief timings. The ST was performed to examine the extent to which the third data depended on a normal distribution, and the results were as ST = 0.96451 and p-value = 0.1373 which means that this data sets follow the normal distribution. However, the new presented GGR-RW model performs better than the normal distribution in modeling this data.
Figure 7. The NKDE panel (the first row, left panel), the TTT panel (the first row, right panel), the box panel (the second row, left panel), the Q-Q panel (the second row, right panel), scattergram panel (the third row, left panel), and the skewness-kurtosis panel (the third row, right panel) for the relief times.
Table 6 lists the statistics CVM, AD, KS, and P υ for each fitted model. Table 7 lists the MLEs and related SEs. The lowest values from Table 6 for the GGR-RW model are CVM = 0.0485, AD = 0.4014, KS = 0.081911, and P υ = 0.8906. As a result, the GGR-RW may be selected as the ideal model. Both the calculated PDF and CDF are shown in Figure 8. The P-P panel and estimated HRF for the relief periods data are shown in Figure 9. We can see from Figure 8 and Figure 9 that the new GGR-RW model offers good fits for the empirical CDFs.
Table 6. CVM, AD, KS and P υ for the relief times data.
Table 7. MLEs and SEs for the relief times data.
Figure 8. Estimated PDF (left) and estimated CDF (right) for the relief times.
Figure 9. P-P panel (left) and estimated HRF (right) for the relief times.

5. Risk Analysis

In this section, we will present a wide range of practical results, whether in the field of actuarial modeling or in the field of insurance applications. First, we introduce synthetic experiments to analyze actuarial risk. Secondly, we will conduct a comprehensive process study on insurance claims data.

5.1. Artificial Risk Analysis

For risk analysis and evaluation, the maximum likelihood estimation (MLE), weighted least squares estimation (WLSQ), ordinary least squares estimation (OLSQ), and Cramér-von Mises estimation (CVME) methods are considered. We will ignore the algebraic derivations and theoretical results of these methods since it is already present in a lot of statistical literature. On the other hand, we will focus a lot on the practical results and statistical applications of mathematical modeling in general and of the risk disclosure, analysis risk, and evaluation of actuarial risks. For computing the above-mentioned KRIs, the following estimation techniques are discussed in these sections: the MLE, OLSQ, WLSQ and CVME methods. Seven CLs (q = 50%, 60%, 70%, 80%, 90%, 95%, 99% and 99%) and N = 1000 with Various sample sizes (n = 20, 50, 100) are considered under α 0 = 2 , b = 0.5 ,   δ 1 = 0.7 ,   δ 2 = 0.3 . All results are reported in Table 8, Table 9 and Table 10. Table 8 shows the KRIs for the GGR-RW under artificial data where n = 20. Table 9 gives the KRIs for the GGR-RW under artificial data where n = 50. Table 10 lists the KRIs for the GGR-RW under artificial data where n = 100. The simulation’s primary objective is to evaluate the efficacy of the five risk analysis methodologies and select the most appropriate and efficient ones. Table 8, Table 9 and Table 10 allow us to display the significant results:
Table 8. The KRIs for the GRR-RW under artificial data where n = 20.
Table 9. The KRIs for the GRR-RW under artificial data where n = 50.
Table 10. The KRIs for the GRR-RW under artificial data where n = 100.
  • The VAR ( Y ; ϕ _ ^ ), TVAR ( Y ; ϕ _ ^ ) and TMV ( Y ; ϕ _ ^ ) increase when q increases for all estimation methods.
  • The TV ( Y ; ϕ _ ^ ) and EL ( Y ; ϕ _ ^ ) decrease when q increases for all estimation methods.
  • The three tables’ results enable us to verify that all approaches are valid and that it is impossible to categorically recommend one approach over another. Given this fundamental finding, we are obliged to develop an application based on real data in the hopes that it would enable us to select one strategy over another and identify the best and most appropriate methods. In other words, even though the results from the five ways to risk assessment were equivalent, the simulation research did not help us decide how to balance the methodologies. These convergent findings comfort us that, when modelling actuarial data and evaluating risk, all methodologies function satisfactorily and within allowable limits.

5.2. Insurance Data for Risk Analysis

The temporal growth of claims over time for each pertinent exposure (or origin) period is typically represented in the chronological insurance real data as a triangle. The year the insurance policy was purchased or the time period during which the loss happened can be considered the exposure period. The origin period need not be annual, as it should be obvious. Origin periods, for instance, might be monthly or quarterly. The length of time it takes for an origin period to develop is referred to as the claim age or claim lag. Data from various insurances are frequently combined to indicate consistent company lines, division levels, or risks. For the purposes of this study, we use a U.K. Motor Non-Comprehensive account as an illustration of the insurance claims payment triangle. We choose to set the origin period between 2007 and 2013 for practical reasons (see Mohamed et al. [12] and Hamed et al. [13]). The insurance claims payment data frame displays the claims data in a manner similar to how a database would normally keep it. The development year, incremental payments, and origin year are all listed in the first column and range from 2007 to 2013. It is crucial to keep in mind that this data on insurance claims were initially examined using a probability-based distribution. The analysis of real data can be carried out numerically, visually, or by combining the two. The numerical method, as well as a few graphical tools, such as the skewness-kurtosis panel (or the Cullen and Frey panel), are taken into consideration when assessing initial fits of theoretical distributions such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull (see Figure 9).
Our negatively skewed actuarial data, with a kurtosis of less than three, are shown in Figure 10. The NKDE method is used to examine the initial shape of the insurance claims density (see Figure 11, top left panel), and the P-P panel is used to assess the “normality” of the current data (see Figure 11, top right panel), the TTT panel is used to assess the initial shape of the empirical HRF (see Figure 11, bottom left panel), and the “box panel” is used to identify the explanatory variables (see Figure 10, the bottom right panel). Figure 11 shows the initial density as an asymmetric function with a left tail (top left panel). Figure 11 does not support any irrational claims (bottom right panel). According to Figure 11’s bottom left panel, the HRF for the models that explain the current data should similarly be monotonically expanding. Figure 12 displays the scattergrams for the data on insurance claims. Figure 13 (left panel) displays the autocorrelation function (ACF) for the data on insurance claims, and Figure 13 (right panel) displays the partial autocorrelation function (partial ACF).
Figure 10. Cullen-Frey panel for the actuarial claims data.
Figure 11. NKDE (top left), Q-Q (top right), TTT (bottom left), and box panels (bottom right) for the claims data.
Figure 12. The scattergrams.
Figure 13. The ACF (left) and the partial ACF (right) for the insurance claims data.
We offer the ACF to show how the correlation between any two signal values changes as a function of the distance between them. The theoretical ACF provides no information about the process’s frequency content; instead, it measures the stochastic process memory in the time domain. Figure 13 illustrates how hills and valleys are distributed across the surface when Lag = k = 1 (the left panel). The theoretical partial ACF with Lag = k = 1 is also presented; see Figure 13 (the right panel). The initial lag value is demonstrated in Figure 13 (the right panel) to be statistically significant, in contrast to the other partial autocorrelations for all other lags. The first NKDE exhibits an asymmetric density with a left tail, as shown in Figure 11’s top left panel. On the other hand, the novel model’s interview, matching, and density are important in statistical modeling since they take into account the left tail shape. Therefore, it is recommended to simulate insurance claim payouts using the GGR-RW model. We present an application for risk analysis under VAR, TVAR, TV, TMV, and EL   that   measures insurance claims data. The risk analysis is performed for some confidence levels as follows:
q = 50 % , 60 % , 70 % , 80 % , 90 % , 95 % , 99 %   and   99 % .
The GGR-RW and RW models estimate the five metrics. The KRIs for the GGR-RW under insurance claims data are shown in Table 11. The estimators and ranks for the GGR-RW model under the claims data are shown in Table 12 for all estimations. The KRIs for the GGR-RW are listed in Table 13 under statistics on insurance claims. The estimators and ranks for the RW model under the claims data are shown in Table 14 for all estimations. The estimators and rankings for each estimating method are shown in the table for the RW model under the claims data. Because it is the foundational distribution upon which the new distribution is based, the RW distribution was chosen. These tables’ conclusions are summarized as follows:
Table 11. The KRIs for the GRR-RW under insurance claims data.
Table 12. The estimations for all estimation methods under the GRR-RW model.
Table 13. The KRIs for the RW under insurance claims data.
Table 14. The estimations for all estimation methods under the RW model.
  • For all risk assessment methods | q = 50 % , ,   99 % :
    VAR ( Y ; ϕ _ ) | q = 50 % < VAR ( Y ; ϕ _ ) | q = 60 % < VAR ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , 60 % , 70 % , 80 % , 90 % , 95 % , 99 %   and   99 % :
    TVAR ( Y ; ϕ _ ) | q = 50 % < TVAR ( Y ; ϕ _ ) | q = 60 % < TVAR ( Y ; ϕ _ ) | q = 99 % .
  • For Most risk assessment methods | q = 60 % , ,   99 % :
    TV ( Y ; ϕ _ ) | q = 50 % < TV ( Y ; ϕ _ ) | q = 60 % < TV ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , ,   99 % :
    TMV ( Y ; ϕ _ ) | q = 50 % > TMV ( Y ; ϕ _ , 0.5 ) | q = 60 % > TMV ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , ,   99 % :
    EL ( Y ; ϕ _ ) | q = 50 % > EL ( Y ; ϕ _ , 0.5 ) | q = 60 % > EL ( Y ; ϕ _ ) | q = 99 % .
  • Under the GGR-RW model and the MLE method, the VAR( Y ; ϕ _ ^ ) is monotonically increasing indicator starting with 2519.589871 | q = 50 % and ending with 9747.670085 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 4267.691033 | q = 50 % and ending with 12,352.578196 | q = 99 % . However, the TV( Y ; ϕ _ ^ ), the TMV( Y ; ϕ _ ^ ) and the EL( Y ; ϕ _ ^ ) are decreasing functions. Under the RW model and the MLE method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 1986.487789 | q = 50 % and ending with 58,937.60432 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 11,961.555933 | q = 50 % and ending with 270,450.59761 | q = 99 % , the TV( Y ; ϕ _ ^ ), the TMV( Y ; ϕ _ ^ ) and the EL( Y ; ϕ _ ^ ) are decreasing function indicators.
  • Under the GGR-RW model and the LSE method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 2428.50708 | q = 50 % and ending with 13,665.84114 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 4865.81245 | q = 50 % and ending with 18,627.28911 | q = 99 % . However, the TV( Y ; ϕ _ ^ ), the TV( Y ; ϕ _ ^ ) and the TMV ( Y ; ϕ _ ^ ) are decreasing functions. Under the RW model and the OLSQ method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 2226.27121 | q = 50 % and ending with 44,780.28947 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 9174.03623 | q = 50 % and ending with 151,748.59629 | q = 99 % . Also, the TV ( Y ; ϕ _ ^ ), the TMV ( Y ; ϕ _ ^ ) and the EL( Y ; ϕ _ ^ ) are decreasing functions.
  • Under the GGR-RW model and the WLSQ method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 2356.17122 | q = 50 % and ending with 9514.1782 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 4108.2022 | q = 50 % and ending with 12,042.68509 | q = 99 % . However, the TV( Y ; ϕ _ ^ ), the TMV( Y ; ϕ _ ^ ) and the EL( Y ; ϕ _ ^ ) are decreasing functions. Under the RW model and the WLSQ method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 1964.20669 | q = 50 % and ending with 19,161.39616 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 4830.43197 | q = 50 % and ending with 41,550.35309 | q = 99 % . However, the TV( Y ; ϕ _ ^ ), the TMV( Y ; ϕ _ ^ ), and the EL( Y ; ϕ _ ^ ) are decreasing functions.
  • Under the GGR-RW model and the CVM method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 2440.48715 | q = 50 % and ending with 12,431.24689 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) in monotonically increasing indicator starting with 4680.62835 | q = 50 % and ending with 16,642.80352 | q = 99 % . However, the TV( Y ; ϕ _ ^ ), the TMV( Y ; ϕ _ ^ ), and the EL( Y ; ϕ _ ^ ) are decreasing function. Under the RW model and the AE method, the VAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 2229.99377 | q = 50 % and ending with 38,705.15155 | q = 99 % , the TVAR( Y ; ϕ _ ^ ) is a monotonically increasing indicator starting with 8126.51599 | q = 50 % and ending with 118,195.6688 | q = 99 % , the TV( Y ; ϕ _ ^ ), the TMV(Z Y ; ϕ _ ^ ) and the EL( Y ; ϕ _ ^ ) are decreasing functions.
The novel Reciprocal Weibull distribution is suitable for insurance claims data since the insurance claims data have a heavy tail and are skewed to the left (see Figure 11, the top right panel). On the other hand, the new distribution also has a heavy tail to the right. This initial fit between the shape of the data and the shape of the distribution is the first starting point through which we recommend a specific probability distribution for modeling specific data. This process is considered part of the data exploration process, and during the data exploration process, the researcher reaches this conclusion. Of course, it is not a requirement that the distribution is suitable for the data, but at least it is a preliminary step through which we can nominate a specific model for a specific data mode. To reach the final results in the distribution selection process, many comparison criteria can be used through which we can be certain of some results, as is the case in Table 2, Table 4 and Table 6. However, in Table 11 and Table 13, the new model is compared with the baseline model in risk analysis process. The new distribution proved to be compatible with the original distribution, according to the aforementioned results.

6. Concluding Remarks

Risk exposure can be correctly described using continuous distributions. To illustrate the level of exposure to a specific threat, it is preferable to use a data point or, at the very least, a restricted range of numbers. These risk exposure numbers, also known as the primary risk indicators, are obviously the product of a certain model. Five key risk indicators—value-at-risk, tail-value-at-risk, tail variance, and tail mean-variance—were also implemented to define the risk exposure under the reinsurance revenue data.
Given that the recommended version was used to make these measurements, this research provides a fresh distribution for this usage. Important statistical characteristics can be derived, such as the generating function, ordinary moments, and incomplete moments. In terms of modeling the breaking data, the new model performs better than the Beta reciprocal Weibull model, the Kumaraswamy reciprocal Weibull model, the McDonald reciprocal Weibull model, the Marshall-Olkin reciprocal Weibull model, the odd Burr reciprocal Weibull model, the odd log-logistic reciprocal Weibull model, the odd log-logistic exponentiated reciprocal Weibull model, the transmuted reciprocal Weibull model.
Five key risk indicators are employed for analyzing the risk level under the reinsurance revenues dataset. An application is provided along with its relevant numerical analysis and panels. Some useful results are identified and highlighted as follows:
  • For all risk assessment methods | q = 50 % , ,   99 % :
    VAR ( Y ; ϕ _ ) | q = 50 % < VAR ( Y ; ϕ _ ) | q = 60 % < VAR ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , 60 % , 70 % , 80 % , 90 % , 95 %   and   99 % :
    TVAR ( Y ; ϕ _ ) | q = 50 % < TVAR ( Y ; ϕ _ ) | q = 60 % < TVAR ( Y ; ϕ _ ) | q = 99 % .
  • For Most risk assessment methods | q = 50 % , ,   99 % :
    TV ( Y ; ϕ _ ) | q = 50 % < TV ( Y ; ϕ _ ) | q = 60 % < TV ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , ,   99 % :
    TMV ( Y ; ϕ _ ) | q = 50 % > TMV ( Y ; ϕ _ ) | q = 60 % > TMV ( Y ; ϕ _ ) | q = 99 % .
  • For all risk assessment methods | q = 50 % , ,   99 % :
    EL ( Y ; ϕ _ ) | q = 50 % > EL ( Y ; ϕ _ ) | q = 60 % > EL ( Y ; ϕ _ ) | q = 99 % .

Author Contributions

H.M.Y.: reviewing and editing, software, validation, writing the original draft preparation, conceptualization, supervision. Y.T.: methodology, conceptualization, software. W.E.: validation, writing the original draft preparation, conceptualization, data curation, formal analysis, software. M.M.A.: reviewing and editing, conceptualization, supervision. M.I.: writing the original draft, software, reviewing, editing, validation, conceptualization. All authors have read and agreed to the published version of the manuscript.

Funding

The study was funded by Researchers Supporting Project number (RSP2023R488), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The dataset can be provided upon request.

Acknowledgments

The study was funded by Researchers Supporting Project Number (RSP2023R488), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Artzner, P. Application of Coherent Risk Measures to Capital Requirements in Insurance. N. Am. Actuar. J. 1999, 3, 11–25. [Google Scholar] [CrossRef]
  2. Wirch, J. Raising Value at Risk. N. Am. Actuar. J. 1999, 3, 106–115. [Google Scholar] [CrossRef]
  3. Tasche, D. Expected Shortfall and Beyond. J. Bank. Financ. 2002, 26, 1519–1533. [Google Scholar] [CrossRef]
  4. Acerbi, C.; Tasche, D. On the coherence of expected shortfall. J. Bank. Financ. 2002, 26, 1487–1503. [Google Scholar] [CrossRef]
  5. Furman, E.; Landsman, Z. Tail variance premium with applications for elliptical portfolio of risks. ASTIN Bull. 2006, 36, 433–462. [Google Scholar] [CrossRef]
  6. Landsman, Z. On the tail mean—Variance optimal portfolio selection. Insur. Math. Econ. 2010, 46, 547–553. [Google Scholar] [CrossRef]
  7. Fréchet, M. Sur la loi de probabilité de lécart maximum. Ann. Soc. Pol. Math. 1927, 6, 93–116. [Google Scholar]
  8. Nadarajah, S.; Kotz, S. The exponentiated Fréchet distribution. Interstate Electron. J. 2003, 14, 1–7. [Google Scholar]
  9. Krishna, E.; Jose, K.K.; Alice, T.; Risti, M.M. The Marshall-Olkin Fréchet distribution. Commun. Stat. Theory Methods 2013, 42, 4091–4107. [Google Scholar] [CrossRef]
  10. Nichols, M.D.; Padgett, W.J. A bootstrap control chart for Weibull percentiles. Qual. Reliab. Eng. Int. 2006, 22, 141–151. [Google Scholar] [CrossRef]
  11. Smith, R.L.; Naylor, J. A comparison of maximum likelihood and Bayesian estimators for the three-parameter Weibull distribution. J. R. Stat. Soc. Ser. C Appl. Stat. 1987, 36, 358–369. [Google Scholar] [CrossRef]
  12. Mohamed, H.S.; Cordeiro, G.M.; Minkah, R.; Yousof, H.M.; Ibrahim, M. A size-of-loss model for the negatively skewed insurance claims data: Applications, risk analysis using different methods and statistical forecasting. J. Appl. Stat. 2022, 1–22. [Google Scholar] [CrossRef]
  13. Hamed, M.S.; Cordeiro, G.M.; Yousof, H.M. A New Compound Lomax Model: Properties, Copulas, Modeling and Risk Analysis Utilizing the Negatively Skewed Insurance Claims Data. Pak. J. Stat. Oper. Res. 2022, 18, 601–631. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.