1. Introduction
Skewed distributions are common in real-world processes, particularly when extreme values or outliers significantly influence standard statistical approaches. The analysis of skewed data is critical in various fields, including finance and insurance [
1,
2,
3,
4], communication and signal processing [
5,
6,
7,
8,
9], and environmental science [
10,
11,
12,
13,
14]. In many instances, data characterized by high positive skewness can exhibit heterogeneity, where a single parametric univariate probability distribution falls short in effectively modeling such data. When dealing with highly right-skewed data, it is necessary to understand the asymmetry of observed data, which can be decomposed into two or more distinct components [
15]. For the case of two components, the first component is the main innovation, representing where the bulk of the data is centered, while the second is the tail component, which captures the few extreme observations above a certain threshold that contribute to the skewness. Failing to account for this heterogeneity could lead to significant loss in efficiency when modeling the data. Moreover, the modeling of this type of data has received widespread attention in recent years, and many models/methods have been put forth, including non-parametric models [
16] and models following the Peak over Threshold (PoT) methodology [
11,
17]. While non-parametric models usually offer good fits to the data, they typically fail to account for a few outlying observations in the tail. On the other hand, the PoT methodology only focuses on the extreme observations beyond a certain threshold while ignoring the remaining observations in the data, and thus, it does not make use of the entire distribution. Given these challenges, hybrid models have emerged as effective solutions for modeling complex distributions. A hybrid model combines two or more probability distributions to adequately fit the characteristics of observed data. In [
15], a general framework for generating multi-component hybrid models was developed. Moreover, several families of two-component hybrid models have been defined and studied in the literature [
3,
13,
18,
19,
20,
21,
22,
23,
24].
In this paper, a new two-component hybrid model for data exhibiting high skewness to the right is introduced. The model links a half-normal distribution for the bulk of the data distribution and a GPD for the tail after a certain threshold point determined when imposing a condition of class
[
3,
25]. The choice of using the half-normal distribution as the bulk distribution, in contrast to using distributions such as the normal, gamma, and lognormal ones, is motivated by the fact that its support on
naturally aligns with strictly positive data with zeroes, common in many real-world measurements, avoiding the conceptual and practical issues of using a normal distribution that extends into negative values. The half-normal distribution also captures moderate skewness with a light, exponentially decaying tail, creating a clear boundary between the bulk and the GPD tail component, a situation crucial for reliable threshold selection and accurate tail index estimation. Compared to the gamma or lognormal distributions, the half-normal distribution is more parsimonious, having only a single scale parameter, which reduces overfitting risk and minimizes parameter identifiability conflicts with the tail component. Its relatively fast tail decay prevents ’tail leakage’, where bulk observations intrude into the extremes, and its closed-form expressions for key properties simplify estimation and inference. Furthermore, its interpretability as the distribution of the absolute value of a normal variate makes it particularly appealing in cases where data represents magnitudes of symmetric underlying processes. These qualities, together, make the half-normal model a robust, stable, and theoretically well-matched partner to the GPD model in hybrid modeling frameworks. Furthermore, the two components of the hybrid half-normal–GPD model are weighted non-uniformly, and an unsupervised iterative and convergent estimation scheme based on the Levenberg–Marquardt (L-M) algorithm [
26,
27] is adopted to estimate the threshold point and other free parameters of the two-component model. In [
15] a maximum likelihood-based algorithm was used to estimate the parameters of multi-component hybrid models, and the challenge of selecting initial values for the parameters before the maximization of the log-likelihood function was bypassed by the use of the unsupervised iterative and convergent estimation scheme presented in this paper, since the scheme is self-calibrating. Also, in standard PoT methodology, the threshold point is usually estimated graphically, whereas in this framework, it is a parameter to be estimated in the hybrid model. This allows us to determine the point beyond which the extremes are observed algorithmically.
In
Section 2, the specification of the two-component non-uniform weights hybrid model framework and the half-normal–GPD model is carried out. A description of the UIA for the estimation of the hybrid model’s free parameters is presented in
Section 3. Results from numerical studies based on Monte Carlo simulations conducted to assess the UIA’s efficiency in estimating the hybrid model’s parameters are reported in
Section 4. In
Section 5, application of the new hybrid model to fit data on the intensity of rainfall that triggered some debris flow events in the South Tyrol region of Italy is performed. The paper closes with discussions and a conclusion in
Section 6.
2. Two-Component Non-Uniform Weight Hybrid Model
Suppose we have data which can be decomposed into two components and the components represent specific behavior of the dichotomized data, while the goal is to use a smooth piecewise probability density function (pdf) to model the data. Assume the data is continuous and follows a non-degenerate distribution. Let
and
be two pdfs, each with parameter vector
and
, in such a manner that each pdf is suitable for modeling specific components of the data and without a loss in generality.
and
are suitable for modeling the first and second components of the data, respectively. Suppose
and
are the respective cumulative distribution functions (cdfs) corresponding to
and
, with respective corresponding quantile functions
and
, where
The general two-component non-uniform weight hybrid model for the data can be specified by the pdf of the form
where
is a vector of free parameters in the model,
and
are weights associated with the respective component of the pdf in Equation (
1), and
u is a junction point or threshold indicating the point of transition from one component or behavior of the data to another.
In modeling complex data distributions, particularly those with heterogeneous characteristics across the range, a hybrid approach offers greater flexibility and accuracy. The two-component model in Equation (
1) assigns non-uniform weights to each component, as it is specifically designed to capture distinct features of the data, and hence,
. To further justify the rationale behind the assignment of non-uniform weights in the hybrid model in Equation (
1), we argue that the first component is weighted more heavily, as it targets the bulk or central mass of the distribution, and the second component is given less weight and is focused on modeling the extreme tail behavior, where outlier events occur. This non-uniform weighting reflects the natural asymmetry in many real-world data sets, where most observations cluster around a central trend and a smaller proportion of data lies in the tail. In contrast, a uniform weighting framework treats both components as contributing equally to the overall model regardless of the relative frequency or importance of the patterns they are meant to capture. While seemingly fair or neutral, this approach introduces several limitations, including under-representation of the central structure of the data if a tail component is given too much influence, ignoring context-specific relevance by treating all observations as equally informative and overall model over-sensitivity to outliers.
Suppose that, in the pdf in Equation (
1), the transition from one component to another is smooth; then, the following assumptions are made:
- (i)
The pdf
f is positive and satisfies
implying that
- (ii)
The distribution of the data has a heavy right tail belonging to the Fréchet maximum domain of attraction.
- (iii)
The pdf f is continuous and differentiable at the threshold u; in addition, it is smooth and -regular, implying that
Given these assumptions, we obtain
The pdf in Equation (
1) has cdf and corresponding quantile function given by the equations below:
Remark 1. One can simulate random samples from the model in Equation (1) using Equation (4) by simply replacing p with the random variable U, where U is uniform on To define a two-component half-normal–GPD hybrid model,
is taken to be the half-normal distribution with pdf, cdf, and quantile function expressed by the following respective equations:
where
is a scale parameter,
is the error function and
is its inverse. Furthermore, take
as the GPD with pdf, cdf, and quantile function expressed by the following respective equations:
where
is a scale parameter and
is the tail index parameter which controls the shape of the GPD. Using assumptions (i)–(iii), we obtain the following relations for some of the parameters of the distribution:
It follows that the vector of parameter
will contain only the free parameters including the threshold
u. Thus,
. These would be the parameters to be estimated using the proposed UIA. Once
has been estimated, the estimates of the parameters
, and
can easily be realized from Equation (
5).
Remark 2. Observe that in the half-normal–GPD hybrid model, there were six parameters whose values we needed to estimate. However, with assumptions (i)–(iii), the number of free parameters to be estimated was reduced to three. The graphs of the density and hazard functions for selected values of free parameters are given by Figure 1. Remark 3. Given that , it follows that the threshold u can be any quantile of the half-normal–GPD model, and thus, the distribution is free of any constraint which is usually imposed for the case when [3,25]. Also, since the half-normal–GPD model is positively skewed, the mode of f is equal to zero and less than the median M. This implies that , and consequently, Furthermore, and Lastly, can be interpreted as the probability of exceeding the threshold u, while is a normalization parameter ensuring that the density f integrates to unity. 3. Unsupervised Iterative Estimation Algorithm
In this section, a description of the UIA employed in estimating the vector of free parameter
is presented. The model described in
Section 2 is taken to belong to the Fréchet maximum domain of attraction (i.e.,
). For each iteration in the UIA, the UIA breaks down the problem of estimating the vector of free parameters
into a double-nested sub-problem, namely the parameter
and
. These are estimated successively. For each iteration of the UIA, the parameter vector
is first estimated by minimizing the Squared Distance (SD) between the empirical cdf based on some sample and the theoretical one based on the value of
, and then,
is replaced by its estimate obtained in this iteration to estimate
in the next step of the algorithm using a similar procedure. Furthermore,
is estimated again by minimizing the SD between the empirical cdf based on some sample and the theoretical one based on the estimate of
obtained from the previous iteration, and then,
is replaced by its estimate obtained in this iteration to estimate
in the next step of the algorithm using a similar procedure. Evidently, the algorithm begins with an initial value for
obtained by minimizing the SD between the empirical cdf based on some sample and the theoretical one based on the initial value chosen for
and only ends when a stop condition is realized.
Consider a sample
from the half-normal–GPD model, and let
be a given realization. Suppose
and
are the initial value and the estimate of the parameter
at the
iteration, respectively. To proceed with the UIA for the two-component hybrid model in
1, we start with the initial value
, rather than beginning by specifying an initial value for
, because the only information we have about
is that it is positive, given that we are dealing with positive data with positive skewness. Moreover,
is chosen, where
represents an empirical quantile of order
associated with
F. Also,
is chosen (as we fit a GPD above
u). Moreover,
since we naturally expect
, the initial value of
is used to estimate
by minimizing the SD between the hybrid cdf given
(fixed) and the empirical cdf
associated with the sample
of size
n, defined, for all
, by
. Moreover, the SD is not evaluated only on the realizations
(because there might turn out to be just a few realizations in the tail), but also on a sequence of generated synthetic data with an increasing property
of size
m (which may not be the same as
n), with a logarithmic step. The synthetic data is added to increase the number of realizations above the tail threshold
u. In particular, for any
,
is defined by
Remark 4. The introduction of new points between the observations of X only has an impact on F by aiding its evaluation on more points, with no impact on the step function .
To obtain
, we solve the following minimization problem using the L-M algorithm [
26,
27]:
where
stands for
for
, and
denotes the
-norm. After realizing
, we proceed with the iterations. Now,
, the
iteration, is divided into two separate minimization problems, which are resolved successively, as described in the following steps.
Step 1: Determine
by minimizing the SD between the hybrid cdf, given
, and the empirical one as follows:
where
denotes
for
(fixed). The L-M algorithm is employed to numerically solve this minimization problem.
Step 2: Determine
by minimizing the SD between the hybrid cdf, given
, and the empirical one by solving the following minimization problem using the L-M algorithm:
where
represents
for
(fixed).
Stop condition: The iterations continue till the following stop conditions are satisfied:
where
is a small positive real number;
stands for the observations above a fixed quantile
of a given order
, which is associated with the cdf
F; and
denotes the distance between
x and
y. The distance
is chosen in this paper to be the Mean Squared Error (MSE), and it can be further interpreted as the Cramér-von-Mises test of goodness of fit.
To guarantee a good fit for all of the data points and not just the data points lying in the area where the bulk of the distribution lies, but also for the tail, the UIA is forced to stop only when the MSE between the hybrid cdf and the empirical one is small enough, and this implies satisfaction of Conditions 1 and 2; otherwise, satisfaction is implied when a fixed maximum number of iterations is attained (Condition 3).
Remark 5. While the hybrid model considered in this paper is assumed to belong to the Fréchet maximum domain of attraction, this algorithm can be extended to the case when the tail index of the GPD is free of any constraints. Also, though the maximum likelihood estimation method appears as a natural estimation method for the parameters of the hybrid model, in practice, this can be very challenging to execute, especially when the number of parameters to be estimated is high. Estimating many free parameters at once can be very challenging, and the challenge starts when selecting initial values for these parameters, as well as obtaining the expression of the gradient function from the likelihood or log-likelihood function. This has informed the choice of the UIA as an alternative method to estimate the parameters of the hybrid model. Nevertheless, when the number of free parameters is relatively small, the maximum likelihood estimation method can be used and is equally robust.
Pseudo-code of the algorithm is given by Algorithm 1.
To study the convergence of the algorithm, it is necessary to provide proof of the existence of stationary points. Moreover, the algorithm, which consists of a sequence of minimizations, does not rely on the optimization of a cost function by seeking a path to reach an extremum of an error surface. For a given realization
and given parameters
with
, we define
and
as the respective domain of
and
. Now, consider the function
where for
,
is defined by
To verify that
is a mapping, it is sufficient to show that
admits a minimum which is unique, for any
, with
and
. To do so,
is fixed at
(the result would be the same for
), and we show that
is a strongly quasiconvex function with respect to
, which is deduced from the strict convexity with respect to
of the function
, denoted by
to simply express it. Thus, we can write
, with
,
, and
:
, which implies
It follows that
,
, and
,
. Thus,
is strongly quasiconvex on
, a compact of
, which ensures the presence of a unique minimum. The strict convexity with respect to
of
F further shows that the second derivative of
F with respect to
is positive.
Algorithm 1 UIA for the hybrid half-normal–GPD parameter estimation |
- 1:
Initialization: Decide start values for and as well as values for and . Proceed to obtain from: - 2:
Iterative process: For
- 3:
Output: Return
|
The two steps of the first iteration of Algorithm 1 can be expressed using
and
, for a fixed
via the following relations:
In general, for any
, one can write
where the function
is defined from
to
by:
. As a consequence, Algorithm 1 can be expressed alternatively as Algorithm 2.
Algorithm 2 Alternative UIA for the hybrid half-normal–GPD parameter estimation |
- 1:
Initialization: Decide start values for and as well as values for and and proceed to obtain - 2:
until or . - 3:
Output: Return
|
To show that stationary points exist for Algorithm 2, it is necessary to prove the existence of fixed points of the function
. To realize this, we proceed from the fixed-point theorem due to Brouwer (see [
28]), which means that
every continuous function from a closed ball of a Euclidean space into itself has a fixed point. This implies that the functional
admits at least one fixed point given the satisfaction of the following two conditions:
Now, as
is a compact set, it is sufficient to show that
is continuous on
to obtain, by the Heine–Cantor theorem, its uniform continuity. For this, it is necessary to study the continuity of
F with respect to
to determine the continuity of
with respect to
, since
. Recall that, by construction,
F is continuous with respect to
x and not to its parameters. Hence, its continuity according to
needs to be established. Since we know that, from Equation (
3),
F is composed of two functions, the half-normal cdf and that of the GPD, it suffices to study the continuity of each of them with respect to
. The continuity of the half-normal cdf
F as a function of
is obvious since it means to look at the continuity of its likelihood with respect to
. Now, for the GPD, its parameter
is expressed as a function of
:
. The GPD parameter
can equally be expressed as a function of
:
, and hence, both are continuous in
. It follows that the GPD is continuous in
as the composition of continuous functions with respect to
. The continuity of the function
on
as a composition, the sum and products of continuous functions on
can be inferred, and from which, one can conclude the continuity of
on
.
In sum, since a fixed point exists for the functional , the algorithm admits at least one stationary point. Moreover, since the method does not follow a path on an error surface, it is free from local minima traps, as found in standard gradient search-based methods. It can also be verified numerically, as performed in the next section, that the algorithm converges to a unique stationary point regardless of its initialization.
4. Numerical Studies
To study the performance of the UIA described in
Section 3, Monte Carlo simulations are carried out. Through simulations, an attempt is made to test the efficiency of the UIA as applied in estimating the parameters of the hybrid half-normal–GPD model.
We proceed with the simulations as follows: We consider
N (
in this paper) training sets
of size
n and
N test sets
of size
l, simulated from the hybrid half-normal–GPD model with a fixed parameters vector
. Using each training set
,
,
is estimated to be
, using the UIA. We denote by
the estimate of the parameter
relative to the
training set. Furthermore, the empirical mean and variance of
over the
N training sets is computed, namely
and
, respectively. The significance of
is determined based on two criteria: the MSE and test of hypothesis on
The MSE is expressed for any parameter
as
A small MSE value indicates the efficiency of the UIA in estimating the parameter
. To test
(with unknown variance), we set up the hypothesis
Because
N is large, a
-test (instead of a
t-test) of size
, with a rejection region of
at risk
described by
is used, where the statistics
are given by
and
denotes the quantile of order
of the standard normal distribution
. Additionally, we compute the
p-value
, with respect to the parameter
, which we will compare to
. Whenever this
p-value is higher than
, we fail to reject
. For any
n and any parameter
, we obtain
and
to confirm high acceptance (at the
level of significance) of
(
), that is, a very high level of similarity between the values obtained through the UIA and the fixed ones. Lastly, the hybrid half-normal GPD pdf
f, given
, is compared with the pdf
estimated on each test set
, given
. For this, the average of the log-likelihood ratio
of
by
over the
N simulations is computed:
A small value of
indicates an efficient estimation of the parameters of the hybrid half-normal–GPD model using the UIA.
We performed many Monte Carlo simulations by varying
and
n in order to ascertain the robustness of the UIA for different values of the parameters and sample sizes. We also set
,
,
,
,
, and
. To conserve space, the results of three such simulations are reported in
Table 1,
Table 2 and
Table 3, and the other unreported simulations follow a fashion similar to the ones reported here. The efficiency of the UIA, in terms of goodness of fit, is shown through the two criteria described above and the average of the log-likelihood ratio
.
We observe from the results in
Table 1,
Table 2 and
Table 3 that, as the sample size increases, the MSE becomes smaller for all parameters. The variance of
is also observed to be smaller than the variances of
u and
for all parameter combinations and sample sizes. We also observe that
is very small for all parameter combinations and sample sizes. This highlights the accuracy and efficiency of the estimation of the parameters using the UIA. Lastly, it can be observed that as the sample size increases, the average execution time of the UIA increases. It should be noted that the estimation algorithm was implemented using the R programming language. For faster programming languages, the average execution time of the estimation algorithm could be significantly reduced.
5. Application
Here, the hybrid half-normal–GPD (HNGPD) model with parameter vector
alongside the hybrid lognormal–GPD (LNGPD), gamma–GPD (GGPD), and the normal GPD (NGPD) models [
30,
31] are used to model the intensity, in millimeters per hour, of rainfall which triggered some 785 debris flow events in the South Tyrol region of Italy between 1987 and 2022. The data set was provided by the Agency for Civil Protection of the Autonomous Province of Bozen-Bolzano, Italy. The skewness and excess kurtosis of the data are
and
, respectively, which clearly shows that the data set is highly skewed to the right with a heavy tail. A violin plot of the data with a superimposed box plot is given in
Figure 2.
The hybrid LNGPD, GGPD, and the NGPD models belong to the class of hybrid models with a cdf with the general form
For the LNGPD model,
is the cdf of the lognormal distribution with parameter vector
, where
and
are the logarithm of the mean and standard deviation, respectively, and
is the cdf of the GPD distribution with parameter
. Thus, the vector of all the parameters in the model is
. For the GGPD model,
is the cdf of the gamma distribution with parameter vector
, where
and
are shape and scale parameters, respectively, and
is the cdf of the GPD distribution with parameter
. Again, the vector of all the parameters in the model is
. Lastly, for the NGPD model,
is the cdf of the normal distribution with parameter vector
, where
and
are the mean and standard deviation, respectively, and
is the cdf of the GPD distribution with parameter
, and the vector of all the parameters in the model is
.
Reported in
Table 4 are the results obtained from using the HNGPD model, estimated using the UIA, in fitting the rainfall intensity data.
Table 4 also contains results obtained from using the LNGPD, GGPD, and the NGPD models, all estimated using the maximum likelihood estimation (MLE) method, to fit the data. The results include estimates of the parameter vector
for all the models, the Akaike Information Criterion (AIC) for all the models, and the Kolmogorov–Smirnov (K-S) statistic, as well as its corresponding
p-value, for all the models.
Figure 3,
Figure 4,
Figure 5 and
Figure 6 show the graph of the fitted density over the histogram of the data and Q-Q plot (on a log-scale) for the HNGPD, LNGPD, GGPD, and NGPD models, respectively. Results from
Table 4 and
Figure 3,
Figure 4,
Figure 5 and
Figure 6 reveal that the HNGPD offered the best fit to that data because it has the highest
p-value for the K-S statistic.
We also compare the algorithmically estimated threshold
using the UIA with some graphical threshold selection approaches, namely the mean excess plot (MEP) [
2], the Hill plot [
32], and the stability plot [
33]. These graphical threshold selection approaches are used to estimate
u from the rainfall intensity data, and by using the method of probability-weighted moments (PWMs) [
34], a GPD is fitted to the observations above the
u selected by each approach. Results regarding the estimated GPD parameters, the selected
u, and the number of observations above each selected
u are contained in
Table 5. A comparison between the empirical cdf and the one estimated using the UIA for the HNGPD model is given by
Figure 7. Also, A comparison between the empirical cdf and the one estimated using model parameters obtained from the MEP, Hill plot, and stability plot for the HNGPD model is given in
Figure 8,
Figure 9 and
Figure 10, respectively. Lastly, a comparison between the empirical upper quantiles and the ones estimated using the UIA and the graphical approaches is given in
Figure 11.
6. Discussion and Conclusions
In
Section 5, it was established from the results derived from the application of the model to rainfall intensity data that the new HNGPD hybrid model, estimated using the unsupervised iterative estimation algorithm discussed in
Section 3, offers the best fit to the data. Moreover, it can be observed that the LNGPD, GGPD, and the NGPD models all underestimated the value of the threshold
u, beyond which a GPD is fitted to the tail of the distribution of the data. The implication of this is that all three models are suggesting the presence of more outliers or extreme observations in the data set than observations around the mean (bulk of the distribution). This can be very misleading from a practical point of view, where the bulk of the distribution is supposed to contain the majority of the observations. In particular,
for the LNGPD model, and thus, the threshold is on the 47.89th quantile. This further implies that a total of 410 observations out of the 785 observations are above the threshold and hence, are classified as extreme observations. For the GGPD model,
and thus, the threshold is on the 38.47th quantile, implying that a total of 483 observations out of the 785 observations are extreme observations. For the NGPD model,
and thus, the threshold is on the 32.10th quantile, implying that a total of 533 observations out of the 785 observations are extreme observations. For the new two-component hybrid HNGPD model estimated using the unsupervised iterative estimation algorithm,
, and thus, the threshold is on the 68.79th quantile, implying that a total of 245 observations out of the 785 observations are extreme observations. This seems more reasonable given the nature of rainfall intensity data, where the bulk of observations of rainfall intensity are clustered around the mean, and a few other extreme rainfall intensity observations are also present in the data set. From the HNGPD model, we also obtained
, which gives the probability of exceeding the threshold
u. Thus, there exists a less than
probability of exceeding the threshold
u, and this makes a lot of sense and justifies our use of non-uniform weights for each component of the model, in contrast to the use of uniform weights.
It can also be observed from
Table 5 that the graphical approaches all selected a lower threshold value
u, beyond which a GPD is fitted to the tail of the data distribution than the one estimated using the UIA. Again, the graphical approaches suggest the presence of many outliers or extreme observations in the data set. In particular,
for the MEP approach, and thus, the threshold is on the 53.89th quantile. This further implies that a total of 362 observations out of the 785 observations are above the threshold, and hence, they are classified as extreme observations. For the Hill plot approach,
, and thus, the threshold is on the 53.50th quantile, implying that a total of 365 observations out of the 785 observations are extreme observations. For the stability plot approach,
, and thus, the threshold is on the 48.92th quantile, implying that a total of 401 observations out of the 785 observations are extreme observations. These threshold points selected by the graphical approaches are significantly lower than the ones selected by the UIA, which offered a more realistic threshold point given the nature of our data.
In conclusion, a new two-component hybrid model, suitable for modeling data with high right skewness and estimated using an unsupervised iterative estimation algorithm, has been introduced in this paper. A demonstration of the hybrid model’s flexibility and robustness in capturing the unique characteristics of such data through application to a real data set has been carried out, and the results obtained have also been compared with those of other hybrid models in its class. Moreover, through empirical analyses on synthetic data sets, the unsupervised iterative estimation algorithm for the estimation of the parameters of the hybrid model has shown high accuracy and efficiency, making it a valuable tool for further practical applications. This new hybrid model and estimation technique hold significant promise for a wide range of fields where right-skewed data sets are prevalent, such as finance, environmental studies, signal processing, biomedicine, etc. Furthermore, given that the estimation algorithms also help in the automatic selection of a threshold beyond which extreme observations are observed in a distribution, it could become an effective tool in tail risk modeling. In finance, it could allow for more accurate estimation of Value-at-Risk (VaR) and Expected Shortfall (ES), ensuring that institutions can quantify potential catastrophic losses with greater reliability. In insurance, it could improve the modeling of extreme claims arising from natural disasters or systemic shocks, supporting fair premium setting and reinsurance strategies. Within the context of environmental studies, it could enhance the prediction of rare but severe climatic events, such as extreme rainfall or heatwaves, critical for climate resilience planning. Future research could extend this work by exploring multi-component extensions and applying the model and the estimation algorithm to more diverse data sets. Additionally, integrating this hybrid model into more complex statistical frameworks, such as regression models and machine learning algorithms, could further enhance its utility and scope of application. In general, our contribution in this paper provides a substantial advancement in the modeling of data exhibiting high skewness to the right, offering a powerful and versatile tool for statisticians, data scientists, and users of statistics.