1. Introduction
In lifetime experiments, censoring is commonly used to address time and cost constraints, especially when testing modern, highly reliable items that require long testing periods in conventional studies. Various censoring techniques are applied in real-world scenarios with different objectives to improve efficiency and reduce unnecessary expenses. Censoring is widely used in fields such as clinical trials, biological studies, and industrial engineering. Two of the most commonly used methods are type-I and type-II censoring schemes, where testing ends at a predetermined time T or upon the occurrence of the r-th failure, respectively.
Epstein [
1] introduced a hybrid Type-I censoring method in which the experiment ends at
, where
represents the failure time of the
r-th unit. In contrast, Childs et al. [
2] proposed a Type-II hybrid censoring scheme that terminates at
. However, both Type-I and Type-II censoring methods share the same limitations as the Type-II hybrid censoring technique. To address these constraints, Chandrasekar et al. [
3] introduced two generalized hybrid censoring methods. Although these approaches were designed to overcome the issues associated with Type-I and Type-II hybrid censoring schemes, they still have certain drawbacks. For instance, in the generalized Type-I hybrid censoring approach, the experimenter may fail to observe the
r-th failure within the specified time frame.
To tackle these issues, Balakrishnan et al. [
4] developed unified hybrid censoring methods (UHCS). The parameters of a life testing experiment with
n objects are predetermined by the experimenter in the UHCS.
k and
, where
, and
, are the preset values. However, if the
k-th failure happens before
, the experiment stops at
. Should the
k-th failure take place between
and
, the experiment is ended at
. If the
k-th failure happens after
, the experiment is terminated at
. Due to the significance of this type of information, a number of authors have studied this topic, namely [
5,
6,
7,
8,
9,
10,
11,
12].
Consequently, we determine the six scenarios in
Table 1 in respect to the UHCS.
The likelihood function in accordance with the UHCS is
where
In this case,
Q and
A stand for the total number of test failures up until the end of time, respectively.
When analyzing real-world data on COVID-19 patient mortality rates, we often encounter information related to the time until a specific event occurs. Such data, commonly referred to as survival time data, typically exhibit a right-skewed distribution. The Gumbel Type-II distribution (G-IID), known for its positive skewness, is well-suited for modeling this type of data. Introduced by the German mathematician Gumbel [
13], the G-IID has proven valuable for modeling extreme events, including floods, earthquakes, and other natural disasters. Its applications extend to life expectancy tables, hydrology, and rainfall studies. Gumbel himself demonstrated the effectiveness of this distribution in simulating expected lifespans during product lifetime comparisons. Additionally, the G-IID plays a crucial role in predicting the probabilities of natural hazards and meteorological phenomena.
Substantial progress has been made in statistical inference methods for the G-IID. For instance, Mousa et al. [
14] investigated Bayesian estimation through simulations based on specific parameter values. Nadarajah and Kotz [
15] enhanced the beta Gumbel distribution and proposed maximum likelihood algorithms, and Miladinovic and Tsokos [
16] examined Bayesian reliability estimation for a modified Gumbel failure model using the square error loss function.
Focusing specifically on the G-IID, Feroze and Aslam [
17] investigated Bayesian estimation methods for doubly censored samples under various loss functions. Abbas et al. [
18] expanded the scope of inference by incorporating Bayesian estimation approaches for the G-IID. The Bayesian estimation of two competing units within this distribution was derived by Feroze and Aslam [
19]. Reyad and Ahmed [
20] estimated unknown shape parameters for joint Type-II censored data using Bayesian and E-Bayesian approaches. Sindhu et al. [
21] examined Bayesian estimators and their associated risks under diverse informative and non-informative priors for left Type-II censored samples. Additionally, Abbas et al. [
22] applied Lindley’s approximation to develop Bayesian estimators for Type-II censored data under a non-informative prior and multiple loss functions. More recently, Qiu and Gui [
23] derived Bayesian estimates for two G-IID under joint Type-II censoring. For
and
, the G-IID’s cumulative distribution function (CDF) is as described below:
additionally, the probability density function (PDF) that corresponds to it is given as follows:
The reliability function is defined as
and the failure rate function
is expressed as follows:
In this context, the parameters
and
denote shape and scale, respectively.
Figure 1 shows the PDF, while
Figure 2 illustrates the CDF of the G-IID for different values of the parameter
, with
held constant. The G-IID was selected for this analysis due to both its theoretical properties and empirical performance in modeling positively skewed, heavy-tailed lifetime data. This type of distribution is well-suited for capturing the nature of COVID-19 mortality data, which often exhibits a long right tail reflecting delayed fatalities or rare high-mortality days during outbreak peaks.
1.1. Why G-IID over Weibull, Gompertz, or GEV?
The Weibull distribution is widely used but often underestimates the probability of extreme outcomes due to its limited tail flexibility.
The Gompertz distribution, while biologically interpretable, is more appropriate for modeling aging-related hazards rather than acute, outbreak-driven data.
The Generalized extreme value (GEV) family is flexible but contains three parameters and may overfit small or moderately censored samples; in contrast, G-IID offers a parsimonious two-parameter alternative with similar right-tail behavior.
The Gumbel Type-II’s cumulative distribution function reflects a heavy right tail and positive skewness. The shape parameter controls the degree of skewness, while the scale parameter affects the spread. This flexibility allows the model to capture rare but significant increases in mortality; such as spikes due to emerging variants or delayed healthcare responses.
1.2. Interpretation of Parameters
Higher values of correspond to sharper declines in the hazard rate, suggesting more concentrated mortality during epidemic peaks.
Lower values of reflect long-tail behavior, i.e., mortality tapering off slowly rather than abruptly.
These properties make the G-IID suitable not only from a statistical fitting perspective but also for capturing the practical dynamics of COVID-19 mortality trends.
This research aims to develop and evaluate estimation methods for the parameters, reliability function, and hazard function of the G-IID under UHCS. It contributes by applying both classical and Bayesian approaches to estimate unknown parameters using real and simulated data. The study derives maximum likelihood estimators (MLEs), constructs asymptotic confidence intervals (ACIs) using the Fisher information matrix, and obtains Bayesian estimates under three distinct loss functions: squared error (SE), LINEX, and generalized entropy (GE). Since analytical Bayesian estimates are not available, the research utilizes Markov Chain Monte Carlo (MCMC) methods, including Metropolis–Hastings within Gibbs sampling, to generate posterior estimates and credible intervals. Through extensive simulation studies and real data application, the study highlights the efficiency and reliability of Bayesian methods, particularly under the GE loss function, in handling censored lifetime data. The work offers a practical and flexible framework for improving reliability analysis and life-testing experiments, providing new insights into the modeling of incomplete or censored data using the G-IID.
This study contributes to the Special Issue’s focus by addressing the challenge of making reliable inferences from censored and incomplete data. Using unified hybrid censoring schemes and real pandemic data, we demonstrate robust statistical estimation techniques that support uncertainty modeling and decision making in applied sciences.
The manuscript makes a substantial contribution to the field of reliability analysis and life-testing experiments by applying a UHCS alongside Bayesian estimation methods. To clarify its novelty, the work integrates both Type-I and Type-II censoring into a single unified framework, offering greater flexibility than previous studies that typically treated these schemes separately. While hybrid censoring has been applied to Gumbel-related distributions before, the unified approach presented here accommodates a wider range of experimental conditions. Additionally, the manuscript distinguishes itself through its application of Bayesian estimation under multiple loss functions—squared error, LINEX, and generalized entropy—unlike prior studies that often relied solely on maximum likelihood estimation (MLE) with a single loss criterion. The rationale for adopting Bayesian methods is well-founded; Bayesian inference allows for the incorporation of prior information and yields credible intervals, enhancing the reliability of estimates, especially under censored data scenarios, such as COVID-19 mortality datasets. These advantages are further highlighted by the limitations of MLE, including its potential bias in small samples and instability under heavy censoring, which can lead to large variances and unreliable estimates. In contrast, Bayesian methods maintain consistent performance under extreme censoring and offer robustness critical for practical applications. To reinforce this point, the manuscript would benefit from addressing the weaknesses of MLE explicitly in the Introduction. Doing so would provide a clearer rationale for the methodological choices made and prepare readers for the simulation-based comparisons discussed later. This framing would enhance the coherence of the argument and highlight the practical and theoretical strengths of the Bayesian approach within the context of the G-IID and UHCS.
This paper proceeds as follows.
Section 2 provides the survival and hazard rate functions, as well as the MLEs for the unknown parameters. The Fisher information matrix (FIM) is used to create ACIs for interval estimation in
Section 3. Bayesian parameter estimation using Gibbs sampling and importance sampling strategies under three distinct loss functions is shown in
Section 4. A real dataset is analyzed in
Section 5 to investigate parameter estimation.
Section 6 presents a Monte Carlo simulation analysis to evaluate the proposed estimates based on average width, mean squared error (MSE), and average value. Finally,
Section 7 provides the conclusions.
5. Application for COVID-19 Mortality Datasets
To illustrate and compare the different estimation methods examined in this study, this section presents real-world COVID-19 death rate data from France, as documented by Almetwally [
26]. The dataset spans a 51-day period, from 1 January to 20 February 2021. The data are as follows in
Table 2. To assess the model’s effectiveness, we conducted several goodness-of-fit tests, including the Anderson–Darling, Kolmogorov–Smirnov (KS), and Cramér–von Mises tests.
To evaluate the model’s suitability for describing the data, we analyze both distance statistics and
p-values. The null hypothesis (
) is rejected (and the alternative
is accepted) at a significance level of
if the
p-value is below 0.05. The test statistics, presented in
Table 3, show relatively high
p-values for all tests. Consequently, we fail to reject
, supporting the conclusion that the data follows the G-IID and confirming the model’s good fit to the real-world dataset.
For visual assessment,
Figure 3 compares the fitted and empirical survival functions of the G-IID, demonstrating close agreement between the two. Further validation is provided by the observed vs. expected probability plots (
Figure 4) and quantile plots (
Figure 5), reinforcing the model’s appropriateness for analyzing the dataset.
Additionally,
Figure 6,
Figure 7 and
Figure 8 depict the histogram, smoothed histogram, and box-and-whisker plot of the real data, respectively.
Figure 9,
Figure 10 and
Figure 11 illustrate the profile log-likelihood function and contour plot for the parameters
and
, confirming that it attains a unique maximum, further validating the model’s fit.
These tests capture different aspects of goodness-of-fit. The K–S test focuses on maximum deviation, the Anderson–Darling test gives more weight to tail discrepancies, and the Cramér–von Mises test assesses integrated squared error across the distribution. Using all three allows for a more comprehensive validation of the model fit.
We analyze the case where censoring is applied to the data. From the dataset, we generate six artificially constructed UHCD sets, as shown in
Table 4. We calculate and present frequentist and Bayesian estimates for the parameters
,
, the survival function
, and the hazard function
at time
, using initial values
and
. These results are displayed in
Table 5,
Table 6,
Table 7 and
Table 8, derived from the dataset in
Table 4. Additionally,
Table 9,
Table 10,
Table 11 and
Table 12 provide the interval lengths for the 95% two-sided ACI and CRI estimates.
Due to the absence of prior data for the G-IID parameters (
,
,
, and
), the Bayesian analysis relies on non-informative priors under the SE, LINEX, and GE loss functions. To mitigate the influence of initial values, we apply a burn-in period of 3000 iterations from a total of 16,000 MCMC samples, following the algorithm detailed in
Section 3. The initial guesses for
and
are based on their frequentist estimates.
7. Conclusions
The methods developed in this paper are directly relevant to the Special Issue theme, particularly in the context of data-driven modeling under censoring and uncertainty using real-life data. This study investigated the estimation of the parameters, reliability function, and hazard function of the G-IID under UHCS. Both classical and Bayesian estimation methods were considered. Bayesian estimates were obtained under three different loss functions: SE, LINEX, and GE. A Monte Carlo simulation study was conducted to evaluate and compare the performance of the proposed estimators.
Several important conclusions can be drawn from the findings:
Bayesian estimation methods produce significantly narrower CRIs compared to the classical asymptotic confidence intervals, indicating that Bayesian techniques provide higher precision in parameter estimation.
Bayesian estimates based on the GE loss function consistently achieve the lowest MSE values when compared to those based on SE and LINEX loss functions. This demonstrates the superior performance of the GE-based Bayesian approach, particularly in terms of efficiency and robustness.
Increasing the values of the censoring parameters r and k leads to a noticeable decrease in the MSEs for the estimates of , , , and . This highlights the importance of larger observed sample sizes in improving estimation accuracy.
Bayesian estimators outperform maximum likelihood estimators across all scenarios analyzed. They produce smaller MSEs for both the parameters and the reliability-related functions. This confirms the advantages of incorporating prior information and a loss function framework in the estimation process.
The CRIs resulting from Bayesian methods are not only narrower but also more stable across different censoring schemes. This reliability enhances the confidence in Bayesian interval estimation, especially in highly censored or small sample environments.
The simulation study revealed that while both classical and Bayesian methods benefit from larger sample sizes, Bayesian methods show a stronger and more consistent improvement in estimation accuracy.
Under extreme censoring conditions, where traditional methods tend to suffer from large variances and unstable behavior, Bayesian estimation maintains reliable and consistent performance, reinforcing its applicability in practical situations involving incomplete data.
Among the three loss functions considered, the GE loss function proved to be the most efficient, yielding the smallest MSEs in almost all scenarios.
Overall, the simulation results strongly support the use of Bayesian estimation techniques, particularly under the GE loss function, when analyzing lifetime data modeled by the G-IID under unified hybrid censoring schemes.
The estimated reliability function and hazard function provide critical insights into the dynamics of COVID-19 mortality. Understanding these functions can help public health officials and researchers make informed decisions regarding interventions and resource allocation. The reliability function represents the probability that an individual survives beyond time t. In the context of COVID-19, it indicates the likelihood of survival for patients diagnosed with the virus over time.
Epidemiological Meaning: A higher value at a given time indicates better survival rates, suggesting effective treatment protocols or lower virulence of the virus.
Monitoring changes in over time can help assess the impact of public health measures, such as vaccination campaigns or social distancing, on mortality rates.
If decreases significantly, it may signal the emergence of more severe variants or the need for updated treatment strategies.
The hazard function quantifies the instantaneous risk of death at time t given that the individual has survived up to that time. It reflects the rate at which deaths occur among the population at risk.
Epidemiological Meaning:
A higher indicates an increased risk of mortality, which can inform healthcare providers about the urgency of care needed for patients at different stages of the disease.
Analyzing can help identify critical time points when patients are most vulnerable, guiding the timing of interventions.
Changes in can also reflect the effectiveness of public health responses, such as the introduction of new treatments or changes in healthcare capacity.