Next Article in Journal
Dual-Performance Multi-Subpopulation Adaptive Restart Differential Evolutionary Algorithm
Next Article in Special Issue
Estimation and Bayesian Prediction for New Version of Xgamma Distribution Under Progressive Type-II Censoring
Previous Article in Journal
Research on the Method of Calculating Anchorage Flaw Length Based on the Propagation Characteristics of Bolt Excitation Stress Waves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation and Bayesian Prediction for the Generalized Exponential Distribution Under Type-II Censoring

School of Mathematics and Statistic, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(2), 222; https://doi.org/10.3390/sym17020222
Submission received: 15 January 2025 / Revised: 27 January 2025 / Accepted: 31 January 2025 / Published: 2 February 2025
(This article belongs to the Special Issue Bayesian Statistical Methods for Forecasting)

Abstract

:
This research focuses on the prediction and estimation problems for the generalized exponential distribution under Type-II censoring. Firstly, maximum likelihood estimations for the parameters of the generalized exponential distribution are computed using the EM algorithm. Additionally, confidence intervals derived from the Fisher information matrix are developed and analyzed alongside two bootstrap confidence intervals for comparison. Compared to classical maximum likelihood estimation, Bayesian inference proves to be highly effective in handling censored data. This study explores Bayesian inference for estimating the unknown parameters, considering both symmetrical and asymmetrical loss functions. Utilizing Gibbs sampling to produce Markov Chain Monte Carlo samples, we employ an importance sampling approach to obtain Bayesian estimates and compute the corresponding highest posterior density (HPD) intervals. Furthermore, for one-sample prediction and, separately, for the two-sample case, we provide the corresponding posterior distributions, along with methods for computing point predictions and predictive intervals. Through Monte Carlo simulations, we evaluate the performance of Bayesian estimation in contrast to maximum likelihood estimation. Finally, we conduct an analysis of a real dataset derived from deep groove ball bearings, calculating Bayesian point predictions and predictive intervals for future samples.

1. Introduction

In 1999, Reference [1] first formally introduced the generalized exponential distribution. It can also be regarded as a particular member of the three-parameter Exponentiated Weibull (EW) distribution, introduced by Reference [2]. Furthermore, its origins can be traced back to the early work of Gompertz and Verhulst, who introduced cumulative distribution functions for modeling human mortality and population growth. Interested readers can refer to References [3,4] for more details.
The probability density function (PDF) and cumulative distribution function (CDF) of generalized exponential distribution can be expressed as:
f ( x ) = α λ e λ x 1 e λ x α 1 , x > 0 , α > 0 , λ > 0 ,
F ( x ) = 1 e λ x α ,
The hazard rate and reliability functions can be expressed as:
h ( x ) = α λ e λ x ( 1 e λ x ) α 1 1 ( 1 e λ x ) α ,
R ( x ) = 1 ( 1 e λ x ) α .
Figure 1 illustrates the PDF plots for the chosen parameter values. Based on Figure 1, the density function demonstrates different behaviors depending on α . For α 1 , it is monotonically decreasing, whereas for α > 1 , it becomes unimodal and right-skewed, similar to Weibull or Gamma distributions. The hazard rate plots with different parameter values are shown in Figure 2. It shows that the hazard function decreases when α < 1 , remains constant when α = 1 , and increases when α > 1 . Notably, the hazard function eventually converges to the value of λ .
There are some interesting physical interpretations for the generalized exponential distribution. It defines the distribution of the maximum among α random variables that are independently and exponentially distributed when α is a positive integer greater than 1. This distribution applies to situations such as the overall lifetime of parallel systems of electronic components, each with an exponential lifetime distribution. In contrast, for a series system, the overall lifetime corresponds to the minimum of the exponentially distributed lifetimes, which follows an exponential distribution with rate parameter α λ . Figure 3 presents a schematic representation of both series and parallel systems.
The generalized exponential distribution has gained significant scholarly interest due to its practical importance, particularly in situations involving censored data. This attention has led to a wealth of research focusing on its statistical properties. Reference [1] conducted a comprehensive study of its properties, addressing maximum likelihood estimation (MLE) along with its asymptotic behavior and issues related to hypothesis testing.
Reference [5] compared the generalized exponential distribution with the Weibull and Gamma distributions, demonstrating its advantages in modeling skewed life data. Further, Reference [6] investigated various parameter estimation methods and evaluated their performance.
Reference [7] studied estimation issues for the generalized exponential distribution under hybrid censoring, utilizing the EM algorithm to compute MLEs and importance sampling to derive Bayesian estimates. The Bayesian estimation for the parameter using doubly censored samples was derived by Reference [8]. Reference [9] applied both classical and Bayesian approaches for the estimation of parameters as well as the reliability and hazard functions based on adaptive Type-II progressive censored data. For more detailed results and examples related to the generalized exponential distribution, please refer to References [10,11,12,13].
However, research on Bayesian prediction problems remains relatively scarce for generalized exponential distribution. Prediction is an important aspect in life-testing experiments, where data from an existing sample is used to make inferences about future observations. Several studies have explored prediction problems in various contexts. For instance, Reference [14] provided a comprehensive and detailed review of Bayesian prediction techniques and their diverse applications. Additionally, Reference [15] studied one-sample and two-sample Bayesian prediction problems for the inverse Weibull distribution under Type-II censoring. Similarly, Reference [16] investigated the Poisson-Exponential distribution by employing both classical and Bayesian approaches, providing point predictions and credible predictive intervals.
This paper focuses on both one-sample and two-sample prediction problems. In the one-sample prediction case, let X 1 < < X r represent the observed informative sample and X r + 1 < < X n denote the future sample. The aim is to predict and make inferences about the future order statistics X k for r < k n . In the two-sample case, the observed sample of size r is denoted by X 1 < < X r , and the future ordered sample of size m is represented by Y 1 < < Y m . These two samples are based on the same distribution and assumed to be independent, enabling predictions and inferences for the order statistics Y 1 < < Y m .
The primary objectives of this study are twofold. Firstly, we aim to obtain the MLEs and Bayesian estimates under the Type-II censoring scheme and to compare their performance. Secondly, we seek to derive the Bayesian posterior predictive density for future order statistics and, based on this posterior predictive density, compute both point predictions and predictive intervals.
This paper proceeds as follows: In Section 2, the EM algorithm is employed to approximate the MLE of the model. For interval estimation, the Fisher information matrix is used to construct asymptotic confidence intervals (ACIs). Additionally, two bootstrap-based interval estimation methods are explored as comparative alternatives to evaluate their performance. Section 3 presents Bayesian parameter estimation under three different loss functions, utilizing Gibbs sampling and importance sampling techniques. HPD credible intervals are further obtained from the Gibbs sampling results. Section 4 focuses on Bayesian point predictions and the computation of predictive intervals for future order statistics. In Section 5, simulation studies are conducted to evaluate the performance of the proposed methodologies, followed by an analysis of a real-world dataset related to deep groove ball bearings, which is discussed in Section 6. Finally, we present conclusions in Section 7.

2. Maximum Likelihood Estimation

Let x 1 < x 2 < < x r denote a sample containing r observations obtained from the generalized exponential distribution under a Type-II censoring scheme characterized by R = ( n , r ) . The likelihood function associated with this sample is given by:
L ( α , λ ) = n ! ( n r ) ! i = 1 r α λ e λ x i 1 e λ x i α 1 · 1 1 e λ x r α n r .
Neglecting the constant term, the log-likelihood function is:
l ( α , λ ) = r ln α λ + ( α 1 ) i = 1 r ln 1 e λ x i λ i = 1 r x i + ( n r ) ln 1 1 e λ x r α .
To compute the MLEs, we differentiate (6) with respect to α and λ and set the derivatives to zero. The partial derivatives are:
l ( α , λ ) α = r α + i = 1 r ln 1 e λ x i ( n r ) 1 e λ x r α ln 1 e λ x r 1 1 e λ x r α = 0 ,
l ( α , λ ) λ = r λ + i = 1 r ( α 1 ) x i e λ x i 1 i = 1 r x i ( n r ) α 1 e λ x r α 1 e λ x r x r 1 1 e λ x r α = 0 .
However, solving these equations analytically is not feasible. Therefore, numerical methods, such as the EM algorithm, are recommended. Initially introduced in Reference [17], the EM algorithm is a suitable approach for this scenario.
Under Type-II censoring framework, the observed data, represented as X = ( x 1 , x 2 , , x r ) , corresponds to the first r observations. The remaining n r data points, which are censored and unobserved, are denoted as Z = ( z 1 , z 2 , , z n r ) . Together, these two components form the complete sample W, where W = ( X , Z ) .
When constants are excluded, the log-likelihood function for W, represented as L c ( α , λ ; W ) , can be expressed as:
L c ( α , λ ; W ) = n ln ( α λ ) λ i = 1 r x i λ i = 1 n r z i + ( α 1 ) i = 1 r ln 1 e λ x i + ( α 1 ) i = 1 n r ln 1 e λ z i .
The EM algorithm iteratively computes the MLEs by alternating between the E-step and M-step. In the E-step, the expectation of the log-likelihood function is calculated given the observed data X and the current parameter estimates. These estimates are updated in the M-step by maximizing the expected complete log-likelihood function. Therefore, we first derive the expression for the pseudo-log-likelihood function as follows:
L s ( α , λ ) = n ln α λ λ i = 1 r x i + i = 1 r ( α 1 ) ln ( 1 e λ x i ) i = 1 n r λ E ( z i | z i > x r ) + i = 1 n r ( α 1 ) E ( ln ( 1 e λ z i ) | z i > x r ) ,
where
E ( z i | z i > x r ) = 1 1 F ( x r ) x r x f ( x ) d x , = Q 1 ( α , λ , x r )
E ( ln ( 1 e λ z i ) | z i > x r ) = 1 1 F ( x r ) x r ln ( 1 e λ x ) f ( x ) d x . = Q 2 ( α , λ , x r )
In the M-step, suppose the estimates of α and λ at j-th iteration are denoted as α ( j ) and λ ( j ) , respectively. The new estimates can be calculated by maximizing the following expression:
L s ( α , λ ) = n ln α λ λ i = 1 r x i + ( α 1 ) i = 1 r ln ( 1 e λ x i ) λ ( n r ) Q 1 ( α ( j ) , λ ( j ) , x r ) + ( α 1 ) ( n r ) Q 2 ( α ( j ) , λ ( j ) , x r ) .
Here, we introduce an application of the iteration method, which is similar to the method proposed by Reference [5]. In order to maximize (11), we first take the partial derivative with respect to α and λ .
L s ( α , λ ) α = n α + i = 1 r ln ( 1 e λ x i ) + ( n r ) Q 2 ( α ( j ) , λ ( j ) , x r ) = 0 ,
L s ( α , λ ) λ = n λ i = 1 r x i + ( α 1 ) i = 1 r e λ x i x i 1 e λ x i ( n r ) Q 1 ( α ( j ) , λ ( j ) , x r ) = 0 .
From (12), the estimate of α at j + 1 -th iteration can be determined as as follows:
α ^ ( λ ) = n ( n r ) Q 2 ( α ( j ) , λ ( j ) , x r ) + i = 1 r ln ( 1 e λ x i ) ,
Thus, the maximization of (11) can be achieved by solving the corresponding fixed-point equation:
λ = ϕ ( λ ) ,
where
ϕ ( λ ) = 1 n i = 1 r x i + n r n Q 1 ( α ( j ) , λ ( j ) , x r ) 1 n i = 1 r α ^ ( λ ) 1 e λ x i x i 1 e λ x i 1 .
After computing λ ( j + 1 ) , the corresponding value of α ( j + 1 ) can be determined using the relation α ( j + 1 ) = α ^ ( λ ( j + 1 ) ) .
The iterative procedure described in Equations (14) and (15) is deemed to have converged when the following inequality holds:
α ( j + 1 ) α ( j ) + λ ( j + 1 ) λ ( j ) < τ ,
where τ denotes a small positive threshold to ensure accuracy and stability. The above process of the EM algorithm is summarized in the following Algorithm 1:
Algorithm 1 EM Algorithm for MLEs under Type-II Censoring
1:
Input: Initial values ( α ( 0 ) , λ ( 0 ) ) , threshold τ .
2:
Initialization: Set j = 0 .
3:
repeat
4:
   Calculate the expectations Q 1 ( x r , α ( j ) , λ ( j ) ) and Q 2 ( x r , α ( j ) , λ ( j ) ) .
5:
   Solve the fixed-point equation to obtain λ ( j + 1 ) using Equation (15).
6:
   Update α ( j + 1 ) using α ( j + 1 ) = α ^ ( λ ( j + 1 ) ) from Equation (14).
7:
   Increment j = j + 1 .
8:
until
α ( j + 1 ) α ( j ) + λ ( j + 1 ) λ ( j ) < τ .
In Section 5, when applying the EM algorithm for MLEs under Type-II censoring, the MLEs obtained from complete samples are used as the initial values.

2.1. Fisher Information Matrix

Reference [18] introduces several core concepts closely related to the Fisher information matrix. Building on this foundation, this subsection describes the approach for obtaining the observed information matrix, following the missing value principles outlined in Reference [19], which is then used to construct confidence intervals.
Specifically, let θ = ( α , λ ) , where I X ( θ ) represents the observed information matrix, I W | X ( θ ) corresponds to the missing information matrix, and I W ( θ ) indicates the complete information matrix. These matrices are related as follows:
I X ( θ ) = I W ( θ ) I W | X ( θ ) .
The missing information matrix, representing the Fisher information associated with the censored data, is expressed as:
I W | X ( θ ) = ( n r ) I Z | X ( θ ) , = ( n r ) E Z | X 2 ln f Z X ( z x r , θ ) θ 2 ,
where
f Z X z x r , θ = f ( z ) 1 F ( x r ) , z > x r .
The complete information matrix, base on the log-likelihood function (7) is then given by:
I W ( θ ) = E 2 ln L c ( θ ; W ) θ 2 .
The details of the elements within these two matrices are elaborated in Appendix A. We can calculate I X ( θ ) using the derived components of I W ( θ ) and I W | X ( θ ) easily. The variance-covariance matrix corresponding to the MLEs, α ^ and λ ^ , is then approximated by taking the inverse of I X ( θ ) , represented as I X 1 ( α ^ , λ ^ ) .
On the basis of this approximation, the 100 ( 1 p ) % asymptotic confidence intervals for α and λ are given as:
α ^ η p / 2 D ( α ^ ) , α ^ + η p / 2 D ( α ^ ) , λ ^ η p / 2 D ( λ ^ ) , λ ^ + η p / 2 D ( λ ^ ) ,
where, D ( α ^ ) and D ( λ ^ ) represent the diagonal elements of the inverse observed information matrix, denoted as I X 1 ( θ ^ ) . The critical value η p / 2 corresponds to the upper p / 2 quantile of the standard normal distribution.

2.2. Bootstrap Methods

The asymptotic confidence interval approach is based on the assumption that the estimators follow a normal distribution, an assumption that holds true when the sample size is adequately large. Nevertheless, in practical, the sample size is often limited. To address this, we propose two bootstrap techniques. The first is the percentile bootstrap method [20], which generates bootstrap samples by resampling the observed data. The second method, the bootstrap-t approach [21], also generates bootstrap samples but computes a bootstrap-t statistic that adjusts the estimates using their standard deviation.
Algorithms 2 and 3 present the procedures for implementing the Boot-p and Boot-t methods, respectively.
One potential direction for future research involves exploring the methods proposed in References [22,23] for identifying extreme values, which may help improve the accuracy of the bootstrap method.
Algorithm 2 Percentile Bootstrap (Boot-p) Method
1:
Input: the Type-II censored sample X; the number of bootstrap replications M; the censoring scheme R = ( n , r ) .
2:
Compute the MLEs ( α ^ , λ ^ ) for the generalized exponential distribution using the censored sample X.
3:
for  j = 1 to M do
4:
   Generate a bootstrap sample from the generalized exponential distribution parameterized by ( α ^ , λ ^ ) under the censoring scheme R.
5:
   Calculate the MLEs ( ( α ^ j * , λ ^ j * ) ) using the bootstrap sample.
6:
end for
7:
Arrange the bootstrap estimates { φ ^ j * } in ascending order to obtain the sorted set { φ ^ ( 1 ) * , φ ^ ( 2 ) * , , φ ^ ( M ) * } , where φ ^ is the MLE obtained from the original sample X and φ ^ j * is the bootstrap estimate of φ .
8:
Determine the percentile values for the confidence interval:
φ ^ Boot - p p 2 = φ ^ ( p 2 M ) * , φ ^ Boot - p 1 p 2 = φ ^ ( ( 1 p 2 ) M ) * ,
where Δ denotes the greatest integer less than or equal to Δ .
9:
Output: The 100 ( 1 p ) % Boot-p confidence interval: φ ^ Boot - p p 2 , φ ^ Boot - p 1 p 2 .
Algorithm 3 Bootstrap-t (Boot-t) Method
1:
Input: the Type-II censored sample X; the number of bootstrap replications M; the censoring scheme R = ( n , r ) .
2:
Compute the MLEs ( α ^ , λ ^ ) for the generalized exponential distribution using the censored sample X.
3:
for  j = 1 to M do
4:
   Generate a bootstrap sample from the generalized exponential distribution parameterized by ( α ^ , λ ^ ) under the censoring scheme R.
5:
   Calculate the MLEs ( ( α ^ j * , λ ^ j * ) ) using the bootstrap sample.
6:
   Calculate the bootstrap-t statistic for the parameter φ :
T j * = φ ^ j * φ ^ D ( φ ^ j * ) ,
7:
end for
8:
Sort the bootstrap-t statistics { T j * } in ascending order to get the ordered set { T ( 1 ) * , T ( 2 ) * , , T ( M ) * } .
9:
Compute T L * = T ( p 2 M ) * , T U * = T ( ( 1 p 2 ) M ) * and pick their corresponding MLEs φ ^ L * , φ ^ U * .
10:
Compute the bounds of the Boot-t confidence interval for φ :
φ ^ Boot - t p 2 = φ ^ + D ( φ ^ L * ) · T L * , φ ^ Boot - t 1 p 2 = φ ^ + D ( φ ^ U * ) · T U * .
11:
Output: The 100 ( 1 p ) % Boot-t confidence interval:
φ ^ Boot - t p 2 , φ ^ Boot - t 1 p 2 .

3. Bayesian Estimation

This section focuses on deriving the Bayes estimates for α , λ using type-II censored data within the framework of a specified loss function. Selecting an appropriate loss function is crucial, as it determines the penalty associated with estimation errors. In this study, we examine three types of loss functions: the general entropy loss function (GELF), the Linex loss function (LILF), and the squared error loss function (SELF). The SELF is widely used due to its symmetric nature and is most appropriate when the estimation error leads to symmetric consequences. For scenarios where the consequences of overestimation and underestimation differ, the Linex loss function, which introduces an asymmetry in the penalties, is more suitable. Meanwhile, the GELF accounts for entropy-based considerations, allowing greater flexibility in modeling uncertainty and estimation inaccuracies. Let θ ^ denote an estimate of the θ . The mathematical expression of the above three loss functions is provided below:
L G E ( θ ^ ) = ( θ ^ / θ ) δ δ ln ( θ ^ / θ ) 1 , δ 0 ,
L L I ( θ ^ ) = e h ( θ ^ θ ) h ( θ ^ θ ) 1 , h 0 ,
L S E ( θ ^ ) = ( θ θ ^ ) 2 .
The Bayes estimates under each of these loss functions are defined as follows:
Bayes estimate for θ under GELF:
θ ^ GE = E 1 θ δ X 1 δ , δ 0 .
Bayes estimate for θ under LILF:
θ ^ LI = 1 h ln E e h θ X , h 0 .
Bayes estimate for θ under SELF:
θ ^ SE = E [ θ X ] .
This setup allows for flexibility in choosing an appropriate loss function based on the characteristics of the estimation problem. Following Reference [24], let us assume that α and λ follow gamma prior distributions, expressed as:
π 1 ( α ) α u 1 e v α , α > 0 ,
π 2 ( λ ) λ s 1 e b λ , λ > 0 .
where all hyper-parameters u, v, s, and b are known and positive.
The joint posterior distribution of α and λ given type-II censored data X is
π ( α , λ X ) = π 1 ( α ) π 2 ( λ ) L ( α , λ ) 0 0 π 1 ( α ) π 2 ( λ ) L ( α , λ ) d α d λ .
Since the denominator is a normalizing constant, we can simplify to obtain
π ( α , λ | X ) α r + u 1 λ r + s 1 e α ( i = 1 r ln ( 1 e λ x i ) + v ) e λ ( i = 1 r x i + b ) × 1 ( 1 e λ x r ) α n r / i = 1 r ( 1 e λ x i ) . h 1 ( λ | X ) h 2 ( α | λ , X ) g ( α , λ | X ) ,
here
g ( α , λ | X ) = ( 1 ( 1 e λ x r ) α ) n r i = 1 r ( 1 e λ x i ) · ( i = 1 r ln ( 1 e λ x i ) + v ) r + u , h 1 ( λ | X ) g a m m a ( r + s , i = 1 r x i + b ) , h 2 ( α | λ , X ) g a m m a ( r + u , i = 1 r ln ( 1 e λ x i ) + v ) .
It is evident that the quantities r + s , i = 1 r x i + b , r + u , and i = 1 r ln ( 1 e λ x i ) + v are all positive.
Therefore, the posterior expectation of any function u ( α , λ ) can be given directly as:
E ( u ( α , λ ) X ) = 0 0 u ( α , λ ) g ( α , λ X ) h 1 ( λ α , X ) h 2 ( α X ) d α d λ 0 0 g ( α , λ X ) h 1 ( λ α , X ) h 2 ( α X ) d α d λ .
Based on (30) and (31), Gibbs sampling is initially applied to draw Monte Carlo Markov Chain (MCMC) samples from π ( α , λ X ) . These samples are subsequently utilized in the importance sampling algorithm outlined in Algorithm 4 to approximate the values in (24)–(26).
Algorithm 4 Importance Sampling for Bayesian Estimation
1:
for  j = 1 to N do
2:
   Sample λ j from h 1 ( λ data ) Gamma r + s , i = 1 r x i + b .
3:
   Sample α j from h 2 ( α λ j , data ) Gamma r + u , i = 1 r ln ( 1 e λ j x i ) + v .
4:
end for
5:
Compute the Bayesian estimate of θ using three different loss functions:
  • General Entropy Loss (GELF):
    θ ^ G E j = 1 N θ j δ g ( α j , λ j | X ) j = 1 N g ( α j , λ j | X ) 1 / δ .
  • Linex Loss (LILF):
    θ ^ L L 1 h ln j = 1 N e h θ j g ( α j , λ j | X ) j = 1 N g ( α j , λ j | X ) .
  • Squared Error Loss (SELF):
    θ ^ S E 1 N j = 1 N θ j g ( α j , λ j | X ) 1 N j = 1 N g ( α j , λ j | X ) .
The credible interval for θ can be constructed using the methodology introduced by Reference [25]. Let π ( θ X ) represent the posterior density function of θ , while Π ( θ X ) corresponds to its cumulative distribution function. The p-quantile of θ , denoted as θ ( p ) , is expressed as:
θ ( p ) = inf { θ : Π ( θ X ) p } ,
For any specific value θ * , the posterior cumulative distribution function is expressed as:
Π ( θ * X ) = E [ 1 θ θ * X ] ,
where 1 θ θ * is an indicator function. An approximation for Π ( θ * X ) can be written as:
Π ( θ * X ) = 1 N j = 1 N 1 θ j θ * g ( α j , λ j X ) 1 N j = 1 N g ( α j , λ j X ) .
Let { θ ( j ) } denote the ordered values of { θ k } , and define ω ( j ) as:
ω ( j ) = g ( α ( j ) , λ ( j ) X ) j = 1 N g ( α ( j ) , λ ( j ) X ) ,
for j = 1 , , N . Using this, Π ( θ * X ) can be estimated as:
Π ^ ( θ * X ) = 0 , if θ * < θ ( 1 ) , j = 1 k ω ( j ) , if θ ( k ) θ * < θ ( k + 1 ) , 1 , if θ * θ ( N ) .
θ ( p ) , can subsequently be estimated using the following expression:
θ ^ ( p ) = θ ( 1 ) , if p = 0 , θ ( k ) , if j = 1 k 1 ω ( j ) < p j = 1 k ω ( j ) .
In order to construct the 100 ( 1 p ) % HPD interval for θ , we first establish the following definition:
R j = θ ^ ( j / N ) , θ ^ ( 1 p + j ) N / N ,
where j = 1 , , p N . The HPD interval is then identified by selecting R j * , the interval with the shortest length.

4. Bayesian Prediction

Bayesian prediction for future observations is explored in this section. In this study, we consider Bayesian prediction within the framework of both one-sample and two-sample cases. We provide the corresponding posterior distributions, along with methods for computing point predictions and predictive intervals.

4.1. One-Sample Prediction

Let X 1 < < X r represent the observed informative sample and X r + 1 < < X n denote the future sample. We aim to predict and make inferences about the future order statistic X k (where r < k n ). Noting that the conditional density function of X k can be written as
f X k ( y | α , λ , X ) = ζ F ( y ) F ( x r ) k r 1 f ( y ) 1 F ( y ) n k 1 F ( x r ) n r ,
where ζ = ( k r ) n r k r . By applying the binomial expansion, the function mentioned above can be reformulated as:
f X k ( y | α , λ , X ) = ζ i = 0 k r 1 k r 1 i ( 1 ) k r 1 i ( 1 F ( y ) ) n r 1 i ( 1 F ( x r ) ) i n + r f ( y ) .
Therefore the corresponding survival function of X k given X is then expressed as:
S X k ( t | α , λ , X ) = t f X k ( y | α , λ , X ) d y x r f X k ( y | α , λ , X ) d y ,
= i = 0 k r 1 k r 1 i ( 1 ) k r 1 i 1 1 e λ t α n r i n r i 1 1 e λ x r α i n + r i = 0 k r 1 k r 1 i ( 1 ) k r 1 i 1 n r i .
The posterior predictive density function of X k can subsequently be expressed as:
f X k * ( y | X ) = 0 0 π ( α , λ | X ) f X k ( y | α , λ , X ) d λ d α ,
and the posterior survival function is
S X k * ( y | X ) = 0 0 π ( α , λ | X ) S X k ( y | α , λ , X ) d λ d α .
Let { ( α j , λ j ) ; j = 1 , , N } represent the samples obtained through Algorithm 4. The corresponding simulation-consistent estimates for f X k * and S X k * can then be approximated by:
f * ^ X k ( y | X ) = 1 N j = 1 N ω j · f X k ( y | α j , λ j , X ) ,
and
S * ^ X k ( y | X ) = 1 N j = 1 N ω j · S X k ( y | α j , λ j , X ) ,
where ω j is defined as
ω j = g ( α j , λ j | X ) j = 1 N g ( α j , λ j | X ) .
The two-sided 100 ( 1 p ) % symmetric predictive interval ( L 0 , U 0 ) for X k is the solution to the following equations:
S X k * ( L 0 | X ) = 1 p 2 and S X k * ( U 0 | X ) = p 2 .
The point prediction for X k can be determined by utilizing the corresponding predictive distribution. Specifically, it is computed as:
X k ^ = x r y f X k * ( y | X ) d y = 0 0 x r y · f X k ( y | X ) π ( α , λ | X ) d y d λ d α = 0 0 I k ( x r ; α , λ ) π ( α , λ | X ) d λ d α ,
where
I k ( x r ; α , λ ) = ζ i = 0 k r 1 k r 1 i ( 1 ) k r 1 i ( 1 F ( x r ) ) i n + r x r y ( 1 F ( y ) ) n r 1 i f ( y ) d y .
We cannot compute the above expression analytically. Therefore, we approximate X k ^ using the previously drawn samples:
X k ^ = 1 N j = 1 N ω j · I k ( x r ; α j , λ j ) .

4.2. Two-Sample Prediction

This section addresses the problem of dealing with two distinct sets of samples, the first, identified as the informative sample, and the second, known as the future sample. The future order statistics, represented as Y 1 < < Y m , are considered to be statistically independent of the informative sample, X 1 < < X r . The main objective is to derive the predictive density for the k-th order statistic Y k , conditioned on the observed sample X. The density function of Y k can be expressed as follows:
g Y k ( y | α , λ , X ) = κ 1 F ( y ) m k F ( y ) k 1 f ( y ) ,
where κ = m ! ( m k ) ! ( k 1 ) ! . If we let g * Y k ( y | X ) denote the posterior predictive density of Y k , then it can be expressed as:
g * Y k ( y | X ) = 0 0 π ( α , λ | X ) g Y k ( y | α , λ , X ) d λ d α .
The function G Y k ( y | α , λ , X ) represents the survival function of g Y k ( y | α , λ , X ) , given by:
G Y k ( y | α , λ , X ) = κ y 1 F ( u ) m k F ( u ) k 1 f ( u ) d u = κ F ( y ) 1 ( 1 ξ ) m k ξ k 1 d ξ .
The predictive survival function of Y k , denoted by G * Y k ( y | X ) , is therefore given by:
G * Y k ( y | X ) = 0 0 π ( α , λ | X ) G Y k ( y | α , λ , X ) d λ d α .
In practical applications, g * Y k ( y | X ) can be approximated based on importance sampling method:
g * ^ Y k ( y | X ) = 1 N j = 1 N ω j · g Y k ( y | α , λ , X ) ,
and G * Y k ( y | X ) can be approximated by:
G * ^ Y k ( y | X ) = 1 N j = 1 N ω j · G Y k ( y | α , λ , X ) ,
where ω j is defined as in (40). The two-sided 100 ( 1 p ) % symmetric predictive interval ( L 1 , U 1 ) for Y ( k ) is the solution to the following equations:
G * Y k ( L 1 | X ) = 1 p 2 and G * Y k ( U 1 | X ) = p 2 .
To obtain the point prediction for Y k , we calculate it based on the corresponding predictive distribution g * Y k ( y | X ) .
Y ^ k = 0 y · g * Y k ( y | X ) d y = 0 0 H k ( α , λ ) π ( α , λ | X ) d λ d α ,
where
H k ( α , λ ) = 0 y · g Y k ( y | α , λ , X ) d y .

5. Simulation

Monte Carlo simulation results are provided to evaluate the performance of Bayesian estimation methods in contrast to classical estimation. The comparisons are conducted under various censoring schemes and a range of parameter settings. The analysis was performed on a system equipped with an 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30 GHz processor. Here, we used R software (version 4.4.1) for all computations.
To begin with, Type-II censored samples are generated under various censoring schemes, assuming that they follow a generalized exponential distribution with the true parameter values α ˜ = 2 and λ ˜ = 1 . The EM algorithm is subsequently applied to calculate the MLEs. Then, we set the hyper-parameters ( u , v , s , b ) the values ( 2 , 1 , 2 , 1 ) . Under these settings, Bayesian estimates are computed using various loss functions through importance sampling techniques. The simulation is performed over 3000 iterations, and the mean estimates of α ^ and λ ^ are computed. Additionally, a comparison is performed between the MLEs and Bayesian estimates, utilizing the mean squared error (MSE) as the primary criterion.
The definitions for the mean and mean squared error of the estimates α ^ and λ ^ are as follows:
M S E α ^ = 1 M j = 1 M ( α ^ j α ˜ ) 2 and M S E λ ^ = 1 M j = 1 M ( λ ^ j λ ˜ ) 2 ,
M e a n α ^ = 1 M j = 1 M α ^ j and M e a n λ ^ = 1 M j = 1 M λ ^ j ,
where α ^ j and λ ^ j are derived from the j-th simulation. Here, j = 1 , 2 , , M , with M set to 3000 for this study.
Furthermore, we explore the effect of different values of h in the LILF and δ in the GELF to analyze their influence on the Bayesian estimates.
Under various Type-II censoring schemes, Table 1 and Table 2 provide the average estimates and MSEs for the parameters α and λ . The results show that the Bayesian estimates outperform the ML method. Notably, when utilizing the GELF with δ = 0.8 , the Bayesian estimates stand out by exhibiting superior performance, achieving the smallest MSEs, and closely approximating the actual parameter values.
Table 3 summarizes the interval estimates for the unknown parameters under various censoring schemes, along with the associated coverage probabilities (CPs) and average lengths (ALs).
The findings reveal that the HPD intervals consistently have the shortest ALs among all the interval estimation methods considered, followed by the asymptotic confidence intervals, while the bootstrap confidence intervals have the longest ALs, ensuring that the intervals are broad enough to cover the asymmetry. Moreover, as n and r increase, the ALs of different interval estimates tend to decrease. In terms of CPs, the asymptotic intervals achieve the highest CPs, with the HPD intervals coming next and the bootstrap intervals showing the lowest CPs. Therefore, considering both AL and CP, the HPD intervals demonstrate superior performance compared to the asymptotic and bootstrap intervals.
Next, we evaluate the feasibility and robustness of the model under various parameter combinations. For the fixed censoring scheme ( n = 40 , r = 30 ) , seven different parameter combinations are specified. For each combination, the MLEs, Bayesian estimates, and the corresponding 95 % interval estimates are computed. Table 4 presents the mean estimates along with the MSEs (in brackets), which are used to evaluate the accuracy of the estimates for different parameter combinations. Table 5 provides the confidence and credible interval estimates corresponding to each parameter combination. Consistent with the previous discussion, the Bayesian estimates exhibit smaller MSEs compared to the MLEs. Additionally, the HPD intervals are characterized by the shortest average lengths.
Then, we explore the impact of prior distribution selection on Bayesian estimation. Similar to the previous discussion, we set the true values to be α ˜ = 2.6 and λ ˜ = 1.2 . We consider two types of prior distributions: the non-informative prior, where the hyper-parameters u, v, s, and b are set to 0, and the informative prior, where u = 2.6 , v = 1 , s = 1.2 , and b = 1 . In this case, the actual value of the parameter is typically considered as the expectation of the prior distribution. The results are provided in Table 6. It can be concluded from the table that Bayesian estimates with informative priors yield smaller MSEs compared to those with non-informative priors. This indicates that incorporating prior information in the Bayesian procedure enhances the precision of the estimates.
Subsequently, we consider the Bayesian prediction problem. First, we employ the inverse transformation method to generate n = 30 random numbers from the generalized exponential distribution, with the true values set as α ˜ = 2 and λ ˜ = 1 . These generated values are presented in Table 7. Following this, we define different censoring schemes, namely ( 30 , 10 ) , ( 30 , 15 ) , ( 30 , 20 ) , and ( 30 , 25 ) . For one-sample prediction, we compute point predictions and 95 % interval predictions for k = 26 , 28, and 30. The detailed results are displayed in Table 8. For two-sample prediction, we set m = 30 and calculated both point predictions and predictive intervals. The results are summarized in Table 9.

6. Data Analysis

This section provides an illustrative example by analyzing a dataset reported by Lawless [26]. The detailed observed data are shown in Table 10.
This dataset has been analyzed in previous studies, with Reference [5] highlighting the effectiveness of the generalized exponential distribution in modeling this data.
To analyze real data, we begin by calculating the MLEs using the complete data, followed by computing the Kolmogorov–Smirnov (K-S) statistic, Akaike Information Criterion (AIC), and Bayesian Information Criterion (BIC). Additionally, in some studies, the Anderson–Darling (A-D) statistic has been found to be more effective than the K-S statistic. For further details on the A-D statistic and its associated probability calculations, please refer to Reference [27]. We also compare the goodness-of-fit for other life distributions, such as the Weibull, Log-Normal, and Gamma distributions. The PDFs of these distributions are provided below:
Gamma : f Gamma ( x ) = λ α x α 1 e λ x Γ ( α ) , x > 0 ,
Weibull : f Weibull ( x ) = α λ x λ α 1 e ( x / λ ) α , x > 0 ,
Log - Normal : f Log - Normal ( x ) = 1 x σ 2 π exp ( ln x μ ) 2 2 σ 2 , x > 0
Table 11 presents the test results for each distribution. Typically, better model fit is indicated by smaller values of K-S, A-D, AIC, and BIC, along with a higher log-likelihood value. As shown in the table, the generalized exponential distribution provides a good fit for this dataset. To further visualize the model performance, Figure 4 includes two plots based on the MLEs. The first plot compares the fitted CDFs of the four distributions mentioned above. The second plot displays the fitted PDFs of these four distributions alongside a histogram of the dataset.
Now, we consider the situation where the final three observations are censored, denoted as r = 20 . To estimate the unknown parameters, both Bayesian estimation and maximum likelihood estimation approaches are utilized. A summary of the estimation results is presented in Table 12.
The corresponding 95% asymptotic confidence intervals for ( α , λ ) are ( 2.5512 , 8.5386 ) and ( 0.0225 , 0.0412 ) , respectively. Meanwhile, the 95% HPD credible intervals are ( 2.5123 , 6.4661 ) for α and ( 0.0222 , 0.03426 ) for λ . Notably, the HPD intervals are shorter than the asymptotic confidence intervals.
Now, we make predictive inference for the 21st, 22nd, and 23rd future order statistics, denoted as X 21 , X 22 , and X 23 , using (41) and (44). We assume that the hyper-parameters are the same as in Section 5. From Table 13, it is evident that the predicted values closely resemble the true values, thereby corroborating the effectiveness of our theoretical model. The Bayesian point prediction for the 21st order statistic is 127.92, with a 95% predictive interval ranging from 106.1134 to 147.1283. This suggests that, based on the observed sample, the 21st deep groove ball bearing is expected to fail between 106.11 and 147.13 million revolutions. Figure 5 illustrates the posterior predictive density functions for the censored observations. Additionally, the predictive survival functions for X 21 , X 22 , and X 23 are displayed in Figure 6. The plots clearly indicate that as k increases, the rate of decline in the predictive survival function decreases.
We now focus on the two-sample prediction problem discussed in Section 4.2. Assume 23 new deep groove ball bearings, denoted as Y 1 , Y 2 , , Y 23 , are subjected to the same test. The goal is to derive the predictive density and make predictive inference for Y k .
Figure 7 illustrates the point predictions and predictive intervals for the future samples derived from the observed dataset. Detailed numerical results are provided in Table 14. The Bayesian point prediction for the median is 65.36, with a 95% predictive interval for the median ranging from 42.1386 to 91.5963. This suggests that, based on the observed sample, the 12th deep groove ball bearing is expected to fail between 42.14 and 91.60 million revolutions. The posterior predictive density functions for Y 21 to Y 23 are shown in Figure 8. It can be observed that as k increases, the expected values of Y k shift rightward, and the variances increase. Additionally, Figure 9 presents the predictive survival functions for Y 21 , Y 22 , and Y 23 .
The results above, including both one-sample and two-sample predictions, can guide maintenance decisions by using the predictive intervals to schedule timely replacement or servicing of bearings, helping to prevent unexpected failures, minimize downtime, and reduce maintenance costs.
We encourage further research utilizing larger sample sizes and real-world datasets to achieve more accurate point and interval predictions, as well as to explore the impact of prior distribution selection on predictive outcomes.

7. Conclusions

In this study, we explore the estimation and prediction problems for the generalized exponential distribution based on Type-II censored data. The MLE is performed using the EM algorithm. Additionally, Bayesian estimation methods are investigated under various loss functions. Using Gibbs sampling and importance sampling, Bayesian estimates and HPD credible intervals are constructed. Monte Carlo simulations are conducted to assess and compare the performance of classical and Bayesian estimation techniques. Notably, the results indicate that Bayesian methods consistently provide lower MSEs and more precise interval estimates. For prediction, the proposed methods are applied to the endurance test data for deep groove ball bearings, where the last three observations are censored. Both one-sample and two-sample prediction scenarios are analyzed, with posterior predictive distributions used to construct predictive intervals. The findings demonstrate that Bayesian approaches offer a robust and reliable framework for both estimation and prediction in the presence of Type-II censored data.
One limitation of the Type-II censoring scheme is that units can only be censored at the final termination time. Future studies are encouraged to investigate more adaptable censoring schemes, such as progressive type-II hybrid censoring and adaptive Type-II progressive censoring. Moreover, exploring different prior distributions, alternative loss functions, or applying the model to other lifetime distributions could further improve the versatility and reliability of the approach.

Author Contributions

Conceptualization and methodology, W.W. and W.G.; software, W.W.; investigation, W.W.; writing—original draft preparation, W.W.; writing—review and editing, W.G.; supervision, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Project 202510004158 which was supported by National Training Program of Innovation and Entrepreneurship for Undergraduates. Wenhao’s work was partially supported by the Science and Technology Research and Development Project of China State Railway Group Company, Ltd. (No. N2023Z020).

Data Availability Statement

The dataset utilized in this study is publicly accessible in reference [26].

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

PDFProbability density function
CDFCumulative distribution function
MLEMaximum likelihood estimation
ACIAsymptotic confidence interval
CIConfidence interval
HPDHighest posterior density
GELFGeneral entropy loss function
LILFLinex loss function
SELFSquared error loss function
MCMCMarkov Chain Monte Carlo
MSEMean squared error
Boot-p CIBootstap-p confidence interval
Boot-t CIBootstap-t confidence interval
ALAverage length
CPCoverage probability
K–SKolmogorov–Smirnov
AICAkaike Information Criterion
BICBayesian Information Criterion
A-DAnderson–Darling

Appendix A

Let a i j ( α , λ ) represent the ( i , j ) th element of the matrix I W ( θ ) , defined as follows:
a 11 = n α 2 , a 22 = n λ 2 + n α ( α 1 ) λ 2 0 x 2 e 2 x 1 e x α 3 d x , a 12 = a 21 = n α λ 0 x e 2 x 1 e x α 2 d x ,
Let b i j represent the ( i , j ) th elements of the matrix I W | X ( θ ) , given by
b 11 = ( n r ) · 1 α 2 1 e λ x r α ln 1 e λ x r 2 ψ 1 2 , b 22 = ( n r ) · 1 λ 2 α x r 2 α e x r λ ψ 1 1 e x r λ α 2 e λ x r ψ 1 2 + ( α 1 ) ψ 2 , b 12 = ( n r ) x r e x r λ 1 e x r λ α 1 α ln 1 e x r λ + ψ 1 ψ 1 2 ( n r ) ψ 3 = b 21 .
where
ψ 1 = 1 ( 1 e x r λ ) α , ψ 2 = 1 λ 2 1 1 e x r λ α 1 e x r λ α 1 ln 1 ξ 1 / α 2 1 ξ 1 / α ξ 2 / α d ξ , ψ 3 = 1 λ 1 1 e x r λ α 1 e x r λ α 1 ln 1 ξ 1 / α 1 ξ 1 / α ξ 1 / α d ξ .

References

  1. Gupta, R.D.; Kundu, D. Theory & methods: Generalized exponential distributions. Aust. N. Z. J. Stat. 1999, 41, 173–188. [Google Scholar]
  2. Mudholkar, G.S.; Srivastava, D.K. Exponentiated Weibull family for analyzing bathtub failure-rate data. IEEE Trans. Reliab. 1993, 42, 299–302. [Google Scholar] [CrossRef]
  3. Gompertz, B. XXIV. On the nature of the function expressive of the law of human mortality, and on a new mode of determining the value of life contingencies. In a letter to Francis Baily, Esq. FRS &c. Philos. Trans. R. Soc. Lond. 1825, 115, 513–583. [Google Scholar]
  4. Verhulst, P.-F. Notice sur la loi que la population suit dans son accroissement. Corresp. Mathématique Phys. 1838, 10, 113–129. [Google Scholar]
  5. Gupta, R.D.; Kundu, D. Exponentiated Exponential Family: An Alternative to Gamma and Weibull Distributions. Biom. J. 2001, 43, 117–130. [Google Scholar] [CrossRef]
  6. Gupta, R.D.; Kundu, D. Generalized exponential distribution: Different method of estimations. J. Stat. Comput. Simul. 2001, 69, 315–337. [Google Scholar] [CrossRef]
  7. Kundu, D.; Pradhan, B. Estimating the parameters of the generalized exponential distribution in presence of hybrid censoring. Commun. Stat.-Theory Methods 2009, 38, 2030–2041. [Google Scholar] [CrossRef]
  8. Kim, C.; Song, S. Bayesian estimation of the parameters of the generalized exponential distribution from doubly censored samples. Stat. Pap. 2010, 51, 583–597. [Google Scholar] [CrossRef]
  9. Mohie El-Din, M.M.M.; Amein, M.M.; Shafay, A.R.; Mohamed, S. Estimation of generalized exponential distribution based on an adaptive progressively type-II censored sample. J. Stat. Comput. Simul. 2017, 87, 1292–1304. [Google Scholar] [CrossRef]
  10. Pradhan, B.; Kundu, D. On progressively censored generalized exponential distribution. Test 2009, 18, 497–515. [Google Scholar] [CrossRef]
  11. Guo, L.; Gui, W. Statistical inference of the reliability for generalized exponential distribution under progressive type-II censoring schemes. IEEE Trans. Reliab. 2018, 67, 470–480. [Google Scholar] [CrossRef]
  12. Madi, M.T.; Raqab, M.Z. Bayesian inference for the generalized exponential distribution based on progressively censored data. Commun. Stat.—Theory Methods 2009, 38, 2016–2029. [Google Scholar] [CrossRef]
  13. Gupta, R.D.; Kundu, D. Generalized exponential distribution: Existing results and some recent developments. J. Stat. Plan. Inference 2007, 137, 3537–3547. [Google Scholar] [CrossRef]
  14. Al-Hussaini, E.K. Predicting observables from a general class of distributions. J. Stat. Plan. Inference 1999, 79, 79–91. [Google Scholar] [CrossRef]
  15. Kundu, D.; Howlader, H. Bayesian inference and prediction of the inverse Weibull distribution for Type-II censored data. Comput. Stat. Data Anal. 2010, 54, 1547–1558. [Google Scholar] [CrossRef]
  16. Arabi Belaghi, R.; Noori Asl, M.; Gurunlu Alma, O.; Singh, S.; Vasfi, M. Estimation and prediction for the Poisson-Exponential distribution based on type-II censored data. Am. J. Math. Manag. Sci. 2019, 38, 96–115. [Google Scholar] [CrossRef]
  17. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar] [CrossRef]
  18. Fisher, R.A. On the mathematical foundations of theoretical statistics. Philos. Trans. R. Soc. Lond. Ser. A Contain. Pap. Math. Phys. Character 1922, 222, 309–368. [Google Scholar]
  19. Louis, T.A. Finding the Observed Information Matrix When Using the EM Algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1982, 44, 226–233. [Google Scholar] [CrossRef]
  20. Efron, B. The Jackknife, the Bootstrap and Other Resampling Plans; SIAM: Philadelphia, PA, USA, 1982. [Google Scholar]
  21. Hall, P. Theoretical comparison of bootstrap confidence intervals. Ann. Stat. 1988, 16, 927–953. [Google Scholar] [CrossRef]
  22. Jäntschi, L. Detecting extreme values with order statistics in samples from continuous distributions. Mathematics 2020, 8, 216. [Google Scholar] [CrossRef]
  23. Jäntschi, L. A test detecting the outliers for continuous distributions based on the cumulative distribution function of the data being tested. Symmetry 2019, 11, 835. [Google Scholar] [CrossRef]
  24. Raqab, M.Z.; Madi, M.T. Bayesian inference for the generalized exponential distribution. J. Stat. Comput. Simul. 2005, 75, 841–852. [Google Scholar] [CrossRef]
  25. Chen, M.H.; Shao, Q.M. Monte Carlo estimation of Bayesian credible and HPD intervals. J. Comput. Graph. Stat. 1999, 8, 69–92. [Google Scholar] [CrossRef]
  26. Lawless, J.F. Statistical Models and Methods for Lifetime Data; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  27. Jäntschi, L.; Bolboacă, S.D. Computation of probability associated with Anderson–Darling statistic. Mathematics 2018, 6, 88. [Google Scholar] [CrossRef]
Figure 1. PDF plots of the generalized exponential distribution for various parameter values.
Figure 1. PDF plots of the generalized exponential distribution for various parameter values.
Symmetry 17 00222 g001
Figure 2. Hazard rate plots of the generalized exponential distribution for various parameter values.
Figure 2. Hazard rate plots of the generalized exponential distribution for various parameter values.
Symmetry 17 00222 g002
Figure 3. Illustration of series and parallel system models.
Figure 3. Illustration of series and parallel system models.
Symmetry 17 00222 g003
Figure 4. Comparison of fitted CDFs and PDFs for various distributions.
Figure 4. Comparison of fitted CDFs and PDFs for various distributions.
Symmetry 17 00222 g004
Figure 5. One-sample posterior predictive density functions of X 21 , X 22 and X 23 .
Figure 5. One-sample posterior predictive density functions of X 21 , X 22 and X 23 .
Symmetry 17 00222 g005
Figure 6. One-sample predictive survival functions of X 21 , X 22 and X 23 .
Figure 6. One-sample predictive survival functions of X 21 , X 22 and X 23 .
Symmetry 17 00222 g006
Figure 7. Two-sample point prediction and interval prediction.
Figure 7. Two-sample point prediction and interval prediction.
Symmetry 17 00222 g007
Figure 8. Two-sample posterior predictive density functions of Y 21 , Y 22 and Y 23 .
Figure 8. Two-sample posterior predictive density functions of Y 21 , Y 22 and Y 23 .
Symmetry 17 00222 g008
Figure 9. Two-sample predictive survival functions of Y 21 , Y 22 and Y 23 .
Figure 9. Two-sample predictive survival functions of Y 21 , Y 22 and Y 23 .
Symmetry 17 00222 g009
Table 1. Mean and corresponding MSE values for α ^ under various censoring schemes.
Table 1. Mean and corresponding MSE values for α ^ under various censoring schemes.
Censoring Scheme α SELFGELFLILFMLE
n r δ = 0 . 6 δ = 0 . 8 h = 0 . 2 h = 0 . 6
3025Mean2.24482.20932.11882.28732.16362.2656
MSE0.36070.33450.27890.40620.28570.6245
4030Mean2.18732.15902.08682.21982.12402.2143
MSE0.27400.25710.22080.30190.22590.5490
5040Mean2.16022.13902.08442.18322.11432.1526
MSE0.20690.19810.17860.22040.18180.3836
6040Mean2.12762.10712.05442.15002.08262.1962
MSE0.27430.26420.24090.29050.24360.4935
6050Mean2.09752.08102.03872.11462.06322.1012
MSE0.13200.12790.11940.13830.12050.3004
Table 2. Mean and corresponding MSE values for λ ^ under various censoring schemes.
Table 2. Mean and corresponding MSE values for λ ^ under various censoring schemes.
Censoring Scheme λ SELFGELFLILFMLE
n r δ = 0 . 6 δ = 0 . 8 h = 0 . 2 h = 0 . 6
3025Mean1.07111.06061.03351.07671.06000.9588
MSE0.04690.04510.04130.04840.04410.0707
4030Mean1.06481.05571.03241.06981.05510.8879
MSE0.04110.03960.03640.04230.03880.0868
5040Mean1.05031.04351.02621.05381.04320.9086
MSE0.03040.02970.02820.03100.02930.0626
6040Mean1.06061.05371.03641.06441.05310.8655
MSE0.04290.04170.03900.04380.04110.1066
6050Mean1.03051.02521.01161.03321.02520.9171
MSE0.02150.02110.02040.02180.02090.0522
Table 3. Estimations of the intervals for parameters α and λ under various censoring schemes.
Table 3. Estimations of the intervals for parameters α and λ under various censoring schemes.
(n, r)ParameterHPD IntervalBoot-t CIBoot-p CIAsymptotic CI
ALCPALCPALCPALCP
(30, 25) α 2.13470.9282.45600.9452.34240.8942.29710.891
λ 0.79600.9250.85300.9120.90810.9060.90980.947
(40, 30) α 1.78270.9241.97120.9061.86820.8871.88040.902
λ 0.71800.9220.72000.8790.80410.9040.78100.915
(50, 40) α 1.48900.9271.75800.9111.61730.9081.60710.922
λ 0.61100.9250.67200.9030.64150.9220.62960.936
(60, 40) α 1.38050.8941.44740.8571.44600.9521.43340.964
λ 0.60080.9180.56910.8740.64760.9580.63850.936
(60, 50) α 1.39100.9271.57300.9431.52680.9161.53250.924
λ 0.52090.9350.58700.9220.59380.9210.58720.927
Table 4. Mean and corresponding MSE values for α ^ , λ ^ under fixed censoring schemes n = 40 , r = 30 with different true values.
Table 4. Mean and corresponding MSE values for α ^ , λ ^ under fixed censoring schemes n = 40 , r = 30 with different true values.
( α , λ )ParameterMLEGELFLILFSELF
δ = 0 . 6 δ = 0 . 8 h = 0 . 2 h = 0 . 6
1.5, 0.5 α 1.67851.75261.69541.79831.73221.7756
(0.2794)(0.2549)(0.2109)(0.3178)(0.2290)(0.2743)
λ 0.44580.57740.56460.58420.57950.5826
(0.0236)(0.0209)(0.0185)(0.0224)(0.0212)(0.0220)
1.5, 1 α 1.68161.68501.63751.72101.67021.7038
(0.2693)(0.1813)(0.1533)(0.2081)(0.1665)(0.1935)
λ 0.89151.13801.11631.15221.13631.1468
(0.0889)(0.0720)(0.0647)(0.0776)(0.0705)(0.0751)
2, 0.5 α 2.16592.17392.10992.22502.14412.1983
(0.3529)(0.2751)(0.2433)(0.3098)(0.2482)(0.2887)
λ 0.43910.53920.52870.54440.54110.5433
(0.0198)(0.0111)(0.0102)(0.0116)(0.0112)(0.0115)
2, 1 α 2.21432.21982.08682.21432.12402.1873
(0.5490)(0.3019)(0.2208)(0.3019)(0.2259)(0.2740)
λ 0.88791.06981.03241.06981.05511.0648
(0.0868)(0.0423)(0.0364)(0.0423)(0.0388)(0.0411)
2, 2 α 2.15132.17962.04192.17962.08102.1458
(0.3299)(0.2975)(0.2237)(0.2975)(0.2236)(0.2699)
λ 1.77432.08111.99512.08112.02352.0617
(0.2977)(0.1545)(0.1402)(0.1545)(0.1375)(0.1481)
3, 1 α 2.89152.90822.81892.92812.84372.9415
(0.4177)(0.3662)(0.3705)(0.3822)(0.3261)(0.3676)
λ 0.83990.98140.96470.99080.98160.9877
(0.0699)(0.0237)(0.0242)(0.0237)(0.0234)(0.0236)
3, 2 α 3.09472.83182.74772.90602.77482.8634
(0.5801)(0.3618)(0.3878)(0.3550)(0.3605)(0.3547)
λ 1.74451.92051.88731.94471.90951.9331
(0.2491)(0.0866)(0.0937)(0.0831)(0.0868)(0.0845)
Table 5. 95% confidence and credible intervals for α , λ under fixed censoring schemes n = 40 , r = 30 .
Table 5. 95% confidence and credible intervals for α , λ under fixed censoring schemes n = 40 , r = 30 .
α , λ ParameterAsymptotic CIBoot-t CIBoot-p CIHPD Interval
ALCPALCPALCPALCP
1.5, 0.5 α 1.39830.9131.53770.9031.52360.9181.41350.924
λ 0.40020.8950.47760.9200.47690.9350.37600.844
1.5, 1 α 1.39720.9171.44690.9181.47470.8941.35440.882
λ 0.79990.8890.88310.8950.86890.8800.73770.894
2, 0.5 α 1.94630.9181.92510.9271.94570.9091.71530.946
λ 0.35940.8890.40720.9270.42140.9180.35120.944
2, 1 α 1.88040.9021.97100.9061.86820.8901.78270.924
λ 0.78100.9040.72000.8790.80410.9040.71800.922
2, 2 α 1.92330.9221.89710.9071.91260.9251.65300.936
λ 1.43020.8851.44150.9071.42730.8941.36080.924
3, 1 α 2.21740.9382.33160.9182.24900.9301.88630.934
λ 0.61650.9410.68230.9180.70020.9200.52020.932
3, 2 α 2.26740.9402.17740.8832.33590.9321.82300.912
λ 1.23540.9421.21790.8831.18810.9251.03990.904
Table 6. Mean and corresponding MSE values of α ^ , λ ^ for Non-informative and informative priors.
Table 6. Mean and corresponding MSE values of α ^ , λ ^ for Non-informative and informative priors.
SchemePrior Information
Non-Informative PriorInformative Prior
LILFGELFSELFLILFGELFSELF
h = 0 . 2 h = 0 . 6 δ = 0 . 6 δ = 0 . 8 h = 0 . 2 h = 0 . 6 δ = 0 . 6 δ = 0 . 8
(30, 25)2.66052.51032.53842.48522.58802.63552.48852.51562.46292.5647
(0.3512)(0.3177)(0.3320)(0.3358)(0.3321)(0.3303)(0.2900)(0.3055)(0.3072)(0.3078)
1.18611.16801.16401.14991.17721.18251.16411.16011.14581.1734
(0.0482)(0.0476)(0.0485)(0.0496)(0.0478)(0.0364)(0.0361)(0.0368)(0.0379)(0.0362)
(40, 30)2.75902.61162.69482.59012.72292.64752.51972.59072.50012.6155
(0.4616)(0.3973)(0.4372)(0.4212)(0.4439)(0.2846)(0.2525)(0.2717)(0.2707)(0.2744)
1.37601.35021.34711.33041.36321.38571.36151.35911.34371.3737
(0.0504)(0.0490)(0.0500)(0.0510)(0.0500)(0.0365)(0.0358)(0.0364)(0.0377)(0.0363)
(50, 40)2.61002.51612.56862.49832.58732.58142.49792.54462.48262.5612
(0.2505)(0.2413)(0.2485)(0.2555)(0.2477)(0.2062)(0.1989)(0.2044)(0.2099)(0.2038)
1.19731.18531.18911.17011.19431.18531.17441.17781.16041.1826
(0.0320)(0.0316)(0.0320)(0.0328)(0.0319)(0.0292)(0.0294)(0.0295)(0.0307)(0.0293)
(60, 40)2.66442.56662.62182.55162.64072.62352.53602.58492.52232.6020
(0.2911)(0.2626)(0.2803)(0.2743)(0.2831)(0.2358)(0.2207)(0.2305)(0.2312)(0.2313)
1.22181.20731.21221.19041.21821.21071.19711.20161.18131.2073
(0.0402)(0.0387)(0.0395)(0.0391)(0.0398)(0.0336)(0.0326)(0.0331)(0.0330)(0.0333)
(60, 50)2.52152.46112.49442.44772.50692.52542.46832.49992.45632.5117
(0.1558)(0.1646)(0.1605)(0.1734)(0.1575)(0.1216)(0.1283)(0.1249)(0.1350)(0.1228)
1.15201.14441.14671.13411.15011.16511.15791.16011.14841.1633
(0.0248)(0.0255)(0.0254)(0.0271)(0.0250)(0.0204)(0.0208)(0.0208)(0.0220)(0.0205)
Table 7. Random numbers generated using the inverse transformation method.
Table 7. Random numbers generated using the inverse transformation method.
0.160.210.310.340.440.470.490.640.700.991.011.111.15
1.231.381.421.431.531.661.681.722.092.482.812.832.95
2.993.013.246.73
Table 8. Point predictions and predictive intervals for k = 26 , 28 , 30 in the one-sample case.
Table 8. Point predictions and predictive intervals for k = 26 , 28 , 30 in the one-sample case.
(n, r)kPredictionsIntervalsInterval Length
(10, 30)262.614(1.804, 4.079)2.275
283.221(2.250, 5.264)3.013
304.709(2.866, 9.704)6.839
(15, 30)262.760(1.993, 4.215)2.221
283.383(2.285, 5.400)3.116
304.970(3.023, 9.375)6.353
(20, 30)262.680(1.988, 3.693)1.705
283.352(2.270, 5.030)2.760
305.054(2.851, 9.219)6.368
(25, 30)263.092(2.840, 3.815)0.975
283.836(3.022, 5.463)2.442
305.728(3.584, 9.842)6.258
Table 9. Point predictions and predictive intervals for k = 26 , 28 , 30 in the two-sample case.
Table 9. Point predictions and predictive intervals for k = 26 , 28 , 30 in the two-sample case.
(n, r)kPredictionsIntervalsInterval Length
(10, 30)262.438(1.704, 4.421)2.717
283.032(2.073, 5.791)3.718
304.535(2.820, 10.123)7.303
(15, 30)262.640(1.741, 4.331)2.590
283.326(2.106, 5.577)3.471
305.070(2.835, 9.556)6.721
(20, 30)262.723(1.624, 4.464)2.840
283.389(1.955, 5.763)3.808
305.075(2.598, 9.649)7.051
(25, 30)262.990(1.787, 4.677)2.890
283.751(2.176, 6.038)3.861
305.683(2.933, 10.264)7.331
Table 10. Endurance Test Data for Deep Groove Ball Bearings (23 Samples).
Table 10. Endurance Test Data for Deep Groove Ball Bearings (23 Samples).
Unit: Million Revolutions
17.8828.9233.0041.5242.1245.6048.8051.8451.9654.12
55.5667.8068.6468.6468.8884.1293.1298.64105.12105.84
127.92128.04173.40
Table 11. Inferences on the goodness-of-fit for real data.
Table 11. Inferences on the goodness-of-fit for real data.
Distribution a ^ λ ^ AICBICK-SA-DLog-Likelihood
Gamma3.98220.0552230.0561232.32710.12200.2157−113.0281
Weibull2.103181.9010231.3773233.64830.15140.3292−113.6887
Log-Normal4.15070.5215230.2574232.52840.09730.1896−113.1287
Generalized Exponential5.29320.0323229.9525232.22350.10550.1874−112.9762
Table 12. Mean estimates of α and λ for censoring scheme n = 23 , r = 20 based on the real dataset.
Table 12. Mean estimates of α and λ for censoring scheme n = 23 , r = 20 based on the real dataset.
ParameterMLEGELFLILFSELF
δ = 0 . 6 δ = 0 . 8 h = 0 . 2 h = 0 . 6
α 5.04395.05064.76495.34094.67065.1462
λ 0.03190.02970.02950.03010.03030.0287
Table 13. Point predictions and predictive intervals for X 21 , X 22 , and X 23 in the one-sample case.
Table 13. Point predictions and predictive intervals for X 21 , X 22 , and X 23 in the one-sample case.
kTrue ValuePredictionsIntervals
21127.92119.1102(106.1134, 147.1283)
22128.04138.6103(109.4465, 194.0252)
23173.40176.8546(118.7290, 289.9706)
Table 14. Point predictions and predictive intervals in the two-sample case.
Table 14. Point predictions and predictive intervals in the two-sample case.
kPoint PredictionLower BoundUpper Bound
117.66084.872936.4397
224.346510.241143.3671
329.421214.449748.8196
433.834818.090753.6814
537.910321.408658.2652
641.805324.527562.7310
745.614627.522667.1804
849.404330.445771.6893
953.227133.336576.3223
1057.129436.228581.1406
1161.156839.152486.2081
1265.357642.138691.5963
1369.786845.219597.3899
1474.453448.4317103.6955
1579.613551.8191110.6520
1685.207255.4372118.4508
1790.662659.3598127.3675
1898.562463.6904137.8241
19106.941368.5841150.5164
20117.129474.2923166.7142
21130.640481.2710189.1098
22150.452890.5097225.0417
23189.4416105.0678308.4021
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Gui, W. Estimation and Bayesian Prediction for the Generalized Exponential Distribution Under Type-II Censoring. Symmetry 2025, 17, 222. https://doi.org/10.3390/sym17020222

AMA Style

Wang W, Gui W. Estimation and Bayesian Prediction for the Generalized Exponential Distribution Under Type-II Censoring. Symmetry. 2025; 17(2):222. https://doi.org/10.3390/sym17020222

Chicago/Turabian Style

Wang, Wei, and Wenhao Gui. 2025. "Estimation and Bayesian Prediction for the Generalized Exponential Distribution Under Type-II Censoring" Symmetry 17, no. 2: 222. https://doi.org/10.3390/sym17020222

APA Style

Wang, W., & Gui, W. (2025). Estimation and Bayesian Prediction for the Generalized Exponential Distribution Under Type-II Censoring. Symmetry, 17(2), 222. https://doi.org/10.3390/sym17020222

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop