Next Article in Journal
Methodological Advances in Temperature Dynamics Modeling for Energy-Efficient Indoor Air Management Systems
Next Article in Special Issue
Research on Heat Transfer Coefficient Prediction of Printed Circuit Plate Heat Exchanger Based on Deep Learning
Previous Article in Journal
Slim-YOLO: An Improved Sugarcane Tail Tip Recognition Algorithm Based on YOLO11n for Complex Field Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parametric Estimation and Analysis of Lifetime Models with Competing Risks Under Middle-Censored Data

School of Mathematics and Statistics, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(8), 4288; https://doi.org/10.3390/app15084288
Submission received: 17 March 2025 / Revised: 10 April 2025 / Accepted: 11 April 2025 / Published: 13 April 2025

Abstract

:
Middle-censoring is a general censoring mechanism. In middle-censoring, the exact lifetimes are observed only for a portion of the units and for others, we can only know the random interval within which the failure occurs. In this study, we focus on statistical inference for middle-censored data with competing risks. The latent failure times are assumed to be independent and follow Burr-XII distributions with distinct parameters. To begin with, we derive the maximum likelihood estimators for the unknown parameters, proving their existence and uniqueness. Additionally, asymptotic confidence intervals are constructed using the observed Fisher information matrix. Furthermore, Bayesian estimates under squared loss function and the corresponding highest posterior density intervals are obtained through the Gibbs sampling method. A simulation study is carried out to assess the performance of all proposed estimators. Lastly, an analysis for a practical dataset is provided to demonstrate the inferential processes developed.

1. Introduction

1.1. Middle-Censored Data and Competing Risks

Censoring occurs frequently in life test studies and cannot be completely avoided because of various practical constraints. Middle-censoring, a general censoring mechanism, was proposed by Jammalamadaka and Mangalam [1]. In middle-censoring, the precise failure time of the tested component cannot be detected when it occurs within a random interval. That is, the exact lifetimes of some units are observable, while for others we only know the intervals.
It is supposed that there are n identical units in a life test. The lifetimes of these units are defined as T 1 , T 2 , , T n , and the i-th corresponding random interval is [ L i , R i ] . If  T i [ L i , R i ] , we cannot know the actual lifetime of the i-th unit, then δ i = 0 . Otherwise, δ i = 1 denotes that the exact lifetime of the i-th unit can be observed directly. We can reorder the data, placing the complete data at the front, then the data can be described as:
( T 1 , δ 1 ) , ( T 2 , δ 2 ) , , ( T n 1 , δ n 1 ) , ( [ L n 1 + 1 , R n 1 + 1 ] , δ n 1 + 1 ) , , ( [ L n 1 + n 2 , R n 1 + n 2 ] , δ n 1 + n 2 ) .
Jammalamadaka and Mangalam [1] not only put forward the middle-censoring scheme, but also developed an algorithm for obtaining the self-consistent estimator for the middle-censored data. The maximum likelihood estimator (MLE) in a non-parametric setting was also obtained. In addition, Jammalamadaka and Iyer [2] developed an approximation to the distribution function of the middle-censored data. Then, they also suggested a straightforward alternative estimator for the lifetime distribution function, analyzed its performance under middle-censored data, and established its consistency and weak convergence.
Moreover, several scholars have delved into middle-censored data within the parametric situation. Iyer et al. [3] considered a middle-censoring scheme in the case where the lifetime of the tested unit follows the exponential distribution. The maximum likelihood estimator was derived and its consistency and asymptotic properties were rigorously established. Additionally, the Bayes estimator was derived under a Gamma prior distribution. Davarzani and Parsia [4] studied a discrete set-up with middle-censored data, where the survival time, the left endpoint of the censoring interval, and the length of the interval are all variables modeled by geometric distributions. They derived the MLE of the unknown parameter using the EM algorithm. Bayesian estimator and the corresponding credible intervals were also obtained. Jammalamadaka and Bapat [5] developed the estimation under the middle-censoring paradigm in a multinomial distribution with k distinct possible outcomes. Abuzaid et al. [6] concentrated on the Weibull and exponential distributions to explore the robustness of their parameter estimates and discovered that estimations for middle-censoring exhibited greater robustness when contrasted with those for right censoring.
In practical applications, failure causes are typically pre-existing risk factors that may interact competitively. For example, in engineering reliability testing, environmental stressors such as temperature and humidity often compete to induce component failure. Statistical literature refers to this competitive failure mechanism as ‘competing risks’, emphasizing how multiple antecedent factors compete to be the first to exceed critical thresholds and cause system breakdown.
To consider the different reasons for failure, many studies combine competing risks with middle-censoring. Based on middle-censored data with two independent competing risks, Wang [7] explored various estimations of the parameter for exponential distribution. Ahmadi et al. [8] considered different estimations under middle-censored independent competing risks data with exponential distribution. Both the fix-point method and the EM algorithm were used to calculate the MLE. In addition, the Bayesian estimate under the Gamma prior was derived using both the Lindley’s approximation method and Gibbs sampling. The reconstruction of the censoring data was also explored. Wang et al. [9] investigated the dependent competing risks model based on middle-censored data by making use of the Marshall–Olkin bivariate Weibull distribution. They derived MLEs, midpoint approximation estimates, and asymptotic confidence intervals (ACIs) for the unknown parameters. They also considered the Bayesian estimates under the Gamma–Dirichlet prior by the acceptance–rejection sampling method. Davarzani et al. [10] analyzed a dependent middle-censoring model with the Marshall–Olkin bivariate exponential distribution.
In addition, other risk models have been explored. Sankaran and Prasad [11] proposed a proportional hazards regression framework to analyze middle-censored lifetime data with Weibull distribution. Rehman and Chandra [12] studied the cumulative incidence function by modeling the cause-specific hazard as a Weibull distribution for middle-censored survival data. Sankaran and Prasad [13] investigated an Exponentiated–Exponential regression model for middle-censored survival time analysis.

1.2. Burr-XII Distribution

The Burr XII distribution is a flexible and versatile distribution that is useful for modeling a wide range of data types. Its parameters allow it to exhibit different kurtosis and tail properties to adapt to different shapes, and the distribution has non-monotone hazard rates. In addition, both the cumulative and reliability function of the Burr-XII distribution have closed-form expressions, making it easier to compute percentiles and the likelihood function based on censored data, which makes it a valuable tool in statistical modeling and analysis [14].
A random variable X is said to follow the Burr ( a , b ) distribution if its probability density function (PDF) and cumulative distribution function (CDF) are given by the following expressions, respectively:
f ( t ; a , b ) = a b t b 1 ( 1 + t b ) ( a + 1 ) , t > 0 ; a , b > 0 ,
F ( t ; a , b ) = 1 ( 1 + t b ) a , t > 0 ,
where a , b are two positive shape parameters of the Burr-XII distribution.
In addition, the reliability and hazard rate function take the form as:
R ( t ; a , b ) = ( 1 + t b ) a , t > 0 ,
h ( t ; a , b ) = a b t b 1 1 + t b , t > 0 .
In the density plots in Figure 1, it can be observed that when b 1, the density curve is strictly decreasing, while the Burr-XII distribution density curve is unimodal when b > 1 and at this time, a larger a corresponds to a higher peak height of the curve.
In addition, Figure 2 shows the hazard rate function plots. As shown in the picture, when b 1 , the hazard rate function decreases with t becoming larger, while the plot has a single peak when b > 1 . The peak is higher when the value of parameter a becomes larger.
Recently, owing to its versatility in applied contexts like reliability analysis, quality assurance, and quantum systems, the Burr-XII distribution has attracted significant interest from several scholars.
There are many studies for Burr-XII distribution based on various life tests and censoring schemes. Soliman [15] derived both the MLE and Bayesian estimation approaches for the Burr-XII distribution parameters under progressive Type-II censoring scenarios. Yan et al. [16] exploited the MLE and Bayes estimation within an improved adaptive Type-II progressive censoring framework under the Burr-XII distribution. Du and Gui [17] investigated the Burr-XII distribution in competitive risk modeling under adaptive Type-II progressive censoring. Bayesian point estimates under various loss functions and the corresponding highest posterior density credible intervals were obtained through MCMC simulations.
Although there are a large amount of statistical inferences related to the Burr-XII distribution along with censored data, they mainly focused on Type-II censoring or other progressive censoring schemes, and studies on the Burr-XII distribution under middle censoring are currently limited. Abuzaid [18] analyzed middle-censored data from the Burr-XII distribution. MLEs of two parameters were obtained, and Bayesian inferences under a Gamma prior were established through the application of Lindley’s approximation technique. However, it did not consider the failure reasons. Thus, based on that, in this paper we will introduce independent competing risks to study the middle-censored data under Burr-XII distribution.
To our knowledge, no existing studies have comprehensively investigated the parameter-estimation challenges for Burr-XII distributions considering both middle-censored data and competing risks models. In this paper, we combine the competing risks with middle-censored data to address the parameter-estimation problems for the Burr-XII distribution. Compared to [17], we introduce a new censoring scheme and consider different numbers of failure causes. Different from [18], we introduce a competing risks model to the analysis framework.
In this paper, we use the perspective of latent failure times for a competing risk model suggested by Cox [19]. An analysis of the independent competing risks model is conducted within a middle-censoring scenario. It is assumed that each latent failure time variable follows the Burr-XII distribution. In Section 2, we make some reasonable assumptions for our model. In Section 3, we prove that the MLEs exist uniquely, implementing both optimization techniques and the EM algorithm to obtain them. In addition, we derive ACIs in accordance with the properties of MLEs. In Section 4, we obtain the Bayesian estimators under exponential and Gamma prior distributions using Gibbs sampling, and HPD credible intervals are also constructed. Section 5 presents the simulation results to validate estimator performance and applies the model to a practical dataset to demonstrate the application value of our model.

2. Model Assumption and Notation

Assume that n N homogeneous units are assigned to a lifetime test. For each unit, there are s failure causes. We suppose that these s latent failure times are mutually independent and conform to the Burr-XII distribution.
Thus, define T i as the exact lifetime of the i-th tested unit, and 
T i = min X i 1 , X i 2 , , X i s ,
where X i j represents the random variable corresponding to the j-th latent failure cause. Moreover, X i 1 , X i 2 , , X i s are independent of each other. We assume that X i j , j = 1 , 2 , , s follow the Burr-XII distributions with the same second parameter ( b > 0 ) and different first parameters ( a 1 , a 2 , , a s > 0 ). The PDF and CDF of X i j are presented as follows, respectively:
f j ( t ; a j , b ) = a j b t b 1 ( 1 + t b ) ( a j + 1 ) , t > 0 ; a j , b > 0 ,
and
F j ( t ; a j , b ) = 1 ( 1 + t b ) a j , t > 0 ,
where a j and b are unknown parameters to be estimated and they are all positive, j = 1 , 2 , , s .
As exhibited in Figure 1, the parameter b affects the overall shape of the density curve. For different failure causes of the same product, it is reasonable to assume that the distributions share the same parameter b. The assumption that the potential failure distributions share a common parameter while differing in the other has been widely used in previous studies on competing risks data, such as [20,21].
Additionally, we use D i to denote the failure cause. D i = j if the i-th unit has failed as a result of the j-th failure cause. We suppose that X i 1 , X i 2 , , X i s are independent and cannot be observed directly, while T i and D i can be observed if the data are not censored. Therefore, taking into account the competing risks data under a middle-censoring framework, we consider a corresponding random censoring interval [ L i , R i ] for the i-th tested unit, and the interval is independent of its actual lifetime T i . That is,
( Y i , D i , δ i ) = ( T i , D i , 1 ) if T i [ L i , R i ] , [ L i , R i ] , D i , 0 if T i [ L i , R i ] .
We suppose L i and R i L i are also independent and exponentially distributed with known parameters.
Through the independence of X i 1 , X i 2 , , X i s , the joint PDF and CDF of lifetime with its corresponding cause ( T , D ) are concluded as follows:
f T , D ( t , j ; a , b ) = f j ( t ; a j , b ) k = 1 , k j s ( 1 F k ( t ; a k , b ) ) = a j b t b 1 ( 1 + t b ) ( a * + 1 ) ,
where a = ( a 1 , a 2 , , a s ) and a * = j = 1 s a j .
Through integration, we derive:
F T , D ( t , j ; a , b ) = a j a * [ 1 ( 1 + t b ) a * ] .
To simplify the notation, we will replace F T , D ( t , j ; a , b ) and f T , D ( t , j ; a , b ) with F ( t , j ; a , b ) and f ( t , j ; a , b ) , respectively, in the following text.

3. Frequentist Estimation

This section begins by deriving the likelihood function and establishing the existence and uniqueness of MLEs of the model parameters. Both optimization-based methods and the EM algorithm are then employed to compute these estimates. ACIs are constructed using the Fisher information matrix.

3.1. Maximum Likelihood Estimators

For simplicity, according to the approach in [8], after reordering the data, the sample is partitioned such that the first n 1 data items are not censored, and the subsequent n 2 data items are middle-censored. Thus, our data can be written as:
{ ( T 1 , D 1 ) , , ( T n 1 , D n 1 ) , ( [ L n 1 + 1 , R n 1 + 1 ] , D n 1 + 1 ) , , ( [ L n 1 + 1 , R n 1 + n 2 ] , D n 1 + n 2 ) } .
Let d a t a represent the observed sample data in the form mentioned above. Based on the sample data, the likelihood function can be derived through (7) and (8) as follows:
L ( a , b ; d a t a ) j = 1 s i = 1 n 1 ( f ( t i , j ; a , b ) ) I ( D i = j ) j = 1 s i = n 1 + 1 n 1 + n 2 ( F ( r i , j ; a , b ) F ( l i , j ; a , b ) ) I ( D i = j ) = b n 1 ( 1 a * ) n 2 j = 1 s a j m j i = 1 n 1 [ t i b 1 ( 1 + t i b ) ( a * + 1 ) ] i = n 1 + 1 n 1 + n 2 [ ( 1 + l i b ) a * ( 1 + r i b ) a * ]
where symbol ∝ denotes ’proportional to’,
I ( D i = j ) = 1 if D i = j , 0 Otherwise ,
and m j = i = 1 n I ( D i = j ) , which equals the number of the items that have failed attributed to the j-th latent risk. For simplicity, we use L ( a , b ) to replace L ( a , b ; d a t a ) .
Through (9), the log-likelihood function takes the form:
ln L ( a , b ) = n 1 ln b n 2 ln a * + j = 1 s m j ln a j + ( b 1 ) i = 1 n 1 ln t i ( a * + 1 ) i = 1 n 1 ln ( 1 + t i b ) + i = n 1 + 1 n 1 + n 2 ln [ ( 1 + l i b ) a * ( 1 + r i b ) a * ] .
To solve for the MLEs, we set the first-order partial derivatives of function ln L ( a , b ) with respect to all unknown parameters a 1 , a 2 , , a s and b to 0, respectively. Consequently, the corresponding likelihood equations are given by:
ln L ( a , b ) a j = m j a j n 2 a * i = 1 n 1 ln ( 1 + t i b ) + K 1 = 0 , j = 1 , 2 , 3 , , s ,
ln L ( a , b ) b = n 1 b + i = 1 n 1 ln t i ( a * + 1 ) i = 1 n 1 t i b ln t i 1 + t i b + a * K 2 = 0 ,
where
K 1 = i = n 1 + 1 n 1 + n 2 ( 1 + r i b ) a * ln ( 1 + r i b ) ( 1 + l i b ) a * ln ( 1 + l i b ) ( 1 + l i b ) a * ( 1 + r i b ) a * ,
and
K 2 = i = n 1 + 1 n 1 + n 2 r i b ( 1 + r i b ) ( a * + 1 ) ln r i l i b ( 1 + l i b ) ( a * + 1 ) ln l i ( 1 + l i b ) a * ( 1 + r i b ) a * .
For j = 1 , 2 , , s , we determine, respectively, the MLEs of a j and a * as a j ^ and a * ^ , and the MLE of b is expressed as b ^ . Based on (11), we can derive that m 1 a 1 = m 2 a 2 = = m s a s . Thus,
a j ^ m j = a * ^ n , j = 1 , 2 , , s .
Before obtaining the solution through (11)–(13), the uniqueness and existence of solutions to the likelihood equations require verification.
Theorem 1.
For competing risks and middle-censored data following the Burr-XII distribution, when parameter b is known, the MLEs of parameters a 1 , a 2 , , a s are always existent and unique.
Proof. 
Through (11)–(13), the likelihood equations are equivalent to the following:
d 1 ( a * , b ) = n 1 a * i = 1 n 1 ln ( 1 + t i b ) + i = n 1 + 1 n 1 + n 2 ψ r i a * ln ψ r i ( ψ l i ) a * ln ψ l i ψ l i a * ψ r i a * = 0 d 2 ( a * , b ) = n 1 b + i = 1 n 1 ln t i ( a * + 1 ) i = 1 n 1 t i b ln t i 1 + t i b + a * i = n 1 + 1 n 1 + n 2 r i b ψ r i ( a * + 1 ) ln r i l i b ψ l i ( a * + 1 ) ln l i ψ l i a * ψ r i a * = 0 ,
where ψ r i = 1 + r i b and ψ l i = 1 + l i b . Therefore,
d 1 ( a * , b ) a * = n 1 a * 2 B ,
where
B = i = n 1 + 1 n 1 + n 2 ψ r i a * ψ l i a * ( ln ψ r i ln ψ l i ) 2 ( ψ l i a * ψ r i a * ) 2 .
When b is a known constant, we just obtain the solution of d 1 ( a * , b ) = 0 to obtain a * ^ . It is obvious that d 1 ( a * , b ) a * < 0 , and that
lim a * 0 d 1 ( a * , b ) = + lim a * + d 1 ( a * , b ) = i = 1 n 1 ln ( 1 + t i b ) i = n 1 + 1 n 1 + n 2 ln ψ l i < 0 ,
so d 1 ( a * , b ) is monotonically decreasing with respect to a * , and takes both positive and negative values, which indicates the existence and uniqueness of the zero point a * ^ . Therefore, through (13), we can conclude that when b is fixed, for parameters a 1 , a 2 , , a s , their MLEs always exist and are unique and finite.    □
Theorem 2.
For competing risks and middle-censored data following the Burr-XII distribution, in the case that a 1 , a 2 , , a s and b are all unknown, when there exists i such that t i < 1 , the MLEs a 1 ^ , a 2 ^ , , a 2 ^ and b ^ all exist and are unique.
Proof. 
According to (9), the joint likelihood function is expressed as:
L ( a , b ) j = 1 s i = 1 n 1 ( f ( t i , j ; a , b ) ) I ( D i = j ) j = 1 s i = n 1 + 1 n 1 + n 2 ( F ( r i , j ; a , b ) F ( l i , j ; a , b ) ) I ( D i = j ) ,
in accordance with the Lagrange mean-value theorem, there is a ξ i [ L i , R i ] satisfying [ F ( r i , j ; a , b ) F ( l i , j ; a , b ) ] = ( R i L i ) f ( ξ i , j ; a , b ) .
So we just need to prove that when the data are uncensored, the MLEs of the parameters a * and b are existent and unique.
Through (24), and for i > n 1 , for convenience, we substitute t i * with t i , then we can obtain that
a * ^ = n / i = 1 n ln ( 1 + t i b ) .
Then, b ^ is the solution of the equation given as follows:
z ( b ) = n b + i = 1 n l n t i ( n i = 1 n ln ( 1 + t i b ) + 1 ) i = 1 n t i b l n t i 1 + t i b = 0 .
We can derive the derivative of z ( b ) as follows:
z ( b ) = n b 2 n ( i = 1 n ln ( 1 + t i b ) ) 2 ( i = 1 n t i b l n t i 1 + t i b ) 2 n i = 1 n ln ( 1 + t i b ) i = 1 n t i b ( ln t i ) 2 ( 1 + t i b ) 2 < 0 ,
and it is obvious that
lim b 0 z ( b ) = + lim b + z ( b ) = lim b + z 1 ( b ) + lim b + z 2 ( b ) ,
where z 1 ( b ) = n i = 1 n ln ( 1 + t i b ) i = 1 n t i b l n t i 1 + t i b and z 2 ( b ) = i = 1 n ln t i 1 + t i b .
For each t i , i = 1 , 2 , , n , if  t i < 1 , lim b + l n t i 1 + t i b = l n t i < 0 and if t i 1 , lim b + l n t i 1 + t i b = 0 , so lim b + z 2 ( b ) < 0 when there exists an i such that t i < 1 . Moreover, it is simple to derive that lim b + z 1 ( b ) 0 . Thus, lim b + z ( b ) < 0 . Subsequently, we are able to prove that the function z ( b ) is strictly monotonically decreasing on the interval ( 0 , + ) . Moreover, as b varies from 0 to + , the sign of the function shifts from positive to negative. Therefore, the solution of z ( b ) = 0 exists and is unique, which indicates that b ^ is unique, and through (17), the uniqueness of a * ^ can be proved. Consequently, the presence and uniqueness of the MLEs of unknown parameters a 1 , a 2 , , a s are demonstrated.    □
Remark 1.
When using our model to deal with data, to ensure the existence, we just need to appropriately transform the units of physical quantities in the sample data to guarantee that there exists t i < 1 .
Since the explicit solution of the equations is difficult to obtain, we use certain iterative algorithms to calculate the MLEs of unknown parameters. In this paper, in order to maximize ln L ( a , b ) , we use the optim function in R language (version 4.3.2).

3.2. EM Algorithm

Given that explicit expressions for a j ^ and b ^ are unavailable, the EM algorithm is considered and employed to find the MLEs. Incomplete data scenarios are typically addressed using the EM algorithm, which is first formulated by Dempster et al. [22].
Define t i * as the exact survival time of the unit that falls within the interval [ l i , r i ], i = n 1 + 1 ,..., n 1 + n 2 .
Based on the complete data ( t 1 , t 2 , , t n 1 , t n 1 + 1 * , , t n 1 + n 2 * ), we can derive the log-likelihood function, which is expressed as:
ln L C ( a , b ) = j = 1 s m j ln a j + n ln b + ( b 1 ) ( i = 1 n 1 ln t i + i = n 1 + 1 n 1 + n 2 ln t i * ) ( a * + 1 ) ( i = 1 n 1 ln ( 1 + t i b ) + i = n 1 + 1 n 1 + n 2 ln ( 1 + t i * b ) ) ,
then by setting the derivatives of ln L C ( a , b ) with respect to a j and b to 0, we can obtain:
ln L c ( a , b ) a j = m j a j ( i = 1 n 1 ln ( 1 + t i b ) + i = n 1 + 1 n 1 + n 2 ln ( 1 + t i * b ) ) = 0 , j = 1 , 2 , , s ,
ln L c ( a , b ) b = n b + ( i = 1 n 1 ln t i + i = n 1 + 1 n 1 + n 2 ln t i * ) ( a * + 1 ) ( i = 1 n 1 t i b ln t i 1 + t i b + i = n 1 + 1 n 1 + n 2 ( t i * ) b ln t i * 1 + t i * b ) = 0 .
Through Equation (21), we can obtain that
m j a j ^ = n a * ^ , j = 1 , 2 , , s ,
and we can derive the system of equations:
n a * ( i = 1 n 1 ln ( 1 + t i b ) + i = n 1 + 1 n 1 + n 2 ln ( 1 + t i * b ) ) = 0 n b + ( i = 1 n 1 ln t i + i = n 1 + 1 n 1 + n 2 ln t i * ) ( a * + 1 ) ( i = 1 n 1 t i b ln t i 1 + t i b + i = n 1 + 1 n 1 + n 2 t i * b ln t i * 1 + t i * b ) = 0 .
For conciseness, we define the conditional expectations as follows:
E a , b ( ln T i | T i [ l i , r i ] , D i = d i ) = l i r i ln t × f ( t , d i ; a , b ) F ( r i , d i ; a , b ) F ( l i , d i ; a , b ) d t ,
E a , b ( ln ( 1 + T i b ) | T i [ l i , r i ] , D i = d i ) = l i r i ln ( 1 + t b ) × f ( t , d i ; a , b ) F ( r i , d i ; a , b ) F ( l i , d i ; a , b ) d t ,
E a , b ( T i b ln T i 1 + T i b | T i [ l i , r i ] , D i = d i ) = l i r i t b ln t 1 + t b × f ( t , d i ; a , b ) F ( r i , d i ; a , b ) F ( l i , d i ; a , b ) d t ,
where the expressions of f ( t , d i ; a , b ) and F ( t , d i ; a , b ) can be seen in (7) and (8).
In each iteration, to perform the E-step, we need to calculate the expectations using the parameter values obtained from the previous iteration by numerical integration methods. For notational convenience, we introduce the symbols as follows:
  • E 1 , i ( k ) = E a ( k 1 ) , b ( k 1 ) ( ln T i | T i [ l i , r i ] , D i = d i ) ,
  • E 2 , i ( k ) = E a ( k 1 ) , b ( k 1 ) ( ln ( 1 + T i b ) | T i [ l i , r i ] , D i = d i ) ,
  • E 3 , i ( k ) = E a ( k 1 ) , b ( k 1 ) ( T i b ln T i 1 + T i b | T i [ l i , r i ] , D i = d i ) ,
where i = n 1 + 1 , n 1 + 2 , , n 1 + n 2 . And for those uncensored data, we define:
F 1 , i ( k ) = ln t i , F 2 , i ( k ) = ln ( 1 + t i b ( k 1 ) ) , F 3 , i ( k ) = t i b ( k 1 ) ln t i 1 + t i b ( k 1 ) ,
where i = 1 , 2 , , n 1 .
When performing the M-step during the iteration process, for the purpose of maximizing the pseudo log-likelihood function ln L C ( a , b ) based on the complete data, through (24), we can derive that
a * ( k ) = n ( i = 1 n 1 F 2 , i ( k ) + i = n 1 + 1 n 1 + n 2 E 2 , i ( k ) ) 1 ,
and
b ( k ) = n [ ( a * ( k ) + 1 ) ( i = 1 n 1 F 3 , i ( k ) + i = n 1 + 1 n 1 + n 2 E 3 , i ( k ) ) ( i = 1 n 1 F 1 , i ( k ) + i = n 1 + 1 n 1 + n 2 E 1 , i ( k ) ) ] 1 .
In addition, we say the process converges when
| a j ( k + 1 ) a j ( k ) ) | < ϵ , j = 1 , 2 , , s ,
and
| b ( k + 1 ) b ( k ) ) | < ϵ ,
where ϵ represents the tolerance threshold fixed in advance, and proper values of ϵ need to be obtained by numerical tests. Based on the above, the sequential steps of the EM algorithm are expressed in Algorithm 1 systematically.
Algorithm 1 The EM algorithm for obtaining MLEs under middle-censored data with competitive risks
  1: Input:  d a t a , initial parameter a * ( 0 ) , b ( 0 ) and a j ( 0 ) = m j a * ( 0 ) n
  2: Initialization:  k = 0
  3: while not converged do
  4:       E-step (Expectation step):
  5:       Under the current parameter estimate ( a ( k 1 ) , b ( k 1 ) ) , calculate the expectations:
E 1 , i ( k ) , E 2 , i ( k ) , E 3 , i ( k ) f o r i = n 1 + 1 , n 1 + 2 , , n 1 + n 2
F 1 , i ( k ) , F 2 , i ( k ) , F 3 , i ( k ) f o r i = 1 , 2 , , n 1

  6:       M-step (Maximization step):
  7:       Update the parameter estimates by using Equations (25) and (26)
  8:        k = k + 1
  9: end while
10: Output: Estimated parameter a j * ( k ) and b ( k )

3.3. Asymptotic Confidence Intervals

To obtain the ACIs of the unknown parameters a j , j = 1 , 2 , , s and b, we use the Fisher information matrix introduced in [23], and the matrix is derived by calculating the corresponding second-order partial derivatives. Then, we present the Fisher information matrix as:
2 ln L ( a , b ) a k a j = m j a j 2 + n 2 a * 2 B if k = j n 2 a * 2 B if k j k , j = 1 , 2 , , s ,
where
B = i = n 1 + 1 n 1 + n 2 ψ r i a * ψ l i a * [ ln ψ r i ln ψ l i ] 2 ( ψ l i a * ψ r i a * ) 2 ,
ψ l i = 1 + l i b , and  ψ r i = 1 + r i b . In addition,
2 ln L ( a , b ) a j b = i = 1 n 1 t i b ln t i 1 + t i b + C + D , j = 1 , 2 , , s ,
where
C = i = n 1 + 1 n 1 + n 2 r i b ψ r i ( a * + 1 ) ln r i l i b ψ l i ( a * + 1 ) ln l i ψ l i a * ψ r i a * ,
D = a * i = n 1 + 1 n 1 + n 2 ( ψ r i ψ l i ) ( a * + 1 ) ( ln ψ r i ln ψ l i ) ψ r i l i b ln l i ψ l i r i b ln r i ( ψ l i a * ψ r i a * ) 2 .
For the second-order derivative of b,
2 ln L ( a , b ) b 2 = n 1 b 2 ( a * + 1 ) i = 1 n 1 t i b ( ln t i ) 2 ( 1 + t i b ) 2 + E F ,
where
E = a * i = n 1 + 1 n 1 + n 2 ψ l i ( a * + 2 ) l i b ( ln l i ) 2 ( a * l i b 1 ) ψ r i ( a * + 2 ) r i b ( ln r i ) 2 ( a * r i b 1 ) ψ l i a * ψ r i a * ,
F = a * 2 i = n 1 + 1 n 1 + n 2 [ ψ l i ( a * + 1 ) l i b ln l i ψ r i ( a * + 1 ) r i b ln r i ] 2 ( ψ l i a * ψ r i a * ) 2 .
For simplicity, the Fisher matrix is given for s = 2 , and the form for other values of s can be obtained similarly. In addition, we denote ln L ( a , b ) by ln L in the following matrix. The Fisher information matrix is given by:
I ( a 1 , a 2 , b ) = E 2 ln L a 1 2 2 ln L a 1 a 2 2 ln L a 1 b 2 ln L a 2 a 1 2 ln L a 2 2 2 ln L a 2 b 2 ln L b a 1 2 ln L b a 2 2 ln L b 2 .
Replace the unknown parameters by their MLEs, then the observed Fisher matrix I o ( a 1 ^ , a 2 ^ , b ^ ) can be given as:
I o ( a 1 ^ , a 2 ^ , b ^ ) = 2 ln L a 1 2 2 ln L a 1 a 2 2 ln L a 1 b 2 ln L a 2 a 1 2 ln L a 2 2 2 ln L a 2 b 2 ln L b a 1 2 ln L b a 2 2 ln L b 2 ( a 1 , a 2 , b ) = ( a 1 ^ , a 2 ^ , b ^ ) .
We define V ( a 1 ^ , a 2 ^ , b ^ ) as the inverse matrix of I o ( a 1 ^ , a 2 ^ , b ^ ) , which is expressed as:
V ( a 1 ^ , a 2 ^ , b ^ ) = I o 1 ( a 1 ^ , a 2 ^ , b ^ ) = v a r ( a 1 ^ ) c o v ( a 1 ^ , a 2 ^ ) c o v ( a 1 ^ , b ^ ) c o v ( a 1 ^ , a 2 ^ ) v a r ( a 2 ^ ) c o v ( a 2 ^ , b ^ ) c o v ( a 1 ^ , b ^ ) c o v ( a 2 ^ , b ^ ) v a r ( b ^ ) .
Let V denote the matrix V ( a 1 ^ , a 2 ^ , , a s ^ , b ^ ) with the same form as V ( a 1 ^ , a 2 ^ , b ^ ) and V j j denote the j-th diagonal element of the covariance matrix V, j = 1 , 2 , , s + 1 . According to the asymptotic properties of MLEs, we subsequently infer that the distributions of ( a j ^ a j ) / V j j for j = 1 , 2 , , s and ( b ^ b ) / V s + 1 , s + 1 can be approximately modeled by the standard normal distribution.
For γ ( 0 , 1 ) , a specified confidence level parameter, the  ( 1 γ ) × 100 % two-sided ACIs for a 1 , a 2 , , a s are formulated as a j ^ ± Z γ 2 V j j . Similarly, the ACI for b can be constructed as b ^ ± Z γ 2 V s + 1 , s + 1 . Here, Z γ 2 refers to the upper γ / 2 percentile point of the standard normal distribution.

4. Bayesian Estimation

Different from frequentist methods, the advantage of Bayesian estimation resides in its capacity to integrate prior knowledge and sample data, enabling more stable parameter inference in scenarios with small samples. In this section, we derive the Bayesian estimates and the corresponding interval estimates in the case where the prior distribution of a j is related to b.
We suppose that the parameter b is assigned an exponential prior distribution with the mean equal to 1 λ , while each a j follows a Gamma prior distribution with the shape parameters θ j , j = 1 , 2 , , s and the rate parameter b. Abuzaid [18] also employed a similar assumption. Moreover, a 1 , a 2 , , a s are independent of each other. The PDFs of the prior distributions of a 1 , a 2 , , a s and b are presented as:
π 0 ( b ) = λ e λ b , b > 0 ; λ > 0 ,
π j ( a j | b ) = e b a j b θ j a j θ j 1 Γ ( θ j ) , a j > 0 ; θ j > 0 , j = 1 , 2 , , s ,
where λ and θ j are known.
So the joint prior density can be given by:
π ( a , b ) = ( j = 1 s π j ( a j | b ) ) π ( b ) = λ e λ b b θ * e b a * j = 1 s a j θ j 1 j = 1 s Γ ( θ j ) ,
where θ * = j = 1 s θ j .
Combining (9) and (30), the PDF of the joint posterior distribution of a 1 , a 2 , , a s and b is derived as:
π ( a , b | d a t a ) b θ * + n 1 e b ( a * + λ ) ( 1 a * ) n 2 j = 1 s a j m j + θ j 1 × i = 1 n 1 [ t i b 1 ( 1 + t i b ) ( a * + 1 ) ] i = n 1 + 1 n 1 + n 2 [ ( 1 + l i b ) a * ( 1 + r i b ) a * ] .
The squared error loss function ( L S E L ( ζ ^ , ζ ) = ( ζ ^ ζ ) 2 ) is widely used in statistics, where ζ represents the true parameter value and ζ ^ is the corresponding estimated value. Squared error loss in Bayesian estimation leads to the posterior mean as the optimal estimator, naturally integrating prior knowledge with observed data to reduce variance and providing inherent uncertainty quantification through the posterior distribution. Under the squared error loss function, the Bayesian estimators for a 1 , a 2 , , a s and b are the posterior conditional expectations, calculated based on (31).

4.1. Gibbs Sampling

Deriving expectations directly through integration from the PDF of joint posterior distribution is often analytically intractable. Hence, Gibbs sampling is employed to draw samples from the posterior distribution. Before using the Gibbs sampling, we need to confirm some theorems.
Theorem 3.
If t i ( i = 1 , 2 , , n ) are observed (i.e., the data are complete) and a = ( a 1 , a 2 , , a s ) is given, the conditional posterior density of b is log-concave.
Proof. 
When the data are complete, the joint posterior density function is expressed as:
π * ( a , b | d a t a * ) b n + θ * e b ( a * + λ ) j = 1 s a j m j + θ j 1 i = 1 n [ t i b 1 ( 1 + t i b ) ( a * + 1 ) ] ,
where d a t a * refers to the complete data.
Therefore, the conditional posterior density of b is given by:
π b ( b | a , d a t a * ) b n + θ * e b ( a * + λ ) i = 1 n [ t i b 1 ( 1 + t i b ) ( a * + 1 ) ] .
After calculation, We calculate the second-order derivative of the logarithmic function with respect to b:
2 ln π b ( b | a , d a t a * ) b 2 = n + θ * b 2 ( a * + 1 ) i = 1 n t i b ( l n t i ) 2 ( 1 + t i b ) 2 < 0 ;
thus, we can see that the conditional posterior density function of b is log-concave, which indicates that we can employ the adaptive rejection sampling [24] to draw samples from the posterior conditional distribution of b.    □
Theorem 4.
When the data are complete and b is given, the conditional posterior of a j ( j = 1 , , s ) follows a Gamma distribution.
Proof. 
Through (32), we know that the conditional posterior density functions of a j , j = 1 , 2 , , s are as follows:
π j * ( a j | a j , b , d a t a * ) e b a j a j m j + θ j 1 i = 1 n ( 1 + t i b ) a j = e a j ( b + i = 1 n ln ( 1 + t i b ) ) a j m j + θ j 1 , j = 1 , 2 , , s ,
where a j represents vector ( a 1 , a 2 , , a j 1 , a j + 1 , , a s ) . Thus, a j follows the Gamma conditional posterior distribution with the shape parameter m j + θ j and the rate parameter b + i = 1 n ln ( 1 + t i b ) .    □
For those censored data, if we have the estimated values of a j and b, we can use t ˜ i to substitute t i , where t ˜ i can be sampled from the conditional distribution of T i . The conditional density function of T i is given by:
f { T | T [ l i , r i ] , D i = j } ( t , j ; a , b ) = f ( t , j ; a , b ) F ( r i , j ; a , b ) F ( l i , j ; a , b ) , l i < t < r i .
According to the two theorems above, in Algorithm 2, the Gibbs sampling procedure is shown.
Algorithm 2 Gibbs sampling method to Bayesian estimation
1: Initialize the values of b ( 0 ) and a ( 0 ) .
    •      Generate b ( 0 ) from E x p ( λ ) ,
    •      Generate a j ( 0 ) from G a m m a ( θ j , b ( 0 ) ) , j = 1 , 2 , , s ,
2: for  k = 1 to N do
3:       Sample t ˜ n 1 + i ( k ) from f T | T [ l i , r i ] ( t ; a ( k 1 ) , b ( k 1 ) ) , i = 1 , 2 , , n 2 , then obtain
d a t a * ( k ) = ( t 1 , , t n 1 , t ˜ n 1 + 1 ( k ) , , t ˜ n 1 + n 2 ( k ) ) .

4:       Sample b ( k ) through π b ( b | a ( k 1 ) , d a t a * ( k ) ) using the adaptive rejection sampling method proposed in [24].
5:       Sample a j ( k ) from G a m m a ( m j + θ j , ω ( k ) ) , j = 1 , 2 , , s , where
ω ( k ) = b ( k ) + i = 1 n 1 ln ( 1 + t i b ( k ) ) + i = 1 n 2 ln ( 1 + ( t ˜ n 1 + i ( k ) ) b ( k ) ) .

6: end for
Remark 2.
The notation E x p ( λ ) refers to the exponential distribution with the expected value 1 λ , and G a m m a ( u , v ) refers to the Gamma distribution with the shape parameter u and the scale parameter 1 v .
Subsequently, based on the N samples generated from the posterior distribution, the Bayesian estimates of a 1 , a 2 , , a s and b under the squared error loss function could be obtained as follows:
  • a j ^ B E = k = d + 1 N a j ( k ) N d , j = 1 , 2 , , s ,
  • b ^ B E = k = d + 1 N b ( k ) N d
where d is the number of the burn-in sample fixed in advance.

4.2. HPD Credible Intervals

We choose to employ the approach proposed by [25] in this subsection to obtain the HPD credible intervals, which contains 100 ( 1 γ ) % of the posterior probability mass, with parameter values within the interval having the highest probability density. Based on the Gibbs sampling in Section 4.1, we perform the following steps to obtain HPD credible intervals.
At the start, rearrange the total N ˜ = N d drawn samples to obtain the order statistics, and they can be written as a j ( 1 ) , a j ( 2 ) , , a j ( N ˜ ) and b ( 1 ) , , b ( N ˜ ) , where j = 1 , 2 , , s . For each parameter,
A j ( k ) = ( a j ( k ) , a j ( k + [ ( 1 γ ) N ˜ ] ) ) , j = 1 , 2 , , s ,
B ( k ) = ( b ( k ) , b ( k + [ ( 1 γ ) N ˜ ] ) ) ,
where [ · ] refers to the floor function and k = 1 , 2 , , γ N ˜ .
Then, traverse all intervals and find the interval with the shortest length:
k j * = arg min 1 k γ N ˜ a j ( k + [ ( 1 γ ) N ˜ ] ) a j ( k )
k b * = arg min 1 k γ N ˜ b ( k + [ ( 1 γ ) N ˜ ] ) b ( k )
Therefore, the HPD credible intervals for each unknown parameter can be expressed as follows:
( a j ( k j * ) , a j ( k j * + [ ( 1 γ ) N ˜ ] ) ) , j = 1 , 2 , , s ,
and
( b ( k b * ) , b ( k b * + [ ( 1 γ ) N ˜ ] ) ) .

5. Simulation and Data Analysis

This section evaluates the performance of previously developed parameter estimators through Monte Carlo simulation studies. Additionally, a real-world data analysis is presented to validate the proposed methodologies.

5.1. Simulation Study

Without losing generality, in the simulation, we set s = 2; that is, there are two latent risks.
To begin with, we generate competing risks sample data from the Burr-XII distributions randomly by the inverse transform sampling method. In this simulation, we set a 1 = 0.5 , a 2 = 1 , and b = 1 . The sample size n takes values of n = 30 , n = 50 , and n = 100 , respectively. In addition, we generate the left endpoints and interval lengths from exponential distributions with mean 1 α and 1 β , respectively. We consider the following combinations for the parameter pairs ( α , β ) : ( 0.5 , 0.5 ) , ( 0.5 , 1 ) , and ( 1 , 0.5 ) . By comparing actual lifetimes with the corresponding interval, we can obtain the middle-censored data.
Then, we can calculate the MLEs and 95% ACIs for a 1 , a 2 , and b by employing the optim function or the EM algorithm.
For Bayesian estimation, regarding the informative prior, to guarantee that the expected value of the prior distribution of b coincides with the true value, we set that b has an exponential prior distribution with λ = 1 . In addition, we select the hyper-parameters as θ 1 = 0.5 and θ 2 = 1 . This method to select hyper-parameters has been used by some scholars such as [8,21].
In practical scenarios, researchers often lack prior knowledge, necessitating the use of non-informative priors in Bayesian methodologies. For parameters a j , j = 1 , 2 , , s , where Gamma distributions are adopted as the prior, it is common practice to set the shape parameter θ j to a decimal value close to 0 in the calculation process. This approach aligns with established methodologies in existing literature, such as references [8,20].
For parameter b, we introduce an improper uniform prior density, defined over the positive real numbers ( b > 0 ). Then, the generalized non-informative prior density function of b and the PDF of the joint posterior distribution function can be expressed as
π 0 0 ( b ) = 1 , b > 0 ,
π 0 ( a , b | d a t a ) b θ * + n 1 e b a * ( 1 a * ) n 2 j = 1 s a j m j + θ j 1 × i = 1 n 1 [ t i b 1 ( 1 + t i b ) ( a * + 1 ) ] i = n 1 + 1 n 1 + n 2 [ ( 1 + l i b ) a * ( 1 + r i b ) a * ] .
Through the derivation combined (32), (33), (35), and (37), in the process of Gibbs sampling, we just need to set λ = 0 in the procedure. It is worth noting that when using a non-informative prior, the method for generating initial values of a 1 , a 2 , and λ needs to be adjusted in the first step of Gibbs sampling to avoid generating 0.
We then derive the Bayesian estimates and the associated HPD credible intervals by Gibbs sampling. The sampling process consists of N = 11 , 000 iterations, during which the first d = 1000 iterations are seen as the burn-in period and are discarded.
After repeating the process above for M = 10 , 000 times, we can compare the performance of all point estimators numerically via bias and mean squared error (MSE). In addition, the outcomes of interval estimators can be assessed by means of average width (AW) and coverage percentages (CP).
In this study, the code was executed in R langurange (version 4.3.2) on a Lenovo XiaoXinPro 16ACH laptop. It is equipped with an 8-core AMD Ryzen 7 5800H with Radeon Graphics processor running at a base frequency, 16 GB of RAM with a speed of 3200 MT/s, and a 512 GB solid-state drive.
For EM algorithm, taking into account both the time cost and accuracy, we need to select a proper value for ϵ . We run one of the schemes and compare the estimation accuracy (according to the MSEs) under different ϵ values to identify an appropriate value. Then, we execute the estimation for the other schemes using this determined ϵ value. As shown in Table 1, reducing ϵ below 0.0001 does not significantly improve the MSE of parameter estimates. Thus, ϵ = 0.0001 is chosen as the optimal threshold.
Before conducting Bayesian estimation, it is essential to verify the convergence of the Gibbs sampler to ensure the reliability of posterior inference. To this end, we systematically evaluate the convergence across multiple representative censoring schemes through trace plots. These visualizations and descriptions demonstrate that the number of realizations (11,000) and the burn-in period (1000) we used are sufficient to ensure the convergence of the sampler. Take the situation where n = 50 and informative priors are used as examples. In Figure 3, Figure 4 and Figure 5, we can see that the trace plots of the three parameters all exhibit the characteristic of fluctuating around a certain value without obvious trend changes after, indicating that with the given number of iterations (N = 11,000 and d = 1000 ), the Gibbs sampling shows a convergent performance when estimating parameters a 1 , a 2 , b .
In addition, to show the details of Gibbs sampler and obtain more information of the posterior distributions, we present the histograms of Gibbs samples for unknown parameters and the estimated lifetimes of the censored data. Due to space limitations, only part of the results are presented here in Figure 6, Figure 7 and Figure 8.
Table 2, Table 3 and Table 4 show the results of the point estimators, including the MLEs and Bayesian estimators for a 1 , a 2 , and b, where ` O P T ’ and ` E M ’, respectively, refer to the MLE results obtained by the optim function and the EM algorithm. And prior 1 refers to the non-informative prior, while prior 2 refers to the informative prior. In addition, pc refers to the censoring proportion under the corresponding scheme and ‘ B E ’ refers to the Bayesian estimator. We discuss different censoring schemes by setting different values of the parameters α and β .
According to Table 2, Table 3 and Table 4 several conclusions are drawn as follows:
(1) Under different values of ( α , β ) , which represent different censoring schemes, the biases and MSEs of each estimator are very similar. This fully demonstrates that the model is robust to middle-censored data.
(2) As anticipated, for every estimator, both the bias and the MSE decline as the sample size n becomes larger. This indicates an improvement in their performance.
(3) When it comes to the results of MLE of each parameter, the performance of the two methods, namely the optimization method and the EM algorithm, is quite similar.
(4) In the situation where the sample size n is limited (e.g., n = 30 ), Bayesian estimators exhibit lower MSEs compared to MLEs, which indicates that, in small-sample scenarios, the Bayesian method outperforms the classical frequentist method.
(5) As expected, the Bayesian estimators under informative prior distribution perform better than those under non-informative prior distribution.
Table 5, Table 6 and Table 7 present the results of interval estimates of a 1 , a 2 , and b, where ‘CP’ refers to the coverage percentages and ‘AW’ refers to the average widths of interval estimates. ‘HPD-P1’ and ‘HPD-P2’ represent the HPD credible intervals under non-informative and informative prior distribution, respectively.
Through Table 5, Table 6 and Table 7, we can observe that:
(1) Similarly to point estimations, the influence of different censoring schemes on the performance of interval estimators is relatively small.
(2) As the sample size n grows, confidence interval widths decrease, and coverage percentages generally rise.
(3) Bayesian interval estimates under informative prior perform better than those under non-informative prior.

5.2. Real Data Analysis

We apply real-world data from [9] in this subsection to demonstrate the inference framework developed in this study.
The real dataset is from the National Eye Institute diabetic retinopathy study, whose purpose is to examine the influence of laser treatment on the delay of the start of blindness in those suffering from diabetic retinopathy. With respect to the i-th participant, X i , 1 represents the blind time of the eye which has received laser treatment, while X i , 2 refers to the blind time of the untreated eye. Then, the time of first onset of blindness T i = min { X i , 1 , X i , 2 } . Therefore, this dataset exhibits characteristics of competing risks data. Original data with corresponding cause are in Table 8.
To ensure that the solution exists, we divide the data by 365 to let the unit be year. It should be noted that, before further analysis, we need to first determine whether the dataset is able to be analyzed using the Burr-XII distribution. Separately for failure cause 1 and 2, initially, the complete dataset is employed to derive MLEs of unknown parameters, followed by the application of the Kolmogorov–Smirnov test. For cause 1, the MLEs of a 1 and b 1 are 0.9033 and 2.3703, respectively. The value of Kolmogorov–Smirnov statistics is 0.1168 with its corresponding p-value = 0.7978 . For cause 2, the MLEs of a 2 and b 2 are 0.9237 and 2.4605, respectively. The value of Kolmogorov–Smirnov statistics equals 0.1470 with its corresponding p-value amounting to 0.4324. Consequently, we can conclude that the Burr-XII distribution can be used to fit the dataset mentioned above.
In addition, b 1 = b 2 = b need to be tested. We use the likelihood ratio test method. Through the analysis and calculation of the sample data, the corresponding p-value is quite close to 1, which indicates that the hypothesis “ b 1 = b 2 = b ” is reasonable.
Then, we generate the artificial data. Both the left endpoint and the interval length are the random variables which follow the exponential distribution with mean 1. The processed data (measured in years) are presented in Table 9.
Since we have no extra information of a 1 , a 2 , and b, we employ the non-informative priors for them.
Using the methods and procedures of estimation outlined above, point and interval estimates for all unknown parameters are derived, with results shown in Table 10, where ‘ M L E O P T ’, ‘ M L E E M ’ represent the MLEs obtained by the optim function or the EM algorithm, respectively, and ‘BE’ refers to the Bayesian point estimator. We can see that the results from two methods to obtain MLEs are same and the widths of HPDs are less than those of ACIs.
Based on the estimated values of parameters a 1 , a 2 , and b, we can draw plots that combine the fitted probability density curves with the frequency histograms of the complete data for the two causes, respectively.
In Figure 9 and Figure 10, ‘MLE’ denotes the density curve based on MLEs under middle-censored data, ‘BE’ denotes the curve based on Bayes estimates under middle-censored data, and ‘MLE_c’ represents the density curve with parameter estimates derived from the complete data. Through Figure 9 and Figure 10, we can conclude that the estimated density curve is generally consistent with the trend of the frequency distribution of the real dataset, and its peak point is close to the peak point of the real dataset, indicating that our model is valuable in practical application.

6. Conclusions

This study mainly concentrates on statistical analysis for the middle-censored data with Burr XII distribution under independent competing risk frameworks. We apply both frequentist and Bayesian approaches to derive parameter estimates.
In the frequentist approach, we present the maximum likelihood estimates and prove their existence and uniqueness. Applying the observed Fisher information matrix, we calculate the ACIs.
Regarding Bayesian estimation, we utilize adaptive rejection sampling and Gibbs sampling techniques. These are employed to compute the Bayesian estimates under the squared error loss function, along with the corresponding HPD credible intervals.
Moreover, a simulation is performed with the aim of observing the performance of various estimations. After comparison, we find that for small-size samples, the Bayesian estimation with prior information performs better. The interval estimate obtained using the Bayesian method has a shorter width. At last, we apply our model to analyze an authentic dataset related to medicine, confirming that our model is reasonable.
Although the current study provides valuable information on the estimation issues regarding the middle-censored data combined with the independent competing risks, there are several areas for future improvement and exploration. One possible direction is to extend the analysis by considering that the value of b for the Burr-XII distribution of each failure cause may vary. Another important aspect for future research is the consideration of dependent competing risks. Within this study, we focus on the basic assumption of independent competitive risks, but in real-world scenarios, the risks may not always be independent. Exploring the implications of dependent competing risks would offer a more comprehensive analysis under complex conditions.

Author Contributions

Conceptualization: S.L. and W.G.; Methodology: S.L. and W.G.; Software: S.L.; Investigation: S.L.; Writing—Original Draft: S.L.; Writing—Review & Editing: W.G.; Supervision: W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Project 202510004172 which was supported by National Training Program of Innovation and Entrepreneurship for Undergraduates. Wenhao’s work was partially supported by the Science and Technology Research and Development Project of China State Railway Group Company, Ltd. (No. N2023Z020).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in [9].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jammalamadaka, S.R.; Mangalam, V. Nonparametric estimation for middle-censored data. J. Nonparametr. Stat. 2003, 15, 253–265. [Google Scholar] [CrossRef]
  2. Jammalamadaka, S.R.; Iyer, S.K. Approximate self consistency for middle-censored data. J. Stat. Plan. Inference 2004, 124, 75–86. [Google Scholar] [CrossRef]
  3. Iyer, S.K.; Jammalamadaka, S.R.; Kundu, D. Analysis of middle-censored data with exponential lifetime distributions. J. Stat. Plan. Inference 2008, 138, 3550–3560. [Google Scholar] [CrossRef]
  4. Davarzani, N.; Parsian, A. Statistical inference for discrete middle-censored data. J. Stat. Plan. Inference 2011, 141, 1455–1462. [Google Scholar] [CrossRef]
  5. Jammalamadaka, S.R.; Bapat, S.R. Middle censoring in the multinomial distribution with applications. Stat. Probab. Lett. 2020, 167, 108916. [Google Scholar] [CrossRef]
  6. Abuzaid, A.H.; El-Qumsan, M.K.A.; El-Habil, A.M. On the robustness of right and middle censoring schemes in parametric survival models. Commun. Stat. Simul. Comput. 2015, 46, 1771–1780. [Google Scholar] [CrossRef]
  7. Wang, L. Estimation for exponential distribution based on competing risk middle censored data. Commun. Stat. Theory Methods 2016, 45, 2378–2391. [Google Scholar] [CrossRef]
  8. Ahmadi, K.; Rezaei, M.; Yousefzadeh, F. Statistical analysis of middle censored competing risks data with exponential distribution. J. Stat. Comput. Simul. 2017, 87, 3082–3110. [Google Scholar] [CrossRef]
  9. Wang, Y.; Shi, Y.; Wu, M. Statistical inference for dependence competing risks model under middle censoring. J. Syst. Eng. Electron. 2019, 30, 209–222. [Google Scholar]
  10. Davarzani, N.; Parsian, A.; Peeters, R. Statistical Inference on Middle-Censored Data in a Dependent Setup. J. Stat. Res. 2014, 9, 646–657. [Google Scholar] [CrossRef]
  11. Sankaran, P.G.; Prasad, S. Weibull Regression Model for Analysis of Middle-Censored Lifetime Data. J. Stat. Manag. Syst. 2014, 17, 433–443. [Google Scholar] [CrossRef]
  12. Rehman, H.; Chandra, N. Inferences on cumulative incidence function for middle censored survival data with Weibull regression. J. Appl. Stat. 2022, 5, 65–86. [Google Scholar] [CrossRef]
  13. Sankaran, P.G.; Prasad, S. Additive risks regression model for middle censored exponentiated-exponential lifetime data. Commun. Stat. Simul. Comput. 2018, 47, 1963–1974. [Google Scholar] [CrossRef]
  14. Zimmer, W.J.; Keats, J.B.; Wang, F.K. The Burr XII Distribution in Reliability Analysis. J. Qual. Technol. 1998, 30, 386–394. [Google Scholar] [CrossRef]
  15. Soliman, A.A. Estimation of Parameters of Life from Progressively Censored Data Using Burr-XII Model. IEEE Trans. Reliab. 2005, 54, 34–42. [Google Scholar] [CrossRef]
  16. Yan, W.; Li, P.; Yu, Y. Statistical inference for the reliability of Burr-XII distribution under improved adaptive Type-II progressive censoring. Appl. Math. Model. 2021, 95, 38–52. [Google Scholar] [CrossRef]
  17. Du, Y.; Gui, W. Statistical inference of Burr-XII distribution under adaptive type II progressive censored schemes with competing risks. Results Math. 2022, 77, 81. [Google Scholar] [CrossRef]
  18. Abuzaid, A.H. The estimation of the Burr-XII parameters with middle-censored data. J. Appl. Probab. Stat. 2015, 4, 101. [Google Scholar] [CrossRef]
  19. Cox, D.R. The analysis of exponentially distributed life - times with two types of failure. J. R. Stat. Soc. Ser. B Stat. Methodol. 1959, 21, 411–421. [Google Scholar] [CrossRef]
  20. Qin, X.; Gui, W. Statistical inference of Burr-XII distribution under progressive Type-II censored competing risks data with binomial removals. J. Comput. Appl. Math. 2020, 378, 112922. [Google Scholar] [CrossRef]
  21. Chacko, M.; Mohan, R. Bayesian analysis of Weibull distribution based on progressive type-II censored competing risks data with binomial removals. Comput. Stat. 2018, 34, 233–252. [Google Scholar] [CrossRef]
  22. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B 1977, 39, 1–22. [Google Scholar] [CrossRef]
  23. Fisher, R.A. On the mathematical foundations of theoretical statistics. Philos. Trans. R. Soc. Lond. Ser. A Contain. Pap. A Math. Phys. Character 1922, 222, 309–368. [Google Scholar]
  24. Gilks, W.R.; Wild, P. Adaptive Rejection Sampling for Gibbs Sampling. Appl. Stat. 1992, 41, 337–348. [Google Scholar] [CrossRef]
  25. Chen, M.-H.; Shao, Q.-M. Monte Carlo Estimation of Bayesian Credible and HPD Intervals. J. Comput. Graph. Stat. 1999, 8, 69–92. [Google Scholar] [CrossRef]
Figure 1. PDF plots of Burr-XII distribution under various parameter configurations.
Figure 1. PDF plots of Burr-XII distribution under various parameter configurations.
Applsci 15 04288 g001
Figure 2. Hazard rate function plots of Burr-XII distribution under various parameter configurations.
Figure 2. Hazard rate function plots of Burr-XII distribution under various parameter configurations.
Applsci 15 04288 g002
Figure 3. Trace of parameters when n = 50 and ( α , β ) = (0.5, 0.5).
Figure 3. Trace of parameters when n = 50 and ( α , β ) = (0.5, 0.5).
Applsci 15 04288 g003
Figure 4. Trace of parameters when n = 50 and ( α , β ) = (0.5, 1).
Figure 4. Trace of parameters when n = 50 and ( α , β ) = (0.5, 1).
Applsci 15 04288 g004
Figure 5. Trace of parameters when n = 50 and ( α , β ) = (1, 0.5).
Figure 5. Trace of parameters when n = 50 and ( α , β ) = (1, 0.5).
Applsci 15 04288 g005
Figure 6. Histograms of Gibbs samples for parameters and censored lifetimes when n = 30 and ( α , β ) = ( 0.5 , 0.5 ) .
Figure 6. Histograms of Gibbs samples for parameters and censored lifetimes when n = 30 and ( α , β ) = ( 0.5 , 0.5 ) .
Applsci 15 04288 g006
Figure 7. Histograms of Gibbs samples for parameters and censored lifetimes when n = 50 and ( α , β ) = ( 0.5 , 0.5 ) .
Figure 7. Histograms of Gibbs samples for parameters and censored lifetimes when n = 50 and ( α , β ) = ( 0.5 , 0.5 ) .
Applsci 15 04288 g007
Figure 8. Histograms of Gibbs samples for parameters and censored lifetimes when n = 100 and ( α , β ) = ( 0.5 , 0.5 ) .
Figure 8. Histograms of Gibbs samples for parameters and censored lifetimes when n = 100 and ( α , β ) = ( 0.5 , 0.5 ) .
Applsci 15 04288 g008
Figure 9. The fitted PDF plots for sample data corresponding to cause 1.
Figure 9. The fitted PDF plots for sample data corresponding to cause 1.
Applsci 15 04288 g009
Figure 10. The fitted PDF plots for sample data corresponding to cause 2.
Figure 10. The fitted PDF plots for sample data corresponding to cause 2.
Applsci 15 04288 g010
Table 1. MSEs of the MLE obtained by EM algorithm with different values of ϵ .
Table 1. MSEs of the MLE obtained by EM algorithm with different values of ϵ .
ϵ a 1 a 2 b
0.0010.0169774480.0348345310.016680585
0.00010.0167345820.0355209780.016437775
0.000010.0163558640.0354803640.016265085
Table 2. Biases and MSEs corresponding to different estimates for a 1 .
Table 2. Biases and MSEs corresponding to different estimates for a 1 .
( α , β ) pcn MLEBE
OPTEMPrior 1Prior 2
(0.5, 0.5)0.1930bias0.02410.0241−0.05030.0099
mse0.03280.03300.03480.0282
50bias0.01010.0100−0.05100.0010
mse0.01660.01660.03100.0174
100bias0.00780.0077−0.0524−0.0049
mse0.00770.00770.02650.0071
(0.5, 1)0.1430bias0.01880.0186−0.04210.0019
mse0.03090.03110.03620.0246
50bias0.01020.0101−0.05010.0016
mse0.01650.01650.03320.0154
100bias0.00740.0073−0.0640−0.0011
mse0.00770.00770.03610.0076
(1, 0.5)0.2830bias0.01950.0195−0.0553−0.0036
mse0.03180.03180.03990.0253
50bias0.01040.0103−0.0509−0.0098
mse0.01690.01690.03170.0148
100bias0.00810.0080−0.0784−0.0111
mse0.00780.00780.03820.0076
Table 3. Biases and MSEs corresponding to different estimates for a 2 .
Table 3. Biases and MSEs corresponding to different estimates for a 2 .
( α , β )pcn MLEBE
OPTEMPrior 1Prior 2
(0.5, 0.5)0.1930bias0.02490.0249−0.08600.0123
mse0.06170.06220.09930.0542
50bias0.01820.0182−0.0969−0.0024
mse0.03610.03610.10020.0352
100bias0.01390.0138−0.0994−0.0061
mse0.01560.01560.09360.0164
(0.5, 1)0.1430bias0.02700.0267−0.08290.0066
mse0.06360.06410.10240.0537
50bias0.01830.0182−0.09630.0011
mse0.03540.03550.10080.0322
100bias0.01320.0129−0.1296−0.0044
mse0.01530.01530.13240.0169
(1, 0.5)0.2830bias0.02750.0275−0.11500.0015
mse0.06360.06360.10780.0593
50bias0.01880.0185−0.1198−0.0171
mse0.03690.03680.10160.0313
100bias0.01450.0143−0.1562−0.0203
mse0.01560.01550.13720.0182
Table 4. Biases and MSEs corresponding to different estimates for b.
Table 4. Biases and MSEs corresponding to different estimates for b.
( α , β )pcn MLEBE
OPTEMPrior 1Prior 2
(0.5, 0.5)0.1930bias0.04370.04400.01450.0756
mse0.02940.02960.08190.0362
50bias0.02650.0267-0.01950.0460
mse0.01630.01640.08920.0236
100bias0.01110.0117-0.04760.0324
mse0.00720.00730.08790.0090
(0.5, 1)0.1430bias0.04070.0420-0.00320.0622
mse0.02880.02900.08660.0364
50bias0.02670.0268-0.03420.0433
mse0.01610.01590.09250.0212
100bias0.01050.0115-0.09160.0243
mse0.00710.00730.12590.0101
(1, 0.5)0.2830bias0.04360.04360.03650.0942
mse0.03050.03050.10370.0477
50bias0.02820.0280-0.00170.0818
mse0.01680.01680.09370.0281
100bias0.01110.0111-0.07350.0618
mse0.00720.00720.13490.0157
Table 5. Average width and coverage percentage for interval estimation of a 1 when γ = 0.05.
Table 5. Average width and coverage percentage for interval estimation of a 1 when γ = 0.05.
( α , β )pcnACIHPD-P1HPD-P2
AWCPAWCPAWCP
(0.5, 0.5)0.19300.64900.92800.53670.85400.59140.9170
500.49380.95000.42040.86000.46100.9090
1000.34800.94600.29930.85400.32880.9430
(0.5, 1)0.14300.64200.92200.54250.86100.58450.9290
500.49200.95300.42080.85000.46110.9320
1000.34660.94300.29050.82800.32990.9410
(1, 0.5)0.28300.64950.93100.52730.84100.57810.9010
500.49660.95000.41740.84200.45210.9300
1000.34980.94700.28230.78500.32420.9310
Table 6. Average width and coverage percentage for interval estimation of a 2 when γ = 0.05.
Table 6. Average width and coverage percentage for interval estimation of a 2 when γ = 0.05.
( α , β )pcnACIHPD-P1HPD-P2
AWCPAWCPAWCP
(0.5, 0.5)0.19300.92640.93800.79190.86400.85710.9390
500.70950.94100.60880.85200.66380.9350
1000.49870.95000.43090.85500.47190.9200
(0.5, 1)0.14300.91910.93700.79390.88000.85190.9320
500.70440.94200.60820.87300.66560.9300
1000.49510.95300.41660.82100.47290.9400
(1, 0.5)0.28300.93740.94600.76970.85000.84840.9100
500.71700.94000.59830.83600.65630.9320
1000.50350.95800.40460.80200.46710.9210
Table 7. Average width and coverage percentage for interval estimation of b when γ = 0.05.
Table 7. Average width and coverage percentage for interval estimation of b when γ = 0.05.
( α , β )pcnACIHPD-P1HPD-P2
AWCPAWCPAWCP
(0.5, 0.5)0.19300.68300.95500.59510.90500.61720.9410
500.51810.96100.44670.86300.47050.9240
1000.35920.95100.30640.86000.32960.9410
(0.5, 1)0.14300.66210.94300.58390.89200.61110.9430
500.50430.95500.43860.87400.46850.9260
1000.34970.96600.29160.82400.32690.9460
(1, 0.5)0.28300.72390.97600.61670.86300.63410.9140
500.54860.98200.45850.84500.49060.9040
1000.37830.97900.30030.78600.34160.8880
Table 8. The original data of the blind time with corresponding cause.
Table 8. The original data of the blind time with corresponding cause.
Unit: Days
(266, 1)(583, 1)(79, 1)(93, 1)(805, 1)(344, 1)(306, 1)(415, 1)
(178, 1)(1484, 1)(315, 1)(1252, 1)(642, 1)(407, 1)(356, 1)(699, 1)
(667, 1)(126, 1)(350, 1)(84, 1)(392, 1)(901, 1)(276, 1)(520, 1)
(503, 1)(584, 1)(355, 1)(1302, 1)(91, 2)(154, 2)(547, 2)(707, 2)
(469, 2)(1313, 2)(790, 2)(125, 2)(777, 2)(307, 2)(637, 2)(577, 2)
(517, 2)(287, 2)(717, 2)(141, 2)(427, 2)(36, 2)(588, 2)(350, 2)
(567, 2)(1140, 2)(448, 2)(904, 2)(485, 2)(248, 2)(423, 2)(285, 2)
(315, 2)(727, 2)(210, 2)(409, 2)(227, 2)
Table 9. The artificial middle-censored data.
Table 9. The artificial middle-censored data.
CauseData (Years)
Cause 10.72880.21640.25482.2055
0.94251.13704.06581.0740
0.86303.43011.75891.1151
1.91511.82740.95890.2301
0.97263.56710.75621.4247
1.6000[0.0911, 5.0378][0.6011, 1.7187][0.3192, 0.6168]
[0.8229, 1.0968][0.0267, 0.5380][2.0658, 4.4214][0.1413, 2.1862]
Cause 20.24931.49861.93701.2849
3.59732.16440.34252.1288
1.74521.58081.96440.3863
1.16990.09861.55341.2274
2.47671.15890.78080.8630
1.99180.57531.12050.6219
[0.1226, 0.9680][0.6727, 2.0312][0.4918, 2.2852][0.4439, 0.9947]
[0.7571, 2.0435][0.0610, 1.6743][0.9411, 3.4598][0.3921, 1.3314]
[0.1619, 1.0728]
Table 10. The result of different point and interval estimates for a 1 , a 2 , b .
Table 10. The result of different point and interval estimates for a 1 , a 2 , b .
MLE OPT MLE EM BEACIHPD
a 1 0.43390.43390.3025[0.2584, 0.6095][0.1832, 0.4357]
a 2 0.51140.51140.3571[0.3182, 0.7047][0.1664, 0.4951]
b2.32922.32922.0231[1.6961, 2.9624][1.5444, 2.5151]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, S.; Gui, W. Parametric Estimation and Analysis of Lifetime Models with Competing Risks Under Middle-Censored Data. Appl. Sci. 2025, 15, 4288. https://doi.org/10.3390/app15084288

AMA Style

Liang S, Gui W. Parametric Estimation and Analysis of Lifetime Models with Competing Risks Under Middle-Censored Data. Applied Sciences. 2025; 15(8):4288. https://doi.org/10.3390/app15084288

Chicago/Turabian Style

Liang, Shan, and Wenhao Gui. 2025. "Parametric Estimation and Analysis of Lifetime Models with Competing Risks Under Middle-Censored Data" Applied Sciences 15, no. 8: 4288. https://doi.org/10.3390/app15084288

APA Style

Liang, S., & Gui, W. (2025). Parametric Estimation and Analysis of Lifetime Models with Competing Risks Under Middle-Censored Data. Applied Sciences, 15(8), 4288. https://doi.org/10.3390/app15084288

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop