Next Article in Journal
Thermodynamic Formalism in Neuronal Dynamics and Spike Train Statistics
Previous Article in Journal
A Discretization Approach for the Nonlinear Fractional Logistic Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Load-Sharing Model under Lindley Distribution and Its Parameter Estimation Using the Expectation-Maximization Algorithm

1
Applied Statistics Laboratory, Department of Industrial Engineering, Pusan National University, Busan 46241, Korea
2
Department of Management Science and Statistics, The University of Texas at San Antonio, San Antonio, TX 78249, USA
3
Mathematical Sciences Department, College of Science, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
Department of Statistics, Al-Azhar University, Cairo 11884, Egypt
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(11), 1329; https://doi.org/10.3390/e22111329
Submission received: 19 October 2020 / Revised: 12 November 2020 / Accepted: 20 November 2020 / Published: 22 November 2020
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
A load-sharing system is defined as a parallel system whose load will be redistributed to its surviving components as each of the components fails in the system. Our focus is on making statistical inference of the parameters associated with the lifetime distribution of each component in the system. In this paper, we introduce a methodology which integrates the conventional procedure under the assumption of the load-sharing system being made up of fundamental hypothetical latent random variables. We then develop an expectation maximization algorithm for performing the maximum likelihood estimation of the system with Lindley-distributed component lifetimes. We adopt several standard simulation techniques to compare the performance of the proposed methodology with the Newton–Raphson-type algorithm for the maximum likelihood estimate of the parameter. Numerical results indicate that the proposed method is more effective by consistently reaching a global maximum.

1. Introduction

We consider a load-sharing system which consists of several components associated in parallel. As each of the components fails one by one in the system, the surviving components will share the load with the remaining surviving components.
The load-sharing system has been studied since the 1940s [1]. Since then, it is widely used in many engineering applications [2,3,4]. The studies in [5,6] drew, in particular, load-sharing parameter inferences in a system with each component with a constant failure rate. Singh et al. [7] presented a more generalized load-sharing system under an assumption of a constant hazard at the beginning and a linearly increasing hazard rate after the failure of several components. Wang et al. [8] studied a more generalized load-sharing system with the work history or memory.
Park [9,10] considered the maximum likelihood estimate (MLE) of the parameters in the system, whose component lifetimes are distributed as exponential, Weibull, normal, and lognormal. Singh et al. [7] analyzed the discrete-type multi-component load-sharing parallel system and later on Deshpandé et al. [11] presented the nonparametric-type load-sharing system.
Singh and Gupta [12,13] and Singh et al. [14] investigated the load-sharing system model, in which the failure times of each component can be either constant or time-dependent. It deserves mentioning that Bayesian analyses of the load-sharing systems with several well-known distributions of lifetime components have also been investigated in the literature [15,16,17]. For instance, Xu et al. [15] studied a test scheme for reliability demonstration test on load-sharing systems with Weibull-distributed (including exponential) lifetime components. More recently, Singh and Goel [16] studied a multi-component parallel system model by assuming a load-sharing tendency on its components. Xu et al. [17] proposed the expectation-maximization (EM) algorithm for estimating the parameters in the load-sharing system with continuous degradation.
It is worthwhile to mention that based on accelerated testing models, the MLE method for the parameter estimation of the load-sharing systems under the classical latent variable approach is not well studied when the normal and lognormal are assumed to be the lifetime distribution of components. As mentioned in [5], a direct generalization of the load-sharing system with these lifetime distributions is usually problematic when making likelihood-based parametric inference, because of following reasons: (a) it may be impossible or extremely difficult to obtain explicit MLEs, (b) standard numerical methods such as the Newton–Raphson-type method can perform poorly when the dimension of parameters is high, or  (c) if the likelihood function is not sharp enough, then the standard numerical methods may fail to reach the global maximizer.
In this paper, we extend the load-sharing system with Lindley-distributed component lifetimes. Since the Lindley distribution was proposed by [18] as a counterexample of fiducial statistics, it has been widely used in reliability engineering due to its non-constant hazard rate in comparison to the exponential distribution. A natural question that arises is about making statistical inference of the parameters in this system. As mentioned above, the Newton–Raphson-type method can be sensitive to starting values so that it often fails to reach a correct solution especially for high-dimensional parameters. An alternative is to consider the EM algorithm particularly in engineering applications. For example, Albert and Baxter [19] derived the EM sequences with several causes, censoring, partial masking and right-censoring for the exponential model. Park [20] generalized their EM-type results in the system with different multi-modal strength distributions. Amari et al. [21] used the EM method to evaluate incomplete warranty data. Recently, Park [22] analyzed interval-data using the quantile implementation of the EM algorithm which improves the Monte Carlo EM (MCEM) algorithm’s convergence and stability properties and Ouyang [23] considered an interval-based model for analyzing incomplete data in a different viewpoint. These observations motivate us to develop an EM-type MLE for the parameters in the considered system to overcome the possible difficulties mentioned above when using the standard numerical methods.
The rest of the paper is as follows. In Section 2, we discuss the construction of the likelihood of the system with Lindley-distributed component lifetimes. In Section 3, we develop an EM algorithm for performing the MLE of the system under consideration. In Section 4, we conduct numerical studies to investigate the performance of the proposed EM-type MLE method, and some concluding remarks are given in Section 5.

2. Likelihood Construction

In this section, we briefly overview the construction of the likelihood function for the load-sharing system in Section 2.1 and then focus on the working likelihood in the load-sharing system with Lindley-distributed component lifetimes in Section 2.2.

2.1. Working Likelihood for the Load-Sharing Model

We here provide a brief description of the likelihood construction needed for a load-sharing framework and refer the interested readers to [9,10] for details on various types of load-sharing models. We assume that in a load-sharing parallel system consisting of J components, the failure rates of the remaining surviving components could change and the lifetimes of the surviving components at each failure step are independent and identically distributed (iid). However, since each of the components fails one by one (without being replaced or repaired) in the system, the parameter of the remaining surviving components should change from θ ( 0 ) to θ ( 1 ) after the occurrence of the first failure. Similarly, it also changes from θ ( j 1 ) to θ ( j ) after the occurrence of the j-th failure. Let Y ( j ) be the lifetime between the failures of the j-th and ( j + 1 ) -st for j = 0 , 1 , , J 1 . We will observe the value of Y ( j ) after the failure of the j-th component. For notational simplicity, we assume that the component fails in the order of the J-th, the  ( J 1 ) -st, etc.
When analyzing the lifetime distribution in a load-sharing system, standard methods are developed based on the use of hypothetical latent random variables. After the occurrence of the j-th failure, there are J j surviving components in the system. These surviving components are denoted by U 1 ( j ) , U 2 ( j ) , , U J j ( j ) , respectively. Then, we follow the traditional latent variable approach and model
Y ( j ) = min U 1 ( j ) , U 2 ( j ) , , U J j ( j ) .
For a parametric model, each lifetime distribution usually includes the unknown parameters. We shall be particularly interested in obtaining the MLEs of the unknown parameters in the system. Given that it is often impossible to obtain the analytic-form MLE of the parameters, we often employ some reliable numerical procedures for obtaining the MLEs. Of particular note is that many lifetimes of the non-observed components can be considered to be right-censoring. For more details, the reader is directed to Equation (1.1) in [24], Section 5.3.1 in [25], Example 6.30 in [26], and the EM algorithm with right-censoring in Park [20]. This observation motivates us to adopt a similar method for constructing the likelihood of the load-sharing system as follows.
We assume that for a random sample of size n, the i-th measurement can be categorized as below
Y i ( j ) = min U i , 1 ( j ) , U i , 2 ( j ) , , U i , J j ( j ) .
Let the realization of the random variables Y i ( j ) be y i ( j ) . By using the formulation in [9,10], we can construct the observed likelihood function from the n observations as follows
L ( Θ ) i = 1 n j = 0 J 1 f Y ( j ) ( y i ( j ) ) ,
where Θ = ( θ ( 0 ) , θ ( 1 ) , , θ ( J 1 ) ) ,
f Y ( j ) ( y ) = = 1 J j S U ( j ) y × = 1 J j h U ( j ) y and h U ( j ) y = f U ( j ) y S U ( j ) y .
Under the equal load-sharing rule, we consider the parameter estimates for the system with Lindley-distributed component lifetimes. For the notion of completeness, we here review the likelihood construction provided in [10]. We assume that for a specific value of j, U ( j ) are iid for = 1 , 2 , , J j . The cumulative distribution function (cdf) of Y ( j ) can then be defined as
P Y ( j ) y = 1 P Y ( j ) > y = 1 S U ( j ) y J j ,
where S U ( j ) ( · ) is the survival function of U ( j ) with the parameter θ ( j ) , where j = 0 , 1 , , J 1 . Note that we omitted the index for S U ( j ) ( · ) since U ( j ) are iid for any .
Let h U ( j ) ( y ) = f U ( j ) ( y ) / S U ( j ) ( y ) be the hazard function. The probability density function (pdf) of Y ( j ) can be easily obtained by differentiating Equation (2), and it is given by
f Y ( j ) ( y ) = ( J j ) h U ( j ) ( y ) S U ( j ) ( y ) J j ,
Then the likelihood and log-likelihood functions for the n observations are, respectively, given by
L ( Θ ) i = 1 n j = 0 J 1 f Y ( j ) ( y i ( j ) )
and
( Θ ) i = 1 n j = 0 J 1 log f Y ( j ) ( y i ( j ) ) .

2.2. Likelihood Construction with the Lindley Distribution

We focus on the load-sharing system with Lindley-distributed component lifetimes. The pdf and cdf of the Lindley distribution with the parameter θ > 0 are given by
f ( x ) = θ 2 ( 1 + x ) e θ x ( θ + 1 ) and F ( x ) = 1 ( θ + 1 + θ x ) e θ x θ + 1 ,
respectively, for  x > 0 . Then the survival and hazard functions of the Lindley distribution are given by S ( x ) = 1 F ( x ) and h ( x ) = f ( x ) / S ( x ) , respectively. Based on the Lindley distribution above, we have the survival and hazard functions of U ( j )
S U ( j ) ( y ) = ( θ ( j ) + 1 + θ ( j ) y ) e θ ( j ) y θ ( j ) + 1 and h U ( j ) ( y ) = θ ( j ) 2 ( 1 + y ) θ ( j ) + 1 + θ ( j ) y .
Thus, using Equations (3) and (7), it follows that
f Y ( j ) ( y ) = ( J j ) θ ( j ) 2 ( 1 + y ) θ ( j ) + 1 + θ ( j ) y ( θ ( j ) + 1 + θ ( j ) y ) e θ ( j ) y θ ( j ) + 1 J j .
Substitution of Equations (8) into (5) leads to
( Θ ) 2 n j = 0 J 1 log θ ( j ) + i = 1 n j = 0 J 1 log ( 1 + y i ( j ) ) + i = 1 n j = 0 J 1 ( J j 1 ) log ( θ ( j ) + 1 + θ ( j ) y i ( j ) ) n j = 0 J 1 ( J j ) log ( θ ( j ) + 1 ) i = 1 n j = 0 J 1 ( J j ) θ ( j ) y i ( j ) .
To the best of our knowledge, it is actually impossible to derive an analytic-expression of the MLE of the likelihood function in Equation (9). Thus, appropriate numerical procedures are usually required to obtain the MLE under this scenario, whereas, as afore-mentioned, the conventional Newton–Raphson-type method is very sensitive to starting values and it could often fail to reach a correct solution especially for high-dimensional parameters. Given that the likelihood in Equation (9) can be easily over-parameterized, we recommend the use of an EM algorithm to overcome these difficulties in maximizing the likelihood in Equation (4) discussed as follows.

3. The Proposed EM Algorithm

We first briefly discuss how to implement an EM algorithm based on the complete-data likelihood function in Section 3.1 and develop the EM-type MLEs for the load-sharing system with Lindley-distributed component lifetimes in Section 3.2.

3.1. An EM Algorithm with the Load-Sharing Model

The EM algorithm [24] is a numerical iterative algorithm to obtain the MLE of parametric models. The EM algorithm consists of two steps which are (i) an expectation step (E-step) and (ii) a maximization step (M-step). It is often employed to solve a difficult or complex likelihood problem to obtain the MLE by iterating two easier steps above. In the E-step, we calculate the expectation of the log-likelihood function under the incomplete data given the observed data. In the M-step, we calculate the maximizer of the expected log-likelihood. Although this algorithm may be slower than the gradient-based methods such as Newton–Raphson, it is very stable. For the sake of completeness, we briefly discuss the EM algorithm when missing values such as censored or masked data are considered. For more details on the term missing values in terms of data coarsening, the reader is referred to the references [26,27,28].
Let θ denote the set of unknown parameters. We assume that the complete data x = ( x 1 , x 2 , , x n ) can be reorganized as x = ( y , z ) , where y = ( y 1 , y 2 , , y m ) represents the fully observed part and z = ( z m + 1 , z m + 2 , , z n ) stands for the missing part (right-censored in the load-sharing) [24]. Then the likelihood function becomes
L c ( θ y , z ) = i = 1 m f ( y i θ ) i = m + 1 n f ( z i θ )
which can also be rewritten as
L c ( θ x ) = i = 1 n f ( x i ) .
For more details on the above, see Section 5.3.1 in [25] and Section 6.3.3 in [26]. Let θ s be the estimate at the s-th step of the algorithm and define
E log L c ( θ y , z ) = log L c ( θ y , z ) f ( z y , θ s ) d z .
The two distinct steps of the EM algorithm can be summarized as follows:
  • E-step:
    Calculate : Q ( θ θ s ) = E log L c ( θ y , z ) .
  • M-step:
    Obtain : θ s + 1 = arg max θ Q ( θ θ s ) .
As yet, it is not clear how to implement the EM algorithm for performing the MLE of the considered load-sharing system. Given that the lifetime between the failures of the j-th and ( j + 1 ) -st is observed from the minimum lifetime among the remaining J j components, we could view this load-sharing system in the setting of competing risks [29]. Consequently, after the occurrence of the j-th failure, let U 1 ( j ) , U 2 ( j ) , , U J j ( j ) be the lifetimes of the remaining J j surviving components, in which these components will compete with each other for the next failure. Also, given that the lifetime between the failures of the j-th and ( j + 1 ) -st are partially observed with the unknown component, we may view the system as masking [30]. Then, it is equivalent to assuming that causes (failed components) are completely masked or missing [20,31]. Thus, analogous to his approach to complete masking, we can easily construct the likelihood function of the system for implementing the EM algorithm as follows.
First, we use an indicator function and a variable to express the cause of failure of the component: let I { C D } be the indicator function of two event C and D and Δ i ( j ) be an indicator variable such that Δ i ( j ) = indicating that the failure of the -th component. Then we define
B i ( j ) ( ) = I { Δ i ( j ) = y i ( j ) } for = 1 , 2 , , J j .
Since B i ( j ) ( ) are the indicator functions of the events for = 1 , 2 , , J j , they have an iid Bernoulli distribution with the probability P B i ( j ) ( ) = 1 , which is obtained as
P B i ( j ) ( ) = 1 = P Δ i ( j ) = y i ( j ) ,
for = 1 , 2 , , J j . Analogous to the approach in Section 8.2.2 of [32], the Bernoulli probability is obtained by
P B i ( j ) ( ) = 1 = 1 J j ,
where i = 1 , 2 , , n and = 1 , 2 , , J j . For the detailed explicit derivation of Equation (11), one can refer to Appendix A in [10].
Next, we incorporate the above Bernoulli random variables and construct the complete-data likelihood function as follows. To be more specific, by using Equation (4), the likelihood function can be constructed as L ( Θ ) i = 1 n j = 0 J 1 f Y ( j ) ( y i ( j ) ) , which can be rewritten as L ( Θ ) j = 0 J 1 L ( θ ( j ) ) , where  L ( θ ( j ) ) = i = 1 n f Y ( j ) ( y i ( j ) ) . Using the above Bernoulli random variables in Equation (10), we can rewrite f Y ( j ) ( y i ( j ) ) for the complete-data likelihood as below
f Y ( j ) ( y i ( j ) ) = = 1 J j f U ( j ) ( y i ( j ) ) B i ( j ) ( ) × S U ( j ) ( y i ( j ) ) 1 B i ( j ) ( ) ,
which results in the following likelihood function
L ( θ ( j ) ) = i = 1 n = 1 J j f U ( j ) ( y i ( j ) ) B i ( j ) ( ) × S U ( j ) ( y i ( j ) ) 1 B i ( j ) ( ) .
For convenience, we rewrite the above likelihood function as
L ( θ ( j ) ) = i = 1 n L ( θ ( j ) y i ( j ) ) ,
where L ( θ ( j ) y i ( j ) ) stands for the likelihood function of the i-th system given by
L ( θ ( j ) y i ( j ) ) = = 1 J j f U ( j ) ( y i ( j ) ) B i ( j ) ( ) × S U ( j ) ( y i ( j ) ) 1 B i ( j ) ( ) .
We observe that the complete-data likelihood function in (13) can now be constructed based on the Bernoulli random variable B i ( j ) ( ) in (10), whereas it still seems difficult to directly derive an explicit Q ( · ) function from Equation (13) in the E-step. We may overcome this difficulty by viewing the censored data as missing. This idea could help us avoid the use of the survival function, since in many practical cases, it is easier to deal with a density than a survival function. We thus provide a method for constructing the complete-data likelihood based on a density rather than a survival function, which could ease the difficulty in calculating the E-step. It should be noted that S U ( j ) ( y i ( j ) ) = y i ( j ) f U ( j ) ( z ) d z . Thus, by introducing the random variable Z i for Z i > y i ( j ) , we can replace S U ( j ) ( y i ( j ) ) in Equation (14) with f U ( j ) ( Z i ( j ) ) in the E-step. That is, L ( θ ( j ) y i ( j ) ) can be re-expressed as
L c ( θ ( j ) y i ( j ) , Z i ( j ) ) = = 1 J j f U ( j ) ( y i ( j ) ) B i ( j ) ( ) × f U ( j ) ( Z i ( j ) ) 1 B i ( j ) ( ) ,
where the pdf of Z i is defined as
f Z ( j ) ( t ) = f U ( j ) ( t ) 1 F U ( j ) ( y i ( j ) ) for t > y i ( j ) .
This technique is widely used for simplifying the calculation of the E-step. The reader is referred to Equation (1.1) in [24], Section 5.3.1 in [25], Example 6.30 in [26], and Park [20] for details. Now, using Equation (15), we can obtain an explicit closed-form Q ( · ) function which will be shown in Equation (21). This brings great simplicity for finding the MLE in the M-step.
Since P [ B i ( j ) ( ) = 1 ] = 1 / ( J j ) for all i and as shown in Equation (11), we just write B ( j ) by ignoring i and for notational simplicity. Then the likelihood function of the complete-data in Equation (15) can be written as
L c * ( θ ( j ) ) = i = 1 n L c ( θ ( j ) y i ( j ) , Z i ( j ) ) = i = 1 n = 1 J j f U ( j ) ( y i ( j ) ) B ( j ) × f U ( j ) ( Z i ( j ) ) 1 B ( j ) = i = 1 n f U ( j ) ( y i ( j ) ) B ( j ) × f U ( j ) ( Z i ( j ) ) 1 B ( j ) J j .
In what follows, using Equation (17), we implement the EM algorithm for finding the MLE of the parameters in the load-sharing system with Lindley-distributed component lifetimes.

3.2. The EM-Type Maximum Likelihood Estimates with the Lindley Distribution

Since the lifetime of the components follows the Lindley distribution with the pdf in Equation (6). Then, the pdf of U ( j ) is defined as
f U ( j ) ( t ) = θ ( j ) 2 θ ( j ) + 1 ( 1 + t ) e θ ( j ) t .
We substitute this function into Equation (17) and take the logarithm of L c * ( θ ( j ) ) , which leads to the complete-data log-likelihood function of θ ( j ) given by
log L c * ( θ ( j ) ) = ( J j ) [ 2 n log θ ( j ) n log ( θ ( j ) + 1 ) + B ( j ) i = 1 n log ( 1 + y i ( j ) ) B ( j ) θ ( j ) i = 1 n y i ( j ) + ( 1 B ( j ) ) i = 1 n log ( 1 + Z i ( j ) ) ( 1 B ( j ) ) θ ( j ) i = 1 n Z i ( j ) ] = ( J j ) 2 n log θ ( j ) n log ( θ ( j ) + 1 ) B ( j ) θ ( j ) i = 1 n y i ( j ) ( 1 B ( j ) ) θ ( j ) i = 1 n Z i ( j ) + C ,
where
C = ( J j ) B ( j ) i = 1 n log ( 1 + y i ( j ) ) + ( 1 B ( j ) ) i = 1 n log ( 1 + Z i ( j ) ) .
It should be noted that C does not include θ ( j ) . Thus, we can treat it as a constant in the M-step, where  Q ( · ) function is differentiated with respect to θ ( j ) .
Let θ s ( j ) be the estimate of θ ( j ) at the s-th EM sequences. We can summarize the proposed EM as follows.
  • E-step:
    Using Equations (11) and (18), we obtain
    Q ( θ ( j ) θ s ( j ) ) = E [ log L c * ( θ ( j ) ) θ s ( j ) ] = ( J j ) [ 2 n log θ ( j ) n log ( θ ( j ) + 1 ) 1 J j θ ( j ) i = 1 n y i ( j ) 1 1 J j θ ( j ) i = 1 n E [ Z i ( j ) θ s ( j ) ] ] + D .
    where D = E [ C θ s ( j ) ] . Since the lifetime of each component follows the Lindley distribution, we observe from Equation (16) that the pdf of Z i ( j ) θ s ( j ) can be written as
    f Z ( j ) ( z ) = θ s ( j ) 2 θ s ( j ) + θ s ( j ) y i ( j ) + 1 ( 1 + z ) e θ s ( j ) ( z y i ( j ) ) ,
    where z > y i ( j ) . Then we obtain
    E Z i ( j ) θ s ( j ) = y i ( j ) z f Z ( j ) ( z ) d z = y i ( j ) + θ s ( j ) + θ s ( j ) y i ( j ) + 2 θ s ( j ) ( θ s ( j ) + θ s ( j ) y i ( j ) + 1 ) .
    It is immediate upon substituting Equation (20) into (19) that we obtain
    Q ( θ ( j ) θ s ( j ) ) = n ( J j ) 2 log θ ( j ) log ( θ ( j ) + 1 ) θ ( j ) w ¯ s ( j ) + D ,
    where
    w ¯ s ( j ) = 1 n i = 1 n y i ( j ) + J j 1 J j θ s ( j ) + θ s ( j ) y i ( j ) + 2 θ s ( j ) ( θ s ( j ) + θ s ( j ) y i ( j ) + 1 ) .
  • M-step:
    We differentiate Q ( θ ( j ) θ s ( j ) ) with respect to θ ( j ) and obtain
    Q ( θ ( j ) θ s ( j ) ) θ ( j ) = n ( J j ) 2 θ ( j ) 1 θ ( j ) + 1 w ¯ s ( j ) .
    Finally, we set the above equation to be zero in order to solve for θ ( j ) and obtain the ( s + 1 ) -st EM sequences, denoted by θ s + 1 ( j ) = θ ( j ) , such that
    θ s + 1 ( j ) = ( w ¯ s ( j ) 1 ) + ( w ¯ s ( j ) 1 ) 2 + 8 w ¯ s ( j ) 2 w ¯ s ( j ) .
We can easily obtain the MLE of the parameter in the system based on the above EM sequences. In addition, to make our procedure be accessible for practitioners, we provide the function written in R language [33] in Appendix A.

4. Numerical Study

4.1. Real Data Analysis

We consider the real data set analyzed in [12,34] to illustrate the usefulness of the proposed procedure. We refer the readers to [12,34] for more details on the description of the data. Singh and Gupta [12] used the load-sharing model with the underlying Lindley distribution and estimated θ ^ ( 0 ) = 0.036246642 , θ ^ ( 1 ) = 0.040587738 , and  θ ^ ( 2 ) = 0.069157598 based on the Newton–Raphson-type method. The corresponding log-likelihood and likelihood values are 340.2890 and 1.638193 × 10 148 , respectively.
We reanalyze the above real data with the proposed EM algorithm in Section 3. The R function for this analysis (Lindley.LS.EM) and the real data set are provided in Appendix A. Using the Lindley.LS.EM function, we obtained θ ^ ( 0 ) = 0.03624714 , θ ^ ( 1 ) = 0.04104211 , and  θ ^ ( 2 ) = 0.06915810 . The corresponding log-likelihood and likelihood values are 340.079 and 2.021003 × 10 148 , respectively. Comparing the two results, we observe that the estimates using the EM algorithm have a greater likelihood value, which clearly implies that it is more effective than the Newton–Raphson-type method by Singh and Gupta [12].

4.2. Sensitivity of Parameter Estimation Due to Starting Values

We compare the sensitivity of starting values to the Newton–Raphson-type method and the EM algorithm. To this end, we generated ten lifetimes ( n = 10 ) from a five-component parallel system with the parameter values of θ ( 0 ) = 0.01 , θ ( 1 ) = 0.02 , θ ( 2 ) = 0.03 , θ ( 3 ) = 0.04 , and  θ ( 4 ) = 0.05 . Please note that Ghitanya et al. [35] used the mixture model to generate the Lindley random sample, whereas the sample can be directly obtained by solving F ( x ) = p for x where p is a uniform random number in ( 0 , 1 ) . In Appendix B, we provided the quantile function of the Lindley distribution by solving F ( x ) = p for x. We provided the generated data in Table 1.
We first need to choose starting values for using both methods to obtain the estimates. We selected starting values at random and obtained the estimates based on the Newton–Raphson-type method and the EM algorithm. For the case of the proposed EM method, we obtained θ ^ ( 0 ) = 0.00837 , θ ^ ( 1 ) = 0.01994 , θ ^ ( 2 ) = 0.03099 , θ ^ ( 3 ) = 0.03908 and θ ^ ( 4 ) = 0.05589 along with the corresponding log-likelihood value ( Θ ^ ) = 223.63 where Θ ^ = ( θ ^ ( 0 ) , θ ^ ( 1 ) , θ ^ ( 2 ) , θ ^ ( 3 ) , θ ^ ( 4 ) ) regardless of starting values. On the other hand, for the case of the Newton–Raphson-type method, the parameter estimates are dependent on the choice of starting values. In Table 2, we reported only some non-convergent cases for brevity. These results clearly show that the Newton–Raphson-type method is sensitive to the choice of starting values.
We investigate the sensitivity of the estimates more thoroughly as follows.
Here, we choose starting values by randomly selecting a value in ( 0 , 1 ) and then estimate the five parameters using both methods. We repeated this procedure 1000 times, which results in 1000 sets of parameter estimates. We draw the box-percentile plots [36] of the estimates under each method in Figure 1. It is easily seen from the figure that the Newton–Raphson-type method shows the wide spread of the estimates indicating that it is sensitive to starting values, while the EM algorithm does not show any spread of the estimates which clearly shows that it is not sensitive to starting values at all. In the figure, we also added the mean (red dashed line) and median (blue dotted line) of the estimates using the Newton–Raphson-type method which indicate that they tend to underestimate the true MLE but their distributions are skewed to the right.
It deserves mentioning that we have more extensive results, but due to the space limitations, we briefly report some important findings. As expected, with an increasing number of parameters, the Newton–Raphson-type method becomes more unstable, and in some cases, it even fails to provide reasonable results. On the other hand, the EM algorithm always provides reliable estimates. It is of interest to note that for a small number of the parameters (e.g., five or smaller), the Newton–Raphson-type method is not so bad. In summary, when the number of the parameter is more than five, we highly recommend the use of the EM algorithm.

5. Conclusions

In this paper, we have introduced a methodology by integrating the conventional procedure under the assumption of the load-sharing system being made up of fundamental hypothetical latent random variables. In addition, we have investigated the problem of estimating the parameter of a load-sharing system with Lindley-distributed component lifetimes. To our knowledge, it is not possible to obtain a closed-form MLE of the parameter under this system. We thus develop an EM algorithm for finding the MLE of the load-sharing system under consideration. The EM algorithm is not only easily implemented by practitioners with the R programs in Appendix A, but also can alleviate some potential issues faced by iterative numerical methods such as the Newton–Raphson-type method.
We have conducted extensive simulations to compare the performance of the EM algorithm with the Newton–Raphson-type method. Numerical results indicated that the Newton–Raphson-type method often fails to converge and is very dependent on the starting values. Also, as the number of load-sharing parameters increases, it becomes increasingly ineffective because the chance of poor convergence becomes greater. Consequently, we do not recommend the use of this method for parameter estimation of the described load-sharing system. Instead, we have a preference for the proposed EM algorithm, because it consistently offers reliable results, even as the number of parameters becomes larger.

Author Contributions

C.P. developed methodology and R functions; M.W. investigated mathematical formulas and wrote the manuscript; and R.M.A. and H.R. carried out literature survey and review. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (NRF-2017R1A2B4004169). Also, this research was funded by the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University through the Fast-track Research Funding Program.

Acknowledgments

The authors wish to thank the anonymous referees for their careful reading and suggestions for improvement of the original manuscript. The authors also thank Mark Leeds for his help for writing an earlier version of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. R Programs for the Proposed EM Algorithm

# Lindley.LS.EM  function
Lindley.LS.EM <- function(Y, start = 1, maxits = 1000 L, eps = 1.0E-15) {
  J = ncol(Y)
  converged = logical(J)
  para = numeric(J)
  THETA = start
  iter = numeric(J)
  if( length(start) < J ) THETA = rep_len(start,J)
  for ( j in 0L:(J-1L) ) {
     thetaj = THETA[j+1L]
     V = (J-j-1L) / (J-j)
     y = Y[,j+1L]
     while ( (iter[j+1L]<maxits)&&(!converged[j+1L]) ) {
      w = mean(y+V*(thetaj+thetaj*y+2)/thetaj/(thetaj+thetaj*y+1))
      newtheta = (1-w+sqrt( (w-1)^2+8*w ))/(2*w)
      converged[j+1L] = abs(newtheta-thetaj) < eps
      iter[j+1L] = iter[j+1L] + 1L
      thetaj = newtheta
     }
     para[j+1L] = newtheta
  }
  list(para=para, iter=iter, conv=converged)
}
# Data
> Y0=c(21.02, 24.25,  6.55, 15.35, 39.08, 16.20, 34.59, 19.10, 28.22,
    32.00, 11.25, 17.39, 28.47, 23.42, 42.06, 28.51, 34.56, 40.33,
    27.56,  9.54, 27.09, 40.36, 41.44, 32.23,  7.53, 28.34, 26.32, 30.47)
> Y1=c(30.22, 45.54, 19.47, 16.37, 30.32,  4.16, 46.44, 38.40, 37.43,
    45.52, 19.09, 25.43, 31.15, 31.28, 23.21, 33.59, 32.53, 15.35,
    46.21, 36.21, 11.11, 33.21, 36.28,  8.17, 37.31, 35.58, 28.02, 40.4)
> Y2=c(43.43, 17.19, 23.28, 25.40, 43.53, 39.52, 16.33, 20.17, 25.41,
    39.11, 11.59, 22.51,  2.41, 40.03, 45.36, 16.20, 40.44, 28.33,
    28.05, 28.12, 23.33, 17.04, 19.13, 41.27, 13.43, 41.48, 29.33, 42.13)
Y = cbind(Y0, Y1, Y2)
# Use of Lindley.LS.EM() function
> Lindley.LS.EM(Y)
$para
[1] 0.03624714 0.04104211~0.06915810
$iter
[1] 58 37~2
$conv
[1] TRUE TRUE TRUE

Appendix B. Generating Lindley-Distributed Random Numbers

The pdf of the Lindley distribution with the parameter θ > 0 is defined as
f ( x ) = θ 2 1 + θ ( 1 + x ) e θ x = θ θ + 1 · θ e θ x + 1 θ θ + 1 · θ 2 x e θ x
for x > 0 . Thus, the Lindley distribution is regarded as the mixture of the exponential with rate θ and the gamma with ( 2 , θ ) with their mixing proportions being θ / ( θ + 1 ) and 1 θ / ( θ + 1 ) [35]. We can thus generate a Lindley random sample based on this mixture model. An alternative way is through the direct use of of the inverse cdf of the Lindley distribution, which depends on the Lambert W function. For the sake of completeness, we provide a brief introduction of the Lambert W function in the appendix and then show how it can be used to generate a Lindley random sample.
Lambert [37] first considered the Lambert’s transcendental equation which is the inverse of
z = W ( z ) e W ( z ) .
The solution W ( z ) of the above equation was first recognized in [38].
Now we consider the cdf of Lindley in (6) and we let p = F ( x ) for notational convenience. Then we have
p = 1 θ + 1 + θ x θ + 1 e θ x ,
which leads to
( θ + 1 ) ( 1 p ) e ( θ + 1 ) = ( θ + 1 + θ x ) e ( θ + 1 + θ x ) .
If we let z = ( θ + 1 ) ( 1 p ) e ( θ + 1 ) , then we can rewrite (A2) as
z = ( θ + 1 + θ x ) e ( θ + 1 + θ x ) .
Then it is immediate from (A1) and (A3) that
z = W ( z ) e W ( z ) ,
where
W ( z ) = ( θ + 1 + θ x ) .
Solving Equation (A4) for x, we obtain
x = ( θ + 1 ) W ( z ) θ .
Substituting Equations (A3) into (A5), we have
x = ( θ + 1 ) W ( θ + 1 ) ( 1 p ) e ( θ + 1 ) θ .
Figure A1. Lambert W function. W 1 ( z ) denotes the lower branch and W 0 ( z ) denotes the upper branch.
Figure A1. Lambert W function. W 1 ( z ) denotes the lower branch and W 0 ( z ) denotes the upper branch.
Entropy 22 01329 g0a1
Please note that the Lambert function W ( z ) is not injective when z is in ( 1 / e , 0 ) and has two branches as seen in Figure A1. The upper branch for the case of W > 1 is denoted by W 0 and the lower branch for the case of W < 1 is by W 1 . Then we can see that the x value in Equation (A6) is always positive in the lower branch and always negative in the upper branch. Thus, the quantile function or inverse cdf function of the Lindley distribution is explicitly expressed as
F 1 ( p ) = ( θ + 1 ) W 1 ( θ + 1 ) ( 1 p ) e ( θ + 1 ) θ .
To ensure that the cdf possesses the standard properties of the cdf, we calculate the cdf values at the endpoints 0 and as follows. As  p 1 , we know W 1 ( 0 + ) and thus we have F 1 ( p ) as expected. When p = 0 , we have
F 1 ( 0 ) = ( θ + 1 ) W 1 ( θ + 1 ) e ( θ + 1 ) θ .
It is immediate from W 1 ( a e 1 ) = a that W 1 ( θ + 1 ) e ( θ + 1 ) = ( θ + 1 ) and thus F 1 ( 0 ) = 0 as expected. The Lambert W function is available in the lamW package [39] of the R language.

References

  1. Daniels, H.E. The Statistical Theory of the Strength of Bundles of Threads I. Proc. R. Soc. Lond. Ser. A 1945, 83, 405–435. [Google Scholar]
  2. Durham, S.D.; Lynch, J.D.; Padgett, W.J. A Theoretical Justification for an Increasing Average Failure Rate Strength Distribution in Fibrous Composites. Nav. Res. Logist. 1989, 36, 655–661. [Google Scholar] [CrossRef]
  3. Durham, S.D.; Lynch, J.D.; Padgett, W.J.; Horan, T.J.; Owen, W.J.; Surles, J. Localized Load-Sharing Rules and Markov-Weibull Fibers: A Comparision of Microcomposite Failure Data with Monte Carlo Simulations. J. Compos. Mater. 1997, 31, 1856–1882. [Google Scholar] [CrossRef]
  4. Durham, S.D.; Lynch, J.D. A Threshold Representation for the Strength Distribution of a Complex Load Sharing System. J. Stat. Plan. Inference 2000, 83, 25–46. [Google Scholar] [CrossRef]
  5. Kim, H.; Kvam, P.H. Reliability Estimation Based on System Data with an Unknown Load Share Rule. Lifetime Data Anal. 2004, 10, 83–94. [Google Scholar] [CrossRef] [Green Version]
  6. Kvam, P.H.; Peña, E.A. Estimating Load-Sharing Properties in a Dynamic Reliability System. J. Am. Stat. Assoc. 2005, 100, 262–272. [Google Scholar] [CrossRef] [Green Version]
  7. Singh, B.; Sharma, K.K.; Kumar, A. A classical and Bayesian estimation of a k-components load-sharing parallel system. Comput. Stat. Data Anal. 2008, 52, 5175–5185. [Google Scholar] [CrossRef]
  8. Wang, D.; Jiang, C.; Park, C. Reliability analysis of load-sharing systems with memory. Lifetime Data Anal. 2019, 25, 341–360. [Google Scholar] [CrossRef]
  9. Park, C. Parameter Estimation for Reliability of Load Sharing Systems. IIE Trans. 2010, 42, 753–765. [Google Scholar] [CrossRef]
  10. Park, C. Parameter Estimation from Load-Sharing System Data Using the Expectation-Maximization Algorithm. IIE Trans. 2013, 45, 147–163. [Google Scholar] [CrossRef]
  11. Deshpandé, J.V.; Dewan, I.; Naik-Nimbalkar, U.V. A family of distributions to model load sharing systems. J. Stat. Plan. Inference 2010, 140, 1441–1451. [Google Scholar] [CrossRef]
  12. Singh, B.; Gupta, P.K. Load-sharing system model and its application to the real data set. Math. Comput. Simul. 2012, 82, 1615–1629. [Google Scholar] [CrossRef]
  13. Singh, B.; Gupta, P.K. Bayesian reliability estimation of a 1-out-of-k load-sharing system model. Int. J. Syst. Assur. Eng. Manag. 2014, 5, 562–576. [Google Scholar] [CrossRef]
  14. Singh, B.; Rathi, S.; Kumar, S. Inferential statistics on the dynamic system model with time dependent failure-rate. J. Stat. Comput. Simul. 2013, 83, 1–24. [Google Scholar] [CrossRef]
  15. Xu, J.; Hu, Q.; Yu, D.; Xie, M. Reliability demonstration test for load-sharing systems with exponential and Weibull components. PLoS ONE 2017, 12, e0189863. [Google Scholar] [CrossRef] [Green Version]
  16. Singh, B.; Goel, R. MCMC estimation of multi-component load-sharing system model and its application. Iran J. Sci. Technol. Trans. Sci. 2019, 43, 567–577. [Google Scholar] [CrossRef]
  17. Xu, J.; Liu, B.; Zhao, X. Parameter estimation for load-sharing system subject to Wiener degradation process using the expectation-maximization algorithm. Qual. Reliab. Eng. Int. 2019, 35, 1010–1024. [Google Scholar] [CrossRef] [Green Version]
  18. Lindley, D.V. Fiducial distributions and Bayes’ theorem. J. R. Stat. Soc. B 1958, 20, 102–107. [Google Scholar] [CrossRef]
  19. Albert, J.R.G.; Baxter, L.A. Applications of the EM algorithm to the Analysis of life length data. Appl. Stat. 1995, 44, 323–341. [Google Scholar]
  20. Park, C. Parameter Estimation of Incomplete Data in Competing Risks Using the EM algorithm. IEEE Trans. Reliab. 2005, 54, 282–290. [Google Scholar] [CrossRef]
  21. Amari, S.V.; Mohan, K.; Cline, B.; Xing, L. Expectation-Maximization Algorithm for Failure Analysis Using Incomplete Warranty Data. Int. J. Perform. Eng. 2009, 5, 403–417. [Google Scholar]
  22. Park, C. A Quantile Variant of the Expectation-Maximization Algorithm and its Application to Parameter Estimation with Interval Data. J. Algorithms Comput. Technol. 2018, 12, 253–272. [Google Scholar] [CrossRef]
  23. Ouyang, L.; Zheng, W.; Zhu, Y.; Zhou, X. An interval probability-based FMEA model for risk assessment: A real-world case. Qual. Reliab. Eng. Int. 2020, 36, 125–143. [Google Scholar] [CrossRef]
  24. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. B 1977, 39, 1–22. [Google Scholar]
  25. Robert, C.P.; Casella, G. Monte Carlo Statistical Methods, 2nd ed.; Springer: New York, NY, USA, 2005. [Google Scholar]
  26. Little, R.J.A.; Rubin, D.B. Statistical Analysis with Missing Data, 2nd ed.; John Wiley & Sons: New York, NY, USA, 2002. [Google Scholar]
  27. Heitjan, D.F.; Rubin, D.B. Inference from coarse data via multiple imputation with application to age heaping. J. Am. Stat. Assoc. 1990, 85, 304–314. [Google Scholar] [CrossRef]
  28. Heitjan, D.F.; Rubin, D.B. Ignorability and coarse data. AOS 1991, 19, 2244–2253. [Google Scholar] [CrossRef]
  29. Crowder, M.J. Classical Competing Risks; Chapman & Hall: London, UK, 2001. [Google Scholar]
  30. Miyakawa, M. Analysis of Incomplete Data in Competing Risks Model. IEEE Trans. Reliab. 1984, 33, 293–296. [Google Scholar] [CrossRef]
  31. Park, C.; Padgett, W.J. Analysis of Strength Distributions of Multi-Modal Failures Using the EM Algorithm. J. Stat. Comput. Simul. 2006, 76, 619–636. [Google Scholar] [CrossRef]
  32. Kalbfleisch, J.D.; Prentice, R.L. The Statistical Analysis of Failure Time Data; John Wiley & Sons: New York, NY, USA, 2002. [Google Scholar]
  33. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2020; Available online: http://www.r-project.org (accessed on 7 June 2020).
  34. Kvam, P.H.; Peña, E.A. Estimating Load-Sharing Properties in a Dynamic Reliability System; Technical Report; Department of Statistics, University of South Carolina: Columbia, SC, USA, 2003; Available online: https://people.stat.sc.edu/pena/TechReports/KvamPena2003.pdf (accessed on 11 November 2020).
  35. Ghitanya, M.E.; Atieha, B.; Nadarajah, S. Lindleydistribution and its application. Math. Comput. Simul. 2008, 78, 493–506. [Google Scholar] [CrossRef]
  36. Esty, W.; Banfield, J. The Box-Percentile Plot. J. Stat. Softw. Artic. 2003, 8, 1–14. [Google Scholar] [CrossRef] [Green Version]
  37. Lambert, J.H. Observations variae in Mathesin Puram. Acta Helvitica Phys. Math. Anat. Bot. Medica 1758, 3, 128–168. [Google Scholar]
  38. Pólya, G.; Szegö. Aufgaben und Lehrsätze der Analysis; Springer: Berlin, Germany, 1925. [Google Scholar]
  39. Adler, A. LamW: Lambert-W Function, R Package Version 1.3.3. 2015. Available online: https://CRAN.R-project.org/package=lamW (accessed on 11 November 2020).
Figure 1. Box-percentile plots of the estimates based on the Newton–Raphson-type method (NR) and the EM algorithm.
Figure 1. Box-percentile plots of the estimates based on the Newton–Raphson-type method (NR) and the EM algorithm.
Entropy 22 01329 g001
Table 1. Lifetimes for load-sharing sample of size n = 10 .
Table 1. Lifetimes for load-sharing sample of size n = 10 .
Y i ( 0 ) 83.1074.91164.7960.1281.91132.6816.3654.71130.9539.10
Y i ( 1 ) 59.2553.0160.3146.3820.6925.5356.1128.9923.0432.21
Y i ( 2 ) 30.1312.0632.9147.9621.644.8019.7917.1149.4366.28
Y i ( 3 ) 47.2910.456.4317.8067.0137.4141.8920.1721.6241.60
Y i ( 4 ) 46.2439.6017.6718.7250.6929.6614.1652.5310.0269.11
Table 2. Parameter estimates with corresponding log-likelihood values when the Newton–Raphson-type method is used.
Table 2. Parameter estimates with corresponding log-likelihood values when the Newton–Raphson-type method is used.
Starting ValuesParameter EstimatesLog-Likelihood
θ ( 0 ) θ ( 1 ) θ ( 2 ) θ ( 3 ) θ ( 4 ) θ ^ ( 0 ) θ ^ ( 1 ) θ ^ ( 2 ) θ ^ ( 3 ) θ ^ ( 4 ) ( Θ ^ )
0.0041.252.983.894.240.008220.020010.031040.038760.05576 223.64
0.0281.312.043.184.180.010590.019590.030470.037640.05875 224.50
0.0211.182.693.384.770.012080.022970.027310.036930.05474 226.24
0.0281.972.663.674.210.041230.024210.069830.029120.05308 307.76
0.0261.722.913.954.070.044520.017810.036170.024400.07059 308.38
0.0021.902.753.244.120.050320.019950.031640.040570.05700 323.65
0.0241.802.833.114.960.054170.019200.033640.043580.05386 336.45
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, C.; Wang, M.; Alotaibi, R.M.; Rezk, H. Load-Sharing Model under Lindley Distribution and Its Parameter Estimation Using the Expectation-Maximization Algorithm. Entropy 2020, 22, 1329. https://doi.org/10.3390/e22111329

AMA Style

Park C, Wang M, Alotaibi RM, Rezk H. Load-Sharing Model under Lindley Distribution and Its Parameter Estimation Using the Expectation-Maximization Algorithm. Entropy. 2020; 22(11):1329. https://doi.org/10.3390/e22111329

Chicago/Turabian Style

Park, Chanseok, Min Wang, Refah Mohammed Alotaibi, and Hoda Rezk. 2020. "Load-Sharing Model under Lindley Distribution and Its Parameter Estimation Using the Expectation-Maximization Algorithm" Entropy 22, no. 11: 1329. https://doi.org/10.3390/e22111329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop