On Estimation of Shannon’s Entropy of Maxwell Distribution Based on Progressively First-Failure Censored Data

: Shannon’s entropy is a fundamental concept in information theory that quantiﬁes the uncertainty or information in a random variable or data set. This article addresses the estimation of Shannon’s entropy for the Maxwell lifetime model based on progressively ﬁrst-failure-censored data from both classical and Bayesian points of view. In the classical perspective, the entropy is estimated using maximum likelihood estimation and bootstrap methods. For Bayesian estimation, two approximation techniques, including the Tierney-Kadane (T-K) approximation and the Markov Chain Monte Carlo (MCMC) method, are used to compute the Bayes estimate of Shannon’s entropy under the linear exponential (LINEX) loss function. We also obtained the highest posterior density (HPD) credible interval of Shannon’s entropy using the MCMC technique. A Monte Carlo simulation study is performed to investigate the performance of the estimation procedures and methodologies studied in this manuscript. A numerical example is used to illustrate the methodologies. This paper aims to provide practical values in applied statistics, especially in the areas of reliability and lifetime data analysis.


Introduction
Information theory offers a feasible way for quantifying uncertainty and reciprocal information of random variables in which entropy plays a significant role.Shannon Shannon [1] established the concept of statistical entropy, namely Shannon's entropy, which measures the average level of missing information in a random source, as a fundamental idea in information theory.The relationship between Shannon's entropy and the information available in classical and quantum systems is well established.In addition, statistical mechanics interpretations can be made through Shannon's entropy.Shannon's entropy is typically used in statistical mechanics to represent the entropy for classical systems with configurations taken from canonical ensembles.Moreover, Shannon's entropy is useful in many disciplines, such as computer science, molecular biology, hydrology, and meteorology, to solve different scientific problems involving uncertainties in random quantities.For example, molecular biologists use the principle of Shannon's entropy to study trends in gene sequences.Recently, Saraiva [2] exemplified Shannon's entropy in biological diversity and student migration studies and provided an intuitive introduction to Shannon's entropy.
For more details, one may refer to the excellent monograph by Cover [3] on the theory and implications of entropy with applications in various disciplines.Shannon's entropy is one of the most widely used entropies in statistics and information theory.Suppose a random variable X has probability density function (pdf) f (•), the Shannon's entropy of the random variable X, is given by Please note that the entropy in Equation ( 1) is a continuous entropy, and the measurement is relative to the coordinate system ([1] Section 20).In other words, the continuous entropy is not invariant under a variable change in general.However, the entropy based on the variable change can be expressed as the original entropy less the expected logarithm of the Jacobian of the variable change.Recently, researchers studied parametric statistical inference for measuring the entropy under different lifetime models based on complete or censored data.For example, entropy for several shifted exponential populations was studied by Kayal and Kumar [4].Another author Cho et al. [5] studied the entropy estimation for Rayleigh distribution under doubly generalized Type-II hybrid censored data.Another paper Du et al. [6] developed statistical inference for the information entropy of the log-logistic distribution based on progressively Type-I interval-censored data, and Liu and Gui [7] investigated the entropy estimation for Lomax distribution based on generalized progressively hybrid censoring.Another paper Yu et al. [8] developed statistical inference on Shannon's entropy for inverse Weibull distribution based on progressively first-failure censored data.The Maxwell-Boltzmann distribution, popularly known as the Maxwell distribution (MWD), was initially developed by James Clerk Maxwell and Ludwig Boltzmann in the late 1800s as a distribution of velocities in a gas at a specific temperature.In this paper, we consider the MWD due to its simplicity and nice physical interpretation.For further information about the reliability characteristics of the MWD, see Bekker and Roux [9].The MWD is frequently employed in physics and chemistry for several reasons, including the fact that the MWD can be used to describe several important gas properties, such as pressure and diffusion.The MWD has become a well-known lifespan model in recent years.Many researchers studied this statistical distribution in depth for modeling lifespan data.As Bekker and Roux [9] pointed out, the MWD is useful in life-testing and reliability studies because of its desirable properties, especially for situations where the assumption of a constant failure rate is not realistic.Recently, in the reliability report for Gallium Nitride (GaN) power devices by Pozo et al. [10], they found that the Maxwell-Boltzmann distribution fits the hot carrier distribution in the high energy regime tails well.
The classical estimation of the model parameters of MWD has been studied by [11,12] for complete and censored samples, respectively.Following this, Krishna and Malik [13] studied the MWD under progressive censoring, Krishna et al. [14] discussed the MWD under randomly censored data, Tomer and Panwar [15] developed an estimation procedure for the MWD based on Type-I progressive hybrid censored data, Panwar and Tomer [16] discussed robust Bayesian analysis for the MWD, and Kumari et al. [17] discussed classical and Bayesian estimation for the MWD based on adaptive progressive Type-II censored data.
The pdf and cumulative distribution function (cdf), respectively, for a random variable X that follows the Maxwell distribution (MWD) with parameter λ (denoted as MWD(λ)) are given by where x 0 e −t t b−1 dt is the incomplete gamma ratio.The MWD has an increasing failure rate function [13].The MWD is a special case of many generalized distributions, such as the generalized gamma distribution [18], generalized Rayleigh distribution ( [19] p.452), and generalized Weibull distribution [20].
Assume that the lifetime of the items of interest follow the MWD with pdf in Equation ( 2), from Equation (1), the Shannon's entropy is given by where γ = 0.5772156649 . . . is the Euler-Mascheconi constant.
Censoring often occurs in reliability engineering or life testing procedures when the actual lifetimes of the items of interest are not observed; for instance, when subjects or experimental units are taken out of the experiments intentionally or accidentally.Preplanned censoring saves time and costs for experiments, and it has been applied in various fields, including but not limited to engineering, survival analysis, clinical trials, medical research, etc.Several censoring schemes, such as Type-I and Type-II censoring schemes (also known as time and item censoring schemes, respectively), are often utilized in lifetesting experiments.When products or items have a long lifespan, a life-testing experiment may require a long time to complete, even with Type-I or Type-II censoring.For these situations, Balasooriya [21] established the first-failure censoring plan that provides a time and cost-saving strategy for life-testing experiments.In this censoring scheme, an experimenter examines k × n units by grouping them into n groups, each having k items, and then conducts all the tests jointly until the first failure is seen in each group.Although the first-failure censoring plan saves time and costs, it does not allow for the periodic removal of surviving units or subjects during the life-testing experiment.Therefore, the progressive censoring scheme proposed by Cohen [22] that allows for removing units throughout the life-testing experiment can be considered.For this reason, first-failure and progressive censoring were combined to create a more flexible life-testing strategy known as the progressive first-failure censoring (PFFC) scheme [23].Due to its compatible features with other censoring plans, the PFFC scheme has gained much attention in the literature.For instance, the Lindley distribution based on PFFC data was studied by Dube et al. [24], the exponentiated exponential distribution based on PFFC data was studied by Mohammed et al. [25], the estimation of stress-strength reliability for generalized inverted exponential distribution based on PFFC data was studied by Krishna et al. [26].One author Kayal et al. [27] studied the PFFC scheme and developed statistical inference on the Chen distribution.Following on this, Saini et al. [28] studied the estimation of stress-strength reliability for generalized Maxwell distribution based on PFFC data, and Kumar et al. [29] discussed the reliability estimation in inverse Pareto distribution based on PFFC data, etc.
The PFFC scheme can be described as follows: suppose there are n independent groups, each with k items, these are placed on a life test, and we have a prespecified number of observed failures m (≤ n) and a prefixed progressive censoring plan G = (G 1 , G 2 , . . ., G m ), where G i (i = 1, 2, . . ., m) are the prefixed number of groups without item failure to be removed at i-th failure such that n In other words, the test will be terminated when the m-th failure is observed.When the first failure occurs at time X 1:m:n:k , G 1 groups without item failure and the group where the first failure is observed are removed from the experiment.When the second failure occurs at time X 2:m:n:k , G 2 groups without item failure and the group containing the second failure are removed from the experiment, and so on.Finally, when the m-th failure occurs at time X m:m:n:k , the remaining G m groups without item failure and the group containing m-th failure are removed from the experiment.Consequently, the observed failure times, m:m:n:k , are called progressively first-failure censored order statistics with the progressive censoring plan G = (G 1 , G 2 , . . ., G m ). Figure 1 represents the schematic diagram of the PFFC scheme.Note that the PFFC scheme bears the following special cases: (i) it reduces to a complete sample case when k = 1, n = m and G i = 0; i = 1, 2, . . .m; (ii) it reduces to a conventional Type-II censoring plan, if k = 1 and G i = 0; i = 1, 2, . . .m − 1, and G m = n − m; (iii) it becomes a progressive Type-II censoring plan when k = 1, and (iv) it reduces to a firstfailure censoring plan, when m = n and G i = 0; i = 1, 2, . . .m.Moreover, the applicability or practicality of PFFC may be a particular setup of a parallel-series system in which n homogeneous system like a batch of electric bulbs are connected parallel, each batch has k bulbs or electronics components connected in series.For the testing procedures or collecting lifetimes, the same mechanism presented in Figure 1 can be used.
m:m:n:k ) be a PFFC sample drawn from a continuous population with cdf F(x) and pdf f (x).For notation convenience, we suppress the notations G and (m : n : k) in the observed data and denote the observed data as x = (x 1 , x 2 , . . . ,x m ).The likelihood function can be expressed as [23] where To establish any information based on the available data, we can consider a statistical probability model involving some parameter(s), and first, we need to estimate the model parameter(s).The two popular estimation methods in the literature are classical and Bayesian.The Bayesian estimation can be used when the prior information is available.In this study, we consider both estimation methods for feasibility purposes.The main objective of this study is to estimate the associate parameter and Shannon's entropy for the MWD based on progressively first-failure censored data.Through the study in this manuscript, we aim to provide practical values in applied statistics, especially in the areas of reliability and lifetime data analysis.
The rest of this paper is organized as follows.Section 2 develops the frequentist estimation techniques, including the maximum likelihood estimation and asymptotic and bootstrap confidence intervals.Section 3 is devoted to Bayesian estimation techniques using the Tierney-Kadane (T-K) approximation and Markov Chain Monte Carlo (MCMC) methods.In Section 4, a Monte Carlo simulation study is used to evaluate the performance of the estimation procedures developed in this manuscript.A numerical example is provided in Section 5 to illustrate the methodologies developed in this manuscript.Finally, in Section 6, some concluding remarks are presented.

Frequentist Estimation Approach
In this section, we develop the expectation-maximization (EM) algorithm to obtain the maximum likelihood estimates (MLEs) of the related parameters λ and Shannon's entropy H(λ).The EM algorithm is one of the popularly used iterative methods for the censored data; for more details, one may refer to Dempster et al. [30], McLachlan and Krishnan [31], and Casella and Berger [32].Additionally, we consider the asymptotic and bootstrap confidence intervals for λ and H(λ).

Maximum Likelihood Estimation
Based on the observed sample under the PFFC scheme with censoring plan G, x = (x 1 , x 2 , . . ., x m ), and assuming that the lifetimes of the items follow the MWD(λ), the likelihood function in Equation ( 5) can be obtained as The MLE of the parameter λ can be obtained by maximizing the likelihood function in Equation ( 6).Instead of directly maximizing the likelihood function, we take advantage of the closed-form MLE of λ based on a complete sample by considering the EM algorithm to obtain the MLE of λ.
Suppose the observed and censored data are X = x = (x 1 , x 2 , . . . ,x m ), and , respectively, the combined forms of complete data set is given by Ỹ = ( X, Z).If a complete sample Ỹ = (y 1 , y 2 , . . . ,y nk ) from MWD is available, the MLE of λ, denoted as λC can be expressed in a closed form as Based on the complete sample Ỹ = ( X, Z), the log-likelihood function can be expressed as The EM algorithm treats the censored data as missing observations and replaces those missing values with their corresponding expected values.In the expectation step (E-step) of the EM algorithm, we replace the functions of the unobserved (censored) data Z with the corresponding expected values given the observed data x. Hence, the pseudo-complete log-likelihood function can be obtained as For given Then, the conditional expectations involved in Equation ( 9) can be computed as In the maximization step (M-step) of the EM algorithm, suppose the estimate of λ in the r-th stage of the EM algorithm is λ(r) , then the updated estimate in the (r + 1)-th stage can be obtained as With an initial estimate λ (0) of parameter λ, the E-step and M-step iterate until convergence occurs.Here, we consider convergence as |λ (r+1) − λ (r) | < , where is a small value.
Once the MLE λ of parameter λ is computed, using the invariance property of MLE, the MLE of Shannon's entropy H(λ) is obtained as

Asymptotic Confidence Interval
In this subsection, we construct confidence intervals for the parameter λ and Shannon's entropy H(λ) based on the asymptotic theory of the MLE.The standard errors of the MLE can be obtained by direct computation or numerical differentiation by evaluating the negative of the second derivative of the observed log-likelihood function Tanner [33] [ Section 4.4].However, the observed information can be difficult to compute because of the incomplete data.Therefore, the missing information principle computes the observed Fisher information by using the difference between the information in the pseudo-complete sample, and the information in the missing observations can be used to obtain the asymptotic variance of the MLE λ.The missing information principle presented by Louis [34] (see also, [33]) can be described as follows: Observe Information = Complete Information − Missing Information (12) Let I X (λ) be the Fisher information based on the observed data, I Y (λ) be the Fisher information based on a complete sample, and I Y|X (λ) be the missing information, then Equation ( 12) can be expressed as For MWD, the Fisher information based on complete data is given by At the time of the i-th failure at time x i , the Fisher information in a censored observation can be obtained as Consequently, the missing information can be obtained as From Equation ( 12), the Fisher information of the observed data I X (λ) can be obtained by using Equations ( 14) and (15).Then, under suitable regularity conditions (e.g., the true param-eter value of λ must be interior to the parameter space, the derivatives of the log-likelihood function with respect to λ exist up to third order and bounded), the asymptotic variance of the MLE of λ can be obtained as Therefore, an asymptotic 100(1 − α)% confidence interval of λ is λ ± z α/2 Var( λ), where z α/2 is the upper (α/2)-th quantile of the standard normal distribution N(0, 1).

Bootstrap Confidence Intervals
The bootstrap method is a well-known resampling technique used in statistics to estimate the sampling distribution of a statistic by repeatedly resampling with replacement from the observed data.The main objective of this method is to provide an empirical approximation of the sampling distribution of a statistic, which can be useful for making inferences and constructing confidence intervals.In the literature, the concept of the bootstrap technique was first introduced by Efron [36], and since then, several applications have been discussed.For more details, one may refer to Efron and Tibshirani [37].Here, we construct confidence intervals for the parameter λ and Shammon's entropy using two parametric bootstrap methods-the percentile bootstrap (boot-p) method [38] and the bootstrap-t (boot-t) method [39].Following Efron and Tibshirani [37], the procedures to obtain the two parametric bootstrap confidence intervals are described as follows.

Percentile Bootstrap (Boot-P) Confidence Intervals
Step A1: Based on the observed PFFC sample X = (x 1 , x 2 , . . . ,x m ) with effective sample size m from n groups each having k items and progressive censoring plan G, compute the MLEs λ and Ĥ of parameter λ and entropy H(λ), respectively.

Boostrap-t (Boot-t) Confidence Intervals
Step B1: Based on the observed PFFC sample X = (x 1 , x 2 , . . . ,x m ) with effective sample size m from n groups each having k items and progressive censoring plan G, compute the MLEs λ and Ĥ of parameter λ and entropy H(λ), respectively.
Step B4: Repeat Step B2 and Step B3 B times to obtain a set of bootstrap statistics as (τ Step B5: Order the bootstrap statistics (τ Step B6: The 100(1 − α)% bootstrap-t confidence intervals for λ and H, respectively, are given by λ − τ Var( Ĥ) .

Bayesian Estimation Approach
In this section, we derive Bayes estimators under the linear exponential (LINEX) loss function and construct the highest posterior density (HPD) credible intervals for the parameters λ and Shannon's entropy H(λ).For details of Bayesian statistical inference and data analysis methods, one may refer to the books by Box and Tiao [40] and Gelman et al. [41].The Bayesian approach to reliability analysis involves previous knowledge of lifespan parameters, technical knowledge of failure mechanisms, as well as experimental data to be incorporated into the inferential procedure.As pointed out by Tian et al. [42], employing the Bayesian approach in reliability analysis has the advantages of making statistical inferences using information from prior experience with the failure mechanism or physics-of-failure and avoiding making inferences based on plausibly inaccurate large-sample theory in frequentist approach.For more details about Bayesian inference methods and specifying prior distribution in reliability applications, one may refer to Tian et al. [42].As a result, Bayesian techniques are frequently applied to small sample data, which is highly advantageous in the case of pricey life testing tests.For Bayesian estimation, the inverted gamma distribution is commonly used as the natural conjugate prior density for the parameter λ of MWD (see, for example, Bekker and Roux [9], Chaudhary et al. [43]).Following Bekker and Roux [9] and Chaudhary et al. [43], we consider the prior distribution of the unknown parameter λ is the inverted gamma distribution with pdf g(λ) ∝ 1 where a and b are hyper-parameters.Thus, by incorporating the prior information in Equation ( 16) to the likelihood function in Equation ( 6), the posterior distribution of λ can be expressed as where K −1 is the normalizing constant given by Here, we consider the LINEX loss function proposed by [44], which is one of the most commonly used asymmetric loss functions.The LINEX loss is defined as where the loss function's scaling parameter is c and λ is an estimate of λ.The LINEX loss function provides greater weight to overestimation or underestimation, depending on whether the value of c is positive or negative, and for small values of c 0, it is virtually identical to the squared error loss function.This loss function is appropriate when overestimation is more expensive than underestimating.Thus, under the LINEX loss function, the Bayes estimator of any function of the parameter λ, say φ(λ), is given by From Equation ( 19), the Bayes estimators take the ratio of two integrals for which has no closed-form solution.To obtain the ratio of the two integrals in Equation ( 19), we suggest using two approximation techniques-the T-K approximation and the MCMC methods.
The T-K approximation method is one of the oldest deterministic approximation techniques, whereas the MCMC method is one of the newest popularized techniques based on the posterior sample algorithm.The MCMC could be expensive to compute, especially for large sample sizes n.Moreover, many MCMC algorithms require a rough estimate of key posterior quantities, such as the posterior variance.Compared to MCMC methods, the T-K approximation cannot be reduced by running the algorithm longer.However, deterministic approximation is typically very fast to compute and sufficiently reliable in several applied contexts.These issues motivate us to develop both techniques in this study.The details of these two approximation techniques are presented in the following subsections.

Tierney-Kadane (T-K) Approximation Technique
According to T-K approximation's technique proposed by Tierney and Kadane [45], the approximation of the posterior mean of the function of the parameter, say φ(λ) is given by where is the log-likelihood function, ρ(λ) = ln g(λ), and |Σ * φ | and |Σ| are the determinants of inverse of the negative Hessian of δ * (λ) and δ(λ) at λδ * and λδ , respectively.Here, λδ and λδ * maximize δ(λ) and δ * (λ), respectively.We observe that To determine the value of λδ , we solve the following non-linear equation: where Now, |Σ| can be obtained from where To calculate the Bayes estimator of λ under the LINEX loss function, we take φ(λ) = e −cλ , consequently the function δ * (λ) becomes and then λ * δ * , is computed as solutions of the following non-linear equation Thus, the approximate Bayes estimator of λ under the LINEX loss function is given by Similarly, the Bayes estimator of Shannon's entropy H(λ) under the LINEX loss function is given by

Markov Chain Monte Carlo (MCMC) Techniques
In this subsection, we use the MCMC techniques to obtain the Bayes estimates of the parameter λ and Shannon's entropy H(λ) under the LINEX loss function.The Metropolis-Hastings (M-H) algorithm was initially established by Metropolis et al. [46] and subsequently extended by Hastings [47] and popularized as one of the most commonly used MCMC techniques.The candidate points are created from a normal distribution to a sample from the posterior distribution of λ using the observed data X in (17).The following steps are used to obtain MCMC sequences: Step C1.For parameter λ, set λ (0) as the initial guess value Step C2.From proposal density η(λ (j) |λ (j−1) ), generate a candidate point λ (j) c .
Step C7.Repeat Steps C2-C6 M times to obtain the sequence of the parameter λ as (λ 1 , λ 2 , . . ., λ M ) and Shannon's entropy H as (H 1 , H 2 , . . ., H M ).To acquire an independent sample from the stationary distribution of the Markov chain, we consider a burn-in period of size M 0 by discarding the first M 0 values in the MCMC sequences.Thus, the Bayes estimators of λ and H(λ) under the LINEX loss function, respectively, are given by Based on the MCMC samples, we can obtain the HPD credible interval for the parameter λ and Shannon's entropy H(λ).Suppose (λ [1] denotes the ordered values of (λ 1 , λ 2 , . . ., λ M * ) and (H 1 , H 2 , . . ., H M * ), respectively, after the burn-in period, where M * = M − M 0 .Then, following Chen and Shao [48], the 100(1 − α)% HPD credible interval for λ can be obtained as λ (j) , λ (j+[(1−α)M * ]) , where j is chosen such that Similarly, the 100(1 − α)% HPD credible interval for H(λ) is given by H where j is chosen such that

Monte Carlo Simulation Study
In this section, a Monte Carlo simulation study is conducted to evaluate the performance of the proposed estimation procedures.The frequentist and Bayesian point estimation procedures for the parameter λ and Shannon's entropy H(λ) are compared by means of the average estimates (AE) and mean squared errors (MSE).For interval estimation procedures, we compare the asymptotic confidence intervals (Asym), the percentile bootstrap confidence intervals (boot-p), the bootstrap-t confidence intervals (boot-t), and the HPD credible intervals in terms of their simulated average lengths (AL) and the simulated coverage probabilities (CP).For the bootstrap confidence intervals, the intervals are obtained based on B = 1000 bootstrap samples.
In the simulation study, we consider that the PFFC samples are generated from the MWD with parameter λ = 0.75 and 1.5 (the corresponding entropy H(λ) are H(0.75)= 0.5057 and H(1.5) = 0.8523, respectively) with various combinations of a number of groups n, effective sample size m, group size k, and censoring scheme G).We consider group sizes k = 3 and 5, the number of groups n = 20 and 50, and effective sample sizes m = 0.4 n and 0.8 n.Three different censoring schemes G for each combination of n and m are considered: from the experiment at the first failure only;  1.Note that simplified notations are used to denote the censoring schemes, for example, (0 7 ) denotes (0, 0, 0, 0, 0, 0, 0) and (4 3 ) stands for (4,4,4).(2 2 , 0 14 ) [11] (5 2 , 0 38 ) [6] (0 15 , 4 1 ) [12] (0 39 , 10 1 ) For the Bayesian estimation approach, the Bayes estimates of Shannon's entropy are computed with informative inverted gamma prior under the LINEX loss function.The hyper-parameters (a, b) are selected for Bayesian computations of the parameter λ and Shannon's entropy in such a manner that the prior mean is precisely identical to the true values of the parameter, i.e., λ = a/b.Specifically, we consider (a, b) = (3, 4) and (3, 2) for λ = 0.75 and 1.5, respectively.When computing Bayes estimators under the LINEX loss function, we consider the loss function parameter c = −0.5 and 0.5.We use M = 10,000 with a burn-in period M 0 = 2000 for the M-H algorithm.The simulation results are based on 1000 repetitions in this study.All the computations are performed using the statistical software R (https://www.r-project.org/)[49].
The simulated results for point estimation are presented in Tables 2 and 3, and the simulation results for interval estimation are presented in Tables 4 and 5.For point estimation, from Tables 2 and 3, we observe that the MLEs and Bayes estimates of Shannon's entropy are performing well with small MSEs.The simulated MSEs decrease as n or m increases.The Bayes estimates perform better than the MLEs in terms of MSEs when the prior information matches the true value of λ.Among the two approaches to obtain the Bayes estimates, the Bayes estimates obtained by the M-H algorithm techniques outperform the Bayes estimates obtained by the T-K approximation in terms of the MSEs.
To illustrate the methodologies developed in this paper, we generate first-failure censored data based on the data set in Table 6.After grouping the 100 carbon fiber into n = 25 groups and k = 4 individuals within each group.The grouped data and the corresponding first-failure censored samples are reported in Table 7.The items with "+" within each group indicate the first failure.Then, we obtain six different PFFC samples using different censoring schemes for m = 10 and 20 based on the first-failure censored data in Table 7.The censoring schemes and the corresponding PFFC samples are presented in Table 8.To avoid ambiguity with the censoring schemes in Table 1, we named these censored schemes [CS1]-[CS6].
Based on each PFFC sample in Table 8, we compute the MLEs and Bayes estimates of the parameter λ and Shannon's entropy H(λ).As we do not have prior information on the parameter λ, we use a non-informative prior for obtaining the Bayes estimates.The Bayes estimates are computed using the T-K approximation and MCMC methods under the LINEX loss function at two values of loss parameter c = −0.5 and 0.5.We construct 95% asymptotic, percentile bootstrap, bootstrap-t confidence intervals and the Bayesian HPD credible intervals for the parameter λ and Shannon's entropy H(λ).The point and interval estimation results are presented in Tables 9 and 10, respectively.
For the Bayesian estimation procedures, we validate the convergence of the generated MCMC sequence of the parameter λ samples from the posterior distribution by using the M-H algorithm for their stationary distributions using graphical diagnostic tools, such as the trace plot, boxplot, and histogram with Gaussian density plots, as shown in Figure 4 for c = −0.5.The trace plot shows a random scatter around the mean (shown by a thick red line) and a fine mixture of parameter chains.The posterior distribution is almost symmetric, as seen by the boxplots and histograms of produced samples, implying that the posterior mean can be used as a reasonable Bayes estimate of the parameter λ.For illustration, we also present the simulated posterior predictive densities in Figure 5.          8.

Concluding Remarks
In this study, we developed statistical inference for the associated parameter and Shannon's entropy of MWD based on PFFC data using frequentist and Bayesian approaches.For frequentist estimation procedures, we applied the EM algorithm to compute the MLEs.In addition, we obtained asymptotic, percentile bootstrap, and bootstrap-t confidence intervals for the parameter and Shannon's entropy of MWD.For Bayesian estimation procedures, we applied two approximation techniques-the T-K approximation and the M-H algorithm-to obtain the Bayes estimates under the LINEX loss function.Moreover, we use the M-H algorithm to obtain the HPD credible intervals for the parameter and Shannon's entropy of MWD.A Monte Carlo simulation study is used to evaluate the performance of the estimation procedures and a real data analysis is used to illustrate the methodologies.Based on the simulation results, we recommend using the Bayesian point and interval estimation of the parameter and Shannon's entropy based on the MCMC method for the MWD when prior information about the model parameter is available.If the prior information about the model parameter is unavailable, then the MLEs are recommended.
For future research, it will be interesting to study the optimal censoring plan (i.e., to determine the values of n, k, and the censoring scheme G = (G 1 , G 2 , . . ., G m )) for specific effective sample size m and the total sample size available for the experiment (i.e., nk) to maximize the information about the Shaanon's entropy can be obtained.On the other hand, we consider MWD due to its simplicity.Considering the trade-off between model

Figure 2 .
Figure 2. TTT plots under consideration of real data set.

Figure 3 .
Figure 3. Empirical and fitted Maxwell distribution plots for real data.

Figure 4 .
MCMC diagnostic plots for different PFFC samples in Table8 .

Figure 5 .
Figure 5. Simulated posterior predictive densities for the Bayesian estimates based on different PFFC samples in Table8.

Table 1 .
Censoring plans used in the simulation study.

Table 7 .
Grouped real data set (Observation with "+ indicates the first failure (FF) in the group).

Table 9 .
MLEs and Bayes estimates of λ and H(λ) for the PFFC samples in Table8with k

Table 10 .
Ninety-five percent asymptotic (Asym), percentile bootstrap (boot-p), bootstrap-t (boot-t) confidence intervals, and the Bayesian HPD credible intervals for the parameter λ and Shannon's entropy H(λ) for the PFFC samples in Table8with k