Next Article in Journal
Transfer Learning for Day-Ahead Load Forecasting: A Case Study on European National Electricity Demand Time Series
Previous Article in Journal
Two Approaches to Estimate the Shapley Value for Convex Partially Defined Games
Previous Article in Special Issue
Quantile-Composited Feature Screening for Ultrahigh-Dimensional Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Estimation of a New Pareto-Type Distribution Based on Mixed Gibbs Sampling Algorithm

Institute of Statistics and Applied Mathematics, Anhui University of Finance & Economics, Bengbu 233000, China
*
Authors to whom correspondence should be addressed.
Mathematics 2024, 12(1), 18; https://doi.org/10.3390/math12010018 (registering DOI)
Submission received: 23 November 2023 / Revised: 15 December 2023 / Accepted: 17 December 2023 / Published: 21 December 2023
(This article belongs to the Special Issue Methods and Applications in Multivariate Statistics)

Abstract

:
In this paper, based on the mixed Gibbs sampling algorithm, a Bayesian estimation procedure is proposed for a new Pareto-type distribution in the case of complete and type II censored samples. Simulation studies show that the proposed method is consistently superior to the maximize likelihood estimation in the context of small samples. Also, an analysis of some real data is provided to test the Bayesian estimation.

1. Introduction

As an important statistical inference method, the Bayesian method has been widely applied and developed in the fields of statistics and artificial intelligence. Especially with the rise of big data and machine learning in the early 21st century, the Bayesian method has received more and more attention. Researchers have improved the efficiency and accuracy of the Bayesian method by developing new algorithms and tools and applying them to various fields, such as computer vision, natural language processing, bioinformatics, and so on. Recently, many scholars [1,2,3] have studied it and obtained some valuable results.
The Pareto distribution [4] is a failure function with a decreasing value and has been widely applied in reliability, personal income, stock price fluctuation, and some other models. The origin and some applications of the Pareto distribution can be seen in Johnson et al. [5]. Bourguignon et al. [6] proposed a new Pareto-type (NP) distribution, which is a generalization of the Pareto distribution and often applied for income and reliability data analysis. Raof et al. [7] used the NP distribution model to simulate the income of upper class groups in Malaysia. Karakaya et al. [8] proposed a new generalization of the Pareto distribution, incorporating a truncation parameter. Sarabia et al. [9] acquired further significant properties of this distribution and derived simpler expressions for relevant inequality indices and risk measures. Nik et al. [10] estimated the parameters of the NP distribution from two different perspectives, namely, classical and Bayesian statistics, in which the Bayesian estimation used importance sampling method. Nik et al. [11] studied the parameter estimation as well as predicting the failure times of the removed units in multiple stages of the progressively censored sample coming from the NP distribution, in which a Bayesian method (i.e., Markov chain Monte Carlo, MCMC) was applied to estimate the unknown parameters involved in the model. Soliman [12] studied the estimation of the reliability function in a generalized life model. The Bayesian estimations of the symmetric loss function (i.e., secondary loss) and asymmetric loss function (i.e., LINEX loss and GE loss) were obtained. These estimations were compared with the maximum likelihood estimation (MLE) with the Burr-XII model using the Bayesian approximation based on Lindley. Furthermore, Soliman [12] studied the Bayesian estimation of the parameter of interest on the premise that some other parameters are known. However, all of the parameters are often unknown in practice. Soliman [13] gave an approximate calculation formula for Bayesian estimation; however, the derivation of the formula is complicated and difficult to understand. Soliman’s work shows that the Bayesian method may encounter high dimensions, complex models, and big data in practical use. Thus, scholars can greatly improve the efficiency and accuracy of the Bayesian method by improving the MCMC algorithm.
The Gibbs sampling algorithm was originally proposed by Geman [14]. In recent years, with the development of technologies such as big data analysis, artificial intelligence, and machine learning, the applications of the Gibbs sampling algorithm to solve Bayesian estimation problems has become increasingly common. Ahmed [15] proposed two prediction methods for predicting the censored units with progressive type II censored samples. The lifetimes under consideration are set to follow a new two-parameter Pareto distribution. Moreover, point and interval estimation of the unknown parameters of the NP distribution can be obtained using the maximum likelihood and Bayesian estimation methods. Since Bayesian estimators cannot be expressed explicitly, the Gibbs sampling and MCMC techniques are utilized for Bayesian calculation. Ibrahim et al. [16] proposed a new discrete analogue of the Weibull class, and gave a Bayesian estimation procedure under the square error loss function. Furthermore, they compared the non-Bayesian estimation with the Bayesian estimation using Gibbs sampling and the Metropolis–Hastings algorithm.
For the NP distribution, Nik et al. [10] proposed some classical estimation methods and a Bayesian estimation based on the importance sampling method. However, the sampling efficiency of the importance sampling method is often low in high-dimensional situations. Thus, some other efficient methods are preferred. In our work, we use the mixed Gibbs sampling algorithm to estimate the parameters of the NP distribution with complete samples and type II censored samples.
The Gibbs sampling algorithm can solve some complex Bayes analysis problems directly, and has become a useful tool in Bayesian statistics. By establishing a Markov chain and iterating to the equilibrium state repeatedly, the Gibbs sampling algorithm can obtain a posterior distribution sample. However, the Gibbs sampling algorithm also has some shortcomings. For instance, the fully conditional posterior distribution of a parameter may be difficult to be directly sampled for some special prior distribution. Thus, the Gibbs sampling algorithm can be combined with other sampling methods to achieve the stability of the efficiency. Nowadays, there are some sampling methods, such as the Metropolis algorithm, importance sampling, accept–reject sampling, and so on. Among them, the Metropolis algorithm is a relatively easy and flexible sampling method. The Metropolis algorithm was improved by Hastings [17], and the result is called the Metropolis–Hastings algorithm. However, the Metropolis–Hastings algorithm is time consuming and inefficient in high-dimensional cases. Sampling may even be rejected after time-consuming calculations. A mixture of the Gibbs sampling algorithm and Metropolis–Hastings algorithm, i.e., the mixed Gibbs sampling algorithm, can overcome these shortcomings of the Metropolis–Hastings algorithm. Furthermore, the mixed Gibbs sampling algorithm can make the Gibbs sampling algorithm and Metropolis–Hastings algorithm perform fully to their own advantages. Thus, in our work, we use the mixed Gibbs sampling algorithm to propose a Bayesian estimation procedure for the NP distribution with complete and type II censored samples.
The rest of this paper is organized as follows. We state the MLE of the NP distribution in Section 2. In Section 3, we propose a Bayesian estimation procedure for the NP distribution with complete and type II censored samples using the mixed Gibbs sampling algorithm. Simulation studies and real data analysis are presented in Section 4. In Section 5, we present a brief discussion and conclusion of the results and methods.

2. Maximum Likelihood Estimation of NP Distribution

Suppose X is a random variable that follows the NP distribution with a shape parameter α and a scale parameter β , that is, X NP ( α , β ) . The cumulative distribution function (CDF) F ( x | α , β ) of X is given as
F ( x | α , β ) = 1 ( β / x ) α 1 + ( β / x ) α = x α β α x α + β α = 1 2 ( β / x ) α 1 + ( β / x ) α = 1 2 β α x α + β α , x β .
The probability density function (PDF) f ( x | α , β ) of X is
f ( x | α , β ) = 2 α ( β / x ) α x [ 1 + ( β / x ) α ] 2 = 2 α ( β / x ) α + 1 β [ 1 + ( β / x ) α ] 2 = 2 α β α x α 1 ( x α + β α ) 2 , x β .
The PDF and CDF diagrams are shown in Figure 1.
It should be highlighted that compared with the Pareto distribution, the NP distribution has the following advantages: (1) It can exhibit either an upside-down bathtub or a decreasing hazard rate function, depending on the values of its parameters. (2) It offers mathematical simplicity, as the probability and distribution functions of the new distribution have a simple form, in contrast to some generalizations of the Pareto distribution which involve special functions such as log, beta, and gamma functions. (3) The proposed distribution has only two parameters, unlike some generalizations of the Pareto distribution which have three or four parameters. Thus, in our work, it is assumed that the life of the product X follows the NP distribution. We present an MLE of the NP distribution below for comparison with the Bayesian estimation.

2.1. Complete Samples Case

In the case of complete samples, assuming that there are n products for a life test, and the order failure data x 1 x 2 x n can be obtained, the log-likelihood function ( α , β ) can be written as
( α , β ) = n log ( 2 α ) n log β + ( α + 1 ) i = 1 n log β x i 2 i = 1 n log 1 + β x i α .
It can be seen that ( α , β ) increases monotonically with β , that is to say, the greater the value of β , the greater the value of ( α , β ) . Since x β , we obtain the MLE β ^ = x 1 . Thus, the MLE α ^ of α can be obtained from the solution of Equation (3):
( α , x 1 ) α = n α + i = 2 n log x 1 x i 2 i = 2 n x 1 x i α log x 1 x i 1 + x 1 x i α = 0 .

2.2. Type II Censored Samples Case

In the case of type II censored samples, assume that there are n products for a life test, then the test stops when there are r product failures, and the order failure data x 1 x 2 x r can be obtained. MLE is used to estimate the unknown parameters α , β below. In order to obtain the likelihood function, it is necessary to know the probability of occurrence as follows.
(1) The probability of a product failing at ( x i , x i + d x i ] is approximately
f ( x i ) d x i = 2 α β α x i α 1 ( x i α + β α ) 2 d x i , i = 1 , 2 , , r .
(2) The probability that the lifetime of the remaining n r products exceed x r is
x r 2 α β α x i α 1 ( x i α + β α ) 2 d x n r = 1 2 β α x i α + β α | x r n r = 2 β α x r α + β α n r .
Thus, the probability of the above observation { x i , i = 1 , 2 , , r } is approximately
n r 2 α β α x 1 α 1 ( x 1 α + β α ) 2 d x 1 2 α β α x 2 α 1 ( x 2 α + β α ) 2 d x 2 2 α β α x r α 1 ( x r α + β α ) 2 d x r 2 β α x r α + β α n r = α r ( 2 β α ) n n r i = 1 r x i α 1 i = 1 r ( x i α + β α ) 2 1 x r α + β α n r d x 1 d x 2 d x r ,
where d x 1 d x 2 d x r is a constant. Ignoring a constant factor does not affect the MLE of α and β , and the likelihood function L ( α , β ) can be given as
L ( α , β ) = α r ( 2 β α ) n i = 1 r x i α 1 i = 1 r ( x i α + β α ) 2 1 x r α + β α n r .
Thus, the log-likelihood function ( α , β ) is as
( α , β ) = r log α + n log 2 + n α log β + ( α 1 ) i = 1 r log x i 2 i = 1 r log ( x i α + β α ) ( n r ) log ( x r α + β α ) .
Similarly, we can obtain the MLE β ^ of β as
β ^ = x 1 .
Therefore, the MLE α ^ of α can be obtained from the solution of Equation (5)
( α , x 1 ) α = r α + n log x 1 + i = 1 r log x i 2 i = 1 r x i α log x i + x 1 α log x 1 x i α + x 1 α ( n r ) x r α log x r + x 1 α log x 1 x r α + x 1 α = 0 .

3. Bayesian Estimation of NP Distribution

The Bayesian estimations of the NP distribution under complete samples and type II censored samples are given below.

3.1. Complete Samples Case

In the case of complete samples, assume that there are n products for a life test, and the order failure data x 1 x 2 x n can be obtained. When the parameters α and β are unknown, we assume that the joint prior distribution π ( α , β ) of α and β is
π ( α , β ) 1 / ( α β ) .
Then, the posterior distribution π ( α , β | x 1 , x 2 , , x n ) is given as (7):
π ( α , β | x 1 , x 2 , , x n ) ( 2 α ) n β n α i = 1 n x i α 1 i = 1 n ( x i α + β α ) 2 · 1 α β = 2 n α n 1 β n α 1 i = 1 n x i α 1 i = 1 n ( x i α + β α ) 2 .
Thus, the fully conditional posterior distribution π ( α | x 1 , x 2 , , x n , β ) of α is
π ( α | x 1 , x 2 , , x n , β ) α n 1 β n α 1 i = 1 n x i α 1 i = 1 n ( x i α + β α ) 2 ,
and the fully conditional posterior distribution π ( β | x 1 , x 2 , , x n , α ) of β is
π ( β | x 1 , x 2 , , x n , α ) β n α 1 i = 1 n ( x i α + β α ) 2 .
According to (8) and (9), both π ( α | x 1 , x 2 , , x r , β ) and π ( β | x 1 , x 2 , , x r , α ) do not have explicit expressions in the case of complete samples. Therefore, it is difficult to conduct sampling directly in practice. However, the mixed Gibbs sampling algorithm does not require an explicit expression of the fully conditional posterior distribution, and can be used to sampling different prior distributions. Thus, in our work, we consider the case of the prior distribution denoted as (6) using the mixed Gibbs sampling algorithm in Section 3.1. For Bayesian estimations for parameters of the NP distribution in the case of complete samples, we state the iterative steps of the mixed Gibbs sampling algorithm as follows.
Step 0 
Choose an initial value ( α ( 0 ) , β ( 0 ) ) of ( α , β ) . Denote the value of the ith iteration as ( α ( i ) , β ( i ) ) . Then, the value of the ( i + 1 ) th iteration ( α ( i + 1 ) , β ( i + 1 ) ) can be obtained using the following Step 1 and Step 2.
Step 1 
Obtain α ( i + 1 ) from π ( α | x 1 , x 2 , , x n , β ( i ) ) using the Metropolis–Hastings algorithm in this step, which consists of the following four steps (1)∼(4).
(1)
Set α N ( α k , σ α 2 ) , where α k is the current state and σ α is the standard deviation. Obtain a sample α of α ; if α 0 , then a new sampling is required.
(2)
Calculate the acceptance probability P accept ( α k , α ) as (10) below:
P accept ( α k , α ) = min 1 , π ( α | x 1 , x 2 , , x n , β ( i ) ) π ( α k | x 1 , x 2 , , x n , β ( i ) ) = min 1 , α n 1 β n α 1 i = 1 n x i α 1 α k n 1 β n α k 1 i = 1 n x i α k 1 · i = 1 n ( x i α k + β α k ) 2 i = 1 n ( x i α + β α ) 2 = min 1 , α α k n 1 · β n · Δ α k · i = 1 n x i Δ α k · i = 1 n ( x i α k + β α k ) 2 i = 1 n ( x i α + β α ) 2 ,
where Δ α k = α α k .
(3)
Generate a random number u from the uniform distribution U ( 0 , 1 ) , and then obtain α k + 1 according to the following conditions (11):
α k + 1 = α , u P accept ( α k , α ) α k , u > P accept ( α k , α ) .
(4)
Set k = k + 1 , then return to Step 1 (1).
The resulting { α ( i ) , i = 1 , 2 , , k } is a Markov chain. Step 1 ends when the Markov chain reaches equilibrium, then we can obtain α ( i + 1 ) .
Step 2 
Obtain β ( i + 1 ) from π ( β | x 1 , x 2 , , x n , α ( i + 1 ) ) using the Metropolis–Hastings algorithm in this step.
(1)
Set β N ( β k , σ β 2 ) , where β k is the current state and σ β is the standard deviation. Obtain a sample β of β ; if β 0 or β > x 1 , then a new sampling is required.
(2)
Calculate the acceptance probability P accept ( β k , β ) as (12):
P accept ( β k , β ) = min 1 , π ( β | x 1 , x 2 , , x n , α ( i + 1 ) ) π ( β k | x 1 , x 2 , , x n , α i + 1 ) = min 1 , β β k n α ( i + 1 ) 1 · i = 1 n ( x i α ( i + 1 ) + β k α ( i + 1 ) ) 2 i = 1 n ( x i α ( i + 1 ) + β α ( i + 1 ) ) 2 .
(3)
Generate a random number u from the uniform distribution of U ( 0 , 1 ) , and then obtain β k + 1 according to the following conditions (13):
β k + 1 = β , u P accept ( β k , β ) β k , u > P accept ( β k , β ) .
(4)
Set k = k + 1 , then return to Step 2 (1).
Similarly, the resulting { β ( i ) , i = 1 , 2 , , k } is a Markov chain. We can obtain β ( i + 1 ) when the Markov chain reaches equilibrium.
Step 3 
Let i = i + 1 , and repeat Step 1 and Step 2 in sequence until the Markov chains reach equilibrium.
It is worth noting that the sample trajectory of the Markov chain can be used as samples of the posterior distributions of α , β when the Markov chain reaches equilibrium, further leading to the Bayesian estimation of α , β . In order to ensure the approximate independence of the posterior distribution samples, subsequent determinations will be made based on the sample autocorrelation coefficients.

3.2. Type II Censored Samples Case

In the case of type II censored samples, assuming that there are n products for a life test, the test stops when there are r products failure, and the order failure data x 1 x 2 x r can be obtained. We still take the joint prior distribution of α , β as π ( α , β ) 1 / ( α β ) , then the posterior distribution π ( α , β | x 1 , x 2 , , x r ) is the following (14):
π ( α , β | x 1 , x 2 , , x r ) i = 1 r f ( x i | α , β ) · i = r + 1 n ( 1 F ( x r | α , β ) ) · 1 α β = ( 2 α ) r β r α i = 1 r x i α 1 i = 1 r ( x i α + β α ) 2 · ( 2 β α ) n r ( x r α + β α ) n r · 1 α β = 2 n α r 1 β n α 1 i = 1 r x i α 1 i = 1 r ( x i α + β α ) 2 ( x r α + β α ) n r .
Thus, the fully conditional posterior distribution π ( α | x 1 , x 2 , , x r , β ) of α , denoted as (15), is
π ( α | x 1 , x 2 , , x r , β ) α r 1 β n α 1 i = 1 r x i α 1 i = 1 r ( x i α + β α ) 2 ( x r α + β α ) n r ,
and the fully conditional posterior distribution π ( β | x 1 , x 2 , , x r , α ) of β , denoted as (16), is
π ( β | x 1 , x 2 , , x r , α ) β n α 1 i = 1 r ( x i α + β α ) 2 ( x r α + β α ) n r .
As can be seen from the above, in the case of type II censored samples, neither (15) nor (16) has an explicit expression, and so they are difficult to sample directly. Similarly, as previously discussed, the mixed Gibbs sampling algorithm does not require a display expression for the complete conditional posterior distribution, so we continue to use the mixed Gibbs algorithm. Similarly, we still only consider the same prior distribution as in Section 3.1. The steps of the mixed Gibbs sampling algorithm for Bayesian estimation of the NP distribution parameters in the case of type II censored samples are stated as follows.
Step 0 
Give an initial value ( α ( 0 ) , β ( 0 ) ) using the Gibbs sampling. For simplicity, the value of the ith iteration is still denoted as ( α ( i ) , β ( i ) ) . Then, we can use the following Step 1 and Step 2 to obtain ( α ( i + 1 ) , β ( i + 1 ) ) .
Step 1 
Obtain a sample α ( i + 1 ) from π ( α | x 1 , x 2 , , x r , β ( i ) ) using the Metropolis–Hastings algorithm in this step.
(1)
Set α N ( α k , σ α 2 ) , where α k is the current state and σ α is the standard deviation. Obtain a sample α of α ; if α 0 , then a new sampling is required.
(2)
Calculate the acceptance probability P accept ( α k , α ) , denoted as (17):
P accept ( α k , α ) = min 1 , π ( α | x 1 , x 2 , , x r , β ( i ) ) π ( α k | x 1 , x 2 , , x r , β ( i ) ) = min 1 , α r 1 β n α 1 i = 1 r x i α 1 α k r 1 β n α k 1 i = 1 r x i α k 1 · i = 1 r ( x i α k + β α k ) 2 · ( x r α k + β α k ) n r i = 1 r ( x i α + β α ) 2 · ( x r α + β α ) n r = min 1 , α α k r 1 β n · Δ α k i = 1 r x i Δ α k i = 1 r ( x i α k + β α k ) 2 i = 1 r ( x i α + β α ) 2 x r α k + β α k x r α + β α n r ,
where Δ α k = α α k .
(3)
Generate a random number u from the uniform distribution of U ( 0 , 1 ) , and then obtain α k + 1 according to the following (18):
α k + 1 = α , u P accept ( α k , α ) α k , u > P accept ( α k , α ) .
(4)
Set k = k + 1 , then return to Step 1(1).
The resulting { α ( i ) , i = 1 , 2 , , k } is a Markov chain, and we can obtain α ( i + 1 ) when the Markov chain reaches equilibrium.
Step 2 
Obtain a sample β ( i + 1 ) from π ( β | x 1 , x 2 , , x r , α ( i + 1 ) ) using the Metropolis–Hastings algorithm in this step.
(1)
Set β N ( β k , σ β 2 ) , where β k is the current state and σ β is the standard deviation. Obtain a sample β of β ; if β 0 or β > x 1 , then a new sampling is required.
(2)
Calculate the acceptance probability P accept ( β k , β ) as (19):
P accept ( β k , β ) = min 1 , π ( β | x 1 , x 2 , , x r , α ( i + 1 ) ) π ( β k | x 1 , x 2 , , x r , α ( i + 1 ) ) = min 1 , β β k n α ( i + 1 ) 1 · i = 1 r ( x i α ( i + 1 ) + β k α ( i + 1 ) ) 2 i = 1 r ( x i α ( i + 1 ) + β α ( i + 1 ) ) 2 .
(3)
Generate a random number u from the uniform distribution of U ( 0 , 1 ) , and then obtain β K + 1 according to the following conditions (20):
β k + 1 = β , u P accept ( β k , β ) β k , u > P accept ( β k , β ) .
(4)
Set k = k + 1 , then return to Step 2(1).
Similarly, the resulting { β ( i ) , i = 1 , 2 , , k } is a Markov chain, and we can obtain β ( i + 1 ) when the Markov chain reaches equilibrium.
Step 3 
Let i = i + 1 , and repeat Step 1 and Step 2 in sequence until the Markov chains reach equilibrium.
The specific approach to obtaining the Bayesian estimation of the target parameters is similar to that in the case of complete samples and will not be further elaborated here.

4. Numerical Studies

In Section 4, we apply the R4.1.3 software in our work.

4.1. Simulation Studies

Case 1. Denote X as the life of a product, and set X NP ( α , β ) with three groups of α and β . A Monte Carlo simulation was carried out in the complete samples case and 20% type II censored samples case. Generate 100 NP random variables using the inverse sampling method in each simulation.
To illustrate the complete implementation process of the mixed Gibbs sampling algorithm, we first present the Bayesian estimation and 95% credible interval of three groups of parameter simulations in both the complete samples case and the 20% type II censored samples case (refer to Table 1 and Table 2). The parameters are estimated using Bayesian estimation based on the sample separately under the square loss function and the LINEX loss function (with a loss parameter of 1). Furthermore, when the parameters ( α , β ) = ( 4 , 1 ) , we demonstrate the process of mixed Gibbs sampling in both the complete samples case and the 20% type II censored samples case. This includes the trace plot, histogram, and autocorrelation coefficient diagram of corresponding parameters (refer to Figure 2 and Figure 3).
As stated in Table 1 and Table 2, the Bayesian estimation of parameters using the mixed Gibbs sampling algorithm has produced relatively satisfactory results for the three groups of selected parameters. The results show that the estimated value is close to the true value, which means that the proposed method has high accuracy and reliability concerning parameter estimation. In addition, the credible interval of the parameters also covers the true value, and the interval length is relatively short, indicating high precision in the estimation of the parameters.
From Figure 2 and Figure 3, the trace plots show that the values of α and β are randomly scattered around the average. Furthermore, the traversal mean of the parameters ( α , β ) tends to stabilize after 200 iterations. It can be considered that the Gibbs sampling process reaches the equilibrium state at this point, and the remaining samples can be regarded as the observed value of the target parameter. As a precaution, the first 500 iterations are discarded, and then 4500 iterations are performed to obtain 4500 samples of each parameter. Additionally, the autocorrelation function indicates that the autocorrelation coefficient of all samples approaches 0 at a lag of 1 order. Therefore, the Bayesian estimation of the parameters is ultimately obtained through the 4500 samples that have been obtained.
Case 2. It is assumed that the life X of a product follows the NP distribution of two parameters ( α , β ) = ( 4 , 1 ) . To investigate the strengths and weaknesses of the two estimation methods with small sample sizes, 20 random variables from the NP distribution were generated in each simulation in the complete samples case. In the type II censored samples case, 100 random variables of the NP distribution were generated for each simulation. Furthermore, 20% of the order failure data for the complete random variables are selected in each simulation. The mean value and mean square error (MSE) of MLE and Bayesian estimation (BE) are given in Table 3. The simulation shows that there is no significant difference between maximum likelihood estimation and Bayesian estimation in the large sample case, so we do not report the results in our article.
As shown in Table 3, the MSEs for Bayesian estimation are consistently smaller than the MLEs. Consequently, we can infer that the BE method is evidently superior to the MLE in the context of small samples. In summary, the findings highlight the superior accuracy and robustness of Bayesian estimation utilizing the mixed Gibbs sampling algorithm compared to MLE.

4.2. Real Data Analysis

Take two reliability datasets from Bourguignon [6]. The first dataset (DataSet I), given by Linhart and Zucchini [18], represents the number of failures of an aircraft’s air conditioning system: 23, 261, 87, 7, 120, 14, 62, 47, 225, 71, 246, 21,42, 20, 5, 12, 120, 11, 3, 14, 71, 11, 14, 11, 11, 90, 1, 16, 52, 95. The second dataset (DataSet II) comes from Murthy et al. [19], which is composed of the number of failures of 20 mechanical components with the following data: 0.067, 0.068, 0.076, 0.081, 0.084, 0.085, 0.085, 0.086, 0.089, 0.098, 0.098, 0.114, 0.115, 0.121, 0.125, 0.131, 0.149, 0.160, 0.485, 0.485. It can be observed that both datasets are positive. Due to the origin of the NP distribution, positive data are ideally modeled from this distribution. Therefore, it is reasonable to use the NP distribution to fit these two datasets. In this section, the Bayesian estimation of two datasets by the NP distribution is presented using the mixed Gibbs sampling algorithm.
DataSet I. For DataSet I, Bourguignon [6] provided the MLEs (i.e., α ^ M L E = 0.433 , β ^ M L E = 1 ) of parameters α and β . Subsequently, Bayesian analysis was conducted for both type II censored data ( r = 24 ) and the complete data ( r = n = 30 ). In our work, 2000 sampling iterations were performed for each case. Testing the sampling mean revealed that the traversal mean of the parameters stabilized within 100 iterations, and the mixed Gibbs sampling algorithm converged in both cases. To mitigate the influence of the transition process, the sampled data from 201 to 2000 iterations were utilized as the mixed Gibbs sampling values for Bayesian estimation (BE), and the posterior estimation of the distribution parameters was performed using these sampling values (refer to Table 4).
In Table 4, the posterior expectation and median represent the Bayesian point estimates of the parameters under square loss and absolute loss. The 95% credible intervals of the parameters are determined by the 2.5th and 97.5th percentiles. The results in Table 4 demonstrate the high accuracy of Bayesian estimation using the mixed Gibbs sampling algorithm for both complete data and type II censored data. Compared to frequency estimation methods, Bayesian estimation proves to be more effective in handling small sample sizes and provides probability distribution estimates of parameters, rather than just point estimates, for reliability data with limited observations. Consequently, Bayesian estimation offers a more precise reflection of parameter uncertainties, particularly when dealing with small datasets.
DataSet II. In the case of complete data, we conducted Bayesian analysis for DataSet II with 2000 sampling iterations. After testing the sampling mean, we found that the traversal mean of the parameters tended to be stable within 100 iterations, and the mixed Gibbs sampling algorithm converged. To eliminate the influence of the transition process, the sampled data of 201 to 2000 times were used as the mixed Gibbs sampling values for Bayesian estimation, and the posterior estimation of the distribution parameters was performed using these sampling values. Then, the Bayesian estimation (BE) results under the square loss were compared with the MLE results of Bourguignon [6] (refer to Table 5).
It can be seen from the above Table 5 that the Bayesian estimation is very close to the MLE result, which further verifies the stability and reliability of the mixed Gibbs sampling algorithm.

5. Discussion and Conclusions

Some scholars have conducted valuable research for the NP distribution in the literature, such as [10,11,15]. However, Bayesian methods may encounter challenges with high-dimensional, complex models, and big data in practice. Contrary to the low efficiency of importance sampling [10] and the complexity of the Lindley approximation [12] methods, a Bayesian estimation procedure is proposed for the NP distribution using the mixed Gibbs sampling algorithm. When the fully conditional distributions in Gibbs sampling steps lack an explicit form, direct sampling from these distributions becomes challenging. In such instances, alternative sampling methods must be proposed. As each step within a loop of Gibbs sampling is itself a Metropolis–Hastings iteration, the algorithm can be used to combine Gibbs sampling with the Metropolis–Hastings algorithm if necessary.
In this paper, we proposed a Bayesian estimation procedure for the NP distribution with complete and type II censored samples using the mixed Gibbs sampling algorithm. For instance, the mean and MSEs of the resulting parameters estimations are given, and the stability and effectiveness of the mixed Gibbs sampling algorithm are demonstrated with simulation studies and real data analysis. It can be seen that the mixed Gibbs sampling algorithm is stable, feasible, and high precision.
In reliability-related product tests, it is often challenging to obtain a sufficient number of samples to assess the quality of the products. Bayesian estimation, however, can assist in obtaining reliable results even in cases of limited sample sizes. By incorporating prior information, Bayesian estimation can improve the accuracy of parameter estimation and provide valuable uncertainty information about the parameters, which is particularly crucial when sample data is scarce. Therefore, in the context of product reliability tests, Bayesian estimation can serve as an effective tool for reliably evaluating and predicting product quality in situations with small sample sizes. Thus, in view of the merits of the Bayesian method with the mixed Gibbs sampling algorithm in this paper, it will become an important Bayesian inference method.

Author Contributions

Methodology, F.L. and S.W.; software, F.L. and S.W.; formal analysis, F.L., S.W. and M.Z.; writing—review and editing, F.L., S.W. and M.Z.; funding acquisition, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Philosophy and Social Science planning project of Anhui Province Foundation (Grant No. AHSKF2022D08).

Data Availability Statement

All relevant data are within the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qian, F.; Zheng, W. An evolutionary nested sampling algorithm for Bayesian model updating and model selection using modal measurement. Eng. Struct. 2017, 140, 298–307. [Google Scholar] [CrossRef]
  2. Kuipers, J.; Suter, P.; Moffa, G. Efficient Sampling and Structure Learning of Bayesian Networks. J. Comput. Graph. Stat. 2022, 31, 639–650. [Google Scholar] [CrossRef]
  3. Alhamzawi, R.; Ali, H.T.M. A New Gibbs Sampler for Bayesian Lasso. Commun. Stat.-Simul. C 2020, 49, 1855–1871. [Google Scholar] [CrossRef]
  4. Pareto, V. Cours d’Économie Politique; F.Rouge: Lausanne, Switzerland, 1897. [Google Scholar]
  5. Jonhson, N.L.; Kotz, S.; Balakrishnan, N. Continuous Univariate Distributions; John Willey and Sons: London, UK, 1994; Volume 1, Chapter 18. [Google Scholar]
  6. Bourguignon, M.; Saulo, H.; Fernandez, R.N. A new Pareto-type distribution with applications in reliability and income data. Physica A 2016, 457, 166–175. [Google Scholar] [CrossRef]
  7. Raof, A.S.A.; Haron, M.A.; Safari, A.; Siri, Z. Modeling the Incomes of the Upper-Class Group in Malaysia using New Pareto-Type Distribution. Sains Malays. 2022, 51, 3437–3448. [Google Scholar] [CrossRef]
  8. Karakaya, K.; Akdoğan, Y.; Nik, A.S.; Kuş, C.; Asgharzadeh, A. A Generalization of New Pareto-Type Distribution. Ann. Data Sci. 2022, 2022, 1–15. [Google Scholar] [CrossRef]
  9. Sarabia, J.M.; Jordá, V.; Prieto, F. On a new Pareto-type distribution with applications in the study of income inequality and risk analysis. Physica A 2019, 527, 121277. [Google Scholar] [CrossRef]
  10. Nik, A.S.; Asgharzadeh, A.; Nadarajah, S. Comparisons of methods of estimation for a new Pareto-type distribution. Statistic 2019, 79, 291–319. [Google Scholar]
  11. Nik, A.S.; Asgharzadeh, A.; Raqab, M.Z. Estimation and prediction for a new Pareto-type distribution under progressive type-II censoring. Math. Comput. Simulat. 2021, 190, 508–530. [Google Scholar]
  12. Soliman, A.A. Reliability estimation in a generalized life-model with application to the Burr-XII. IEEE Trans. Reliab. 2002, 51, 337–343. [Google Scholar] [CrossRef]
  13. Soliman, A.A. Estimation of parameters of life from progressively censored data using Burr-XII model. IEEE Trans. Reliab. 2005, 54, 34–42. [Google Scholar] [CrossRef]
  14. Geman, S.; Geman, D. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. 1984, 6, 721–741. [Google Scholar] [CrossRef] [PubMed]
  15. Ahmed, H.H. Best Prediction Method for Progressive Type-II Censored Samples under New Pareto Model with Applications. J. Math. 2021, 2021, 1355990. [Google Scholar]
  16. Ibrahim, M.; Ali, M.M.; Yousof, H.M. The Discrete Analogue of the Weibull G Family: Properties, Different Applications, Bayesian and Non-Bayesian Estimation Methods. Ann. Data Sci. 2023, 10, 1069–1106. [Google Scholar] [CrossRef]
  17. Hastings, W.K. Monte Carlo Sampling Methods Using Markov Chains and Their Applications. Biometrika 1970, 57, 97–109. [Google Scholar] [CrossRef]
  18. Linhart, H.; Zucchini, W. Model Selection; John Wiley and Sons: New York, NY, USA, 1986. [Google Scholar]
  19. Murthy, D.N.P.; Xie, M.; Jiang, R. Weibull Models; John Wiley and Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
Figure 1. PDF and CDF for X NP ( α , β ) with some different α and β = 1 .
Figure 1. PDF and CDF for X NP ( α , β ) with some different α and β = 1 .
Mathematics 12 00018 g001
Figure 2. Trace plots, histogram, and ACF of NP distribution parameters with complete samples case.
Figure 2. Trace plots, histogram, and ACF of NP distribution parameters with complete samples case.
Mathematics 12 00018 g002
Figure 3. Trace plots, histogram, and ACF of NP distribution parameters with type II censored samples case.
Figure 3. Trace plots, histogram, and ACF of NP distribution parameters with type II censored samples case.
Mathematics 12 00018 g003
Table 1. Bayesian estimation and 95% credible interval for parameters in complete samples case under different loss functions.
Table 1. Bayesian estimation and 95% credible interval for parameters in complete samples case under different loss functions.
Loss Function ( α , β ) α ^ β ^ 95% Credible Interval for α 95% Credible Interval for β
LINEX loss(2, 1)1.82710.9938[1.5392, 2.1614][0.9631, 1.0052]
(3, 1.5)2.96471.4964[2.5138, 3.4898][1.4691, 1.5067]
(4, 1)3.92840.9955[3.3583, 4.6654][0.9819, 1.0006]
Square loss(2, 1)2.09860.9977[1.7768, 2.4561][0.9718, 1.0067]
(3, 1.5)2.87591.4952[2.4206, 3.3609][1.4685, 1.5056]
(4, 1)3.98400.9977[3.3579, 4.6719][0.9845, 1.0027]
Table 2. Bayesian estimation and 95% credible interval for parameters in 20% type II censored samples case under different loss functions.
Table 2. Bayesian estimation and 95% credible interval for parameters in 20% type II censored samples case under different loss functions.
Loss Function ( α , β ) α ^ β ^ 95% Credible Interval for α 95% Credible Interval for β
LINEX loss(2, 1)1.98960.9908[1.2957, 3.0046][0.9630, 1.0008]
(3, 1.5)2.93611.4974[1.9785, 4.5394][1.4681, 1.5077]
(4, 1)4.11260.9988[2.8326, 6.5555][0.9855, 1.0037]
Square loss(2, 1)1.95760.9946[1.2310, 2.8416][0.9637, 1.0051]
(3, 1.5)3.13651.4977[1.9785, 4.5394][1.4681, 1.5077]
(4, 1)4.31410.9955[3.5397, 6.2427][0.9816, 1.0002]
Table 3. Mean and mean square error of Bayesian estimation and maximum likelihood estimation.
Table 3. Mean and mean square error of Bayesian estimation and maximum likelihood estimation.
Sample CasenEstimation Method(Mean of α , MSE α )(Mean of β , MSE β )
Complete samples20BE(4.0457, 0.0389)(0.9896, 0.0002)
MLE(4.4374, 0.4789)(1.0287, 0.0014)
Type II censored samples (20%)100BE(4.1136, 0.1671)(0.9989, 1.6214 × 10 5 )
MLE(4.0985, 0.4564)(1.0056, 5.3148 × 10 5 )
Table 4. Numerical results of the real data analyses.
Table 4. Numerical results of the real data analyses.
rParametersBE2.5% Percentile50% Percentile97.5% Percentile
r = 24 α 0.33290.21520.33040.4645
β 0.76090.31800.79630.9923
r = n = 30 α 0.42790.31520.42640.5590
β 0.87230.58960.90160.9956
Table 5. Comparison of two estimation methods.
Table 5. Comparison of two estimation methods.
Estimation MethodSample CaseParameterEstimation Results
MLEComplete samples α 2.871
β 0.067
BEComplete samples α 2.823
β 0.065
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, F.; Wei, S.; Zhao, M. Bayesian Estimation of a New Pareto-Type Distribution Based on Mixed Gibbs Sampling Algorithm. Mathematics 2024, 12, 18. https://doi.org/10.3390/math12010018

AMA Style

Li F, Wei S, Zhao M. Bayesian Estimation of a New Pareto-Type Distribution Based on Mixed Gibbs Sampling Algorithm. Mathematics. 2024; 12(1):18. https://doi.org/10.3390/math12010018

Chicago/Turabian Style

Li, Fanqun, Shanran Wei, and Mingtao Zhao. 2024. "Bayesian Estimation of a New Pareto-Type Distribution Based on Mixed Gibbs Sampling Algorithm" Mathematics 12, no. 1: 18. https://doi.org/10.3390/math12010018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop