Next Article in Journal
Integrating Copula-Based Random Forest and Deep Learning Approaches for Analyzing Heterogeneous Treatment Effects in Survival Analysis
Previous Article in Journal
Probabilistic Basis of Parametrized Relativistic Quantum Theory in Curved Spacetime
Previous Article in Special Issue
Bayesian Model Selection for Addressing Cold-Start Problems in Partitioned Time Series Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Empirical Bayes Estimators for Mean Parameter of Exponential Distribution with Conjugate Inverse Gamma Prior Under Stein’s Loss

1
Department of Statistics and Actuarial Science, College of Mathematics and Statistics, Chongqing University, Chongqing 401331, China
2
Yunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming 650500, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(10), 1658; https://doi.org/10.3390/math13101658
Submission received: 20 February 2025 / Revised: 23 April 2025 / Accepted: 15 May 2025 / Published: 19 May 2025
(This article belongs to the Special Issue Bayesian Statistical Analysis of Big Data and Complex Data)

Abstract

:
A Bayes estimator for a mean parameter of an exponential distribution is calculated using Stein’s loss, which equally penalizes gross overestimation and underestimation. A corresponding Posterior Expected Stein’s Loss (PESL) is also determined. Additionally, a Bayes estimator for a mean parameter is obtained under a squared error loss along with its corresponding PESL. Furthermore, two methods are used to derive empirical Bayes estimators for the mean parameter of the exponential distribution with an inverse gamma prior. Numerical simulations are conducted to illustrate five aspects. Finally, theoretical studies are illustrated using Static Fatigue 90% Stress Level data.

1. Introduction

The Empirical Bayes analysis uses a conjugate prior model, estimating hyperparameters from observations and using this estimate as a conventional prior for further inference [1,2]. It should be noted that prior distribution remains unknown because of the unknown hyperparameters. However, marginal distribution could be employed to recover the prior distribution based on observed data. For further literature about empirical Bayes, researchers are referred to [1,3,4,5,6,7,8,9,10].
The motivations behind this paper could be summarized below. Firstly, to our knowledge, no one in the empirical Bayes analysis literature has considered the exponential-inverse gamma model. Additionally, as highlighted by [11], when dealing with a parameter space Θ = , in the location case, it is appropriate that the squared error loss equally penalizes both overestimation and underestimation. However, in cases where we have a restricted positive space of parameters Θ = 0 , with lower bound setting at 0 as well as an asymmetric problem of estimation, selecting Stein’s loss [12,13] instead of squared error loss is more suitable since it equally penalizes both gross overestimation and underestimation—actions that tend towards 0 or ∞ will incur infinite losses. In our specific context of the exponential-inverse gamma model (1), where our interested parameter is θ n + 1 > 0 representing a mean parameter value, we opt for Stein’s loss. Relevant literature on Stein’s loss can be found for interested readers [9,14,15,16].
The key findings of this paper are outlined below. First of all, the Bayes estimator of the mean parameter for the exponential distribution is derived under Stein’s loss, and its Posterior Expected Stein’s Loss (PESL) is calculated within an exponential-inverse gamma model. Additionally, we obtain a Bayes estimator of a mean parameter using squared error loss as well as its associated PESL. Furthermore, we employ two methods, namely the method of moments and the method of Maximum Likelihood Estimation (MLE), to estimate a mean parameter of an exponential distribution, which is assumed with a prior distribution of inverse gamma in an empirical Bayes framework. In our simulations, we investigate five aspects. The simulations suggest that MLEs are superior to moment estimators in estimating hyperparameters. At last, we illustrate our theoretical studies by applying them to analyze Static Fatigue 90% Stress Level data.
The remaining part of the paper is structured below. In Section 2, four theorems will be presented. Section 3 will provide an illustration of five aspects in simulations. In Section 4, we employ the Static Fatigue 90% Stress Level data to demonstrate the empirical Bayes estimators’ calculations for a mean parameter of an exponential distribution with an inverse gamma prior. Finally, Section 5 offers conclusions and discussions.

2. Main Results

The exponential-inverse gamma model is assumed to generate observations X 1 , X 2 , , X n + 1 in our study:
X i θ i independent E x p ( θ i ) , θ i iid I G ˜ ( α , β ) , i = 1 , 2 , , n + 1 .
The hyperparameters α > 0 and β > 0 need to be estimated, while θ n + 1 > 0 represents a parameter of interest that is unknown. E x p ( θ i ) denotes an exponential distribution with a mean parameter of θ i , and I G ˜ α , β denotes an inverse gamma distribution characterized by two hyperparameters. According to [1], when observing data ( x 1 , x 2 , , x n + 1 ) = x ( n + 1 ) , statisticians aim to make inferences about θ n + 1 . Therefore, x n + 1 directly provides information regarding the parameter θ n + 1 , while Supplementary Information ( x 1 , x 2 , , x n ) = x ( n ) also offered. A relationship between the primary data x n + 1 as well as Supplementary Information x ( n ) is established through their common distributions E x p ( θ i ) and I G ˜ ( α , β ) .
There are two probability density function (pdf) forms for the exponential distribution. One form uses the mean (or scale) θ as the parameter (see [11]), and X E x p θ with pdf f X x | θ = 1 θ exp x θ , for x > 0 and θ > 0 . Another form utilizes the rate λ as the parameter (see [2,17]), and X E x p ˜ λ with pdf f X x | λ = λ exp λ x , for x > 0 and λ > 0 . The two pdfs are the same with a relationship of the parameters θ = 1 / λ .
Other distributions that have two alternative forms are the gamma and inverse gamma distributions. Similar to [18], suppose that X G α , β and Y = 1 / X I G α , β . The pdfs of X and Y are, respectively, given by
f X x | α , β = 1 Γ α β α x α 1 exp x β , x > 0 , α , β > 0 , f Y y | α , β = 1 Γ α β α 1 y α + 1 exp 1 β y , y > 0 , α , β > 0 .
Moreover, suppose that X ˜ G ˜ α , β and Y ˜ = 1 / X ˜ I G ˜ α , β . The pdfs of X ˜ and Y ˜ are, respectively, given by
f X ˜ x | α , β = β α Γ α x α 1 exp β x , x > 0 , α , β > 0 , f Y ˜ y | α , β = β α Γ α 1 y α + 1 exp β y , y > 0 , α , β > 0 .
The exponential-gamma model is assumed to generate observations X 1 , X 2 , , X n + 1 :
X i λ i independent E x p ˜ ( λ i ) , λ i iid G ˜ ( α , β ) , i = 1 , 2 , , n + 1 .
Firstly, the exponential-inverse gamma model (1) and the exponential-gamma model (2) are equivalent in the sense of their marginal pdfs. For convenience, now let E x p θ , E x p ˜ λ , G α , β , I G α , β , G ˜ α , β , and I G ˜ α , β be random variables with the corresponding distributions. For example, E x p θ is a random variable with the E x p θ distribution. It is easy to see that
E x p θ = G 1 , θ = G ˜ 1 , 1 θ = G ˜ 1 , λ = E x p ˜ λ = E x p ˜ 1 θ .
Therefore, the two hierarchical models (1) and (2) are equivalent. It is straightforward to derive that the two marginal pdfs of the two hierarchical models are the same, and they are equal to
m x | α , β = α β α β + x α + 1 , x > 0 , α , β > 0 .
Since the two marginal pdfs are the same, the moment estimators (displayed in Theorem 2) and the MLEs (see Theorem 3) of the hyperparameters α and β for the two hierarchical models (1) and (2) are the same.
Another reason to use (1) is that it motivates us to consider 16 hierarchical models of the gamma and inverse gamma distributions (see [18]). It is easy to see that I G ˜ ( α , β ) is a conjugate prior for the E x p θ = G 1 , θ distribution. Writing in the form of likelihood-prior, it is G 1 , θ I G ˜ ( α , β ) . Similarly, the G ˜ ( α , β ) is a conjugate prior for the E x p ˜ λ = G ˜ 1 , λ distribution. Writing in the form of likelihood-prior, it is G ˜ 1 , λ G ˜ ( α , β ) . The G 1 , θ I G ˜ ( α , β ) expression motivates us to consider G ν , θ , G ˜ ν , θ , I G ν , θ , and I G ˜ ν , θ as the likelihood, and G α , β , G ˜ α , β , I G α , β , and I G ˜ α , β as the prior, leading to 16 combinations of the likelihood-prior.
The introduction highlights the importance of calculating and utilizing a Bayes estimator for a mean parameter θ n + 1 under Stein’s loss, considering a prior I G ˜ ( α , β ) . This particular estimator is denoted as δ s π , θ n + 1 ( x n + 1 ) . The supplement provides further justifications for why Stein’s loss outperforms squared error loss on a restricted positive space of parameters Θ = 0 , . For clarity, Stein’s loss function is given by
L s θ , a = a θ 1 log a θ ,
while the squared error loss function is given by
L 2 θ , a = θ a 2 .

2.1. Bayes Estimators and PESLs

In the exponential-inverse gamma model, the following theorem provides a summary of the posterior of θ n + 1 and a marginal distribution of X. The proof of the theorem can be found in the supplement.
Theorem 1. 
The posterior of θ n + 1 for the exponential-inverse gamma model is obtained as
π θ n + 1 | x n + 1 I G ˜ α * , β * ,
in which
α * = α + 1 > 1 a n d β * = β + x n + 1 > 0 .
The marginal distribution of X, furthermore, is
m x | α , β = α β α β + x α + 1 , for x > 0 and α , β > 0 .
Two inequalities will be derived for any hierarchical model, assuming the existence of posterior expectations. The same as [15], under Stein’s loss, a Bayes estimator of θ n + 1 is
δ s π , θ n + 1 x n + 1 = E 1 θ n + 1 | x n + 1 1 ,
and under squared error loss, a Bayes estimator of θ n + 1 is
δ 2 π , θ n + 1 x n + 1 = E θ n + 1 | x n + 1 .
The same as [15], we can show that
δ s π , θ n + 1 x n + 1 δ 2 π , θ n + 1 x n + 1
by leveraging the application of Jensen’s inequality. The same as [15], the PESL at δ s π , θ n + 1 x n + 1 is
P E S L s π , θ n + 1 x n + 1 = log E 1 θ n + 1 | x n + 1 + E log θ n + 1 | x n + 1 .
The PESL at δ 2 π , θ n + 1 x n + 1 is
P E S L 2 π , θ n + 1 x n + 1 = E θ n + 1 | x n + 1 E 1 θ n + 1 | x n + 1 1 log E θ n + 1 | x n + 1 + E log θ n + 1 | x n + 1 .
The following should be noted
P E S L s π , θ n + 1 x n + 1 P E S L 2 π , θ n + 1 x n + 1 .
This is a straightforward result of a method for deriving a Bayes estimator, because δ s π , θ n + 1 x n + 1 is constructed to minimize the PESL.
The Bayes estimators δ s π , θ n + 1 x n + 1 and δ 2 π , θ n + 1 x n + 1 , as well as the PESLs P E S L s π , θ n + 1 x n + 1 and P E S L 2 π , θ n + 1 x n + 1 , are analytically calculated under the exponential-inverse gamma model. According to (3), we possess
θ n + 1 | x n + 1 I G ˜ α * , β * and 1 θ n + 1 | x n + 1 G ˜ α * , β * .
The Bayes estimator of θ n + 1 is elegantly expressed by the following equation under Stein’s loss (reminiscent of [15])
δ s π , θ n + 1 ( x n + 1 ) = β * α * ,
in which α * and β * are determined by (4). Additionally, under squared error loss, we obtain
δ 2 π , θ n + 1 ( x n + 1 ) = β * α * 1 , for α * > 1 .
It can be observed that
δ s π , θ n + 1 ( x n + 1 ) = β * α * < β * α * 1 = δ 2 π , θ n + 1 ( x n + 1 ) ,
which illustrates the theoretical investigation in (6). The analytical calculation of P E S L s π , θ n + 1 x n + 1 and P E S L 2 π , θ n + 1 x n + 1 requires the analytical calculation of E log θ n + 1 | x n + 1 , which is given by
E ( log θ n + 1 | x n + 1 ) = log β * ψ ( α * ) ,
in which ψ x is a digamma function. The PESL at δ s π , θ n + 1 x n + 1 and δ 2 π , θ n + 1 x n + 1 can be obtained through algebraic operations, similar to the approach presented in [15],
P E S L s π , θ n + 1 ( x n + 1 ) = log α * ψ ( α * ) ,
and
P E S L 2 π , θ n + 1 ( x n + 1 ) = 1 α * 1 + log ( α * 1 ) ψ ( α * ) , for α * > 1 .
The demonstration of this fact is straightforward
P E S L s π , θ n + 1 ( x n + 1 ) = log α * ψ ( α * ) log ( α * 1 ) ψ ( α * ) + 1 α * 1 = P E S L 2 π , θ n + 1 ( x n + 1 ) .
The sentence exemplifies the theoretical investigation of (7). Notice that the PESLs solely depend on α * , disregarding β * . Hence, the dependency of the PESLs lies exclusively on α , without considering β and x n + 1 . In both simulations and real data sections, we will provide examples illustrating (6) and (7). Furthermore, we shall demonstrate that the dependence of the PESLs is solely on α , excluding any influence from β and x n + 1 .
The Bayes estimators, as well as the PESLs considered in this subsection, assume prior knowledge of the hyperparameters α and β . In other words, the PESLs and Bayes estimators presented here are computed using the oracle method, which will be further discussed in Section 2.3.

2.2. Empirical Bayes Estimators of θn+1

Empirical Bayes estimators of θ n + 1 can be obtained by estimating the hyperparameters from the Supplementary Information x 1 , x 2 , , x n = x n . There are two commonly used methods for estimating the hyperparameters: the method of moments and the method of MLE.
The method of moments offers estimators for the hyperparameters of the exponential-inverse gamma model, and their consistencies are succinctly summarized below. The proof of the theorem can be found in the supplement.
Theorem 2. 
The hyperparameters of an exponential-inverse gamma model can be estimated using the method of moments, where
α 1 n = 2 A 2 A 1 2 A 2 2 A 1 2 , β 1 n = A 1 A 2 A 2 2 A 1 2 .
Here, A k = 1 n i = 1 n X i k , k = 1 , 2 represents X’s kth sample moment. Additionally, it should be noted that the moment estimators exhibit consistency as estimators for these hyperparameters.
The estimators of hyperparameters for the exponential-inverse gamma model, obtained through the method of MLE, are presented below, along with an overview of their consistent properties. The proof of the theorem can be found in the supplement.
Theorem 3. 
The hyperparameters’ estimators of an exponential-inverse gamma model obtained through the method of MLE, denoted as α 2 n and β 2 n , are determined by solving equations:
f 1 α , β = α 1 + log β 1 n i = 1 n log ( β + x i ) = 0 ,
f 2 α , β = α β 1 ( α + 1 ) 1 n i = 1 n β + x i 1 = 0 .
Furthermore, it is worth noting that these MLEs serve as consistent estimators.
It is infeasible to analytically calculate the MLEs for α and β through solving nonlinear Equations (12) and (13), necessitating the utilization of numerical solutions. Newton’s method can be employed to solve Equations (12) and (13) and find the MLEs. It should be noted that the MLEs exhibit a high degree of sensitivity to initial estimates; thus, moment estimators are commonly utilized as reliable initial estimates.
The estimators of the parameter in the exponential-inverse gamma model are enhanced through empirical Bayes methodology, obtained through both moment and MLE methods, and they are summarized below.
Theorem 4. 
Under Stein’s loss and using the method of moments, the empirical Bayes estimator of the parameter in the exponential-inverse gamma model is (8) with estimated hyperparameters according to Theorem 2. Alternatively, when employing the MLE method and under Stein’s loss, the empirical Bayes estimator of the parameter in the exponential-inverse gamma model is (8) with numerically determined hyperparameters as stated in Theorem 3.
Remark 1. 
In Theorem 4 and under Stein’s loss, the empirical Bayes estimator of the parameter in the exponential-inverse gamma model is given by (8), where the hyperparameters ( α , β ) are estimated from observed data through marginal likelihood. More specifically, using the method of moments, the hyperparameters are estimated from observed data according to Theorem 2. Alternatively, when employing the MLE method, the hyperparameters are numerically determined from observed data as stated in Theorem 3.

2.3. Theoretical Analysis of PESLs and Bayes Estimators of Three Methods

PESLs and Bayes estimators for three methods (oracle, moment, and MLE) will be theoretically compared in this subsection. It is worth noting that their numerical comparisons can be found in the supplement.
Subscripts 0, 1, and 2 correspond to oracle, moment, and MLE, respectively. Following calculations as before [9,15], the PESLs of these three methods are given by ( i = 0 , 1 , 2 )
P E S L i a | x n + 1 = a E i 1 θ n + 1 | x n + 1 1 log a + E i log θ n + 1 | x n + 1 ,
in which
E i 1 θ n + 1 | x n + 1 = α i * n β i * n = α i n + 1 β i n + x n + 1 ,
E i log θ n + 1 | x n + 1 = log β i * n ψ α i * n = log β i n + x n + 1 ψ α i n + 1 ,
α 0 * n = α 0 n + 1 = α + 1 = α * , β 0 * n = β 0 n + x n + 1 = β + x n + 1 = β * , α 1 * n = α 1 n + 1 , β 1 * n = β 1 n + x n + 1 , α 2 * n = α 2 n + 1 , β 2 * n = β 2 n + x n + 1 .
The hyperparameters are unknown, while the moment estimators of α and β in Theorem 2, and the numerically determined MLEs of α and β in Theorem 3, are presented.
Under Stein’s loss, a Bayes estimators of θ n + 1 minimize an associated PESL (see [1,11,15]), i.e.,
δ s , i π , θ n + 1 x n + 1 = arg min a A E i L s θ n + 1 , a | x n + 1 ,
in which A = a = a x n + 1 : a x n + 1 > 0 represents a space of actions. Herein a is an action (estimator), L s θ n + 1 , a denotes Stein’s loss, and θ n + 1 > 0 refers to the unknown interested parameter. The Bayes estimators can be easily derived in a manner similar to [15]:
δ s , i π , θ n + 1 x n + 1 = β i n + x n + 1 α i n + 1 .
The PESLs at δ s , i π , θ n + 1 x n + 1 for i = 0 , 1 , 2 exhibit similarities to those discussed in Zhang’s work [15]:
P E S L s , i π , θ n + 1 x n + 1 = log α i n + 1 ψ α i n + 1 .
Under squared error loss, a Bayes estimators of θ n + 1 are provided in a similar manner as described by [15]:
δ 2 , i π , θ n + 1 x n + 1 = β i n + x n + 1 α i n ,
for α i n > 0 , i = 0 , 1 , 2 . Additionally, the PESLs at δ 2 , i π , θ n + 1 x n + 1 can be derived in a manner similar to that of [15] for i = 0 , 1 , 2 :
P E S L 2 , i π , θ n + 1 x n + 1 = 1 α i n + log α i n ψ α i n + 1 .
The primary objectives of our study involve estimating the PESLs and Bayes estimators using the method of oracle. However, the hyperparameters remain unknown. The method of oracle possesses knowledge of these hyperparameters only in simulations, whereas in reality, they are undisclosed.
The positive aspect is that the PESLs and Bayes estimators can be obtained through the method of moments and method of MLE once the data x n = x 1 , , x n are provided. To compare the moment and MLE methods in simulations, we can assess the PESLs and Bayes estimators generated by these two methods against those obtained using the method of oracle. In simulations, the method that yields PESLs and Bayes estimators closer to those derived from the method of oracle is considered superior.

3. Simulations

We conduct simulations to examine the exponential-inverse gamma model (1) in this section. We present five key aspects of our findings. Firstly, we demonstrate an inequality (6) of Bayes estimators and an inequality (7) of PESLs using the method of oracle. Secondly, we show that both moment estimators and MLEs consistently estimate hyperparameters. Thirdly, we calculate the model’s goodness-of-fit with respect to simulated data. Fourthly, we compare numerically the PESLs and Bayes estimators obtained from three methods: oracle, moment, and MLE. Finally, the model’s marginal densities are plotted for different hyperparameters.
The simulated data are produced based on the model (1), with hyperparameters set as α = 3 and β = 2 . These values are selected because the conditions α > 2 and β > 0 are necessary for the moment-based estimation of the hyperparameters. Alternative numerical values for the hyperparameters can also be defined.

3.1. Two Inequalities Related to Bayes Estimators and PESLs

In this subsection, we aim to demonstrate numerically the two inequalities (6) and (7) related to Bayes estimators and PESLs using the oracle approach.
First, we fix α = 3 , β = 2 , and x n + 1 = 1 . Figure 1 presents the histogram for θ n + 1 | x n + 1 along with the estimated density curve of π ( θ n + 1 | x n + 1 ) . Our goal is to determine δ s π , θ n + 1 x n + 1 in a way that minimizes the PESL based on π ( θ n + 1 | x n + 1 ) . Numerical results show that
δ s π , θ n + 1 x n + 1 = 0.75 < 1 = δ 2 π , θ n + 1 x n + 1
and
P E S L s π , θ n + 1 x n + 1 = 0.1301767 < 0.1758280 = P E S L 2 π , θ n + 1 x n + 1 ,
which exemplify the theoretical studies of (6) and (7).
Next, consider allowing one of the variables α , β , or x n + 1 to vary while keeping the others constant. As depicted in Figure 2, both the Bayes estimators and the PESLs are represented as functions of α , β , and x n + 1 . From the left panels of the figure, it is evident that the Bayes estimators are influenced by α , β , and x n + 1 , illustrating the relationship (6). More precisely, the Bayes estimators decrease with increasing α , increase linearly with β , and also increase linearly with x n + 1 . On the other hand, the right panels reveal that the PESLs are solely dependent on α , showing no dependence on β or x n + 1 , thus exemplifying the inequality (7). Specifically, the PESLs exhibit a decreasing trend as α increases. In conclusion, the findings presented in Figure 2 validate the theoretical analyses outlined in (6) and (7).
Third, given that the Bayes estimators δ 2 π , θ n + 1 x n + 1 and δ s π , θ n + 1 x n + 1 , as well as the PESLs P E S L 2 π , θ n + 1 x n + 1 and P E S L s π , θ n + 1 x n + 1 , rely on α * and β * , where α * > 1 and β * > 0 , we can visualize their surfaces over the domain α * , β * ( 1 , 10 ] × ( 0 , 10 ] = D using the ‘persp3d()’ function from the R (1.0) package ‘rgl’ (see [16,19]). It is worth noting that while the ‘persp()’ function in the R package ‘graphics’ cannot overlay additional surfaces onto an existing one, ‘persp3d()’ has this capability. Furthermore, ‘persp3d()’ enables users to rotate the perspective plots of the surfaces according to their preferences. Figure 3 displays the surfaces of the Bayes estimators and the PESLs, along with the surfaces representing the differences between these estimators and PESLs. From the first two plots on the left, it is evident that δ s π , θ n + 1 x n + 1 < δ 2 π , θ n + 1 x n + 1 for all α * , β * in D. Similarly, the last two plots on the right indicate that P E S L s π , θ n + 1 x n + 1 < P E S L 2 π , θ n + 1 x n + 1 for all α * , β * in D. The findings presented in Figure 3 align with the theoretical results outlined in (6) and (7).

3.2. Consistencies of Moment Estimators and MLEs

In this subsection, we aim to demonstrate through numerical examples that both the moment estimators and the MLEs serve as consistent estimators for the hyperparameters α and β of the hierarchical exponential and inverse gamma model (1). It is important to note that in this subsection, only the data x n = x 1 , x 2 , , x n are utilized.
The estimators for the hyperparameters ( α and β ) based on the method of moments and MLE are denoted as ( α 1 n and β 1 n ) and ( α 2 n and β 2 n ), respectively. Consistency implies that α i n converges to α in probability as n approaches infinity, and β i n converges to β in probability as n approaches infinity, for i = 1 , 2 . Here, i = 1 corresponds to the moment estimator, and i = 2 corresponds to the MLE.
Let θ represent either hyperparameter α or β . Then, consistency can be expressed as θ i n converging to θ in probability as n tends to infinity, for i = 1 , 2 . Alternatively, consistency means that for every ε > 0 and every θ Θ , the following holds:
lim n P θ i n θ ε = 0 ,
The probabilities can be approximated by their corresponding frequencies:
P θ i n θ ε 1 M m = 1 M I θ i m n θ ε F θ i ε , M n ,
for i = 1 , 2 , where I ( A ) is the indicator function of A, equaling 1 if A is true and 0 otherwise, and M represents the number of simulations. Thus, the convergence of the frequencies ( F θ i ε , M n , with θ being α or β , and i = 1 , 2 ) toward 0 indicates that the estimators are consistent.
The occurrences of the moment estimators ( α 1 n and β 1 n ) and the MLEs ( α 2 n and β 2 n ) for the hyperparameters ( α and β ) as n changes, given M = 100 and ε = 1 , 0.5 , and 0.1 , are presented in Table 1. By examining Table 1, we can identify the following observations.
  • For ε values of 1, 0.5, or 0.1, the frequencies of the estimators, denoted as F θ i ε , M n where θ = α or β and i = 1 , 2 , approach 0 as n grows toward infinity. This indicates that both the moment estimators and the MLEs are consistent estimators for the hyperparameters. When ε is set to 0.1, the frequencies of these estimators remain relatively high. Nevertheless, we can still observe a decreasing trend toward 0 as n increases to infinity.
  • By examining the frequencies associated with ε = 1 , 0.5, and 0.1, it can be seen that as ε decreases, the frequencies become higher. This occurs because the conditions
    θ i m n θ ε , for θ = α or β , i = 1 , 2 ,
    become simpler to satisfy.
  • By comparing the moment estimators with the MLEs for the hyperparameters α and β , it can be observed that the frequencies of the MLEs are lower than those of the moment estimators. This indicates that the MLEs outperform the moment estimators in terms of consistency when estimating the hyperparameters.

3.3. Model Fit Assessment: Kolmogorov–Smirnov Test

In this subsection, we aim to evaluate how well the hierarchical exponential and inverse gamma model (1) fits the simulated data (see [9,20,21]). It is important to note that in this subsection, we only utilize x n = x 1 , x 2 , , x n . Typically, the chi-square test serves as an indicator of fit quality. Nevertheless, due to its high sensitivity—stemming from challenges in determining the number of groups and identifying appropriate cut-points—we opt for an alternative approach. Instead of relying on the chi-square test, we will employ the Kolmogorov–Smirnov test (KS test) to assess the goodness-of-fit.
The Kolmogorov–Smirnov (KS) statistic is defined as the maximum distance D between the empirical cumulative distribution function (cdf), denoted as F n x , and the theoretical population cdf, F 0 x . Mathematically, this can be expressed as
D = sup < x < F n x F 0 x .
In the R programming environment, the KS test can be performed using the built-in function ‘ks.test()’, as detailed in references such as [22,23,24]. It is important to note that the KS test is primarily applicable for one-dimensional continuous distributions. However, adaptations of the KS test for discrete data have also been explored in recent literature, including works by [25,26,27].
In the context of our study, the null hypothesis assumes that the random variable X follows an exponential-inverse gamma distribution with parameters α and β , written as H 0 : X E x p I G ˜ ( α , β ) . The E x p I G ˜ ( α , β ) represents the marginal distribution derived from the hierarchical exponential and inverse gamma model (1). The marginal density of the E x p I G ˜ ( α , β ) distribution, as given in (5), is clearly one-dimensional and continuous. Consequently, the KS test serves as a suitable tool for assessing the goodness-of-fit in this scenario.
The outcomes of the KS test for assessing the goodness-of-fit of the model (1) to the simulated data are presented in Table 2. It is important to note that the data were generated based on the hierarchical exponential and inverse gamma model (1) with parameters α = 3 and β = 2 . In this table, the hyperparameters α and β are estimated using three distinct approaches. The first approach is the oracle method, where the true values of the hyperparameters α = 3 and β = 2 are known. The second approach is the moment estimation method, in which α and β are estimated using their respective moment estimators (as described in Theorem 2). The third approach is the MLE method, where α and β are estimated using their MLEs (as outlined in Theorem 3). In this table, the sample size is set to n = 1000 , and the number of simulations conducted is M = 100 . To evaluate the performance, we employ five metrics proposed in [9]: D ¯ , p - value ¯ , (min D)%, (max p-value)%, and Accept H 0 % . Here, D ¯ represents the average D values (14) across M simulations, with smaller values indicating better performance. The p - value ¯ denotes the average p-values from M simulations, where larger values are preferred. (min D)% refers to the proportion of M simulations in which the minimum D value is achieved among the three methods, with higher percentages being more favorable. The sum of the three (min D)% values should equal 1. (max p-value)% indicates the percentage of M simulations where the maximum p-value is attained among the three methods, with higher percentages also being desirable. Similarly, the sum of the three (max p-value)% values should equal 1. Lastly, Accept H 0 % reflects the proportion of M simulations in which the null hypothesis H 0 (defined as p-value > 0.05 ) is accepted for each method, with high values being preferable. Each Accept H 0 % value should fall within the range of 0–100%.
By examining Table 2, the following observations can be made.
  • The D ¯ values obtained using the three methods are 0.0270, 0.0230, and 0.0205, respectively. This indicates that the MLE method performs the best, followed by the moment method as the second-best, while the oracle method ranks last. One plausible explanation for this outcome is that in (14), the empirical cdf F n x relies on data, and the population cdfs F 0 x for both the MLE and moment methods also depend on data. In contrast, the population cdf F 0 x for the oracle method does not rely on data.
  • The p - value ¯ corresponding to the three methods are 0.5102, 0.6683, and 0.7693, respectively. These results align with the previous findings, ranking the MLE method first, the moment method second, and the oracle method third. The explanation for this phenomenon has been discussed in the preceding paragraph.
  • The (min D)% values for the three methods are 0.16, 0.15, and 0.69, respectively. Notably, the (min D)% value of the MLE method represents more than half of the M simulations. Based on these results, the order of preference among the three methods is the MLE method, the oracle method, and the moment method.
  • The (max p-value)% values for the three methods are also 0.16, 0.15, and 0.69, respectively. A smaller D value corresponds to a larger p-value, meaning the smallest D value correlates with the largest p-value. Consequently, the (min D)% and (max p-value)% values for the three methods are identical. Thus, the preferred order remains the MLE method, the oracle method, and the moment method.
  • Finally, the Accept H 0 % values for the three methods are 0.97, 0.99, and 1.00, respectively. Once again, the ranking of the methods in terms of preference is the MLE method, the moment method, and the oracle method.
Figure 4 presents the boxplots of D values for the three methods (on the left) and the boxplots of p-values for the three methods (on the right). Based on this figure, we can identify the following observations.
  • The D values obtained using the oracle method are considerably higher than those of the other two methods. Given that smaller D values are preferable, the ranking of the three methods in terms of performance is: the MLE method, followed by the moment method, and lastly the oracle method.
  • Additionally, the p-values of the oracle method are notably lower than those of the other two methods. Since larger p-values are more desirable, the preference order for the three methods based on p-values remains the same: the MLE method, the moment method, and finally the oracle method.
  • It is also worth noting that there is an inverse relationship between D values and p-values; smaller D values correspond to larger p-values, while larger D values result in smaller p-values.
  • Overall, the MLE method outperforms the moment method in both D values and p-values.

3.4. Numerical Comparisons Between Bayes Estimators and PESLs of Three Methods

In this subsection, we will conduct a numerical comparison of the Bayes estimators and PESLs for three approaches: the oracle method, the moment method, and the MLE method. It should be highlighted that the theoretical comparisons among these three methods regarding their Bayes estimators and PESLs are presented in Section 2.3.
It is important to highlight that the data have been generated based on the hierarchical exponential and inverse gamma model (1), using α = 3 and β = 2 . Additionally, the oracle method is aware of the hyperparameters α = 3 and β = 2 during the simulation process.
The comparisons of α , β , δ s , δ 2 , P E S L s , and P E S L 2 for the three methods, given a sample size of n = 10 , 000 and a number of simulations M = 100 , are presented in Figure 5. By examining the figure, we can identify that the MLE method demonstrates superior performance compared to the method of moments in several aspects. Specifically:
  • For the estimators of α and β , the MLE approach aligns significantly closer with the oracle method than the moment method.
  • Regarding the Bayes estimators δ s , i π , θ n + 1 x n + 1 and δ 2 , i π , θ n + 1 x n + 1 i = 0 , 1 , 2 , the MLE method exhibits a slight advantage over the moment method in terms of proximity to the oracle method. The differences among these estimators are minimal, making their respective curves nearly indistinguishable.
  • In the context of the PESL functions P E S L s , i π , θ n + 1 x n + 1 and P E S L 2 , i π , θ n + 1 x n + 1 i = 0 , 1 , 2 , the MLE method again shows a marked improvement over the moment method in its alignment with the oracle method.
  • Overall, the graphical results consistently suggest that the MLE method outperforms the moment method, as demonstrated by the hyperparameter estimators, Bayes estimators, and PESL values obtained through the MLE method being more consistent with those derived from the oracle method than those from the moment method.
Figure 6 presents the boxplots of absolute errors obtained using the oracle method for both the moment method and the MLE method, with a sample size of n = 10,000 and M = 100 simulations. The visualizations consistently demonstrate that the MLE method outperforms the moment method. This is because the absolute errors associated with the oracle method for the hyperparameter estimators, Bayes estimators, and PESLs are significantly smaller when using the MLE method compared to the moment method.
The summary of the averages and proportions of absolute errors for the oracle method, using both the moment method and the MLE method, is presented in Table 3. Specifically, these averages represent the sample means of the absolute error vectors derived from the oracle method via the moment and MLE methods, where smaller values indicate superior performance. The proportions reflect the sample ratios of absolute errors achieved by each method that are equal to the minimum value between the two methods, with higher proportions being more desirable. From the table, it can be observed that the MLE method yields significantly smaller average absolute errors compared to the moment method. Additionally, the proportions of absolute errors obtained through the MLE method are notably larger than those from the moment method. Overall, the data in the table demonstrates that the MLE method outperforms the moment method in terms of both the averages and proportions of absolute errors as determined by the oracle method.
Let θ equal α or β . Suppose θ ^ i represents an estimator of θ , where i = 1 corresponds to the method of moments estimator and i = 2 corresponds to the MLE. The Mean Square Error (MSE) for the estimator θ ^ i is given by (as referenced in [11]):
MSE θ ^ i = E θ ^ i θ 2 1 M m = 1 M θ ^ i m θ 2 .
In a similar manner, the Mean Absolute Error (MAE) for the estimator θ ^ i is defined as (also from [11]):
MAE θ ^ i = E θ ^ i θ 1 M m = 1 M θ ^ i m θ .
It is important to note that both MSE and MAE serve as criteria for comparing two distinct estimators. These metrics function as risk functions (or expected loss functions) associated with the estimators of hyperparameters, providing a measure of the estimators’ performance. Generally, smaller values of MSE and MAE indicate superior performance of the estimator.
The MSE and MAE for the hyperparameter estimators using both the moment method and the MLE method are presented in Table 4. As shown in the table, when estimating the hyperparameters α and β , the MLE method significantly outperforms the moment method, since the MSE and MAE values obtained via the MLE method are considerably lower than those achieved through the moment method.

3.5. Marginal Densities for Different Hyperparameters

In this subsection, we aim to illustrate the marginal densities of the hierarchical exponential and inverse gamma model (1) using different hyperparameters α and β . It is important to note that the marginal density of X is defined by (5), which depends on the two hyperparameters α and β . Here, we will investigate the variations in the marginal densities in relation to the density where the hyperparameters are set as α = 3 and β = 2 . Additionally, other specific values for the hyperparameters can also be considered.
Figure 7 illustrates the marginal densities for different values of α ( 1 , 3 , 5 , 7 ), with β remaining constant at 2. As observed in the figure, an increase in α corresponds to a higher peak value of the curve. Additionally, all marginal densities exhibit a decreasing relationship with x and display right-skewed distributions.
Figure 8 illustrates the marginal densities for different values of β ( 1 , 2 , 4 , 6 ), with α kept constant at 3. As observed in the figure, an increase in β corresponds to a reduction in the peak value of the curve. Additionally, all marginal densities exhibit a decreasing trend as functions of x and display right-skewed distributions.

4. An Illustrative Data Instance

In this section, we employ the data of Static Fatigue 90% Stress Level to exemplify our methodologies (refer to [28]). Kevlar Epoxy is a material utilized in the National Aeronautics and Space Administration (NASA) space shuttle. Strands of this epoxy were subjected to testing at 90% breaking strength. The provided data represents the time until failure in hours at the 90% stress level for a randomly selected sample of 50 epoxy strands.
The sample is x 1 , x 2 , , x n + 1 = x n + 1 (the data for Static Fatigue 90% Stress Level). The sample’s histogram is illustrated in Figure 9, along with its corresponding curve of density estimation. From the figure, it can be observed that the histogram exhibits a gradual decrease, indicating that the exponential-inverse gamma model would be suitable (refer to subsection “Marginal densities for various hyperparameters” in the supplement for more details).
The hyperparameters’ estimators, model’s goodness-of-fit, empirical Bayes estimators of a mean parameter for an exponential-inverse gamma model, PESLs, and mean and variance of Static Fatigue 90% Stress Level data obtained through both moment and MLE methods are reported in Table 5. We see several findings from this table.
  • The hyperparameters’ moment estimators and MLEs exhibit significant disparities. However, it does not imply that the exponential-inverse gamma model fails to fit the actual data or that both types of estimators are inconsistent in estimating the hyperparameters. The substantial divergence between these two estimation methods can be attributed to the relatively small sample size of n = 49 . Nevertheless, note that the MLEs for the hyperparameters are more dependable, as demonstrated by the simulations results in the supplement.
  • To measure goodness-of-fit, use a KS test. A p-value obtained from the method of moments, 0.4389 > 0.05 , indicates that the E x p I G ˜ ( α , β ) model with hyperparameters calculated by the method of moments well fits the sample x 1 , x 2 , , x n = x n . The same goes for it, a p-value derived from the method of MLE, 0.4708 > 0.05 , suggests that the E x p I G ˜ ( α , β ) model with hyperparameters calculated by MLEs also fits the sample well. Comparing both methods reveals that MLEs’ D values are observed to be smaller, while their corresponding p-values tend to be larger than those of moment estimators; thus indicating that estimating α and β using MLEs results in a better fit for the E x p I G ˜ ( α , β ) distribution to sample x n .
  • When estimating the hyperparameters using the MLE method, we observe that δ s π , θ n + 1 x n + 1 is 1.179281, which is less than δ 2 π , θ n + 1 x n + 1 at 1.229669. Additionally, P E S L s π , θ n + 1 x n + 1 is calculated as 0.020628 and it is lower than P E S L 2 π , θ n + 1 x n + 1 , which equals to 0.021516. The same trend can be observed when estimating the hyperparameters using the method of moments for both PESLs and Bayes estimators. Therefore, we have exemplified (6) and (7).
  • The E X , which represents the Static Fatigue 90% Stress Level data, is estimated as i = 1 , 2 :
    E X = β α 1 β i n α i n 1 .
    Meanwhile, Var X can be evaluated as i = 1 , 2 :
    Var X = α β 2 α 1 2 α 2 α i n β i 2 n α i n 1 2 α i n 2 .
    It is noteworthy that despite different estimators for hyperparameters, both methods yield very similar results by E X and Var X . Furthermore, it can be observed that
    E X 1.252418 and Var X 1.715116
    using the method of MLE. The method of moments yields similar values for E X and Var X .

5. Conclusions and Discussions

The posterior and marginal densities for the exponential-inverse gamma model are calculated in Theorem 1. Subsequently, Bayes estimators as well as PESLs of θ n + 1 are computed. Furthermore, these estimators satisfy two inequalities (6) and (7). The hyperparameters of the exponential-inverse gamma model are estimated using the method of moments as embodied in Theorem 2. Additionally, the method of MLE provides estimators for these hyperparameters, which are embodied in Theorem 3. At last, under Stein’s loss, both methods yield empirical Bayes estimators for a mean parameter of an exponential-inverse gamma model as presented in Theorem 4.
In the simulation studies, we systematically investigate five critical dimensions of the proposed framework. Firstly, we rigorously validate the inequality relationships between Bayes estimators (6) and PESLs (7) under oracle hyperparameter settings. Secondly, we establish the asymptotic consistency of both moment estimators and MLEs for hyperparameter recovery. Thirdly, the model’s fit to simulated data is quantitatively evaluated using Kolmogorov–Smirnov statistics. Fourthly, comprehensive numerical comparisons of PESLs and Bayes estimators are conducted across three approaches—oracle, moment-based, and MLE-driven methods. Finally, we characterize the model’s behavior by visualizing marginal density variations under distinct hyperparameter configurations.
The methods are illustrated using the Static Fatigue 90% Stress Level data. Table 5 summarizes hyperparameters’ estimators, model’s goodness-of-fit, empirical Bayes estimators of mean parameter for exponential-inverse gamma model, PESLs, and mean and variance of Static Fatigue 90% Stress Level data obtained through methods of moment and MLE. The exponential-inverse gamma model, when hyperparameters are calculated by the method of MLE, exhibits a superior goodness-of-fit to real data compared to that calculated by the method of moments. Furthermore, we demonstrate (6) and (7) for the real data.
The numerical results imply that MLEs outperform moment estimators in calculating hyperparameters α and β . Nevertheless, this improvement has a cost. In comparison to moment estimators, MLEs require more computational resources, are numerically unstable, necessitate β greater than 0 throughout the process of iteration, and lack closed-form answers. Hyperparameters’ MLEs are highly ingenious for starting estimations, whereas estimators from the method of moments often serve as reliable initial estimates. Furthermore, if there are cases where MLEs are virtually nonexistent, the presence of a valid reason allows us to utilize estimators from the method of moments.
The inverse gamma prior was chosen for its conjugacy with the exponential likelihood, ensuring analytical tractability in deriving Bayes estimators and PESL—a critical first step in establishing theoretical properties for this underexplored model. While focusing on a single prior limits generality, it provides a foundational framework for future extensions to broader classes (e.g., non-conjugate or robust priors), which we acknowledge as valuable next steps. Our goal here was to rigorously address the uniqueness of Stein’s loss in this specific hierarchical setting, but we agree that exploring prior sensitivity would enrich the analysis and plan to investigate this in subsequent work.
The exponential model was chosen for its conjugacy with the inverse gamma prior, enabling closed-form derivations of Bayes estimators and PESL under Stein’s loss—a key focus of our theoretical analysis. This simplicity allowed us to rigorously isolate the impact of Stein’s loss in restricted parameter spaces, with direct applications to reliability (e.g., fatigue data). While the exponential case serves as a foundational step, extending this framework to other models (e.g., Weibull, gamma) would broaden its utility. Such extensions, though computationally challenging, are a promising direction for future work, and we appreciate the reviewer’s emphasis on this potential.
While our focus was on the conjugate inverse gamma prior for analytical tractability, the methodology can theoretically accommodate non-conjugate or loss-informed priors. However, this would require numerical methods (e.g., Markov Chain Monte Carlo (MCMC)) for posterior computation and alter the closed-form Bayes estimators and PESL expressions.
Direct numerical comparison with MCMC is challenging due to fundamental methodological differences: MCMC requires hyperpriors to ensure posterior integrability (full Bayesian), while our empirical Bayes approach avoids hyperpriors by estimating hyperparameters from data. However, our method leverages conjugacy to achieve closed-form estimators, eliminating MCMC’s sampling/convergence steps. This yields 10–100× faster computation (e.g., seconds vs. minutes) with comparable accuracy in conjugate settings. While full MCMC excels in generality, our framework prioritizes efficiency for exponential-inverse gamma models.
Stein’s loss was chosen due to its asymmetric penalization of extreme errors (e.g., θ 0 or ∞), which aligns with our restricted parameter space ( θ > 0 ). This ensures robustness in reliability contexts (e.g., underestimating failure rates risks catastrophic outcomes). In the supplement, we have highlighted scenarios where Stein’s loss is preferable (positive restricted parameters) versus inappropriate (unrestricted parameters) than the squared error loss function. In this manuscript, the parameter of interest is θ > 0 , the mean parameter of the exponential distribution. Therefore, Stein’s loss function is more appropriate than the squared error loss function. Consequently, we consider the PESL as a criterion to compare Bayes estimators.
The empirical Bayes framework fundamentally differs from full hierarchical Bayesian methods by estimating hyperparameters ( α , β ) via marginal likelihood maximization from observed data (Section 2.2), rather than assigning hyperpriors. This eliminates hyperprior specification, enhancing computational efficiency and reducing subjectivity—key advantages in conjugate settings like our exponential-inverse gamma model. While empirical Bayes sacrifices full probabilistic coherence, it provides tractable, data-driven estimates ideal for applications requiring rapid inference (e.g., reliability analysis). We emphasize the role of empirical Bayes in balancing computational practicality with robust hyperparameter estimation.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/math13101658/s1.

Author Contributions

Conceptualization, Y.-Y.Z.; Funding acquisition, Y.-Y.Z.; Methodology, Z.L. and Y.-Y.Z.; Software, Z.L. and Y.-G.S.; Supervision, Y.-Y.Z.; Validation, Z.L. and Y.-G.S.; Writing—original draft, Z.L. and Y.-Y.Z.; Writing—review and editing, Z.L., Y.-Y.Z. and Y.-G.S. All authors have read and agreed to the published version of the manuscript.

Funding

National Social Science Fund of China (21XTJ001) supports this research.

Data Availability Statement

The data utilized to substantiate findings of the study could be accessed at https://college.cengage.com/mathematics/brase/understandable_statistics/7e/students/datasets/svls/frames/frame.html (accessed on 20 February 2025).

Acknowledgments

The authors would like to thank the Editor and the reviewers for their constructive comments, which have led to a substantial improvement of the paper. During the preparation of this work, the authors used Youdao AIBox (with version number 11.1.2) in order to improve language and readability and to remove duplicates from the paper. Moreover, the authors used DeepSeek (web version) to assist us in responding to reviewers’ comments. After using these tools, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Robert, C.P. The Bayesian Choice: From Decision-Theoretic Motivations to Computational Implementation, 2nd paperback ed.; Springer: New York, NY, USA, 2007. [Google Scholar]
  2. Gelman, A.; Carlin, J.B.; Stern, H.S.; Dunson, D.B.; Vehtari, A.; Rubin, D.B. Bayesian Data Analysis, 3rd ed.; Chapman & Hall: London, UK, 2014. [Google Scholar]
  3. Efron, B. Large-Scale Inference: Empirical Bayes Methods for Estimation, Testing, and Prediction; Cambridge University Press: New York, NY, USA, 2010. [Google Scholar]
  4. van Houwelingen, H.C. The role of empirical Bayes methodology as a leading principle in modern medical statistics. Biom. J. 2014, 56, 919–932. [Google Scholar] [CrossRef]
  5. Ghosh, M.; Kubokawa, T.; Kawakubo, Y. Benchmarked empirical Bayes methods in multiplicative area-level models with risk evaluation. Biometrika 2015, 102, 647–659. [Google Scholar] [CrossRef]
  6. Satagopan, J.M.; Sen, A.; Zhou, Q.; Lan, Q.; Rothman, N.; Langseth, H.; Engel, L.S. Bayes and empirical Bayes methods for reduced rank regression models in matched case-control studies. Biometrics 2016, 72, 584–595. [Google Scholar] [CrossRef] [PubMed]
  7. Martin, R.; Mess, R.; Walker, S.G. Empirical Bayes posterior concentration in sparse high-dimensional linear models. Bernoulli 2017, 23, 1822–1847. [Google Scholar] [CrossRef]
  8. Mikulich-Gilbertson, S.K.; Wagner, B.D.; Grunwald, G.K.; Riggs, P.D.; Zerbe, G.O. Using empirical Bayes predictors from generalized linear mixed models to test and visualize associations among longitudinal outcomes. Stat. Methods Med Res. 2019, 28, 1399–1411. [Google Scholar] [CrossRef] [PubMed]
  9. Sun, J.; Zhang, Y.Y.; Sun, Y. The empirical Bayes estimators of the rate parameter of the inverse gamma distribution with a conjugate inverse gamma prior under Stein’s loss function. J. Stat. Comput. Simul. 2021, 91, 1504–1523. [Google Scholar] [CrossRef]
  10. Iqbal, A.; Shad, M.Y.; Yassen, M.F. Empirical E-Bayesian estimation of hierarchical poisson and gamma model using scaled squared error loss function. Alex. Eng. J. 2023, 69, 289–301. [Google Scholar] [CrossRef]
  11. Casella, G.; Berger, R.L. Statistical Inference, 2nd ed.; Duxbury: Pacific Grove, CA, USA, 2002. [Google Scholar]
  12. James, W.; Stein, C. Estimation with Quadratic Loss. In Breakthroughs in Statistics: Foundations and Basic Theory; Springer: New York, NY, USA, 1961; Volume 1, pp. 361–380. [Google Scholar]
  13. Brown, L.D. Comment on the Paper by Maatta and Casella. Stat. Sci. 1990, 5, 103–106. [Google Scholar]
  14. Bobotas, P.; Kourouklis, S. On the estimation of a normal precision and a normal variance ratio. Stat. Methodol. 2010, 7, 445–463. [Google Scholar] [CrossRef]
  15. Zhang, Y.Y. The Bayes rule of the variance parameter of the hierarchical normal and inverse gamma model under Stein’s loss. Commun. Stat.-Theory Methods 2017, 46, 7125–7133. [Google Scholar] [CrossRef]
  16. Sun, Y.; Zhang, Y.Y.; Sun, J. The empirical Bayes estimators of the parameter of the uniform distribution with an inverse gamma prior under Stein’s loss function. Commun. Stat.-Simul. Comput. 2024, 53, 3027–3045. [Google Scholar] [CrossRef]
  17. Mao, S.S.; Tang, Y.C. Bayesian Statistics, 2nd ed.; China Statistics Press: Beijing, China, 2012. [Google Scholar]
  18. Zhang, L.; Zhang, Y.Y. The Bayesian posterior and marginal densities of the hierarchical gamma-gamma, gamma-inverse gamma, inverse gamma-gamma, and inverse gamma-inverse gamma models with conjugate priors. Mathematics 2022, 10, 4005. [Google Scholar] [CrossRef]
  19. Adler, D.; Murdoch, D.; Nenadic, O.; Urbanek, S.; Chen, M.; Gebhardt, A.; Bolker, B.; Csardi, G.; Strzelecki, A.; Senger, A.; et al. rgl: 3D Visualization Using OpenGL, R package version 0.98.1. 2017. Available online: https://cran.r-project.org/web/packages/rgl/index.html (accessed on 20 February 2025).
  20. Xue, Y.; Chen, L.P. Statistical Modeling and R Software; Tsinghua University Press: Beijing, China, 2007. [Google Scholar]
  21. Ross, S. Simulation, 5th ed.; Mechanical Industry Press: Beijing, China, 2013. [Google Scholar]
  22. Conover, W.J. Practical Nonparametric Statistics; John Wiley & Sons: New York, NY, USA, 1971; Pages 295–301 (one-sample Kolmogorov test), 309–314 (two-sample Smirnov test). [Google Scholar]
  23. Durbin, J. Distribution Theory for Tests Based on the Sample Distribution Function; SIAM: Philadelphia, PA, USA, 1973. [Google Scholar]
  24. Marsaglia, G.; Tsang, W.W.; Wang, J.B. Evaluating Kolmogorov’s distribution. J. Stat. Softw. 2003, 8, 1–4. [Google Scholar] [CrossRef]
  25. Dimitrova, D.S.; Kaishev, V.K.; Tan, S. Computing the Kolmogorov-Smirnov distribution when the underlying cdf is purely discrete, mixed or continuous. J. Stat. Softw. 2020, 95, 1–42. [Google Scholar] [CrossRef]
  26. Aldirawi, H.; Yang, J.; Metwally, A.A. Identifying appropriate probabilistic models for sparse discrete omics data. In Proceedings of the IEEE EMBS International Conference on Biomedical and Health Informatics (BHI), Chicago, IL, USA, 19–22 May 2019. [Google Scholar]
  27. Santitissadeekorn, N.; Lloyd, D.J.B.; Short, M.B.; Delahaies, S. Approximate filtering of conditional intensity process for Poisson count data: Application to urban crime. Comput. Stat. Data Anal. 2020, 144, 106850. [Google Scholar] [CrossRef]
  28. R. E. Barlow University of California, Berkeley. Static Fatigue 90% Stress Level. 2021. Available online: https://college.cengage.com/mathematics/brase/understandable_statistics/7e/students/datasets/svls/frames/frame.html (accessed on 19 December 2021).
Figure 1. The histogram of θ n + 1 | x n + 1 and the density estimation curve of π ( θ n + 1 | x n + 1 ) .
Figure 1. The histogram of θ n + 1 | x n + 1 and the density estimation curve of π ( θ n + 1 | x n + 1 ) .
Mathematics 13 01658 g001
Figure 2. Left: The estimators expressed as functions of α , β , and x n + 1 . Right: The PESLs represented as functions of α , β , and x n + 1 .
Figure 2. Left: The estimators expressed as functions of α , β , and x n + 1 . Right: The PESLs represented as functions of α , β , and x n + 1 .
Mathematics 13 01658 g002
Figure 3. The domain of α * , β * is defined as D = 1 , 10 × 0 , 10 for all the graphs. In the axes of these graphs, “a” corresponds to α * and “b” corresponds to β * . In the upper two plots, the red surface represents δ 2 π , θ n + 1 x n + 1 , while the blue surface represents δ s π , θ n + 1 x n + 1 . (upper left) The estimators as functions of α * and β * . δ 2 π , θ n + 1 x n + 1 > δ s π , θ n + 1 x n + 1 for all α * , β * on D. (upper right) The PESLs as functions of α * and β * . P E S L 2 π , θ n + 1 x n + 1 > P E S L s π , θ n + 1 x n + 1 for all α * , β * on D. (lower left) The surface of δ 2 π , θ n + 1 x n + 1 δ s π , θ n + 1 x n + 1 which is positive for all α * , β * on D. (lower right) The surface of P E S L 2 π , θ n + 1 x n + 1 P E S L s π , θ n + 1 x n + 1 which is also positive for all α * , β * on D.
Figure 3. The domain of α * , β * is defined as D = 1 , 10 × 0 , 10 for all the graphs. In the axes of these graphs, “a” corresponds to α * and “b” corresponds to β * . In the upper two plots, the red surface represents δ 2 π , θ n + 1 x n + 1 , while the blue surface represents δ s π , θ n + 1 x n + 1 . (upper left) The estimators as functions of α * and β * . δ 2 π , θ n + 1 x n + 1 > δ s π , θ n + 1 x n + 1 for all α * , β * on D. (upper right) The PESLs as functions of α * and β * . P E S L 2 π , θ n + 1 x n + 1 > P E S L s π , θ n + 1 x n + 1 for all α * , β * on D. (lower left) The surface of δ 2 π , θ n + 1 x n + 1 δ s π , θ n + 1 x n + 1 which is positive for all α * , β * on D. (lower right) The surface of P E S L 2 π , θ n + 1 x n + 1 P E S L s π , θ n + 1 x n + 1 which is also positive for all α * , β * on D.
Mathematics 13 01658 g003
Figure 4. The boxplots display the D values for the three methods on the left and the p-values for the three methods on the right.
Figure 4. The boxplots display the D values for the three methods on the left and the p-values for the three methods on the right.
Mathematics 13 01658 g004
Figure 5. A comparison of α , β , δ s , δ 2 , P E S L s , and P E S L 2 across the three methods for a sample size of n = 10,000 and a simulation count of M = 100 .
Figure 5. A comparison of α , β , δ s , δ 2 , P E S L s , and P E S L 2 across the three methods for a sample size of n = 10,000 and a simulation count of M = 100 .
Mathematics 13 01658 g005
Figure 6. The boxplots display the absolute errors obtained from the oracle method, using both the moment method and the MLE method, for a sample size of n = 10,000 and a number of simulations totaling M = 100 .
Figure 6. The boxplots display the absolute errors obtained from the oracle method, using both the moment method and the MLE method, for a sample size of n = 10,000 and a number of simulations totaling M = 100 .
Mathematics 13 01658 g006
Figure 7. The marginal densities for different values of α ( 1 , 3 , 5 , 7 ) while keeping β constant at 2.
Figure 7. The marginal densities for different values of α ( 1 , 3 , 5 , 7 ) while keeping β constant at 2.
Mathematics 13 01658 g007
Figure 8. The marginal densities for different values of β ( 1 , 2 , 4 , 6 ) while keeping α constant at 3.
Figure 8. The marginal densities for different values of β ( 1 , 2 , 4 , 6 ) while keeping α constant at 3.
Mathematics 13 01658 g008
Figure 9. The sample’s histogram.
Figure 9. The sample’s histogram.
Mathematics 13 01658 g009
Table 1. The frequencies of the moment estimators and the MLEs for the hyperparameters are examined as n changes, with M set to 100 and ε taking values of 1, 0.5, and 0.1.
Table 1. The frequencies of the moment estimators and the MLEs for the hyperparameters are examined as n changes, with M set to 100 and ε taking values of 1, 0.5, and 0.1.
Moment EstimatorsMLEs
n F α 1 ε , M n F β 1 ε , M n F α 2 ε , M n F β 2 ε , M n
ε = 11 × 1040.000.000.000.00
2 × 1040.000.000.000.00
4 × 1040.000.000.000.00
8 × 1040.000.000.000.00
ε = 0.51 × 1040.140.110.000.00
2 × 1040.020.010.000.00
4 × 1040.010.010.000.00
8 × 1040.000.000.000.00
ε = 0.11 × 1040.770.740.400.31
2 × 1040.700.690.240.15
4 × 1040.640.620.130.09
8 × 1040.480.460.030.01
Table 2. The outcomes of the KS test for assessing the model (1) fit to the simulated data.
Table 2. The outcomes of the KS test for assessing the model (1) fit to the simulated data.
OracleMomentMLE
D ¯ 0.02700.02300.0205
p - value ¯ 0.51020.66830.7693
(min D)%0.160.150.69
(max p-value)%0.160.150.69
(Accept H0)%0.970.991.00
Table 3. The means and ratios of absolute errors derived from the oracle method, using both the moment method and the MLE method, for the estimators of hyperparameters, Bayes estimators, and PESLs.
Table 3. The means and ratios of absolute errors derived from the oracle method, using both the moment method and the MLE method, for the estimators of hyperparameters, Bayes estimators, and PESLs.
AveragesProportions
MomentMLEMomentMLE
α0.24580.10990.280.72
β0.23540.09310.210.79
δ s 0.01990.01060.220.78
δ 2 0.02100.01670.370.63
P E S L s 0.00770.00370.280.72
P E S L 2 0.01380.00670.280.72
Table 4. The MSE and MAE of the hyperparameter estimators obtained using the moment method and the MLE method.
Table 4. The MSE and MAE of the hyperparameter estimators obtained using the moment method and the MLE method.
MSEMAE
MomentMLEMomentMLE
α 0.091780.018770.245820.10993
β 0.082040.013630.235380.09313
Table 5. The hyperparameters’ estimators, model’s goodness-of-fit, empirical Bayes estimators of a mean parameter for an exponential-inverse gamma model, PESLs, and mean and variance of Static Fatigue 90% Stress Level data obtained through both moment and MLE methods.
Table 5. The hyperparameters’ estimators, model’s goodness-of-fit, empirical Bayes estimators of a mean parameter for an exponential-inverse gamma model, PESLs, and mean and variance of Static Fatigue 90% Stress Level data obtained through both moment and MLE methods.
MomentMLE
Estimators of hyperparameters α 17.56223.404
β 20.75028.059
Model’s goodness-of-fitD0.12400.1209
p-value0.43890.4708
Empirical Bayes estimators and PESLs δ s π , θ n + 1 x n + 1 1.1566551.179281
δ 2 π , θ n + 1 x n + 1 1.2225161.229669
P E S L s π , θ n + 1 x n + 1 0.0271790.020628
P E S L 2 π , θ n + 1 x n + 1 0.0287410.021516
Mean and variance E X 1.2528571.252418
Var X 1.7713801.715116
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Zhang, Y.-Y.; Shi, Y.-G. Empirical Bayes Estimators for Mean Parameter of Exponential Distribution with Conjugate Inverse Gamma Prior Under Stein’s Loss. Mathematics 2025, 13, 1658. https://doi.org/10.3390/math13101658

AMA Style

Li Z, Zhang Y-Y, Shi Y-G. Empirical Bayes Estimators for Mean Parameter of Exponential Distribution with Conjugate Inverse Gamma Prior Under Stein’s Loss. Mathematics. 2025; 13(10):1658. https://doi.org/10.3390/math13101658

Chicago/Turabian Style

Li, Zheng, Ying-Ying Zhang, and Ya-Guang Shi. 2025. "Empirical Bayes Estimators for Mean Parameter of Exponential Distribution with Conjugate Inverse Gamma Prior Under Stein’s Loss" Mathematics 13, no. 10: 1658. https://doi.org/10.3390/math13101658

APA Style

Li, Z., Zhang, Y.-Y., & Shi, Y.-G. (2025). Empirical Bayes Estimators for Mean Parameter of Exponential Distribution with Conjugate Inverse Gamma Prior Under Stein’s Loss. Mathematics, 13(10), 1658. https://doi.org/10.3390/math13101658

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop