Next Article in Journal
Robust Parameter Designs Constructed from Hadamard Matrices
Previous Article in Journal
Rank-Based Control Charts Under Non-Overlapping Counting with Practical Applications in Logistics and Services
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Bell Regression Model for Fitting of Overdispersed Count Data with Application

by
Ameer Musa Imran Alhseeni
1 and
Hossein Bevrani
1,2,*
1
Department of Statistics, University of Tabriz, Tabriz 51666-15648, Iran
2
Department of Statistics, University of Kurdistan, Sanandaj 66177-15175, Iran
*
Author to whom correspondence should be addressed.
Stats 2025, 8(4), 95; https://doi.org/10.3390/stats8040095
Submission received: 5 September 2025 / Revised: 4 October 2025 / Accepted: 7 October 2025 / Published: 10 October 2025
(This article belongs to the Section Computational Statistics)

Abstract

The Bell regression model (BRM) is a statistical model that is often used in the analysis of count data that exhibits overdispersion. In this study, we propose a Bayesian analysis of the BRM and offer a new perspective on its application. Specifically, we introduce a G-prior distribution for Bayesian inference in BRM, in addition to a flat-normal prior distribution. To compare the performance of the proposed prior distributions, we conduct a simulation study and demonstrate that the G-prior distribution provides superior estimation results for the BRM. Furthermore, we apply the methodology to real data and compare the BRM to the Poisson and negative binomial regression model using various model selection criteria. Our results provide valuable insights into the use of Bayesian methods for estimation and inference of the BRM and highlight the importance of considering the choice of prior distribution in the analysis of count data.

1. Introduction

Count regression models are valuable for understanding the relationships between predictor variables and count outcomes in various domains, offering a flexible and powerful framework for analyzing discrete, non-negative data. Count data regression analysis has a wide range of important applications. In biological and genetic studies, it is used to analyze data such as the number of genes, genetic mutations, or disease occurrences over time [1]. In epidemiology and public health, this analysis helps assess disease incidence, mortality rates, and the frequency of health events in specific populations [2]. Social scientists employ count data regression to model phenomena such as criminal offenses, births, and deaths within particular areas. Economic research utilizes this method to examine events including business failures, patents, and the frequency of various economic activities [3]. In insurance and actuarial science, it is applied to model claim frequencies and policyholder behavior [4]. Finally, environmental studies benefit from count data regression to analyze ecological counts, such as species diversity in habitats and wildlife populations [5].
Undoubtedly, the Poisson regression model (PRM) is the primary choice for analyzing count data. However, this model has a significant limitation: it assumes that the variance in the count variable is equal to its mean. This assumption is often violated in real-world datasets due to overdispersion, where the variance exceeds the mean. Consequently, the PRM’s applicability is limited in such scenarios, prompting the need for alternative models. One such alternative is the Bell regression model (BRM), introduced by [6], which has been well-received. BRMs have been widely discussed in various contexts, for example, in the presence of multicollinearity [7,8,9,10,11,12,13], excess zeros [14,15], and shrinkage strategies [16,17].
In statistical modeling, Bayesian inference has emerged as a powerful approach for data analysis. By combining prior knowledge with observed data, Bayesian methods enable the quantification of uncertainty and the estimation of model parameters. Although the Bell regression model has been applied in many contexts, most existing work has focused on frequentist estimation or shrinkage techniques. A Bayesian treatment of the BRM that utilizes informative priors is largely absent from the literature. In particular, no study has explored the use of G-prior distributions in this setting, even though the choice of prior is known to strongly influence Bayesian inference. This paper aims to fill that gap with several contributions. First, we propose a Bayesian formulation of the Bell regression model and introduce the use of G-priors, with hyperparameters specified through KL-divergence. Second, we develop a tailored Metropolis–Hastings algorithm for efficient posterior inference. Third, we conduct a simulation study that directly compares G-priors with the commonly used flat-normal prior, showing that G-priors consistently yield more accurate and stable estimates. Finally, we demonstrate the practical value of our approach through an application to mine fracture data, where the BRM with G-priors provides a better fit than both the Poisson and negative binomial regression models, according to several Bayesian model selection criteria. Together, these results highlight how incorporating G-priors into Bayesian BRMs can strengthen inference and offer new insights for applied researchers working with overdispersed count data. This paper was previously published as a preprint [18].
The remainder of the paper is structured as follows: Section 2 provides a detailed explanation of the BRM, including prior specification, posterior inference, and model selection criteria. In Section 3, a simulation study is carried out to compare the proposed prior distributions for the BRM. The methodology is then illustrated in Section 3.2, where the BRM is compared to the Poisson regression model (PRM) and the negative binomial regression model (NBRM) using Bayesian inference. Finally, Section 4 provides a discussion and concluding remarks.

2. Material and Methods

2.1. Bell Regression Model

The discrete Bell distribution is introduced by [6] based on a series expansion due to [19,20]. Its probability mass function is defined as follows:
f ( y ) = θ y e 1 e θ B y y ! , y = 0 , 1 , 2 , ; θ > 0 ,
where B y = e 1 k = 0 k y k ! indicates the Bell numbers. The important characteristics of the Bell distribution are given by the following:
E ( Y ) = θ e θ , V ( Y ) = θ e θ ( 1 + θ ) = E ( Y ) ( 1 + θ ) .
Since θ > 0 , the variance value is greater than its mean, which is known as overdispersion. Consequently, the regression model is defined based on Bell distribution, which will be suitable for count data with overdispersion. In regression contents, it is common to assume that the mean of the distribution depends on the vector of covariates. Let y = ( y 1 , y 2 , , y n ) T be a random sample from B e l l ( θ i ) . We relate the μ i : = E ( Y i ) to p covariates by the log link function, i.e.,
log ( μ i ) = X i β , i = 1 , 2 , , n
where β = ( β 1 , β 2 , , β p ) is the model parameter vector and X i = ( x i 1 , x i 2 , , x i p ) is the ith observation for p model covariates. Due to (2), the parameter of the Bell distribution will be θ i = W 0 ( μ i ) ( i = 1 , 2 , , n ) , where W 0 ( . ) is the Lambert function [21]. Hence, the model in (1) can be reparameterized as follows:
f ( y i | X i ) = B y i y i ! e 1 e W 0 ( μ i ) [ W 0 ( μ i ) ] y i .

2.2. Bayesian Inference

Consider n observations of the response–covariates pair as D = { ( y i , X i ) , i = 1 , 2 , , n } . The likelihood function of the BRM is given as follows:
L ( D | β ) = i = 1 n B y i y i ! i = 1 n [ W 0 ( μ i ) ] y i e n i = 1 n e W 0 ( μ i ) ,
where μ i = exp X i β . Bayesian regression allows for the incorporation of prior knowledge about the parameters, which is particularly useful when such information is available or when making use of expert knowledge. The inference is based on the posterior distribution of BRM, i.e.,
π ( β | D ) = L ( D | β ) π ( β ) = i = 1 n B y i y i ! i = 1 n [ W 0 ( μ i ) ] y i e n i = 1 n e W 0 ( μ i ) × π ( β ) .
Thus, we first need to specify the prior distribution.

2.2.1. Specification of Priors

First, we consider a common prior on β = ( β 1 , β 2 , , β p ) for our model, the flat-normal prior N p ( 0 , τ 2 I p ) , where τ > 0 . When we set a large value for τ 2 , the resulting flat prior becomes a diffuse prior. This diffuse prior assigns equal probability to all possible values of the regression coefficients. Unfortunately, this can lead to overestimating the magnitude of the regression coefficient and being overconfident about its sign. This is problematic because when it comes to regression coefficients (other than the intercept), we are typically more interested in knowing the magnitude and sign of the effect. Thus, we consider the second prior on model parameters based on the idea presented in [22].
In a proposed regression model, if the mean μ of y i is assumed to be in the range of ( 0 , ) , and a subject matter expert has information on the marginal distribution of μ characterized by inverse-gamma distribution with parameters a μ and b μ , i.e., I G ( a μ , b μ ) , where a μ > 0 and b μ > 0 are known, the objective is to formulate a prior on β that incorporates this prior information while adjusting for covariates. Following [23], we consider a G-prior distribution for the regression parameters. This choice is particularly suitable for BRMs because it allows for prior information on the marginal distribution of the mean response to be incorporated while adjusting for covariates, making it a natural and flexible prior structure for regression models with overdispersed count data. The G-prior can be considered as follows:
β N p M u , g n ( X X ) 1 ,
where u = ( 1 , 0 , , 0 ) , M is a prior mean for the intercept and g > 0 is a scaling constant. Suppose that X 1 , X 2 , , X n i i d H ( x ) with the mean A and covariance matrix Σ . With X i including the intercept in the first element, the first element of A is one and entries in the first row and those in the first column of Σ all are zero. For the new subject with covariates X H and response y, we can see that the mean of y equals μ ( X i ) = h 1 ( X i β ) and also assume that X i and β are independent [22]. Therefore, we have
E ( X i β ) = E E X i β | X i = E X i M u = M ,
and
V ( X i β ) = E V X i β | X i + V E X i β | X i = E g n X i ( X X ) 1 X i + V M = g . t r n ( X X ) 1 + A A .
Since n ( X X ) 1 converges in probability to Σ + A A 1 ,
V ( X i β ) p g p .
Ref. [24] found that for a various H ( . ) considered in their simulations, X i β is approximately distributed as normal for any given value of X i . Therefore, it is reasonable to consider that X i β N ( M , g p ) . Using these results, we can select the value of M and g in G-prior distribution so that the induced distribution of μ ( X i ) = h 1 ( X i β ) matches the marginal prior distribution μ I G ( a μ , b μ ) . We minimize the Kullback–Leibler divergence, specifically D K L ( P Q ) , since this measures how well the distribution induced by our model (P) approximates the prior information distribution (Q). Thus, we have M = E h ( μ ) and g = V h ( μ ) / p . In our regression model, we considered h ( . ) as the log link function; hence,
M = ψ ( a μ ) + log ( b μ ) and g = 1 p ψ ( 1 ) ( a μ )
where ψ ( . ) and ψ ( 1 ) ( . ) are digamma and trigamma function, respectively. When the values for a μ and b μ are not available, we use a μ = b μ = 1 as the defaults, yielding relatively weak prior information on the location of μ . It is worth noting that, under standard conditions, G-priors are known to yield consistent Bayesian estimators [23,24].

2.2.2. MCMC Algorithm

Consider the posterior distribution by using flat-normal prior and G-prior distribution in (6), which is analytically intractable. Monte Carlo Chain Markov (MCMC) simulation methods, such as the Gibbs sampler and Metropolis–Hastings algorithm, are utilized to obtain a sample from the posterior distribution [25]. To implement the Metropolis–Hastings algorithm, the following steps are taken:
  • Start with any point β ( 0 ) and stage indicator k = 0 .
  • Generate β ˜ according to the transitional kernel K ( β ˜ , β ( k ) ) = N p ( β ( k ) , Σ ˜ ) , where Σ ˜ is a known symmetric positive defined matrix.
  • Accept β ˜ as β ( k + 1 ) with the following probability:
    min 1 , π ( β ˜ | D ) π ( β ( k ) | D ) .
  • By increasing the stage indicator, repeat steps (1) to (3) until the process reaches a stationary distribution.
The computational program is available upon request from the authors.

2.3. Model Selection Criteria

There are multiple techniques available to compare and select the best-fitting model among several competing models for a given dataset. One commonly used technique in applied research is the conditional predictive ordinate (CPO) statistic. To learn more about the CPO and its applications in model selection, see [26,27]. Suppose that D is the full data, D ( i ) for i = 1 , 2 , , n denotes the data with the ith observation deleted, and the posterior distribution based on D ( i ) is denoted by π β | D ( i ) . For the ith observation, C P O i is defined as follows:
C P O i = β π ( β | D ) f ( y i | β ) d β 1 .
The low CPO values indicate poor model fit, but a closed-form CPO is unavailable for the proposed model. However, a Monte Carlo estimate of the C P O i can be obtained by using a single MCMC sample { β ( 1 ) , β ( 2 ) , , β ( T ) } from the posterior distribution π ( β | D ) . Thus, the C P O i can be approximated by
C P O ^ i = 1 T k = 1 T 1 f ( y i | β ( k ) ) 1 .
The statistic for model comparison is the log-marginal pseudo-likelihood (LMPL) defined as follows:
L M P L = i = 1 n log ( C P O ^ i ) .
Therefore, the largest value of L M P L indicates that the data is well-fitted by the model under consideration.
The second criterion which is proposed by [28] is called the deviance information criterion (DIC). Based on MCMC samples, the DIC can be estimated as follows:
D I C ^ = 2 2 1 T k = 1 T log L ( D | β ( k ) ) + log L ( D | β ¯ ) ,
where β ¯ is the mean of MCMC samples. The model with the lowest D I C ^ value is considered as the best-fitted model.
The final two criteria considered here are the expected Akaike information criterion (EAIC) by [29] and the expected Bayesian information criterion (EBIC) by [30]. These criteria can also be estimated using the following:
E A I C ^ = 2 k = 1 T log L ( D | β ( k ) ) + 2 p ,
and
E B I C ^ = 2 k = 1 T log L ( D | β ( k ) ) + p log ( n ) ,
where p is the number of model parameters. The model that exhibits the lowest value of these criteria, similar to the DIC, is considered to be a better fit for the data.

3. Results

3.1. Simulation Study

In this section, we conduct a simulation study to illustrate the implementation of the proposed regression methodology.
The model we consider here is given by
μ i = exp β 1 x i 1 + β 2 x i 2 + + β p x i p , i = 1 , 2 , , n .
Here, we consider the model with intercept; thus, we set x i 1 = 1 . The observations for the covariates, i.e., x i 2 , , x i p , are generated from standard normal distribution. The real value of the parameters in model (13) is considered as follows:
β = ( 0 , 0.5 , 1 , 1 , , 1 p 2 ) .
In summary, we obtained the response observations from B e l l W 0 ( μ i ) . In order to evaluate the effectiveness of our proposed method, we tested it with different sample sizes n = 50 , 100 , 200 to represent low, medium, and high sample sizes, and with different numbers of covariates p = 3 , 6 . For the prior distributions, we consider flat-normal distribution with τ = 10 2 and G-prior defined in (6) with hyperparameters obtained from (7) by setting a μ = b μ = 1 .
We generate two parallel independent MCMC runs of size T = 50,000 for each posterior distribution and discard the 10,000 first generated sample as the burn-in to eliminate the impact of initial values. For computing the posterior estimates, we used every 20th sample to reduce the autocorrelations of the generated chains and yield better convergence results. In the Metropolis–Hastings algorithm, the proposal covariance was set equal to the covariance matrix of the maximum likelihood estimator. The convergence of the MCMC chains was monitored using the trace and autocorrelation function (ACF), Heidelberger–Welch, and Gelman–Rubin convergence diagnostics [31]. Additionally, we performed a small sensitivity study to evaluate the robustness of the model with respect to the choice of hyperparameters in the prior distributions by testing different values of τ and M for prior distributions.
The summaries of the parameters in the posterior distribution exhibit minimal differences and do not impact the results presented in Table 1 and Table 2. These tables display the results of posterior inference for both prior distributions. These tables show the estimated values using the squared loss function, posterior standard deviations (PSD), and 95 % highest posterior density (HPD) intervals. The results show that as the sample size grows, both the posterior standard deviations (PSDs) and the widths of the 95 % HPD intervals become smaller. Across all cases, the G-prior gives tighter intervals and lower PSDs than the flat-normal prior, which means that it provides more precise estimates. In addition, the acceptance rate of the algorithm stayed high ( 90 % to 95 % ) for both priors, indicating that the sampling procedure worked efficiently.
Table 3 presents the mean squared errors (MSEs) and the mean absolute errors (MAEs) for the estimates found in Table 1 and Table 2. Table 3 shows that the MSE and MAE of the G-prior are consistently lower than those of the flat-normal prior, except in the case p = 6 and n = 200 . This suggests that the G-prior generally produces more accurate and stable parameter estimates, with particularly clear benefits when the sample size is small or moderate. For both prior choices, we also observe that increasing the number of covariates leads to higher MSE and MAE values.

3.2. Application

In this section, we apply our methodology to a mine fracture dataset that was primarily analyzed by [32]. The dataset includes four variables: the thickness of the inner burden in feet ( x 1 ), the percentage of extraction from the previously mined lower seam ( x 2 ), the height of the lower seam ( x 3 ), and the time since the mine has been opened ( x 4 ). The number of fractures in the mine is denoted by y, and there are 44 observations available for each variable.
We examine three count regression models: the Bell regression model (BRM), the Poisson regression model (PRM), and the negative binomial regression model (NBRM). To begin, we assess whether the response variable follows each distribution using a chi-square goodness-of-fit test at the 95 % confidence level. The results, presented in Table 4, indicate that the Poisson distribution does not provide an adequate fit for this dataset, despite its previous use in similar analyses by [32,33]. In contrast, both the Bell and negative binomial distributions demonstrate a good fit, highlighting their suitability for modeling this data.
As in the Simulation Section, MCMC runs of size 50,000 were generated for each posterior distribution of BRM, PRM and NBRM under the G-prior distribution. To reduce the autocorrelations of the generated chains and obtain better convergence results, every 20th sample was used after discarding the first 10,000 generated samples as burn-in. The posterior means, medians, PSDs, and 95 % HPD intervals for both regression models are presented in Table 5. The results indicate that the BRM provides estimations with the lowest PSD and narrowest HPD intervals. Moreover, the HPD intervals of the BRMs indicate that x 1 and x 3 are not significant while only x 2 is significant based on the HPD intervals of PRM and NBRM.
We compared the BRM, PRM, and NBRM using the LMPL, DIC, EAIC, and EBIC criteria in Table 6. The results show that the BRM consistently performs best across all measures. In the real data application (Table 5 and Table 6), the BRM with a G-prior not only yields narrower HPD intervals but also achieves stronger model selection scores. While the NBRM offers an improvement over the PRM, the BRM still provides the most reliable fit, even with a small sample size. This highlights the practical usefulness of the Bayesian BRM for applied researchers working with limited or overdispersed count data.

4. Discussion and Conclusions

This paper has presented a comprehensive Bayesian framework for Bell regression models (BRMs). To the best of our knowledge, it is the first study to introduce G-priors in this context, underscoring the critical importance of prior choice when modeling overdispersed count data. Alongside the G-prior, we considered a flat-normal prior, with inference carried out using a tailored MCMC algorithm. Our simulation study demonstrated that the G-prior consistently yields more precise parameter estimates, as evidenced by smaller posterior standard deviations, narrower HPD intervals, and lower MSE and MAE compared to the flat-normal prior. In the real-data application, the BRM with a G-prior also outperformed both Poisson and negative binomial regression models across several Bayesian model selection criteria, including the LMPL, DIC, EAIC and EBIC variants. These results highlight not only the methodological contribution but also the practical utility of the proposed approach. Unlike earlier studies that focused primarily on frequentist estimation, this work provides a Bayesian alternative that performs well even with modest sample sizes. Although a full asymptotic analysis is beyond the scope of this paper, our empirical findings align with the established theoretical advantages of G-priors in the literature [23,24]. Future work could involve a detailed examination of the asymptotic properties and an extension of the method to scenarios such as zero-inflated BRMs or the use of other informative prior families to further enhance the model’s flexibility and performance.

Author Contributions

Conceptualization, H.B.; methodology, A.M.I.A. and H.B.; software, A.M.I.A. and H.B.; validation, A.M.I.A.; formal analysis, A.M.I.A.; investigation, A.M.I.A.; resources, H.B.; data curation, A.M.I.A.; writing—original draft preparation, A.M.I.A.; writing—review and editing, H.B.; visualization, A.M.I.A.; supervision, H.B.; project administration, H.B. All authors have read and agreed to the published version of the manuscript.

Funding

The authors received no specific funding for this work.

Data Availability Statement

The data supporting this paper are from previously reported studies and datasets, which have been cited.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Montesinos-López, O.A.; Montesinos-López, A.; Pérez-Rodríguez, P. Genomic Prediction Models for Count Data. J. Agric. Biol. Environ. Stat. 2015, 20, 533–554. [Google Scholar] [CrossRef]
  2. Du, J.; Park, Y.T.; Theera-Ampornpunt, N.; McCullough, J.S.; Speedie, S.M. The use of count data models in biomedical informatics evaluation research. J. Am. Med. Inform. Assoc. 2015, 19, 39–44. [Google Scholar] [CrossRef] [PubMed]
  3. Cameron, A.C.; Trivedi, P.K. Regression Analysis of Count Data, 2nd ed.; Cambridge University Press: New York, NY, USA, 2013. [Google Scholar]
  4. Frees, E.W. Regression Modeling with Actuarial and Financial Applications; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
  5. Lee, D.; Neocleous, T. Bayesian quantile regression for count data with application to environmental epidemiology. J. R. Stat. Soc. C Appl. Stat. 2010, 59, 905–920. [Google Scholar] [CrossRef]
  6. Castellares, F.; Ferrari, S.L.P.; Lemonte, A.J. On the Bell distribution and its associated regression model for count data. Appl. Math. Model. 2018, 56, 172–185. [Google Scholar] [CrossRef]
  7. Abduljabbar, L.A.; Algamal, Z.Y. Jackknifed K–L estimator in Bell regression model. Math. Stat. Eng. Appl. 2022, 71, 267–278. [Google Scholar]
  8. Algamal, Z.Y.; Lukman, A.; Golam, B.M.K.; Taofik, A. Modified Jackknifed Ridge Estimator in Bell Regression Model: Theory, Simulation and Applications. Iraqi J. Comput. Sci. Math. 2023, 4, 146–154. [Google Scholar] [CrossRef]
  9. Amin, M.; Akram, M.N.; Majid, A. On the estimation of Bell regression model using ridge estimator. Commun. Stat. Simul. Comput. 2021, 52, 854–867. [Google Scholar] [CrossRef]
  10. Bulut, Y.M.; Lukman, A.F.; Işilar, M.; Adewuyi, E.T.; Algamal, Z.Y. Modified ridge estimator in the Bell regression model. J. Inverse Ill-Posed Probl. 2024, 32, 1081–1091. [Google Scholar] [CrossRef]
  11. Ertan, E.; Algamal, Z.Y.; Erkoç, A.; Ulaş Akay, K. A new improvement Liu-type estimator for the Bell regression model. Commun. Stat. B Simul. Comput. 2023, 54, 603–614. [Google Scholar] [CrossRef]
  12. Majid, A.; Amin, M.; Akram, M.N. On the Liu estimation of Bell regression model in the presence of multicollinearity. J. Stat. Comput. Simul. 2021, 92, 262–282. [Google Scholar] [CrossRef]
  13. Shewaa, G.A.; Ugwuowo, F.I. Combating the Multicollinearity in Bell Regression Model: Simulation and Application. J. Niger. Soc. Phys. Sci. 2022, 4, 713. [Google Scholar] [CrossRef]
  14. Algamal, Z.Y.; Lukman, A.F.; Abonazel, M.R.; Awwad, F.A. Performance of the Ridge and Liu Estimators in the zero-inflated Bell Regression Model. J. Math. 2022, 2022, 9503460. [Google Scholar] [CrossRef]
  15. Lemonte, A.J.; Moreno-Arenas, G.; Castellares, F. Zero-inflated Bell regression models for count data. J. Appl. Stat. 2020, 47, 265–286. [Google Scholar] [CrossRef]
  16. Seifollahi, S.; Bevrani, H.; Algamal, Z.Y. Improved estimators in Bell regression model with application. J. Stat. Comput. Simul. 2024, 94, 2710–2726. [Google Scholar] [CrossRef]
  17. Seifollahi, S.; Bevrani, H.; Algamal, Z.Y. Shrinkage estimators in zero-inflated Bell regression model with application. J. Stat. Theory Pract. 2025, 19, 1. [Google Scholar] [CrossRef]
  18. Imran Alhaseeni, A.M.; Bevrani, H. Bayesian Bell regression model for fitting of overdispersed count data with application. arXiv 2024, arXiv:2403.07067. [Google Scholar] [CrossRef]
  19. Bell, E.T. Exponential numbers. Am. Math. Mon. 1934, 41, 419. [Google Scholar] [CrossRef]
  20. Bell, E.T. Exponential polynomials. Ann. Math. 1934, 35, 258. [Google Scholar] [CrossRef]
  21. Corless, R.M.; Gonnet, G.H.; Hare, D.E.; Jeffrey, D.J.; Knuth, D.E. On the Lambert W function. Adv. Comput. Math. 1996, 5, 329–359. [Google Scholar] [CrossRef]
  22. Zhou, H.; Huang, X. Bayesian beta regression for bounded responses with unknown supports. Comput. Stat. Data Anal. 2022, 167, 107345. [Google Scholar] [CrossRef]
  23. Zellner, A. On assessing prior distributions and Bayesian regression analysis with g-prior distributions. In Bayesian Inference and Decision Techniques: Essays in Honor of Bruno de Finetti; North-Holland/Elsevier: Amsterdam, The Netherlands, 1986; pp. 233–243. [Google Scholar]
  24. Hanson, T.E.; Branscum, A.J.; Johnson, W.O. Informative g-priors for logistic regression. Bayesian Anal. 2014, 9, 597–612. [Google Scholar] [CrossRef]
  25. Gamerman, D.; Lopes, H.F. Markov Chain Monte Carlo: Stochastic Simulation for Bayesian Inference; CRC Press: Boca Raton, FL, USA, 2006. [Google Scholar]
  26. Geisser, S.; Eddy, W. A predictive approach to model selection. J. Am. Stat. Assoc. 1979, 74, 153–160. [Google Scholar] [CrossRef]
  27. Gelfand, A.; Dey, D.; Chang, H. Model Determination Using Predictive Distributions with Implementation via Sampling Based Methods (with Discussion); Bernardo, J.M., Berger, J.O., Dawid, A.P., Smith, A.F.M., Eds.; Oxford University Press: Oxford, UK, 1992; Volume 1, pp. 7–167. [Google Scholar] [CrossRef]
  28. Spiegelhalter, D.J.; Best, N.G.; Carlin, B.P.; Van der Linde, A. Bayesian measures of model complexity and fit. J. R. Stat. Soc. B Stat. Methodol. 2002, 64, 583–639. Available online: http://www.jstor.org/stable/3088806 (accessed on 23 October 2002). [CrossRef]
  29. Brooks, S.P.; Smith, J.; Vehtari, A.; Plummer, M.; Stone, M.; Robert, C.P.; Titterington, D.M.; Nelder, J.A.; Atkinson, A.; Dawid, A.P.; et al. Discussion on the paper by Spiegelhalter, Best, Carlin, and van der Linde. J. R. Stat. Soc. B Stat. Methodol. 2002, 64, 616–639. [Google Scholar]
  30. Carlin, B.; Louis, T. Bayes and Empirical Bayes Methods for Data Analysis, 2nd ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2001. [Google Scholar]
  31. Chen, M.; Shao, Q.; Ibrahim, J. Monte Carlo Methods in Bayesian Computation; Springer: New York, NY, USA, 2000. [Google Scholar]
  32. Myers, R.H.; Montgomery, D.C.; Vining, G.G.; Robinson, T.J. Generalized Linear Models: With Applications in Engineering and the Sciences; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  33. Lukman, A.F.; Arashi, M.; Prokaj, V. Robust biased estimators for Poisson regression model: Simulation and applications. Concurr. Comput. Pract. Exp. 2023, 35, e7594. [Google Scholar] [CrossRef]
Table 1. Bayesian inference of model (13) based on prior distributions when p = 3 .
Table 1. Bayesian inference of model (13) based on prior distributions when p = 3 .
G-PriorFlat Normal
95% HPD 95% HPD
True Value Estimate PSD Lower Upper Estimate PSD Lower Upper
n = 50
β 1 0−0.09840.2022−0.47950.3073−0.10880.2135−0.52450.3017
β 2 −0.5−0.47430.1830−0.8319−0.0946−0.49900.1874−0.8604−0.1053
β 3 11.03690.14400.75471.31251.06290.14820.77261.3460
n = 100
β 1 0−0.16090.1430−0.42540.1273−0.16360.1465−0.44490.1175
β 2 −0.5−0.58250.1168−0.8042−0.3377−0.59030.1172−0.8164−0.3505
β 3 11.02400.08760.84791.19001.03290.08910.85691.2067
n = 200
β 1 0−0.11710.1051−0.31330.0911−0.12020.1081−0.32030.0926
β 2 −0.5−0.49770.0817−0.6432−0.3273−0.50070.0822−0.6459−0.3300
β 3 11.04040.10670.83291.23421.05270.10910.83691.2478
Table 2. Bayesian inference of model (13) based on prior distributions when p = 6 .
Table 2. Bayesian inference of model (13) based on prior distributions when p = 6 .
G-PriorFlat Normal
95% HPD 95% HPD
True Value Estimate PSD Lower Upper Estimate PSD Lower Upper
n = 50
β 1 00.00450.2046−0.40250.3899−0.08010.2286−0.54060.3511
β 2 −0.5−0.55170.1238−0.8012−0.3143−0.57420.1282−0.8326−0.3259
β 3 10.93970.10270.72571.12820.97980.10890.75351.1779
β 4 10.82490.11670.58381.05480.87500.12020.63261.1170
β 5 10.92140.10460.71681.13180.92830.10870.70991.1381
β 6 11.08210.16150.76071.39631.15150.17700.80251.4978
n = 100
β 1 0−0.14250.1471−0.43460.1352−0.21180.1588−0.51300.0938
β 2 −0.5−0.44480.0803−0.6019−0.2862−0.44910.0810−0.6077−0.2882
β 3 11.03780.05710.91991.15071.05500.05850.93291.1725
β 4 10.83370.11120.60461.04050.87840.11470.64201.0970
β 5 10.86800.08660.70291.04040.90070.08890.73561.0770
β 6 11.07710.10610.86261.28121.12230.11320.88991.3387
n = 200
β 1 0−0.01520.0981−0.20780.1788−0.04060.1016−0.23210.1737
β 2 −0.5−0.45730.0542−0.5611−0.3551−0.46700.0546−0.5710−0.3645
β 3 10.87290.06740.74551.01970.88810.06840.75561.0359
β 4 10.97380.05770.86471.09380.98370.05960.86771.1038
β 5 11.05070.05520.94961.16001.05650.05620.95201.1654
β 6 10.93950.06580.81721.07270.95390.06690.83291.0912
Table 3. MSEs and MAEs for the estimated values reported in Table 1 and Table 2.
Table 3. MSEs and MAEs for the estimated values reported in Table 1 and Table 2.
MSEMAE
G-Prior Flat Normal G-Prior Flat Normal
np = 3 p = 6 p = 3 p = 6 p = 3 p = 6 p = 3 p = 6
500.00360.00870.00520.00930.05370.07540.05760.0871
1000.01110.01260.01200.01500.08910.10180.09560.1102
2000.00510.00420.00570.00350.05330.05370.05790.0507
Table 4. Goodness-of-fit test for the mine fracture dataset.
Table 4. Goodness-of-fit test for the mine fracture dataset.
CountObservedBellPoissonNegative Binomial
01010.1494.7446.640
179.16410.56710.752
288.27411.76710.173
386.2268.7367.342
444.2164.8644.475
≥575.9703.3214.618
χ 2 1.21612.5234.813
p-value 0.9430.0280.439
Table 5. Bayesian inference of models for the mine fracture dataset.
Table 5. Bayesian inference of models for the mine fracture dataset.
95% HPD
Model Parameter Mean Median PSD Lower Upper
Bell β 1 −3.5991−3.58641.0666−5.6350−1.6295
β 2 −0.0015−0.00150.0009−0.00320.0002
β 3 0.063100.06340.01270.03800.0879
β 4 −0.0032−0.00310.0055−0.01480.0069
β 5 −0.0323−0.03160.0164−0.0666−0.0032
Poisson β 1 −4.1262−4.05151.4028−7.0078−1.7307
β 2 −0.0015−0.00140.0011−0.00350.0006
β 3 0.06950.06890.01680.03860.1031
β 4 −0.0031−0.00270.0074−0.01740.0116
β 5 −0.0314−0.03040.0227−0.07640.0117
Negative Binomial β 1 −3.9884−3.96001.2110−6.6541−1.8616
β 2 −0.0014−0.00140.0010−0.00320.0005
β 3 0.06750.06740.01480.04140.0987
β 4 −0.0027−0.00240.0060−0.01390.0095
β 5 −0.0314−0.03150.0186−0.06650.0059
Table 6. Bayesian Criteria for the fitted models to the mine fracture dataset.
Table 6. Bayesian Criteria for the fitted models to the mine fracture dataset.
ModelLMPLDICEAICEBIC
Bell−72.4172144.4801149.3860158.3070
Poisson−78.1424156.8615161.5732170.4942
Negative Binomial−75.7465149.3917154.5123163.4332
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Imran Alhseeni, A.M.; Bevrani, H. Bayesian Bell Regression Model for Fitting of Overdispersed Count Data with Application. Stats 2025, 8, 95. https://doi.org/10.3390/stats8040095

AMA Style

Imran Alhseeni AM, Bevrani H. Bayesian Bell Regression Model for Fitting of Overdispersed Count Data with Application. Stats. 2025; 8(4):95. https://doi.org/10.3390/stats8040095

Chicago/Turabian Style

Imran Alhseeni, Ameer Musa, and Hossein Bevrani. 2025. "Bayesian Bell Regression Model for Fitting of Overdispersed Count Data with Application" Stats 8, no. 4: 95. https://doi.org/10.3390/stats8040095

APA Style

Imran Alhseeni, A. M., & Bevrani, H. (2025). Bayesian Bell Regression Model for Fitting of Overdispersed Count Data with Application. Stats, 8(4), 95. https://doi.org/10.3390/stats8040095

Article Metrics

Back to TopTop