Next Article in Journal
Signatures of Extreme Events in Cumulative Entropic Spectrum
Previous Article in Journal
Identification of Important Nodes Based on Local Effective Distance-Integrated Gravity Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Discrete Analogue of the Continuous Muth Distribution for Over-Dispersed Data: Properties, Estimation Techniques, and Application

by
Howaida Elsayed
1,* and
Mohamed Hussein
1,2
1
Department of Business Administration, College of Business, King Khalid University, Abha 61421, Saudi Arabia
2
Department of Mathematics and Computer Science, Alexandria University, Alexandria 21544, Egypt
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(4), 409; https://doi.org/10.3390/e27040409
Submission received: 6 March 2025 / Revised: 30 March 2025 / Accepted: 8 April 2025 / Published: 10 April 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
We present a new one-parameter discrete Muth (DsMuth) distribution, a flexible probability mass function designed for modeling count data, particularly over-dispersed data. The proposed distribution is derived through the survival discretization approach. Several of the proposed distribution’s characteristics and reliability measures are investigated, including the mean, variance, skewness, kurtosis, probability-generating function, moments, moment-generating function, mean residual life, quantile function, and entropy. Different estimation approaches, including maximum likelihood, moments, and proportion, are explored to identify unknown distribution parameters. The performance of these estimators is assessed through simulations under different parameter settings and sample sizes. Additionally, a real dataset is used to emphasize the significance of the proposed distribution compared to other available discrete probability distributions.

1. Introduction

Modeling count data plays an essential role in various fields, including medicine, public health, sociology, medicine, agriculture, epidemiology, and applied science. Various probability distributions have been established to model count data, particularly count data with over-dispersion. Nonetheless, we need to become more adaptable in terms of discrete distributions to effectively model count data sets with high over-dispersion. The Poisson distribution is a widely used model for count data. It is well known that in the Poisson distribution, the mean and variance are equal. This characteristic of the Poisson distribution poses challenges in modeling actual life datasets [1]. In practical applications, count data is often over-dispersed, meaning that the empirical variance exceeds the empirical mean. In this case, using the Poisson distribution for these types of data sets yields misspecification of the underlying probability distribution. This occurs because real-life count data often exhibit either over-dispersion, where the variance is greater than the mean, or under-dispersion, where the variance is lower than the mean. Negative binomial distribution is commonly used as the primary choice for modeling over-dispersed count data. Consequently, several discrete distributions have been developed based on well-known continuous models used in reliability analysis, failure time studies, and related fields.
Discretizing continuous distributions can be achieved through various approaches, including survival discretization, mixed-Poisson models, and infinite series methods. Among these, the survival discretization approach is the most commonly used [2]. The discretization of continuous probability distributions has attracted considerable attention in recent years. Various discrete distributions have been proposed and analyzed in the literature based on the discretization of the survival function. For instance, various discrete distributions have been introduced and studied, including the discrete Rayleigh [3], discrete half-normal [4], discrete Burr and Pareto [5], discrete exponential [6], discrete inverse-Weibull [7], discrete Power Ailamujia [8], discrete Rayleigh generator [9], new generalization of the geometric [10], discrete Lindley [11], discrete Burr-XII [12], discrete Burr III [13], two-parameter discrete Lindley [14], discrete log-logistic [15], new discrete distribution [16], Poisson Ailamujia [17], discrete alpha power inverse Lomax [18], Uniform Poisson and Ailamujia [19], discrete Half-Logistic [20], discrete Marshall-Olkin Weibull [21], discrete Gompertz-G family [22], Discrete Bilal [23], discrete Burr-Hatke [24], three-parameter discrete Lindley [25], new discrete Lindley [26], exponentiated discrete Lindley [27], discrete generalized Lindley [28], discrete Gumble [29], discrete inverted Topp-Leone [30], discrete Ramos-Louzada distribution [31], generalized exponential type II [32], and discrete inverse-Rayleigh [33] distributions.
Although numerous distributions exist in the literature for analyzing lifetime data, there remains a need for more flexible and adaptable distributions that can effectively model data under different conditions. The continuous Muth distribution, introduced by [34], is a one-parameter lifetime distribution that has proven useful in modeling various reliability phenomena. It may be valuable to explore the development of a discrete version of the Muth distribution, enabling its implementation in modeling discrete data. An essential purpose of this paper is to propose a new discrete analog based on the continuous Muth distribution, referred to as the discrete Muth (DsMuth) distribution. As a fundamental mathematical definition, the proposed distribution is characterized by the subsequent probability density function (PDF) and cumulative distribution function (CDF), respectively.
f ( x ; α ) = ( e α x α ) e α x 1 α ( e α x 1 ) , x 0 and α ( 0,1 ]
F ( x ; α ) = P ( X < x ) = 1 e α x 1 α ( e α x 1 ) , x > 0 and α ( 0,1 ]
The corresponding survival function (SF) and hazard rate function (HRF) related to Equation (2) can be formulated as:
S ( x ; α ) = P ( X x ) = e α x 1 α ( e α x 1 )
h ( x ; α ) = ( e α x α ) e α x 1 α ( e α x 1 ) e α x 1 α ( e α x 1 )
In this article, we propose a discrete from of the Muth distribution using the survival discretization method, referred to as the discrete Muth distribution. The DsMuth distribution is a feasible alternative for modeling the over-dispersed count data set. A comprehensive study of the statistical characteristics of the DsMuth distribution is conducted. Additionally, different simulation studies are presented to estimate the unknown parameter of the DsMuth distribution and study the behavior of the maximum likelihood estimation (MLE), moments estimation (MOM), and proportion estimation (PE) methods. The suggested distribution demonstrates superior effectiveness in modeling over-dispersed count data compared to existing competing distributions.
The article is structured as outlined below: Section 2 is dedicated to deriving the discrete version of the Muth distribution, accompanied by graphical representations of its probability mass function (PMF) and hazard rate function (HRF). Section 3 presents the derivation of various mathematical characteristics of the proposed distribution. Section 4 introduces the DsMuth distribution’s entropy, which measures its uncertainty and randomness. Section 5 introduces quantiles of a discrete random variable. Section 6 presents the parameter estimation employing the maximum likelihood method, the method of moments, and the proportion estimation method. Additionally, Section 7 presents simulation experiments to assess the effectiveness of the estimation methods under various sample sizes and parameter values. Section 8 utilizes a real dataset to validate the applicability of the DsMuth distribution in modeling count data and to demonstrate the flexibility of the proposed distribution. Finally, Section 9 provides a summary of the results and presents the conclusions.

2. The DsMuth Distribution

Let X be a continuous random variable with survival function S ( x ) = P ( X x ) . Using the survival discretization approach, we define a discrete random variable Y = X (i.e., the greatest integer less than or equal to X). Then, the probability mass function (pmf) of the discrete random variable Y follows as:
P ( y ) = P ( y X < y + 1 ) = P ( X y ) P ( X y + 1 ) = S ( y ) S ( y + 1 ) , y = 0 , 1 , 2 ,
By employing the survival discretization approach in Equation (5), a random variable X is said to follow the DsMuth distribution. If the survival function of the Muth distribution with parameter λ = e α is expressed as:
S ( x ; λ ) = λ x exp 1 ln λ ( λ x 1 ) , x = 0 , 1 , 2 ,
The corresponding probability mass function (pmf) and the cumulative distribution function (CDF) can be represented as:
p ( x ; α ) = e α x 1 α ( e α x 1 ) e α ( x + 1 ) 1 α ( e α ( x + 1 ) 1 ) , x = 0 , 1 , 2 ,
By performing the variable transformation λ = e α (i.e., α = ln λ ), we obtain
p ( x ) = λ x exp 1 ln λ ( λ x 1 ) λ x + 1 exp 1 ln λ ( λ x + 1 1 ) , x = 0 , 1 , 2 ,
The cumulative distribution function (CDF) of the DsMuth distribution is defined based on the survival function as follows:
F ( x ; λ ) = P ( X x ; λ ) = 1 S ( x ; λ ) + P ( X = x ; λ ) = 1 e α ( x + 1 ) 1 α e α ( x + 1 ) 1 F ( x ; λ ) = 1 λ x + 1 exp 1 ln λ λ x + 1 1 , x = 0 , 1 , 2 ,
where λ = e α , 1 < λ e , and x = 0 , 1 , 2 , 3 ,
The survival function (SF) of the DsMuth model can be written as
S ( x ; α ) = λ x + 1 exp 1 ln λ λ x + 1 1 .
The probability mass function (PMF) of the DsMuth distribution, as stated in Equation (7), is log-concave for all values of λ ( 1 , e ] . Figure 1, Figure 2 and Figure 3 illustrate various possible shapes of the PMF of the DsMuth distribution for various values of the parameter λ . It has been noted that the PMF of DsMuth is suitable for modeling positively skewed data with a unimodal shape for any value of λ .
The DsMuth distribution’s hazard rate function (HRF), reversed hazard rate (RH), and second failure rate ( r * h ) can be formulated as follows:
h ( x ; α ) = p ( x ; α ) 1 F ( x 1 ; α ) = 1 λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) λ x exp 1 ln λ ( λ x 1 )
r h ( x ; α ) = p ( x ; α ) F ( x ; α ) = λ x exp 1 ln λ ( λ x 1 ) λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) 1 λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 )
r * h ( x ; α ) = log S ( x ; α ) S ( x + 1 ; α ) = log λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) λ ( x + 2 ) exp 1 ln λ ( λ ( x + 2 ) 1 ) , x = 0 , 1 , 2 ,
Figure 4 and Figure 5 illustrate that the hazard rate function (HRF) of the DsMuth distribution is consistently increasing with respect to x for all values of the parameter λ .

3. Some Properties of the DsMuth Distribution

In this section, the probability generating function (pgf), the rth moment, and the quantile function of the DsMuth distribution are examined.

3.1. Probability Generating Function (PGF)

Let X be a discrete random variable defined on the set of non-negative integers { 0 , 1 , 2 , } . The probability generating function G X ( s ) of the DsMuth distribution is obtained as:
G X ( s ) = E ( s X ) = x = 0 s x P ( X = x ) = x = 0 s x [ S ( x ) S ( x + 1 ) ] . G X ( s ) = 1 + ( s 1 ) x = 1 s ( x 1 ) S ( x ) G X ( s ) = 1 + ( s 1 ) x = 1 s ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
The factorial moments can be derived from the probability generating function (PGF) of the DsMuth distribution. To find the factorial moments using the PGF, we use the relationship between the derivatives of the PGF and the factorial moments. Specifically, the n-th factorial moment μ n is given by:
μ n = d n d s n G X ( s ) s = 1
The first four factorial moments, denoted as ( μ 1 , μ 2 , μ 3 , μ 4 ) , can be determined utilizing the factorial moments as:
  • First factorial moment of the DsMuth Distribution
The first factorial moment (mean) of the DsMuth distribution is obtained by differentiating the probability generating function G X ( s ) with respect to s and then setting s = 1 :
E [ X ] = G X ( s ) | s = 1 = d d s G X ( s ) | s = 1 . G X ( s ) = x = 1 s ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) + ( s 1 ) x = 1 ( x 1 ) s ( x 2 ) λ x exp 1 ln λ ( λ x 1 ) G X ( 1 ) = E [ X ] = x = 1 λ x exp 1 ln λ ( λ x 1 )
  • Second factorial moment: Differentiating G X ( s ) once more with respect to and setting s = 1 , we obtain
G X ( s ) = x = 1 ( x 1 ) s ( x 2 ) λ x exp 1 ln λ ( λ x 1 ) + x = 1 ( x 1 ) s ( x 2 ) λ x exp 1 ln λ ( λ x 1 ) + ( s 1 ) x = 1 ( x 1 ) ( x 2 ) s ( x 3 ) λ x exp 1 ln λ ( λ x 1 ) G X ( 1 ) = μ 2 = 2 x = 1 ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
  • Third factorial moment: Differentiating G X ( s ) once more with respect to and setting s = 1 , we have
G X ( 1 ) = μ 3 = 3 x = 1 ( x 1 ) ( x 2 ) λ x exp 1 ln λ ( λ x 1 ) .
  • Fourth factorial moment: Differentiating G X ( s ) once more with respect to and putting s = 1 , we obtain
G X ( 1 ) = μ 4 = 4 x = 1 ( x 1 ) ( x 2 ) ( x 3 ) λ x exp 1 ln λ ( λ x 1 ) .
By substituting s by e s ,( i.e., s = e s ), in Equation (13), the moment generating function (MGF) can be expressed as:
M X ( e s ) = 1 + ( e s 1 ) x = 1 ( e s ) ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
The first four moments around the origin of the DsMuth distribution can be computed using the moment generating function as follows:
Differentiating the moment generating function (MGF) M X ( e s ) with respect to s and setting s = 0 , we obtain the mean of the DsMuth distribution:
E ( X ) = d M X ( e s ) d s = e s x = 1 ( e s ) ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) + ( e s 1 ) x = 1 ( x 1 ) ( e s ) ( x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
By setting s = 0 , we obtain the expected value:
E ( X ) = d M X ( e s ) d s s = 0 = x = 1 λ x exp 1 ln λ ( λ x 1 ) .
The second moment is derived from the second derivative of M X ( e s ) :
E ( X 2 ) = d 2 M X ( e s ) d s 2 s = 0 = x = 1 ( 2 x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
The third moment is obtained by differentiating M X ( e s ) three times:
E ( X 3 ) = d 3 M X ( e s ) d s 3 s = 0 = x = 1 ( 3 x 2 3 x + 1 ) λ x exp 1 ln λ ( λ x 1 ) .
Finally, the fourth moment follows from the fourth derivative:
E ( X 4 ) = d 4 M X ( e s ) d s 4 s = 0 = x = 1 ( 4 x 3 6 x 2 + 4 x 1 ) λ x exp 1 ln λ ( λ x 1 ) .
Based on the derived moments, the variance, skewness, and kurtosis can be expressed as follows:
σ 2 = E ( X 2 ) [ E ( X ) ] 2 .
Skewness = E ( X 3 ) 3 E ( X 2 ) E ( X ) + 2 [ E ( X ) ] 3 [ Var ( X ) ] 3 / 2 .
Kurtosis = E ( X 4 ) 4 E ( X 3 ) E ( X ) + 6 E ( X 2 ) [ E ( X ) ] 2 3 [ E ( X ) ] 4 [ Var ( X ) ] 2 .
The dispersion index (DI) is described as the ratio of variance to the mean:
DI = Variance ( X ) E ( X ) .
  • If DI < 1, the model is appropriate for under-dispersed data.
  • If DI > 1, the model is appropriate for over-dispersed data.
Table 1 provides quantitative values for the mean, variance, dispersion index (DI), skewness, and kurtosis of the DsMuth distribution for various values of the model parameter λ . Based on Table 1, the following key observations can be observed:
  • As the parameter λ increases, the mean and variance of the DsMuth distribution gradually decline.
  • The dispersion index (DI) rises as λ increases, implying that the distribution becomes more dispersed with higher values of λ .
  • Skewness reduces as λ increases, implying that the distribution becomes less positively skewed for larger values of λ .
  • Kurtosis declines with increasing λ , meaning the distribution becomes less peaked (less leptokurtic) as λ grows.
Table 1. Descriptive statistics for the DsMuth distribution.
Table 1. Descriptive statistics for the DsMuth distribution.
Measure λ
1.1 1.2 1.3 1.4 1.6 1.8 2.0 2.2 2.4
Mean0.5750.5680.5620.5560.5450.5350.5260.5170.510
Variance0.7720.6710.5970.5400.4560.3980.3560.3240.300
Dispersion Index (DI)0.7440.8460.9411.0301.1941.3421.4761.5961.700
Skewness1.8061.5321.3371.1870.9640.7990.6640.5390.418
Kurtosis13.1411.23910.0419.1767.9166.9546.1105.2464.330

3.2. Mean Residual Life (MRL)

The Mean Residual Life function is a crucial reliability characteristic used in modeling burn-in and maintenance policies. It provides the expected remaining lifetime given that a system or component has already survived until a specific point in time.
For a discrete random variable X with cumulative distribution function (CDF) F ( x ) , the mean residual life function is described as:
MRL = ε ( i ) = E ( X i X i ) = 1 1 F ( i 1 , λ ) j = i + 1 1 F ( j 1 , λ ) .
Accordingly, the Mean Residual Life (MRL) function of the DsMuth model simplifies to:
MRL = 1 1 F ( i 1 , λ ) j = i + 1 1 F ( j 1 , λ ) .
Further, substituting F ( x ) from the DsMuth distribution:
MRL = 1 λ i exp 1 ln λ ( λ i 1 ) j = i + 1 λ j exp 1 ln λ ( λ j 1 ) .

4. Entropy

Entropy measures the fluctuation in the uncertainty in physical systems. The Shannon entropy is a well-known entropy measurement [35]. In this section, we obtain the discrete Shannon entropy of a random variable. The entropy H ( X ) of a discrete random variable X, with its probability mass function (pmf), is given by Equation (7):
H ( X ) = x = 0 p ( x ) log p ( x )
The entropy of the DsMuth distribution can be computed using Mathcad software© var.15. The following formula is derived as follows:
H ( X ) = x = 0 p ( x ) log λ x exp 1 ln λ ( λ x 1 ) λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) H ( X ) = log λ x = 0 x p ( x ) + 1 ln λ x = 0 ( λ x 1 ) p ( x ) + λ x = 0 1 1 ln λ ( λ ( x + 1 ) λ x ) p ( x ) .
See Appendix A for the proof.
Table 2 provides numerical values for the entropy of a DsMuth distribution under various parameter choices, computed using Mathcad software© version 15. Additionally, Figure 6 illustrates the relationship between entropy, H ( X ) , and the parameter values. Notably, the entropy function H ( X ) exhibits a monotonic decrease as the parameter λ varies within the interval λ ( 1 , 2.7 ] .

5. Quantiles of a Discrete Random Variable

Definition 1. 
The point y u is referred to as the u t h quantile of a discrete random variable Y if it satisfies the following conditions [36]:
P ( Y y u ) u and P ( Y y u ) 1 u .
This can be represented in terms of the cumulative distribution function (CDF) F ( Y ) as:
F ( y u 1 ) < u F ( y u ) .
Proposition 1. 
The u t h quantile of the DsMuth distribution can be obtained by:
y u = 1 α log α w 1 α ( 1 u ) exp 1 α , α ( 0 , 1 ]
where w ( x ) is the Lambert function and · is the ceiling function, which rounds a given number up to the nearest integer. In other words, for any real number x, the ceiling function returns the smallest integer that is greater than or equal to x.
See Appendix B for the proof.
Quartiles of the DsMuth Distribution: The quartiles of the DsMuth distribution can be calculated using the formulas below:
y 0.25 = 1 α log α w 1 α ( 1 0.25 ) exp 1 α
y 0.75 = 1 α log α w 1 α ( 1 0.75 ) exp 1 α
y 0.5 = 1 α log α w 1 α ( 1 0.5 ) exp 1 α
A random integer can be generated from the proposed model using the inverse transformation method. This method converts a uniformly distributed random variable into a sample from the target distribution. Let U be a random number selected from a uniform distribution on ( 0 , 1 ) . Then, a random number Y that follows the DsMuth( α ) distribution is generated using Equation (24).

6. Estimation Methods

In this section, three different estimation techniques are employed to estimate the unknown parameter of the DsMuth distribution. The estimation techniques considered are the Maximum Likelihood Estimation (MLE) method, the Moments Method (MOM), and the Proportion Estimation (PE) method.

6.1. Maximum Likelihood Estimation

Assume a random sample x 1 , x 2 , , x n from the DsMuth distribution. The log-likelihood function can be expressed as:
L = i = 1 n p ( x ¯ ; α ) L = i = 1 n λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 )
Taking the logarithm:
log L = log i = 1 n λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 ) l = i = 1 n log λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 )
By differentiating Equation (25) with respect to the parameter λ , we obtain the non-linear equation:
d l d λ = d d λ i = 1 n log λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 ) = 0 d l d α = i = 1 n exp 1 ln λ ( λ x i 1 ) x i λ x i 1 x i λ 2 x i 1 ln λ + λ x i ( λ x i 1 ) λ ( ln λ ) 2 λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 ) ( x i + 1 ) λ x i ( x i + 1 ) λ 2 x i 1 ln λ + λ ( x i + 1 ) ( λ ( x i + 1 ) 1 ) λ ( ln λ ) 2 λ x i exp 1 ln λ ( λ x i 1 ) λ ( x i + 1 ) exp 1 ln λ ( λ ( x i + 1 ) 1 ) = 0
The likelihood equation does not have a closed-form solution; however, the maximum likelihood estimate (MLE) of λ can be determined through numerical methods. By applying a numerical method to solve Equation (26), the estimators can be efficiently obtained using Mathcad software.

6.2. Method of Moment Estimation

Utilizing the method of moments (MOM) to estimate the parameter λ of the DsMuth distribution, we equate the sample mean to the theoretical population mean and solve for λ . This results in a non-linear equation that must be solved numerically. Thus, the estimator λ is obtained by solving the following equation with respect to λ :
x = 1 x λ x exp 1 ln λ ( λ x 1 ) = 1 n i = 1 n x i
Since Equation (27) does not have a closed-form solution, numerical methods are required to determine its value. Therefore, a symbolic computation tool like Mathcad software should be used to solve the equation numerically, utilizing observed data points x i , i = 1 , 2 , , n .

6.3. Method of Proportion Estimation

Assume a random sample x 1 , x 2 , , x n drawn from the DsMuth distribution, which is characterized by a single parameter. Therefore, an indicator function is defined as follows:
I ( x i ) = 1 , if x i = 0 0 , otherwise
Assume that W = i = 1 n I ( x i ) represents the total count of zero observations in the sample. Using Equations (8) and (28), the probability of obtaining a zero observation is given by:
P ( X 0 ) = W n
Based on this, the Proportion Estimation (PE) method is used to estimate the parameter λ . The estimate λ ^ is obtained by solving the following equation:
1 λ ^ exp 1 ln λ ^ ( λ ^ 1 ) W n = 0 .
Since W n is an unbiased and consistent empirical estimator of probability P ( X 0 ) , it follows that the estimator λ ^ also possesses consistency and biasedness in estimating λ . Further details on this can be found in [24,37].

7. Simulation Study

To verify the accuracy of the performance of the proposed estimators (MLE method, MME method, and PE method), we conducted a numerical simulation to assess the estimation based on the DsMuth model concerning the sample size n. The evaluation depends on the simulation procedure outlined below:
  • Generate 10,000 samples of size n = 50 , 100 , 150 , 200 , 300 , 500 from DsMuth( λ ), considering different values: DsMuth(1.1), DsMuth(1.2), DsMuth(1.7), DsMuth(1.8), DsMuth(2.2), and DsMuth(2.5), respectively. This simulation is performed using Mathcad Software.
  • A general expression to generate a random variable X from the DsMuth distribution is to first generate a value Y from the continuous distribution and then discretize this value to obtain X. Here, X = [Y], which represents the largest integer less than or equal to Y [38,39].
  • Compute the MLEs, MMEs, and PEs for the 10,000 samples, denoted as λ ^ i for i = 1 , 2 , , 10,000 .
  • Compute the bias (B), mean square errors (MSEs), and mean relative error (MRE) of λ by employing three methods, specifically MLE, MME, and PE, using the following formulas:
    Bias ( λ ) = 1 10,000 i = 1 10,000 ( λ ^ i λ )
    MSE ( λ ) = 1 10,000 i = 1 10,000 ( λ ^ i λ ) 2
    MRE ( λ ) = 1 10,000 i = 1 10 , 000 | λ ^ i λ | λ
The results obtained from the empirical analysis are presented in Table 3. The bias, MSE, and MRE of the parameter were computed using the Mathcad program with the three estimation methods. The following observations can be made from Table 3:
  • Improved Accuracy with Larger Sample Sizes:
    The estimates of λ get closer to their true values as the sample size n increases across all estimation methods. This demonstrates the asymptotic property of the estimators, meaning they improve as more data become available.
  • Bias Reduction:
    The bias of the parameter decreases toward zero as the sample size increases in all estimation methods. This indicates that the estimators are unbiased or asymptotically unbiased, ensuring greater accuracy in large samples.
  • Mean Squared Error (MSE) Decrease:
    The MSE values decrease as n increases for all estimation methods. This confirms the consistency of the estimators, indicating that as more data is used, the estimates become more precise and less variable.
  • Mean Relative Error (MRE) Decrease:
    The MRE also declines as n increases, further supporting the consistency of the estimators. This metric highlights how estimation errors become smaller in proportion to the true parameter value.
  • MLE as the Most Effective Method:
    Among the three estimation techniques, Maximum Likelihood Estimation (MLE) performs best. It consistently provides estimates with the lowest bias, smallest MSE, and smallest MRE compared to the Method of Moments (MME) and Proportion Estimation (PE) methods. This suggests that MLE is the most efficient method for estimating λ in the DsMuth distribution.
Table 3. Simulation results of DsMuth distribution for several parameter values.
Table 3. Simulation results of DsMuth distribution for several parameter values.
ParameternAEBiasMSEMRE
MLE MME PE MLE MME PE MLE MME PE MLE MME PE
λ = 1.1 501.0540.7730.931−0.046−0.327−0.1690.0420.3230.1470.1370.3890.238
1001.0540.8061.011−0.046−0.294−0.0890.0200.2970.0700.0990.3630.158
1501.0750.7951.027−0.025−0.305−0.0730.0100.2960.0560.0710.3670.151
2001.0790.8811.022−0.021−0.219−0.0780.0060.2030.0450.0600.2900.131
3001.0910.9051.057−0.009−0.195−0.0430.0040.1840.0270.0470.2760.112
5001.0980.9381.079−0.002−0.162−0.0210.0020.1290.0130.0340.2340.081
λ = 1.2 501.1530.8731.028−0.047−0.327−0.1720.0450.3220.1490.1430.3890.239
1001.1670.8991.087−0.033−0.301−0.1130.0170.2920.0850.0870.3640.180
1501.1800.9031.123−0.020−0.297−0.0770.0120.2740.0600.0760.3520.151
2001.1840.9591.138−0.016−0.241−0.0620.0080.2240.0400.0630.3070.131
3001.1890.9761.158−0.011−0.224−0.0420.0050.2000.0270.0490.2900.108
5001.1941.0141.172−0.006−0.186−0.0280.0030.1580.0170.0380.2570.089
λ = 1.7 501.6461.6551.753−0.054−0.0450.0530.1120.4390.1550.1310.3740.217
1001.6701.6371.717−0.030−0.0630.0170.0330.4010.1190.0830.3510.185
1501.6841.6091.727−0.016−0.0910.0270.0220.3750.1060.0710.3350.173
2001.6841.6311.704−0.016−0.0690.0040.0170.3570.0940.0620.3260.164
3001.7031.6531.697−0.003−0.047−0.0030.0100.3070.0770.0460.2940.145
5001.6981.6551.698−0.002−0.045−0.0020.0060.2500.0590.0380.2580.127
λ = 1.8 501.7131.7711.998−0.087−0.0290.1980.1990.4270.1950.1410.3410.193
1001.7551.8261.911−0.0450.0260.1110.0460.3870.1190.0890.3170.156
1501.7751.8231.891−0.0250.0230.0910.0260.3640.0950.0690.3010.141
2001.7791.7791.881−0.021−0.0210.0810.0200.3380.0920.0590.2900.140
3001.7841.7991.852−0.016−0.0010.0520.0130.2720.0630.0490.2480.114
5001.7131.7711.998−0.009−0.0000.0040.1990.4270.1950.1410.3410.193
λ = 2.2 501.9472.3562.673−0.2530.1560.4730.8450.7520.3370.2310.3610.213
1002.1322.2312.627−0.0680.0310.4270.1350.5730.2680.1040.3060.192
1502.1602.2332.591−0.0400.0330.3910.0680.5120.2150.0840.2840.176
2002.1802.2272.557−0.0200.0270.3570.0380.4280.1750.0660.2540.161
3002.1832.2222.547−0.0170.0220.3470.0250.4190.1560.0550.2570.156
5002.1852.1702.507−0.015−0.0300.3070.0140.2960.1130.0410.2080.138
λ = 2.5 502.1132.7333.214−0.3870.2330.7141.0490.7040.6300.2690.2980.290
1002.2952.6843.131−0.2050.1840.6310.4130.5960.4700.1550.2690.256
1502.4062.6343.108−0.0940.1340.6080.2010.4870.4250.1100.2430.247
2002.4452.6123.075−0.0550.1120.5750.0810.4050.3710.0840.2230.234
3002.4692.5683.063−0.0310.0680.5630.0410.3410.3520.0630.2030.229
5002.4812.5393.020−0.0190.0390.5200.0200.2540.2860.0450.1740.211

8. Empirical Study

This section demonstrates the importance and superiority of the DsMuth distribution based on application to a real dataset. We compare the fits of the DsMuth distribution with several competitive distributions that have one or two parameters, including the discrete Rayleigh (DsR) [3], discrete Inverse-Rayleigh (DsIR) [33], discrete Lindley (DsLi) [10], Poisson (Poi) [40], discrete Poisson-Lindley (PoiLi) [41], discrete Lindley-Two Parameter (DLi-II) [14], discrete Linear Failure Rate (DLFR) [42], discrete Inverse Weibull (DIW) [7], and discrete Log-logistic (DLog-L) [15] distributions, which are listed in Table 4. The comparison of the fitted models is made using several criteria, including the negative maximum log-likelihood (-log-Lik.), Akaike information criterion (AIC), Bayesian information criterion (BIC), Hannan-Quinn information criterion (HQIC), and Chi-square ( χ 2 ) with its corresponding p-value.
The dataset includes the count of carious teeth among the four deciduous molars, with a sample size of 100. Detailed information about the dataset can be found in Krishna and Pundir [5,24]. The maximum likelihood estimates, standard errors (Std-er), confidence intervals (C.I), and goodness-of-fit measures for the dataset are provided in Table 5.
According to Table 5, the DsMuth distribution proves to be the best choice for evaluating these data compared to other competitive distributions, as it has the lowest values for AIC, BIC, and HQIC, in addition to the highest p-value among all the tested distributions. Figure 7 and Figure 8 illustrate the fitted PMF plots for all tested distributions, further supporting the empirical results presented in Table 5.

9. Conclusions

This paper proposes a novel and flexible discrete distribution, known as the discrete Muth distribution, designed to effectively model count data that typically exhibit over-dispersion. The proposed distribution offers several advantageous properties that enhance its performance compared to various existing discrete distributions, especially in analyzing over-dispersed count data. Various structural characteristics of the DsMuth distribution are explored, including its mean, variance, skewness, kurtosis, probability generating function, moment generating function, mean residual life, moments, quantile function, and entropy. The model parameter is estimated to be employing different estimation techniques. Several simulation studies are conducted with various parameter settings, estimation techniques, and sample sizes. The results indicate that the model’s performance improves as the sample size increases, and it was observed that the maximum likelihood method is efficient for estimating the DsMuth parameters. Additionally, the flexibility and applicability of the proposed distribution are demonstrated through its application to a real dataset. The goodness of fit measures and graphical representations indicate that the DsMuth distribution is both effective and appealing when compared to other competing discrete distributions. Finally, the proposed distribution can be utilized as a replacement for existing distributions in the literature when modeling count data.

Author Contributions

Conceptualization, H.E. and M.H.; methodology, H.E. and M.H.; software, H.E. and M.H.; investigation, H.E. and M.H.; writing—original draft preparation, H.E. and M.H.; writing—review and editing, H.E. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding from Deanship of Scientific Research at King Khalid University through General Research Project under grant number GRP/99/45.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Stated in the text.

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through a General Research Project under grant number GRP/99/45.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The discrete Shannon entropy H ( X ) of a discrete random variable X with probability mass function (PMF) is given by Equation (7):
H ( X ) = x = 0 p ( x ) log p ( x ) . H ( X ) = x = 0 p ( x ) log λ x exp 1 ln λ ( λ x 1 ) λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) .
For small values where 0 < b < a , we use:
log a b log a + log ( 1 b / a ) .
For small z, the logarithmic expansion is:
log ( 1 z ) z .
This leads to the following. Using the logarithm of subtraction property (approximation for small differences), we define:
p 1 ( x ) = λ x exp 1 ln λ ( λ x 1 ) , p 2 ( x ) = λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) .
Thus, entropy simplifies as:
H ( X ) = x = 0 p ( x ) log p 1 ( x ) p 2 ( x ) .
Using the logarithmic identity approximation:
log p 1 ( x ) p 2 ( x ) log p 1 ( x ) + log 1 p 2 ( x ) p 1 ( x ) .
log p 1 ( x ) p 2 ( x ) log λ x exp 1 ln λ ( λ x 1 ) + log 1 λ ( x + 1 ) exp 1 ln λ ( λ ( x + 1 ) 1 ) λ x exp 1 ln λ ( λ x 1 ) .
Simplifying further:
log p ( x ) = log p 1 ( x ) p 2 ( x ) x log λ 1 ln λ ( λ x 1 ) + log 1 λ exp 1 ln λ ( λ ( x + 1 ) λ x ) .
Since:
log ( 1 z ) z ,
we obtain:
H ( X ) = x = 0 p ( x ) x log λ 1 ln λ ( λ x 1 ) λ exp 1 ln λ ( λ ( x + 1 ) λ x ) .
For small λ ( x + 1 ) λ x , using Taylor expansion:
exp ( z ) 1 z ,
Thus:
exp 1 ln λ ( λ ( x + 1 ) λ x ) 1 1 ln λ ( λ ( x + 1 ) λ x ) .
Final entropy expression:
H ( X ) = log λ x = 0 x p ( x ) + 1 ln λ x = 0 ( λ x 1 ) p ( x ) + λ x = 0 1 1 ln λ ( λ ( x + 1 ) λ x ) p ( x ) .

Appendix B

Proof of Proposition A1. 
The proof involves the following algebraic operations:
P ( Y y u ) = F ( y u ) u
1 exp α ( y u + 1 ) 1 α e α ( y u + 1 ) 1 u
exp α ( y u + 1 ) 1 α e α ( y u + 1 ) 1 1 u
By setting Φ = e α ( y u + 1 ) , we obtain:
Φ = e α ( y u + 1 )
Φ exp 1 α Φ exp 1 α 1 u
Φ exp 1 α Φ ( 1 u ) exp 1 α
Multiplying by 1 α :
1 α Φ exp 1 α Φ 1 α ( 1 u ) exp 1 α
Here, w ( z ) represents the Lambert function, which satisfies:
w ( z ) e w ( z ) = z .
Thus, we obtain:
1 α Φ w 1 α ( 1 u ) exp 1 α
Φ α w 1 α ( 1 u ) exp 1 α
e α ( y u + 1 ) α w 1 α ( 1 u ) exp 1 α
y u 1 α log α w 1 α ( 1 u ) exp 1 α 1
Similarly, solving for P ( Y y u ) = 1 P ( Y y u 1 ) 1 u gives:
1 1 e α ( y u ) 1 α ( e α ( y u ) 1 ) 1 u
e α ( y u ) 1 α ( e α ( y u ) 1 ) 1 u
Setting Φ = e α ( y u ) , we obtain:
Φ exp 1 α Φ exp 1 α 1 u
Φ exp 1 α Φ ( 1 u ) exp 1 α
Multiplying by 1 α :
1 α Φ exp 1 α Φ 1 α ( 1 u ) exp 1 α
Using the Lambert function:
1 α Φ w 1 α ( 1 u ) exp 1 α
Φ α w 1 α ( 1 u ) exp 1 α
y u 1 α log α w 1 α ( 1 u ) exp 1 α
Combining [ A 1 ] and [ A 2 ] , we get:
1 α log α w 1 α ( 1 u ) exp 1 α < y u 1 α log α w 1 α ( 1 u ) exp 1 α 1
Therefore, y u is an integer given by:
y u = 1 α log α w 1 α ( 1 u ) exp 1 α .

References

  1. Lawless, J.F. Statistical Models and Methods for Lifetime Data; John Wiley and Sons, Inc.: Hoboken, NJ, USA, 2003. [Google Scholar]
  2. Roy, D. The discrete normal distribution. Commun. Stat. Theory Methods 2003, 32, 1871–1883. [Google Scholar] [CrossRef]
  3. Roy, D. Discrete Rayleigh Distribution. IEEE Trans. Reliab. 2004, 53, 255–260. [Google Scholar] [CrossRef]
  4. Kemp, A.W. The discrete half-normal distribution. In Advances in Mathematical and Statistical Modeling; Springer: Berlin/Heidelberg, Germany, 2008; pp. 353–360. [Google Scholar] [CrossRef]
  5. Krishna, H.; Pundir, P.S. Discrete Burr and discrete Pareto distributions. J. Stat. Methodol. 2009, 6, 177–188. [Google Scholar] [CrossRef]
  6. Afify, A.Z.; Ahsan-ul-Haq, M.; Aljohani, H.M.; Alghamdi, A.S.; Babar, A.; Gòmez, H.W. A new one-parameter discrete exponential distribution: Properties, inference, and applications to COVID-19 data. J. King Saud Univ.-Sci. 2022, 34, 102199. [Google Scholar] [CrossRef]
  7. Jazi, M.A.; Lai, C.; Alamatsaz, M.H. A discrete inverse Weibull distribution and estimation of its parameters. Stat. Methodol. 2010, 7, 121–132. [Google Scholar] [CrossRef]
  8. Alghamdi, A.S.; Ahsan-ul-Haq, M.; Babar, A.; Aljohani, H.M.; Afify, A.Z. The discrete power-Ailamujia distribution: Properties, inference, and applications. AIMS Math. 2022, 7, 8344–8360. [Google Scholar] [CrossRef]
  9. Aboraya, M.A.; Yousof, H.Y.; Hamedani, G.G.; Ibrahim, M. New Family of Discrete Distributions with Mathematical Properties, Characterizations, Bayesian and Non-Bayesian Estimation Methods. Mathematics 2020, 8, 1648. [Google Scholar] [CrossRef]
  10. Gómez-Déniz, E. Another generalization of the geometric distribution. Test 2010, 19, 399–415. [Google Scholar] [CrossRef]
  11. Gómez-Déniz, E.; Calderín-Ojeda, E. The discrete Lindley distribution: Properties and applications. J. Stat. Comput. Simul. 2011, 81, 1405–1416. [Google Scholar] [CrossRef]
  12. Para, B.A. Discrete Generalized Burr-Type XII Distribution. J. Mod. Appl. Stat. Methods 2014, 13, 244–258. [Google Scholar] [CrossRef]
  13. AL-Huniti, A.A.; AL-Dayian, G.R. Discrete Burr type Iii distribution. Am. J. Math. Stat. 2012, 2, 145–152. [Google Scholar] [CrossRef]
  14. Hussain, T.; Aslam, M.; Ahmad, M. A Two Parameter Discrete Lindley Distribution. Rev. Colomb. Estad. 2016, 39, 45–61. [Google Scholar] [CrossRef]
  15. Para, B.A.; Jan, T.R. Discrete Version of Log-Logistic Distribution and its Applications in Genetics. Int. J. Mod. Math. Sci. 2016, 14, 407–422. [Google Scholar]
  16. Bakouch, H.S.; Jazi, M.A.; Nadarajah, S. A new discrete distribution. J. Theor. Appl. Stat. 2012, 48, 200–240. [Google Scholar] [CrossRef]
  17. Hassan, A.; Shalbaf, G.A.; Sheikh, S.B.; Rashid, A. A New Flexible Discrete Distribution with Applications to Count Data. J. Stat. Theory Appl. 2020, 19, 102–108. [Google Scholar] [CrossRef]
  18. Almetwally, E.M.; Ibrahim, G.I. Discrete Alpha Power Inverse Lomax Distribution with Application of COVID-19 Data. J. Appl. Math. Stat. Sci. 2020, 9, 11–22. [Google Scholar]
  19. Aljohani, H.M.; Akdoğan, Y.; Cordeiro, G.M.; Afify, A. The Uniform Poisson—Ailamujia Distribution: Actuarial Measures and Applications in Biological Science. Symmetry 2021, 13, 1258. [Google Scholar] [CrossRef]
  20. Barbiero, A.; Hitaj, A. A Discrete Analogue of the Half-Logistic Distribution. In Proceedings of the International Conference on Decision Aid Sciences and Application (DASA), Sakheer, Bahrain, 8–9 November 2020; pp. 64–67. [Google Scholar] [CrossRef]
  21. Opone, F.C.; Izekor, E.A.; Akata, I.U.; Osagiede, F.E.U. A Discrete Analogue of the Continuous Marshall-Olkin Weibull Distribution with Application to Count Data. J. Math. Sci. 2021, 5, 415–428. [Google Scholar] [CrossRef]
  22. Eliwa, M.S.; Alhussain, Z.A.; El-Morshedy, M. Discrete Gompertz-G Family of Distributions for Over-and Under-Dispersed Data with Properties, Estimation, and Applications. Mathematics 2020, 8, 358. [Google Scholar] [CrossRef]
  23. Altun, E.; El-Morshedy, M.; Eliwa, M.S. A Study on Discrete Bilal Distribution with Properties and Applications on Integer-Valued Autoregressive Process. Revstat-Stat. J. 2022, 20, 501–528. [Google Scholar] [CrossRef]
  24. El-Morshedy, M.; Eliwa, M.S.; Altun, E. Discrete Burr-Hatke Distribution with Properties, Estimation Methods and Regression Model. IEEE Access 2020, 8, 74359–74370. [Google Scholar] [CrossRef]
  25. Eliwa, M.S.; Altun, E.; El-Dawoody, M.; El-Morshedy, M. A New Three-Parameter Discrete Distribution with Associated INAR(1) Process and Applications. IEEE Accesss 2020, 8, 91150–91162. [Google Scholar] [CrossRef]
  26. Al-Babtain, A.A.; Ahmed, A.H.N.; Afify, A.Z. A New Discrete Analog of The Continuous Lindley Distribution, with Reliability Applications. Entropy 2020, 22, 603. [Google Scholar] [CrossRef]
  27. El-Morshedy, M.; Eliwa, M.S.; Nagy, H. A new two-parameter exponentiated discrete Lindley distribution: Properties, estimation and applications. J. Appl. Stat. 2020, 47, 354–375. [Google Scholar] [CrossRef] [PubMed]
  28. El-Morshedy, M.; Altun, E.; Eliwa, M.S. A new statistical approach to model the counts of novel coronavirus cases. Math. Sci. 2022, 16, 37–50. [Google Scholar] [CrossRef]
  29. Chakraborty, S.; Chakravarty, D.; Mazucheli, J.; Bertoli, W. A discrete analog of Gumbel distribution: Properties, parameter estimation and applications. J. Appl. Stat. 2021, 48, 712–737. [Google Scholar] [CrossRef]
  30. Eldeeb, A.S.; Ahsan-Ul-Haq, M.; Babar, A.A. Discrete Analog of Inverted Topp-Leone Distribution: Properties, Estimation and Applications. Int. J. Anal. Appl. 2021, 19, 695–708. [Google Scholar] [CrossRef]
  31. Eldeeb, A.S.; Ahsan-Ul-Haq, M.; Eliwa, M.S. A discrete Ramos-Louzada distribution for asymmetric and over-dispersed data with leptokurtic-shaped: Properties and various estimation techniques with inference. AIMS Math. 2021, 7, 1726–1741. [Google Scholar] [CrossRef]
  32. Nekoukhou, V.; Alamatsaz, M.H.; Bidram, H. A Discrete generalized exponential distribution of a second type. Statistics 2013, 47, 876–887. [Google Scholar] [CrossRef]
  33. Hussain, T.; Ahmad, M. Discrete inverse Rayleigh distribution. J. Stat. 2014, 30, 203–222. [Google Scholar]
  34. Muth, E.J. Reliability models with positive memory derived from the mean residual life function. In The Theory and Applications of Reliability; Tsokos, C.P., Shimi, I., Eds.; Academic Press, Inc.: Cambridge, MA, USA, 1977; Volume 2, pp. 401–435. [Google Scholar]
  35. Shannon, C.E. A Mathematical Theory of Communication. Bell Labs Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  36. Rohatgi, V.K.; Saleh, A.K.M.E. An Introduction to Probability and Statistics; John Wiley and Sons: Hoboken, NJ, USA, 2015. [Google Scholar] [CrossRef]
  37. Khaliqu, K.A.; Abouammoh, A.M. On estimating parameters in a discrete Weibull distribution. IEEE Trans. Reliab. 1989, 38, 348–350. [Google Scholar] [CrossRef]
  38. Sarhan, A.M.; Tadj, L.; Hamilton, D.C. A New Lifetime Distribution and its Power Transformation. J. Probab. Stat. 2014, 6, 532024. [Google Scholar] [CrossRef]
  39. Eliwa, M.S.; El-Morshedy, M. A one-parameter discrete distribution for over dispersed data: Statistical and reliability properties with applications. J. Appl. Stat. 2022, 49, 2467–2487. [Google Scholar] [CrossRef]
  40. Poisson, S.D. Probabilité des Jugements en Matière Criminelle et en Matière Civile: Précédées des Règles Générales du Calcul des Probabilités; Bachelier: Paris, France, 1837; pp. 206–207. [Google Scholar]
  41. Sankaran, M. The Discrete Poisson-Lindley Distribution. Biometrics 1970, 26, 145–149. [Google Scholar] [CrossRef]
  42. Kumar, C.; Tripathi, Y.M.; Rastogi, M.K. On a Discrete Analogue of Linear Failure Rate Distribution. Am. J. Math. Manag. Sci. 2017, 36, 229–246. [Google Scholar] [CrossRef]
Figure 1. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 1.1 (left) and λ = 1.3 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Figure 1. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 1.1 (left) and λ = 1.3 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Entropy 27 00409 g001
Figure 2. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 1.6 (left) and λ = 1.005 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Figure 2. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 1.6 (left) and λ = 1.005 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Entropy 27 00409 g002
Figure 3. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 2.1 (left) and λ = 2.5 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Figure 3. Graphical representations of the probability mass function (pmf) for the discrete Muth (DsMuth) distribution: λ = 2.1 (left) and λ = 2.5 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Entropy 27 00409 g003
Figure 4. Plots of the hazard rate function (HRF) for the DsMuth distribution: λ = 1.2 (left) and λ = 1.3 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Figure 4. Plots of the hazard rate function (HRF) for the DsMuth distribution: λ = 1.2 (left) and λ = 1.3 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Entropy 27 00409 g004
Figure 5. Plots of the hazard rate function (HRF) for the DsMuth distribution: λ = 1.5 (left) and λ = 1.7 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Figure 5. Plots of the hazard rate function (HRF) for the DsMuth distribution: λ = 1.5 (left) and λ = 1.7 (right). 1. Red dots represent the actual values of p(x) at specific points. 2. Blue line represents the probability distribution showing how the probability changes with different values of x.
Entropy 27 00409 g005
Figure 6. Entropyof H ( X ) versus parameter λ .
Figure 6. Entropyof H ( X ) versus parameter λ .
Entropy 27 00409 g006
Figure 7. The probability mass functions (PMFs) fitted to the models with a single parameter for the dataset.
Figure 7. The probability mass functions (PMFs) fitted to the models with a single parameter for the dataset.
Entropy 27 00409 g007
Figure 8. The probability mass functions (PMFs) fitted to the models with two parameters for the dataset.
Figure 8. The probability mass functions (PMFs) fitted to the models with two parameters for the dataset.
Entropy 27 00409 g008
Table 2. Entropy of X versus parameter λ .
Table 2. Entropy of X versus parameter λ .
Parameter λ H ( X ) Parameter λ H ( X )
1.11.0301.70.900
1.21.0142.00.859
1.30.9962.10.841
1.40.9772.20.823
1.50.9572.30.807
1.60.9372.40.791
Table 4. The competitive distributions of the DsMuth distribution with their pmfs.
Table 4. The competitive distributions of the DsMuth distribution with their pmfs.
DistributionAbbreviationAuthorPmf
Discrete-RayleighDsR [3] p ( x ) = e x 2 2 λ 2 e ( x + 1 ) 2 2 λ 2
Discrete Inverse-RayleighDsIR [33] p ( x ) = e λ ( 1 + x ) 2 e λ x 2
Discrete LindleyDsLi[11] p ( x ) = λ x [ λ log λ + ( 1 λ ) ( 1 log λ x + 1 ) ] 1 log λ
PoissonPoi [40] p ( x ) = e λ λ x x !
Discrete Poisson-LindleyPoiLi[41] p ( x ) = θ 2 ( θ + 2 + x ) ( θ + 1 ) x 3
Discrete Lindley-Two ParameterDLi-II [14] p ( x ) = ( 1 + p ) 2 ( 1 + B x ) p x 1 + p ( B 1 )
Discrete linear failure rateDLFR [42] p ( x ) = λ + β ( x 1 ) 1 + β ( x 1 )
Discrete Inverse WeibullDIW [7] p ( x ) = β λ x λ β 1 exp x λ β
Discrete Log-logistic DLog-L  [15 p ( x ) = 1 1 + x + 1 λ δ 1 1 + x λ δ
Table 5. The MLE, Std-er, C.I., and goodness of fit measures for distributions with one or two parameters for the dataset.
Table 5. The MLE, Std-er, C.I., and goodness of fit measures for distributions with one or two parameters for the dataset.
XObs.
Freq.
One-ParameterTwo-Parameter
DsMuth DR DIR DLi Poi PoiLi DLi-II DLFR DIW DLLogL
06463.2133.5062.5057.1351.1737.5059.8859.9063.3062.73
11723.2546.9426.4126.8834.2825.0024.0224.0122.4822.42
2108.55617.015.9910.4511.4915.639.649.636.447.01
363.1472.392.193.712.579.383.873.862.762.98
≥431.5830.162.911.830.4912.492.592.605.024.86
Total100100100100100100100100100100
MLE1.0010.6650.6250.2740.6701.9980.4010.4010.6330.745
Std-er0.010.0290.0490.0290.0820.2630.2690.0560.0490.101
L.C.I1.0000.6080.5290.2170.5091.4810.0000.2910.5370.546
U.C.I1.0190.7220.7210.3310.8312.5140.9280.5110.7290.944
χ 2 3.93466.079.0566.63823.6530.8893.3473.3403.5032.783
DF2222221111
p-value0.140<0.0010.0110.036<0.001<0.0010.0670.0680.0610.095
−log Lik.112.8205.3118.4113.1120.3112.1112.47112.47116.27115.47
AIC227.7412.6238.9230.8242.7226.2228.95228.94236.55234.94
BIC230.3415.2241.5230.8245.3228.8234.16234.15241.76240.15
HQIC228.7413.7240.0229.2243.8227.3231.06231.05238.66237.04
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elsayed, H.; Hussein, M. A New Discrete Analogue of the Continuous Muth Distribution for Over-Dispersed Data: Properties, Estimation Techniques, and Application. Entropy 2025, 27, 409. https://doi.org/10.3390/e27040409

AMA Style

Elsayed H, Hussein M. A New Discrete Analogue of the Continuous Muth Distribution for Over-Dispersed Data: Properties, Estimation Techniques, and Application. Entropy. 2025; 27(4):409. https://doi.org/10.3390/e27040409

Chicago/Turabian Style

Elsayed, Howaida, and Mohamed Hussein. 2025. "A New Discrete Analogue of the Continuous Muth Distribution for Over-Dispersed Data: Properties, Estimation Techniques, and Application" Entropy 27, no. 4: 409. https://doi.org/10.3390/e27040409

APA Style

Elsayed, H., & Hussein, M. (2025). A New Discrete Analogue of the Continuous Muth Distribution for Over-Dispersed Data: Properties, Estimation Techniques, and Application. Entropy, 27(4), 409. https://doi.org/10.3390/e27040409

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop