Next Article in Journal
Numerical Solution of Nonlinear Backward Stochastic Volterra Integral Equations
Next Article in Special Issue
Cumulative Entropy of Past Lifetime for Coherent Systems at the System Level
Previous Article in Journal
Improvement in Some Inequalities via Jensen–Mercer Inequality and Fractional Extended Riemann–Liouville Integrals
Previous Article in Special Issue
Modeling Bland–Altman Limits of Agreement with Fractional Polynomials—An Example with the Agatston Score for Coronary Calcification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Estimation of Variance-Based Information Measures and Their Application to Testing Uniformity

Department of Mathematical & Computational Sciences, University of Toronto Mississauga, Toronto, ON L5L 1C6, Canada
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(9), 887; https://doi.org/10.3390/axioms12090887
Submission received: 16 August 2023 / Revised: 13 September 2023 / Accepted: 15 September 2023 / Published: 17 September 2023

Abstract

:
Entropy and extropy are emerging concepts in machine learning and computer science. Within the past decade, statisticians have created estimators for these measures. However, associated variability metrics, specifically varentropy and varextropy, have received comparably less attention. This paper presents a novel methodology for computing varentropy and varextropy, drawing inspiration from Bayesian nonparametric methods. We implement this approach using a computational algorithm in R and demonstrate its effectiveness across various examples. Furthermore, these new estimators are applied to test uniformity in data.

1. Introduction

In this section, we begin by reviewing the concepts of entropy and extropy, along with some existing estimators from the literature. Subsequently, we introduce estimators of varentropy and varextropy developed using a frequentist approach.
Entropy is a fundamental concept in information theory that was originally introduced in [1]. It has found numerous applications in various fields, such as thermodynamics, communication theory, reliability, computer science, biology, economics, and statistics [2,3]. Let X be a continuous random variable with support on S , cumulative distribution function (CDF) F, and probability density function (PDF) f. The entropy H ( F ) of X is defined as
H ( F ) = E f log f ( x ) = S f ( x ) log f ( x ) d x ,
where log is the natural logarithm.
Extropy, a concept introduced by [4], represents a relatively recent development in the field of statistics that serves as a dual counterpart to entropy. Its significance has been demonstrated in various studies, including those conducted by [5,6], where it has found applications in the context of goodness-of-fit tests.
For a continuous random variable X supported on S with CDF F and PDF f, the extropy of X, denoted by J ( F ) , is defined as
J ( F ) = 1 2 E f f ( x ) = 1 2 S f 2 ( x ) d x
In most practical scenarios, the true PDF f is unknown, and hence, we need to estimate the entropy (1) and extropy (2) from the available data, which can be a challenging task. Several frequentist methods are available for entropy estimation in the literature. Among different approaches, Ref. [7] estimator has gained wide popularity due to its simplicity. Ref. [7] derived the expression for (1) in terms of the inverse of the distribution function F, given by
H ( F ) = 0 1 log d d t F 1 ( t ) d t .
Using the empirical distribution function F n instead of the unknown F, Ref. [7] proposed an estimator for H ( F ) based on the difference operator rather than the differential operator. The derivative of F 1 ( t ) is estimated using a function of the order statistics. Specifically, if  X 1 , X 2 , , X n is a sample from F, then [7] estimator is given by
H 1 m , n = n 1 i = 1 n log X ( i + m ) X ( i m ) 2 m / n ,
where m is a positive integer smaller than n / 2 , and  X ( 1 ) X ( 2 ) X ( n ) are the order statistics of X 1 , X 2 , , X n with X ( i m ) = X ( 1 ) if i m and X ( i + m ) = X ( n ) if i n m . Ref. [7] showed that H 1 m , n p H ( F ) as n , m , and  m n 0 , where p denotes convergence in probability. Note that the expression inside the log in (3) is the slope of the straight line that passes through the points i + m n , X ( i + m ) and i m n , X ( i m ) , where F n ( X ( i + m ) ) = i + m n and F n ( X ( i m ) ) = i m n . Ref. [8] proposed a modification to the estimator (3) as it does not provide the correct formula for the slope when i m or i n m + 1 . The proposed estimator, denoted by H 2 m , n , is given by
H 2 m , n = n 1 i = 1 n log X ( i + m ) X ( i m ) c i m / n ,
where
c i = m + i 1 m if 1 i m 2 if m + 1 i n m n + m i m if n m + 1 i n .
As for extropy estimation, Ref. [5] noticed that the extropy, first defined in (2) can be rewritten as
J ( F ) = 1 2 0 1 d d t F 1 ( t ) 1 d t
and proposed the following estimator for J ( F ) :
J 1 m , n = 1 2 n i = 1 n 2 m / n X ( i + m ) X ( i m ) ,
As in H 1 m , n , Ref. [5] found that J 1 m , n gives incorrect estimates for i m or i n m + 1 . Therefore, they proposed the revised estimator J 2 m , n , where
J 2 m , n = 1 2 n i = 1 n c i m / n X ( i + m ) X ( i m ) .
Here, c i is defined as specified in Equation (5). Ref. [5] proved that J 1 m , n and J 2 m , n converges in probability to J ( F ) under the same conditions as for H 1 m , n or H 2 m , n .
Other frequentist nonparametric estimators of entropy include those proposed by  [9,10,11,12,13,14]. A comprehensive review of nonparametric entropy estimators can be found in [15]. For extropy estimation, alternative approaches are provided by [16] as well as [17].
Bayesian estimation of entropy has not received as much attention as the frequentist approach. However, Ref. [18] developed a Bayes estimator of H ( F ) based on the Dirichlet process [19]. Recently, Refs. [6,20,21] proposed an estimator of entropy and extropy based on an approximation of the Dirichlet process D P ( a , G ) introduced by [22], where a > 0 and G is a known CDF. The approximation is defined as
P N ( · ) = i = 1 N w i , N δ Y i ( · ) ,
In this equation, the weights ( w 1 , N , , w N , N ) follow a Dirichlet distribution with parameters ( a / N , , a / N ) , while Y 1 , , Y N are independent and identically distributed from the distribution G. The notation δ Y i represents the Dirac measure at the point Y i . The sequences ( W i , N ) 1 i N and ( Y i ) 1 i N are independent. We refer to ( Y i ) 1 i N as the data points of P N . Let
H m , N , a = 1 N i = 1 N log Y ( i + m ) Y ( i m ) c i , a
and
J m , N , a = 1 2 N i = 1 N c i , a Y ( i + m ) Y ( i m ) ,
where
c i , a = k = 2 i + m w k , N 1 i m , k = i m + 1 i + m w k , N m + 1 i N m , k = i m + 1 N w k , N N m + 1 i N .
As N , m , m N 0 and a , Refs. [6,20] showed that
H m , N , a p H ( G ) = S g ( x ) log g ( x ) d x
and
J m , N , a p J ( G ) = 1 2 S g 2 ( x ) d x ,
where G ( x ) = g ( x ) . Observe that the slope of the straight line connecting the two points P N ( Y ( i m ) ) , Y ( i m ) and P N ( Y ( i + m ) ) , Y ( i + m ) is
Y ( i + m ) Y ( i m ) P N ( Y ( i + m ) ) P N ( Y ( i m ) ) = Y ( i + m ) Y ( i m ) c i , a .
Let X = ( X 1 , X 2 , , X n ) be a sample from F and D P ( a , G ) be a prior of F. Consider H m , N , a | X to be the posterior version of H m , N , a as defined in (7) with P N replaced by P N | X , an approximation of D P ( a + n , G X ) , where
G X = a ( a + n ) 1 G + n ( a + n ) 1 F n .
Then, as  N , m , n , m N 0 , and  a n 0 , we have [6,20]
H m , N , a | X p H ( F ) = S f ( x ) log f ( x ) d x ,
and
J m , N , a | X p J ( F ) = 1 2 S f 2 ( x ) d x ,
where F ( x ) = f ( x ) .
Recently, there has been significant interest in studying the variability of information measures in the literature. In certain situations, it is possible to encounter two random variables with identical entropy or extropy. As a result of these circumstances, researchers ponder whether entropy or extropy would be the most suitable criterion for measuring uncertainty. One way of determining which probability distribution is more suitable is to check its variance. This scenario serves as a motivation for exploring two variance measures associated with entropy and extropy, known as varentropy and varextropy, respectively.
For a random variable X, the varentropy, denoted by V H ( F ) , is defined as follows:
V H ( F ) = Var f ( log ( f ( X ) ) ) = S f ( x ) [ log f ( x ) ] 2 d x S f ( x ) log f ( x ) d x 2 = E f log ( ( f ( x ) ) 2 ) [ H ( F ) ] 2 .
Ref. [23] introduced varentropy as a compelling alternative to the kurtosis measure, particularly for comparing heavy-tailed distributions. In fact, varentropy has proven to be a valuable tool for assessing heavy-tailed distributions instead of relying solely on kurtosis. Subsequently, varentropy has found diverse applications across various fields. In computer science, varentropy plays an instrumental role in data compression, studying the variability of uncertainty measures [24], testing uniformity [25], and analyzing statistical issues [26]. Moreover, researchers have effectively employed varentropy to explore the variability of interval entropy measures [27]. Additionally, varentropy has proven valuable in applications related to proportional hazard rate models [24] and residual lifetime distributions [26].
Ref. [25] presented six estimators for computing varentropy. In this context, we will focus on two specific estimators based on the estimators (3) and (4). Specifically, V H ( F ) can be expressed as
V H ( F ) = 0 1 log 2 d d t F 1 ( t ) d t 0 1 log d d t F 1 ( t ) d t 2 ,
the two estimators of [25] are
V H 1 m n = 1 n i = 1 n log 2 X ( i + m ) X ( i m ) 2 m / n H 1 m , n 2 = 1 n i = 1 n log 2 X ( i + m ) X ( i m ) 1 n i = 1 n log ( X ( i + m ) X ( i m ) ) 2
and
V H 2 m n = 1 n i = 1 n log 2 X ( i + m ) X ( i m ) c i m / n H 2 m , n 2 ,
where H 1 m , n , H 2 m , n , and  c i are defined in (3), (4), and (5), respectively. Noughabi and Noughabi (2023) showed that V H 1 m n and V H 2 m n converge in probability to V H ( F ) under the same conditions as for H 1 m , n or H 2 m , n .
Another measure of information variability is the varextropy. Let X be an absolutely continuous random variable, then its varextropy, denoted as V J ( X ) , is defined as follows [28]:
V J ( X ) = Var f ( 1 2 f ( X ) ) = 1 4 E f ( ( f ( X ) ) 2 ) ( J ( X ) ) 2
= 1 4 S f 3 ( x ) d x 1 4 S f 2 ( x ) d x 2 .
Unlike the varentropy V H ( F ) , the estimation of V J ( F ) has not been intensively discussed in the literature. Notice that
V J ( F ) = 1 4 0 1 d d t F 1 ( t ) 2 d t 1 4 0 1 d d t F 1 ( t ) 1 d t 2 .
Accordingly, our proposed estimators of V J ( X ) based on the two estimator of [5] are:
V J 1 m n = 1 4 n i = 1 n 2 m / n x ( i + m ) x ( i m ) 2 J 1 m , n 2
and
V J 2 m n = 1 4 n i = 1 n c i m / n x ( i + m ) x ( i m ) 2 J 2 m , n 2 ,
where c i is defined in (5). The convergence in probability of V J 1 m n and V J 2 m n to V J ( F ) straightforwardly follows from the consistency of J 1 m , n and J 2 m , n .
The rest of this paper is structured as follows. Section 2 presents the proposed Bayesian estimator based on the Dirichlet process. Section 3 details the proposed approach, including a computational algorithm. In Section 4, a test for uniformity is developed. Section 5 presents several examples to illustrate the approach. Finally, Section 6 contains concluding remarks and discussions.

2. Bayesian Estimation of Varentropy and Varextropy

In this section, we derive Bayesian nonparametric estimators for varentropy and varextropy. Define the following two quantities:
V H m , N , a = 1 N i = 1 N log 2 Y ( i + m ) Y ( i m ) c i , a H m , N , a 2
and
V J m , N , a = 1 4 N i = 1 N c i , a Y ( i + m ) Y ( i m ) 2 J m , N , a 2 ,
where c i , a , H m , N , a , and J m , N , a are defined in (9), (7), and (8), respectively. The following proposition presents the prior formulation of varentropy and varextropy. The proof follows from the consistency of H m , N , a and J m , N , a .
Lemma 1.
Let V H m , N , a and V J m , N , a be defined as in (18) and (19), respectively. Let P N be an approximation of the D P ( a , G ) as defined (6). As  N , m , m N 0 and a , we have
V H m , N , a p V H ( G )
and
V J m , N , a p V J ( G ) ,
where G ( x ) = g ( x ) .
The following proposition demonstrates that as the sample size increases (with the concentration parameter a being relatively small compared to the sample size n), the posterior distributions of V H m , N , a and V J m , N , a converge in probability to V H ( F ) and V J ( F ) , respectively. The proof follows from consistency of H m , N , a | X and J m , N , a | X .
Lemma 2.
Let X = ( X 1 , , X n ) be a sample from F. Let the prior on F be D P ( a , H ) . Let V H m , N , a and V J m , N , a be as defined in (18) and (19), respectively. As  N , m , n , m N 0 , and  a n 0 , we have
V H m , N , a | X p V H ( F )
and
V J m , N , a | X p V J ( F ) ,
where V H ( F ) and V J ( F ) are defined in (11) and (14), respectively, with  F ( x ) = f ( x ) .

3. Computational Algorithms

Let X = ( X 1 , , X n ) be a sample from a continuous distribution F. The objective is to approximate V H ( F ) and J H ( F ) using the approximation discussed in Section 2. To proceed with this approximation, it is important to determine the values of m, a, and G. We begin by considering the choice of m. A commonly used formula, proposed by [29], is given by
m = N + 0.5 ,
where ⌊y⌋ denotes the largest integer less than or equal to y. Note that the value of m in (20) is used for the prior. For the posterior, the value of N should be replaced with the number of distinct data points in P N | X , an approximation of F | X . It is worth noting that, from (10), if  a / n is close to zero, the number of distinct data points in P N | X will be approximately n.
Regarding the hyperparameters a and G of the Dirichlet process, their selection depends on the specific application of interest. For varentropy and varextropy estimation, any choice of a such that a / n is close to zero should be suitable, regardless of the choice of G. This property is evident from (10), as when a / n approaches 0, the sample will dominate the prior guess G. Consequently, the approach becomes invariant to the choice of G. As an illustrative example, by setting a = 0.01 and n = 20 in (10), we obtain
G X = 0.0005 G + 0.9995 F n .
This suggests a 99.95% likelihood of selecting a sample from the gathered data rather than drawing a new sample from G. To facilitate estimation, we will consider G as the uniform distribution over ( 0 , 1 ) and set a = 0.01 , although alternative choices are certainly possible. Within Section 4, we will include an example that investigates the sensitivity of the approach to the choices of a and G.
For a given observed data set X = ( X 1 , X 2 , , X n ) , we employ the following computational algorithm to estimate V H ( F ) and V J ( F ) based on Equations (18) and (19).
Algorithm 1.
(Nonparametric Estimation of Varentropy and Varextropy):
(i) 
Generate a sample from P N , where P N is an approximation of D P ( a = 0.01 , G = U ( 0 , 1 ) ) .
(ii) 
Generate a sample from P N | X , where P N | X is an approximation of ∼ D P ( a + n , G X ) .
(iii) 
Compute V H m , N , a | X and V J m , N , a | X as specified in Lemma 2.
(iv) 
Repeat steps (i) and (iii) to obtain a sample of r values from V H m , N , a | X and V J m , N , a | X . As r increases, the average of the generated r values becomes the estimator for varentropy and varextropy.

4. Testing for Uniformity

Suppose that X = ( X 1 , , X n ) is a sample from an unknown continuous distribution F. The objective is to test the hypothesis H 0 : F ( x ) = F 0 ( x ) , for all x R , where F 0 represents a fully specified distribution. By utilizing the probability integral transform property, we can deduce that F ( X 1 ) , , F ( X n ) follows the uniform distribution on the interval ( 0 , 1 ) . Therefore, testing the null hypothesis is equivalent to testing H 0 : U ( x ) = x for all x ( 0 , 1 ) , where U ( x ) represents the CDF of a uniform ( 0 , 1 ) random variable. For further details on testing uniformity, please consult the work of [30].
For any random variable X, it holds that V H ( F ) 0 and V J ( F ) 0 . The following proposition demonstrates that equality is achieved when F corresponds to the CDF of a uniform distribution on the interval ( 0 , 1 ) . This property plays a crucial role in testing the hypothesis H 0 .
Proposition 1.
Let f be a probability density function with support in [ 0 , 1 ] , we have
(i) 
V H ( F ) = 0 if and only if f ( x ) = 1 for all x ( 0 , 1 ) (i.e., f is the PDF of the uniform random variable on ( 0 , 1 ) ).
(ii) 
V J ( F ) = 0 if and only if f ( x ) = 1 for all x ( 0 , 1 ) (i.e., f is the PDF of the uniform random variable on ( 0 , 1 ) ).
Proof. 
For the proof of (i), see Theorem 4.1 of [25]. For (ii), if  f ( x ) = 1 for all x ( 0 , 1 ) , then V J ( F ) = Var ( 0.5 f ( X ) ) = 0 . Also, if  V J ( F ) = 0 , then f ( x ) = c , for all x ( 0 , 1 ) . Since 0 1 f ( x ) d x = 1 , we have c = 1 . Hence, f ( x ) = 1 for all x ( 0 , 1 ) .    □
The proposed test for uniformity involves comparing V H m , N , a | X and V J m , N , a | X . When the null hypothesis H 0 is true, it is expected that V H m , N , a | X V J m , N , a | X 0 . Conversely, if there is evidence that V H m , N , a | X or V J m , N , a | X deviates significantly from zero, H 0 is rejected.

5. Examples

5.1. Simulation Study

In this subsection, we focus on investigating the efficiency and robustness of the proposed estimator for varentropy and varextropy. Additionally, we demonstrate the implementation of the uniformity test using these estimations. To evaluate the performance of our proposed Bayesian estimator, we compare it with the non-Bayesian counterparts obtained from (12), (13), (16), and (17). The comparison between Bayesian and non-Bayesian methods holds particular significance in this context, particularly when we consider the scenario where a = 0.01 . In this case, the estimator remains unaffected by the selection of the prior guess G, as demonstrated in Section 4.
To carry out the computations, we implemented the required program codes in the programming language R, and these codes are made available by the authors. For demonstration purposes, we constructed a demo for the algorithms and presented them using R Shiny, as shown here: https://annaly.shinyapps.io/BayesianVarentropyVarextropy/ (accessed on 15 July 2023). In Algorithm 1, we set the parameters r = 1000 and N = 500 to ensure accurate and sufficient evaluations. Additionally, to ensure reproducibility, the set.seed(100) function in R was utilized for all examples.
Throughout this section, we use the following notation: N ( μ , σ 2 ) denotes the normal distribution with mean μ and standard deviation σ , t r represents the t distribution with r degrees of freedom, Exp ( λ ) corresponds to the exponential distribution with mean 1 / λ , U ( a , b ) signifies the uniform distribution over the interval ( a , b ) , and beta ( α , β ) denotes the beta distribution with parameters α and β .
In Table 1 and Table 2, for each sample size ( n = 20 , 50 , 100 ), 1000 samples were generated. We have considered three distributions: uniform on ( 0 , 1 ) (exact varentropy and varextopy are both 0), exponential with mean 1 (exact varentropy and varextropy are 1 and 1 / 48 0.0208 , respectively), N ( 0 , 1 ) (exact varentropy and exact varextropy are 0.5 and ( 2 3 ) / 16 π 3 0.0031 , respectively). The estimators and their root mean squared errors are computed and reported in Table 1 and Table 2. The reported value of the estimator (Est) is the average of the 1000 estimates. On the other hand, the root mean squared error (RMSE) is computed as follows: i = 1 1000 ( Est i true value ) 2 / 1000 , where Est i is the estimated value based on the ith sample.
Based on the findings presented in Table 1 and Table 2, it is evident that the estimators for varentropy and varextropy demonstrate overall good performance.
It is also of interest to examine the impact of utilizing different base measures G and concentration parameters a on the methodology. To explore this, we consider two distinct values for a, namely 0.01 and 5, and examine various choices for G. In our analysis, we utilize a dataset generated from the exponential distribution with a mean of 1. Based on the findings presented in Table 3, it can be concluded that the estimators exhibit robustness to the choice of G when a = 0.01 .
In this last example, we generated samples of sizes n = 20 , 50 , and 100 from the uniform distribution on the interval ( 0 , 1 ) . The goal is to test the hypothesis H 0 : F ( x ) = F 0 ( x ) using the proposed test of uniformity. To achieve this, we considered a range of candidate distribution functions F 0 ( x ) as outlined in Table 4, where
A k : F 0 ( x ) = 1 ( 1 x ) k , 0 x 1 ( for k = 1.5 , 2 ) ; B k : F 0 ( x ) = 2 k 1 x k 0 x 0.5 , 1 2 k 1 ( 1 x ) k 0.5 x 1 ( for k = 1.5 , 2 , 3 ) ; C k : F 0 ( x ) = 0.5 2 k 1 ( 0.5 x ) k 0 x 0.5 , 0.5 + 2 k 1 ( 0.5 x ) k 0.5 x 1 ( for k = 1.5 , 2 ) .
These candidate distribution functions A k , B k , and C k of F 0 ( x ) have been previously studied by various authors, including [25,31]. The distributions of Exp ( 2 ) and N ( 0 , 1 ) are included here to explore cases with support different from [ 0 , 1 ] . The results are presented in Table 4, where we also included the p-values obtained from the Kolmogorov–Smirnov test.
Using Monte Carlo simulation, we can determine an appropriate cut-off for both V H m , n , a | X and V J m , n , a | X in the test of uniformity under H 0 . We recommend using Q 3 , the third quartile, as a suitable threshold. In Table 4, we present the thresholds for n = 20 , 50 , and 100, based on 5000 values of V H m , n , a | X and V J m , n , a | X . If both estimates are less than their respective thresholds, it is advisable not to reject the null hypothesis H 0 . However, if one of the estimates is greater than its threshold, it is recommended to reject H 0 .
For example, when n = 50 and F 0 = A 1.5 , as V H m , n , a | X > 0.1469 (or V J m , n , a | X > 0.0475 ), H 0 is rejected. Conversely, when F 0 = B 1.5 , as both V H m , n , a | X < 0.1469 and V J m , n , a | X < 0.0475 , H 0 is not rejected.

5.2. Real Data Examples

Military Personnel Carriers Dataset [32]: The following data represents mileages for 19 military personnel carriers that failed in service. The mileages are as follows:
  • 162, 200, 271, 320, 393, 508, 539, 629, 706, 778, 884, 1003, 1101, 1182, 1463, 1603, 1984, 2355, 2880.
The aim of the study is to test whether the data follows an exponential distribution with a mean of 998. Employing Algorithm 1 with parameters n = 19 , N = 500 , and r = 1000 , we obtained V H m , n , a | X = 0.1124 and V J m , n , a | X = 0.0331 . Since both of these values are significantly lower than their respective thresholds (0.2509 for V H m , n , a | X and 0.0969 for V J m , n , a | X ), we cannot reject the hypothesis that the failure time is exponentially distributed with a mean of 998. This conclusion aligns with the findings of [33].
Chick Dataset [30]: The dataset below represents the weights of 20 chicks in grams:
  • 156, 162, 168, 182, 186, 190, 190, 196, 202, 210, 214, 220, 226, 230, 230, 236, 236, 242, 246, 270.
The goal of this study is to test whether the data follow a normal distribution with a mean of 200 and a variance of 1225. Using Algorithm 1 with parameters n = 20 , N = 500 , and r = 1000 , we obtained V H m , n , a | X = 0.1396 and V J m , n , a | X = 0.0000 . Both of these values are significantly lower than their respective thresholds (0.2461 for V H m , n , a | X and 0.0955 for V J m , n , a | X ). Consequently, we cannot reject the hypothesis that the data follow a normal distribution with a mean of 200 and a variance of 1225. This conclusion is consistent with the findings of [30].

6. Conclusions

In this paper, we introduced a novel estimator for varentropy and varextropy, drawing inspiration from Bayesian nonparametric statistical methods. This method exhibits flexibility as it does not rely on any specific assumptions about the underlying distribution. Furthermore, we also presented a goodness-of-fit test. Extensive testing and validation of our estimator were conducted using multiple simulated examples and a real-life application. The results clearly demonstrate that our estimator displays favorable and accurate performance.
Moreover, the applicability of our approach is not limited to varentropy and varextropy alone. It is possible to extend the results presented in this paper to study other dispersion indices. For instance, dispersion indices based on Kerridge inaccuracy measure and Kullback–Leibler divergence, as studied by [34], can be explored using a similar Bayesian nonparametric framework.

Author Contributions

Methodology, L.A.-L. and A.L.; Software, L.A.-L., A.L. and M.H.; Resources L.A.-L.; Writing—review & editing, L.A.-L., A.L. and M.H.; Supervision, L.A.-L. All authors have read and agreed to the published version of the manuscript.

Funding

The Natural Sciences and Engineering Research Council of Canada (NSERC).

Data Availability Statement

There are no data associated with this paper.

Acknowledgments

We would like to express our sincere gratitude to the Editor and the anonymous referees for their valuable and constructive comments. Their insightful feedback and suggestions have greatly contributed to the improvement of this paper.

Conflicts of Interest

On behalf of all authors, there are no conflict of interest.

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons, Inc.: New York, NY, USA, 1991. [Google Scholar]
  3. Kamavaram, S.; Goseva-Popstojanova, K. Entropy as a measure of uncertainty in software reliability. In Proceedings of the 13th International Symposium on Software Reliability Engineering, Annapolis, MD, USA, 12–15 November 2002; pp. 209–210. [Google Scholar]
  4. Lad, F.; Sanfilippo, G.; Agro, G. Extropy: Complementary dual of entropy. Stat. Sci. 2015, 30, 40–58. [Google Scholar] [CrossRef]
  5. Qiu, G.; Jia, K. Extropy estimators with applications in testing uniformity. J. Nonparametr. Stat. 2018, 30, 182–196. [Google Scholar] [CrossRef]
  6. Al-Labadi, L.; Berry, S. Bayesian estimation of extropy and goodness of fit tests. J. Appl. Stat. 2020, 49, 357–370. [Google Scholar] [CrossRef]
  7. Vasicek, O. A test for normality based on sample entropy. J. R. Stat. Soc. B 1976, 38, 54–59. [Google Scholar] [CrossRef]
  8. Ebrahimi, N.; Habibullah, M.; Soofi, E.S. Testing exponentiality based on Kullback-Leibler information. J. R. Stat. Soc. Ser. B Stat. Methodol. 1992, 54, 739–748. [Google Scholar] [CrossRef]
  9. Van Es, B. Estimating functionals related to a density by a class of statistics based on spacings. Scand. J. Stat. 1992, 19, 61–72. [Google Scholar]
  10. Correa, J.C. A new estimator of entropy. Commun. Stat.—Theory Methods 1995, 24, 2439–2449. [Google Scholar] [CrossRef]
  11. Wieczorkowski, R.; Grzegorzewski, P. Entropy estimators-improvements and comparisons. Commun. Stat.—Simul. Comput. 1999, 28, 541–567. [Google Scholar] [CrossRef]
  12. Alizadeh Noughabi, H. A new estimator of entropy and its application in testing normality. J. Stat. Comput. Simul. 2010, 80, 1151–1162. [Google Scholar] [CrossRef]
  13. Alizadeh Noughabi, H.; Arghami, N.R. A new estimator of entropy. J. Iran. Stat. Soc. 2010, 9, 53–64. [Google Scholar]
  14. Al-Omari, A.I. A new measure of entropy of continuous random variable. J. Stat. Theory Pract. 2016, 10, 721–735. [Google Scholar] [CrossRef]
  15. Beirlant, J.; Dudewicz, E.J.; Györia, L.; van der Meulen, E.C. Nonparametric entropy estimation: An overview. Int. J. Math. Stat. 1997, 6, 17–39. [Google Scholar]
  16. Noughabia, H.A.; Jarrahiferizb, J. Extropy of order statistics applied to testing symmetry. Commun. Stat.—Simul. Comput. 2020, 51, 3389–3399. [Google Scholar] [CrossRef]
  17. Källberg, D.; Seleznjev, O. Estimation of entropy-type integral functionals. Commun. Stat.—Theory Methods 2016, 45, 887–905. [Google Scholar] [CrossRef]
  18. Mazzuchi, T.A.; Soofi, E.S.; Soyer, R. Bayes estimate and inference for entropy and information index of fit. Econom. Rev. 2008, 27, 428–456. [Google Scholar] [CrossRef]
  19. Ferguson, T.S. A Bayesian analysis of some nonparametric problems. Ann. Stat. 1973, 1, 209–230. [Google Scholar] [CrossRef]
  20. Al-Labadi, L.; Patel, V.; Vakiloroayaei, K.; Wan, C. A Bayesian nonparametric estimation to entropy. Braz. J. Probab. Stat. 2021, 35, 421–434. [Google Scholar] [CrossRef]
  21. Al-Labadi, L.; Patel, V.; Vakiloroayaei, K.; Wan, C. Kullback-Leibler divergence for Bayesian nonparametric model checking. J. Korean Stat. Soc. 2021, 50, 272–289. [Google Scholar] [CrossRef]
  22. Ishwaran, H.; Zarepour, M. Exact and approximate sum representations for the Dirichlet process. Can. J. Stat. 2002, 30, 269–283. [Google Scholar] [CrossRef]
  23. Song, K.S. Rényi information, log likelihood and an intrinsic distribution measure. J. Stat. Plan. Inference 2001, 93, 51–69. [Google Scholar] [CrossRef]
  24. Saha, S.; Kayal, S. Weighted (residual) varentropy with properties and applications. arXiv 2023, arXiv:2305.00852. [Google Scholar]
  25. Noughabi, H.A.; Noughabi, M.S. Varentropy estimators with applications in testing uniformity. J. Stat. Comput. Simul. 2023, 93, 2582–2599. [Google Scholar] [CrossRef]
  26. Maadani, S.; Mohtashami Borzadaran, G.R.; Rezaei Roknabadi, A.H. A new generalized varentropy and its properties. Ural. Math. J. 2020, 6, 114–129. [Google Scholar] [CrossRef]
  27. Sharma, A.; Kundu, C. Varentropy of doubly truncated random variable. Probab. Eng. Inf. Sci. 2022, 7, 852–871. [Google Scholar] [CrossRef]
  28. Vaselabadi, N.M.; Tahmasebi, S.; Kazemi, M.R.; Buono, F. Results on varextropy measure of random variables. Entropy 2021, 23, 356. [Google Scholar] [CrossRef]
  29. Grzegorzewski, P.; Wieczorkowski, R. Entropy-based goodness-of-fit test for exponentiality. Commun. Stat.—Theory Methods 1999, 28, 1183–1202. [Google Scholar] [CrossRef]
  30. D’Agostino, R.B.; Stephens, M.A. Goodness-of-Fit Techniques; Marcel Dekker: New York, NY, USA, 1986. [Google Scholar]
  31. Stephens, M.A. EDF statistics for goodness of fit and some comparisons. J. Am. Stat. Assoc. 1974, 69, 730–737. [Google Scholar] [CrossRef]
  32. Grubbs, F.E. Fiducial bounds on reliability for the two-parameter negative exponential distribution. Technometrics 1971, 13, 873–876. [Google Scholar] [CrossRef]
  33. Ebrahimi, N.; Pflughoeft, K.; Soofi, E.S. Two measures of sample entropy. Stat. Probab. Lett. 1994, 20, 225–234. [Google Scholar] [CrossRef]
  34. Balakrishnan, N.; Buono, F.; Calì, C.; Longobardi, M. Dispersion indices based on Kerridge inaccuracy measure and Kullback-Leibler divergence. Commun. Stat.—Theory Methods 2023. [Google Scholar] [CrossRef]
Table 1. Varentropy Measure Estimates.
Table 1. Varentropy Measure Estimates.
VH m , n , a | X VH 1 m , n VH 2 m , n
Distribution n m Est(RMSE)
U(0, 1)2040.2055(0.2248)0.1749(0.218)0.1099(0.1465)
5070.1266(0.1334)0.1059(0.1214)0.0641(0.0769)
100100.0919(0.0954)0.0783(0.0873)0.0469(0.0536)
Exp(1)2040.7446(0.4699)0.6663(0.5105)0.6586(0.5258)
5070.9025(0.3148)0.7929(0.3454)0.8344(0.3396)
100100.9671(0.2315)0.8641(0.2533)0.9203(0.2417)
N ( 0 , 1 ) 2040.2646(0.2621)0.1221(0.3885)0.1620(0.3574)
5070.3051(0.2245)0.1449(0.3633)0.2351(0.2864)
100100.3666(0.1677)0.2136(0.2971)0.319(0.2076)
Table 2. Varextropy Measure Estimates.
Table 2. Varextropy Measure Estimates.
VJ m , n , a | X JQ 1 m , n JQ 2 m , n
Distribution n m Est(RMSE)
U(0, 1)2040.0828(0.1159)0.1698(0.3129)0.0502(0.0875)
5070.0404(0.0460)0.0642(0.0943)0.0217(0.0291)
100100.0262(0.0283)0.0352(0.0460)0.0135(0.0166)
Exp(1)2040.0589(0.1451)0.1419(0.5516)0.0505(0.1466)
5070.0370(0.0291)0.0621(0.0696)0.0311(0.0238)
100100.0337(0.0197)0.0514(0.0435)0.0294(0.0161)
N ( 0 , 1 ) 2040.0063(0.0065)0.0062(0.0136)0.0043(0.0047)
5070.0053(0.0034)0.0033(0.0022)0.0041(0.0026)
100100.0069(0.0044)0.0037(0.0018)0.0043(0.0021)
Table 3. Analysis of the impact of different values of a and G on the proposed estimators.
Table 3. Analysis of the impact of different values of a and G on the proposed estimators.
Ga VH m , n , a | X VJ m , n , a | X
N ( 0 , 1 ) 0.010.87580.0663
52.49170.0255
N ( 3 , 9 ) 0.010.88770.0667
53.23280.0073
t 1 0.010.89320.0681
57.73350.0215
Exp ( 1 ) 0.0010.85870.0657
51.61130.0725
U ( 0 , 1 ) 0.010.86990.0668
50.75250.0776
Table 4. Goodness-of-Fit Test.
Table 4. Goodness-of-Fit Test.
n F 0 Est. of VH Est. of VJ Threshold of VH Threshold of VJ p-Value
20 A 1.5 0.15730.04230.24610.09550.2953
A 2 0.24940.0839 0.0442
B 1.5 0.13220.0337 0.7472
B 2 0.15530.0421 0.4646
B 3 0.29340.1163 0.1469
C 1.5 0.30820.1687 0.1958
C 2 0.55430.6276 0.0745
U ( 0 , 1 ) 0.15290.0484 0.6307
Exp ( 2 ) 0.26040.1169 0.1300
beta ( 3 , 1 ) 1.14382.6605 0.0000
N ( 0 , 1 ) 0.19440.3892 0.0000
50 A 1.5 0.23780.10440.14690.04750.0073
A 2 0.61050.7342 0.0000
B 1.5 0.11760.0391 0.4119
B 2 0.25640.1421 0.0857
B 3 0.76991.6625 0.0048
C 1.5 0.24490.1046 0.0430
C 2 0.547170.4817 0.0045
U ( 0 , 1 ) 0.08760.0232 0.3349
Exp ( 2 ) 0.38190.2553 0.0013
beta ( 3 , 1 ) 1.01891.6964 0.0000
N ( 0 , 1 ) 0.13610.2849 0.0000
100 A 1.5 0.20730.08730.10420.031070.0019
A 2 0.59060.8231 0.0000
B 1.5 0.09580.0329 0.6042
B 2 0.26020.1665 0.0764
B 3 0.91353.0551 0.0011
C 1.5 0.28770.1323 0.0080
C 2 0.70740.8427 0.0001
U ( 0 , 1 ) 0.07550.0193 0.2672
Exp ( 2 ) 0.35070.1926 0.0000
beta ( 3 , 1 ) 1.36694.1632 0.0000
N ( 0 , 1 ) 0.10060.1951 0.0000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-Labadi, L.; Hamlili, M.; Ly, A. Bayesian Estimation of Variance-Based Information Measures and Their Application to Testing Uniformity. Axioms 2023, 12, 887. https://doi.org/10.3390/axioms12090887

AMA Style

Al-Labadi L, Hamlili M, Ly A. Bayesian Estimation of Variance-Based Information Measures and Their Application to Testing Uniformity. Axioms. 2023; 12(9):887. https://doi.org/10.3390/axioms12090887

Chicago/Turabian Style

Al-Labadi, Luai, Mohammed Hamlili, and Anna Ly. 2023. "Bayesian Estimation of Variance-Based Information Measures and Their Application to Testing Uniformity" Axioms 12, no. 9: 887. https://doi.org/10.3390/axioms12090887

APA Style

Al-Labadi, L., Hamlili, M., & Ly, A. (2023). Bayesian Estimation of Variance-Based Information Measures and Their Application to Testing Uniformity. Axioms, 12(9), 887. https://doi.org/10.3390/axioms12090887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop