Next Article in Journal
Information Hiding Based on Statistical Features of Self-Organizing Patterns
Previous Article in Journal
Effectiveness of Biologically Inspired Neural Network Models in Learning and Patterns Memorization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymptotic Normality for Plug-In Estimators of Generalized Shannon’s Entropy

Department of Mathematics and Statistics, Mississippi State University, Mississippi State, MS 39762, USA
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(5), 683; https://doi.org/10.3390/e24050683
Submission received: 14 April 2022 / Revised: 3 May 2022 / Accepted: 11 May 2022 / Published: 12 May 2022
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
Shannon’s entropy is one of the building blocks of information theory and an essential aspect of Machine Learning (ML) methods (e.g., Random Forests). Yet, it is only finitely defined for distributions with fast decaying tails on a countable alphabet. The unboundedness of Shannon’s entropy over the general class of all distributions on an alphabet prevents its potential utility from being fully realized. To fill the void in the foundation of information theory, Zhang (2020) proposed generalized Shannon’s entropy, which is finitely defined everywhere. The plug-in estimator, adopted in almost all entropy-based ML method packages, is one of the most popular approaches to estimating Shannon’s entropy. The asymptotic distribution for Shannon’s entropy’s plug-in estimator was well studied in the existing literature. This paper studies the asymptotic properties for the plug-in estimator of generalized Shannon’s entropy on countable alphabets. The developed asymptotic properties require no assumptions on the original distribution. The proposed asymptotic properties allow for interval estimation and statistical tests with generalized Shannon’s entropy.

1. Introduction

1.1. Introduction and Related Work

Shannon’s entropy, introduced by [1], is one of the building blocks of Information Theory and a key aspect of Machine Learning (ML) methods (e.g., Random Forests). It is one of the most popular quantities on countable alphabet (An countable alphabet is a space that could be either finite, or countably infinite; the elements in an alphabet can be either ordinal (e.g., numbers) or non-ordinal (e.g., letters)), particularly on non-ordinal space with categorical data. For example, in [2], all reviewed feature selection methods on non-ordinal space boiled down to a function of Shannon’s entropy. In addition, Shannon’s entropy is one of the most important foundations for all tree-based ML algorithms, sometimes substitutable with the Gini impurity index [3,4,5].
As one of the essential information-theoretical quantities, Shannon’s entropy and its estimation are widely studied in the past decades [6,7,8,9,10,11,12]. In particular, [9] proved that an unbiased estimator of Shannon’s entropy does not exist. Current state-of-art Shannon’s entropy point estimator was provided in [10] with the fastest bias decaying rate (exponentially-decaying).
Nevertheless, Shannon’s entropy is only finitely defined for distributions with fast decaying tails [13].
It is never known if the real distribution yields a finite Shannon’s entropy in practice. Furthermore, all existing results on Shannon’s entropy require it to be finitely defined, which results in a usage restriction when adopting the entropy-based methods. This is, in fact, a void in the foundation of all Shannon’s entropy-related results.
Example 1
(Unbounded Shannon’s Entropy). Let a distribution P be P k = c / ( k ln 2 k ) for k 2 , where c is the constant that makes P a valid probability distribution. Such c uniquely exists because k = 2 [ 1 / ( k ln 2 k ) ] converges. Then Shannon’s entropy of P , H ( P ) is unbounded because
H ( P ) = k = 2 P k ln P k = k = 2 c k ln 2 k ln c k ln 2 k = k = 2 c k ln 2 k ln c + k = 2 c k ln 2 k ln k + k = 2 c k ln 2 k ln ln 2 k = k = 2 c k ln 2 k ln c + k = 2 c k ln k + k = 2 2 c ln ln k k ln 2 k = A Finite Value + + A Finite Value = .
The effort to generalize Shannon’s entropy has been long and extensive in the existing literature. As summarized in [14], the main perspective in the generalization in the existing literature is based on axiomatic characterization of Shannon’s entropy [15,16]. For example, Refs. [17,18] are efforts with respect to the functional form, H = k 1 h p k , under certain desirable axioms, h ( p ) = p log p is uniquely determined up to a multiplicative constant; if the strong additivity axiom is relaxed to be one of the weaker versions, say α -additivity or composability, then h ( p ) may be of other forms, which give rise to Rényi’s entropy [19], and the Tsallis entropy [20]. However, all such generalization effort does not seem to lead to an information measure on a joint alphabet that would possess all the desirable properties of mutual information, which is supported by an argument via the Kullback–Leibler divergence [21]. Interested readers may refer to [14] for details.
To further address the deficiency of Shannon’s entropy [14] proposed generalized Shannon’s entropy (GSE) and showed that GSE enjoys all properties of a finite Shannon’s entropy. In addition, GSE is finitely defined on all distributions. Due to the advantages of GSE and the deficiency of Shannon’s entropy, the use of Shannon’s entropy should eventually be transited to GSE.

1.2. Summary and Contribution

To aid the transition, the estimation of GSE needs to be studied. In practice, the plug-in estimator is one of the most popular estimation approaches. For plug-in estimation of GSE, asymptotic properties are required for statistical tests and confidence intervals. This article studies the asymptotic properties for plug-in estimators of GSE.
As a summary of the article’s results, Theorem 1 and Corollary 1 provide asymptotic normality properties for the plug-in estimators of GSE for all orders (An explanation of the order is given in Definition 2) on countably infinite alphabet. Corollary 2 provides the asymptotic normality properties for the plug-in estimators of GSE for all orders on finite alphabet, except the underlying distribution being uniform (Under a uniform distribution, the estimation of GSE is reduced to an estimation of population size. Interested readers can read [22]). The presented asymptotic normality properties immediately allow interval estimation and hypothesis testing with plug-in estimators of GSE. The numerical results in Section 3 show that the developed asymptotic properties converge fast, especially when the order is 2.
The presented properties and performance of GSE plug-in estimators suggest that GSE’s use is full of promising potential. One may be concerned the construction of CDOTC (Defined in Definition 1) would bring additional estimation challenges to the already-difficult estimation of Shannon’s entropy. Yet, the convergence speed for GSE plug-in estimators is fast. To further unlock the potentials of GSE, additional estimation methods of GSE and asymptotic properties of functions of GSE (e.g., Generalized Mutual Information, also originated in [14]) shall be visited. This article’s results and proofs’ approaches provide a solid direction toward that end.
The rest of this paper is organized as follows. Section 2 formally states the problem and gives our main results. In Section 3, we provide a small-scale simulation study. In Section 4, we discuss the potential of GSE. Proofs are postponed to Section 5.

2. Main Results

Let Z be a random element on a countable alphabet Z = z k ; k 1 with an associated distribution p = p k ; k 1 . Let the cardinality or support on Z be denoted K = k 1 1 p k > 0 , where 1 [ · ] is the indicator function. K is possibly finite or infinite. Let P denote the family of all distributions on Z . Shannon’s entropy, H, is defined as
H = H ( Z ) = k 1 p k ln p k .
To state our main result, we need to state Definitions 1 and 2 given by [14], and Definition 3.
Definition 1
(Conditional Distribution of Total Collision (CDOTC)). Given Z = z k ; k 1 and p = p k , consider the experiment of drawing an identically and independently distributed (iid) sample of size m ( m 2 ) . Let C m denote the event that all observations of the sample take on a same letter in Z , and let C m be referred to as the event of total collision. The conditional probability, given C m , that the total collision occurs at letter z k is
p m , k = p k m i 1 p i m ,
where m 2 . p m = p m , k is defined as the m-th order CDOTC.
Definition 2
(Generalized Shannon’s Entropy (GSE)). Given Z = z k ; k 1 , p = p k , and p m = { p m , k } , generalized Shannon’s entropy (GSE) is defined as
H m ( Z ) = k 1 p m , k ln p m , k ,
where p m , k is defined in Definition 1, and m = 2 , 3 , is the order of GSE. GSE with order m is referred to as the m-th order GSE.
It is clear that p m is a probability distribution induced from p = p k . Furthermore, for each m, m 2 , p and p m uniquely determined each other (Lemma 1 in [14]). To help understand Definitions 1 and 2, Examples 2 and 3 are provided as follows.
Example 2
(The 2nd order CDOTC). Given Z = z k ; k 1 and p = p k = { 6 k 2 / π 2 ; k = 1 , 2 , 3 , } , the 2nd order CDOTC is then defined as
p 2 = { p 2 , k } ,
where
p 2 , k = p k 2 i 1 p i 2 = 36 k 4 / π 4 i 1 36 i 4 / π 4 = k 4 i 1 i 4
for k = 1 , 2 , 3 , .
Example 3
(The 2nd order GSE). Given Z = z k ; k 1 , p = p k = { 6 k 2 / π 2 ; k = 1 , 2 , 3 , } , and p 2 = { p 2 , k } = { k 4 i 1 i 4 ; k = 1 , 2 , } , the 2nd order GSE, H 2 ( Z ) , is then defined as
H 2 ( Z ) = k 1 p 2 , k ln p 2 , k ,
where p 2 , k is given in Example 2.
The definition of the plug-in estimator of GSE is stated in Definition 3.
Definition 3
(Plug-in estimator of GSE). Let Z 1 , Z 2 , , Z n be independent and identically distributed (iid) random variables taking values in Z = z k ; k 1 with distribution p . For each k = 1 , 2 , , let Y k = i = 1 n 1 [ Z i = z k ] be the sample count of observations in category z k , and let p ^ k = Y k / n be the sample proportion. The plug-in estimator for the m-th order GSE, H ^ m ( Z ) , is defined as
H ^ m ( Z ) = k 1 p ^ m , k ln p ^ m , k = k 1 p ^ k m i 1 p ^ i m ln p ^ k m i 1 p ^ i m .
Our main results are stated in Theorem 1, Corollary 1 and 2.
Theorem 1.
Let p = { p k } be a probability distribution supported by a countably infinite alphabet, without any further conditions,
n H ^ m ( Z ) H m ( Z ) d N ( 0 , σ m 2 ) ,
where
σ m 2 = k = 1 m 2 p k p m , k ln p m , k + p m , k H m ( Z ) 2 .
Corollary 1.
Let p = { p k } be a probability distribution supported by a countably infinite alphabet, without any further conditions,
n H ^ m ( Z ) H m ( Z ) σ ^ m d N ( 0 , 1 ) ,
where
σ ^ m 2 = k = 1 m 2 p ^ k p ^ m , k ln p ^ m , k + p ^ m , k H ^ m ( Z ) 2 .
Corollary 2.
Let p = { p k ; k = 1 , 2 , , K } be a non-uniform probability distribution on a finite alphabet, without any further conditions,
n H ^ m ( Z ) H m ( Z ) σ ^ m d N ( 0 , 1 ) ,
where
σ ^ m 2 = k = 1 K m 2 p ^ k p ^ m , k ln p ^ m , k + p ^ m , k H ^ m ( Z ) 2 .
Corollary 2 is a special case of Theorem 1. All proofs are provided in Section 5.

3. Simulations

One of the main applications of our results is the ability to construct confidence intervals, and hence testing hypothesis. Specifically, Corollary 1 implies that an asymptotic ( 1 α ) 100 % confidence interval for H m is given by
H ^ m z α / 2 σ ^ m n , H ^ m + z α / 2 σ ^ m n ,
where σ ^ m is given by (1) and z α / 2 is a number such that P Z > z α / 2 = α / 2 and Z N ( 0 , 1 ) . In this section, we give a small scale simulation study to check the finite sample performance of this confidence interval.
We consider Zeta distribution
P ( x = k ) = 1 ζ ( s ) k s , k { 1 , 2 , }
with s = 1.5 and 2.5, where ζ ( s ) is the Riemann zeta function given by
ζ ( s ) = n = 1 1 n s .
The simulations were performed as follows. For the given distribution, we obtained a random sample of size n and used it to evaluate a 95 % confidence interval for a given index using (2). We then checked to see if the true value of the H m was in the interval or not. This was repeated 5000 times, and the proportion of times when the true value was in the interval was calculated. When the asymptotics works well, this proportion should be close to 0.95 . We repeated this for sample sizes ranging from 10 to 1000 in increments of 10. The results for s = 1.5 , order m = 2 and m = 3 are given in Figure 1 and Figure 2; the results for s = 2.5 , order m = 2 and m = 3 are given in Figure 3 and Figure 4.
The results suggest that convergence is fast, particularly when the order is m = 2 . We conjecture that this may be caused by the fact that, when m is larger, the probabilities in the corresponding CDOTC are smaller and hence require a larger sample size for convergence. For the same reason, the results with s = 1.5 converge faster than that of s = 2.5 , because s = 2.5 yields a thinner tail distribution which requires a larger sample size for convergence. Although GSE with order m 3 may shed some light on specific information, GSE with order m = 2 is enough to well exist with asymptotic properties for any valid underlying probability distribution p .

4. Discussion

The proposed asymptotic properties in Corollary 1 and 2 make it possible for interval estimation and statistical tests. Based on the simulation results, the convergence is quite fast, particularly under order m = 2 . Note that a GSE with order m = 2 already enjoys all asymptotic properties without any assumption on original distribution p .
We recommend using GSE with order m = 2 in place of Shannon’s entropy in all entropy-based methods when applicable. By replacing Shannon’s entropy with GSE, one still enjoys all the benefits of Shannon’s entropy with a fast convergence speed. Moreover, using GSE is risk-free compared to Shannon’s entropy because Shannon’s entropy (1) does not exist on some thick-tailed distributions and (2) requires thinner tail distribution for some asymptotic properties [11]. Additional research is required to aid the transition. The proposed asymptotic results allow interval estimation and statistical tests on the modified entropy-based methods that replaced Shannon’s entropy with GSE. Future research should aim to provide additional estimation methods of GSE and statistical properties of functions of GSE, such as GMI. The proposed asymptotic properties in this article directly provide asymptotic normality for the plug-in estimator of GMI when the real underlying GMI is not 0. The asymptotic behavior for the plug-in estimator of GMI when the real underlying GMI is 0 remains an open question, which we will address in future work.

5. Proofs

The proofs require several lemmas. The first lemma is state below.
Lemma 1
([11,23]). Assume that k = 1 p k log p k 2 < and that there is a deterministic sequence K ( n ) with K ( n ) such that lim n K ( n ) / n 0 and
lim n n k = K ( n ) p k log p k = 0 .
In this case
n H ^ n H d N 0 , σ 2 ,
where
σ 2 = k = 1 p k log p k 2 k = 1 p k log p k 2 .
Furthermore, if σ > 0 ,
n H ^ n H σ ^ n d N ( 0 , 1 )
where
σ ^ n 2 = k = 1 p ^ k log p ^ k 2 k = 1 p ^ k log p ^ k 2 .
Different proofs of Lemma 1 are provided in [11,23].
The spirit for proof of Theorem 1 is to regard CDOTC as an original distribution and utilize the result from Lemma 1. Toward that end, several lemmas are needed and stated below.
Lemma 2
(Equivalent conditions in Lemma 1). For any valid distribution p , let the corresponding CDOTC with order m be denoted as p m , then
k = 1 p m , k log p m , k 2 <
and that there is a deterministic sequence K ( n ) with K ( n ) such that lim n K ( n ) / n 0 and
lim n n k = K ( n ) p m , k log p m , k = 0 .
Lemma 3
( σ m 2 in Theorem 1). In Theorem 1,
σ m 2 = k = 1 m 2 p k p m , k ln p m , k + p m , k H m ( Z ) 2 .
Lemma 4
( σ ^ m 2 in Corollary 1). In Corollary 1,
σ ^ m 2 = k = 1 m 2 p ^ k p ^ m , k ln p ^ m , k + p ^ m , k H ^ m ( Z ) 2 .
Proof of Lemma 2.
Note that for any p to be a valid distribution, the tail of p must be thicker than 1 / ( k ln k ) because k 2 1 / ( k ln k ) diverges. Hence p m is thicker than 1 / ( k 2 ln 2 k ) for any m 2 by definition. It is shown in Example 3 of [11] that such tail satisfies the mentioned conditions. □
Proof of Lemma 3.
Because of Lemma 2, σ 2 can be obtained under finite K and then let K . For a finite K, it can be verified that for i = 1 , , K 1 ,
H m p i = ln p m , K ln p m , i m p m , i p i m p m , i p i p m , K p K H m + ln p m , K .
Let
v = p 1 , , p K 1 τ , v ^ = p ^ 1 , , p ^ K 1 τ .
We have n ( v ^ v ) L M V N ( 0 , Σ ( v ) ) , where Σ ( v ) is the ( K 1 ) × ( K 1 ) covariance matrix given by
Σ ( v ) = p 1 1 p 1 p 1 p 2 p 1 p K 1 p 2 p 1 p 2 1 p 2 p 2 p K 1 p K 1 p 1 p K 1 p 2 p K 1 1 p K 1
According to the first-order Delta method,
σ K 2 = H m T Σ H m = k = 1 K m 2 p k p m , k ln p m , k + p m , k H m ( Z ) 2 .
Given Lemma 2, let K ,
σ 2 = k = 1 m 2 p k p m , k ln p m , k + p m , k H m ( Z ) 2 .
 □
Proof of Lemma 4.
Lemma 4 is because of σ ^ m 2 p σ m 2 . □
Proof of Theorem 1 and Corollary 1.
With Lemmas 1–4, and Slutsky’s theorem, Theorem 1 and Corollary 1 are proved. □
Proof of Corollary 2.
Corollary 2 is a directly result of Theorem 1, except under uniform distribution when H m = 0 for all m 2 . □

Author Contributions

Conceptualization, J.Z. and J.S.; methodology, J.Z. and J.S.; simulation, J.Z. and J.S.; writing—original draft preparation, J.Z.; writing—review and editing, J.Z. and J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

We thank Zhiyi Zhang at the University of North Carolina at Charlotte for their profound discussions on the subject of Entropic Statistics.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CDOTCConditional Distribution of Total Collision
GMIGeneralized Mutual Information
GSEGeneralized Shannon’s Entropy
MLMachine Learning

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  2. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef]
  3. Banerjee, M.; Reynolds, E.; Andersson, H.B.; Nallamothu, B.K. Tree-based analysis: A practical approach to create clinical decision-making tools. Circ. Cardiovasc. Qual. Outcomes 2019, 12, e004879. [Google Scholar] [CrossRef] [PubMed]
  4. Mienye, I.D.; Sun, Y.; Wang, Z. Prediction performance of improved decision tree-based algorithms: A review. Procedia Manuf. 2019, 35, 698–703. [Google Scholar] [CrossRef]
  5. Hssina, B.; Merbouha, A.; Ezzikouri, H.; Erritali, M. A comparative study of decision tree ID3 and C4. 5. Int. J. Adv. Comput. Sci. Appl. 2014, 4, 13–19. [Google Scholar]
  6. Miller, G.A.; Madow, W.G. On the Maximum Likelihood Estimate of the Shannon-Weiner Measure of Information; Operational Applications Laboratory, Air Force Cambridge Research Center: Bedford, MA, USA, 1954. [Google Scholar]
  7. Harris, B. The Statistical Estimation of Entropy in the Non-Parametric Case; Technical Report; Wisconsin Univ-Madison Mathematics Research Center: Madison, WI, USA, 1975. [Google Scholar]
  8. Esty, W.W. A normal limit law for a nonparametric estimator of the coverage of a random sample. Ann. Stat. 1983, 11, 905–912. [Google Scholar] [CrossRef]
  9. Paninski, L. Estimation of entropy and mutual information. Neural Comput. 2003, 15, 1191–1253. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, Z. Entropy estimation in Turing’s perspective. Neural Comput. 2012, 24, 1368–1389. [Google Scholar] [CrossRef] [PubMed]
  11. Zhang, Z.; Zhang, X. A normal law for the plug-in estimator of entropy. IEEE Trans. Inf. Theory 2012, 58, 2745–2747. [Google Scholar] [CrossRef]
  12. Zhang, Z. Asymptotic normality of an entropy estimator with exponentially decaying bias. IEEE Trans. Inf. Theory 2013, 59, 504–508. [Google Scholar] [CrossRef]
  13. Baccetti, V.; Visser, M. Infinite shannon entropy. J. Stat. Mech. Theory Exp. 2013, 2013, P04010. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, Z. Generalized Mutual Information. Stats 2020, 3, 158–165. [Google Scholar] [CrossRef]
  15. Amigó, J.M.; Balogh, S.G.; Hernández, S. A brief review of generalized entropies. Entropy 2018, 20, 813. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Csiszár, I. Axiomatic characterizations of information measures. Entropy 2008, 10, 261–273. [Google Scholar] [CrossRef] [Green Version]
  17. Khinchin, A.Y. Mathematical Foundations of Information Theory; Courier Corporation: Chelmsford, MA, USA, 2013. [Google Scholar]
  18. Chakrabarti, C.; Chakrabarty, I. Shannon entropy: Axiomatic characterization and application. Int. J. Math. Math. Sci. 2005, 2005, 2847–2854. [Google Scholar] [CrossRef] [Green Version]
  19. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkley, CA, USA, 1 January 1961; University of California Press: Los Angeles, CA, USA, 1961; Volume 4, pp. 547–562. [Google Scholar]
  20. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  21. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  22. Zhang, Z.; Chen, C.; Zhang, J. Estimation of population size in entropic perspective. Commun.-Stat.-Theory Methods 2020, 49, 307–324. [Google Scholar] [CrossRef]
  23. Grabchak, M.; Zhang, Z. Asymptotic normality for plug-in estimators of diversity indices on countable alphabets. J. Nonparametr. Stat. 2018, 30, 774–795. [Google Scholar] [CrossRef]
Figure 1. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 1.5 and GSE with order m = 2 . The horizontal dashed line is at 0.95.
Figure 1. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 1.5 and GSE with order m = 2 . The horizontal dashed line is at 0.95.
Entropy 24 00683 g001
Figure 2. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 1.5 and GSE with order m = 3 . The horizontal dashed line is at 0.95.
Figure 2. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 1.5 and GSE with order m = 3 . The horizontal dashed line is at 0.95.
Entropy 24 00683 g002
Figure 3. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 2.5 and GSE with order m = 2 . The horizontal dashed line is at 0.95.
Figure 3. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 2.5 and GSE with order m = 2 . The horizontal dashed line is at 0.95.
Entropy 24 00683 g003
Figure 4. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 2.5 and GSE with order m = 3 . The horizontal dashed line is at 0.95.
Figure 4. Effectiveness of the 95% confidence intervals as a function of sample size. Simulations from Zeta distribution with s = 2.5 and GSE with order m = 3 . The horizontal dashed line is at 0.95.
Entropy 24 00683 g004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, J.; Shi, J. Asymptotic Normality for Plug-In Estimators of Generalized Shannon’s Entropy. Entropy 2022, 24, 683. https://doi.org/10.3390/e24050683

AMA Style

Zhang J, Shi J. Asymptotic Normality for Plug-In Estimators of Generalized Shannon’s Entropy. Entropy. 2022; 24(5):683. https://doi.org/10.3390/e24050683

Chicago/Turabian Style

Zhang, Jialin, and Jingyi Shi. 2022. "Asymptotic Normality for Plug-In Estimators of Generalized Shannon’s Entropy" Entropy 24, no. 5: 683. https://doi.org/10.3390/e24050683

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop