Next Article in Journal
On the Performance of Video Resolution, Motion and Dynamism in Transmission Using Near-Capacity Transceiver for Wireless Communication
Next Article in Special Issue
Information Theory Based Evaluation of the RC4 Stream Cipher Outputs
Previous Article in Journal
Entropy Based Student’s t-Process Dynamical Model
Previous Article in Special Issue
Refined Young Inequality and Its Application to Divergences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy

by
Lianet Contreras Rodríguez
1,
Evaristo José Madarro-Capó 
1,
Carlos Miguel Legón-Pérez 
1,
Omar Rojas
2,* and
Guillermo Sosa-Gómez
2
1
Facultad de Matemática y Computación, Instituto de Criptografía, Universidad de la Habana, Habana 10400, Cuba
2
Facultad de Ciencias Económicas y Empresariales, Universidad Panamericana, Álvaro del Portillo 49, Zapopan, Jalisco 45010, Mexico
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(5), 561; https://doi.org/10.3390/e23050561
Submission received: 5 April 2021 / Revised: 26 April 2021 / Accepted: 28 April 2021 / Published: 30 April 2021
(This article belongs to the Special Issue Types of Entropies and Divergences with Their Applications)

Abstract

:
Entropy makes it possible to measure the uncertainty about an information source from the distribution of its output symbols. It is known that the maximum Shannon’s entropy of a discrete source of information is reached when its symbols follow a Uniform distribution. In cryptography, these sources have great applications since they allow for the highest security standards to be reached. In this work, the most effective estimator is selected to estimate entropy in short samples of bytes and bits with maximum entropy. For this, 18 estimators were compared. Results concerning the comparisons published in the literature between these estimators are discussed. The most suitable estimator is determined experimentally, based on its bias, the mean square error short samples of bytes and bits.

1. Introduction

Entropy allows the measurement of the uncertainty about an information source from the distribution of its output symbols [1]. The Shannon’s entropy of a discrete source of information will be maximum if the source transmits symbols that follow a Uniform distribution; that is, the highest level of uncertainty.
In most cases, the data for the measurements are obtained but the distribution is unknown, so it is necessary to estimate the parameters. Even knowing the data distribution, it is necessary to determine which is the best performing entropy estimator since there is no single ideal estimator for all the distributions that may occur. On the other hand, on many situations, there are not enough samples with sufficient sizes compared to the source alphabet, in certain scenarios called undersample regime, where many estimators underestimate the real entropy [2,3]. It is well known that there is not an unbiased estimator [4], and the convergence rate of a consistent estimator can be arbitrarily slow [5]. However, there are many estimators, but deciding which estimator to use depending on the scenario turns out to be a practical problem [6].
Estimating a random variable’s entropy is of great importance and with many applications, having received much attention in recent years [7,8,9,10,11]. Most of the known estimators are generally applied for the estimation of mutual information [12], but in this scenario entropy estimators will perform similarly, and thus the selection between different estimators is less of an issue. However, entropy has great application in cryptography as a vital tool for designing and analyzing encryption methods [13,14,15,16,17]. Generally, in cryptography, one usually deals with large sample volumes [18,19,20,21,22], so the estimation of entropy is often useless, as a very similar behavior is obtained for all estimators. However, it is possible to find scenarios in cryptography where the estimation of entropy is vital since it works with smaller samples [23,24]. In these cases, it is necessary to have the estimator with the highest convergence rate and the lowest mean square error to increase the result’s precision.
There is a wide variety of entropy estimators [3,5,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45] tested in dissimilar scenarios. However, it is not easy to find results of applying or selecting entropy estimators in samples uniformly distributed composed of bytes and bits. In this work, the most suitable method to estimate entropy in short samples of bytes and bits uniformly distributed is determined experimentally from the bias and mean square error characteristics. For this, 18 estimators reported in the literature were compared, and the results are discussed concerning the literature found. The structure of the paper is as follows: Section 2 presents some preliminaries about entropy and comparison criteria for the estimators; Section 3 discusses some of the entropy estimators from the literature; Section 4 presents the main results, which have to do with the selection of the entropy estimators; finally, Section 5 presents some conclusions and possible future lines of work.

2. Preliminaries

2.1. Shannon Entropy

Shannon’s entropy is a measure of a random variable’s uncertainty. Let X be a discrete random variable with the alphabet K and the probability function p x = P r { X = x } , x K , then we have the following
Definition 1.
The entropy H X of a discrete random variable X is defined as [1]:
H X = x K p x log 2 p x .
A discrete source (source with a finite alphabet) with alphabet K reaches its maximum entropy H max if its output symbols follow a uniform distribution. In this case, any discrete random variable on this alphabet has an entropy no greater than log 2 | K | [1]. On the other hand, the lowest value of the entropy H min = 0 is reached when, for some x K one has that p x = 1 . The maximum possible value for the entropy of a discrete random variable with k different values (where | K | = k ) H max = log 2 k is reached when the random variable has a uniform probability distribution. Thus, since the objective is to select an effective estimator of entropy for sequences of bytes and bits with maximum entropy, the selected estimator’s results will be compared with the expected value of entropy for uniformly distributed samples, H max .
In [46], a lower bound on the convergence rate of an optimal estimator is provided. In there, they theoretically demonstrated that with a sample of size k / log 2 k the entropy can be estimated, where k is the alphabet’s size. The undersample regime happens when the size n of the samples is less than the size k of the alphabet and is denoted: n < k [47]. The size of the alphabet for the case of byte samples is 256 and for the case of samples of 2-bit. Thus, the entropy value can vary from H min = 0 to H max = 8 and from H min = 0 to H max = 1 , for byte and bit samples, respectively. This work focuses on samples with sizes greater than or equal to n = 8 bytes and bits, respectively. In the case of bit samples, it was decided to work with 8 bits because this constitutes the smallest value of n most frequent in current encryption algorithms.

2.2. Comparison Criterion between the Estimators of H

As mentioned before, it is well known that there is not an unbiased estimator [4] and of minimum variance, while the convergence rate of a consistent estimators can be arbitrarily slow [5]. Two well-known statistical tools of estimation theory will be used to experimentally evaluate the selected estimators, bias and mean square error [48]. The bias S H ^ of an estimator H ^ measures the deviation of the estimate with respect to the expected real value and is calculated from the difference S H ^ = E H H ^ , where E H ^ is the expected value of H ^ . In our scenario, we will work with uniformly distributed samples of bytes and bits so the data follows a uniform distribution and we have that E H ^ = H max .
The mean square error (MSE) of an estimator measures the average of the squared errors, that is, the difference between the estimator and what is estimated. The difference occurs by two possible reasons—by randomness or because the estimator does not consider the information that could produce a more accurate estimate. The MSE is a measure to determine the quality of an estimator with a non-negative value. If the value of the MSE is close to zero, then it implies that the quality of the estimator is better. For an unbiased estimator, the MSE is the variance of the estimator. That is, if the variance of an estimator is smaller among all the estimators, then it will be considered as the best unbiased estimator. The MSE is equal to the sum of the variance and the square of the estimator bias,
MSE H ^ = V a r H ^ + S H ^ 2 .
Thus, the MSE assesses an estimator’s quality in terms of its variance and the degree of bias.

3. Entropy Estimators

Various entropy estimators have been proposed in the literature with the aim of approaching its theoretical value in practice. This section briefly describes 18 entropy estimators present in several works published in the literature. The empirical approach estimates entropy from the observed individual and joint frequencies for each bit. The maximum likelihood estimator (ML) of entropy is defined as follows
H ^ M L = x K p ^ x M L log 2 p ^ x M L .
The probability of appearance in the sample with respect to the observed frequencies y x for each x K is specified as p ^ x M L = y x / n , where n is the sample size. Table 1 describes the others 17 entropy estimators that will be used.
Apart from the entropy estimators described, there are many others in the literature, such as the KDE [36], KNN [39], B-Spline [35], Edqeworth [37], Polynomial [38] and the Bayesian PYM and DPM [6]. This shows that entropy is a very well studied information theory tool in the current literature.

3.1. Theoretical Approximations between Estimators of Entropy

It is well known that there are relationships between various estimators depending on the scenarios that are being worked on. These can coincide based on their formulation and on the values of their parameters, such as the probability distribution, the size of the alphabet, or eigenvalues used in the formulation of the estimator. There is a relation between the SHR and Bayes estimators [32]. Using t k = a k / A and λ = A / n + A , then p ^ x S H R = p ^ x B a y e s . This show that the SHR estimator is an empirical Bayes estimator based on the t k and λ selected values. For a choice with no shrinkage ( λ = 0 ), p ^ x S H R = p ^ x M L is obtained and ML estimator is achieved. On the other hand, if  a k = 0 is selected for Bayes estimator, then p ^ x B a y e s = p ^ x M L and again ML estimator is achieved. In addition, for bit samples the SG estimator must coincide with the JEFF estimator because in this case k = 2 . In [50], the Zhang estimator was classified within the family of entropy estimators originally introduced in [42]. Also, in the asymptotic regime y x 1 , the estimator given for the choice ξ = 1 within the family of entropy estimators in [42] leads to the MM estimator using the asymptotic relation ψ x log x 1 / 2 x . The BUB estimator leads exactly to a version of the Miller-Madow correction and gives another angle on why this correction fails in the N m regime [4]. In this perspective, the MLE estimator can be expressed as
H ^ M L E = i = 0 n a i h i ,
where a i = i n log i n .

3.2. Previous Work on Comparison of Entropy Estimators

Many of the entropy estimators were thought for specific contexts, so there has been a tendency to compare which estimator would be the most suitable to be used in each scenario. The following reports present various comparisons between entropy estimators described in the literature. This allows a vision of the best behavior estimators even when the used scenarios do not coincide with those of this paper. In [4], four different estimators were analyzed from various statistical points of view, such as the central limit theorem’s consistency, bias, and variance. In samples of real and simulated spike trains, it was obtained that the behavior of the estimation with the BUB approach is better than with the use of the ML, MM, and Jackknife methods.
In [32], nine estimators (MM, ML, NSB, CS, Shrink, MIN, SG, JEFF and LAP) on the Dirichlet distribution, with different structures and values of the parameter a, and the Zipf-type power law were compared. In summary, the three best-performing estimators are the NSB, CS and Shrink estimators, with very similar behavior in the scenarios presented. However, in his simulations, he emphasizes that the Shrink estimator is the only one that estimates probabilities with high precision for the use of Shannon entropy, even for small samples. On the other hand, the rest of the estimators can be considered with similar behavior in several scenarios. Specifying that the Bayesian estimators can have better or worse behavior than the ML estimator, depends on selecting the prior and the sample scenario but without a marked difference. In the case of Bayesian estimators, there is no general agreement on which allocation of a is better a priori. However, as discussed in [32], choosing an inappropriate a can cause the resulting estimator to perform worse than the ML estimator, influencing the practical results.
In [33], six entropy estimators—ML, MM, Jackknifed, CS, BUB and Unseen—were compared. The samples were drawn from six different classes of distributions—the uniform distribution, a mixture of two uniform distributions; the Zipf distribution, which is commonly used to model naturally occurring “power law” distributions, particularly in natural language processing, a modified Zipf distribution with 0.6 power-law exponent, the geometric distribution, and finally, a uniform mixture of Geometric and Zipf. For each distribution, three configurations of the size k of the alphabet were considered: 1000, 10,000, and 100,000. The sample size, n, varied in the interval k 0.6 , k 1.25 . It was found, as a result, that the three best-performing estimators were the Unseen, the CS, and the BUB, where the Unseen estimator being ranked above these three estimators. Specifically for larger sample sizes and in samples from the Zipf distribution with respect to the CS estimator and for the uniform and Zipf distributions with respect to the BUB estimator. The Unseen estimator performance on all classes of distributions is significant since the algorithm proposed for its computation is designed to compute a representation of the distributions rather than specifically adapted to estimate entropy.
In [51], the effects of entropy estimation on mutual information and data discretization in the inference of gene regulatory networks were investigated. Here, they evaluated the performance of the MM, ML, SG, and Shrink estimators with three different discretization approaches (equal frequency, equal amplitude, equal global amplitude) and stated that changing the discretization approach influences the result more than changing the estimators. They indicated that using the MM estimator with the discretizations of equal amplitude and equal global amplitude achieves the best inference result. Samples of size 50, 100, 200, 500, 1000 were used in the experiments. In [34], some experiments were run with 6 estimators ML, Jackknife, NSB, Zhang, and two corrections of the Zhang estimator (correction proposed in [33]) and (Jackknife correction). The comparison of the estimators was made in three categories: finite alphabets with known cardinalities, finite alphabets with unknown cardinalities, and infinite alphabets. For each case in the simulation, the bias reported was based on the average of 5000 simulated samples. The sample size varied from ten to several hundred. The jackknifing was done using all subsamples of sizes m, where 5 < m < n . The experiments were carried out on six distributions, Zipf, Poisson, Gauss and variations of these.
The ML estimator is dominated by the Zhang estimator and the Jackknife estimator by the correction of the Zhang estimator with the Jackknife. The NSB estimator has a better behavior but very similar to the Zhang estimator with the Jackknife correction. It is highlighted that in this case the Jackknife correction provides a better performance to the Zhang estimator than with its proposed correction. It is reported that the NSB has a tendency to overestimate the entropy value and they expose it in cases with the Poisson distribution and the Gaussian distribution with certain characteristics. In [43], five estimators (Jackknife, Zhang with correction, CS, CWJ and GSB) were compared, modeling species abundance distributions. The number of species was fixed to be 100, 500 or 1000 and seven species abundance models were considered, a homogeneous model, the log-normal model, the Zipf–Mandelbrot model, the broken-stick model, power-decay model, Poisson model and exponential-decay model. For each fixed model, considered a range of sample sizes ( n = 25 500 in an increment of 25 if S < 500 , and  n = 50 1000 in an increment of 50 if S 500 ) . For each combination of abundance model and sample size, 5000 simulated data sets were generated from the model. As results, the GSB, Jackknife and Zhang’s estimators are biased downwards and the three estimators have similar trends. The GSB estimator is less biased than the Jackknife, which is less biased than the Zhang’s estimator. The RMSE shows the same pattern. On the other hand, CWJ estimator in nearly all cases is better than the GSB, Jackknife and Zhang’s estimators based on the criteria of bias and RMSE. The CS estimator generally has smaller bias and RMSE than the GSB estimator.
The results described in this section corroborate the problem of selecting the ideal entropy estimator for a particular practical case. In summary, these results highlight how factors such as sample size, sample distribution, alphabet size, sample discretization method, and optimal selection of distribution parameters can influence the behavior of the estimators or the optimal selection of parameters in the calculation of the estimator. The comparison of entropy estimators presented in this work differs from the previous ones in several respects. First, 18 estimators are compared, a figure much higher than in previous studies. Second, it is limited to discrete samples, specifically of bits and bytes from the uniform distribution, which are of great cryptographic importance, so discretization is unnecessary. Third, the greatest interest is in the short sample regime for byte samples when the sample size is less than the alphabet’s size. However, for the bit case, it has also experimented with short samples. Fourth, of the 17 compared estimators, 15 are present in previous works, while 2 (CDM and BN) provide new comparisons. Fifth, an experimental correlation study between entropy estimators in this scenario is presented for the first time.

4. Selecting an Effective Entropy Estimator through Experimental Evaluation

In this section, the 18 estimators were applied to short uniformly distributed samples of bytes and bits obtained using the Linux urandom random number generator [52]. Then, the results obtained through their bias and their mean square error were compared, illustrating the behavior of their characteristics using some plots. Also, the selection of the most effective estimator as a result of the comparison made is discussed. To carry out the experiments, 1000 samples of uniformly distributed sequences of bytes and bits were generated for each of the sizes 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, and 16,384. In [53], it shows how the output sequences of random number generators must follow a uniform distribution. Thus, the values of the estimators will be compared with the expected theoretical value for this distribution (for uniformly distributed samples in bytes, k = 256 and H m a x = 8 , and for the case of uniformly distributed bit samples k = 2 and H m a x = 1 ). Cases in the regime of short samples were analyzed, and situations when the sample size exceeds the size of the alphabet, that is,  n > k , managing to visualize the convergence of the estimators.

4.1. Implementation of Entropy Estimators

The estimators ML, MM, JEF, LAP, SG, MIN, CS, SHR are available in the R software package Entropy [53]. The entropy function of that package allows estimating entropy from observed counts. The Zhang estimator’s estimates were made with the EntropyEstimation package [54] of the R software. The Matlab implementation of the BUB estimator provided by the author [4] was used, and its numerical adjustment parameters were left. Likewise, for the Unseen estimator, the author’s Matlab implementation in [33] was used. For the NSB estimator, the Python implementation proposed in [55] was used. The SHU, UnveilJ, GSB, CWJ, and BON estimators can be found in the package Entropart [56] for R. While the CDM estimator is part of the CDMEntropy project implemented in Matlab [57].

4.2. Analysis of Bias between Estimators

In this section, the bias behavior of the 18 estimators is analyzed. Figure 1 shows the behavior of the estimated mean of the estimators for each sample size for uniformly distributed samples of bytes. The straight red line represents the expected theoretical value, that is, H m a x = 8 .
The behavior of the 18 estimators in the uniformly distributed samples of bytes is similar in terms of convergence towards the expected theoretical value. However, there are differences between them in terms of the speed at which they converge. When the size n of the samples is less than the alphabet’s size k = 256 , there is a deviation in many of the estimates from the expected theoretical value. Figure 1 shows three estimators (SHR, LAP, JEF) that are considerably close, when n < k , to the expected theoretical value H m a x = 8 . In this group, the SHR estimator stands out for reaching values very close to H m a x = 8 for all sample sizes, as corresponds to a uniformly distributed sample with a uniform distribution. Table 2 shows the average value of H ^ for each estimator in all the sizes used. The LAP and JEF estimators show values very close to H m a x up to n = 32 , however, for n = 256 , their value decreases a little.
The rest of the estimators show a similar trend towards the expected H m a x value. However, the behavior of the second group of estimators (CS, CJ, BUB, NSB, Unseen, CDM) stands out in terms of the speed of convergence towards H m a x . On the other hand, the BN estimator’s unusual behavior is highlighted from n > k = 256 , coincidentally the size of the alphabet, which confirms that the estimator was constructed for short samples [40]. Furthermore, the CS and CJ estimator show a tendency to overestimate the expected theoretical value in some sample sizes. Starting from n = 2048 approximately, the values of all estimators adjust to the expected theoretical value H m a x = 8 . Table 3 shows the difference ( H m a x H ^ ) that represents an estimate of the bias in each of the estimators for byte samples. This measure gives the deviation of the expected value of an estimator concerning the real value.
Table 3 confirms the SHR estimator as the one with the best performance in this scenario, with almost zero bias for all sizes. In bit samples, a very similar convergence speed is observed for all estimators (see Figure 2). However, a group of estimators stands out, the cases of Zhang, CJ, GSB, SHU, and MM, with values very close to the expected value and faster convergence to it.
To accompany Figure 2 in the visual description of each estimator’s behavior, Table 4 presents the average H ^ value of each of the estimators for all sample sizes.
This group is followed by the SHR and BUB estimators, highlighting the BUB estimator with a very fast convergence towards H m a x . Approximately, starting at n = 512 , the values of all estimators adjust to the expected theoretical value H m a x = 1 . Table 5 shows the difference ( H m a x H ^ ) that represents an estimate of the bias in each of the estimators for bit samples.
Table 5 shows the good behavior of all estimators in this scenario. Unlike byte samples in this case, as mentioned above, the estimators show very similar behavior. It is known that the estimated mean can hide behaviors. For this reason, it is necessary to perform an analysis of the variance of these estimators to confirm that the average value of H ^ correctly describes the behavior of the estimator in each scenario.
Figure 3 and Figure 4 show the dispersion of the estimated values in the samples of uniformly distributed sequences of bytes and bits, respectively, for each of the chosen sizes. Figure 3 confirms the overestimation of the estimators CS and CJ expected theoretical value for various sizes of n (see Figure 3c–e). Furthermore, it shows how the Unseen estimator also has some over estimated values of H ^ that are not visible through the average value of H ^ . Figure 4 shows how for bit samples, the estimators have greater variance. The SHR estimator with the lowest variance and with a concentration of its values closer to H m a x stands out. On the contrary, the GSB estimator is highlighted as the highest variance with a very unstable dispersion. Although this estimator performs well on average, the variance shows a considerable dispersion of its values.

4.3. Comparison of Estimators in Terms of Mean Square Error

Another of the most used tools to measure an estimator’s quality is the mean square error (MSE). Figure 5 illustrates the estimate’s mean square error in uniformly distributed samples of bytes and bits for each size.
Figure 5a shows a group of seven estimators (LAP, JEF, BUB, NSB, SHR, CS and CDM) with a greater tendency towards MSE = 0 . Table 6 shows the MSE values for each of the sample sizes.
The SHR estimator has the best behavior, with an MSE very close to 0 for all sizes. This result establishes the SHR estimator as the most suitable, lowest bias, and lowest MSE for uniformly distributed samples of bytes. On the other hand, in Figure 5b, a clear difference is not observed, but 5 estimators can be distinguished (LAP, JEF, CS, MIN, SHR) with better convergence (see Table 7).
It is more difficult to select a suitable single estimator in bit samples since all estimators show good results in terms of bias and MSE value. However, the MM and SHR estimators stand out in both. Therefore, in this scenario, the use of either of these two estimators is suggested, specifying that MM has less bias and SHR has less MSE. In this case, the estimators have similar efficiency. It would be of interest to compare its efficiency since looking for a balance between effectiveness and efficiency, it would be possible to define which estimator to select.

4.4. Correlation between Estimators of Entropy Using Bias

Figure 1 suggests a trend or approach in the distribution of the estimated means of various estimators, which could suggest the existence of a correlation between them. Thus, in this section, the correlation between the estimators will be measured using Kendall’s correlation coefficient. In Figure 6, the Kendall correlation coefficient values between the estimators for random samples of bytes can be observed.
The size of the circles and the intensity of the color are in correspondence with the correlation value existing between the estimators. From the results shown in Figure 6, groups of estimators with higher correlation can be analytically determined. Considering the values of Kendall’s correlation coefficient (CCK) [58], the following four groups are obtained, grouping them by having a correlation value greater than 0.6 or less than −0.6.
  • { JEF, LAP, BN},
  • { MM, BUB, ML, SG, MIN, Zhang, NSB, CDM, UnveilJ, GSB, SHU},
  • { CS, Unseen, CJ}.
  • { SHR}.
The third and fourth group’s estimators stand out for their rapid convergence in byte samples; therefore, it is suggested in terms of bias to use one of these estimators in this scenario. However, the SHR estimator stands out for its excellent behavior for all sample sizes. This grouping of the estimators was corroborated using the hierarchical grouping method to determine the highest correlation groups. Hierarchical grouping algorithms are used for grouping patterns whose internal organization is unknown; that is, there is no knowledge about the class label to which they belong. In data mining, these data groups are called clusters. Figure 7 shows the clustering performed using the full link agglomerative method [59,60] with the distance based on the correlation, in this case, the Kendall correlation.
The BN estimator is in a different group, possibly because its correlation with the JEF and LAP estimators is in the opposite direction, that is, negative. Furthermore, some of these groups can be divided consecutively into subsets such that the value of the coefficient has less variation within each subset. Figure 8 shows the dendrogram corresponding to the hierarchical grouping of the estimators where this characteristic is displayed.
For bit samples, the behavior between all estimators is much more similar. However, a similar correlation analysis was performed. In Figure 9, the Kendall correlation coefficient values between the estimates for uniformly distributed samples of bits can be observed.
Figure 9 shows a division into fewer subgroups than for byte samples; this confirms no significant distinction in terms of bias in this scenario between the estimators. However, when applying the full link agglomerative method with correlation-based distance, it is obtained that it identifies four groups of estimators:
  • { NSB, CDM, BN},
  • { ML, SG, MIN, JEF, LAP, UnveilJ, CS, Unseen},
  • { GSB}.
  • { SHU, MM, CJ, Zhang, BUB, SHR}.
Figure 10 confirms the discussion raised about four of the estimators (CJ, SHU, MM, Zhang) with less bias in uniformly distributed samples of bits and, in turn, shows that although the average of H ^ for the GSB estimator has satisfactory behavior, the values of this estimator are far from the behavior of the estimators of this group. This is the consequence of its high variance value. This case illustrates how the observation’s average can hide inappropriate behaviors and the need to use various statistical tools together.
The SHR and BUB estimators are not displayed in Figure 1 within the estimator’s group with the best bias behavior. However, the agglomerative method detects that these two estimator’s values have a trajectory more similar to this group and superior to the other estimators. The relationship between the Bayesian estimators JEF, SG, MIN, and LAP is highlighted, which can be seen in Figure 11. Highlighting the relationship mentioned on the SG and JEF estimators in this scenario in Section 3.1.
The estimators { S H U , M M , C J , Z h a n g , B U B , S H R } have similar behavior. It would be of interest to compare their efficiency because, in applications that require an extensive estimation of entropy, it is possible to evaluate using the most efficient of all of them, seeking a balance between effectiveness and efficiency. This comparison is left as an open problem for future work as it requires a unique own implementation that includes all these estimators.

5. Conclusions

In this work, a comparison was made between 18 estimators of entropy in short uniformly distributed samples of bytes and bits, based on their bias, variance, and mean square error. It was concluded that to estimate entropy in short uniformly distributed samples of bytes, the SHR estimator will be the most effective, while for bit samples, it is proposed to use the MM and SHR estimators. To estimate the entropy of random samples of bytes and bits, even considering the regime in short samples, based on the results obtained in this work, it is recommended to use the SHR estimator. The SHR estimator was the estimator with the least bias in random samples of bytes and the one with the highest convergence towards a mean square error equal to 0. For random samples of bits, it is within the set with the least bias, and within these, it is the estimator with the lowest dispersion in the values of the estimates with greater convergence towards an MSE = 0. It is noteworthy that the estimator is effective even for samples with the smallest sizes in both scenarios. This result is of great importance in cryptographic applications where scenarios with tiny samples often arise.
Additionally, the 18 estimators were grouped in clusters (different for each scenario) such that all the estimators of a cluster have similar effectiveness. This suggests that in applications that require intensive estimation of entropy, a selected estimator could be substituted for another correlated one, but of higher efficiency, which suggests comparing the efficiency of the estimators { C S , U n s e e n , C } { S H R } in uniformly distributed samples of bytes and the estimators { S H U , M M , C J , Z h a n g , B U B , S H R } in uniformly distributed samples of bits, which will be addressed in future works.

Author Contributions

Conceptualization, L.C.R. and E.J.M.-C.; methodology, L.C.R., E.J.M.-C., C.M.L.-P., O.R. and G.S.-G.; software, L.C.R.; validation, L.C.R., E.J.M.-C., C.M.L.-P., O.R. and G.S.-G.; formal analysis, E.J.M.-C., C.M.L.-P., O.R. and G.S.-G.; investigation, L.C.R., E.J.M.-C., C.M.L.-P. and G.S.-G.; writing—original draft preparation, L.C.R. and E.J.M.-C.; writing—review and editing, L.C.R., E.J.M.-C., C.M.L.-P., O.R. and G.S.-G.; visualization, L.C.R.; supervision, E.J.M.-C., C.M.L.-P., O.R. and G.S.-G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  2. Verdú, S. Empil Estimation of Information Measures: A Literature guide. Entropy 2019, 21, 720. [Google Scholar] [CrossRef] [Green Version]
  3. Vu, V.Q.; Yu, B.; Kass, R.E. Coverage-adjusted entropy estimation. Stat. Med. 2007, 26, 4039–4060. [Google Scholar] [CrossRef]
  4. Paninski, L. Estimation of entropy and mutual information. Neural Comput. 2003, 15, 1191–1253. [Google Scholar] [CrossRef] [Green Version]
  5. Antos, A.; Kontoyiannis, I. Convergence properties of functional estimates for discrete distributions. Random Struct. Algorithms 2001, 19, 163–193. [Google Scholar] [CrossRef]
  6. Archer, E.; Park, I.M.; Pillow, J.W. Bayesian entropy estimation for countable discrete distributions. J. Mach. Learn. Res. 2014, 15, 2833–2868. [Google Scholar]
  7. Timme, N.M.; Lapish, C. A tutorial for information theory in neuroscience. eNeuro 2018, 5. [Google Scholar] [CrossRef] [PubMed]
  8. Sechidis, K.; Azzimonti, L.; Pocock, A.; Corani, G.; Weatherall, J.; Brown, G. Efficient feature selection using shrinkage estimators. Mach. Learn. 2019, 108, 1261–1286. [Google Scholar] [CrossRef] [Green Version]
  9. Choudhury, P.; Kumar, K.R.; Nandi, S.; Athithan, G. An empirical approach towards characterization of encrypted and unencrypted VoIP traffic. Multimed. Tools Appl. 2020, 79, 603–631. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Fu, H.; Knill, E. Efficient randomness certification by quantum probability estimation. Phys. Rev. Res. 2020, 2, 13016. [Google Scholar] [CrossRef] [Green Version]
  11. Meyer, P.E.; Lafitte, F.; Bontempi, G. Minet: A r/bioconductor package for inferring large transcriptional networks using mutual information. BMC Bioinform. 2008, 9, 1–10. [Google Scholar] [CrossRef]
  12. Kurt, Z.; Aydin, N.; Altay, G. Comprehensive review of association estimators for the inference of gene networks. Turk. J. Electr. Eng. Comput. Sci. 2016, 24, 695–718. [Google Scholar] [CrossRef]
  13. Schulman, J.S., Jr. Entropy: An Essential Component of Cryptographic Security. J. Cybersecur. Aware. Educ. 2019, 1, 29–38. [Google Scholar]
  14. Dai, S.; Guo, D. Comparing security notions of secret sharing schemes. Entropy 2015, 17, 1135–1145. [Google Scholar] [CrossRef] [Green Version]
  15. Austrin, P.; Chung, K.M.; Mahmoody, M.; Pass, R.; Seth, K. On the Impossibility of Cryptography with Tamperable Randomness. Algorithmica 2017, 79, 1052–1101. [Google Scholar] [CrossRef] [Green Version]
  16. Yasser, I.; Mohamed, M.A.; Samra, A.S.; Khalifa, F. A chaotic-based encryption/decryption framework for secure multimedia communications. Entropy 2020, 22, 1253. [Google Scholar] [CrossRef]
  17. Lu, Q.; Zhu, C.; Deng, X. An Efficient Image Encryption Scheme Based on the LSS Chaotic Map and Single S-Box. IEEE Access 2020, 8, 25664–25678. [Google Scholar] [CrossRef]
  18. Knuth, D. The Art of Computer Programming: Volume 2, Seminumerical Algoritms; Addison-Wesley Professional: Reading, MA, USA, 1981. [Google Scholar]
  19. Pseudorandom Number Sequence Test Program. 2011. Available online: http://www.fourmilab.ch/random/ (accessed on 18 April 2021).
  20. Marsaglia, George; The Marsaglia Random Number CDROM Including the Diehard Battery of Tests of Randomness; Natl. Sci. Found. (Grants DMS-8807976 DMS-9206972). 1995. Available online: http://stat.fsu.edu/pub/diehard/ (accessed on 18 April 2021).
  21. Rukhin, A.; Soto, J.; Nechvatal, J.; Miles, S.; Barker, E.; Leigh, S.; Levenson, M.; Vangel, M.; Banks, D.; Heckert, A.; et al. SP800-22: A Statistical Test Suite for Random and Pseudorandom Number Generators for Cryptographic Applications. Booz-Allen and Hamilton Inc Mclean VA; 2010. Available online: http://csrc.nist.gov/groups/ST/toolkit/rng/documents/SP800-22rev1a.pdf (accessed on 22 April 2021).
  22. L’ecuyer, P.; Simard, R. TestU01: A C library for empirical testing of random number generators. ACM Trans. Math. Softw. 2007, 33. [Google Scholar] [CrossRef]
  23. Madarro-Capó, E.J.; Legón-Pérez, C.M.; Rojas, O.; Sosa-Gómez, G.; Socorro-Llanes, R. Bit independence criterion extended to stream ciphers. Appl. Sci. 2020, 10, 7668. [Google Scholar] [CrossRef]
  24. Madarro Capó, E.J.; Cuellar, O.J.; Legón Pérez, C.M.; Gómez, G.S. Evaluation of input—Output statistical dependence PRNGs by SAC. In Proceedings of the 2016 International Conference on Software Process Improvement (CIMPS), Aguascalientes, Mexico, 12–14 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
  25. Miller, G. Note on the bias of information estimates. Inf. Theory Psychol. Probl. Methods 1955, 71, 108. [Google Scholar]
  26. Nemenman, I.; Shafee, F.; Bialek, W. Entropy and Inference, Revisited. arXiv 2001, arXiv:0108025. [Google Scholar]
  27. Schürmann, T.; Grassberger, P. Entropy estimation of symbol sequences. Chaos 1996, 6, 414–427. [Google Scholar] [CrossRef] [PubMed]
  28. Chao, A.; Shen, T.J. Nonparametric estimation of Shannon’s index of diversity when there are unseen species in sample. Environ. Ecol. Stat. 2003, 10, 429–443. [Google Scholar] [CrossRef]
  29. Holste, D.; Große, I.; Herzel, H. Bayes’ estimators of generalized entropies. J. Phys. A. Math. Gen. 1998, 31, 2551–2566. [Google Scholar] [CrossRef] [Green Version]
  30. Krichevsky, R.E.; Trofimov, V.K. The Performance of Universal Encoding. IEEE Trans. Inf. Theory 1981, 27, 199–207. [Google Scholar] [CrossRef]
  31. Trybula, S. Some problems of simultaneous minimax estimation. Ann. Math. Stat. 1958, 29, 245–253. [Google Scholar] [CrossRef]
  32. Hausser, J.; Strimmer, K. Entropy inference and the james-stein estimator, with application to nonlinear gene association networks. J. Mach. Learn. Res. 2009, 10, 1469–1484. [Google Scholar]
  33. Valiant, G.; Valiant, P. Estimating the unseen: Improved estimators for entropy and other properties. J. ACM 2017, 64, 1–41. [Google Scholar] [CrossRef]
  34. Zhang, Z. Entropy estimation in Turing’s perspective. Neural Comput. 2012, 24, 1368–1389. [Google Scholar] [CrossRef] [PubMed]
  35. Daub, C.O.; Steuer, R.; Selbig, J.; Kloska, S. Estimating mutual information using B-spline functions—An improved similarity measure for analysing gene expression data. BMC Bioinform. 2004, 5, 118. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Margolin, A.A.; Nemenman, I.; Basso, K.; Wiggins, C.; Stolovitzky, G.; Favera, R.D.; Califano, A. ARACNE: An algorithm for the reconstruction of gene regulatory networks in a mammalian cellular context. BMC Bioinform. 2006, 7, 1–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Van Huile, M.M. Edgeworth approximation of multivariate differential entropy. Neural Comput. 2005, 17, 1903–1910. [Google Scholar] [CrossRef] [PubMed]
  38. Vinck, M.; Battaglia, F.P.; Balakirsky, V.B.; Vinck, A.J.; Pennartz, C. Estimation of the entropy on the basis of its polynomial representation. IEEE Int. Symp. Inf. Theory Proc. 2012, 85, 1054–1058. [Google Scholar] [CrossRef]
  39. Kozachenko, L.F.; Leonenko, N.N. Sample Estimate of the Entropy of a Random Vector. Probl. Inf. Transm. 1987, 23, 95–101. [Google Scholar]
  40. Bonachela, J.A.; Hinrichsen, H.; Mũoz, M.A. Entropy estimates of small data sets. J. Phys. A Math. Theor. 2008, 41, 202001. [Google Scholar] [CrossRef]
  41. Grassberger, P. Entropy estimates from insufficient samplings. arXiv 2003, arXiv:0307138. [Google Scholar]
  42. Schürmann, T. Bias analysis in entropy estimation. J. Phys. A. Math. Gen. 2004, 37, L295. [Google Scholar] [CrossRef] [Green Version]
  43. Chao, A.; Wang, Y.T.; Jost, L. Entropy and the species accumulation curve: A novel entropy estimator via discovery rates of new species. Methods Ecol. Evol. 2013, 4, 1091–1100. [Google Scholar] [CrossRef]
  44. Burnham, K.P.; Overton, W.S. Estimation of the Size of a Closed Population when Capture Probabilities vary Among Animals. Biometrika 1978, 65, 625. [Google Scholar] [CrossRef]
  45. Archer, E.; Park, I.M.; Pillow, J.W. Bayesian entropy estimation for binary spike train data using parametric prior knowledge. Adv. Neural Inf. Process. Syst. 2013, 15, 1700–1708. [Google Scholar]
  46. Valiant, G.; Valiant, P. Estimating the unseen: An n/log(n)-sample estimator for entropy and support size, shown optimal via new CLTs. Proc. Annu. ACM Symp. Theory Comput. 2011, 685–694. [Google Scholar] [CrossRef]
  47. Nemenman, I. Coincidences and estimation of entropies of random variables with large cardinalities. Entropy 2011, 13, 2013–2023. [Google Scholar] [CrossRef] [Green Version]
  48. Al-Omari, A.I. New entropy estimators with smaller root mean squared error. J. Mod. Appl. Stat. Methods 2015, 14, 88–109. [Google Scholar] [CrossRef] [Green Version]
  49. Wolpert, D.H.; Wolf, D.R. Estimating functions of probability distributions from a finite set of samples. Phys. Rev. E 1995, 52, 6841. [Google Scholar] [CrossRef] [PubMed]
  50. Schürmann, T. A note on entropy estimation. Neural Comput. 2015, 27, 2097–2106. [Google Scholar] [CrossRef] [Green Version]
  51. de Matos Simoes, R.; Emmert-Streib, F. Influence of Statistical Estimators on the Large-Scale Causal Inference of Regulatory Networks. Stat. Mach. Learn. Approaches Netw. Anal. 2012, 6, 131–152. [Google Scholar] [CrossRef]
  52. Müller, S. Linux Random Number Generator-A New Approach. 2018. Available online: http://www.chronox.de/lrng/doc/lrng.pdf (accessed on 22 April 2021).
  53. Marton, K.; Suciu, A.; Ignat, I. Randomness in digital cryptography: A survey. Rom. J. Inf. Sci. Technol. 2010, 13, 219–240. [Google Scholar]
  54. Zhang, Z.; Grabchak, M. Nonparametric estimation of Küllback-Leibler divergence. Neural Comput. 2014, 26, 2570–2593. [Google Scholar] [CrossRef]
  55. GitHub—Simomarsili/ndd: Bayesian Entropy Estimation in Python—Via the Nemenman-Schafee-Bialek Algorithm. Available online: https://github.com/simomarsili/ndd (accessed on 15 March 2021).
  56. Marcon, E.; Herault, B. entropart: An R package to measure and partition diversity. J. Stat. Softw. 2015, 11, 1–26. [Google Scholar]
  57. GitHub—Pillowlab/CDMentropy: Centered Dirichlet Mixture Entropy Estimator for Binary Data. Available online: https://github.com/pillowlab/CDMentropy (accessed on 15 March 2021).
  58. Rosenblad, A. The Concise Encyclopedia of Statistics; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011; Volume 38, pp. 867–868. [Google Scholar] [CrossRef]
  59. Yim, O.; Ramdeen, K.T. Hierarchical Cluster Analysis: Comparison of Three Linkage Measures and Application to Psychological Data. Quant. Methods Psychol. 2015, 11, 8–21. [Google Scholar] [CrossRef]
  60. Ma, X.; Dhavala, S. Hierarchical clustering with prior knowledge. arXiv 2018, arXiv:1806.03432. [Google Scholar]
Figure 1. Behavior of the estimated mean of the 18 entropy estimators for different uniformly distributed sample sizes of bytes.
Figure 1. Behavior of the estimated mean of the 18 entropy estimators for different uniformly distributed sample sizes of bytes.
Entropy 23 00561 g001
Figure 2. Behavior of the estimated mean of the 18 entropy estimators for different uniformly distributed sample sizes of bits.
Figure 2. Behavior of the estimated mean of the 18 entropy estimators for different uniformly distributed sample sizes of bits.
Entropy 23 00561 g002
Figure 3. Behavior of the mean and observed variance of the estimators in byte samples using box plots for all sizes. Figure 3 (a) corresponds to the sample size n = 8 , so on until Figure 3 (l) for n = 16,384.
Figure 3. Behavior of the mean and observed variance of the estimators in byte samples using box plots for all sizes. Figure 3 (a) corresponds to the sample size n = 8 , so on until Figure 3 (l) for n = 16,384.
Entropy 23 00561 g003aEntropy 23 00561 g003b
Figure 4. Behavior of the mean and the observed variance of the estimators in bit samples using box plots for all sizes. Figure 4 (a) corresponds to the sample size n = 8 , so on until Figure 4 (l) for n = 16,384.
Figure 4. Behavior of the mean and the observed variance of the estimators in bit samples using box plots for all sizes. Figure 4 (a) corresponds to the sample size n = 8 , so on until Figure 4 (l) for n = 16,384.
Entropy 23 00561 g004aEntropy 23 00561 g004b
Figure 5. Mean Square Error of Entropy Estimates in Byte (a) and Bit (b) Samples.
Figure 5. Mean Square Error of Entropy Estimates in Byte (a) and Bit (b) Samples.
Entropy 23 00561 g005
Figure 6. Correlation matrix between the estimators for random samples of bytes. The correlation is represented based on its numerical value (a) and the intensity of the color (b).
Figure 6. Correlation matrix between the estimators for random samples of bytes. The correlation is represented based on its numerical value (a) and the intensity of the color (b).
Entropy 23 00561 g006
Figure 7. Clustering of estimators in uniformly distributed byte samples using the full-link agglomerative method.
Figure 7. Clustering of estimators in uniformly distributed byte samples using the full-link agglomerative method.
Entropy 23 00561 g007
Figure 8. Dendrogram corresponding to the hierarchical grouping of the estimators for uniformly distributed samples of bytes.
Figure 8. Dendrogram corresponding to the hierarchical grouping of the estimators for uniformly distributed samples of bytes.
Entropy 23 00561 g008
Figure 9. Correlation matrix between the estimators for random samples of bytes. The correlation is represented based on its numerical value (a) and the intensity of the color (b).
Figure 9. Correlation matrix between the estimators for random samples of bytes. The correlation is represented based on its numerical value (a) and the intensity of the color (b).
Entropy 23 00561 g009
Figure 10. Clustering of estimators in uniformly distributed bit samples using the full link agglomerative method.
Figure 10. Clustering of estimators in uniformly distributed bit samples using the full link agglomerative method.
Entropy 23 00561 g010
Figure 11. Dendrogram corresponding to the hierarchical grouping of the estimators for uniformly distributed samples of bits.
Figure 11. Dendrogram corresponding to the hierarchical grouping of the estimators for uniformly distributed samples of bits.
Entropy 23 00561 g011
Table 1. Description of entropy estimators.
Table 1. Description of entropy estimators.
Known asNotationEstimator
Miller-Madow
correction [25]
MM H ^ M M = h a t H M L + m 1 2 n ,
with m the number of x K such that y j > 0
Jackknife [44]UnveilJ H ^ J N = n H ^ M L n 1 n H ^ i M L ,
donde H ^ i M L is the entropy of the sample original without
the i-th symbol
Best Upper Bound [4]BUB H ^ BUB = i = 0 n a i h i ,
where h i = x = 1 k 1 y x = i
and a i = i n log i n + 1 i n 2 n
Grassberger [41]GSB H ^ GSB = log n 1 n x = 1 k y x ψ y x + ( 1 ) y x 0 1 t y x 1 t + 1 dt
, where ψ ( · ) is the digamma function
Schürmann [42]SHU H ^ SHU = ψ n 1 n x = 1 k y x ψ y x + ( 1 ) y x 0 1 ξ 1 t y x 1 1 + t dt
Chao-Chen [28]CS H ^ CS = x K p ^ x C S log 2 p ^ x CS 1 1 p ^ x CS n ,
where p ^ x CS = 1 m n p ^ x M L
James-Stein [32]SHR H ^ SHR = x K p ^ x S H R log 2 p ^ x S H R ,
where p ^ x SHR = λ ^ t x + 1 λ ^ p ^ x M L
with λ ^ = 1 x = 1 k ( p ^ x M L ) 2 ( n 1 ) x = 1 k ( t x p ^ x M L ) 2
and t x = 1 / k
Bonachela [40]BN H BN = 1 n + 2 x = 1 k ( y x + 1 ) j = y x + 2 n + 2 1 j
Zhang [34]Zhang H ^ Zhang = v = 1 n 1 1 v Z v ,
where Z v = n v + 1 n v + 1 ! n ! x K p ^ x i = 0 v 1 1 p ^ x i n
Chao-Wang-Jost [43]CWJ H ^ CWJ = 1 y x n 1 y x n k = y x n 1 1 k +
f 1 n 1 A n + 1 log A r = 1 n 1 1 r 1 A r
with
A = 2 f 2 n 1 f 1 + 2 f 2 if f 2 > 0 2 n 1 f 1 1 + 2 if f 2 = 0 , f 1 0 1 if f 2 = f 1 = 0 ,
where f 1 denote the number of singletons and
f 2 denote the number of doubletons in the sample.
Jeffrey [30]JEF H ^ Bayes = x K p ^ x Bayes log 2 p ^ x Bayes ,
where p ^ x Bayes = y x + a x n + A
with A = x = 1 k a x
a x = 1 / 2
Laplace [29]LAP a x = 1
Schürmann-
Grassberger [27]
SG a x = 1 / k
Minimax prior [31]MIN a x = n / k
NSB [26]NSB H NSB = p ξ , n H β m ( n ) d ξ p ξ , n d ξ ,
where p ( ξ , n ) = Γ [ k β ( ξ ) ] Γ [ n + k β ( ξ ) ] x K Γ [ y x + β ξ ] Γ [ β ξ ] ,
with
ξ = ψ 0 k β + 1 ψ 0 ( β + 1 ) , ψ m x = d / dx m + 1 log 2 Γ ( x )
and H β m ( n ) is the expectation value
of the m-th entropy moment at fixed β ;
exact expression for m = 1 , 2 is given in [49].
CDM [45]CDM H C D M = ψ 0 ( N + a + 1 ) x = 1 K y x + a μ ^ x N + a ( y x + a μ ^ x + 1 )
where μ ^ x = p ^ x M L
Unseen [33]UnseenThe authors propose to compute its value
algorithmically.
Table 2. Estimated mean of estimates for byte samples.
Table 2. Estimated mean of estimates for byte samples.
EstimatorsSample Sizes
8163264128256512102420484096819216,384
CS6.1427.6358.2518.1268.0648.0998.1168.0037.9147.9407.9637.974
ML2.9693.9434.8805.7686.5507.1787.5917.8097.9087.9557.9777.989
MM3.5904.5985.5366.3977.1137.6337.9027.9857.9987.9998.0008.000
SG3.9994.4705.1365.8846.5987.1947.5957.8097.9087.9557.9777.989
SHR7.9407.9717.9827.9917.9957.9987.9997.9998.0008.0008.0008.000
JEF7.9477.9027.8347.7517.6907.6967.7677.8547.9197.9577.9787.989
LAP7.9837.9697.9437.9047.8597.8327.8427.8837.9287.9607.9797.989
MIN5.0485.3855.8356.3576.8757.3267.6437.8237.9127.9567.9787.989
BUB7.0087.0217.4767.7327.8688.0617.8967.9717.9837.9857.9857.985
NSB6.4277.0657.4747.7267.8557.9227.9537.9697.9777.9817.9837.984
Unseen5.2026.7517.7397.9257.9457.9577.9507.9487.9537.9607.9667.971
Zhang3.6904.6965.6276.4787.1787.6747.9167.9797.9857.9857.9857.985
CJ5.3497.0288.0898.0938.0047.9967.9877.9857.9857.9857.9857.985
UnveilJ3.0364.1245.3166.6457.7497.957.8557.8327.8947.9407.9637.974
BN2.9634.4335.8156.9817.7267.8897.416.6626.1155.8295.6865.616
GSB4.6465.6136.4647.1827.6707.9217.9817.9857.9857.9857.9857.985
SHU4.3475.3296.2106.9787.5417.8687.9727.9857.9857.9857.9857.985
CDM6.1656.9087.3917.6897.8447.9227.9617.9817.9907.9957.9987.999
Table 3. The bias of the estimates for the byte samples.
Table 3. The bias of the estimates for the byte samples.
Estimator8163264128256512102420484096819216,384
CS1.8580.365−0.251−0.126−0.064−0.099−0.116−0.0030.0860.0600.0370.026
ML5.0314.0573.1202.2321.4500.8220.4090.1910.0920.0450.0230.011
MM4.4103.4022.4641.6030.8870.3670.0980.0150.0020.0010.0000.000
SG4.0013.5302.8642.1161.4020.8060.4050.1910.0920.0450.0230.011
SHR0.0600.0290.0180.0090.0050.0020.0010.0010.0000.0000.0000.000
JEF0.0530.0980.1660.2490.3100.3040.2330.1460.0810.0430.0220.011
LAP0.0170.0310.0570.0960.1410.1680.1580.1170.0720.0400.0210.011
MIN2.9522.6152.1651.6431.1250.6740.3570.1770.0880.0440.0220.011
BUB0.9920.9790.5240.2680.132-0.0610.1040.0290.0170.0150.0150.015
NSB1.5730.9350.5260.2740.1450.0780.0470.0310.0230.0190.0170.016
Unseen2.7981.2490.2610.0750.0550.0430.0500.0520.0470.0400.0340.029
Zhang4.3103.3042.3731.5220.8220.3260.0840.0210.0150.0150.0150.015
CJ2.6510.972−0.089−0.093−0.0040.0040.0130.0150.0150.0150.0150.015
UnveilJ4.9643.8762.6841.3550.2510.0500.1450.1680.1060.0600.0370.026
BN5.0373.5672.1851.0190.2740.1110.5901.3381.8852.1712.3142.384
GSB3.3542.3871.5360.8180.3300.0790.0190.0150.0150.0150.0150.015
SHU3.6532.6711.791.0220.4590.1320.0280.0150.0150.0150.0150.015
CDM1.8351.0920.6090.3110.1560.0780.0390.0190.0100.0050.0020.001
Table 4. Estimated mean of the estimates for the bit samples.
Table 4. Estimated mean of the estimates for the bit samples.
EstimatorSample Sizes
8163264128256512102420484096819216,384
CS0.9350.950.9750.9870.9930.9950.9970.9970.9980.9980.9980.998
ML0.8950.9510.9770.9890.9950.9970.9980.9991.0001.0001.0001.000
MM0.9840.9960.9991.0001.0001.0001.0001.0001.0001.0001.0001.000
SG0.9190.9560.9780.9890.9950.9970.9980.9991.0001.0001.0001.000
SHR0.9520.9810.9920.9960.9980.9990.9991.0001.0001.0001.0001.000
JEF0.9190.9560.9780.9890.9950.9970.9980.9991.0001.0001.0001.000
LAP0.9350.9610.9790.990.9950.9970.9980.9991.0001.0001.0001.000
MIN0.9450.9690.9830.9910.9950.9980.9990.9991.0001.0001.0001.000
BUB0.9340.9780.9970.9980.9980.9980.9980.9980.9980.9980.9980.998
NSB0.8720.9290.9620.9790.9880.9930.9950.9970.9970.9980.9980.998
Unseen0.8940.9490.9750.9870.9930.9950.9970.9970.9980.9980.9980.998
Zhang0.9890.9950.9980.9980.9980.9980.9980.9980.9980.9980.9980.998
CJ0.9890.9950.9980.9980.9980.9980.9980.9980.9980.9980.9980.998
UnveilJ0.8930.9490.9750.9870.9930.9950.9970.9970.9980.9980.9980.998
BN0.8620.9210.9570.9770.9870.9930.9950.9970.9970.9980.9980.998
GSB0.9900.9931.0000.9980.9980.9980.9980.9980.9980.9980.9980.998
SHU0.9900.9950.9980.9980.9980.9980.9980.9980.9980.9980.9980.998
CDM0.8470.9190.9590.9790.9900.9950.9970.9990.9991.0001.0001.000
Table 5. Estimation bias for bit samples.
Table 5. Estimation bias for bit samples.
Estimator8163264128256512102420484096819216,384
CS0.0650.050.0250.0130.0070.0050.0030.0030.0020.0020.0020.002
ML0.1050.0490.0230.0110.0050.0030.0020.0010.0000.0000.0000.000
MM0.0160.0040.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
SG0.0810.0440.0220.0110.0050.0030.0020.0010.0000.0000.0000.000
SHR0.0480.0190.0080.0040.0020.0010.0010.0000.0000.0000.0000.000
JEF0.0810.0440.0220.0110.0050.0030.0020.0010.0000.0000.0000.000
LAP0.0650.0390.0210.010.0050.0030.0020.0010.0000.0000.0000.000
MIN0.0550.0310.0170.0090.0050.0020.0010.0010.0000.0000.0000.000
BUB0.0660.0220.0030.0020.0020.0020.0020.0020.0020.0020.0020.002
NSB0.1280.0710.0380.0210.0120.0070.0050.0030.0030.0020.0020.002
Unseen0.1060.0510.0250.0130.0070.0050.0030.0030.0020.0020.0020.002
Zhang0.0110.0050.0020.0020.0020.0020.0020.0020.0020.0020.0020.002
CJ0.0110.0050.0020.0020.0020.0020.0020.0020.0020.0020.0020.002
UnveilJ0.1070.0510.0250.0130.0070.0050.0030.0030.0020.0020.0020.002
BN0.1380.0790.0430.0230.0130.0070.0050.0030.0030.0020.0020.002
GSB0.010.0070.0000.0020.0020.0020.0020.0020.0020.0020.0020.002
SHU0.010.0050.0020.0020.0020.0020.0020.0020.0020.0020.0020.002
CDM0.1530.0810.0410.0210.010.0050.0030.0010.0010.0000.0000.000
Table 6. MSE estimates for byte samples.
Table 6. MSE estimates for byte samples.
Estimator8163264128256512102420484096819216,384
CS3.7220.7161.0240.3120.0540.0220.0170.0010.0070.0040.0010.001
ML25.31516.4699.7444.9892.1060.6780.1680.0370.0090.0020.0010.000
MM19.46611.5836.0852.5810.7930.1390.0110.0010.0000.0000.0000.000
SG16.01212.4658.2084.4831.9710.6530.1650.0370.0090.0020.0010.000
SHR0.0410.0050.0020.0010.0000.0000.0000.0000.0000.0000.0000.000
JEF0.0030.010.0280.0620.0970.0930.0550.0210.0070.0020.0000.000
LAP0.0000.0010.0030.0090.020.0280.0250.0140.0050.0020.0000.000
MIN8.7196.8414.6932.7041.2680.4560.1290.0320.0080.0020.0000.000
BUB1.1001.0430.390.1250.0360.0130.0120.0010.0000.0000.0000.000
NSB2.6220.9440.3150.0870.0250.0070.0020.0010.0010.0000.0000.000
Unseen8.0001.8590.710.3280.0790.0180.0060.0040.0030.0020.0010.001
Zhang18.59510.935.6452.3290.6840.1110.0090.0010.0000.0000.0000.000
CJ7.1341.1290.680.3490.0610.0120.0020.0010.0000.0000.0000.000
UnveilJ24.65815.0517.2461.9290.0920.0230.0240.0280.0110.0040.0010.001
BN25.39112.7444.8111.080.110.0330.3541.7893.5544.7155.3535.685
GSB11.3095.7452.4060.7060.1310.0150.0030.0010.0000.0000.0000.000
SHU13.3867.1643.2331.070.2270.0250.0030.0010.0000.0000.0000.000
CDM3.4981.2550.4040.1080.0270.0070.0020.0000.0000.0000.0000.000
Table 7. MSE estimates for bit samples.
Table 7. MSE estimates for bit samples.
Estimator8163264128256512102420484096819216,384
CS0.0160.0070.0020.0000.0000.0000.0000.0000.0000.0000.0000.000
ML0.0330.0080.0020.0000.0000.0000.0000.0000.0000.0000.0000.000
MM0.0240.0050.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
SG0.0190.0060.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
SHR0.0190.0030.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
JEF0.0190.0060.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
LAP0.0120.0050.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
MIN0.0080.0030.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
BUB0.0260.0070.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
NSB0.0280.0080.0020.0010.0000.0000.0000.0000.0000.0000.0000.000
Unseen0.0330.0080.0020.0000.0000.0000.0000.0000.0000.0000.0000.000
Zhang0.0240.0050.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
CJ0.0240.0050.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
UnveilJ0.0340.0080.0020.0000.0000.0000.0000.0000.0000.0000.0000.000
BN0.0320.0090.0030.0010.0000.0000.0000.0000.0000.0000.0000.000
GSB0.0630.0150.0030.0010.0000.0000.0000.0000.0000.0000.0000.000
SHU0.0240.0050.0010.0000.0000.0000.0000.0000.0000.0000.0000.000
CDM0.0410.0110.0030.0010.0000.0000.0000.0000.0000.0000.0000.000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Contreras Rodríguez, L.; Madarro-Capó , E.J.; Legón-Pérez , C.M.; Rojas, O.; Sosa-Gómez, G. Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy. Entropy 2021, 23, 561. https://doi.org/10.3390/e23050561

AMA Style

Contreras Rodríguez L, Madarro-Capó  EJ, Legón-Pérez  CM, Rojas O, Sosa-Gómez G. Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy. Entropy. 2021; 23(5):561. https://doi.org/10.3390/e23050561

Chicago/Turabian Style

Contreras Rodríguez, Lianet, Evaristo José Madarro-Capó , Carlos Miguel Legón-Pérez , Omar Rojas, and Guillermo Sosa-Gómez. 2021. "Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy" Entropy 23, no. 5: 561. https://doi.org/10.3390/e23050561

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop