Abstract
In molecular sciences, the estimation of entropies of molecules is important for the understanding of many chemical and biological processes. Motivated by these applications, we consider the problem of estimating the entropies of circular random vectors and introduce non-parametric estimators based on circular distances between n sample points and their k th nearest neighbors (NN), where k () is a fixed positive integer. The proposed NN estimators are based on two different circular distances, and are proven to be asymptotically unbiased and consistent. The performance of one of the circular-distance estimators is investigated and compared with that of the already established Euclidean-distance NN estimator using Monte Carlo samples from an analytic distribution of six circular variables of an exactly known entropy and a large sample of seven internal-rotation angles in the molecule of tartaric acid, obtained by a realistic molecular-dynamics simulation.
1. Introduction
Estimation of entropies of molecules is an important problem in molecular sciences. Internal configurational entropy of a molecule is the entropy of the joint distribution of the internal molecular coordinates (bond lengths, bond angles, and dihedral angles), and as such it is a measure of random fluctuations in these coordinates. Most significant contribution to the internal configurational entropy of a molecule comes from the fluctuations in dihedral angles (also called internal-rotation angles). Many important properties of complex molecules, such as their stability and adopted conformation, depend on random fluctuations in their internal coordinates. Estimation of the internal configurational entropy of molecules is therefore important for understanding many chemical and biological processes, such as the spontaneity of a chemical reaction, protein folding, intermolecular protein-protein interactions, and protein-ligand interactions. It is also a key in the design of drugs that can stabilize the normally folded molecular structure or correct a misfolded structure, since protein misfolding is a cause of several diseases such as Alzheimer disease, mad cow disease, cystic fibrosis, and some types of cancer.
Estimation of the internal entropy of macromolecules, such as proteins, is a challenging problem because of the large number of correlated internal molecular coordinates. A commonly used method of estimating the internal entropy of a molecule, known as the quasi-harmonic approach, is based on the assumption of a multivariate normal distribution for the internal molecular coordinates []. Misra et al. [] discussed the decision theoretic estimation of the entropy of a multivariate normal distribution and obtained improvements over the best affine equivariant estimator under the squared error loss function. However, the assumption of a multivariate normal distribution for the internal coordinates of a molecule is appropriate only at low temperatures, when the fluctuations in its internal coordinates are small. At higher temperatures, the dihedral angles of a complex molecule exhibit multimodes and skewness in their distributions, and the multivariate normal distribution becomes inadequate.
Demchuk and Singh [] discussed a circular probability approach for modeling the dihedral angles of a molecule in the estimation of internal rotational entropy. As an illustration, they modeled the torsional angle of the methanol molecule by a trimodal von Mises distribution and derived a bath-tub-shaped distribution for the torsional potential energy of the molecule. Singh et al. [] introduced a torus version of a bivariate normal distribution for modeling two dihedral angles. The marginal distributions of the model are symmetric unimodal or symmetric bimodal depending on the configurations of the parameters. A multivariate generalization of this bivariate model has been proposed by Mardia et al. []. Hnizdo et al. [] and Darian et al. [] used a Fourier series expansion approach for modeling univariate and bivariate distributions of molecular dihedral angles. Complex molecules, however, have many significantly correlated dihedral angles, whose joint distribution can take an arbitrary form. For this reason, a non-parametric approach for estimating the entropy of a circular random vector of arbitrary dimensions is desirable.
Several non-parametric estimators of the entropy of an m-dimensional random variable X have been discussed in the literature. A common approach is to replace the probability density function (pdf) in the definition of the differential entropy,
by its non-parametric kernel or histogram density estimator [,]. However, in most practical situations, implementation of such estimates in higher dimensions becomes difficult. In one dimension (), several authors have proposed estimates of entropy in the context of testing goodness of fit [,]. Singh et al. [] proposed the following asymptotically unbiased and consistent nearest-neighbor (NN) estimator of the entropy :
Here, is the Euclidean distance of a point to its , nearest (in the Euclidean-distance sense) neighbor in a random sample from the distribution ; is Euler’s constant, , , and is the usual gamma function. For , the estimator reduces to the NN estimator proposed by Kozachenko and Leonenko []. Results similar to those of [] have been also reported by Goria et al. []. For the purpose of the estimation of the information-theoretic quantity of mutual information, Kraskov et al. [] generalized the first-nearest-neighbor estimator of [] in terms of nearest-neighbor distances in a general metric, giving, however, explicit expressions only for the maximum and Euclidean metrics and without providing formal proofs of asymptotic unbiasedness and consistency. For , Tsybakov and van der Meulen [] established the mean-square-root-n consistency of a truncated version of . Earlier, Loftsgaarden and Quesenberry [] had used NN distances to construct non-parametric estimates of a multivariate pdf. Recently, Mnatsakanov et al. [] studied k-NN estimators of entropy in which the parameter k is assumed to be a function of the sample size.
The NN entropy estimator (2) uses Euclidean distances between the sample points. However, when the random variable X is circular, it is natural to base an NN estimate of entropy on a circular distance rather than the Euclidean distance. A circular observation can be regarded as a point on a circle of unit radius. Once an initial direction and an orientation of the circle have been chosen, each circular observation can be specified by the angle from the initial direction to the point on the circle corresponding to the observation. In this paper, we construct estimates of the entropy of an m-dimensional circular random vector based on two different definitions of circular distances. Let and be two observations on an m-dimensional circular random vector Θ. We define two circular distance functions and as follows:
and
Note that , is the arc length between the points and on the unit circle . On the other hand, is the Euclidean distance between the points and on the unit circle .
In Section 2 and Section 3, we propose explicit expressions for NN estimators of entropy based on the circular distance functions and , respectively, and prove their asymptotic unbiasedness and consistency (with some mathematical details given in an Appendix). In Section 4, we compare the performance of the estimator based on the circular distance with that of the Euclidean-distance estimator (2) using Monte Carlo simulations from an analytic 6-dimensional circular distribution, where the exact joint entropy is known. We there also apply the -distance estimator to the problem of estimating the entropy of a 7-dimensional joint distribution of the internal-rotation angles in the molecule of tartaric acid, using a large sample of these angles obtained by a realistic molecular-dynamics simulation.
2. Nearest Neighbor Estimates of Entropy Based on Circular Distance
For constructing nearest neighbor estimates of the entropy of a circular random vector based on the distance function (3), we first derive the expression for the volume of a ball
centered at and having a radius , being the maximum value of .
Lemma 2.1. Let , and let be the volume of the ball , defined by (5). Then
where denotes the cumulative distribution function of the sum of m independent and identically distributed random variables each having a beta distribution with parameters , .
Proof. Without loss of generality, we may take . Then
where , and are independent and identically distributed uniform random variables over the interval . Define , . Then are independent and identically distributed beta random variables, having parameters , . Hence the result follows.
Remark 2.1. (i) For and , we have . For and , it can be verified that
(ii) For and , it can be verified that
(iii) For a general m () and , it can be verified that
(iv) For any and for , satisfies the following recursion relation
where is the pdf of a beta random variable with parameters and .
The circular distance (3) becomes the Euclidean distance when . Circular-distance balls , where and , thus tend to the corresponding Euclidean-distance balls as . We can therefore apply the Lebesgue differentiation theorem [] to the probability density function of a circular random variable in the form
where is given by (6). Equation (7) suggests that, given a sufficiently large random sample from the distribution of Θ, the probability density function can be approximated at almost all by
where denotes the cardinality of the set and r is sufficiently small.
Guided by this insight, we will now construct nearest neighbor estimates of entropy for an m-variate circular random vector Θ, having a probability density function . Let be a random sample from the distribution of Θ and let be a given positive integer. For , let denote the circular distance of from its closest neighbor, with respect to the circular distance , i.e.,
Assume that the sample size n is sufficiently large, so that the distances are small, on the average. Then, based on approximation (8), a reasonable estimator of is
and thus a reasonable estimator of the entropy is
In the following theorem, we derive the expression for the asymptotic mean of the estimator . Apart from the arguments for the interchange of the limit and the integral signs, the proof is similar to that of Theorem 8 of Singh et al. [], who in their proof interchange the limit and the integral signs without mentioning the conditions under which it is allowed.
Theorem 2.1. Suppose that there exists an , such that
and
Then, for a fixed (not depending on n)
where is defined by (9), , and is Euler’s constant.
Proof. Let
Then are identically distributed random variables. Therefore,
where, for a given , is a random variable having the same distribution as that of the conditional distribution of given .
where denotes the inverse function of . Using standard arguments, we get
where is defined by (5). For a fixed , and for almost all values of , using Lemma 2.1, we have
Therefore, using the Poisson approximation to the binomial distribution, we get
for almost all values of . For a fixed , let be a random variable having the pdf
Then, in view of (13),
where stands for the convergence in distribution. For each fixed , it can be verified that
Under the condition (10), it can be shown (for details, see the Appendix) that, for almost all values of , there exists a constant C (not depending on n) such that for all sufficiently large values of n
Then, in view of (14) and the moment convergence theorem, it follows that
for almost all values of Using Fatou’s lemma, we get
by (10). Therefore,
Now the result follows from (12) and (16).
Since the estimator is not asymptotically unbiased, we propose the following (asymptotic) bias corrected estimator for estimating the entropy :
Thus, we have the following corollary to Theorem 2.1.
Corollary 2.1. Under the assumptions of Theorem 2.1, the estimator is asymptotically unbiased for estimating the entropy .
The following theorem provides conditions under which the estimator is consistent for estimating the entropy .
Theorem 2.2. Suppose that there exists an , such that
and
Then, for a fixed (not depending on n),
and thus is a consistent estimator of the entropy .
Under conditions (18) and (19), in the proof of Theorem 2.2, the steps involved in justifying the interchange of the limit and the integral sign are tedious but virtually identical to the arguments used in the proof of Theorem 2.1. The remaining part of the proof is identical to the proof of Theorem 11 of Singh et al. []. We therefore omit the proof of Theorem 2.2.
Remark 2.2. For small values of , the function involved in the evaluation of estimate can be computed using numerical integration. For moderate and large values of m, which is the case with many molecules encountered in molecular sciences, using the central limit theorem one can get a reasonable approximation of by the cumulative distribution function of a normal distribution having mean and variance .
3. Nearest Neighbor Estimates of Entropy Based on Circular Distance
Let . In order to construct nearest neighbor estimates of entropy based on the distance function , defined by (4), we require the volume of the ball to be
centered at and having radius .
Lemma 3.1. Let , and let be the volume of the ball , defined by (20). Then
where denotes the cumulative distribution function of the sum of m independent and identically distributed random variables each having a beta distribution with parameters and
Proof. We have
where are independent and identically distributed uniform random variables over the interval .
Define . Then, are independent and identically distributed beta random variables, having parameters . Hence the result follows.
Remark 3.1. (i) For and ,
(ii) satisfies a similar recursion relation as that satisfied by and given in Remark 2.1 (iv).
For , let denote the circular distance of from its closest neighbor with respect to the circular distance , defined by (4). Assume that the sample size n is sufficiently large, so that on average the distances are small. Then, based on approximation (7), a reasonable estimator of is
and thus a reasonable estimator of the entropy is
The proof of the following theorem is identical to the proof of Theorem 2.1 and therefore it is omitted.
Theorem 3.1. Suppose that there exists an such that (10) holds and
Then, for a fixed (not depending on n),
where is defined by (21).
Since the estimator is not asymptotically unbiased, we propose the following (asymptotic) bias corrected estimator for estimating the entropy :
Thus, we have the following corollary to Theorem 3.1.
Corollary 3.1. Under the assumptions of Theorem 3.1, the estimator is asymptotically unbiased for estimating the entropy .
The following theorem provides conditions under which the estimator is consistent for estimating the entropy . The proof of the theorem follows using the arguments similar to the one given for the proof of Theorem 2.2.
Theorem 3.2. Suppose that there exists an such that (18) holds and
Then, for a fixed (not depending on n),
and therefore, under conditions (18) and (24), is a consistent estimator of the entropy .
Remark 3.2. (i) Using Remark 3.1 (i) and the fact that is a decreasing function of , for and , we have
since
Therefore, for and , we have
Thus for , estimators and , based on circular distance functions and respectively, are identical.
(ii) For small values of , the function involved in the evaluation of estimate can be computed using numerical integration. For moderate and large values of m, which is the case with many molecules encountered in molecular sciences, using the central limit theorem one can get a reasonable approximation of by the cumulative distribution function of a normal distribution having mean and variance .
With , (17) and (23) define two classes of estimators for the entropy . The biases and variances of these estimators depend on k, the sample size n, and the pdf and its dimensions m. It would be useful to have the knowledge of the biases and variances as functions of k, n, m, and some characteristic of , such as , where is a function such that a reliable estimate of μ can be obtained using the available data on Θ. We have not been able to derive any meaningful expressions for the biases and variances of the proposed estimators, and this problem is under further investigation.
4. Monte Carlo Results and a Molecular Entropy Example
The performance of an entropy estimator can be investigated rigorously by using Monte Carlo samples from a distribution for which the entropy is known exactly. While analytic distributions of more than two correlated circular variables with exactly calculable entropic attributes do not seem available, one may construct a distribution of higher dimensionality as a product of a suitable number of bivariate distributions. To test the performance of the circular-distance estimator (17), we used an analytic 6-dimensional circular distribution given as the product of three bivariate circular distributions, each of the form []
which, as a circular analogue of the bivariate normal distribution, can be called the bivariate von Mises distribution. Details pertaining to the 6-dimensional distribution used and the Monte Carlo sampling are given in [], where the same circular distribution was used in an investigation of the combined mutual-information-expansion and Euclidean-distance-NN method of entropy estimation.
Table 1.
Circular- and Euclidean-distance estimates and , respectively, from samples of size n of the analytic distribution of 6 circular variables; the exact entropy value is (to 4 decimals) .
| 0.05 | 1.86851 | 1.92019 | 1.96379 | 2.00395 | 2.03653 |
| 2.01225 | 2.09929 | 2.16540 | 2.22380 | 2.27390 | |
| 0.10 | 1.81503 | 1.86968 | 1.90481 | 1.92899 | 1.95164 |
| 1.95640 | 2.01952 | 2.07339 | 2.11255 | 2.14865 | |
| 0.20 | 1.80070 | 1.83713 | 1.85672 | 1.87498 | 1.89122 |
| 1.92305 | 1.96691 | 2.00110 | 2.03086 | 2.05650 | |
| 0.40 | 1.8007 | 1.81582 | 1.82840 | 1.83904 | 1.84862 |
| 1.89363 | 1.92652 | 1.95001 | 1.96939 | 1.98668 | |
| 0.60 | 1.80066 | 1.80693 | 1.81496 | 1.82296 | 1.83097 |
| 1.88696 | 1.90952 | 1.92764 | 1.94358 | 1.95783 | |
| 0.80 | 1.79837 | 1.80297 | 1.81063 | 1.81660 | 1.82222 |
| 1.87936 | 1.89971 | 1.91649 | 1.92946 | 1.94125 | |
| 1.00 | 1.79539 | 1.80017 | 1.80566 | 1.81030 | 1.81566 |
| 1.87317 | 1.89238 | 1.90694 | 1.91850 | 1.92915 | |
| 2.00 | 1.79660 | 1.79533 | 1.79736 | 1.79970 | 1.80266 |
| 1.86471 | 1.87555 | 1.88493 | 1.89298 | 1.90073 | |
| 4.00 | 1.79673 | 1.79383 | 1.79404 | 1.79480 | 1.79613 |
| 1.85696 | 1.86477 | 1.87136 | 1.87702 | 1.88211 | |
| 6.00 | 1.79795 | 1.79491 | 1.79385 | 1.79419 | 1.79458 |
| 1.85403 | 1.86071 | 1.86552 | 1.87029 | 1.87414 | |
| 8.00 | 1.79893 | 1.79484 | 1.79373 | 1.79322 | 1.79337 |
| 1.85240 | 1.85745 | 1.86197 | 1.86545 | 1.86891 | |
| 10.00 | 1.80036 | 1.79562 | 1.79426 | 1.79350 | 1.79329 |
| 1.85170 | 1.85578 | 1.85969 | 1.86287 | 1.86583 |
Table 1 presents the circular-distance estimates , obtained from samples of sizes in the range –, together with the corresponding Euclidean-distance estimates . Figure 1 displays the estimates and as functions of the sample size n. Noting that the exact entropy value here is, to 4 decimal places, , we observe that as n increases the circular-distance estimates initially “undershoot” the exact value and then start to approach it slowly from below. In contrast, the Euclidean-distance estimates approach the exact value monotonically from above. Interestingly, the biases of the two kinds of estimates at sample sizes million are approximately equal in absolute value. The behavior of the circular-distance estimates at is similar to that at , and the estimate values at different k’s become very close at million.
Figure 1.
Plots of the circular- and Euclidean-distance estimates and , respectively, as functions of the sample size n, for the analytic distribution of 6 circular variables; the exact entropy value is (to 4 decimals) .
To investigate the usefulness of circular-distance NN estimators in the problem of evaluating the configurational entropy of internal rotations in molecules, we used the circular-distance estimator to estimate the entropy of the joint distribution of internal-rotation angles in the molecule of tartaric acid, where the number of variables is . Samples of size n up to 14.4 million of the internal-rotation angles were obtained from a molecular dynamics simulation of the (R,S) stereoisomer of this molecule []. Figure 2 shows marginal histograms, smoothed using a Gaussian kernel, of the seven internal-rotation angles of tartaric acid; note that these marginals display markedly non-Gaussian features. The code ANN [] (with our modification for the circular distance ), which utilizes a k-d tree algorithm [], was used for finding the k-th NN distances between sample points. Figure 3 presents the estimates , as functions of the sample size n. The values of decrease as n increases, while, at a fixed value of million, they increase as k increases; at greater values of n, the estimates at different k’s become quite close in value. Figure 4 compares the circular-distance estimates and with the corresponding Euclidean-distance estimates . We note that an extrapolated Euclidean-distance estimate was obtained for this entropy in []. Again, as in the case of the analytic circular distribution, this value approximately equals the arithmetic mean of the circular and Euclidean distance estimates at sample sizes million.
Figure 2.
Smoothed marginal histograms of the internal-rotation angles , of the (R,S) isomer of tartaric acid obtained by molecular dynamics simulations.
Figure 3.
Circular-distance nearest-neighbor estimates , of the entropy of the 7-dimensional joint distribution of internal-rotation angles in the (R,S) isomer of tartaric acid as functions of the sample size n. The estimates at a fixed million increase in value as k increases.
Figure 4.
Circular-distance nearest-neighbor estimates and of the internal-rotation entropy of tartaric acid as functions of the sample size n compared with the Euclidean-distance nearest-neighbor estimates and . An extrapolated estimate is [].
Perhaps surprisingly, the results of both the analytic-distribution and molecular-simulation studies undertaken here indicate that only when relatively small data samples are available, the use of a circular-distance estimator has some advantage over the Euclidean-distance estimator. On samples of large size, needed for sufficient convergence of an NN estimate of the entropy of a multivariate distribution, the circular-distance estimates obtained did not have a significantly smaller bias than the Euclidean-distance estimates. In view of such findings, one may question whether the additional computational complexity of a circular-distance estimate is worth the effort. However, we observed that as the sample size increased, the circular NN distances in the sample became quickly so small that the circular-distance estimator coincided in value with the simpler Euclidean-distance estimator in which the same NN-distance values were used. This is explained by the fact that when the circular NN distances , the estimator can be replaced with the estimator in which the NN distances are substituted for the Euclidean NN distances ; this fact follows directly from Remark 2.1 (iii). The only extra computational effort is then expended in finding the circular, instead of Euclidean, NN distances in a given sample.
Acknowledgments
The authors are thankful to Jun Tan for carrying out the circular-distance NN calculations, and to E. James Harner, Cecil Burchfiel, Robert Mnatsakanov and Dan S. Sharp for helpful discussions. The findings and conclusions in this paper are those of the authors and do not necessarily represent the views of the National Institute for Occupational Safety and Health.
Appendix
Here we provide the proof of (15). For a fixed , let be the distribution function of the random variable . To establish (15), we will first show that, for and ,
for almost all values of . In order to establish (25), consider
Thus, to establish (25), it is enough to show that the second term in the above expression is finite. Note that, for ,
and thus
In view of assumption (11), it follows that
for almost all values of . Therefore, (25) is established.
Now we will establish (15). Consider
We can write
We have,
For , we have Therefore,
where denotes the beta random variable with parameter , . For , we have Therefore, (28) yields
Note that . Thus, on using (7), we get
for almost all values of .
For a , for which , choose . Then, for sufficiently large values of n,
and therefore, for sufficiently large values of n,
Therefore, for sufficiently large values of n and for almost all values of , (29) yields
On making the change of variable in the integral in (30), we get
We also have
and
where .
Note that, for ,
Therefore (33) yields
In view of (25), we have
Therefore, using (30)-(35), we conclude that
for almost all values of θ.
Now consider
For , we have
Therefore, for and for almost all values of , we have
uniformly in u.
For , let . Then, for sufficiently large n, and for almost all values of , we have
and therefore
Using (39) in (37) we conclude that, for sufficiently large values of n, and for almost all values of θ
Using (36) and (40) in (27), we conclude further that there exists a constant such that, for sufficiently large values of n,
for almost all values of .
Now consider
Note that, for each and for almost all values of ,
uniformly in , i.e., for almost all values of , and every
for sufficiently large values of n. Therefore,
Using (43) in (42), we conclude that, for sufficiently large values of n and for almost all values of ,
Finally, on using (41) and (44) in (26), we conclude (15).
References
- Karplus, M.; Kushick, J.N. Method for estimating the configurational entropy of macromolecules. Macromolecules 1981, 14, 325–332. [Google Scholar] [CrossRef]
- Misra, N.; Singh, H.; Demchuk, E. Estimation of the entropy of a multivariate normal distribution. J. Multivar. Anal. 2005, 92, 324–342. [Google Scholar] [CrossRef]
- Demchuk, E.; Singh, H. Statistical thermodynamics of hindered rotation from computer simulations. Mol. Phys. 2001, 99, 627–636. [Google Scholar] [CrossRef]
- Singh, H.; Hnizdo, V.; Demchuk, E. Probabilistic modeling of two dependent circular variables. Biometrika 2002, 89, 719–723. [Google Scholar] [CrossRef]
- Mardia, K.V.; Hughes, G.; Taylor, C.C.; Singh, H. A multivariate von Mises distribution with applications to bioinformatics. Can. J. Stat. 2008, 36, 99–109. [Google Scholar] [CrossRef]
- Hnizdo, V.; Fedorowicz, A.; Singh, H.; Demchuk, E. Statistical thermodynamics of internal rotation in a hindering potential of mean force obtained from computer simulations. J. Comput. Chem. 2003, 24, 1172–1183. [Google Scholar] [CrossRef] [PubMed]
- Darian, E.; Hnizdo, V.; Fedorowicz, A.; Singh, H.; Demchuk, E. Estimation of the absolute internal-rotation entropy of molecules with two torsional degrees of freedom from stochastic simulations. J. Comput. Chem. 2005, 26, 651–660. [Google Scholar] [CrossRef] [PubMed]
- Beirlant, J.; Dudewicz, E.J.; Gyorfi, L.; van der Meulen, E.C. Nonparametric estimation of entropy: An overview. Internat. J. Math. Stat. 1997, 6, 17–39. [Google Scholar]
- Scott, D. Multivariate Density Estimation: Theory, Practice and Visualization; Wiley: New York, NY, USA, 1992. [Google Scholar]
- Vasicek, O. On a test for normality based on sample entropy. J. R. Stat. Soc. Series B 1976, 38, 54–59. [Google Scholar]
- Dudewicz, E.J.; van der Meulen, E.C. Entropy-based tests of uniformity. J. Am. Stat. Assoc. 1981, 76, 967–974. [Google Scholar] [CrossRef]
- Singh, H.; Misra, N.; Hnizdo, V.; Fedorowicz, E.; Demchuk, E. Nearest neighbor estimates of entropy. Am. J. Math. Manag. Sci. 2003, 23, 301–321. [Google Scholar] [CrossRef]
- Kozachenko, L.F.; Leonenko, N.N. Sample estimates of entropy of a random vector. Prob. Inf. Trans. 1987, 23, 95–101. [Google Scholar]
- Goria, M.N.; Leonenko, N.N.; Novi Inveradi, P.L. A new class of random vector entropy estimators and its applications. Nonparam. Stat. 2005, 17, 277–297. [Google Scholar] [CrossRef]
- Kraskov, A.; Stögbauer, H.; Grassberger, P. Estimating mutual information. Phys. Rev. E 2004, 69, 066138-1–066138-16. [Google Scholar] [CrossRef]
- Tsybakov, A.B.; van der Meulen, E.C. Root-n consistent estimators of entropy for densities with unbounded support. Scan. J. Stat. 1996, 23, 75–83. [Google Scholar]
- Loftsgaarden, D.O.; Quesenberry, C.P. A non-parametric estimate of a multivariate density function. Ann. Math. Stat. 1965, 36, 1049–1051. [Google Scholar] [CrossRef]
- Mnatsakanov, R.M.; Misra, N.; Li, Sh.; Harner, E.J. kn-Nearest neighbor estimators of entropy. Math. Meth. Stat. 2008, 17, 261–277. [Google Scholar] [CrossRef]
- Lebesgue, H. Sur l’intégration des fonctions discontinues. Ann. Ecole Norm. 1910, 27, 361–450. [Google Scholar]
- Hnizdo, V.; Tan, J.; Killian, B.J.; Gilson, M.K. Efficient calculation of configurational entropy from molecular simulations by combining the mutual-information expansion and nearest-neighbor methods. J. Comput. Chem. 2008, 29, 1605–1614. [Google Scholar] [CrossRef] [PubMed]
- Hnizdo, V.; Darian, E.; Fedorowicz, A.; Demchuk, E.; Li, S.; Singh, H. Nearest-neighbor nonparametric method for estimating the configurational entropy of complex molecules. J. Comput. Chem. 2006, 28, 655–668. [Google Scholar] [CrossRef] [PubMed]
- Arya, S.; Mount, D.M. Approximate nearest neighbor searching. In the Proceedings of the Fourth Annual ACM-SIAM Symposium on Discrete Algorithms, 25–27 January 1993; p. 271. Available online: http:// www.cs.umd.edu/∼mount/ANN/ (accessed on 5 May 2010).
- Friedman, J.H.; Bentley, J.L.; Finkel, R.A. An algorithm for finding best matches in logarithmic expected time. ACM Trans. Math. Software 1977, 3, 209–226. [Google Scholar] [CrossRef]
© 2010 by the authors; licensee MDPI, Basel, Switzerland. This article is an open-access article distributed under the terms and conditions of the Creative Commons Attribution license http://creativecommons.org/licenses/by/3.0/.