Next Article in Journal
Entropy of Mixing and the Glass Transition of Amorphous Mixtures
Previous Article in Journal
Entropic Behavior of Binary Carbonaceous Mesophases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Calculation of Differential Entropy for a Mixed Gaussian Distribution

by
Joseph V. Michalowicz
1,
Jonathan M. Nichols
2,* and
Frank Bucholtz
2
1
U. S. Naval Research Laboratory, Optical Sciences Division, Washington, D.C. 20375, USA
2
U. S. Naval Research Laboratory, Optical Sciences Division, Washington, D.C. 20375, USA
*
Author to whom correspondence should be addressed.
Entropy 2008, 10(3), 200-206; https://doi.org/10.3390/entropy-e10030200
Submission received: 16 June 2008 / Revised: 14 August 2008 / Accepted: 18 August 2008 / Published: 25 August 2008

Abstract

:
In this work, an analytical expression is developed for the differential entropy of a mixed Gaussian distribution. One of the terms is given by a tabulated function of the ratio of the distribution parameters.

1. Introduction

The concept of entropy for a random process was introduced by Shannon [1] to characterize the irreducible complexity in a particular process beyond which no compression is possible. Entropy was first formulated for discrete random variables, and was then generalized to continuous random variables in which case it is called differential entropy. By definition, for a continuous random variable X with probability density function p(x), the differential entropy is given by
h ( X ) = S p ( x ) log ( p ( x ) ) d x
where S = {x|p(x) > 0} is the support set of X. The log function may be taken to be log2, and then the entropy is expressed in bits; or as ln, in which case the entropy is in nats. We shall use the latter convention for the computations in this paper.
Textbooks (e.g. Cover & Thomas [2]) which discuss the concept of entropy often do not provide analytic calculations of differential entropy for many probability distributions; specific cases are usually limited to the uniform and Gaussian distributions. Cover & Thomas [2], (pg. 486-487) does provide a table of entropies for a large number of the probability density functions usually listed in a table of statistical distributions. This table was extracted from a paper by Lazo & Rathie [3]. In addition, a very detailed computation of these entropies may be found in Michalowicz et al. [4]. (Note: There are two typographical errors in the Cover & Thomas list; please double check by using the other two references, both of which have the correct formulas).
In this paper we calculate the differential entropy for a case not appearing in the lists cited above; namely, for a mixed Gaussian distribution with the probability density function
p ( x ) = 1 2 2 π σ [ e ( x μ ) 2 / 2 σ 2 + e ( x + μ ) 2 / 2 σ 2 ] < x <
Clearly this distribution is obtained by just splitting a Gaussian distribution N (0, σ2) into two parts, centering one half about +µ and the other about −µ and summing the resultants. Such a density function is depicted in Figure 1. This distribution has a mean of zero and a variance given by σ m g 2 = σ2 + µ2. This is because the second moment of the mixed Gaussian is 1/2 the sum of the second moments for the Gaussian components, each of which is σ2 + µ2. It can also be written in the more compact form
p ( x ) = 1 2 π σ e ( x + μ ) 2 / 2 σ 2 cosh ( μ x / σ 2 ) .
The mixed Gaussian distribution is often considered as a noise model in a number of signal processing applications. This particular noise model is used in describing co-channel interference, for example, where thermal, Gaussian distributed noise is combined with man-made “clutter” e.g., signals from communication systems [5]. Wang and Wu [6] considered a mixed-Gaussian noise model in a nonlinear signal detection application. Mixed Gaussian noise was also used for modeling purposes in Tan et al. [7]. Additional works on mixed Gaussian noise include that of Bhatia and Mulgrew [5], who looked at a non-parametric channel estimator for this type of noise, and Lu [8], who looked at entropy regularized likelihood learning on Gaussian mixture models. It has also been demonstrated that entropy-based parameter estimation techniques (e.g. mutual information maximization) are of great utility in estimating signals corrupted by non-Gaussian noise [9, 10], particularly when the noise is mixed-Gaussian [11]. However, these works relied on non-parametric estimation of signal entropy due to the absence of a closed-form expression. Our work is therefore aimed at providing an analytical expression for signal entropy in situations where the corrupting noise source is mixed-Gaussian.
The calculation of the differential entropy, in terms of nats, proceeds as follows
Entropy 10 00200 i001
If we let y = µx/σ2 in this integral, the above expression becomes
h e ( X ) = ln ( 2 π σ ) + μ 2 σ 2 + 1 2 1 2 π σ e μ 2 / 2 σ 2 ( σ 2 / μ ) e σ 2 y 2 / 2 μ 2 cosh ( y ) ln ( cosh ( y ) ) d y .
Noting that the integrand is an even function, we obtain
h e ( X ) = 1 2 ln ( 2 π e σ 2 ) + μ 2 σ 2 2 2 π e μ 2 / 2 σ 2 ( σ / μ ) 0 e σ 2 y 2 / 2 μ 2 cosh ( y ) ln ( cosh ( y ) ) d y .
Let α = µ/σ. Then
h e ( X ) = 1 2 ln ( 2 π e σ 2 ) + α 2 2 2 π α e α 2 / 2 0 e y / 2 α 2 cosh ( y ) ln ( cosh ( y ) ) d y .
The first term is recognized as the entropy in nats of a Gaussian distribution. When µ = 0 (and so α = 0), our distribution reduces to a Gaussian distribution and the entropy reduces to just this first term.
An analytic expression for the integral in Eqn. (7) could not be found. However, there are analytic bounds for the integral term which are derived by noting that
y − ln 2 ≤ ln(cosh(y)) ≤ yy ≥ 0.
Thus, for the upper bound to the integral term we have
Entropy 10 00200 i002
by means of formula 3.562 (4) in [12], where erf denotes the error function, defined as
erf ( z ) = 2 π 0 z e u 2 d u .
Likewise, for the lower bound we have
Entropy 10 00200 i003
by means of formula 3.546 (2) in [12].
Since the integrand in I is always greater than or equal to 0, we know that I 0, so we can write
h e ( X ) = 1 2 ln ( 2 π e σ 2 ) + α 2 I
where
max ( 0 ; α 2 erf ( α / 2 ) + 2 / π α e σ 2 / 2 ln 2 ) I α 2 erf ( α / 2 ) + 2 / π α e σ 2 / 2
for all α = µ/σ 0.
The graph of I as a function of α is shown in Figure 2, along with the analytic upper and lower bounds. Clearly I converges rapidly to the lower bound as α increases. A tabulation of numerically computed values of I is presented in Table 1, together with corresponding values of α2I. As is clear in the Table, (α2I) monotonically increases from 0 to ln 2 = 0.6931. Hence the differential entropy, in nats, of a mixed Gaussian distribution, as depicted in Figure 1, can be expressed as
h e ( X ) = 1 2 ln ( 2 π e σ 2 ) + ( α 2 I )
where (α2I) is a function of α = µ/σ (tabulated in Table 1) which is equal to zero at α = 0 (in which case the distribution is Gaussian) and monotonically increases to ln 2 as α increases to α > 3.5 (in which case the distribution is effectively split into two separate Gaussians). In particular, if σ = 1, he(X) is a monotonically increasing function of µ which has the value 1.419 for µ = 0 and converges to the value 2.112 as µ is increased and the two parts of the mixed Gaussian distribution are split apart.
To express the differential entropy in bits, Eqn. (12) needs to be divided by ln 2, which gives
h ( X ) = 1 2 log 2 ( 2 π e σ 2 ) + ( α 2 I ln 2 )
where the second term is a monotonically increasing function of α = µ/σ which goes from 0 at α = 0 to 1 for α > 3.5. In particular, for σ = 1, the differential entropy in bits goes from 2.05 to 3.05 depending on the value of µ; that is, depending on how far apart the two halves of the mixed Gaussian distribution are.

2. Conclusions

This paper calculates the differential entropy for a mixed Gaussian distribution governed by the pa- rameters µ and σ. A closed form solution was not available for one of the terms, however, this term was calculated numerically and tabulated, as well as estimated by analytic upper and lower bounds. For µ = 0 the entropy corresponds to the entropy for a pure Gaussian distribution; it monotonically increases to a well-defined limit for two well-separated Gaussian distribution halves (µ >> 0). Parameter estimation techniques based on information theory are one area where such calculations are likely to be useful.

Acknowledgements

The authors acknowledge the Naval Research Laboratory for providing funding for this work.

References

  1. Shannon, C. E. A Mathematical Theory of Communication. The Bell System Technical Journal 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
  2. Cover, T. M.; Thomas, J. A. Elements of Information Theory; John Wiley and Sons: New Jersey, 2006. [Google Scholar]
  3. Lazo, A. C. G. V.; Rathie, P. N. On the Entropy of Continuous Probability Distributions. IEEE Transactions on Information Theory 1978, IT-24(1), 120–122. [Google Scholar] [CrossRef]
  4. Michalowicz, J. V.; Nichols, J. M.; Bucholtz, F. Calculation of Differential Entropy for Continuous Probability Distributions. Technical Report MR/5650/, U. S. Naval Research Laboratory Technical Report. 2008. [Google Scholar]
  5. Bhatia, V.; Mulgrew, B. Non-parametric Likelihood Based Channel Estimator for Gaussian Mixture Noise. Signal Processing 2007, 87, 2569–2586. [Google Scholar] [CrossRef]
  6. Wang, Y. G.; Wu, L. A. Nonlinear Signal Detection from an Array of Threshold Devices for Non-Gaussian Noise. Digital Signal Processing 2007, 17(1), 76–89. [Google Scholar] [CrossRef]
  7. Tan, Y.; Tantum, S. L.; Collins, L. M. Cramer-Rao Lower Bound for Estimating Quadrupole Resonance Signals in Non-Gaussian Noise. IEEE Signal Processing Letters 2004, 11(5), 490–493. [Google Scholar] [CrossRef]
  8. Lu, Z. W. An Iterative Algorithm for Entropy Regularized Likelihood Learning on Gaussian Mixture with Automatic Model Selection. Neurocomputing 2007, 69(13–15), 1674–1677. [Google Scholar] [CrossRef]
  9. Mars, N. J. I.; Arragon, G. W. V. Time Delay Estimation in Nonlinear Systems. IEEE Transactions on Acoustics, Speech, and Signal Processing 1981, ASSP-29(3), 619–621. [Google Scholar] [CrossRef]
  10. Hild, K. E.; Pinto, D.; Erdogmus, D.; Principe, J. C. Convolutive Blind Source Separation by Minimizing Mutual Information Between Segments of Signals. IEEE Transactions on Circuits and Systems I 2005, 52(10), 2188–2196. [Google Scholar] [CrossRef] [Green Version]
  11. Rohde, G. K.; Nichols, J. M.; Bucholtz, F.; Michalowicz, J. V. Signal Estimation Based on Mutual Information Maximization. In ‘Forty-First Asilomar Conference on Signals, Systems, and Computers’, IEEE; 2007. [Google Scholar]
  12. Gradshteyn, I. S.; Ryzhik, I. M. Table of Integrals, Series and Products, 4th ed.; Academic Press: New York, 1965. [Google Scholar]
Figure 1. Probability Density Function of a Mixed Gaussian Distribution (µ = 2.0, σ = 1.0)
Figure 1. Probability Density Function of a Mixed Gaussian Distribution (µ = 2.0, σ = 1.0)
Entropy 10 00200 g001
Figure 2. Lower and upper bounds for I(α) vs. α
Figure 2. Lower and upper bounds for I(α) vs. α
Entropy 10 00200 g002
Table 1. Tabulated values for I(α) and α2I
Table 1. Tabulated values for I(α) and α2I
αI(α)α2I αI(α)α2I
0.00.0000.000 (continued)
0.10.0050.005 2.13.7650.645
0.20.0200.020 2.24.1850.656
0.30.0470.043 2.34.6260.664
0.40.0860.074 2.45.0890.671
0.50.1390.111 2.55.5740.676
0.60.2070.153 2.66.0800.680
0.70.2920.198 2.76.6070.683
0.80.3960.244 2.87.1540.686
0.90.5190.291 2.97.7220.688
1.00.6630.337 3.08.3110.689
1.10.8290.381 3.18.9200.690
1.21.0180.422 3.29.5490.691
1.31.2300.460 3.310.1980.692
1.41.4650.495 3.410.8680.692
1.51.7230.527 3.511.5580.692
1.62.0050.555 3.612.2670.693
1.72.3110.579 3.712.9970.693
1.82.6400.600 3.813.7470.693
1.92.9920.618 3.914.5170.693
2.03.3670.633 4.015.3070.693

Share and Cite

MDPI and ACS Style

Michalowicz, J.V.; Nichols, J.M.; Bucholtz, F. Calculation of Differential Entropy for a Mixed Gaussian Distribution. Entropy 2008, 10, 200-206. https://doi.org/10.3390/entropy-e10030200

AMA Style

Michalowicz JV, Nichols JM, Bucholtz F. Calculation of Differential Entropy for a Mixed Gaussian Distribution. Entropy. 2008; 10(3):200-206. https://doi.org/10.3390/entropy-e10030200

Chicago/Turabian Style

Michalowicz, Joseph V., Jonathan M. Nichols, and Frank Bucholtz. 2008. "Calculation of Differential Entropy for a Mixed Gaussian Distribution" Entropy 10, no. 3: 200-206. https://doi.org/10.3390/entropy-e10030200

Article Metrics

Back to TopTop