Abstract
This work is an extension of our earlier article, where a well-known integral representation of the logarithmic function was explored and was accompanied with demonstrations of its usefulness in obtaining compact, easily-calculable, exact formulas for quantities that involve expectations of the logarithm of a positive random variable. Here, in the same spirit, we derive an exact integral representation (in one or two dimensions) of the moment of a nonnegative random variable, or the sum of such independent random variables, where the moment order is a general positive non-integer real (also known as fractional moments). The proposed formula is applied to a variety of examples with an information-theoretic motivation, and it is shown how it facilitates their numerical evaluations. In particular, when applied to the calculation of a moment of the sum of a large number, n, of nonnegative random variables, it is clear that integration over one or two dimensions, as suggested by our proposed integral representation, is significantly easier than the alternative of integrating over n dimensions, as needed in the direct calculation of the desired moment.
1. Introduction
In mathematical analyses associated with many problems in information theory and related fields, one is often faced with the need to compute expectations of logarithmic functions of composite random variables (see, e.g., [1,2,3,4,5,6,7,8]), or moments of such random variables, whose order may be a general positive real, not even necessarily an integer (see, e.g., [9,10,11,12,13,14,15,16,17,18,19,20,21,22]).
In the case of the logarithmic function, the common practice is either to resort to approximate evaluations, provided by upper and lower bounds on the desired expression (for example, by using Jensen’s inequality) or to approximate the calculations by using the Taylor series expansion of the function . More recently, it has become popular to use the replica trick (see, e.g., Chapter 8 in [23]), which is a non-rigorous, but useful technique, borrowed from statistical physics.
In our earlier work [6], we demonstrated how the following well-known integral representation of the logarithmic function,
can be useful in a variety of application areas in the field of information theory, including both source and channel coding, as well as other aspects of this field. To calculate the expectation, , where X is a positive random variable, the idea is simply to invoke the integral representation (1) and to commute the expectation and integration operators, i.e.,
thereby replacing the calculation of by the calculation of the moment-generating function (MGF), for all , which is often much easier to express in closed form. Moreover, in frequently encountered situations where X is given by the sum of n independently identically distributed (i.i.d.) random variables, the MGF of X is given by the power of the MGF of a single random variable in the sum that forms X. This reduces the dimension of the integration from n (in the original expression) to a single dimension of the integration over u. Interestingly, this integral representation has also been used in the statistical physics literature (see, e.g., [23] (p. 140) [24,25]), but not as much as the replica trick.
In this paper, we proceed in the same spirit as in [6], and we extend the scope to propose an integral representation of a general moment of a nonnegative random variable, X, namely the expectation, for a given real . Obviously, when is an integer, this moment is simply given by the order derivative of the MGF of X, calculated at the origin, as is very well known. However, the integral representation we propose, in this work, applies to any non-integer, positive , and here too, it replaces the direct calculation of by integration of an expression that involves the MGF of X. We refer to this representation as an extension of (2), as the latter can be obtained as a special case of the formula for , by invoking one of the equivalent identities
While the proposed integral representation of can be readily obtained from [26] (p. 363, Identity (3.434.1)) in the range , the nontrivial extension we propose for a non-integer and real is new to the best of our knowledge.
Fractional moments have been considered in the mathematical literature (see, e.g., [27,28,29,30]). A relationship between fractional and integer-order moments was considered in [27] by expressing a fractional moment as an infinite series, which depends on all the positive integer-order moments, followed by an algorithm for numerical calculations of fractional moments.
As in [6], the proposed integral representation is applied to a variety of examples with an information-theoretic motivation, and it is shown how it facilitates the numerical evaluations. In particular, similar to the case of the logarithmic function, when applied to the calculation of a moment of the sum of a large number, n, of nonnegative random variables, it is clear that integration over one or two dimensions, as suggested by our proposed integral representation, is significantly easier than the alternative of integrating over n dimensions, as needed in the direct calculation of the desired moment. Furthermore, single- or double-dimensional integrals can be instantly and accurately calculated using built-in numerical integration procedures.
Fractional moments have been considered in the mathematical literature (see, e.g., [27,28,29,30]). A relationship between fractional and integer-order moments was considered in [27] by expressing a fractional moment as an infinite series that depends on all the positive integer-order moments, which was followed by an algorithm for numerical calculations of fractional moments.
The outline of the remainder part of this paper is as follows. In Section 2, we provide the mathematical background associated with the integral representation in general. In Section 3, we demonstrate this integral representation in applications, including: moments of guesswork, moments of estimation errors, differential Rényi entropies of generalized multivariate Cauchy distributions, and mutual information calculations of a certain model of a jammed channel. Each one of these examples occupies one subsection of Section 3. The integral representations in this paper are not limited to the examples in Section 3, and such representations can be proven useful in other information-theoretic problems (see, e.g., [6] and the references therein for the integral representation of the logarithmic expectation and some of its information-theoretic applications).
2. Statistical Moments of Arbitrary Positive Orders
It is well known that integer-order moments of a random variable X are calculable from its MGF
by using its derivatives, calculated at , i.e.,
Quite often, however, there is a theoretical and practical interest to calculate fractional moments of nonnegative random variables. We next obtain a closed-form integral expression of the moment of a nonnegative random variable X, as a functional of its MGF, for any positive real . Before we proceed, it should be noted that for , such an expression is available in handbooks of standard tables of integrals, for example in [26] (p. 363, Identity (3.434.1)). The first innovation here, however, is in a nontrivial extension of this formula for all as an expression that involves a one-dimensional integral. It should be noted that although the definition of a fractional moment of a random variable (RV) is also given by a one-dimensional integral (or a sum, depending on whether the RV is discrete or continuous), the utility of our formula is, e.g., in expressing the moment of a sum of nonnegative and independent random variables as a one-dimensional integral, instead of an n-dimensional integral, which is obtained by the direct definition. This new formula serves as the basic building block in all of our information-theoretic applications throughout this paper.
We first define the Beta and Gamma functions (see, e.g., Section 8.3 in [26] and Chapter 5 in [31]):
where these functions are related by the equality
Theorem 1.
Let X be a nonnegative random variable, and let be a non-integer real. Then,
where, for all ,
Proof.
See Appendix A. □
Remark 1.
The proof of (9) in Appendix A does not apply to (see (A7) and (A8) etc., where the denominators vanish for ). In the latter case, by referring to the second term on the right-hand side of (9), we get , and also, the integral diverges (specifically, for , the integrand scales like for u that is sufficiently close to zero), yielding an expression of the type . However, taking a limit in (9) where we let ρ tend to an integer and applying L’Hôpital’s rule can reproduce the well-known result in (5).
Corollary 1.
For any ,
Proof.
Remark 2.
Corollary 1 also follows from [26] (p. 363, Identity (3.434.1)) (see Section 4 in [6]).
Corollary 2.
[6] Let X be a positive random variable. Then,
A proof of (15) was presented in Section 2 in [6], based on the integral representation of the logarithmic function in (1), and by interchanging the integration and the expectation (due to Fubini’s theorem). It can be alternatively proven by using Corollary 1, the identity
and swapping the order of the expectation and limit by the dominated convergence theorem.
Identity (15) has many useful information-theoretic applications in its own right, as demonstrated in [6], and here, we add even some more. The current work is an extension and further development of [6], whose main theme is exploiting Theorem 1 and studying its information-theoretic applications, as well as some more applications of the logarithmic expectation.
3. Applications
In this section, we exemplify the usefulness of the integral representation of the moment in Theorem 1 and the logarithmic expectation in several problem areas in information theory and statistics. These include analyses of randomized guessing, estimation errors, Rényi entropy of n-dimensional generalized Cauchy distributions, and finally, calculations of the mutual information for channels with a certain jammer model. To demonstrate the direct computability of the relevant quantities, we also present graphs of their numerical calculations.
3.1. Moments of Guesswork
Consider the problem of guessing the realization of a random variable, which takes on values in a finite alphabet, using a sequence of yes/no questions of the form “Is ?”, “Is ?”, etc., until a positive response is provided by a party that observes the actual realization of X. Given a distribution of X, a commonly used performance metric for this problem is the expected number of guesses or, more generally, the moment of the number of guesses until X is guessed successfully. When it comes to guessing random vectors, say, of length n, minimizing the moments of the number of guesses by different (deterministic or randomized) guessing strategies has several applications and motivations in information theory, such as sequential decoding, guessing passwords, etc., and it is also strongly related to lossless source coding (see, e.g., [9,10,11,12,13,19,20,21,22,32,33,34]). In this vector case, the moments of the number of guesses behave as exponential functions of the vector dimension, n, at least asymptotically, as n grows without bound. For random vectors with i.i.d. components, the best achievable asymptotic exponent of the guessing moment is expressed in [9] by using the Rényi entropy of X of order . Arikan assumed in [9] that the distribution of X is known and analyzed the optimal deterministic guessing strategy, which orders the guesses according to nonincreasing probabilities. Refinements of the exponential bounds in [9] with tight upper and lower bounds on the guessing moments for optimal deterministic guessing were recently derived in [19]. In the sequel, we refer to randomized guessing strategies, rather than deterministic strategies, and we aim to derive exact, calculable expressions for their associated guessing moments (as is later explained in this subsection).
Let the random variable X take on values in a finite alphabet . Consider a random guessing strategy where the guesser sequentially submits a sequence of independently drawn random guesses according to a certain probability distribution, , defined on . Randomized guessing strategies have the advantage that they can be used by multiple asynchronous agents, which submit their guesses concurrently (see [33,34]).
In this subsection, we consider the setting of randomized guessing and obtain an exact representation of the guessing moment in the form of a one-dimensional integral. Let be any realization of X, and let the guessing distribution, , be given. The random number, G, of independent guesses until success has the geometric distribution
and so, the corresponding MGF is equal to
In view of (9)–(11) and (18), for and non-integer ,
with , and for all
In (20), is a polylogarithm (see, e.g., Section 25.12 in [31]), which is given by
with denoting differentiation with respect to x and multiplication of the derivative by x, repeatedly j times. In particular, we have
and so on. The function is a built-in function in the MATLAB and Mathematica software, which is expressed as . By Corollary 1, if , then (19) is simplified to
Let P denote the distribution of X. Averaging over X to get the unconditional moment using (23), one obtains for all ,
where (24) is obtained by using the substitution . A suitable expression of such an integral is similarly obtained, for all , by averaging (19) over X. In comparison, a direct calculation of the moment gives
The double sum in (25) involves a numerical computation of an infinite series, where the number of terms required to obtain a good approximation increases with and needs to be determined. The right-hand side of (24), on the other hand, involves integration over . For every practical purpose, however, definite integrals in one or two dimensions can be calculated instantly using built-in numerical integration procedures in MATLAB, Maple, Mathematica, or any other mathematical software tools, and the computational complexity of the integral in (24) is not affected by .
As a complement to (19) (which applies to a non-integral and positive ), we obtain that the moment of the number of randomized guesses, with , is equal to
where (26) follows from (20) and since . By averaging over X,
To conclude, (19) and its simplification in (23) for give calculable one-dimensional integral expressions for the guessing moment with any . This refers to a randomized guessing strategy whose practical advantages were further explained in [33,34]. This avoids the need for numerical calculations of infinite sums. A further simplification for is provided in (26) and (27), expressed in closed form as a function of polylogarithms.
3.2. Moments of Estimation Errors
Let be i.i.d. random variables with an unknown expectation to be estimated, and consider the simple estimator,
For given , we next derive an easily-calculable expression of the moment of the estimation error.
Let and . By Theorem 1, if is a non-integral multiple of two, then
where
, and for all (see (11))
By Corollary 1 and (29), if in particular , then the right-hand side of (30) is simplified to
and, for all ,
In view of (29)–(34), obtaining a closed-form expression for the moment of the estimation error, for an arbitrary , hinges on the calculation of the right side of (31) for all . To this end, we invoke the identity
which is the MGF of a zero-mean Gaussian random variable with variance . Together with (31), it gives (see Appendix B.1)
where X is a generic random variable with the same distribution as for all i.
The combination of (30)–(34) enables calculating exactly the moment , for any given , in terms of a two-dimensional integral. Combining (33) and (36) yields, for all ,
where we have used the identity in the derivation of the first term of the integral on the right-hand side of (37).
As an example, consider the case where are i.i.d. Bernoulli random variables with
where the characteristic function is given by
Thanks to the availability of the exact expression, we can next compare the exact moment of the estimation error , with the following closed-form upper bound (see Appendix B.2) and thereby assess its tightness:
which holds for all , , and , with
Figure 1 and Figure 2 display plots of as a function of and n, in comparison to the upper bound (40). The difference in the plot of Figure 1 is significant except for the boundaries of the interval , where both the exact value and the bound vanish. Figure 2 indicates that the exact value of , for large n, scales like ; this is reflected by the apparent parallelism of the curves in both graphs and by the upper bound (40).
To conclude, this subsection provides an exact, double-integral expression for the moment of the estimation error of the expectation of n i.i.d. random variables. In other words, the dimension of the integral does not increase with n, and it is a calculable expression. We further compare our expression with an upper bound that stems from concentration inequalities. Although the scaling of the bound as a polynomial of n is correct, the difference between the exact expression and the bound is significant (see Figure 1 and Figure 2).
3.3. Rényi Entropy of Extended Multivariate Cauchy Distributions
Generalized Cauchy distributions, their mathematical properties, and applications are of interest (see, e.g., [6,35,36,37]). The Shannon differential entropy of a family of generalized Cauchy distributions was derived in Proposition 1 in [36], and also, a lower bound on the differential entropy of a family of extended multivariate Cauchy distributions (cf. Equation (42) in [37]) was derived in Theorem 6 in [37]. Furthermore, an exact single-letter expression for the differential entropy of the different family of extended multivariate Cauchy distributions was recently derived in Section 3.1 in [6]. Motivated by these studies, as well as the various information-theoretic applications of Rényi information measures, we apply Theorem 1 to obtain the Rényi (differential) entropy of an arbitrary positive order for the extended multivariate Cauchy distributions in Section 3.1 in [6]. As we shall see in this subsection, the integral representation for the Rényi entropy of the latter family of extended multivariate Cauchy distributions is two-dimensional, irrespective of the dimension n of the random vector.
Let be a random vector whose probability density function is of the form
for a certain function and a positive constant q such that
We refer to this kind of density (see also Section 3.1 in [6]) as a generalized multivariate Cauchy density because the multivariate Cauchy density function is the special case pertaining to the choices and . The differential Shannon entropy of the generalized multivariate Cauchy density was derived in Section 3.1 in [6] using the integral representation of the logarithm (1), where it was presented as a two-dimensional integral.
We next extend the analysis of [6] to differential Rényi entropies of an arbitrary positive order (recall that the differential Rényi entropy is specialized to the differential Shannon entropy at [38]). We show that, for the generalized multivariate Cauchy density, the differential Rényi entropy can be presented as a two-dimensional integral, rather than an n-dimensional integral. Defining
we get from (42) (see Section 3.1 in [6]) that
For , with a fixed , (44) implies that
In particular, for and , we get the multivariate Cauchy density from (42). In this case, it follows from (46) that for , and from (45)
For , the (differential) Rényi entropy of order is given by
Using the Laplace transform relation,
we obtain that, for (see Appendix C),
If , we distinguish between the following two cases:
- (a)
- If for some , thenwith
- (b)
- Otherwise (i.e., if ), thenwhere , and for all ,
The proof of the integral expressions of the Rényi entropy of order , as given in (50)–(54), is provided in Appendix C.
Once again, the advantage of these expressions, which do not seem to be very simple (at least on the face of it), is that they only involve one- or two-dimensional integrals, rather than an expression of an n-dimensional integral (as it could have been in the case of an n-dimensional density).
3.4. Mutual Information Calculations for Communication Channels with Jamming
Consider a channel that is fed by an input vector and generates an output vector , where and are either finite, countably infinite, or continuous alphabets, and and are their order Cartesian powers. Let the conditional probability distribution of the channel be given by
where and are given conditional probability distributions of Y given X, and . This channel model refers to a discrete memoryless channel (DMC), which is nominally given by
where one of the transmitted symbols is jammed at a uniformly distributed random time, i, and the transition distribution of the jammed symbol is given by instead of . The restriction to a single jammed symbol is made merely for the sake of simplicity, but it can easily be extended.
We wish to evaluate how the jamming affects the mutual information . Clearly, when one talks about jamming, the mutual information is decreased, but this is not part of the mathematical model, where the relation between r and q has not been specified. Let the input distribution be given by the product form
The mutual information (in nats) is given by
For the simplicity of notation, we henceforth omit the domains of integration whenever they are clear from the context. We have,
By using the logarithmic expectation in (15) and the following equality (see (55) and (56)):
we obtain (see Appendix D.1)
where, for ,
Moreover, owing to the product form of , it is shown in Appendix D.2 that
Combining (60), (62), and (65), we express as a double integral over , independently of n (rather than an integration over ):
We next calculate the differential channel output entropy, , induced by . From Appendix D.3,
where, for all ,
By (1), the following identity holds for every positive random variable Z (see Appendix D.3):
where . By setting where are i.i.d. random variables with the density function v, some algebraic manipulations give (see Appendix D.3)
where
Combining (58), (66), and (71), we obtain the mutual information for the channel with jamming, which is given by
We next exemplify our results in the case where q is a binary symmetric channel (BSC) with crossover probability and p is a BSC with a larger crossover probability, . We assume that the input bits are i.i.d. and equiprobable. The specialization of our analysis to this setup is provided in Appendix D.4, showing that the mutual information of the channel , fed by the binary symmetric source, is given by
where is the binary entropy function
with the convention that , and
denotes the binary relative entropy. By the data processing inequality, the mutual information in (75) is smaller than that of the BSC with crossover probability :
Figure 3 refers to the case where and . Here, , and is decreased by 2.88 nats due to the jammer (see Figure 3).
Figure 3.
The degradation in mutual information for . The jammer-free channel q is a binary symmetric channel (BSC) with crossover probability , and r is a BSC with crossover probability . The input bits are i.i.d. and equiprobable. The degradation in (nats) is displayed as a function of .
Figure 4 refers to the case where and (referring to complete jamming of a single symbol, which is chosen uniformly at random), and it shows the difference in the mutual information , as a function of the length n, between the jamming-free BSC with crossover probability and the channel with jamming.
Figure 4.
The degradation in mutual information as a function of n. The jammer-free channel is a BSC with crossover probability , and for the jammed symbol is a BSC with crossover probability . The input bits are i.i.d. and equiprobable.
To conclude, this subsection studies the change in the mutual information due to jamming, relative to the mutual information associated with the nominal channel without jamming. Due to the integral representations provided in our analysis, the calculation of the mutual information finally depends on one-dimensional integrals, as opposed to the original n-dimensional integrals, pertaining to the expressions that define the associated differential entropies.
Author Contributions
Both authors contributed equally to this research work, as well as to the write-up of this article. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A. Proof of Theorem 1
Let be a non-integer real, and define the function as follows:
with the convention that . By the Taylor series expansion of as a function of around , we find that for small positive u, the integrand of (A1) scales as with . Furthermore, for large u, the same integrand scales as . This guarantees the convergence of the integral, and so, is well defined and finite in the interval .
From (A1), (for , the integrand of (A1) is identically zero on ). Differentiation ℓ times with respect to , under the integration sign with , gives according to Leibniz integral rule
which implies that
We next calculate for and . By invoking the Leibniz integral rule,
Hence, from (A3) and (A4),
By integrating both sides of (A6) with respect to , successively k times, the equalities in (A5) imply that
with some suitable constants . Since (see (A5)), (A7) implies that
and since (by assumption) is a non-integer, the denominator on the right-hand side of (A8) is nonzero. Moreover, since for all (see (A5)), differentiation of both sides of (A7) ℓ times at yields
Substituting (A8) and (A9) into (A7) gives
Combining (A1) with (A10) and rearranging the terms, we obtain
Setting and taking expectations of both sides of (A11) imply, by swapping the order of integration and expectation (which is validated due to Fubini’s theorem), that the following equality holds (see (4) and (10)):
We next rewrite and simplify both terms in the right side of (A12) as follows:
and
Equations (A13), (A14), (A17), and (A19) are based on the recursion (see, e.g., [26] (Identity (8.331)))
(A15) relies on the relation between the Beta and Gamma functions in (6); (A16) is based on the following equality (see (6), (A20), and recall that ):
and finally, (A19) holds by using the identity (see, e.g., [26] (page 905, Identity (8.334)))
with (since, by assumption, is a non-integer). Combining (A12)–(A19) gives (9) (recall that , and holds by (A6)).
We finally prove (11). By (10), for all ,
Appendix B. Complementary Details of the Analysis in Section 3.2
Appendix B.1. Proof of Equation (36)
Appendix B.2. Derivation of the Upper Bound in (40)
For all ,
We next use the Chernoff bound for upper bounding for all ,
with , and
We now use an upper bound on for every . By Theorem 3.2 and Lemma 3.3 in [39] (see also Lemma 2.4.6 in [40]), we have
with
Combining (A30) and (A32) yields
Similarly, it is easy to show that the same Chernoff bound applies also to , which overall gives
Inequality (A35) is a refined version of Hoeffding’s inequality (see Section 2.4.4 in [40]), which is derived for the Bernoulli distribution (see (A30)) and by invoking the Chernoff bound; moreover, (A35) coincides with Hoeffding’s inequality in the special case (which, from (A33), yields ). In view of the fact that (A35) forms a specialization of Theorem 2.4.7 in [40], it follows that the Bernoulli case is the worst one (in the sense of leading to the looser upper bound) among all probability distributions whose support is the interval and whose expected value is . However, in the Bernoulli case, a simple symmetry argument applies for improving the bound (A35) as follows. Since are i.i.d., Bernoulli with mean , then obviously, are Bernoulli, i.i.d. with mean and (from (28))
which implies that the error estimation is identical in both cases. Hence, is symmetric around . It can be verified that
which follows from (A33) and since for all (see Figure 2.1 in [40]). In view of (A37) and the above symmetry consideration, the upper bound in (A35) is improved for values of , which therefore gives
From (28), the probability in (A38) vanishes if or . Consequently, for ,
where (A39)–(A43) hold, respectively, due to (A29), (A38), the substitution , (7), and (41).
Appendix C. Complementary Details of the Analysis in Section 3.3
We start by proving (50). In view of (48), for ,
where . For , we get
where (A45) holds due to (42); (A46) follows from (49), and (A47) holds by swapping the order of integrations, which is validated by Fubini’s theorem. Furthermore, from (42) and (49),
and it follows from (A48) and by swapping the order of integrations (due to Fubini’s theorem),
where (A49) holds by the definition of in (44). Finally, combining (45), (A44), (A47), and (A49) gives (50).
Appendix D. Calculations of the n-Dimensional Integrals in Section 3.4
Appendix D.1. Proof of Equations (62)–(64)
Appendix D.2. Proof of Equation (65)
Appendix D.3. Proof of Equations (67)–(73)
We next prove (70), which is used to calculate the entropy of with the density in (A64). In view of the integral representation of the logarithmic function in (1) and by interchanging the order of the integrations (due to Fubini’s theorem), we get that for a positive random variable Z
which proves (70). Finally, we prove (71). In view of (A64),
A calculation of the first integral on the right-hand side of (A66) gives
For , the inner integral on the right-hand side of (A67) satisfies
and for ,
Therefore, combining (A67)–(A69) gives
Finally, we calculate the second integral on the right-hand side of (A66). Let be the probability density function defined as
and let
where are i.i.d. -valued random variables with a probability density function v. Then, in view of (70), the second integral on the right-hand side of (A66) satisfies
The MGF of Z is equal to
where
and consequently, (A75) yields
and
Therefore, combining (A73)–(A77) gives the following single-letter expression for the second multi-dimensional integral on the right-hand side of (A66):
where the functions and are defined in (72) and (73), respectively. Combining (A66), (A70), and (A78) gives (71).
Appendix D.4. Specialization to a BSC with Jamming
In the BSC example considered, we have , and
where is the indicator function that is equal to one if the relation holds, and to zero otherwise. Recall that we assume . Let
be the binary symmetric source (BSS). From (63) and (64), for ,
Furthermore, we get from (A79), (A80), and (A81) that
and
Substituting (A82)–(A85) into (66) (where the integrals in (66) are replaced by sums) gives
Since the input is a BSS, due to the symmetry of the channel (55), the output is also a BSS. This implies that (in units of nats)
As a sanity check, we verify it by using (71). From (68) and (69), for ,
and from (72) and (73), it consequently follows that
and also
It can be verified that substituting (A88)–(A91) into (71) reproduces (A87). Finally, subtracting (A86) from (A87) gives (75).
References
- Dong, A.; Zhang, H.; Wu, D.; Yuan, D. Logarithmic expectation of the sum of exponential random variables for wireless communication performance evaluation. In 2015 IEEE 82nd Vehicular Technology Conference (VTC2015-Fall); IEEE: Piscataway, NJ, USA, 2015. [Google Scholar] [CrossRef]
- Appledorn, C.R. The entropy of a Poisson distribution. SIAM Rev. 1988, 30, 314–317. [Google Scholar] [CrossRef]
- Knessl, C. Integral representations and asymptotic expansions for Shannon and Rényi entropies. Appl. Math. Lett. 1998, 11, 69–74. [Google Scholar] [CrossRef]
- Lapidoth, A.; Moser, S. Capacity bounds via duality with applications to multiple-antenna systems on flat fading channels. IEEE Trans. Inf. Theory 2003, 49, 2426–2467. [Google Scholar] [CrossRef]
- Martinez, A. Spectral efficiency of optical direct detection. J. Opt. Soc. Am. B 2007, 24, 739–749. [Google Scholar] [CrossRef]
- Merhav, N.; Sason, I. An integral representation of the logarithmic function with applications in information theory. Entropy 2020, 22, 51. [Google Scholar] [CrossRef]
- Rajan, A.; Tepedelenlioǧlu, C. Stochastic ordering of fading channels through the Shannon transform. IEEE Trans. Inf. Theory 2015, 61, 1619–1628. [Google Scholar] [CrossRef]
- Zadeh, P.H.; Hosseini, R. Expected logarithm of central quadratic form and its use in KL-divergence of some distributions. Entropy 2016, 18, 288. [Google Scholar] [CrossRef]
- Arikan, E. An inequality on guessing and its application to sequential decoding. IEEE Trans. Inf. Theory 1996, 42, 99–105. [Google Scholar] [CrossRef]
- Arikan, E.; Merhav, N. Guessing subject to distortion. IEEE Trans. Inf. Theory 1998, 44, 1041–1056. [Google Scholar] [CrossRef]
- Arikan, E.; Merhav, N. Joint source-channel coding and guessing with application to sequential decoding. IEEE Trans. Inf. Theory 1998, 44, 1756–1769. [Google Scholar] [CrossRef]
- Boztaş, S. Comments on “An inequality on guessing and its application to sequential decoding”. IEEE Trans. Inf. Theory 1997, 43, 2062–2063. [Google Scholar] [CrossRef]
- Bracher, A.; Hof, E.; Lapidoth, A. Guessing attacks on distributed-storage systems. IEEE Trans. Inf. Theory 2019, 65, 6975–6998. [Google Scholar] [CrossRef]
- Bunte, C.; Lapidoth, A. Encoding tasks and Rényi entropy. IEEE Trans. Inf. Theory 2014, 60, 5065–5076. [Google Scholar] [CrossRef]
- Campbell, L.L. A coding theorem and Rényi’s entropy. Inf. Control 1965, 8, 423–429. [Google Scholar] [CrossRef]
- Courtade, T.; Verdú, S. Cumulant generating function of codeword lengths in optimal lossless compression. In Proceedings of the 2014 IEEE International Symposium on Information Theory, Honolulu, HI, USA, 29 June–4 July 2014; pp. 2494–2498. [Google Scholar] [CrossRef]
- Courtade, T.; Verdú, S. Variable-length lossy compression and channel coding: Non-asymptotic converses via cumulant generating functions. In Proceedings of the 2014 IEEE International Symposium on Information Theory, Honolulu, HI, USA, 29 June–4 July 2014; pp. 2499–2503. [Google Scholar] [CrossRef]
- Merhav, N. Lower bounds on exponential moments of the quadratic error in parameter estimation. IEEE Trans. Inf. Theory 2018, 64, 7636–7648. [Google Scholar] [CrossRef]
- Sason, I.; Verdú, S. Improved bounds on lossless source coding and guessing moments via Rényi measures. IEEE Trans. Inf. Theory 2018, 64, 4323–4346. [Google Scholar] [CrossRef]
- Sason, I. Tight bounds on the Rényi entropy via majorization with applications to guessing and compression. Entropy 2018, 20, 896. [Google Scholar] [CrossRef]
- Sundaresan, R. Guessing under source uncertainty. IEEE Trans. Inf. Theory 2007, 53, 269–287. [Google Scholar] [CrossRef]
- Sundaresan, R. Guessing based on length functions. In Proceedings of the 2007 IEEE International Symposium on Information Theory, Nice, France, 24–29 June 2007; pp. 716–719. [Google Scholar] [CrossRef]
- Mézard, M.; Montanari, A. Information, Physics, and Computation; Oxford University Press: New York, NY, USA, 2009. [Google Scholar]
- Esipov, S.E.; Newman, T.J. Interface growth and Burgers turbulence: The problem of random initial conditions. Phys. Rev. E 1993, 48, 1046–1050. [Google Scholar] [CrossRef]
- Song, J.; Still, S.; Rojas, R.D.H.; Castillo, I.P.; Marsili, M. Optimal work extraction and mutual information in a generalized Szilárd engine. arXiv 2019, arXiv:1910.04191. [Google Scholar]
- Gradshteyn, I.S.; Ryzhik, I.M. Tables of Integrals, Series, and Products, 8th ed.; Academic Press: New York, NY, USA, 2014. [Google Scholar]
- Gzyl, H.; Tagliani, A. Stieltjes moment problem and fractional moments. Appl. Math. Comput. 2010, 216, 3307–3318. [Google Scholar] [CrossRef]
- Gzyl, H.; Tagliani, A. Determination of the distribution of total loss from the fractional moments of its exponential. Appl. Math. Comput. 2012, 219, 2124–2133. [Google Scholar] [CrossRef]
- Tagliani, A. On the proximity of distributions in terms of coinciding fractional moments. Appl. Math. Comput. 2003, 145, 501–509. [Google Scholar] [CrossRef]
- Tagliani, A. Hausdorff moment problem and fractional moments: A simplified procedure. Appl. Math. Comput. 2011, 218, 4423–4432. [Google Scholar] [CrossRef]
- Olver, F.W.J.; Lozier, D.W.; Boisvert, R.F.; Clark, C.W. NIST Handbook of Mathematical Functions; NIST and Cambridge University Press: New York, NY, USA, 2010. [Google Scholar]
- Hanawal, M.K.; Sundaresan, R. Guessing revisited: A large deviations approach. IEEE Trans. Inf. Theory 2011, 57, 70–78. [Google Scholar] [CrossRef]
- Merhav, N.; Cohen, A. Universal randomized guessing with application to asynchronous decentralized brute-force attacks. IEEE Trans. Inf. Theory 2020, 66, 114–129. [Google Scholar] [CrossRef]
- Salamatian, S.; Huleihel, W.; Beirami, A.; Cohen, A.; Médard, M. Why botnets work: Distributed brute-force attacks need no synchronization. IEEE Trans. Inf. Foren. Sec. 2019, 14, 2288–2299. [Google Scholar] [CrossRef]
- Carrillo, R.E.; Aysal, T.C.; Barner, K.E. A generalized Cauchy distribution framework for problems requiring robust behavior. EURASIP J. Adv. Signal Process. 2010, 2010, 312989. [Google Scholar] [CrossRef]
- Alzaatreh, A.; Lee, C.; Famoye, F.; Ghosh, I. The generalized Cauchy family of distributions with applications. J. Stat. Distrib. App. 2016, 3, 12. [Google Scholar] [CrossRef]
- Marsiglietti, A.; Kostina, V. A lower bound on the differential entropy of log-concave random vectors with applications. Entropy 2018, 20, 185. [Google Scholar] [CrossRef]
- Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 8–9 August 1961; pp. 547–561. [Google Scholar]
- Berend, D.; Kontorovich, A. On the concentration of the missing mass. Electron. Commun. Probab. 2013, 18, 3. [Google Scholar] [CrossRef]
- Raginsky, M.; Sason, I. Concentration of Measure Inequalities in Information Theory, Communications and Coding. In Foundations and Trends in Communications and Information Theory, 3rd ed.; NOW Publishers: Hanover, MA, USA; Delft, The Netherlands, 2019. [Google Scholar] [CrossRef]
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).