Next Article in Journal
Quantum Misuse Attack on Frodo
Next Article in Special Issue
Rényi Entropies of Multidimensional Oscillator and Hydrogenic Systems with Applications to Highly Excited Rydberg States
Previous Article in Journal
Transactional Interpretation and the Generalized Poisson Distribution
Previous Article in Special Issue
Rényi Entropy, Signed Probabilities, and the Qubit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rényi Cross-Entropy Measures for Common Distributions and Processes with Memory

Department of Mathematics and Statistics, Queen’s University, Kingston, ON K7L 3N6, Canada
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(10), 1417; https://doi.org/10.3390/e24101417
Submission received: 9 September 2022 / Revised: 2 October 2022 / Accepted: 2 October 2022 / Published: 4 October 2022
(This article belongs to the Special Issue Rényi Entropy: Sixty Years Later)

Abstract

:
Two Rényi-type generalizations of the Shannon cross-entropy, the Rényi cross-entropy and the Natural Rényi cross-entropy, were recently used as loss functions for the improved design of deep learning generative adversarial networks. In this work, we derive the Rényi and Natural Rényi differential cross-entropy measures in closed form for a wide class of common continuous distributions belonging to the exponential family, and we tabulate the results for ease of reference. We also summarise the Rényi-type cross-entropy rates between stationary Gaussian processes and between finite-alphabet time-invariant Markov sources.

1. Introduction

The Rényi entropy [1] of order α of a probability mass function p with finite support S is defined as
H α ( p ) = 1 1 α ln x S p ( x ) α
for α > 0 , α 1 . The Rényi entropy generalizes the Shannon entropy,
H ( p ) = x S p ( x ) ln p ( x ) ,
in the sense that as α 1 , H α ( p ) H ( p ) . Several other Rényi-type information measures have been put forward, each obeying the condition that their limit as α goes to one reduces to a Shannon-type information measure. This includes the Rényi divergence (of order α ) between two discrete distributions p and q with common finite support S , given by
D α ( p | | q ) = 1 α 1 ln x S p ( x ) α q ( x ) 1 α
which reduces to the familiar Kullback–Leibler divergence,
D KL ( p | | q ) = x S p ( x ) ln p ( x ) q ( x ) ,
as α 1 . Note that in some cases [2], there may exist multiple Rényi-type generalisations for the same information measure (particularly for mutual information).
Many of these definitions admit natural counterparts in the case when the involved distributions have a probability density function (pdf). This gives rise to information measures such as the Rényi differential entropy for pdf p with support S ,
h α ( p ) = 1 1 α ln S p ( x ) α d x ,
and the Rényi differential divergence between pdfs p and q with common support S ,
D α ( p q ) = 1 α 1 ln S p ( x ) α q ( x ) 1 α d x .
The Rényi cross-entropy between distributions p and q is an analogous generalization of the Shannon cross-entropy
H ( p ; q ) = x S p ( x ) ln q ( x ) .
Two definitions for this measure have been recently suggested. In light of the fact that Shannon’s cross-entropy satisfies H ( p ; q ) = D ( p q ) + H ( p ) , a natural definition of the Rényi cross-entropy is:
H ˜ α ( p ; q ) D α ( p | | q ) + H α ( p ) .
This definition was indeed proposed in [3] in the continuous case, with the differential cross-entropy measure given by
h ˜ α ( p ; q ) D α ( p | | q ) + h α ( p ) .
In contrast, prior to [3], the authors of [4] introduced the Rényi cross-entropy in their study of shifted Rényi measures expressed as the logarithm of weighted generalized power means. Specifically, upon simplifying Definition 6 in [4], their expression for the Rényi cross-entropy between distributions p and q is given by
H α p ; q 1 1 α ln x S p x q x α 1 .
For the continuous case, (10) can be readily converted to yield the Rényi differential cross-entropy between pdfs p and q:
h α ( p ; q ) 1 1 α ln S p ( x ) q ( x ) α 1 d x .
Note that both (8) and (10) reduce to the Shannon cross-entropy H ( p ; q ) as α 1 [5]. A similar result holds for (9) and (11), where the Shannon differential cross-entropy,
h ( p ; q ) = S p ( x ) ln q ( x ) d x ,
is obtained. Further, the Rényi (differential) entropy is recovered in all equations when p = q (almost everywhere). These properties alone make these definitions viable extensions of the Shannon (differential) cross-entropy.
Finding closed-form expressions for the cross-entropy measure in (9) for continuous distributions is direct, since the Rényi divergence and the Rényi differential entropy were already calculated for numerous distributions in [6,7], respectively. However, deriving the measure in (11) is more involved. We hereafter refer to the measures H ˜ α ( p ; q ) in (8) and h ˜ α ( p ; q ) in (9) as the Natural Rényi cross-entropy and the Natural Rényi differential cross-entropy, respectively; while we plainly call the measures H α ( p ; q ) in (10) and h α ( p ; q ) (11) as the Rényi cross-entropy and the Rényi differential cross-entropy, respectively.
In a recent conference paper [8], we showed how to calculate the Rényi differential cross-entropy h α ( p ; q ) between distributions of the same type from the exponential family. Building upon the results shown there, the purpose of this paper is to derive in closed form the expression of h α ( p ; q ) for thirteen commonly used univariate distributions from the exponential family, as well as for multivariate Gaussians, and tabulate the results for ease of reference. We also analytically derive the Natural Rényi differential cross-entropy h ˜ α ( p ; q ) for the same set of distributions. Finally, we present tables summarising the Rényi and Natural Rényi (differential) cross-entropy rate measures, along with their Shannon counterparts, for two important classes of sources with memory, namely stationary Gaussian sources and finite-state time-invariant Markov sources.
Motivation for determining formulae for the Rényi cross-entropy originates from the use of the Shannon differential cross-entropy as a loss function for the design of deep learning generative adversarial networks (GANs) in [9]. The parameter α , ubiquitous to all Rényi information measures, allows one to fine-tune the loss function to improve the quality of the GAN-generated output. This can be seen in [3,5,10], which used the Rényi differential cross-entropy, and the Natural Rényi differential cross-entropy measures, respectively, to generalize the original GAN loss function (which is recovered as α 1 ), resulting in both improved GAN system stability and performance for multiple image datasets. It is also shown in [5,10] that the introduced Rényi-centric generalized loss function preserves the equilibrium point satisfied by the original GAN via the so-called Jensen-Rényi divergence [11], a natural extension of the Jensen–Shannon divergence [12] upon which the equilibrium result of [9] is established. Other GAN systems that utilize different generalized loss functions were recently developed and analysed in [13,14,15] (see also the references therein for prior work).
The rest of this paper is organised as follows. In Section 2, the formulae for the Rényi differential cross-entropy and Natural Rényi differential cross-entropy for distributions from the exponential family are given. In Section 3, these calculations are systematically carried out for fourteen pairs of distributions of the same type within the exponential family, and the results are presented in two tables. The Rényi and Natural Rényi differential cross-entropy rates are presented in Section 4 for stationary Gaussian sources; furthermore, the Rényi and Natural Rényi cross-entropy rates are provided in Section 5 for finite-state time-invariant Markov sources. Finally, the paper is concluded in Section 6.

2. Rényi and Natural Rényi Differential Cross-Entropies for Distributions from the Exponential Family

An exponential family is a class of probability distributions over a support S R n defined by a parameter space Θ R m and functions b : S R , c : Θ R , T : S R m , and η : Θ R m such that the pdf in this family has the form
f ( x ) = c ( θ ) b ( x ) exp η ( θ ) , T ( x ) , x S
where · , · denotes the standard inner product in R m . Alternatively, by using the (natural) parameter η = η ( θ ) , the pdf can also be written
f ( x ) = b ( x ) exp η , T ( x ) + A ( η ) ,
where A ( η ) : η ( Θ ) R with A ( η ) = ln c ( θ ) . Examples of important pdfs we consider from the exponential family are included in Appendix A.
In [8], the cross-entropy between pdfs f 1 and f 2 of the same type from the exponential family was proven to be
h α f 1 ; f 2 = A η 1 A η h + ln E h 1 α A η 2 ,
where
E h = E f h b ( X ) α 1 = b ( x ) α 1 f h ( x ) d x .
Here, f h refers to a distribution of the same type as f 1 and f 2 within the exponential family with natural parameter
η h η 1 + ( α 1 ) η 2 .
It can also be shown that the Natural Rényi differential cross-entropy between f 1 and f 2 is given by
h ˜ α f 1 ; f 2 = A η α A α η 1 + ln E α 1 α A η 2 ,
where
η α = α η 1 + ( 1 α ) η 2 ,
and
E α = E f α 1 b ( X ) α 1 = b ( x ) α 1 f α 1 ( x ) d x
where f α 1 refers to a distribution of the same type as f 1 and f 2 within the exponential family with natural parameter α η 1 .

3. Tables of Rényi and Natural Rényi Differential Cross-Entropies

Table 1 and Table 2 list Rényi and Natural differential cross-entropy expressions, respectively, between common distributions of the same type from the exponential family (which we describe in Appendix A for convenience). The closed-form expressions were derived using (15) and (18), respectively. In the tables, the subscript i is used to denote that a parameter belongs to pdf f i , i = 1 , 2 .

4. Rényi and Natural Rényi Differential Cross-Entropy Rates for Stationary Gaussian Processes

In [8], the Rényi differential cross-entropy rate between two stationary zero-mean Gaussian processes is derived. We present its formula alongside the expressions of the Shannon and Natural Rényi differential cross-entropy rates in Table 3. Here, f ( λ ) is the spectral density of the first zero-mean Gaussian process, g ( λ ) is the spectral density of the second Gaussian process,
h ( λ ) = f ( λ ) + ( α 1 ) g ( λ ) ,
and
j ( λ ) = α f ( λ ) + ( 1 α ) g ( λ ) .

5. Rényi and Natural Rényi Cross-Entropy Rates for Markov Sources

In [8], the Rényi cross-entropy rate between finite-state time-invariant Markov sources was established, using, as in [16], tools from the theory of non-negative matrices and Perron–Frobenius theory (e.g., cf. [17,18]). This measure, as well as the Shannon and Natural Rényi differential cross-entropy rates, are derived and summarised in Table 4. Here, P and Q are the m × m (stochastic) transition matrices associated with the first and second Markov sources, respectively, where both sources have a common alphabet of size m. To allow any value of the Rényi parameter α in ( 0 , 1 ) ( 1 , ) , we assume that the transition matrix Q of the second Markov chain has positive entries ( Q > 0 ); however, the transition matrix P of the first Markov chain is taken to be an arbitrary stochastic matrix. For simplicity, we assume that the initial distribution vectors, p and q, of both Markov chains also have positive entries ( p > 0 and q > 0 ). This condition can be relaxed via the approach used to prove Theorem 1 in [16]. Moreover, π p T denotes the stationary probability row vector associated with the first Markov chain, and 1 is an m-dimensional column vector in which each element equals one. Furthermore, ⊙ denotes element-wise multiplication (i.e., the Hadamard product operation), and ln ˙ is the element-wise natural logarithm.
Finally, the definition of λ ( R ) : R m × m R for a matrix R is more involved. If R is irreducible, λ ( R ) is its largest positive eigenvalue. Otherwise, rewriting R in its canonical form as detailed in Proposition 1 in [16], we have that λ ( R ) = max ( λ * , λ * ) , where λ * is the maximum of all largest positive eigenvalues of (irreducible) sub-matrices of R corresponding to self-communicating classes, and λ * is the maximum of all largest positive eigenvalues of sub-matrices of R corresponding to classes reachable from an inessential class.

6. Conclusions

We have derived closed-form formulae for the Rényi and Natural Rényi differential cross-entropies of commonly used distributions from the exponential family. This is of potential use to further studies in information theory and machine learning, particularly problems where deep neural networks, trained according to a Shannon cross-entropy loss function, can be improved via generalized Rényi-type loss functions in virtue of the extra degree of freedom provided by the Rényi ( α ) parameter. In addition, we have provided formulae for the Rényi and Natural Rényi differential cross-entropy rates for stationary zero-mean Gaussian processes and expressions for the cross-entropy rates for Markov sources. Further work includes expanding the present collection by considering distributions such as Levy or Weibull and investigating cross-entropy measures based on the f-divergence [19,20,21], starting with Arimoto’s divergence [22].

Author Contributions

Conceptualization, investigation, and manuscript preparation, all authors; derivation, F.C.T.; validation and supervision, F.A. and T.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Sciences and Engineering Research Council (NSERC) of Canada.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In Table A1, we describe the distributions of Table 1 and Table 2 (for the multivariate Gaussian, μ R n is a mean vector, and Σ 2 R n × R n is a positive definite covariance matrix).
Table A1. Distributions listed in Table 1 and Table 2.
Table A1. Distributions listed in Table 1 and Table 2.
NameParameters ( Θ )PDF f ( x ) Support
Beta( a > 0 , b > 0 ) B ( a , b ) x a 1 ( 1 x ) b 1 S = ( 0 , 1 )
χ ( scaled ) ( k > 0 , σ > 0 ) 2 1 k / 2 x k 1 e x 2 / 2 σ 2 σ k Γ k 2 S = R +
χ ( non scaled ) ( k > 0 ) 2 1 k / 2 x k 1 e x 2 / 2 Γ k 2 S = R +
χ 2 ( ν > 0 ) 1 2 ν 2 Γ ν 2 x ν 2 1 e x 2 S = R +
Exponential( λ > 0 ) λ e λ x S = R +
Gamma( k > 0 , θ > 0 ) 1 θ k Γ k x k 1 e k θ S = R +
Gaussian ( univariate ) ( μ , σ 2 > 0 ) 1 2 π σ 2 e 1 2 x μ σ 2 S = R
Gaussian ( multivariate ) ( μ , Σ 2 ) 1 ( 2 π ) n | Σ | e 1 2 x μ T Σ 1 x μ S = R n
Half-Normal( σ 2 > 0 ) 2 π σ 2 e 1 2 x σ 2 S = R +
Gumbel( μ , β > 0 ) 1 β exp x μ β + e x μ β S = R
Pareto( m > 0 , a > 0 ) a m a x ( 1 + m ) S = ( m , )
Maxwell Boltzmann( σ > 0 ) 2 x 2 π σ 6 e 1 2 x σ 2 S = R +
Rayleigh( σ 2 > 0 ) x σ 2 e 1 2 x σ 2 S = R +
Laplace( μ , b 2 > 0 ) 1 2 b e | x μ | b S = R
Notes
  • B ( a , b ) = 0 1 t a 1 ( 1 t ) b 1 d t is the Beta function.
  • Γ ( z ) = 0 x z 1 e x d x is the Gamma function.

References

  1. Rényi, A. On measures of entropy and information. Fourth Berkeley Symp. Math. Stat. Probab. 1961, 1, 547–561. [Google Scholar]
  2. Verdú, S. α-Mutual Information. In Proceedings of the 2015 Information Theory and Applications Workshop (ITA), San Diego, CA, USA, 1–6 February 2015; pp. 1–6. [Google Scholar]
  3. Sarraf, A.; Nie, Y. RGAN: Rényi generative adversarial network. SN Comput. Sci. 2021, 2, 17. [Google Scholar] [CrossRef]
  4. Valverde-Albacete, F.J.; Peláez-Moreno, C. The Case for shifting the Rényi entropy. Entropy 2019, 21, 46. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Bhatia, H.; Paul, W.; Alajaji, F.; Gharesifard, B.; Burlina, P. Least kth-Order and Rényi generative adversarial networks. Neural Comput. 2021, 33, 2473–2510. [Google Scholar] [CrossRef] [PubMed]
  6. Gil, M.; Alajaji, F.; Linder, T. Rényi divergence measures for commonly used univariate continuous distributions. Inf. Sci. 2013, 249, 124–131. [Google Scholar] [CrossRef]
  7. Song, K.S. Rényi information, loglikelihood and an intrinsic distribution measure. J. Statist. Plann. Inference 2001, 93, 51–69. [Google Scholar] [CrossRef]
  8. Thierrin, F.C.; Alajaji, F.; Linder, T. On the Rényi cross-entropy. In Proceedings of the 17th Canadian Workshop on Information Theory, Ottawa, ON, Canada, 5–8 June 2022; pp. 1–5. [Google Scholar] [CrossRef]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  10. Bhatia, H.; Paul, W.; Alajaji, F.; Gharesifard, B.; Burlina, P. Rényi generative adversarial Networks. arXiv 2020, arXiv:2006.02479v1. [Google Scholar]
  11. Kluza, P.A. On Jensen-Rényi and Jeffreys-Rényi type f-divergences induced by convex functions. Phys. Stat. Mech. Its Appl. 2019, 548, 122527. [Google Scholar] [CrossRef]
  12. Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 1991, 31, 145–151. [Google Scholar] [CrossRef] [Green Version]
  13. Pantazis, Y.; Paul, D.; Fasoulakis, M.; Stylianou, Y.; Katsoulakis, M. Cumulant GAN. arXiv 2020, arXiv:2006.06625. [Google Scholar] [CrossRef]
  14. Kurri, G.R.; Sypherd, T.; Sankar, L. Realizing GANs via a tunable loss function. In Proceedings of the IEEE Information Theory Workshop (ITW), Kanazawa, Japan, 17–21 October 2021; pp. 1–6. [Google Scholar]
  15. Kurri, G.R.; Welfert, M.; Sypherd, T.; Sankar, L. α-GAN: Convergence and estimation guarantees. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), Espoo, Finland, 26 June–1 July 2022; pp. 312–317. [Google Scholar]
  16. Rached, Z.; Alajaji, F.; Campbell, L.L. Rényi’s divergence and entropy rates for finite alphabet Markov sources. IEEE Trans. Inf. Theory 2001, 47, 1553–1561. [Google Scholar] [CrossRef]
  17. Seneta, E. Non-Negative Matrices and Markov Chains; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  18. Gallager, R.G. Discrete Stochastic Processes; Springer: Berlin/Heidelberg, Germany, 1996; pp. 31–55. [Google Scholar]
  19. Csiszar, I. Eine Informationstheoretische Ungleichung und ihre Anwendung auf den Bewis der Ergodizitat on Markhoffschen Ketten. Publ. Math. Inst. Hung. Acad. Sci. Ser. A 2006, 8, 85–108. [Google Scholar]
  20. Csiszár, I. Information-type measures of difference of probability distributions and indirect observations. Stud. Sci. Math. Hung. 1967, 2, 299–318. [Google Scholar]
  21. Ali, S.M.; Silvey, S.D. A general class of coefficients of divergence of one distribution from another. J. R. Stat. Society. Ser. B (Methodol.) 1966, 28, 131–142. [Google Scholar] [CrossRef]
  22. Liese, F.; Vajda, I. On divergences and informations in statistics and information theory. IEEE Trans. Inf. Theory 2006, 52, 4394–4412. [Google Scholar] [CrossRef]
Table 1. Rényi differential cross-entropies.
Table 1. Rényi differential cross-entropies.
Name h α ( f 1 ; f 2 )
Beta ln B ( a 2 , b 2 ) + 1 α 1 ln B ( a h , b h ) B ( a 1 , b 1 )
a h : = a 1 + ( α 1 ) ( a 2 1 ) , a h > 0
b h : = b 1 + ( α 1 ) ( b 2 1 ) , b h > 0
χ ( scaled ) 1 2 k 2 ln σ 2 2 σ h 2 ln 2 σ h 2 + ln Γ k 2 2
+ 1 α 1 ln Γ k h 2 ln Γ k 1 2 k 1 2 ln σ 1 2 σ h 2
σ h 2 : = 1 σ 1 2 + α 1 σ 2 2 , σ h 2 > 0
k h : = k 1 + ( α 1 ) ( k 2 1 ) , k h > 0
χ ( non-scaled ) 1 2 k 2 ln α ln 2 α + ln Γ k 2 2
+ 1 α 1 ln Γ k h 2 ln Γ k 1 2 k 1 2 ln α
k h : = k 1 + ( α 1 ) ( k 2 1 ) , k h > 0
χ 2 1 1 α ν 1 2 ln α ln Γ ν 1 2 + ln Γ ν h 2
+ 2 ν 2 2 ln α + ln 2 Γ ν 2 2
ν h : = ν 1 + ( α 1 ) ( ν 2 2 ) , ν h > 0
Exponential 1 1 α ln λ 1 λ h ln λ 2
λ h : = λ 1 + ( α 1 ) λ 2 , λ h > 0
Gamma ln Γ ( k 2 ) + k 2 ln θ 2
+ 1 1 α ln Γ ( k h ) Γ ( k 1 ) k h ln θ h k 1 ln θ 1
θ h : = θ 1 + ( a 1 ) θ 2 ( α 1 ) θ 1 θ 1 , θ h > 0
k h : = k 1 + ( α 1 ) k 2 , k h > 0
Gaussian ( univariate ) 1 2 ln ( 2 π σ 2 2 ) + 1 1 α ln σ 2 2 ( σ 2 ) h + ( μ 1 μ 2 ) 2 ( σ 2 ) h
( σ 2 ) h : = σ 2 2 + ( α 1 ) σ 1 2 , ( σ 2 ) h > 0
Gaussian ( multivariate ) 1 2 2 α ln | A | | Σ 1 | + 1 α ln 2 π n | Σ 2 | d
A : = Σ 1 1 + ( α 1 ) Σ 2 1 , A 0
d : = μ 1 T Σ 1 1 μ 1 + α 1 μ 2 T Σ 2 1 μ 2
( μ 1 T Σ 1 1 + α 1 μ 2 T Σ 2 1 ) A 1 ( Σ 1 1 μ 1 + α 1 Σ 2 1 μ 2 )
Gumbel ( β 1 = β 2 = β ) 1 1 α ln Γ ( 2 α ) β μ 1 β α ln η h + μ 2 β
η h : = e μ 1 / β + ( α 1 ) e μ 2 / β , η h > 0
Half-Normal 1 2 ln ( π σ 2 2 2 ) + 1 1 α ln σ 2 2 ( σ 2 ) h
( σ 2 ) h : = σ 2 2 + ( α 1 ) σ 1 2 , ( σ 2 ) h > 0
Laplace ( μ 1 = μ 2 = 0 ) ln ( 2 b 2 ) + 1 1 α ln b 2 2 b h
b h : = b 2 + ( 1 α ) b 1 , b h > 0
Maxwell Boltzmann 1 2 ln 2 π + 3 ln σ 2 2 + ln σ h 2
+ 1 1 α ln Γ ( 2 α ) Γ ( α ) 3 2 ln σ 1 2 σ h 2
σ h 2 : = σ 1 2 + ( α 1 ) σ 2 2 , σ h 2 > 0
Pareto ( m 1 = m 2 = m ) ln m ln λ 2 + 1 1 α ln λ 1 λ h
λ h : = λ 1 + α 1 λ 2 + 1 , λ h > 0
Rayleigh ln σ 1 2 α ln σ h 2 + ln Γ ( 1 α 2 ) 1 α + ln 2 σ 2 2
σ h 2 : = σ 1 2 + ( α 1 ) σ 2 2 , σ h 2 > 0
Table 2. Natural Rényi differential cross-entropies.
Table 2. Natural Rényi differential cross-entropies.
Name h ˜ α ( f 1 ; f 2 )
Beta ln B ( a 2 , b 2 ) + 1 α 1 ln B ( a α , b α ) B α a 1 1 + 1 , α b 1 1 + 1
a α : = α a 1 + ( 1 α ) a 2 , a α > 0
b α : = α b 1 + ( 1 α ) b 2 , b α > 0
χ ( scaled ) 1 2 ln 2 σ 1 2 α + k 2 ln σ 2 2 σ α 2 + ln Γ k 2 2
+ 1 1 α α k 1 ln σ α 2 σ 1 2 α 2 ln Γ ( k α 2 ) + ln Γ α ( k 1 1 ) + 1 2
σ α 2 : = α σ 1 2 + 1 α σ 2 2 , σ α 2 > 0
k α : = α k 1 + ( 1 α ) k 2 , k α > 0
χ ( non-scaled ) ln 2 α 2 + ln Γ ( k 2 2 )
+ 1 1 α ln Γ ( k α 2 ) α k 1 ln α 2 + ln Γ ( α ( k 1 1 ) + 1 2 )
k α : = α k 1 + ( 1 α ) k 2 , k α > 0
χ 2 1 1 α ln Γ ν α 2 + α ln Γ ν 1 2 + ln Γ ν 2 2
ν α : = α ν 1 + ( 1 α ) k , ν α > 0
Exponential 1 1 α ln λ 1 α λ α ln λ 2
λ α : = α λ 1 + ( 1 α ) λ 2 , λ α > 0
Gamma ln Γ ( k 2 ) + k 2 ln θ 2
+ 1 1 α ln Γ ( k 1 ) Γ ( k α ) k α ln θ α α 2 k 1 ln θ 1
θ α : = α θ 1 1 + ( 1 α ) θ 2 1 , k α : = α k 1 + ( 1 α ) k 2 , θ α > 0
Gaussian ( univariate ) 1 2 ln ( 2 π σ 2 2 ) + ( μ 1 μ 2 ) 2 ( σ 2 ) α + 1 1 α ln α σ 2 2 ( σ 2 ) α
( σ 2 ) α : = α σ 2 2 + ( 1 α ) σ 1 2 , ( σ 2 ) α > 0
Gaussian ( multivariate ) 1 2 2 α ln | α | + ln | A | | Σ 1 | + d + 1 2 ln ( 2 π ) n | Σ 1 | 2 | Σ 2 |
A : = α Σ 1 1 + ( 1 α ) Σ 2 1 , A 0
d : = ( μ 1 μ 2 ) T Σ 1 A Σ 2 ( μ 1 μ 2 )
Gumbel ( β 1 = β 2 = β ) μ 2 + α μ 1 β + 1 1 α ln Γ ( 2 α ) η α α β + μ 1 β
η α : = α e μ 1 / β + ( 1 α ) e μ 2 / β , η α > 0
Half-Normal 1 2 ln ( π σ 2 2 2 ) + 1 1 α ln α σ 2 2 ( σ 2 ) α
( σ 2 ) α : = α σ 2 2 + ( 1 α ) σ 1 2 , ( σ 2 ) α > 0
Laplace ( μ 1 = μ 2 = 0 ) ln b α + ln α b 1 1 α + ln 2 b 2
b α : = α b 1 + 1 α b 2 , b α > 0
Maxwell Boltzmann ln 2 + 3 ln σ 2 2 2 + ln α σ 1 2
+ 1 1 α 3 2 ln σ α σ 1 2 α α ln π 2 + ln Γ α + 1 2
σ α 2 : = α σ 1 2 + 1 α σ 2 2 , σ α 2 > 0
Pareto ( m 1 = m 2 = m ) 1 1 α ln λ α ln 1 α ( λ 1 1 ) ln λ 2 m
λ α : = α λ 1 + 1 α λ 2 , λ α > 0
Rayleigh ln σ 1 2 ( σ 2 ) α + ln α + ln Γ ( 1 α 2 ) 1 α + ln 2 σ 1 2
( σ 2 ) α : = α σ 1 2 + 1 2 ln 2 σ 1 4 σ 2 4 α , ( σ 2 ) α > 0
Table 3. Differential cross-entropy rates for stationary zero-mean Gaussian sources.
Table 3. Differential cross-entropy rates for stationary zero-mean Gaussian sources.
Information MeasureRateConstraint
Shannon Differential Cross-Entropy 1 2 ln 2 π + 1 4 π 0 2 π ln g ( λ ) + f ( λ ) g ( λ ) d λ g ( λ ) > 0
Natural R é nyi Differential Cross-Entropy 1 2 ln 4 π 2 α 1 α 1 + 1 4 π ( 1 α ) 0 2 π ln j ( λ ) g ( λ ) α d λ j ( λ ) g ( λ ) > 0
R é nyi Differential Cross-Entropy ln 2 π 2 + 1 4 π ( 1 α ) 0 2 π ( 2 α ) ln g ( λ ) ln h ( λ ) d λ g ( λ ) h ( λ ) > 0
Table 4. Cross-entropy rates for time-invariant Markov sources.
Table 4. Cross-entropy rates for time-invariant Markov sources.
Information MeasureRate
Shannon Cross-Entropy π p T P ln ˙ Q 1
Natural R é nyi Cross-Entropy 1 α 1 ln λ ( P α Q 1 α ) λ ( P α )
R é nyi Cross-Entropy 1 1 α ln λ ( P Q α 1 )
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Thierrin, F.C.; Alajaji, F.; Linder, T. Rényi Cross-Entropy Measures for Common Distributions and Processes with Memory. Entropy 2022, 24, 1417. https://doi.org/10.3390/e24101417

AMA Style

Thierrin FC, Alajaji F, Linder T. Rényi Cross-Entropy Measures for Common Distributions and Processes with Memory. Entropy. 2022; 24(10):1417. https://doi.org/10.3390/e24101417

Chicago/Turabian Style

Thierrin, Ferenc Cole, Fady Alajaji, and Tamás Linder. 2022. "Rényi Cross-Entropy Measures for Common Distributions and Processes with Memory" Entropy 24, no. 10: 1417. https://doi.org/10.3390/e24101417

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop