From Rényi Entropy Power to Information Scan of Quantum States

In this paper, we generalize the notion of Shannon’s entropy power to the Rényi-entropy setting. With this, we propose generalizations of the de Bruijn identity, isoperimetric inequality, or Stam inequality. This framework not only allows for finding new estimation inequalities, but it also provides a convenient technical framework for the derivation of a one-parameter family of Rényi-entropy-power-based quantum-mechanical uncertainty relations. To illustrate the usefulness of the Rényi entropy power obtained, we show how the information probability distribution associated with a quantum state can be reconstructed in a process that is akin to quantum-state tomography. We illustrate the inner workings of this with the so-called “cat states”, which are of fundamental interest and practical use in schemes such as quantum metrology. Salient issues, including the extension of the notion of entropy power to Tsallis entropy and ensuing implications in estimation theory, are also briefly discussed.


Introduction
The notion of entropy is undoubtedly one of the most important concepts in modern science. Very few other concepts can compete with it in respect to the number of attempts to clarify its theoretical and philosophical meaning [1]. Originally, the notion of entropy stemmed from thermodynamics, where it was developed to quantify the annoying inefficiency of steam engines. It then transmuted into a description of the amount of disorder or complexity in physical systems. Though many such attempts were initially closely connected with the statistical interpretation of the phenomenon of heat, in the course of time, they expanded their scope far beyond their original incentives. Along those lines, several approaches have been developed in attempts to quantify and qualify the entropy paradigm. These have been formulated largely independently and with different applications and goals in mind. For instance, in statistical physics, entropy counts the number of distinct microstates compatible with a given macrostate [2], in mathematical statistics, it corresponds to the inference functional for an updating procedure [3], and in information theory, it determines a limit on the shortest attainable encoding scheme [2,4].
Particularly distinct among these are the information-theoretic entropies (ITEs). This is not only because they discern themselves through their firm operational prescriptions in terms of coding theorems and communication protocols [5][6][7][8][9], but because they also offer an intuitive measure of disorder phrased in terms of missing information about a system. Apart from innate issues in communication theory, ITEs have also proved to be indispensable tools in other branches of science. Typical examples are provided by chaotic dynamical systems and multifractals (see, e.g., [10] and citations therein). Fully developed turbulence, earthquake analysis, and generalized dimensions of strange attractors provide √ τ being concurrence) [61], N 1/2 is related to both fidelity F and robustness R of a pure state [62], N ∞ quantifies the Bures distance to the closest separable pure state [63], etc. Even though our main focus here will be on REs and REPs since they are more pertinent in information theory, we will include some discussion related to Tsallis entropy powers at the end of this paper.
The aim of this paper is twofold. First, we wish to appropriately extend the notion of SE-based EP to the RE setting. In contrast to our earlier works on the topic [13,64], we will do it now by framing REP in the context of RE-based estimation theory. This will be done by judiciously generalizing such key notions as the De Bruijn identity, isoperimetric inequality (and ensuing Cramér-Rao inequality), and Stam inequality. In contrast to other similar works on the subject [65][66][67][68], our approach is distinct in three key respects: (a) we consistently use the notion of escort distribution and escort score vector in setting up the generalized De Bruijn identity and Fisher information matrix, (b) we generalize Stam's uncertainty principle, and (c) Rényi EP is related to variance of the reference Gaussian distribution rather than the Rényi maximizing distribution. As a byproduct, we derive within such a generalized estimation theory framework the Rényi-EP-based quantum uncertainty relations (REPUR) of Schrödinger-Roberston type. The REPUR obtained coincides with our earlier result [13] that was obtained in a very different context by means of the Beckner-Babenko theorem. This in turn serves as a consistency check of the proposed generalized estimation theory. Second, we identify interesting new playgrounds for the Rényi EPs obtained. In particular, we asked ourselves a question: assuming one is able in specific quantum protocols to measure Rényi EPs of various orders, how does this constrain the underlying quantum state distribution? To answer this question, we invoke the concept of the information distribution associated with a given quantum state. The latter contains a complete "information scan" of the underlying state distribution. We set up a reconstruction method based on Hausdorff's moment problem [69] to show explicitly how the information probability distribution associated with a given quantum state can be numerically reconstructed from EPs. This is a process that is analogous to quantum-state tomography. However, whereas tomography extracts the full density matrix from an ensemble using many measurements on a tomographically complete basis, the EP reconstruction method extracts the probability density on a given basis. This is an alternative approach that may be advantageous, for example, in quantum metrology schemes, where only knowledge of the local probability density rather than the full quantum state is needed [70].
The paper is structured as follows. In Section 2, we introduce the concept of Rényi's EP. With quantum metrology applications in mind, we discuss this in the framework of estimation theory. First, we duly generalize the notion of Fisher information (FI) by using a Rényi entropy version of De Bruijn's identity. In this connection, we emphasize the role of the so-called escort distribution, which appears naturally in the definition of higher-order score functions. Second, we prove the RE-based isoperimetric inequality and ensuing Cramér-Rao inequality and find how the knowledge of Fisher information matrix restricts possible values of Rényi's EP. Finally, we further illuminate the role of Rényi's EP by deriving (through the Stam inequality) Rényi's EP-based quantum uncertainty relations for conjugate observables. To flesh this out, the second part of the paper is devoted to the development of the use of Rényi EPs to extract the quantum state from incomplete data. This is of particular interest in various quantum metrology protocols. To this end, we introduce in Section 3 the concepts of information distribution, and, in Section 4, we show how cumulants of the information distribution can be obtained from knowledge of the EPs. With the cumulants at hand, one can reconstruct the underlying information distribution in a process which we call an information scan. Details of how one could explicitly realize such an information scan for quantum state PDFs are provided in Section 5. There we employ generalized versions of Gram-Charlier A and the Edgeworth expansion. In Section 6, we illustrate the inner workings of the information scan using the example of a so-called cat state. This state is of interest in applications of quantum physics such as quantumenhanced metrology, which is concerned with the optimal extraction of information from measurements subject to quantum mechanical effects. The cat state we consider is a superposition of the vacuum state and a coherent state of the electromagnetic field; two cases are studied comprising different probabilistic weightings of the superposition state corresponding to balanced and unbalanced cat states. Section 7 is dedicated to EPs based on Tsallis entropy. In particular, we show that Rényi and Tsallis EPs coincide with each other. This, in turn, allows us to phrase various estimation theory inequalities in terms of TE. In Section 7, we end with conclusions. For the reader's convenience, we relegate some technical issues concerning the generalized De Bruijn identity and associated isoperimetric and Stam inequalities to three appendices.

Rényi Entropy Based Estimation Theory and Rényi Entropy Powers
In this section, we introduce the concept of Rényi's EP. With quantum metrology applications in mind, we discuss this in the framework of estimation theory. This will not only allow us to find new estimation inequalities, such as the Rényi-entropy-based De Bruijn identity, isoperimetric inequality, or Stam inequality, but it will also provide a convenient technical and conceptual frame for deriving a one-parameter family of Rényientropy-power-based quantum-mechanical uncertainty relations.

Fisher Information-Shannon's Entropy Approach
First, we recall that the Fisher information matrix J(X ) of a random vector {X i } in R D with the PDF F (x) is defined as [38] J(X ) = cov(V (X )) , where the covariance matrix is associated with the random zero-mean vector-the so-called score vector, as A corresponding trace of J(X ), i.e., is known as the Fisher information. Both the FI and FI matrix can be conveniently related to Shannon's differential entropy via De Bruijn's identity [66,67]. De Bruijn's identity: Let {X i } be a random vector in R D with the PDF F (x) and let {Z G i } be a Gaussian random vector (noise vector) with zero mean and unit-covariance matrix, independent of {X i }. Then, where is Shannon's differential entropy (measured in nats). In the case when the independent additive noise {Z i } is non-Gaussian with zero mean and covariance matrix Σ = cov(Z ), then the following generalization holds [67]: The key point about De Bruijn's identity is that it provides a very useful intuitive interpretation of FI, namely, FI quantifies the sensitivity of transmitted (Shannon type) information to an arbitrary independent additive noise. An important aspect that should be stressed in this context is that FI as a quantifier of sensitivity depends only on the covariance of the noise vector, and thus it is independent of the shape of the noise distribution. This is because De Bruijn's identity remains unchanged for both Gaussian and non-Gaussian additive noise with the same covariance matrix.

Fisher Information-Rényi's Entropy Approach
We now extend the notion of the FI matrix to the Rényi entropy setting. A natural way to do it is via an extension of De Bruijn's identity to Rényi entropies. In particular, the following statement holds: Generalized De Bruijn's identity: Let {X i } be a random vector in R D with the PDF F (x) and let {Z i } be an independent (generally non-Gaussian) noise vector with the zero mean and covariance matrix Σ = cov(Z ), then, for any q > 0 where is Rényi's differential entropy (measured in nats) with I 1 = H. The ensuing FI matrix of order q has the explicit form with the score vector Here, ρ q = F q / R D F q dx is the so-called escort distribution [71]. The "cov q " denotes the covariance matrix computed with respect to ρ q . Proofs of both the conventional (i.e., Shannon entropy based) and generalized (i.e., Rényi entropy based) De Bruijn's identity are provided in Appendix A. There we also discuss some further useful generalizations of De Bruijn's identity. Finally, as in the Shannon case, we define the FI of order q-denoted as J q (X ), as

Rényi's Entropy Power and Generalized Isoperimetric Inequality
Similarly as in conventional estimation theory, one can expect that there should exist a close connection between the FI matrix J q (X ) and the corresponding Rényi entropy power N p (X ). In Shannon's information theory, such a connection is phrased in terms of isoperimetric inequality [67]. Here, we prove that a similar relationship works also in Rényi's information theory.
Let us start by introducing the concept of Rényi's entropy power. This is defined as the solution of the equation [13,64] where {Z G i } represents a Gaussian random vector with a zero mean and unit covariance matrix. Thus, N p (X ) denotes the variance of a would be Gaussian distribution that has the same Rényi information content as the random vector {X i } described by the PDF F (x). Expression (12) was studied in [13,64,72], where it was shown that the only class of solutions of (12) is with 1/p + 1/p = 1 and p ∈ R + . In addition, when p → 1 + , one has N p (X ) → N(X ), where N(X ) is the conventional Shannon entropy power [5]. In this latter case, one can use the asymptotic equipartition property [55,73] to identify N(X ) with "typical size" of a state set, which in the present context is the effective support set size for a random vector. This, in turn, is equivalent to Einstein's entropic principle [74]. In passing, it should be noted that the form of the Rényi EP expressed in (13) is not universally accepted version. In a number of works, it is defined merely as an exponent of RE, see, e.g., [75,76]. Our motivation for the form (13) is twofold: first, it has a clear interpretation in terms of variances of Gaussian distributions and, second, it leads to simpler formulas, cf. e.g., Equation (22). Generalized isoperimetric inequality: Let {X i } be a random vector in R D with the PDF F (x). Then, where the Rényi parameter q ≥ 1. We relegate the proof of the generalized isoperimetric inequality to Appendix B. It is also worth noting that the relation (14) implies another important inequality. By using the fact that the Shannon entropy is maximized (among all PDF's with identical covariance matrix Σ) by the Gaussian distribution, we have N 1 (X ) ≤ det(Σ) 1/D (see, e.g., [77]). If we further employ that I q is a monotonously decreasing function of q, see, e.g., [31,78], we can write (recall that q ≥ 1) The isoperimetric inequality (14) then implies We can further use the inequality (valid for any positive semi-definite D × D matrix A) to write where σ 2 is an average variance per component. Relations (16)-(18) represent the q-generalizations of the celebrated Cramér-Rao information inequality. In the limit of q → 1, we recover the standard Cramér-Rao inequality that is widely used in statistical inference theory [38,79]. A final logical step needed to complete the proof of REPURs is represented by the so-called generalized Stam inequality. To this end, we first define the concept of conjugate random variables. We say that random vectors {X i } and {Y i } in R D are conjugate if their respective PDF's F (x) and G(y) can be written as where the (generally complex) probability amplitudes ϕ F (x) ∈ L 2 (R D ) and ϕ G (y) ∈ L 2 (R D ) are mutual Fourier images, i.e., and analogously for ϕ G (y) =φ F (y). With this, we can state the generalized Stam inequality. Generalized Stam inequality (Stam's uncertainty principle): is valid for any r ∈ [1, ∞) and q ∈ [1/2, 1] that are connected via the relation 1/r + 1/q = 2.
In particular, if we define r = 2r and q = 2q, then r and q are Hölder conjugates. A proof of the generalized Stam inequality is provided in Appendix C.
Let us now consider Hölder conjugate indices p and q with p ∈ [2, ∞) (so that q ∈ [1,2]). Combining the isoperimetric inequality (14) together with the generalized Stam inequality (21), we obtain the following one-parameter class of REP-based inequalities By symmetry, the role of q and p can be reversed. In Refs. [13,64], we presented an alternative derivation of inequalities (22) that was based on the Beckner-Babenko theorem.
There it was also proved that the inequality saturates if and only if the distributions involved are Gaussian. The only exception to this rule is for the asymptotic values p = 1 and q = ∞ (or vice versa) where the saturation happens whenever the peak of F (x) and tail of G(y) (or vice versa) are Gaussian. The passage to quantum mechanics is quite straightforward. First, we realize that, in QM, the Fourier conjugate wave functions are related via two reciprocal relations The Plancherel (or Riesz-Fischer) equality implies that, when ||ψ F || 2 = 1, then also automatically ||ψ G || 2 = 1 (and vice versa). Thus, the connection between amplitudes ϕ F and ϕ G from (19) and amplitudes ψ F and ψ G from (23) is The factor (2πh) D/4 ensures that also ϕ F and ϕ G functions are normalized (in the sense of || . . . || 2 ) to unity; however, due to Equation (19), it might be easily omitted. The corresponding Rényi EPs change according to and hence REP-based inequalities (22) acquire in the QM setting a simple form This represents an infinite tower of mutually distinct (generally irreducible) REPURs [13]. At this point, some comments are in order. First, historically, the most popular quantifier of quantum uncertainty has been variance because it is conceptually simple and relatively easily extractable from experimental data. The variance determines the measure of uncertainty in terms of the fluctuation (or spread) around the mean value, which, while useful for many distributions, does not provide a sensible measure of uncertainty in a number of important situations including multimodal [12, 13,64] and heavy-tailed distributions [13,14,64]. To deal with this, a multitude of alternative (non-variance based) measures of uncertainty in quantum mechanics (QM) have emerged. Among these, a particularly prominent role is played by information entropies such as the Shannon entropy [63], Rényi entropy [63,64], Tsallis entropy [80], associated differential entropies, and their quantum-information generalizations [13,15,64]. REPURs (26) fit into this framework of entropic QM URs. In connection with (26), one might observe that the conventional URs based on variances-so-called Robertson-Schrödinger URs [81,82]) and Shannon differential entropy based URs (e.g., Hirschman or Białynicki-Birula URs [15,83]) naturally appear as special cases in this hierarchy. Second, the ITEs enter quantum information theory typically in three distinct ways: (a) as a measure of the quantum information content (e.g., how many qubits are needed to encode the message without loss of information), (b) as a measure of the classical information content (e.g., amount of information in bits that can be recovered from the quantum system) and (c) to quantify the entanglement of pure and mixed bipartite quantum states. Logarithms in base 2 are used because, in quantum information, one quantifies entropy in bits and qubits (rather than nats). This in turn also modifies Rényi's EP as In the following, we will employ this QM practice.

Information Distribution
To put more flesh on the concept of Rényi's EP, we devote the rest of this paper to the development of the methodology and application of Rényi EPs in extracting quantum states from incomplete data. The technique of quantum tomography is widely used for this purpose and involves making many different measurements on an ensemble of identical copies of a quantum state with a tomographically complete measurement basis [84,85]. This process is very measurement-intensive, scaling exponentially with the number of particles and so methods have been developed to approximate it with fewer measurements [86].
However, the method of Rényi EPs provides an efficient alternative approach. Instead of reconstructing the full quantum state, this process extracts the PDF of the quantum state in a given basis. For a broad class of quantum metrology problems, local rather than global approaches are preferred [70] and, for these, the local PDF of the state at each sensor is needed rather than the full density matrix. With this in mind, we first start with the notion of the information distribution.
Let F (x) be the PDF for the random variable X . We define the information random variable i X (X ) so that i X (x) = log 2 1/F (x). In other words, i X (x) represents the information in x with respect to F (x). In this connection, it is expedient to introduce the cumulative distribution function for i X (X ) as The function ℘(y) thus represents the probability that the random variable i X (X ) is less than or equal to y. We have denoted the corresponding probability measure as d℘(i X ).
Taking the Laplace transform of both sides of (28), we get where E[· · · ] denotes the mean value with respect to F . By assuming that ℘(x) is smooth, then the PDF associated with i X (X )-the so-called information PDF-is Setting s = (p − 1) log 2, we have The mean here is taken with respect to the PDF g. Equation (31) can also be written explicitly as Note that, when F p is integrable for p ∈ [1, 2], then (32) ensures that the moment-generating function for g(x) PDF exists. Thus, in particular, the moment-generating function exists when F (x) represents Lévy α-stable distributions, including the heavy-tailed stable distributions (i.e, PDFs with the Lévy stability parameter α ∈ (0, 2]). The same holds forF and p ∈ [2, ∞) due to the Beckner-Babenko theorem [13,87,88].

Reconstruction Theorem
Since L{g}(s) is the moment-generating function of the random variable i X (X ), one can generate all moments of the PDF g(x) (if they exist) by taking the derivatives of L{g} with respect to s. From a conceptual standpoint, it is often more useful to work with cumulants rather than moments. Using the fact that the cumulant generating function is simply the (natural) logarithm of the moment-generating function, we see from (32) that the differential RE is a reparametrized version of the cumulant generating function of the information random variable i X (X ). In fact, from (31), we have To understand the meaning of REPURs, we begin with the cumulant expansion (33), i.e., where κ n (X ) ≡ κ n (i X ) denotes the n-th cumulant of the information random variable i X (X ) (in units of bits n ). We note that i.e., they represent the Shannon entropy and varentropy, respectively. By employing the identity we can rewrite (34) in the form From (37), one can see that κ n (X ) = nD 2 (log 2 e) n−1 d n−1 log 2 N 1−p (X ) dp n−1 p=0 where δ 1n is the Kronecker delta function that has a value of one if n = 1, or zero otherwise. In terms of the Grünwald-Letnikov derivative formula (GLDF) [89], we can rewrite (38) as Thus, in order to determine the first m cumulants of i X (X ), we need to know all N 1 , N 1+∆ , . . . , N 1+(m−1)∆ entropy powers. In practice, ∆ corresponds to a characteristic resolution scale for the entropy index which will be chosen appropriately for the task at hand, but is typically of the order 10 −2 . Note that the last term in (38) and (39) can be also written with Y being the random variable distributed with respect to the Gaussian distribution Z 1I G with the unit covariance matrix. When all the cumulants exist, then the problem of recovering the underlying PDF for i X (X ) is equivalent to the Stieltjes moment problem [90]. Using this connection, there are a number of ways to proceed; the PDF in question can be reconstructed e.g., in terms of sums involving orthogonal polynomials (e.g., the Gram-Charlier A series or the Edgeworth series [91]), the inverse Mellin transform [92], or via various maximum entropy techniques [93]. Pertaining to this, the theorem of Marcinkiewicz [94] implies that there are no PDFs for which κ m = κ m+1 = . . . = 0 for m ≥ 3. In other words, the cumulant generating function cannot be a finite-order polynomial of degree greater than 2. The important exceptions, and indeed the only exceptions to Marcinkiewicz's theorem are the Gaussian PDFs that can have the first two cumulants nontrivial and κ 3 = κ 4 = . . . = 0. Thus, apart from the special case of Gaussian PDFs where only N 1 and N 1+∆ are needed, one needs to work with as many entropy powers N 1+k∆ , k ∈ N (or ensuing REPURs) as possible to receive as much information as possible about the structure of the underlying PDF. In theory, the whole infinite tower of REPURs would be required to uniquely specify a system's information PDF. Note that, for Gaussian information PDFs, one needs only N 1 and N 1+∆ to reconstruct the PDF uniquely. From (37) and (39), we see that knowledge of N 1 corresponds to κ 1 (X ) = H(X ) while N 1+∆ further determines κ 2 , i.e., the varentropy. Since N 1 is involved (via (39)) in the determination of all cumulants, it is the most important entropy power in the tower. Thus, the entropy powers of a given process have an equivalent meaning to the PDF: they describe the morphology of uncertainty of the observed phenomenon.
We should stress that the focus of the reconstruction theorem we present is on cumulants κ n which can be directly used for a shape estimation of g(x) but not F (x). However, by knowing g(y), we have a complete "information scan" of F (x). Such an information scan is, however, not unique, indeed, two PDFs that are rearrangements of each other-i.e., equimeasurable PDFs, have identical ℘(y) and g(y). Even though equimeasurable PDFs cannot be distinguished via their entropy powers, they can be, as a rule, distinguished via their respective momentum-space PDFs and associated entropy powers. Thus, the information scan has a tomographic flavor to it. From the multi-peak structure of g(y), one can determine the number and height of the stationary points. These are invariant characteristics of a given family of equimeasurable PDFs. This will be further illustrated in Section 6.

Information Scan of Quantum-State PDF
With knowledge of the entropy powers, the question now is how we can reconstruct the information distribution g(x). The inner workings of this will now be explicitly illustrated with the (generalized) Gram-Charlier A expansion. However, other-often more efficient methods-are also available [91]. Let κ n be cumulants obtained from entropy powers and let G(x) be some reference PDF whose cumulants are γ k . The information PDF g(x) can be then written as [91] g With hindsight, we choose the reference PDF G(x) to be a shifted gamma PDF, i.e., with a < x < ∞, β > 0, α > 0. In doing so, we have implicitly assumed that the F (y) PDF is in the first approximation equimeasurable with the Gaussian PDF. To reach a corresponding matching, we should choose a = log 2 (2πσ 2 )/2, α = 1/2 and β = log 2 e.
Using the fact that [95] ( (where L δ k is an associated Laguerre polynomial of order k with parameter δ) and given that κ 1 = γ 1 = αβ + a = log 2 (2πσ 2 e)/2, and γ k = Γ(k)αβ k = (log 2 e) k /2 for k > 1 we can write (41) as If needed, one can use a relationship between the moments and the cumulants (Faà di Bruno's formula [94]) to recast the expansion (44) into more familiar language. For the Gram-Charlier A expansion, various formal convergence criteria exist (see, e.g., [91]). In particular, the expansion for nearly Gaussian equimeasurable PDFs F (y) converges quite rapidly and the series can be truncated fairly quickly. Since in this case one needs fewer κ k 's in order to determine the information PDF g(x), only EPs in the small neighborhood of the index 1 will be needed. On the other hand, the further the F (y) is from Gaussian (e.g., heavy-tailed PDFs), the higher the orders of κ k are required to determine g(x), and hence a wider neighborhood of the index 1 will be needed for EPs.

Example-Reconstruction Theorem and (Un)Balanced Cat State
We now demonstrate an example of the reconstruction in the context of a quantum system. Specifically, we consider cat states that are often considered in the foundations of quantum physics as well as in various applications, including solid state physics [96] and quantum metrology [97]. The form of the state we consider is | ψ = N (| 0 + ν| α/ν ), where N = [1 + 2ν exp(−α 2 /2ν 2 ) + ν 2 ] −1/2 is the normalization factor, | 0 is the vacuum state, ν ∈ R a weighting factor, and | α is the coherent state given by (taking α ∈ R). For ν = 1, we refer to the state as a balanced cat state (BCS) and for ν = 1, as an unbalanced cat state (UCS). Changing the basis of | ψ to the eigenstates of the general quadrature operatorŶ whereâ andâ † are the creation and annihilation operators of the electromagnetic field, we find the PDF for the general quadrature variable y θ to be where N is the normalization constant. Setting θ = 0 and ν = 1 returns the PDF of the BCS for the position-like variable y 0 . With this, the Rényi EPs N 1−p (χ) are calculated and found to be constant across varying p. This is because F (y 0 ) for the BCS is in fact a piecewise rearrangement of a Gaussian PDF (yet has an overall non-Gaussian structure) as depicted in Figure 1, thus N 1−p (χ) = σ 2 for all p, where σ 2 is the variance of the 'would be Gaussian'.
Taking the reference PDF to be G(x) = G(x|a, α, β), with a = log 2 (2πσ 2 )/2, α = 1/2 and β = log 2 (e), it is evident that (κ k − γ k ) = 0 for all k ≥ 1, and from the Gram-Charlier A series (41), a perfect matching in the reconstruction is achieved. Furthermore, it can be shown that the variance of (47) increases with α, i.e., the variance increases as the peaks of the PDF diverge, which is in stark contrast to the Rényi EPs which remain constant for increasing α. This reveals the shortcomings of variance as a measure of uncertainty for non-Gaussian PDFs. The peaks, located at F (y θ ) = 2 −a + j , where j is an index labelling the distinct peaks, give rise to sharp singularities in the target g(x). With regard to the BCS position PDF, distributions of the conjugate parameter F (y π/2 ) distinguish F (y 0 ) from its equimeasurable Gaussian PDF and hence the Rényi EPs also distinguish the different cases. The number of available cumulants k is computationally limited, but, as this grows, information about the singularities will be recovered in the reconstruction. In the following, we show how the tail convergence and location of a singularity for g(x) can be reconstructed using k = 5. We consider the case of a UCS with ν = 0.97, α = 10 and we take θ = 0 in Equation (47) to find the PDF in the y 0 quadrature which is non-Gaussian for all piecewise rearrangements.
As such, all REPs N 1−p vary with p and consequently all cumulants κ k carry information on g(x). Here, we choose to reconstruct the UCS information distribution by means of the Edgeworth series [91] so that where the reference PDF G(x) is again the shifted gamma distribution. Using the Edgeworth series, the information PDF is approximated by expanding in orders of n, which has the advantage over the Gram-Charlier A expansion discussed above of bounding the errors of the approximation. For the particular UCS of interest, expanding to order n −3/2 reveals convergence toward the analytic form of the information PDF shown as the target line in Figure 2. This shows that, for a given characteristic resolution, control over the first five Rényi EPs can be enough for a useful information scan of a quantum state with an underlying non-Gaussian PDF. In the example shown in Figure 2, we see that the information scan accurately predicts the tail behavior as well as the location of the singularity, which corresponds to the second (lower) peak of F (y 0 ).  The Edgeworth expansion has been used here to order n −3/2 requiring control of the first five REPs. Good convergence of the tail behavior is evident as well as the location of the singularity corresponding to the second peak; a + 2 corresponds to the value of x at the point of intersection with the second (lower) peak of F (y 0 ).

Entropy Powers Based on Tsallis Entropy
Let us now briefly comment on the entropy powers associated with yet another important differential entropy, namely Tsallis differential entropy, which is defined as [47] where, as before, the PDF F (x) is associated with a random vector {X i } in R D .
Similarly to the RE case, the Tsallis entropy power N T p (X ) is defined as the solution of the equation The ensuing entropy power has not been studied in the literature yet, but it can be easily derived by observing that the following scaling property for differential Tsallis entropy holds, namely where a ∈ R and the q-deformed sum and logarithm are defined as [11]: x ⊕ q y = x + y + (1 − q)xy and ln q x = (x 1−q − 1)/(1 − q), respectively. Relation (51) results from the following chain of identities: We can further use the simple fact that Here, q and q satisfy 1/q + 1/q = 1 with q ∈ R + . By combining (50), (51), and (53), we get where we have used the sum rule from the q-deformed calculus: ln q x ⊕ q ln q y = ln q xy. Equation (54) can be resolved for N T p by employing the q-exponential, i.e., e x q = [1 + (1 − q)x] 1/(1−q) , which (among others) satisfies the relation e ln q x q = ln q (e x q ) = x. With this, we have In addition, when q → 1 + , one has where N(X ) is the conventional Shannon entropy power and H(X ) is the Shannon entropy [5].
In connection with Tsallis EP, we might notice one interesting fact, namely by starting from Rényi's EP (considering RE in nats), we can write Here, we have used a simple identity Thus, we have obtained that Rényi and Tsallis EPs coincide with each other. In particular, Rényi's EPI (22) can be equivalently written in the form Similarly, we could rephrase the generalized Stam inequality (21) and generalized isoperimetric inequality (14) in terms of Tsallis EPs. Though such inequalities are quite interesting from a mathematical point of view, it is not yet clear how they could be practically utilized in the estimation theory as there is no obvious operational meaning associated with Tsallis entropy (e.g., there is no coding theorem for Tsallis entropy). On the other hand, Tsallis entropy is an important concept in the description of entanglement [98]. For instance, Tsallis entropy of order 2 (also known as linear entropy) directly quantifies state purity [63].

Conclusions
In the first part of this paper, we have introduced the notion of Rényi's EP. With quantum metrology applications in mind, we carried out our discussion in the framework of estimation theory. In doing so, we have generalized the notion of Fisher information (FI) by using a Rényi entropy version of De Bruijn's identity. The key role of the escort distribution in this context was highlighted. With Rényi's EP at hand, we proved the RE-based isoperimetric and Stam inequalities. We have further clarified the role of Rényi's EP by deriving (through the generalized Stam inequality) a one-parameter family of Rényi EP-based quantum mechanical uncertainty relations. Conventional variance-based URs of Robertson-Schrödinger and Shannon differential entropy-based URs of Hirschman or Białynicki-Birula naturally appear as special cases in this hierarchy of URs. Interestingly, we found that the Tsallis entropy-based EP coincided with Rényi's EP provided that the order is the same. This might open quite a new, hitherto unknown role for Tsallis entropy in estimation theory.
The second part of the paper was devoted to developing the application of Rényi's EP for extracting quantum states from incomplete data. This is of particular interest in various quantum metrology protocols. To that end, we introduced the concepts of information distribution and showed how cumulants of the information distribution can be obtained from knowledge of EPs of various orders. With cumulants thus obtained, one can reconstruct the underlying information distribution in a process which we call an information scan. A numerical implementation of this reconstruction procedure was technically realized via Gram-Charlier A and Edgeworth expansion. For an explicit illustration of the information scan, we used the non-Gaussian quantum states-(un)balanced cat states. In this case, it was found that control of the first five significant Rényi EPs gave enough information for a meaningful reconstruction of the information PDF and brought about non-trivial information on the original balanced cat state PDF, such as asymptotic tail behavior or the heights of the peaks.
Finally, let us stress one more point. Rényi EP-based quantum mechanical uncertainty relations (26) basically represent a one-parameter class of inequalities that constrain higherorder cumulants of state distributions for conjugate observables [13]. In connection with this, the following two questions are in order. Assuming one is able to control Rényi EPs of various orders: (i) how do such Rényi EPs constrain the underlying state distribution and (ii) how do the ensuing REPURs restrict the state distributions of conjugate observables? The first question was tackled in this paper in terms of the information distribution and reconstruction theorem. The second question is more intriguing and has not yet been properly addressed. Work along these lines is presently under investigation.

Conflicts of Interest:
The authors declare no conflict of interest.

Abbreviations
The following abbreviations are used in this manuscript:

Appendix A
Here, we provide an intuitive proof of the generalized De Bruijn identity. Generalized De Bruijn identity I: By denoting the PDF associated with a random vector {X i } as F (x) and the noise PDF as G(z), we might write the LHS of (7) as It should be noted that our manipulations make sense only for any q > 0, as only in these cases are RE and escort distributions well defined. The right-hand-side of (A1) can also be equivalently written as where the mean E q {. . .} is performed with respect to the escort distribution ρ q , while E with respect to G distribution. We note in passing that the conventional De Bruijn's identity (6) emerges as a special case when q → 1. For the Gaussian noise vector, we can generalize the previous derivation in the following way: Generalized De Bruijn's identity II: Let {X i } be a random vector in R D with the PDF F (x) and let {Z i } be an independent Gaussian noise vector with the zero mean and covariance matrix Σ = cov(Z G ), then, The right-hand-side is equivalent to To prove the identity (A3), we might follow the same line of reasoning as in (A1). The only difference is that, while in (A1) we had a small parameter in which one could expand to all orders of correlation functions and easily perform differentiation and limit → 0 for any noise distribution (with zero mean), the same procedure can not be done in the present context for a generic noise distribution. In fact, only the Gaussian distribution has the property that the higher-order correlation functions and their derivatives with respect to Σ ij are small when Σ is small. The latter is a consequence of the Marcinkiewicz theorem [99].

Appendix B
Here, we prove the Generalized isoperimetric inequality from Section 2. The starting point is the entropy-power inequality (EPI) [64]: Let X 1 and X 2 be two independent continuous random vectors in R D with probability densities F (1) ∈ q (R D ) and F (2) ∈ p (R D ), respectively. Suppose further that λ ∈ (0, 1) and r > 1, and let then the following inequality holds: Let us now consider a Gaussian noise vector Z G (independent of X ) with zero mean and covariance matrix Σ. Within this setting, we can write the following EPIs: with ≡ det(Σ) 1/D . Here, we have used the simple fact that N r (Z G ) = det(Σ) 1/D , irrespective of the value of r.
Let us now fix r and maximize the RHS of inequality (A7) with respect to λ and q provided we keep the constraint condition (A5). This yields the condition extremum With this, q turns out to be which in the limit → 0 reduces to q = r ≥ 1. The latter implies that p = 1. The result (A10) implies that the RHS of (A7) reads Should we have started with the p index, we would arrive at an analogous conclusion. To proceed, we stick, without loss of generality, to the inequality (A7). This implies that To proceed, we employ the identity log N r (X ) representing a Gaussian random vector with zero mean and unit covariance matrix, and the fact that I r is monotonously decreasing function of r, i.e., dI r /dr ≤ 0 (see, e.g., Ref. [78]).
With this, we have Consequently, Equation (A12) can be rewritten as At this stage, we are interested in the Σ ij → 0 limit. In order to find the ensuing leading order behavior of /Σ ij , we can use L'Hospital's rule, namely Now, we neglect the sub-leading term of order O( 2 /Σ ij ) in (A14) and take det(. . .) 1/D on both sides. This gives det dN r (X + Z G ) dΣ ij or equivalently N r (X )[det(J r (X ))] 1/D ≥ 1 .
At this stage, we can use the inequality of arithmetic and geometric means to write (note that J r = cov r (V r ) is a positive semi-definite matrix) 1 D Tr(J r (X )) ≥ [det(J r (X ))] 1/D .
Here, 2r and 2q are Hölder conjugates so that r ∈ [1, ∞]. The inequality employed is due to the Hausdorff-Young inequality (which in turn is a simple consequence of the Hölder inequality [64]). We further have where a ∈ R D is an arbitrary x-independent vector, ∇ i ≡ ∂/∂x i and V D denotes a regularized volume of R D -D-dimensional ball of a very large (but finite) radius R. In the first line of (A22), we have employed the triangle inequality |E r e ia·x | ≤ 1 (with equality if and only if a = 0), namely The inequality in the last line holds for a i = π/(2R) (for all i), since, in this case, cos(a · x) ≥ 0 for all x from the D-dimensional ball. In this case, one may further estimate the integral from below by neglecting the positive integrand (r − 1)ρ r (x)[V i (x)] 2 . Note that (A22) implies with equality if and only if a → 0 (to see this, one should apply L'Hospital's rule). Equation (A24) allows for writing where γ > 0 is some as yet unspecified constant and C D = π D/2 /Γ(D/2 + 1). In deriving (A25), we have used the Hölder inequality Here, and also in (A22) and (A25), V D = C D R D denotes the regularizated volume of R D . As already mentioned, the best estimate of the inequality (A25) is obtained for a → 0. As we have seen, a i goes to zero as π/(2R) which allows for choosing the constant γ so that the denominator in (A25) stays finite in the limit R → ∞. This implies that γ = D/2 − D/(2r). Consequently, (A25) acquires in the large R limit the form ||F || r ≥ With this, we can write [see Equations (A20)-(A21)] where, in the last inequality, we have used the fact that q 1/(1−q) ≥ 1/4 for q ∈ [1/2, 1] and that [Γ(D/2 + 1)] 2/D ≥ π/4. As a final step, we employ Equations (A18) and (A28) to write N q (Y ) ≥ 1 16π 2 D Tr(J r (X )) ≥ 1 16π 2 [det(J r (X ))] 1/D , which completes the proof of the generalized Stam's inequality.