Equivalence of Partition Functions Leads to Classification of

We derive a two-parameter family of generalized entropies, Spq, and means mpq. To this end, assume that we want to calculate an entropy and a mean for n non-negative real numbers {x1,…,xn}. For comparison, we consider {m1,…,mk} where mi = m for all i = 1,…,k and where m and k are chosen such that the lp and lq norms of {x1,…,xn} and {m1,…,mk} coincide. We formally allow k to be real. Then, we define k, log k, and m to be a generalized cardinality kpq, a generalized entropy Spq, and a generalized mean mpq respectively. We show that this family of entropies includes the Shannon and Renyi entropies and that the family of generalized means includes the power means (such as arithmetic, harmonic, geometric, root-mean-square, maximum, and minimum) as well as novel means of Shannon-like and Renyi-like forms. A thermodynamic interpretation arises from the fact that the lp norm is closely related to the partition function at inverse temperature β = p. Namely, two systems possess the same generalized entropy and generalized mean energy if and only if their partition functions agree at two temperatures, which is also equivalent to the condition that their Helmholtz free energies agree at these two temperatures.


Introduction
Two of the most basic concepts of thermodynamics are: (a) the average of measurement outcomes and (b) the uncertainty or entropy about measurement outcomes.Consider, for example, a physical system, A, that is in contact with a heat bath at some fixed temperature, i.e., a canonical ensemble.The measurement of the system's energy can return any one of its energy eigenstates.What then is (a) the mean energy to expect and (b) how uncertain is the prediction of the measured energy state?
We notice that, in principle, many different notions of energy mean and many different measures of entropy could be employed here.Of course, in thermodynamics, the Boltzmann factor weighted mean as well as the Shannon/von Neumann entropy are of foremost importance.In this paper, we show that also other important notions of average such as the harmonic mean, the geometric mean and the arithmetic mean arise naturally, along with generalized notions of entropy including Rényi entropies [1], all unified in a two-parameter family of notions of means and notions of entropies.
To this end, consider systems (canonical ensembles) in a heat bath.We begin by considering the simplest kind of system, namely the type of system which possesses only one energy level, E. Let us denote its degeneracy by k.Unambiguously, we should assign that system the mean : m E  and the entropy : log S k  . Let us denote these simple one-level systems by the term reference system.Now, let X be a system with arbitrary discrete energy levels.Our aim is to assign X a mean and an entropy by finding that reference system M which is in some sense equivalent to X. Then we assign X the same value for the mean and entropy as the reference system M.
But how do we decide if a reference system is in some sense equivalent to system X?Given that we want the reference system M and system X to share two properties, namely a mean and an entropy, we expect any such condition for the equivalence of two systems to require two equations to be fulfilled.Further, since the properties of systems are encoded in their partition function ( ) Z X , we expect that these two equations can be expressed in terms of the partition functions of the two systems in question.
To this end, let us adopt what may be considered the simplest definition.We choose two temperatures, T 1 and T 2 and we define that a reference system is (T 1 , T 2 )-equivalent to system X if the partition functions of the two systems coincide with each other at these two temperatures.Since the Helmholtz free energy H A obeys log( ) , where B K is the Boltzmann constant, this is the same as saying that two systems are put in the same equivalence class if their Helmholtz free energies coincide at these two temperatures.This allows us now to assign any system X a mean and an entropy.We simply find its unique (T 1 and T 2 )-equivalent reference system M. Then the mean and entropy of X are defined to be the mean and the entropy of the reference system M.
Clearly, the so-defined mean and entropy of a system X now actually depend on two temperatures, namely (T 1 , T 2 ).As we will show below, in the limit when we let the two temperatures become the same temperature, we recover the usual Boltzmann factor-weighted mean, i.e., the usual mean energy, along with the usual Shannon/von Neumann entropy.
For general (T 1 , T 2 ), we cover more, however.Namely, we naturally obtain a unifying 2-parameter family of notions of mean that includes for example the geometric, the harmonic, the arithmetic and the root-mean-square (RMS) means.And we obtain a unifying 2-parameter family of notions of entropy that, for example, includes the Rényi family of entropies.
To be precise, let us assume that a system X has only discrete energy levels, {E i }, where i enumerates all the energy levels counting also possible degeneracies.Notice that {E i } is formally what is called a multiset, because its members are allowed to occur more than once.Similarly, let us also collect the exponentiation of the negative energies : exp( ) x   in the multiset { } i x .Either multiset can be used to describe the same thermodynamic system X.Let : denote the inverse temperature, where B K is the Boltzmann constant.The partition function of system X, i.e., the sum of its Boltzmann factors, then reads: For later reference, note that the partition function is therefore related to the l p norm of Now the key definition is that we call two physical systems 1 2 ( , )   -equivalent if their partition functions coincide at the two inverse temperatures 1 2 ( , )   , i.e., systems X and M are 1 2 ( , ) . To be more explicit, one may also call such systems 1 2 ( , )   -partition function equivalent, or also 1 2 ( , )   -Helmholtz free energy equivalent, but we will here use the term 1 2 ( , )   -equivalent for short.
In particular, for any given system X, let us consider the 1 2 ( , )   -equivalent reference system M which possesses just one energy level, with energy E 0 and degeneracy k, where we formally allow k to be any positive number.E 0 and k are then determined by the two conditions that the partition function of M is to coincide with that of X at the two inverse temperatures 1  and 2  .Then, we define  to be the generalized entropy, and   to be the generalized mean energy of system X with respect to the temperatures 1 2 ( , )   .
We will explore the properties of these families of generalized entropies and means in the subsequent sections.First, however, let us consider the special limiting case when the two temperatures coincide (i.e., 1 2      ).As will be detailed in the subsequent sections of the manuscript, in this limiting case, the two equivalence conditions of partition functions can be shown to reduce to: which can be shown to be equivalent to: The conditions (3a) and (3b) physically mean that systems X and M have the same partition function and average energy, respectively, at the inverse temperature  .Notice that this is also the same as saying that the two systems have the same average energy and the same Helmholtz free energy at the inverse temperature  .Now by employing either pair of conditions, (2) or (3), we then recover indeed the usual thermodynamic entropy of the system X, which is given in the Shannon form at the inverse temperature  by ( ) The proofs of Equations ( 2)-( 4) are straightforward.We will spell them out in detail in the subsequent sections where the setting is abstract mathematical.
Before we begin the mathematical treatment, let us remark that entropy is not only a cornerstone of thermodynamics but it is also crucial in information theory.Due to its universal significance, measures of uncertainty in the form of an entropy have been proposed by physicists and mathematicians for over a century [2].Our approach here for deriving a generalized family of entropies was originally motivated by basic questions regarding the effective dimensionality of multi-antenna systems (e.g., [3][4][5]).After initial attempts in [6] and later in [5,7], we here give for the first time a comprehensive derivation with the proofs and we also include the family of generalized means.
The manuscript is organized as follows.In Section 2, we introduce the proposed family of entropies and means mathematically, and also show some special cases thereof.The axiomatic formulation is presented in Section 3, followed by the study of resulting properties in Section 4. Proofs are provided in the appendices.

Let
be a multiset of real non-negative numbers, where X K denotes the cardinality of X.We assume that X possesses at least one non-zero element.Further, let , p q be arbitrary fixed real numbers obeying 0 pq  .Let { : , 0, 1, 2, , } be a reference multiset possessing exactly one real positive element m , which is of multiplicity 1 k  .We introduce the following definitions: . To simplify notation, the subscript X is omitted when dealing with one multiset at hand.
is the multiplicity of the maximum elements of X  min X K is the multiplicity of the minimum elements of X Our objective is to determine suitable values for m and k , possibly non-integer, that can serve as mean and effective cardinality of X, respectively, namely by imposing a suitable criterion for the equivalence of X to a reference multiset M. Having two unknowns ( m and k ) in M, we need two equivalence conditions.We choose to impose the equivalence of the p-norms and the q-norms: Here, the p-norm p  is defined as usual through: with the proviso that ( ) p i x is replaced by 0 if 0 i x  and 0 p  .We remark that, for 1 p  , ( 6) is merely a quasi-norm since the triangle inequality does not hold.Note the singularity Solving for k and m in (5), we obtain: We call ( ) pq k X and ( ) pq m X the norm-induced effective cardinality and generic mean of order , p q for the multiset X, respectively.Let us now express (7) in a logarithmic form and define the entropy ( ) pq S X as follows: where the last step is obtained by straightforward manipulations.Similarly, we find for ( ) qq m X : pq qq q p q p q qq p q q q q q q q q q K i i where in the second to last step, we used the fact that ( ) ( ) log We remark that in the early phase of this work [5], each author independently suggested either ( ) pq k X or ( ) qq k X as two possible distinct notions for the effective cardinality.In [7], it was reported that the average energy and Shannon entropy of a thermodynamic system are obtained by starting from equivalence of partition functions of two systems at two temperatures when the two temperatures coincide as mentioned in the introduction.Clearly, the limiting operation in (9) makes the connection and establishes (7) as the general definition of this norm-induced family of entropies and means.In fact, for the case of degenerate order, ( p q  ), the quantities ( ) qq k X , ( ) qq S X , and ( ) qq m X , could have been obtained as well through a differential equivalence of the q -norm.To see this, we impose the following two conditions: After employing and solving for k and m , we ultimately obtain ( ) qq k X and ( ) qq m X as given by ( 9) and (10).The condition (12) is the mathematical equivalent of the aforementioned physical condition (2) imposed on the two thermodynamic systems, which yielded the Shannon entropy form (4).
From (9), it is obvious that ( ) qq S X is the Shannon entropy of the distribution   . On the other hand, ( ) pq S X is a more general expression of the Rényi entropy of order  .For a probability distribution { } i P p  , the Rényi entropy of order  is given by [9]: By setting the order 1 p  in pq S , we obtain from (8): By comparing ( 13) and ( 14), we readily identify 1 ( ) q S X as the Rényi entropy of order q for a complete statistical distribution given by , where the multiset elements add to 1. Formally: In the degenerate case (when 1 q  ), 11 ( ) S X is the Shannon entropy of the latter distribution.For 1 p  , ( ) pq S X from ( 8) can be rearranged as a generalization of ( 13): which can be viewed as the Rényi entropy of order q p for the th p order escort distribution Rényi defined his entropy for , 0 p q  .We relax this condition further and allow ( ) pq S X and ( ) pq m X to be defined for any real indices , p q such that 0 pq  .Accordingly, we obtain the following properties: , When at least one order , p q is zero, we find the interesting results: We recognize 0 ( ) q S X in (19) as the Hartley entropy [10], which will be shown later to be the maximum value of any entropy.From (20), we obtain a famous family of generic p-means of the non-zero elements of X: particularly 0,  are the minimum, harmonic mean, arithmetic mean, root-mean-square mean, and maximum, respectively.In the limiting case 0 q  , we obtain 0,0 m , which is the geometric mean.In Table 1, we summarize these and other particular cases of means and entropies at specific , p q .The key point is that each , p q uniquely defines an entropy pq S with a corresponding mean pq m , such that each pair of pq S and pq m is coupled in this sense.
A typical plot for p X , pp S and pp m is shown in log scale in Figure 1, illustrating some of the properties to be discussed hereafter.In particular, we notice: is non-decreasing/non-increasing for negative/positive p , respectively, and is guaranteed to be maximized at 0 p  .This is discussed more generally in Property 4.6. pp m ranges from min( ) X to max( ) X and is always non-decreasing with respect to p .This is discussed more generally in Property 4.

An Axiomatic Approach to the Generalized Entropies and Means
In order to simplify the conceptual underpinnings, let us now describe the generalized entropies pq S and means pq m , through two simple axioms.

Axioms for the Generalized Entropy
Let , p q be fixed real numbers obeying 0 pq  .Consider a map, pq S , which maps multisets of positive real numbers into the real numbers.We call pq S a generalized entropy of order , p q if it obeys the following two axioms: Entropy Axiom 1: (where the base of the logarithm is arbitrarily chosen), i.e., Entropy Axiom 2: If p q  , the map pq S depends only on the ratio of the multiset's p and q norms, i.e., pq S is some function, pq f , of this ratio: ( ): If p q  , the map qq S depends only on the ratio of the multiset's q norm to its derivative, i.e., pq S is some function, q f , of this ratio: ( ): To see that (22b) arises in the limit from (22a) we notice that, since the logarithm is strictly monotone, (22a) is equivalent to saying that pq S is some function pq h of some finite number times the logarithm of the ratio of norms: Choosing p q    and taking the limit 0   we obtain (22b) with q q q f h  .

Proof:
Entropy Axiom 2 implies that for any two multisets X and Y : ( ) ( ) Choosing forY the uniform multiset M of Axiom 1, and taking the logarithm of both sides yields: , we now uniquely obtain the formulas for ( ) pq S X and ( ) qq S X given in ( 8) and ( 9), respectively.We note that the functions pq f and q f are therefore:

Remarks:
 Even though k is treated as an integer representing the multiplicity in Axiom 1, this condition is tacitly relaxed in Axiom 2 to include non-integer values, which we may call the effective cardinality (or effective dimensionality) of order , p q . The logarithmic measure of Axiom 1 is directly connected to the celebrated Boltzmann entropy formula ( ) log , where B K is the Boltzmann constant and W is the number of the microstates in the system.The logarithmic measure is also connected to the so-called "Hartley's measure" [11,12], which indicates the non-specificity [13] and does not require a probability distribution assumption.In Axiom 1, a multiset of equal positive numbers is all that is required.In fact, Axiom 1 encompasses the additivity and monotonicity axioms [12,13], which are equivalent to the earlier Khinchin's axioms of additivity, maximality and expansibility [14]. In Axiom 2, note that the p-norm definition is relaxed to include the values 1 p  , which would result in the triangle inequality to be violated should the multisets be treated as vectors.

Axioms for the Generalized Mean
We define the p th moment of the multiset { } i X x  as: ( ): The nomenclature "p th moment" is motivated by the fact that for the density function ( ) : ( ) , where ( ) x  is the Dirac delta function, the p th moment is indeed: p q be fixed real numbers obeying 0 pq  .Consider a map, pq m , which maps multisets of positive real numbers into the real numbers.We call pq m a generalized mean of order , p q if it obeys the following two axioms: Mean Axiom 1: Mean Axiom 2: If p q  , the map pq m depends for any multiset X only on the ratio of the multiset's p th and q th moments, || || p p X and || || q q X , i.e., pq m is some function, pq g , of their ratio: ( ): If p q  , the map qq m is a function only of the ratio: ( ): The fact that (30b) is the limit of (30a) follows by the same reasoning as in (23).
Proof: Axiom 2 implies for any two multisets X and Y that: Choosing the multiset Y to be the uniform multiset M from (29) we obtain 0 0 log p p p q q q q q q q X m p q X We can now use that qq m X as given in ( 7) and (10), respectively.Accordingly the functions pq g and g are found to be We have obtained axiomatizations of the generalized entropies and means which revealed, in particular, that the generalized entropies can be characterized as those entropies that cover the reference multiset case (the multiset of equal elements) and that are functions of only the ratio of the multisets' p l and q l norms.Similarly, the axiomatization also revealed that the generalized means can be characterized as those means which cover the reference multiset case and which are functions of only the ratio of the multisets' p th and q th moments || || p p X and || || q q X .We will now develop an axiomatization that links up with Section 2, yielding simultaneously a unique family of generalized entropies and means.

Unifying Axioms for Generalized Entropies and Means
We notice that, as is straightforward to verify: This means that we can describe the generalized entropies and means also through a unifying set of axioms.To this end, let , p q be fixed real numbers obeying 0 pq  .Consider maps pq S and pq m , which map multisets of positive real numbers into the real numbers.We call pq S and pq m generalized entropies and means of order , p q respectively, if they obey the following two axioms: Proposition: The maps pq S and pq m are unique and given by Equations ( 7)- (10).
Proof: The proofs are straightforward and proceed similarly to the proofs of the propositions related to the entropy and mean axioms.

Properties of pq S and pq m
We list in this section useful properties of pq m , pq S and pq k with proofs in the appendix.The definitions of pq m , pq S and pq k are given by ( 7)- (10).We also add two plots for an example multiset {10;9;8;7;6;0.5;0.4;0.3;0.2;0.1}X  in Figures 2 and 3 in order to provide some numerical illustration of the properties hereunder.  , we have: That is the entropy is invariant of the scaling, whereas the mean varies linearly with scaling.The proof is straightforward, based on (7).

Symmetry with Respect to the Elements of X ( )
pq S X does not depend on the order of the elements of X .

Symmetry with Respect to the Order , p q
By exchanging p and q in (8), we readily find that:

Sign Change of the Order , p q
From (17), for 0 pq  , we obtain: Let 1 X N  be the number of the non-zero elements in X .Therefore: The minimum value occurs when X has exactly one non-zero element.For , 0 p q  , the maximum value occurs when all the non-zero elements of X are equal.When either p or q is zero, ( ) pq S X yields the maximum value, log X N , for any distribution of X , which is physically intuitive since the zero th order renders all the non-zero multiset elements to an equal value and thus we reach the equiprobable case leading to maximum entropy.The proof of this property is in Appendix A.3.

Monotonicity of ( )
pq S X and ( ) pq m X with respect to , p q We have the following results for the monotonicity of ( ) pq S X and ( ) pq m X with respect to , p q .
The proofs are in Appendix B and Appendix C, respectively: where equality holds when all the non-zero elements i x are equal.Accordingly, for p q  and 0 pq  , by fixing one order (say q ), pq m is always non-decreasing with respect to the other order ( p ); whereas pq S is non-decreasing for 0 p  , non-increasing for 0 p  , with a maximum value at 0 p  .
The result is true when switching , p q from the symmetry Property 4.3.Similarly, for the degenerate case p q  , pp m is always non-decreasing with respect to p ; whereas pp S is non-decreasing for 0 p  , non-increasing for 0 p  , with a maximum value at 0 p  .In all cases, both the mean and entropy are invariant with respect to the order , p q if and only if all the non-zero elements i x are equal.
This property explains the monotonicity of the curves in Figures 1 and 3.
, which is an intuitive range for a mean.This is also true in the degenerate case p q  , i.e., min( ) where, as usual, equality holds when all the non-zero elements i x are equal.

Additivity of the Joint Multiset Entropy
For the two probability distribution multisets we define the joint multiset 1, 2, , : : 1, 2, , . Therefore, we have: The proof is straightforward by using the fact that along with ( 7)-( 10), Property 4.8 is true for both p q  and for the degenerate case p q  .

Sub-Additivity of the Effective Cardinality Subject to the Multiset Additive Union Operation
Let  denote a multiset additive union operation [15] (page 50), e.g., {2, 2} {1, be two multisets of non-negative real numbers.Moreover, let  and  be two positive real scaling factors of the elements of X and Y , respectively.Then: where the equality holds under the following condition for the value /   : This property is a generalization of the effective alphabet size of two disjoint alphabets mixture as discussed in [16] (Problem 2.10).To see this, set 1 p q   and note that

Effective Rank of a Matrix
Let X be the multiset of the singular values of a matrix M .Then, 0 ( ) rank( ) Accordingly, for general , p q , ( ) pq k X can be viewed as a biased effective rank of M corresponding to the order , p q .From Property 4.5, we have 1 ( ) rank( ) The minimum value occurs when M has exactly one non-zero singular value (the rank of M is 1).The maximum value, rank( ) M , is reached for any 0 pq  when all the non-zero singular values are equal.The effective rank can be helpful to determine, in a well-defined , p q sense, how to view an ill-conditioned matrix, which is a full-rank from a mathematical perspective, but is effectively behaving as if possessing a lower rank.Such ill-conditioned matrices often arise in problems of oversampling or determining the degrees of freedom, where the singular values, ordered in non-increasing order by definition, exhibit some sort of "knee cut-off", similar to{ } i x in Figure 2. A biased effective rank can help to compare matrices when the knee cut-off is not sharp, thus giving more weight to the small or large singular values according to the order , p q in a consistent manner for different matrices.An example thereof is the evaluation of the degrees of freedom of some applications such as multi-antenna systems [3][4][5][6], optical imaging systems [20], or in general any case of similar limitation to space-bandwidth product [21].

Geometrical Interpretation of pq S and pq m on Log Scale, with Thermodynamics Analogy
, after taking the logarithm of both sides of (5), we obtain a simple relation between q X , pq S and pq m as in (11): Accordingly, a secant cutting the function ( ) log q q f q X  at 0 q p  and 0 q q  will have a slope and intercept of 0 0 log p q m and 0 0 p q S , respectively, as shown in Figure 4. Based on the discussion of Section 1, (46) readily yields the following analogous expressions for a thermodynamics system described by the Boltzmann factors { } i e  : In the limiting case, when , p q     , the secant in Figure 4 becomes a tangent and we get the Gibbs-Shannon entropy at the inverse temperature  .Consequently, (47) can be re-written after introducing the Boltzmann constant B K and using the absolute temperature 1/ B T K   as: 0 log : where H A is the Helmholtz free energy of the system.X versus q .
order q log |X| q q Geometrical interpretation for S pq and m pq log |X| q q secant S p0,q0 slope = log m p0,q0 p 0 q 0

Discussion and Conclusions
A two-parameter family of cardinalities, entropies and means has been derived for multisets of non-negative elements.Rather than starting from thermodynamic or information theoretic considerations to derive entropies and means, see e.g., [9,[17][18][19], we here defined the generalized entropies and means through simple abstract axioms.There are other families of entropies in the literature (e.g., [23]), which are a generalization of Shannon entropy.The generalized entropy in this manuscript is shown to preserve the additivity (Property 4.8), which is not the case with the generalized entropies based on Tsallis non-additive entropy as in [23].
Our first two axiomatizations treat the generalized entropies and means separately.It revealed that the generalized entropies are exactly those entropies that are functions of only the ratio of the multisets' p l and q l norms.It also revealed that the generalized means are exactly those means that are functions only of the ratio of the multisets' p th and q th moments, || || p p X and || || q q X .Subsequently, our unifying axiomatization characterized the generalized entropies and means together.This showed that if two multisets have exactly the same p l and q l norms, then they share the same generalized entropy and mean.
We presented several key features of the new families of generalized entropies and means, for example, that the family of generalized entropies contains and generalizes the Rényi family of entropies, of which the Shannon entropy is a special case, thus including some of the desiderata for entropies [22].We also showed the monotonicity with respect to , p q , extreme values, symmetry with respect to , p q , and additivity preservation.The effective cardinality pq k measures the distribution uniformity of the multiset elements in the sense of the pand q-norm equivalence to a reference flat multiset.From an information theory perspective, pq S and pq k represent a two-parameter entropy of order , p q and its corresponding effective alphabet size, respectively, when a probability distribution is constructed after proper normalization of the multiset elements.Furthermore, we recall that knowing the p  and q  norms of a multiset is to know the multiset's p th and q th moments.Our findings here therefore imply that knowledge of a multiset's p th and q th moments is exactly enough information to deduce the multiset's (p,q) entropy and (p,q) mean.Further, knowledge of sufficiently many moments of a multiset can be sufficient to reconstruct the multiset.Conversely, it should be interesting to examine how many (p,q)-entropies and/or (p,q)-means are required to completely determine the multiset.
Regarding the thermodynamic interpretation, we noticed that to require that the p  and q  norms of multisets coincide is mathematically equivalent to requiring that the two partition functions of two thermodynamic systems coincide at two temperatures.This in turn is equivalent to requiring that the Helmholtz free energy of the two thermodynamic systems coincide at two temperatures.The Helmholtz free energy represents the maximum mechanical work that can be extracted from a thermodynamic system under certain idealized circumstances.This suggests that there perhaps exists a thermodynamic interpretation of the generalized entropies and means in terms of the extractability of mechanical work.In this case, the fact that the generalized entropies and means depend on two rather than one temperature could be related to the fact that the maximum efficiency of a heat engine, obtained in Carnot cycles, is a function of two temperatures.We did show that in the limiting case, when the two temperatures become the same, one recovers the usual Boltzmann factor weighted mean energy as well as the usual Shannon/von Neumann entropy.
for p<q.The equality occurs only when p p i i x x   , i.e., when { }

A.2. Non-Decreasing Generalized p-Mean with Respect to p
Let X be a multiset of N strictly positive numbers.For p < q, we want to show that . To this end, we define / ( ) , where 0 x  and / 0 p q  , i.e., , 0 p q  have the same sign.Clearly, , where . Accordingly,   ( ) , the proof is complete.Equality holds when all i x are equal.We start by proving that 0 pq S  .From Lemma A.1, for p < q, we have log log p q X X  .
Therefore, log log 0 p q X X   , , 0 p q  and have the same sign.Since 1/ 1/ 0 p q   , we get . Following the same steps for p > q yields the same result.Finally, when p q  , we have log 0 . In all cases, the equality holds when { } .From Lemma A.2, assuming both , 0 p q  have the same sign, for p < q, we have . Taking the logarithm yields . For p > q, we obtain the same result.
When p q  , we have: x q p S q q p x q p x q x p x p x x q q p x q x q x x x q q p x q where we inserted another summation index for the last two terms in the last step.Therefore: where we have multiplied the last two terms in the second step by , where we have used the fact that log ( 1) x x   .The equality occurs only when for every i x , implying that each i x is either zero or some fixed value 0 x .
Accordingly, after multiplying the last inequality by 2 2 ( ) q q p  , we obtain: where the equality holds when all the non-zero elements i x are equal.Since p and q are assumed to have the same sign (from the condition 0 pq  ), we deduce that, with respect to either p or q (while fixing the other order), pq S is non-decreasing for negative order p,q, non-increasing for positive order p,q, with a maximum value at either , 0 p q  (note that 0 ). pq S is invariant to p,q if and only if all the non-zero elements i x are equal.

B.2. Degenerate Case p q 
From Appendix B. , we obtain the following two conditions: 0 0 0 0 0 0 q q p p i i i i q q q q p p p p i j i j i j i j x x x y x y 0 0 0 0 0 0 q q p p j j j j q q q q p p p p i j i j i j i j y y x y x y From ( 50) and (51), we have: 0 0 0 0 q q q q j i j i p p p p i j i j y x x y Accordingly, the required proportionality condition is: 1 0 0 q p q p j j j j q p i i i i y y x x From (49), we have: p p q q q q q i j i j pq i j q p p p p i j i j x y k x y x y q p q q q p q i i q q q q pq i j i j q i j p p i i q q q p q i i q q q q i j q i j p p p q i i q q q q q q p q p q i i i q q i j q p p j p p i p q i i i p q q p q i i q p p p i i x k x y x y x     0 0 0 0 0 0 0 q q q p q j j q q j p p j j q j p p q q p q q q p q j i j i q q p p p p p q p q i j i j y y y y x x y where we employed (52) to obtain the second to last step.Consequently, since 0  and 0  cancel out in the last step, we get: p p q p p q q q j i j i pq q q p p i j i j pq pq In order to confirm that (53) is indeed a maximizing condition of By taking the limit when p q  , (53) becomes 1/ 0 0 ( ) ( ) and ( ) ( ) ( ) 6.  1 p  has a specific property of making pp pp p m k X  .

Figure 2 . 12 Figure 3 .
Figure 2. Numerical example showing the multiset elements { } i x versus their index i ; with their corresponding mean pq m and effective cardinality pq k for different values of , p q .

Figure 4 .
Figure 4.A secant of log non-zero element.Note that there is a singularity for log

Table 1 .
Special cases of pq S and pq m .Note that pq