Abstract
We consider the concept of generalized Kolmogorov–Sinai entropy, where instead of the Shannon entropy function, we consider an arbitrary concave function defined on the unit interval, vanishing in the origin. Under mild assumptions on this function, we show that this isomorphism invariant is linearly dependent on the Kolmogorov–Sinai entropy.
1. Introduction
The Boltzmann-Gibbs entropy and the Shannon entropy are measures that lie on the foundations of statistical physics and modern information theory, respectively. The dynamical counterparts of the latter—the dynamical and the Kolmogorov-Sinai entropy—had a great impact on the modern theory of dynamical systems and mathematical physics (see, e.g., [1,2]). They were extensively studied and successfully applied, among others, to statistical physics and quantum information. It appeared to be an exceptionally powerful tool for exploring nonlinear systems. One of the biggest advantages of the Kolmogorov-Sinai entropies lies in the fact that they make it possible to distinguish the formally regular systems (those with the measure-theoretic entropy equal to zero) from the chaotic ones (with positive entropy, which implies the positivity of topological entropy [3]).
The Kolmogorov-Sinai entropy of a given transformation T acting on a probability space (X, Σ, μ) is defined as the supremum over all finite measurable partitions of the dynamical entropy of T with respect to , denoted by h(T, ). As a dynamical counterpart of Shannon entropy, the entropy of transformation T with respect to a given partition is defined as the limit of the sequence
, where:
with η being the Shannon function given by η(x) = −x log x for x > 0 with η(0) : = 0, and n is the join partition of partitions T−i for i = 0, ..., n – 1. The existence of the limit in the definition of the dynamical entropy follows from the subadditivity of η. The most common interpretation of this quantity is the average (over time and the phase space) one-step gain of information about the initial state. Taking the supremum over all finite partitions, we obtain an isomorphism invariant that measures the rate of producing randomness (chaos) by the system.
Since Shannon’s seminal paper [4], several refinements of the concept of Shannon static entropy were considered (see, e.g., Rényi [5], Arimoto [6], Wu, Verdú [7] and Csiszár’s survey article [8]). Their dynamical and measure-theoretic counterparts were also considered by a few authors. De Paly [9] proposed generalized dynamical entropies based on the concept of the relative static entropies. Unfortunately, it appeared that, despite some special cases [9,10], the explicit calculation of this invariant may not be possible. Grassberger and Procaccia proposed in [11] a dynamical counterpart of the well-known generalization of Shannon entropy—the Rényi entropy—and its measure-theoretic counterpart were considered by Takens and Verbitski. They showed that for ergodic transformations with positive measure-theoretic entropy, Rényi entropies of a measure-theoretic transformation are either infinite or equal to the measure-theoretic entropy [12]. The answer for non-ergodic aperiodic transformations is different: for Rényi entropies of order α > 1, they are equal to the essential infimum of measure-theoretic entropies of measures forming the decomposition of a given measure into ergodic components, while for α < 1, they are still infinite [13]. In particular, this means that Rényi entropies of order α < 1 are metric invariants sensitive to ergodicity. A similar generalization was made by Mesón and Vericat [14–16] for so-called Havrda–Charvát–Tsallis entropy [17], and their results were similar to the ones obtained by Takens and Verbitski in [12].
In this paper, we present a generalization of the concept of dynamical entropy. This was made for few reasons. First of all, the entropy used in the theory of dynamical systems is a natural isomorphism invariant. However, outside of the class of Bernoulli automorphisms, it is not a complete invariant, i.e., two systems with the same entropy need not be isomorphic. In particular, it is not applicable to a wide class of systems with zero entropy. Moreover, we can ask whether considering of a function different than the Shannon function (at the expense of losing its interpretations from the information theory) can give a significantly new invariant. This can show which properties of the Shannon function are crucial for the usage of entropy in dynamical systems. Finally, different dynamical generalizations of entropy have been used by the physicists for years [18–22]. However, up till now, there have not been many rigorous results. We hope that this note will help to fill in this gap.
Our approach is based on the generalization suggested by Rényi and extended by Arimoto applied to the dynamical case. Instead of the Shannon function η, we consider a concave function g :[0, 1] ↦ ℝ, such that
and define the dynamical g-entropy of the finite partition as:
The behavior of the quotient g(x)/η(x) as x converges to zero appears to be crucial for our considerations. Defining:
we will prove that:
In the case of Ci(g) = ∞, we will show that in every aperiodic system and for every γ ≥ 0, there exists a finite partition , such that h(g, T, ) ≥ γ.
Considering the supremum over all partitions, we obtain a Kolmogorov entropy-like isomorphism invariant, which we will call the measure-theoretic g-entropy of a transformation with respect to an invariant measure. One might ask whether this invariant may give any new information about the system. We will prove (Theorem 5) that for g with Cs(g) < ∞, this new invariant is linearly dependent on Kolmogorov–Sinai entropy. It shows that Shannon entropy function is the special one from the point of view of the theory of dynamical entropies—it is the most natural one—not only does it have all of the properties that the entropy function should have [1], but also considering different entropy functions, we will not obtain an essentially different invariant. This result might have the other interpretation. Ornstein and Weiss showed in [23] that every finitely observable invariant for the class of all ergodic processes has to be a continuous function of the entropy. It is easy to see that any continuous function of the entropy is finitely observable; one simply composes the entropy estimators with the continuous function itself. In other words, an isomorphism invariant is finitely observable if and only if it is a continuous function of the Kolmogorov-Sinai entropy. Therefore, our result implies that the generalized measure-theoretic entropy is, in fact, finitely observable. It should be possible to give a more direct proof of the finite observability of the generalized measure-theoretic entropy, but the proof cannot be easier than the proof that entropy itself is finitely observable [24,25]. On the other hand, different entropy functions might be still of use, e.g., in the case of zero entropy systems, where we may consider generalizations of concepts used in this case: entropy convergence rates [26,27], generalized topological entropies [28] or entropy dimensions [29]. The generalization of entropy convergence rates can be found in [30]. The result implies that, from the point of view of the theory of entropy in dynamical systems, the crucial property of the Shannon function η is its behavior in the neighborhood of zero.
The note is organized as follows: in the next section, we introduce the dynamical g-entropy and establish its basic properties. The subsequent section is devoted to the construction of a zero dynamical entropy process with sufficiently large g-entropy. Finally, in the last section, we define a measure-theoretic g-entropy of a transformation and show connections between this new invariant and the Kolmogorov-Sinai entropy.
2. Results
2.1. Basic Facts and Definitions
Let (X, Σ, μ) be a Lebesgue space, and let g : [0, 1] ↦ ℝ be a concave function with
(We might assume only that g(0) = 0, but then, the idea of the dynamical g-entropy would fail, since if n+1 ≠ n for every n and
, then the dynamical g-entropy of the partition would be infinite. Therefore, if g is not well-defined at zero, we will assume that
.). By 0, we will denote the set of all such functions. Every g ∈0 is subadditive, i.e., g(x + y) ≤ g(x)+ g(y) for every x, y, x + y ∈ [0, 1], and quasi-homogenic, i.e., φg :(0, 1] → ℝ defined by φg(x):= g(x)/x is decreasing (see [31]). If g is fixed, we will omit the index, writing just φ.
Any finite family of pairwise disjoint subsets of X, such that
, is called a partition. The set of all finite partitions will be denoted by 𝔅. Foragiven ∈ 𝔅, we define the g-entropy of the partition as:
For g = η, the latter is equal to the Shannon entropy of the partition . For , ∈ 𝔅 of the space X, we define a new partition ∨ (the joint partition of and ) consisting of the subsets of the form B ∩ C, where B ∈ and C ∈ .
2.1.1. Dynamical g-Entropies
For an automorphism T : X ↦ X and a partition = {E1, ..., Ek}, we put:
and
Now, for a given g ∈ 0 and a finite partition , we can define the dynamical g-entropy of the transformation T with respect to as:
Alternatively we will call it the g-entropy of the process (X, Σ, μ, T, ). If the dynamical system (X, Σ, T, μ) is fixed, then we omit T, writing just h(g, ). As in the case of Shannon dynamical entropies, we are interested in the existence of the limit of
. If g = η, we obtain the Shannon dynamical entropy h(T, ). However, in the general case, we cannot replace an upper limit in Equation (2) by the limit, since it might not exist. The existence of the limit in the case of the Shannon function follows from the subadditivity of static Shannon entropy. This property has every subderivative function, i.e., a function for which the inequality g(xy) ≤ xg(y) + yg(x) holds for any x, y ∈ [0, 1], but this is not true in general (an appropriate example will be given in Section 2.1.3). Therefore, we propose more general classes of functions for which this limit exists:
It is easy to show that if g is subderivative, then the limit
is finite. Moreover, we will see that values of dynamical g-entropies depend on the behavior of g in the neighborhood of zero. We will prove that if
, then there is a linear dependence between the dynamical g-entropy and the Shannon dynamical entropy of a given partition. Before we give the general result (Theorem 1), we will state a few facts, which we will use in the proof of this theorem. We give the following lemmas, omitting their elementary proofs.
Lemma 1. Let bi > 0, ai ∈ ℝ for i = 1,...,m, then
Lemma 2. If ∈ 𝔅, δ > 0 and g :[0, 1] ↦ ℝ, then:
The following lemma states that the value of the dynamical g-entropy is given by the behavior of g in the neighborhood of zero.
Lemma 3. If g1,g2 ∈0 and there exists c > 0, such that g1(x) = g2(x) for x ∈ [0, c], then for every ∈ 𝔅 h(g1, ) = h(g2, ).
Proof. Let ∈ 𝔅 and g1, g2 ∈0, c > 0, fulfill the assumptions. Because g ∈ 0 is bounded, we have:
Dividing by n and converging to infinity, we obtain:
We may state now the main theorem of this section.
Theorem 1. Let ∈ 𝔅.
- (1)
- If g ∈ 0 is such that g′(0) < ∞, then h(g, ) = 0.
- (2)
- If g1,g2 ∈ 0 are such that g′1(0) = g′2(0) = ∞,and h(g2, ) < ∞, then:If, additionally, , then:
- (3)
- If h(g2, ) = ∞ and, then h(g1, ) = ∞.
Remark 1. Whenever g2 :[0, 1] ↦ ℝ is a nonnegative concave function satisfying g2(0) = 0 and g′2(0) = ∞, we can have any pair 0 < a ≤ b ≤ ∞ as a limit inferior and limit superior of g1/g2 in zero, choosing a suitable function g1. The idea is as follows: construct g1 piecewise linear. To do so, define inductively a strictly decreasing sequence xk → 0 and a decreasing sequence of values yk = g1(xk) → 0, thus defining intervals Jk : = [xk+1, xk], where g is affine. The only constraint to get a concave function is that the slope of g on each interval Jk has to be smaller than yk/xk and increasing with respect to k; this is not an obstruction to approach any limit inferior and limit superior for g1(x)/g2(x), provided that xk+1 > 0 is chosen small enough.
Proof of Theorem 1. Let ∈ 𝔅. Suppose that g ∈ 0 and g′(0) < ∞. Then:
which completes the proof of Point 1. To show Point 2, let g1, g2 ∈ 0 be such that g′1(0) = g′2(0) = ∞ and h(g2, ) < ∞. W.l.o.g, we can assume that g1(x), g2(x) > 0 for x ∈ (0, 1), since if there exists x0 ∈ (0, 1), such that gi(x0) = 0 for i = 1 or i = 2, then we can define g̃i :[0, 1] ↦ ℝ as:
where si ∈ (0, 1] is such that
. Then, g̃ is strictly positive, and by Lemma 3 we have:
Let us assume that:
Since g is subadditive, the sequence
is nondecreasing, and there exists the limit of H(g2, n). If it is finite, then h(g2, ) = 0, and by Equation (3) and Lemma 1, we have:
Because
, there exists M > 0, such that g1(x)/g2(x) < M for x < 1/2. Therefore,
, and by Lemma 1, we obtain:
Thus, we can assume that
.
Fix ε > 0. There exists δ > 0, such that, for x ∈ (0, δ], we have:
Lemma 1 implies that:
Using Equation (3) for every n > 0, we get:
where
for i = 1, 2. Therefore:
and
. Dividing sums by
and from Equation (4), we obtain:
Converging with n to infinity, we obtain:
Therefore:
Thus, we obtain the assertion. In the case of the infinite upper limit of the quotient g1(x)/g2(x), we can repeat the above reasoning, just omitting the upper bound for the considered expressions.
If
and h(g2, ) = ∞, then
, and using similar arguments, we obtain Point 3.
Similar arguments lead us to the following statement:
Theorem 2. Let g1, g2 ∈0 be such that, and let a finite partition have positive g2-entropy. Then, h(g1, ) is infinite.
Theorems 1 and 2 imply a few corollaries:
Corollary 1. If there exists the limit, then
Let
. If g1 = g, g2 = η, then we have the following corollary:
Corollary 2. Let ∈ 𝔅 and g ∈ 0, then:
- (1)
- If Ci(g) < ∞, then h(g, ) ≥ Ci(g) · h().
- (2)
- If Cs(g) < ∞, then h(g, ) ∈ (Ci(g) · h(), Cs(g) · h()).
- (3)
- If, then h(g, )=C(g) · h().
- (4)
- If and h() > 0, then h(g, ) = ∞.
Corollary 3. If (X, Σ, μ, T) has positive Kolmogorov-Sinai entropy and g ∈0 then: Cs(g) < ∞ ⇒ g-entropy of any process (X, Σ, μ, T, ) is finite ⇒ Ci(g) < ∞.
Corollary 4. If, then.
2.1.2. Case of
We will show that for every
, any aperiodic automorphism T and every γ ∈ ℝ, there exists a partition ∈ 𝔅 such that h(g, ) ≥ γ. Since we omit the assumption of ergodicity, we will use different techniques, mainly based on the well-known Rokhlin Lemma, which guarantees the existence of so-called Rokhlin towers of a given height, covering a sufficiently large part of X. Using such towers, we will find lower bounds for the g-entropy of a process.
We will assume that we have an aperiodic system, i.e., (X, Σ, μ, T) for which
If M0,..., Mn−1 ⊂ X are pairwise disjoint sets of equal measure, then τ = (M0, M1,..., Mn−1) is called a tower. If additionally Mk = T−(n−k−1) Mn−1 for k = 1,..., n – 1, then τ is called a Rokhlin tower (It is also known as a Rokhlin–Halmos or Rokhlin–Kakutani tower.). By the same bold letter τ, we will denote the set
. Obviously μ(τ) = nμ(Mn−1). Integer n is called the height of tower τ. Moreover, for i < j, we define a sub-tower:
In aperiodic systems, there exist Rokhlin towers of a given length that cover a sufficiently large part of X:
Lemma 4 ([32]). If T is an aperiodic and surjective transformation of Lebesgue space (X, Σ, μ), then for every ε > 0 and every integer n ≥ 2, there exists a Rokhlin tower τ of height n with μ(τ) > 1 – ε.
Our goal is to find a lower bound for the dynamical g-entropy of a given partition. For this purpose, we will use Rokhlin towers, and we will calculate dynamical g-entropy with respect to a given Rokhlin tower. This leads to the following quantity: Let be a finite partition of X and F ∈ Σ then, we define the (static) g-entropy of restricted to F as:
The following lemma gives the estimation for H(g, ) from below by the value of the g-entropy restricted to a subset of X.
Lemma 5. Let g ∈0. Let be a finite partition, such that there exists a set E ∈ with 0 < μ(E) < 1. If F ∈ Σ, then:
where.
Proof. Let A ∈ . Three-slope inequality implies that:
Thus, for sets of measure
, we have:
Therefore, we obtain:
and:
which implies that:
The following lemma will play an important role in the proof of the main theorem of this section.
Lemma 6. Let n ∈ ℕ, E ∈ Σ. Let E : = {E, X\E}. Suppose that g ∈ 0 is nonnegative in [0,α], where α is some positive number. Then, there exist δ > 0 and s ∈ (0,α), such that:
for every F ∈ Σ s.t. μ(EΔF) < δ (Δ denotes the symmetric difference), where.
Proof. It is easy to show that for every n ∈ ℕ, E ∈ Σ, ε > 0, there exists δ > 0, such that:
W.l.o.g., we may assume that
(adding empty sets if necessary). The nonnegativity of g for x ∈ [0, α] and its concavity imply that there exists s ∈ (0, α), such that g is nondecreasing in [0, s]. Fix n ∈ ℕ and E ∈ Σ. There exists ε ∈ (0, s/2), such that:
It is easy to see that for x ∈ [0, s], the monotonicity and subadditivity of g implies that:
Let F ∈ Σ be such that μ(EΔF) ≤ δ. Define
. | max{μ(Ai), μ(Bi)} < s}. From Equations (5) and (6) and the monotonicity of g in [0, s], we obtain:
To find the lower bound for the g-entropy of a partition, we will use so-called independent sets. We construct the independent set in the following way: Let τ be a tower of height m. We divide the highest level of this tower (Mm−1) into two sets of equal measure, let us say I(m−1) and Mm−1\I(m−1). Next, we consider T−1I(m−1) and T−1(Mm−1\I(m−1)). We divide each of them into two sets of equal measure, obtaining sets
, and define set I(m−2) as the algebraic sum of two of those sets—one subset of T−1I(m−1) and one of T−1(Mm−1\I(m−1)). We perform this algorithm, until we achieve the lowest level of the tower — M0 (see Figure 1). Eventually, we define
. We call this set an independent set in τ.
Figure 1.
Set I (with dashes) in a tower a height of five.
We can make this construction, because every aperiodic system does not have atoms of positive measure, and in every non-atomic Lebesgue space for every measurable set A and every α ∈ [0, α], there exists B ⊂ A, such that μ(B) = α.
We are able to give an explicit formula for the g-entropy of the partition generated by the independent set in τ.
Lemma 7. Let τ = (M, TM,..., T2n−1M) be the Rokhlin tower of height 2n and I ∈ Σ be an independent set in τ. If, then:
where φ(x) = g(x)/x for x > 0.
Proof. The independence of I in τ implies that the partition:
is a partition of
into 2n sets of equal measure 2−n. Therefore:
Theorem 3. Let and T be an aperiodic, surjective automorphism of a Lebesgue space (X, Σ, μ), and let γ ∈ ℝ, then, there exists a partition ∈ 𝔅, such that:
Proof. We will prove that for any γ > 0, there exists a partition E = {E, X\E}, such that h(g, ) ≥ γ. We define recursively a sequence of sets En ∈ Σ. Let:
Let n > 0, and assume that we have already defined En−1, Nn−1 and δn−1. Using Lemma 6, we can choose δn > 0, such that:
for any F ∈ Σ, for which μ(En−1 ΔF) < 2δn.
Since:
we can choose such Nn ∈ ℕ that:
By Lemma 4, there exists Mn ∈ Σ, such that τn = (Mn, TMn,...,T2Nn−1Mn) is a Rokhlin tower of measure μ(τn) = δn. Let In ⊂ τn be an independent set in τn and:
Then:
for all positive integers n. By Equation (7), we have δn < 2−n, and we conclude that
is a Cauchy sequence in L1(X). Therefore, there exist E ∈ Σ, such that 1En converges to 1E. For this set, we have:
Since En ∩ τn = In, applying Equation (8) and Lemmas 5 and 7, we obtain that for Nn, such that δn · 2−Nn−1 < s:
From Equation (9), we obtain that:
Thus,
2.1.3. Bernoulli Shifts
Let 𝒜 = {1,...,k} be a finite alphabet. Let
and σ be a left shift:
For any s ≤ t and block [ω0, ... ,ωt−s] with ai ∈𝒜, we define a cylinder:
We consider a Borel σ-algebra with respect to the metric, which is given by d(x, y) = 2−N, where N = min{|i| : xi ≠ yi}. Let p = (p1,..., pk) be a probability vector. We define a measure ρ = ρ(p) on 𝒜 by setting ρ({i}) = pi. Then, μp is a corresponding product measure on X = 𝒜. Thus, the static g-entropy of a partition 𝒜 = {[1], [2],..., [k]} is equal to:
where ω = (ω0,...,ωn−1). By the concavity of g, we have:
where equality holds only when
. Before calculating the dynamical g-entropy of 𝒜 with respect to μp*, we give the following lemma, the proof of which will be given later:
Lemma 8. If g ∈0, then:
for any κ > 1.
Therefore, applying Lemma 8 for the partition 𝒜 and κ = k, we obtain:
Remark 2. If we consider the lower limit instead of the upper limit, we would obtain:
Therefore, we cannot replace an upper limit by the limit in the definition of the dynamical g-entropy.
Proof of Lemma 8. We will show just the equality for the upper limit, since the equality for the lower limit may be obtained analogously. Let
and
be such that
and xn ∈ (κ−mn,κ−mn+1) for every n ∈ ℕ. Then, –log xn ≥ – log κ−mn+1. Every g ∈ 0 is quasi-homogenic; so, for every positive integer n occurs:
Therefore:
and:
2.2. Kolmogorov-Sinai Entropy-Like Invariant
The basic tool considered in the ergodic theory is the Kolmogorov–Sinai entropy, which is a supremum of Shannon dynamical entropies over all finite partitions:
It is invariant under metric isomorphism. Following the Kolmogorov proposition, we take the supremum over all partitions of the dynamical g-entropy of a partition. For a given system (X, Σ, μ, T), we define:
and call it the measure-theoretic g-entropy of transformation T with respect to measure μ.
It is easy to see that it is an isomorphism invariant. Ornstein and Weiss [23] showed the striking result that measure-theoretic entropy is the only finitely observable invariant for the class of all ergodic processes. More precisely, every finitely observable invariant for a class of all ergodic processes is a continuous function of entropy. Of course, in the case of
by Corollary 2, we have:
We will show that for a wider class of functions, namely for those for which:
we have:
for any ergodic transformation T. This shows that the measure-theoretic g-entropy is, in fact, finitely observable: one might simply compose the entropy estimators [25] with the linear function itself. Our proof will be similar to the proof of ([12], Theorem 1.1) where Takens and Verbitski showed that for ergodic transformations, the supremum over all finite partitions of dynamical Rényi entropies of order α > 1 are equal to the measure-theoretic entropy of T with respect to measure μ.
Let us introduce necessary definitions and facts. Let Ti be automorphisms of Lebesgue space (Xi, Σi, μi) for i = 1, 2, respectively. Then, we say that T2 is a factor of transformation T1, if there exists a homomorphism ϕ: X1 ↦ X2, such that:
Suppose that T2 is a factor of T1 under homomorphism ϕ. Then, for an arbitrary finite partition of X2, we have:
Hence, h(g, T2, ) = h(g, T1, ϕ−1 ). Therefore:
This implies the following proposition:
Proposition 1. If T2 is a factor of T1, then for every function g ∈ 0:
2.2.1. Measure-Theoretic g-Entropies for Bernoulli Automorphisms
An automorphism T on (X, Σ, μ) is called Bernoulli automorphism if it is isomorphic to some Bernoulli shift. The crucial role in the proof of the main theorem of this section (Theorem 5) will be played by a well-known Sinai theorem:
Theorem 4 (Sinai [33]). Let T be an arbitrary ergodic automorphism of some Lebesgue space (X, Σ, μ). Then, each Bernoulli automorphism with hμ(T1) ≤ hμ(T) is a factor of the automorphism T.
We start proving the following proposition:
Proposition 2. Let T be an arbitrary ergodic automorphism with hμ(T) ≥ log M for some integer M ≥ 2. Then, for every g ∈0:
Proof. Consider a shift σ over the alphabet 𝒜 = {0, 1 ..., M – 1} with the corresponding Bernoulli measure generated by
. It is easy to see that hμ(σ) = log M. From Theorem 4, we conclude that σ is a factor of T . Therefore, applying formula (10), we obtain:
Applying Lemma 8 completes the proof.
2.2.2. Main Theorem
Our goal in this section is the following result:
Theorem 5. Let T be an ergodic automorphism of Lebesgue space (X, Σ, μ), and g ∈0 be such that Cs(g) ∈ (0, ∞). Then:
If, then hμ(g, T) = 0. If g ∈0 is such that Cs(g) = ∞ and T has a positive measure-theoretic entropy, then hμ(g, T) = ∞.
Moreover, for
from Theorem 3, we have:
Corollary 5. Let. If (X, T) is aperiodic and surjective, then hμ(g, T) = ∞.
To prove Theorem 5, we need the first few preliminary lemmas.
Lemma 9. If T is an automorphism of the Lebesgue space (X, Σ, μ), then for every g ∈0:
Proof. Let ∈ 𝔅, m ∈ ℕ. We have:
Fix k ∈ ℕ. Then,
is a refinement of ∨ T−m ∨ ...∨ T−m(k−1) for n = km,..., k(m+1) − 1.
Therefore:
for n = km,..., k(m +1) – 1. Let us introduce the following notation:
Then, we can rewrite Equation (12) in the form:
for n = km,..., km + m – 1. Taking the supremum in Equation (13), we obtain:
Therefore:
and this is equivalent to the statement:
Taking the supremum over all finite partitions, we obtain the assertion.
Next, the lemma will be just a weaker version of Theorem 5.
Lemma 10. If an automorphism Tm of a Lebesgue space (X, Σ, μ) is ergodic for every m ∈ ℕ, then for every g ∈0, such that Cs(g) < ∞ holds:
If, then hμ(g, T) = 0. If g ∈0 is such that Cs(g) = ∞ and T has a positive Kolmogorov-Sinai entropy, then hμ(g, T) = ∞.
Proof. The case of
follows from Corollary 2. Suppose that there exists such
, which fulfills the assumptions of the lemma and for which we have:
Then, applying Lemma 9 to the transformation Tm and using equality hμ(Tm) = mhμ(T) (see [2], Theorem 4.3.16), we obtain:
Therefore, for sufficiently large m, there exists an integer M for which:
Proposition 2 applied to the transformation Tm guarantees that for every g ∈ 0 with positive (finite) Cs(g), we have:
If Cs(g) = ∞ and hμ(T) > 0, then there exists such an integer m > 0 that:
and by Proposition 2 and Lemma 9:
which completes the proof.
Proof of Theorem 5. If hμ(T) = 0, the statement is true, because for any ∈ 𝔅, we have:
Suppose that 0 < hμ(T) < ∞. Automorphism T is ergodic. Thus, it has a factor, which is a Bernoulli automorphism T′ with hμ(T) = hμ(T′). Every Bernoulli automorphism is mixing, so Tm is ergodic for each m. Applying Lemma 10, we obtain:
Since T′ is a factor of T, Proposition 1 implies that:
which completes the proof of the case of finite hμ(T). If hμ(T) = ∞, then Proposition 2 implies that:
for every M > 0.
2.2.3. Generator Theorem Counterpart
In the case of
, there is no counterpart of a Kolmogorov-Sinai generator theorem, which states that the measure-theoretic entropy of the transformation T is realized on every generator of the σ-algebra Σ. Let us consider Sturm shifts, shifts that model translations of the circle 𝕋 = [0, 1). Let β ∈ [0, 1) and consider the translation ϕβ :[0, 1) ↦ [0, 1) defined by ϕβ(x) = x + β (mod 1). Let denote the partition of [0, 1) given by = {[0,β), [β, 1)}. Then, we associate a binary sequence to each t ∈ [0, 1) according to its itinerary relative to ; that is, we associate to t ∈ [0, 1) the bi-infinite sequence x defined by xi =0 if
(t) ∈ [0,β) and xi =1 if
(t) ∈ [β, 1). The set of such sequences is not necessary closed, but it is shift-invariant and so its closure is a shift space called the Sturmian shift. If β is irrational, then Sturmian shift is minimal, i.e., there is no proper subshift. Moreover, for a minimal Sturmian shift, the number of n-blocks that occur in an infinite shift space is exactly n +1. Therefore for zero-coordinate partition 𝒜, which is a finite generator of σ-algebra Σ, and for any function g ∈0, we have:
where μS is the unique invariant measure for Sturm shift. Thus,
On the other hand, since it is strictly ergodic (thus aperiodic), Theorem 3 implies that for any
:
Therefore, we have a finite generator, for which the supremum is not attained.
3. Discussion
In this note, we discussed the generalization of the dynamical and Kolmogorov-Sinai entropy based on the idea of considering in the definition of the dynamical entropy a concave function vanishing at the origin instead of the Shannon entropy function η. The connections between dynamical entropies and g-entropies show that the crucial property of η for applications of the Shannon entropy in the dynamical systems is the behavior of η in the neighborhood of zero. Additionally, the main result of the paper obtained for the generalization of the KS entropy states that, usually, there is a linear dependence between the obtained invariant and the KS entropy. It also implies (due to the fact that it is a continuous function of the entropy) that the measure-theoretic g-entropy is finitely observable (see, e.g., [23]). Moreover, considering functions that behave in the neighborhood of zero differently than the Shannon function usually trivializes the theory. On the other hand, we showed that if the limit of g/η at zero is infinite, then for every positive number γ, there exists a partition for which the g-entropy will be greater than or equal to γ. Thus, the measure-theoretic g-entropy will be infinite. The example from Section 2.2.3, based on this result, implies that there is no counterpart of the generator theorem (e.g., Tsallis entropies for α< 1, which we obtain considering
with α ∈ (0, 1), fit into this scheme). However, the concept of g-entropies is still of use, e.g., considering the rate of convergence of partial g-entropies may give additional information about the system [30]. The most promising direction in this context seems to be considering functions for which C(g) = 0 and g′(0) = ∞.
Conflicts of Interest
The author declares no conflict of interest.
References
- Downarowicz, T. Entropy in Dynamical Systems; Cambridge University Press: New York, NY, USA, 2011. [Google Scholar]
- Katok, A.; Hasselblatt, B. Introduction to the Modern Theory of Dynamical Systems; Cambridge University Press: New York, NY, USA, 1997. [Google Scholar]
- Misiurewicz, M. A short proof of the variational principle for n+ action on a compact space. Astérisque 1976, 40, 147–157. [Google Scholar]
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J 1948, 27. [Google Scholar]
- Rényi, A. On measures of entropy and information. In Proceedings of the 4th Berkeley Symposium on Mathematical Statistics and Probability, 20 June–30 July 1960; University of California Press: Berkeley, CA, USA, 1961; Volume 1. Statistical Laboratory of the University of California: Berkeley, CA, USA; pp. 547–561.
- Arimoto, S. Information-theoretical considerations on estimation problems. Inf. Control 1971, 19, 181–194. [Google Scholar]
- Wu, Y.; Verdú, S. Rényi Information Dimension: Fundamental Limits of Almost Lossless Analog Compression. IEEE Trans. Inf. Theory 2010, 56, 3721–3748. [Google Scholar]
- Csiszár, I. Axiomatic characterization of information measures. Entropy 2008, 10, 261–273. [Google Scholar]
- De Paly, T. On entropy-like invariants for dynamical systems. Z. Anal. Anwend 1982, 1, 69–79. [Google Scholar]
- De Paly, T. On a class of generalized K-entropies and Bernoulli shifts. Z. Anal. Anwend 1982, 1, 87–96. [Google Scholar]
- Grassberger, P.; Procaccia, I. Estimation of the Kolmogorov entropy from a chaotic signal. Phys. Rev. A 1983, 28, 2591–2593. [Google Scholar]
- Takens, F.; Verbitski, E. Generalized entropies: Rényi and correlation integral approach. Nonlinearity 1998, 11, 771–782. [Google Scholar]
- Takens, F.; Verbitski, E. Rényi entropies of aperiodic dynamical systems. Isr. J. Math 2002, 127, 279–302. [Google Scholar]
- Liu, Q.; Cao, K.-F.; Peng, S.-L. A generalized Kolmogorov–Sinai-like entropy under Markov shifts in symbolic dynamics. Physica A 2009, 388, 4333–4344. [Google Scholar]
- Mesón, A.M.; Vericat, F. Invariant of dynamical systems: A generalized entropy. J. Math. Phys 1996, 37, 4480–4483. [Google Scholar]
- Mesón, A.M.; Vericat, F. On the Kolmogorov-like generalization of Tsallis entropy, correlation entropies and multifractal analysis. J. Math. Phys 2002, 43, 904–918. [Google Scholar]
- Havrda, J.; Charvàt, F. Quantification method of classification processes. Concept of structural α-entropy. Kybernetika 1967, 3, 30–35. [Google Scholar]
- Abe, S. Tsallis entropy: How unique? Contin. Mech. Thermodyn 2004, 16, 237–244. [Google Scholar]
- Furuichi, S. Information theoretical properties of Tsallis entropies. J. Math. Phys 2006, 47, 023302. [Google Scholar]
- Tsallis, C. Possible generalization of Boltzmann–Gibbs statistics. J. Stat. Phys 1988, 52, 479–487. [Google Scholar]
- Tsallis, C. Entropic nonextensivity: A possible measure of complexity. Chaos Solitons Fractals 2002, 13, 371–391. [Google Scholar]
- Tsallis, C.; Plastino, A.R.; Zheng, W.-M. Power-law sensitivity to initial conditions—New entropic representation. Chaos Solitons Fractals 1997, 8, 885–891. [Google Scholar]
- Ornstein, D.S.; Weiss, B. Entropy is the only finitely observable invariant. J. Mod. Dyn 2007, 1, 93–105. [Google Scholar]
- Weiss, B. (The Hebrew University of Jerusalem); 2013; personal communication.
- Weiss, B. Single Orbit Dynamics; American Mathematical Society: Providence, RI, USA, 2000. [Google Scholar]
- Blume, F. The Rate of Entropy Convergence, University of North Carolina, Chapel Hill, NC, USA, 1995.
- Blume, F. Possible rates of entropy convergence. Ergod. Theory Dyn. Syst 1997, 17, 45–70. [Google Scholar]
- Galatolo, S. Global and local complexity in weakly chaotic dynamical systems. Discret. Contin. Dyn. Syst 2003, 9, 1607–1624. [Google Scholar]
- Ferenczi, S.; Park, K.K. Entropy dimensions and a class of constructive examples. Discret. Contin. Dyn. Syst 2007, 17, 133–141. [Google Scholar]
- Falniowski, F. Possible g-entropy convergence rates. 2013, arXiv. 1309.6246. [Google Scholar]
- Rosenbaum, R.A. Sub-additive functions. Duke Math. J 1950, 17, 227–247. [Google Scholar]
- Heinemann, S.-M.; Schmitt, O. Rokhlin’s Lemma for non-invertible maps. Dyn. Syst. Appl 2001, 10, 201–214. [Google Scholar]
- Sinai, Y.G. Weak isomorphism of transformation with an invariant measure. Sov. Math 1962, 3, 1725–1729. [Google Scholar]
© 2014 by the authors; licensee MDPI, Basel, Switzerland This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).