1. Introduction
According to the central limit theorem (CLT), the sum of independent variables with finite first and second moments is governed by a Gaussian distribution when the number of summands is asymptotically large. The mean value and the variance of the Gaussian equal the sum of the individual mean values and variances, respectively. The Gaussian distribution has maximal entropy for a given variance and is reached independently of the distributions from which the summands are sampled. The convergence to the Gaussian limit, therefore, can be viewed as a loss of information about the original data. Extension to sums of variables with diverging first and second moments have been derived in [
1,
2,
3,
4]; the asymptotic distributions of there are no longer Gaussian, but are still members of a family of so-called
stable distributions.
Experience shows that many systems are successfully modeled by stable distributions, for example, in the theory of errors and propagation of uncertainty. This is often justified by the fact that errors, as well as many other quantities of interest, can be conceived as the sum of a large number of variables representing disparate magnitudes that appear to be unrelated. Yet, Physics, for instance, dictates that all the variables describing a system of interacting particles (as opposed to an ensemble of free particles) are correlated to one another. Therefore, the independence condition is no more than an approximation.
To improve this approximation, extensions of the CLT have also been developed for variables that bear different degrees of statistical dependencies, including those obtained through the subsequent application of a deterministic rule that produces ergodicity and aperiodicity [
5,
6,
7,
8,
9,
10,
11]. Here, we analyze several systems of this type. As discussed in the next section, a conveniently modified version of the CLT exists for appropriately scaled sums of variables deterministically related to one another. Notably, the family of stable distributions for these cases coincides with the ones obtained for independent variables. These extensions provide mathematical certainty that sums of strongly correlated variables, if produced through a chaotic dynamical system, lose all memory of their original distribution, and asymptotically approach a distribution that also happens to be the limit of sums of independent variables sampled from a certain distribution. The strong statistical dependencies governing the physical world, therefore, may be legitimately ignored when describing the probability distributions of macroscopic variables, and it is legitimate to conceive the latter as a sum of a large number of microscopic, independent variables. This property greatly simplifies the description of macroscopic systems and has probably played a crucial role in the development of the theory of probability.
In practical situations, however, it is important to know how many terms a sum needs to include for its distribution to be well described by the asymptotic result. To shed light on this question, we study in this paper the convergence of the probability distribution of a sum of perfectly correlated variables, generated through the iteration of a chaotic, deterministic map, towards the asymptotic distribution predicted using the extensions of the CLT. The aim is to characterize how the loss of information about the deterministic nature of the map depends on the number of variables that are summed together. Since previous theoretical results do not predict the rate of convergence towards asymptotic distributions in deterministic systems, our analysis is based on numerical simulations of several paradigmatic examples, and on a comparison with the behavior of randomly sampled systems with the same distributions.
The paper is organized as follows. In
Section 2, we present the main theoretical tools to be employed later; these include the extension of the CLT to variables that are strongly correlated, the information–theoretical measures that quantify the differences between probability distributions, and the behavior of the variance of a sum of variables that are correlated. The following three sections apply these tools to the analysis of a chaotic dynamical system with a uniform marginal distribution and varying Lyapunov exponent (the Bernoulli map,
Section 3), a chaotic dynamical system with a highly nonuniform marginal distribution and several types of orbits (the logistic map,
Section 4), and an example of a process with fat-tailed distribution (
Section 5). Our main conclusions are summarized in
Section 6.
2. Central Limit Theorem for Deterministic Maps
We consider a generic one-dimensional map,
, with a well-defined invariant measure
, determined using the identity
where
is the composition of functions
and
, and the prime indicates differentiation with respect to
x. We assume that the mean value
is finite over the distribution
, and—for now—we assume the variance
of
x is also finite:
where the integrals run over the whole domain of variation of
x. In
Section 5, we study a case where we relax the condition that
is finite. A central limit theorem (CLT) for this kind of system applies [
6,
7,
8,
9,
10,
11] when the map under study is ergodic and aperiodic. We recall that a map is ergodic if all its invariant sets are null or co-null, and it is aperiodic if its periodic orbits form a null set [
11]. The combination of ergodicity and aperiodicity is typically equivalent to the dynamics being chaotic [
12]. In this case, the CLT states that the distribution of the (centered, suitably normalized) sums of
N successive values of
,
becomes normal for
:
for some value of the variance
. Here,
denotes the Gaussian centered at zero, with standard deviation
.
For each value of
N, the variables
and
can be integrated into a single two-dimensional map:
where
is the
N-th self-composition of
. Thus, for
, the marginal invariant measures of the variables
x and
in map (
5) are, respectively,
and the Gaussian
of Equation (
4).
In contrast with the sums of statistically independent random variables drawn from a given distribution, in the limit
, the variance of the sums
does not necessarily coincide with that of the summands,
. The difference arises from the correlations between successive values of
, induced by the map
, with the ensuing mutual correlations between the values of
. For a finite number of summands
N, the variance of
is given by the Green–Kubo formula [
13]:
where the overline indicates the average with respect to the distribution
. The value of
becomes independent of
t when the process
has reached a stationary regime. For
, the variance is
Provided that the sum converges, this formula gives the variance of the asymptotic normal distribution
of increasingly long sums
.
In the following, we study the process of convergence towards the asymptotic distribution predicted by the above CLT for some selected deterministic maps, as the number of terms in the sums
grows. For each
N, we numerically iterate Equation (
5) and estimate the distribution of the sums
,
, as a suitably normalized
-column histogram built from, typically,
values of
. To quantify the difference between
and the expected asymptotic Gaussian distribution
, we use the Kullback–Leibler divergence (KLD). Recall that the KLD between two distributions
and
is defined as
This quantity measures the inefficiency with which the data
s is represented by a code optimized to be maximally compact under the assumption that the distribution is
when, in reality, the data are generated from
. The inefficiency equals the mean number of extra bits per sample [
14]. The divergence only vanishes when the two distributions coincide, and is otherwise positive. For brevity, we hereafter denote as
the KLD between the distribution
and the asymptotic normal distribution
:
.
Additionally, for each
N, it is interesting to compare
with a normal distribution with the variance
given by Equation (
6), namely, the same variance as the sums
. Since
as
, this is an alternative way of characterizing the convergence to the asymptotic Gaussian
. For this comparison, we introduce the KLD
.
Finally, in order to contrast the deterministic dynamics of the chaotic map under study with a genuinely aleatory process, we calculate the KLD for the distribution of sums of the same form as in Equation (
3), but with the
N values of the variable
x drawn at random from the invariant measure
. According to the standard CLT for statistically independent variables, as
N grows, the distribution
of these random sampling sums is expected to asymptotically converge to a Gaussian with variance
. To quantify this convergence, we compute
.
The measures
,
and
reflect three different aspects of the convergence of
to
. The process by which
tends to zero describes how independent variables, when summed together, lose the memory of the distribution from which they are sampled and approach a Gaussian. The Gaussian distribution is the one with maximal entropy among those with fixed variance. By acquiring a Gaussian shape, therefore, the distribution of the sum maximizes uncertainty. In
Appendix B, we show that, for large
N, the divergence
decays as
if
is not symmetric around its mean value, and at least as fast as
if there is symmetry.
A steep decay of
with
N, at a faster rate than
, implies a rapid evolution of
towards a bell-shaped distribution, whose variance may still have to evolve to its asymptotic value
. The convergence process can therefore be conceived as a sequence of two stages, the first one consisting of shedding all the structure in
and becoming Gaussian-like, and the second, adjusting the variance. Once
is approximately Gaussian, its KLD with the asymptotic distribution
can be analytically calculated in terms of their respective variances:
3. The Bernoulli Map
As a first case of study, we take the generalized Bernoulli map
where
indicates a fractional part, and
is an integer factor. This map has been extensively studied since long ago as a paradigm of deterministic chaotic systems, due to its combination of complex behavior and analytical traceability. Its Lyapunov exponent equals
. The invariant measure of
is particularly simple:
for all
m, with
and
. We show in
Appendix A that the variances of the sums
can be explicitly calculated:
Note that for
, this variance takes the approximate form
with
. For
, in turn,
We first consider the Bernoulli map for
. Dark full lines on the left column of
Figure 1 show numerical results for the distributions of the sums
,
, for three small values of
N. Light-colored curves stand for the asymptotic Gaussian
, and dashed curves are the Gaussians
for each
N. Their respective variances,
and
, are given by Equation (
12). On the right column, dark- and light-colored curves, respectively, show the distributions of the sums of randomly sampled values of
x,
, calculated analytically as
N-th order self-convolutions of
, and the expected asymptotic Gaussian
. A comparison of the two columns illustrates the difference between the distributions of the sums generated by map iteration on one side and by random sampling on the other. It also shows that convergence to the asymptotic distribution is faster in the latter case.
The main panel of
Figure 2 shows, with different symbols, the KLDs
,
, and
, defined in the preceding section. For
, by definition,
. For large
N (
Appendix B),
decays as
. The straight lines in the log–log plot of the figure have slope
, suggesting that the decay of the divergence
approximately follows the same asymptotic dependence on
N. The inset in
Figure 2 shows, as dots, the numerical estimation of the variance of
over the distribution
as a function of
N. The dashed curve corresponds to the analytical expression of Equation (
12).
In the range shown in the figure, for
,
is larger than
by a factor of around 14. Meanwhile, in the same range,
decays faster, approximately as
. As discussed at the end of
Section 2, this faster decay of
suggests that
is rapidly approaching a Gaussian distribution, with a KLD with the asymptotic distribution
as given by Equation (
9). Replacing Equation (
13) into Equation (
9) and expanding up to second order in
yields
For
we have
so that, according to the above equation,
. A power–law fitting of the data for
for
gives
, which fits the prediction of Equation (
15) remarkably well. This agreement provides strong evidence in favor of the hypothesis that
converges to
in two stages, acquiring a Gaussian shape in the first, and adjusting the variance in the second. The transition from the first stage to the second, however, does not imply that
is strictly speaking a Gaussian distribution.
What are the implications of the fact that after the initial transient
and
both decay with the same power law, approximately proportional to
? In this regime,
which means that, for each
N,
is approximately equal to
. By increasing the number of random samples drawn from the invariant measure (
11),
diminishes by the same amount as
diminishes when running the Bernoulli deterministic mapping a re-scaled, larger number of samples, with a scaling factor of
. In other words,
samples of the deterministic map are as informative about the asymptotic distribution as a single sample in the random drawing. The presence of correlations makes each new sample from the deterministic dynamics less informative (by a factor of
) than from purely independent draws.
The factor
may also be semi-quantitatively associated with the relation between the asymptotic variance
and the original variance
. In Equation (
3), the normalization factor
compensates for the fact that the variance of a sum of
N independent samples is proportional to
N. Yet, when the summands bear statistical interdependence, the intended compensation need not be attained. The higher the correlations in the deterministic map, the less informative each new datum is, the more unsuccessful the compensation, and the larger the increase in the asymptotic variance. In the present case, the variance increases threefold, from
to
, which is similar to the factor relating
and
, namely,
.
Considering now the other values of
m in the Bernoulli map (
10), the numerical results presented in
Figure 3 show that the dependence of
on
N is similar to that obtained for
, with the only difference that
becomes progressively smaller as
m grows. As before, the convergence may be conceived as consisting of two stages, with Equation (
9) approximately holding for the second stage. According to the results of
Figure 3, the second state is reached faster for larger values of
m. As expected from the large-
N asymptotic behavior of
predicted by Equation (
15) with
[cf. Equation (
13)], it approaches
for large
m. This implies that the effect of the statistical dependencies induced by the deterministic nature of the map decreases as
m grows. The KLD
is not shown in
Figure 3, but its behavior is similar to that of the case of
(
Figure 2).
In summary, in the Bernoulli map, decreases faster than during the first stage of the convergence process, where acquires a Gaussian-like shape. Only later is the variance adjusted towards its final value. The second stage can be modeled analytically, providing a good qualitative description of the asymptotic behavior inferred from numerical results.
4. The Logistic Map: Full Chaos and Intermittency
We now turn our attention to the logistic map [
17,
18]
with
. Much like Bernoulli’s, the logistic map hardly needs any presentation. We first consider the case
, which we call the regime of “full chaos”. For this value of
, the dynamics are chaotic and therefore comply with the hypotheses of the CLT for deterministic systems discussed in
Section 2. Moreover, due to the existence of a nonlinear change of variables that transforms the logistic map with
into the Bernoulli map of Equation (
10) with
, several analytical results for the latter can be extended to the former. In spite of this connection, as we show below, the statistics of the sums
are qualitatively different between the two maps.
For
, the invariant measure of the logistic map can be written explicitly as [
19]
for
, and 0 otherwise. The mean value is
and the variance is
. As we show in
Appendix A, the correlations between iterations of the map,
, vanish for all
k. From Equations (
6) and (
7), this implies that the variances of the sums
are the same for all
N, and therefore coincide with both the variance of
x and with the limit for
:
. Therefore—in contrast with the Bernoulli map studied in the preceding section—it is not possible to discern between a first stage of convergence to a Gaussian profile and a second stage of adjustment of the variance.
In
Figure 4, the left column shows numerical estimations of the distributions
of the sums of
N consecutive iteration of the logistic map with
, for three values of
N. The light-colored curve corresponds to the expected asymptotic Gaussian.
In addition to the sharp peaks in the profile of
for small
N, an important difference with the Bernoulli map (
Figure 1) is that
is no longer symmetric with respect to zero. This asymmetry may come as a surprise, taking into account that both
and
are symmetric around the mean value
. The asymmetry, however, originates from the fact that the functions
,
,
, …, which ultimately determine the distributions of the sums
, are
not symmetric around
.
On the right column of
Figure 4 we show, for the same values of
N, the distributions
of sums of
N random values of
x sampled from
. In contrast with the case of the Bernoulli map,
is here estimated numerically. As expected, the distributions of random sampling sums are now symmetric with respect to zero, and exhibit a fast convergence to the asymptotic Gaussian.
Figure 5 shows
and
for the fully chaotic logistic map, as functions of
N. Since, as explained above,
equals
for all
N, now
coincides with
. Due to the symmetry of
with respect to its mean value, the arguments given in
Appendix B apply to this case, and
decays as
for large
N. The full straight line in the log–log plot of the figure has slope
, confirming this prediction in the plotted range. Yet, the behavior of
is considerably different. It starts with a small increment between
and 2, where it attains a maximum, and thereafter decays rapidly up to
. This decay corresponds to the interval of
N for which the distribution
displays identifiable singularities. For
, the singularities start to overlap, and the distribution
varies more smoothly and displays a well-defined asymmetric bell-shaped profile. In this zone, the decay of
is slower and approximately behaves as
, as illustrated by the dashed straight segment of slope
. As shown in
Appendix B, a decay as
is expected for the KLD of the distribution of random sampling sums when the distribution of the individual summands is not symmetric with respect to the mean value. If the disparate dependence on
N between
and
persists as
N grows beyond the range considered here, their relative difference would increase indefinitely for
.
Although still chaotic, other values of
in Equation (
16) give rise to qualitatively different dynamical features in the logistic map. For
, which is our next case of study, the dynamics are intermittent. Just above this value of
, at
, the logistic map enters the largest stability window within its chaotic regime, where
becomes asymptotically locked in a period-3 orbit. For
, the vicinity of the critical point manifests itself in the form of intermittent behavior for
. Namely, the dynamics alternate intermittently between intervals of “turbulent” evolution, where its behavior is conspicuously chaotic, and “laminar” evolution, where
remains temporarily close to the period-3 orbit, but eventually departs away from it. The left panel of
Figure 6 shows 900 successive iterations of
for the above value of
, illustrating both kinds of behavior.
For
, no analytical description of the logistic map exists, and we must resort to numerical techniques. As inferred from the left panel of
Figure 6, in this case,
covers only a portion of the interval
, between
and
, and displays three peaks near the values of
x in the period-3 orbit. Our numerical estimations for the mean value and the variance are
and
. In principle, the variance of the sums
could be obtained from Equations (
6) and (
7) by numerically computing the correlations
. These quantities, however, exhibit sharp oscillations and slow convergence as
k grows, as well as persistent fluctuations for large
k. The right panel in
Figure 6 shows
up to
. In practice, such features make impossible the evaluation of the variances
and
using the sums in Equations (
6) and (
7). We therefore resort to their direct numerical calculation using the values of
obtained from successive map iterations. In particular, our estimation for the variance of the sums in the limit
is
.
Colored symbols in the main panel of
Figure 7 stand for
in the case of the logistic map with
, as a function of
N. As with full chaos (cf.
Figure 5), two distinct decay regimes are identifiable. Moreover, the behavior for
now contains signatures of the pseudo-periodic nature of the mapping in the “laminar” intervals, namely, the relatively large values of
when
N is a multiple of 3 (triangles). In fact, for those values of
N, the distributions
are narrower and sharper than for the remaining values, giving rise to higher KLDs. This is clearly illustrated by the dependence of the variance
on
N, shown in the inset of the figure. After an abrupt initial decay,
displays oscillations of period 3, which progressively damp out as
N grows. For
, the difference in
for multiples of 3 rapidly smooths out, as the KLD enters a regime where it decays approximately as
, as indicated by the dashed segment of slope
.
For this case of intermittent dynamics, we have also calculated
, finding qualitatively the same behavior as for
. As a matter of fact,
and
typically differ from each other in just about a 10%. Thus, for the sake of clarity, the numerical estimations of
were not included in
Figure 7. As for the KLD of the distribution of random sampling sum,
, the results of
Appendix B indicate that it should decay as
for large
N. This behavior, however, has not yet been reached in the range of values displayed in
Figure 7. Assuming nevertheless that this is the asymptotic dependence of
, our results suggest that the KLD for random sampling sums is no less than three orders of magnitude smaller than
for large
N.
In summary, both for
and
, the main difference between the statistics of the sums
obtained from the iteration of the logistic map and from a random sampling of the corresponding invariant measures, as
N grows, resides in their disparate rates of approach towards the asymptotic distribution. Within the range of
N considered in our numerical calculations, the decay of
as
can be qualitatively understood by the lack of symmetry in the invariant measures although, strictly speaking, the corresponding result in
Appendix B holds for random sampling only.
Both when and , for , the difference between and is well above two orders of magnitude. In the intermittent case, moreover, the pseudo-periodic character of the “laminar” dynamics reveals itself in the form of oscillations in for small N, which are naturally absent in . Plausibly, pseudo-periodicity is also responsible for the slow decrease in during the oscillatory regime. Intermittency degrades the mixing properties of the mapping since, during the pseudo-periodic intervals, the dynamics only explore a reduced portion of the available range in x.
5. A Fat-Tailed Invariant Distribution
Much like the standard CLT, the CLT for deterministic systems can be generalized to the situation where the variance of the relevant variable
x diverges [
11]. In particular, this is the case of invariant distributions with a sufficiently slow algebraic decay for large
:
with
. Under the same hypotheses of ergodicity and aperiodicity stated in
Section 2, and assuming for simplicity that
—for instance, due to the symmetry of
around zero—the distribution of the sums
[cf. Equation (
3)] converges to a stable distribution given by the Fourier anti-transform of
, for some value of the dispersion parameter
. The result for distributions with finite variance is re-obtained in the limit
, with
as defined in Equation (
7).
In this section, we give an example of convergence toward a stable distribution different from a Gaussian in the case of a map with a fat-tailed invariant distribution decaying as
for large
(i.e.,
). This specific case has the analytical advantage that the stable distribution predicted by the CLT can be explicitly written out, namely,
which is nothing but the Cauchy (or Lorentzian) distribution. Like the Gaussian, the Cauchy distribution is a maximum entropy distribution, but with a different constraint.
To obtain a deterministic chaotic map with a variable distributed following a fat-tailed function, we use the ad hoc procedure of applying a suitable transformation to a map whose invariant distribution is known in advance. Specifically, we take the Bernoulli map of Equation (
10) with
, for which we know that the invariant distribution is the function given by Equation (
11), and introduce a change of variables that transforms this function into the desired fat-tailed profile. This is formally achieved by defining the two-variable map
where
transforms a variable
u with uniform distribution in
into a variable
with distribution
. By construction, thus, the invariant measure of variable
x in map (
20) is
with
x varying from
to
∞.
By analyzing the behavior of the Fourier transform of
near the origin, it is possible to obtain the dispersion parameter for the Cauchy distribution of sums of independently chosen values of
x, which turns out to be
. Unfortunately, the value of
when the summands are successive iterations of
x in map (
20) cannot be found analytically in an explicit way. However, we have numerically found that, for
, the dispersion parameter again coincides with
to a high precision. This is the value of
that we use to compute the KLD
between the distribution of the sums
of Equation (
18) and the Cauchy distribution (
19). In addition, we do not have a practical procedure to assign a value to the dispersion parameter when the number of summands
N is finite. Therefore, in the present case, we do not calculate a quantity analogous to the KLD
of
Section 3 and
Section 4. Regarding
, due to the non-analytic behavior of the Fourier transform of
at the origin, it is now not possible to use the procedure of
Appendix B to predict how this KLD decreases as
N grows. Our analysis must thus rely on numerical results.
In
Figure 8, we show the distributions
(left column) and
(right column) for three small values of
N. Light-colored curves correspond to the expected asymptotic Cauchy distribution, given by Equation (
19) with
. Note that for
, due to the peak at
, the difference between
and the asymptotic distribution seems to be larger than that of
. The KLDs, however, reveal that
is slightly closer to the Cauchy distribution (see
Figure 9). For
, it is already clear that the approach to the Cauchy distribution is faster for the random sampling sums. Comparison with the results for the Bernoulli and the logistic maps (cf.
Figure 1 and
Figure 4) suggest however that, in the present situation, the convergence to the corresponding asymptotic distribution is considerably slower than for those cases.
Figure 9 presents numerical results for the KLDs
and
. In order to have significant statistics in the construction of the 1000-column histogram that represents
from
samples of the sums
, we have cut off the interval of variation of
to (
), disregarding samples outside that interval. Otherwise, for the fat-tailed distributions involved in the present case, the calculation of the KLDs would be dominated by sampling fluctuations for large values of
. Along most of the range of
N spanned by the figure, both
and
exhibit rather well defined power–law decays. Their different exponents, however, make that they progressively diverge from each other as
N grows, while
approximately decays as
; as illustrated by the full straight line of slope
, a linear fitting of
for
, shown as a dashed line, points to a slower decay with a nontrivial exponent:
. This result suggests that the convergence to an asymptotic distribution for the sums
in the case of fat-tailed invariant measures may generally be characterized by unusual exponents in the decay of the KLD. This conjecture will be thoroughly explored in future work, through both analytical and numerical means.
6. Conclusions
We here analyzed the convergence to the asymptotic probability density distribution of a succession of distributions for a conveniently scaled sum of N samples obtained from iterations of a deterministic map. Previous analytical studies had established that a modified version of the central limit theorem (CLT) exists for these cases. Yet, as far as we know, the convergence to the limit had not yet been characterized. Here, we studied several archetypal examples that expose a variety of ways the limiting distributions are approached.
Our characterization was based on the behavior of the Kullback–Leibler divergence (KLD) between and , in that specific order. With this choice, the KLD equals the number of extra bits required to encode a sample from if the code has been optimized for . The CLT for sums of random samples with finite variance predicts a KLD that decreases as if each sample is drawn from a distribution that is symmetric around its mean value, and as if it is not. This is a bold statement, since an infinitesimal modification may suffice to turn a symmetric distribution into an asymmetric one, so even a minute modification would suffice to change the entire asymptotic behavior of the KLD—the change, however, would only become relevant at increasingly larger values of N, as the asymmetry tended to disappear. We are not aware of an analogous theoretical prediction for the case of correlated samples, but the results presented have revealed similar behaviors: decreased as for the Bernoulli map, for which the sums are distributed symmetrically around their mean value, and as for the logistic map, where the distributions are asymmetric.
In both the Bernoulli and the logistic map, the rates at which approached the asymptotic distribution increased with the strength of mixing. Moreover, for the intermittent logistic map, where mixing is virtually absent during pseudo-periodic intervals, convergence to the asymptotic distribution was particularly slow. Therefore, even though all the explored examples were equally deterministic, their behavior differed considerably. Details in the chaotic dynamics are crucial to the behavior of for large N.
The convergence of in the Bernoulli map could be divided into two stages, one in which the distribution acquired an approximately Gaussian profile, and a subsequent one, in which the variance was adjusted to approach its asymptotic value. Remarkably, in the second stage and for sufficiently large N, the divergence was equal to with , implying that each sample of the deterministic map was as informative about the asymptotic distribution as random samples. This equivalence could not be established in the other explored examples, since in all of them, and decreased with N with different power laws. No re-scaling procedure, hence, could transform one into the other.
The last example involved variables with divergent variance. In this case, the derivation of
Appendix B is no longer valid, and no theoretical formulation describing how
tends to zero is known to us. Our numerical explorations revealed a behavior proportional to
for
, even for samples drawn from distributions that are symmetric around their mean values. The deterministic counterpart
exhibited an even slower evolution, at a rate that is also slower than the one observed in the cases of finite variance.
In conclusion, in all the examples explored here, the asymptotic trend of the KLD behaved as a power law. Different deterministic maps yielded different exponents, displaying a variety of behaviors. The factors that influenced the exponents were (a) the strength of mixing in the chaotic map, (b) the tendency of the system to evolve near periodic orbits, and (c) the tails of the distribution of individual variables. We stress that the open question of establishing a quantitative connection between the rate of mixing, on one hand, and of KLD decay, on the other, remains an interesting problem for future work. Remarkably, except for the logistic map in the intermittent regime, all the maps explored here are related to each other by simple, nonlinear transformations. Despite these deterministic functional relations, their nonlinear nature determines differences in the statistical behavior of the sums of samples drawn from each map, with a large impact on the convergence towards their asymptotic distributions.