Entropic Approach to the Detection of Crucial Events

In this paper, we establish a clear distinction between two processes yielding anomalous diffusion and 1/f noise. The first process is called Stationary Fractional Brownian Motion (SFBM) and is characterized by the use of stationary correlation functions. The second process rests on the action of crucial events generating ergodicity breakdown and aging effects. We refer to the latter as Aging Fractional Brownian Motion (AFBM). To settle the confusion between these different forms of Fractional Brownian Motion (FBM) we use an entropic approach properly updated to incorporate the recent advances of biology and psychology sciences on cognition. We show that although the joint action of crucial and non-crucial events may have the effect of making the crucial events virtually invisible, the entropic approach allows us to detect their action. The results of this paper lead us to the conclusion that the communication between the heart and the brain is accomplished by AFBM processes.


Introduction
In the last 30 years or so, the concept of anomalous diffusion has been widely adopted to deal with processes, ranging from biology to sociology, departing from the conditions of thermodynamic equilibrium that the Boltzmann principle [1] establishes for physical processes. For instance, in 1992, Peng et al. [2] introduced the concept of DNA walks that became a very popular way to study fluctuations of biological processes. They studied nucleotide sequences and assigned the symbol 1 to purines and the symbol −1 to pyrimidines. The position of a nucleotide is thought of as a time and the random walker at that time makes a step forward or backward according to whether the nucleotide is a purine or a pyrimidine. Since a nucleotide sequence is unique and it is not possible to adopt the conventional Gibbs prescription of making averages over many identical copies, the authors of this important paper adopted the method of a moving window of size l. The window moves along the nucleotide sequence and the observer records the space traveled by the walker in the "time" interval l, namely the distance from the position of the walker at time l to the position she had at the beginning of the window, assumed to be zero. In the case of random walk the fluctuations from 1 to −1 and back are totally uncorrelated and the resulting scaling η is equal to 0.5. These authors found that the scaling is larger than 0.5, thereby suggesting that the DNA nucleotides are correlated; or, said differently, the random walk is persistent.
What is the origin of this anomalous behavior? A widely shared conjecture is that the source of this correlation is properly described by means of the Fractional Brownian Motion (FBM) proposed by Mandelbrot [3]. This is a generalization of ordinary Brownian diffusion, yielding the following relation where H is the symbol adopted by Mandelbrot to denote scaling. Ordinary Brownian motion is a singularity of this formula, corresponding to H = 0.5. This relation implies that FBM has memory of the infinitely distant past, since no limit is set on the magnitude of t. However, it has been noticed [4] that if we go beyond this mathematical formalism and adopt a dynamical derivation of FBM from the traditional diffusion equationẋ (t) = ξ(t), (2) yielding in the integral we obtain the auto-correlation function where Φ ξ (|t 1 − t 2 )) is the stationary auto-correlation function ξ(t 2 )ξ(t 2 ) / ξ 2 . With a proper choice of this correlation function, setting t ≡ t 2 − t 1 and sending t, t 1 and t 2 to ∞ has the effect of recovering Equation (1). For this reason, we define this form of infinite memory as Stationary Fractional Brownian Motion (SFBM).
Herein we adopt the symbol H to denote the SFBM scaling and the symbol η to denote the scaling of a differently described form of anomalous diffusion process.
The technological progress allowing the observation of the diffusion of single molecules in biological cells has attracted the general interest for a form of fractional diffusion that we term Aging Fractional Brownian Motion (AFBM) to stress its non-stationary nature [5]. The non-stationarity of this form of diffusion is not a consequence of the physical rules behind diffusion changing with time. However, rather, depends on the occurrence of crucial events that are responsible for the breakdown of ergodicity.
A good way to introduce the readers to crucial events is by adopting the engineering language of [6], which defines the age dependent failure rate g(t) through where Ψ(t) is the survival probability, namely the probability that a machine keeps working for a time interval t from the time at which it was created. Let us imagine that a team of engineers acts the moment a machine fails. They instantaneously correct the machine's ill-functioning making it brand new. This has the effect of extending the working life of the machine to the next failure, when it will again require the instantaneous action of the team of engineers. The time distance between two consecutive failures has the waiting time distribution density This leads to Assuming that g(t) decays in time as 1/t, namely at the limit of integrability, and more precisely according to when inserted into Equation (7) and integrating yields where T = 1/r 1 and µ = 1 + r 0 /r 1 . Equation (9) readily yields the waiting time distribution density given by We define the occurrence of these failures as the crucial events when the inverse power index µ fits the condition 1 < µ < 3. We note that the average waiting time, defined using Equation (10) is given by, when µ > 2, To stress the non-stationary nature of this process, in the whole interval µ < 3, as well as in the sub-interval µ < 2, where the ergodicity breaking is made evident by the divergence of τ , let us assume that the laminar region between consecutive crucial events is filled with either 1 or −1, according to a coin tossing procedure. In this case, if we observe the process with the mobile window prescribed by the authors of Reference [2], with the constraint of locating the beginning of the window where we see an abrupt transition from 1 to −1 or from −1 to 1, we find We adopt the symbol Φ(l), rather than Ψ(τ), to stress the adoption of the mobile window to evaluate the brand-new survival probability. This is a consequence of the fact that when the window of length l overlaps more than one laminar region, different laminar regions may have opposite signs making the survival probability Φ(l) vanish. If we do not adopt the above constraint and we move the left end of laminar region in a continuous way along the sequence, Φ(l) becomes the equilibrium correlation function defined by [7]: Using Equation (11) and inserting Equation (10) under the integral, direct integration yields This is a result of the non-stationary nature of this process that as effect of aging changes the power index µ of Equation (12) into µ − 1. If we try to interpret this result by using SFBM we are immediately led to make the conjecture that the scaling index of Mandelbrot H fluctuates. In fact, the stationary correlation function Φ ξ (t) yielding the surprisingly extended memory of Equation (1) has an inverse power law (IPL) tail with index δ related to H by Consequently, if we identify the exponents of Equation (12) and Equation (14) with a non-stationary δ, we get the non-stationary H given by which, for instance, in the case µ = 3, would change from H = 0 (µ = 3) to H = 0.5 (µ = 2). The main purpose of the present paper is to propose an entropic approach to the analysis of time series that will establish if the anomalous diffusion emerging from the use of the mobile window of Reference [2] is due to the Mandelbrot infinite memory or to crucial events. We also posit suggestions to establish if both sources of anomalous diffusion are jointly acting on the complex process under study. We show that this entropic approach to the analysis of physiological data, on the dynamics of heart and the brain, settles the ambiguity about the 1/ f noise generated by these physiological processes leading to the conclusion that they are driven by crucial events.
The outline of the present paper is as follows. In Section 2 we show that the observation of 1/ f noise does not afford a clear-cut criterion to establish whether SFBM or AFBM applies. In Section 3 we illustrate the entropy concepts that are used in this paper to detect crucial events, by assessing if the experimental signal under analysis is a SFBM or an AFBM. Section 4 illustrates the main result of this paper, namely, how to prove that an anomalous scaling is a manifestation of crucial events. Section 5 affords detail on the method of stripes. In Section 6 we discuss the consequences of the main results of this paper for the dynamics of heart and the brain. Finally, in Section 7 we argue that the results of this paper go much beyond the limit of a single discipline and can be beneficial for psychology and sociology as well as for physiology and biology.

Pink Noise
In the literature there exists general consensus concerning the connection between criticality and 1/ f noise (pink noise). In addition to the original work of Bak and coworkers [8], see also, for instance, the review paper [9] emphasizing the connection between criticality, 1/ f noise and cognition. Here we show that 1/ f noise can be derived from both SFBM and AFBM.

Stationary Fractional Brownian Motion
As pointed in Section 1, SFBM is usually referred to in the literature as FBM. The dynamical approach to FBM [4] yielding Equation (4), generates the following expression for the second moment of the diffusion process According to Mandelbrot, the second moment scales in the following way, x(t) 2 ∝ t 2H , thereby entailing the stationary correlation function Φ ξ (t) to have, for H > 0.5, an IPL tail with IPL index B related to H by In summary, the experimental signal ξ(t), which is the object of our statistical analysis, generates the stationary correlation function where B 1 when H is very close to the maximal value of 1. According to the Wiener-Khinchine theorem [10], the power spectrum is the Fourier transform of the auto-correlation function: Inserting the correlation function from Equation (19) into Equation (20) we obtain where We see that the ideal 1/ f noise is obtained by setting the condition B → 0, forcing H to be very close to the maximum value of 1. Using Equation (18) we obtain the widely used relation

Aging Fractional Brownian Motion
In the case when the process is driven by crucial events with an IPL index in the range 2 < µ < 3, in the long-time limit a stationary correlation function identical to that of Equation (19) is generated. To avoid confusion, let us adopt the following notation and not Equation (19) since we have used the IPL for crucial events and not Equation (17). Please note that in this case, due to Equation (14), We understand the temptation to assume that C = B , yielding This relation is, however, not correct because, as proven elsewhere, crucial events generate [5] the scaling index Please note that when the fluctuation ξ(t) hosts crucial events whose power spectrum has the IPL index: this equation is valid throughout the whole range 1 < µ < 3. In the case µ > 2, Equation (28) is recovered by using Equation (22) with The entropic analysis proposed by this paper, is a significant refinement of the technique of Diffusion Entropy Analysis (DEA) proposed in 2001, [11,12]. The DEA of Reference [11] was proposed as method of scaling evaluation based on observations of the experimental signal with no focus on the detection of crucial events. The DEA of [12] was proposed for the explicit purpose of detecting the scaling generated by crucial events, and was applied to surrogate sequences of crucial events with the explicit rule of forcing the random walker to make a jump ahead with the occurrence of an event. The refined form of DEA proposed in this paper is based on the DEA of [12]. In the case where we must analyze an experimental fluctuation and we do not use the method of stripes, the DEA of [12] is equivalent to the DEA of Reference [11]. Using the method of stripes, we define events, which may be either crucial or not, and we use the DEA of [12] to assess the fraction of crucial events imbedded in the set of events generated by the stripes. We shall see that this procedure affords a method to assess whether 1/ f noise does host crucial events, or does not.

Entropy Concept
The concept of entropy has a thermodynamic origin and is closely connected to the second law of thermodynamics. The well-known expression [1] can be fruitfully used to explain why the free expansion of a gas in a container is an irreversible process [13]. However, the attempts to describe the time evolution towards equilibrium resting on the Gibbs entropy forced investigators to adopt the concept of a Gibbs ensemble average and, consequently, to assume ergodicity [14], namely that an average over infinitely many copies of the same system is identical to averaging a single system of the ensemble over time. At the same time, these attempts led the investigators to assume that chaos is an important ingredient to generate an irreversible transition to equilibrium, even if chaos may not be completely random [15]. The main problem of reconciling the second law of thermodynamics and irreversibility with the reversible nature of both classical and quantum mechanics led the investigators to establish a connection between the second law and information theory [16], as explained in the work of Landauer [17] and in the more recent work of Reference [18]. This reconciliation attempt becomes even harder when we move from quantum physics to the second law insofar as it raises the still unresolved problem of deriving classical from quantum physics [19]. These contributions to the field of non-equilibrium statistical physics, although generating fruitful applications, are based on the Gibbs ensemble perspective and consequently do not shed light into the dynamics of the individual systems of a statistical ensemble, if no recourse is done to the ergodicity assumption.
In the last ten years, however, increasing attention has been devoted to the ergodicity breaking, in two different fields of research. The first field of investigation is molecular diffusion, with the tracking of single molecules in living cells [5], making it possible to do time averages over the motion of a single molecule, and the latter is the field of complex networks, where the discovery of cooperation-induced criticality has proven to yield temporal complexity, namely non-ergodic fluctuations of the complex network's mean field [20]. A reasonable conjecture currently being made is that in both cases non-ergodic behavior is a signature of the transition from a non-cooperative to a self-organized state [21]. The brain is an example of a complex system of this kind and its non-ergodic nature raises the important question of how to measure its complexity using time averages, or, equivalently, how to define the entropy of a single trajectory. We therefore explore the important issue of Kolmogorov complexity [22], which is expected to shed light into the entropy of a single time series and thus on the entropy of a single trajectory, which is called Kolmogorov-Sinai (KS) entropy. We discuss the joint use of compression and diffusion [23], two methods of analyzing of time series based on the KS entropy. While the former procedure establishes the amount of order by the numerical evaluation of algorithmic compressibility of a time series, the latter assesses the amount of order by forcing the time series to generate diffusion, the scaling of which is sensitive to the deviation from randomness.
We assume that complexity is generated by the occurrence of renewal non-Poisson processes, and on the basis of this assumption we propose an approach to calculating the entropy of a single trajectory based on the theoretical perspective of a continuous random walk [24]. Randomness is a property of events occurring in the operational time ψ, related to the clock time t by the relation ψ = t α . This observation leads to a generalization of the Pesin's identity [25] that we propose to adopt to define the entropic complexity of non-ergodic trajectories. We conclude this section arguing that non-ergodic fluctuations may be incompressible despite their vanishing Lyapunov exponent and we discuss to what extent this theoretical perspective may afford a useful technique of analysis of non-stationary time series.

External Entropy
The authors of [26] shed light on the paradoxical macroscopic effects generated by chaotic trajectories exploring regions with different Lyapunov coefficients [15] used the concept of external entropy that was later adopted in Reference [23] to discuss the Kolmogorov complexity in terms of compression and diffusion. To evaluate external entropy for non-ergodic trajectories, we can benefit from the time series generated by the idealized Manneville map [23]. Imagine a particle moving in the interval [0, 1] from an initial condition 1 > y > 0, with uniform probability, according to the dynamical prescriptionẏ with z > 1 and λ << 1. The time τ necessary to arrive at the border y = 1 moving from an initial value 0 < y < 1 is given by with When the particle reaches the border y = 1 it is injected back to a new initial position by randomly selecting a new number y. It is important to notice that the waiting time probability distribution function (PDF) ψ(τ) separating consecutive activations of the external entropy has the same analytical form as the waiting time PDF ψ(τ) of Equation (10). Notice that when µ > 2, we recover the mean waiting time given by Equation (11) with Please note that the limiting case µ = ∞ is realized by replacing the procedure of Equation (31) with which is easily shown to generate the Poisson waiting time PDF ψ(τ) The main remaining question is whether DEA can properly address the condition µ < 2, which is characterized by perennial ergodicity breaking.
The authors of Reference [27] proved that the Kolmogorov-Sinai entropy, η KS , reads: where, according to Equation (32), The land of crucial events ranges from z = 1.5, µ = 1.5, to z = ∞, µ = 1, with η KS vanishing from z = 2 to z = ∞. We can recover the proposal of Korabel and Barkai [25] by the following conjectures on the computational cost. The computation cost, in the case µ > 2 increases linearly in time, while in the case µ < 2 it increases as t α , with In fact, the number of random drawings n of the initial condition y is for µ > 2. This is the condition making it possible to realize a Lévy walk [28,29]. In the case µ < 2, as earlier mentioned, the number of random drawings is This simple heuristic prediction has the effect of defining the incompressibility of the time series also for µ < 2 [25]. In fact, adopting the generalized form of KS entropy proposed by Korabel and Barkai [25], based on observing that for µ < 2, it is necessary to use a new definition of time, taking into account the transition from n proportional to t to n proportional to t α , a singularity appears at z = 2. The generalized h KS of Korabel and Barkai [25] is a decreasing function of z for z < 2 and it becomes an increasing function of z for z > 2. These intuitive arguments can be used to attract attention to the rigorous work of References [30][31][32][33][34]. Here we limit ourselves to point out that informational compressibility implies the existence of a message carried by the time series, thereby implying a connection with cognition.
We emphasize that according to the statistical analysis of the brain in the awake condition [35] the brain is controlled by an AFBM process with µ = 2. This means that the brain operates at the border between the perennial aging condition, µ < 2, and the temporary lack of stationarity condition, µ > 2. The observation of 1/ f noise does not make it possible to realize the singularity of this condition because the IPL index β of the spectrum obeys Equation (28), making the spectrum become that of an ideal 1/ f noise at µ = 2, with β moving in a continuous way from values slightly larger than 1 to values slightly smaller. The singularity of this condition, made evident by the entropic analysis of Korabel and Barkai, should be properly taken into account by the analysis of real physiological data.

Entropic Treatment of the Scale Detection Issue
The existence of a message with a meaning motivates us to move from the Boltzmann to the Wiener/Shannon entropy. Let us imagine that we have a sequence of events and that all of them are crucial. We follow the prescription of Reference [12]. For any crucial event, the random walker makes a step ahead by the fixed quantity 1 and thereby builds up a diffusion trajectory x(t). We then observe this trajectory with the moving window of size l so as to generate the histogram of the PDF p(x, l). According to the generalized central limit theorem [12] we obtain where the power law index is if 1 < µ < 2, and becomes if 2 < µ < 3. Finally, we have if µ > 3.
It is important to stress that these rules in the anomalous case generate the asymmetric Lévy diffusion [12].
To appreciate the main result of this paper, namely, how a refined version of DEA makes it possible to distinguish AFBM from SFBM, it is convenient to discuss DEA in action on an experimental signal ξ(t), when we do not know if the single fluctuations are crucial or not crucial. Let us consider, for instance, the case when the laminar regions between two consecutive crucial events are filled with either 1's or −1's, with a coin tossing prescription, and 2 < µ < 3. This is the celebrated Lévy walk [28,29]. We create the diffusion trajectory We then observe this trajectory with the mobile window of size l. We evaluate the difference between the value of x(t) at the end of the window and the value of x(t) at the beginning of the window. This allows us to create the probability density function In this case, the scaling η is identical to that of the asymmetric Lévy scaling of Equation (43). Let us also discuss the case where the signal does not host any crucial event. If ξ(t) is a generator of FBM the observation of the trajectory of Equation (45) with the method of the mobile window of length l yields the probability density function The main problem with the use of the mobile window of size l [2] is that in the case of crucial events, Equation (41), the long, slow tails of the PDF make their second moment divergent, and the scaling evaluation is affected by the numerical truncation that cannot be easily controlled. The second moment technique works in the case of SFBM because in this case the PDFs are Gaussian, with fast exponential tails.
This difference in the second moment is the reason the adoption of DEA [11,12] turns to be very convenient. In fact, DEA lead us to evaluate the Wiener/Shannon entropy of the diffusion process, namely Inserting Equation (41) into Equation (48) yields where A is a constant. When we insert Equation (47) into Equation (48) and integrate, we obtain where B is a different constant from A. The scaling parameter η is properly defined even if the probability density function of Equation (41) has a diverging second moment. The DEA does perceive the correct scaling η even if the long-time limit leads to numerical statistical inaccuracy, since the crucial events are rare in this limit. However, this use of DEA does not allow us to assess if we are dealing with a SFBM or an AFBM phenomenon. In fact, if η > 0.5 it is impossible to establish with the use of DEA alone if we are observing super-diffusion generated by SFBM or by AFBM.

Refined Diffusion Entropy Analysis
This section is devoted to proving that DEA becomes an efficient detector of crucial events when it is supplemented by the method of stripes and we compare its results to those obtained without using stripes.

Without Stripes
Here we explain why DEA detects crucial events even when the ordinary methods fail. We need to explain why crucial events are invisible to the ordinary methods of statistical analysis. To accomplish this, we create a time series where the crucial events are imbedded in a dense set of Poisson events, and using the conventional method of correlation functions it turns out to be difficult to evaluate the complexity of crucial events. To prove this important property, we generate a surrogate sequence ξ(t) according to the prescription The fluctuation ξ C (t) is obtained by generating the crucial events using Equation (31) and filling the laminar regions with either 1 or −1 using a coin tossing procedure. The fluctuation ξ P (t) is obtained by generating the crucial events using Equation (34) and filling again the laminar regions with either 1 or −1 using a coin tossing procedure. We are studying the case 2 < µ < 3, namely the Lévy walk case, which is characterized in the long-time limit by a stationary correlation function. The equilibrium correlation function of the fluctuation ξ(t) is expected to be given by where the second term is the equilibrium correlation function of the signal generated by crucial events.
In the case where 1 the crucial events generate a weak tail that can be mistaken for noise. In fact, the left panel of Figure 1 shows that after the fast decay due to the Poisson events a weak tail appears. This tail contains information about crucial events, but it is virtually impossible to derive the crucial index 2 < µ < 3 from it. The adoption of DEA without using the method of stripes, on the contrary, makes it possible to derive with sufficient accuracy the value of µ using which is illustrated by the panel on the right of Figure 1. The reason for this important property is that DEA determines the scaling in the time asymptotic limit. The scaling generated by the Poisson events is η = 0.5. The scaling generated by the crucial events for µ > 2 is given by Equation (43). The adoption of the coin tossing prescription for the extended laminar region has the effect of generating a Lévy walk [36]. This is so because in this case we are allowed to use Equation (39) with the Lévy walk generating super-diffusion. It is surprising that despite many changes of sign due to the high density of Poisson events the DEA technique, without the use of stripes, in the long-time region is sensitive only to crucial events. The intuitive explanation is that the crucial events, yielding a diffusion scaling η > 0.5, despite their low density generate a broader distribution density and therefore a predominant contribution to the scaling parameter η.

With Stripes
In this section, we show that the joint use of DEA with the method of stripes and without the method of stripes makes it possible for us to establish if super-diffusion is generated by SFBM or AFFBM. Let us denote with the symbol ξ(t) the experimental time series under study. Rather than converting it into the diffusion process x(t) directly, we follow the method adopted by Allegrini et al., 2002 [37]. We divide the ordinate axis into bins of size s and record the times at which the experimental signal crosses the axis. In this way we obtain a time series {t i }. At any of these times an event occurs. We do not know if this event is crucial or not. We replace the experimental fluctuation ξ(t) with a time series z(t) defined as follows. If time t coincides with one of the times t i , we set z(t) = 1 and z(t) = 0 otherwise. In other words, if the time t corresponds to the occurrence of an event, the random walker makes a step ahead by the fixed quantity 1. The diffusion trajectory x(t) is obtained from Figure 2 refers to the case where H = 0.6, namely, the case of anomalous super-diffusion due to SFBM. The right-hand panel of Figure 2 shows that the ordinary DEA with no stripes yields η = H > 0.5, thereby correctly recognizing super-diffusion. The left-hand panel of Figure 2 shows that the use of stripes yields η = 0.5, thereby killing the contribution of SFBM to super-diffusion. This is the important property that we use to establish if the anomalous scaling is due to AFBM or to SFBM.
The theoretical reason for this result is connected to the recrossing of the origin of FBM ξ(t). According to Sinn [38] the time distance between consecutive origin crossings is described by the exponential waiting time survival probability We make the plausible conjecture that the exponential structure is a general property of the FBM crossing. The method of stripes generated many events that can be approximately interpreted as Poisson. Of course, the crossing is characterized also by memory that generates a weak deviation from ordinary scaling.
What about embedding crucial events into a dense set of non-crucial events generated by a SFBM? To answer this question, we create the surrogate sequence where ξ FBM (t) is generated according to the Mandelbrot algorithm of Reference [39]. In this case, it is necessary to exorcise the misleading influence of SFBM and we must use DEA with the method of stripes, which is proved by Figure 2 to kill the anomalous scaling generated by SFBM. In this case, rather than filling the laminar regions between two crucial events with either 1 or −1 , as done for Figures 1 and 2, we adopt a different method that will make it easier for us to use the detailed illustration of the method of stripes of Section 5, to show that this method is only sensitive to crucial events. When a crucial event occurs, we assign to it an intensity proportional to the time distance from the occurrence of earlier crucial events. To explain the effect of this procedure on the scaling evaluated using the ordinary method of ensemble average we must make an excursion to the earlier work of References [40,41]. The intensity of the fluctuations has a divergent second moment, thereby yielding the diffusion scaling However, the method of stripes depends only on the statistics of crucial events and it is independent of the anomalous intensity of the crucial events as well as of SFBM. All this is described by Figure 3, where the left panel refers to the time series of Equation (51) and the right panel to the time series of Equation (56). We see that the stripes kill the anomalous SFBM scaling, H = 0.7, which is larger than the AFBM scaling, η = 0.625. It is interesting to notice that the density of non-crucial events in both figures is the same, and the scaling detected by DEA with stripes is in both cases the scaling generated by crucial events. In other words, the method of stripes makes the AFBM fluctuations equivalent to Poisson fluctuations. Finally, we detect the time distances τ i between two consecutive crossings of the origin for both the case (51) and (56). We evaluate the cross-correlation function This correlation function is stationary, namely, C ij = C(t), with t ≡ |i − j|. In the scale of Figure 4 the integer variable t is virtually equivalent to a continuous variable. Figure 4 shows that when the non-crucial events are a form of SFBM the tail of C(t) reveals the existence of a SFBM process to the time series under study. The analysis of heartbeats of References [37,42] in addition to finding the correct crucial scaling of heartbeats, corresponding to the parameter µ slightly larger than 2, with the method of stripes, yields for C(t) the slow tail that is a signature of SFBM.

Details on the Action of the Stripes
We devote this Section to more details on the method of the stripes. We afford details in the case of the results of Figure 3. This is expected to help the readers to also get a better understanding of the method used to get the results of the left panel of Figure 2. In Figure 5 we show the fluctuation ξ(t) generated by Equation (56), which is the combination of crucial and non-crucial events adopted by us the generate the diffusional trajectory of Equation (54).
The ordinate axis is divided in many stripes of size s = 1. The time at which the signal ξ(t) moves from one stripe to one of the two nearest neighbor stripes is recorded as the time of occurrence of an event, which can be either crucial, if it depends on ξ C (t), or not, if it depends on ξ FBM (t). In the case of fluctuations of intensity much larger than the size s of the stripes many events occur at the same time and they are recorded as a single event. This explains why the large intensities of the crucial fluctuations do not influence the scaling, but only their temporal occurrence does.
The adoption of the always stepping ahead rule [12] to define the diffusional trajectory x(t) of Equation (54) is essential to making the method of stripes efficient. In fact, if all the events are crucial, this diffusion process generates the scaling η = 1/(µ − 1), which is larger than 0.5 for µ < 3. If all the events are not crucial, η = 05. When the events detected using the stripes are a mixture of crucial and non-crucial events, the long-time limit of the diffusion process is dominated by the faster scaling of crucial events. This explains why the adoption of the method of stripes makes the DEA so efficient for the detection of crucial events.
The fluctuation ξ(t) of Figure 5 is too dense to see the single events. Therefore, we invite the readers to look at Figure 6. Also Figure 6 refers to the analysis of the surrogate sequence of Equation (56) and to the results of Figure 3. However, the realization ξ(t) of Figure 6, is of size 100, much shorter than the realization of ξ(t) of Figure 5: it is short enough as to illustrate the details of this fluctuation. The times at which the signal ξ(t) crosses one of the border lines are recorded to signal the occurrence of an event, which can be either crucial or not, according to the component of ξ of Equation (56), either ξ C (t) or ξ FBM , determining the crossing. We convert these events into the diffusional trajectory x(t). This is shown in detail in Figure 6.

Criticality and Physiological Processes
Self-organized criticality (SOC) and the connection between SOC and 1/ f noise have been the subject of sometimes heated debates since the publication of the original work of Bak and coworkers [8]. The connection between criticality and 1/ f noise has been emphasized [43,44] and questioned [45,46], and it remains an open question especially because whether or not 1/ f noise itself has a unique origin is still unsettled.
On one side we have proposals that advocate to some extent the adoption of SFBM arguments [47,48] and introduce the adoption of multifractality to take into account the fluctuations of H. On the other side there are approaches to 1/ f noise in systems with ergodicity breaking [49] including the extreme case of perennial aging [50] with µ < 2. The analysis of the brain dynamics of Reference [35] led these authors to the conclusion that the brain in the awake state is a generator of an ideal 1/ f noise, corresponding to µ = 2.
Another important physiological process is the dynamics of the heart and significant interest exists on the correlation between heart rate variability and brain activity [51][52][53]. However, the theoretical explanation of the correlation between these two important physiological processes remains a difficult problem due to their different frequency scales.
The main theoretical result of the present paper allows us to reach some compelling conclusions about the nature of these two physiological processes. In fact, the heartbeats have been analyzed with the method of the stripes [37,42] leading the conclusion that the IPL index µ in the case of healthy individuals is close to the condition generating ideal noise, µ = 2. In a more recent paper, Bohara et al. [54] using the method of the stripes analyzed the EEGs of healthy subjects in the same awake conditions as the human subjects of Reference [35] and confirmed that they have a power law index µ very close to the ideal condition µ = 2. They also established a bridge between waves and crucial events leading to the important conjecture that for the brain-heart correlation the tuning of frequencies is not as important as the tuning of temporal complexity, namely the brain and heart dynamics sharing the same IPL index µ.
On the basis of the present results, we conclude that the frequencies of the brain wave may affect the brain physiological process with SFBM contributions that are killed by the DEA analysis resting on the adoption of the method of stripes and that the form of 1/ f noise generated by the brain in the awake condition is due to ASBM, thereby supporting the conjecture that the brain-heart correlation is a consequence of complexity matching [55], mainly resting on tuning temporal complexity rather than frequencies.
It is important to stress that the theoretical perspective of this paper affords strong support to our conviction that crucial events play a fundamental role for cognition, without ruling out coherence. The authors of the recent work of Reference [56] found that meditation has the remarkable role of enhancing coherence. We believe that SFBM is closely related to coherence. On the other hand, the earlier work of Reference [37], where the method of stripes has been used for the first time, suggests that crucial events are imbedded in a cloud of Poisson events, ruling out coherence-induced SFBM. This paper shows that the crucial events may be imbedded in a cloud of SFBM fluctuations and that also in this case, thanks to the method of stripes, the action of crucial events can be revealed. The right panel of Figure 4 illustrates SFBM effects that the DEA with stripes does not perceive.

Concluding Remarks
We believe with the authors of the recent paper [57] that 1/ f noise is a manifestation of Self-Organized Temporal Criticality (SOTC) [58], a new form of self-organization generating crucial events and, consequently, the pink noise generated by an AFBM process. There is general agreement that the emergence of life is signaled by a departure from the ordinary scaling η = 0.5 [2]. Whether this departure from ordinary statistical physics rests on SFBM, or AFBM, or a joint action of both processes is still a subject of research and possible debate. The DEA approach using the method of stripes illustrated in this paper will contribute to progress in this field of research, which is not limited to applications in biological and physiological processes. The discussion on the origin of cognition moves from the elementary biological process, for instance cell to cell communication [59][60][61], generating the conjecture that this is based on quantum mechanical coherence, one form of non-commutative probability theory, a theoretical perspective that we believe to be related to SFBM processes on the basis of the fact that the dynamical approach to FBM [4] was derived by the quantum mechanical theory of quantum dissipative processes of Weiss [62]. We think that at this elementary biological level crucial events, namely form of compressible randomness, are generated by SOTC. This fundamental problem goes far beyond elementary biological processes. In fact, the fast and slow thinking aspects of decision-making processes [63,64] establish a complex interaction [65] between the brain of the single individuals and the society networks.
Author Contributions: Numerical and computational work by G.C.; Supervision by P.G. and B.J.W.; Draft preparation by all authors.
Funding: G.C. and P.G. warmly thank ARO for the financial support of this work through grant W911NF1901.

Conflicts of Interest:
The authors declare no conflict of interest.

Abbreviations
The following abbreviations are used in this manuscript: