Market Efficiency , Roughness and Long Memory in PSI 20 Index Returns : Wavelet and Entropy Analysis

In this study, features of the financial returns of the PSI20index, related to market efficiency, are captured using waveletand entropy-based techniques. This characterization includes the following points. First, the detection of long memory, associated with low frequencies, and a global measure of the time series: the Hurst exponent estimated by several methods, including wavelets. Second, the degree of roughness, or regularity variation, associated with the Hölder exponent, fractal dimension and estimation based on the multifractal spectrum. Finally, the degree of the unpredictability of the series, estimated by approximate entropy. These aspects may also be studied through the concepts of non-extensive entropy and distribution using, for instance, the Tsallis q-triplet. They allow one to study the existence of efficiency in the financial market. On the other hand, the study of local roughness is performed by considering wavelet leader-based entropy. In fact, the wavelet coefficients are computed from a multiresolution analysis, and the wavelet leaders are defined by the local suprema of these coefficients, near the point that we are considering. The resulting entropy is more accurate in that detection than the Hölder exponent. These procedures enhance the capacity to identify the occurrence of financial crashes.


Introduction
The purpose of this paper is to analyze two main issues concerning the Portuguese Index PSI20 daily returns, in the period from 2000 to 2013.The first issue is market efficiency (see Kristoufek and Vorvsda [1,2]).A market is called efficient if prices are adjusted so that they reflect the new information (see Fama [3] and Samuelson [4]).It is assumed that no investor can predict any information that is not already available in prices.As for the correlation structure of the series, there should be neither long memory nor local persistence or anti-persistence reflected in less rough or rougher paths, respectively.As a consequence, three aspects of the series are symptoms of efficiency: unpredictability, no long memory and the roughness of the series path (irregularity).These will be evaluated for the PSI20 series to check for the existence of deviations from efficiency.
Long memory reflects a tendency for a slow decay of the magnitude of the time series correlations as a function of lag size, but still preserving stationarity.In a strict sense, it is defined as reflecting a trend-like behavior, that is, persistence (positive long memory).In a broader sense, it may reflect either persistence or anti-persistence, that is, a switching behavior more pronounced than that of a random process (negative long memory); see [5].Roughness is a time series feature that describes its tendency for an erratic behavior with frequent and heterogeneous changes.
While long memory, measured by the Hurst coefficient, H, is a global characteristic of the series, roughness, measured by the fractal dimension, D, is a local one.For a self-affine process, given by: it is verified that D + H = 2 (see, for instance, [1]).In these processes, the global long memory characteristic of the series is a reflection of its local roughness characteristic.
A generalization is given by multifractal processes, given by: E[X(t)] = c(q)t τ (q)+1 where τ (q) is a concave function.In a monofractal process, we have: τ (q) + 1 = Hq, so that a linear scaling is attained.The definition of entropy as a measure of uncertainty, or lack of information, is used not only to measure unpredictability, but also to reflect indirectly the degree of roughness in the path of the series.
As an alternative in the analysis of market efficiency, we will use the concept of the q-triplet, created by Tsallis [6] in the context of nonextensive statistical mechanics.It is used for the characterization of non-integrable dynamical systems, where all Lyapunov coefficients vanish.It consists of a threefold determination of an entropic coefficient, q, in the context of: (1) the sensitivity to initial conditions in a dynamical system, which may be seen as reflecting unpredictability; (2) relaxation of macroscopic variable towards a stationary state, which in a time series, may be taken to detect the existence of long memory; (3) the stationary distribution obtained after the constrained optimization of an entropy function, where the departure of this distribution from the Gaussian distribution is the result of self-organization in the market leading to a rougher path.For several empirical approaches, see Pavlos et al. [7], Ferri et al. [8], de Freitas et al. [9], De Sousa and Rostirolla [10] and, in the finance area, Queirós et al. [11].
The second issue is local regularity, which is related to the identification of crisis events.Financial time series evolve showing patterns, such as time varying volatility and abrupt changes.Irregularity presented by a signal gives information about its behavior.The characterization of irregularity is given by the quantification of the local regularity of a function, f .The pointwise Hölder exponent is one of the quantifiers proposed to measure the local regularity; in fact, a low pointwise Hölder exponent reflects a highly irregular path around the point, whereas a high pointwise Hölder exponent is related to a smooth behavior.Jaffard [12] proposed a characterization of the local regularity using the wavelet coefficients obtained from the wavelet decomposition of the signal; see also [13].Rosenblatt et al. [14] studied the local regularity of a time series applying an entropy measure based on the wavelet leaders.
The rest of this paper is organized as follows.In Section 2, we present the methods for studying market efficiency, using the concepts of long memory, unpredictability and path roughness (Section 2.1) and the concept of q-triplet (Section 2.2).The treatment of the wavelet approach for measuring irregularity is given in Section 3. Results are presented in Section 4. We provide a brief conclusion in Section 5.

Market Efficiency
Analysis of the existence of efficiency in the financial markets is an important issue in financial analysis.A capital market is considered efficient if prices, at each moment, reflect all relevant information.Three types of efficiency are considered according to the degree of information incorporated: weak efficiency if information includes only past prices; semi-strong efficiency if information includes also the information publicly available in the market; and strong efficiency if it includes all information, public or private.

Three Markers for Market Efficiency
We analyze three aspects of the return series, which result from the existence of efficiency.The empirical testing of those aspects correspond to the check for the existence of efficiency.
The assurance of efficiency in the market is attained by an automatic elimination of arbitrage opportunities.The absence of arbitrage implies that there will be no long memory, described as a power decay of correlations, in the return series-the first aspect that characterizes the return series.As a consequence of the absence of arbitrage opportunities, it is expected that the price series are modeled as a martingale and the corresponding returns as a martingale difference that is unpredictable-the second aspect of the return series.This implies an erratic behavior of prices, which is quantified as a high degree of roughness-the third aspect of the return series.

Long Memory
A long memory process was originally defined and motivated, respectively, by McLeod and Hipel [15] and Hall [16], as a stationary process for which autocorrelations absolute values are not summable in the discrete case, that is: where ρ(h) is the lag, h, autocorrelation.
In what follows, we assume that the parameter, H, verifies 1 2 < H < 1.The restriction H < 1 assures that the process is stationary.Additionally, imposing H > 1  2 assures that the process is persistent (for H = 1  2 , there is short memory, while for 0 < H < 1 2 , there is long memory, although negative; see [5]).We retain the case of positive long memory, which is the one usually observed in the financial time series representing volatility.
An alternative definition of a long memory process is attained by the asymptotic characterization of autocorrelations (see [17]): where 1 is a slow variation function (that is, a measurable function, which is positive in a neighborhood of ∞ and for which: , and H is called the Hurst coefficient.We can see that the autocorrelation function has a slow decay, as a power law function.In the discrete case, this implies the non-summability of autocorrelations.
A third alternative definition may be stated in the frequency domain; see [18]: where 2 is a slow variation function, λ denotes a frequency and f is the spectral density.Therefore, the spectral density tends to infinity as the frequency approaches zero.

Remark 1. (i)
There is no equivalence between this and the precedent definition, but it is implied by it if 1 is almost monotone, that is: A last definition of a discrete time long memory process, {y t }, based on Wold's decomposition: where (Ψ k ) 2 < +∞ and {ε t } is a white noise, states that: 3 being a slow variation function.
(iii) If H = 1 2 , we have a short memory process; if H ≥ 1, the process is non-stationary.Therefore, the long memory case is an intermediate case between these two.For H ≤ 1 2 , we have an antipersistent behavior.
(iv) Long memory is characterized by a power law asymptotic behavior, where the Hurst coefficient plays a prominent role.
In alternative definitions of long memory, we consider asymptotic relations based on power functions, either in the time or in the frequency domain.By applying a logarithm transformation on the variables, we obtain linear regressions, which may be fitted by a least squares approach.This is the mechanism taken to build several estimation methods for the Hurst coefficient, H.
Considering a long memory process in the form , where {ε t } is a white noise, u = E(y t ), λ denotes a frequency and L is the lag operator.Taking logarithms: . Therefore, we have the linear regression: where β = ln f ε (λ) and η f (T ) is the number of frequencies considered.This is the regression used to obtain the Geweke and Porter-Hudak estimator [19].The periodogram estimator is obtained after (3), which allows one to obtain a linear regression of ln f y on ln λ, for frequencies λ near zero, with the slope given by 1 − 2H.
The R/S (range over standard deviation) statistic, originally proposed by Hurst [20], is given by: where y and S T are, respectively, the sample mean and the sample standard deviation.This is an increasing measure of long memory, and for i.i.d.Gaussian random variables, y t , we have: where {V t } is a Brownian bridge and "⇒" stands for weak convergence.Lo [17] proposed a more robust version, which assures the convergence for short memory processes; here, S T is substituted by the long run Newey-West standard deviation [21].Estimation based on the R/S statistic is obtained after a linear regression of ln Q T on ln T , where T is the size of the subsamples used to estimate the R/S statistic.This regression has slope H.Other approaches may be taken, for instance, the Whittle estimator [22] is obtained by maximum likelihood on the frequency domain, considering frequencies near zero.

Fractal Dimension
The fractal dimension is a measure of roughness, and in opposition to the Long memory characterization, it measures the local memory of the series (Kristoufek and Vorvsda [1]).When modeling the dynamic behavior of a variable, by the solution of deterministic equations, the set of all instantaneous states of the system is the phase space.The subset of the phase space towards which the system converges, called the attractor, may be a fractal (Theiler [23]).A fractal is an irregular geometric form, the patterns of evolution being similar at different time scales (self-similar).In this context, the fractal dimension of an attractor measures the number of degrees of freedom of the system.When the self-similarity of the geometric form through scales is not perfect, it is called statistical.The fractal dimension can be obtained as an exponent of a scaling behavior of a quantity, measuring the bulk of an object (here, bulk may correspond to the mass of the object) with respect to another measuring the corresponding size (linear distance): bulk∼ size dimension (see Theiler [23]) from which we obtain: so that the dimension is given by: This is a local quantity from which a global definition of the fractal dimension can be found by averaging.
(1) The classical box-counting dimension is defined as follows.We take a partition of the state space in a grid where each box has size ε.Then, count the non-empty boxes (that is, those containing points attained by the attractor).The scaling of this counting number, N (ε), with respect to size ε leads to dimension: which is an upper bound of the Hausdorff dimension (under weak regularity conditions, they coincide).This definition is global, since the bulk, 1 N (ε) , is the average proportion that each non-empty box has of the whole fractal.
(2) The Hall-Wood estimator is a version of the box-counting estimator, this being obtained from (10) (see Gneiting et al. [37]).In fact, considering the boxes of size (scale) ε that intersect with the linearly interpolated data graph {(t, X t ) : t = i n , i = 0, 1, . . ., n}, there are N (ε) such boxes with a total area of A(ε) ∝ N (ε)ε 2 .Therefore, the dimension is given by: The Hall-Wood estimator is based on an ordinary least squares regression fit of ln A( n ) on ln n : where n + 1 is the number of observations and takes the values given by: and n denotes the greatest integer smaller or equal to n , s = L =1 s L and s = ln n , L ≥ 2. It is recommended that L = 2, so that bias is minimized: (see Hall et al. [24]).
(3) The Genton estimator is based on the variogram given by 2γ 2 (t) where: for which we have γ 2 (t) ∝ |ct| α , as t → 0. The graph of a sample path has the fractal dimension given by D = d + 1 − α 2 , where d is the dimension of the considered random vector.When we study a single random variable, we have d = 1, so that: The ordinary least squares regression fit of: on ln( n ) leads to the following estimator for α: Remark that , L, s and s are defined as in (12).Substituting ( 16) into (15), we obtain the variogram estimator for the fractal dimension: The mean squared error of the estimator is minimized for L = 2, so the following estimator is chosen: When the method of moments estimator V 2 ( n ) is replaced by the highly robust variogram estimator proposed by Genton [25], we have the Genton estimator for the γ 2 (h), γ 2 (h) = ( 2 ), (h ∈ d ), where: which is approximately equal to 1  4 for large N h , V (h) = X(t + h) − X(t), N h is the number of points (x i , x j ), such that {(x i , x j ) : x i − x j = h}, and {•} (k) denotes the k-th quantile of the quantity inside the brackets.

Approximate Entropy (ApEn)
As Kristoufek and Vosvrda [1] state, entropy can be seen as measuring the complexity of the system, so that the greater the entropy, the greater the randomness.Approximate entropy was introduced by Pincus [26].First, we consider a time series u(1), . . ., u(N ) of observations equally spaced in time.Then, we fix the parameters, m and r, where m (integer) is the length of runs of data considered and r is an upper threshold for a distance defined below.A sequence: x(1), . . ., x(N − m + 1) is built by making x(i) = [u(i), . . ., u(i + m − 1)].Then, approximate entropy (ApEn) is defined as: where: being: and d a distance between x and x * , given by: As referred to by Pincus et al. [27], a heuristic interpretation of ApEn is that it measures the logarithm likelihood that runs of patterns that are close for m observations remain close on the next incremental comparisons.Note that ApEn may be written as: Typically, it is chosen m = 2 or m = 3 (these values aim to obtain a good estimation for the conditional probability measured by the ApEn), the number of input data points, N , between 10 m and 30 m , and the parameter, r, depends on the application (see [27]).
Remark 2. This entropy is related to the more abstract Kolmogorov-Sinai entropy, given by: Remark 3. Pincus and Kalman [28] point out that the irregularity or unpredictability of the time series is another way by which it may deviate from constancy, as an alternative to volatility, which refers to the magnitude of variations from observation to observation.

q-Triplet
The concept of the q-triplet created by Tsallis [6] arose in the context of nonextensive statistical mechanics for the characterization of non-integrable dynamical systems, where all Lyapunov coefficients vanish, concerning: (i) sensitivity to initial conditions; (ii) relaxation of macroscopic variable towards an anomalous stationary state; and (iii) the stationary distribution obtained after the constrained optimization of an entropy function.These three aspects lead to a threefold determination of the entropic coefficient, q, and are related to the three aspects of efficiency referred to before, respectively: (i) unpredictability; (ii) long memory; and (iii) roughness.

q-Sens
This indicator allows one to stress the power-law sensitivity to initial conditions (Lyra and Tsallis [29]).This sensitivity represents the deviation of two initially nearby paths: In the exponential deviation case, we have, ξ(t) ∼ e λ 1 t , where λ 1 is the Liapunov exponent.The power-law sensitivity to initial conditions is given by: It is a generalization of the classical exponential case; the limit case as q → 1 recovers the exponential sensitivity.The expression for ξ(t) is the solution of dξ dt = λ q ξ q , while, in the exponential case, it is the solution of dξ dt = λ 1 ξ.The entropic index, q, is expressed as a function of the fractal scaling properties of the attractor.These properties are expressed through the multifractal formalism.In this context, we take for each scale a partition of the attractor with N boxes for which a probability measure is defined (p i is the probability attributed to box i, given by the proportion of points of the path in box i).As N → ∞, we have, for a generic q, a subset of boxes visited by the trajectory (at least once) for which: the number of such boxes, N q , scales as N q ∝ N f (q) ; the partition function, χ q (N ) = N i=1 p q i , scales as χ q (N ) ∝ N −τ (q) ; the content of each box is roughly constant and scales as p q ∝ N −α(q) .Note that α(q) is the local Hölder exponent in the scaling relation between the probability, p q , and the size of the box.Then, we have: so that N −τ (q) = N f (q)−qα(q) , and we obtain the Legendre transformation: Considering f * (α(q)) = f (q), f * is defined as the multifractal spectrum, i.e., the fractal dimension of the subset of the boxes with Hölder coefficient α, that is the subset of boxes whose number scales with N as N f * (α) .Note that τ (q) is equal to (q − 1)D * q , where D * q is the generalized fractal dimension of Renyi; see [23]: and ε is the scale.We take the α values at the end points of the multifractal spectrum: α min = α(q = +∞) (resp.α max = α(q = −∞))) is associated with the most concentrated (resp.rarefied) region of the set.Our goal is to measure the power-law divergence of nearby orbits.Let B be the number of time steps over which the set of points in the attractor are generated.The measure on the i-box is roughly 1  B , and the typical size of a box in the most concentrated (resp.rarefied) regions in the attractor is +∞ (resp.−∞ ).Note that for a given q, we have p = B −1 .Then, recalling that α is the exponent in the scaling relation between the probability of a cell and its size, we have: and consequently: Therefore, ln ∝ − ln B α .Taking into account that the smallest splitting between two nearby orbits is of order +∞ , and it can become at most a splitting of order −∞ , we can express (24) as: On the other hand, after (28), we have: Finally, 1 1−q = 1 α min − 1 αmax , which is the relation that allows one to obtain q sens after the multifractal spectrum.The power-law sensitivity to initial conditions may be seen as a mechanism generating a certain degree of uncertainty, associated with the divergence of nearby orbits.

q-Rel
This indicator is defined in the context of the relaxation of an observable variable, Z t , towards a stationary state.The variable, Ω(t), is defined as: which behaves as a function of time t: Ω(τ ) e −bτ q rel , where e x q is the q-exponential function given by e x q = [1 + (1 − q)x] 1 1−q .Then, where ln q x = x 1−q −1 1−q is the inverse function of e x q .In a time series context, the relaxation variable is the autocorrelation function of Z given by: To estimate q rel , we fit the regressions of ln q C(τ ) on τ , for each of the values of q in the interval [1, 1.5] with δ q = 0.01, and choose the q-value for which the corresponding coefficient of determination is maximized (see Ferri et al. [8]; Pavlos et al. [7]).A high value for q rel is a symptom of long-range memory.

q-Stat
This is the q parameter associated with a probability distribution, which arises by maximizing the Tsallis entropy S q = k 1− [P (x)] q dx q−1 (continuous version) under some adequate constraints (see Tsallis [6]).In the original context, considered by Tsallis (continuous version), these constraints are: (the normalizing condition, so that we have a probability distribution) and: (the mean value under the escort distribution [P (x)] q [P (x)] q = P q (x) is known to be µ q ).In this case, a qexponential distribution is obtained; it is a generalization of the standard exponential distribution, which arises when we make q → 1.
In the financial context (see Queirós et al. [11]), it makes sense to add a constraint concerning the variance under P q (x): In this case, we attain the q-Gaussian distribution, with density function: where: and β q = [(3 − q)σ 2 q ] −1 .In the limit case, q → 1, we have the standard Gaussian distribution.
Remark 4. When maximizing the classical Boltzmann-Gibbs-Shannon entropy, we obtain the Gaussian distribution.
Remark 5.In the context of information theory, the entropy measures the uncertainty associated with the variable, X (see Daroneeh et al. [30]).The Tsallis entropy, which is given, in the discrete case, by: is nonextensive, that is, it does not, in general, verify the additivity axiom: where P and Q are probability distributions P * Q : p i * q j , i = 1 . . ., m, k = 1, . . ., m.This relation is verified in the classical Boltzmann-Gibbs-Shannon entropy: the limiting case of S q as q → 1.
In the general case, (q = 1), we have: Darooneh et al. [30] interpret this nonextensive case as reflecting the incompleteness of our knowledge represented by the escort distribution given by: P q i N i=1 P q i .Remark 6. Queirós et al. [11] refer to the fact that returns r follow a q-Gaussian law if its underlying dynamics are represented by the stochastic differential equation: where W t is a Wiener process and p(r, t) the probability density function of r.The deterministic term reflects a mean-reversion mechanism, while the stochastic term reflects, for q > 1, the inverse relation between volatility and the density of the returns, so that the occurrence of rare returns (with high magnitude) causes higher instabilities in the market.
By reparametrization, it can be seen that the q-Gaussian distribution is, in fact, a t-student distribution where, n being the (non-integer) degrees of freedom: Parameters can be estimated by maximum likelihood or minimizing the mean square deviation between this distribution and the empirical distribution (see Cortines et al. [31]).
Alternatively, the following procedure may be applied.The range of values for X is subdivided into small cells of width δx centered at x i , and we find the relative frequency of each cell.We estimate the probability distribution for x, p(x i ) through the properly normalized histogram.For 1 < q < 3, we Therefore, we have for c = √ βq Cq and z = −β q (x − µ q ) 2 .
where ln q (x) = x q −1 1−q .Taking a grid of values for q, with δ q = 0.01 for q ∈ [1, 1.5], we obtain the best linear fit of ln q (p(x)) over (x − x) 2 (that is, the one with a higher coefficient of determination).Then, we select the β-value, which minimizes The q-Gaussian distribution may be seen as the one associated with a stationary state.The q-statcoefficient reflects the sensibility of the volatility to the occurrence of higher variations, such as crashes, resulting from self-organization in the market (see Pavlos et al. [7] and Ferri et al. [8]).
Remark 7. sde (38) expresses a mechanism to explain the clustering of high volatility based on a leverage effect, which may be linked to the fat tails of the q-exponential distribution.On the other hand, underlying this mechanism is the idea that markets have some self-organization, which causes a "roughness" on the financial time series.

Wavelets
A wavelet transform is a possible representation for a time series, so that the information given by the data can be captured in a clarified way.In order to capture the features of the time series, a basic function, called the mother wavelet, is used.The mother wavelet is shifted and stretched, so that the different frequencies, at different times, can be revealed and the events that are local in time are captured.This enables the wavelet transform to study nonstationary time series.
Wavelets can be used to decompose a time series showing its different components.The analysis using wavelets converts the original signal into different domains with different levels of resolution, so that the time series can be analyzed and processed.In fact, while Fourier transforms decompose the signal as a linear combination of sine and cosine functions, the wavelet transform explains the signal as a sum of flexible functions that allow a localization in frequency and time.Depending on the purposes of the study, we have different wavelet transforms: continuous and discrete.In particular, the discrete wavelet transform (DWT) allows one to decompose a time series, originating a set of coefficients that are obtained using the shifted and stretched versions of the mother wavelet.The DWT of a time series can be a way to represent a signal using a small number of terms.General references on wavelet transforms include, among others, [32][33][34][35][36].
The multiresolution pyramidal decomposition allows one to decompose a signal into detailed and approximated signals.The detailed signals express the high frequency components, while the approximated signals express the low frequency components.In order to check for the regularity, we should consider an orthogonal decimated discrete wavelet transform with fast decay derivatives and an appropriate number of vanishing moments.
In order to quantify the local regularity of a function, f , we can use, among others, the pointwise Hölder exponent.If we have a low pointwise Hölder exponent, it means that there is high irregularity; on the other hand, a high Hölder exponent is related to the smooth behavior of the function.
Jaffard [12] proposed a new way to characterize the regularity variation of f through the local suprema of the wavelet coefficients; this information is summarized in the wavelet leaders coefficients.
Consider Ψ 0 , a real valued function with compact support and: Define the number N ≥ 1, such that: (1) t k Ψ 0 (t)dt = 0, ∀k = 0, 1, 2, . . ., N − 1 N is the number of vanishing moments of Ψ 0 .Let us consider translations and dilations of Ψ 0 : The set {Ψ j,k (t) : j ∈ N, k ∈ N} forms an orthonormal basis of L 2 ().Given a signal, X(t), t ∈ [0, n[, the wavelets coefficients, C X (j, k), are given by the inner products: and the signal can be written as: Assuming that Ψ 0 (t) has a compact time support, let us consider the interval I j,k = [k2 j , (k + 1)2 j [ and the union of the three adjacent intervals: giving the path of daily log prices changes.

Market Efficiency
Fractal dimension results are presented in Table 1, for the log prices series and the corresponding estimated values of the integrated variances.These are estimated through the realized variances (RV ), which for log-prices are given by: We consider different approaches, such as Hall-Wood, Genton and box-counting estimators (see [23][24][25]37]), applied to the PSI20 log-prices and to the estimated series of integrated variance.The reference value for the fractal dimension is 1.5, which stands for the absence of either local persistence or local anti-persistence.If the fractal dimension is greater (resp.smaller) than 1.5, it means that there is local anti-persistence (resp.persistence), and the series path is rougher (resp.less rough) than in the reference case.
For log-prices, the values we found are slightly less than 1.5, indicating that there is no significant local persistence.As for the estimated series of integrated variance, we obtain values close to unity, which means that there is strong local persistence.
The Hurst coefficient was computed with different approaches, such as Geweke Porter-Hudak, periodogram and R/Sestimators; see [17,19,20,38].We considered the PSI20 returns and the square of these returns; the results are shown in Table 2.
The reference value for the Hurst coefficient is 0.5, which stands for the absence of either positive long memory or negative long memory.If the Hurst coefficient is greater (resp.smaller) than 0.5, then there is positive (resp.negative) long memory in the series.In our data, there is long memory in the squared return series (representing volatility), but not in the returns series (as observed in general, in the empirical literature; see [39][40][41]).
In Figure 2, we compare the approximate entropy from the returns series to the approximate entropy from a white noise process, for m = 2 and r = 1.The nearness of the curves indicates that PSI20 returns have a high degree of unpredictability.For the q-triplet, the reference value is one (see, for instance, [6,9]).In our data, the q-triplet assumes the following results: q rel = 1 for the PSI20 returns (the fact that there are negative correlations in the returns indicates that there is no long memory) and q rel = 1.5 in the squared returns, indicating long memory in volatility; q stat = 1.44 (either using a maximum likelihood method or using the scaling based method), which indicates rougher paths than in the Gaussian reference case; q sens = 0.541 (obtained using the multifractal spectrum estimated for the PSI20 returns and a third order polynomial approximation, presented in Figure 3); we have q sens < 1, which means that the return series is sensitive to initial conditions (as opposed to insensitivity if q sens > 1).

Wavelet Leader Entropy
There is an inverse relation between the pointwise Hölder exponent and the pointwise wavelet leader entropy, as we pointed out before.We computed the wavelet coefficients considering an orthogonal decimated discrete wavelet transform; our mother wavelet is a Daubechies, with three vanishing moments.After computing the wavelet leader coefficients for the return series, we estimate the temporal evolution of the pointwise wavelet leader entropy (see Figure 4).
The entropy is considered to be the maximum when we have the most uncertain situation.Wavelet leader values near log

Concluding Remarks
In our estimations, we find that the PSI20 returns series is highly unpredictable, rougher than a normally distributed series and has no long memory (but has persistent volatility).These characteristics are typical findings in an efficient market.The analysis of local regularity, using wavelet leaders, allows one to identify the moments of a crash in the Portuguese market as those where a peak, in the degree of irregularity, is attained.Therefore, this technique may be seen as a means to identify those kinds of events.

2 ( 8 )
(eight is the number of scales from the wavelet decomposition) indicate high regularity in the signal, while values near zero indicate low regularity.The temporal evolution of regularity allows one to identify a crisis; the financial crisis of 2008 and subsequent local minimums in PSI20 returns.Those dates are indicated in Figure4.In fact, this entropy presents sharp peaks at these temporal moments (when we obtain values near log 2 8 for the pointwise wavelet leader entropy, we expect an irregularity in the signal).