Reducing the Bias of the Smoothed Log Periodogram Regression for Financial High-Frequency Data

: For typical sample sizes occurring in economic and ﬁnancial applications, the squared bias of estimators for the memory parameter is small relative to the variance. Smoothing is therefore a suitable way to improve the performance in terms of the mean squared error. However, in an analysis of ﬁnancial high-frequency data, where the estimates are obtained separately for each day and then combined by averaging, the variance decreases with the sample size but the bias remains ﬁxed. This paper proposes a method of smoothing that does not entail an increase in the bias. This method is based on the simultaneous examination of di ﬀ erent partitions of the data. An extensive simulation study is carried out to compare it with conventional estimation methods. In this study, the new method outperforms its unsmoothed competitors with respect to the variance and its smoothed competitors with respect to the bias. Using the results of the simulation study for the proper interpretation of the empirical results obtained from a ﬁnancial high-frequency dataset, we conclude that signiﬁcant long-range dependencies are present only in the intraday volatility but not in the intraday returns. Finally, the robustness of these ﬁndings against daily and weekly periodic patterns is established.


Introduction
After Mandelbrot (1971) had discussed the possibility that the strength of the statistical dependence of stock prices decreases very slowly, several researchers investigated this issue empirically. For example, Greene and Fielitz (1977) found indications of long-range dependence when they applied a technique called range over standard deviation (R/S) analysis (Hurst 1951;Mandelbrot and Wallis 1969;Mandelbrot 1972Mandelbrot , 1975 to daily stock return series. This technique is based on the R/S statistic Q n , which is defined as the range of all partial sums of a time series of length n from its mean divided by its standard deviation. For a large class of short-range dependent processes, Q n /n H converges to a non-degenerate random variable if H = 0.5. An analogous result with H 0.5 holds for long-range dependent processes. The parameter H is called the Hurst coefficient and is used as a measure of long-range dependence. However, Lo (1991) pointed out that the results obtained with this technique may be misleading because of the sensitivity of Q n to short-range dependence (see also Davis and Harte 1987;Hauser and Reschenhofer 1995) and proposed, therefore, a Newey and West (1987) type modification for the denominator of the R/S statistic, which is appropriate for general forms of short-range dependence. Contrary to the findings of Greene and Fielitz (1977) and others, he found no evidence of long-range dependence in daily and monthly index returns once the possible short-range dependence was properly taken care of. A disadvantage of Lo's (1991) modified R/S analysis is its dependence on an important tuning parameter, namely the truncation lag q, which determines the number of included autocovariances. The general conditions that ensure the consistency of the Newey and West estimator provide little guidance in selecting q in finite samples. Additionally, Andrews's (1991) data-dependent rule for choosing q is based on asymptotic arguments.
Long-range dependence can not only be characterized by a Hurst coefficient H 0.5 but also by a slowly decaying autocorrelation function ρ or a spectral density f that is steep in a small neighborhood of frequency zero, i.e., ρ(k)k 1−2d → c as k → ∞, c > 0, d 0, and: respectively. The parameter d is called a memory parameter (or fractional differencing parameter) and is related to H by d = H − 0.5. It can be estimated by replacing the unknown spectral density f in (2) by the periodogram (Geweke and Porter-Hudak 1983) or a more sophisticated estimate of f (Hassler 1993;Peiris and Court 1993;Reisen 1994), taking the log of both sides, and regressing the log estimate on a deterministic regressor. Robustness against short-range dependence can be achieved by using only the K = n α lowest Fourier frequencies in the regression. A popular choice for the tuning parameter α is 0.5. For the purpose of testing, the asymptotic error variance is used. Applying the log periodogram regression method of Geweke and Porter-Hudak (1983) to the daily returns of the 30 components of the Dow Jones Industrials index and several indices, Barkoulas and Baum (1996) found no convincing evidence in favor of long-range dependence, which is not surprising in light of the finding of Mangat and Reschenhofer (2019) that the test based on the asymptotic error variance has very low power. Unfortunately, using the standard variance formula of the least squares estimator of the slope in a simple linear regression instead of the asymptotic error variance is also problematic because it leads to overrejecting the true null hypothesis (see Mangat and Reschenhofer 2019). The negative results of Lo (1991) and Barkoulas and Baum (1996) are in line with the results obtained by Cheung and Lai (1995) with both modified R/S analysis and log periodogram regression for stock return data from eighteen countries and by Crato (1994) with fractionally integrated ARMA (ARFIMA) models (Granger and Joyeux 1980;Hosking 1981) for stock indices of the G-7 countries. Using not only the log periodogram regression with the asymptotic error variance but additionally also nonparametric techniques such as R/S analysis and modified R/S analysis as well as parametric techniques, Grau-Carles (2000) also found little evidence of long-range dependence in index returns but strong evidence of persistence in volatility measured as squared returns and absolute returns, respectively, which corroborates earlier findings of Crato and de Lima (1994) and Lobato and Savin (1998). In general, results obtained with ARFIMA models must be treated with caution. Firstly, the true model dimension is unknown in practice and reliable inference after automatic model selection is illusory. Secondly, Pötscher (2002) has shown that the problem of estimating the memory parameter d falls into the category of ill-posed estimation problems when the class of data generating processes is too rich. For example, Grau-Carles (2000) considered all ARFIMA(p,q) processes with p ≤ 3 and q ≤ 3, which is possibly an unnecessarily large class for return series.
While the bulk of empirical research focused on major capital markets, Barkoulas et al. (2000) examined an emerging capital market, namely the Greek stock market, with the log periodogram regression and obtained significant estimates of d in the range between 0.20 and 0.30 for values of the tuning parameter α between 0.5 and 0.6. However, their sample period is relatively short and the sampling frequency is weekly rather than daily. Even less confidence-inspiring are the positive results obtained by Henry (2002) with monthly data from several international stock markets. Clearly, methods that have been designed for large samples should not be applied to small and medium samples. Recently, small-sample tests for testing hypotheses about the memory parameter d have been proposed (Mangat and Reschenhofer 2019;. When applied to asset returns, these tests produced negative results throughout. Cajueiro and Tabak (2004), Carbone et al. (2004), Batten and Szilagyi (2007), Batten et al. (2008), Souza et al. (2008), Batten et al. (2013), and Auer (2016aAuer ( , 2016b observed time-variability of the Hurst exponent in stock returns, currency prices, and the prices of precious metals, respectively. These apparent changes were occasionally interpreted as indications of changing market efficiency or even used for the construction of trading strategies. Although it cannot be ruled out that some erratic estimator for the memory parameter d catches signals that are useful for trading purposes even when in fact there is no long-range dependence, there still seems to be a need for a more efficient estimator that actually allows to get some information about the true nature of the data generating process. In general, there is always a trade-off between bias and variance. Estimators for the memory parameter d that are based on a smooth estimate of the spectral density have typically a smaller variance and a larger bias than those based on the periodogram (Chen et al. 1994;, which is advantageous in situations where the squared bias is small relative to the variance. However, in the case of high-frequency financial data, there are usually gaps between the individual trading sessions, which make it necessary to estimate d separately for each trading session and compute the final estimate by averaging the individual estimates. Here, the variance decreases with the number of trading sessions but the bias remains fixed; hence, conventional smoothing methods, which achieve a reduction in the variance at the expense of an increase in the bias, are of no use. The goal of this paper is therefore to introduce a new method of smoothing that does not systematically have a negative impact on the bias. This method will be described in detail in the next section. Section 3 presents the results of an extensive simulation study, which compares the performance of various estimators for the memory parameter in terms of bias, variance, and root-mean-square error (RMSE). Using limit order book data obtained from Lobster, Section 4 searches for indications of long-range dependence both in the intraday volatility and in the intraday returns. Section 5 provides a conclusion.

Log Periodogram Regression
Fractionally integrated white noise satisfies the difference equation: where L is the lag operator and u t is white noise with mean zero and variance σ 2 (Adenstedt 1974). Its spectral density is given by: The memory parameter d, which represents the degree of long memory if d 0, can be estimated by regressing the log periodogram of the time series y 1 , . . . , y n on a deterministic regressor (Geweke and Porter-Hudak 1983). Indeed, we have: where: is the periodogram, are the first K Fourier frequencies between 0 and π, is a deterministic regressor, is a constant, and v j = log I ω j / f ω j .
are random perturbations. Choosing K m rather than K = m is advisable when it is suspected that not only long-term dependencies are present but also short-term dependencies, e.g., when the data come from an ARFIMA process: (Granger and Joyeux 1980;Hosking 1981), where the parameter d takes care of the former dependencies and the parameters φ 1 , . . . , φ p , θ 1 , . . . , θ q take care of the latter. It is assumed that d < 0.5 (stationarity condition), d > −0.5 (invertibility condition), and all roots of the lag operator polynomials Φ(L) = 1 − φ 1 L − . . . − φ p L p and Θ(L) = 1 + θ 1 L + . . . + θ q L q lie outside the unit circle (causality condition and invertibility condition, respectively).
In the special case of d = p = q = 0 and Gaussianity, the ratios I ω j / f ω j are independent and identically distributed (i.i.d.) standard exponential and v 1 , . . . , v m are, therefore, i.i.d. Gumbel with mean −γ and variance π 2 /6, where γ = 0.57721 . . . is Euler's constant. The variance of the Geweke Porter-Hudak (GPH) estimatord GPH is then identical to the variance of the ordinary least squares (OLS) estimator for the slope in a simple regression model, i.e., where: In a neighborhood of frequency zero: Hence: Furthermore: Indeed, we have: If: (see Hurvich and Beltrao 1993), hence, the variance formula (10) becomes: in line with the asymptotic result: derived by Hurvich et al. (1998) under the assumption that K = o n 4/5 and log 2 (n) = o(K) for a class of stationary Gaussian long-memory processes with spectral densities of the form: which includes all stationary ARFIMA processes. If d 0, the ratios I ω j / f ω j are neither independent nor identically distributed, not even asymptotically (Robinson 1995). The problem is the irregular behavior of the spectral density in the neighborhood of frequency zero, i.e., Robinson (1995), therefore, proposed to remove the lowest Fourier frequencies from the log periodogram regression. Künsch (1986) showed that in the case of ARFIMA processes, the ratios I ω j / f ω j , j = H + 1, . . . , H + K are indeed asymptotically i.i.d. standard exponential provided that (H + 1)/ √ n → ∞ and (H + K)/n → 0 . However, Reisen et al. (2001) and Mangat and Reschenhofer (2019) found that even the removal of only the first Fourier frequency already has a negative effect on the performance of the estimatord GPH .

Smoothing the Periodogram
An obvious possibility to further develop the estimatord GPH is to smooth the periodogram before it is used in the regression (5) (Hassler 1993;Peiris and Court 1993;Reisen 1994). In order to illustrate the effect of smoothing, we consider the simple case of K/3 non-overlapping averages: In this case, the sample size is divided by three but at the same time the variance of the error term decreases approximately from: var to the variance of the log chi-square distribution with 6 degrees of freedom because: var Noting that the mean (first cumulant) and the variance (second cumulant) of the log chi-square distribution with k degrees of freedom are given by: and: respectively. we obtain for k = 6, κ 1 = 1.615932 and κ 2 = 0.3949341. Here, ψ is the digamma function and ψ is its first derivative. Overall, the (approximate) variance of the least squares estimator of the memory parameter d decreases from where we have assumed that The little practical relevance of asymptotic results such as (20) can be seen when the asymptotic values are confronted with the actual values obtained by simulations. In the simplest case of Gaussian white noise, we do not have to safeguard against short-range dependence and can therefore choose a value of α slightly below 4/5. Choosing α = 0.7 and K ≈ n α , we obtain 0.00857 (27) and 0.00617 (28) vs. 0.01148 and 0.00885 (simulated) for n = 250 and K = 48, 0.00326 and 0.00235 vs. 0.00381 and 0.00282 for n = 1000 and K = 126, 0.00065 and 0.00047 vs. 0.00068 and 0.00050 for n = 10, 000 and K = 630, and 0.00021 and 0.00015 vs. 0.00021 and 0.00015 for n = 50, 000 and K = 1947. Obviously, huge sample sizes are required for good agreement. In the case of a nontrivial ARFIMA process, this problem will become even more serious because a smaller value of α must be chosen.
More sophisticated further developments of the estimatord GPH are obtained by using more than three periodogram ordinates, allowing for overlaps, and introducing weights, or, equivalently, by using a lag-window estimator of the form: whereγ(s) denotes the sample autocovariance at lag s and the lag window w satisfies w(0) = 1, w(s) ≤ 1, and w(−s) = w(s) (see Hassler 1993;Peiris and Court 1993;Reisen 1994). A disadvantage of these estimation procedures is that they require the specification of a second tuning parameter, namely the length of the weighted averages in the former case and m ≤ n − 1 in the latter case, in addition to K. Of course, suitable weights and a suitable lag window, respectively, must be chosen too. Carrying out an extensive simulation study to compare various frequency-domain estimators for d,  found that too strong smoothing, e.g., caused by choosing a too small value for m, entails an extremely large bias. Hunt et al. (2003) derived an approximation for the bias and observed generally a good agreement between their approximation and the corresponding value obtained by simulations when d > 0. However, the practical relevance of this approximation is limited because of its dependence on characteristics of the data generating process, which are unknown in practice.

Using Subsamples
A simple method of smoothing without introducing a bias is to average estimates obtained from different subsamples. Assume, for example, that the final estimated is obtained by averaging over N preliminary estimatesd 1 , . . . ,d N obtained from independent subsamples y 11 , . . . , y n1 , . . . , y 1N , . . . , y nN ; then, the variance ofd vanishes as N increases while the bias remains unchanged. Of course, artificially splitting a long, homogeneous time series into non-overlapping subseries does not necessarily have a positive effect. For illustration, consider the simplest case where the time series y 1 , . . . , y n is split into two disjoint subseries y 1 , . . . , y n/2 and y n/2+1 , . . . , y n of equal length. To allow a fair comparison, the frequency range (0, ω K ], is kept constant, which implies that in the case of the two subseries the number of used Fourier frequencies is K/2. Under the simplistic and mostly unrealistic assumption that the two subseries are independent, the (approximate) variance of the mean of the two GPH estimators based on the two subseries is given by: which is, therefore, of the same size as that of the original estimator, which is based on the whole time series. However, there is still room for improvement. A reduction in the variance may be achieved by allowing for overlaps between the subseries, e.g., with a rolling estimation window or a combination of different partitions. At first glance, the idea of improving an OLS estimator by averaging the OLS estimators obtained from the whole sample and the first and second halves, respectively, seems to be at odds with the Gauß-Markov theorem because the combined estimator is still linear. However, the crucial point here is that only the observations are partitioned and not the log periodogram, which is used as dependent variable in the regression and is obtained from the observations through nonlinear transformations. For illustration, consider an estimator of the form: whered 1 ,d 21 ,d 22 are the OLS estimators for d based on the log periodograms L 1 , L 21 , L 22 of the whole sample and the first and second halves, respectively. In the special case of Gaussian white noise with variance 2π, the constant c in the regression (3) vanishes, and we may, therefore, use the simpler estimators:d and:d where x j = x j − x. For the variances of the simplistic estimatorsd 1 and: we obtain approximately: and: respectively, where we have used that cov(d 1 ,d 21 ) = cov(d 1 ,d 22 ) and cov(d 21 ,d 22 ) = 0 as well as the rough approximations: (see Table 1), and: For a further reduction of the variance, we may consider more general estimators of the form: which are based on k partitions. The next section examines whether this possible reduction actually materializes and whether it is accompanied by an increase in the bias. All computations are carried out with the free statistical software R (R Core Team 2018).

Simulations
In this section, we compare the new estimator d k (41) for k = 2, 3, 5, 10 with Geweke and Porter-Hudak's (1983) estimatord GPH , which is based on the log periodogram regression (5), and the estimatorsd sm andd β smP , which are obtained by replacing the periodogram ordinates in (5) by simple moving averages of neighboring periodogram ordinates and lag-window estimates of the form (30) with truncation lags m = n β , β = 0.5, 0.7, 0.9, 1, respectively. In the latter case, the Parzen window is used, which is given by: With a view to the later application of the estimators to 1-min intraday returns in Section 4, the sample size n = 390 is chosen for our simulation study because there are 390 min in a regular trading session for U.S. stocks, which starts at 9:30 a.m. and ends at 4:00 p.m. The number K of Fourier frequencies included in the log periodogram regression is defined by setting K = 20 ≈ [n α ], α = 0.5. For k = 2, the first K/k = 10 Fourier frequencies of the two disjoint subseries of length n/k = 195 are given by ω 2 , ω 4 , . . . , ω K , and for k = 10, the first K/k = 2 Fourier frequencies of the 10 disjoint subseries of length n/k = 39 are given by ω 10 , ω K . Clearly, we cannot go beyond k = 10 because at least two frequencies are required to carry out the log periodogram regression. Additionally, using frequencies outside the interval (0, ω K ] is not an option because this would amount to an unfair advantage, particularly when there are no short-term dependencies which have to be taken into account.
With the help of the R-package 'fracdiff', 10,000 realizations of length n = 390 of ARFIMA(1,d,0) processes with standard normal innovations and parameter values d = −0.25, −0.1, 0, 0.1, 0.25 and φ 1 = −0.25, −0.1, 0, 0.1, 0.25, respectively, are generated using a burn-in period of 10,000. For each realization, the estimatorsd GPH ,d sm ,d β smP , β = 0.5, 0.7, 0.9, 1, d k , k = 2, 3, 5, 10, are employed for the estimation of the memory parameter d. The competing estimators are compared with respect to bias (Table 2), variance (Table 3), and RMSE (Table 4). Table 3 shows that d 2 has indeed a smaller variance thand GPH = d 1 . The variance keeps decreasing as the number of partitions increases from two to 10. Table 2 shows that this improvement does in general not come at the cost of a greater bias. In contrast, the reduction in the variance achieved in the case of the estimatord β smP by increasing the degree of smoothing from β = 0.9 to β = 0.5 is for d 0 accompanied by a dramatic increase in the bias. Overall, in terms of the RSME, the best results are obtained withd 0.5 smP for small values of d and withd 0.7 smP for larger value of d. However, this is only relevant in the standard case where only a single time series is available. When a large number of time series are examined simultaneously (as in the empirical study of Section 4), the bias is the decisive factor and the new estimators d k are therefore more appropriate than the conventional estimatorsd β smP . Since values of β such as 0.5, 0.7, or 0.9 are usually chosen to minimize the MSE for a single sample, we may suspect that the estimatord β smP becomes more competitive in the case of multiple samples when the averaging is taken into account. This can be done by further reducing the degree of smoothing. Unfortunately, there is a limit to what can be achieved by increasing the value of β. Table 2 shows that large biases are still obtained with the maximum possible value of β, i.e., β = 1. This is due to the fact that global smoothing inevitably causes local distortions and cutting off higher-order sample autocovariances is not the only source of smoothing. Downweighting the sample autocovariances with the Parzen window also has a strong smoothing effect, even when all sample autocovariances are used.     Table 4. RMSE of the estimatorsd GPH (log periodogram regression),d sm (simple smoothing),d β smP , β = 1, 0.9, 0.7, 0.5 (smoothing with Parzen window and truncation lag m = [n β ]), and d k , k = 2, 3, 5, 10 (k partitions) obtained from 10,000 realizations (length: n = 390, number of used

Empirical Results
In this section, we employ the estimators discussed in the previous sections for the search of possible long-range dependencies in intraday returns and absolute intraday returns. For this purpose, the limit order book data from 27 June 2007 to 30 April 2019 (2980 trading days) of the iShares Core S&P 500 ETF (IVV) are downloaded from Lobster (https://lobsterdata.com). In the process of data cleaning, 27 early-closure days (the day before Independence Day, the day after Thanksgiving, and Christmas Eve) are removed as well as 9 January 2019 because of a large number of missing values. For each of the remaining days, the first mid-quotes (midpoints of the best bid and ask quotes) in each minute and the last mid-quote in the last minute are computed and subsequently used to obtain 1-min log returns. Finally, another three days are omitted because of extreme returns, namely 19 September 2008, 6 May 2010, and 24 August 2015, which leaves 2949 days for our analysis. Estimates are computed for each day, divided by the number of days, and plotted cumulatively; hence, the last values correspond to the averages of the estimates. The validity of these values is reinforced by the striking linearity of the curves. This linearity also implies that the possible long-range dependence is not changing over time; hence, there appears to be no such thing as fractal dynamics. Figure 1a suggests that d is close to zero in case of the 1-min log returns. The large negative values obtained withd 0.9 smP andd 0.7 smP as well as the comparatively inconspicuous values obtained withd 0.5 smP can be explained with the help of the results of our simulation study. According to Table 2, they are indicative for d = 0. In contrast, there is strong evidence of long-range dependence in the volatility (see Figure 1b). Most estimators suggest that the memory parameter d is approximately in the range between 0.3 and 0.4. Only the estimator d 0.5 smP , which is severely downward biased in case of positive d (see Table 2), favors a smaller value.

Discussion
In this paper, we have introduced a new estimator for the memory parameter d, which is based on running a log periodogram regression repeatedly for different partitions of the data. In contrast to conventional smoothing methods, which manage to achieve a reduction in the variance at the expense of an increase in the bias, our approach does not systematically have a negative impact on the bias, which makes it particularly useful for applications where the bias is the decisive factor. For example, intraday returns are usually only available during trading hours and estimation must therefore be carried out separately for each trading day. When the individual estimates are eventually combined by averaging, the variance decreases as the sample size increases, but the bias does not change. The results of an extensive simulation study confirm the good performance of the new estimator. It outperforms all of its competitors when both bias and variance are taken into account, but the bias is weighted more heavily.
The importance of results obtained with the help of simulations is due to the fact that reliable inference on the memory parameter is not possible under general conditions. Some asymptotic Visual significance of the differences between certain estimates can be ascertained just by observing the large differences between the slopes of the corresponding lines in Figure 1 and noting the striking stability of these lines over time. However, we still might want to augment our visual analysis with a formal statistical test. A simple way to accomplish that is to calculate the difference between two estimates separately for each trading day and compare the number of positive differences to the number of negative differences (sign test). Not surprisingly, the resulting p-values are infinitesimal. For example, even in the case of the two neighboring lines corresponding tod GPH and d 2 in Figure 1b, the p-value is less than 2.2 × 10 −16 . It is still less than 9.7 × 10 −8 when we omit most of the trading days and use only Wednesdays in order to ensure approximate independence of the subsamples. Note that there are 4 × 390 = 1560 1-min returns between the last 1-min return of some Wednesday and the first 1-min return of the next Wednesday plus five overnight breaks and a whole weekend. Even for a relatively large value of the memory parameter such as d = 0.3, the autocorrelation of an ARFIMA(0,d,0) process at lag j = 1561 is quite small, i.e., Finally, in order to check the robustness of our findings against daily and weekly periodic patterns, we repeat the graphical analyses with suitably transformed data. Replacing the 1-min log returns r t (s), s = 1, . . . , 390, by the daily differences r t (s) − r t−1 (s) and the weekly differences r t (s) − r t−5 (s), respectively, ensures that any daily or weekly periodic patterns are erased while long-range dependencies remain unaffected. Figure 1c,e are very similar to Figure 1a, which shows that the insights gained from Figure 1a are genuine. Analogously, comparing Figure 1d,f with Figure 1b, we see that the same is true for the absolute returns

Discussion
In this paper, we have introduced a new estimator for the memory parameter d, which is based on running a log periodogram regression repeatedly for different partitions of the data. In contrast to conventional smoothing methods, which manage to achieve a reduction in the variance at the expense of an increase in the bias, our approach does not systematically have a negative impact on the bias, which makes it particularly useful for applications where the bias is the decisive factor. For example, intraday returns are usually only available during trading hours and estimation must therefore be carried out separately for each trading day. When the individual estimates are eventually combined by averaging, the variance decreases as the sample size increases, but the bias does not change. The results of an extensive simulation study confirm the good performance of the new estimator. It outperforms all of its competitors when both bias and variance are taken into account, but the bias is weighted more heavily.
The importance of results obtained with the help of simulations is due to the fact that reliable inference on the memory parameter d is not possible under general conditions. Some asymptotic results can be obtained under very restrictive conditions though. Unfortunately, convergence is typically very slow (recall the discussion in Section 2.2). Indeed, Pötscher (2002) showed that many common estimation problems in statistics and econometrics, which include the estimation of d, are ill-posed in the sense that the minimax risk is bounded from below by a positive constant independent of n and does, therefore, not converge to zero as n → ∞ . In particular, he found that for any estimatord n for d based on a sample of size n from a Gaussian process with spectral density f : where 1 ≤ r < ∞ and F is the set of all ARFIMA spectral densities (p ≥ 0, q ≥ 0), ARFI spectral densities (p ≥ 0, q = 0), or FIMA spectral densities (p = 0, q ≥ 0). Furthermore, he showed that for every f 0 ∈ F , (44) holds also "locally," when the supremum is taken over an arbitrarily small L 1 -neighborhood of f 0 . Finally, he established that confidence intervals for d coincide with the entire parameter space for d with high probability and are therefore uninformative. Nevertheless, it may be possible to formally derive the statistical properties of our new estimator for a rather narrow class of processes such as low order ARFI processes. However, this is left for future research. The current paper provides just a proof of concept.
In our empirical investigation of high-frequency data of an index ETF, we have applied the competing estimators to 1-min log returns and absolute 1-min log returns separately for each day. The results are quite stable over time and across estimation methods. The few deviations are due to conventional smoothing methods and can easily be explained by the size of their bias as shown in Table 2. We may, therefore, safely conclude that significant long-range dependencies are present only in the intraday volatility but not in the intraday returns. These findings are genuine and not just due to daily or weekly periodic patterns because similar results are obtained when daily and weekly differences are investigated instead of the original intraday returns.
Author Contributions: Both authors contributed equally to the paper. All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.