Next Article in Journal
An Expectation-Maximization Algorithm for the Exponential-Generalized Inverse Gaussian Regression Model with Varying Dispersion and Shape for Modelling the Aggregate Claim Amount
Previous Article in Journal
Minimal Expected Time in Drawdown through Investment for an Insurance Diffusion Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bayesian Approach to Measurement of Backtest Overfitting

Department of Banking and Insurance, Faculty of Finance and Accounting, University of Economics, W. Churchill Sq. 4, 130 67 Prague, Czech Republic
Risks 2021, 9(1), 18; https://doi.org/10.3390/risks9010018
Submission received: 19 November 2020 / Revised: 16 December 2020 / Accepted: 1 January 2021 / Published: 8 January 2021

Abstract

:
Quantitative investment strategies are often selected from a broad class of candidate models estimated and tested on historical data. Standard statistical techniques to prevent model overfitting such as out-sample backtesting turn out to be unreliable in situations when the selection is based on results of too many models tested on the holdout sample. There is an ongoing discussion of how to estimate the probability of backtest overfitting and adjust the expected performance indicators such as the Sharpe ratio in order to reflect properly the effect of multiple testing. We propose a consistent Bayesian approach that yields the desired robust estimates on the basis of a Markov chain Monte Carlo (MCMC) simulation. The approach is tested on a class of technical trading strategies where a seemingly profitable strategy can be selected in the naïve approach.

1. Introduction

The problem of backtest overfitting appears mainly in two important econometric research areas: testing and selection of factors explaining asset returns (see, e.g., De Prado 2015; Harvey and Liu 2013, 2014, 2015, 2017, 2020) and selection of investment strategies (see, e.g., White 2000; Bailey et al. 2016; López de Prado and Lewis 2019). Our focus is the investment strategy selection problem arising when many strategies are developed and tested on historical data in order to find a “performing” one. The selection process can be realized by an individual researcher or institution or latently by a set of researchers investigating various strategies and publishing only the promising ones (see, e.g., Andrews and Kasy 2019; Chen and Zimmermann 2020). The latter approach is more common for theoretical research, while the former, easier to control, would be typical for a quantitative investment firm.
We formalize and investigate the problem of strategy selection on the basis of a large set of candidates. Consider investment strategies1 S 1 , , S K that are backtested and evaluated over a historical period with T 1 (e.g., daily) returns r k , t ,   k = 1 , , K ,   t = 1 , , T 1 . Note that the strategies could have been developed on another preceding training period and backtested or validated on the { 1 , , T 1 } period. Another possibility that we use in the empirical study is that one considers a number of expertly proposed, e.g., technical, strategies that are just evaluated on the backtest period. On the basis of the historical data, we estimate the (annualized) sample means m k , standard deviations   s k , or Sharpe ratios S R k , and, given a criterion, we select the “best” strategy S b . Of course, the key question is what can be realistically expected from the best strategy if implemented in the future period of length T 2 (see Figure 1).
Specifically, the questions usually asked include the following:
  • First, is it sufficient to apply the standard single p-test to the best strategy?
  • If not, how should we modify the test to incorporate the multiple test effects?
  • What is the expected future, i.e., true out-of-sample (OOS) performance (return, SR, etc.) of the best strategy selected on the in-sample (IS), i.e., historical dataset?
  • What is the haircut, i.e., the percentage reduction, of the expected OOS performance compared to the IS performance?
  • What is the probability of loss, if the strategy is implemented over a future period?
  • What is the expected OOS rank of the IS best strategy among the candidate strategies?
  • What is the probability that the selected model will in fact underperform most of the candidate models?
  • What is the probability that we have selected a false model (false discovery rate—FDR)?
We propose a Bayesian methodology that allows us to simulate many times the IS selection and OOS, i.e., future, realization process (Figure 1) in order to address the questions formulated above. We provide an overview of several methods proposed in the literature that are compared with our method in an empirical study on a set of technical strategies. Since the strategies are basically random and, in fact, we know their performance following the backtesting period, none should be selected. On the other hand, we also artificially modify one of the strategies making it a “true” discovery and analyze results of the strategy selection methods.

2. An Overview of the Existing Approaches

There are several relatively simple classical methods for adjusting p-values in order to accommodate the multiple test (see, e.g., Streiner and Norman 2011). More advanced and computationally demanding methods are based on various approaches to bootstrapping and simulation of the past and future data.

2.1. Classical Approaches

To test significance of a single strategy, for example, S b with sample mean return m b and standard deviation s b observed over T periods, the classical approach is to calculate the t-ratio,
T R = m b s b / T   ,
and the two-sided2 p-value,
p S = Pr [ | X | > T R ] ,
where X is a random variable following the t-distribution with T 1 degrees of freedom. The implicit assumption is that the returns are i.i.d. normal. If the p-value p S happens to be small enough, e.g., below 5% or 1%, then one tends to jump to the conclusion that a strategy with significantly positive returns has been discovered.
The problem with the process of selecting the best strategy or, alternatively, testing a number of strategies until we find a significant one is that the correct p-value should (Harvey and Liu 2015) reflect the fact that we selected the strategy with the best t-ratio T R out of K proposal strategies.
p M = Pr [ max { | X k | , k = 1 , , K } > T R ] ,
where X k are K random variables following the t-distribution with T 1 degrees of freedom (corresponding to the sample t-ratios of the K strategies). It can be noted (Harvey and Liu 2015) that, if the variables were independent, then we could find a simple relationship (also called Šidák’s adjustment) between the single and multiple test p-values.
p M = 1 k Pr [ | X k | T R ] = 1 ( 1 p S ) K = K p S ( K 2 ) ( p S ) 2 + .
Harvey and Liu (2015) provided an overview of simple adjustment methods, such as Bonferroni’s adjustment p M = min { K p S , 1 } , Holm’s, or Benjamini, Hochberg, Zekutieli (BHY) adjustments, using the ordered sequence of the single-test p-values p 1 S , , p K S for all the strategies. The weak point of all those methods is the assumption of independence since the tested strategies are often closely related (e.g., of the same type with varying parameters).
We also propose and test a numerically relatively simple and efficient method on the basis of an estimation of the covariance matrix Ω of the returns and numerically generate the distribution of max { | X k | , k = 1 , , K } conditional on the null hypothesis m k = 0 for all k where X k are K random variables following the t-distribution with T 1 degrees of freedom (or alternatively standard normal for a large T ) and with covariances given by Ω .
Note that the classical (corresponding to the basic period, e.g., daily) Sharpe ratio can be easily calculated given the t-ratio and vice versa.
S R k = m k s k = T R k T .
The Sharpe ratio is usually annualized as follows:
S R k a = m k s k T a = T R k T a T ,
where T a is the number of observation periods in a year, e.g., 252 in the case of daily returns. According to Equation (1), the maximal acceptable p-value level can be easily translated to a minimum required Sharpe ratio.
Generally, given a selected strategy with an in-sample (according to the backtest data) Sharpe ratio S R I S , the question is what the expected (out-of-sample) Sharpe ratio E 0 [ S R O O S ] over a future, e.g., 1-year period, is. Here, E 0 [ . ] denotes the expectation given all the information available today, in particular, given the in-sample performance such as S R I S , the number of strategies from which the best one was selected, the relationship between the strategies, and the underlying asset return process properties. The Sharpe ratio haircut is then defined as the percentage we need to deduct from the in-sample Sharpe ratio to get a realistic estimate of the future performance,
H C = 1 E 0 [ S R O O S ] S R I S .
Harvey and Liu (2015) noted that the rule-of-thumb haircut used by the investment industry is 50%, but that, according to their analysis, it significantly depends on the level of the in-sample Sharpe ratio and the number of strategies. They proposed to use the relationship between the single- and multiple-test p-values in order to derive the haircut Sharpe ratio. Their estimate of the annualized expected Sharpe ratio E S R H L is based on the idea that its corresponding single-test p-value should be equal to the adjusted multiple test p-value p M , i.e.,
p M = Pr [ | X | > E S R H L T T a ] ,   i . e . , E S R H L = F 1 ( p M / 2 ) T a T ,
where X is a random variable following the t-distribution with T 1 degrees of freedom and F is its cumulative distribution function. The haircut is then calculated according to Equation (2). The haircut estimation, of course, depends on the p-value adjustment method as Bonferroni, Holm’s, BHY, or the general one we suggested above. Although the estimation is obviously directionally correct, it is not obvious why this approach should yield a consistent estimate of the expected Sharpe ratio E 0 [ S R O O S ] and of the corresponding haircut. We compare the different haircut estimates in the simulation study outlined below.

2.2. Stationary Bootstrap

In order to simulate the past and the future returns, we consider a bootstrapping and a cross-validation approach. The stationary bootstrap proposed and analyzed in White (2000), Sullivan et al. (1999), and Politis and Romano (1994) is applied to the underlying asset returns assumed to be strictly stationary and weakly dependent time-series to generate a pseudo time series that is again stationary (Figure 2). The tested strategies S 1 , , S K are based only on the single underlying using its historical returns to take long, short, or possibly zero positions.
Formally, we generate new sequences of the underlying asset returns { u Θ ( i ) ; i = 1 , .. , T ˜   } where u 1 , , u T 1 is the original series of observed returns and Θ ( i ) { 1 , , T 1 } . In order to implement the bootstrap, we need to select a smoothing parameter 0 < q = 1 / b < 1 , where b corresponds to the mean length of the bootstrapped blocks, for example, q = 0.1 , proposed by Sullivan et al. (1999). A bootstrapped sequence is obtained by drawing randomly Θ ( 1 ) { 1 , , T } , and, for i = 2 , ,   T ˜ , setting Θ ( i ) = Θ ( i 1 ) + 1 with probability 1 q or randomly drawing a new block starting position Θ ( i ) { 1 , , T 1 } with probability q . If it happens that Θ ( i ) > T 1 then we draw random Θ ( i ) { 1 , , T 1 } .
Next, given a bootstrapped sequence of the underlying asset returns, we need to apply strategies S 1 , , S K to get the strategies’ bootstrapped returns r ˜ k , t ,   k = 1 , , K ,   t = 1 , , T 1 . Note that, since the strategies’ decisions are built on the basis of past returns, we generally need to have a longer series of the bootstrapped asset returns, T ˜ > T 1 , where T 1 is the length of the in-sample period. Then, we evaluate our desired performance indicator values (mean return, Sharpe ratio, etc.) f ˜ k . Let f k * denote the performance indicators of the original series of returns. According to White (2000), under certain mild theoretical assumptions, the empirical distribution of the B bootstrapped values V j ˜ = max k = 1 , , K ( f ˜ k f k * ) for j = 1 , , B asymptotically converges to the distribution of the best strategy performance indicator under the null hypothesis H 0 that all the strategies have zero performance. That is, obtaining B bootstrapped values { V ˜ j ; j = 1 , , B } , we can test H 0 by calculating the empirical p-value Pr [ | V ˜ j | > f b * ] , where b is the index of the best strategy and f b * = max k f k * .
The bootstrap technique can be also used to analyze the relationship between IS and OOS Sharpe ratio (or another indicator), generating series of strategy returns over a time period 1 , , T 1 , selecting the best strategy S b with an in-sample performance S R I S , and then looking at its out-of-sample performance S R O O S   over the following period T 1 + 1 , , T 1 + T 2 . Note that, in this case, the original bootstrapping must be done over a period of length T ˜ > T 1 + T 2 . Then, we can compare the mean S R O O S   against the mean S R I S or conditional on certain level of S R I S . We may also bootstrap the OOS returns for the actually selected strategy S b (according to the real dataset). However, in this case, it is especially obvious that even a truly positive strategy that is using medium-term or long-term trends to make good predictions does not have to work on the bootstrapped series of returns where the future and past returns of the original series are to a large extent mixed up. Therefore, the estimated conditional S R O O S   may easily lead to a false rejection of a positive strategy.

2.3. Combinatorial Symmetric Cross-Validation

Another disadvantage of the stationary bootstrap technique is that it cannot be applied if we are given only the strategy returns but not details on the strategies themselves. The stationary bootstrap is also problematic if the strategies are not technical ones and use a number of additional, possibly lagged, explanatory factors. This is not the case for the combinatorial symmetric cross-validation (CSCV) (Bailey et al. 2014, 2016) utilizing only the matrix of the strategies’ returns M = { r k , t ,   k = 1 , , K ,   t = 1 , , T 1 } . The idea is to split the time window of length T 1 = S N into S blocks of length N , where S is even and draws combinations of S / 2 blocks (Figure 3). The submatrix J formed by joining T 1 / 2 rows of M corresponding to the selected time indices in the original order then represents an in-sample dataset of returns where the best performing strategy can be selected while the complementary K × T 1 / 2 submatrix J ¯   represents the out-of-sample returns. The sampling can be done with or without replacement. Since there are ( S S / 2 ) combinations, we can sufficiently form many different combinations with replacement as long as S is sufficiently large, e.g., at least 16.
Bailey et al. (2014, 2016) proposed using the technique to specifically estimate the probability of backtest overfitting (PBO) defined as the probability that the best IS selected strategy performs below the average OOS. More precisely, for K strategies S 1 , , S K ,
P B O = Pr [ R a n k O O S ( X ) < K / 2 | R a n k I S ( X ) = 1 ] .
The PBO indicator, as well as the Sharpe ratio haircut, can be estimated using sufficient cross-validation pairs of the IS/OOS datasets J , J ¯ . However, it is obvious that the estimates are biased, introducing a negative drift into the OOS order of the strategies. For example, if all the strategies represented just pure noise with mean returns over the full time interval { 1 , , T 1 } close to zero, then, for an IS/OOS combination J , J ¯ , the best strategy IS return r ¯ b , J implies that the complementary OOS return r ¯ b , J ¯ r ¯ b , J would probably be the worst on J ¯ . We demonstrate the effect in the empirical section. The cross-validation technique also cannot be used, due to this property, to estimate the OOS Sharpe ratio or mean for a particular selected strategy. We can just estimate the overall P B O or Sharpe ratio haircut keeping in mind that the estimations incorporate a conservative bias. The cross-validation, as well as the bootstrapping approach, cannot be easily used to estimate the false discovery rate (FDR) since it is not clear how to identify true and false discoveries given a CSCV simulation. This could be possibly done by testing the significance of OOS performance involving an ad hoc probability level.
We show that all the indicators of interest can be consistently estimated in the Bayesian setup outlined below.

3. Bayesian Simulation Approach

The Bayesian approach is based on the scheme given in Figure 4.
First, of course, we need to specify a model defining the return generating process with unknown parameters Θ for the observed strategy returns { r k , t ,   k = 1 , , K ,   t = 1 , , T } , where T = T 1 or T = T 1 + T 2 . Then, the plan is to use a Bayesian technique, specifically, the Markov Chain Monte Carlo (MCMC) simulation, in order to extract the posterior distribution of the model parameters Θ . Finally, we simulate the matrices of IS and OOS returns over desired time intervals 1 , , T 1 and T 1 + 1 , , T 1 + T 2 . The Monte Carlo (MC) is done in two steps, always selecting the parameters Θ from the posterior distribution and then generating K series of T 1 + T 2 returns according to the model. The simulated IS returns can be used to select the best strategy and the OOS returns to measure its future performance. The average haircut or average relative rank can be easily estimated as in the case of the stationary bootstrap (see Figure 4).
We consider two models; the simple one assumes that the returns are multivariate normal with unknown covariance matrix and means, while the second incorporates unknown indicators of truly profitable strategies allowing us to consistently estimate the false discovery rate (FDR). The second model follows an idea of Scott and Berger (2006), also mentioned in Harvey et al. (2016); nevertheless, in both cases, the model was formulated only for observed mean returns and without considering a correlation structure of returns. It should be emphasized that our focus is to analyze the impact of backtest overfitting assuming that the strategies’ cross-sectional returns behave in a relatively simple and stable way over time similarly to the classical, bootstrapping, or cross-validation approaches. One could certainly come up with state-of-the-art models incorporating jumps, switching regimes, stochastic variances, or even dynamic correlations. These improvements would make the methodology computationally difficult to manage with results probably even more conservative compared to the approaches we consider below.

3.1. The Naïve Model 1

To set up the naïve model, we assume that the cross-sectional strategy returns are multivariate normal,
r t = r 1 , t , , r K , t   N ( μ , Σ ) ,
and that the observations over time are independent.
Given data =   r t , i.e., the matrix of back test returns, and possibly some priors for μ and Σ , we can find the posterior distribution p ( μ , Σ | data ) using the standard Gibbs MCMC sampler.
Specifically, the iterative sampling is given by
p ( μ | Σ , data ) = φ ( μ ; 1 T t = 1 T r t , 1 T Σ ) ,   and p ( Σ | μ , data ) = I W ( Σ ; T , S ) ,
where S = t = 1 T ( r t μ ) ( r t μ ) is the scale matrix (i.e., the covariance matrix times T ), and I W is the inverse Wishart distribution. For example, Matlab allows sampling from the distributions, and the posterior distribution may be obtained quite efficiently (e.g., 10,000 runs of the sampler).
Remark 1.
The sampler above assumes the noninformative prior on the means, p ( μ ) 1 , and the standard improper prior on the covariance matrix,
p ( Σ ) | Σ | K + 1 2 .
Given the extracted posterior distribution p ( μ , Σ | data ) , the parameters μ , Σ can be now easily sampled in order to get the empirical distribution of the selected strategy performance. However, in the process of selecting the best strategy, we do not know μ , Σ but only a time series of the backtested returns with cross-sections from N ( μ , Σ ) . On the basis of the time series, the “best” strategy   S b is selected. Our key question is about its expected forward-looking performance, e.g., μ b   or S R b . Therefore, we need to run the following Monte Carlo simulation in order to faithfully sample the empirical distribution of the performance indicators:
  • Sample μ , Σ from p ( μ , Σ | data ) .
  • Independently sample T 1 + T 2 cross-sections R t N ( μ , Σ ) .
  • Determine the index of the best strategy b on the basis of the backtest statistics calculated from the matrix of backtested returns R = R t for t = 1 , .. , T 1 .
  • Calculate and store the performance indicators, μ ^ b , S R b ^ , in the OOS period T 1 + 1 , , T 2 . Alternatively, store the selected strategy “true” performance indicators, i.e., μ b , S R b .
The MCMC estimation of the multivariate normal distribution parameters is known to converge relatively fast (see e.g., Lynch 2007); nevertheless, we apply a burn-out period according to simple diagnostics, e.g., the average of the vector of mean returns μ . The simulated posterior distribution of the desired performance indicators (after removing the burn-out period) then tells us the mean, median, confidence intervals, or Bayesian probabilities where the true performance is positive or above any given minimum threshold. The ratio between the ex post and ex ante performance indicators also gives us an estimate of the “backtest overfitting haircut”.

3.2. Model 2—Bimodal Means Distribution

In order to capture the situation when most strategies are random and only some positive (nonzero), assume that there are, in addition, latent indicators γ i { 0 , 1 } so that the mean of strategy i is μ i * = γ i μ i . Therefore, the row vector of returns has the distribution
r t = [ r 1 , t , , r K , t ]   N ( μ * , Σ ) .
Here, we need to assume a prior distribution for γ i B e r n ( 1 p 0 ) and μ i N ( m 0 , V 0 ) . It means that the Bayesian distribution of the means is bimodal with a large probability mass on 0, and the other mode is normal with prior mean m 0 > 0 and variance   V 0 . The Gibb’s sampler can be modified as follows:
  • Given μ , γ , set μ i * = γ i μ i , and estimate Σ as above, i.e.,
    p ( Σ | μ , γ ,   data ) = I W ( Σ ; T , S ) ,   where   S = t = 1 T ( r t μ * ) ( r t μ * ) .
  • Given Σ ,   γ , estimate μ . Set A = 1 T Σ , Γ = d i a g ( γ ) , m = 1 T t = 1 t r t , m 0 = [ m 0 , , m 0 ] , and D = d i a g ( [ V 0 , , V 0 ] ) , where d i a g creates a matrix with diagonal elements given by the vector in the argument, and sample
    p ( μ | Σ , γ ,   data ) = φ ( μ ; ( Γ A 1 Γ + D 1 ) ( Γ A 1 m + D 1 m 0 ) ,   ( Γ A 1 Γ + D 1 ) 1 ) .
  • Given Σ and μ , estimate γ . For i = 1 , , K set Γ 0 equal to Γ with the exception of the diagonal element Γ 0 ( i , i ) = 0 , and Γ 1 setting Γ 1 ( i , i ) = 1 . Let
    L 0 = exp ( 1 2 ( ( Γ 0 μ m ) A 1 ( Γ 0 μ m )   ) ) ( 1 p 0 ) ,
    L 1 = exp ( 1 2 ( ( Γ 1 μ m ) A 1 ( Γ 1 μ m )   ) ) p 0 ,
    p ˜ = L 1 L 0 + L 1 ,   and   finally ,   sample   γ i B e r n ( p ˜ ) .
Proofs of the formulas used in steps 2 and 3 can be found in Appendix A. Again, we apply a burn-out period according to a convergence diagnostic of the averages of γ   and of μ .
Remark 2.
There are certain possible extensions.
  • We may allow γ i { 1 , 0 , 1 } encoding negative significant mean return, zero return, or significant positive return. In this case, the mean parameter of the prior distribution μ i N ( m 0 , V 0 ) must be strictly positive. In the Gibb’s sampler above, we just need to modify step 3 in a straightforward manner.
  • The hyper-parameters p 0 , m 0 , V 0 for γ i B e r n ( 1 p 0 ) and μ i N ( m 0 , V 0 ) might be estimated within the MCMC procedure. In this case, the Gibb’s sampler can be extended as follows:
    4.
    Sample p 0 given γ :
    p ( p 0 | γ ) p 0 n 1 ( 1 p 0 ) 1 n 1 p ( p 0 ) B e t a ( p 0 ; n 1 + k 1 + 1 , K n 1 + k 2 + 1 ) ,
    where n 1 = # { i ; γ i = 1 } and p ( p 0 ) = B e t a ( p 0 ; k 1 + 1 , k 2 + 1 ) is a conjugate prior distribution (e.g., k 1 = k 2 = 1 ) .
    5
    Sample m 0 , V 0 given μ and γ . Here, we just use the means { μ i | γ i = 1 } where the signal is positive and the normal Gibb’s sampler. Since the set may be empty, we need to use proper conjugate priors, e.g., p ( m 0 ) = φ ( m 0 ; 0 , m P ) and
    ( V 0 ) I G ( V 0 , k 0 2 , k 0 V P 2 ) .
    For K ˜ = # { μ i | γ i = 1 } 0 , set μ ˜ = { μ i | γ i = 1 } / K ˜ and V ˜ = { ( μ i m 0 ) 2 | γ i = 1 } / K ˜ . Then,
    p ( m 0 | μ ,   γ , V 0 ) φ ( m 0 ; μ ˜ , V 0 / K ˜ )   φ ( m 0 ; 0 , V P )   φ ( m 0 ; μ ˜ K ˜ V P K ˜ V P + V 0 , V P V 0 K ˜ V P + V 0 )   and p ( V 0 | μ ,   γ , m 0 ) I G ( 0 ; K ˜ 2 , K ˜ V ˜ 2 )   I G ( V 0 , k 0 2 , k 0 V P 2 ) I G ( V 0 , K ˜ + k 0 + 1 2 , K ˜ V ˜ + k 0 V P 2 ) .
    If K ˜ = 0 . then we have to sample on the basis of conjugate priors p ( m 0 ) and p ( V 0 ) only.

4. Numerical Study

Following Sullivan et al. (1999) and other studies, we compare and illustrate the proposed Bayesian methods on a set of technical strategies’ returns. We also artificially modify the mean returns of the strategies in order to test the methods, on one hand, if there is a clearly extraordinary strategy or, on the other hand, if the returns of the returns of all the strategies are very low.

4.1. Technical Strategies Selection

We used 1000 daily S&P 500 values and returns for the period 5 June 2009–24 May 2013 (1000 trading days). The period was selected with the purpose of finding at least one strategy with a higher mean return. As in Sullivan et al. (1999), we applied the filter, moving average, support, and resistance rules with varying parameters. We randomly selected 200 strategies with the condition that the daily return series are not collinear (it may even happen that the series are identical if the parameters do not differ too much).
The means and Sharpe ratios of the individual strategies’ return series and their densities over the period 5 June 2009–24 May 2013 are shown in Figure 5. It should not be surprising that the strategies’ returns are mostly positively correlated with the average pairwise correlation 23.32%. Note that strategy 7 is apparently the best with the annualized ( n y = 252 ) mean return over 21% p.a. and Sharpe ratio approximately 1.2 (it is a filter strategy with x = y = 1%, i.e., a long or short position is taken if the previous daily return is over 1% or below −1%, respectively, and the minimum number of days to stay in a position is 20). The strategy returns look attractive; nevertheless, looking forward, it turns out that the realized mean return of the strategy in the 1000 days following 24 May 2013 is negative (−5.21%).

4.2. Single and Multiple p-Value Testing

The single-test annualized t-ratio and the p-value of the best strategy 7 with S R = 1.1987 are
T R = S R n o b s n y = 1.1987 1000 / 252 = 2.3878   and   p S = 0.0171 .
The multiple test p-value after Bonferroni adjustment is simply
p M = min { 200 × 0.171 , 1 } = 1 .
Thus, the adjusted expected Sharpe ratio is 0 and the haircut is 100%. Šidák’s adjustment yields only a slightly more optimistic result with p M = 0.9685 , adjusted expected S R = 0.02 , and haircut 98.3%. Similar adjustments can be obtained by the Holm or Benjamini, Hochberg, Yekutieli (BHY) methods using, for example, the package provided by Harvey and Liu (2015). The simple adjustment methods allow to easily estimate the minimum return of the best strategy (keeping the same covariance structure) in order to get the multiple test p-value at most 5%. The estimated minimum return using the same package is around 36%.

4.3. Multivariate Normal Simulation and the Stationary Bootstrap According to the Null Hypothesis

Another relatively simple possibility is to estimate the return covariance matrix and simulate the future multivariate returns on the basis of the return covariance matrix and conditional on zero means. Figure 6 below shows the density of 1000 simulated annualized SR as a function of a 1000 day period. The adjusted p-value of the best strategy with S R = 1.1987 is then relatively optimistic 0.352 and the adjusted expected SR is 0.4683 (i.e., the implied haircut is just 61%).
An analogous distribution (Figure 7) can be obtained via the much more computationally demanding stationary bootstrap (White’s reality check). The p-value according to 1000 bootstrap simulations for a 1000-day time period and with q = 0.1 turns out to be 0.728, the corresponding adjusted expected SR is 0.175, and the SR haircut is 85.4%.

4.4. Stationary Bootstrap Two-Stage Simulation

The stationary bootstrap method can also be used to simulate the backtest period of length T 1 = 1000 , as well as the future period with T 2 = 1000 . The number of stationary bootstrap iterations is again 1000 according to the 5 June 2009–24 May 2013 window of S&P returns, and the parameter is set to q = 0.1 with the corresponding average length of the bootstrapped blocks being 10. Figure 8 shows the typical strong shift of the ex ante performance density to the left-hand side and the wider ex post performance density. The results show that the best IS selected strategy performs poorly OOS with 32.8% probability of loss, PBO around 0.44 (see also Figure 9), and SR haircut over 73%. For detailed results including the ex ante and ex post SR or mean return values, see the summary in Table 1. Note that the row “stationary bootstrap” shows values obtained via the two-stage simulation except the p-value estimated by White’s reality check.

4.5. Combinatorial Symmetric Cross-Validation

In order implement the CSCV algorithm, we chose the number of blocks 20 corresponding to
( 20 10 ) = 184 , 756
combinations of 10 blocks of length 50. However, we sampled only 1000 combinations. In this case, we always split the 1000-day time into the IS and OOS parts on the same length, i.e., T 1 = 500 and T 2 = 500 . The results shown in Table 1 are quite like the stationary bootstrap only with PBO being slightly lower (0.323). The slightly better performance is reflected in the bimodal ex post densities in Figure 10 where the right-hand side positive mode corresponds to the selected strategy that performs well IS, as well as OOS. Figure 11 also indicates that, in this case, compared to Figure 9, the best IS strategy remains the best OOS quite often.

4.6. The Naïve Bayes Model 1

In the Bayesian approach, we firstly extract the multivariate normal model means and covariance given the observed data. This can be done in 1500 iterations using the standard Gibbs sampler including 500 burn-out simulations, i.e., only the last 1000 iterations are used to estimate the posterior parameters (see Figure A1, Appendix A, for the fast MCMC convergence). In the MC simulation, we can choose the length of the backtest (IS) period T 1 = 1000   and the OOS forward looking period T 2 = 1000 . We then generated 1000 scenarios (after removing the burned-out simulations) sampling the parameters and the cross-sectional returns, selecting the best IS strategy, and measuring its OOS performance. We have to keep in mind that the sampled posterior means may differ quite significantly from the observed mean returns due to the high return volatility3 and, thus, the observed best strategy 7 may look weak in the simulations, while other strategies are selected as the best. The detailed results given in Table 1 and Figure 12 confirm that the naïve multivariate normal model indeed appears too optimistic. Indeed, the best IS strategy remains the best quiet often as shown in Figure 13 and the PBO is quite low (0.168). This can be explained by the simple multivariate normal model and relatively long IS window allowing to identify the truly positive strategy.
In reality, if we generate a large number of models and the best model performance is still poor, then we are probably not going to enroll it for real trading. Therefore, we might also consider a minimum hurdle at which we choose or reject the best selected model. This is quite easy given the simulation outputs. For example, if we set the minimum SR to 1.2, then the condition will be satisfied in 96.6% of the simulations with average ex ante SR 2.14 and ex post SR 1.00, i.e., again with the haircut slightly over 53%. It is interesting that the haircut is not much sensitive to the hurdle, e.g., if the minimum SR was 2, then the corresponding average haircut would be even higher (37%). Nevertheless, the probability of loss can be reduced by setting the minimum SR higher, e.g., if we set the hurdle to 2, then the conditional probability of loss would decline to 16.5% (conditional on the strategy selection) and the unconditional probability would decline to 9.1% since 44.9% of the proposed models would be rejected in the simulation.

4.7. Bayes Bimodal Mean Returns Model 2

In this case, in addition to the multivariate normal distribution with unknown parameters, we also consider latent indicators of zero and nonzero models. In order to extract the posterior distribution of the parameters and the latent indicators, we again run 1500 iterations and use only the last 1000 iterations to estimate the parameters (removing first 500 burn-out iterations—see Figure A1, Appendix A, for the relative MCMC convergence) of the Gibbs sampler outlined in Section 3.2. Analogously to Bayes Model 1, we then run 1000 Monte Carle simulations with T1 = 1000 and T2 = 1000. Since, in this case, the Bayesian model incorporates the uncertainty whether the model is a true discovery or not, the results should be more conservative compared to the naïve model. Indeed, the PBO turns out to be 0.395, substantially lower compared to model 1, the SR haircut is 82.5%, and the probability of loss is 38% (see Table 1, Figure 14 and Figure 15).
The model also provides posterior averages of γi for each individual model i (see Figure 16). The averages can be interpreted as Bayesian probabilities that the models are true discoveries. There are a few models with the averages over 80%, including the model 7 with the value over 86%. The complements of these Bayesian probabilities to 100% can be in a certain sense compared to the frequentist single-test p-values. However, the Bayesian model also allows us to answer the key question we are asking: Given the observed data and the general model assumptions, what is the probability that the best strategy b selected on the basis of observed data is a true discovery, i.e., γb = 1? This can be estimated as the mean of γb which turns out to be only 0.459. This means that, applying the selection process, we identify the true discovery only in 45.1% of cases and we make a false discovery in 54.9% of cases, i.e., FDR = 54.9% can be shown instead of the adjusted p-value in Table 1.
We may again test whether a higher SR hurdle reduces the high SR haircut. The results are similar for the naïve model 1, i.e., the SR haircut stays around 82% more or less independently of the hurdle. The probability of loss can be reduced only slightly, e.g., for the hurdle of 1.5, the conditional probability of loss declines to 32.1%, but the unconditional probability of loss goes significantly down to 5.2% as 83.8% of the best strategies are rejected in the simulation.
It is also interesting to look at the dependence of the average posterior gamma depending on the number of strategies tested, e.g., the first 10, 20, …, 200 (with the number of simulations again being 1000 after removing 500 burn-out iterations, T1 = 1000, T2 = 1000). Note that the best observed strategy is included in the first ten strategies; however, as expected, the posterior expected gamma, mean, or SR of the best strategy declines with the number of strategies tested (Figure 17). The result of the simulation is expected, i.e., with an increasing number of strategies from which the best one is selected, the expected out-of-sample performance decreases.

4.8. Testing with Modified Mean Returns

In order to better compare the methods, we modify the vector of returns of the strategies while keeping the “natural” correlation structure. Firstly, we increase the strategy 7 return by 19% p.a. while keeping the other returns unchanged so that the strategy 7 with mean over 40% and SR 2.27 stands out among the others (Figure 18), and one expects that it should be identifiable as significant using the various methods.
Table 2 shows the results (for simplicity, focusing only on SR values). Note that, in this case, we are not able to implement the stationary bootstrap since there is no real strategy behind the modified returns of “strategy” 7 and, thus, the row is missing.
All methods confirm that a positive strategy can be selected with CSCV being the most optimistic in terms of SR haircut or probability of loss. Figure 19 indicates that, in this case, there is a fairly good coincidence between the ex ante and ex post SR distributions for all the methods with CSCV again looking the best. Bayes model 2 provides a reasonable estimate of the haircut and the probability of loss, but the estimated “p-value”, i.e., the probability that the selected model is a false discovery is surprisingly high (11%). Nevertheless, it should be noted that, in the MC simulations based on Bayesian posterior parameters, the SR of the strategy 7 might be quite lower than the “observed” value of 40% due to the high return volatility as already mentioned above.
Lastly, we modify the returns of the strategies by deducting the observed mean returns (from the daily strategy returns) and adding random noise means with standard deviation 1% p.a. (Figure 20). Therefore, in this case, we expect the methods to reject the existence of a positive strategy.
All the methods, except for the Bayes Model 1, clearly refute the existence of a positive strategy (Table 3). The surprisingly optimistic results of Bayes Model 1 can be again explained by the volatility incorporated into the Bayes parameter MCMC estimation leading to sampling of models with higher positive means in the MC part of the simulations. The first graph in Figure 21 also clearly demonstrates the strong negative bias of the CSCV method where the best IS model tends to the worst OOS not because of the models but due to the design of the method. See also the IS/OOS scatter plots in Figure 22.

5. Conclusions

The classical methods to adjust single-test p-values for the effect of multiple testing when selecting a trading strategy out of many possibilities such as Bonferroni, Holms, or BHY work relatively well, but provide very conservative estimations due to their approximate nature. Certain improvement can be achieved by applying the independence-based multiple-test p-value (Šidák’s) adjustment or the proposed multivariate normal MC simulation method. The derived expected SR and the related haircut proposed by Harvey and Liu (2015) are rather heuristic and, in our view, not well theoretically founded. The stationary bootstrap method proposed by Sullivan et al. (1999) provides a consistent p-value adjustment. However, if used in a two-stage simulation, it may damage functionality of a positive strategy depending on medium/long-term trends due to the mixing bootstrap algorithm. It also turns out to be the most computationally demanding, since all strategies must be replicated for each sequence of bootstrapped asset prices. Moreover, it cannot be used if the strategies are not known or depend on other economic series. The CSCV method (Bailey et al. 2016) is relatively computationally efficient and provides good results if the mean returns of the strategies are well diversified. However, if the strategies’ mean returns are all close to zero, then the method gives negatively biased results. On the other hand, it appears overoptimistic if one strategy stands high above the others.
Lastly, we proposed and investigated two Bayesian methods, the naïve one based on the simple assumption that the returns are multivariate normal, and the second extended with latent variables indicating zero and nonzero mean return strategies. While the naïve model gives mixed results, the second provides, according to our empirical study, the most consistent results, and it is a useful tool to properly analyze the issue of backtest overfitting. In addition to the probability of loss and backtest overfitting (PBO), it estimates the posterior probabilities of whether each individual strategy is a true discovery and, at the same time, the probability of making a true discovery (and the complementary FDR) and selecting the best one. We believe that the proposed method (Bayesian model 2) provides an efficient way to analyze the effect of backtest overfitting, keeping relatively parsimonious assumptions on the underlying data-generating model. More advanced multivariate stochastic models of the underlying returns might be considered in further research.

Funding

This research was supported by the Czech Science Foundation Grant 18-05244S and by the VSE institutional grant IP 100040.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The publicly available data (daily S&P 500 index values and returns for the period 5 June 2009–24 May 2013) were obtained from https://finance.yahoo.com.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

The following are proofs of the formulas used for Model 2 in Section 3:
Proof of Step 2.
p ( μ | Σ , γ ,   data ) φ ( Γ μ ; m , A ) φ ( μ ; m 0 , D )
exp ( 1 2 ( μ ( Γ A 1 Γ + D 1 ) μ 2 μ ( Γ A 1 m + D 1 m 0 ) ) )
φ ( μ ; ( Γ A 1 Γ + D 1 ) ( Γ A 1 m + D 1 m 0 ) ,   ( Γ A 1 Γ + D 1 ) 1 ) .
Proof of Step 3.
Again
p ( γ | Σ , μ ,   data ) φ ( Γ μ ; m , A ) p ( γ ) ,
where Γ = d i a g ( γ ) and p ( γ ) = i p 0 γ i ( 1 p 0 ) 1 γ i .
Since we can sample γ i { 0 , 1 } step by step given γ j , for j i , it is enough to calculate
p ( γ i = 0 | ) φ ( Γ 0 μ ; m , A ) p ( γ ) exp ( 1 2 ( ( Γ 0 μ m ) A 1 ( Γ 0 μ m )   ) ) ( 1 p 0 ) ,
and similarly for p ( γ i = 1 | ) . We prefer the expression on the right-hand side of the relation above in order to avoid a numerical underflow problem that appears for a higher dimension if the full multivariate density function is used. □
Figure A1. MCMC convergence diagnostics for the Models 1 and 2.
Figure A1. MCMC convergence diagnostics for the Models 1 and 2.
Risks 09 00018 g0a1aRisks 09 00018 g0a1b

References

  1. Andrews, Isaiah, and Maximilian Kasy. 2019. Identification of and correction for publication bias. American Economic Review 109: 2766–94. [Google Scholar]
  2. Bailey, David H., Jonathan Borwein, Marcos Lopez de Prado, and Qiji Jim Zhu. 2014. Pseudomathematics and financial charlatanism: The effects of backtest over fitting on out-of-sample performance. Notices of the AMS 61: 458–71. [Google Scholar] [CrossRef] [Green Version]
  3. Bailey, David H., Jonathan Borwein, Marcos Lopez de Prado, and Qiji Jim Zhu. 2016. The Probability of Backtest Overfitting. Available online: https://ssrn.com/abstract=2840838 (accessed on 21 June 2016).
  4. Chen, Andrew Y., and Tom Zimmermann. 2020. Publication bias and the cross-section of stock returns. The Review of Asset Pricing Studies 10: 249–89. [Google Scholar] [CrossRef] [Green Version]
  5. De Prado, Marcos Lopez. 2015. The future of empirical finance. The Journal of Portfolio Management 41: 140–44. [Google Scholar] [CrossRef]
  6. Harvey, Campbell R., and Yan Liu. 2013. Multiple Testing in Economics. Available online: https://ssrn.com/abstract=2358214 (accessed on 28 April 2017).
  7. Harvey, Campbell R., and Yan Liu. 2014. Evaluating trading strategies. The Journal of Portfolio Management 40: 108–18. [Google Scholar] [CrossRef]
  8. Harvey, Campbell R., and Yan Liu. 2015. Backtesting. The Journal of Portfolio Management 42: 13–28. [Google Scholar] [CrossRef]
  9. Harvey, Campbell R., and Yan Liu. 2017. Lucky Factors. Available online: https://ssrn.com/abstract=2528780 (accessed on 28 June 2017). [CrossRef]
  10. Harvey, Campbell R., and Yan Liu. 2020. False (and missed) discoveries in financial economics. The Journal of Finance 75: 2503–53. [Google Scholar] [CrossRef]
  11. Harvey, Campbell R., Yan Liu, and Heqing Zhu. 2016. … and the cross-section of expected returns. The Review of Financial Studies 29: 5–68. [Google Scholar] [CrossRef] [Green Version]
  12. López de Prado, Marcos, and Michael J. Lewis. 2019. Detection of false investment strategies using unsupervised learning methods. Quantitative Finance 19: 1555–65. [Google Scholar] [CrossRef]
  13. Lynch, Scott M. 2007. Introduction to Applied Bayesian Statistics and Estimation for Social Scientists. Berlin/Heidelberg: Springer, p. 359. [Google Scholar]
  14. Politis, Dimitris N., and Joseph P. Romano. 1994. The stationary bootstrap. Journal of the American Statistical Association 89: 1303–13. [Google Scholar]
  15. Scott, James G., and James O. Berger. 2006. An exploration of aspects of Bayesian multiple testing. Journal of Statistical Planning and Inference 136: 2144–62. [Google Scholar] [CrossRef]
  16. Streiner, David L., and Geoffrey R. Norman. 2011. Correction for multiple testing: Is there a resolution? Chest 140: 16–18. [Google Scholar] [CrossRef] [PubMed]
  17. Sullivan, Ryan, Allan Timmermann, and Halbert White. 1999. Data-snooping, technical trading rule performance, and the bootstrap. The Journal of Finance 54: 1647–91. [Google Scholar] [CrossRef]
  18. White, Halbert. 2000. A reality check for data snooping. Econometrica 68: 1097–126. [Google Scholar] [CrossRef]
1
By an investment strategy, we mean a rule that dynamically determines a portfolio of assets that can be long or short according to information available at the beginning of the period t over which the investment is held. The return over the period is net of the cost of borrowings needed to set up the portfolio (i.e., the strategy is self-financing). For example, an investment strategy may just determine at the end of each trading day whether a long or short position is taken in a specific index futures contract for the next day.
2
A strategy with a significant negative t-ratio can be considered as a discovery as well since we can revert it in order to achieve systematic positive returns.
3
For example, 18% annualized volatility of returns is translated to 18 % 252 / 100 9 % volatility of the posterior annualized mean return.
Figure 1. Future performance of the best strategy selected on the basis of past data.
Figure 1. Future performance of the best strategy selected on the basis of past data.
Risks 09 00018 g001
Figure 2. Stationary bootstrap process where the arrows indicate sampling of returns from the original series.
Figure 2. Stationary bootstrap process where the arrows indicate sampling of returns from the original series.
Risks 09 00018 g002
Figure 3. An example of the combinatorial symmetric cross-validation (CSCV) combination for S = 6 where green blocks of returns are sampled as in-sample and blue blocks are sampled as out-of-sample.
Figure 3. An example of the combinatorial symmetric cross-validation (CSCV) combination for S = 6 where green blocks of returns are sampled as in-sample and blue blocks are sampled as out-of-sample.
Risks 09 00018 g003
Figure 4. Two-step Bayesian simulation (Markov chain Monte Carlo (MCMC) parameter estimation and Monte Carlo (MC) data simulation).
Figure 4. Two-step Bayesian simulation (Markov chain Monte Carlo (MCMC) parameter estimation and Monte Carlo (MC) data simulation).
Risks 09 00018 g004
Figure 5. Annualized means of the selected 200 strategies and the Sharpe ratios’ probability distributions.
Figure 5. Annualized means of the selected 200 strategies and the Sharpe ratios’ probability distributions.
Risks 09 00018 g005
Figure 6. Sharpe ratio’s probability density according to the null hypothesis and the multivariate normal MC simulation.
Figure 6. Sharpe ratio’s probability density according to the null hypothesis and the multivariate normal MC simulation.
Risks 09 00018 g006
Figure 7. Sharpe ratio’s probability density according to White´s reality check.
Figure 7. Sharpe ratio’s probability density according to White´s reality check.
Risks 09 00018 g007
Figure 8. Stationary bootstrap simulation Sharpe ratio and the mean return ex ante (blue) and ex post (red) probability densities.
Figure 8. Stationary bootstrap simulation Sharpe ratio and the mean return ex ante (blue) and ex post (red) probability densities.
Risks 09 00018 g008
Figure 9. Stationary bootstrap simulation histogram of the OOS relative rank of the best IS strategy.
Figure 9. Stationary bootstrap simulation histogram of the OOS relative rank of the best IS strategy.
Risks 09 00018 g009
Figure 10. CSCV simulation of the ex ante (blue) and ex post (red) probability densities.
Figure 10. CSCV simulation of the ex ante (blue) and ex post (red) probability densities.
Risks 09 00018 g010
Figure 11. CSCV histogram of the out-of-sample relative rank of the best IS strategy.
Figure 11. CSCV histogram of the out-of-sample relative rank of the best IS strategy.
Risks 09 00018 g011
Figure 12. Bayes model 1 simulation of the ex ante and ex post probability densities.
Figure 12. Bayes model 1 simulation of the ex ante and ex post probability densities.
Risks 09 00018 g012
Figure 13. Bayes model 1 histogram of the out-of-sample relative rank of the best IS strategy.
Figure 13. Bayes model 1 histogram of the out-of-sample relative rank of the best IS strategy.
Risks 09 00018 g013
Figure 14. Bayes model 2 simulation of the ex ante (blue) and ex post (red) probability densities.
Figure 14. Bayes model 2 simulation of the ex ante (blue) and ex post (red) probability densities.
Risks 09 00018 g014
Figure 15. Bayes model 2 histogram of the out-of-sample relative rank of the best IS strategy.
Figure 15. Bayes model 2 histogram of the out-of-sample relative rank of the best IS strategy.
Risks 09 00018 g015
Figure 16. Posterior gamma average values for the 200 strategies.
Figure 16. Posterior gamma average values for the 200 strategies.
Risks 09 00018 g016
Figure 17. Estimated values of the average ex post gamma, mean return, and SR depending on the number of strategies tested.
Figure 17. Estimated values of the average ex post gamma, mean return, and SR depending on the number of strategies tested.
Risks 09 00018 g017
Figure 18. The modified annualized mean returns and Sharpe ratios of the 200 strategies.
Figure 18. The modified annualized mean returns and Sharpe ratios of the 200 strategies.
Risks 09 00018 g018
Figure 19. The simulations of the ex ante (blue) and ex post (red) SR probability densities.
Figure 19. The simulations of the ex ante (blue) and ex post (red) SR probability densities.
Risks 09 00018 g019
Figure 20. The modified annualized mean returns and Sharpe ratios of the 200 strategies.
Figure 20. The modified annualized mean returns and Sharpe ratios of the 200 strategies.
Risks 09 00018 g020
Figure 21. The simulations of the ex ante and ex post SR probability densities.
Figure 21. The simulations of the ex ante and ex post SR probability densities.
Risks 09 00018 g021
Figure 22. Scatter plots of IS versus OOS Sharpe ratios generated by the three models.
Figure 22. Scatter plots of IS versus OOS Sharpe ratios generated by the three models.
Risks 09 00018 g022
Table 1. Summary of the backtest overfitting tests’ results.
Table 1. Summary of the backtest overfitting tests’ results.
Adjusted p-Value (FDR)Ex Ante av. SR/MeanAdjusted Expected SR/MeanSR/Mean Hair CutProbability of LossMean OOS RankPBO
Boferroni method1.001.1990100%--
Šidák’s correction0.9681.1990.0298.3%--
Mult. norm. MC adj.0.3521.1990.466861%--
Stationary bootstrap0.7281.110/0.1940.297/0.05173.2%/74%0.32855%0.444
CSCV-1.382/0.2440.336/0.05875.7%/76.4%0.37166.8%0.323
Bayes mod. 1-2.102/0.3711.014/0.18051.8%/51.5%0.17175.7%0.168
Bayes mod. 20.5491.201/0.2130.211/0.03782.5%/82.5%0.38060%0.395
Table 2. Summary of the test results.
Table 2. Summary of the test results.
Adjusted p-Value (FDR)Ex Ante av. SRAdjusted Expected SRHair CutProbability of LossMean RankPBO
Boferroni method0.00142.27071.612129%--
Šidák’s correction0.00142.27071.612229%--
Multivariate norm. MC adj.0.00042.27071.78321.5%--
CSCV-2.2572.1962.7%0.01498.5%0.01
Bayes mod. 1-2.5491.88726%0.09287.1%0.087
Bayes mod. 20.111.7761.44218.8%0.06792.1%0.059
Table 3. Summary of the test results.
Table 3. Summary of the test results.
Adjusted p-Value (FDR)Ex Ante av. SRAdjusted Expected SRHair CutProbability of LossMean RankPBO
Boferroni method10.15360100%--
Šidák’s correction10.15360100%--
Multivariate norm. MC adj.0.9970.15360.00298.7%--
CSCV-1.027−0.960193.5%1.001.2%1
Bayes mod. 1-1.9750.60769.3%0.28667.1%0.265
Bayes mod. 20.8871.1270.06294.5%0.44753.2%0.450
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Witzany, J. A Bayesian Approach to Measurement of Backtest Overfitting. Risks 2021, 9, 18. https://doi.org/10.3390/risks9010018

AMA Style

Witzany J. A Bayesian Approach to Measurement of Backtest Overfitting. Risks. 2021; 9(1):18. https://doi.org/10.3390/risks9010018

Chicago/Turabian Style

Witzany, Jiří. 2021. "A Bayesian Approach to Measurement of Backtest Overfitting" Risks 9, no. 1: 18. https://doi.org/10.3390/risks9010018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop