Abstract
This is a simulation-based warning note for practitioners who use the unit root tests in the context of structural change using different selection lag length criteria. With , we find severe oversize problems when using some criteria, while other criteria produce an undersizing behavior. In view of this dilemma, we do not recommend using these tests. While such behavior tends to disappear when , it is important to note that most empirical applications use smaller sample sizes such as or . The test does not present an oversizing or undersizing problem. The only disadvantage of the test arises in the presence of negative correlation, in which case the tests are preferable, but in all other cases they are very undersized. When there is a break in the series, selecting the breakpoint using the Supremum method greatly improves the results relative to the Infimum method.
Keywords:
unit root tests; structural change; truncation lag; GLS detrending; information criteria; sequential general to specific t-sig method JEL Classification:
C22; C52
1. Introduction
Testing for the presence of a unit root in a time series (i.e., whether or not a structural change can be identified) is now a common starting point in advanced models frequently used in macroeconomics and finance. Recent efficient unit root tests are the and the tests proposed by Elliott et al. (1996), and the tests proposed by Ng and Perron (2001).1 All these (GLS-based) tests have been extended to the unit root with one unknown structural change as suggested by Perron and Rodríguez (2003), who show that these tests enjoy the same efficiency characteristics. tests have become increasingly popular in the literature. For example, Haldrup and Jansson (2006) argue that practitioners should abandon the use of tests altogether in favor of tests because of their excellent size properties and nearly optimal power properties. However, this note arrives at the opposite conclusion, suggesting that the choice of the most suitable testing method should be carefully assessed.
Currently, it is widely accepted that the selection of the lag length (denoted by k) has important implications for the (size and power) behavior of the different unit root tests. See, for instance, Schwert (1989), Ng and Perron (1995), Agiakloglou and Newbold (1992), Agiakloglou and Newbold (1996), Elliott et al. (1996), Ng and Perron (2001), Del Barrio Castro et al. (2011), and Fossati (2012). The consensus is to use data-dependent methods. These rules include AIC (Akaike Information Criterion), BIC (Bayesian Information Criterion), Modified AIC (MAIC), Modified BIC (MBIC), and the t-sig method, which are briefly explained below.
Recently, we performed a routine empirical application of the tests and obtained strange results. For example, applying the and the AIC method to the labor market of the Spanish region of Cantabria,2 we obtained an unemployment rate of −3’140,463, a huge (explosive) negative value with . Using the t-sig procedure , we obtained −50’078,041 with , which is even more impressive. A straightforward interpretation implies an overwhelming rejection of the null hypothesis, given any of the asymptotic or finite critical values tabulated in Perron and Rodríguez (2003). However, it is clear that the magnitude of this value is counter-intuitive and inadmissible, because its magnitude is very far from standard values. In contrast, other rules yield opposite results (very small values in absolute value). When applied to other three time series (unemployment rates in the Spanish regions of Galicia and Murcia, and to Peru’s monetary policy rate), similar results are obtained.3 In consequence, we consider that it is worth analyzing the source of the poor behavior of the tests in the cases mentioned above. Hence, we perform extensive finite sample simulations for the tests using different lag-length criteria, where the size performance is our primary interest.
This note (to our best knowledge) represents the first simulation-based attempt to study the size and the eccentric behavior of the unit root tests in the context of structural change. We do not pretend to perform an exhaustive analysis of each rule. Rather, this document is only a simulation-based note of caution for users of these unit root tests.4
This note is structured as follows. In Section 2, the GLS approach with structural break, the test statistics, the rules used to select k, and the two methods to select the break date are briefly reviewed. In Section 3 we present simulation evidence about the size of the test linking the results with an explosive behavior of the test. Section 4 provides some conclusions.
2. DGP, GLS Detrending, Tests with Structural Change, Rules for Selecting the Lag Length, and Methods for Selecting the Breakpoint
2.1. The DGP
Following Perron and Rodríguez (2003), the data generating process (DGP) is:
for , where , , that is, is an unobserved stationary zero-mean process, where and is a martingale difference sequence. We assume that throughout, although the results generally hold for the weaker requirement that (even as . The process has a non-normalized spectral density at frequency zero given by , where .
In the first equation of (1), , where is a set of deterministic components. Perron and Rodríguez (2003) consider two models in the context of an unknown structural break: (i) Model I, where there is a single structural change in the slope, that is, where is the indicator function and is the time of change and can be expressed as a fraction of the whole sample as for some ; and (ii) Model II, which includes a single structural change in intercept and slope, that is, 5
2.2. GLS Detrending and Statistics
The class of tests are due to Stock (1999) and further analyzed by Perron and Ng (1996). These tests are shown to have far less size distortions in the presence of important negative serial correlation. The tests are constructed using , where , with , and , for , and for a chosen and where has been defined in Section 2.1. We also use the test, as defined in Perron and Rodríguez (2003). Hence, defining for , the and the are:
Following Perron and Rodríguez (2003), we use .6 The statistics are modified versions of the test of Phillips and Perron (1988), Bhargava (1986)’s statistic, and the test proposed by Phillips and Perron (1988), respectively. The term is an autoregressive estimate of ( times) the spectral density at frequency zero of , suggested by Perron and Ng (1998), and defined by , where , with and obtained from the autoregression:
Another test is the so-called test, which is the t-statistic for testing the null hypothesis that in (2).
2.3. Rules for Selecting the Lag Length (k)
In the derivation of the asymptotic distributions of the different unit root tests, the theoretical conditions provide little practical guidance for choosing k. The literature suggests to use data-dependent rules like the and the where k is chosen by minimizing: , where , is the penalty attached to an additional regressor, and is the number of observations effectively available.7 The and the are obtained when and , respectively. Another procedure is the sequential t-sig procedure described in Campbell and Perron (1991).8 Selecting a value for , the lag k is selected in a general to specific recursive procedure based on a two-tailed t-statistic on the coefficient associated with the last lag in (2). This approach is denoted by t-sig(10). In a more recent contribution, Ng and Perron (2001) proposed a class of Modified Information Criteria () that selects k satisfying: arg, with . The modified Akaike () is obtained when , and the modified () is obtained when . Recently, in order to improve finite (size and power) sample performance, Perron and Qu (2007) have proposed a hybrid approach consisting of two steps: (i) detrended data are used to select k using , , or ; and (ii) estimating (2) using detrended data to construct . In the simulations, we consider this hybrid approach and the methods used are classified as , , and , respectively.
2.4. Selecting the Breakpoint
Given that the break date () is considered to be unknown, we follow Perron and Rodríguez (2003) using two methods for selecting the break date. The first is to define the break date as the point that minimizes the statistic in (2). This procedure is known as the Infimum method; see Zivot and Andrews (1992) and Perron and Rodríguez (2003) for further details. The second method is based on the maximum absolute value of the t-statistic associated with the dummy variable of the break in the slope. This procedure is known as the Supremum method, which is equivalent to minimizing the SSR; see Perron (1997) and Perron and Rodríguez (2003) for further details.
3. Finite Sample Simulations
3.1. Setup
The DGP is with three scenarios for the autocorrelation of : (i) the case: ; (ii) the AR(1) case: ; and (iii) the MA(1) case: . For all cases, , 1000 replications, and 250, and and (null hypothesis). We performed extensive simulations for all tests, using both models and both ways to select the break point. We present a selected set of results. We have selected the test as the representative test for the entire family of the tests. Furthermore, the Infimum method is used to select the break date and results are only reported for Model I. All other results or Tables are available upon request.9
3.2. The Problem of Size
Table 1 shows the size of the test for and for the different criteria for selecting k. The , that is, . For the case, the results indicate that the test constructed using and have a size around 3.0%, suggesting an undersized test. Testing based on all ( and versions) seems to be extremely conservative (with an exact size of 0.0%). On the other side, testing constructed with , and the t-sig(10) present values implying an extremely oversized test (22%, 27% and 63%, respectively). This same result appears when we use some fixed values of k (), where sizes go from 43% to 82%. Indeed, the size is greater when the selected k is higher. For the case, very similar results are found. In the case, we observe the standard result that the test is oversized. In fact, when , all selection criteria yield an oversized test. Even when using and , the sizes are 23% and 24%, respectively.
Table 1.
Size of the Test, Model I, .
In Table 2, the results are presented for , where . The values of the distortions decrease, meaning that the explosiveness (oversizing) problem decreases. For the case, the tests constructed with and yield 2.6% and 2.7%, respectively which are very similar when . With and , the test has sizes of 1.7% and 1.6%, respectively which are better than for , but are still very undersized. Tests using the , and t-sig(10) have sizes of 9%, 11.2%, and 37.9%, respectively, which are smaller than the values for , but they still indicate an oversized test, in particular the t-sig(10) criterion. With a fixed k (), sizes are greater when k is higher, although smaller compared with .
Table 2.
Size of the Test, Model I, .
If we increase , the size of the test for higher k values increases considerably. We may emphasize this issue comparing with the same class of test, but without a structural change, that is, with some of the results obtained by Ng and Perron (2001). If we observe Table II.B of Ng and Perron (2001), the for using yields a size of 18% with . In our case, for the same values, we have a size of 62%. With , Ng and Perron (2001) obtain 3.6%, a size close to the nominal size (5%). However, in our case, for this sample we have a size of 19% (Table 2, ). In fact, our simulations suggest that we need in order to obtain a size close to 5% when . The results are surely due to the higher number of deterministic components in our models compared with Ng and Perron (2001). However, our conclusion is that practitioners interested in applying the need a non-trivial number of observations.
A further comparison with Ng and Perron (2001) is possible if we select k using different criteria. Again, in the case, where and , the test constructed with and yields sizes of 23% and 24%, respectively. The versions of these criteria yield 32% and 33%, respectively (see Table 1). However, in the case shown in Table VI.A of Ng and Perron (2001), sizes of 5.9% are obtained using and 12.3% using (). In Table 2, for , the tests constructed with and yield sizes of 3.8% and 4.6%, respectively. In the case of Ng and Perron (2001), and yield 1.2% and 1.6%, respectively.
3.3. Some Additional Results10
Two values are used in the construction of : and . Available simulations show that the reason why is . That is, when a higher k is selected, it is possible to incur in overparameterization in (2) and . If tends to , then the and statistics tend to and and converge in probability to zero.
Additional simulations show a link between the excessive size of the test and a high probability of selecting higher values of k. Following Ng and Perron (1995), we examine the number of times that is selected by each rule for and . In the case, the results show that , , and have probabilities to select of 56.2%, 93.2%, 74.4%, and 81.6%, respectively. The t-sig(10) criterion has probabilities of selecting lag lengths that are equally distributed for all values of k. For instance, the recursive t-sig(10) has a probability of around 53% of selecting . Until now, a basic conclusion is that the , , and t-sig(10) methods are not recommended, as they have high probabilities of selecting higher values of k, which are associated with the size distortions observed in Table 1 and Table 2.
When we calculate the mean value for (in the case), explosive negative values are obtained for in , , and t-sig(10). In contrast, reduced values of the test (in absolute value) are given by , , and . We also examine the number of times that the test is smaller than a threshold. We consider six possible values: , and the case. For all thresholds considered, we find that the number of explosive values of increases as the value of k is larger. For example, for , the probability of getting a value of is 13.4%; and the probabilities for and are 31% and 40.2%, respectively. Furthermore, the probabilities of finding values of are 18% and 22.7% for and , respectively.
All previous results are less severe when . Among other things, the probabilities of finding elevated k values are lower. In this regard, the oversizing problem is attenuated (see Table 2). Moreover, when a break is included in the simulations, the improvement is greater when . However, explosive negative values are still observed when the lag is selected with , , and t-sig (10).
3.4. The Statistic
While the test (and the entire family of the tests) shows either oversizing or undersizing problems, depending on the criteria used to choose k, the statistic works well. In the available Tables, we find that the mean value for is not explosive irrespective of the selection criterion used. There are some slightly large negative values when , but it is a standard result in the literature.
Table 3 shows the exact size of the statistic when . For the case, the tests constructed with and yield sizes of 3.1% and 3.4%, respectively; that is, they are slightly undersized, but closer to 5%. A similar observation is valid for and . Other information criteria, like , and t-sig(10), generate oversized tests; but the values are much smaller compared with Table 1 for the test. For example, for the t-sig (10) procedure, Table 1 ( case) shows that the statistic has a size of 63%, which is poor. However, this value is reduced to 14.6% in the case of the test (Table 3). In general, the values in all scenarios are smaller compared with Table 1 for . The only difference (as expected) arises when . In this case, the test has sizes of 23% and 24% for the and , respectively, while for the test the values are 31.5% and 32.6%, respectively.
Table 3.
Size () of Test, Model I, .
Table 4 shows the exact size of the test when . Again, the size distortions are clearly smaller compared to those of the test (Table 2). As in Table 3, the results using the test are better when . In Table 4, the test yields 11.5% and 12.8% when and are used, respectively. In the case of the test, the values are 3.8% and 4.6%, respectively. Furthermore, our calculations show that the test will have a size closer to 5% for when . This sample size is even more prohibitive for most empirical applications.
Table 4.
Size () of Test, Model I, .
3.5. The Supremum Method and a Single Breakpoint
The results change favorably when the Supremum method is used to select the breakpoint. Several simulations have been performed under the setup of Section 3.1 for Model I: with two scenarios: (i) , that is, no break; and (ii) with . Similar experiments have been performed for Model II. In the first case, the test still has explosive values, although less frequently; and the values are negative but of a smaller magnitude (in absolute value) than when using the Infimum method. In the second case, the results show considerable improvement, especially when . The explosive values of the test practically disappear for the and rules, although the cost is to have small values (in absolute value), which produce a conservative test. On the other hand, the rules , , and t-sig(10) continue to present an test with explosive values which, however, are very small compared to the previous cases, and occur only when a higher k is selected.
The best results with the Supremum method are important, since this method is recommended in the literature to select the break date. For instance, Vogelsang and Perron (1998) argue that this method is to be preferred, since it allows a consistent estimate of the breaking point, a matter that the Infimum method cannot do.
The evidence suggests that, in the empirical applications, the Supremum method should be used to select the breakpoint along with the and rules, although the potential cost is to have a conservative test. The evidence suggests avoiding the use of rules such as , , and t-sig(10) to select k, as well as the use of the Infimum method to select the breakpoint.
4. Conclusions
This note aims to examine the performance of the size of the statistics to test for the presence of a unit root using different lag length selection criteria in the context of an unknown structural change. In particular, we have focused on the size performance of the test. Overall, the results show that there is a strong relationship between the explosive negative values of the test and the values of the selected k. Using the Infimum method to select the break point jointly with some rule, such as , or t-sig(10), produces the worst scenario, in the sense that the test yields explosive negative values, which generates severe oversizing problems. On the opposite side, using other criteria for k implies conservative tests. These issues seem to improve when (relative to ) or more, which creates sample size difficulties for most macroeconomic applications, especially in Latin American countries.
The results indicate that should be used, because it does does not result in explosiveness. Although for other reasons, this recommendation is in the same vein as Harvey et al. (2013). The advantage of the test is that it is intrinsically conservative. So, if we obtain a good size when , this is achieved at the cost of having an undersized test in the other cases, including the case. Our results are in line with those obtained in Del Barrio Castro et al. (2011), Del Barrio Castro et al. (2013), and Del Barrio Castro et al. (2015)11.
The results change for the better when using the Supremum method (minimizing the SSR) to select the breakpoint. However, this result only occurs when there is a break in the series. With this method, the test values are reduced (in absolute value) and no explosiveness is observed. Furthermore, the advantage is that the method offers a consistent breakpoint estimator which is currently suggested in the literature. Although a possible undersizing problem is addressed, then a possible best scenario is to use the Supremum method together with rules for selecting k such as MIC. This potential need to perform a pre-testing to see the existence of a break is similar to what is proposed by Kim and Perron (2009) when there is only one break and the proposal of Carrión-i-Silvestre et al. (2009) when there are multiple breaks.
Author Contributions
Both authors contributed eaully to the paper.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Agiakloglou, Christos, and Paul Newbold. 1992. Empirical Evidence on Dickey-Fuller-Type Tests. Journal of Time Series Analysis 13: 471–83. [Google Scholar] [CrossRef]
- Agiakloglou, Christos, and Paul Newbold. 1996. The balance between size and power in Dickey-Fuller tests with data-dependent rules for the choice of truncation lag. Economics Letters 52: 229–34. [Google Scholar] [CrossRef]
- Bhargava, Alok. 1986. On the theory of testing for unit root in observed time series. Review of Economic Studies 53: 369–84. [Google Scholar] [CrossRef]
- Campbell, John Y., and Pierre Perron. 1991. Pitfalls and Opportunities: What Macroeconomists Should Know about Unit Roots. In NBER Macroeconomics Annual. Edited by Olivier J. Blanchard and Stanley Fischer. New York: MIT Press, Volume 6, pp. 141–201. [Google Scholar]
- Carrión-i-Silvestre, Josep Lluís, Dukpa Kim, and Pierre Perron. 2009. GLS-based Unit Root Tests with Multiple Structural Breaks both Under the Null and Alternative Hypotheses. Econometric Theory 25: 1754–92. [Google Scholar] [CrossRef]
- Choi, In. 2015. Almost All about Unit Roots. Series: Themes in Modern Econometrics; New York: Cambridge University Press. [Google Scholar]
- Del Barrio Castro, Tomás, Paulo M. M. Rodrigues, and A. M. Robert Taylor. 2011. The Impact of Persistent Cycles on Zero Frequency Unit Root Tests. Working Paper 24. Lisbon: Banco de Portugal. [Google Scholar]
- Del Barrio Castro, Tomás, Paulo M. M. Rodrigues, and A. M. Robert Taylor. 2013. The Impact of Persistent Cycles on Zero Frequency Unit Root Tests. Econometric Theory 29: 1289–313. [Google Scholar] [CrossRef]
- Del Barrio Castro, Tomás, Paulo M. M. Rodrigues, and A. M. Robert Taylor. 2015. On the Behaviour of Phillips-Perron Tests in the Presence of Persistent Cycles. Oxford Bulletin of Economics and Statistics 77: 495–511. [Google Scholar] [CrossRef]
- Elliott, Graham, Thomas J. Rothenberg, and James H. Stock. 1996. Efficient tests for an autoregressive unit root. Econometrica 64: 813–39. [Google Scholar] [CrossRef]
- Fossati, Sebastian. 2012. Covariate unit root tests with good size and power. Computational Statistics and Data Analysis 56: 3070–79. [Google Scholar] [CrossRef]
- Haldrup, Niels, and Michael Jansson. 2006. Improving Size and Power in Unit Root Testing. In Palgrave Handbook of Econometrics, Volume 1: Econometric Theory. Edited by Terence C. Mills and Kerry Patterson. Basingstoke: Palgrave Macmillan, pp. 252–77. [Google Scholar]
- Hall, Alastair R. 1994. Testing for a Unit Root in Time Series with Pretest Data-Based Model Selection. Journal of Business & Economic Statistics 12: 461–70. [Google Scholar] [CrossRef]
- Harvey, David I., Stephen J. Leybourne, and A. M. Robert Taylor. 2013. Testing for unit roots in the possible presence of multiple trend breaks using minimum Dickey-Fuller statistics. Journal of Econometrics 177: 265–84. [Google Scholar] [CrossRef]
- Kim, Dukpa, and Pierre Perron. 2009. Unit root tests allowing for a break in the trend function at an unknown time under both the null and alternative hypotheses. Journal of Econometrics 148: 1–13. [Google Scholar] [CrossRef]
- Maddala, Gangadharrao S., and In-Moo Kim. 1998. Unit Roots Cointegration, and Structural Change. Cambridge: Cambridge University Press. [Google Scholar]
- Ng, Serena, and Pierre Perron. 1995. Unit root tests in ARMA models with data dependent methods for the selection of the truncation lag. Journal of the American Statistical Association 90: 268–81. [Google Scholar] [CrossRef]
- Ng, Serena, and Pierre Perron. 2001. Lag length selection and the construction of unit root tests with good size and power. Econometrica 69: 1519–54. [Google Scholar] [CrossRef]
- Ng, Serena, and Pierre Perron. 2005. A note on the selection of time series models. Oxford Bulletin of Economics and Statistics 67: 115–34. [Google Scholar] [CrossRef]
- Perron, Pierre. 1989. The great crash, the oil price shock and the unit root hypothesis. Econometrica 57: 1361–401. [Google Scholar] [CrossRef]
- Perron, Pierre. 1997. Further evidence of breaking trend functions in macroeconomic variables. Journal of Econometrics 80: 355–85. [Google Scholar] [CrossRef]
- Perron, P. 2006. Dealing with Structural Breaks. In Palgrave Handbook of Econometrics, Volume 1: Econometric Theory. Edited by Terence C. Mills and Kerry Patterson. Basingstoke: Palgrave Macmillan, pp. 278–352. [Google Scholar]
- Perron, Pierre, and Serena Ng. 1996. Useful Modifications to Unit Root Tests with Dependent Errors and their Local Asymptotic Properties. Review of Economic Studies 63: 435–65. [Google Scholar] [CrossRef]
- Perron, Pierre, and Serena Ng. 1998. An Autoregressive Spectral Density Estimator at Frequency Zero for Nonstationarity Tests. Econometric Theory 14: 560–603. [Google Scholar] [CrossRef]
- Perron, Pierre, and Zhongjun Qu. 2007. A simple modification to improve the infinite sample properties of Ng and Perron’s unit root tests. Economics Letters 94: 12–19. [Google Scholar] [CrossRef]
- Perron, Pierre, and Gabriel Rodríguez. 2003. GLS Detrending, Efficient Unit Root Tests and Structural Change. Journal of Econometrics 115: 1–27. [Google Scholar] [CrossRef]
- Phillips, Peter C. B., and Pierre Perron. 1988. Testing for a unit root in time series regression. Biometrika 75: 335–46. [Google Scholar] [CrossRef]
- Phillips, Peter C. B., and Zhijie Xiao. 1998. A primer on unit root testing. Journal of Economic Surveys 12: 423–70. [Google Scholar] [CrossRef]
- Quineche, Ricardo, and Gabriel Rodríguez. 2015. Data-Dependent Methods for the Lag Length Selection in Unit Root Tests with Structural Change. Working Paper 404. Lima, Peru: Department of Economics, Pontificia Universidad Católica del Perú. [Google Scholar]
- Rodríguez, Gabriel. 2007. Finite Sample Behaviour of the Level Shift Model using Quasi-Differenced Data. Journal of Statistical Computation and Simulation 77: 889–905. [Google Scholar] [CrossRef]
- Schwert, G. William. 1989. Tests for unit roots: A Monte Carlo Investigation. Journal of Business and Economics Statistics 7: 147–59. [Google Scholar] [CrossRef]
- Stock, James H. 1994. Unit Roots, Structural Breaks and Trends. In Handbook of Econometrics. Edited by Daniel L. McFaden and Robert F. Engle. Amsterdam: Elsevier, vol. IV, pp. 2740–841. [Google Scholar]
- Stock, James H. 1999. A Class of Tests for Integration and Cointegration. In Cointegration, Causality and Forecasting. A Festschrift in Honour of Clive W. J. Granger. Edited by Robert F. Engle and Halbert White. Oxford: Oxford University Press, pp. 137–67. [Google Scholar]
- Vogelsang, Timothy J., and Pierre Perron. 1998. Additional Tests for a Unit Root Allowing the Possibility of Breaks in the Trend Function at an Unknown Time. International Economic Review 39: 1073–100. [Google Scholar] [CrossRef]
- Zivot, Eric, and Donald W. K. Andrews. 1992. Further evidence on the great crash, the oil-price shock and the unit root hypothesis. Journal of Business and Economics Statistics 10: 251–70. [Google Scholar] [CrossRef]
| 1. | For excellent surveys, see Stock (1994), Maddala and Kim (1998), Phillips and Xiao (1998), Haldrup and Jansson (2006), Perron (2006), and Choi (2015). |
| 2. | Quartely data covering the period Q3 1976–Q2 2012 ( observations). |
| 3. | The sample size for Galicia and Murcia are the same as for Cantabria. For Peru’s monetary policy rate, the data are monthly for February 2002–August 2010 ( observations). |
| 4. | We recognize the limitations of this note, which is only based on simulations. We agree with a Referee that formal proofs are needed in the spirit of Del Barrio Castro et al. (2013). Hence, further work in the direction of a formal treatment will be addressed in a future research project. |
| 5. | See Rodríguez (2007) for the crash model proposed by Perron (1989). |
| 6. | Following Elliott et al. (1996) and Ng and Perron (2001), the parameter is selected in such a way that 50% of the Gaussian power envelope is attained. |
| 7. | Note that in all experiments we use as the available number of observations, which is fixed, as suggested by Ng and Perron (2005). |
| 8. | See also Hall (1994) and Ng and Perron (1995). |
| 9. | We are agree with the Editor that our scenario is the worst possible scenario because we are using the Infimum method jointly (in some cases) with the t-sig(10) rule. However, this worst scenario is widely used in typical empirical applications. Furthermore, it is a regular or natural option in many statistical packages used by practitioners. Minimizing SSR (or Supremum) is better, as we mention later. |
| 10. | We present a summary of the Tables from the Working Paper version of this Note (see Quineche and Rodríguez (2015)). All other tables are available upon request. |
| 11. | They explain both issues (in particular the undersizing feature) in the context of time series admitting for (near-) unit roots at cyclical frequencies. They suggest that the degree of undersizing is worst when is used. The problem is aggravated if detrended data are used. See these references for further details. |
© 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).