Abstract
This paper considers estimation and inference concerning the autoregressive coefficient () in a panel autoregression for which the degree of persistence in the time dimension is unknown. Our main objective is to construct confidence intervals for that are asymptotically valid, having asymptotic coverage probability at least that of the nominal level uniformly over the parameter space. The starting point for our confidence procedure is the estimating equation of the Anderson–Hsiao (AH) IV procedure. It is well known that the AH IV estimation suffers from weak instrumentation when is near unity. But it is not so well known that AH IV estimation is still consistent when . In fact, the AH estimating equation is very well-centered and is an unbiased estimating equation in the sense of Durbin (1960), a feature that is especially useful in confidence interval construction. We show that a properly normalized statistic based on the AH estimating equation, which we call the statistic, is uniformly convergent and can be inverted to obtain asymptotically valid interval estimates. To further improve the informativeness of our confidence procedure in the unit root and near unit root regions and to alleviate the problem that the AH procedure has greater variation in these regions, we use information from unit root pretesting to select among alternative confidence intervals. Two sequential tests are used to assess how close is to unity, and different intervals are applied depending on whether the test results indicate to be near or far away from unity. When is relatively close to unity, our procedure activates intervals whose width shrinks to zero at a faster rate than that of the confidence interval based on the statistic. Only when both of our unit root tests reject the null hypothesis does our procedure turn to the statistic interval, whose width has the optimal rate of shrinkage when the underlying process is stable. Our asymptotic analysis shows this pretest-based confidence procedure to have coverage probability that is at least the nominal level in large samples uniformly over the parameter space. Simulations confirm that the proposed interval estimation methods perform well in finite samples and are easy to implement in practice. A supplement to the paper provides an extensive set of new results on the asymptotic behavior of panel IV estimators in weak instrument settings.
JEL Classification:
C23; C26
1. Introduction
Due to the many challenges that arise in estimating and conducting statistical inference for dynamic panel data models, a vast literature has emerged studying these models over the past three decades. Much has been learnt about the large sample properties and finite sample performance of various estimation procedures in stable dynamic panel models, not only in univariate but also in multivariate contexts. Important contributions to this literature began with (Nickell 1981; Anderson and Hsiao 1981, 1982), followed by (Arellano and Bond 1991; Ahn and Schmidt 1995; Arellano and Bover 1995; Kiviet 1995; Blundell and Bond 1998; Hahn and Kuersteiner 2002; Alvarez and Arellano 2003), amongst many others. Progress has also been made recently in studying such models when more persistent behavior, such as unit root or near unit root behavior, is present. (Phillips and Moon 1999) provided methods that opened up the rigorous development of asymptotics in such models for both stationary and nonstationary cases and with multidimensional joint and sequential limits. Many subsequent contributions to this nonstationary panel literature have considered more complex regressions, analyzing the effects of incidental trends, serial dependence and cross section dependence; e.g., (Chang 2002, 2004; Phillips and Sul 2003; Moon and Phillips 2004; Moon et al. 2014, 2015; Pesaran 2006; Pesaran and Tosetti 2011).
While this literature has greatly enhanced our understanding of the panel data sampling behavior of point estimators and of associated test statistics, such as the Studentized t statistic or the Wald statistic, what have not been studied are confidence interval procedures which areasymptotically valid in the sense that asymptotic coverage probabilities are at least that of the nominal level uniformly over the parameter space. The development of theoretically justified confidence intervals is especially important in cases where the empirical researcher may not have good prior information about the degree of persistence in the data, since in such situations interval estimates can serve as indispensible supplements to point estimates by providing additional information about sampling uncertainty and about the range of possible values of the autoregressive parameter that are consistent with the observed data. Moreover, we know from the unit root time series literature that constructing an asymptotically valid confidence interval for the autoregressive parameter of an process is a challenging task when the parameter space is taken to be large enough to include both the stable and the unit root cases. This is because the Studentized statistic based on OLS estimation is not uniformly convergent in this case, so that an asymptotically correct confidence interval cannot be constructed by inverting the Studentized statistic in the usual way. To address this problem in the time series literature, (Stock 1999) proposed a confidence procedure based on local-to-unity asymptotics, while simulation and bootstrap type methods have been introduced by (Andrews 1993; Hansen 1999). Recent results by (Mikusheva 2007, 2012) and by (Phillips 2014) have shown that the methods of (Andrews 1993; Hansen 1999), as well as a recentered version of Stock’s method, all give the correct asymptotic coverage probability uniformly over the parameter space. Extending these procedures to the panel data setting does not seem to be straightforward, and panel data versions of these methods are currently unavailable.
To address this need, the present paper proposes simple, asymptotically correct confidence procedures for the autoregressive coefficient of a panel autoregression1. We take as our starting point the estimating equation of the (Anderson and Hsiao 1981) IV procedure. Although much has been written about the Anderson–Hsiao IV estimator having a weak instrument problem when is unity or very nearly unity, it should be noted that this estimator is still consistent even when and that the weak instrument problem primarily manifests itself in the form of the asymptotic distribution having greater dispersion2. In fact, the Anderson–Hsiao estimating equation is very well-centered and is an unbiased estimating equation in the sense of (Durbin 1960), a property that is of particular importance in constructing asymptotically valid confidence intervals. Exploiting this unbiasedness property, we then show that a properly normalized statistic based on this estimating equation is uniformly convergent over the parameter space . This statistic, which we refer to as the statistic since it is based on the (empirical) IV moment function, can be easily and analytically inverted to obtain an asymptotically correct confidence interval. However, because of the weak instrument problem, when the true is unity or very near unity, confidence intervals obtained by inverting this statistic may be less informative in the sense that they may be relatively wide in finite samples and, asymptotically, their width shrinks toward zero at the slower rate of even when both the cross section (N) and the time series (T) sample sizes approach infinity. A similar drawback applies to the GMM procedure of (Han and Phillips 2010), which achieves uniforminference with shrinkage rate over the full domain
To obtain more informative interval estimates, we introduce a new confidence procedure which uses information from two different unit root tests, with different power properties, to assess the proximity of the true autoregressive parameter from the exact unit root null hypothesis . More precisely, we infer that the true parameter value is unity or very close to unity if the more powerful of the two unit root tests fails to reject , and we use, in this case, an interval that is localized at , with width that shrinks at a faster rate3. Second, if the more powerful test rejects but the less powerful test fails to reject, we use another interval that is still localized at but with greater width which shrinks at the rate , a rate that is still faster than that of the width of the confidence interval based on the statistic in the vicinity of . Finally, if both tests reject , then we conclude that the true parameter value is far enough away from unity that we can use the confidence interval based on the statistic, whose width shrinks at the optimal rate in the stable region of the parameter space. We show that the asymptotic size of this pretest based procedure can be uniformly controlled, so that this procedure is asymptotically valid, albeit slightly conservative when the underlying process is stable. The degree of conservatism under our procedure is also controllable and can be kept small by carefully controlling the probability of a Type II error under a local-to-unity parameter sequence. Moreover, in addition to providing informative and asymptotically correct confidence intervals, our procedure has the further advantage that it is given in analytical form and, hence, is computationally simple to implement. Simulations confirm that the proposed method performs well in finite samples.
The remainder of the paper proceeds as follows. Section 2 briefly describes the model, assumptions, and notation. Section 3 introduces two new ways of constructing uniform confidence intervals for the parameter . The first is based on inverting the statistic, and the second is the pre-test based confidence interval. Results given in this section show that both confidence procedures are asymptotically valid. Section 4 reports the results of a Monte Carlo study comparing our proposed confidence procedures with some alternative procedures. We provide a brief conclusion in Section 5. Proofs of the main theorems are given in the Appendix A to this paper. Proofs of additional supporting lemmas, as well as additional Monte Carlo results, are reported in an online supplement to this paper (Chao and Phillips 2019). The supplement provides an extensive set of results for panel estimation limit theory in unit root and near unit root cases that help deliver the main results in the paper but are of wider interest regarding asymptotic behavior of panel IV estimators, particularly in weak instrument settings. The supplement also includes additional simulation results concerning the performance of the estimation procedures considered in the paper.
A word on notation. We use ⇒ for convergence in distribution or weak convergence, for convergence in probability, denotes a chi-square random variable with degrees of freedom, denotes the standard normal random variable, and is the standard Brownian motion on the unit interval for each i. For two sequences and , we take to mean and to mean that and , as . Similarly, for random variables and , we take to mean that and , as . In addition, the notations and denote, respectively, a probability measure indexed by the fixed parameter and one indexed by the local-to-unity parameter . Finally, we use wid to denote the width of the confidence interval .
2. Model and Assumptions
We work with the following dynamic panel data model written in unobserved components form
for and . Here, is the observed data, is generated by a latent AR(1) process, denotes an (unobserved) individual effect, and denotes the panel autoregressive parameter, which is assumed to belong in the parameter space . In this paper, we will show that certain properties of our procedure hold uniformly for . We do this by making use of a result (Lemma 2.6.2) from (Lehmann 1999) which establishes the equivalence of uniform convergence and convergence for every parameter sequence belonging to a given parameter space. For this purpose, it is convenient for us to consider a general class of local-to-unity parameter sequences of the form , where is a non-negative function of T such that as 4. Moreover, parameter sequences for stable AR processes can also be written in this general form by considering parameter sequences which belong to the collection
Note also that, in the case where such parameter sequences are considered, the AR process given in expression (2) will depend on an indexed parameter , as opposed to a fixed autoregressive parameter, so that, strictly speaking, the observed data and the latent process, in this case, will be strictly triangular indexed arrays , which depend additionally on T. However, for notational ease, we shall suppress this additional dependence and simply write in what follows.
It is sometimes convenient to rewrite the model (1) and (2) in the alternate familiar form as a first-order autoregressive process in , viz.,
where . The following assumptions are made on the model.
Assumption 1 (Errors):
(a) across i and t, ; (b) .
Assumption 2 (Random Effects):
(a) across i, ; (b) ; (c) and , are mutually independent for all and for all
Assumption 3 (Initialization):
Let. Suppose thatis independent across i. Suppose also that there exists a positive constant C such that, and thatandare independent for alland for all
Note that Assumption 3 on the initial condition does not impose mean stationarity, i.e., the condition that , which in our setup is equivalent to the restriction that In addition, observe that Assumption 3 allows for the case where the initial condition is fixed, i.e., for some sequence of constants such that . It is also general enough so that we may specify to be fixed in the unit root case but allow to be a draw from its unconditional distribution with variance when the underlying process is stationary.
In lieu of Assumption 2, we also consider in this paper a fixed-effects specification given by the following assumption.
Assumption 2* (Fixed Effects):
Let be a nonrandom sequence. Suppose that there exists a positive constant C such that for all N.
All our theoretical results hold under either the random-effects specification given by Assumption 2 or the fixed-effects specification given by Assumption 2*.
3. Uniform Asymptotic Confidence Intervals
3.1. Confidence Intervals Based on the Anderson–Hsiao IV Procedure
A primary objective of this paper is to develop confidence procedures with asymptotic coverage probability that is at least that of the nominal level uniformly over the parameter space . As a first step, we consider a statistic based on the empirical moment function of the Anderson–Hsiao IV procedure, but properly standardized by an appropriate estimator of the scale parameter. In particular, let
where and and where and with . In addition, we let denote any preliminary estimator of that satisfies the following conditions
Assumption 4: Let be an estimator of . Suppose that the following conditions hold for this estimator as such that , for and .
- (a)
- , if for all T sufficient large or if such that ;
- (b)
- , if such that but ;
- (c)
- if , where is given in expression (3) above.
The asymptotic properties of under different parameter sequences are given by the following result.
Theorem 1.
Let Assumptions 1, 3, 4, and either 2 or 2* hold. The following statements hold as such that , for and .
- (a)
- Suppose that , where . Then,
- (b)
- Suppose that , where . Then,
- (c)
- Suppose that , where . Then,where .
- (d)
- Suppose that , where . Then,
- (e)
- Suppose that , where . Then,
To provide some intuition about the statistic and about the conditions placed on the preliminary estimator (i.e., Assumption 4), we set to be the unstandardized version of From the proof of Theorem 1, given in Appendix A, it is evident that can be decomposed into several terms whose orders of magnitude change depending on how close the parameter sequence is to unity. In consequence, the lead term of is not the same in the stable (panel) autoregression case as it is in the case where is very close to unity. On the other hand, when appropriately normalized, this statistic will converge to a standard normal distribution in each case, but this requires a scale estimator that will adapt to variation in the normalization factor under alternative parameter sequences. The estimator
turns out to have these adaptive properties, as shown in Lemma SC-13 and its proof (given in the supplement). An important component in the construction of a proper normalization factor is to have a preliminary estimator with a fast enough rate of convergence, so that the resulting estimator of is consistent under every possible parameter sequence in the parameter space . Examination of the proof of Lemma SC-12 reveals that the conditions needed on are precisely those given in Assumption 45.
It should be noted that the Anderson–Hsiao IV estimator, which we will denote by in this paper, does not satisfy the conditions of Assumption 46. This is because, as shown in Theorem SA-1 of the supplement to this paper, when for all T sufficient large or when such that , so that its rate of convergence is not fast enough in the unit root and near unit root regions of the parameter space. Furthermore, the pooled OLS (POLS) estimator, which we will denote by , also does not satisfy these conditions since it is inconsistent in the stationary region as shown in Theorem SA-2 of the supplement. Hence, in Appendix SA of the supplement to our paper, we introduce a new point estimator, , which is an average of and where the average is taken using a data-dependent weight function that, in turn, depends on a unit root statistic. This estimator turns out to satisfy the conditions of Assumption 4 because it exploits the differential strengths of and in different parts of the parameter space and can place more or less weight on one or the other of these two estimators, depending on the information provided by a preliminary unit root test on the true value of . We use in constructing the scale estimator for the Monte Carlo results reported Section 4 of this paper, but we note that is not the only estimator which satisfies Assumption 4, as both the within-group OLS estimator and the bias-corrected within-group estimator proposed by (Hahn and Kuersteiner 2002) could also be used to obtain a scale estimator with the desired properties, although does have a faster rate of convergence than the uncorrected within-group estimator in the unit root and near unit root cases. Because the focus of this paper is on confidence procedures, and not on point estimation, we will not give technical details of in the body of this paper, but will instead refer interested readers to Appendix SA of the supplement for more details, as well as for formal results, on the rate of convergence of under alternative parameter sequences.
The following theorem shows the uniform convergence of the statistic over the parameter space .
Theorem 2.
Let denote the cdf of a standard normal random variable. Suppose that Assumptions 1, 3, 4, and either 2 or 2* hold. Then, for each ,
as such that for constants and 7.
Remark 1.
- (i)
- Let denote the quantile of the standard normal distribution. A level confidence interval based on the statistic can be taken to beIt is immediate from Theorem 2 that the confidence procedure defined by (5) is asymptotically valid in the sense that its coverage probability is equal to the nominal level in large samples, uniformly over .
- (ii)
- The uniform limit result given in Theorem 2 above is established under a pathwise asymptotic scheme where we take such that for constants κ and . Note that the asymptotic framework employed here does not restrict N and T to follow a specific diagonal expansion path, but rather allows for a whole range of possible paths indexed by κ ; and, hence, our results do not require the kind of restrictions on the relative magnitudes of N and T that are often imposed in other asymptotic analysis of panel data models. Indeed, by allowing T to grow as any positive (real-valued) power of N, our framework can accommodate a wide variety of settings where T may be of smaller, larger, or similar order of magnitude as N.
- (iii)
- As noted earlier and as is evident from the proof of Theorem 2 given in the Appendix A, uniform convergence here is established by showing convergence to the same distribution under every parameter sequence in the parameter space. To the best of our knowledge, the use of this approach in statistics originated with the book on large sample theory by (Lehmann 1999). Important extensions, as well as applications, of this approach to a variety of econometric models and inferential procedures have also been made more recently in the papers by (Andrews and Guggenberger 2009; Andrews et al. 2011).
- (iv)
- As we noted earlier in the Introduction, a primary reason why the statistic is well-behaved is that the (empirical) IV moment function is well-centered as an unbiased estimating equation. In this sense, our approach relates to early work by (Durbin 1960) on unbiased estimatingequations which was applied to time series regression in his original study. Importantly, in dynamic panel data models with individual effects, estimating equations associated with least squares procedures tend not to be as well-centered as the IV estimating equations explaining the need for IV in this context (c.f., Han and Phillips 2010).
- (v)
- A drawback of is that the rate at which the width of this confidence interval shrinks toward zero as sample sizes grow is relatively slow for parameter sequences that are very close to unity. As also noted in the Introduction, this is due to the well-known ‘weak instrument’ problem which induces a slow rate of convergence for the Anderson–Hsiao IV procedure in this case. More precisely, using the results given in Lemmas SA-1, SC-1, and SC-13 in the supplement to this paper, we can easily show that wid when such that , so that the rate of shrinkage here does not even depend on N, even as both N and T go to infinity (see also Phillips 2018). This slower rate of convergence is also reflected in the Monte Carlo results reported in Section 4 below, as the results there show that the average interval width of can be a very substantial fraction of the width of the entire parameter space when . To improve on the performance of , the next subsection introduces a pretest-based confidence procedure which is similarly asymptotically valid but which in addition provides more informative intervals when the underlying process has a unit root or a near unit root.
3.2. A Pretest-Based Confidence Procedure
To enhance the informativeness of the confidence procedure when there is a unit root, we use a pretest approach. The idea is to apply two different unit root tests sequentially to assess the proximity of to unity and then implement different confidence intervals depending on the information about the location of that emerges from these tests. More precisely, we propose the following level confidence interval of the form
where is as defined in (5) above,
and where , , is an indicator function, and we again take to be the quantile of a standard normal distribution for some , with and similarly defined. In addition, we take
with , to be the unit root test statistic based on the POLS estimator; and
with , is a unit root test statistic based on the IV estimator
which was introduced by (Arellano and Bover 1995) and further analyzed in (Blundell and Bond 1998). From expression (6), it is apparent that the confidence procedure follows a sequential tree structure. We first pretest for the presence of a unit root using . If the result of this first test fails to reject the unit root null hypothesis, then we employ the tighter unit root interval Otherwise, we conduct a second test of the unit root null hypothesis using a less powerful test . If this second test fails to reject the null hypothesis, we use the wider unit root interval . On the other hand, if both tests reject the unit root null hypothesis, we then use the interval , which is asymptotically valid but less informative unless the true value of is sufficiently far away from unity.
The next theorem shows that this confidence procedure is asymptotically valid in the sense that its non-converage probability is at most the nominal significance level uniformly over the parameter space under pathwise asymptotics.
Theorem 3.
Let be the specified significance level and let such that for constants and . Set and . Suppose that Assumptions 1, 3, 4, and either 2 or 2* hold. Then,
Remark 2.
- (i)
- The pre-test based confidence procedure proposed here is inspired by the work of (Lepski 1999) who used information from a test procedure to increase the accuracy of confidence sets. The original Lepski paper and subsequent extensions of that paper focused on problems of nonparametric function estimation and canonical versions of such problems, as represented by the many normal means model. Because we deal with a model that differs from the one studied in (Lepski 1999) and because we use a dual pre-test framework, the construction and analysis of our procedure also differ, even though we use the same idea to improve set estimation accuracy.
- (ii)
- Sinceit follows that the result obtained in Theorem 3, i.e., , is equivalent to , so that the proposed confidence interval has asymptotic coverage probability that is at least the nominal level uniformly over .
- (iii)
- In the procedure given by (6), is the significance level for the confidence interval . It is, of course, also the asymptotic non-coverage probability of , since is asymptotically valid.
- (iv)
- As noted in the Introduction and in Remark 3.1 (v) above, a drawback of is that its width shrinks slowly for parameter sequences that are very close to unity. The pre-test confidence procedure seeks to improve on this rate by applying two different unit root tests sequentially and by using the information from these tests to determine whether to use local-to-unity intervals whose width shrinks at a faster rate than when the autoregressive parameter value is in close proximity of unity. To see how this improvement is achieved, note that when the true parameter value is within an neighborhood of unity then, aside from the relatively small probability event of a Type I error, the first unit root test will fail to reject , resulting in the use of the interval When the parameter is this close to unity, wid whereas wid, so that the use of leads to significant improvement over . The reason for a second unit root test using the statistic is that for parameter sequences such that , the first unit root test will reject with probability approaching one as sample sizes grow, but the less powerful unit root test based on will not, subject again to the relatively small probability event of a Type I error. For parameter sequences in this region, wid The result is that we can make further improvement by using the interval which has width wid. Finally, if both these unit root tests reject , then our procedure will infer that the parameter is far enough away from unity to use . Of course, the two unit root tests are subject to Type II errors; but, as explained in Remark 3.2(vi) below, the probability of Type II errors could also be properly controlled under our procedure8.
- (v)
- and , on the other hand, are the significance levels for the unit root tests based on and . Note that, especially in large samples, the specification of and really has more of an impact on the width of the resulting interval than it does on the coverage probability, so that and are not significance levels in the traditional sense. For example, consider the choice of . Observe that a smaller value of leads to a wider . However, the effect of on the width of the interval adopted by the overall procedure could be ambiguous, since, if the null hypothesis of an exact unit root is true, an increase in would reduce the width of but could also lead to a greater chance that will falsely reject the null hypothesis and switch to either or , both of which are wider than in large samples. A similar argument shows that it is also difficult to predict a priori the effect of varying on the width of the resulting interval. On the other hand, note that, except for pathological specifications where and/or (ruled out by our assumption), varying either or or both does not lead to a material distortion in the (asymptotic) coverage probability of the proposed procedure. To see why this is so, consider the case where the unit root specification is true. Then, even when both and are set to be large so that the null hypothesis is falsely rejected with high probability leading to the use of , we will still end up with asymptotic coverage probability greater than the nominal level since is asymptotically valid and, by design, . On the other hand, if the underlying process is stable, then both of the unit root tests will reject the null hypothesis withprobability approaching one asymptotically, as long as neither nor is set equal to zero, and our procedure will switch to which controls the asymptotic coverage probability properly.
- (vi)
- Pre-testing leads to the possibility of errors whose probability needs to be controlled. In particular, there may be parameter sequences which lie just outside of , for which may fail to reject even in large samples. In addition, there may be parameter sequences which lie just outside of , for which is rejected by but for which may not reject even in large samples. In both of these scenarios, there is the possibility that none of our intervals will cover the true parameter sequence. However, in the proof of Lemma A1 given in the Appendix SB of the technical supplement, we show that, under our procedure, the probability of committing such Type II errors can be no greater than asymptotically 9. Hence, by constructing and in the manner suggested above, we can properly control the probability of not switching to when it is preferable to make that switch. In consequence, the asymptotic non-coverage probability under our procedure is always less than or equal to . Given a particular significance level α, different combinations of and involve trade-offs where a smaller leads to a smaller probability of committing a Type II error but also leadsto a larger and, thus, to having a smaller asymptotic coverage probability.
- (vii)
- An advantage of our pretest based confidence procedure is its computational simplicity, as it is given in analytical form and, thus, does not require the use of bootstrap or other types of simulation-based methods for its computation. Moreover, the fact that , the interval used under our procedure in the stable case, is based on the Anderson–Hsiao procedure has the further benefit that its validity does not depend on imposing the assumption of mean stationarity of the initial condition. Hence, the design of our procedure has taken into consideration certain trade-offs on the competing goals of interval accuracy, computational simplicity, and the relaxation of the assumption of initial condition stationarity.
4. Monte Carlo Study
This section reports the results of a Monte Carlo study comparing the finite sample performance of alternative confidence procedures. For the simulation study, we consider data generating processes of the form
where and . We vary , , , , , and and 2. In addition, we let 200. When , we take , 100; and when , we consider , 200. We take throughout, so that the (nominal) confidence level is always kept at . Four versions of the pre-test based confidence interval (PCI) given by expression (6) above are considered, with different specifications of , , , and , as summarized in the following tables.
Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 below provide simulation results comparing the four PCI procedures described above with the procedure given in (5) and with confidence intervals obtained by inverting Studentized statistics associated with the POLS and IVD estimators. More specifically, Table 1, Table 2, Table 3 and Table 4 give the empirical coverage probabilities while Table 5, Table 7, Table 9 and Table 11 report the average width of the confidence intervals under each of forty-eight experimental settings, obtained by varying , N, T, and . In addition, in Table 6, Table 8, Table 10 and Table 12 we report the number of instances out of 10,000 simulationrepetitions that a particular confidence procedure leads to an empty interval, which occurs when the intersection of the (unrestricted) interval and the parameter space is the null set. For example, in the case of the procedure, an empty interval would arise if
Table 1.
Coverage Probabilities (nominal level = 0.95).
Table 2.
Coverage Probabilities (nominal level = 0.95).
Table 3.
Coverage Probabilities (nominal level = 0.95).
Table 4.
Coverage Probabilities (nominal level = 0.95).
Table 5.
Average Width of Confidence Intervals.
Table 6.
Number of Empty Intervals (out of 10,000).
Table 7.
Average Width of Confidence Intervals.
Table 8.
Number of Empty Intervals (out of 10,000).
Table 9.
Average Width of Confidence Intervals.
Table 10.
Number of Empty Intervals (out of 10,000).
Table 11.
Average Width of Confidence Intervals.
Table 12.
Number of Empty Intervals (out of 10,000).
Glancing at Table 1, Table 2, Table 3 and Table 4, we see that, consistent with our theory, the empirical coverage probabilities of the procedure show the greatest degree of uniformity across different experiments. On the other hand, all four PCIs have empirical coverage probabilities that are uniformly better than the procedure across all forty-eight experiments. An intuitive explanation for this result can be given as follows. When the unit root null hypothesis is true, application of the pre-test procedure will lead to the use of either or , except in the small probability event where a Type I error is committed by both of the unit root tests and . Since both of these intervals cover the point by construction, the overall procedure in this case should cover this point with very high probability. On the other hand, when the unit root hypothesis is false, the pre-test procedure switches to the interval but with set at a level strictly less than , resulting again in coverage probabilities which are greater than that of the procedure.
A possible deficiency of the procedure as shown in Table 5, Table 7, Table 9 and Table 11 is that the average widths of intervals obtained from this procedure are substantially wider than that of the other procedures when . Moreover, in the case, the use of the procedure results in empty intervals in roughly of the times, ranging from a low of 215 empty intervals (out of repetitions) in the case with , and to a high of 295 empty intervals (out of repetitions) in the case where , and 10. In contrast, no empty interval is observed for any of the pre-test procedures in any of the 48 experiments, including experiments where . It should also be noted that, outside of the unit root case, the results of Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 do show to provide informative intervals with average widths that are much smaller than those in the case. In addition, as the true value of moves significantly away from unity, such as in the cases where , , , and ; empty intervals were no longer observed for .
For the four alternative specifications of PCI, there does not seem to be a great deal of difference in their performance across the experiments, although some minor trade-offs in coverage probability vis-à-vis average interval width can be discerned. For example, looking at PCI1, we see that this procedure provides very tight intervals in the case where . In fact, the average interval width for this procedure in the unit root case is ≤, except in the smaller sample size case with and , where it is still around . Moreover, amongst the seven procedures examined in our study, the empirical coverage probability of PCI1 is the highest, or is at least tied for the highest, almost across the board, for the 48 experiments whose results are reported in Table 1, Table 2, Table 3 and Table 4. Although the higher coverage probability of PCI1 in the stable region is due at least in part to the fact that it is designed to be conservative with when the true process is stable, it should be noted that the informativeness of PCI1, as measured by its average width, does not seem to have suffered significantly as a result. Note, in particular, that over the 48 experiments the widest average interval width recorded for PCI1 was only , or approximately of the width of the entire parameter space ; and this occurred with the smaller sample sizes of and . In addition, PCI1 has average width strictly less than in 38 of the experiments. On the other hand, PCI2 sets and is, thus, less conservative relative to PCI1, particularly in the stable region. In consequence, PCI2 tends to have not only smaller interval widths but also lower coverage probabilities relative to PCI1 when the underlying process is stable. The results for PCI1 and PCI2 are illustrative of how the pre-test procedures can greatly improve upon in terms of accuracy in the unit root and near unit root cases while maintaining coverage probability at a high level throughout the parameter space, with the only downside being that they yield slightly wider intervals when the true process is stable.
Table 1, Table 2, Table 3 and Table 4 also show that confidence intervals constructed by inverting Studentized statistics associated with and are decidedly inferior to the pre-test based confidence procedures. Consistent with our theory, Table 1, Table 2, Table 3 and Table 4 show that these confidence intervals have highly non-uniform coverage probabilities across different (true) parameter values . More specifically, the coverage probabilities of the IV-based confidence intervals are especially poor when is unity or near-unity, whereas the coverage probabilities for the POLS-based confidence intervals begin to deviate dramatically from the nominal level when or less. Moreover, from the results reported in Table 6, Table 8, Table 10 and Table 12, we note that , the confidence procedure based on inverting the Studentized statistic associated with , leads to an empty interval in more than of the simulation runs when . This is perhaps not surprising since, as shown in Theorem SA-1 in Appendix SA of the supplement to this paper, is not uniformly convergent over the parameter space and does not have an asymptotic normal distribution when the true equals unity. Hence, when , the procedure, which is designed to achieve the correct asymptotic coverage in the stationary case, will not only exhibit poor coverage probabilities but will often deliver intervals that lie entirely outside the parameter space. Interestingly, even though the procedure is based on the correct asymptotics when , it nevertheless produces some empty intervals in the unit root case as shown in Table 6, Table 8, Table 10 and Table 12. This suggests a need to modify the usual t-ratio based confidence procedure in cases where there is interest in a point on the boundary of a bounded parameter space, such as .
5. Conclusions
The uniform inference procedure proposed here utilizes information from pretesting the unit root hypothesis to aid the construction of confidence intervals in panel autoregression by means of data-based selection among intervals that are well suited to particular regions of the parameter space. The construction is asymptotically valid in the sense that the large sample coverage probability is at least that of the nominal level uniformly over a wide parameter space that includes unity. The method is particularly simple to implement in practical work and simulations provide encouraging evidence that the method produces confidence intervals with good finite sample accuracy, as measured by the combination of empirical coverage probability and average interval width. The panel AR model considered here is a simple model. But it is the kernel of all dynamic panel models and embodies all the characteristics that make uniform inference and confidence interval construction difficult. Even in the time series case these problems are well known to be challenging. In the panel case, the challenges are accentuated by additional issues arising from the presence of incidental effects and multi-index limit theory. The pre-test confidence interval solution proposed here addresses these challenges and has potential for application in more complex models.
Supplementary Materials
The following are available at https://www.mdpi.com/2225-1146/7/4/45/s1.
Author Contributions
Conceptualization: J.C.C. 70% and P.C.B.P. 30%; Methodology: J.C.C. 80% and P.C.B.P. 20%; Software: J.C.C. 100% and P.C.B.P. 0%; Writing: J.C.C. 60% and P.C.B.P. 40%; Editing: J.C.C. 50% and P.C.B.P. 50%.
Funding
This research received no external funding.
Acknowledgments
We thank Chris Sims for a thought-provoking comment which led us to a substantial improvement in some of our results. We also thank participants at the 2017 Greater New York Metropolitan Area Econometrics Colloquium, the New York Camp Econometrics XI, the 2016 Australasia Meeting of the Econometric Society, and the 2016 NBER/NSF Time Series Conference, as well as workshop participants at Boston College, Indiana University Bloomington, Pennsylvania State University, and the University of Pennsylvania for comments. Phillips acknowledges support from the NSF under Grant SES 1258258, the Kelly Fund at the University of Auckland, and an LKC Fellowship at Singapore Management University.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A. Proofs of the Main Results
The proofs given here rely on a large number of technical results that are established in the Technical Supplement (Chao and Phillips 2019). These results are designated in the derivations that follow by use of the prefix S. Lemmas A1 and A2 are stated in Appendix A and their proofs are given in the Technical Supplement. The proofs rely on functional limit theory for integrated and near integrated processes in conjunction with joint limit theory arguments for multi-indexed asymptotics (Phillips 1987a, 1987b; Phillips and Moon 1999).
Proof of Theorem 1.
Let , and note that
Applying partial summation, we have
so that
We turn first to part (a). In this case, by assumption, for all T sufficiently large. Under the random-effects specification given by Assumption 2, we can apply parts (g) and (i) of Lemma SD-11, part (a) of Lemma SD-25, and part (a) of Lemma SC-13 to obtain
It follows from applying Lemma SD-24 that , as required. Moreover, it is easily seen that, by applying part (b) of Lemma SE-1 in lieu of part (g) of Lemma SD-11 in the argument given above, the same result can be obtained under the fixed-effects specification given by Assumption 2*.
Next consider part (b), where we take such that . In the case of the random-effects specification given by Assumption 2, we can use the results in parts (g) and (i) of Lemma SD-11, part (b) of Lemma SD-25, and part (b) of Lemma SC-13 to deduce that
It follows from part (a) of Lemma SD-22 that . Moreover, it is easily seen that, by applying part (b) of Lemma SE-1 in lieu of part (g) of Lemma SD-11 in the argument given above, the same result can be obtained under the fixed-effects specification given by Assumption 2*.
Consider part (c), where we take such that . Under the random-effects specification given by Assumption 2, we can apply parts (g) and (i) of Lemma SD-11, part (c) of Lemma SD-25, and part (c) of Lemma SC-13 to deduce that
where . It follows from part (b) of Lemma SD-22 that . Moreover, it is easily seen that, by applying part (b) of Lemma SE-1 in lieu of part (g) of Lemma SD-11 in the argument given above, the same result can be obtained under the fixed-effects specification given by Assumption 2*.
For part (d), we consider the case where such that but . Here, we first apply part (d) of Lemma SC-13 and part (d) of Lemma SD-21 to obtain
Under the random-effects specification given by Assumption 2, we further apply parts (g) and (i) of Lemma SD-11 and part (d) of Lemma SD-25 to obtain
By part (c) of Lemma SD-22, we then deduce that , as required for (d). Moreover, it is easily seen that, by applying part (b) of Lemma SE-1 in lieu of part (g) of Lemma SD-11 in the argument given above, the same result can be obtained under the fixed-effects specification given by Assumption 2*.
Finally, to show part (e), we first consider the random-effects specification given by Assumption 2. In this case, note that, by applying parts (g) and (i) of Lemma SD-11, part (e) of Lemma SD-21, and part (e) of Lemma SC-13, we obtain
where and . It follows by Lemma SD-23 that . Moreover, it is easily seen that, by applying part (b) of Lemma SE-1 in lieu of part (g) of Lemma SD-11 in the argument given above, the same result can be obtained under the fixed-effects specification given by Assumption 2*. □
Proof of Theorem 2.
To proceed, note that, in the pathwise asymptotics considered here, N grows as a monotonically increasing function of T, so that the asymptotics can be taken to be single-indexed with . Now, let be the collections of parameter sequences defined in the statement of Theorem 1. Moreover, let (for ), i.e., is a sequence belonging to the collection 11. Define , with , where is an increasing function in its argument, and let denote a subsequence of . Note that every parameter sequence can be represented as , where , with for and where , with denoting the set of natural numbers.
Next, note that . Theorem 1 implies that, for any and for each , there exists positive integer such that for every positive integer ,
where . Moreover, for any positive integer , we have by Lemma SD-33 (given in Appendix SD in the technical supplement to this paper), from which we further deduce that
Next, let . Consider any positive integer ; we must have for some and for some . Given that , we also deduce that by Lemma SD-33 since is an increasing function of its argument. It follows that for every sequence and for all
The desired result then follows from (Lepski 1999) Lemma 2.6.2. □
Lemma A1.
Let , and suppose that Assumptions 1, 3, 4, and either 2 or 2* hold. Then, the following statements are true as such that , for constants and .
- (a)
- Let , and set and . Then, for ,
- (b)
- Let , and set and . Then, for ,
- (c)
- Letand set and . Then, for ,
- (d)
- Letand set and . Then, for ,
- (e)
- Let , and set and . Then, for ,
- (f)
- Letand set and . Then, for ,
- (g)
- Letand set and . Then, for ,
- (h)
- Let , and set and . Then, for ,
- (i)
- Let , and set and . Then, for ,
- (j)
- Letand set and . Then, for ,
- (k)
- Letand set and . Then, for ,
- (l)
- Letand set and . Then, for ,
- (m)
- Letand set and . Then, for ,
- (n)
- Letand set and . Then, for ,
- (o)
- Letand set and . Then, for ,
- (p)
- Let , and set and . Then, for ,
- (q)
- Let , and set and . Then, for ,
- (r)
- Letand set and . Then, for ,
- (s)
- Let , and set and . Then, for ,
- (t)
- Letand set and . Then, for ,
- (u)
- Let , and set and . Then, for ,
- (v)
- Letand set and . Then, for ,
The proof of Lemma A1 is given in Appendix SB of the technical supplement.
Lemma A2.
Suppose that Assumptions 1, 3, 4, and either 2 or 2* hold, and let
Also, let such that , for constants and , so that we can set and . Then,
for .
The proof of Lemma A2 is also given in Appendix SB of the technical supplement.
Proof of Theorem 3.
In the pathwise asymptotics considered here, N grows as a monotonically increasing function of T, so that the asymptotics can be taken to be single-indexed with . Hence, we can set and simplify notation by writing .
To proceed, note that, by property of a supremum, there exists a sequence such that
Thus, for some fixed significance level , to show that
it suffices to show that
for every sequence . To proceed, let be the collections of parameter sequences defined in Lemmas A1 and A2 given above. Moreover, let (for ), i.e., is a sequence belonging to the collection . Define , with , where is an increasing function in its argument, and let denote a subsequence of . Note that every parameter sequence can be represented as , where , with for and where
with denoting the set of natural numbers . Moreover, define
and . It is clear from the definition of and that for each ; or, more formally, for any , there exists positive integer such that, for all , , from which it follows, using the results of Lemma A1, that, for any and for each , there exists a positive integer such that, for all , . Now, for any and for any positive integer , we have, by Lemma SD-33 given in Appendix SD of the technical supplement to this paper, that , so that , for any subsequence of , from which we further deduce that
Next, let . Consider any positive integer ; then, (A1) implies that for some and for some . By the fact that is an increasing function of its argument, we have that , from which it follows that for every positive integer
Hence, for any sequence ,
Since is arbitrary, we deduce that
for any sequence , which gives the desired conclusion. □
References
- Ahn, S. C., and P. Schmidt. 1995. Efficient Estimation of Models for Dynamic Panel Data. Journal of Econometrics 68: 29–51. [Google Scholar] [CrossRef]
- Alvarez, J., and M. Arellano. 2003. The Time Series and Cross-Section Asymptotics of Dynamic Panel Data Estimators. Econometrica 71: 1121–59. [Google Scholar] [CrossRef]
- Anderson, T. W., and C. Hsiao. 1981. Estimation of Dynamic Models with Error Components. Journal of American Statistical Association 76: 598–606. [Google Scholar] [CrossRef]
- Anderson, T. W., and C. Hsiao. 1982. Formulation and Estimation of Dynamic Models Using Panel Data. Journal of Econometrics 18: 47–82. [Google Scholar] [CrossRef]
- Andrews, D. W. K. 1993. Exactly Median-Unbiased Estimation of First-Order Autoregressive/Unit Root Models. Econometrica 61: 139–65. [Google Scholar] [CrossRef]
- Andrews, D. W. K., and P. Guggenberger. 2009. Validity of Subsampling and ‘Plug-in Asymptotic’ Inference for Parameters Defined by Moment Inequalities. Econometric Theory 25: 669–709. [Google Scholar] [CrossRef]
- Andrews, D. W. K., X. Cheng, and P. Guggenberger. 2011. Generic Results for Establishing the Asymptotic Size of Confidence Sets and Tests. Cowles Foundation Discussion Paper No. 1813. New Haven: Yale University. [Google Scholar]
- Arellano, M., and S. Bond. 1991. Some Tests of Specification for Panel Data: Monte Carlo Evidence and an Application to Employment Equations. Review of Economic Studies 58: 277–97. [Google Scholar] [CrossRef]
- Arellano, M., and O. Bover. 1995. Another Look at the Instrumental Variable Estimation of Error-components Models. Journal of Econometrics 68: 29–51. [Google Scholar] [CrossRef]
- Blundell, R., and S. Bond. 1998. Initial Conditions and Moment Restrictions in Dynamic Panel Data Models. Journal of Econometrics 87: 115–43. [Google Scholar] [CrossRef]
- Bun, M. J. G., and F. Kleibergen. 2014. Identification and Inference in Moments Based Analysis of Linear Dynamic Panel Data Models. UvA Econometrics Discussion Paper. Amsterdam: University of Amsterdam-Econometrics Discussion. [Google Scholar]
- Chang, Y. 2002. Nonlinear IV Unit Root Tests in Panel with Cross-sectional Dependency. Journal of Econometrics 110: 261–92. [Google Scholar] [CrossRef]
- Chang, Y. 2004. Bootstrap Unit Root Tests in Panels with Cross-sectional Dependency. Journal of Econometrics 120: 263–93. [Google Scholar] [CrossRef]
- Chao, J., and P. C. B. Phillips. 2019. Technical Supplement to: Uniform Inference in Panel Autoregression. Working paper. College Park: University of Maryland. [Google Scholar]
- Durbin, J. 1960. Estimating of Parameters in Time Series Regression Models. Journal of the Royal Statistical Society, Series B 22: 139–53. [Google Scholar]
- Gorodnichenko, Y., A. Mikusheva, and S. Ng. 2012. Estimators for Persistent and Possibly Nonstationary Data with Classical Properties. Econometric Theory 28: 1003–36. [Google Scholar] [CrossRef]
- Hahn, J., and G. Kuersteiner. 2002. Asymptotically Unbiased Inference for a Dynamic Panel Model with Fixed Effects When Both N and T are Large. Econometrica 70: 1639–57. [Google Scholar] [CrossRef]
- Han, C., and P. C. B. Phillips. 2010. GMM Estimation for Dynamic Panels with Fixed Effects and Strong Instruments at Unity. Econometric Theory 26: 119–51. [Google Scholar] [CrossRef]
- Han, C., P. C. B. Phillips, and D. Sul. 2011. Uniform Asymptotic Normality in Stationary and Unit Root Autoregression. Econometric Theory 27: 1117–51. [Google Scholar] [CrossRef]
- Han, C., P. C. B. Phillips, and D. Sul. 2014. X-Differencing and Dynamic Panel Model Estimation. Econometric Theory 30: 201–51. [Google Scholar] [CrossRef]
- Hansen, B. E. 1999. The Grid Bootstrap and the Autoregressive Model. Review of Economics and Statistics 81: 594–607. [Google Scholar] [CrossRef]
- Kiviet, J. F. 1995. On Bias, Inconsistency, and Effciency of Various Estimators in Dynamic Panel Data Models. Journal of Econometrics 68: 53–78. [Google Scholar] [CrossRef]
- Lehmann, E. L. 1999. Elements of Large Sample Theory. New York: Springer. [Google Scholar]
- Lepski, O. V. 1999. How to Improve the Accuracy of Estimation. Mathematical Methods in Statistics 8: 441–86. [Google Scholar]
- Mikusheva, A. 2007. Uniform Inference in Autoregressive Models. Econometrica 75: 1411–52. [Google Scholar] [CrossRef]
- Mikusheva, A. 2012. One-dimensional Inference in Autoregressive Models with the Potential Presence of a Unit Root. Econometrica 80: 173–212. [Google Scholar]
- Moon, H. R., and P. C. B. Phillips. 2004. GMM Estimation of Autoregressive Roots Near Unity with Panel Data. Econometrica 72: 467–522. [Google Scholar] [CrossRef]
- Moon, H. R., B. Perron, and P. C. B. Phillips. 2014. Point Optimal Panel Unit Root Tests with Serially Correlated Data. Econometrics Journal 17: 338–72. [Google Scholar] [CrossRef]
- Moon, H. R., B. Perron, and P. C. B. Phillips. 2015. Incidental Parameters and Dynamic Panel Modeling. In The Oxford Handbook of Panel Data. Edited by B. Baltagi. New York: Oxford University Press. [Google Scholar]
- Nickell, S. 1981. Biases in Dynamic Models with Fixed Effects. Econometrica 49: 1417–26. [Google Scholar] [CrossRef]
- Pesaran, M. H. 2006. Estimation and Inference in Large Heterogeneous Panels with a Multifactor Error Structure. Econometrica 74: 967–1012. [Google Scholar] [CrossRef]
- Pesaran, M. H., and E. Tosetti. 2011. Large Panels with Common Factors and Spatial Correlation. Journal of Econometrics 161: 182–202. [Google Scholar] [CrossRef]
- Phillips, P. C. B. 1987a. Time Series Regression with a Unit Root. Econometrica 55: 277–301. [Google Scholar] [CrossRef]
- Phillips, P. C. B. 1987b. Towards a Unified Asymptotic Theory of Autoregression. Biometrika 74: 535–47. [Google Scholar] [CrossRef]
- Phillips, P. C. B. 2014. On Confidence Intervals for Autoregressive Roots and Predictive Regression. Econometrica 82: 1177–95. [Google Scholar] [CrossRef]
- Phillips, P. C. B. 2018. Dynamic Panel Anderson–Hsiao Estimation With Roots Near Unity. Econometric Theory 34: 253–276. [Google Scholar] [CrossRef]
- Phillips, P. C. B., and H. R. Moon. 1999. Linear Regression Limit Theory for Nonstationary Panel Data. Econometrica 67: 1057–111. [Google Scholar] [CrossRef]
- Phillips, P. C. B., and D. Sul. 2003. Dynamic Panel Estimation and Homogeneity Testing under Cross Sectional Dependence. Econometrics Journal 6: 217–59. [Google Scholar] [CrossRef]
- Stock, J. 1999. Confidence Intervals for the Largest Autoregressive Root in US Macroeconomic Time Series. Journal of Monetary Economics 28: 435–59. [Google Scholar] [CrossRef]
| 1. | We do not consider in this paper issues related to incidental trends, cross section dependence, and slope parameter heterogeneity discussed earlier. While these complications are important and empirically relevant, they are beyond the scope of the current paper and considering them here would divert from the main point of this paper which concerns the development of uniform inference procedures. |
| 2. | For readers interested in the asymptotic properties of the Anderson–Hsiao IV estimator, we would like to refer them to Theorem SA-1 in the Technical Supplement to this paper. There, we present a very extensive set of results on the large sample behavior of this estimator under various parameter sequences both near and far away from unity. In addition, the proof of Theorem SA-1 is provided in Appendix SB of the Technical Supplement. |
| 3. | Other approaches for achieving uniform inference in estimation have been proposed recently in the time series literature by (Han et al. 2011) using partial aggregation methods and by (Gorodnichenko et al. 2012) using quasi-differencing. In the unit root and very near unit root cases, extending these approaches to the panel data setting leads to confidence intervals whose width shrinks at a slower rate than the optimal rate obtained here. (Han et al. 2014) developed a panel estimator using X-differencing which has good bias properties and limit theory but has different limit theory in unit root and stationary cases, complicating uniform inference. |
| 4. | The reason we consider indexed parameter which depends on T only, and not on both N and T, is because our main results are obtained under a general pathwise asymptotic scheme where N can grow as an arbitrary positive real-valued power of T. In such a framework, the asymptotics are effectively single-indexed. Hence, it suffices to consider parameter sequences that depend only on T. |
| 5. | The proof of Lemma SC-12 is also given in Appendix SC of the technical supplement. |
| 6. | We use the notation to denote the Anderson–Hsiao IV estimator because it is a procedure where IV estimation is performed on a first-differenced equation. Later, we use to denote the IV estimator introduced by (Arellano and Bover 1995) since, in that procedure, IV is performed on the panel autoregression in levels. |
| 7. | Note that we use the notation instead of perhaps the more familiar notation to denote a probability measure indexed by the parameter because, in this paper, we will often consider somewhat complicated local-to-unity parameters and subsequences of such parameters, which are less conveniently expressed in terms of subscripts. |
| 8. | A recent paper by (Bun and Kleibergen 2014) also considers, amongst other things, combining elements of the approach of (Anderson and Hsiao 1981, 1982; Arellano and Bond 1991), which uses lagged levels of as instruments for equations in first differences, with the approach by (Arellano and Bover 1995; Blundell and Bond 1998) which uses lagged differences of as instruments for equations in levels. The focus of the (Bun and Kleibergen 2014) paper differs substantially from that of the present paper. In particular, they consider test procedures which attain the maximal attainable power curve under worst case setting of the variance of the initial conditions, whereas our procedure uses pretest based information to aggressively increase the power of our inferential procedure in certain regions of the parameter space. Moreover, unlike our paper, they do not provide results on confidence procedures whose asymptotic coverage probability is explicitly shown to be at least that of the nominal level uniformly over the parameter space; and their analysis is conducted within a fixed T framework. |
| 9. | The statement of Lemma A1 is given in the Appendix A of the paper. Its proof is lengthy and it is therefore placed in the technical supplement. |
| 10. | It might initially seem strange in Table 6 and Table 8 that in the cases where and , the number of empty intervals for actually increased as the sample size in the time dimension increased from to . However, there is an intuitive explanation for this result. As noted earlier, in the unit root case, the rate of concentration of the width of the interval is , so that intervals obtained under this procedure are wider in the case than in the case, leading to a higher chance of a non-null intersection with the parameter space. |
| 11. | The reason for using the notation , as opposed to , is so that we can refer to a particular collection of sequences amongst without necessarily being , for example. |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).