Abstract
Typically, the traditional bootstrap methods for parameter inference of spatial error models suffer from high computational costs, so this study proposes fast cluster bootstrap methods for spatial error models to deal with the dilemma. The key idea is to calculate the sufficient statistics for each cluster before performing the bootstrap loop of the spatial error model, and based on these sufficient statistics, all quantities needed for bootstrap inference can be computed. Furthermore, this study performed Monte Carlo simulations, and the result reveals that compared with traditional bootstrap methods, our proposed methods can reduce the computational cost substantially and improve the reliability for obtaining the bootstrap test statistics and confidence intervals of the parameters for spatial error models.
MSC:
62F40; 62J05
1. Introduction
Traditionally, maximum likelihood estimation is typically used to perform parameter inference for spatial regression models. Lim et al. [1] proposed the maximum block independent likelihood estimator based on independent sub-blocks in spatial models. As for the spatial autoregressive model, Wang and Song [2] developed a quasi-maximum likelihood method with a penalty to achieve parameter estimation in the presence of missing responses. Al-Momani and Arashi [3] constructed the maximum likelihood estimator for the parameters of the spatial error model. However, the computational cost of parameter inference for spatial regression models based on maximum likelihood estimation is relatively high, especially for high-dimensional data. Furthermore, spatial dependencies generally increase the complexity of the computation.
Typically, the objects or observations collected from the same geographic region are more comparable than those collected from distant regions, which can be regarded as a cluster, and the objects or observations within the same cluster are more comparable to each other than to those in other clusters. Spatial regression models, incorporating various spatial dependencies, provide an opportunity to analyze the objects or observations collected from the same geographic location, known as a site [3], which is similar to the idea of clusters described above.
Moreover, inference based on cluster-robust variance estimators (CRVEs) performs well when large-sample theory serves as a good guide to the finite-sample properties of the CRVE. Nevertheless, when the number of clusters is small, the size of the clusters and the characteristics of the regressors and regressand are not fairly homogeneous, the CRVE becomes seriously unreliable, and there is a great deal of theoretical evidence and simulation results in the latest papers about this problem [4,5,6,7,8,9].
In fact, one of the alternative methods for spatial regression models is the bootstrap method, which involves generating a number of bootstrap samples mimicking the actual sample distribution, and the bootstrap test statistics and confidence intervals for each sample are acquired with the same test procedure as for the original sample [10,11,12,13,14]. There are two bootstrap methods based on clustered data, including pairs cluster bootstrap and the wild cluster bootstrap, which were originally proposed by Freedman [15], Liu [16], and Mammen [17], respectively, and further extended by Davidson and Flachaire [18] and Cameron et al. [19] and have been proved to be asymptotically valid [8,20]. Furthermore, these bootstrap methods are widely applied to various problems. Bouzebda et al. [21] extended the existing theory on the bootstrap of the M-estimators, and they proposed an exchangeably weighted bootstrap for function-valued estimators defined as a zero point of a function-valued random criterion function subsequently [22].
However, applying these traditional bootstrap methods to perform bootstrapping on linear regression models for clustered data still suffers from high computational costs. MacKinnon [23] has proposed an efficient computational algorithm for bootstrapping linear regression models based on clustered data. Inspired by MacKinnon [23], this study uses a similar idea to deal with the high computational costs of parameter inference for spatial regression models.
In addition, one of the most widely used spatial regression models is the spatial error model (SEM), which models the mean of the spatial response variable by using a linear regression with a spatially lagged autoregressive error component [24].
Considering the notable gap in the literature, this study proposes fast cluster bootstrap methods for SEM to reduce the computational cost and improve the reliability for obtaining the bootstrap test statistics and confidence intervals of SEM. Specifically, before performing the bootstrap loop of SEM, the sufficient statistics are computed for each of the clusters, and based on these sufficient statistics, it is possible to calculate all of the quantities needed for bootstrap inference. In addition, the test statistics and the bootstrap confidence intervals are associated with the samples solely by these sufficient statistics.
The key contributions of this study are presented as follows:
Firstly, this is the first study to propose fast cluster bootstrap methods to overcome the high computational cost of parameter inference for SEM. Secondly, this study simulates the computational cost, two-sided equal-tailed coverage frequencies, and the precision of the bootstrap confidence intervals of the parameter with various bootstrap methods, respectively. The result reveals that the computational cost of our proposed methods is substantially reduced, and the optimal coverage of the parameter is higher than the nominal frequency. Finally, compared with traditional methods, the methods proposed in this study are conceptually simple, easier to understand, and can also be used in other spatial models.
The remainder of this paper is organized as follows. In Section 2, we present an overview of the SEM. In Section 3, we describe bootstrap computations for maximum likelihood estimation of SEM, including the fast pairs cluster bootstrap method for SEM and the fast unrestricted and restricted wild cluster bootstrap methods for SEM, respectively. In Section 4, we perform simulation experiments and show that using these methods can yield substantial computational savings. In Section 5, we investigate the finite-sample performance of the two-sided equal-tailed coverage frequencies and average widths of the bootstrap confidence intervals of the parameter for SEM based on various bootstrap methods. In Section 6, we give the conclusions. Finally, in Section 7, we present the future work of this study. In addition, in order to improve clarity, we provide a summary of key notations in Table 1.
Table 1.
Summary of key notations.
2. Spatial Error Model
Let represent a set of disjoint clusters, which are often referred to as locations, regions, etc. The cluster is assumed to have observations, and the sample size is . The SEM can be written as
with
where is an matrix of exogenous regressors, represents a vector of unknown regression parameters, is an vector of observations collected for the cluster, and denotes an vector of spatially autocorrelated remainder error. The error term involves the spatial dependence parameter , the matrix of spatial weight matrix , and an vector of the noise .
Furthermore, we make the following assumptions:
Assumption 1.
the noise has a normal distribution , where is the identity matrix.
Assumption 2.
the elements of the main diagonal of spatial weight matrix are zeros, and if position is adjacent to position , the off-diagonal element ; otherwise, for , and the normalization of is row normalized as .
Equivalently, under these assumptions, the SEM in (1) can be rewritten as follows,
Let , Equation (2) is equivalent to
and it can be rewritten in matrix form, as follows,
Let . The maximum likelihood estimator (MLE) of can be obtained by maximizing the concentrated likelihood function [25]. Specifically, we first fix and find the MLEs of as a function of as follows,
Secondly, we plug and into the log-likelihood to obtain the MLE of . Finally, the MLEs for and are acquired by substituting for in Equations (5) and (6), respectively.
Thus, when the data generation process (DGP) is a special case of (1), the following formular holds:
where is the true value of , and represents the score vector corresponding to the cluster. Assuming that
where the expectations here are conditional on the , the matrix denotes a positive semi-definite and symmetric matrix, which is the (conditional) covariance matrix of the score vector corresponding to the cluster. Based on Equations (7) and (8), we provide a more informative representation of the (conditional) covariance matrix of , which makes it more explicit that the key to estimate is in terms of the covariance matrix of the score vectors, and it follows
When estimating , it is natural to apply the outer products of the empirical score vectors:
By correcting for degrees of freedom, we acquire the most widely used CRVE [24]:
Note that given the matrix and the vectors , can be acquired without computing the residual subvector , which can reduce the computational costs. Since each of the matrices has rank at most 1, then (11) has rank at most (it has rank in many situations), which indicates that asymptotic inference derived from the above (11) may not be reliable when is not large, and particularly when there are several restrictions. Consequently, the bootstrap methods are essential.
3. Bootstrap Computations for Maximum Likelihood Estimation of SEM
Bootstrapping is an alternative method for inferring statistical parameters where computational complexity arises [10]. In addition, different cluster bootstrap methods (to illustrate the asymptotic properties of the estimator proposed in our paper, we cite an existing theorem in Appendix A and briefly explain why this theorem is applicable to our estimator.) for SEM yield bootstrap samples in different ways, which incur different computational costs. In this part, we discuss the fast pairs cluster bootstrap method for SEM and the fast unrestricted and restricted wild cluster bootstrap methods for SEM, respectively.
3.1. The Fast Pairs Cluster Bootstrap Method for SEM
Bootstrapping is an alternative method for inferring statistical parameters where computational complexity arises [10]. In addition, different cluster bootstrap methods for SEM yield bootstrap samples in different ways, which incur different computational costs. In this part, we discuss the fast pairs cluster bootstrap method for SEM and the fast unrestricted and restricted wild cluster bootstrap methods for SEM, respectively.
The original pairs cluster bootstrap method for SEM is to group the data for each cluster into a pair, and then to resample the data from the original pairs with replacement, such that each bootstrap sample is composed of the original pairs with equal probability . The big difference between the fast pairs cluster bootstrap method (pcb) and the original method is the first step. As for pcb, firstly, we group the data for each cluster into a pair, and then we resample from the pairs. Consequently, each bootstrap sample is constructed by randomly choosing pairs with equal probability , and the bootstrap sample is represented as follows,
Thus, the fast pairs cluster bootstrap estimates of and are given by the following, respectively:
Typically, and are small, and the computational cost in both (13) and (14) is . Furthermore, the CRVE for is
and the bootstrap test statistic for , where is a known -vector:
Let and represent the and quantiles of the , respectively, and based on these statistics, the bootstrap percentile-t confidence interval for is constructed as follows,
where is the cluster-robust standard error of .
In terms of the pcb, there are several limitations. Firstly, the pcb generally involves higher computational costs; specifically, the matrix in the above calculation is different for each bootstrap sample and requires to be generated and inverted times. Secondly, a technical problem that is typically ignored is that it is possible for to be singular in a simulated bootstrap sample, in which case the least squares estimator cannot be defined, especially when our problem involves treatment at the level of the cluster and few clusters are treated, and obtained based on this method may be inaccurate [5,6,26,27]. Finally, since the null hypothesis is not imposed on bootstrap samples, many of the existing studies indicate that pcb may not yield satisfactory results, especially for samples with different sizes [19].
3.2. The Fast Wild Cluster Bootstrap Methods for SEM
The original wild cluster bootstrap methods use auxiliary random variables for each cluster, which are i.i.d. with zero mean and unit variance, and the most popular choice for is the Rademacher random variables, which satisfy and [8,16,18,28]. The bootstrap observations are then generated as with , where the regressors are held fixed at their sample values, denotes the sample estimator, and are the least squares residuals, which are also held fixed at their sample values. The big difference between fast wild cluster bootstrap methods (WCB) and the original method is the first step. As for WCB, the bootstrap disturbances affect the estimates only through scores with , and based on whether generating bootstrap samples depends on the restrictions to be tested, WCB is divided into the fast unrestricted and restricted wild cluster bootstrap methods.
3.2.1. The Fast Unrestricted Wild Cluster Bootstrap Method for SEM
The fast unrestricted wild cluster bootstrap method (UWC) generates bootstrap samples that do not depend on the restrictions to be tested; furthermore, it is based on different steps of the computation, which is divided into UWC1, UWC2, and UWC3.
The detailed steps for applying UWC1 to generate the bootstrap sample and statistics are described below:
Step 1: Generate from Rademacher distribution, and furthermore, generate the bootstrap score vectors as follows,
where .
Step 2: Based on the bootstrap scores from (18), we can easily obtain
And Equation (10) is equivalent to
Similarly, the empirical bootstrap score are given as follows,
Step 3: We can obtain the CRVE for and the bootstrap test statistic for , respectively:
Similar to the bootstrap percentile-t confidence interval constructed by pcb in Section 3.1, we can acquire the bootstrap percentile-t confidence interval constructed by UWC1.
UWC1 involves some unnecessary work in order to obtain t-statistics and further construct confidence intervals, which can be improved by UWC2.
In fact, when we calculate the numerator of the t-statistic, it is possible to avoid calculating . To this end, before the bootstrap loop, form the matrices
generate the vector
and compute the numerator of the t-statistic as follows,
where
and
The detailed steps for applying UWC2 to generate the bootstrap sample and statistics are described below.
Step 1: Generate the bootstrap score vectors as follows,
where .
Step 2: Based on Equation (24), we can obtain the empirical bootstrap scores as follows,
Step 3: We can calculate the CRVE for and the bootstrap t-statistic, respectively:
In fact, in terms of calculating the denominator of the bootstrap t-statistic, a trick proposed by Roodman et al. [29] can further reduce the computational cost, and based on this trick, a method called UWC3 is derived. Specifically, before the bootstrap loop, in addition to (24), (25), (26), and (27), we can define the matrix with the typical element:
The detailed steps for applying UWC3 to generate the bootstrap sample and statistics are described below.
Step 1: For each bootstrap replication, based on (25) and (27), we can calculate the instead of in the denominator of the bootstrap t-statistic for UWC2 as follows,
Step 2: We can obtain the denominator of the bootstrap t-statistic and the bootstrap t-statistic, respectively, as follows,
Furthermore, we can obtain the bootstrap percentile-t confidence interval constructed by UWC2 and UWC3, respectively, which are similar to that of UWC1.
Obviously, by applying Equation (18), UWCs have two advantages over pcb; firstly, the computational cost of applying Equation (18) is , rather than that would be required by the traditional methods [30]. Secondly, Equation (18) can preserve the covariance matrix of the scores evidently; specifically, is determined by the bootstrap scores, while the covariance matrix is determined by the empirical bootstrap scores, which captures the key feature of WCB.
Moreover, when calculating the denominator of the bootstrap t-statistic, the computational cost required to form in UWC1 or UWC2 is . Meanwhile, when the initial work of UWC3 is completed, the effort required for each bootstrap sample is and does not rely on or . Therefore, in general, UWC3 is less computationally expensive than UWC2, except for very large .
3.2.2. The Fast Restricted Wild Cluster Bootstrap Method for SEM
The difference between the restricted wild cluster bootstrap method (RWC) and UWC is that the former uses the restricted empirical score vectors instead of unrestricted ones .
Let us partition as [, where is a -vector, and is a scalar, and consider a restriction of the form . When is partitioned conformably with , the SEM can be rewritten as
where represents the matrix, and denotes the -vector, with . It can also be rewritten in matrix form as follows,
and the restrict MLEs of is given by
where consists of the upper-left matrix block of , and consists of the first elements of .
Before the bootstrap loop, form the matrix with the typical element,
generate the vector
and compute the numerator of the t-statistic as follows,
where
and .
The detailed steps for applying RWC to generate the bootstrap sample and statistics are described below.
Step 1: For each bootstrap replication, we can calculate the in the denominator of the bootstrap t-statistic as follows,
Step 2: We can obtain the denominator of the bootstrap t-statistic and the bootstrap t-statistic, respectively:
However, the bootstrap percentile-t confidence interval constructed by RWC differs from the one by UWC. Typically, inverting a RWC bootstrap test is a good way to obtain a confidence interval for one of the parameters in a linear model. In specific, the confidence interval of the parameter is obtained by performing two RWC bootstrap tests: one for the upper limit of the interval, and the other for the lower limit of the interval [30,31]. For example, if we intend to form a bootstrap confidence interval for the parameter based on RWC, it is necessary to acquire the upper and lower limits of the interval through an iterative process. Specifically, let represent one of the limits of the confidence interval, the upper limit or the lower limit , and in such a case, we could apply the equal-tail bootstrap p-value as follows,
where
and is an indicator function that takes 1 if its argument is true and 0 otherwise.
Let denotes the significance level. Obviously, the lower limit is virtually certain to be less than ; thus, is less than for and greater than for . Similarly, it is certain that , which is greater than , also satisfies these two inequalities.
During the iterative process, to determine the bootstrap confidence interval of based on RWC, is needed to be evaluated several times with the same set of realized values of the auxiliary random variables. Specifically, before generating the bootstrap samples and statistics for any value , we need to form the vector and the matrix for testing , respectively, which both depend on . Let and , where is a vector with 1 for the element and 0 for other elements, then and can be rewritten as
and
Moreover, for each bootstrap sample, the bootstrap t-statistic for testing is available based on (39), (41), and (42), and it involves less computational cost. However, since is not a continuous function of its parameter, it is not possible to obtain the optimal value using the methods that rely on derivatives. Based on the properties discussed above, the bisection method is a reliable, easy to implement, and guaranteed convergence method to determine the real roots of such problems, and the steps for applying the bisection method to determine the limits of the confidence interval are as follows,
Step1: Define choose initial values of , including and , such that , and tolerance rate .
Step 2: If , then the root does not lie in this interval.
Step 3: Find the midpoint, set ,
(i) if the function value of the midpoint , then is the root. Go to Step 5.
(ii) if ) () , the root lies between and , then set , .
(iii) or else set , .
Step 4: If the is higher than , go to step 3, otherwise display the interval , where and .
Step 5: Display as the approximate root, which satisfies .
The bisection process is repeated, and finally, the interval is available.
4. Computing Costs
Obviously, the methods presented in Section 3 are much faster than the bootstrap method that directly generates the full bootstrap samples. Specifically, if we perform the full maximum likelihood estimation of SEM for each bootstrap sample, the cost of each bootstrap replication is . In contrast, by pre-computing some sufficient statistics, we can calculate all of the quantities needed for bootstrap inference without performing a computationally expensive process of . Thus, all calculations with computational cost are performed only once, not for each bootstrap sample, and the computational cost of each bootstrap replication depends solely on and , but not on .
In this section, we support the above claims by simulating the computational cost of calculating the bootstrap t-statistics with pcb, UWC1, UWC2, UWC3, and RWC, as proposed in Section 3, respectively. In addition, for comparative purposes, this study also reports a benchmark number that is times the computing time of a single test statistic. Table 2 illustrates the average computing times (in seconds) of 1000 Monte Carlo simulations for various bootstrap methods with or , or 20, or 9999, , varying in the set , and having the rook-based contiguity neighborhood. In addition, the consuming time to generate the data (which is typically larger than the computing time of bootstrapping) is not included in all experimental results.
Table 2.
Computing times (in seconds) for various bootstrap methods.
Firstly, it is obvious that pcb is much less computationally expensive than the baseline method; nevertheless, pcb is more computationally expensive than UWCs since UWCs pre-compute the same matrix for all bootstrap samples, while pcb involves constructing the matrix for each bootstrap sample. Secondly, compared to UWC1, UWC2 is faster, which may be attributed to the fact that the matrices and and vectors are formed before the bootstrap loop, which makes computing the t-statistic less expensive. Moreover, UWC3 applies a trick to compute the denominator of the bootstrap t-statistic, which further reduces computational cost. Thirdly, applying RWC is marginally faster than UWC3, and it performs the fastest; therefore, it can be used for large samples in applications. Finally, when , increasing the times of has less impact on the computational cost, which is different from the performance when , and this is because that the initial computation takes up a large proportion of the total time for large samples.
5. Monte Carlo Simulations
To demonstrate our theoretical findings, we perform 1000 Monte Carlo simulations and 999 bootstrap iterations to investigate the empirical coverage frequencies of nominal 95% confidence intervals constructed by pcb, UWC1, UWC2, UWC3, and RWC, respectively. Specifically, the Monte Carlo analysis was performed with a simple SEM, which includes one explanatory variable and a constant:
where , , is generated from a normal distribution with zero mean and unit variance, and is drawn from another normal distribution with zero mean and variance 2. We consider a rook-based contiguity neighborhood for the spatial weight matrix , which has been normalized, and the parameter varies over the set .
In each of these experiments, there are a total of observations, and these observations are grouped into clusters given by the following equations,
where denotes the integer of its argument. The parameter determines the heterogeneity of the cluster size; specifically, when , for all there holds , while the difference in cluster size grows as increases. Table 3 reports the two-sided equal-tailed coverage frequencies of with the precisions of the bootstrap confidence intervals (in parentheses) when or .
Table 3.
Two-sided equal-tailed coverage frequencies of with 10 clusters.
From Table 3, we can find that firstly, when , (i.e., all clusters contain 100 observations), the two-sided equal-tailed coverage frequencies and the precisions of the bootstrap confidence intervals of obtained with each method perform better than (i.e., the cluster sizes vary from 18 to 277). Secondly, the bootstrap confidence intervals of constructed by pcb severely under-cover for all values of , and the UWCs improve on the pcb but still under-cover for all . Thirdly, the RWC outperforms all UWCs but still gives coverage less than the nominal frequency for all values of , which is consistent with the findings that the under-coverage of bootstrap confidence interval based on UWCs is more severe than that of bootstrap confidence interval based on RWC in some cases [21,29]. Finally, it is obvious that the RWC obtains the optimal performance with and = 0.6; in particular, the two-sided equal-tailed coverage frequency of is 0.9345.
Table 4 shows the two-sided equal-tailed coverage frequencies of with the precisions of the bootstrap confidence intervals (in parentheses) when and or .
Table 4.
Two-sided equal-tailed coverage frequencies of with 20 clusters.
Firstly, with a bigger cluster size, the two-sided equal-tailed coverage frequencies and the precisions of the bootstrap confidence intervals of based on each method with perform better than those with (i.e., the cluster sizes vary from 16 to 304). Secondly, when , the results of all experiments largely outperform those when ; in particular, the optimal coverage of based on RWC is around 0.96 with and , which is higher than the nominal frequency of 0.95. Finally, the best performances of based on UWCs and RWC are obtained with and , respectively, which suggests that choosing an optimal value of for different methods may result in coverage closer to the nominal frequency.
6. Conclusions
This study proposes fast cluster bootstrap methods for SEM, which significantly reduce the computational cost and improve the reliability of parameter inference for SEM with large samples compared with traditional bootstrap methods. Specifically, before performing the bootstrap loop of SEM, the sufficient statistics are computed for each of the clusters, and based on these sufficient statistics, we can calculate all of the quantities needed for bootstrap inference. In addition, the test statistics and the bootstrap confidence intervals are associated with the samples solely by these sufficient statistics.
This study simulates the computational cost of calculating bootstrap t-statistics with various bootstrap methods, and the result reveals that the computational cost of our proposed methods is substantially reduced compared with traditional bootstrap methods, particularly with a large sample size.
Moreover, an extensive Monte Carlo simulation study indicates that, firstly, when all clusters have the same size, the two-sided equal-tailed coverage frequencies and the precisions of the bootstrap confidence intervals of the parameter obtained with the various methods outperform those of the clusters with varied sizes. Secondly, when , the results of all experiments largely outperform those when , and the optimal coverage of the parameter based on RWC is higher than the nominal frequency of 0.95. Finally, choosing an optimal value of for different methods may result in coverage closer to the nominal frequency. Table 5 shows the main empirical findings of this study.
Table 5.
The main empirical findings of this study.
7. Future Work
This study proposes fast cluster bootstrap methods for SEM and has the future research work as follows: firstly, our methods can be extended to other spatial models such as the Spatial Lag Model (SLM). In specific, the SLM can be written as Equivalently, the SLM can be expressed as Furthermore, let and the SLM is equivalent to Similarly, it can be rewritten in matrix form with Obviously, the bootstrap computations for maximum likelihood estimation of SLM, including the fast pairs cluster bootstrap method for SLM and the fast unrestricted and restricted wild cluster bootstrap methods for SLM, are similar to the calculation process used for SEM in the above sections.
Secondly, our experiments are limited in scope. We suppose that the error of the model with the classical assumption has a normal distribution. In fact, more realistic data-generating processes (e.g., non-normal errors, spatial heteroscedasticity) have more challenging problems, which will be our focus of research in the future.
Author Contributions
Y.Z. helped with conceptualization, methodology, software, investigation, and writing—original draft. H.F. was involved in submitting, writing—review and editing, and validation. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by National Natural Science Foundation of China (grant no. 72472150), and Fundamental Research Funds for the Central Universities (grant no. 292024082).
Data Availability Statement
Data is contained within the article. The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
Appendix A
We introduce the following Theorem 1 [32] to illustrate the theoretical properties of the methods proposed in this study.
Theorem A1.
Let be a stationary sequence of real random variables with mar ginal distribution , and let be a class of functions in Also, assume that are generated by the moving blocks bootstrap (MBB) procedure with block size , as , and that there exists , , and such that
- (a)
- ,
- (b)
- is permissible, VC, and has envelop F satisfying , and
- (c)
- .
Then, the MBB empirical process converges to the same limiting process as the original empirical process in
MBB works as follows for a stationary sample For a chosen block length , extend the sample by defining , and let be the smallest integer such that . Now, define blocks (as row vectors) for and sample from the with replacement to obtain blocks . The bootstrapped sample consists of the first observations from the row vector (). The bootstrapped empirical measure indexed by the class is then defined as .
The fast cluster bootstrap proposed in our study constructs estimators from cluster-wise sufficient statistics that are asymptotically equivalent to their full-data counterparts, while the resampling scheme fully preserves the underlying spatial dependence. Consequently, our procedure falls squarely within the theoretical framework of the moving blocks bootstrap for spatial data, satisfies all of its regularity conditions, and, therefore, inherits the conclusions of Theorem 1, guaranteeing asymptotic validity under dependence.
References
- Lim, J.; Lee, K.; Yu, D.; Liu, H.Y.; Sherman, M. Parameter estimation in the spatial auto-logistic model with working independent subblocks. Comput. Stat. Data Anal. 2012, 56, 4421–4432. [Google Scholar] [CrossRef]
- Wang, Y.F.; Song, Y.Q. Variable selection via penalized quasi-maximum likelihood method for spatial autoregressive model with missing response. Spat. Stat. 2024, 59, 100809. [Google Scholar] [CrossRef]
- Al-Momani, M.; Arashi, M. Ridge-Type Pretest and Shrinkage Estimation Strategies in Spatial Error Models with an Application to a Real Data Example. Mathematics 2024, 12, 390. [Google Scholar] [CrossRef]
- MacKinnon, J.G.; Webb, M.D. Wild bootstrap inference for wildly different cluster sizes. J. Appl. Econom. 2017, 32, 233–254. [Google Scholar] [CrossRef]
- MacKinnon, J.G.; Webb, M.D. Pitfalls when estimating treatment effects using clustered data. Political Methodol. 2017, 24, 20–31. [Google Scholar]
- MacKinnon, J.G.; Webb, M.D. The wild bootstrap for few (treated) clusters. Econom. J. 2018, 21, 114–135. [Google Scholar] [CrossRef]
- Pustejovsky, J.E.; Tipton, E. Small sample methods for cluster-robust variance estimation and hypothesis testing in fixed effects models. J. Bus. Econ. Stat. 2018, 36, 672–683. [Google Scholar] [CrossRef]
- Djogbenou, A.A.; MacKinnon, J.G.; Nielsen, M.Ø. Asymptotic theory and wild bootstrap inference with clustered errors. J. Econom. 2019, 212, 393–412. [Google Scholar] [CrossRef]
- Canay, I.A.; Santos, A.; Shaikh, A. The wild bootstrap with a ‘small’ number of ‘large’ clusters. Rev. Econ. Stat. 2021, 103, 346–363. [Google Scholar] [CrossRef]
- Hong, H.; Li, J. The numerical bootstrap. Ann. Stat. 2020, 48, 397–412. [Google Scholar] [CrossRef]
- Taconeli, C.A.; de Lara, I.A.R. Improved confidence intervals based on ranked set sampling designs within a parametric bootstrap approach. Comput. Stat. 2022, 37, 2267–2293. [Google Scholar] [CrossRef]
- Varmann, L.; Mouriño, H. Clustering Empirical Bootstrap Distribution Functions Parametrized by Galton–Watson Branching Processes. Mathematics 2024, 12, 2409. [Google Scholar] [CrossRef]
- Bongiorno, C. Bootstraps regularize singular correlation matrices. J. Comput. Appl. Math. 2024, 449, 115958. [Google Scholar] [CrossRef]
- Febrero-Bande, M.; Galeano, P.; García-Portugués, E.; González-Manteiga, W. Testing for linearity in scalar-on-function regression with responses missing at random. Comput. Stat. 2024, 39, 3405–3429. [Google Scholar] [CrossRef]
- Freedman, D.A. Bootstrapping Regression Models. Ann. Stat. 1981, 9, 1218–1228. [Google Scholar] [CrossRef]
- Liu, R.Y. Bootstrap procedures under some non-I.I.D. models. Ann. Stat. 1988, 16, 1696–1708. [Google Scholar] [CrossRef]
- Mammen, E. Bootstrap and wild bootstrap for high dimensional linear models. Ann. Stat. 1993, 21, 255–285. [Google Scholar] [CrossRef]
- Davidson, R.; Flachaire, E. The wild bootstrap, tamed at last. J. Econom. 2008, 146, 162–169. [Google Scholar] [CrossRef]
- Cameron, A.C.; Gelbach, J.B.; Miller, D.L. Bootstrap-based improvements for inference with clustered errors. Rev. Econ. Stat. 2008, 90, 414–427. [Google Scholar] [CrossRef]
- Flachaire, E. Bootstrapping heteroskedastic regression models: Wild bootstrap vs. pairs bootstrap. Comput. Stat. Data Anal. 2005, 49, 361–376. [Google Scholar] [CrossRef]
- Bouzebda, S.; El-hadjali, T.; Ferfache, A.A. Central limit theorems for functional Z-estimators with functional nuisance parameters. Commun. Stat.-Theory Methods 2022, 53, 2535–2577. [Google Scholar] [CrossRef]
- Bouzebda, S.; Elhattab, I.; Ferfache, A.A. General M-Estimator Processes and their m out of n Bootstrap with Functional Nuisance Parameters. Methodol. Comput. Apply Probab. 2022, 24, 2961–3005. [Google Scholar] [CrossRef]
- Mackinnon, J.G. Fast cluster bootstrap methods for linear regression models. Econom. Stat. 2023, 26, 52–71. [Google Scholar] [CrossRef]
- Rüttenauer, T. Spatial Regression Models: A Systematic Comparison of Different Model Specifications Using Monte Carlo Experiments. Sociol. Methods Res. 2022, 51, 728–759. [Google Scholar] [CrossRef]
- Cressie, N. Statistics for Spatial Data; John Wiley & Sons: Nashville, TN, USA, 1993. [Google Scholar]
- Mackinnon, J.G.; Webb, M.D. When and How to Deal with Clustered Errors in Regression Models; Queen’s Economics Department Working Paper; Queen’s University: Kingston, ON, USA, 2020. [Google Scholar]
- Hansen, B.E. Econometrics; University of Wisconsin Department of Economics: Madison, WI, USA, 2020. [Google Scholar]
- Wu, C.F.J. Jackknife bootstrap and other resampling methods in regression analysis. Ann. Stat. 1986, 14, 1261–1295. [Google Scholar] [CrossRef]
- Roodman, D.; MacKinnon, J.G.; Nielsen, M.Ø.; Webb, M.D. Fast and wild: Bootstrap inference in Stata using boottest. Stata J. 2019, 19, 4–60. [Google Scholar] [CrossRef]
- MacKinnon, J.G. Wild cluster bootstrap confidence intervals. L’Actualité Économique 2015, 91, 11–33. [Google Scholar] [CrossRef]
- Mackinnon, J.G.; Nielsen, M.Ø.; Webb, M.D. Cluster-robust inference: A guide to empirical practice. J. Econom. 2023, 232, 272–299. [Google Scholar] [CrossRef]
- Radulović, D. The bootstrap for empirical processes based on stationary observations. Stoch. Process. Their Appl. 1996, 65, 259–279. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).