Abstract
We explore an exchangeably weighted bootstrap of the general function-indexed empirical U-processes in the Markov setting, which is a natural higher-order generalization of the weighted bootstrap empirical processes. As a result of our findings, a considerable variety of bootstrap resampling strategies arise. This paper aims to provide theoretical justifications for the exchangeably weighted bootstrap consistency in the Markov setup. General structural conditions on the classes of functions (possibly unbounded) and the underlying distributions are required to establish our results. This paper provides the first general theoretical study of the bootstrap of the empirical U-processes in the Markov setting. Potential applications include the symmetry test, Kendall’s tau and the test of independence.
Keywords:
bootstrap; Markov chains; regenerative processes; empirical processes; VC classes of functions; U-processes; Donsker classes; weak convergence MSC:
60F05; 60G15; 60K05; 60K15; 62F40
1. Introduction
U-statistics are a class of estimators, initially explored in association with unbiased estimators by [1] and officially introduced by [2], and are defined as follows: let be a sequence of random variables defined on a measurable space , and let be a measurable function, the U-statistics of order m and kernel h based on the sequence are
where
The empirical variance, Gini’s mean difference or Kendall’s rank correlation coefficient are common examples of U-estimators, while a classical test based on a U-statistic is Wilcoxon’s signed rank test for the hypothesis of the location at zero (see, e.g., [3], Example 12.4).The authors in [1,2,4] provided, amongst others, the first asymptotic results for the case in which the underlying random variables have independent and identical distributions. Extensive literature works have treated the theory of U-statistics, for instance, see [5,6,7,8], etc. Complex statistical issues are also amenable to being solved using U-processes. Examples include tests for goodness-of-fit, nonparametric regression and density estimation. U-processes are a set of U-statistics that are indexed by a family of kernels. U-processes might be viewed as infinite-dimensional variants of U-statistics with a single kernel function or as nonlinear stochastic extensions of empirical processes. Both thoughts have the following advantages: first, considering a large group of statistics rather than a single statistic is more statistically interesting. Second, we may use ideas from the theory of empirical processes to construct limit or approximation theorems for U-processes. Nevertheless, achieving results in U-processes is not easy. Extending U-statistics to U-processes necessitates a significant effort and distinct methodologies; generalizing empirical processes to U-processes is quite challenging, especially when U-processes are presented in the stationary setting. We highlight that the U-processes are used often in statistics, such as when higher order terms are a part of von Mises expansions. Particularly, the study of estimators (including function estimators) with various smoothness degrees involves U-statistics. For instance, Ref. [9] applied almost-sure uniform bounds for -canonical U-processes to analyze the product limit estimator for truncated data. Two new tests for normality based on U-processes were also presented in [10]. Inspired by [11,12,13], they developed other tests for normality that employed weighted -distances between the standard normal density and local U-statistics based on standardized observations as test statistics. Estimating the mean of multivariate functions in the case of possibly heavy-tailed distributions was explored by [14]; they presented the median-of-means too, and both explorations were based on U-statistics. Moreover, other researchers emphasized the importance of U-processes; refs. [15,16,17] used them for testing qualitative features of functions in nonparametric statistics, ref. [18] represented the cross-validation for density estimation using U-statistics, in addition to [6,7,19], where the authors established limiting distributions of M-estimators. Since then, this discipline has made significant advancements, and the results have been broadly interpreted. Asymptotic behaviors were demonstrated under weak dependence assumptions, for example, in the works of [20,21,22] or more recently in [23] as well as more generally in [24,25]. However, in practice, explicit computation is not always possible due to the complexity of the U-processes’ limiting distributions or their functionals. We suggest a general bootstrap of the U-processes in the Markov setting to solve this issue, which is a challenging problem. The concept of the bootstrap, given by [26], in the case of independent and identically distributed (iid) random variables, is to resample from an original sample of observations of an unknown marginal distribution function , , a new i.i.d sample with the marginal distribution function , which represents the empirical distribution function constructed from the original sample. Moreover, it is commonly known that the bootstrap approach gives a better approximation to the statistic’s distribution, mainly when the sample size is small [27]. Bootstraps for U-statistics of independent observations were studied by [28,29,30,31]. However, the bootstrap technique is not the same for dependent variables because the dependence structure cannot be conserved in the new sample. For this reason, other blockwise bootstrap methods were introduced, aiming to keep the structure of dependence. Among those methods, we can cite the circular block bootstrap introduced by [32] and the nonoverlapping block bootstrap introduced by [33]. In [34], the authors proposed a bootstrap method related to the weakly dependent stationary observation, the stationary bootstrap. This latter can be seen as an expansion of the circular block bootstrap, where a random variable, such as a geometric random variable, can be used for the block length. It is important to note that Efron’s initial bootstrap formulation (see [26]) had a few flaws. To be more precise, certain observations might be sampled several times while others might not be at all. A more generalized version of the bootstrap, the weighted bootstrap, was developed to get around this issue and was also demonstrated to be computationally more appealing in some applications. This resampling strategy was initially described in [35] and thoroughly investigated by [28], who coined the name “weighted bootstrap”. For example, Bayesian bootstrap when the weighted vector
is equal to the vector of n spacings of ordered uniform random variables in distributions, that is, follows a Dirichlet distribution of parameters For more details, see [36]. This diversity of resampling approaches necessitates the use of a uniform approach, commonly known as general weighted resampling, which was first described by [37] and has since been developed by [38,39]. In [40], the authors investigated the almost-sure rate of convergence of strong approximation for the weighted bootstrap process by a sequence of Brownian bridge processes; refer to [41] for the multivariate setting and [42] for recent references. The concept of the generalized bootstrap, introduced by [37], was extended to the class of nondegenerate U-statistics of degree two and the corresponding Studentized U-statistics by [43]; refer to [44,45]. In [46], the author generalized this theory for a higher order. In his work, he developed a multiplier inequality of a U-process for i.i.d. random variables. We mention that the multiplier processes’ theory is directly and strongly related to the symmetrization inequalities investigated by [6,7].
This paper aims to investigate the exchangeable bootstrap for U-processes in the same way that [46] did but without the restriction of the independence setting. The previous reference focused on U-processes in an independent framework, whereas this paper considers U-processes in the dependent setting of Markov chains. We believe we are the first to present a successful consideration in this general context. We combine the techniques of the renewal bootstrap with the randomly weighted bootstrap in a nontrivial way. We mention a connection between moving-blocks bootstrap and its modification, matched-block bootstrap, at this point. Instead of artificially splitting a sample into fixed-size blocks and then resampling them, the latter seeks to match the blocks to create a smoother transition; for more information, see [47]. The main difficulties in proving Theorem 3 are due to the random size of the resampled blocks. This randomness generates a problem with the random stopping times, which cannot be removed by replacing a random stopping time with its expectation. In the present setting, the bootstrap random variables are generated by resampling from a random number of blocks. One can think that using the conditioning arguments can overcome the problem, but the answer is negative. Our proof uses some arguments from [46,47] by verifying bootstrap stochastic equicontinuity by comparing it to the original process in a similar way as in [48]. However, as we shall see later, integrating concepts from these papers is not enough to solve the problem. To deal with U-processes in the Markov framework, sophisticated mathematical derivations are necessary. We present the first complete theoretical justification of the bootstrap consistency. This justification requires the efficient use of large sample theoretical approaches established for U-empirical processes.
The rest of this paper is organized as follows. Section 2 is devoted to the introduction of the Markov framework, the U-process, the bootstrap weights and the definitions needed in our work. In Section 3, we recall the necessary ingredient for U-statistics and U-processes in the Markov setting. Furthermore, we provide some asymptotic results including the weak convergence of U-processes in Theorem 1. In Section 4, we derive the main results concerning the bootstrap of the U-processes. In Section 5, we collect some examples of weighted U-statistics. Some concluding remarks and possible future developments are relegated to Section 6. To prevent the interruption of the flow of the presentation, all proofs are gathered in Section 7. Appendix A contains a few pertinent technical findings and proofs.
2. Notation and Definitions
In what follows, we aim to properly define our settings. For this reason, we have collected the definitions and notation needed.
2.1. Markov Chain
Let be an homogeneous -irreducible Markov chain, that means that the chain has stationary transition probabilities, defined on a measurable space , where is a separable -algebra. Let be the transition probability and the initial probability. Therefore, we denote by or just the probability measure for . Likewise, denotes the integration with respect to . In our framework, let be a probability measure such that , and is the -expectation. We further assume that the Markov chain is Harris positive recurrent with an atom .
Definition 1
(Harris recurrent). A Markov chain is said to be Harris recurrent if there exists a σ-finite measure such that, for ψ a positive measure on a countable generated measurable space , and if for all with
Recall that a chain is positive Harris recurrent and aperiodic if and only if it is ergodic ([49] Proposition 6.3), i.e., there exists a probability measure , called the stationary distribution, such that, in total variation distance,
Definition 2
(Small sets). A set is said to be Ψ-small if there exists a positive probability measure Ψ supported by S and an integer , such that
Definition 3.
Let be a Markov chain taking value in . We say that is positive recurrent if
- 1.
- is recurrent (or Harris recurrent if is countably generated), where is a set, , m is an integer and ν is a probability measure.
- 2.
- , where is the hitting time of A by the m-step chain, roughly speaking, .
Definition 4.
A ψ-irreducible aperiodic chain X is called regenerative or atomic if there exists a measurable set called an atom, in such a way that and for all we have . Roughly speaking, an atom is a set on which the transition probabilities are the same. If a finite number of states or subsets are visited from the chain, then any state or any subset of the states is actually an atom.
Definition 5
(Aperiodicity). Assuming ψ-irreducibility, there exists and disjoints sets (set ) positively weighted by ψ such that
and
The period of the chain is the greatest common divisor d of such integers, it is said to be aperiodic if .
Definition 6
(Irreducibility). The chain is ψ-irreducible if there exists a σ-finite measure ψ such that, for all set , when , for any , there exists such that .
One of the most important properties of Harris recurrent Markov chains is the existence of an invariant distribution which we is called (a limiting probability distribution, also called occupation measure). Furthermore, Harris recurrent Markov chains can always be embedded in a certain Markov chain on an extended sample space with a recurrent atom. The existence of a recurrent atom gives an immediate consequence for the construction of a regenerative extension of this chain. The time that the chain hits a given atom (recurrent state) is seen as the regenerative time. In [50,51], the authors give the construction of such a regenerative extension. The development of a regenerative extension makes the use of regenerative techniques possible in order to study this type of Markov chain. As we mentioned above, we assume in this work that the Harris recurrent chain is atomic, i.e., the set which is infinitely almost sure is well-defined and accessible, this set is called an atom. By definition, an atom is a set in , where , and for all , . Let (respectively, ) be the probability measure on the underlying space such that (respectively, the -expectation).
The conditions imposed on the Markov chain ensure that the defined atom (or the constructed one in the case of a nonatomic chain) is one recurrent class, and let us define the following terms.
2.1.1. Hitting Times
Define by
A well-known property of the hitting time is that for all , , ([52], chap. I14).
2.1.2. Renewal Times
Using the hitting times, we can define the renewal times as
Similar to the regenerative process, the sequence of renewal times is i.i.d. and it is independent of the choice of the initial probability. All over this work, we set and .
Definition 7
(Strong Markov property). Let be a Markov chain, with T the stopping time of . Then, conditionally on and , is a sequence of a Markov chain and is independent of .
2.1.3. Regenerative Blocks
Let be the number of visits to the atom . Using the strong property of a Markov chain, it is possible to divide the given sample into a sequence of blocks such that:
where is the total number of blocks. The length of each block is denoted by
2.2. Exchangeable Weights
In what follows, represents a real-valued random variable, are independent from . For , we denote the p-norm by
Definition 8
(Exchangeability). Let be a sequence of random variables with joint distribution and let be the group of all permutations acting on . We say that is exchangeable if, for all ,
Assuming the following:
- (A1)
- are exchangeable non-negative, symmetric and for all n
- (A2)
- in -probability which is satisfied by the assumption of the moment
- (A3)
- There exists such that, in -probability,
- (A4)
- Assume
2.3. The U-Process Framework
Let be a sequence of random variables with values in a measurable space . Let be a measurable function symmetric in its arguments. The U-statistic of order (or degree) m and kernel is defined as:
Accordingly, a U-process is the collection , where is the class of kernels of m variables. The decoupling inequality of U-statistics and U-processes plays a central role in the latest developments in the asymptotic theory. As a result, the decoupling inequality can give a relation between the quantities
where is a non-negative function and are independent copies of the original sequence . One of the useful reasons for decoupling is randomization, which is frequently used in the study of the asymptotic theory of U-statistics, and was studied by [6,7]. The main idea of randomization is to compare the tail probabilities or moments of the original U-statistic or process, , with the tail probabilities or moments of the statistic
where are independent Rademacher variables, independent from , and the variables depend on the degree of degeneracy (centering) of the kernel .
Definition 9
([6]). A symmetric -integrable kernel is -degenerate of order if and only if
holds for any , whereas
is not a constant function. If h is furthermore -centered, that is, , we write . For notational simplicity, we usually write .
Moreover, is said to be canonical or completely degenerated if the integral with respect to one variable is equal to zero, i.e.,
The fact that the kernel is completely degenerate with the condition is used for the orthogonality of the different elements of the Hoeffding decomposition of the U-statistics.
Definition 10
(Covering number). The covering number is defined as the minimal number of balls with radius ε that are needed to cover a class of functions in the norm , where Q is the measure on E with finite support.
We can associate some distances to the covering numbers, where
In this work, we use the two distances defined afterward
For decoupled statistics, we also associate covering numbers, well-known as and a distance, which can be defined for as follows:
Definition 11.
A class of measurable functions is said to be of VC type (or Vapnik–Chervonenkis type) for an envelope F and admissible characteristic (positive constants) such that and , if for all probability measure on with and every ,
We assume that the class is countable to avoid measurability issues (the noncountable case may be handled similarly by using an outer probability and additional measurability assumptions, see [53]).
Definition 12
(Stochastic equicontinuity, ([54])). Let be a sequence of stochastic processes. Call stochastically equicontinuous at if for each , there exists a neighborhood D of such that
In the context of the U-process , the stochastic equicontinuity at a function implies generally that should be uniformly small for all close enough to , with high probability and for all n large enough.
2.4. Gaussian Chaos Process
Definition 13.
Let H denote a real separable Hilbert space with scalar product . We say that a stochastic process defined in a complete probability space is an isonormal Gaussian process (or a Gaussian process on H) if is a centered Gaussian family of random variables such that for all .
Define the mapping . Under the assumption mentioned above, this map is linear and it provides a linear isometry of H onto a closed subspace which contains a zero mean Gaussian random variables as its elements. Let be the isonormal Gaussian chaos process associated with determined by:
where and is a polynomial defined as a sum of monomials of degree m; ref. [6] give us a simple expression of this polynomial, extracted from Newton’s identity given by
Therefore,
Hence, by the continuous mapping theorem, we can see that CLT and LLN give:
Under the linearity of the kernel, we only need to show that:
to hold the weak convergence. The limit is useful in the case of degenerate U-statistics and it provides a convergence of all moments, which in turn plays a crucial role because it is due to the hypercontractivity, which makes the uniform integrability better. For a good explanation of , readers are invited to see ([6] Chapter 4, Section 4.2).
2.5. Technical Assumptions
For our results, we need the following assumptions.
- (C.1)
- (Block-length assumption) For all and ,
- (C.2)
- (Nonregenerative blocks) For , we haveand
- (C.3)
- (Block-sum: moment assumptions) For , we haveand
- (C.4)
- For , we have
- (C.5)
- (Nondegeneracy.) We suppose also that
Remark 1
(Moment assumptions). In practice, we recall that block-moment assumptions for the split Markov chain can be generally checked by establishing drift conditions of Lyapunov’s type for the original chain; see Chapter 11 in [55,56], as well as All these moment conditions are discussed in detail in ([57], Chapters 11 and 17). There is a key condition in the proof of ergodic theorems in the Markovian context, which is the fact that , for any that is a set in , such that . In fact, when there is a finite invariant measure and an atom , then this condition is readily found. We also refer to [58] for an explicit check of such conditions on several important examples and to §4.1.2 of [59] for sufficient conditions expressed in terms of a uniform return rate to small sets. Finally, as discussed in Chapter 8 of [60], similar conditions can be expressed in potential kernels. Observe that, in the positive recurrent case, the assumptions of (C.1) are not independent when : from the basic renewal theory, one has for all . Hence, conditions and are equivalent.
3. Preliminary Results
A significant issue was detected in recovering the estimation of our parameter of interest using the U-process. The given shape of this parameter is as follows:
where is a kernel function. The estimation of this parameter should be possible using the U-statistics of the form:
As the parameter of interest is defined and based on Kac’s theorem for the occupation measure, in the regeneration setup can be written as follows:
In the Markovian context and since the variables are not independent, the approximation related to the i.i.d. blocks and the regenerative case is introduced below:
Definition 14
(Regenerative kernel). Let a kernel. We define the regenerative kernel as follows:
It is not necessary that the kernel be symmetric, as soon as . In fact, we can use the symmetrization of in the following way
where the first sum is over all permutations of . Next, we consider the U-statistic formed by the regenerative data.
Definition 15
(Regenerative U-statistic). Let a kernel such that and set The regenerative U-statistic associated with the sequence of regenerative blocks , generated by the Markov chain is given by
Hence, is a standard U-statistic with mean zero.
Proposition 1.
Let us define
Then, under conditions (C.1), (C.2), (C.3) and (C.4), we have the following stochastic convergences:
Before stating the weak convergence in the next theorem, we define the corresponding U-processes related to the U-statistic and the regenerative U-statistic , respectively:
Theorem 1.
Let be a positive recurrent Harris Markov chain, with an accessible atom , satisfies the conditions (C.1) and (C.2) (moments assumptions), (C.3), (C.4), (C.5) and, for a fixed , . Let be a uniform bounded class of functions with an envelope H square-integrable such that:
Then, the process converges weakly in probability under to a Gaussian process indexed by whose sample paths are bounded and uniformly continuous with respect to the metric .
The Bootstrapped U-Processes
Trying to facilitate the bootstrap technique, we write the detailed steps of the regenerative block construction and the weighted bootstrap method in Algorithm 1:
| Algorithm 1 Regenerative block and weighted bootstrap construction. |
|
In what follows, we denote by and , respectively, the conditional probability and the conditional expectation given the sample . The same notation is used for the sample . Define the bootstrapped U-statistic as
and the regenerative bootstrapping
and the U-processes are:
and
Given , a real-valued function, defined on the product probability space, we say that is of an order in -probability if for any
and that is of an order in -probability if for any , there exists a such that
We must comment here that the bootstrap works in probability if
where
and
and is the measurable envelope of . In addition, for any measurable random elements, and Y, the convergence in law of to Y is in the sense of Hoffman–Jorgensen, which is defined as
for g bounded and continuous. This weak convergence is metrizable by Theorem A1 in Appendix A.
Proposition 2.
Suppose that the bootstrap weights satisfy Assumptions (A1)–(A4). Let
Then, we have
The proof of Proposition 2 is postponed until Section 7.
Now, in the following lemma, there are some instrumental results needed later.
Lemma 1.
Let be a Markov chain defined in Section 2.1. Define . Then, for any initial probability ν, we have:
- (i)
- For some and :
- (ii)
- in -probability.
- (iii)
- Let be a sequence of random variables. Ifthen for any integer valued sequence of random variables,
The proof of Lemma 1 is postponed until Section 7.
4. Weighted Bootstrap Weak Convergence
In this section, we extend some existing results concerning the multiplier U-process to prove the bootstrap uniform weak convergence. Most of these results can be found in [46], generalizing the empirical process work of [38] in the i.i.d. setting. The weak convergence is proved for degenerate U-processes, as we mentioned before, and under the weighted regenerative bootstrap schemes described in Algorithm 1. Before stating the weak convergence theorem, we recall the following important results. The next theorem, proved in [46], is a sharp multiplier inequality, which is essential in the study of the multiplier U-process. These results are based on the decoupling symmetrized U-process, a basic framework of U-statistics. In [47], the author solved these problems for the empirical process settings in the Markov setting (multinomial bootstrap), which we generalize to the U-process by considering more general weights, i.e., the exchangeable weighted bootstrap.
Theorem 2
([46]). Let be a random vector independent of . Then, there exists some measurable function such that the expected supremum of the decoupled (Here “decoupled” refers to the fact that are independent copies of , and are independent copies of the Rademacher sequence .) U-processes
for all , consequently,
Furthermore, if there exists a concave and nondecreasing function such that , then
Here, is a constant depending on m only and can be taken as for .
Lemma 2
([46]). Let be function classes such that for all . Suppose that the ’s have the same marginal distributions with . Suppose that there exists some bounded measurable function with as , such that the expected supremum of the decoupled U-processes satisfies
for all . Then,
The main result of this paper is represented in the following theorem. It is worth noting here that it is not easy to prove the stochastic equicontinuity in the present setting as explained in the introduction.
Definition 16
(Permissible classes of function). Let be a measurable space ( a Borel σ-field on E). Let be a class of functions indexed by a parameter x that belongs to a set E. is called permissible if it can be indexed by a E such that:
- There exists a function defined from to in such a way that this function is measurable function, where is the Borel σ-algebra generated by the metric on .
- E is a Suslin measurable space whose mean E is an analytic subset of a compact metric space E from which it inherits its metric and Borel σ-field.
Theorem 3.
Suppose Assumptions (A1) to (A4), and Conditions (C.1)–(C.5) hold. Let be permissible and admit a -square integrable envelope F such that
where the supremum is taken over all discrete probability measures. Then,
where c is the constant in (A3), and the convergence in probability is with respect to the outer probability of defined on .
The proof of Theorem 3 is postponed until Section 7.
4.1. Bootstrap Weights Examples
Let be a class of real random variables satisfying Assumptions (A1)–(A4). We give some examples of bootstrap weights; for instance, refer to [38,61] for more explanations.
4.1.1. Bayesian Resampling Scheme
In this case, are positive i.i.d. random variables with mean and finite variance . The weights satisfy , and we define
The Bayesian bootstrapped weight can be defined as:
satisfying
For or , the Bayesian weights are distributionally equivalent with Dirichlet weights. For the value of , we have:
4.1.2. Efron’s Resampling Scheme
For Efron’s bootstrap, we have
Condition (A1) follows directly. Condition (A3) follows from ([37] Lemma 4.1), and Condition (A2) is detailed in [43].
4.1.3. The Delete h-Jackknife
In [62], the authors permute deterministic weights , where
in order to build new bootstrap weights, and they defined the new weights where is a random permutation uniformly distributed over . These weights are called the delete h-Jackknife. In order to achieve Assumption (A3), we must assume that , as and .
4.1.4. The Multivariate Hypergeometric Resampling Scheme
As its name indicates, the bootstrap weights of this scheme follow the multivariate hypergeometric distribution with density:
where K is a positive integer. Assumption (A3) is satisfied with .
Remark 2.
As was pointed out in [38], the preceding mentioned bootstraps are “smoother” in some way than the multinomial bootstrap because they place some (random) weight on all elements in the sample, whereas the multinomial bootstrap applies the positive weight at a proportion of about of each element of the sample, on average. Notice that when , the are equivalent to four spacings from a sample of Uniform random variables. In [63,64], it was noticed that in addition to being four times more expensive to implement, the choice of four spacings depends on the functional of interest and is not universal.
Remark 3.
It is noteworthy that choosing the bootstrap weights properly implies a smaller limit variance, that is, is smaller than 1. A typical example is the multivariate hypergeometric bootstrap ([38] Example 3.4) and the subsample bootstrap, ([65] Remark 2.2-(3)). A thorough treatment of the weight selection is undoubtedly outside the scope of the current work; for review, we refer the readers to [66].
Remark 4.
In the present paper, we considered a renewal type of bootstrap for atomic Markov chains under minimal moment conditions on renewal times. The atomic Markov chains assumption can be dropped by mimicking the ideas of [50,51] by introducing an artificial atom and deriving the bootstrap procedure that applies to nonatomic Markov chains. Precisely, in the case of a general irreducible chain X with a transition kernel satisfying a minorization condition:
for an accessible measurable set S, a probability measure ψ and (note that such a minorization condition always holds for Π or an iterate when the chain is irreducible), an atomic extension of the chain may be explicitly constructed by the Nummelin splitting technique (see [49]) from the parameters and the transition probability Π, see for instance [47,67]. From a practical viewpoint, the size of the first block may be large compared to the size n of the whole trajectory, for instance, in the case where the expected return time to the (pseudo-)atom when starting with the initial probability distribution is large. The effective sample size for constructing the data blocks and the corresponding statistic is then dramatically reduced. However, in [68], some simulations were given together with examples including content-dependent storage systems and general AR models supporting the method discussed in this work.
5. Applications
Example 1
(Symmetry test). This example gives an application for the bootstrap U-statistics, inspired by the goodness-of-fit tests in [69], where they considered the symmetry test for the distribution of . Let be a stationary mixing process with the Lebesgue density. We test the hypothesis:
The estimator of is:
where is a kernel function and is a smoothing parameter or the bandwidth. An appropriate estimator of the integrated squared difference represent the symmetry test:
According to [69], I can be estimated by
where with , for . Clearly, is a degenerate U-statistic with kernel varying with the sample size n. Thus, the stationary bootstrap test,
can be shown to have the same limit as .
Example 2
(Kendall’s tau). The covariance matrix quantifies the linear dependency in a random vector. The rank correlation is another measure of the nonlinear dependency in a random vector. Two generic vectors and in are said to be concordant if . For , define
Then, Kendall’s tau rank correlation coefficient matrix is a matrix-valued U-statistic with a bounded kernel. It is clear that quantifies the monotonic dependency between and and it is an unbiased estimator of
that is, the probability that and are concordant.
Example 3
(Test of independence). In [2] the author introduced the parameter
where and is the distribution function of and . The parameter Δ has the property that if and only if and are independent. From [8], an alternative expression for Δ can be developed by introducing the functions
and
We have
The corresponding U-statistics may be used to test the independence.
6. Conclusions
The present paper was concerned with the randomly weighted bootstrap of the U-process in a Markov framework. A large number of bootstrap resampling schemes emerged as special cases of our setting, in particular, the multinomial bootstrap, which is the best-known bootstrap scheme introduced by [26]. One of the main tools was the approximation of the Markov U-process by the corresponding regenerative one. We looked to mimic this result in Proposition 2, in order to approximate the weighted-bootstrap U-process to the regenerative weighted-bootstrap U-process . Other technical arguments were given in Lemma 1 extended from the work of [47]. These intricate tools were used to reach the full independence of regenerative block variables by proving that a deterministic one could substitute the random size of blocks, which was the main problem for the extension of the bootstrap results to the Markov framework. After a lengthy proof to arrive at independence, we used the results of [46]. All the above steps led us to prove the weak convergence of the regenerative-block weighted-bootstrap U-process, which implied the weak convergence of the weighted-bootstrap U-process. It will be of interest to consider the extension of the paper to the semi-Markov setting. A more delicate problem is to consider the setting of incomplete data such as censored cases or missing data. To the best of our knowledge, this problem has not been considered, even for the original sample (without bootstrap) in the Markov framework. It would be interesting to extend our work to the case of the local stationary process, which requires nontrivial mathematics; this would go well beyond the scope of the present paper.
7. Mathematical Development
This section is devoted to the proof of our results. The previously defined notations continue to be used in what follows.
Proof of Proposition 2.
We have
Given is not excluded), and , we set to be the point of obtained from i by deleting the coordinates in the places not in J (e.g., if , then . Furthermore, indicates the sum over ; for instance, if and , then
By convention, . Notice that
In a similar way, we have
Making use of Proposition 1 and the law of large numbers, we infer that
Hence, the proof is completed. □
Proof of Lemma 1.
The proof of part and part follows from ([47] Lemma 3.1 and Lemma 3.2). In order to prove , we need to show that, for every ,
which follows if, conditioned on the sample,
We have:
We denote by the expectation conditionally on . By the fact that are i.i.d. and using Chebyshev’s inequality, we have:
The last inequality follows using , which implies that and where
for . For we have:
The last equality converges to zero by the fact that and by iii)
This proves Lemma 1. □
Proof of Theorem 3.
For the weak convergence, we need to show the finite-dimensional convergence and the asymptotic equicontinuity. According to Proposition 2 and [6], the finite-dimensional convergence is considered if, for every fixed finite collection of functions ,
where is the Gaussian chaos process. According to Cramér–Wold and the countability of , we only need to show that for any ,
By ([6] Section 4.2) and ([29] Section 2A), any can be expanded in by , where is a sequence of real numbers and
for some bounded . Fix . Then, there exists such that with ,
The left-hand side of (25) can be further bounded by
Let ; noting that is bounded by one and using Lemma 1, we can replace by which is deterministic. In the following, we denote by a random permutation uniformly distributed over , the set of all permutations over We have
We have, according to [43], for a non-negative sequence of variables such that and for a random permutation of , for any and ,
Furthermore, according to [70,71], we have:
Hence we have
Now for the second term, we have:
where is the polynomial of degree m (see [6], p. 175):
As we mentioned before, this polynomial follows from Newton’s inequality and allows us to show a polynomial function as a sum of monomials. All we need now is to check each argument of this polynomial function. □
For : We first recall the following lemma from [53].
Lemma 3
([53]). Let be a vector and be a vector of exchangeable random variables. Suppose that
and
Then,
Applying Lemma 3 with and replaced by , we can see that
where is a Gaussian process defined on with covariance
For : Note that
Furthermore,
The first inequality in the above display follows, since
This shows that
For :
This shows that
Then, we have
where is the Gaussian chaos process defined on (⊕ is the orthogonal sum in )
Hence, it follows that, by linearity of ,
The last term in (26) follows from the definition of
All these final results give the finite-dimensional convergence.
Now, we take a step-by-step approach to establish stochastic equicontinuity. We assume that the class of functions must be bounded, so we suppose that , for H an envelope. Throughout the following, we denote by
- Step 1
Let
and
In this step, we must prove that the stochastic equicontinuity of the U-process implies that of the regenerative U-process. This is a consequence of 1, and for the weighted bootstrap Proposition 2 and part ii) of Lemma 1.
- Step 2
Define
and
Hypothesis: The stochastic equicontinuity of implies the stochastic equicontinuity of .
Proof.
In order to prove the previous implication, we only need to show that:
Suppose that , the opposite case can be treated in a similar way. We have
However, by Lemma 1, part i). Then, the exists a constant , such that for every ,
and the first expression in the previous expression is bounded by:
The last expression follows from the Montgomery–Smith inequality. Since
the last expression matches the stochastic equicontinuity condition for . This proves this step. □
Before passing to the next step, we introduce a new bootstrap sample. Define for . Now, apply the weighted bootstrap procedure on the sample . This new procedure is the same as the old one for , but we aim here to replace the random quantity with a deterministic one, which is .
- Step 3
Define:
Hypothesis: The stochastic equicontinuity of implies the stochastic equicontinuity of .
Proof.
First case: :
In this case, all of the terms in the following computation should be multiplied with . We leave it out to keep the already complex notation simple. Define
is well defined, i.i.d., and has the same distribution as and . Hence, if we show that:
then the stochastic equicontinuity of is established. However, we aim to approximate the one of . In order to achieve our goal, it is sufficient to estimate:
For : Let
conditioned on the sample, we have:
Hence,
where
For n large enough, we need to show that there exists such that
As are i.i.d and bounded,
therefore, we can find such that
However,
by Lemma 1 i), then
Then, we only need to estimate the first part in (33). Define the following bootstrap procedure: let and let be a class of function, related to the class of functions , such that, for every :
It is classical that are i.i.d., applying the same bootstrap method of Algorithm 1. This new sample allows us to enlarge and bound (33) by
where and the corresponding class , with envelope and F, respectively. To estimate the last expression, we use bracketing. Define the bracket by:
and the bracketing entropy number by , which denotes the minimal number for which there exist functions and such that:
For the class of functions , consider the bracket , such that , where and it is determined later. In this framework, the bracketing entropy number is , for
Hence, we have the following inequalities
Treating each term, keeping in mind Condition (A.1), i.e., , we have
and
yet, are i.i.d. and and , so for any , we have
Using the same argument as in part iii) of Lemma 1, we can prove that
Then, it remains to find that, for every fixed , is bounded in probability, as the last expression in (38) does not depend on k. It is interesting to note that is finite, due to the boundness of by with and the fact that are i.i.d. and discrete random variables. Under the norm , define brackets and . Observe that
converges to zero in probability, and does not depend on n. That implies that in probability. Replacing by , is identical to , i.e., also converges to zero in probability. This proves the convergence of to zero in probability.
For : In the same manner, let
Define a new bootstrap sample in . Clearly, the new sample is well-defined since we assumed at the beginning that , and it is defined independently from and . In this case:
Hence, as in (33), we have:
where
Using the same bootstrap procedure defined previously for , let
for , and let be a class of function such that, for every :
It is classical that are i.i.d., applying the same bootstrap method of Algorithm 1. This new sample allows us to enlarge and bound (33) by
where
corresponding to the class
with envelope and F, respectively. As before, for the class of functions , consider the bracket , such that
where and it is determined later. In this framework, the bracketing entropy number is , for
Following the same arguments from Equations (37) through (38), we can find that (42) is
Here, we must pay attention to the randomness of which depends on n. According to Lemma 1 i), we can see that in probability, under the assumption that . Now, using the same treatment of , and for ( to provide the convergence of to ∞), as in [47], this allows the convergence of (43) to zero in probability. Estimating now by considering the same brackets and , we have , which does not depend on n. Then, is proved. Following the same footsteps, we can prove the case where . This proves Step 3. □
The end of the previous step yields that we only need to show the stochastic equicontinuity of , where the number of blocks is replaced by the deterministic quantity . In order to achieve the equicontinuity of this statistic, Lemma 2 shows that it is sufficient to prove that:
for all . We begin to define the distance:
defined in , associated with the Rademacher process
Take and
Using Corollary A1, we have
Assuming that , the upper bound in the integral can be replaced by . The following proposition is necessary for the following.
Proposition 3
([46]). Let be i.i.d. random variables with law . Let be a class of measurable real-valued functions defined on with an -integrable envelope such that the following holds: for any fixed ,
holds for any . Here for and ,
and
where is an envelope for . Then,
in as . The above equation can be replaced by the decoupled version.
By this proposition, as , therefore, it suffices to get as and . It is obvious that all that is left to do now is to demonstrate that
Verifying condition (45)
The shift from the second to the third line is true because
As the condition is verified, as well as , (46) follows directly using the previous proposition. Hence, there exists some sequence , in a way that for any sequence with both under , such that:
An application of Lemma 2 proves that
This completes the proof for the asymptotic equicontinuity.
Author Contributions
I.S. and S.B.: conceptualization, methodology, investigation, writing—original draft, writing—review and editing. All authors contributed equally to the writing of this paper. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
Not applicable.
Acknowledgments
The authors would like to thank the Special Issue Editor of the Special Issue on “Current Developments in Theoretical and Applied Statistics”, Christophe Chesneau for the invitation. The authors are indebted to the Editor-in-Chief and the three referees for their very generous comments and suggestions on the first version of our article, which helped us to improve the content, presentation and layout of the manuscript.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
This appendix contains supplementary information that is an essential part of providing a more comprehensive understanding of the paper. We also refer to [46] for more details.
Proof of Proposition 1.
Let and the possibly empty non-regenerative blocks of observations. Note that, for , the demonstration can be viewed directly in [59], under the assumptions (C1), (C2) and (C3), we can see that
Otherwise, for , we can write as follows:
where
where
the complement of index set, with cardinal equal to . To prove the convergence of to zero in probability, we must fulfill the convergence of (I) and (II) to zero in probability.
where and . We apply the SLLN for Harris Markov chains to find the convergence of
to
Using the conditions, all terms in and are finite and we can prove the convergence of to zero. Now, for , applying the SLLN and by Lemma 3.2 in [47] part i), we can see that
We have
We obtain, in turn, that
Hence, also converges to zero a.s under as . □
Proof of Theorem 1.
In what follows, let denote the number of blocks observed. We find that
where
where represents the conditional expectation of given the c of the coordinates, for all . The U-statistics is obtained by truncating the Hoeffding decomposition after the first term . Then, we just need to show that:
For , introduce
Using (A1), we can replace the random variable with the deterministic quantity and we write
where In order to establish the weak convergence for the empirical process , it is sufficient and necessary to prove the finite dimensional convergence and the stochastic equicontinuity. For the finite multidimensional convergence, we have to prove that converges weakly to , for every fixed finite collection of functions
In order to fix this, it is enough to show that for every fixed ,
where
By linearity, and in the same footsteps of the arguments of ([57], Chapter 17), we can prove that
where, under Condition (C5),
We readily infer that we have
Now, to verify the equicontinuity, we need to check that for every ,
where is a pseudo distance for which the class is totally bounded, and belong to . According to [72], we have
where For the left-hand part in the last inequality, we have
Dividing the last inequality by and using the convergence result in ([72] Lemma 2.11) with Condition (C1), we obtain the desired result. The right-hand part in the inequality is treated using ([72] Lemma 4.2) providing that for and the hypothesis of a finite uniform entropy integral. To complete the weak convergence of the regenerative U-statistic, we must treat the remaining terms of its Hoeffding decomposition. For , let us introduce
Once can see that is centered, that is
By the randomization theorem, according to [7] (for ):
Hence, for a constant:
It is sufficient now to use the theorem hypothesis of a uniform entropy integral to complete the proof of the theorem. □
Proof of Theorem 2.
We have
By decoupling of the U-process, due to [6],
By symmetrization, due to [6], we have
for a sequence independent and with the same distribution as . By the invariance of and the fact that is independent of , we have that
using the reversed order statistics of , , and the permutations between the different sequences of random variables, and in the same footsteps as [46],
substituting by , with , we have
Now, suppose that . Then, we may further bound the above equation by
where the last inequality follows from the generalized Hölder inequality and the assumption that is nondecreasing. □
Proof of Lemma 2.
For
Theorem 2 implies that:
Here
as long as none of vanishes. The claim now follows from the dominated convergence theorem. □
Corollary A1
([6]). Let , be a (weak) Gaussian or Rademacher chaos process of degree m and let
If
then there is a version of X, which we keep denoting X, with almost all of its sample paths in and such that
and
for all , where K is a universal constant and D is the diameter of T for the pseudodistance . In fact, every separable version of X satisfies these properties.
Theorem A1
([73]). For any random elements with values in a metric space , where Y is measurable and has a separable range, the following are equivalent:
- converge in law to Y;
- as ;
References
- Halmos, P.R. The theory of unbiased estimation. Ann. Math. Stat. 1946, 17, 34–43. [Google Scholar] [CrossRef]
- Hoeffding, W. A class of statistics with asymptotically normal distribution. Ann. Math. Stat. 1948, 19, 293–325. [Google Scholar] [CrossRef]
- van der Vaart, A.W. Asymptotic Statistics; Cambridge Series in Statistical and Probabilistic Mathematics; Cambridge University Press: Cambridge, MA, USA, 1998; Volume 3, p. xvi+443. [Google Scholar]
- Mises, R.V. On the asymptotic distribution of differentiable statistical functions. Ann. Math. Stat. 1947, 18, 309–348. [Google Scholar] [CrossRef]
- Serfling, R.J. Approximation Theorems of Mathematical Statistics; John Wiley & Sons: Hoboken, NJ, USA, 2009; Volume 162. [Google Scholar]
- de la Peña, V.H.; Giné, E. Decoupling; From dependence to independence, Randomly stopped processes. U-statistics and processes. Martingales and beyond; Probability and its Applications (New York); Springer: New York, NY, USA, 1999; p. xvi+392. [Google Scholar]
- Arcones, M.A.; Giné, E. Limit theorems for U-processes. Ann. Probab. 1993, 21, 1494–1542. [Google Scholar] [CrossRef]
- Lee, A.J. U-statistics; Volume 110, Statistics: Textbooks and Monographs; Theory and practice; Marcel Dekker, Inc.: New York, NY, USA, 1990; p. xii+302. [Google Scholar]
- Stute, W. Almost sure representations of the product-limit estimator for truncated data. Ann. Statist. 1993, 21, 146–156. [Google Scholar] [CrossRef]
- Arcones, M.A.; Wang, Y. Some new tests for normality based on U-processes. Statist. Probab. Lett. 2006, 76, 69–82. [Google Scholar] [CrossRef]
- Giné, E.; Mason, D.M. Laws of the iterated logarithm for the local U-statistic process. J. Theoret. Probab. 2007, 20, 457–485. [Google Scholar] [CrossRef]
- Giné, E.; Mason, D.M. On local U-statistic processes and the estimation of densities of functions of several sample variables. Ann. Statist. 2007, 35, 1105–1145. [Google Scholar] [CrossRef]
- Schick, A.; Wang, Y.; Wefelmeyer, W. Tests for normality based on density estimators of convolutions. Statist. Probab. Lett. 2011, 81, 337–343. [Google Scholar] [CrossRef]
- Joly, E.; Lugosi, G. Robust estimation of U-statistics. Stochastic Process. Appl. 2016, 126, 3760–3773. [Google Scholar] [CrossRef]
- Lee, S.; Linton, O.; Whang, Y.J. Testing for stochastic monotonicity. Econometrica 2009, 77, 585–602. [Google Scholar]
- Ghosal, S.; Sen, A.; van der Vaart, A.W. Testing monotonicity of regression. Ann. Statist. 2000, 28, 1054–1082. [Google Scholar] [CrossRef]
- Abrevaya, J.; Jiang, W. A nonparametric approach to measuring and testing curvature. J. Bus. Econom. Statist. 2005, 23, 1–19. [Google Scholar] [CrossRef]
- Nolan, D.; Pollard, D. U-processes: Rates of convergence. Ann. Statist. 1987, 15, 780–799. [Google Scholar] [CrossRef]
- Sherman, R.P. Maximal inequalities for degenerate U-processes with applications to optimization estimators. Ann. Statist. 1994, 22, 439–459. [Google Scholar] [CrossRef]
- Yoshihara, K.i. Limiting behavior of U-statistics for stationary, absolutely regular processes. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 1976, 35, 237–252. [Google Scholar] [CrossRef]
- Borovkova, S.; Burton, R.; Dehling, H. Limit theorems for functionals of mixing processes with applications to U-statistics and dimension estimation. Trans. Amer. Math. Soc. 2001, 353, 4261–4318. [Google Scholar] [CrossRef]
- Denker, M.; Keller, G. On U-statistics and v. Mises’ statistics for weakly dependent processes. Z. Wahrsch. Verw. Gebiete 1983, 64, 505–522. [Google Scholar] [CrossRef]
- Leucht, A. Degenerate U- and V-statistics under weak dependence: Asymptotic theory and bootstrap consistency. Bernoulli 2012, 18, 552–585. [Google Scholar] [CrossRef]
- Leucht, A.; Neumann, M.H. Degenerate U- and V-statistics under ergodicity: Asymptotics, bootstrap and applications in statistics. Ann. Inst. Statist. Math. 2013, 65, 349–386. [Google Scholar] [CrossRef]
- Bouzebda, S.; Nemouchi, B. Weak-convergence of empirical conditional processes and conditional U-processes involving functional mixing data. In Statistical Inference for Stochastic Processes; Springer: New York, NY, USA, 2022; pp. 1–56. [Google Scholar]
- Efron, B. Bootstrap methods: Another look at the jackknife. Ann. Statist. 1979, 7, 1–26. [Google Scholar] [CrossRef]
- Hall, P. The Bootstrap and Edgeworth Expansion; Springer Series in Statistics; Springer: New York, NY, USA, 1992; p. xiv+352. [Google Scholar]
- Bickel, P.J.; Freedman, D.A. Some asymptotic theory for the bootstrap. Ann. Statist. 1981, 9, 1196–1217. [Google Scholar] [CrossRef]
- Arcones, M.A.; Giné, E. On the bootstrap of U and V statistics. Ann. Statist. 1992, 20, 655–674. [Google Scholar] [CrossRef]
- Dehling, H.; Mikosch, T. Random quadratic forms and the bootstrap for U-statistics. J. Multivariate Anal. 1994, 51, 392–413. [Google Scholar] [CrossRef]
- Leucht, A.; Neumann, M.H. Consistency of general bootstrap methods for degenerate U-type and V-type statistics. J. Multivariate Anal. 2009, 100, 1622–1633. [Google Scholar] [CrossRef]
- Politis, D.N.; Romano, J.P. A circular block-resampling procedure for stationary data. In Exploring the Limits of Bootstrap (East Lansing, MI, 1990); Wiley Ser. Probab. Math. Statist. Probab. Math. Statist.; Wiley: New York, NY, USA, 1992; pp. 263–270. [Google Scholar]
- Carlstein, E. The use of subseries values for estimating the variance of a general statistic from a stationary sequence. Ann. Statist. 1986, 14, 1171–1179. [Google Scholar] [CrossRef]
- Politis, D.N.; Romano, J.P. The stationary bootstrap. J. Amer. Statist. Assoc. 1994, 89, 1303–1313. [Google Scholar] [CrossRef]
- Rubin, D.B. The Bayesian bootstrap. Ann. Statist. 1981, 9, 130–134. [Google Scholar] [CrossRef]
- Lo, A.Y. A Bayesian method for weighted sampling. Ann. Statist. 1993, 21, 2138–2148. [Google Scholar] [CrossRef]
- Mason, D.M.; Newton, M.A. A rank statistics approach to the consistency of a general bootstrap. Ann. Statist. 1992, 20, 1611–1624. [Google Scholar] [CrossRef]
- Præstgaard, J.; Wellner, J.A. Exchangeably weighted bootstraps of the general empirical process. Ann. Probab. 1993, 21, 2053–2086. [Google Scholar] [CrossRef]
- van der Vaart, A. New Donsker classes. Ann. Probab. 1996, 24, 2128–2140. [Google Scholar] [CrossRef]
- Alvarez-Andrade, S.; Bouzebda, S. Strong approximations for weighted bootstrap of empirical and quantile processes with applications. Stat. Methodol. 2013, 11, 36–52. [Google Scholar] [CrossRef]
- Bouzebda, S. On the strong approximation of bootstrapped empirical copula processes with applications. Math. Methods Statist. 2012, 21, 153–188. [Google Scholar] [CrossRef]
- Bouzebda, S.; Elhattab, I.; Ferfache, A.A. General M-Estimator Processes and their m out of n Bootstrap with Functional Nuisance Parameters. In Methodology and Computing in Applied Probability; Springer: New York, NY, USA, 2022; pp. 1–45. [Google Scholar]
- Huskova, M.; Janssen, P. Consistency of the generalized bootstrap for degenerate U-statistics. Ann. Stat. 1993, 21, 1811–1823. [Google Scholar] [CrossRef]
- Janssen, P. Weighted bootstrapping of U-statistics. J. Stat. Plan. Inference 1994, 38, 31–41. [Google Scholar] [CrossRef]
- Alvarez-Andrade, S.; Bouzebda, S. Cramér’s type results for some bootstrapped U-statistics. Statist. Papers 2020, 61, 1685–1699. [Google Scholar] [CrossRef]
- Han, Q. Multiplier U-processes: Sharp bounds and applications. Bernoulli 2022, 28, 87–124. [Google Scholar] [CrossRef]
- Radulović, D. Renewal type bootstrap for Markov chains. Test 2004, 13, 147–192. [Google Scholar] [CrossRef]
- Giné, E.; Zinn, J. Bootstrapping general empirical measures. Ann. Probab. 1990, 18, 851–869. [Google Scholar] [CrossRef]
- Nummelin, E. General Irreducible Markov Chains and Nonnegative Operators; Cambridge Tracts in Mathematics; Cambridge University Press: Cambridge, UK, 1984; Volume 83, p. xi+156. [Google Scholar]
- Athreya, K.B.; Ney, P. A new approach to the limit theory of recurrent Markov chains. Trans. Amer. Math. Soc. 1978, 245, 493–501. [Google Scholar] [CrossRef]
- Nummelin, E. A splitting technique for Harris recurrent Markov chains. Z. Wahrsch. Verw. Gebiete 1978, 43, 309–318. [Google Scholar] [CrossRef]
- Chung, K.L. Markov Chains with Stationary Transition Probabilities, 2nd ed.; Die Grundlehren der mathematischen Wissenschaften, Band 104; Springer: New York, NY, USA, 1967; p. xi+301. [Google Scholar]
- van der Vaart, A.W.; Wellner, J.A. Weak Convergence and Empirical Processes; Springer Series in Statistics; With applications to statistics; Springer: New York, NY, USA, 1996; p. xvi+508. [Google Scholar]
- Pollard, D. Convergence of Stochastic Processes; Springer Series in Statistics; Springer: New York, NY, USA, 1984; p. xiv+215. [Google Scholar]
- Douc, R.; Guillin, A.; Moulines, E. Bounds on regeneration times and limit theorems for subgeometric Markov chains. Ann. Inst. Henri Poincaré Probab. Stat. 2008, 44, 239–257. [Google Scholar] [CrossRef]
- Meyn, S.; Tweedie, R.L. Markov Chains and Stochastic Stability, 2nd ed.; Cambridge University Press: Cambridge, MA, USA, 2009. [Google Scholar]
- Meyn, S.P.; Tweedie, R.L. Markov Chains and Stochastic Stability; Communications and Control Engineering Series; Springer: London, UK, 1993; p. xvi+548. [Google Scholar]
- Bertail, P.; Clémençon, S. Regeneration-based statistics for Harris recurrent Markov chains. In Dependence in Probability and Statistics; Springer: New York, NY, USA, 2006; Volume 187, pp. 3–54. [Google Scholar]
- Bertail, P.; Clémençon, S. A renewal approach to Markovian U-statistics. Math. Methods Statist. 2011, 20, 79–105. [Google Scholar] [CrossRef]
- Revuz, D. Markov Chains, 2nd ed.; North-Holland Publishing Co.: Amsterdam, The Netherlands, 1984; Volume 11. [Google Scholar]
- Cheng, G. Moment consistency of the exchangeably weighted bootstrap for semiparametric M-estimation. Scand. J. Stat. 2015, 42, 665–684. [Google Scholar] [CrossRef]
- Shao, J.; Wu, C. Heteroscedasticity-robustness of jackknife variance estimators in linear models. Ann. Statist. 1987, 15, 1563–1579. [Google Scholar] [CrossRef]
- Weng, C.S. On a second-order asymptotic property of the Bayesian bootstrap mean. Ann. Statist. 1989, 17, 705–710. [Google Scholar] [CrossRef]
- van Zwet, W.R. The Edgeworth expansion for linear combinations of uniform order statistics. In Second Prague Symposium on Asymptotic Statistics (Hradec Králové, 1978); North-Holland: Amsterdam, NY, USA, 1979; pp. 93–101. [Google Scholar]
- Pauly, M. Consistency of the subsample bootstrap empirical process. Statistics 2012, 46, 621–626. [Google Scholar] [CrossRef]
- Shao, J.; Tu, D.S. The Jackknife and Bootstrap; Springer Series in Statistics; Springer: New York, NY, USA, 1995; p. xviii+516. [Google Scholar]
- Bertail, P.; Clémençon, S. Regenerative block bootstrap for Markov chains. Bernoulli 2006, 12, 689–712. [Google Scholar] [CrossRef]
- Bertail, P.; Clémençon, S. Approximate regenerative-block bootstrap for Markov chains. Comput. Statist. Data Anal. 2008, 52, 2739–2756. [Google Scholar] [CrossRef]
- Fan, Y.; Ullah, A. On goodness-of-fit tests for weakly dependent processes using kernel method. J. Nonparametr. Statist. 1999, 11, 337–360, First NIU Symposium on Statistical Sciences (De Kalb, IL, 1996). [Google Scholar] [CrossRef]
- Frees, E.W. Infinite order U-statistics. Scand. J. Statist. 1989, 16, 29–45. [Google Scholar]
- Rempala, G.; Gupta, A. Weak limits of U-statistics of infinite order. Random Oper. Stochastic Equations 1999, 7, 39–52. [Google Scholar] [CrossRef]
- Levental, S. Uniform limit theorems for Harris recurrent Markov chains. Probab. Theory Related Fields 1988, 80, 101–118. [Google Scholar] [CrossRef]
- Dudley, R.M. Nonlinear functionals of empirical measures and the bootstrap. In Probability in Banach Spaces, 7 (Oberwolfach, 1988); Progr. Probab.; Birkhäuser Boston: Boston, MA, USA, 1990; Volume 21, pp. 63–82. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).