Next Article in Journal
Optimal Control for an Epidemic Model of COVID-19 with Time-Varying Parameters
Previous Article in Journal
Streamlining Ocean Dynamics Modeling with Fourier Neural Operators: A Multiobjective Hyperparameter and Architecture Optimization Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Almost Sure Central Limit Theorem for Error Variance Estimator in Pth-Order Nonlinear Autoregressive Processes

School of Mathematic, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(10), 1482; https://doi.org/10.3390/math12101482
Submission received: 2 April 2024 / Revised: 6 May 2024 / Accepted: 9 May 2024 / Published: 10 May 2024
(This article belongs to the Section Probability and Statistics)

Abstract

:
In this paper, under some suitable assumptions, using the Taylor expansion, Borel–Cantelli lemma and the almost sure central limit theorem for independent random variables, the almost sure central limit theorem for error variance estimator in the pth-order nonlinear autoregressive processes with independent and identical distributed errors was established. Four examples, first-order autoregressive processes, self-exciting threshold autoregressive processes, threshold-exponential AR progresses and multilayer perceptrons progress, are given to verify the results.

1. Introduction

Over the past twenty years, there has been an increasing interest in the nonlinear time series literature, for example, the monograph by Tong [1] represents a good account of nonlinear time series models. Compared to linear models, studying the properties of estimators in nonlinear time series models is technically more complex and difficult. In this paper, we will investigate the properties of estimators in nonlinear autoregressive processes.
Throughout this paper, we always assume that  { ε i , i Z }  is a sequence of independent and identically distributed random variables with mean zero, finite variance  σ 2 { X i , i Z }  is a sequence of strictly stationary real random variables which satisfies nonlinear autoregressive processes of order p
X i = r θ X i 1 , , X i p + ε i ,
for some  θ = θ 1 , , θ q Θ R q , where  r θ θ Θ , is a family of known measurable functions from  R p R . Obviously,  X i 1 , , X i s  are independent of  { ε j , j i } .
In recent years, many authors have studied the properties of estimators for the error sequence. One research interest is the error density estimator, for example, Liebscher [2] proved the law of logarithm and the law of iterated logarithm of the M-estimator in the nonlinear autoregressive processes of order p with independent errors. Cheng and Sun [3] studied the goodness-of-fit test of the errors in the nonlinear autoregressive processes of order p with independent and identical distributed errors. Fu and Yang [4] obtained the asymptotic normality of error kernel density estimators in the pth-order nonlinear autoregressive processes with independent and identical distributed errors. Cheng [5] obtained the asymptotic distribution of the maximum of a suitably normalized deviation of the density estimator from the expectation of the kernel error density. Li [6] established the asymptotic normality of the  L p -norms of error density estimators in the pth-order nonlinear autoregressive processes with independent and identical distributed errors. Kim et al. [7] considered the goodness-of-fit test of the errors in the nonlinear autoregressive processes of order p with a stationary  α -mixing error. Cheng [8] considered the uniform strong consistency of the classical Glivenko–Cantelli Theorem for the residual-based empirical error in the pth-order nonlinear autoregressive processes with independent and identical distributed errors. Liu and Zhang [9] established the law of the iterated logarithm for error density estimators in the pth-order nonlinear autoregressive processes with independent and identical distributed errors.
The other research interest is the error variance estimator. Cheng [10] obtained the consistency and asymptotic normality of the variance estimator in the pth-order nonlinear autoregressive processes with independent and identical distributed errors. As we know, there are few results about the error variance estimators except for Cheng [10], and there are no results for the almost sure central limit theorem for the error variance estimator, and therefore, we will study the almost sure central limit theorem for the error variance estimator in this paper.
The almost sure central limit theorem (ASCLT, for short) has been first introduced independently by Brosamler [11] and Schatte [12]. Since then many interesting results have been discovered in this field. The classical ASCLT for a sequence  { X , X n ; n 1 }  of i.i.d. random variables with zero means states that when  Var ( X ) = σ 2 ,
lim n 1 D n k = 1 n d k I S k k σ x = Φ ( x ) a . s .
for all  x R  with the logarithmic averages  d k = 1 / k  and  D n = k = 1 n d k S k = j = 1 k X j . However, logarithmic averaging is not the only one providing a.s. convergence for partial sums of i.i.d. random variables. Peligrad and Révész [13] showed that (2) holds with  d k = ( log k ) α / k α > 1 . Berkes and Csáki [14] showed that (2) holds also if  d k = exp { ( log k ) α } / k 0 α < 1 / 2 . To compare these results, Hörmann [15], Tong et al. [16], Miao [17], Li [18], Zhang [19,20], Wu and Jiang [21], and Li and Zhang [22,23,24] showed that the a.s. limit (2) holds for any weight sequence  { d k }  satisfying a mild growth condition similar to Kolmogorov’s condition on the law of iterated logarithm.
The paper is organized as follows. In Section 1, the significance and background of research is introduced. Some assumptions and main results are stated in Section 2. Several useful lemmas are listed in Section 3. The proofs are listed in Section 4. Examples are stated in Section 5. In the sequel, we denote with  C , C 1 , C 2 ,  generic constants that may be different in each of its appearances.  I { A }  denotes the indicator function of the set A Φ ( · )  denotes the distribution function of the standard normal random variable  N .

2. Main Results

The goal of this paper is to study the properties of the estimator of the error variance  σ 2  by means of the observations  { X 1 , X 2 , , X n }  in model (1). The main difficulty is that we do not observe the error  { ε 1 , ε 2 , , ε n } , the structure of estimation of parameters is complex, and the residuals are unknown. We need to use Taylor’s expansion and many other techniques to deal with it. This is the greatest contribution of this paper. We will follow the following steps. Firstly, we compute an estimator  θ ˆ = ( θ ˆ 1 , , θ ˆ q )  of unknown parameter  θ . Secondly, based on the estimator  θ ˆ  and model (1), we calculate the following residuals
ε ˆ i = X i r θ ˆ X i 1 , , X i p , i = 1 , 2 , , n .
Finally, using the above residuals, we estimate the error variance  σ 2  by using the following equation
σ ˆ n 2 = 1 n i = 1 n ε ˆ i 2 .
Before giving the main results, we need the following basic assumptions for model (1) which will be used throughout the paper. For  1 i n  and  1 j q , let
Y i j θ j r θ ( X i 1 , , X i p ) , Z i j l 2 θ j θ l r θ * ( X i 1 , , X i p ) ,
where  θ * = θ + λ ( θ ˆ θ )  for some  λ ( 0 , 1 ) . By the fact that  X i 1 , , X i s  are independent of  { ε j , j i } , we conclude that  ε i  is independent of  Y i j .
Assumption 1.
Let  U Θ R q  be an open neighborhood of θ. For any  y R p θ = ( θ 1 , , θ q ) U j , l = 1 , , q , assume that
θ j r θ ( y ) M 1 ( y ) , 2 θ j θ l r θ ( y ) M 2 ( y ) ,
where  E [ M 1 4 ( X i 1 , , X i p ) ] <  and  E [ M 2 4 ( X i 1 , , X i p ) ] <  for each  i 1 .
Assumption 2.
Let  θ ˆ = ( θ ˆ 1 , , θ ˆ q )  be a strong consistent estimator for θ satisfying the following law of iterated logarithm
lim sup n n log log n θ ˆ θ C , a . s . ,
where  | θ ˆ θ | = j = 1 q ( θ ˆ j θ j ) 2  and C is a positive constant.
Remark 1.
By Corollary 2.2 of Klimko and Nelson [25], we know that the least square estimator for the stochastic process under some suitable conditions satisfies (5). For the first-order autoregressive progresses, Wang et al. [26] proved that the least square estimator of the unknown parameters meets (5). For smooth threshold autoregressive progresses, Chan and Tong [27] obtained the conditional least square estimators of the unknown parameters that satisfy (5). For general nonlinear autoregressive progresses of order p, Liebscher [2] established M-estimators for the unknown parameters that satisfy (5), Yao [28] obtained (5) for least square estimators of nonlinear autoregressive progresses.
Now, we will state the main result for the almost sure central limit theorem of the error variance estimator  σ ˆ 2 .
Theorem 1.
Suppose that  { d k }  is a sequence of positive numbers satisfying the following conditions:
(C1) 
lim sup k k d k ( log D k ) ρ / D k <  for some  ρ > 1 , where  D n = k = 1 n d k .
(C2) 
D n D n = o ( n ϵ ) , for any  ϵ > 0 .
For model (1), under the Assumptions 1 and 2, if  E ε 1 4 < , for all  x R , we have
lim n 1 D n k = 1 n d k I k V a r ( ε 1 2 ) σ ˆ k 2 σ 2 x = Φ ( x ) a . s .
Corollary 1.
Let  c n > 0  with  c n  and  lim n c n + 1 c n = 1  and  k l ( c k c l ) γ , k < l  for some constant  γ > 0 . Denote
d k = log c k + 1 c k exp log β c k , D n = k = 1 n d k , 0 β < 1 / 2 .
Then, under the assumptions of Theorem 1, (6) also holds.
Remark 2.
If the conditions (C1) and (C2) of Theorem 1 is satisfied for some sequence  { D n } , then it is also satisfied for any other sequence  D n * = Ψ ( D n ) , provided that  Ψ : R + R +  is differentiable,  Ψ ( x ) = O ( Ψ ( x ) / x )  and  log Ψ ( x )  is uniformly continuous on  ( B , )  for some  B > 0 . Typical examples are  Ψ ( x ) = x γ Ψ ( x ) = ( log x ) γ Ψ ( x ) = ( log log x ) γ  with some suitable  γ > 0 .
Remark 3.
It is easy to show that  d k = l ( k ) / k , where  l ( x )  is slowly varying at infinity and  D n , satisfies the conditions  ( C 1 )  and  ( C 2 ) . So typical examples including  d k = 1 / k d k = log θ k / k , θ > 1 d k = exp ( log γ k ) / k , 0 γ < 1 / 2 , 1 < ρ < ( 1 γ ) / γ .

3. Preliminary Lemmas

Some useful lemmas which are needed to prove the main result are given in the following section.
Lemma 1
(Hall and Heyde [29], Theorem 2.11, P.23). Let  X 1 = S 1  and  X i = S i S i 1 2 i n  denote the differences of the sequences  { S i , 1 i n } . If  { S i , F i , 1 i n }  is a martingale and  p > 0 , then there exists constant C depending only on p such that
E max 1 i n | S i | p C E i = 1 n E ( X i 2 | F i 1 ) p / 2 + E max 1 i n | X i | p .
Lemma 2.
For  1 i n 1 j l q , then for any  2 t 4 , one can obtain
E | Y i j | t E M 1 t ( X i 1 , , X i p ) E M 1 4 ( X i 1 , , X i p ) t / 4 < , E | Z i j l | t E M 2 t ( X i 1 , , X i p ) E M 2 4 ( X i 1 , , X i p ) t / 4 < .
Proof. 
The proof of Lemma 2 is obvious by Assumption 1 and the Hölder inequality. □
Lemma 3.
Assume that  { G n , n 1 }  is a sequence of random variables satisfying the ASCLT with the weight  { d k }  defined as in Theorem 1, that is
x R , lim n 1 D n k = 1 n d k I { G k x } = Φ ( x ) a . s .
Let  { R n , n 1 }  be a sequence of random variables converging almost surely to zero. Then,  { G n + R n , n 1 }  also satisfies the ASCLT. That is
x R , lim n 1 D n k = 1 n d k I { G k + R k x } = Φ ( x ) a . s .
Proof. 
For fixed  x R  and  η > 0 , recall that  { G n , n 1 }  satisfies the ASCLT, then we have
T n , η : = 1 D n k = 1 n d k I { G k x + η } Φ ( x + η ) 0 , a . s .
and
W n , η : = 1 D n k = 1 n d k I { G k x η } Φ ( x η ) 0 , a . s .
Remark that
{ G n + R n x } { G n x + η } { | R n | > η } ,
{ G n x η } { G n + R n x } { | R n | > η } .
Then, we can conclude that
1 D n k = 1 n d k I { G k + R k x } Φ ( x ) 1 D n k = 1 n d k I { G k x + η } Φ ( x + η ) + 1 D n k = 1 n d k I { | R k | > η } + | Φ ( x + η ) Φ ( x ) | T n , η + 1 D n k = 1 n d k I { | R k | > η } + x x + η 1 2 π e t 2 2 d t T n , η + 1 D n k = 1 n d k I { | R k | > η } + η 2 π ,
and
1 D n k = 1 n d k I { G k + R k x } Φ ( x ) 1 D n k = 1 n d k I { G k x η } Φ ( x η ) 1 D n k = 1 n d k I { | R k | > η } + Φ ( x η ) Φ ( x ) W n , η 1 D n k = 1 n d k I { | R k | > η } η 2 π .
Noting that  { R n , n 1 }  is a sequence of random variables converging almost surely to zero and the arbitrariness of  η , the desired conclusion follows from above discussion. □
Lemma 4
(Zhang [30], Lemma 2.10, P.391). Let  { ζ n , n 1 }  be a sequence of uniformly bounded random variables and  { d n } { D n }  be defined as in Theorem 1. If there exist constants  C > 0  and  δ > 0  and a sequence of positive numbers  { a ( k ) }  such that  n = 1 a ( 2 n ) <  and
E | ζ k ζ j | C ( ( k / j ) δ + a ( k ) ) , for j / k > b n = ( log D n ) ρ / δ ,
then
lim n 1 D n k = 1 n d k ζ k = 0 a . s .
Lemma 5.
Let  { ε i , i 1 }  be a sequence of independent and identically distributed random variables with mean zero, finite variance  σ 2  and  E ε 1 4 < . Let  { d n } { D n }  be defined as in Theorem 1. Then for all  x R ,
lim n 1 D n k = 1 n d k I 1 k V a r ( ε 1 2 ) i = 1 k ε i 2 σ 2 x = Φ ( x ) a . s .
Proof. 
Denote  T k = i = 1 k ε i 2 σ 2 . Suppose that f is a bounded Lipschitz function. By classical central limit theorem, we have
E f T k k V a r ( ε 1 2 ) E f ( N ) as k .
By the conclusions in Section 2 of Peligrad and Shao [31] and Theorem 7.1 of Billingsley [32], we know that (7) is equivalent to
lim n 1 D n k = 1 n d k f T k k V a r ( ε 1 2 ) = E f ( N ) , a . s .
Hence, to prove (7), it suffices to show that
1 D n k = 1 n d k f T k k V a r ( ε 1 2 ) E f T k k V a r ( ε 1 2 ) 0 , a . s . n .
For convenience, let  W k = f ( T k k V a r ( ε 1 2 ) ) E f ( T k k V a r ( ε 1 2 ) ) . Notice that  { ε i , i Z }  are independent, both f and  f  are bounded, then we conclude that for  1 k < j n ,
| E W k W j | = Cov f T k k V a r ( ε 1 2 ) , f T j j V a r ( ε 1 2 ) = Cov f T k k V a r ( ε 1 2 ) , f T j j V a r ( ε 1 2 ) f T j T k j V a r ( ε 1 2 ) C 1 E f T j j V a r ( ε 1 2 ) f T j T k j V a r ( ε 1 2 ) C 2 E | T k | j V a r ( ε 1 2 ) C 3 ( E T k 2 ) 1 / 2 j V a r ( ε 1 2 ) C 4 k V a r ( ε 1 2 ) j V a r ( ε 1 2 ) C 5 ( k j ) 1 / 2 ,
then by Lemma 4 with  δ = 1 / 2  and  a ( k ) 0 , (8) holds, and therefore, the proof of (7) is completed. □
Lemma 6.
Under the assumptions of Theorem 1, for any  1 j q , we have
lim sup n log log n n 3 / 2 i = 1 n Y i j 2 = 0 a . s .
Proof. 
Let  n k = [ k α ] α > 2 . By Lemma 2 and the Markov inequality, for any  ϵ > 0 , it is easy to known that
k = 1 P log log n k + 1 n k 3 / 2 i = 1 n k Y i j 2 > ϵ C 1 k = 1 log log n k + 1 n k 3 / 2 i = 1 n k E Y i j 2 C 2 k = 1 log log n k + 1 n k 3 / 2 · n k C 3 k = 1 log log ( k + 1 ) k α / 2 < .
Then by the Borel–Cantelli lemma, we obtain
log log n k + 1 n k 3 / 2 i = 1 n k Y i j 2 0 a . s . a s k .
Similarly, by Lemma 2 and the Markov inequality, for any  ϵ > 0 , one can obtain
k = 1 P log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Y i j 2 > ϵ C 1 k = 1 log log n k + 1 n k 3 / 2 E max n k < n n k + 1 i = n k + 1 n Y i j 2 C 2 k = 1 log log n k + 1 n k 3 / 2 i = n k + 1 n k + 1 E Y i j 2 C 3 k = 1 log log n k + 1 n k 3 / 2 · [ n k + 1 n k ] C 4 k = 1 log log ( k + 1 ) k 3 α / 2 · ( k + 1 ) α k α C 5 k = 1 log log ( k + 1 ) k α / 2 + 1 < .
By the Borel–Cantelli lemma, it follows that
log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Y i j 2 0 a . s . a s k .
Then combining (9) with (10), for  n k < n n k + 1 , one can obtain
lim sup n log log n n 3 / 2 i = 1 n Y i j 2 lim sup k log log n k + 1 n k 3 / 2 i = 1 n k Y i j 2 + lim sup k log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Y i j 2 0 a . s . a s k .
Thus, the proof of Lemma 6 is completed. □
Lemma 7.
Under the assumptions of Theorem 1, for any  1 j , l q , we have
lim sup n ( log log n ) 2 n 5 / 2 i = 1 n Z i j l 2 = 0 a . s .
Proof. 
By Lemma 2 and the Markov inequality, for any  ϵ > 0 , it is easy to see that
n = 1 P ( log log n ) 2 n 5 / 2 i = 1 n Z i j l 2 > ϵ C 1 n = 1 ( log log n ) 2 n 5 / 2 i = 1 n E Z i j l 2 C 2 n = 1 ( log log n ) 2 n 5 / 2 n C 3 n = 1 ( log log n ) 2 n 3 / 2 < .
By the Borel–Cantelli lemma, one can obtain
( log log n ) 2 n 5 / 2 i = 1 n Z i j l 2 0 a . s . a s n .
The proof of Lemma 7 is completed. □
Lemma 8.
Under the assumptions of Theorem 1, for any  1 j q , we have
lim sup n ( log log n ) 1 / 2 n i = 1 n Y i j ε i = 0 a . s .
Proof. 
Let
Y m = i = 1 m Y i j ε i , 1 m n .
Let  F m  be the  σ -algebra generated by the random variables  ε i , 1 i m . By the fact that  Y i j  and  ε i  are independent, it is easy to compute that the process  Y m , F m , 1 m n  is a martingale. By Lemmas 1 and 2, for some  2 < t < 4 , we know
E | Y n | t = E i = 1 n Y i j ε i t C 1 E i = 1 n E Y i j 2 ε i 2 | F i 1 t / 2 + C 2 E max 1 i n | Y i j ε i | t C 3 E i = 1 n E Y i j 2 t / 2 · [ E ε 1 2 ] t / 2 + C 4 i = 1 n E | Y i j | t · E | ε 1 | t C 5 n t / 2 + C 6 n C 7 n t / 2 .
By the Markov inequality and (11), for any  ϵ > 0 , it is easy to obtain
n = 1 P ( log log n ) 1 / 2 n | i = 1 n Y i j ε i | > ϵ C 1 n = 1 ( log log n ) t / 2 n t E i = 1 n Y i j ε i t C 2 n = 1 ( log log n ) t / 2 n t n t / 2 C 3 n = 1 ( log log n ) t / 2 n t / 2 < .
By the Borel–Cantelli lemma, we can obtain
( log log n ) 1 / 2 n i = 1 n Y i j ε i 0 a . s . a s n .
The proof of Lemma 8 is completed. □
Lemma 9.
Under the assumptions of Theorem 1, for any  1 j , l q , we have
lim sup n log log n n 3 / 2 i = 1 n Z i j l ε i = 0 a . s .
Proof. 
Let  n k = [ k α ] α > 2 . By Lemma 2, the Markov inequality,  C r  inequality and Cauchy–Schwarz inequality, for any  ϵ > 0 , it is easy to see that
k = 1 P log log n k + 1 n k 3 / 2 | i = 1 n k Z i j l ε i | > ϵ C 1 k = 1 ( log log n k + 1 ) 2 n k 3 E i = 1 n k Z i j l ε i 2 C 2 k = 1 ( log log n k + 1 ) 2 n k 3 · n k i = 1 n k E Z i j l 2 ε i 2 C 3 k = 1 ( log log n k + 1 ) 2 n k 2 i = 1 n k ( E Z i j l 4 ) 1 / 2 ( E ε i 4 ) 1 / 2 C 4 k = 1 ( log log n k + 1 ) 2 n k C 5 k = 1 ( log log ( k + 1 ) ) 2 k α < .
Then by the Borel–Cantelli lemma, we obtain
log log n k + 1 n k 3 / 2 i = 1 n k Z i j l ε i 0 a . s . a s k .
Similarly, By Lemma 2 and the Markov inequality and  C r  inequality, for any  ϵ > 0 , one can obtain
k = 1 P log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Z i j l ε i > ϵ C 1 k = 1 ( log log n k + 1 ) 2 n k 3 E max n k < n n k + 1 | i = n k + 1 n Z i j l ε i | 2 C 2 k = 1 ( log log n k + 1 ) 2 n k 3 E i = n k + 1 n k + 1 | Z i j l ε i | 2 C 3 k = 1 ( log log n k + 1 ) 2 n k 3 · ( n k + 1 n k ) i = n k + 1 n k + 1 E Z i j l 2 ε i 2 C 4 k = 1 ( log log n k + 1 ) 2 n k 3 · ( n k + 1 n k ) i = n k + 1 n k + 1 ( E Z i j l 4 ) 1 / 2 ( E ε i 4 ) 1 / 2 C 5 k = 1 ( log log n k + 1 ) 2 n k 3 · ( n k + 1 n k ) 2 C 6 k = 1 ( log log ( k + 1 ) ) 2 k 3 α · k 2 ( α 1 ) C 7 ( log log ( k + 1 ) ) 2 k α + 2 < .
By the Borel–Cantelli lemma, it follows that
log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Z i j l ε i 0 a . s . a s k .
Then combining (12) with (13), for  n k < n n k + 1 , one can obtain
lim sup n log log n n 3 / 2 i = 1 n Z i j l ε i lim sup k log log n k + 1 n k 3 / 2 i = 1 n k Z i j l ε i + lim sup k log log n k + 1 n k 3 / 2 max n k < n n k + 1 i = n k + 1 n Z i j l ε i 0 a . s . a s k .
The proof of Lemma 9 is completed. □
Lemma 10.
Under the assumptions of Theorem 1, for any  1 j , l , k q , we have
lim sup n ( log log n ) 3 / 2 n 2 i = 1 n Y i j Z i l k = 0 a . s .
Proof. 
By Lemma 2, the Markov inequality,  C r  inequality and Cauchy–Schwarz inequality, for any  ϵ > 0 , it is easy to see that
n = 1 P ( log log n ) 3 / 2 n 2 i = 1 n Y i j Z i l k > ϵ C 1 n = 1 ( log log n ) 3 t / 4 n t E i = 1 n Y i j Z i l k t / 2 C 2 n = 1 ( log log n ) 3 t / 4 n t n t / 2 1 i = 1 n E Y i j t / 2 Z i l k t / 2 C 3 n = 1 ( log log n ) 3 t / 4 n t / 2 + 1 i = 1 n E | Y i j | t 1 / 2 E | Z i l k | t 1 / 2 C 4 n = 1 ( log log n ) 3 t / 4 n t / 2 < .
where  2 < t 4  is defined in Assumption 1.
By the Borel–Cantelli lemma, we obtain
( log log n ) 3 / 2 n 2 i = 1 n Y i j Z i l k 0 a . s . a s n .
The proof of Lemma 10 is completed. □

4. Proof

Recall (1) and (3), by Taylor’s expansion expansion with the Lagrange remainder, there exists  λ ( 0 , 1 ) , and  θ * = θ + λ ( θ ˆ θ )
ε ˆ i = ε i [ r θ ˆ ( X i 1 , , X i p ) r θ ( X i 1 , , X i p ) ] = ε i j = 1 q ( θ ˆ j θ j ) Y i j 1 2 j = 1 q l = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) Z i j l .
Then by (14), we can obtain
n V a r ( ε 1 2 ) σ ˆ n 2 σ 2 = n V a r ( ε 1 2 ) 1 n i = 1 n ε ˆ i 2 1 n i = 1 n ε i 2 + 1 n i = 1 n ε i 2 1 n i = 1 n E ε i 2 = 1 n V a r ( ε 1 2 ) i = 1 n ( ε ˆ i 2 ε i 2 ) + 1 n V a r ( ε 1 2 ) i = 1 n ( ε i 2 E ε i 2 ) = 1 n V a r ( ε 1 2 ) i = 1 n j = 1 q ( θ ˆ j θ j ) Y i j 2 + 1 16 n V a r ( ε 1 2 ) i = 1 n j = 1 q l = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) Z i j l 2 4 n V a r ( ε 1 2 ) i = 1 n j = 1 q ( θ ˆ j θ j ) Y i j ε i 1 n V a r ( ε 1 2 ) i = 1 n j = 1 q l = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) Z i j l ε i + 1 n V a r ( ε 1 2 ) i = 1 n j = 1 q l = 1 q k = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) ( θ ˆ k θ k ) Y i j Z i l k + 1 n V a r ( ε 1 2 ) i = 1 n ( ε i 2 E ε i 2 ) = : I n 1 + I n 2 I n 3 I n 4 + I n 5 + I n 6 .
Recall the elementary inequality
i = 1 q a i b i 2 i = 1 q a i 2 i = 1 q b i 2 .
For  I n 1 , by (5) and (16) and Lemma 6, it is easy to know that
I n 1 = 1 n V a r ( ε 1 2 ) i = 1 n j = 1 q ( θ ˆ j θ j ) Y i j 2 1 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) 2 · j = 1 q i = 1 n Y i j 2 = 1 V a r ( ε 1 2 ) n log log n j = 1 q ( θ ˆ j θ j ) 2 · j = 1 q log log n n 3 / 2 i = 1 n Y i j 2 0 a . s . a s n .
For  I n 2 , by (5) and (16) and Lemma 7, one can obtain
I n 2 = 1 16 n V a r ( ε 1 2 ) i = 1 n j = 1 q l = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) Z i j l 2 1 16 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) 2 · l = 1 q ( θ ˆ l θ l ) 2 · j = 1 q l = 1 q i = 1 n Z i j l 2 = 1 4 V a r ( ε 1 2 ) n log log n j = 1 q ( θ ˆ j θ j ) 2 2 · j = 1 q l = 1 q ( log log n ) 2 n 5 / 2 i = 1 n Z i j l 2 0 a . s . a s n .
For  I n 3 , by (5) and (16) and Lemma 8, one can obtain
I n 3 = 4 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) · i = 1 n Y i j ε i 4 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) 2 · j = 1 q i = 1 n Y i j ε i 2 1 / 2 = 2 V a r ( ε 1 2 ) n log log n j = 1 q ( θ ˆ j θ j ) 2 · j = 1 q ( log log n ) 1 / 2 n i = 1 n Y i j ε i 2 1 / 2 0 a . s . a s n .
For  I n 4 , by (5) and (16) and Lemma 9, we have
I n 4 = 1 n V a r ( ε 1 2 ) j = 1 q l = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) i = 1 n Z i j l ε i 1 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) 2 · l = 1 q ( θ ˆ l θ l ) 2 · j = 1 q l = 1 q i = 1 n Z i j l ε i 2 1 / 2 = 1 V a r ( ε 1 2 ) n log log n j = 1 q ( θ ˆ j θ j ) 2 2 · j = 1 q l = 1 q log log n n 3 / 2 i = 1 n Z i j l ε i 2 1 / 2 0 a . s . a s n .
For  I n 5 , by (5) and (16) and Lemma 10, we know
I n 5 = 1 n V a r ( ε 1 2 ) j = 1 q l = 1 q k = 1 q ( θ ˆ j θ j ) ( θ ˆ l θ l ) ( θ ˆ k θ k ) i = 1 n Y i j Z i l k 1 n V a r ( ε 1 2 ) j = 1 q ( θ ˆ j θ j ) 2 · l = 1 q ( θ ˆ l θ l ) 2 · k = 1 q ( θ ˆ k θ k ) 2 · j = 1 q l = 1 q k = 1 q i = 1 n Y i j Z i l k 2 1 / 2 = 1 V a r ( ε 1 2 ) n log log n j = 1 q ( θ ˆ j θ j ) 2 3 · j = 1 q l = 1 q k = 1 q ( log log n ) 3 / 2 n 2 i = 1 n Y i j Z i l k 2 1 / 2 0 a . s . a s n .
Combing (17)–(21), one can obtain
I n 1 + I n 2 I n 3 I n 4 + I n 5 0 a . s . a s n .
For  I n 6 , by Lemma 5, it is obviously that
lim n 1 D n k = 1 n d k I 1 k V a r ( ε 1 2 ) i = 1 k ( ε i 2 E ε i 2 ) x = Φ ( x ) a . s .
Finally, (6) follows by combining (15), (22) with (23) and Lemma 3, thus the proof of Theorem 1 is completed.

5. Examples

Some examples are given in this section to verify the almost sure central limit theorem for the error variance estimator for some special nonlinear autoregressive models. The first example is a degenerate model, that is, AR(1) progresses.
Example 1.
An AR(1) model is a family of  { X i }  of random variables such that for every  i 1
X i = θ X i 1 + ε i ,
where  { ε i , i 1 }  is a collection of i.i.d. random variables with zero mean and finite variance  σ 2 . We also assume that  E exp { γ | ε i ε j | } <  for some  γ > 0  and any  i , j 1 . It is obviously that  { X i }  is a stationary model under the condition  | θ | < 1 .
It is easy to check that the Assumption 1 holds naturally. For Assumption 2, by Theorem 1 of Wang et al. [26] and  E exp { γ | ε i ε j | } < , (5) holds for the least squares estimator  θ ˆ . Therefore, we have the following statement for AR(1) progression due to Theorem 1.
Theorem 2.
Suppose  { d k }  is a sequence of positive numbers satisfying conditions (C1) and (C2). For the above AR(1) model, if  E exp { γ | ε i ε j | } <  for some  γ > 0  and any  i , j 1 , for any  x R , one can obtain
lim n 1 D n k = 1 n d k I k V a r ( ε 1 2 ) σ ˆ k 2 σ 2 x = Φ ( x ) a . s .
The next example concerns the self-exciting threshold autoregressive (SETAR) progresses.
Example 2.
Let  { X i , i p }  be a sequence of stationary and geometrically ergodic random variable satisfying the following continuous SETAR( p , l , d ) progresses.
X i = a 0 + m = 1 p a m X i m + ε i , if X i d R 1 , a 0 + m = 1 p a j X i m + k = 2 j b k ( X i d r k 1 ) + ε i , if X i d R j , j = 2 , , l
where  { ε i }  is a collection of i.i.d. random variables with zero mean and finite variance  σ 2 R 1 , , R l  are the different regions with  R s = ( r s 1 , r s ]  for  1 s l , and  = r 0 < r 1 < r 2 < < r l 1 < r l = +  are the thresholds. Let  θ 0 = ( a 0 , , a p , b 2 , , b l , r 1 , , r l 1 ) Θ R q  be the true parameters of the progresses and  θ = ( a ¯ 0 , , a ¯ p , b ¯ 2 , , b ¯ l , r ¯ 1 , , r ¯ l 1 ) X ˜ i = ( X i , , X i p + 1 ) q = p + 2 l 1 .
Condition  C Suppose that  { ε i }  has the density h and the density f of  X i  is continuous and has a support including the interval  [ r min η , r max + η ] , η > 0  where  r min = min { r ¯ 1 : θ Θ } r max = max { r ¯ l 1 : θ Θ } . There is some  ε > 0  such that  r ¯ k 1 r ¯ k ε  for all  θ Θ  and  k = 2 , , l .
By Corollary 3.1 of Liebscher [2], under Condition  C  and  E | ε i | γ < E | | X ˜ i | | γ < γ > 4 , the Assumption 2 holds. Therefore, we have the following result for SETAR progresses due to Theorem 1.
Theorem 3.
Suppose  { d k }  is a sequence of positive numbers satisfying conditions (C1) and (C2). For the above SETAR( p , l , d ) progresses, under Assumption 1 and Condition  C  and  E | ε i | γ < E | | X ˜ i | | γ < γ > 4 , for any  x R , one can obtain
lim n 1 D n k = 1 n d k I k V a r ( ε 1 2 ) σ ˆ k 2 σ 2 x = Φ ( x ) a . s .
Next, we will consider the threshold-exponential AR progresses.
Example 3.
Let  R j , j = 1 , , K  be non-overlapping and non-empty intervals of  R  such that  j R j = R . A combined threshold-exponential AR progresses is defined by
X i = j = 1 K ( α j + β j X i 1 ) I { X i 1 R j } + c e γ X i 1 2 X i 1 + ε i ,
with  X 0 = x 0  and  { ε i }  is a collection of i.i.d. random variables with zero mean. Let the true parameters  θ 0 = ( α 1 , , α K , β 1 , , β K , c , γ ) Θ R q  and the parameters  θ = ( α ¯ 1 , , α ¯ K , β ¯ 1 , , β ¯ K , c ¯ , γ ¯ )  with  q = 2 K + 2 .
For Assumption 2, if  c 0 γ > 0 | β j | < 1 j = 1 , , K  and  E | ε i | 2 + δ <  for some  δ > 0 , by Theorem 4 of Yao [28], (5) holds for the least squares estimator  θ ˆ . Therefore, we have the following statement for threshold-exponential AR progresses due to Theorem 1.
Theorem 4.
Suppose  { d k }  is a sequence of positive numbers satisfying conditions (C1) and (C2). For the above threshold-exponential AR progresses, if  c 0 γ > 0 | β j | < 1 j = 1 , , K  and  E | ε i | 2 + δ <  for some  δ > 0  and  E ε 1 4 < , then under Assumption 1, for any  x R , one can obtain
lim n 1 D n k = 1 n d k I k V a r ( ε 1 2 ) σ ˆ k 2 σ 2 x = Φ ( x ) a . s .
Next, we will consider the multilayer perceptrons progress.
Example 4.
Multilayer perceptrons progressesw have become popular in nonlinear modeling due to its universal approximation ability. Such an example is the model described below which has p input units feeding by variables  X i 1 , , X i p  at time i, a hidden layer with K units and one output unit which provides the variable  X i
X i = j = 1 K α j ψ l = 1 p β l j X i l + β 0 j + α 0 + ε i ,
where  { ε i }  is a collection of i.i.d. random variables with zero mean. Let the true parameters  θ 0 = ( α 0 , , α K , β l j , 0 l p , 1 j K ) Θ R q  and the parameters  θ = ( α ¯ 0 , , α ¯ K , β ¯ l j , 0 l p , 1 j K )  with  q = 1 + K ( p + 1 ) .
For Assumption 2 and sigmoid map  ψ ( x ) = t a n h ( x ) = e x e x e x + e x , if for all  θ  are different from  θ 0 , there exists  x R p  such that  r θ ( x ) r θ 0 ( x ) E | ε i | 6 + δ <  for some  δ > 0  and the matrix  I 0  is regular, where
I 0 = 2 R p M θ 0 ( x ) μ θ 0 ( d x ) , M θ ( x ) = ( r θ ( x ) θ i · r θ ( x ) θ j ) 1 i , j q ,
then by Theorem 5 of Yao [28], (5) holds for the least squares estimator  θ ˆ . Therefore, we have the following statement for multilayer perceptrons due to Theorem 1.
Theorem 5.
Suppose  { d k }  is a sequence of positive numbers satisfying conditions (C1) and (C2). For the univariate multilayer perceptrons progress with  ψ ( x ) = t a n h ( x ) , if for all θ different from  θ 0 , there exists  x R p  such that  r θ ( x ) r θ 0 ( x ) E | ε i | 6 + δ <  for some  δ > 0  and the matrix  I 0  is regular, then under Assumption 1, for any  x R , one can obtain
lim n 1 D n k = 1 n d k I k V a r ( ε 1 2 ) σ ˆ k 2 σ 2 x = Φ ( x ) a . s .

6. Conclusions

In this paper, using Taylor’s expansion, the Borel–Cantelli lemma and the classical almost sure central limit theorem for independent random variables, the authors establish the almost sure central limit theorem for the error variance estimator for nonlinear autoregressive progresses with independent and identical distributed errors. The results extend the almost sure central limit theorem for the error variance estimator to the nonlinear autoregressive progresses. Four examples, first-order autoregressive processes, self-exciting threshold autoregressive processes, threshold-exponential AR progresses and multilayer perceptrons progress, are given to verify the results. In the future, we will try to investigate the almost sure central limit theorem for the error variance estimator for nonlinear autoregressive progresses with dependent errors and the moderate deviation principle for the error variance estimator for nonlinear autoregressive progresses with independent errors.

Author Contributions

Conceptualization, K.L.; formal analysis, K.L.; methodology, Y.Z.; validation, K.L. and Y.Z.; visualization, K.L. and Y.Z.; writing—original draft preparation, K.L.; writing—review and editing, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 12171198, 11771178); the Science and Technology Development Program of Jilin Province (Grant No. 20210101467JC); and the Science and Technology Program of Jilin Educational Department during the “14th Five-Year” Plan Period (Grant No. JJKH20241239KJ).

Data Availability Statement

Data sharing is not applicable to this article as no datasets were generated or analyzed during the current (theoretical) study.

Acknowledgments

We would like to thank the anonymous referees for a very careful reading of the article.

Conflicts of Interest

The authors declare no conflicts of interest in this paper.

References

  1. Tong, H. Non-Linear Time Series: A Dynamical Approach; Oxford University Press: New York, NY, USA, 1990. [Google Scholar]
  2. Liebscher, E. Strong convergence of estimators in nonlinear autoregressive models. J. Multivar. Anal. 2003, 84, 247–261. [Google Scholar] [CrossRef]
  3. Cheng, F.; Sun, S. A goodness-of-fit test of the errors in nonlinear autoregressive time series models. Stat. Probab. Lett. 2008, 78, 50–59. [Google Scholar] [CrossRef]
  4. Fu, K.; Yang, X. Asymptotics of kernel density estimators in nonlinear autoregressive models. J. Math. Chem. 2008, 44, 831–838. [Google Scholar] [CrossRef]
  5. Cheng, F. Global property of error density estimation in nonlinear autoregressive time series models. Stat. Inference Stoch. Process. 2010, 13, 43–53. [Google Scholar] [CrossRef]
  6. Li, J. Asymptotics of the Lp-norms of density estimators in the nonlinear autoregressive models. Commun. Stat. Theory Methods 2014, 43, 4845–4855. [Google Scholar] [CrossRef]
  7. Kim, K.; Sin, M.; Kim, O. A goodness-of-fit test of the errors in nonlinear autoregressive time series models with stationary α-mixing error terms. ROMAI J. 2014, 10, 63–70. [Google Scholar]
  8. Cheng, F. Strong consistency of the distribution estimator in the nonlinear autoregressive time series. J. Multivar. Anal. 2015, 142, 41–47. [Google Scholar] [CrossRef]
  9. Liu, T.; Zhang, Y. Law of the iterated logarithm for error density estimators in nonlinear autoregressive models. Commun. Stat. Theory Methods 2020, 49, 1082–1098. [Google Scholar] [CrossRef]
  10. Cheng, F. Variance estimation in nonlinear autoregressive time series models. J. Stat. Plann. Inference 2011, 141, 1588–1592. [Google Scholar] [CrossRef]
  11. Brosamler, G. An almost everywhere central limit theorem. Math. Proc. Camb. Philos. Soc. 1988, 104, 561–574. [Google Scholar] [CrossRef]
  12. Schatte, P. On strong versions of the central limit theorem. Math. Nachrichten 1988, 137, 249–256. [Google Scholar] [CrossRef]
  13. Peligrad, M.; Reévész, P. On the almost sure central limit theorem. In Almost Everywhere Convergence; Academic Press: Boston, MA, USA, 1989; Volume II, pp. 209–225. [Google Scholar]
  14. Berkes, I.; Csáki, E. A universal result in almost sure central limit theory. Stoch. Process. Appl. 2001, 94, 105–134. [Google Scholar] [CrossRef]
  15. Hörmann, S. An extension of almost surecentral limit theory. Stat. Probab. Lett. 2006, 76, 191–202. [Google Scholar]
  16. Tong, B.; Peng, Z.; Saralees, N. An extension of almost surecentral limit theorem for order statistics. Extremes 2009, 12, 201–209. [Google Scholar]
  17. Miao, Y. An extension of almost surecentral limit theory for the product of partial sums. J. Dyn. Syst. Geom. Theor. 2009, 7, 49–60. [Google Scholar]
  18. Li, Y. An extension of the almost sure central limit theorem for products of sums under association. Commun. Stat. Theory Methods 2013, 42, 478–490. [Google Scholar] [CrossRef]
  19. Zhang, Y. A universal result in almost sure central limit theorem for products of sums of partial sums under mixing sequence. Stochastics 2016, 88, 803–812. [Google Scholar] [CrossRef]
  20. Zhang, Y. An extension of almost sure central limit theorem for self-normalized products of sums for mixing sequences. Commun. Stat. Theory Methods 2016, 45, 6625–6640. [Google Scholar] [CrossRef]
  21. Wu, Q.; Jiang, Y. Almost sure central limit theorem for self-normalized partial sums and maxima. Rev. R. Acad. Cienc. Exactas Fís. Nat. Ser. A Mat. RACSAM 2016, 110, 699–710. [Google Scholar] [CrossRef]
  22. Li, J.Y.; Zhang, Y. An almost sure central limit theorem for the stochastic heat equation. Stat. Probab. Lett. 2021, 177, 109149. [Google Scholar] [CrossRef]
  23. Li, J.; Zhang, Y. An almost sure central limit theorem for the parabolic Anderson model with delta initial condition. Stochastics 2023, 95, 483–500. [Google Scholar] [CrossRef]
  24. Li, J.; Zhang, Y. Almost sure central limit theorems for stochastic wave equations. Electron. Commun. Probab. 2023, 28, 9. [Google Scholar] [CrossRef]
  25. Klimko, L.; Nelson, P. On conditional least squares estimation for stochastic processes. Ann. Stat. 1978, 6, 629–642. [Google Scholar] [CrossRef]
  26. Wang, Y.; Mao, M.; Hu, X.; He, T. The law of iterated logarithm for autoregressive processes. Math. Probl. Eng. 2014, 2014, 972712. [Google Scholar] [CrossRef]
  27. Chan, K.; Tong, H. On estimating thresholds in autoregressive models. J. Time Ser. Anal. 1986, 7, 179–190. [Google Scholar] [CrossRef]
  28. Yao, J. On least squares estimation for stable nonlinear AR processes. Ann. Inst. Stat. Math. 2000, 52, 316–331. [Google Scholar] [CrossRef]
  29. Hall, P.; Heyde, C. Martingale Limit Theory and its Application; Academic Press: New York, NY, USA, 1980. [Google Scholar]
  30. Zhang, Y. Further research on limit theorems for self-normalized sums. Commun. Stat. Theory Methods 2020, 49, 385–402. [Google Scholar] [CrossRef]
  31. Peligrad, M.; Shao, Q. A note on the almost sure central limit theorem. Stat. Probab. Lett. 1995, 22, 131–136. [Google Scholar] [CrossRef]
  32. Billingsley, P. Convergence of Probability Measures; Wiley: New York, NY, USA, 1968. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, K.; Zhang, Y. Almost Sure Central Limit Theorem for Error Variance Estimator in Pth-Order Nonlinear Autoregressive Processes. Mathematics 2024, 12, 1482. https://doi.org/10.3390/math12101482

AMA Style

Liang K, Zhang Y. Almost Sure Central Limit Theorem for Error Variance Estimator in Pth-Order Nonlinear Autoregressive Processes. Mathematics. 2024; 12(10):1482. https://doi.org/10.3390/math12101482

Chicago/Turabian Style

Liang, Kaiyu, and Yong Zhang. 2024. "Almost Sure Central Limit Theorem for Error Variance Estimator in Pth-Order Nonlinear Autoregressive Processes" Mathematics 12, no. 10: 1482. https://doi.org/10.3390/math12101482

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop