Next Article in Journal
Sharp Unified Smoothness Theory for Cavalieri Estimation via Fourier Decay
Previous Article in Journal
Algorithms for Solving the Resolvent of the Sum of Two Maximal Monotone Operators with a Finite Family of Nonexpansive Operators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Law of the Iterated Logarithm for the Error Distribution Estimator in First-Order Autoregressive Models

1
School of Big Data and Statistics, Anhui University, Hefei 230601, China
2
Irving K. Barber Faculty of Science, University of British Columbia, Kelowna, BC V1V 1V7, Canada
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(11), 784; https://doi.org/10.3390/axioms14110784
Submission received: 29 August 2025 / Revised: 22 October 2025 / Accepted: 23 October 2025 / Published: 26 October 2025

Abstract

This paper investigates the asymptotic behavior of kernel-based estimators for the error distribution in a first-order autoregressive model with dependent errors. The model assumes that the error terms form an α -mixing sequence with an unknown cumulative distribution function (CDF) and finite second moment. Due to the unobservability of true errors, we construct kernel-smoothed estimators based on residuals obtained via least squares. Under mild assumptions on the kernel function, bandwidth selection, and mixing coefficients, we establish a logarithmic law of the iterated logarithm (LIL) for the supremum norm difference between the residual-based kernel estimator and the true distribution function. The limiting bound is shown to be 1 / 2 , matching the classical LIL for independent samples. To support the theoretical results, simulation studies are conducted to compare the empirical and kernel distribution estimators under various sample sizes and error term distributions. The kernel estimators demonstrate smoother convergence behavior and improved finite-sample performance. These results contribute to the theoretical foundation for nonparametric inference in autoregressive models with dependent errors and highlight the advantages of kernel smoothing in distribution function estimation under dependence.

1. Introduction

Suppose that the sequence { X i } satisfies the first-order autoregressive process
X i = ρ X i 1 + ϵ i ,
where ϵ i , i = 0 , ± 1 , ± 2 , , are random errors with mean 0, variance σ 0 2 > 0 , and they form a weakly stationary sequence of α -mixing with an unknown cumulative distribution function (CDF) F. Henceforth, we use the term stationary to mean weakly stationary. The definition of α -mixing will be provided later. It is well-known that the parameter ρ characterizes the properties of the process X i . In this paper, we suppose | ρ | < 1 and sup i 1 V a r ( X i ) < . Hence, X i is a so-called stationary process (see [1]). Throughout this paper, we consider the stationary solution of the first-order autoregressive process, which can be represented as
X i = j = 0 ρ j ϵ i j .
If ϵ 1 , ϵ 2 , , ϵ n , were observed, then the estimation of their CDF, F, can be obtained by the empirical CDF as follows:
F n ( t ) = 1 n i = 1 n I ( ϵ i t ) , t R .
For any t, F n ( t ) is the uniformly minimum variance unbiased estimator of F ( t ) for all continuous distribution functions, F. It is also well known as the Glivenko–Cantelli theorem that
sup t R | F n ( t ) F ( t ) | 0 , a . s . ,
where “a.s.” stands for “almost surely” (see Gut [2]). The law of the iterated logarithm (LIL) for F n ( t ) , i.e.,
lim sup n n 2 log ( log n ) sup t R | F n ( t ) F ( t ) | = 1 2 , a . s .
was obtained by Smirnov [3] and, independently, Chung [4]. Cai and Roussas [5] extended LIL of F n ( t ) to the case that ϵ 1 , ϵ 2 , , ϵ n were α -mixing. Since the empirical distribution function is a step function and thus discontinuous, smooth estimators are often preferred to provide continuous and differentiable approximations that are more suitable for statistical inference and further analysis. Therefore, Yamato [6] proposed the following kernel estimator:
F n h ( t ) = 1 n i = 1 n t K h ( u ϵ i ) d u , t R ,
where K h ( u ) = K ( u / h ) / h , h is the bandwidth and K is a kernel probability density function (PDF). Cai and Roussas [5] also studied the uniform strong rate of convergence for sup t R | F n h ( t ) F ( t ) | = O ( ( log ( log n ) / n ) 1 / 2 ) ; a.s., Cheng [7] obtained the uniform strong rate of convergence for sup t R n α | F n h ( t ) F ( t ) | 0 , a.s., for any 0 α < 1 / 2 .
But in autoregressive model (1), only the random variables X 1 , X 2 , , X n are observed. Therefore, in order to obtain an estimator of F ( t ) , we modify the definition F n h ( t ) in (4) by plugging in the residuals
ϵ ^ i = X i ρ ^ n X i 1 , 1 i n ,
where ρ ^ n = ρ ^ n ( X 1 , , X n ) is an estimator of ρ . For example, the least squares estimator of ρ is
ρ ^ n = i = 1 n X i X i 1 i = 1 n X i 1 2 .
Therefore, the residual empirical CDF is given by
F ˜ n ( t ) = 1 n i = 1 n I ( ϵ ^ i t ) , t R ,
and the residual kernel estimation of F is
F ^ n h ( t ) = 1 n i = 1 n t K h ( u ε ^ i ) d u , t R .
Under the i.i.d. errors, Cheng [8] considered the first-order autoregressive model (1) and verified that the Glivenko–Cantelli theorem still holds for the estimators F ^ n ( t ) and F ^ n h ( t ) , i.e.,
sup t R | F ^ n ( t ) F ( t ) | 0 , sup t R | F ^ n h ( t ) F ( t ) | 0 , a . s .
Beyond the classical LIL under the supremum norm, recent research has extended the LIL framework to other normed settings. In particular, Cheng [9] considered the first-order autoregressive model (1) and obtained the integrated absolute error (i.e., L 1 -norm) of F ^ n h ,
lim sup n n 2 log ( log n ) | F ^ n h ( t ) F ( t ) | d F ( t ) = 3 6 , a . s .
Motivated by these findings, our work focuses on the LIL for the residual kernel estimator F ^ n h ( t ) in the first-order autoregressive model (1) with α -mixing errors, i.e.,
lim sup n n 2 log ( log n ) sup t R | F ^ n h ( t ) F ( t ) | = 1 2 , a . s .
Let us recall the definition of an α -mixing sequence.
Definition 1.
Denote N = { 1 , 2 , , n , } . Let F m n = σ ( ϵ i , m i n , i N ) be the σ-field generated by random variables { ϵ m , ϵ m + 1 , , ϵ n } , 1 m n . For n 1 , we define
α ( n ) = sup m N sup A F 1 m , B F m + n | P ( A B ) P ( A ) P ( B ) | .
If α ( n ) 0 as n , { ϵ n , n 1 } is called a strong mixing or α-mixing sequence.
The α -mixing condition is a type of weak dependence commonly used in time series analysis. It measures how quickly the dependence between past and future observations decays as the time gap increases. For more properties of α -mixing sequences, one can refer to Györfi [10], Roussas [11], Fan and Yao [12], Wang et al. [13], etc.
For more research on LIL, one can refer to Gajek et al. [14] and Cheng [15] for L p -norms of the empirical distribution function F n ( t ) and kernel distribution function F n h ( t ) ; Li and Wang [16] and Petrov [17] investigated the LIL for the sequences of dependent random variables; Liu and Zhang [18] studied the LIL for error density estimators in nonlinear autoregressive models, etc.
In the context of estimating the error distribution function, this paper makes three main contributions. First, we consider a linear first-order autoregressive model in which the autoregressive coefficient ρ must be estimated rather than assumed known. Second, we study kernel-based nonparametric estimators for the error distribution, providing smoother approximations than the empirical distribution function and enabling refined probabilistic analysis. Third, we explicitly allow for dependent errors modeled by a stationary α -mixing sequence, extending classical results beyond the independent error setting.
It is worth noting that for the i.i.d. case, the LIL for the empirical distribution function case has been well established by Smirnov [3] and Chung [4]. For kernel estimators of the error distribution, related results were obtained by Niu [19] and Cheng [20]. Our work extends these classical results to the dependent error setting, motivated in part by earlier studies on autoregressive processes such as that by Wang et al. [21], bridging the gap between the i.i.d. theory and time series models with weakly dependent innovations.
We also remark that alternative weak dependence structures could be considered. For example, if the errors are independent, the convergence rates simplify, and some technical conditions can be relaxed. If α -mixing is replaced by other forms of weak dependence, such as β -mixing, ρ -mixing, or martingale difference sequences or the weak dependence framework of Doukhan and Louhichi [22], we expect similar LIL results to hold under appropriate moment and dependence assumptions, though the technical details and proof techniques would require adaptation. A full exploration of these generalizations is beyond the scope of the present paper, but they constitute interesting directions for future research.
The rest of this paper is organized as follows: in Section 2, we list some basic assumptions for the first-order autoregressive model (1). The main results of a uniform strong rate of sup t R | F ^ n h ( t ) F n h ( t ) | and LIL of (9) are presented in Section 3. In order to check our results, some simulations are presented in Section 4. Some technical Lemmas and proofs are presented in Section 5. Last, we provide a conclusion and discussion in Section 6. Throughout the paper, we assume that limits are taken as n unless otherwise specified. The C , C 1 , and C 2 denote some positive constants not depending on n, which may be different in various places.

2. Assumptions

This section presents the basic assumptions required for model (1) as well as the main theorems.
(A1)
In model (1), let the sequence { ϵ i , < i < } of errors form a stationary sequence of α -mixing with unknown CDF F with a bounded second-order derivative, i.e., there exists a positive constant C ( 0 < C < ) such that | F ( t ) | C for all t R . Assume E ϵ 1 = 0 and E | ϵ 1 | p < for some p > 2 . The mixing coefficient satisfies α ( n ) = O ( n r ) , where r > 5 p + 2 p 2 .
(A2)
The kernel function K ( t ) , t K ( t ) , and t 2 K ( t ) are integrable over the real line and satisfy
0 inf t R K ( t ) sup t R K ( t ) < , + K ( t ) d t = 1 , + t K ( t ) d t = 0 , + t 2 K ( t ) d t < .
(A3)
The bandwidth h satisfies the following conditions:
h 0 , n h 4 / log ( log n ) 0 , n h 2 / ( log n ) 2 .
(A4)
In model (1) with | ρ | < 1 , let the ρ ^ n be an estimator of ρ with the following almost-surely (a.s.) property: there exists a positive constant C such that
| ρ ^ n ρ | C n 1 / 2 ( log ( log n ) ) 1 / 2 , a . s .
Remark 1.
We discuss these assumptions as follows:
(i) 
Assumption (A1) contains the smoothness of the error CDF F and the boundness of its second derivative F . It also requires the mixing coefficients satisfy α ( n ) = O ( n r ) , where r > 5 p + 2 p 2 and p > 2 (p-th moment of error). If p 2 , then r . It is a strong condition of mixing coefficients, which require the errors to be asymptotically independent. If p , then r 5 . This requires a strong p-th moment of errors. In future research, we will try to relax the mixing coefficient condition. For more properties of α-mixing sequences, one can refer to Györfi [10], Roussas [11], Fan and Yao [12], Wang et al. [13], etc.
(ii) 
Assumption (A2) is the common condition of kernel K ( · ) about non-negatively, symmetry, and integrability. Obviously, the probability density functions Gaussian kernel K ( t ) = ( 2 π ) 1 2 exp ( t 2 2 ) and Epanechnikov kernel K ( t ) = 3 20 5 ( 5 t 2 ) I ( t 5 ) conform to Assumption (A2). See for example, Györfi et al. [10], Roussas [11], Fan and Yao [12], Li and Racine [23], etc.
(iii) 
Assumption (A3) is a standard condition for bandwidth selection in kernel density estimation. For example, similar conditions are used by Cheng et al. [7,9,15], particularly in the proof of Lemma 5. A commonly used bandwidth is h = n 1 / 4 .
(iv) 
Assumption (A4) requires that the estimator ρ ^ of the autoregressive parameter ρ converges almost surely at the rate O ( n 1 / 2 ( log ( log n ) ) 1 / 2 ) . This condition plays a key role in our theoretical results, as it ensures that the effect of estimating ρ is asymptotically negligible. Without such a rate, the additional error introduced by residuals could dominate the behavior of the kernel estimator and invalidate the uniform LIL established in this paper. When ϵ 1 , ϵ 2 , , ϵ n are i.i.d. errors, Koul and Zhu [24] considered a generalized M-estimator for p-th order autoregression models and obtained a strong rate of convergence for M-estimators ρ ^ n , i.e., | | ρ ^ n ρ | | C n 1 / 2 ( log ( log n ) ) 1 / 2 , a.s., including the least-squares estimator; Cheng [9] used Assumption (A4) to study the L 1 -norm of F ^ n h in (8). When ϵ 1 , ϵ 2 , , ϵ n in the first-order autoregression model (1) are α-mixing random variables with common density function f and | ρ | < 1 , Gao et al. [25] used the condition | ρ ^ n ρ | = O p ( n 1 / 2 ) to study the asymptotic normality of the kernel density estimator f ^ n ( t ) = 1 n h i = 1 n K h ( u ε ^ i ) for the error density function f. Wu et al. [26] extended the work to nonlinear autoregressive models. In the proof by Gao et al. [25], moment inequalities were used to establish the convergence in probability. However, to extend these results to almost-sure convergence, it would be necessary to employ exponential inequalities. This in turn requires additional conditions on the α-mixing coefficients, ensuring that the covariance structure of the α-mixing sequence satisfies certain summability and decay properties. The proof of such a result is highly technical and involves delicate handling of the dependence structure. Because this is beyond the scope of the present paper, we treat Assumption (A4) as a standing assumption rather than providing a full proof here, and we leave this as an important topic for our future research.

3. Main Results

We now present the first main result of this section, which describes the uniform strong rate of convergence between F ^ n h ( t ) and F n h ( t ) in the stationary autoregressive model (1).
Theorem 1.
In the model (1), let Assumptions (A1)–(A4) hold. Then
n log ( log n ) sup t R | F ^ n h ( t ) F n h ( t ) | = o ( 1 ) , a . s .
where F ^ n h ( t ) is defined in (7) and F n h ( t ) is defined by (6).
By Theorem 1, one can get the uniform strong rate
sup t R | F ^ n h ( t ) F n h ( t ) | = o ( ( log ( log n ) / n ) 1 / 2 ) , a . s .
Combining the LIL of F n h ( t ) in Lemma 6, i.e.,
lim sup n n 2 log ( log n ) sup t R | F n h ( t ) F ( t ) | = 1 2 , a . s . ,
we present the LIL for the residual kernel estimator F ^ n h ( t ) as follows:
Theorem 2.
Under the same conditions as in Theorem 1, we have
lim sup n n 2 log ( log n ) sup t R | F ^ n h ( t ) F ( t ) | = 1 2 , a . s .
Remark 2.
In Theorem 3.2 of Cai and Roussas’s study [5], Cai and Roussas obtained the LIL with α-mixing samples such that
lim sup n n 2 log ( log n ) sup t R | F n ( t ) F ( t ) | = 1 , a . s .
They had a typo error in estimation (17) that was equal to 1 a.s. In fact, they mistakenly cited [27] [Corollary 1.15.1] and claimed that
lim sup n [ 1 n n 2 log ( log n ) sup t R | K ( t , n ) | ] = 1 , a . s .
But it is
lim sup n [ 1 n n 2 log ( log n ) sup t R | K ( t , n ) | ] = 1 2 , a . s . ,
(see the proof of Lemma 4). Thus, Theorem 2 extends the conclusion of LIL to the kernel estimator F ^ n h ( t ) of error CDF F ( t ) in autoregressive model (1) based on α-mixing errors.

4. Simulations

To verify the theoretical results, we conduct simulation studies to assess the uniform convergence behavior of four distribution estimators under different error distribution settings, namely Gaussian and Gamma. The goal is to compare the accuracy of these estimators in approximating the true error distribution function, particularly in terms of the iterated logarithm convergence rate.
We consider the first-order autoregressive model:
X i = 0.5 X i 1 + ϵ i , 1 i n ,
where { ϵ i } is a dependent error sequence following either a Gaussian distribution or a Gamma distribution.
  • Gaussian errors: ϵ N ( 0 , Σ n ) , where Σ n is a Toeplitz covariance matrix with entries | i j | 3 , ensuring that the errors form an α -mixing sequence. A Toeplitz matrix is a matrix in which each descending diagonal from left to right is constant. For more details about the Toeplitz matrix, we can refer to Trench [28]. As discussed in [29], it is easy to check that ϵ 1 , ϵ 2 , are α -mixing with α ( n ) = O ( n 3 ) .
  • Gamma errors: we adopt a Gaussian copula-based approach to generate random vectors with exact Gamma marginals and an approximately specified covariance matrix. Specifically, we first generate a multivariate normal vector Z N ( 0 , R ) , where R is the correlation matrix corresponding to the target covariance matrix Σ n . Each component of Z is then transformed to a uniform random variable via the standard normal CDF and is subsequently mapped to a Gamma(4,1) random variable using the Gamma inverse CDF. This method guarantees the correct marginal Gamma distribution, while the resulting sample covariance matrix is very close to Σ n , with only small differences. An exact covariance match can be achieved using the NORTA (Normal To Anything) method, which iteratively adjusts the Gaussian correlation matrix to perfectly reproduce the target covariance after the nonlinear transformation. However, NORTA is computationally intensive, so we do not adopt it here, given that the Gaussian copula-based approach already provides sufficient accuracy and efficiency.
For each sample size n { 300 , 600 , 2000 , 3000 , 4000 , 5000 } , we perform 100 Monte Carlo simulations. In each replication, the following four estimators of the error distribution function are constructed:
  • F n ( t ) : empirical distribution function based on the true errors;
  • F n h ( t ) : kernel-smoothed distribution estimator based on true errors;
  • F ˜ n ( t ) : empirical distribution based on residuals (from least squares estimation);
  • F ^ n h ( t ) : kernel-smoothed distribution based on residuals (from least-squares estimation).
To measure the estimation accuracy, we compute the Kolmogorov–Smirnov (KS) distance between each estimator and the true distribution function and normalize it using the theoretical upper bound under the law of the iterated logarithm:
N K S = n 2 log ( log n ) KS distance .
Define the NKS for the estimators,
N K S 1 = n 2 log ( log n ) sup t R | F n ( t ) F ( t ) | , N K S 2 = n 2 log ( log n ) sup t R | F n h ( t ) F ( t ) | , N K S 3 = n 2 log ( log n ) sup t R | F ˜ n ( t ) F ( t ) | , N K S 4 = n 2 log ( log n ) sup t R | F ^ n h ( t ) F ( t ) | .
A Gaussian kernel is used for smoothing, with bandwidth h = n 1 / 3 . The estimators are evaluated over the range t [ 3 , 3 ] .
To investigate the effect of the bandwidth h on the performance of the proposed kernel estimator, we conducted a sensitivity analysis for two representative sample sizes, n = 600 and n = 3000 . For n = 600 , the bandwidth h was taken from the set { 0.01 , 0.05 , 0.10 , 0.15 , 0.20 , 0.202 , 0.25 , 0.30 , 0.35 , 0.40 } , where h = 0.202 corresponds exactly to the suggested choice n 1 / 4 with n = 600 . For n = 3000 , the bandwidth h was chosen from the set { 0.01 , 0.05 , 0.10 , 0.135 , 0.15 , 0.20 , 0.25 , 0.30 , 0.35 , 0.40 } , where h = 0.13 corresponds to the suggested choice n 1 / 4 with n = 3000 .
Figure 1 and Figure 2 present the corresponding boxplots of the statistic N K S 4 for these different bandwidths under the Gaussian error setting, with Figure 1 corresponding to the sample size n = 600 and Figure 2 corresponding to the sample size n = 3000 . For n = 600 , although there is mild fluctuation across different bandwidths, the boxplot corresponding to h = 0.202 = n 1 / 4 shows that its median is closer to zero and the interquartile range is under 0.5, indicating that the estimator performs more stably and with less bias at this recommended bandwidth. For n = 3000 , the variability of N K S 4 across bandwidths is smaller. Again, the boxplot for h = 0.135 = n 1 / 4 is closer to zero, further supporting the theoretical bandwidth selection. These findings confirm that N K S 4 is not sensitive to bandwidth variation, and they provide strong empirical evidence that the bandwidth h = n 1 / 4 offers a good kernel estimator for CDF F. Therefore, all subsequent simulation results reported in this paper are based on the bandwidth h = n 1 / 4 to ensure consistency and comparability across experiments. Figure 3, Figure 4, Figure 5 and Figure 6 display boxplots of N K S across different sample sizes and estimation methods. Each sample size includes 100 simulation replicates per method. The vertical axis represents the NKS, and a horizontal dashed line at level 0.5 is included to indicate the asymptotic upper bound suggested by the iterated logarithm law.
Figure 3 presents the NKS distances for the empirical estimator F n ( t ) and the kernel estimator F n h ( t ) , both based on true Gaussian errors. Each boxplot summarizes the distribution of NKS across 100 replications. We observe that the upper edge of the boxes (the 75th percentile) for both estimators stays well below the 0.5 theoretical boundary, even at the smallest sample size n = 300 . As the sample size increases, the entire box shrinks downward, indicating uniform improvement in estimator accuracy. The kernel estimator F n h ( t ) consistently shows a tighter spread and a lower upper quartile than F n ( t ) , highlighting the advantage of kernel smoothing in reducing estimation variance under Gaussian error structures.
Figure 4 shows the NKS distances for the residual-based estimators F ˜ n ( t ) and F ^ n h ( t ) , derived from least-squares residuals under the same Gaussian error process. The results are similar to those in Figure 3. The upper quartile of the NKS remains consistently below 0.5 for both methods across all sample sizes. Again, the kernel estimator F ^ n h ( t ) shows superior concentration, with its upper edge often well below the 0.5 threshold, even at n = 300 . This suggests that residual-based estimators, particularly when smoothed, maintain excellent uniform convergence properties under Gaussian dependence, nearly matching the performance of error-based methods.
Figure 5 and Figure 6 report the results under Gamma-distributed and dependent errors. Although the Gamma distribution is skewed and heavy-tailed, the overall performance of the estimators remains similar to that observed under Gaussian errors. All estimators exhibit convergence as the sample size increases, and kernel-based methods continue to outperform empirical ones in terms of reduced dispersion and tighter concentration around the theoretical upper bound.
The simulation results confirm the validity of the iterated logarithm law for both error-based and residual-based estimators under dependent data. Among all the methods, the kernel-based estimators consistently demonstrate superior stability and accuracy across different error distributions and sample sizes.

5. Proofs of the Main Results

Lemma 1
(Hall and Heyde [30], Corollary A.2). Suppose that ξ and η are random variables which are F 1 t and F t + τ -measurable, respectively, t , τ N , and that E | η | p < , E | ξ | q < , where p , q > 1 and 1 p + 1 q < 1 . Then,
| E ξ η E ξ E η | 8 ( E | ξ | p ) 1 / p ( E | η | q ) 1 / q ( α ( G , H ) ) 1 1 p 1 q .
Lemma 2
(Liebscher [31], Proposition 5.1). Let { Z n } n 1 be a stationary α-mixing sequence with mixing coefficient α ( n ) . Assume that E Z i = 0 and | Z i | S < , a . s . , i = 1 , 2 , , n . Then, for n , h N , 0 < h n / 2 and every ε > 0 ,
P | i = 1 n Z i | > ε 4 exp ε 2 16 ( n h 1 D h + 1 3 ε S h ) 1 + 32 S ε n α ( h ) ,
where D h = max 1 j 2 h Var ( i = 1 j Z i ) .
Lemma 3.
In the first-order autoregression (1), let Assumption (A1) hold. Then for any permutation { j 1 , j 2 , , j n } of { 1 , 2 , , n } , we have
max 1 m < n | k = 1 m X j k | = O ( n 1 / 2 log n ) , a . s .
Proof of Lemma 3.
Since | ρ | < 1 , by (2), the stationary process { X i } i = 1 n in the first-order autoregression (1) has a representation of X i = j = 0 ρ j ϵ i j , n 1 . Let I i ( m ) = 1 if i { j 1 , , j m } and I i ( m ) = 0 otherwise. Then,
max 1 m n | i = 1 n X j i | max 1 m n | i = 1 n I i ( m ) j = 0 n ρ j ϵ i j | + max 1 m n | i = 1 n I i ( m ) j > n ρ j ϵ i j | .
By p > 2 , E ϵ 1 = 0 , E | ϵ 1 | p < , it has V a r ( ϵ 1 ) = σ ϵ 2 < . According to the Chebyshev inequality and Cauchy inequality, we have
n = 1 P max 1 m n | i = 1 n I i ( m ) j > n ρ j ϵ i j | n 1 / 2 log n n = 1 1 n ( log n ) 2 m = 1 n E i = 1 n I i ( m ) j > n ρ j ϵ i j 2 n = 1 n 2 σ ϵ 2 n ( log n ) 2 j > n ρ 2 j + 2 k = 1 n 1 j > n , j + k > n | ρ 2 j + k | n = 1 σ ϵ 2 n ( log n ) 2 n Σ j > n | ρ | j 2 < ,
by using the fact | ρ | < 1 . So, it follows from the Borel–Cantelli Lemma that
max 1 m n | i = 1 n I i ( m ) j > n ρ j ϵ i j | = O ( n 1 / 2 log n ) , a . s .
For p > 2 , set
p n l = ϵ l I ( | ϵ l | ( 2 n ) 1 / p ) , q n l = ϵ l p n l = ϵ l I ( | ϵ l | > ( 2 n ) 1 / p ) .
Then,
max 1 m n i = 1 n I i ( m ) j = 0 n ρ j ϵ i j = max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ϵ l max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l p n l + max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l q n l .
Observe that
max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l q n l j = 0 | ρ | j l = 1 n n | ϵ l | I ( | ϵ l | > ( 2 n ) 1 / p ) C i = 1 2 n | ϵ i n | I ( | ϵ i n | > ( 2 n ) 1 / p ) .
In the following, we show that i = 1 2 n | ϵ i n | I ( | ϵ i n | > ( 2 n ) 1 / p ) converges to a positive and finite random variable as n . Let { a n , n 1 } be a positive constant sequence with n = 1 a n < . Let { ξ n , n 1 } be a sequence of random variables satisfying
n = 1 P ( | ξ n + 1 ξ n | > a n ) < .
Then ξ n ξ a.s. n , where ξ is a positive and finite random variable. For more details, see Wang [32] [Proposition 2.4.2]. Note that E | ϵ 1 | p < if and only if i = 1 P ( | ϵ 1 | > i 1 / p ) < . Let ξ n = i = 1 2 n ϵ i n I | ϵ i n | > ( 2 n ) 1 / p and a n = 1 / n 2 , which satisfies n = 1 a n < . Then we have
n = 1 P | ξ n + 1 ξ n | > a n = n = 1 P i = 1 2 n + 2 ϵ i n 1 I | ϵ i n 1 | > ( 2 n + 2 ) 1 / p i = 1 2 n ϵ i n I | ϵ i n | > ( 2 n ) 1 / p > a n n = 1 P i = 1 2 n + 2 ϵ i n 1 I | ϵ i n 1 | > ( 2 n ) 1 / p i = 1 2 n ϵ i n I | ϵ i n | > ( 2 n ) 1 / p > a n = n = 1 P ϵ n I | ϵ n | > ( 2 n ) 1 / p + ϵ n + 1 I | ϵ n + 1 | > ( 2 n ) 1 / p > a n n = 1 P ϵ n I | ϵ n | > ( 2 n ) 1 / p + ϵ n + 1 I | ϵ n + 1 | > ( 2 n ) 1 / p > a n n = 1 P ϵ n I | ϵ n | > ( 2 n ) 1 / p > a n 2 + P ϵ n + 1 I | ϵ n + 1 | > ( 2 n ) 1 / p > a n 2 = n = 1 P ϵ n > a n 2 , | ϵ n | > ( 2 n ) 1 / p + P ϵ n + 1 > a n 2 , | ϵ n + 1 | > ( 2 n ) 1 / p n = 1 P | ϵ n | > ( 2 n ) 1 / p + P | ϵ n + 1 | > ( 2 n ) 1 / p 2 n = 1 P | ϵ 1 | > n 1 / p < .
Then there exists a positive and finite random variable ξ such that ξ n = i = 1 2 n | ϵ i n | I ( | ϵ i n | > ( 2 n ) 1 / p ) ξ , a.s., n . Consequently, we can obtain
max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l q n l = O ( n 1 / 2 log n ) , a . s .
In addition, as E ϵ 1 = 0 and E | ϵ 1 | p < , we have
max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l E p n l C j = 0 | ρ | j l = 1 n n E | ϵ l | I ( | ϵ l | > ( 2 n ) 1 / p ) = O ( n 1 / 2 log n ) .
To prove (10), we will show that
max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) = O ( n 1 / 2 log n ) , a . s .
Note that
max 1 n l n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) j = 0 | ρ | j 2 n 1 / p C 1 n 1 / p , a . s . ,
and by Assumption (A1) and Lemma 1, we have
V a r l = 1 k k i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) 1 + i = 1 α ( i ) p 2 p ( E | ϵ 1 | p ) 2 / p j = 0 | ρ | j 2 2 k C 2 k , for k n .
It implies that
D h = max 1 k h V a r l = 1 k k i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) C 3 h ,
where 0 < h n . Since p > 2 and r > 5 p + 2 p 2 according to Assumption (A1), it has 3 / 2 + 1 / p ( 1 / 2 1 / p ) r < 1 . Then, by α ( n ) = O ( n r ) , we take h = n 1 / 2 1 / p in Lemma 2 and have that
n = 1 P max 1 m n l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) C n 1 / 2 log n n = 1 m = 1 n P l = 1 n n i = max ( 1 , l ) min ( n , n + l ) I i ( m ) ρ i l ( p n l E p n l ) C n 1 / 2 log n n = 1 4 n exp C 2 n ( log n ) 2 16 n h 1 D h + 1 3 C n 1 / 2 log n C 1 n 1 / p h 1 + 32 n C 1 n 1 / p C n 1 / 2 log n n α ( h ) n = 1 4 n exp C 2 n ( log n ) 2 16 C 4 n + C 5 n log n 1 + C 6 n 2 + 1 / p n 1 / 2 log n n ( 1 / 2 1 / p ) r n = 1 4 n exp C 7 log n + C 6 n 3 / 2 + 1 / p ( 1 / 2 1 / p ) r log 1 n < ,
for some positive C 7 > 2 . So, by the Borel–Cantelli Lemma, the proof of (11) is completed.
Remark 3.
Gao [33] and Sun et al. [34] considered a moving-average process, denoted by M A ( ) , i.e.,
X n = l = 0 ϕ l ε n l ,
with l = 0 | ϕ l | < and i.i.d. { ε n } errors; they obtained that for any permutation ( l 1 , , l n ) of ( 1 , , n ) ,
max 1 k n | i = 1 k X l i | = O ( n 1 / 2 log n ) , a . s .
Lemma 4.6 by Liang et al. [35] extended the result (13) of the linear process (12) with the i.i.d. case to the negatively associated case. In Lemma 3, we extend the result (13) to the stationary process of the first-order autoregression model (1) with α-mixing errors.
Before proceeding, we recall the definition of the Kiefer process [36].
Definition 2.
A separable Gaussian process K ( s , t ) on [ 0 , 1 ] × [ 0 , ) is called a Kiefer process if it satisfies
K ( 0 , t ) = K ( 1 , t ) = K ( s , 0 ) = 0 , 0 s 1 , t 0 ,
with mean zero
E [ K ( s , t ) ] = 0 ,
and covariance function
E { K ( s , t ) K ( s , t ) } = min ( t , t ) Γ ( s , s ) ,
where Γ ( s , s ) is the covariance function of a separable Gaussian process G ( s ) on [ 0 , 1 ] , with G ( 0 ) = G ( 1 ) = 0 .
The Kiefer process generalizes the Brownian bridge to two parameters and naturally arises in the asymptotic study of empirical and residual processes, including LIL.
Lemma 4.
Let ϵ 1 , ϵ 2 , . . . , ϵ n be an α-mixing sample with PDF F. Let Assumption (A1) hold. Then the LIL for the empirical distribution function F n ( t ) is
lim sup n n 2 log ( log n ) sup t R | F n ( t ) F ( t ) | = 1 2 , a . s . ,
where F n ( t ) is defined in (3).
Proof of Lemma 4.
Similarly to Cai and Roussas [5], define the empirical process:
R ( t , s ) = k s g k ( t ) , t R , s 0 ,
where g k ( t ) = I ( ϵ k t ) F ( t ) . Then the covariance function is
Γ ( t , t ) = E [ g 1 ( t ) g 1 ( t ) ] + k = 2 E [ g 1 ( t ) g k ( t ) ] + k = 2 E [ g k ( t ) g 1 ( t ) ] , t , t R
On an enriched probability space, by Theorem 3 by Dhompongsa [36], we can redefine the process to admit a Kiefer process { K ( t , s ) , t R , s 0 } with covariance
E [ K ( t , s ) K ( t , s ) ] = Γ ( t , t ) min ( s , s )
such that for some λ > 0 , depending only on r,
sup 0 s T sup t R | R ( t , s ) K ( t , s ) | = O ( T 1 2 ( log T ) λ ) , a . s . , T > 0 .
On the one hand, it follows from (14) and (15) that
lim sup n n 2 log ( log n ) sup t R | F n ( t ) F ( t ) | = lim sup n 1 n n 2 log ( log n ) sup t R | K ( t , n ) | ,
(see Cai and Roussas [5]). On the other hand, one uses Corollary 1.15.1 in [27] and obtains
lim sup n 1 n n 2 log ( log n ) sup t R | K ( t , n ) | = 1 2 , a . s .
Combining (16) with (17), we complete the proof of Lemma 4. □
Lemma 5.
In model (1), let Assumptions (A1)–(A3) be satisfied. Then
sup t R n log ( log n ) | F n h ( t ) F n ( t ) | 0 , a . s .
Proof of Lemma 5.
Cheng [15] used the Kiefer process to obtain (18) under the i.i.d sample assumption. By the Kiefer process, Cai and Roussas [5] obtained the uniform approximation of the empirical process F n ( t ) to F ( t ) under the α -mixing samples (see Lemma 4). It can be seen that the proof of (18) in Cheng [15] also holds for the case of α -mixing. □
Lemma 6.
In model (1), let Assumptions (A1)–(A3) be satisfied. Then
lim sup n n 2 log ( log n ) sup t R | F n h ( t ) F ( t ) | = 1 2 , a . s .
Proof of Lemma 6.
By the triangle inequality,
| F n h ( t ) F ( t ) | | F n h ( t ) F n ( t ) | + | F n ( t ) F ( t ) | ,
| F n h ( t ) F ( t ) | | F n h ( t ) F n ( t ) | + | F n ( t ) F ( t ) | .
Thus,
sup t R | F n h ( t ) F ( t ) | sup t R | F n h ( t ) F n ( t ) | + sup t R | F n ( t ) F ( t ) | ,
sup t R | F n h ( t ) F ( t ) | sup t R | F n h ( t ) F n ( t ) | + sup t R | F n ( t ) F ( t ) | .
Combining (20) and (21) with Lemmas 4 and 5, we obtain the result of (19). □
Proof of Theorem 1.
From (1) and (5), it follows ϵ ^ i ϵ i = ( ρ ^ n ρ ) X i 1 , 1 i n and X 0 = 0 . Substituting this into the definitions of F n h ( t ) and F ^ n h ( t ) , we use the mean-value theorem and obtain
| F ^ n h ( t ) F n h ( t ) | = 1 n i = 1 n t K h ( t ϵ ^ i ) d t t K h ( t ϵ i ) d t = 1 n i = 1 n ( ρ ^ n ρ ) X i 1 h K ( η i t ) | ρ ^ n ρ | n h i = 1 n X i 1 K ( η i t ) ,
where K ( · ) is defined in Assumption (A2) and η i t is a point between t ϵ ^ i h and t ϵ i h . We now consider the term | i = 1 n X i 1 K ( η i t ) | in (22). Let Y i t = K ( η i t ) , 1 i n . By Assumption (A2), Y 1 t , , Y n t are non-negative and bounded. So, we sort Y 1 t , , Y n t by Y ( 1 t ) Y ( n t ) and denote the corresponding parts of X 0 , , X n 1 as X 0 * , , X n 1 , * . Obviously, one can use Abel’s inequality (see [37]) and establish that
| i = 1 n X i 1 K ( η i t ) | = | i = 1 n Y i t X i 1 | = | i = 1 n Y ( i t ) X i 1 , * | max 1 k n | Y ( k t ) | max 1 k n | i = 1 k X i 1 , * | sup t R K ( t ) | max 1 k n | i = 1 k X i 1 , * | .
Thus, it follows from Assumption (A4), Lemma 3, and (23) that
sup t R | F ^ n h ( t ) F n h ( t ) | | ρ ^ n ρ | n h | sup t R | i = 1 n X i 1 K ( η i t ) | = O ( n 1 / 2 ( log ( log n ) ) 1 / 2 ) n h O ( n 1 / 2 log n ) = O ( log n ) 2 n h 2 log ( log n ) n = o log ( log n ) n ,
proving n h 2 / ( log n ) 2 . Consequently,
n log ( log n ) sup t R | F ^ n h ( t ) F n h ( t ) | = o ( 1 ) ,
i.e., (1) holds. □
Proof of Theorem 2.
Similarly to the proof of Lemma 6, by the triangle inequality,
| F ^ n h ( t ) F ( t ) | | F ^ n h ( t ) F n h ( t ) | + | F n h ( t ) F ( t ) | ,
| F ^ n h ( t ) F ( t ) | | F ^ n h ( t ) F n h ( t ) | + | F n h ( t ) F ( t ) | ,
sup t R | F ^ n h ( t ) F ( t ) | sup t R | F ^ n h ( t ) F n h ( t ) | + sup t R | F n h ( t ) F ( t ) | ,
sup t R | F ^ n h ( t ) F ( t ) | sup t R | F ^ n h ( t ) F n h ( t ) | + sup t R | F n h ( t ) F ( t ) | .
Combining Lemma 6 with Theorem 1, we complete the proof of Theorem 2. □

6. Conclusions

In this paper, we investigated the asymptotic behavior of kernel estimators for the error distribution in first-order autoregressive models (1) when the error sequence is a stationary α -mixing process with an unknown distribution. Due to the unobservable nature of the true errors, we proposed a residual kernel estimator constructed from residuals and examined its convergence properties under a set of mild regularity conditions.
Our main theoretical contribution lies in establishing the LIL for the residual kernel distribution estimator. Specifically, we proved that under appropriate mixing, moment, and bandwidth conditions, the supremum norm of the deviation between the residual kernel estimator and the true error distribution function satisfies
lim sup n n 2 log ( log n ) sup t R | F ^ n h ( t ) F ( t ) | = 1 2 , a . s .
which parallels the classical LIL for i.i.d. sequences.
Moreover, we derived an intermediate result demonstrating that the difference between the residual kernel estimator and its infeasible counterpart based on true errors converges to zero at the same logarithmic rate. This ensures that the influence of estimation error in residuals is asymptotically negligible under our conditions. These theoretical findings were further corroborated by extensive simulation studies under both Gaussian and Gamma distributed α -mixing errors, which consistently showed that kernel estimators exhibit smoother behavior and smaller fluctuations compared to their empirical counterparts, especially in moderate-to-large sample sizes.
In addition to the simulation study, our proposed method has strong potential for practical applications in real data analysis. For instance, in economics and finance, AR(1) models are widely used to model interest rates, exchange rates, and stock returns, where the error terms often exhibit temporal dependence and heavy-tailed behavior. The residual-based kernel estimator provides a flexible tool for accurately estimating the error distribution in such contexts, which can improve model diagnostics, forecasting, and risk management. Future work will include applying this methodology to real-world datasets, such as macroeconomic indicators or climate time series, to demonstrate its practical utility.
The proposed approach has several strengths. It extends classical nonparametric distribution estimation to dependent data settings while maintaining desirable asymptotic properties. The kernel smoothing step improves finite-sample stability and produces smoother estimates that are more suitable for inference and visualization. However, the method also has limitations. For example, the theoretical results rely on the assumption of α -mixing errors with specific decay rates, which may not hold for certain strongly dependent processes. Additionally, while the study focuses on AR(1) models, many practical systems are better described by higher-order AR(p), MA(q), and ARMA( p , q ) processes.
An important open question is how to extend the present results to general AR(p), MA(q), and ARMA( p , q ) processes. The main difficulty lies in handling the increased complexity of parameter estimation and residual dependence when multiple lagged terms are present. Developing a unified theoretical framework for these models would significantly broaden the applicability of the method and provide deeper insights into the asymptotic behavior of residual-based estimators. Other promising directions include exploring data-driven bandwidth selection procedures, incorporating robust kernel functions to handle outliers, and applying the methodology to multivariate or nonlinear time series.
In summary, the residual kernel estimator studied in this paper provides a theoretically sound and practically useful tool for error distribution estimation in AR(1) models with dependent errors. With further research and development, it has the potential to be extended to more complex autoregressive structures and to offer valuable insights for both theoretical and applied settings.

Author Contributions

Supervision W.Y.; software B.W. and Y.J.; writing—original draft preparation, L.W., X.S. and W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (12301327) and the Anhui Province University Research Project (2023AH050096).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Brockwell, P.J.; Davis, R.A. Time Series: Theory and Methods, 2nd ed.; Springer: New York, NY, USA, 1991. [Google Scholar]
  2. Gut, A. Probability: A Graduate Course, 2nd ed.; Springer: New York, NY, USA, 2013. [Google Scholar]
  3. Smirnov, N.V. An approximation to distribution laws of random quantities determined by empirical data. Uspekhi Mat. Nauk. 1944, 10, 179–206. [Google Scholar]
  4. Chung, K.L. An estimate concerning the Kolmogorov limit distribution. Am. Math. Soc. 1949, 67, 36–50. [Google Scholar]
  5. Cai, Z.W.; Roussas, G.G. Uniform strong estimation under α-mixing, with rates. Stat. Probab. Lett. 1992, 15, 47–55. [Google Scholar] [CrossRef]
  6. Yamato, H. Uniform convergence of an estimator of a distribution function. Bull. Math. Statist. 1973, 15, 69–78. [Google Scholar] [CrossRef] [PubMed]
  7. Cheng, F. Strong uniform consistency rates of kernel estimators of cumulative distribution functions. Commun. Stat. Theory Methods 2017, 46, 6803–6807. [Google Scholar] [CrossRef]
  8. Cheng, F. Glivenko-Cantelli Theorem for the kernel error distribution estimator in the first-order autoregressive mode. Stat. Probab. Lett. 2018, 139, 95–102. [Google Scholar] [CrossRef]
  9. Cheng, F. The integrated absolute error of the kernel error distribution estimator in the first-order autoregression model. Stat. Probab. Lett. 2024, 214, 110215. [Google Scholar] [CrossRef]
  10. Györfi, L.; Härdle, W.; Vieu, P. Nonparametric Curve Estimation from Time Series; Springer: New York, NY, USA, 1989. [Google Scholar]
  11. Roussas, G.G. Nonparametric regression estimation under mixing conditions. Stoch. Process. Their Appl. 1990, 36, 107–116. [Google Scholar] [CrossRef]
  12. Fan, J.Q.; Yao, Q.W. Nonlinear Time Series: Nonparametric and Parametric Methods; Springer: New York, NY, USA, 2003. [Google Scholar]
  13. Wang, J.Y.; Liu, R.; Cheng, F.X.; Yang, L.J. Oracally efficient estimation of autoregressive error distribution with simultaneous confidence band. Ann. Statist. 2014, 42, 654–668. [Google Scholar] [CrossRef]
  14. Gajek, L.; Kahszka, M.; Lenic, A. The law of the iterated logarithm for Lp-norms of empirical processes. Stat. Probab. Lett. 1996, 28, 107–110. [Google Scholar] [CrossRef]
  15. Cheng, F. The law of the iterated logarithm for Lp-norms of kernel estimators of cumulative distribution functions. Mathematics 2024, 12, 1063. [Google Scholar] [CrossRef]
  16. Li, Y.X.; Wang, J.F. The law of the iterated logarithm for positively dependent random variables. J. Math. Anal. Appl. 2008, 339, 259–265. [Google Scholar] [CrossRef][Green Version]
  17. Petrov, V.V. On the law of the iterated logarithm for sequences of dependent random variables. Vestnik St. Petersb. Univ. Math. 2017, 50, 32–34. [Google Scholar] [CrossRef][Green Version]
  18. Liu, T.Z.; Zhang, Y. Law of the iterated logarithm for error density estimators in nonlinear autoregressive models. Commun. Stat. Theory Methods 2019, 49, 1082–1098. [Google Scholar] [CrossRef]
  19. Niu, S.L. LIL for kernel estimator of error distribution in regression model. J. Korean Math. Soc. 2007, 44, 1082–1098. [Google Scholar] [CrossRef]
  20. Cheng, F. A law of the iterated logarithm for error density estimator in censored linear regression. J. Nonparametr. Stat. 2022, 34, 283–298. [Google Scholar] [CrossRef]
  21. Wang, Y.; Mao, M.; Hu, X.; He, T. The Law of Iterated Logarithm for Autoregressive Processes. Math. Probl. Eng. 2014, 2014, 972712. [Google Scholar] [CrossRef]
  22. Doukhan, P.; Louhichi, S. A new weak dependence condition and applications to moment inequalities. Stochastic Processes Appl. 1999, 84, 312–342. [Google Scholar] [CrossRef]
  23. Li, Q.; Racine, J.S. Nonparametric Econometrics: Theory and Practice; Princeton University Press: Princeton, NJ, USA, 2007. [Google Scholar]
  24. Koul, H.L.; Zhu, Z.W. Bahadur-Kiefer representations for M-estimators in autoregression models. Stochastic Process. Appl. 1995, 57, 167–189. [Google Scholar] [CrossRef]
  25. Gao, M.; Yang, W.Z.; Wu, S.P.; Yu, W. Asymptotic normality of residual density estimator in stationary and explosive autoregressive models. Comput. Stat. Data An. 2022, 175, 107549. [Google Scholar] [CrossRef]
  26. Wu, S.P.; Yang, W.Z.; Gao, M.; Fang, H.Y. Asymptotic results of error density estimator in nonlinear autoregressive models. J. Korean Stat. Soc. 2024, 53, 563–582. [Google Scholar] [CrossRef]
  27. Csörgo, M.; Révész, P. Strong Approximation in Probability and Statistics; Academic Press: New York, NY, USA, 1981. [Google Scholar]
  28. Gray, R.M. On the asymptotic eigenvalue distribution of Toeplitz matrices. IEEE Trans. Inf. Theory 1972, 18, 725–730. [Google Scholar] [CrossRef]
  29. Withers, C.S. Conditions for linear processes to be strong-mixing. Z. Wahrscheinlichkeitstheorie Verw. Gebiete 1981, 57, 477–480. [Google Scholar] [CrossRef]
  30. Hall, P.; Heyde, C.C. Martingale Limit Theory and Its Application; Academic Press: New York, NY, USA, 1980. [Google Scholar]
  31. Liebscher, E. Estimation of the density and the regression function under mixing conditions. Stat. Decis. 2001, 19, 9–126. [Google Scholar] [CrossRef]
  32. Wang, J.G. Fundamentals of Modern Probability Theory; Fudan University Press: Shanghai, China, 2005. [Google Scholar]
  33. Gao, J.T. Asymptotic theory for partly linear models. Commun. Stat. Theory Methods 1995, 24, 1985–2009. [Google Scholar] [CrossRef]
  34. Sun, X.Q.; You, J.H.; Chen, G.M.; Zhou, X. Convergence rates of estimators in partial linear regression models with MA() error process. Commun. Stat. Theory Methods 2002, 31, 2251–2273. [Google Scholar] [CrossRef]
  35. Liang, H.Y.; Mammitzsch, V.; Steinebach, J. On a semiparametric regression model whose errors form a linear process with negatively associated innovations. Statistics 2006, 40, 207–226. [Google Scholar] [CrossRef]
  36. Dhompongsa, S. A note on the almost sure approximation of the empirical process of weakly dependent random vectors. Yokohama Math. J. 1984, 32, 113–121. [Google Scholar]
  37. Mitrinovic, D.S. Analytic Inequalities; Springer: New York, NY, USA, 1970. [Google Scholar]
Figure 1. Boxplot of N K S 4 under different bandwidth choices for sample size n = 600 in the Gaussian error setting.
Figure 1. Boxplot of N K S 4 under different bandwidth choices for sample size n = 600 in the Gaussian error setting.
Axioms 14 00784 g001
Figure 2. Boxplot of N K S 4 under different bandwidth choices for sample size n = 3000 in the Gaussian error setting.
Figure 2. Boxplot of N K S 4 under different bandwidth choices for sample size n = 3000 in the Gaussian error setting.
Axioms 14 00784 g002
Figure 3. Boxplot of NKS distance for error-based estimators under Gaussian errors. N K S 1 and N K S 2 are the NKS distances for F n ( t ) and F n h ( t ) , respectively.
Figure 3. Boxplot of NKS distance for error-based estimators under Gaussian errors. N K S 1 and N K S 2 are the NKS distances for F n ( t ) and F n h ( t ) , respectively.
Axioms 14 00784 g003
Figure 4. Boxplot of NKS distance for residual-based estimators under Gaussian errors. N K S 3 and N K S 4 are the NKS distances for F ˜ n ( t ) and F ^ n h ( t ) , respectively.
Figure 4. Boxplot of NKS distance for residual-based estimators under Gaussian errors. N K S 3 and N K S 4 are the NKS distances for F ˜ n ( t ) and F ^ n h ( t ) , respectively.
Axioms 14 00784 g004
Figure 5. Boxplot of NKS distance for error-based estimators under Gamma errors. N K S 1 and N K S 2 are the NKS distances for F n ( t ) and F n h ( t ) , respectively.
Figure 5. Boxplot of NKS distance for error-based estimators under Gamma errors. N K S 1 and N K S 2 are the NKS distances for F n ( t ) and F n h ( t ) , respectively.
Axioms 14 00784 g005
Figure 6. Boxplot of NKS distance for residual-based estimators under Gamma errors. N K S 3 and N K S 4 are the NKS distances for F ˜ n ( t ) and F ^ n h ( t ) , respectively.
Figure 6. Boxplot of NKS distance for residual-based estimators under Gamma errors. N K S 3 and N K S 4 are the NKS distances for F ˜ n ( t ) and F ^ n h ( t ) , respectively.
Axioms 14 00784 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, B.; Jin, Y.; Wang, L.; Shi, X.; Yang, W. The Law of the Iterated Logarithm for the Error Distribution Estimator in First-Order Autoregressive Models. Axioms 2025, 14, 784. https://doi.org/10.3390/axioms14110784

AMA Style

Wang B, Jin Y, Wang L, Shi X, Yang W. The Law of the Iterated Logarithm for the Error Distribution Estimator in First-Order Autoregressive Models. Axioms. 2025; 14(11):784. https://doi.org/10.3390/axioms14110784

Chicago/Turabian Style

Wang, Bing, Yi Jin, Lina Wang, Xiaoping Shi, and Wenzhi Yang. 2025. "The Law of the Iterated Logarithm for the Error Distribution Estimator in First-Order Autoregressive Models" Axioms 14, no. 11: 784. https://doi.org/10.3390/axioms14110784

APA Style

Wang, B., Jin, Y., Wang, L., Shi, X., & Yang, W. (2025). The Law of the Iterated Logarithm for the Error Distribution Estimator in First-Order Autoregressive Models. Axioms, 14(11), 784. https://doi.org/10.3390/axioms14110784

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop