Next Article in Journal
Online Classification of High Gamma Dose Rate Incidents
Previous Article in Journal
Deep Representation Learning for Cluster-Level Time Series Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Expectation-Maximization Algorithm for Autoregressive Models with Cauchy Innovations †

Department of Mathematics, Indian Institute of Technology Ropar, Rupnagar 140001, India
*
Author to whom correspondence should be addressed.
Presented at the 8th International Conference on Time Series and Forecasting, Gran Canaria, Spain, 27–30 June 2022.
Eng. Proc. 2022, 18(1), 21; https://doi.org/10.3390/engproc2022018021
Published: 21 June 2022
(This article belongs to the Proceedings of The 8th International Conference on Time Series and Forecasting)

Abstract

:
In this paper, we study the autoregressive (AR) models with Cauchy distributed innovations. In the AR models, the response variable y t depends on previous terms and a stochastic term (the innovation). In the classical version, the AR models are based on normal distribution which could not capture the extreme values or asymmetric behavior of data. In this work, we consider the AR model with Cauchy innovations, which is a heavy-tailed distribution. We derive closed forms for the estimates of parameters of the considered model using the expectation-maximization (EM) algorithm. The efficacy of the estimation procedure is shown on the simulated data. The comparison of the proposed EM algorithm is shown with the maximum likelihood (ML) estimation method. Moreover, we also discuss the joint characteristic function of the AR(1) model with Cauchy innovations, which can also be used to estimate the parameters of the model using empirical characteristic function.

1. Introduction

Autoregressive (AR) models with stable and heavy-tailed innovations are of great interest in time series modeling. These distributions can easily assimilate the asymmetry, skewness, and outliers present in time series data. The Cauchy distribution is a special case of stable distribution with undefined expected value, variance, and higher order moments. The Cauchy distribution and its mixture has many applications in the field of economics [1], seismology [2], biology [3], and various other fields, but only a few studies have been conducted concerning time series models with Cauchy errors. In [4], the maximum likelihood (ML) estimation of AR(1) model with Cauchy errors is studied.
The standard estimation techniques for the AR(p) model with Cauchy innovations, particularly the Yule–Walker method and conditional least squares method, cannot be used due to the infinite second order moments of the Cauchy distribution. Therefore, it is worthwhile to study and assess the alternate estimation techniques for the AR(p) model with Cauchy innovations. In the literature, several estimation techniques have been proposed to estimate the parameters of AR models with infinite variance errors (see e.g., [5,6,7]).
In this paper, we propose to use the EM algorithm to estimate the parameters of the distribution and model simultaneously. It is a general iterative algorithm for model parameter estimation which iterates between two steps, namely the expectation step (E-step) and the maximization step (M-step) [8]. It is an alternative to the numerical optimization of the likelihood function which is proven to be numerically stable [9]. We also provide the formula based on the characteristic function (CF) and empirical characteristic function (ECF) of Cauchy distribution for AR(p) model estimation. The idea to use ECF in a time series stable ARMA model has been discussed in [10].
The remainder of the paper is organized as follows. In Section 2, we present a brief overview of the Cauchy AR(p) model, followed by a discussion of estimation techniques, namely the EM algorithm and estimation by CF and ECF. Section 3 checks the efficacy of the estimation procedure on simulated data. We also present the comparative study, where the proposed technique is compared with ML estimation for Cauchy innovations. Section 4 concludes the paper.

2. Cauchy Autoregressive Model

We consider the AR(p) univariate stationary time-series { Y t } , t N with Cauchy innovations defined as
Y t = i = 1 p ρ i Y t i + ε t = ρ T Y t 1 + ε t ,
where ρ = ( ρ 1 , ρ 2 , , ρ p ) T is a p-dimensional column vector, Y t 1 = ( Y t 1 , Y t 2 , , Y t p ) T is a vector of p lag terms, and { ϵ t } , t N are i.i.d. innovations distributed as Cauchy ( α , γ ) . The pdf of Cauchy ( α , γ ) [11] is
f ( x ; α , γ ) = γ π 1 γ 2 + ( x α ) 2 , γ > 0 , α R , x R .
The conditional distribution of Y t given the preceding data F t 1 = ( Y t 1 , Y t 2 , , Y 1 ) T is given by [4]
p ( Y t | F t 1 ) = f ( y t ρ T y t 1 ; α , γ ) = γ π 1 γ 2 + ( y t ρ T y t 1 α ) 2 ,
where y t 1 is the realization of Y t 1 . In the next subsection, we propose the methods to estimate the model parameters ρ and innovation parameters α and γ simultaneously.

2.1. Parameter Estimation Using EM Algorithm

We estimate the parameters of the AR(p) model using an EM algorithm which maximizes the likelihood function iteratively. Further, we discuss the time series { Y t } using the characteristic function (CF) and estimation method using CF and ECF. Recently [7], the exponential-squared estimator for AR models with heavy-tailed errors was introduced and proven to be n -consistent under some regularity conditions; similarly, the self-weighted least absolute deviation estimation method was also studied for the infinite variance AR model [12]. The ML estimation of AR models with Cauchy errors with intercept and with linear trend is studied, and the AR coefficient is shown to be n 3 / 2 -consistent under some conditions [4]. For the AR(p) model with Cauchy innovations with n samples, the log likelihood is defined as
l ( Θ ) = ( n p ) log ( γ ) ( n p ) log ( π ) t = p + 1 n log ( γ 2 + ( y t ρ T Y t 1 α ) 2 ) ,
where Θ = α , γ , ρ T .
Proposition 1.
Consider the AR(p) time-series model given in Equation (1) where error terms follow Cauchy ( α , γ ) . The maximum likelihood estimates of the model parameters using EM algorithm are as follows
ρ ^ T = t = p + 1 n ( y t α ) Y t 1 T s t ( k ) t = p + 1 n Y t 1 Y t 1 T s t ( k ) 1 ; α ^ = t = p + 1 n ( y t ρ T Y t 1 ) s t ( k ) t = p + 1 n 1 s t ( k ) ; a n d γ ^ = n p 2 t = p + 1 n 1 s t ( k ) ,
where s t ( k ) = ( y t ρ T ( k ) Y t 1 α ( k ) ) 2 + γ ( k ) 2 .
Proof. 
Consider the AR(p) model
Y t = ρ T Y t 1 + ε t , t = p + 1 , 2 , , n ,
where ε t follows Cauchy distribution Cauchy ( α , γ ) . Let ( ε t , V t ) for t = 1 , 2 , , n denote the complete data for innovations ε . The observed data ε t are assumed to be from Cauchy ( α , γ ) and the unobserved data V t follow inverse gamma IG ( 1 / 2 , γ 2 / 2 ) . A random variable V∼IG ( a , b ) if the pdf is given by
f V ( v ; a , b ) = b a Γ ( a ) e b / v v a + 1 , a > 0 , b > 0 , v > 0 .
We can rewrite ε t as ε t = Y t ρ T Y t 1 , for t = 1 , 2 , , n . The stochastic relation ε = α + V Z with Z N ( 0 , 1 ) i.e., standard normal and V IG ( 1 / 2 , γ 2 / 2 ) is used to generate Cauchy ( α , γ ) distribution. Then, the conditional distribution is
f ( ε = ϵ t | V = v t ) = 1 2 π v t exp 1 2 v t ( ϵ t α ) ) 2 .
Now, we need to estimate the unknown parameters Θ = ( α , γ , ρ T ) . To apply the EM algorithm for estimation, we first find the conditional expectation of log-likelihood of complete data ( ε , V ) with respect to the conditional distribution of V given ε . As the unobserved data are assumed to be from IG ( 1 / 2 , γ 2 / 2 ) , the posterior distribution is again an inverse gamma i.e.,
V | ε = e , Θ IG 1 , ( e α ) 2 + γ 2 2 .
The following conditional inverse first moment and E ( log ( V ) | ε = e ) will be used in calculating the conditional expectation of the log-likelihood function:
E ( V 1 | ε = e ) = 2 ( e α ) 2 + γ 2 , E ( log ( V ) | ε = e ) = log ( ( e α ) 2 + γ 2 ) log 2 + 0.5776 .
The complete data likelihood is given by
L ( Θ ) = t = p + 1 n f ( ϵ t , v t ) = t = p + 1 n f ε | V ( ϵ t | v t ) f V ( v t ) = t = p + 1 n γ 2 π v t 2 exp ( ϵ t α ) 2 + γ 2 2 v t .
The log likelihood function will be
l ( Θ ) = ( n p ) log ( γ ) ( n p ) log ( 2 π ) 2 t = p + 1 n log ( v t ) t = p + 1 n ( ϵ t α ) 2 + γ 2 2 v t .
Now, we will use the relation ϵ t = Y t ρ T Y t 1 in further calculations. In the first step at kth iteration, E-step of EM algorithm, we need to compute the expected value of the complete data log likelihood known as Q ( Θ | Θ ( k ) ) , which is expressed as
Q ( Θ | Θ ( k ) ) = E V | ε , Θ ( k ) [ log f ( ε , V | Θ ) | ϵ t , Θ ( k ) ] = E V | ε , Θ ( k ) [ l ( Θ | Θ ( k ) ) ] = ( n p ) log γ ( n p ) log 2 π t = p + 1 n E ( log v t | ϵ t , Θ ( k ) ) γ 2 2 t = p + 1 n E ( v t 1 | ϵ t , Θ ( k ) ) 1 2 t = p + 1 n ( ϵ t α ) 2 E ( v t 1 | ϵ t , Θ ( k ) ) = ( n p ) log γ ( n p ) log 2 π ( n p ) log 2 + 0.5776 ( n p ) t = p + 1 n log ( ( ϵ t α ( k ) ) 2 + γ ( k ) 2 ) γ 2 2 t = p + 1 n 1 ( ϵ t α ( k ) ) 2 + γ ( k ) 2 t = p + 1 n ( ϵ t α ) 2 ( ϵ t α ( k ) ) 2 + γ ( k ) 2 = ( n p ) log γ ( n p ) log 2 π ( n p ) log 2 + 0.5776 ( n p ) t = p + 1 n log ( s t ( k ) ) γ 2 2 t = p + 1 n 1 s t ( k ) t = p + 1 n ( ϵ t α ) 2 s t ( k ) .
where s t = ( y t ρ T Y t 1 α ) 2 + γ 2 . In the next M-step, we estimate the parameters α , γ , and ρ T by maximizing the Q function using the equations below:
Q ρ = 4 t = p + 1 n ( y t ρ T Y t 1 α Y t 1 T ) s t ( k ) , Q α = t = p + 1 n ( y t ρ T Y t 1 α ) s t ( k ) , Q γ = n p γ 2 γ t = p + 1 n 1 s t ( k ) .
Solving the above equations at each iteration, we find the following closed form estimates of the parameters at k + 1 th iteration:
ρ ^ T = t = p + 1 n ( y t α ) Y t 1 T s t ( k ) t = p + 1 n Y t 1 Y t 1 T s t ( k ) 1 ; α ^ = t = p + 1 n ( y t ρ T Y t 1 ) s t ( k ) t = p + 1 n 1 s t ( k ) ; a n d γ ^ = n p 2 t = p + 1 n 1 s t ( k ) ,
where s t ( k ) = ( y t ρ T ( k ) Y t 1 α ( k ) ) 2 + γ ( k ) 2 . □

2.2. Characteristic Function for Estimation

Thus far, we have considered the conditional distribution of Y t given the preceding data F t 1 . Now, we include the dependency of time series { Y t } by defining the variable d j = ( y j , , y j + p ) for j = 1 , , n p . In each variable { d j } , there are p terms the same as adjacent variable. The distribution of { d j } will be multivariate Cauchy with dimension r = p + 1 .
The CF of each d j is c ( Θ , s ) = E ( exp ( i s T d j ) ) and the ECF is c n ( s ) = 1 n j = 1 n p exp ( i s T d j ) where s = ( s 1 , , s p + 1 ) T .
To estimate the parameters using CF and ECF, we make sure that the joint CF of the AR(p) model has a closed form. In the next result, the closed form expression for the joint CF of the AR(1) model with Cauchy innovations is given.
Proposition 2.
The joint CF of stationary AR(1) model with Cauchy innovations is
c ( s 1 , s 2 ; Θ ) = exp i α ( s 1 + s 2 ) 1 1 ρ × exp γ | s 2 | + | s 1 + ρ s 2 | 1 1 | ρ | .
Proof. 
For stationary AR(1) model y t = ρ y t 1 + ε t , we can rewrite it as
y t = ε t + ρ ε t 1 + ρ 2 ε t 2 + ρ 3 ε t 3 + .
Note that { ε t } are i.i.d from Cauchy ( α , γ ) distribution, and the CF of Cauchy ( α , γ ) is E ( exp ( i s ε t ) ) = exp ( i α s γ | s | ) [11]. Then, the joint CF of ( y t , y t 1 ) is calculated as follows:
c ( s 1 , s 2 ; Θ ) = E [ exp ( i s 1 y t 1 + i s 2 y t ) ] = E [ exp ( i s 1 ( ε t 1 + ρ ε t 2 + ρ 2 ε t 3 + ) ) + i s 2 ( ε t + ρ ε t 1 + ρ 2 ε t 2 + ) ] = E [ exp ( i s 2 ε t + i ( s 1 + s 2 ρ ) ε t 1 + i ρ ( s 1 + ρ s 2 ) ε t 2 + ) ] = exp ( i α s 2 γ | s 2 | ) × exp ( i α ( s 1 + ρ s 2 ) γ | s 1 + ρ s 2 | ) × exp ( i α ( ρ s 1 + ρ 2 s 2 ) γ | ρ s 1 + ρ 2 s 2 | ) × . = exp ( i α ( s 1 + s 2 ) ( 1 + ρ + ρ 2 + ρ 3 + ) ) × exp ( γ ( | s 2 | + | s 1 + ρ s 2 | + | ρ | | s 1 + ρ s 2 | + | ρ 2 | | s 1 + ρ s 2 | + ) ) = exp ( i α ( s 1 + s 2 ) 1 1 ρ × exp ( γ ( | s 2 | + | s 1 + ρ s 2 | ( 1 + | ρ | + | ρ 2 | + ) ) ) = exp i α ( s 1 + s 2 ) 1 1 ρ × exp γ | s 2 | + | s 1 + ρ s 2 | 1 1 | ρ | .
The joint CF for a higher dimension can be obtained in similar manner. Now, the model parameters can be estimated by solving the following integral with CF and ECF as defined in [10]:
w Θ ( s ) ( c n ( s ) c ( s ; Θ ) ) d s = 0 .
where optimal weight function
w Θ ( s ) = 1 2 π exp ( i s T d j ) log f ( y j + p | y j , , y j + p 1 ) Θ d y j d y j + p .
Remark 1.
For a stationary AR(l) process { Y t } with p = l , the ECF estimator defined by Equation (5) with optimal weight function defined in Equation (6) is a conditional ML (CML) estimator and hence asymptotically efficient. The conditional log pdf for Cauchy distribution is:
log f ( y j + p | y j , , y j + p 1 ) = log γ log π log γ 2 + ( y t ρ T Y t 1 α 2 ) 2 .
The proof is similar to the proof of Proposition 2.1 in [10].

3. Simulation Study

In this section, we assess the proposed model and the introduced estimation technique using a simulated data set. We discuss the estimation procedure for the AR(2) model with Cauchy innovations. The AR ( 2 ) model defined in (1) is simulated with ρ 1 and ρ 2 as model parameters. We generate 1000 trajectories, each of size N = 500 of Cauchy innovations using the normal variance-mean mixture form ε = α + V Z with Z N ( 0 , 1 ) , i.e., standard normal and V IG ( 1 / 2 , γ 2 / 2 ) . We then use the following simulation steps to generate the Cauchy innovations:
  • step 1: Generate standard normal variate Z;
  • step 2: Generate inverse gamma random variate IG ( 1 / 2 , γ 2 / 2 ) with γ = 2 ;
  • step 3: Using the relation ε = α + V Z , we simulate the Cauchy innovations with α = 1 ;
  • step 4: The time series data y t is generated with model parameters ρ 1 = 0.5 and ρ 2 = 0.3 .
The exemplary time series data plot and scatter plot of innovation terms are shown in Figure 1. We apply the discussed EM algorithm to estimate the model parameters and distribution parameters. The relative change in the parameters is considered to terminate the algorithm. The following is the stopping criteria which is commonly used in literature:
max | α ( k + 1 ) α ( k ) α ( k ) | , | γ ( k + 1 ) γ ( k ) γ ( k ) | , | ρ 1 ( k + 1 ) ρ 1 ( k ) ρ 1 ( k ) | , | ρ 2 ( k + 1 ) ρ 2 ( k ) ρ 2 ( k ) | < 10 4 .
We compare the estimation results of Cauchy ( α , γ ) with the EM algorithm and maximum likelihood (ML) estimation. The ML estimates are computed using the inbuilt function “mlcauchy” in R, which uses the exponential transform of the location parameter and performs non-linear minimization by a Newton-type algorithm. The comparison of the estimates of Cauchy( α , γ ) are shown in boxplots in Figure 2. From the boxplots, we find that the EM algorithm converges near to the true value of the Cauchy ( α , γ ) as compared to the ML estimation. There is a possibility of achieving a better result from ML method if a different algorithm or inbuilt function for optimization are used for estimation.

4. Conclusions and Future Scope

In this work, we derive the closed form of estimates of AR model with Cauchy innovations using an EM algorithm. The performance of the proposed algorithm is compared with the ML method using simulated data. The ML estimation is found using an inbuilt function in R. Another benefit of using the EM algorithm is that it calculates the model as well as the innovation parameters simultaneously. It is evident from the boxplot that the EM algorithm outperforms the ML method. Further, we discuss another approach based on CF to estimate the AR model parameters with stable distribution. In the future, we plan to study and compare the proposed algorithm and ECF based estimation method with the existing techniques in [5,6,7] for an AR model with infinite variance. Further, the real life phenomena can be studied using the proposed model and methods.

Author Contributions

Conceptualization, A.K. and M.S.D.; Methodology, A.K. and M.S.D.; WritingOriginal Draft Preparation, M.S.D. and A.K.; WritingReview & Editing, A.K. and M.S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

M.S.D. would like to thank her institute Indian Institute of Technology Ropar, India and the Ministry of Education (MoE), Government of India, for supporting her research work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, T.; Zhang, P.; Dai, W.S.; Xie, M. An intermediate distribution between Gaussian and Cauchy distributions. Phys. A Stat. Mech. Appl. 2012, 391, 5411–5421. [Google Scholar] [CrossRef] [Green Version]
  2. Kagan, Y.Y. Correlations of earthquake focal mechanism. Geophys. J. Int. 1992, 110, 305–320. [Google Scholar] [CrossRef] [Green Version]
  3. Bjerkedal, T. Acquisition of resistance in guinea pigs infected with different doses of virulent tubercle bacilli. Am. J. Hyg. 1960, 72, 130–148. [Google Scholar] [PubMed]
  4. Choi, J.; Choi, I. Maximum likelihood estimation of autoregressive models with a near unit root and Cauchy errors. Ann. Inst. Stat. Math. 2019, 71, 1121–1142. [Google Scholar] [CrossRef]
  5. Jiang, Y. An exponential-squared estimator in the autoregressive model with heavy-tailed errors. Stat. Interface 2016, 9, 233–238. [Google Scholar] [CrossRef]
  6. Li, J.; Liang, W.; He, S.; Wu, X. Empirical likelihood for the smoothed LAD estimator in infinite variance autoregressive models. Statist. Probab. Lett. 2010, 80, 1420–1430. [Google Scholar] [CrossRef]
  7. Tang, L.; Zhou, Z.; Wu, C. Efficient estimation and variable selection for infinite variance autoregressive models. J. Appl. Math. Comput. 2012, 40, 399–413. [Google Scholar] [CrossRef]
  8. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via EM Algorithm. J. R. Stat. Soc. Ser. B 1977, 39, 1–22. [Google Scholar]
  9. McLachlan, G.J.; Krishnan, T. The EM Algorithm and Extensions, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  10. Yu, J.; Knight, J.L. Empirical characteristic function in time series estimation. Econom. Theory 2002, 18, 691–721. [Google Scholar]
  11. Feller, W. An Introduction to Probability Theory and Its Applications, 2nd ed.; John Wiley & Sons: New York, NY, USA, 1991; Volume 2. [Google Scholar]
  12. Ling, S. Self-weighted least absolute deviation estimation for infinite variance autoregressive models. J. R. Stat. Soc. Ser. B 2005, 67, 381–393. [Google Scholar] [CrossRef]
Figure 1. (a) The data plot of exemplary time series of length N = 500 and (b) the scatter plot of the corresponding innovation terms of the AR(2) model with Cauchy innovations. The chosen parameters of the model are ρ 1 = 0.5 , ρ 2 = 0.3 , α = 1 , and γ = 2 .
Figure 1. (a) The data plot of exemplary time series of length N = 500 and (b) the scatter plot of the corresponding innovation terms of the AR(2) model with Cauchy innovations. The chosen parameters of the model are ρ 1 = 0.5 , ρ 2 = 0.3 , α = 1 , and γ = 2 .
Engproc 18 00021 g001
Figure 2. Boxplots of the estimates of the AR(2) model’s parameters with theoretical values (a) α = 1 and (b) γ = 2 represented with blue dotted lines. The boxplots are created using 1000 trajectories, each of length 500.
Figure 2. Boxplots of the estimates of the AR(2) model’s parameters with theoretical values (a) α = 1 and (b) γ = 2 represented with blue dotted lines. The boxplots are created using 1000 trajectories, each of length 500.
Engproc 18 00021 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dhull, M.S.; Kumar, A. Expectation-Maximization Algorithm for Autoregressive Models with Cauchy Innovations. Eng. Proc. 2022, 18, 21. https://doi.org/10.3390/engproc2022018021

AMA Style

Dhull MS, Kumar A. Expectation-Maximization Algorithm for Autoregressive Models with Cauchy Innovations. Engineering Proceedings. 2022; 18(1):21. https://doi.org/10.3390/engproc2022018021

Chicago/Turabian Style

Dhull, Monika S., and Arun Kumar. 2022. "Expectation-Maximization Algorithm for Autoregressive Models with Cauchy Innovations" Engineering Proceedings 18, no. 1: 21. https://doi.org/10.3390/engproc2022018021

Article Metrics

Back to TopTop