1. Introduction
In this paper, we investigate the relationship between a set of linear restrictions on the parameters of a Vector Autoregressive Moving Average (VARMA) model (see [
1]) and the autoregressive metric (AR-metric hereafter), a notion of the distance between two univariate ARMA models introduced by Piccolo [
2]. In particular, we show that these linear restrictions are satisfied if and only if the distance
d between the two given ARMA models (say
and
) is zero. This result provides the logical basis for using
as null hypothesis for testing this set of restrictions. Moreover, we show that the set of linear restrictions considered is sufficient for the condition of Granger noncausality ([
3]), while in the VAR framework it becomes also a necessary condition (see [
4]). This theoretical result allows the implementation of an inferential procedure and a bootstrap algorithm. Our procedure is verified by some Monte Carlo experiments also in a quite small sample. The paper is organized as follows.
Section 2 introduces the notion of the distance between ARMA models and specifies the relationship between the AR metric and the set of linear restrictions considered for a VARMA model.
Section 3 presents the inferential implication.
Section 4 provides some Monte Carlo evidence about the finite sample behavior of our testing procedure.
Section 5 contains two empirical illustrations.
Section 6 gives some concluding remarks.
2. Linear Restrictions in a VARMA Model and AR-Metric
Let
be a zero mean invertible ARMA model defined as
where
and
are polynomials in the lag operator
L, with no common factors, and
is a white noise process with constant variance
. It is well-known that this process admits the following representation:
where the AR(∞) operator is defined by
with
.
Let
be the class of ARMA invertible models. If
and
, following Piccolo [
2], the AR-metric is defined as the Euclidean distance between the corresponding
π-weights sequence,
,
The AR-metric
d has been widely used in time series analysis (see, e.g., [
5,
6,
7,
8,
9,
10]). We observe that Equation (1) is a well-defined measure because of the absolute convergence of the
π-weights sequences.
Now, we consider the following VARMA model of order
, for an
vector time series
:
where
and
are two
matrices of polynomials in the lag operator
L, and
is an
vector white noise process with positive definite covariance matrix Σ. We assume that det
for
. This condition allows non-stationarity for the series, in the sense that the characteristic polynomial of the VARMA model described by the equation det
may have roots on the unit circle. Condition det
for
, however, excludes explicitly explosive processes from our consideration. We further assume that the model Equation (2) satisfies the usual identifiability conditions. If
, we obtain a pure vector autoregressive (VAR) model of order
p. If
, we obtain a pure vector moving average (VMA) model of order
q. Consider the partition
where
is a scalar time series and
is an
vector of time series. Accordingly, the model Equation (2) for the partition of
can be rewritten as:
where
and
are matrix polynomials in the lag operator
L, with
. In this framework it is well-known (see, for example, [
11]) that
does not Granger-cause
if and only if
and that a sufficient condition for Equation (4) to hold is
We note that if the condition Equation (5) holds then
follows a univariate ARMA model given by:
The main aim of this paper is to establish the implications of the set of linear restrictions Equation (5), using the notion of the distance between ARMA models measured by Equation (1). In particular, we will consider the distance between the ARMA() model Equation (6) (denoted ) and the ARMA model for the subprocess implied by the VARMA() model Equation (2) (denoted ).
Following Lütkepohl [
1], the implied ARMA model
can be obtained as follows. Premultiplying both sides of Equation (2) by the adjoint of
, denoted as
, we obtain
We note that each component of
is a sum of finite order MA processes, thus it is a finite order MA process (see Proposition 11.1 in [
1]). Hence, the subprocess
follows an ARMA model given by:
where
is univariate white noise and
is an invertible polynomial in the lag operator
L. More precisely,
and
are such that
where
denotes the first row of the matrix
. Finally, we observe that
has also the following autoregressive representation of infinite order:
where
2.1. Theoretical Results
We consider the distance according to Equation (1) between the model Equations (6) and (8)
and
:
where
The following proposition provides a necessary and sufficient condition for the set of linear restrictions Equation (5) in terms of the distance .
Proposition 1. if and only if .
Proof of Proposition 1. (⇒) We have
and the first row the matrix
is such that
where
and
with
If
, then
and
Thus we have that
(where this equality between random variables means equality with probability 1) and
. It follows that
and hence
.
(⇐) We have to show that if , then . We may have two cases: or .
First case: .
If
, then
On the other hand, we have
and hence
Using the Schur’s formula, we get
Thus
assume the following expression
where
.
Since the degree of polynomial
is finite
Equation (9) implies that
Since
it follows for Equation (10) that it must be
Since by hypothesis
, it follows that
and this in turn implies that
and
On the other hand
is such that
and hence
where this equality is with probability 1. Since
is a white noise, Equation (11) implies that
.
Second case: .
By hypothesis
, this implies that
and the first row of the matrix
is given by
where
If
, then
and hence
The following equality then occurs with probability 1:
Since is a white noise, this implies that and .
☐
We have also the following corollaries.
Corollary 1. Let be a pure VAR(p) process. y does not Granger-cause x if and only if .
Proof of Corollary 1. If y does not Granger-cause x, then . By hypothesis, . Hence we have . It follows from Proposition 1 that .
If , by Proposition 1, it follows that and this, in a VAR framework, implies that y does not Granger-cause x. ☐
Corollary 2. Let be a pure VMA(q) process. y does not Granger-cause x if and only if .
Proof of Corollary 2. It is similar to the proof of Corollary 1. ☐
3. Inferential Implications
Proposition 1 allows us to test the set of linear restrictions Equation (5) considering the null hypothesis
. Further, we observe that if the process
follows a VAR model, Corollary 1 establishes that the Granger noncausality from
to
is equivalent to the condition
. Thus, in a VAR framework, we can test for Granger noncausality from
to
using the null hypothesis
without considering the nature of the involved variables. In fact, it is well-known that the use of non-stationary data in causality tests can yield spurious causality results (see, e.g., [
12]). Thus, before testing for Granger causality, it is important to establish the properties of the time series involved because different model strategies must be adopted when: the series are I(0), the series are partly I(0) and partly I(1), the series are determined I(1) but not cointegrated, or the series are cointegrated. Of course, the weakness of this strategy is that incorrect conclusions drawn from preliminary analysis might be carried over into the causality tests. In the VAR framework an alternative method is the so-called
lag-augmented Wald test (see [
13,
14]), which is a modified Wald test that requires the knowledge of the maximum order of integration of the involved variables. In this way, the proposed test based on the AR-metric can be a valid alternative for a Granger noncausality test (see [
4]), since it does not require the exact knowledge of the series properties or the knowledge of the maximum order of integration.
To conduct inference on the basis of Proposition 1, we need an asymptotic distribution for
. In the class of ARMA processes, the asymptotic distribution of the maximum likelihood estimator
has been studied, among others, in [
5,
15]. In this case, for two independent ARMA(
) processes
X and
Y, under the null hypothesis
, the maximum likelihood estimator
has the following asymptotic distribution:
where
are independent
-distributions with
degrees of freedom,
are the eigenvalues of the covariance matrix of
and
. The evaluation of this distribution can be cumbersome; hence approximations, as well as evaluation algorithms, have been proposed (see [
15]). Anyhow, in our framework, the ARMA models implied by Equation (6) and by the VARMA model Equation (8) under the null hypothesis
are equal, so they cannot be considered independent. Then, to conduct the inferential procedures, we suggest the bootstrap algorithm proposed by Di Iorio and Triacca [
4], which is described in the next section.
3.1. The Bootstrap Test Procedure
For an easy illustration of our bootstrap procedure, let us consider a bivariate VARMA() model simply denoted as where , with covariance matrix Σ and, based on Proposition 1, we want to test the null hypothesis using
Estimate on the observed data the VARMA() and obtain , , and the residuals ;
using the estimated parameters from Step 1, obtain the univariate ARMA implied by the estimated VARMA for the subprocess ;
evaluate the AR(∞) representation truncated at some suitable lag of the ARMA model in Step 2 (model );
estimate for , using the observed data, an ARMA() model under the null hypothesis and evaluate its AR(∞) representation truncated at some suitable lag (model );
evaluate the distance between the AR() and the AR() obtained in Steps 3 and 4;
estimate the VARMA() model under the null hypothesis to obtain the estimates , and ;
apply bootstrap to the re-centered residuals and obtain the pseudo-residuals ;
generate the pseudo-data obeying the null hypothesis using with ;
using the pseudo-data , repeat Steps 1–5 to obtain the bootstrap estimate of the distance ;
repeat Steps 7–9 for b times;
evaluate the bootstrap p-value as the proportion of the b estimated bootstrap distance that exceeds the same statistic evaluated on the observed data , that is, .
When this procedure is applied, two remarks concerning the pseudo-data generation and the modeling of the dependency across the subprocess are in order. Firstly, in a well-specified model framework (as well as during a simulation exercise), the estimated residuals
do not show any autocorrelation structure, so we do not need any particular resampling scheme for dependent data to obtain pseudo-error terms
, and we can then apply a simple resampling procedure. Besides, for empirical studies the pseudo-data can be obtained considering several resampling strategies, as a block bootstrap algorithm (see [
16]). Secondly, in order to reproduce the dependency across the subprocess expressed by Σ in the pseudo-data, we simply have to apply the resampling algorithm to the entire
matrix of the estimated residuals
.
5. Empirical Applications
In this section we present two empirical examples to illustrate the application of the test suggested in the paper. First, we consider a VAR model and in particular we examine the causal relationship between the log of real per capita income and the inflation. Then, we consider a VARMA example based on the SCC dataset discussed in [
22].
To take into account any possible dependence structure in the residuals of the estimated models, we use the Stationary Bootstrap ([
23]) as resampling algorithm. The Stationary Bootstrap is a block bootstrap scheme where the resampled pseudo-series are stationary; this scheme chains blocks of observations of the original series starting at random locations, and the length of each block is randomly chosen from a geometric distribution. Following Palm
et al. [
24], the mean block length can be computed as a function of the length of the time sample; by some exploratory simulations we verify the robustness of the tests to different block sizes, so we report results for blocks
.
To discuss the possible causal relationship between the log of real per capita income (
y) and inflation (
) we re-examined the dataset used by Ericsson
et al. [
25]. The dataset refers to United States over the period 1953–1992 and can be downloaded from the
Journal of Applied Econometrics Data Archive. The VAR order selection is based on Bayesian Information Criterion and the following model is estimated.
The computed
-statistic is equal to 0.35 with a bootstrap
p-value 0. This result indicates the presence of Granger causality from output to inflation. This finding is in accordance with the results of Ericsson
et al. [
25]. The same result is obtained using the lag-augmented Wald test.
The
SCC dataset discussed by Tiao and Box [
22] considers the quarterly time series of the U.K. Financial Time Ordinary Share Index, the U.K. Car Production and the U.K. Financial Time Commodity Price from the III Quarter 1952 to the IV Quarter 1967. The goal is verify the possibility of predicting the first variable from the lagged values of the last two. According to Tiao and Box [
22], a VARMA(
) is the best model for this data, then a null hypothesis following Equation (5) will be the inferential base to test just a sufficient condition on the predictability hypothesis. The VARMA(
) maximum likelihood parameter estimates using the Kalman filter procedure implemented in Gretl (ver. 1.9.9) are the following (standard errors in bracket):
The estimates are quite similar to the values reported as “full model” in the Table 10 in [
22], taking into account the difference in the estimation algorithm and software. The computed
-statistic is equal to 4.58 with a bootstrap
p-value 0.225, evaluated on 500 bootstrap replications, and this finding is in accordance with the results of “final model” in the Table 10 in Tiao and Box [
22]. We perform also a Wald test on the same null hypothesis, the value is 36.684, which asymptotically rejects the null, but with a bootstrap
p-value 0.146 that sustains the results of our test.