Next Article in Journal
A Nonlinear Maximum Correntropy Information Filter for High-Dimensional Neural Decoding
Next Article in Special Issue
Upper Bounds for the Capacity for Severely Fading MIMO Channels under a Scale Mixture Assumption
Previous Article in Journal
Dynamic Risk Measures for Processes via Backward Stochastic Differential Equations Associated with Lévy Processes
Previous Article in Special Issue
On Representations of Divergence Measures and Related Quantities in Exponential Families
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Designing Bivariate Auto-Regressive Timeseries with Controlled Granger Causality

Japan Advanced Institute of Science and Technology, 1-1 Asahidai, Nomi 923-1292, Ishikawa, Japan
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(6), 742; https://doi.org/10.3390/e23060742
Submission received: 12 May 2021 / Revised: 1 June 2021 / Accepted: 8 June 2021 / Published: 12 June 2021
(This article belongs to the Special Issue Measures of Information)

Abstract

:
In this manuscript, we analyze a bivariate vector auto-regressive (VAR) model in order to draw the design principle of a timeseries with a controlled statistical inter-relationship. We show how to generate bivariate timeseries with given covariance and Granger causality (or, equivalently, transfer entropy), and show the trade-off relationship between these two types of statistical interaction. In principle, covariance and Granger causality are independently controllable, but the feasible ranges of their values which allow the VAR to be proper and have a stationary distribution are constrained by each other. Thus, our analysis identifies the essential tri-lemma structure among the stability and properness of VAR, the controllability of covariance, and that of Granger causality.

1. Introduction

1.1. Background and Motivation

In the field of cognitive psychology, the human perception of the life-likeness (called animacy perception) of one or multiple moving geometric patterns has been studied for decades [1,2,3,4,5]. There are multiple findings on the effect of “synchrony” or “temporal contingency” between multiple moving points on animacy perception. Findings from one line of research [2] have suggested that a higher degree of “temporal contingency” of the moving objects is related to a higher likelihood of animacy perception. Findings from the other line of research [6] have suggested that the highest “temporal contingency”, presented in the form of perfect synchronization, would decrease the likelihood of animacy perception.
These two lines of research have together suggested the existence of multiple types of “temporal contingency”. Nevertheless, this past research does not appear to clarify these types. Further, confusion surrounding these two distinct types of effects have led to two lines of apparently conflicting effects of “temporal contingency”.
With this potential conflict in the literature on animacy perception in mind, we explore a theoretical framework which can generate timeseries of multiple random variables with multiple distinct types of statistical dependency. One such system, which is sufficiently simple and readily manipulable, is vector auto-regression (VAR). Vector auto-regression is a random process for generating multivariate timeseries for a given set of parameters. In this manuscript, we specifically consider only bivariate VAR, which is a minimal system with interaction between two moving points.

1.2. Vector Auto-Regressive Model, Granger Causality, and Transfer Entropy

Importantly, bivariate VAR, a series of paired random variables ( x t , y t ) for t = 0 , 1 , , has two types of statistical dependency—that is, the correlation and Granger causality of a timeseries [7], which has been identified as transfer entropy [8] of the timeseries generated by a Gaussian process by [9]. The correlation between univariate series x and y is a statistical dependency between x t and y t in the limit t (if it exists), while a Granger causality from y to x is that between x t and y t 1 given x t 1 in the limit t (if it exists). Conceptually, correlation captures a type of similarity between two timeseries, whereas Granger causality captures the “reactiveness” of one timeseries to another.
Thus, given these differences in theory, our goal is to propose a theoretical method to generate a bivariate random timeseries with a desired correlation and two types. The Granger causality of VAR has also been considered in other fields. In econometrics, the field in which it was originally proposed [7], Granger causality has been used as a measure of interaction between a pair of economic timeseries [10,11,12]. It has also been used in general behavioral sciences [13,14], particularly in computational neuroscience [15,16,17,18]. Such an in-principle data generation technique would be vital to the testing of any hypothesis on the empirical nature of timeseries (e.g., animacy perception) in the empirical sciences using VAR and Granger causality, as mentioned above. To our knowledge, however, there has been no mathematical analysis of the theoretical limitation of such a data generation technique for a given statistics.
Thus, we first need to explore the mathematical relationship among the parameters in a VAR with the correlation and Granger causality in a timeseries generated from it. In this paper, we therefore explore a theoretical structure of bivariate VAR from the designer’s perspective, and analyze a mathematical limit to the extent which we can simultaneously control correlation and Granger causality of a bivariate timeseries.
This paper is written with the following structure. In Section 2, the VAR model is defined, from which a set of basic statistical properties of the VAR model are derived, such as Granger causality (Section 3) with a set of parameters. In Section 4, the existence of the stationary distribution of the VAR is analyzed. This is a foundation which sets the limit of a controllable set of parameters. In Section 5, we give a method to derive the parameters of VAR for a given set of statistics in bivariate timeseries. In Section 6, the mathematical analysis provided in this paper is summarized and a remark on the design principle of bivariate timeseries generated by VAR is added.

2. Vector Auto-Regression (VAR)

In theory, Granger causality (GC) is the transfer entropy of random variables in a bivariate vector auto-regression (VAR) model up to a constant factor of 2, if the VAR model has a stationary distribution [9]. Thus, it is straightforward to start with the bivariate VAR and derive its transfer entropy. In this way, we can derive a rich mathematical relationship between GC and the properties of VAR, rather than just a statistics of the bivariate timeseries.
Definition 1.
For some real vector μ R 2 and some positive-definite matrix Σ R 2 × 2 , suppose the random variable ϵ t for every integer t = 0 , 1 , is drawn from the bivariate normal distribution N ( ϵ t | μ , Σ ) with mean μ and variance Σ. Define the initial vector by v 0 = x 0 y 0 , and for t 0 and a given coefficient matrix with real entries A : = a 0 , 0 a 0 , 1 a 1 , 0 a 1 , 1 R 2 × 2 , define the random variable v t by
v t + 1 : = A v t + ϵ t .
Then, bivariate vector auto-regression is defined by the semi-infinite series of random variables V = ( v 0 , v 1 , v 2 , ) .
In general, one can generate a timeseries v 0 , v 1 , by fixing a set of the VAR parameters, the coefficient matrix A, and the base covariance matrix Σ , where the base mean vector μ is omitted as its effect is lost in the limit t when the VAR is stationary. The stationary correlation (covariance) Σ ^ and Granger causality G 0 and G 1 defined later are the statistics of the timeseries generated by a VAR model (Figure 1). In what follows, we first explore the forward relationship of how the statistics Σ ^ and G 0 , G 1 are given by the VAR parameters ( A , Σ ) . We then consider the backward relationship in which the VAR parameters ( A , Σ ) suffice to generate a timeseries with given desired timeseries statistics ( G 0 , G 1 , Σ ^ ) .

2.1. Marginal Distribution of the VAR at Each Step

Lemma 1
(Marginal distribution of the VAR random variables at each step). The VAR model with the initial vector v 0 R 2 and the coefficient matrix A R 2 × 2 has the bivariate normal distribution
N ( v t | μ t , Σ t )
as its marginal distribution of the random variable v t at each step t = 0 , 1 , , where
μ t : = A t v 0 , Σ t = s = 0 t A s Σ ( A s ) .
Proof. 
By Definition 1, Lemma 1 holds for t = 0 . For t + 1 > 0 , we prove Lemma 1 by assuming that it holds up to t 0 . By this assumption held for t, we have the distribution of v t R 2 as the bivariate normal distribution
N ( v t | μ t , Σ t ) = ( 2 π ) 1 | Σ t | 1 2 e 1 2 ( v t μ t ) Σ t 1 ( v t μ t )
with its mean μ t and its covariance matrix Σ t . Then, the random variable A v t is distributed by the normal distribution
N ( A v t | A μ t , A Σ t A ) = ( 2 π ) 1 | A Σ t A | 1 2 × e 1 2 ( A ( v t μ t ) ) ( A Σ t A ) 1 ( A ( v t μ t ) )
with its mean A μ t and the covariance matrix A Σ t A . The random variable ϵ t is distributed by the following normal distribution:
N ( ϵ t | 0 , Σ ) = ( 2 π ) 1 | Σ | 1 2 e 1 2 ϵ t Σ 1 ϵ t .
Thus, by VAR, Equation (1), we have the random variable
v t + 1 : = A v t + ϵ t
which has a distribution calculated by the following integral:
P ( v t + 1 ) = ϵ t R 2 N ( v t + 1 ϵ t | A μ t , A Σ t A ) N ( ϵ t | 0 , Σ ) d ϵ t .
Calculating this, we have
P ( v t + 1 ) = N v t + 1 | A μ t , A Σ t A + Σ .
Thus defining by μ t + 1 : = A μ t and Σ t + 1 = A Σ t A + Σ , Lemma 1 holds for t + 1 . By expanding this, we have the Lemma 1 for any integer t 0 . □

2.2. Stability of VAR: Lyapunov Equation

By Lemma 1, the mean and covariance matrix of the random variable at the t th step are
μ t = A t v 0 and Σ t = s = 0 t A s Σ ( A s ) .
From this, we have the stationary distribution
lim t N ( v t | μ t , Σ t ) ,
if and only if the absolute values of all the eigenvalues λ 0 , λ 1 C of the coefficient matrix A are less than 1. If there is such a stationary distribution, we call the VAR stable, and its stationary distribution is the bivariate normal distribution
N ( v ^ | 0 2 , Σ ^ ) ,
where the stationary mean vector v ^ R 2 and stationary covariance matrix Σ ^ R 2 × 2 are defined as follows. If the VAR is stable, we have the following Lyapunov equation of the stationary covariance matrix Σ ^ R 2 × 2 :
Σ ^ = Σ + A Σ ^ A .
The Lyapunov equation is solved analytically by
vec Σ ^ = I 4 A A 1 vec Σ ,
where I d R d × d is the d th order identity matrix, ⊗ denotes the Kronecker product, and vec X for any matrix X = ( x i , j ) i = 1 , , n , j = 1 , , m is the vectorization operator vec ( ) : R n × m R n m × 1 defined by
vec X : = ( x 1 , 1 , x 2 , 1 , , x n , 1 , , x 1 , m , x 2 , m , , x n , m ) .
The Lyapunov Equation (4) has the solution for Σ ^ if the VAR is stable, but not vice versa. This is shown by Lemma 4.

3. Transfer Entropy and Granger Causality

In [9], the transfer entropy of an appropriate triplet of variables in the VAR model is shown to be equivalent to Granger causality up to the constant factor 2. Following this guide, we define this quantity as the Granger causality of the VAR model, below.
Although this relationship has been known in a more general form [9], we re-derive it for bivariate VAR in order to later analyze the structure of VAR and GCs in depth—for example, its upper and lower bounds (Lemma 3), stability (Section 4), and design principle (Section 5).
Definition 2.
If VAR with its random variables v t = x t , y t R 2 for t = 0 , 1 , is stable, transfer entropy from y to x is defined by
T y x : = lim t ( H ( x t + 1 | x t ) H ( x t + 1 | x t , y t ) ) ,
and the transfer entropy from x to y is defined by
T x y : = lim t ( H ( y t + 1 | y t ) H ( y t + 1 | y t , x t ) ) ,
where the differential entropy of random variable x with its probability density function P is
H ( x ) : = x Ω P ( x ) log P ( x ) d x ,
and the conditional entropy is
H ( x | y ) : = H ( x , y ) H ( y ) .
In particular, we call two times of transfer entropy Granger causality denoted by
G 0 = 2 T y x and G 1 = 2 T x y .
Specifically, GCs are specifically written by the terms of the VAR parameters in the following lemma.
Lemma 2
(Granger causality). If a stable VAR has its covariance matrix, coefficient matrix, and stationary matrix
Σ = σ 0 , 0 σ 0 , 1 σ 1 , 0 σ 1 , 1 , A = a 0 , 0 a 0 , 1 a 1 , 0 a 1 , 1 , Σ ^ = σ ^ 0 , 0 σ ^ 0 , 1 σ ^ 1 , 0 σ ^ 1 , 1 ,
each Granger causality of this VAR for i = 0 , 1 is
G i = log 1 + a i , 1 i 2 det Σ ^ σ ^ i , i σ i , i .
Proof. 
In general, the differential entropy of multivariate normal distribution N ( v | μ , Σ ) is
H ( v ) = 1 2 log | 2 π e Σ | ,
where e 2.71 is Napier’s constant. For the joint probability distribution of v t = ( x t , y t )
P ( v t + 1 | v t ) = N ( v t + 1 | A v t , Σ ) ,
the two marginal probability distributions of x t , y t are
P ( x t + 1 | v t ) = N ( x t + 1 | ( 1 , 0 ) A v t , σ 0 , 0 ) and P ( y t + 1 | v t ) = N ( y t + 1 | ( 0 , 1 ) A v t , σ 1 , 1 ) .
Thus, the conditional entropy of x t + 1 and y t + 1 given v t = ( x t , y t ) are
H ( x t + 1 | x t , y t ) = 1 2 log | 2 π e σ 0 , 0 | and H ( y t + 1 | x t , y t ) = 1 2 log | 2 π e σ 1 , 1 | .
With the conditional probability distribution and marginal probability distribution
P ( v t + 1 | v t ) = N ( v t + 1 | A v t , Σ ) and P ( v t ) = N ( v t | A t v 0 , Σ t ) ,
the joint probability distribution of v t and v t + 1 is their product
P ( v t + 1 , v t ) = N ( v t + 1 | A v t , Σ ) N ( v t | A t v 0 , Σ t ) .
Specifically, this quad-variate normal distribution is
P ( v t + 1 , v t ) = e 1 2 ( v t + 1 A v t ) Σ 1 ( v t + 1 A v t ) 1 2 ( v t A t v 0 ) Σ t 1 ( v t A t v 0 ) ( 2 π ) 2 | Σ | 1 2 | Σ t | 1 2 .
Applying the identities
v t + 1 A v t = v t + 1 A t + 1 v 0 A ( v t A t v 0 ) ,
Σ t : = Σ + A Σ t A A Σ t Σ t A Σ t = Σ 1 Σ 1 A A Σ 1 Σ t 1 + A Σ 1 A 1 ,
and | Σ | | Σ t | = | Σ t | to P ( v t + 1 , v t ) , we have
P ( v t + 1 , v t ) = N ( v t | μ t , Σ t ) ,
where
v t : = v t + 1 v t , μ t : = A t + 1 v 0 A t v 0 , Σ t : = Σ + A Σ t A A Σ t Σ t A Σ t .
From this joint probability distribution P ( v t + 1 , v t ) , we drive the marginal distributions
P ( x t + 1 , x t ) = N ( x t + 1 , x t | μ t , 0 , Σ t , 0 ) , P ( y t + 1 , y t ) = N ( y t + 1 , y t | μ t , 1 , Σ t , 1 ) ,
where for i = 0 , 1 the mean vectors and covariance matrices are defined as follows:
μ t , i : = e i A t + 1 v 0 e i A t v 0 = ( I 2 e i ) μ t , Σ t , i : = e i ( Σ + A Σ t A ) e i e i A Σ t e i e i Σ t A e i e i Σ t e i = ( I 2 e i ) Σ t ( I 2 e i ) ,
with the unit vectors e 0 : = ( 1 , 0 ) , e 1 : = ( 0 , 1 ) .
Thus, we have the joint entropy of x t and x t + 1
H ( x t + 1 , x t ) = 1 2 log | 2 π e Σ t , 0 |
= 1 2 log ( 2 π e ) 2 | e 0 ( Σ + A Σ t A ) e 0 e 0 Σ t e 0 ( e 0 A Σ t e 0 ) 2 | ,
and the marginal distribution of x t
H ( x t ) = 1 2 log 2 π e | e 0 Σ t e 0 | .
Using these, we have the conditional entropy
H ( x t + 1 | x t ) = 1 2 log 2 π e 0 ( Σ + A Σ t A ) e 0 e 0 Σ t e 0 ( e 0 A Σ t e 0 ) 2 | e 0 Σ t e 0 | .
By the stability of VAR, the Lyapunov Equation (4) holds, and this conditional entropy in the limit t is
lim t H ( x t + 1 | x t ) = 1 2 log 2 π ( e 0 Σ ^ e 0 ) 2 ( e 0 A Σ ^ e 0 ) 2 | e 0 Σ ^ e 0 | .
Applying Definition 2 and denoting by entries in the stationary covariance matrix Σ ^ = σ ^ 0 , 0 σ ^ 0 , 1 σ ^ 1 , 0 σ ^ 1 , 1 , we have
G 0 = 2 T y x = log ( σ ^ 0 , 0 ) 2 ( a 0 , 0 σ ^ 0 , 0 + a 0 , 1 σ ^ 1 , 0 ) 2 σ ^ 0 , 0 σ 0 , 0 .
Similarly, we have
G 1 = 2 T x y = log ( σ ^ 1 , 1 ) 2 ( a 1 , 0 σ ^ 0 , 1 + a 1 , 1 σ ^ 1 , 1 ) 2 σ ^ 1 , 1 σ 1 , 1 .
Let us define for i = 0 , 1
δ i : = ( σ ^ i , i ) 2 ( a i , i σ ^ i , i + a i , 1 i σ ^ 1 i , i ) 2 and δ i : = σ ^ i , i σ i , i .
By the Lyapunov equation, σ i , i = σ ^ i , i e i A Σ ^ A e i . Applying this to δ i , we have
δ i = σ ^ i , i 2 e i A σ ^ i , i 2 σ ^ i , i σ ^ 1 i , i σ ^ i , i σ ^ 1 i , i σ ^ 1 i , i 2 A e i ,
δ i = σ ^ i , i 2 e i A σ ^ i , i σ ^ 0 , 0 σ ^ i , i σ ^ 0 , 1 σ ^ i , i σ ^ 1 , 0 σ ^ i , i σ ^ 1 , 1 A e i .
As δ i δ i = a i , 1 i 2 det Σ ^ and G 0 = log 1 + δ i δ i δ i , we have
G i = log 1 + a i , 1 i 2 det Σ ^ σ ^ i , i σ i , i .
 □
The Granger causality has its lower and upper bounds in theory. Although these bounds may be further narrowed by considering the stability of the VAR, what follows below are the theoretical bounds regardless of the stability of the VAR.
Lemma 3
(The upper and lower bound for Granger causality). For each i = 0 , 1 , Granger causality G i has the following bounds:
0 G i log γ i ,
where γ i : = σ ^ i , i σ i , i 1 due to the Lyapunov Equation (4). The lower bound G i = 0 is given only if
a i , 1 i 2 det Σ ^ = 0 .
The upper bound G i = log γ i is given only if
a i , i σ ^ i , i + a i , 1 i σ ^ 1 i , i = 0 .
Proof. 
As the stationary covariance matrix is (semi-)positive definite, det Σ ^ 0 . Thus, the lower bound of Granger causality is G i log ( 1 ) = 0 and this bound is only reacheable when a i , 1 i 2 det Σ ^ = 0 .
Modifying (13) and (14), for i = 0 , 1 we have
( a i , i σ ^ i , i + a i , 1 i σ ^ 1 i , i ) 2 = σ ^ i , i σ ^ i , i σ i , i e G i .
As ( a i , i σ ^ i , i + a i , 1 i σ ^ 1 i , i ) 2 0 and σ ^ i , i > 0 ,
G i log γ i .
This upper bound holds only if a i , i σ ^ i , i + a i , 1 i σ ^ 1 i , i = 0 . □
The upper bound Lemma 3 can also be obtained by the following information-theoretic identity:
lim t I ( x t 1 ; x t ) + I ( x t ; x t 1 | x t 1 ) = lim t I ( x t ; x t 1 , x t 1 ) ,
where lim t I ( x t ; x t 1 | x t 1 ) = 1 2 G 0 is the transfer entropy, lim t I ( x t ; x t 1 , x t 1 ) = 1 2 log σ ^ 0 , 0 σ 0 , 0 , and
lim t I ( x t 1 ; x t ) = 1 2 log σ ^ 0 , 0 2 Σ ^ Σ ^ A A Σ ^ Σ ^ = 1 2 log γ 0 G 0 .

4. Stability and Constraints of VAR

In this study, we primarily consider the class of stable VAR models with a proper set of parameters. In this class, the statistical nature of any VAR is characterized by the base covariance matrix Σ R 2 × 2 , coefficient matrix A R 2 × 2 , and stationary covariance matrix Σ ^ R 2 × 2 . Let us denote the set of (strictly) positive definite matrices by
R + 2 × 2 : = M R 2 × 2 | det M > 0 and tr M > 0 ,
and the set of coefficient matrices of stable VAR models
R * 2 × 2 : = M R 2 × 2 | 1 + | tr M | < det M < 1 .
We will briefly show that the stable set R * 2 × 2 includes all and only coefficient matrices of stable bivariate VAR models.
With this notation of the set of matrices, the two conditions that any proper VAR model needs to satisfy are as follows.
Stability 
Any stable VAR model has both of the eigenvalues λ 0 , λ 1 of its coefficient matrix A meeting | λ 0 | , | λ 1 | < 1 .
Properness 
To have a proper (non-degenerated) bivariate normal distribution in a VAR model, its base covariance matrix Σ and stationary covariance matrix Σ ^ need to satisfy Σ , Σ ^ R + 2 × 2 . The set of positive-definite matrices is equivalently written with the entries of the following matrix:
R + 2 × 2 = C R 2 × 2 C 0 , 0 > 0 , C 1 , 1 > 0 , and C 0 , 0 C 1 , 1 C 0 , 1 C 1 , 0 > 0 .

4.1. Stability of VAR

As stated previously in Section 2.2, the stability of VAR is primarily characterized by the eigenvalues of the coefficient matrix A. However, this condition is equivalent to A R * 2 × 2 , as shown by the following lemma.
Lemma 4.
A given bivariate VAR model with its coefficient matrix A R 2 × 2 is stable if and only if
| tr A | 1 < det A < 1 .
Proof. 
Let λ be an eigenvalue of the coefficient matrix A. Such an eigenvalue then satisfies
f ( λ ) = A λ I 2 = λ 2 tr A λ + det A = 0 .
If a VAR is stable, this eigenvalue needs to satisfy | λ | < 1 . As (24) is rewritten by
f ( λ ) = λ 1 2 tr A 2 1 4 tr A 2 4 det A ,
we analyze this condition on (24) for the following two cases with λ being real or non-real:
  • If λ is real, this stability condition is equivalent to
    tr A 2 4 det A , f ( 1 ) > 0 , f ( 1 ) > 0 , | tr A | < 2 .
  • If λ is not real, this stability condition is equivalent with
    tr A 2 < 4 det A , | λ | 2 < 1 .
If λ of (24) is non-real (Case 2), λ (and its conjugate) is
λ = 1 2 tr A ± j 2 | tr A 2 4 det A | ,
with the imaginary unit denoted by j.
With the inequality (27), the stability condition in this case is
tr A 2 2 < | λ | 2 = det A < 1 .
If λ of (24) is real, tr A 2 4 det A 0 and
f ( 1 ) = 1 tr A + det A > 0
f ( 1 ) = 1 + tr A + det A > 0
| tr A | < 2 .
Combining (30) and (31), we have | tr A | 1 < det A . This inequality with (26),
C 0 < det A tr A 2 2 < C 1 ,
where C 0 : = | tr A | 1 and C 1 : = min 1 , 1 + det A 2 2 . Find for an arbitrary A R 2 × 2 we have the following two inequalities:
| tr A | 1 1 2 tr A 2 ,
and
det A 1 + det A 2 2 .
The inequality (34) holds equality for and only for tr A = 2 , and the inequality (34) holds equality for and only for det A = 1 . As both of these equality conditions do not hold under (29), (29) is equivalent to
C 0 < tr A 2 2 < det A < C 1 .
Integrating the two inequalities (33) for real λ and (36) for non-real λ , the VAR with the coefficient matrix A is stable if
C 0 < det A < C 1
and
C 0 < tr A 2 2 < C 1 .
As the inequality (37) implies 0 < 1 + det A 2 < 1 and tr A < 2 , (38) is equivalent to
tr A 2 2 < 1 + det A 2 2 .
As the upper bound for det A in (37) can be implied by det A < 1 , it is equivalent to
| tr A | 1 < det A < 1 .
Thus, the pair of inequalities (37) and (38) for A is equivalent to the single inequality (40) for A. □

4.2. Stability and Existence of the Solution for the Lyapunov Equation

Intuitively, it would be reasonable if there was a stationary covariance matrix Σ ^ R + 2 × 2 satisfying the Lyapunov Equation (4), if the coefficient matrix is A R * 2 × 2 . However, this is not trivial, as the opposite may not be always true: the existence of Σ ^ R + 2 × 2 does not imply A R * 2 × 2 . This relationship between A and Σ ^ is stated by the following Theorem 1.
Theorem 1.
There is a stationary covariance matrix Σ ^ R + 2 × 2 satisfying the Lyapunov Equation (4), if the coefficient matrix is A R * 2 × 2 . However, the existence of Σ ^ R + 2 × 2 does not imply A R * 2 × 2 .
Proof. 
Find the identity
det I 4 A A = det I 2 a 0 , 0 A I 2 a 1 , 1 A a 0 , 1 a 1 , 0 A 2 = det I 2 A tr ( A ) + A 2 det ( A ) = 1 det ( A ) 2 1 a 0 , 1 a 0 , 0 2 1 a 0 , 1 a 1 , 1 2 a 0 , 1 a 1 , 0 tr ( A ) = 1 det ( A ) 2 1 + det ( A ) 2 tr ( A ) 2 .
By Lemma 4 and (41), we have det I 4 A A > 0 . Thus, Lyapunov Equation (5) has the solution for Σ ^ , as the matrix I 4 A A is invertible. The converse of this theorem does not hold, as we construct a counter-example of the coefficient matrix A such that det I 4 A A < 0 , with which there is a Σ ^ R + 2 × 2 , but such a VAR is not stable. □

5. Design of Bivariate Timeseries Given GCs

The goal of this study was to derive a design principle of bivariate timeseries generated by a VAR model with the desired correlation and two types of Granger causality. In this section, we explore the inter-dependent relationships among the variables in the VAR. This analysis revealed a trade-off limitation in designing these variables of timeseries. Specifically, a timeseries with a certain range of desired Granger causality cannot be realized by a stable VAR, in which no stationary covariance is defined in theory.
The set of parameters in any stable VAR model includes
  • The coefficient matrix A;
  • The base covariance matrix Σ ;
  • The stationary covariance matrix Σ ^ ; and
  • The two types of Granger causality G 0 , G 1 .
There are equality constraints on these variables:
  • The variables A , Σ , Σ ^ need to satisfy the Lyapunov Equation (4).
  • Granger causality G i ( i = 0 , 1 ) is the function of a i , 1 i , σ i , i , and Σ ^ (Lemma 2).
Besides, it is important to know the feasibility of a set of parameters in VAR, which constrains the range of these variables:
  • Stability: A R * 2 × 2 (Section 4);
  • Properness: Σ , Σ ^ R + 2 × 2 (Section 4) and σ i , i σ ^ i , i due to the existence of a solution for the Lyapunov equation; and
  • The bound for each Granger causality: G i [ 0 , log γ i ] (Lemma 3).
The Lyapunov Equation (4) on the matrices can be decomposed into the three equations on the scalar variables as follows. For a coefficient matrix A = a 0 , 0 a 0 , 1 a 1 , 0 a 1 , 1 , let us define two vectors by
a 0 : = a 0 , 0 a 0 , 1 , a 1 : = a 1 , 0 a 1 , 1 .
The Lyapunov equation is then equivalently written with these vectors a 0 , a 1 by the set of the three equations
σ ^ 0 , 0 σ 0 , 0 = a 0 Σ ^ a 0
σ ^ 1 , 1 σ 1 , 1 = a 1 Σ ^ a 1
σ ^ 0 , 1 σ 0 , 1 = a 0 Σ ^ a 1 .
Equations (42) and (43) above imply that each of the vectors a 0 and a 1 are on an ellipsis on each of their planes. This gives the lower bound for σ ^ i , i σ i , i (i.e., one condition of the properness above), as x Σ ^ x 0 for any x R 2 with a positive-definite matrix Σ ^ .
Fixing G 0 and G 1 imposes each of the two vectors a 0 and a 1 on the two parallel lines by
( a i σ ^ i ) 2 = τ i 2 ,
where
σ ^ i : = ( σ ^ i , 0 , σ ^ i , 1 ) , τ i 2 : = σ ^ i , i 2 1 γ i 1 e G i .
Thus, the solution of a i which satisfies the Lyapunov equation and the fixed Granger causality is the four intersections of the ellipsis and the two parallel lines (Figure 2). This ellipsis is obtained by scaling and shearing transformation to the standard circle a 0 , 0 2 + a 0 , 1 2 = 1 . This observation gives the angular parametrization of the solution vector ( a 0 , 0 , a 0 , 1 ) , which is explicitly stated by Lemma 5 in the next section.

5.1. Solution A of the Lyapunov Equality Given Σ ^ , G 0 , and G 1

In what follows, we start with the derivation of the coefficient matrix A as a root of the equality constraint by the Lyapunov Equation (4) and the Granger causality, for a fixed proper Σ ^ , σ i , i and G i for each i = 0 , 1 . The following Lemma 5 gives a necessary condition for the coefficient matrix A R 2 × 2 to satisfy the equality conditions above. Note, however, that such a solution A in this equation does not guarantee the stability of the corresponding VAR (i.e., A R * 2 × 2 ). This sufficiency is explored in Section 5.2.
Lemma 5.
For a given set of parameters, a positive-definite matrix Σ ^ = σ ^ 0 , 0 σ ^ 0 , 1 σ ^ 1 , 0 σ ^ 1 , 1 R + 2 × 2 , σ i , i ( 0 , σ ^ i , i ) , G i [ 0 , log γ i ] for each i = 0 , 1 , suppose that a coefficient matrix A = a 0 , 0 a 0 , 1 a 1 , 0 a 1 , 1 R 2 × 2 , satisfies the set of the equations
a 0 Σ ^ a 0 = σ ^ 0 , 0 σ 0 , 0 a 1 Σ ^ a 1 = σ ^ 1 , 1 σ 1 , 1 σ ^ 0 a 0 2 = τ 0 2 σ ^ 1 a 1 2 = τ 1 2 ,
where for i = 0 , 1
σ ^ i : = ( σ ^ i , 0 , σ ^ i , 1 ) , τ i 2 : = σ ^ i , i 2 1 γ i 1 e G i .
Any coefficient matrix A of a root of this Equation (46) is in the form
A = S 0 cos θ 0 sin θ 0 P 2 S 1 cos θ 1 sin θ 1 ,
where each pair of the angles θ 0 [ 0 , 2 π ) and θ 1 [ 0 , 2 π ) takes one of the two or four pairs satisfying for each i = 0 , 1
sin 2 θ i = e G i 1 γ i 1
and
P 2 : = 0 1 1 0 , S i : = 1 γ i 1 1 σ ^ i , 1 i σ ^ i , i 0 1 1 0 0 σ ^ i , i det Σ ^ .
Proof. 
Find that the following pair of equations in (46) is symmetric under exchange of i = 0 , 1 :
a i Σ ^ a i = σ ^ i , i σ i , i σ ^ i a i 2 = τ i 2 .
Thus, we solve this for i = 0 below, and it holds for i = 1 .
Solving the second equation of (49) for a 0 , 0 , we have
a 0 , 0 = ± τ 0 a 0 , 1 σ ^ 0 , 1 σ ^ 0 , 0 .
Inserting this into the first equation of (49), we have
a 0 , 1 2 = σ ^ 0 , 0 2 ( 1 γ 0 1 ) τ 0 2 det Σ ^ .
Inserting τ 0 2 = σ ^ 0 , 0 2 1 γ 0 1 e G 0 , we have
a 0 , 1 = ± σ ^ 0 , 0 σ 0 , 0 e G 0 1 det Σ ^ .
Inserting this into (50), we have at most four vectors a 0 = ( a 0 , 0 , a 0 , 1 ) as the solution of (49) for i = 1 :
a 0 = c 0 σ ^ 0 , 1 σ ^ 0 , 0 d 0 d 0 , c 0 + σ ^ 0 , 1 σ ^ 0 , 0 d 0 d 0 , c 0 σ ^ 0 , 1 σ ^ 0 , 0 d 0 d 0 , c 0 + σ ^ 0 , 1 σ ^ 0 , 0 d 0 d 0 ,
where for i = 0 , 1
c i : = 1 γ i 1 e G i , d i : = σ ^ i , i σ i , i e G i 1 det Σ ^ .
By symmetry to i = 0 , 1 , there are at most four vectors as the solution of (49) for i = 1 :
P 2 a 1 = a 1 , 1 a 1 , 0 = c 1 σ ^ 1 , 0 σ ^ 1 , 1 d 1 d 1 , c 1 + σ ^ 1 , 0 σ ^ 1 , 1 d 1 d 1 , c 1 σ ^ 1 , 0 σ ^ 1 , 1 d 1 d 1 , c 1 + σ ^ 1 , 0 σ ^ 1 , 1 d 1 d 1 .
Find that these four solution vectors parameterized by
a 0 = S 0 cos θ 0 sin θ 0 and a 1 = P 2 S 1 cos θ 1 sin θ 1 ,
satisfy (49), if ( θ 0 , θ 1 ) holds (48), with the trigonometric identity cos 2 θ i + sin 2 θ i = 1 and
S 0 Σ ^ S 0 = σ ^ 0 , 0 σ 0 , 0 I 2 and S 1 P 2 Σ ^ P 2 S 1 = σ ^ 1 , 1 σ 1 , 1 I 2 .

5.2. Sufficiency of the Solution

For a given set of parameters, Lemma 5 in the previous section gives a set of solutions of the coefficient matrix A for the Lyapunov equation. Note that not all of these solutions A are feasible, in the sense that they satisfy all constraints such as the stability of A R * 2 × 2 and the properness of Σ R + 2 × 2 , in which Σ can be derived from Lyapunov Equation (4) given A and Σ ^ . The following lemmas provide the sufficient condition for a solution A by checking the properness of Σ and the stability of A.
Lemma 6.
Suppose A is a solution of Equation (46) in Lemma 5, represented by a pair of ( θ 0 , θ 1 ) . In this case, Σ R + 2 × 2 , if and only if
cos η ^ γ 0 γ 1 1 2 γ ^ 0 γ ^ 1 cos η ^ θ 0 θ 1 cos η ^ + γ 0 γ 1 1 2 ,
where η ^ [ 0 , 2 π ] is the angler parametrization of the correlation coefficient defined by cos η ^ : = σ ^ 0 , 1 σ ^ 0 , 0 σ ^ 1 , 1 and γ ^ i = 1 γ i 1 .
Proof. 
Applying the polar representation of the a 0 , a 1 in (47) in Lemma 5 to the third Equation (44) of the Lyapunov equation, we have
σ ^ 0 , 1 σ 0 , 1 = σ ^ 0 , 0 σ 0 , 0 σ ^ 1 , 1 σ 1 , 1 cos η ^ θ 0 θ 1 .
By the positive definiteness of Σ , σ 0 , 1 2 σ 0 , 0 σ 1 , 1 . This inequality applied to (57) gives the lemma. □
If we have
γ ^ 0 γ ^ 1 cos η ^ + γ 0 γ 1 1 2 and cos η ^ γ 0 γ 1 1 2 γ ^ 0 γ ^ 1 ,
Equation (56) holds for any pair of angles ( θ 0 , θ 1 ) . This condition is
| cos η ^ | γ 0 γ 1 1 2 γ ^ 0 γ ^ 1 ,
or
| σ ^ 0 , 1 | σ 0 , 0 σ 1 , 1 σ ^ 0 , 0 σ 0 , 0 σ ^ 1 , 1 σ 1 , 1 .
On the other hand, (56) holds for any 0 < η ^ < π (equivalently 1 < cos η ^ < 1 ) if
γ 0 γ 1 1 ( γ 0 1 ) ( γ 1 1 ) cos ( θ 0 + θ 1 ) γ 0 γ 1 + 1 ( γ 0 1 ) ( γ 1 1 ) .
These bounds (58) and (59) mean that the range of feasible GCs ( θ 0 , θ 1 ) and the range of feasible correlation cos η ^ are in a trade-off relationship in general.
Lemma 7.
The VAR with the correlation | cos ( η ^ ) | < 1 and Granger causality θ 0 , θ 1 in the angular form is stable if and only if
sin η ^ + | γ ^ 0 sin η ^ θ 0 + γ ^ 1 sin η ^ θ 1 | < γ ^ 0 γ ^ 1 sin η ^ θ 0 θ 1 < sin η ^ .
Proof. 
Using the angular notation of the solution A with θ 0 , θ 1 , we have
det A = γ ^ 0 γ ^ 1 sin η ^ θ 0 θ 1 sin η ^ and tr A = γ ^ 0 sin η ^ θ 0 + γ ^ 1 sin η ^ θ 1 sin η ^ .
Inserting these into stability condition (23), we have the inequality (60). □
As well as Lemma 6, Lemma 7 leads the trade-off relationship between Σ ^ in the angle from η ^ and A in the angle from θ 0 , θ 1 . In general, this bound further narrows the upper and lower bounds given by Lemma 3. In general, correlation is limited to close to zero if one wishes for higher Granger causality. On the other hand, the two types of Granger causality are limited to close to zero if one wishes for a higher correlation in the absolute value.

6. Concluding Remarks

6.1. Summary and Potential Usage of the Algorithm

In this paper, we explored the relationship between the VAR parameters and timeseries statistics (Figure 1), and identified the trade-off limitation between the stationary covariance Σ ^ and Granger causality G 0 , G 1 (Lemma 6). This suggests that the following Algorithm 1 will generate a timeseries with desired statistics.
Algorithm 1: Compute a VAR parameter set for the desired statistics
Data: Desired timeseries statistics ( G 0 , G 1 , Σ ^ , σ 0 , 0 , σ 1 , 1 ) in the feasible range satisfy both inequalities (56) in Lemma 6 and (60) in Lemma 7.
1 Derive four sets of VAR parameters ( A , Σ ) for the given timeseries statistics ( G 0 , G 1 , Σ ^ , σ 0 , 0 , σ 1 , 1 ) by Lemma 5
2 Choose one of the four sets of VAR parameters ( A , Σ ) .
Result: The VAR parameters ( A , Σ ) .
This data-generation algorithm can be used to generate surrogate data [19], which can be used to test whether an empirical timeseries is a sample from a VAR with a given correlation and Granger causality. This algorithm is also useful in analyzing to what extent a class of VAR timeseries varies under the same statistics.

6.2. Validity of Granger Causality Estimated on Empirical Timeseries

Our analysis also warns that not all Granger causality (or transfer entropy) is “valid”, in the sense that its underlying VAR model is not stable and thus the Granger causality is undefined in theory. In theory, we can identify some value of Granger causality for a finite empirical time series, which is generated by an underlying unstable VAR model without any stationary statistics. Such timeseries statistics will diverge in the long run, but it may be difficult to identify this with a finite empirical timeseries. This asymmetry—namely, that the Granger causality can be calculated numerically but does not guarantee the stability of the underlying VAR—is explicitly demonstrated by Theorem 1. For a given empirical timeseries v = ( v 0 , v 1 , , v T ) R 2 × ( T + 1 ) , one should calculate not just the Granger causality but also its validity by checking (1) A R * 2 × 2 , (2) Σ R + 2 × 2 , and (3) Σ 0 , 1 Σ ^ 0 , 1 , by calculating the maximum likelihood estimator of ( A ( v ) , Σ ( v ) ) , such as
A ( v ) : = V 1 , 0 V 0 , 0 1 and Σ ( v ) = V 1 , 1 V 1 , 0 V 0 , 0 1 V 0 , 1 ,
where
V i , j : = T 1 t = 1 T v t 1 + i v t 1 + j .
In fact, Σ ( v ) is always (semi-)positive definite, as it takes the form of the Schur complement of the (semi-)positive-definite matrix V 0 , 0 V 0 , 1 V 1 , 0 V 1 , 1 . Thus, this maximum likelihood estimator readily satisfies condition (2).

6.3. Future Work

In this paper, we limit the VAR model to be bivariate for simplicity of analysis. We expect it is possible to generalize the current result to any higher dimensional VAR model. In such a generalization, feasible boundaries for the stable VAR models may require further effort to understand.

Author Contributions

Conceptualization, S.H. and T.T.; Methodology, S.H. and T.T.; Software, S.H.; Validation, S.H. and T.T.; Formal Analysis, S.H.; Investigation, S.H.; Writing—Original Draft Preparation, S.H.; Writing—Review and Editing, S.H. amd T.T.; Visualization, S.H.; Supervision, S.H.; Project Administration, S.H.; Funding Acquisition, S.H. Both authors have read and agreed to the published version of the manuscript.

Funding

Japan Society for the Promotion of Science: JP20H04994, Japan Science and Technology Agency: JPMJPR20C9.

Institutional Review Board Statement

No human participants are involved with this study.

Informed Consent Statement

No human participants are involved with this study.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

This work was supported by JSPS KAKENHI Grant Number JP 20H04994 and JST, PRESTO Grant Number JPMJPR20C9, Japan.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Heider, F.; Simmel, M. An experimental study of apparent behavior. Am. J. Psychol. 1944, 57, 243–259. [Google Scholar] [CrossRef]
  2. Bassili, J.N. Temporal and spatial contingencies in the perception of social events. J. Personal. Soc. Psychol. 1976, 33, 680–685. [Google Scholar] [CrossRef]
  3. Dittrich, W.H.; Lea, S.E. Visual perception of intentional motion. Perception 1994, 23, 253–268. [Google Scholar] [CrossRef] [PubMed]
  4. Tremoulet, P.D.; Feldman, J. Perception of animacy from the motion of a single object. Perception 2000, 29, 943–951. [Google Scholar] [CrossRef] [PubMed]
  5. Scholl, B.J.; Tremoulet, P.D. Perceptual causality and animacy. Trends Cogn. Sci. 2000, 4, 299–308. [Google Scholar] [CrossRef]
  6. Takahashi, K.; Watanabe, K. Synchronous motion modulates animacy perception. J. Vis. 2015, 15, 1–17. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Granger, C.W. Investigating causal relations by econometric models and cross-spectral methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
  8. Schreiber, T. Measuring information transfer. Phys. Rev. Lett. 2000, 85, 461–464. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Barnett, L.; Barrett, A.B.; Seth, A.K. Granger causality and transfer entropy are equivalent for Gaussian variables. Phys. Rev. Lett. 2009, 103, 238701. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Eichler, M. Granger causality and path diagrams for multivariate time series. J. Econom. 2007, 137, 334–353. [Google Scholar] [CrossRef]
  11. Freeman, J.R. Granger causality and the times series analysis of political relationships. Am. J. Political Sci. 1983, 27, 327–358. [Google Scholar] [CrossRef]
  12. Joerding, W. Economic growth and defense spending: Granger causality. J. Dev. Econ. 1986, 21, 35–40. [Google Scholar] [CrossRef]
  13. Seth, A.K. Measuring autonomy and emergence via Granger causality. Artif. Life 2010, 16, 179–196. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Okazaki, S.; Hirotani, M.; Koike, T.; Bosch-Bayard, J.; Takahashi, H.K.; Hashiguchi, M.; Sadato, N. Unintentional interpersonal synchronization represented as a reciprocal visuo-postural feedback system: A multivariate autoregressive modeling approach. PLoS ONE 2015, 10, e0137126. [Google Scholar] [CrossRef] [PubMed]
  15. Bressler, S.L.; Seth, A.K. Wiener–Granger causality: A well established methodology. Neuroimage 2011, 58, 323–329. [Google Scholar] [CrossRef] [PubMed]
  16. Ding, M.; Chen, Y.; Bressler, S.L. 17 Granger causality: Basic theory and application to neuroscience. In Handbook of Time Series Analysis: Recent Theoretical Developments and Applications; Wiley Online Library: Hoboken, NJ, USA, 2006; Volume 437. [Google Scholar]
  17. Porta, A.; Faes, L. Wiener–Granger causality in network physiology with applications to cardiovascular control and neuroscience. Proc. IEEE 2015, 104, 282–309. [Google Scholar] [CrossRef]
  18. Seth, A.K.; Barrett, A.B.; Barnett, L. Granger causality analysis in neuroscience and neuroimaging. J. Neurosci. 2015, 35, 3293–3297. [Google Scholar] [CrossRef] [PubMed]
  19. Prichard, D.; Theiler, J. Generating surrogate data for time series with several simultaneously measured variables. Phys. Rev. Lett. 1994, 73, 951. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Schematic diagram of the organization of this paper.
Figure 1. Schematic diagram of the organization of this paper.
Entropy 23 00742 g001
Figure 2. The ellipsis (42) and two parallel lines (45) (forming the parallelogram touching the ellipsis) on the plane ( a 0 , 0 , a 0 , 1 ) R 2 . The solution ( a 0 , a 1 ) is four intersections of these two (depicted by the colored points). Granger causality takes its maximum with the largest | a 0 , 1 | on the ellipsis and its minimum with | a 0 , 1 | = 0 .
Figure 2. The ellipsis (42) and two parallel lines (45) (forming the parallelogram touching the ellipsis) on the plane ( a 0 , 0 , a 0 , 1 ) R 2 . The solution ( a 0 , a 1 ) is four intersections of these two (depicted by the colored points). Granger causality takes its maximum with the largest | a 0 , 1 | on the ellipsis and its minimum with | a 0 , 1 | = 0 .
Entropy 23 00742 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hidaka, S.; Torii, T. Designing Bivariate Auto-Regressive Timeseries with Controlled Granger Causality. Entropy 2021, 23, 742. https://doi.org/10.3390/e23060742

AMA Style

Hidaka S, Torii T. Designing Bivariate Auto-Regressive Timeseries with Controlled Granger Causality. Entropy. 2021; 23(6):742. https://doi.org/10.3390/e23060742

Chicago/Turabian Style

Hidaka, Shohei, and Takuma Torii. 2021. "Designing Bivariate Auto-Regressive Timeseries with Controlled Granger Causality" Entropy 23, no. 6: 742. https://doi.org/10.3390/e23060742

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop