1. Introduction
Consider a multivariate time series
of length
n and dimension
r. It is supposed to be generated by an array process
. There is now a vast body of literature on ARMA and VARMA models with time-dependent or time-varying coefficients. For a recent review of VARMA models, including their time-dependent variants, see [
1].
The history of time-dependent models for time series started with the works of [
2,
3,
4,
5,
6]. Several of them, [
2,
4,
5] and also [
7,
8,
9,
10,
11,
12], focussed on the temporal aspects, while others [
3,
6] were more interested in the spectral point of view.
We will not consider the numerous studies related to the time-dependent spectral approaches, except for the following one that had considerable attraction. We mean the theory based on local stationarity processes (LSP) due to Dahlhaus (see [
13,
14,
15,
16,
17]). We will not repeat here the many contributions about it since they are nicely summarized in [
18], and a few more recent references are mentioned in [
19]. See also [
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30].
Other somewhat related approaches to ours include [
31] (Chapter 17), which treats tdVAR models by Gaussian maximum likelihood (but does not discuss asymptotic properties), generalized autoregressive score (GAS) models of [
32], testing parameter constancy against deterministically time-varying parameters, e.g., ref. [
33] (Section 6.3) and references therein, generalized to VAR models in [
34], smooth online parameter estimation approach [
35], deep-learning approaches [
36], using state-space methods [
37], a non-parametric approach [
38], or by using an explicit representation for ARMA recursions with either deterministically or stochastically varying coefficients [
39].
If a few of the above-mentioned papers treat multivariate time-dependent models, like [
25,
31,
36], they are often tdVAR, and, apparently, never tdVMA and tdVARMA models, with the notable exception of [
38], which is semi-parametric and uses a kernel-density estimator. Ref. [
40] explains that, although in theory a tdMA process can be written as an infinite tdAR process, it is not efficient to fit a high-order model. Since the first edition of [
41] in 1970 and the practical studies that followed, it is well known that most time series in many fields like economics, sociology, tourism, agriculture, energy, and so on, are better fitted by ARIMA models, possibly on transformed data, rather than by AR models. Based on 13,238 monthly series taken from the Industrial Short-Term Indicator section of the EUROSTAT database on 15 Member States of the European Union and a few series from the United States and Japan, ref. [
42] observed that the airline model, i.e., the seasonal ARIMA(0,1,1)(0,1,1
model on the log-transformed data, was best-fitting for 61% of the series, whereas only 2% were best fitted by an autoregression. There is no reason why the primacy of moving averages would not extend to multivariate time series and time-dependent models, although there is no empirical study at this stage. A few of the previously mentioned papers for univariate ARMA models with time-dependent coefficients include MA coefficients like [
7,
10,
11,
12]. In the multivariate case, ref. [
17] treats time-varying or time-dependent VARMA (tdVARMA) models in the context of local stationarity but only for Gaussian processes and under the assumption that all the eigenvalues of the true spectral density matrix are uniformly bounded from below, a condition that is difficult to verify in practice. More recently, ref. [
38] has provided results for a semi-parametric estimator for tdVARMA models.
In a recent paper [
43], estimation results were produced for a wide class of vector ARMA array processes with time-dependent coefficients, denoted tdVARM
, which includes as special cases both local stationary processes [
18] and cyclically time-dependent processes [
11]. The assumptions are rather general but complex at first sight, so it would be worthwhile to demonstrate their applicability. Previously, ref. [
44] has already treated examples for cyclically tdVARMA stochastic processes. Ref. [
43] could also be seen as a generalization to multivariate processes of [
12] devoted to univariate ARMA models with time-dependent coefficients, thereby generalizing the autoregressive moving average (ARMA) models popularized by Box and Jenkins, ref. [
41]. In [
12], the two cases where the coefficients depend only on time
t and both
t and
n were considered with an accent on the former case. The case where the coefficients depend on
t but not series length
n was generalized to VARMA models by [
44]. Ref. [
12] contained very simple univariate examples where the theoretical assumptions for the asymptotic properties were checked. We will see that developing simple examples is much more complex in a multivariate setting.
Previously, ref. [
45] had shown that, when specialized to VARMA models with constant coefficients, these assumptions coincide with the assumptions for the standard asymptotic properties of the parametric estimation for these models. The problem is that the assumptions in [
43] are complex. For instance, that paper contains remarks to address requests from reviewers who could not believe that these assumptions would work, despite the provided proofs. Consequently, we consider here the case of relatively simple array processes and will check analytically a representative sample of the assumptions for bivariate processes. By representative sample, we mean that the other assumptions can be verified using the same arguments. To simplify, we will restrain ourselves to first-order processes, tdVA
(1), when
and
, and tdVM
(1), when
and
. Even for these two special cases, we drastically limit the form of time-dependency and the number
m of parameters. We believe, however, that verifying the assumptions of [
43] in these two cases is exemplary for more complex models. These models were used, for specific values of the parameters, in a simulation study presented by [
43], so that the present paper completes the information. Not only does estimation give results that are predicted by the theory for sufficiently large series, but we will also see that the values selected for the parameters in these simulation experiments fulfill the requirements.
We do not consider high-dimensional models in this paper for two reasons: first, there is little hope to be able to offer simple examples for
; and second, since the size of the matrices in the coefficients will be
, the number of parameters
m increases with
. With
,
, and
, the number of parameters is already
, plus
in the error covariance matrix, leaving only
degrees of freedom. Moreover, the dimension of the information matrix (defined below) grows with
, so in that case, it is already
, possibly implying serious computational problems with matrix inversion. A solution would be to extend the sparse identification and estimation approach proposed for VARMA models by [
46] and implemented by [
47] in the R package
bigtime 0.2.3 for R 3.6.0 and above. This is done by using sparsity-inducing convex regularizers. It works even for large-scale VARMA models, under sufficient regularity conditions expressed by the condition
.
This article is organized as follows. In
Section 2, we introduce the general marginally heteroscedastic tdVARM
array model with the main notations, describe the crucial assumptions under which [
43] have proved the asymptotic properties of the Gaussian maximum likelihood estimator, and present approximations of the true information matrix.
Section 3 contains our results: in
Section 3.1, we consider a tdVA
(1) process and a tdVM
(1) process in
Section 3.2. In both cases, after reducing the number of parameters progressively, we prove analytically that the assumptions can be verified, provide the constraints on the true values, and simulation results with non-Gaussian errors to assess the quality of the estimates and their standard errors.
Section 3.3 and
Section 3.4 are short attempts for generalization but, of course, the model complexity does not permit a complete analytical treatment. The paper ends with a discussion of the results in
Section 4. There are three appendices for details about
Section 3.1 and
Section 3.2.
3. Results
To begin with, we consider here two first-order models, i.e., the cases of the tdVA
(1) model where
and
, on the one hand, and the tdVM
(1) model where
and
, on the other hand. To make it practical, and allow for a fully analytical treatment, we will quickly move to bivariate processes, i.e., the case where
in (
1). Then, we will consider a tdVARM
model for which we will provide an expression for the coefficients
in the moving average representation of the first-order derivatives of the residuals. Finally, for the general case of a tdVARM
model, we will only provide indications on sufficient conditions to demonstrate how it is possible to proceed.
3.1. Treatment of a tdVA(1) Model
In this section, we consider a tdVA
(1) model, first in the general case before taking
. The model is defined by
with
,
, and
having all moments of order
,
, and such that
is bounded. Let also
. Let us define
It can be checked that
,
, and
,
. Note that the coefficients of the pure autoregressive decomposition of the process are
, and
,
.
To be more specific, assume a bivariate process such that the elements of the matrices
are linear functions of time, and the diagonal elements of
are exponential functions of time. More precisely, we suppose that
Remark 1. We have taken linear functions of time for illustrative purposes but it should be clear that the theory works in whole generality. The case of a linear function of time with the divisor appearing in (10) is compatible with Dahlhaus LSP theory. This is why we consider array processes instead of stochastic processes. We will come back to this in the discussion. Now we start examining the typical assumptions of the theory stated in
Section 2.2.
3.1.1. Assumptions (i) and (ii)
Assumption
(i) is clearly satisfied. Denote
for
and
for
. It is obvious that
for all
. This is because
so that we have preferred the denominator
instead of
n used in the theory of locally stationary processes. Using the definition,
are matrices whose Frobenius norms are bounded uniformly in
t and
n, from below by a strictly positive number and also from above, hence Assumption
(ii) is satisfied.
3.1.2. Assumption (iii)
Only to simplify the analytical expressions, assume that
in
, that the element
of
is identically zero, and the element
of
is a constant
. This is to have an upper-triangular form (facilitating the analytical treatment) that does not degenerate into a diagonal matrix (as the latter would imply uncorrelated components since
is diagonal). To simplify further the details (although it is not necessary for the principles), in addition to
, instead of the full vector of parameters
we put
and
is fixed to
, so that the vector of parameters to estimate reduces to
, hence
and there is one parameter of each kind. The presence of
makes sure that the two components of
are not independent. There are assumptions about the true value
that will be stated later. Then, for
, we can check by induction that
and
where
, for
, and
, for
, generalizing relations in [
44] (Appendix S1). To simplify the notations, we will henceforth omit
(n) in the entries of
.
From the definition of
in (
3), we deduce
and
We define the constant
such that
. Therefore, we assume
and
. From (
13), since
, we deduce
, where
. It is more delicate for
for which
The sum of the first term is also bounded by
because
. Thanks to (
12), an upper bound of
equals
hence
, where
(see the details in
Appendix A.1). We could have used [
54] to simplify the investigation.
3.1.3. Assumption (iv)
First, to evaluate the elements
,
, of
V in (
4) with (
5), we have to take the limit for
of the sum for
of
, which can be computed using (
7). But in the tdVA
(1) case, using (
2) with
, it is more simply expressed as
The most difficult case is for
: the product of derivatives and
has the element
equal to
. Assume
. Then it can be seen that
converges to a limit when
. We will now prove that
tends to a limit when
. Denoting
, it is a sum for
to
of matrices
where
. Using the bounds already presented when discussing
,
, and given that
, it is easy to see that the elements of (
16) behave like terms of a geometric series (or terms of two sums of geometric series) for large
k, so that, for large
t, their sum is finite, strictly positive, and convergent when
. The existence of the limit
then follows from Theorem 2.5 of [
55]. The two other cases (
, and
) are straightforward.
The computation of
involves only the second term of the definition in (
5). The derivative of
with respect to
yields a zero matrix, except the element
, which is
. The product of
by that derivative gives a matrix of 0 except for the element
which is equal to
. We have, therefore, to take one-half of the limit for
of the sum for
of
. However, the sum of
is the variance in a discrete uniform distribution over
which equals
. Hence, we obtain
whose limit for
equals
. We can check that the elements not discussed vanish. Note that the factor
appeared already in a similar univariate example shown by [
12] (Example 3).
3.1.4. Assumption (v)
For a Gaussian process,
in (
6). For a Laplace or a Student distribution (the latter with at least 5 degrees of freedom), in particular, the entries of
are not smaller than those for a normal distribution [
51] (Section 4), hence
.
3.1.5. Assumption (vi)
In order to verify Assumption (vi), for example for , the simplest case, we have to show that multiplied by is . First equals , which is bounded by . Then, we take an upper bound of by and thus of by . The sum for of the product is bounded by times a constant . By exchanging the two outside summations, we have to find an upper bound of by times the sum for of a constant . Dividing by , we have . The case where is more delicate and will not be detailed but the principle is identical.
Example 1. In [43] (Section 4.1.2), simulations results were shown for artificial Gaussian time series generated by a specific case of model Equations (8)–(10), under all the restrictions mentioned above and with the following values for . Moreover, is a constant as above. First, note that these values satisfy the requirements. Indeed and . Let us consider 1000 time series of length obtained using multivariate Laplace, on the one hand, and Student with 5 degrees of freedom, on the other hand. The empirical estimation results are shown, respectively, in Table 1 and Table 2. In both cases (multivariate Laplace and multivariate Student), the estimates in column (a)
(emp.est.) are close to the true value, and nearly always closer when n increases; the sample standard errors in column (b)
(emp.s.e.) decrease with n and are very close to the averages across simulations of the estimated standard errors (obtained using the sandwich formula and estimates and of V and W, respectively, see [52] for details) shown in column (c)
(est.s.e.) and the approximate theoretical standard errors in column (d)
(theor.s.e.), also based on a sandwich formula but now on the two finite averages of (4) evaluated at , as discussed in Section 2.3; the percentages of simulations where the hypothesis is rejected at the 5% level in column
(e)
(% rej.) are, of course, close to 5%. The results look better for the Laplace than the Student distribution, especially for column (d)
and row . As a consequence of these simulations, we see that, for n large enough, the estimates become close to the true value, the standard errors that are the by-product of estimation correspond broadly to the empirical results (although less well for the parameter related to heteroscedasticity), coincide relatively well with the approximated values derived from the theory, and, finally, the level of the test that the parameter equals the true value is relatively close to 5%. Thanks to the sandwich correction, the standard errors are not underestimated. For instance, for , , and the Laplace distribution, the average of the standard errors using instead of would lead to instead of as shown in column (c)
and instead of in column (d). These numbers are far from the empirical standard deviation of . Finally, the results in [43] for a normal distribution in an otherwise similar setup are closer to the expected values than those shown here. For instance, the last row shows the results , , , , and in columns (a) to (e), respectively, with a much better agreement for the empirical, estimated, and theoretical standard errors. 3.2. Treatment of a tdVM(1) Model
Moving average models are more difficult to study than autoregressive ones. We consider a tdVM
(1) model defined by
with the same notations as in
Section 2, and
. For any
, the pure autoregressive representation of
is
hence
Replacing
,
by (
17) for
, given that
, we obtain
with
and
, like before. Let us denote
with elements
,
. Hence, from (
3),
Assume now in . To proceed more in detail, in order to simplify the analytic computations (and this is not necessary for numerical computations), we assume also that , a constant.
We can now examine the typical assumptions of the theory stated in
Section 2.2.
3.2.1. Assumptions (i) and (ii)
To obtain nice analytic expressions, we suppose that the diagonal elements of the matrices
and
are exponential functions of time. The elements
and
of
are supposed to be different from zero so that the correlation between the two components of
varies with time. More precisely, using again
, we suppose that
where
and
are here constants, with
and
. Denoting
an element of
and omitting the argument
,
is equal to
It is easy to see that Assumptions (i) and (ii) are verified.
3.2.2. Assumption (iii)
To simplify the example, since a correlation between the two components is already guaranteed by the presence of a non-diagonal covariance matrix
, we will assume that
so that
defined in (
18) is diagonal.
To simplify the discussion, we will further suppose, similarly to what we did in
Section 3.1, in addition of course to the already introduced parameters
and
, we have only one parameter of each type, e.g.,
and
, assuming that
and
are fixed constants, respectively denoted
and
. To summarize,
. Contrary to
Section 3.1, we recourse to the results of [
54] here, although the more lengthy direct analysis is given in
Appendix A.2. Indeed, a sufficient condition for
(iii) is stated in [
54]. It can be shown that the eigenvalues of the MA polynomial are the solutions of the equation
. They should be smaller than 1 in absolute value at
. Therefore, we assume that the true value
of
satisfies
,
. We denote
Assumption
(iii) is, therefore, satisfied. In practice, we don’t assume zero initial values for the process, but that it is invertible before time 1. This leads to additional conditions that
and
. If we were interested in forecasting up to time
, we should replace above
with
to have stronger conditions in (
22).
3.2.3. Assumption (iv)
Now, we consider the existence of
V in (
4) with (
5) for the 4-parameter model described in the previous paragraphs. Let us start with the elements
,
, corresponding to the parameters
and
. The formulas in [
51] (Theorem 3) are applicable for computing the terms
,
. Given the expression of
in (
20), the terms in these formulas are all of the forms
where
is equal to
,
, and
, or
, and
. See
Appendix A.3 for indications on how to prove the existence of the limit of averages of these terms, and even their computation using integration.
Using (
7), the elements of
,
are sums for
to
of terms of the kind
, where
are elements of
and
are elements of
. Given (
21), the special form of
,
is a quadratic polynomial of the diagonal elements
,
. More precisely,
can be written as a finite sum of the form
, with
and
integers such that
,
, and
, and
is a constant. However,
is equal to
. Consequently,
is composed of terms
equal to
. Hence
,
, is a sum for
to
of terms of the kind
where we denote
where
and
are obtained using (
A2) or/and (
A3). The terms in (
24) are not all positive because of
. However,
for all
and all
n, and, therefore,
so that the absolute value of the product
can be bounded by
, given (
22). We want to show that the triangular array
converges to a constant (dependent on
i,
j,
, and
) when
, and hence
. First, the absolute value of the
kth term of (
24) is bounded by
, since
, and the limit of the sum of these terms is a convergent geometric series, ensuring convergence of the sequence (
24) when
. Now, for
,
is a sum of terms that are proportional to
, where
We have defined
as elements of
, so as elements of the matrix of cofactors of the matrix
divided by its determinant
, where
is
. Hence,
is a finite sum with terms that are proportional to
where
and
are integers. We obtain
for
to show that
exists, is finite, and can be evaluated using an integral. Therefore, we can invoke Theorem 2.5 of [
55] to show that
exists and is finite. This proves the existence of
for
.
3.2.4. Assumption (v)
The existence of
W in (
4) with (
6) can be studied similarly using
, depending on the distribution of
, see [
52] (Section 3.5).
3.2.5. Assumption (vi)
It is also easy to check Assumption (vi), for example for
. We have to show that
multiplied by
is
. First, it can be shown that
is equal to
, and this is less than or equal to
. Then, we are exactly in the same situation as in
Section 3.1.
Example 2. In [43] (Section 4.1.2), simulations results were shown for artificial Gaussian or Student (with 5 degrees of freedom) time series generated by model Equations (17)–(20) with the following values for but with a linear expression for that we keep here instead of the exponential that was easier for the analytic developments: and . Moreover, , , and are taken as constants, as well as . First, note that these values satisfy the requirements after a linear approximation of the exponentials. Indeed, and . Let us consider 1000 time series of length obtained using multivariate Laplace distribution. The empirical estimation results are shown in Table 3. In all cases, the estimates in column (a)
(emp.est.) are often close to the true value, and closer when n increases; the sample standard errors in column (b)
(est.s.e.) decrease with n and are very close to the averages across simulations of the estimated standard errors (obtained using the sandwich formula and estimates and of V and W, respectively, see [52] for details) shown in column (c)
(emp.s.e.) and the approximate standard errors in column (d)
(theor.s.e.), also based on a sandwich formula but now on the two finite averages of (4) evaluated at ; the percentages of simulations where the hypothesis is rejected at the 5% level in column (e)
(% rej.) are close to at least for large n. Like in Example 1, for n large enough, the estimates become close to the true value, the estimated standard errors are close to the empirical results (but also for the parameter related to heteroscedasticity) and the approximated values derived from the theory, and, finally, the percentage of rejection in the tests are slightly closer to 5%. Thanks to the sandwich correction, the standard errors are not underestimated. For instance, for and , the average of the standard errors using instead of would lead to instead of shown in column (c)
and instead of in column (d)
, far from the empirical standard error of shown in column (b). Note that the results are slightly better for a normal distribution of the errors. In Table 4, we present the normal results obtained for as we had only shown them for in [43]. The agreement is better for with in row (b)
(emp.s.e.), in row (c1)
(est.s.e.) or in row (c2)
(est.s.e.) or in row (d)
(theor.s.e.). In the normal case, we have computed an estimate of using the integration approach in [55] (Procedure 6). Using this, refer to the details in Appendix A.3, and the corresponding standard error for when is , in full agreement with the empirical, estimated, and theoretical approximation standard errors in the last column of Table 4. 3.3. Treatment of a tdVARM(1,1) Model
Starting from here, we consider only homoscedastic models. Let
be an
r-vector time series satisfying
and
. Hence,
(ii) and
(v) have no object since
does not depend on
and
. We will not cover the tdVARM
model in detail, simply show how the coefficients
can be computed. It is a special case of the model defined in (
1) with
and
. Now, following [
44], in this special case, the coefficients of the pure moving average representation are given by:
where a product for
to
is set to
. The coefficients of the pure autoregressive form are
so, for
, their derivatives are given by
These results correct the findings presented in the univariate case by [
12].
From this expression, it is possible to check (iii) and (vi) relatively easily. Indeed, again let and . If we assume that and , for all t, where , since only a few factors involve sums like and the others are bounded by , the Frobenius norm of the and the are bounded by . Of course, checking (iii), the existence of V, depends heavily on the parametrization. For example, if it exists, V will not be invertible if for all t (or even for most t).
3.4. Treatment of a More General tdVARM Model
Ref. [
45] indicates how to handle more generally homoscedastic tdVARM
(
) models, with
. Indeed, it is shown in [
54] that
(iii) and
(vi) are valid if the determinants of the tdVAR and tdVMA polynomials, respectively,
and
, do not vanish when
. Of course, it is only a sufficient condition. That argument was used in
Section 3.2.2 to simplify the treatment, whereas
Appendix A.2 does not use that argument for the same model and is therefore lengthier. Again,
(ii) and
(v) have no object, while
(i) and
(iii) depend on the specific parametrization.
4. Discussion
The results presented in
Section 3.1 and
Section 3.2 confirm that the theory exposed in [
43] is applicable and that its assumptions can be checked analytically, at least in the two simple bivariate tdVA
(1) and tdVM
(1) models. The results have already been exploited in the simulation experiments in [
43] (Sections 4.1.2 and 4.2.2) as the values of the parameters used there meet the conditions stated here. The treatment of other models by analytical methods is certainly more challenging, although we discussed it in some detail for a tdVARM
(1, 1) model. For more complex models, the approach of [
54] can also be considered, as it allows us to put the model in tdVAR form, albeit with a higher dimension.
As mentioned in Remark 1, we have taken linear functions of time for illustrative purposes, but it should be clear that the theory works in whole generality. First, a linear function of time with the divisor
appearing in (
10) is compatible with Dahlhaus’ LSP theory (see [
16]), since the (Frobenius norm of the) coefficient can easily be bounded from above by 1. Anyway, we can consider that linearity is a first attempt if constancy cannot be retained. Second, in [
19], the authors have shown a univariate tdA
process of order 1, where the coefficient can be greater than 1 during an interval of time that shrinks to 0 as
n increases. We can have the same change here, meaning that the process does not need to be locally stationary. What is essential is that the coefficients of the pure tdMA representation are bounded by a decreasing exponential function of the lag. Third, the tdA
coefficient does not need to be differentiable with respect to time. The theory in [
43] is valid when there are any number of breaks, provided they do not add too many parameters, for instance, if there is a periodicity of 2, with a coefficient equal to
for
t odd, and
for
t even. Indeed, it is a generalization of [
12] where such an example is shown. We can even have a periodic behavior, even with an incommensurable period for the different matrix entries, as shown by [
44].
The lag of one can be replaced with another integer without any change. For instance, for quarterly data, we can replace it with 4, or 12 for monthly data. The results in
Section 3.1 and
Section 3.2 are easily adapted. Because the inference is infill, meaning that more and more observations are assumed to be made between the first and last ones, the LSP theory is no longer valid in the case of periodicity. On the contrary, the inference in [
43,
44] is of type outfill or increasing domain, that is to say, that more years are supposed to be observed, therefore preserving the period of the periodic behavior. Ref. [
19] has also shown a tdA
(1) process with a coefficient that varies linearly with time but where heteroscedasticity (so
) is a periodic function of time. Finally, ref. [
19] indicates that the so-called multiplicative seasonal ARMA models of [
41] cannot be generalized in the LSP framework for the same reason. On the contrary, there is no problem working with those models in our context. That remark made here for the tdVA
(1) model extends to the tdVMA(1) model and all the tdVARM
models, as well.
It is possible to obtain the coefficients of the pure MA representations by simple recursive relations (see [
54] for algorithms). Ref. [
39] has even proposed an explicit solution, although it is limited to univariate processes. With these coefficients, everything can be computed, at least for finite
n, and the assumptions can be checked. Of course, when we have a finite multivariate time series, it is not immediately apparent to suggest a tdVARMA model. What we propose in practice (see the examples treated in [
43,
53]) is first fitting a VARMA model, and then adding slopes for the linearly varying coefficients and, possibly, heteroscedasticity, before removing non-significant parameters one by one. At this time, there is no large-scale study on real data using time-dependent multivariate models. Still, the results for univariate models in [
21] (with marginal heteroscedasticity only) and [
56] seem promising.
We have specified in the introduction that our approach is not adequate for high-dimensional tdVARM
models, i.e., when
r is greater than a very small integer because the number of parameters grows too quickly. Here, we have the entries in the VAR and VMA polynomials but also the coefficients of polynomials if we extend the linear dependence in (
10) or (
20) to a polynomial dependence, for instance. In principle, the concept in [
46] should be valid in our framework of time-dependent VARMA models, including the case of polynomial dependence, multiple breaks, or threshold models with multiple regimes, although
bigtime in [
47] should be strongly changed for that purpose. This is left for future research.
It should be noted that, although it is interesting to see that the assumptions of [
43] can be verified, they are not particularly helpful in the face of a real multivariate time series. We note that, unfortunately, verifying the same assumptions using the data (one realization of length
n) is impossible. The same criticism, however, holds for the entire literature on models with time-dependent coefficients.