Next Article in Journal
Mathematical Modeling and Analytical Solution of Thermoelastic Stability Problem of Functionally Graded Nanocomposite Cylinders within Different Theories
Next Article in Special Issue
Nonlinear Frequency-Modulated Waveforms Modeling and Optimization for Radar Applications
Previous Article in Journal
Transforming Private Pensions: An Actuarial Model to Face Long-Term Costs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Proper ARMA Modeling and Forecasting in the Generalized Segre’s Quaternions Domain

by
Jesús Navarro-Moreno
,
Rosa M. Fernández-Alcalá
* and
Juan C. Ruiz-Molina
Department of Statistics and Operations Research, University of Jaén, Paraje Las Lagunillas, 23071 Jaén, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(7), 1083; https://doi.org/10.3390/math10071083
Submission received: 16 February 2022 / Revised: 23 March 2022 / Accepted: 24 March 2022 / Published: 28 March 2022
(This article belongs to the Special Issue Novel Mathematical Methods in Signal Processing and Its Applications)

Abstract

:
The analysis of time series in 4D commutative hypercomplex algebras is introduced. Firstly, generalized Segre’s quaternion (GSQ) random variables and signals are studied. Then, two concepts of properness are suggested and statistical tests to check if a GSQ random vector is proper or not are proposed. Further, a method to determine in which specific hypercomplex algebra is most likely to achieve, if possible, the properness properties is given. Next, both the linear estimation and prediction problems are studied in the GSQ domain. Finally, ARMA modeling and forecasting for proper GSQ time series are tackled. Experimental results show the superiority of the proposed approach over its counterpart in the Hamilton quaternion domain.

1. Introduction

Time series theory is a very well-known statistical methodology for the analysis of stochastic processes that has found widespread applications in signal processing, communications engineering, economics, astronomy, meteorology, seismology, oceanography, among other fields [1]. The most extensively used class of stationary process models are the so-called autoregressive moving average (ARMA) processes. The relevance of ARMA models stems from their power to approximate any stationary process arbitrarily well [2,3]. In addition, ARMA models can be cast in a state-space framework and also have enormous significance in prediction problems, since they facilitate the derivation of several efficient prediction algorithms [3].
The traditional settings for developing time series analysis have been the real and complex numbers, while other number systems have demonstrated a high potential to outperform them in some applications. The most paradigmatic example is the system of Hamilton quaternions (HQs), which has proven to be of great value for the description of 3-dimensional (3D) movements, in fields such as graphic design or automated control [4,5]. Actually, HQs have become the most popular and researched system of hypercomplex numbers [6]. In spite of its relevance, the use of this system in time series analysis is scarce [7].
HQs belong to a more general family of algebras, named hypercomplex systems, which extends the real and complex numbers to higher dimensions. Among them, 4D hypercomplex systems are of special relevance since four coordinates are usually considered for describing the physical world [6]. Further, depending on how the product is defined, they can be classified into two types: noncommutative and commutative systems. HQs are of the first type, i.e., they form a normed division algebra, and this property makes them well suited to deal with signal processing benchmark problems, such as the signal estimation problem [8]. Though commutative hypercomplex numbers, which includes the generalized Segre’s quaternions (GSQs), are less well known, their interest is increasing within the signal processing community due to their availability to extend the methodologies developed in the complex and real field to the 4D case [9,10,11,12,13,14,15,16,17,18,19,20,21]. These commutative hypercomplex algebras possess zero divisors, which implies that the Euclidean norm is not multiplicative [13]. This drawback, however, has essentially no impact on signal and image processing applications [10].
Interestingly, a recent line of research has discovered that the HQs algebra is not always the best option for processing 4D hypercomplex signals [22,23,24,25]. Specifically, some investigations have revealed that properness, a structural property of the hypercomplex signals, determines the performance of the used processing [24,25]. In all of these papers, the analyses have been performed by comparing HQs with tessarines, and several scenarios where the latter outperforms the former have been shown. It should be noted that tessarines are a subset of GSQs obtained by defining a particular product. In general, hypercomplex algebras are characterized by its product, which is usually provided by a multiplication table (Appendix A in [6]). From an intuitive point of view, it is to be expected that more general commutative hypercomplex algebras than tessarines can attain better performance results. On that basis, special attention is paid to such systems in this paper.
As indicated above, properness is a crucial property of hypercomplex signals. This property has two important effects: it determines which type of processing should be used, and also it achieves a significant decrease in the computational complexity involved in comparison to the real processing. Properness has been investigated in detail in the HQ domain [26,27,28,29,30], and for tessarines [24,25]. To our knowledge, however, properness has not been studied for GSQs yet.
Inspired by all the above discussed, the main objective of this paper is to present the time series analysis in the GSQs domain while analyzing the properness properties in this setting. Specifically, the concept of properness is introduced for GSQ signals and their properties studied. Remarkably, two tests to experimentally check whether real data are proper in the GSQs field are proposed and an estimate of the parameter characterizing the product in the hypercomplex algebra is provided. Such an estimate gives the best value of the parameter under which the properness properties could be fulfilled by the signal of interest. It should be highlighted that, unlike the existing approaches, a specific hypercomplex algebra is not fixed in advance in the proposed strategy, but a broad family of commutative hypercomplex algebras is considered. So, the suggested estimate allows us to select the particular algebra in which it is most likely to achieve, if possible, the properness properties. A simulation example illustrates how this approach is effective.
The linear estimation problem is also tackled. Although GSQ numbers are not a normed algebra and thus, the classical projection theorem cannot be applied to obtain the optimum linear estimate, we exploit its metric space feature to prove the existence and uniqueness of the projection and an extension of the innovations algorithm for proper GSQ signals is devised. As could be expected, the proposed algorithms attain a notable reduction in the computational burden. Finally, ARMA modeling for proper GSQ time series is introduced. For that, we first investigate an extension of the Yule–Walker equations and then, the prediction problem is addressed. The particular structure of the ARMA models allows us to improve the innovations algorithm giving rise to a more efficient prediction method. The steady-state properties of the suggested predictor are also studied. To conclude, the superiority in terms of performance of our proposal in comparison with its counterpart in the HQ domain is numerically analyzed.
The rest of this paper is organized as follows. Section 2 presents both the GSQ random variables and signals. Then, Section 3 introduces the concepts of properness in the GSQs domain and provides the statistical inference methods for testing the properness as well as estimating the associated parameter that characterizes the product in the hypercomplex algebra. This section ends with a simulation example. The linear estimation problem for GSQ random variables and the prediction problem for GSQ random signals are analyzed in Section 4. Section 5 focuses on the analysis of time series in the GSQs setting. It deals with the problems of modeling and forecasting time series data. A final example is included to show the efficiency of the proposed strategy. To improve the readability of the manuscript, all proofs have been deferred to Appendix B, Appendix C, Appendix D, Appendix E and Appendix F.

Notation

We designate matrices with boldface uppercase letters, column vectors with lowercase letters, and scalar quantities with lightface lowercase letters. Moreover, the symbols Z , R , C , and S represent the set of integer, real, complex, and GSQ numbers, respectively. In this regard, A R p × q (respectively, A S p × q ) indicates that A is a real (respectively, a GSQ) matrix of dimension p × q . Similarly r R p (respectively, r S p ) means that r is a p-dimensional real (respectively, a GSQ) vector. In particular, 0 p × q represents the zero matrix of dimension p × q , and I p is the identity matrix of dimension p. In addition, A i j denotes the i j element of the matrix A and diag ( · ) stands for a diagonal (or block diagonal) matrix with entries (or block entries) specified in the operator argument.
Furthermore, superscript “ T ” represents the transpose. “⊗”stands for the Kronecker product. E [ · ] represents the expectation operator. R { · } denotes the real part of a GSQ, and | z | is either the complex modulus if z C or the absolute value if z R . Finally, δ n l denotes the Kronecker delta function.
Unless otherwise stated, all random variables are assumed to have a zero-mean throughout this paper.
A list of symbols that are used in the paper is given in Table 1.

2. Generalized Segre’s Quaternion Random Variables

This section establishes the basic concepts and properties of interest in the GSQ domain. For that, we first introduce the GSQ random variables and then the GSQ random signals are considered.
A random variable x S can be defined as x = a + i b + j c + k d where a , b , c , d R are random variables and the imaginary units ( i , j , k ) satisfy the following multiplication rules:
ij = k ,       jk = i ,       ki = α j ,       i 2 = k 2 = α ,       j 2 = 1
where α is a nonzero real number.
As is well-known, the form in which the product is defined is an essential feature of hypercomplex numbers. The product defined in (1) is commutative, associative, and distributive over addition; however, these algebras have divisors of zero. In contrast to GSQs, the product in the HQs is defined in such a way that the algebra becomes noncommutative and does not have divisors of zero [6]. Further, two special algebras in GSQs are of interest: for α > 0 , S is called the hyperbolic quaternions system and for α < 0 , S is the elliptic quaternions system [6]. In particular, if α = 1 then, S is named the tessarine numbers.
The real vector of x S is defined by x r = [ a , b , c , d ] T . Moreover, we define the following auxiliary GSQ random variables:
x α = a + i b α + j c + k d α   x i = a + i b j c k d x j = a i b + j c k d   x k = a i b j c + k d
The triplet ( x i , x j , x k ) are the so-called principal conjugations and obey the following properties.
Property 1.
Let x , y S be random variables, then
  • ( x η ) η = x ,   η = i , j , k
  • ( x η 1 ) η 2 = x η 3 ,   η 1 , η 2 , η 3 = i , j , k ,   η 1 η 2 η 3
  • x η y η = ( x y ) η ,   η = i , j , k
Consider the vector random variables x S p and y S q . We define the following functions:
Γ x y 1 ( α ) = E [ x y T ] Γ x y 2 ( α ) = E [ x y j T ] Γ x y 3 ( α ) = E [ x y α T ]
where y = [ y 1 , , y q ] T , y j = [ y 1 j , , y q j ] T , and y α = [ y 1 α , , y q α ] T ; however, for reasons that will become clear, they are denoted by Γ x y 1 ( α ) E [ x y H 1 ] , Γ x y 2 ( α ) E [ x y H 1 ] , and Γ x y 3 ( α ) E [ x y H α ] . Likewise, we denote Γ x x ν ( α ) Γ x ν ( α ) , ν = 1 , 2 , 3 .
Now, we extend some of the above concepts to the case of random signals. Consider a random signal vector x ( t ) = [ x 1 ( t ) , , x p ( t ) ] T S p , t Z , where x i ( t ) = a i ( t ) + i b i ( t ) + j c i ( t ) + k d i ( t ) , i = 1 , , p , with a i ( t ) , b i ( t ) , c i ( t ) , and d i ( t ) real random signals. The real vectors associated to x ( t ) is denoted by
x r ( t ) = [ a T ( t ) , b T ( t ) , c T ( t ) , d T ( t ) ] T ,   t Z
where a ( t ) = [ a 1 ( t ) , , a p ( t ) ] T , b ( t ) = [ b 1 ( t ) , , b p ( t ) ] T , c ( t ) = [ c 1 ( t ) , , c p ( t ) ] T , and d ( t ) = [ d 1 ( t ) , , d p ( t ) ] T .
In a similar way to (2), we define the following functions for the random signal vectors x ( t ) S p and y ( t ) S q :
Γ x ν ( α , t , s ) = E [ x ( t ) x H 3 2 ν ( s ) ] ,     Γ x y ν ( α , t , s ) = E [ x ( t ) y H 3 2 ν ( s ) ] ,     ν = 1 , 2 Γ x 3 ( α , t , s ) = E [ x ( t ) x H α ( s ) ] ,     Γ x y 3 ( α , t , s ) = E [ x ( t ) y H α ( s ) ]
It is easy to check, for α 0 , that
Γ x y ν ( α , t , s ) = Γ y x ν H 3 2 ν ( α , s , t ) ,         t , s Z ,         ν = 1 , 2
We define the augmented vector of x ( t ) as
x ¯ ( t ) = [ x T ( t ) , x i T ( t ) , x j T ( t ) , x k T ( t ) ] T ,       t Z
with x η ( t ) = [ x 1 η ( t ) , , x p η ( t ) ] T , for η = i , j , k . The following relationship between the augmented vector and the real vector x r ( t ) given in (3) can be established
x ¯ ( t ) = T p x r ( t )
where T p = B I p with
B = [ 1 i j k 1 i j k 1 i j k 1 i j k ]
Moreover, from Property 1, the next result can be stated.
Property 2.
Γ x ¯ ν ( α , t , s ) = ( Γ x ν ( α , t , s ) Γ x x i ν ( α , t , s ) Γ x x j ν ( α , t , s ) Γ x x k ν ( α , t , s ) Γ x x i ν i ( α , t , s ) Γ x ν i ( α , t , s ) Γ x x k ν i ( α , t , s ) Γ x x j ν i ( α , t , s ) Γ x x j ν j ( α , t , s ) Γ x x k ν j ( α , t , s ) Γ x ν j ( α , t , s ) Γ x x i ν j ( α , t , s ) Γ x x k ν k ( α , t , s ) Γ x x j ν k ( α , t , s ) Γ x x i ν k ( α , t , s ) Γ x ν k ( α , t , s ) )
for α 0 , t , s Z and ν = 1 , 2 , 3 . The functions Γ x ν ( α , t , s ) , Γ x x i ν ( α , t , s ) , Γ x x j ν ( α , t , s ) and Γ x x k ν ( α , t , s ) , ν = 1 , 2 , 3 , are obtained in Table A1 (see Appendix A) in function of the correlation functions (denoted by Γ ) of the components of the real vector (3).

3. Properness Conditions

This section is devoted to analyze the concepts of properness for GSQ random signals. After introducing the notion of properness, a characterization of this property based on the correlations of the real vectors is presented.
Definition 1.
A random signal vector x ( t ) S p is said to be H α 0 -proper if, and only if, there exists a value α 0 > 0 such that the functions Γ x x η 1 ( α 0 , t , s ) , η = i , j , k , vanish t , s Z .
In a similar manner, two random signal vectors x ( t ) S p and y ( t ) S q are cross H α 0 -proper if, and only if, there exists a value α 0 > 0 such that the functions Γ x y η 1 ( α 0 , t , s ) , η = i , j , k , vanish t , s Z . Finally, x ( t ) and y ( t ) are jointly H α 0 -proper if, and only if, are H α 0 -proper and cross H α 0 -proper.
A similar definition about E α 0 -properness can be given by relplacing α 0 > 0 by α 0 < 0 , Γ x x η 1 ( α 0 , t , s ) by Γ x x η 2 ( α 0 , t , s ) , and Γ x y η 1 ( α 0 , t , s ) by Γ x y η 2 ( α 0 , t , s ) .
Remark 1.
The T 1 -properness condition introduced in [24] is a particular case of E α 0 -properness for α 0 = 1 .
From Table A1, we have the following characterization result.
Proposition 1.
Let x ( t ) S p be a random signal vector. Then
  • x ( t ) is H α 0 -proper if, and only if, the correlation of their real vectors verifies
    Γ a ( t , s ) = Γ c ( t , s ) = α 0 Γ d ( t , s ) ,   Γ b ( t , s ) = Γ d ( t , s ) Γ a b ( t , s ) = Γ b a ( t , s ) = Γ c d ( t , s ) = Γ d c ( t , s ) Γ a c ( t , s ) = Γ c a ( t , s ) = α 0 Γ d b ( t , s ) ,   Γ b d ( t , s ) = Γ d b ( t , s ) Γ a d ( t , s ) = Γ b c ( t , s ) = Γ c b ( t , s ) = Γ d a ( t , s )
    with α 0 > 0 .
  • x ( t ) is E α 0 -proper if, and only if, the correlation of their real vectors verifies
    Γ a ( t , s ) = Γ c ( t , s ) = α 0 Γ d ( t , s ) ,   Γ b ( t , s ) = Γ d ( t , s ) Γ a b ( t , s ) = Γ b a ( t , s ) = Γ c d ( t , s ) = Γ d c ( t , s ) Γ a c ( t , s ) = Γ c a ( t , s ) = α 0 Γ d b ( t , s ) ,   Γ b d ( t , s ) = Γ d b ( t , s ) Γ a d ( t , s ) = Γ b c ( t , s ) = Γ c b ( t , s ) = Γ d a ( t , s )
    with α 0 < 0 .
From an experimental standpoint, it becomes necessary to develop a method for checking whether a random vector x S p is proper and devise an estimate of the associated α 0 . Specifically, based on the information supplied by a random sample, we aim to test if x is H α 0 -proper (or E α 0 -proper) and we wish to estimate the suitable value of α 0 under which such properness properties could be satisfied by x , if possible. To this end, we consider the following statistical hypotheses test:
H 0 : α 0 such that x is H α 0 - proper ( respectively , x is E α 0 - proper ) H 1 : α 0 such that x is H α 0 - proper ( respectively , x is E α 0 - proper )
The following result derives two generalized likelihood ratio tests (GLRT), Wilks’ statistics, for the problem (8). These statistics are denoted by ϕ ν ( x 1 , , x n ) . The value assigned to the letter “ ν ” as subscript, superscript or in any other role will indicate when H α 0 -properness ( ν = 1 ) or E α 0 -properness ( ν = 2 ) should be assumed. This convention is applied henceforth. Notice that, these tests generalize the one given in ([24], Theorem 1) to the GSQs domain. The main difference between them lies in the choice of the appropriate α 0 to achieve properness conditions.
Theorem 1.
Given n independent and identically distributed random samples x 1 , , x n of a random vector x S p such that x r follows a Gaussian distribution then, the GLRT statistics for (8) are given by
ϕ ν ( x 1 , , x n ) = n [ ln ( det ( Γ ^ x ¯ ν ( α 0 ) ) α 0 ) ln ( det ( Γ S ν ( α 0 ) ) α 0 ) ] ,         ν = 1 , 2
where det ( · ) α 0 denotes the determinant under the product associated to α 0 ,
Γ ^ x ¯ ν ( α 0 ) = T p Γ ^ x r T p H 3 2 ν ,         ν = 1 , 2
Γ ^ x r is the sample autocorrelation matrix and
Γ S ν ( α 0 ) = diag ( Γ ^ x ν ( α 0 ) , Γ ^ x ν i ( α 0 ) , Γ ^ x ν j ( α 0 ) , Γ ^ x ν k ( α 0 ) ) ,     ν = 1 , 2
The parameter α 0 can be estimated as
α 0 = max α , α > 0 [ c 1 n 2 [ ln ( det ( Γ S 1 ( α ) ) α ) + 4 p ] ] ,           ν = 1 α 0 = max α , α < 0 [ c 1 n 2 [ ln ( det ( Γ S 2 ( α ) ) α ) + 4 p ] ] ,         ν = 2
with c 1 = 2 n p ln ( 2 π ) + n p ln ( 16 | α | ) .
Further, assuming that H 0 is true, the distributions of the statistics ϕ 1 ( x 1 , , x n ) and ϕ 2 ( x 1 , , x n ) tend to a chi-squared distribution with degrees of freedom equal to 6 p 2 1 and 2 p ( 3 p + 1 ) 1 , respectively, as the sample size tends to infinity.

Example

Consider a random vector x = [ x 1 , x 2 ] T S 2 such that the real vectors of x 1 and x 2 given by (3) verify
Γ x i r = ( | α 0 | 6 λ | α 0 | 0.4 ν λ 6 μ 0.4 | α 0 | 0.4 μ | α 0 | 6 λ μ 0.4 λ 6 ) ,       i = 1 , 2
and
Γ x 1 r x 2 r = ( 2 ( 1 ) ν + 1 0.5 1.2 ( 1 ) ν + 1 0.4 0.5 1 0.4 0.6 1.2 ( 1 ) ν + 1 0.4 2 ( 1 ) ν + 1 0.5 0.4 0.6 0.5 1 )
The following two cases are studied:
  • Case 1: λ = 0.7 and μ = 0.8 in an hyperbolic quaternions algebra with α 0 = 2 . From Proposition 1, we have that the vector x is H 2 -proper under these conditions. Moreover, from Theorem 1, the GLRT statistic ϕ 1 ( x 1 , , x n ) converges to a χ 2 ( 23 ) .
  • Case 2: λ = μ = 0 in an elliptic quaternions algebra with α 0 = 2 . It follows that x is E 2 -proper. Further, ϕ 2 ( x 1 , , x n ) converges to a χ 2 ( 27 ) .
On the one hand, the behavior of the cumulative distribution functions (CDFs) of ϕ ν ( x 1 , , x n ) , ν = 1 , 2 , is analyzed in both cases in terms of the sample size. For that, and through Montecarlo simulations, 2000 values of the statistics ϕ ν ( x 1 , , x n ) , ν = 1 , 2 , have been generated for n = 50 , 100 , 300 , 500 . The corresponding empirical CDFs are displayed in Figure 1. At a glance, this figure shows the rapid convergence of the empirical CDFs to the target chi-squared distributions.
On the other hand, we assess the accuracy of the estimations of α 0 depending on the sample size n. The maximization process in (9) can be achieved using the optimization toolbox in Matlab (The Nelder–Mead search algorithm is used). Table 2 and Table 3 depict the 95% confidence intervals for α 0 obtained from 2000 simulated runs. As can be expected, the amplitudes of the confidence intervals become narrower as the sample size grows.

4. Linear Estimation and Prediction

The objective of this section is to solve the linear estimation problem for proper GSQ random variables. Since it is not possible to define a norm in the set of GSQs then, we cannot apply the projection theorem to obtain the optimum linear estimator. So, we first need to define a suitable metric and thus, we prove the existence and the uniqueness of the projection.
Given the random variables x , y S , we define the product
x , y α = E [ x y α ] = Γ x y 3 ( α )
where Γ x y 3 ( α ) is the scalar version of Γ x y 3 ( α ) given in (2).
For any random variables x , y , u , v S and deterministic coefficients λ , β S , γ , δ R , it follows that
  • λ x + β y , γ u + δ v α = λ γ x , u α + λ δ x , v α + β γ y , u α + β δ y , v α .
  • x , x α = 0 x = 0 .
Thus, we can define a distance as
d α ( x , y ) = x y α
where x α 2 = R { Γ x 3 ( α ) } .
Consider the set of random vectors C α = { x 1 , , x m } , with x i S p , i = 1 , , m , and denote by G C α the linear span of C α , i.e., the set of all finite linear combinations of elements of C α . The issue of existence and uniqueness of the projection of a given random variable y S onto the set G C α is now briefly investigated. This projection is denoted by y ^ α .
In a metric space, neither the existence nor the uniqueness of y ^ α is assured; however, following a reasoning similar to [24], the existence and uniqueness of y ^ α can be proved. Further, y ^ α is the projection of y onto the set G C α if, and only if, y y ^ α , x α = 0 , x G C α . As a consequence, if { ϑ 1 , , ϑ r } is a basis of G C α then,
y ^ α = i = 1 r h i ϑ i
where the deterministic coefficients h i S are obtained from the system
y , ϑ j α = i = 1 r h i ϑ i , ϑ j α ,         j = 1 , , r
Now, we are prepared to introduce the notion of linear estimator in the GSQ domain.
Definition 2.
The projection of y onto the set G C α , y ^ α , is called the S α estimator of y respect to C α and, from now on, it is denoted by y ^ α S . If the information set is formed by augmented vectors defined in (5), i.e., if we consider C ¯ α = { x ¯ 1 , , x ¯ m } with x ¯ i S 4 p , i = 1 , , m , then, the projection of y onto the set G C ¯ α is called SWL estimator of y respect to C α and it is denoted by y ^ α S W L .
These concepts are easily extended to the vectorial case. For example, the S α estimator of the random vector y S q respect to C α is y ^ α S = [ y ^ 1 α S , , y ^ q α S ] , where y ^ j α S is the projection of the component jth of y onto the set G C α .
Theorem 2.
Consider a random variable y S and the set of random vectors C α = { x 1 , , x m } , with x i S p , i = 1 , , m . Then,
  • The SWL estimator of y respect to C α is obtained as
    y ^ α S W L = i = 1 m ( h 1 i T ( α ) x i + h 2 i T ( α ) x i i + h 3 i T ( α ) x i j + h 4 i T ( α ) x i k )
    where the deterministic vectors h j i ( α ) S p are computed through the equation
    [ h 11 T ( α ) , , h 1 m T ( α ) , h 21 T ( α ) , , h 2 m T ( α ) , h 31 T ( α ) , , h 3 m T ( α ) , h 41 T ( α ) , , h 4 m T ( α ) ] = Γ y z ¯ 3 ( α ) Γ z ¯ 3 1 ( α )
    with z ¯ = [ x 1 T , , x m T , x 1 i T , , x m i T , x 1 j T , , x m j T , x 1 k T , , x m k T ] T .
    Moreover, this estimator and its associated error are independent of the value of α, i.e., y ^ S W L = y ^ α S W L and ϵ S W L = ϵ α S W L , α 0 , where ϵ α S W L = y y ^ α S W L α 2 .
  • The S α estimator of y respect to C α is calculated as
    y ^ α S = i = 1 m g i T ( α ) x i
    where the deterministic vectors g i ( α ) S p are computed through the equation
    [ g 1 T ( α ) , , g m T ( α ) ] = Γ y z 3 ( α ) Γ z 3 1 ( α )
    with
    z = [ x 1 T , , x m T ] T
  • ϵ S W L ϵ α S = y y ^ α S α 2 , α 0 .
  • If x 1 , , x m are jointly H α 0 -proper (respectively, E α 0 -proper) and y is cross H α 0 -proper (respectively, E α 0 -proper) with each element of { x 1 , , x m } then, y ^ S W L = y ^ α 0 S .
Remark 2.
Theorem 2 shows that whatever hypercomplex algebra you choose provides the same SWL estimator. Likewise, the computational burden of the SWL estimator can be notably reduced whenever hyperbolic or elliptic conditions of properness are fulfilled (compare (13) with (14)). Effectively, while the computational complexity of SWL estimators is of order O ( 64 p 3 m 3 ) , this is of order O ( p 3 m 3 ) under hyperbolic or elliptic properness. More importantly, this reduction in computational complexity cannot be achieved in the real numbers domain.
Now, we are concerned with the prediction problem for random signals. Consider a random signal vector x ( t ) S p . Our aim is to obtain the S α one-stage predictor of x ( t + 1 ) from the information contained in the set C α t = { x ( 1 ) , , x ( t ) } , denoted by x ^ α S ( t + 1 | t ) . Analogously, the SWL one-stage predictor is denoted by x ^ SWL ( t + 1 | t ) .
Note that a similar result to Theorem 2 can be stated for random signals, but we do not include it here for the sake of brevity. It should be remarked, however, that a computational problem remains still unsolved. Although the estimation problem is indeed simplified due to the reduction in dimension implied by the properness conditions, we have to cope with the inverse in (14) as the number of observations t grows. The following result overcomes this problem for proper signals.
Theorem 3
(The innovations algorithm). Consider a H α 0 -proper (respectively, E α 0 -proper) random signal vector x ( t ) S p . If Γ x ν ( α 0 , t , s ) is nonsingular for ν = 1 , 2 then, x ^ S W L ( t + 1 | t ) = x ^ α 0 S ( t + 1 | t ) and can be recursively obtained in the following way:
x ^ α 0 S ( t + 1 | t ) = i = 1 t Λ ( t , i ) ( x ( t + 1 i ) x ^ α 0 S ( t + 1 i | t i ) )   t 1
with x ^ α 0 S ( 1 | 0 ) = 0 p × 1 and where
Λ ( t , t k ) = ( Γ x ν ( α 0 , t + 1 , k + 1 ) i = 0 k 1 Λ ( t , t i ) Π ( i ) Λ H 3 2 ν ( k , k i ) ) Π 1 ( k )
for k = 1 , , t 1 , with Λ ( t , t ) = Γ x ν ( α 0 , t + 1 , 1 ) Π 1 ( 0 ) and
Π ( t ) = Γ x ν ( α 0 , t + 1 , t + 1 ) i = 0 t 1 Λ ( t , t i ) Π ( i ) Λ H 3 2 ν ( t , t i )
with the initial condition Π ( 0 ) = Γ x ν ( α 0 , 1 , 1 ) . Moreover, the error associated to the component x j ( t + 1 ) is given by
ϵ j S W L ( t + 1 | t ) = 2 | α 0 | + 2 4 | α 0 | R { Π j j ( t ) }

5. Proper ARMA Models

This section constitutes the core of the paper. Starting from the definition of proper ARMA models in the GSQs domain we then solve the one-stage prediction problem.

5.1. ARMA Modeling

Next, we adapt the proper ARMA signal concept to the GSQs domain and extend the Yule–Walker equations to this kind of signals. To do so, we first extend the concept of stationarity to the GSQs domain in the proper signal scenario.
Definition 3.
A H α 0 -proper random signal vector x ( t ) S p is said to be stationary if Γ x 1 ( α 0 , t + h , t ) = Γ x 1 ( α 0 , s + h , s ) , t , s , h Z , i.e., Γ x 1 only depends on the lag h for a given α 0 . In this case, we denote Γ x 1 ( α 0 , h ) = Γ x 1 ( α 0 , t + h , t ) . The concept of a E α 0 -proper stationary signal is defined analogously by replacing Γ x 1 by Γ x 2 .
Definition 4.
A random signal vector x ( t ) S p is said to be a H α 0 -proper ARMA signal of orders p and q (ARMA( p , q )) if it is H α 0 -proper, stationary, and can be modeled by the following system:
x ( t ) = i = 1 p F ( i ) x ( t i ) + i = 0 q G ( i ) v ( t i )
where G ( 0 ) = I p , F ( i ) , G ( i ) S p × p are deterministic matrices and v ( t ) is a H α 0 -proper white noise verifying
Γ v 1 ( α 0 , t , s ) = Ω δ t s
with Ω S p × p a deterministic matrix.
The concept of a E α 0 -proper ARMA( p , q ) signal is defined analogously by replacing H α 0 -proper by E α 0 -proper, and Γ v 1 by Γ v 2 in (21).
The following result extends the Yule–Walker equations to the case of proper ARMA signals.
Proposition 2.
Let x ( t ) S p be a H α 0 -proper (or E α 0 -proper) ARMA( p , q ) signal. Consider the matrix-valued polynomial F ( z ) = I 4 p i = 1 p F ¯ ( i ) z i with F ¯ ( i ) = diag ( F ( i ) , F i ( i ) , F j ( i ) , F k ( i ) ) , i = 1 , , p . If det ( F ( z ) ) α 0 0 , z C such that | z | 1 then, Γ x ν ( α 0 , h ) , ν = 1 , 2 , obeys the equation
Γ x ν ( α 0 , h ) i = 1 p F ( i ) Γ x ν ( α 0 , h i ) = h i q G ( i ) Ω L H 3 2 ν ( i h ) ,         h 0
where the matrices L ( t ) are determined by
L ( 0 ) = I p ,     L ( t ) = 0 p × p ,         t < 0 L ( t ) = i = 1 t F ( i ) L ( t i ) + G ( t ) ,         t 1
with F ( t ) = 0 p × p , t > p and G ( t ) = 0 p × p , t > q .

5.2. One-Stage Prediction

We aim to improve the recursive algorithm for x ^ α 0 S ( t + 1 | t ) given in Theorem 3 by making use of the particular structure of proper ARMA( p , q ) models.
Theorem 4.
Let x ( t ) S p be a H α 0 -proper (or E α 0 -proper) ARMA( p , q ) signal. If det ( F ( z ) ) α 0 0 , z C such that | z | 1 then, x ^ S W L ( t + 1 | t ) = x ^ α 0 S ( t + 1 | t ) and this predictor can be calculated through the following recursive procedure:
x ^ α 0 S ( t + 1 | t ) = { i = 1 t Λ ( t , i ) ( x ( t + 1 i ) x ^ α 0 S ( t + 1 i | t i ) ) ,   1 t < m i = 1 p F ( i ) x ( t + 1 i ) + i = 1 q Λ ( t , i ) ( x ( t + 1 i ) x ^ α 0 S ( t + 1 i | t i ) ) ,   t m
with x ^ α 0 S ( 1 | 0 ) = 0 p × 1 , m = max ( p , q ) and the matrices Λ ( t , i ) are calculated by replacing Γ x ν in (17) and (18) by the following function:
Υ ν ( t , s ) = { Γ x ν ( α 0 , t s ) ,     1 t s m Γ x ν ( α 0 , t s ) i = 1 p Γ x ν ( α 0 , t + i s ) F H 3 2 ν ( i ) ,     1 t m s 2 m 0 p × p ,     1 t m , s > 2 m i = 0 q G ( i ) Ω G H 3 2 ν ( i + s t ) ,       m < t s t + q 0 p × p ,     m < t , t + q < s Υ ν H 3 2 ν ( s , t ) ,     s < t
with G ( i ) = 0 p × p for i > q , Ω given in (21), and Γ x ν ( α 0 , t s ) obeys (22). Moreover, the error associated to (23) coincides with (19).
Remark 3.
The advantage of working with this algorithm is that Υ ν ( t , s ) = 0 p × p for | s t | > q and thus, unlike the innovations algorithm given in Theorem 3, only Λ ( t , i ) , for i = 1 , , q , have to be computed in each recursion.
Corollary 1.
For an ARMA( 1 , 1 ) signal it follows that
Υ ν ( t , s ) = { Γ x ν ( α 0 , 0 ) ,     t = s = 1 Ω G H 3 2 ν ( 1 ) , 1 t , s = t + 1 Ω + G ( 1 ) Ω G H 3 2 ν ( 1 ) ,     1 < t = s 0 p × p ,     1 t , t + 1 < s Υ ν H 3 2 ν ( s , t ) ,     s < t ,         ν = 1 , 2
and the one-stage predictor is computed through
x ^ α 0 S ( t + 1 | t ) = F ( 1 ) x ( t ) + Λ ( t , 1 ) ( x ( t ) x ^ α 0 S ( t | t 1 ) ) ,     t 1
where
Λ ( t , 1 ) = G ( 1 ) Ω Π 1 ( t 1 ) Π ( t ) = Ω + G ( 1 ) Ω G H 3 2 ν ( 1 ) Λ ( t , 1 ) Π ( t 1 ) Λ H 3 2 ν ( t , 1 )
and Π ( 0 ) = Γ x ν ( α 0 , 0 ) is computed solving the following system of matrix equations:
Γ x ν ( α 0 , 0 ) F ( 1 ) Γ x ν H 3 2 ν ( α 0 , 1 ) = Ω + G ( 1 ) Ω ( F H 3 2 ν ( 1 ) + G H 3 2 ν ( 1 ) ) Γ x ν ( α 0 , 1 ) F ( 1 ) Γ x ν ( α 0 , 0 ) = G ( 1 ) Ω
To conclude, a steady-state performance analysis for the one-stage predictor (23) and its associated error is fulfilled.
Corollary 2.
Under conditions of Theorem 4, if G ( z ) given in (A4) satisfies det ( G ( z ) ) α 0 0 , z C such that | z | 1 , then,
Λ ( t , i ) t G ( i ) ,     ϵ j S W L ( t + 1 | t ) t ϵ j
with
ϵ j = 2 | α 0 | + 2 4 | α 0 | R { Ω j j }
being Ω j j the jth diagonal entry of Ω given in (21).

5.3. Example

By way of illustration, we consider a proper ARMA(1,1) model. With this model, we first carry out a performance comparative analysis between the S α 0 one-stage predictor (25), and its counterpart in the HQ domain, x ^ α 0 Q S L ( t + 1 | t ) . Our aim is to show the superiority in terms of performance of the proposed approach respect to the usual HQ methodology under different initial conditions. Finally, a steady-state performance analysis is carried out.
Let the ARMA(1,1) system be
x ( t ) = f x ( t 1 ) + v ( t ) + g v ( t 1 )
with f = 0.3 0.05 j + 0.1 k , g = 0.1 + 0.2 j + 0.4 k and v ( t ) a white noise whose real vector (3) verifies Γ v r ( t , s ) = Ω δ t s where Ω is the matrix given in (10).
Two scenarios are contemplated: in the first one, we consider that λ = 0.7 and μ = 0.8 in Ω , and three different hyperbolic quaternion algebras are studied ( α 0 = 1 , 2 , 3 ). In the second one, we set the value of the parameters λ and μ to be both 0 and three elliptic quaternion algebras are analyzed ( α 0 = 1 , 2 , 3 ). These scenarios present an interesting characteristic: from Proposition 1, x ( t ) is either H α 0 -proper (in scenario 1) or E α 0 -proper (in scenario 2); however, the degree of improperness of x ( t ) in the HQ system increases as | α 0 | grows.
Note that the conditions of Proposition 2 and Corollary 2 are fulfilled and hence, x ^ α 0 S ( t + 1 | t ) can be obtained through the recursions (25) and (26) and the steady-state results in (27) hold.
The errors associated to x ^ α 0 S ( t + 1 | t ) and x ^ α 0 QSL ( t + 1 | t ) depend on the value α 0 and thus, they are denoted by ϵ α 0 SWL ( t + 1 | t ) and ϵ α 0 QSL ( t + 1 | t ) , respectively. To assess the performance of both approaches, the following measure is computed:
D P α 0 ( t ) = ϵ α 0 QSL ( t + 1 | t ) ϵ α 0 SWL ( t + 1 | t )
Figure 2 depicts the measure D P α 0 ( t ) in (29), for t = 1 , , 100 , in the above two scenarios. As can be seen, the S α 0 one-stage predictor outperforms the HQ one-stage predictor in both scenarios. Furthermore, the measure becomes greater as | α 0 | grows, i.e., as the degree of HQ improperness grows; being D P α 0 ( t ) slightly greater for the E α 0 -proper case.
Now, we analyze the steady-state performance of both the estimate (25) and its error. The errors (28) computed for different values of α 0 appear in Table 4. Further, Figure 3 shows as ϵ α 0 SWL ( t + 1 | t ) converges toward their respective steady-state errors when t grows. These plots indicate that the convergence is rapid.
Finally, Figure 4 displays the convergence performance of the components of Λ ( t , 1 ) = Λ 1 ( t , 1 ) + Λ 2 ( t , 1 ) i + Λ 3 ( t , 1 ) j + Λ 4 ( t , 1 ) k given in (26). As could be expected from (27), we also obtain here a rapid convergence toward the target values and, in general, this convergence is slightly slower as | α 0 | grows as well as in the H α 0 -proper case.

6. Discussion

Recent investigations have shown that HQs are not always the best choice for processing 4D hypercomplex signals. These results suggest distinct strategies depending on whether properness conditions are complied with. As the main characteristic, they offer a significant computational saving in relation to the real-valued processing. Unlike earlier research, the approach described in this paper has the additional advantage of the flexibility in the choice of the hypercomplex algebra. In our opinion, this research topic opens a broad range of possibilities for studying new applications in signal processing. This paper precisely proposes the time series analysis as a possible field of application of the proper GSQ signal processing. Other potential areas include dynamic linear models, filtering and smoothing, signal detection, etc.

Author Contributions

Conceptualization, J.C.R.-M.; Formal analysis, J.N.-M.; Methodology, J.N.-M.; Investigation, J.N.-M., R.M.F.-A. and J.C.R.-M.; Visualization, J.N.-M., R.M.F.-A. and J.C.R.-M.; Writing—original draft preparation, J.N.-M.; Writing—review and editing, J.N.-M., R.M.F.-A. and J.C.R.-M.; Funding acquisition, J.N.-M. and R.M.F.-A.; Project administration, J.N.-M. and R.M.F.-A.; Software, J.N.-M.; Supervision, J.N.-M., R.M.F.-A. and J.C.R.-M.; Validation, J.N.-M., R.M.F.-A. and J.C.R.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported in part by I+D+i project with reference number 1256911, under ‘Programa Operativo FEDER Andalucía 2014–2020’, Junta de Andalucía, and Project EI-FQM2-2021 of ‘Plan de Apoyo a la Investigación 2021–2022’ of the University of Jaén.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study: in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Table A1. Expressions for Γ x ν ( α , t , s ) and Γ x x k ν ( α , t , s ) , ν = 1 , 2 , 3 , in function of the correlation functions of the components of x r ( t ) (3).
Table A1. Expressions for Γ x ν ( α , t , s ) and Γ x x k ν ( α , t , s ) , ν = 1 , 2 , 3 , in function of the correlation functions of the components of x r ( t ) (3).
Γ x 3 ( α , t , s ) = Γ a ( t , s ) + Γ b ( t , s ) + Γ c ( t , s ) + Γ d ( t , s ) + i ( 1 α Γ a b ( t , s ) + Γ b a ( t , s ) + 1 α Γ c d ( t , s ) + Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) + Γ b d ( t , s ) + Γ c a ( t , s ) + Γ d b ( t , s ) ) + k ( 1 α Γ a d ( t , s ) + Γ b c ( t , s ) + 1 α Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x 1 ( α , t , s ) = Γ x x j 2 ( α , t , s ) = Γ a ( t , s ) + α Γ b ( t , s ) + Γ c ( t , s ) + α Γ d ( t , s ) + i ( Γ a b ( t , s ) + Γ b a ( t , s ) + Γ c d ( t , s ) + Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) + α Γ b d ( t , s ) + Γ c a ( t , s ) + α Γ d b ( t , s ) ) + k ( Γ a d ( t , s ) + Γ b c ( t , s ) + Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x i 3 ( α , t , s ) = Γ a ( t , s ) + Γ b ( t , s ) Γ c ( t , s ) Γ d ( t , s ) + i ( 1 α Γ a b ( t , s ) + Γ b a ( t , s ) 1 α Γ c d ( t , s ) Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) Γ b d ( t , s ) + Γ c a ( t , s ) + Γ d b ( t , s ) ) + k ( 1 α Γ a d ( t , s ) Γ b c ( t , s ) + 1 α Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x i 1 ( α , t , s ) = Γ x x k 2 ( α , t , s ) = Γ a ( t , s ) + α Γ b ( t , s ) Γ c ( t , s ) α Γ d ( t , s ) + i ( Γ a b ( t , s ) + Γ b a ( t , s ) Γ c d ( t , s ) Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) α Γ b d ( t , s ) + Γ c a ( t , s ) + α Γ d b ( t , s ) ) + k ( Γ a d ( t , s ) Γ b c ( t , s ) + Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x j 3 ( α , t , s ) = Γ a ( t , s ) Γ b ( t , s ) + Γ c ( t , s ) Γ d ( t , s ) + i ( 1 α Γ a b ( t , s ) + Γ b a ( t , s ) 1 α Γ c d ( t , s ) + Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) Γ b d ( t , s ) + Γ c a ( t , s ) Γ d b ( t , s ) ) + k ( 1 α Γ a d ( t , s ) + Γ b c ( t , s ) 1 α Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x j 1 ( α , t , s ) = Γ x 2 ( α , t , s ) = Γ a ( t , s ) α Γ b ( t , s ) + Γ c ( t , s ) α Γ d ( t , s ) + i ( Γ a b ( t , s ) + Γ b a ( t , s ) Γ c d ( t , s ) + Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) α Γ b d ( t , s ) + Γ c a ( t , s ) α Γ d b ( t , s ) ) + k ( Γ a d ( t , s ) + Γ b c ( t , s ) Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x k 3 ( α , t , s ) = Γ a ( t , s ) Γ b ( t , s ) Γ c ( t , s ) + Γ d ( t , s ) + i ( 1 α Γ a b ( t , s ) + Γ b a ( t , s ) + 1 α Γ c d ( t , s ) Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) + Γ b d ( t , s ) + Γ c a ( t , s ) Γ d b ( t , s ) ) + k ( 1 α Γ a d ( t , s ) Γ b c ( t , s ) 1 α Γ c b ( t , s ) + Γ d a ( t , s ) )
Γ x x k 1 ( α , t , s ) = Γ x x i 2 ( α , t , s ) = Γ a ( t , s ) α Γ b ( t , s ) Γ c ( t , s ) + α Γ d ( t , s ) + i ( Γ a b ( t , s ) + Γ b a ( t , s ) + Γ c d ( t , s ) Γ d c ( t , s ) ) + j ( Γ a c ( t , s ) + α Γ b d ( t , s ) + Γ c a ( t , s ) α Γ d b ( t , s ) ) + k ( Γ a d ( t , s ) Γ b c ( t , s ) Γ c b ( t , s ) + Γ d a ( t , s ) )

Appendix B

Proof of Theorem 1.
The proof is similar to that of Theorem 1 in [24]. □

Appendix C

Proof of Theorem 2.
Point 1 is obtained from (12) and the fact that the SWL and the real processing are isomorphic. Again, point 2 follows from (12). The proof of point 3 is immediate due to G C α G C ¯ α , α 0 .
Next, point 4 is demonstrated under the H α 0 -properness condition. The proof of the E α 0 -proper case is similar. From (2) and (6) we obtain
Γ y z ¯ 3 ( α 0 ) = [ 1 , i , j , k ] Γ y r z r T p H α 0 ,     Γ z ¯ 3 ( α 0 ) = T p Γ z r T p H α 0 Γ y z ¯ 1 ( α 0 ) = [ 1 , i , j , k ] Γ y r z r T p T ,     Γ z ¯ 1 ( α 0 ) = T p Γ z r T p T
with z given in (15). Then, since Γ y z ¯ 3 ( α 0 ) Γ z ¯ 3 1 ( α 0 ) = Γ y z ¯ 1 ( α 0 ) Γ z ¯ 1 1 ( α 0 ) , the proof follows from (A1), Definition 1 and (7). □

Appendix D

Proof of Theorem 3.
We carry out the demonstration for the H α 0 -proper case. The E α 0 -proper case is analogous. Firstly, since x ( t ) is H α 0 -proper and applying Theorem 2 we have that x ^ SWL ( t + 1 | t ) = x ^ α 0 S ( t + 1 | t ) . Now, consider the next product respect to α 0
x , y 1 = E [ x y ] = Γ x y 1 ( α 0 ) ,     x , y S
and define the distance
d 1 ( x , y ) = x y 1
where x 1 2 = R { Γ x 1 ( α 0 ) } . Following a similar procedure to the one used with (11), the existence and uniqueness of the projection of x ( t + 1 ) onto the set G C α 0 t associated to the metric (A2) is guaranteed. Further, this projection coincides with x ^ α 0 S ( t + 1 | t ) , which is obtained from (11).
Define the innovations as x ˜ α 0 S ( i ) = x ( i ) x ^ α 0 S ( i | i 1 ) . Thus, the set { x ˜ α 0 S ( 1 ) , , x ˜ α 0 S ( t ) } spans G C α 0 t and, from (4), it follows that
x ˜ α 0 S ( i ) , x ˜ α 0 S ( j ) 1 = 0 ,     i j
Then, (16) holds. Now, denoting Π ( k ) = Γ x ˜ α 0 S 1 ( α 0 , k + 1 , k + 1 ) , we obtain
Γ x x ˜ α 0 S 1 ( α 0 , t + 1 , k + 1 ) = Γ x ^ α 0 S x ˜ α 0 S 1 ( α 0 , t + 1 , k + 1 ) = Λ ( t , t k ) Π ( k ) ,     0 k < t
By using (16), Equation (A3) can be rewritten as
Λ ( t , t k ) Π ( k ) = Γ x 1 ( α 0 , t + 1 , k + 1 ) i = 0 k 1 Γ x x ˜ α 0 S 1 ( α 0 , t + 1 , i + 1 ) Λ T ( k , k i )
Taking into account that Γ x 1 ( α 0 , t , s ) is nonsingular and hence, Π ( k ) is also nonsingular, we obtain (17).
Now, since Π ( t ) = Γ x 1 ( α 0 , t + 1 , t + 1 ) Γ x ^ α 0 S 1 ( α 0 , t + 1 , t + 1 ) and from (16), we obtain (18). Finally, taking into account that x ˜ S ( t + 1 ) is H α 0 -proper and
x ˜ j S ( t + 1 ) α 0 2 = 2 α 0 + 2 4 α 0 R { Γ x ˜ j S 1 ( α 0 , t + 1 , t + 1 ) }
then (19) follows. □

Appendix E

Proof of Proposition 2.
From (20), we have
F ( B ) x ¯ ( t ) = G ( B ) v ¯ ( t )
where B denotes the backward shift operator, G ( z ) = I 4 p + i = 1 q G ¯ ( i ) z i and G ¯ ( i ) = diag ( G ( i ) , G i ( i ) , G j ( i ) , G k ( i ) ) , i = 1 , , q . Thus, post-multiplying (A4) by x ¯ H 3 2 ν ( t h ) , taking expectations and taking into account the H α 0 -properness (respectively, E α 0 -properness), (22) holds. The recursive expression for the matrices L ( t ) can be derived as follows. Taking into account that det ( F ( z ) ) α 0 0 z C such that | z | 1 , we have that (Theorem 11.3.1 in [3]).
x ¯ ( t ) = F 1 ( B ) G ( B ) v ¯ ( t ) = j = 0 L ¯ ( j ) v ¯ ( t j )
By comparing coefficients of z j in the power series j = 0 L ¯ ( j ) z j after multiplying through by F ( z ) and from the block diagonality of F ¯ ( i ) and G ¯ ( i ) , the result follows. □

Appendix F

Proof of Proposition 4.
The proof is given for the H α 0 -proper case and the E α 0 -proper case is similarly obtained. Since x ( t ) is H α 0 -proper and from Theorem 2 we have that x ^ SWL ( t + 1 | t ) = x ^ α 0 S ( t + 1 | t ) .
Now, consider the transformed signal
w ( t ) = x ( t ) ,     1 t m w ( t ) = x ( t ) i = 1 p F ( i ) x ( t i ) ,     t > m
It is trivial that G C α 0 t = G D α 0 t with C α 0 t = { x ( 1 ) , , x ( t ) } and D α 0 t = { w ( 1 ) , , w ( t ) } . Then, defining w ^ α 0 ( 1 | 0 ) = 0 p × 1 and denoting by w ^ α 0 ( t + 1 | t ) the S α 0 one-stage predictor of w ( t + 1 ) respect to D α 0 t under (A2), we have
x ^ α 0 S ( t + 1 | t ) = w ^ α 0 ( t + 1 | t ) ,     0 t < m x ^ α 0 S ( t + 1 | t ) = i = 1 p F ( i ) x ( t i ) + w ^ α 0 ( t + 1 | t ) ,     t m
Thus, the problem of computing x ^ α 0 S ( t + 1 | t ) is equivalent to calculating w ^ α 0 ( t + 1 | t ) . It is not difficult to prove that Γ w 1 ( α 0 , t , s ) = Υ 1 ( t , s ) given in (24). Since Υ 1 ( t , s ) = 0 p × p if t > m and | t s | > q , then, from Theorem 3, we obtain Λ ( t , i ) = 0 p × p for t m and i > q , and hence,
w ^ α 0 ( t + 1 | t ) = { i = 1 t Λ ( t , i ) ( w ( t + 1 i ) w ^ α 0 ( t + 1 i | t i ) ) ,   1 t < m i = 1 q Λ ( t , i ) ( w ( t + 1 i ) w ^ α 0 ( t + 1 i | t i ) ) ,   t m
with w ^ α 0 ( 1 | 0 ) = 0 p × 1 and the matrices Λ ( t , i ) calculated from (17) and (18) for Υ 1 ( t , s ) .
Finally, from (A5), (A6) and the following equality
x ( t + 1 ) x ^ α 0 S ( t + 1 | t ) = w ( t + 1 ) w ^ α 0 ( t + 1 | t ) ,     t 0
(23) follows. □

References

  1. Pollock, D.S.G. A Handbook of Time-Series Analysis, Signal Processing and Dynamics; Academic Press: San Diego, CA, USA, 1999. [Google Scholar]
  2. Manolakis, D.; Bosowski, N.; Ingle, V.K. Count time-series analysis: A signal processing perspective. IEEE Signal Process. Mag. 2019, 36, 64–81. [Google Scholar] [CrossRef]
  3. Brockwell, P.J.; Davis, R.A. Time Series: Theory and Methods, 2nd ed.; Springer: New York, NY, USA, 1991. [Google Scholar]
  4. Kuipers, J.B. Quaternions and Rotation Sequences: A Primer with Applications to Orbits, Aerospace and Virtual Reality; Princeton University Press: Princeton, NJ, USA, 2002. [Google Scholar]
  5. Hanson, A.J. Visualizing Quaternions; The Morgan Kaufmann Series in Interactive 3D Technology; Elsevier: Amsterdam, The Netherlands, 2005. [Google Scholar]
  6. Catoni, F.; Boccaletti, D.; Cannata, R.; Catoni, V.; Nichelatti, E.; Zampetti, P. The Mathematics of Minkowski Space-Time: With an Introduction to Commutative Hypercomplex Numbers; Birkhaüser Verlag: Basel, Switzerland, 2008. [Google Scholar]
  7. Ginzberg, P.; Walden, A.T. Quaternion VAR modelling and estimation. IEEE Trans. Signal Process. 2013, 61, 154–158. [Google Scholar] [CrossRef] [Green Version]
  8. Tobar, F.; Mandic, D.P. Quaternion reproducing kernel Hilbert spaces: Existence and uniqueness conditions. IEEE Trans. Inform. Theory 2014, 60, 5736–5749. [Google Scholar] [CrossRef]
  9. Schütte, H.D.; Wenzel, J. Hypercomplex numbers in digital signal processing. In Proceedings of the International Symposium on Circuits and Systems, New Orleans, LA, USA, 1–3 May 1990; pp. 1557–1560. Available online: https://ieeexplore.ieee.org/document/112431 (accessed on 15 February 2022).
  10. Pei, S.H.; Chang, J.H.; Ding, J.J. Commutative reduced biquaternions and their Fourier transform for signal and image processing applications. IEEE Trans. Signal Process. 2004, 52, 2012–2031. [Google Scholar] [CrossRef]
  11. Labunets, V. Clifford algebras as unified language for image processing and pattern recognition. In Computational Noncommutative Algebra and Applications; NATO Science Series II: Mathematics, Physics and Chemistry; Byrnes, J., Ed.; Springer: Dordrecht, The Netherlands, 2004; Volume 136, pp. 197–225. [Google Scholar]
  12. Alfsmann, D. On families of 2N-dimensional hypercomplex algebras suitable for digital signal processing. In Proceedings of the 14th European Signal Processing Conference (EUSIPCO 2006), Florence, Italy, 4–8 September 2006; pp. 1–4. [Google Scholar]
  13. Alfsmann, D.; Göckler, H.G.; Sangwine, S.J.; Ell, T.A. Hypercomplex algebras in digital signal processing: Benefits and drawbacks. In Proceedings of the 15th European Signal Processing Conference, Poznan, Poland, 3–7 September 2007; pp. 1322–1326. [Google Scholar]
  14. Pei, S.H.; Chang, J.H.; Ding, J.J.; Chen, M.Y. Eigenvalues and singular value decompositions of reduced biquaternion matrices. IEEE Trans. Circuits Syst. I 2008, 525, 2673–2685. [Google Scholar]
  15. Guo, L.; Zhu, M.; Ge, X. Reduced biquaternion canonical transform, convolution and correlation. Signal Process. 2011, 91, 2147–2153. [Google Scholar] [CrossRef]
  16. Hahn, S.L.; Snopek, K.M. Complex and Hypercomplex Analytic Signals: Theory and Applications; Artech House: Norwood, MA, USA, 2016. [Google Scholar]
  17. Kamal, A.T.; El-Melegy, M.T. Color image processing using reduced biquaternions with applications to face recognition in a PCA framework. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 3039–3046. [Google Scholar]
  18. Esposito, A.; Faundez-Zanuy, M.; Morabito, F.C.; Pasero, E. Neural Advances in Processing Nonlinear Dynamic Signals; Springer: Cham, Switzerland, 2018. [Google Scholar]
  19. Kobayashi, M. Twin-multistate commutative quaternion Hopfield neural networks. Neurocomputing 2018, 320, 150–156. [Google Scholar] [CrossRef]
  20. Zanetti de Castro, F.; Eduardo Valle, M. A broad class of discrete-time hypercomplex-valued Hopfield neural networks. Neural Netw. 2020, 122, 54–67. [Google Scholar] [CrossRef] [Green Version]
  21. Takahashi, K. Comparison of high-dimensional neural networks using hypercomplex numbers in a robot manipulator control. Artif. Life Robot. 2021, 26, 367–377. [Google Scholar] [CrossRef]
  22. Ortolani, F.; Scarpiniti, M.; Comminiello, D.; Uncini, A. On the influence of microphone array geometry on the behavior of hypercomplex adaptive filters. In Proceedings of the 5th IEEE Microwaves, Radar and Remote Sensing Symposium (MRRS), Kiev, Ukraine, 29–31 August 2017; pp. 37–42. [Google Scholar]
  23. Ortolani, F.; Scarpiniti, M.; Comminiello, D.; Uncini, A. On 4-dimensional hypercomplex algebras in adaptive signal processing. In Neural Advances in Processing Nonlinear Dynamic Signals; Springer: Berlin, Germany, 2017; pp. 131–140. [Google Scholar]
  24. Navarro-Moreno, J.; Fernández-Alcalá, R.M.; López-Jiménez, J.D.; Ruiz-Molina, J.C. Tessarine signal processing under the T-properness condition. J. Frankl. Inst. 2020, 357, 10100–10126. [Google Scholar] [CrossRef]
  25. Navarro-Moreno, J.; Ruiz-Molina, J.C. Wide-sense Markov signals on the tessarine domain. A study under properness conditions. Signal Process. 2021, 183C, 108022. [Google Scholar] [CrossRef]
  26. Took, C.C.; Mandic, D.P. Augmented second-order statistics of quaternion random signals. Signal Process. 2011, 91, 214–224. [Google Scholar] [CrossRef] [Green Version]
  27. Navarro-Moreno, J.; Fernández-Alcalá, R.M.; Ruiz-Molina, J.C. Semi-widely simulation and estimation of continuous-time Cη-proper quaternion random signals. IEEE Trans. Signal Process. 2015, 63, 4999–5012. [Google Scholar] [CrossRef]
  28. Navarro-Moreno, J.; Ruiz-Molina, J.C. Semi-widely linear estimation of Cη-proper quaternion random signal vectors under Gaussian and stationary conditions. Signal Process. 2016, 119, 56–66. [Google Scholar] [CrossRef]
  29. Nitta, T.; Kobayashi, M.; Mandic, D.P. Hypercomplex widely linear estimation through the lens of underpinning geometry. IEEE Trans. Signal Process. 2019, 67, 3985–3994. [Google Scholar] [CrossRef]
  30. Grassucci, E.; Comminiello, D.; Uncini, A. An information-theoretic perspective on proper quaternion variational autoencoders. Entropy 2021, 23, 856. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Wilks’ statistic for case 1 (left) and case 2 (right).
Figure 1. Wilks’ statistic for case 1 (left) and case 2 (right).
Mathematics 10 01083 g001
Figure 2. D P α 0 ( t ) in scenarios 1 (left) and 2 (right).
Figure 2. D P α 0 ( t ) in scenarios 1 (left) and 2 (right).
Mathematics 10 01083 g002
Figure 3. ϵ α 0 SWL ( t + 1 | t ) for α 0 = 3 , 2 , 1 , 1 , 2 , 3 .
Figure 3. ϵ α 0 SWL ( t + 1 | t ) for α 0 = 3 , 2 , 1 , 1 , 2 , 3 .
Mathematics 10 01083 g003
Figure 4. Values of the components Λ i ( t , 1 ) , i = 1 , , 4 , for α 0 = 3 , 2 , 1 , 1 , 2 , 3 .
Figure 4. Values of the components Λ i ( t , 1 ) , i = 1 , , 4 , for α 0 = 3 , 2 , 1 , 1 , 2 , 3 .
Mathematics 10 01083 g004aMathematics 10 01083 g004b
Table 1. List of symbols.
Table 1. List of symbols.
SymbolMeaning
x r ( t ) Real vector formed with the components of x S p
x ¯ ( t ) Augmented GSQ signal vector formed by x S p and its conjugations
Γ a ( t , s ) Correlation function of a real random signal vector a ( t ) R p
Γ a b ( t , s ) Correlation function of the real random signal vectors a ( t ) R p and b ( t ) R q
Γ x ν ( α , t , s ) E [ x ( t ) x H 3 2 ν ( s ) ] , for x S p , ν = 1 , 2
Γ x y ν ( α , t , s ) E [ x ( t ) y H 3 2 ν ( s ) ] , for x S p , y S q , and ν = 1 , 2
Γ x 3 ( α , t , s ) E [ x ( t ) x H α ( s ) ] , for x S p
Γ x y 3 ( α , t , s ) E [ x ( t ) y H α ( s ) ] , for x S p and y S q
Γ ^ x Sample autocorrelation matrix of a random vector x
C α Set of the random vectors { x 1 , , x m } , with x i S p , i = 1 , , m
C ¯ α Set of the augmented random vectors { x ¯ 1 , , x ¯ m } , with x ¯ i S 4 p , i = 1 , , m
G C α Set of all finite linear combinations of elements of C α
G C ¯ α Set of all finite linear combinations of elements of C ¯ α
y ^ SWL SWL estimator of y respect to C α
y ^ α S S α estimator of y respect to C α
y ^ QSL QSL estimator of y respect to C α
ϵ SWL SWL estimation error associated to y ^ SWL
ϵ α S S α estimation error associated to y ^ α S
ϵ QSL QSL estimation error associated to y ^ QSL
Table 2. The 95% confidence intervals for α 0 in case 1.
Table 2. The 95% confidence intervals for α 0 in case 1.
n50100300500
C.I.[1.506, 2.615][1.642, 2.433][1.788, 2.236][1.835, 2.190]
Table 3. The 95% confidence intervals for α 0 in case 2.
Table 3. The 95% confidence intervals for α 0 in case 2.
n50100300500
C.I.[−2.650, −1.512][−2.443, −1.633][−2.233, −1.777][−2.181, −1.846]
Table 4. Steady-state errors of the one-stage predictor for the different values of α 0 .
Table 4. Steady-state errors of the one-stage predictor for the different values of α 0 .
| α 0 | 123
ϵ 243648
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Navarro-Moreno, J.; Fernández-Alcalá, R.M.; Ruiz-Molina, J.C. Proper ARMA Modeling and Forecasting in the Generalized Segre’s Quaternions Domain. Mathematics 2022, 10, 1083. https://doi.org/10.3390/math10071083

AMA Style

Navarro-Moreno J, Fernández-Alcalá RM, Ruiz-Molina JC. Proper ARMA Modeling and Forecasting in the Generalized Segre’s Quaternions Domain. Mathematics. 2022; 10(7):1083. https://doi.org/10.3390/math10071083

Chicago/Turabian Style

Navarro-Moreno, Jesús, Rosa M. Fernández-Alcalá, and Juan C. Ruiz-Molina. 2022. "Proper ARMA Modeling and Forecasting in the Generalized Segre’s Quaternions Domain" Mathematics 10, no. 7: 1083. https://doi.org/10.3390/math10071083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop