Next Article in Journal
The Real-Life Application of Differential Evolution with a Distance-Based Mutation-Selection
Next Article in Special Issue
Evaluating the Efficiency of Off-Ball Screens in Elite Basketball Teams via Second-Order Markov Modelling
Previous Article in Journal
Dirichlet and Neumann Boundary Value Problems for the Polyharmonic Equation in the Unit Ball
Previous Article in Special Issue
On State Occupancies, First Passage Times and Duration in Non-Homogeneous Semi-Markov Chains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

State Space Modeling with Non-Negativity Constraints Using Quadratic Forms

by
Ourania Theodosiadou
*,† and
George Tsaklidis
*,†
Department of Mathematics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2021, 9(16), 1908; https://doi.org/10.3390/math9161908
Submission received: 11 July 2021 / Revised: 3 August 2021 / Accepted: 5 August 2021 / Published: 10 August 2021

Abstract

:
State space model representation is widely used for the estimation of nonobservable (hidden) random variables when noisy observations of the associated stochastic process are available. In case the state vector is subject to constraints, the standard Kalman filtering algorithm can no longer be used in the estimation procedure, since it assumes the linearity of the model. This kind of issue is considered in what follows for the case of hidden variables that have to be non-negative. This restriction, which is common in many real applications, can be faced by describing the dynamic system of the hidden variables through non-negative definite quadratic forms. Such a model could describe any process where a positive component represents “gain”, while the negative one represents “loss”; the observation is derived from the difference between the two components, which stands for the “surplus”. Here, a thorough analysis of the conditions that have to be satisfied regarding the existence of non-negative estimations of the hidden variables is presented via the use of the Karush–Kuhn–Tucker conditions.

1. Introduction

State space modeling is used for estimating—revealing the dynamic evolution of hidden variables’ processes. In some cases, the state vector, which includes the hidden components, is subject to constraints, which are derived either due to the physical meaning of the states or because of the mathematical properties that have to be satisfied. For example, state space models with constraints are used in camera surveillance [1,2], navigation issues [3], and biological systems [4]. Especially, in finance, the hidden variables are often subject to non-negative constraints or in general have to be bounded. For example, in the Vasicek model [5] and its extension [6], the interest rates are considered to be hidden random variables subject to non-negative constraints, while in [7,8], the eigenvalues of the VAR process were restricted within the unit circle. Considering the use of state space models in the domain of finance, a discrete state space model could be implemented for the estimation of the hidden jump components of asset returns [9,10]. The use of jumps has been proposed for the description of the dynamics of asset prices since they can explain some of the empirical characteristics of the asset prices, e.g., the lack of a normal distribution or the existence of leptokurticity (see for example [11]).
When dealing with state space models that are subject to constraints, the Kalman filtering algorithm [12] can no longer be used, since it assumes linearity in the model. In the domain of nonlinear filters, the particle filtering approach (see for example [13,14,15,16]) has wide applicability, and it adopts resampling techniques for the estimation of the state vector at every time t. However, the use of resampling techniques adds considerable computational cost in the estimation procedure.
In this work, the observation is defined as the difference between the two-sided components under noise inclusion. The components are considered to be hidden random variables, and therefore, a state space model is established, where the state equation describes the dynamic evolution of the two hidden components. This equation represents a first-order Markov process, i.e., all the information needed for the estimation of the components at time t is derived by the components at time t 1 , and no other information from past times is needed. Moreover, the state vector is subject to non-negative constraints that have to be taken into account for its estimation in time. Such a model could describe, for example, the evolution of a system where the positive component represents “gain”, while the negative one represents “loss”; the observation is derives from the difference between the two components, which stands for the “surplus”, under noise inclusion. In asset pricing, an asset return can be defined as the difference between the two-sided non-negative return jump components under noise inclusion, and the jump components are considered to be hidden variables. Another example could be the one-dimensional random walk, where a positive jump could represent (the measure of) a move to the right and a negative jump (the measure of) a move to the left, while the observation could be a function of the two jump components given at discrete times. To handle such kinds of problems, non-negative definite quadratic forms are adopted in the state equation for the dynamic evolution of the two-sided components. In this case, the recursive equations of the Kalman filter cannot be used for the estimation of the state vector, since this filter assumes linearity in the measurement and state equation. To this end, this work first derives the recursive equations for the estimation of the state vector based on the state space model representation with non-negative definite quadratic forms in the state equation and their Taylor expansions. Then, a thorough analysis of the necessary conditions that have to be satisfied in order to obtain the non-negative estimations at every time t is provided. In Proposition 1, the stationary points of the optimization problem with the non-negative constraints are given by using the Karush–Kuhn–Tucker conditions, while in Proposition 2, the necessary conditions for the existence of feasible solutions in the constrained optimization problem are provided.
Overall, this work proposes a method in state space modeling representation, which can be used when dealing with hidden components that are subject to non-negativity constraints. The method results in the formulation of a constrained optimization problem for which the stationary points are derived via Proposition 1, and the necessary conditions for the existence of feasible solutions in this optimization problem are provided via Proposition 2; to that end, the iterative formulas for the minimum variance a posteriori estimators for the (hidden) state vector are illustrated. Moreover, the proposed method has a low computational burden compared to other nonlinear filtering methods that can be used in state space modeling with inequality constraints and are based on resampling techniques (e.g., particle filtering).
The paper is organized as follows. In Section 2, the state space model proposed for the estimation of the two jump components is established. Two non-negative quadratic forms are adopted to describe the dynamic evolution of the two-sided components subject to their non-negative restrictions. In Section 3, the recursive equations of the second-order Kalman filter are presented, while in Section 4, a thorough analysis of the conditions that have to be fulfilled so as to have non-negative estimations is presented. The results of this analysis are summarized in Propositions 1 and 2. In Section 5, an illustrative example concerning the evolution of positive and negative jumps of asset returns is presented to demonstrate the theoretical results. Finally, Section 6 concludes on the findings and provides suggestions for future work.

2. State Space Model

In this section, a state space model representation is illustrated considering the case where there are two hidden processes subject to non-negativity constraints. The state equation that describes the dynamic evolution of the hidden components adopts the use of non-negative definite quadratic forms, while the measurement equation is linear.
The state equation is given by:
X t = ( z t 1 + w t 1 ) G ( 1 ) ( z t 1 + w t 1 ) = f 1 ( z t 1 , w t 1 ) Y t = ( z t 1 + w t 1 ) G ( 2 ) ( z t 1 + w t 1 ) = f 2 ( z t 1 , w t 1 )
or equivalently:
z t = k = 1 2 ϕ k ( z t 1 + w t 1 ) G ( k ) ( z t 1 + w t 1 )
where:
  • z t = ( X t , Y t ) = ( z t , 1 , z t , 2 ) stands for the state vector;
  • w t stands for the noise, and it is assumed that w t N ( 0 , Q ) , where Q = σ x 2 0 0 σ y 2 ;
  • G ( k ) , k = 1 , 2 , is a (symmetric, 2 × 2 ) non-negative definite matrix, i.e.,
    g 11 ( k ) > 0 and g 11 ( k ) g 22 ( k ) ( g 12 ( k ) ) 2 > 0 , k = 1 , 2 .
The vector ϕ k is a ( 2 × 1 ) column vector, where the k-th element equals 1, and the other element equals 0. The measurement equation is given by the relation:
R t = H z t + e t ,
where H = 1 1 and e t N ( 0 , V ) . Moreover, it is assumed that E ( e k w j T ) = 0 .
Apparently, state Equation (2) describes a (nonobservable) first-order non-negative valued Markovian process, the evolution of which and its characteristics (e.g., periodicity, convergence etc.) depend on the structure (values) of the associated noisy observation sequence. The aim of our study here was to estimate (reveal) the Markovian process (2) (i.e., the matrices G ( k ) , k = 1 , 2 , and Q ), through the observation Equation (3), if the components of the state vector have to be non-negative. For this purpose, Model (2) and (3) adopts the use of non-negative definite quadratic forms to describe the dynamic evolution of the hidden two-sided components; that is, to ensure that the estimations of the components will be non-negative. To that end, the extended Kalman filter of second order is proposed in order to estimate at every time t the state vector z t that incorporates the hidden jump components. It is noticed here that the noise component in Relation (2) is multiplicative and not additive.
Next, the extended Kalman filter of second order is described and its iterative equations for the estimation of the state vector are presented.

3. Extended Kalman Filter of Second Order

Model (2) and (3) presented in Section 2 is nonlinear, and subsequently, the recursive standard algorithm of the Kalman filter cannot be used for the estimation of the state vector. Aiming to derive the recursive equations for the estimation of the hidden states taking into consideration that the state Equation (2) is a quadratic form, the following notation is used:
  • z ^ t : the a priori estimation of the state vector z t , i.e., without taking into consideration the measurement at time t;
  • z ^ t + : the a posteriori estimation of the state vector z t , i.e., by considering the measurement at time t;
  • P t , P t + : the variance–covariance matrices of the a priori and a posteriori error estimations of z t , respectively, i.e.,
    P t = E [ ( z t z ^ t ) ( z t z ^ t ) ] and P t + = E [ ( z t z ^ t + ) ( z t z ^ t + ) ] .
According to (2), z t , k , k = 1 , 2 is a function of the random variables z t 1 , and w t 1 , i.e., z t , k = z t , k ( z t 1 , w t 1 ) . Then, using the Taylor expansion of second order of z t , k at ( z ^ t 1 + , 0 ) , it is derived that:
z t , k = f k ( z ^ t 1 + , 0 ) + ( f k ( z ^ t 1 + , 0 ) z t 1 ) ( z t 1 z ^ t 1 + ) + ( f k ( z ^ t 1 + , 0 ) w t 1 ) w t 1 + 1 2 ( z t 1 z ^ t 1 + ) 2 f k ( z ^ t 1 + , 0 ) z t 1 2 ( z t 1 z ^ t 1 ) + 1 2 w t 1 2 f k ( z ^ t 1 + , 0 ) w t 1 2 w t 1 + ( z t 1 z ^ t 1 + ) 2 f k ( z ^ t 1 + , 0 ) z t 1 w t 1 w t 1 , k = 1 , 2
where functions f k = f k ( z t 1 , w t 1 ) , k = 1 , 2 , are given in (1). By equating the mean values in Relation (4), the a priori estimation of z t (prediction stage) is derived, that is:
z ^ t , k = f k ( z ^ t 1 + , 0 ) + 1 2 tr ( 2 f k ( z ^ t 1 + , 0 ) z t 1 2 P t 1 + ) + 1 2 tr ( 2 f k ( z ^ t 1 + , 0 ) w t 1 2 Q ) , k = 1 , 2
and the entries of the respective variance–covariance matrix P t are given by the relation,
( P t ) k , m = ( f k ( z ^ t 1 + , 0 ) z t 1 ) P t 1 + f m ( z ^ t 1 + , 0 ) z t 1 + ( f k ( z ^ t 1 + , 0 ) w t 1 ) Q f m ( z ^ t 1 + , 0 ) w t 1 + 1 2 tr ( 2 f k ( z ^ t 1 + , 0 ) z t 1 2 P t 1 + 2 f m ( z ^ t 1 + , 0 ) z t 1 2 P t 1 + ) + 1 2 tr ( 2 f k ( z ^ t 1 + , 0 ) w t 1 2 Q 2 f m ( z ^ t 1 + , 0 ) w t 1 2 Q ) , k , m = 1 , 2
where ( P t ) k , m denotes the ( k , m ) -element of matrix P t and tr ( . ) denotes the trace of the respective matrix. Taking into consideration the properties of the trace of a matrix, it is derived after some algebraic manipulations on Relations (5) and (6) that:
z ^ t , k = z ^ t 1 + T G ( k ) z ^ t 1 + + tr ( G ( k ) P t 1 + ) + tr ( G ( k ) Q ) , k = 1 , 2
( P t ) k , m = 4 z ^ t 1 + T G ( k ) P t 1 + G ( m ) z ^ t 1 + + 4 z ^ t 1 + T G ( k ) Q G ( m ) z ^ t 1 + + 2 tr ( G ( k ) P t 1 + G ( m ) P t 1 + ) + 2 tr ( G ( k ) Q G ( m ) Q ) , k , m = 1 , 2 .
Regarding the a posteriori estimations of z t , it is taken into account that the joint distribution of z t and R t is normal, based on the relation:
z t R t N ( z ^ t H z ^ t , P t P t H T H P t T H P t H T + V ) .
Then, we make use of the following Lemma (see for example [17]):
Lemma 1.
Let x , y be two random variables that are jointly normally distributed with:
E ( x y ) = μ x μ y and Σ = Σ 11 Σ 12 Σ 21 Σ 22 .
Then, ( x / y ) N ( μ , Σ ) , where:
μ = μ x + Σ 11 Σ 22 1 ( y μ x ) and Σ = Σ 11 Σ 12 Σ 22 1 Σ 21 .
Based on Lemma 1, the a posteriori estimation of z t (update stage) and the related variance–covariance matrix P t ( + ) are given by,
z ^ t + = z ^ t + K t ( R t H z ^ t ) ,
P t + = ( I K t H ) P t ,
where K t = P t H T ( HP t H T + V ) 1 . By using the recursive Relations (7)–(10), we can estimate the hidden components at every time t.
Next, a detailed investigation regarding the existence of non-negative solutions (i.e., non-negative a posteriori estimations of z t ) derived from (9) is presented.

4. Investigation of the State Space Model

In what follows, we present an investigation concerning the conditions that have to be satisfied so as to derive non-negative a posteriori estimations of the state vector z t . Obviously, Relation (7) ensures the existence of non-negative a priori estimations of z t at every time t. However, the a posteriori estimations of z t given by (9) may not fulfil the non-negativity condition. We note that the solutions depend on the term K t ( R t H z ^ t ) , the sign of which is not time invariant. To this end, in order to ensure that the a posteriori unbiased estimator z ^ t + will be a minimum variance estimator under the non-negativity restrictions that its components must satisfy, the following optimization problem arises,
m i n z ^ t + { tr ( P t + ) = E [ ( z t z ^ t + ) ( z t z ^ t + ) T ] } where z ^ t + 0
Symbol ⪰ (or ⪯) is used for the elementwise inequality, while z t = ( X t , Y t ) T is given by Equation (1) (or (2)). The following Proposition 1 provides the set of stationary points related to the optimization problem (11), subject to the non-negativity restrictions. This set includes the optimal solution, i.e., the unbiased minimum variance estimator z ^ t + . In what follows, we use the following notations:
a t = R t H z ^ t ,   b t = H P t H T and K t = ( K t , 1 , K t , 2 )
.
Remark 1.
Notice that, if a t = 0 , then Relation (9) leads to z ^ t + = z ^ t 0 , and consequently, the solution is acceptable.
Taking into consideration Remark 1, it is assumed in the sequel that a t 0 for every t.
Proposition 1.
The weight matrix K t and the stationary points related to the optimization problem (11) are given by the relations:
(i) 
K t = ( b t + R t ) 1 P t H T , which leads to the solution:
z ^ t + = z ^ t + a t ( b t + R t ) 1 P t H T ;
(ii) 
K t = ( b t + R t ) 1 ( P t H T ) 1 a t 1 z ^ t , 2 which leads to the solution:
z ^ t , 1 + = z ^ t , 1 + a t ( b t + R t ) 1 ( P t H T ) 1 ,
and:
z ^ t , 2 + = 0 ;
(iii) 
K t = a t 1 z ^ t , 1 ( b t + R t ) 1 ( P t H T ) 2 , which leads to the solution:
z ^ t , 1 + = 0
and:
z ^ t , 2 + = z ^ t , 2 + a t ( b t + R t ) 1 ( P t H T ) 2 ;
(iv) 
K t = a t 1 z ^ t , which leads to the solution:
z ^ t + = 0 ,
where ( P t H T ) i denotes the i t h -row of matrix P t H T , i = 1 , 2 .
Proof. 
The Lagrangian function related to the optimization problem (11) is defined as:
Λ = tr ( P t + ) + λ 1 ( z ^ t , 1 + ) + λ 2 ( z ^ t , 2 + ) = tr ( E [ ( z t z ^ t + ) ( z t z ^ t + ) T ] ) + λ 1 ( z ^ t , 1 + ) + λ ( z ^ t , 2 + ) , λ 1 , λ 2 0 .
Based on (10), it is derived that:
tr ( P t + ) = tr [ ( I K t H ) P t ( I K t H ) T + K t V K t T ]
while (by assuming the dependence of z ^ t , i + on R t and z ^ t , i , i = 1 , 2 , as provided in Kalman filtering):
z ^ t , 1 + = z ^ t , 1 + K t , 1 ( R t H z ^ t ) = z ^ t , 1 + a t K t , 1 ,
z ^ t , 2 + = z ^ t , 2 + K t , 1 ( R t H z ^ t ) = z ^ t , 2 + a t K t , 2 .
By calculating the first derivative of the Lagrangian function and equating it to 0, it is derived that:
d Λ d K t = d d K t [ tr [ ( I K t H ) P t ( I K t H ) T + K t V K t T ] λ 1 ( z ^ t , 1 + a t K t , 1 ) λ 2 ( z ^ t , 2 + a t K t , 2 ) ] = 2 P t T H T + 2 K t H P t H T + 2 K t V a t λ = 0
where λ = ( λ 1 , λ 2 ) T . Thus, matrix K t has to satisfy the following condition (by noticing that P t is symmetric):
2 P t H T + 2 K t H P t H T + 2 K t V = a t λ
based on the constraints [18]:
λ 1 ( z ^ t , 1 + a t K t , 1 ) = 0 , λ 2 ( z ^ t , 2 + a t K t , 2 ) = 0 λ 1 , λ 2 0
.
The following cases have to be considered:
(i)
The two constraint conditions are inactive. Then, λ 1 = λ 2 = 0 , and the optimization problem, leading to (14), is transformed into the unconstrained one considered in the case of the Kalman filter. It is derived that:
K t = P t H T ( H P t H T + V ) 1 ,
which is the well-known Kalman gain matrix. The related solution in terms of the a posteriori estimator z ^ t + is:
z ^ t + = z ^ t + a t ( b t + R t ) 1 P t H T .
Relation (16) constitutes a possible solution of the optimization problem (11), and it has to satisfy the constraint z ^ t + 0 ;
(ii)
The first constraint condition is inactive (i.e., λ 1 = 0 ), while the second one is active. Then, the following two cases are considered:
(a)
If λ 2 = 0 , then we are led to the unconstrained optimization problem presented in Case (i), and the solution must satisfy the non-negative restrictions, i.e., z ^ t + 0 ;
(b)
If z ^ t , 2 + a t K t , 2 = 0 , it is derived via the active constraint condition that:
K t , 2 = a t 1 z ^ t , 2 .
By using (17), Relation (14) is transformed into:
( P t H T ) 1 ( P t H T ) 2 + K t , 1 a t 1 z ^ t , 2 H P t H T + K t , 1 a t 1 z ^ t , 2 V = a t λ .
Consequently,
K t , 1 = ( b t + R t ) 1 ( P t H T ) 1 ,
where b t = H P t H T 0 . By using (17) and (18), it is derived that:
K t = ( b t + R t ) 1 ( P t H T ) 1 a t 1 z ^ t , 2 .
Thus,
z ^ t , 1 + = z ^ t , 1 + a t ( b t + R t ) 1 ( P t H T ) 1
and:
z ^ t , 2 + = 0 ;
(iii)
The first constraint condition is active, while the second one is inactive (i.e., λ 2 = 0 ). The following two cases are considered:
(a)
If λ 1 = 0 , then we obtain the unconstrained optimization problem presented in Case (i), and the solution must fulfil the nonnegative restrictions, i.e., z ^ t + 0 ;
(b)
If z ^ t , 1 + a t K t , 1 = 0 and λ 1 = 0 , then it is derived that:
K t , 1 = a t 1 z ^ t , 1 ,
and Relation (14) is transformed into:
( P t H T ) 1 ( P t H T ) 2 + K t , 1 a t 1 z ^ t , 2 H P t H T + K t , 1 a t 1 z ^ t , 2 V = a t λ .
Then,
K t , 2 = ( b t + R t ) 1 ( P t H T ) 2
where b t = H P t H T 0 . By using (19) and (20), it is derived that:
K t = a t 1 z ^ t , 1 ( b t + R t ) 1 ( P t H T ) 2 ,
and consequently:
z ^ t , 1 + = 0
and:
z ^ t , 2 + = z ^ t , 2 + a t ( b t + R t ) 1 ( P t H T ) 2 ;
(iv)
The two constraint conditions are active, i.e., z ^ t , 1 + a t K t , 1 = 0 and z ^ t , 2 + a t K t , 2 = 0 . In this case, we have to seek solutions such that λ 1 , λ 2 0 .
Based on the active constraint conditions, it is derived that:
K t , 1 = a t 1 z ^ t , 1   and   K t , 2 = a t 1 z ^ t , 2
i.e., K t = a t 1 z ^ t , resulting in the relation,
z ^ t + = z ^ t + a t K t = z ^ t a t a t 1 z ^ t = 0 .
The state vector z ^ t + = 0 constitutes a feasible solution, and it has to be checked whether Relation (14) is satisfied with λ 1 , λ 2 0 . □
In what follows, Proposition 2 provides the necessary conditions for the existence of feasible solutions regarding the constrained filter.
Proposition 2.
The solutions given in Proposition 1 regarding the optimization problem (11) are feasible upon the following conditions (necessary conditions):
(i) 
z ^ t + = z ^ t + a t ( b t + R t ) 1 P t H
constitutes a feasible solution, if:
a t P t H T ( b t + R t ) z ^ t ;
(ii) 
z ^ t + = z ^ t , 1 + a t ( b t + R t ) 1 ( P t H T ) 1 0
constitutes a feasible solution, if:
a t ( P t H ) 1 ( b t + R t ) z ^ t , 1
and:
a t ( P t H ) 2 < ( b t + R t ) z ^ t , 2 ;
(iii) 
z ^ t + = 0 z ^ t , 2 + a t ( b t + R t ) 1 ( P t H T ) 2
constitutes a feasible solution, if:
a t ( P t H T ) 1 < ( b t + R t ) z ^ t , 1
and:
a t ( P t H T ) 2 ( b t + R t ) z ^ t , 2 ;
(iv) 
z ^ t + = 0 constitutes a feasible solution, if:
a t ( P t H T ) 1 < ( b t + R t ) z ^ t , 1
and:
a t ( P t H T ) 2 < ( b t + R t ) z ^ t , 2 .
Proof. 
Similar to the proof of Proposition 1, four cases are considered:
(i)
The two constraint conditions are inactive. Then, λ 1 = λ 2 = 0 , and the optimization problem is transformed into the unconstrained one that is met in the case of the Kalman filter. In this case, based on Proposition 1, we obtain that K t = ( b t + R t ) 1 P t H T , resulting in the estimation:
z ^ t + = z ^ t + a t ( b t + R t ) 1 P t H T ,
where z ^ t + is a feasible solution of the optimization problem with the nonnegative constraints, if:
z ^ t + a t ( b t + R t ) 1 P t H T 0 .
Consequently, the necessary condition is formulated as follows:
a t P t H T a t ( b t + R t ) z ^ t ;
(ii)
The first constraint condition is inactive, while the second one is active, i.e, λ 1 = 0 and z ^ t , 2 + a t K t , 2 = 0 , respectively. The following two cases are considered:
(a)
If λ 2 = 0 , then based on (14), the solution is given by (15), which is related to the Kalman filter and the unconstrained optimization problem. This solution is acceptable if it is aligned with the active constraint condition. Otherwise, it is rejected;
(b)
If λ 2 > 0 , matrix K t has to be in such a form so that z ^ t , 1 + 0 .
It is derived via the active constraint condition that K t , 2 = a t 1 z ^ t , 2 where a t 0 based on Remark 1. Then, (14) results in:
a t λ = ( P t H T ) 1 ( P t H T ) 2 + K t , 1 a t 1 z ^ t , 2 H P t H T + K t , 1 a t 1 z ^ t , 2 V λ 1 λ 2 = a t 1 ( P t H T ) 1 a t 1 ( P t H T ) 2 + a t 1 K t , 1 a t 2 z ^ t , 2 H P t H T + a t 1 K t , 1 a t 2 z ^ t , 2 V = a t 1 ( P t ( ) H T ) 1 + a t 1 b t K t , 1 + a t 1 K t , 1 V a t 1 ( P t H T ) 2 + a t 2 b t z ^ t , 2 + a t 2 z ^ t , 2 V .
Since λ 1 = 0 and for a t 0 , Relation (21) implies:
0 λ 2 = ( P t H T ) 1 + b t K t , 1 + K t , 1 V a t 1 ( P t H T ) 2 a t 2 b t z ^ t , 2 a t 2 z ^ t , 2 V
It is derived via (22) that if λ 2 > 0 , then:
λ 2 = a t ( P t H T ) 2 b t z ^ t , 2 z ^ t , 2 V > 0 .
Consequently,
a t ( P t H T ) 2 < ( b t + V ) z ^ t , 2 ,
resulting in a t ( P t H T ) 2 < 0 . Moreover, taking into consideration that:
z ^ t , 1 + = z ^ t + a t K t , 1 = z ^ t , 1 a t ( b t + V ) ( P t H T ) 1
and z ^ t , 1 + 0 , it is derived that:
a t ( P t H T ) 1 < ( b t + V ) z ^ t , 1 ;
(iii)
The first constraint condition is active, while the second one is inactive, i.e., z ^ t , 1 + a t K t , 1 = 0 and λ 2 = 0 , respectively. The following two cases are considered:
(a)
λ 1 = 0 and λ 2 = 0 ;
(b)
λ 1 > 0 and λ 2 = 0 .
Similar to Case (ii), the third part of Proposition 2 can be derived;
(iv)
The two constraint conditions are active, i.e., z ^ t , 1 + a t K t , 1 = 0 and z ^ t , 2 + a t K t , 2 = 0 . In this case, we have to search for solutions where λ 1 , λ 2 0 . It is derived via Proposition 1 that:
K t = a t 1 z ^ t ,
which leads to the solution z ^ t + = 0 .
The following subcases are considered:
(a)
If λ 1 = λ 2 = 0 , Solution (15) is derived via Relation (14), and it is accepted if it coincides with Relation (25). Otherwise, it is rejected;
(b)
If λ 1 = 0 and λ 2 > 0 , then by taking into consideration Case (iib), it is concluded that the solution z t ( + ) = 0 is accepted, if:
a t ( P t H T ) 2 < ( b t + V ) z ^ t , 2 .
Otherwise, it is rejected since it is not aligned with the conditions of the considered case (i.e., λ 1 = 0 λ 2 > 0 );
(c)
If λ 1 > 0 and λ 2 = 0 , similar to Case (iv)-c, the solution z t + = 0 is accepted, if:
a t ( P t H T ) 1 < ( b t + V ) z ^ t , 1 ;
(d)
If λ 1 > 0 and λ 2 > 0 , then by taking into consideration Relations (23) and (25), it turns out that the necessary condition in order for z t + = 0 to be accepted as a feasible solution (which satisfies the conditions of the considered case, i.e., λ 1 > 0 λ 2 > 0 ) is:
a t ( P t H T ) 1 < ( b t + V ) z ^ t , 1   and   a t ( P t H T ) 2 < ( b t + V ) z ^ t , 2
.
In conclusion, in Case (iv), the vector z t + = 0 is a possible optimal solution, if at least one of the following conditions holds:
a t ( P t ( ) H T ) 1 < ( b t + V ) z ^ t , 1   or   a t ( P t H T ) 2 < ( b t + V ) z ^ t , 2
. □
Remark 2.
Based on the low computational cost, the four possible solutions of the constrained optimization problem (11) can be examined one-to-one, aiming to find the optimal solution. In any case, the necessary conditions presented in Proposition 2 can be examined simultaneously to have a more comprehensive view in the process of searching for the optimal solution.
Next, an illustrative application of the described methodology is presented regarding the estimation (revelation) of the two-sided jump components of asset returns.

5. Application; Estimation of the Two-Sided Jump Components of the NASDAQ Index

In this section, an application example of the proposed methodology analyzed in Section 4 is illustrated concerning the estimation of the hidden two-sided jump components of the NASDAQ index for the 3 y period 2006–2008. To estimate the parameters of the model, i.e., the parameter set ϕ = ( G ( 1 ) , G ( 2 ) , σ x 2 , σ y 2 , V ) , the maximum likelihood estimation method is used taking into consideration that the distribution of R t conditioned on z t is normal, i.e.,
R t | z t N ( H z ^ t , HP t H T + V ) .
Therefore, the log-likelihood function, LogL, is of the form:
L o g L ( R 1 , , R n ) = n / 2 log ( 2 π ) 0.5 t = 1 n ( l o g ( | ω t | ) + u t T ω t 1 u t )
where,
u t = R t H z ^ t and ω t = HP t H T + V .
The estimations derived by maximizing L o g L , given in (26), are as follows:
G ( 1 ) = 5.4741 2.8498 2.8498 7.3474 , G ( 2 ) = 7.4368 1.4909 1.4909 2.8304
and:
σ x 2 = 0.9897 × 10 3 , σ y 2 = 0.86281 × 10 3 , V = 4.961 × 10 11 ,
with L o g L = 995.9854. Based on the estimated parameters, the estimated two-sided jump components of the NASDAQ index are showcased in Figure 1.

6. Conclusions

In this work, the topic of state space modeling with non-negative constraints was considered. For that purpose, a state space model was constructed where the state equation that describes the dynamic evolution of the components of the hidden state vector was expressed via non-negative definite quadratic forms and represents a non-negative valued Markovian stochastic process of order one. Due to the inequality conditions, a constrained optimization problem arises to derive estimators for the states, which are unbiased and of minimum variance. Towards this direction, a thorough analysis was illustrated via Propositions 1 and 2, concerning the stationary points of the optimization problem along with the special conditions that have to be satisfied in order to derive non-negative estimations for the state vectors at every time. Thus, in Proposition 2, necessary conditions were derived for a stationary point to constitute a feasible solution. The proposed method constitutes an alternative for handling state space models with non-negativity constraints, and it has a low computational burden compared to resampling methods for the estimation procedure.
Regarding future work, the generalization of the proposed method for the case of an n-dimensional non-negative state vector, n > 2 , could be examined. This is a challenging problem in many applications. For example, in navigation problems, for n = 3 , state space models with non-negativity constraints are suitable to describe the distance covered during the motion of a vehicle, if we let the three non-negative components of the state vector represent the measures of the velocities (speeds) along the axes in R 3 .

Author Contributions

Methodology, O.T. and G.T.; Writing—original draft, O.T. and G.T.; Writing—review & editing, O.T. and G.T. Both authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used to illustrate the applicability of the proposed method is publicly available at www.finance.yahoo.com accessed on 11 June 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Julier, S.; LaViola, J. On kalman filtering with nonlinear equality constraints. IEEE Trans. Signal Process. 2007, 55, 2774–2784. [Google Scholar] [CrossRef] [Green Version]
  2. Loumponias, K.; Vretos, N.; Tsaklidis, G.; Daras, P. An Improved Tobit Kalman Filter with Adaptive Censoring Limits. Circuits Syst. Signal Process. 2020, 39, 5588–5617. [Google Scholar] [CrossRef]
  3. Wang, L.; Chiang, Y.; Chang, F. Filtering method for nonlinear systems with constraints. Control. Theory Appl. IEE Proc. 2002, 149, 525–531. [Google Scholar] [CrossRef]
  4. Chia, T.; Chow, P.; Chizeck, H. Recursive parameter identification of constrained systems: An application to electrically stimulated muscle. IEEE Trans. Biomed. Eng. 1991, 38, 429–442. [Google Scholar] [CrossRef] [PubMed]
  5. Vasicek, O. An equilibrium characterization of the term structure. J. Financ. Econ. 1997, 5, 177–188. [Google Scholar] [CrossRef]
  6. Hull, J.; White, A. Pricing interest rate derivatives. Rev. Financ. Stud. 1990, 3, 573–592. [Google Scholar] [CrossRef]
  7. Cogleya, T.; Sargent, T. Drifts and volatilities: Monetary policies and outcomes in the post WWII US. Rev. Econ. Dyn. 2005, 8, 262–302. [Google Scholar] [CrossRef] [Green Version]
  8. Cogley, T.; Sargent, T. Evolving post World War II US inflation dynamics. Nber Macroecon. Annu. 2001, 16, 331–373. [Google Scholar] [CrossRef]
  9. Theodosiadou, O.; Tsaklidis, G. Estimating the Positive and Negative Jumps of Asset Returns via Kalman Filtering: The Case of NASDAQ Index. J. Methodol. Comput. Appl. Probab. 2017, 19, 1123–1134. [Google Scholar] [CrossRef]
  10. Theodosiadou, O.; Skaperas, S.; Tsaklidis, G. Change Point Detection and Estimation of the Two-Sided Jumps of Asset Returns using a Modified Kalman Filter. J. Risks 2017, 5, 15. [Google Scholar] [CrossRef] [Green Version]
  11. Cont, R. Empirical properties of asset returns: Stylized facts and statistical issues. Quant. Financ. 2001, 1, 223–236. [Google Scholar] [CrossRef]
  12. Kalman, R. A new approach to linear filtering and prediction problems. J. Basic Eng. Trans. ASME Ser. 1960, 82, 35–45. [Google Scholar] [CrossRef] [Green Version]
  13. Gordon, N.; Salmond, D.; Smith, A. Novel Approach to Nonlinear/Non-Gaussian Bayesian State Estimation. Radar Signal Process. IEE Proc. F 1993, 140, 107–113. [Google Scholar] [CrossRef] [Green Version]
  14. Doucet, A.; Johansen, A. A Tutorial on Particle Filtering and Smoothing: Fifteen years later. In Handbook of Nonlinear Filtering; Oxford University Press: Oxford, UK, 2008. [Google Scholar]
  15. Doucet, A.; de Freitas, N.; Gordon, N. (Eds.) Sequential Monte Carlo Methods in Practice; Springer: New York, NY, USA, 2001. [Google Scholar]
  16. Urteaga, I.; Bugallo, M.F.; Djurić, P.M. Sequential Monte Carlo methods under model uncertainty. In Proceedings of the 2016 IEEE Statistical Signal Processing Workshop (SSP), Palma de Mallorca, Spain, 26–29 June 2016; pp. 1–5. [Google Scholar]
  17. Kotz, S.; Balakrinshnan, N.; Johnson, N. Continouous Multivariate Distributions; Wiley: New York, NY, USA, 1995; Volume 1. [Google Scholar]
  18. Griva, I.; Nash, S.; Sofer, A. Linear and Nonlinear Optimization, 3rd ed.; SIAM: Philadelphia, PA, USA, 2008. [Google Scholar]
Figure 1. (a) Estimated positive return jumps of the NASDAQ index during 2006–2008. (b) Estimated negative return jumps of the NASDAQ index during 2006–2008.
Figure 1. (a) Estimated positive return jumps of the NASDAQ index during 2006–2008. (b) Estimated negative return jumps of the NASDAQ index during 2006–2008.
Mathematics 09 01908 g001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Theodosiadou, O.; Tsaklidis, G. State Space Modeling with Non-Negativity Constraints Using Quadratic Forms. Mathematics 2021, 9, 1908. https://doi.org/10.3390/math9161908

AMA Style

Theodosiadou O, Tsaklidis G. State Space Modeling with Non-Negativity Constraints Using Quadratic Forms. Mathematics. 2021; 9(16):1908. https://doi.org/10.3390/math9161908

Chicago/Turabian Style

Theodosiadou, Ourania, and George Tsaklidis. 2021. "State Space Modeling with Non-Negativity Constraints Using Quadratic Forms" Mathematics 9, no. 16: 1908. https://doi.org/10.3390/math9161908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop