Next Article in Journal
Generalized Multiscale Finite Element Method and Balanced Truncation for Parameter-Dependent Parabolic Problems
Next Article in Special Issue
A Novel Approach to Modeling Incommensurate Fractional Order Systems Using Fractional Neural Networks
Previous Article in Journal
Certain Quantum Operator Related to Generalized Mittag–Leffler Function
Previous Article in Special Issue
Continuous-Time Subspace Identification with Prior Information Using Generalized Orthonormal Basis Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parameter Identification of the Discrete-Time Stochastic Systems with Multiplicative and Additive Noises Using the UD-Based State Sensitivity Evaluation

by
Andrey Tsyganov
1,*,† and
Yulia Tsyganova
2,†
1
Department of Mathematics, Physics and Technology Education, Ulyanovsk State University of Education, Ulyanovsk 432071, Russia
2
Department of Mathematics, Information and Aviation Technology, Ulyanovsk State University, Ulyanovsk 432017, Russia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2023, 11(24), 4964; https://doi.org/10.3390/math11244964
Submission received: 24 November 2023 / Revised: 6 December 2023 / Accepted: 13 December 2023 / Published: 15 December 2023
(This article belongs to the Special Issue New Trends on Identification of Dynamic Systems)

Abstract

:
The paper proposes a new method for solving the parameter identification problem for a class of discrete-time linear stochastic systems with multiplicative and additive noises using a numerical gradient-based optimization. The constructed method is based on the application of a covariance UD filter for the above systems and an original method for evaluating state sensitivities within the numerically stable, matrix-orthogonal MWGS transformation. In addition to the numerical stability of the proposed algorithm to machine roundoff errors due to the application of the MWGS-UD orthogonalization procedure at each step, the main advantage of the obtained results is the possibility of analytical calculation of derivatives at a given value of the identified parameter without the need to use finite-difference methods. Numerical experiments demonstrate how the obtained results can be applied to solve the parameter identification problem for the considered stochastic system model.

1. Introduction

This paper addresses a parameter identification problem for a class of discrete-time linear stochastic systems represented by state-space difference equations with additive and multiplicative noises. A distinctive feature of this class of dynamic systems is multiplicative noise, which can be included in both the state and measurement equations [1]. The reasons for the appearance of multiplicative noise are different; for example, it can be due to linearization, quantization, and modeling errors, or physical phenomena such as fading in communication channels. Most often, systems with additive and multiplicative noises are considered when solving problems related to various kinds of measurement processing.
The parameter identification problem consists of determining the unknown parameters of a mathematical model of the system belonging to a selected class of models using known input and output measurement data [2]. As noted in [3,4], the main approaches to solving parameter identification problems, which remain of current interest today, are subspace identification methods and minimum prediction error (MPE) methods. The first approach is based on the application of projections in Euclidean space, and the second one is based on minimizing the identification criterion depending on the system parameters. The foundation of these approaches was laid in the seminal works [5,6]. A great contribution to the development of MPE methods was made by Lennart Ljung, who defined the basic concepts of MPE methodology [2,4].
Linear discrete-time stochastic systems with associated Kalman-type filtering algorithms have been extensively used in practice. As a rule, application of the filter equations assumes a complete a priori knowledge of the system model parameters, but it is a rare case. The classical way of solving the parameter identification problem is to use adaptive filters where the model parameters are estimated together with the dynamic state [7]. This requires determination of the sensitivities of the system state to unknown parameters, i.e., partial derivatives of the state estimates. Straightforward differentiation of the filter equations is a direct approach to compute the state sensitivities. This leads to a set of filter sensitivity equations. It is well known that, for discrete-time linear stochastic systems with additive noises, a conventional Kalman filter may suffer from numerical instability caused by machine roundoff errors [8]. This is also true for such systems with multiplicative noises [9]. Nevertheless, there are currently many numerically stable modifications of the conventional Kalman filtering algorithm, which are based on various methods for factorizing the estimation error covariance matrices [8,10]. It is worth noting that, for discrete-time stochastic systems with multiplicative noises, such modifications were developed relatively recently (see, for example, [9,11]).
In this paper, we chose a numerically stable UD-based modification of the Kalman-type filtering algorithm [9] for solving the parameter identification problem, since it has the following attractive numerical properties: numerical stability to machine roundoff errors, absence of square root and matrix inversion operations, and the compactness and simplicity of the block matrix array form for implementation on a computer, including parallel computing [10].
Thus, the purpose of our work is to develop a numerically stable UD-based method for solving the parameter identification problem for a class of discrete-time linear stochastic systems with multiplicative and additive noises using a numerical, gradient-based optimization of the identification criterion in the form of a negative logarithmic likelihood function.
We propose the following solution steps to achieve the goal:
  • Replace the conventional Kalman-type filtering algorithm with a UD-based covariance filtering algorithm that is numerically stable to machine roundoff errors.
  • Construct a new method for calculating the state sensitivity values in the adaptive UD-based filter.
  • Apply this method to the gradient-based minimization of the identification criterion.
The paper is organized as follows. Section 2 provides the problem statement of parameter identification for a considered class of stochastic systems. Then, the UD-based covariance filtering algorithm for discrete-time stochastic systems with multiplicative and additive noises is described. Lemma 1 provides the method of obtaining the main results presented in the next section. Section 3 contains the new UD-based state sensitivity evaluation method, which is described in Proposition 1. Proposition 2 explains how the values of the identification criterion and its gradient can be calculated using the suggested algorithm. Section 4 demonstrates how these two methods can be applied for solving the parameter identification problem of the considered stochastic system model. Section 5 concludes the paper.

2. Methods

2.1. The Problem Statement

Consider a discrete-time linear stochastic system with multiplicative and additive noises
x k = ( F k 1 + F ˜ k 1 ξ k 1 ) x k 1 + G k 1 w k 1 , z k = ( H k + H ˜ k ζ k ) x k + v k , k = 1 , 2 , , M ,
where x k R n is the system state vector; z k R m is the measurement vector; matrices F k , F ˜ k R n × n ; H k , H ˜ k R m × n ; G R n × q ; M is the number of measurements; x 0 N ( x ¯ 0 , Π 0 ) is the initial state; ξ k R N ( 0 , σ ξ 2 ) and w k R q N ( 0 , Q k )   are multiplicative and additive noises in the state equation, respectively; ζ k R N ( 0 , σ ζ 2 ) and v k R m N ( 0 , R k )  are multiplicative and additive noises in the measurement equation, respectively; covariance matrices Q k and R k of noises w k and v k , respectively, are positive definite; and all noises and the initial state are mutually independent.
The considered systems find their application in wireless sensor networks, navigation, and other fields. Multiplicative noises are commonly used to account for stochastic uncertainties in system dynamics and measurements. In this paper, we assume that system (1) additionally contains parametric uncertainty, i.e., the matrices defining the equations of system (1) depend on the unknown parameter θ R p . Consequently, all elements of the system matrices F k , F ˜ k , G k , H k , H ˜ k , noise covariance matrices Q k and R k , and variances σ ξ 2 , σ ζ 2 , as well as the initial conditions x ¯ 0 and Π 0 , may depend on the unknown parameter θ ; i.e., F k = F k ( θ ) , F ˜ k = F ˜ k ( θ ) , G k = G k ( θ ) , etc.
Thus, the parameter identification problem of system (1) arises. It is known that the performance of the Kalman filter degrades in the presence of modeling uncertainties. Solving the problem of parameter identification makes it possible to cope with this problem. The well-known approaches involve numerical minimization by θ of the identification criterion [12]
θ ^ m i n = argmin θ D ( θ ) J ( θ ; Z 1 M ) ,
which implies solving an optimization problem with constraints, where D ( θ ) represents some compact set, such as a segment in the scalar case.
To solve the problem of parameter identification, we chose an identification criterion (2) in the form of the negative logarithmic likelihood function [13]:
J ( θ ; Z 1 M ) = M m 2 ln 2 π + 1 2 k = 1 M ln det Σ k ( θ ) + | | ν k ( θ ) | | Σ k 1 ( θ ) 2 ,
depending on the available measurement information, which includes both measurements Z 1 M = { z 1 , , z k , , z M } themselves and the observed measurement residuals ν k ( θ ) = z k H k x ^ k ( θ ) , calculated by conventional Kalman-type filtering algorithm (Algorithm 1 in [9]). Here, | | ν k ( θ ) | | Σ k 1 ( θ ) 2 = ν k T ( θ ) Σ k 1 ( θ ) ν k ( θ ) .
In this paper, our first task is to replace the conventional Kalman-type filtering algorithm with a numerically stable, UD-based covariance filtering algorithm.

2.2. The UD-Based Covariance Filtering Algorithm for Discrete-Time Stochastic Systems with Multiplicative and Additive Noises

Originally, two algorithms based on the U D U T decomposition of covariance and information matrices for discrete-time stochastic systems with multiplicative and additive noises were proposed in [9]. The first of these algorithms is the UD-based covariance filter, and the second one is the UD-based information filter. Both of these filters have an extended array form, and their computational schemes allow for updating all required filter quantities with the use of the numerically stable MWGS orthogonalization procedure.
The UD-based implementations imply the decomposition of the error covariance matrix in the form of P = U P D P U P T , where U P is an upper triangular matrix with 1’s on the main diagonal, and D P is a diagonal matrix. To recursively update the resulting UD-factors U P and D P , we use the modified weighted Gram–Schmidt UD-based (MWGS-UD) orthogonalization [14] as follows: given a pair of the pre-arrays { A , D A } , compute a pair of the post-arrays { U , D U } by means of the MWGS-UD orthogonalization, i.e.,
A T = U M T and A T D A A = U D U U T ,
where A R r × s , r > s , and M R r × s is the MWGS-UD transformation that produces the block upper triangular matrix U R s × s . The diagonal matrices D A R r × r and D U R s × s satisfy M T D A M = D U and D A > 0 (see Lemma VI.4.1 in [14] for an extended explanation).
In this paper, we consider a modification of the UD-based covariance filter [9], whose equations are presented in Algorithm 1. The proof of Algorithm 1 is similar to the proof presented in [9].   
Algorithm 1: UD-Based Covariance Filter (UD-CF)
Input:  x ¯ 0 , Π 0 .
Initialization
1. Set x ^ 0 = x ¯ 0 , X 0 = Π 0 + x ¯ 0 x ¯ 0 T , Π 0 = U Π 0 D Π 0 U Π 0 T , X 0 = U X 0 D X 0 U X 0 T ,
U P 0 = U Π 0 , D P 0 = D Π 0 .
For  k = 1 , 2 , , M  do
Time Update
2. x ^ k | k 1 = F k 1 x ^ k 1 .
3. Q k 1 = U Q k 1 D Q k 1 U Q k 1 T .
4. A 1 T = F ˜ k 1 U X k 1 G k 1 U Q k 1 , D A 1 = σ ξ 2 D X k 1 0 0 D Q k 1 .
  Using MWGS-UD A 1 T , D A 1 U 1 , D U 1 where  U 1 = U Q ˜ k 1 , D U 1 = D Q ˜ k 1 .
5. A 2 T = F k 1 U P k 1 U Q ˜ k 1 , D A 2 = D P k 1 0 0 D Q ˜ k 1 .
  Using MWGS-UD A 2 T , D A 2 U 2 , D U 2 where  U 2 = U P k | k 1 , D U 2 = D P k | k 1 .
6. A 3 T = F k 1 U X k 1 U Q ˜ k 1 , D A 3 = D X k 1 0 0 D Q ˜ k 1 .
  Using MWGS-UD A 3 T , D A 3 U 3 , D U 3 where  U 3 = U X k , D U 3 = D X k .
Measurement Update
7. R k = U R k D R k U R k T .
8. A 4 T = H ˜ k U X k U R k , D A 4 = σ ζ 2 D X k 0 0 D R k .
  Using MWGS-UD A 4 T , D A 4 U 4 , D U 4 where  U 4 = U R ˜ k , D U 4 = D R ˜ k .
9. A 5 T = U P k | k 1 0 H k U P k | k 1 U R ˜ k , D A 5 = D P k | k 1 0 0 D R ˜ k .
  Using MWGS-UD A 5 T , D A 5 U 5 , D U 5 where  U 5 = U P k ( K k U Σ k ) 0 U Σ k ,
D U 5 = D P k 0 0 D Σ k .
10. e ¯ k = U Σ k 1 ( z k H k x ^ k | k 1 ) .
11. x ^ k = x ^ k | k 1 + ( K k U Σ k ) e ¯ k .
End For
Output:  x ^ k , P k = U P k D P k U P k T , k = 1 , 2 , , M .
Remark 1.
Steps 3 and 7 of Algorithm 1 require the application of the U D U T modified Cholesky decomposition [15]. Steps 4–6, 8, and 9 require the application of the MWGS-UD orthogonalization to a pair of block matrices A i T , D A i that produces another pair of block matrices U i , D U i ( i = 1 , , 5 ) so that the equalities (4) are satisfied.
Let us rewrite the identification criterion (3) in terms of the UD-CF algorithm. Taking into account that det Σ k ( θ ) = det D Σ k ( θ ) and | | ν k ( θ ) | | Σ k 1 ( θ ) 2 = | | e ¯ k ( θ ) | | D Σ k ( θ ) 1 2 , we obtain
J U D ( θ ; Z 1 M ) = M m 2 ln 2 π + 1 2 k = 1 M ln det D Σ k ( θ ) + | | e ¯ k ( θ ) | | D Σ k ( θ ) 1 2 ,
where diagonal matrix D Σ k and normalized residual vector e ¯ k are calculated in Steps 9 and 10 of Algorithm 1.

2.3. Derivative Evaluation of the MWGS-Based Array of Block Matrices

Solving the parameter identification problem requires minimization of the identification criterion (5) with respect to unknown system parameters. It is often done by using the gradient approach [7], where the computation of θ J U D ( θ ; Z 1 M ) is necessary. For the discrete-time stochastic system (1), the J U D ( θ ; Z 1 M ) and θ J U D ( θ ; Z 1 M ) evaluation demands an implementation of the UD-CF and of the so-called “differentiated” UD-CF to determine the state sensitivities of the system state to the unknown system parameters [7,16,17]. The θ J U D ( θ ; Z 1 M ) computation will lead to a set of p filter sensitivity equations for computing x ^ k / θ and a set of p matrix Riccati-type sensitivity equations for computing P k / θ .
Such a method of state sensitivity evaluation within the UD-based array covariance filter was proposed in [17] for a class of discrete-time linear stochastic systems with only additive noises. In this paper, we extend this approach to systems with multiplicative and additive noises. Solving the problem of state sensitivity evaluation, we augment the numerical scheme of the UD-CF (Algorithm 1) with a procedure for numerically efficient evaluation of the derivatives of the UD-based filter variables with respect to unknown system parameters.
We use the following basic result of [17] that gives a simple and convenient technique which naturally augments any MWGS-based array of block matrices for computing derivatives of its elements.
Lemma 1
([17]). Let entries of the pre-arrays A , D A in (4) be known differentiable functions of a parameter θ. Consider the transformation in (4). Given the derivatives of the pre-arrays A θ and ( D U ) θ , the following formulas calculate the corresponding derivatives of the post-arrays:
U θ = U L ¯ 0 T + U ¯ 0 + U ¯ 2 D U 1 a n d D U θ = 2 D 0 + D 2 ,
where the quantities L ¯ 0 , D 0 , and U ¯ 0 are, respectively, the strictly lower triangular, diagonal, and strictly upper triangular parts of the matrix product M T D A A θ U T . Additionally, D 2 and U ¯ 2 are the diagonal and strictly upper triangular parts of the product M T ( D A ) θ M , respectively.

3. Main Results

One can see from Algorithm 1 that elements x ^ k and the UD factors U P k and D P k are readily available from this UD-based filter. Hence, our aim is to augment equations of the UD-CF so that the derivatives x ^ k ( θ ) / θ i and U P k ( θ ) / θ i , D P k ( θ ) / θ i , i = 1 , , p can be computed using quantities available from this UD-CF algorithm.

3.1. The New UD-Based State Sensitivity Evaluation Method

Now, we are ready to present our new result—the UD-based state sensitivity evaluation method for a class of discrete-time linear stochastic systems with multiplicative and additive noises.
Proposition 1.
Let the elements of matrices defining system (1) be known differentiable functions of a parameter θ. Then, for a given value of parameter θ, the estimates of state vector x k , their sensitivity values x ^ k / θ , and the UD factors U P k , D P k of the error covariance matrices and their sensitivity values U P k / θ , D P k / θ can be evaluated simultaneously using the subsequent Algorithm 2.
Algorithm 2: State Sensitivity Evaluation within Adaptive UD-CF
Input:  x ¯ 0 , Π 0 , θ = θ i , i = 1 , , p .
Initialization
1. Set x ^ 0 = x ¯ 0 , X 0 = Π 0 + x ¯ 0 x ¯ 0 T , Π 0 = U Π 0 D Π 0 U Π 0 T , X 0 = U X 0 D X 0 U X 0 T .
( x ^ 0 ) θ = x ^ 0 θ , ( U P 0 ) θ = U P 0 θ , ( D P 0 ) θ = D P 0 θ , ( U X 0 ) θ = U X 0 θ , ( D X 0 ) θ = U X 0 θ .
For  k = 1 , 2 , , M  do
Time Update
2. x ^ k | k 1 = F k 1 x ^ k 1 , ( x ^ k | k 1 ) θ = F k 1 θ x ^ k 1 + F k 1 ( x ^ k 1 ) θ .
3. Q k 1 = U Q k 1 D Q k 1 U Q k 1 T , ( U Q k 1 ) θ = U Q k 1 θ , ( D Q k 1 ) θ = D Q k 1 θ .
4. A 1 T = F ˜ k 1 U X k 1 G k 1 U Q k 1 , ( A 1 T ) θ = A 1 T θ ;
D A 1 = σ ξ 2 D X k 1 0 0 D Q k 1 , ( D A 1 ) θ = D A 1 θ .
  By Lemma 1 A 1 T , ( A 1 T ) θ , D A 1 , ( D A 1 ) θ U 1 , ( U 1 ) θ , D U 1 , ( D U 1 ) θ
  where  U 1 = U Q ˜ k 1 , ( U 1 ) θ = ( U Q ˜ k 1 ) θ , D U 1 = D Q ˜ k 1 , ( D U 1 ) θ = ( D Q ˜ k 1 ) θ .
5. A 2 T = F k 1 U P k 1 U Q ˜ k 1 , ( A 2 T ) θ = A 2 T θ ;
D A 2 = D P k 1 0 0 D Q ˜ k 1 , ( D A 2 ) θ = D A 2 θ .
  By Lemma 1 A 2 T , ( A 2 T ) θ , D A 2 , ( D A 2 ) θ U 2 , ( U 2 ) θ , D U 2 , ( D U 2 ) θ
  where  U 2 = U P k | k 1 , ( U 2 ) θ = ( U P k | k 1 ) θ , D U 2 = D P k | k 1 , ( D U 2 ) = ( D P k | k 1 ) θ .
6. A 3 T = F k 1 U X k 1 U Q ˜ k 1 , ( A 3 T ) θ = A 3 T θ ;
D A 3 = D X k 1 0 0 D Q ˜ k 1 , ( D A 3 ) θ = D A 3 θ .
  By Lemma 1 A 3 T , ( A 3 T ) θ , D A 3 , ( D A 3 ) θ U 3 , ( U 3 ) θ , D U 3 , ( D U 3 ) θ
  where  U 3 = U X k , ( U 3 ) θ = ( U X k ) θ , D U 3 = D X k , ( D U 3 ) θ = ( D X k ) θ .
Measurement Update
7. R k = U R k D R k U R k T , ( U R k ) θ = U R k θ , ( D R k ) θ = D R k θ .
8. A 4 T = H ˜ k U X k U R k , ( A 4 T ) θ = A 4 T θ ; D A 4 = σ ζ 2 D X k 0 0 D R k , ( D A 4 ) θ = D A 4 θ .
  By Lemma 1 A 4 T , ( A 4 T ) θ , D A 4 , ( D A 4 ) θ U 4 , ( U 4 ) θ , D U 4 , ( D U 4 ) θ
  where  U 4 = U R ˜ k , ( U 4 ) θ = ( U R ˜ k ) θ , D U 4 = D R ˜ k , ( D U 4 ) θ = ( D R ˜ k ) θ .
9. A 5 T = U P k | k 1 0 H k U P k | k 1 U R ˜ k , ( A 5 T ) θ = A 5 T θ ;
D A 5 = D P k | k 1 0 0 D R ˜ k , ( D A 5 ) θ = D A 5 θ .
  By Lemma 1 A 5 T , ( A 5 T ) θ , D A 5 , ( D A 5 ) θ U 5 , ( U 5 ) θ , D U 5 , ( D U 5 ) θ
  where  U 5 = U P k ( K k U Σ k ) 0 U Σ k , ( U 5 ) θ = ( U P k ) ( K k U Σ k ) θ 0 ( U Σ k ) θ ,
D U 5 = D P k 0 0 D Σ k , ( D U 5 ) θ = ( D P k ) θ 0 0 ( D Σ k ) θ .
10. e ¯ k = U Σ k 1 ( z k H k x ^ k | k 1 ) ,
( e ¯ k ) θ = ( U Σ k 1 ) θ ( z k H k x ^ k | k 1 ) U Σ k 1 H k θ x ^ k | k 1 + H k ( x ^ k | k 1 ) θ .
11. x ^ k = x ^ k | k 1 + ( K k U Σ k ) e ¯ k , ( x ^ k ) θ = ( x ^ k | k 1 ) θ + ( K k U Σ k ) θ e ¯ k + ( K k U Σ k ) ( e ¯ k ) θ .
End For
Output:  x ^ k , x ^ k / θ , U P k , U P k / θ , D P k , D P k / θ , k = 1 , 2 , , M .
Proof. 
Steps 1, 3, and 7 of Algorithm 2 are obtained from the corresponding steps of Algorithm 1 by direct differentiation of the vector x ^ 0 and the UD factors of the covariance matrices P 0 , Q k 1 , and R k with respect to parameter θ .
Steps 2, 10, and 11 of Algorithm 2 are obtained from the corresponding steps of Algorithm 1 by direct differentiation of the equations with respect to parameter θ .
Let us consider the equations that define Steps 4–6, 8, and 9 in Algorithm 2. They have the same form and are obtained using Lemma 1 as follows. At each step, the MWGS-UD procedure orthogonalizes the columns of the block matrix A i with respect to the weight matrix D A i so that the equalities in (4) are satisfied. To calculate sensitivity values ( U i ) θ and ( D U i ) θ for a given value of parameter θ , it is necessary and sufficient to find the values of the partial derivatives of the elements of the block matrices A i and D A i , i.e., calculate the matrices ( A i ) θ and ( D A i ) θ . Lemma 1 with A : = A i , D A : = D A i , and U : = U i , D U : = D U i ( i = 1 , , 5 ) allows for achieving this goal. As a result, we find block matrices ( U i ) θ and ( D U i ) θ , from which we obtain the required sensitivity values.    □
Thus, Proposition 1 gives the method for state sensitivity evaluation using the UD-based covariance filter for discrete-time linear stochastic systems with multiplicative and additive noises.

3.2. The UD-Based Computation of the Identification Criterion and Its Gradient

Algorithm 2 can be considered as an adaptive filter, in which parameter θ is adjusted according to the minimum of criterion J U D ( θ ; Z 1 M ) . When solving the parameter identification problem using gradient-based algorithms, adaptive filters are used, supplemented by a sensitivity model equations [18] to calculate the gradient of the identification criterion.
Implementation of gradient-based numerical methods for minimizing identification criterion (5) requires calculating the values of its gradient θ J U D ( θ ; Z 1 M ) .
The conventional gradient-based method has the following iterative form:
θ j + 1 = θ j β j θ J U D ( θ j ; Z 1 M ) ,
where θ j is the parameter vector at the jth iteration. In (7), θ denotes the gradient operator / θ 1 | | / θ p T ( θ R p ), which is applied here to (5) at point θ = θ j . Scalar step size parameter β j is designed to ensure that J U D ( θ j + 1 ; Z 1 M ) J U D ( θ j ; Z 1 M ) + e , where e is a positive number that can be chosen in a variety of ways [19].
Let us write the equation we could use to evaluate the gradient of the identification criterion (3) in terms of the UD-CF algorithm. Let θ R p . Then, from (5), we have
J U D ( θ ; Z 1 M ) θ i = 1 2 k = 1 M ln det D Σ k ( θ ) θ i + e ¯ k T ( θ ) D Σ k ( θ ) 1 e ¯ k ( θ ) θ i , i = 1 , , p .
Taking into account matrix differentiation rules, we rewrite (8) to obtain the expression for θ J U D ( θ ; Z 1 M ) evaluation ( i = 1 , , p ):
J U D ( θ ; Z 1 M ) θ i = 1 2 k = 1 M tr D Σ k ( θ ) θ i D Σ k ( θ ) 1 + 2 e ¯ k T ( θ ) θ i D Σ k ( θ ) 1 e ¯ k ( θ ) e ¯ k T ( θ ) D Σ k ( θ ) 2 D Σ k ( θ ) θ i e ¯ k ( θ ) .
Proposition 2.
Let the elements of matrices defining system (1) be known differentiable functions of a parameter θ R p . Then, for a given value of parameter θ, the values of identification criterion J U D ( θ ; Z 1 M ) and its gradient θ J U D ( θ ; Z 1 M ) can be evaluated according to Equations (5) and (9) using Algorithm 2.
Proof. 
Calculating the values of identification criterion J U D ( θ ; Z 1 M ) and its gradient θ J U D ( θ ; Z 1 M ) according to (5) and (9) requires the values of e ¯ k ( θ ) , D Σ k ( θ ) , and their sensitivities e ¯ k ( θ ) / θ i , D Σ k ( θ ) / θ i for k = 1 , , M ; i = 1 , , p . The values of these quantities are available in Steps 10 and 11 of Algorithm 2.    □
Thus, for a given value of parameter θ , Algorithm 2 allows us to obtain all the quantities necessary to calculate the values of the identification criterion and its gradient that are used in the parameter identification gradient-based method.

4. Discussion

As a practical application of the proposed method, let us consider the parameter identification problem for a nearly constant velocity model of the uniform motion augmented with multiplicative noises [11]:
x k = 1 θ 0 1 + 0 0 0 1 ξ k 1 x k 1 + θ 2 2 θ w k 1 , z k = 1 0 0 1 + 0 0 0 1 ζ k x k + v k , k = 1 , , 100 ,
where x k = [ x 1 , x 2 ] k T , x 1 = x is the coordinate of the object, x 2 = v x is its velocity, x 0 N ( [ 0 , 1 ] T , 10 I 2 ) , w k N ( 0 , 10 2 ) , v k N ( 0 , σ 2 I 2 ) ( σ = 0.1 , 0.5 , 1.0 ), ξ k N ( 0 , 10 4 ) , ζ k N ( 0 , 10 4 ) , and θ is the model parameter to be identified. Let us put the “true” value of the parameter equal to θ * = 0.3 .
To demonstrate the validity of the proposed approach, we have conducted numerical experiments in MATLAB, which is a common tool for simulation, optimization, control, and filtering [8,20,21]. We have implemented all necessary functions for simulating system dynamics and measurements, as well as functions for calculating the identification criterion (5) and its gradient (9). Figure 1 shows the identification criterion and its gradient for the considered problem, with noise covariance matrix R = 0 . 5 2 I 2 averaged over 100 runs.
The following MATLAB functions were used for numerical minimization of the identification criterion: simulannealbnd, ga, and fmincon. The first two functions implement gradient-free metaheuristic algorithms SA (Simulated Annealing) and GA (Genetic Algorithm), respectively. The third function was configured to use two different gradient-based algorithms: interior-point (IP) and trust-region-reflective (TRR). The first algorithm estimates the gradient using finite differences, and the second algorithm uses a user-provided gradient of the objective function.
A series of 100 numerical experiments was conducted for each value of the noise level σ . In each experiment, numerical identification of parameter θ was performed based on the results of simulated measurements. The solution, θ * , was searched on the segment [ 0.1 ; 1 ] . The initial point for SA and both gradient-based algorithms was chosen randomly in each experiment.
Table 1 presents the main settings of the optimizers used in the numerical experiments. The remaining settings are taken by default. All experiments were conducted on the following platform: MATLAB R2017a, Windows 11, Intel Core i3-1115G4 CPU @ 3.00 GHz, 8 GB of RAM.
Table 2 provides the average number of iterations and running times for all optimizers. It can be seen that, at the chosen settings, gradient-based algorithms run significantly faster than non-gradient algorithms. The TRR algorithm is about 10% faster and requires fewer iterations than the IP algorithm. The SA algorithm works on average two times faster than GA, whose performance largely depends on the population size, which in this series of experiments was taken to be 10. Reducing the population size can proportionally reduce the running time of this algorithm, but it can also lead to a decrease in accuracy.
The results of numerical identification of parameter θ are summarized in Table 3. They show that, for the problem under consideration with the selected settings, all algorithms demonstrate approximately the same mean accuracy. RMSE and MAPE values decrease with decreasing noise levels, but for the SA algorithm, they remain slightly larger than for other algorithms.

5. Conclusions

In this paper, we have proposed a new method for solving the parameter identification problem for discrete-time linear stochastic systems with multiplicative and additive noises based on the application of the covariance UD-filter and the original method for the state sensitivity evaluation within the numerically stable, matrix-orthogonal MWGS-UD transformation.
The main theoretical results of the paper are the UD-based state sensitivity evaluation method and the method of calculating the values of identification criterion J U D ( θ ; Z 1 M ) and its gradient, which are formulated in Propositions 1 and 2, respectively. Both methods use Algorithm 2 to calculate the state vector estimates and their sensitivity values. In addition to the numerical stability of the MWGS orthogonalization procedure to machine roundoff errors, the main advantage of the proposed method is the possibility of analytical calculation of derivative values at a given value of the identified parameter without the need to use finite-difference methods.
Numerical experiments confirm that the obtained results can be used for solving the parameter identification problems of the considered stochastic systems. The gradient-based minimization of the identification criterion that uses the proposed method outperforms both non-gradient algorithms and the algorithm that estimates gradients using finite differences.

Author Contributions

Conceptualization, A.T. and Y.T.; methodology, A.T. and Y.T.; software, A.T.; validation, Y.T.; formal analysis, A.T. and Y.T.; investigation, A.T. and Y.T.; resources, A.T. and Y.T.; data curation, A.T. and Y.T.; writing—original draft preparation, Y.T.; writing—review and editing, A.T. and Y.T.; visualization, A.T. and Y.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Russian Science Foundation at Ulyanovsk State University of Education, grant no. 22-21-00387, https://rscf.ru/en/project/22-21-00387/ (accessed on 24 November 2023).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MWGSModified weighted Gram–Schmidt orthogonalization
MWGS-UDMWGS based on the UD factorization
KFKalman filter
TUTime update
MUMeasurement update
MPEMinimum prediction error
GAGenetic algorithm
SASimulated annealing
RMSERoot mean square error
MAPEMean absolute percentage error
IPInterior point algorithm
TRRTrust region reflective algorithm

References

  1. Wu, Y.; Zhang, Q.; Shen, Z. Kalman filtering with multiplicative and additive noises. In Proceedings of the 12th World Congress on Intelligent Control and Automation (WCICA 2016), Guilin, China, 12–15 June 2016; pp. 483–487. [Google Scholar]
  2. Ljung, L. System Identification: Theory for the User, 2nd ed.; Prentice Hall PTR: Upper Saddle River, NJ, USA, 1999. [Google Scholar]
  3. Gevers, M. A personal view of the development of system identification: A 30-year journey through an exciting field. IEEE Control Syst. Mag. 2006, 26, 93–105. [Google Scholar]
  4. Ljung, L. Perspectives on system identification. Annu. Control 2010, 34, 1–12. [Google Scholar] [CrossRef]
  5. Åström, K.-J.; Bohlin, T. Numerical Identification of Linear Dynamic Systems from Normal Operating Records. In Proceedings of the Second IFAC Symposium on the Theory of Self-Adaptive Control Systems, Teddington, UK, 14–17 September 1966; pp. 96–111. [Google Scholar]
  6. Ho, B.L.; Kalman, R.E. Effective construction of linear state-variable models from input/output functions. Regelungstechnik 1966, 14, 545–548. [Google Scholar]
  7. Gupta, N.K.; Mehra, R.K. Computational aspects of maximum likelihood estimation and reduction in sensitivity function calculations. IEEE Trans. Autom. Control 1974, AC-19, 774–783. [Google Scholar] [CrossRef]
  8. Grewal, M.S.; Andrews, A.P. Kalman Filtering: Theory and Practice Using MATLAB, 4th ed.; John Wiley & Sons, Inc.: New York, NY, USA, 2015. [Google Scholar]
  9. Tsyganov, A.V.; Tsyganova, J.V.; Kureneva, T.N. UD-based Linear Filtering for Discrete-Time Systems with Multiplicative and Additive Noises. In Proceedings of the 19th European Control Conference, Saint Petersburg, Russia, 12–15 May 2020; pp. 1389–1394. [Google Scholar]
  10. Kailath, T.; Sayed, A.; Hassibi, B. Linear Estimation; Prentice Hall: Upper Saddle River, NJ, USA, 2000. [Google Scholar]
  11. Tsyganov, A.; Tsyganova, Y. SVD-Based Identification of Parameters of the Discrete-Time Stochastic Systems Models with Multiplicative and Additive Noises Using Metaheuristic Optimization. Mathematics 2023, 11, 4292. [Google Scholar] [CrossRef]
  12. Gibbs, B.P. Advanced Kalman Filtering, Least-Squares and Modeling: A Practical Handbook; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2011. [Google Scholar]
  13. Åström, K.-J. Maximum Likelihood and Prediction Error Methods. Automatica 1980, 16, 551–574. [Google Scholar] [CrossRef]
  14. Bierman, G.J. Factorization Methods for Discrete Sequential Estimation; Academic Press: New York, NY, USA, 1977. [Google Scholar]
  15. Golub, G.H.; Van Loan, C.F. Matrix Computations; Johns Hopkins University Press: Baltimore, MD, USA, 1983. [Google Scholar]
  16. Bierman, G.J.; Belzer, M.R.; Vandergraft, J.S.; Porter, D.W. Maximum likelihood estimation using square root information filters. IEEE Trans. Autom. Control 1990, 35, 1293–1298. [Google Scholar] [CrossRef]
  17. Tsyganova, J.V.; Kulikova, M.V. State sensitivity evaluation within UD based array covariance filters. IEEE Trans. Autom. Control 2013, 58, 2944–2950. [Google Scholar] [CrossRef]
  18. Tsypkin, Y.Z. Information Theory of Identification; Fizmatlit: Moscow, Russia, 1995. (In Russian) [Google Scholar]
  19. Nocedal, J.; Wright, S.J. Numerical Optimization; Springer Series in Operations Research and Financial Engineering; Springer Nature: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  20. Dorf, R.C.; Bishop, R.H. Modern Control Systems, 13th ed.; Pearson: Upper Saddle River, NJ, USA, 2016. [Google Scholar]
  21. Mu, D.; Xu, C.; Liu, Z.; Pang, Y. Further Insight into Bifurcation and Hybrid Control Tactics of a Chlorine Dioxide-Iodine-Malonic Acid Chemical Reaction Model Incorporating Delays. MATCH Commun. Math. Comput. Chem. 2023, 89, 529–566. [Google Scholar] [CrossRef]
Figure 1. Identification criterion (a) and its gradient (b).
Figure 1. Identification criterion (a) and its gradient (b).
Mathematics 11 04964 g001
Table 1. Optimizer settings.
Table 1. Optimizer settings.
OptimizerSettings
SA‘TimeLimit’ = Inf, ‘MaxFunEvals’ = Inf, ‘MaxIter’ = Inf,
‘StallIterLimit’ = 100, ‘ReannealInterval’ = 100, ‘Display’ = ‘off’
‘TimeLimit’ = Inf, ‘Generations’ = Inf, ‘StallGenLimit’ = 20,
GA‘PopulationSize’ = 10, ‘PopInitRange’ = [0.1; 1],
‘MutationFcn’ = @mutationadaptfeasible, ‘Display’ = ‘off’
FMINCON‘SpecifyObjectiveGradient’ = false, ‘Algorithm’ = ‘interior-point’,
(IP)‘MaxFunctionEvaluations’ = Inf, ‘Display’ = ‘off’
FMINCON‘SpecifyObjectiveGradient’ = true, ‘Algorithm’ = ‘trust-region-reflective’,
(TRR)‘MaxFunctionEvaluations’ = Inf, ‘Display’ = ‘off’
Table 2. Average number of iterations and time, in seconds.
Table 2. Average number of iterations and time, in seconds.
SAGAFMINCON
(IP)
FMINCON
(TRR)
σ = 1.0 Iterations18733137
Time2.8575.0440.6610.499
σ = 0.5 Iterations17636127
Time2.7395.5820.5680.509
σ = 0.1 Iterations18742127
Time2.8806.5060.6070.540
Table 3. Identification results.
Table 3. Identification results.
SAGAFMINCONFMINCON
(IP)(TRR)
Mean0.2874240.2878230.2878030.287802
σ = 1.0 RMSE0.0294940.0289420.0289620.028962
MAPE7.9672217.8189637.8279297.827960
Mean0.2992970.2998020.2998050.299805
σ = 0.5 RMSE0.0160560.0159950.0159960.015996
MAPE4.2177314.1836244.1838794.183878
Mean0.2981750.2994360.2994380.299438
σ = 0.1 RMSE0.0056880.0034320.0034290.003429
MAPE1.4194620.8810470.8801590.880159
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsyganov, A.; Tsyganova, Y. Parameter Identification of the Discrete-Time Stochastic Systems with Multiplicative and Additive Noises Using the UD-Based State Sensitivity Evaluation. Mathematics 2023, 11, 4964. https://doi.org/10.3390/math11244964

AMA Style

Tsyganov A, Tsyganova Y. Parameter Identification of the Discrete-Time Stochastic Systems with Multiplicative and Additive Noises Using the UD-Based State Sensitivity Evaluation. Mathematics. 2023; 11(24):4964. https://doi.org/10.3390/math11244964

Chicago/Turabian Style

Tsyganov, Andrey, and Yulia Tsyganova. 2023. "Parameter Identification of the Discrete-Time Stochastic Systems with Multiplicative and Additive Noises Using the UD-Based State Sensitivity Evaluation" Mathematics 11, no. 24: 4964. https://doi.org/10.3390/math11244964

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop