Next Article in Journal
On the Eventually Periodic Continued β-Fractions and Their Lévy Constants
Next Article in Special Issue
Unbiased Identification of Fractional Order System with Unknown Time-Delay Using Bias Compensation Method
Previous Article in Journal
An Algebraic Approach to Clustering and Classification with Support Vector Machines
Previous Article in Special Issue
A Simple Analytical Method for Estimation of the Five-Parameter Model: Second-Order with Zero Plus Time Delay
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Algorithmic Differentiation of the MWGS-Based Arrays for Computing the Information Matrix Sensitivity Equations within the Problem of Parameter Identification

by
Andrey Tsyganov
1,† and
Julia Tsyganova
2,3,*,†
1
Department of Mathematics, Physics and Technology Education, Ulyanovsk State University of Education, 432071 Ulyanovsk, Russia
2
Department of Mathematics, Information and Aviation Technology, Ulyanovsk State University, 432017 Ulyanovsk, Russia
3
Department of Academic Policy and Organization of Educational Activities, Innopolis University, 420500 Innopolis, Russia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(1), 126; https://doi.org/10.3390/math10010126
Submission received: 6 November 2021 / Revised: 19 December 2021 / Accepted: 28 December 2021 / Published: 2 January 2022
(This article belongs to the Special Issue New Trends on Identification of Dynamic Systems)

Abstract

:
The paper considers the problem of algorithmic differentiation of information matrix difference equations for calculating the information matrix derivatives in the information Kalman filter. The equations are presented in the form of a matrix MWGS (modified weighted Gram–Schmidt) transformation. The solution is based on the usage of special methods for the algorithmic differentiation of matrix MWGS transformation of two types: forward (MWGS-LD) and backward (MWGS-UD). The main result of the work is a new MWGS-based array algorithm for computing the information matrix sensitivity equations. The algorithm is robust to machine round-off errors due to the application of the MWGS orthogonalization procedure at each step. The obtained results are applied to solve the problem of parameter identification for state-space models of discrete-time linear stochastic systems. Numerical experiments confirm the efficiency of the proposed solution.

1. Introduction

Matrix orthogonal transformations are widely used in solving various problems of computational linear algebra [1].
The problem of calculating the values of the derivatives in matrix orthogonal transformations arises in automatic differentiation [2], perturbation and control theories, differential geometry when solving such problems as calculating Lyapunov exponents [3,4], numerical solutions of the matrix differential Riccati equation [5,6], and the Riccati sensitivity equation [7,8], computing higher-order derivatives in the experiment planning [9]. In the theory of Kalman filtering [10,11,12,13,14], orthogonal transformations are used to efficiently compute the solution of the matrix difference Riccati equation.
The methods for calculating the values of derivatives in matrix orthogonal transformations are similar in their properties to automatic (algorithmic) differentiation methods. Three methods for calculating derivatives are currently most common:
  • symbolic (analytical) differentiation;
  • numerical differentiation;
  • automatic (algorithmic) differentiation.
Symbolic differentiation allows one to obtain exact analytical formulas for the derivatives of the elements of a parameterized matrix, but this approach requires significant computational costs, and is not suitable for solving problems in real-time. With numerical differentiation, the result depends significantly on many factors, for example, on the step size. On the contrary, algorithmic (automatic) differentiation does not allow one to calculate expressions for derivatives or their tabular approximation, but does allow one to calculate the values of the derivatives at a given point for given values of the function arguments. This requires knowledge of the expression for the function, or at least a computer program to calculate it.
Consider a rectangular parameterized matrix A ( θ ) and a diagonal matrix D A ( θ ) , where θ is a scalar real parameter. The problem of algorithmic differentiation of matrix MWGS transformation (MWGS—the modified weighted Gram–Schmidt orthogonalization [10]) is to find, at a given point θ = θ ^ a triangular (upper or lower) matrix of derivatives B θ θ = θ ^ and the diagonal derivative matrix ( D B ) θ θ = θ ^ using known parameterized matrices A ( θ ) , D A ( θ ) and obtained as the result of MWGS transformation A T = B W T (where A T D A A = B D B B T ) triangular (upper or lower) matrix B with ones on the diagonal and diagonal matrix D B .
Suppose that the domain of the parameter θ is such that the matrix A has full column rank and the diagonal matrix D A > 0 . In what follows, for convenience of presentation, we denote A A ( θ ^ ) and D A D A ( θ ^ ) . Then, we will call two pairs of matrices A , D A and B , D B as MWGS-based arrays.
In our recent papers [15,16], we have proposed two methods for algorithmic differentiation of the MWGS-based arrays. These computational methods are based on the forward MWGS-LD orthogonalization ([16], p. 66) and the backward MWGS-UD orthogonalization procedure ([10], p. 127).
In this paper, we further develop our recently obtained results. Our research aims to construct a novel computational algorithm for evaluating the derivatives of MWGS factors of information matrix Y. Firstly, we show how our early suggested methods for algorithmic differentiation of the MWGS-based arrays can be applied to construct a new MWGS-based array algorithm for computing the information matrix sensitivity equations. Secondly, we demonstrate how the proposed algorithm can be efficiently applied to solve the parameter identification problem when gradient-based optimization methods are used.
The paper is organized as follows. Section 2 provides basic definitions associated with the information form of the Kalman filter, discusses the MWGS-based array algorithm for computing the information matrix, and presents two algorithms for computing derivatives of the MWGS-based arrays. Section 3 contains the main result of the paper—the new MWGS-based array algorithm for computing the information matrix sensitivity equations. Section 4 discusses the implementation details of the proposed algorithm and demonstrates how it can be applied for solving the parameter identification problem of the one practical stochastic system model. Finally, conclusions are made in Section 5.

2. Methodology

2.1. Information Kalman Filter and Information Matrix

The information Kalman filter (IKF) is an alternative formulation of the well-known Kalman filter (KF) [17]. IKF differs from KF in that it computes not an error covariance matrix P but its inverse matrix Y known as the information matrix. When there is no a priori information about the initial state value, the IKF is particularly useful because it easily starts from Y 0 = 0 . In the same case, the initial error covariance matrix Π 0 is not defined. Additionally, an implementation of the IKF could be computationally cheaper when the size of the measurements vector is greater than the size of the state vector [11]. Information filter does not use the same state vector representation as KF. They utilize the so-called information state d Y x instead ([11], p. 263).
Consider a discrete-time linear stochastic system
x k + 1 = F k x k + G k w k , k 0 ,
z k + 1 = H k + 1 x k + 1 + v k + 1
where x k R n is the state and z k R m are the measurements; k is a discrete-time instant. The process noise w k R q and the measurement noise v k R m are Gaussian white-noise processes with zero mean and covariance matrices Q k > 0 and R k > 0 . That means
E w k v k w j T v j T = Q k 0 0 R k δ k j
where δ k j denotes the Kronecker delta function. The initial state vector x 0 N ( x ¯ 0 , Π 0 ) .
Suppose that matrices F k are invertible [11]. Consider the problem of information filtering. It is to calculate at each discrete-time moment k the information state estimate, d k | k = Y k | k x ^ k | k given Z 1 k = { z 1 , , z k } . The solution is obtained by the conventional IKF equations, which are as follows [11]:
I. Time Update: The predicted information state estimate and the predicted information matrix obey the difference equations
d ^ k + 1 = [ I L k G k T ] F k T d ^ k | k , d ^ 0 | 0 = Y 0 x ¯ 0 .
Y k + 1 = [ I L k G k T ] A k , Y 0 | 0 = Π 0 1
where
A k F k T Y k | k F k 1 ,
L k = A k G k C k 1 , C k = G k T A k G k + Q k 1 .
II. Measurement Update: The updated (filtered) information state estimate d ^ k + 1 | k + 1 obeys
d ^ k + 1 | k + 1 = d ^ k + 1 + H k + 1 T R k + 1 1 z k + 1 .
The filtered information matrix satisfies the difference equation
Y k + 1 | k + 1 = Y k + 1 + H k + 1 T R k + 1 1 H k + 1 .
Equations (4)–(9) can be derived from the KF formulas by taking into account the definitions of the information matrix and the information state.
Furthermore, we will use the notations. Let B be a triangular matrix which can be either a unit upper triangular matrix, i.e., B : = U (with 1’s on the main diagonal) or a unit lower triangular matrix, i.e., B : = L . D is a diagonal matrix.

2.2. The MWGS-Based Array Algorithm for Computing the Information Matrix

Let us consider Equations (5)–(7) and (9). They allow one to compute information matrix Y k | k at each discrete-time instant k. To improve the numerical robustness to machine round-off errors, we have proposed in [18] a new MWGS-based array algorithm for computing the information matrices and the information states in IKF.
The MWGS-based array computations imply the use of numerically stable modified weighted Gram–Schmidt (MWGS) orthogonalization procedure for updating the required quantities. (The MWGS outperforms the conventional Gram–Schmidt algorithm for computational accuracy [19].) In [18], we have used both the forward MWGS-LD and the backward MWGS-UD orthogonalization procedures.
Each iteration of these IKF implementations has the following form: given a pair of matrices { A , D A } (so-called the pre-arrays), compute a pair of matrices { B , D B } (the post-arrays) using the MWGS orthogonalization procedure
A T = B W T
where a rectangular matrix A R r × s , the MWGS transformation matrix W R r × s ( r s ) produces the block triangular matrix B with 1’s on the main diagonal. A matrix B R s × s is either an upper triangular block matrix U or a lower triangular block matrix L such that  
A T D A A = B D B B T a n d W T D A W = D B
where the diagonal matrices are D A R r × r , D B R s × s , and D A > 0 ; see ([10], Lemma VI.4.1) for details.
We have proved ([18], Statement 1) that Algorithm 1 is algebraically equivalent to the IKF given by Equations (4)–(9). Thus, we have obtained an alternative method for calculating the information matrix Y within the MWGS-based array algorithm. A significant advantage of Algorithm 1 is its numerical robustness to machine round-off errors. A detailed discussion can be found in [18,20].
Algorithm 1 [18]. The MWGS-based array IKF.
Initialization. Let Y 0 = Π 0 1 . Compute the modified Cholesky decomposition. (The modified Cholesky decomposition has the form A = B A D A B A T where A is a symmetric positive definite matrix, D A is a diagonal matrix, and B A is a unit triangular (lower or upper) matrix [1,11].) Y 0 = B Y 0 D Y 0 B Y 0 T . Set the initial values d ^ 0 | 0 = Y 0 x ¯ 0 and B Y 0 | 0 = B Y 0 , D Y 0 | 0 = D Y 0 .
▹ For k = 0 , , K 1 do
I. Time Update. Apply the modified Cholesky decomposition for the process noise covariance matrix Q k = B Q k D Q k B Q k T . Compute matrices B Q k 1 and D Q k 1 . Find the MWGS factors { B Y k + 1 , D Y k + 1 } of matrix Y k + 1 as follows:
I.A. In the case of the forward MWGS-LD factorization (i.e., B = L ), the following steps should be done:
L Q k T G k T F k T L Y k | k 0 F k T L Y k | k Pre - array : A T U T = L C k 0 L k L C k L Y k + 1 Post - array : L T U W T U T ,
W T U T D Q k 1 0 0 D Y k | k Pre - array : D A T U W T U = D C k 0 0 D Y k + 1 Post - array : D L T U
where W T U R ( q + n ) × ( q + n ) is the MWGS-LD transformation, the block unit triangular (lower) post-array L T U R ( q + n ) × ( q + n ) , the diagonal post-array D L T U R ( q + n ) × ( q + n ) . The resulted MWGS-LD factors { L Y k + 1 , D Y k + 1 } are simply read off from the resulting post-arrays L T U and D L T U .
I.B. In the case of the backward MWGS-LD factorization (i.e., B = U ), one has to follow the next steps:
F k T U Y k | k 0 G k T F k T U Y k | k U Q k T Pre - array : D A T U = U Y k + 1 L k U C k 0 U C k Post - array : U T U W T U T ,
W T U T D Y k | k 0 0 D Q k 1 Pre - array : D A T U W T U = D Y k + 1 0 0 D C k Post - array : D U T U
where W T U R ( n + q ) × ( n + q ) is the MWGS-UD transformation, the block unit triangular (upper) post-array U T U R ( n + q ) × ( n + q ) , the diagonal post-array D U T U R ( n + q ) × ( n + q ) . Again, we can easily extract the required MWGS-UD factors { U Y k + 1 , D Y k + 1 } from the resulting post-arrays U T U and D U T U .
Given d ^ k | k , find the predicted information state estimate:
d ^ k + 1 = I L k B C k B C k 1 G k T F k T d ^ k | k
where matrix product L k B C k is directly extracted either from the post-arrays L T U in (12) or from U T U in (14).
II. Measurement Update. Apply the modified Cholesky factorization for the measurement noise covariance matrix R k = B R k D R k B R k T . Compute matrices B R k 1 and D R k 1 . Find the filtered MWGS factors { B Y k + 1 | k + 1 , D Y k + 1 | k + 1 } :
B Y k + 1 H k + 1 T B R k + 1 T Pre - array : A M U T = B Y k + 1 | k + 1 Post - array : B M U W M U T ,
W M U T D Y k + 1 0 0 D R k + 1 1 Pre - array : D A M U W M U = D Y k + 1 | k + 1 Post - array : D B M U
where W M U R ( n + m ) × n is the MWGS transformation matrix, the block triangular post-array B M U R n × n , the diagonal post-array D B M U R n × n . The resulting post-arrays B M U and D B M U are the MWGS factors { B Y k + 1 | k + 1 , D Y k + 1 | k + 1 } .
Next, compute the filtered estimate d ^ k + 1 | k + 1 by (8).
▹ End.

2.3. Algorithmic Differentiation of the MWGS-Based Arrays

When solving practical problems of parameter identification [21], the discrete linear stochastic model (1),(2) is often parameterized. The latter means that the system matrices could depend on the unknown parameter θ . Therefore, it should be estimated together with the hidden state vector x k given measurements z k . In this case, any parameters’ estimation scheme includes two components, namely: the filtering method for computing an identification criterion and the chosen optimization algorithm to identify the optimal value θ ^ . Altogether, it is called the adaptive filtering scheme [22].
It is well-known that the gradient-based optimization algorithms converge fast and, therefore, they are the preferred methods for practical implementation [8]. They require the computation of the gradient of identification criterion. The latter leads to the problem of the adaptive filter derivatives computation. The related vector- and matrix-type equations are called the filter sensitivity equations with respect to unknown parameter θ .
Consider conventional information Kalman filter, presented by Equations (4)–(9), and MWGS-based IKF (Algorithm 1). We can construct matrix sensitivity equations for information matrix evaluating in the conventional IKF by direct differentiation of (5)–(7), and (9). This solution is not hard, and it is as follows:
( Y k + 1 ) θ = [ I L k G k T ] ( A k ) θ [ ( L k ) θ G k T + L k ( G k T ) θ ] A k ,
( A k ) θ = ( F k T ) θ Y k | k F k 1 + F k T ( Y k | k ) θ F k 1 + F k T Y k | k ( F k 1 ) θ ,
( L k ) θ = ( A k ) θ G k C k 1 + A k ( G k ) θ C k 1 + A k G k ( C k 1 ) θ ,
( C k ) θ = ( G k T ) θ A k G k + G k T ( A k ) θ G k + G k T A k ( G k ) θ + ( Q k 1 ) θ ,
( Y k + 1 | k + 1 ) θ = ( Y k + 1 ) θ + ( H k + 1 T ) θ R k + 1 1 H k + 1 + H k + 1 T ( R k + 1 1 ) θ H k + 1 + H k + 1 T R k + 1 1 ( H k + 1 ) θ .
However, a corresponding solution is not obvious for the MWGS-based array IKF (Algorithm 1). Finding new computational methods for evaluating the derivatives of MWGS-factors of information matrix Y is the aim of our research.
Let us consider two of our methods for algorithmic differentiation of the MWGS-based arrays. They were proposed in [15,16].
Case 1. Consider the forward MWGS-LD orthogonalization procedure (10) and (11), where B : = L (lower triangular matrix), D B : = D L (diagonal matrix).
Lemma 1
([16]). Let entries of the pre-arrays A , D A in (11) be known differentiable functions of a parameter θ. Consider the transformation (11) in Case 1. Given the derivatives of the pre-arrays A θ and ( D A ) θ , we can calculate the corresponding derivatives of the post-arrays:
L θ = L L ¯ 0 + L ¯ 2 + U ¯ 0 T D L 1 , D L θ = 2 D 0 + D 2
where L ¯ 0 , D 0 , U ¯ 0 are strictly lower triangular, diagonal and strictly upper triangular parts of the matrix product W T ( D A ) θ L T , respectively. D 2 and L ¯ 2 are diagonal and strictly lower triangular parts of the product W T ( D A ) θ W .
Case 2. Consider the backward MWGS-UD orthogonalization procedure (10) and (11) where B : = U (upper triangular matrix), D B : = D U (diagonal matrix).
Lemma 2
([15]). Let entries of the pre-arrays A , D A in (11) be known differentiable functions of a parameter θ. Consider the transformation (11) in Case 2. Given the derivatives of the pre-arrays A θ and ( D A ) θ , we can calculate the corresponding derivatives of the post-arrays:
U θ = U L ¯ 0 T + U ¯ 0 + U ¯ 2 D U 1 , D U θ = 2 D 0 + D 2
where L ¯ 0 , D 0 , U ¯ 0 are strictly lower triangular, diagonal and strictly upper triangular parts of the matrix product W T ( D A ) θ U T , respectively. D 2 and U ¯ 2 are diagonal and strictly upper triangular parts of the product W T ( D A ) θ W .
Applying Lemmas 1 and 2, we construct the corresponding algorithms for calculating the values of derivatives in the MWGS-based arrays for given parametrized matrices A ( θ ) and D A ( θ ) .
Remark 1.
FunctionMWGS-LD( A , D A ) implements the forward MWGS orthogonalization procedure.
Remark 2.
FunctionMWGS-UD( A , D A ) implements the backward MWGS orthogonalization procedure.
Thus, computational Algorithms 2 and 3 have the following properties:
  • They allow calculating, at a given point, the values of derivatives of elements of the matrix factors obtained by MWGS transformation of the pair of parameterized matrices. In this case, there is no need to calculate values of the derivatives of elements of the MWGS transformation matrix.
  • These algorithms require simple addition and multiplication matrix operations, and only one triangular and one diagonal matrix inversion operation. Therefore, they have a simple structure to easily implement in program code.
  • Their correctness is strictly mathematically proved [15,16].
  • Their performance has been confirmed by practical examples [15,23].
It should be noted here that the results of Lemma 2 and Algorithm 3 have been successfully applied in [24] for constructing an efficient UD-based algorithm for the computation of maximum likelihood sensitivity of continuous-discrete systems.
Algorithm 2. Diff_LD (LD-based derivative computation).
 ▹ Input data:  A ( θ ) R r × s , D A ( θ ) R r × r , θ = θ ^ R p ,
   A θ i = a k j ( θ ) θ i , ( D A ) θ i = d k ( θ ) θ i , i = 1 , , p .
 ▹ Begin
 1 evaluate A A ( θ ^ ) , D A D A ( θ ^ ) ;
 2 evaluate A θ ^ i A θ i θ i = θ ^ i , ( D A ) θ i ^ ( D A ) θ i θ i = θ ^ i ;
 3. compute [ L , D L , W ] MWGS-LD( A , D A );
▹ For i = 1 , , p do
 4 compute X W T D A A θ ^ i L T ;
 5 split X into three parts  [ L ¯ 0 + D 0 + U ¯ 0 ] X;
 6 compute V W T ( D A ) θ ^ i W ;
 7 split V into three parts  [ L ¯ 2 + D 2 + L ¯ 2 T ] V;
 8 obtain result  D L θ ^ i 2 D 0 + D 2 ;
 9 obtain result  L θ ^ i L L ¯ 0 + L ¯ 2 + U ¯ 0 T D L 1 .
▹ End for
 ▹ End.
 ▹ Output data:  L R s × s , D L R s × s ; L θ ^ i , ( D U ) θ ^ i , i = 1 , , p .
Algorithm 3. Diff_UD (UD-based derivative computation).
 ▹ Input data:  A ( θ ) R r × s , D A ( θ ) R r × r , θ = θ ^ R p ,
   A θ i = a k j ( θ ) θ i , ( D A ) θ i = d k ( θ ) θ i , i = 1 , , p .
 ▹ Begin
 1 evaluate A A ( θ ^ ) , D A D A ( θ ^ ) ;
 2 evaluate A θ ^ i A θ i θ i = θ ^ i , ( D A ) θ i ^ ( D A ) θ i θ i = θ ^ i ;
 3 compute [ U , D U , W ] MWGS-UD( A , D A );
▹ For i = 1 , , p do
 4 compute X W T D A A θ ^ i U T ;
 5 split X into three parts  [ L ¯ 0 + D 0 + U ¯ 0 ] X;
 6 compute V W T ( D A ) θ ^ i W ;
 7 split V into three parts  [ U ¯ 2 T + D 2 + U ¯ 2 ] V;
 8 obtain result  D U θ ^ i 2 D 0 + D 2 ;
 9 obtain result  U θ ^ i U L ¯ 0 T + U ¯ 0 + U ¯ 2 D U 1 .
▹ End for
 ▹ End.
 ▹ Output data: U R s × s , D U R s × s ; U θ ^ i , ( D U ) θ ^ i .

3. Main Result

The New MWGS-Based Array Algorithm for Computing the Information Matrix Sensitivity Equations

Now, we are ready to present the main result—the new MWGS-based array algorithm for computing the information matrix sensitivity equations. We are extending the functionality of Algorithm 1 so that it is able to calculate not only the values of information matrix Y using MWGS-based arrays, but also the values of their derivatives.
Let us consider the given value of parameter θ = θ ^ .
The new Algorithm 4 naturally extends any MWGS-based IKF implementation on the information matrix sensitivities evaluation.
Algorithm 4. The differentiated MWGS-based array.
Initialization. Let θ = θ ^ . Evaluate the initial value of information matrix Y 0 = Y 0 ( θ ^ ) . Find ( Y 0 ) θ ^ i , i = 1 , , p . Apply the modified Cholesky factorization Y 0 = B Y 0 D Y 0 B Y 0 T . Find ( B Y 0 ) θ ^ i , ( D Y 0 ) θ ^ i , i = 1 , , p . Set the initial values { B Y 0 | 0 = B Y 0 , D Y 0 | 0 = D Y 0 } and { ( B Y 0 | 0 ) θ ^ i = ( B Y 0 ) θ ^ i , ( D Y 0 | 0 ) θ ^ i = ( D Y 0 ) θ ^ i } .
▹ For k = 0 , , K 1 do
I. Time Update.
I.1 Evaluate matrices F ^ k = F k ( θ ^ ) , G ^ k = G k ( θ ^ ) , and Q ^ k = Q k ( θ ^ ) . Find ( F k ) θ ^ i , ( G k ) θ ^ i , and ( Q k ) θ ^ i , i = 1 , , p .
I.2 Use the modified Cholesky decomposition for matrices Q ^ k and ( Q k ) θ ^ i to find { B Q ^ k , D Q ^ k } , and ( B Q k ) θ ^ i , ( D Q k ) θ ^ i , i = 1 , , p .
I.3 Given the MWGS factors { B Y k | k , D Y k | k } and their derivatives
{ ( B Y k | k ) θ ^ i , ( D Y k | k ) θ ^ i } , find their predicted values { B Y k + 1 , D Y k + 1 } and their derivatives { ( B Y k + 1 ) θ ^ i , ( D Y k + 1 ) θ ^ i } ( i = 1 , , p ) as follows:
I.A. In the case of the forward MWGS-LD factorization (i.e., B = L ), the following steps should be taken:
  • Form the pre-arrays { A T U T , D A T U } and their derivatives { ( A T U T ) θ ^ i , ( D A T U ) θ ^ i } , ( i = 1 , , p ):
    A T U T = L Q ^ k T G ^ k T F ^ k T L Y k | k 0 F ^ k T L Y k | k , D A T U = D Q ^ k 1 0 0 D Y k | k ;
    ( A T U T ) θ ^ i = ( L Q ^ k T ) θ ^ i ( G ^ k T F ^ k T L Y k | k ) θ ^ i 0 ( F ^ k T ) θ ^ i ( L Y k | k ) θ ^ i , ( D A M U ) θ ^ i = ( D Q ^ k 1 ) θ ^ i 0 0 ( D Y k | k ) θ ^ i .
  • Apply Algorithm 2 where A : = A T U T , D A : = D A T U ; L : = L T U , D L : = D L T U .
  • Obtain the post-arrays { L T U , D L T U } and their derivatives { ( L T U ) θ ^ i , ( D L T U ) θ ^ i } , ( i = 1 , , p ):
    L T U = L C k 0 L k L C k L Y k + 1 , D L T U = D C k 0 0 D Y k + 1 ;
    ( L T U ) θ ^ i = ( L C k ) θ ^ i 0 ( L k L C k ) θ ^ i ( L Y k + 1 ) θ ^ i , ( D L T U ) θ ^ i = ( D C k ) θ ^ i 0 0 ( D Y k + 1 ) θ ^ i .
  • Extract matrices { L Y k + 1 , D Y k + 1 } and { ( L Y k + 1 ) θ ^ i , ( D Y k + 1 ) θ ^ i } ( i = 1 , , p ) from the post-arrays.
I.B. In the case of the backward MWGS-LD factorization (i.e., B = U ), one has to take the next steps:
  • Form the pre-arrays { A T U T , D A T U } and their derivatives { ( A T U T ) θ ^ i , ( D A T U ) θ ^ i } , ( i = 1 , , p ):
    A T U T = F ^ k T U Y k | k 0 G ^ k T F ^ k T U Y k | k U Q ^ k T , D A T U = D Y k | k 0 0 D Q ^ k 1 ;
    ( A T U T ) θ ^ i = ( F ^ k T U Y k | k ) θ ^ i 0 ( G ^ k T F ^ k T U Y k | k ) θ ^ i ( F ^ k T ) θ ^ i ( U Q ^ k T ) θ ^ i , ( D A M U ) θ ^ i = ( D Y k | k ) θ ^ i 0 0 ( D Q ^ k 1 ) θ ^ i .
  • Apply Algorithm 3, where A : = A T U T , D A : = D A T U ; L : = U T U , D U : = D U T U .
  • Obtain the post-arrays { U T U , D U T U } and their derivatives { ( U T U ) θ ^ i , ( D U T U ) θ ^ i } , ( i = 1 , , p ):
    U T U = U Y k + 1 L k U C k 0 U C k , D U T U = D Y k + 1 0 0 D C k ;
    ( U T U ) θ ^ i = ( U Y k + 1 ) θ ^ i ( L k U C k ) θ ^ i 0 ( U C k ) θ ^ i , ( D U T U ) θ ^ i = ( D Y k + 1 ) θ ^ i 0 0 ( D C k ) θ ^ i .
  • Extract matrices { U Y k + 1 , D Y k + 1 } and { ( U Y k + 1 ) θ ^ i , ( D Y k + 1 ) θ ^ i } ( i = 1 , , p ) from the post-arrays.
II. Measurement Update.
II.1 Evaluate matrices H ^ k = H k ( θ ^ ) and R ^ k = R k ( θ ^ ) . Find ( H k ) θ ^ i and ( R k ) θ ^ i , i = 1 , , p .
II.2 Use the modified Cholesky decomposition for matrices R ^ k and ( R k ) θ ^ i to find { B R ^ k , D R ^ k } , and ( B R k ) θ ^ i , ( D R k ) θ ^ i , i = 1 , , p .
II.3 Given the MWGS factors { B Y k + 1 , D Y k + 1 } and their derivatives { ( B Y k + 1 ) θ ^ i , ( D Y k + 1 ) θ ^ i } , find the corresponding pairs of matrices { B Y k + 1 | k + 1 , D Y k + 1 | k + 1 } and { ( B Y k + 1 | k + 1 ) θ ^ i , ( D Y k + 1 | k + 1 ) θ ^ i } ( i = 1 , , p ) as follows:
  • Form the pre-arrays { A M U T , D A M U } and their derivatives { ( A M U T ) θ ^ i , ( D A M U ) θ ^ i } , ( i = 1 , , p ):
    A M U T = B Y k + 1 H ^ k + 1 T B R ^ k + 1 T , D A M U = D Y k + 1 0 0 D R ^ k + 1 1 ;
    ( A M U T ) θ ^ i = ( B Y k + 1 ) θ ^ i ( H ^ k + 1 T B R ^ k + 1 T ) θ ^ i , ( D A M U ) θ ^ i = ( D Y k + 1 ) θ ^ i 0 0 ( D R ^ k + 1 1 ) θ ^ i .
  • Apply Algorithm 2 in Case 1 or Algorithm 3 in Case 2, where A : = A M U T , D A : = D A M U ; B : = B M U , D B : = D B M U .
  • Obtain the post-arrays { B M U , D B M U } and their derivatives { ( B M U ) θ ^ i , ( D B M U ) θ ^ i } , ( i = 1 , , p ):
    B M U = B Y k + 1 | k + 1 , D B M U = D Y k + 1 | k + 1 ;
    ( B M U ) θ ^ i = ( B Y k + 1 | k + 1 ) θ ^ i , ( D B M U ) θ ^ i = ( D Y k + 1 | k + 1 ) θ ^ i .
  • Extract matrices { B Y k + 1 | k + 1 , D Y k + 1 | k + 1 } , { ( B Y k + 1 | k + 1 ) θ ^ i , ( D Y k + 1 | k + 1 ) θ ^ i } ( i = 1 , , p ) from the post-arrays.

▹ End.

4. Discussion

4.1. Implementation Details of Algorithm 4

Let us consider the computational scheme of the constructed algorithm in detail. The new Algorithm 4 is built based on one of two variants of the MWGS transformation and the corresponding method of algorithmic differentiation of such an orthogonal transformation. Therefore, Algorithms 2 and 3 can be considered as basic computational tools (or technologies) for the implementation of Algorithm 4. From this point of view, the technology of software implementation of the new algorithmic differentiation algorithm seems to be simple and understandable. It consists of only three simple steps:
  • Fill in block pre-arrays with available data.
  • Execute an algorithm for calculating derivatives in a matrix MWGS transformation of one of the types corresponding to Case 1 or Case 2.
  • As a result, get block post-arrays and read off the required results from them in the form of matrix blocks.
The implementation scheme of the measurement update step in Algorithm 4 based on the MWGS-UD transformation is shown in Figure 1. Similarly, a general scheme for the MWGS-LD transformation can also be represented.
The computational complexity of the novel Algorithm 4 is mainly determined by the computational complexity of Algorithm 1, i.e., the MWGS-based information-form Kalman filtering algorithm. A detailed analysis of its computational complexity is given in ([18], Section 5.2).
It was shown that conventional IKF and Algorithm 1 have the complexity of the same order. However, IKF requires four full matrix inversions while calculating the information matrix Y. At the same time, Algorithm 1 requires only one full matrix inversion of the matrix F k . Besides, if the matrices Q k and R k are positively definite and do not depend on k, then the modified Cholesky decomposition needs to be performed only once at the initialization step of the MWGS-based algorithm. If matrix F k is not singular and also does not depend on k, then the inversion of matrix F also needs to be performed only once, i.e., at the initialization step of the MWGS-based algorithm. Algorithm 4 requires additionally only one inversion of the unit triangular matrix and one inversion of the diagonal matrix (see Algorithm 2 or 3). Therefore, no additional inversions of the full matrices are required.
To summarize, the information matrix sensitivity evaluation based on the conventional IKF (Equations (5)–(7) and (19)–(23)) requires eight inversions of the full matrices, while the new Algorithm 4 avoids the full matrix inversion operations and requires the inverse of the unit triangular and diagonal matrices, only. Thus, we can conclude that the newly proposed algorithm is computationally efficient compared to the conventional IKF.

4.2. Application of the Results Obtained to the Problem of Parameter Identification

In practice, the matrices characterizing discrete-time linear stochastic system (1)–(2) are often known up to certain parameters. Consider an important problem of parameter identification [21]. Assume that the elements of system matrices F k R n × n , G k R n × q , H k R m × n , Q k R q × q , R k R m × m , and Π 0 R n × n are functions of unknown system parameters vector θ R p . It needs to be identified. For the sake of simplicity, instead of F k ( θ ) , G k ( θ ) , H k ( θ ) etc., we will write F k , G k , H k , etc.
We wish to demonstrate how the new Algorithm 4 can be applied to solve the parameter identification problem of the practical stochastic system model.
Consider the instrument error model for one channel of the INS (Inertial Navigation System) given as follows [25]:
Δ v x β m A x n G y k + 1 = 1 τ g τ 0 τ / a 1 0 τ 0 0 b 1 0 0 0 0 1 Δ v x β m A x n G y k + 0 0 a 1 0 w k ,
z k + 1 = ( Δ v x ) k + 1 + v k + 1
where w k N ( 0 , 1 ) , v k N ( 0 , 0.01 ) , x 0 N ( 0 , I 4 ) and subscripts x, y, A, G denote “axis O x ”, “axis O y ”, “Accelerometer”, and “Gyro”, respectively.
The state vector x k = Δ v x , β , m A x , n G y T where the first element is the random error in reading velocity along axis O x of a GSP, the second element is the angular error in determining the local vertical, the third one is the accelerometer reading random error, and the fourth one is the gyro constant drift rate.
The constants τ , g, a are, respectively, equal to the rate of data arrival and processing, the gravity acceleration, and the semi-major axis of the Earth. The quantities a 1 = H 1 1 b 1 2 H 1 2 γ τ and b 1 = e γ τ 1 γ τ . The constants H 1 and γ are elements of the accepted model of the correlation function R m A x = H 1 2 e γ | τ | . Numerical values of the model parameters are given in Table 1.
Equations (38) and (39) correspond to the general model (1)–(2). Note that, in our case, all system model matrices do not depend on k.
Let us suppose that parameter γ is unknown and needs to be identified. This means that the model parameter θ = γ , and therefore F = F ( θ ) , G = G ( θ ) .
Solving the problem of parameter identification, we use the Active Principle of Adaptation (APA) [26,27,28,29], which consists of constructing an Auxiliary Performance Index (API) [23,28,30] and minimizing it with the use of a gradient-based numerical procedure.
The APA approach to system adaption within the parameter uncertainty differs in the fact that it suggests an indirect state prediction error control in the API form. It has to satisfy two main requirements:
  • it depends on the system observable values only;
  • it attains its minimum coincidently with the Original Performance Index (OPI).
The API satisfies a relation
API = OPI + c o n s t ,
if the OPI is defined as the expected (Euclidean) norm of the state prediction error. Thus API and OPI have the one and the same minimizing argument θ .
In order to construct the API, we build a Standard Observable Model (SOM), i.e., we perform the corresponding transformation of the basis in the state space from the representation (1)–(2). Model
x k + 1 * = F * ( θ ) x k * + G * ( θ ) w k , k 0 ,
z k + 1 = H * x k + 1 * + v k + 1
is equivalent to the original model (1)–(2) and is its canonical representation, where x k * is the new state vector; F * ( θ ) , G * ( θ ) , and H * are matrices of the following form:
G * = τ a 1 0 1 1 + b 1 1 ρ + b 1 + b 1 2 , H * = 1 0 0 0 , ρ = τ 2 g / a ,
F * = 0 1 0 0 0 0 1 0 0 0 0 1 α 4 α 3 α 2 α 1 , α 4 = b 1 ( 1 + ρ ) , α 3 = b 1 ( 3 + ρ ) ( 1 + ρ ) , α 2 = 3 b 1 + ( 3 + ρ ) , α 1 = b 1 3 .
From representation (42), (43) it follows that the maximum observability index of the system is = 4 .
Using the results of [26,27,28], we construct the auxiliary performance index (API)
J ε ( θ ) 1 2 E ε k ( θ ) T ε k ( θ ) ,
for which the auxiliary process is written in the form:
ε k = S ( Z k + 1 k ) x ^ k + 1 = [ z k 3 , z k 2 , z k 1 , z k ] T x ^ k 3
where x ^ k is the prediction state estimate obtained in an adaptive filter.
Next, to get a reasonable estimate for (44), we replace it by the realizable (workable) performance index
J ε ( θ , K ) = 1 2 K j = k K + 1 k ε j ( θ ) T ε j ( θ ) .
With the purpose of finding the optimal value θ of the unknown parameter θ and to minimize the API (46), we can use existing methods of numerical optimization. Moreover, all non-search methods require the calculation of the API gradient. Assume that the gradient-based optimization method is chosen for the parameter identification procedure. Then, from (46), we can write the expression for calculating the API gradient:
J ε ( θ , K ) θ = 1 K j = k K + 1 k ε j T ( θ ) ε j ( θ ) θ .
Evaluating (47) requires the computation of sensitivities (partial derivatives) of the auxiliary process ε j with respect to the adjustable parameter θ . Various methods for calculating the sensitivities based on the standard Kalman filter are discussed in detail in [8].
Now, we are ready to demonstrate how the new Algorithm 4 can be applied for constructing a computational scheme for the numerical identification of the parameter θ based on the API approach.
Consider again the system model (42)–(43). Denote for simplicity F = F * ( θ ) , G = G * ( θ ) , H = H * , x ¯ 0 = x ¯ 0 * ( θ ) , Π 0 = Π 0 * ( θ ) . Let θ = γ , Y 0 = Π 0 1 , d ^ 0 | 0 = Y 0 x ¯ 0 .
The identification of unknown parameter θ and the estimation of state vector x k of the system (38)–(39) according to the criterion
θ = θ ^ min = argmin D ( θ ) J ε ( θ , K )
can be performed simultaneously by the following algorithm.
Let us consider the practical application of Algorithm 5 to identify the unknown value of γ in the model (38)–(39). We have simulated the sequence of output signals Z 1 200 with the “true” value of γ = 0.0002 . All computer codes were constructed in MATLAB.
In order to conduct our numerical experiments, we have implemented Algorithms 1–5 as the corresponding MATLAB functions. Then, we have calculated the API (46) and the API gradient (47) depending on different values of γ . Results are illustrated by Figure 2 and Figure 3.
Algorithm 5. The API-based parameter identification computational scheme.
 BEGIN
✓ Assign an initial parameter estimate θ ^ i for i = 0 .
   REPEAT
Take the current point θ ^ i .
Evaluate the system matrices for (42)–(43) at θ ^ i :
F * : = F ( θ ^ i )  , G : = G * ( θ ^ i ) .
Evaluate the matrix derivatives ( F * ) θ and ( G * ) θ at θ ^ i .
Given the output data Z 1 K = [ z 1 , , z K ] T , compute a value of the API at θ ^ i using (46) where the auxiliary process ε j ( θ ^ i ) (45) can be evaluated as follows:
ε j ( θ ^ i ) = Z j 3 j x ^ j 3 = Z j 3 j Y j 3 1 d ^ j 3 .
The information matrix Y j = B Y j D Y j B Y j T and information state estimate d ^ j can be computed according to Algorithm 1 (the MWGS-based array IKF).
Evaluate the API gradient (47) at θ i using the results of Step 3 and the Algorithm 4. Wherein
ε j ( θ ) θ = Y j 3 θ 1 d ^ j 3 ( Y j 3 ) 1 d ^ j 3 θ ,
Y j θ = ( B Y j D Y j B Y j T ) θ are partial derivatives of information matrix Y j . The partial derivative vector d ^ j θ can be evaluated by direct differentiation of (16).
Find θ i + 1 by the gradient-based optimization procedure
θ ^ i + 1 = θ ^ i β i J ε ( θ , K ) θ θ = θ ^ i
where scalar step size parameter β j is designed to ensure that J ε ( θ ^ i + 1 , K ) J ε ( θ ^ i , K ) + e (small e > 0 ).
i : = i + 1 UNTIL a stopping criterion is satisfied.
 END
As can be seen from these figures, the minimum point of the API coincides with the true value of parameter θ = γ . Furthermore, a plot of the API gradient has negative values to the left and positive values to the right of the zero point, which correspond to the minimum of the API. All this evidence substantiates our theoretical derivations.
Further to solve a parameter identification problem, we apply the MATLAB Optimization Toolbox with the built-in function fminunc, which implements the gradient-type method. Algorithm 5 was incorporated into the optimization method fminunc to compute the API and its gradient. We have chosen the initial value θ ^ 0 = 0.001 and the stopping criteria epsf= 10 6 , epsx= 10 6 .
Results summarized in Table 2 show that the computed estimate θ ^ comes close to the true parameter value γ .
So, we conclude that the newly constructed Algorithm 4 can be efficiently applied to solve the parameter identification problem when the gradient-based optimization method is used.

5. Conclusions

This paper presents the new MWGS-based array algorithm for computing the information matrix sensitivity equations. We have extended the functionality of the MWGS-based information-form Kalman filtering methods so that they are able to calculate not only the values of the information matrix using the MWGS-based arrays, but also the values of their derivatives. The proposed algorithm is robust to machine round-off errors due to the application of the MWGS orthogonalization procedure at each step.
Moreover, we have demonstrated how the new Algorithm 4 can be applied for solving the parameter identification problem of the one practical stochastic system model, i.e., a simplified version of the instrument error model of the INS. We also have suggested the new API-based parameter identification computational scheme. Numerical experiments conducted in MATLAB confirm the efficiency of the proposed solution.

Author Contributions

Conceptualization, A.T. and J.T.; methodology, A.T. and J.T.; software, A.T.; validation, J.T.; formal analysis, A.T. and J.T.; investigation, A.T. and J.T.; resources, A.T. and J.T.; data curation, A.T. and J.T.; writing—original draft preparation, J.T.; writing—review and editing, A.T. and J.T.; visualization, A.T. and J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

MWGSModified weighted Gram–Schmidt transformation
MWGS-LDForward MWGS procedure
MWGS-UDBackward MWGS procedure
KFKalman filter
IKFInformation Kalman filter
TUTime update step
MUMeasurement update step
APAActive principle of adaptation
APIAuxiliary Performance Index

References

  1. Golub, G.H.; Van Loan, C.F. Matrix Computations; Johns Hopkins University Press: Baltimore, MD, USA, 1983. [Google Scholar]
  2. Giles, M. An Extended Collection of Matrix Derivative Results for Forward and Reverse Mode Algorithmic Differentiation; Report 08/01; Oxford University Computing Laboratory: Oxford, UK, 2008; 23p. [Google Scholar]
  3. Dieci, L.; Eirola, T. Applications of Smooth Orthogonal Factorizations of Matrices. In The IMA Volumes in Mathematics and Its Applications; Springer: New York, NY, USA, 2000; Volume 119, pp. 141–162. [Google Scholar]
  4. Dieci, L.; Russell, R.D.; Van Vleck, E.S. On the Computation of Lyapunov Exponents for Continuous Dynamical Systems. SIAM J. Numer. Anal. 1997, 34, 402–423. [Google Scholar] [CrossRef] [Green Version]
  5. Kunkel, P.; Mehrmann, V. Smooth factorizations of matrix valued functions and their derivatives. Numer. Math. 1991, 60, 115–131. [Google Scholar] [CrossRef]
  6. Dieci, L. On smooth decompositions of matrices. SIAM J. Matrix Anal. Appl. 1999, 20, 800–819. [Google Scholar] [CrossRef]
  7. Åström, K.-J. Maximum Likelihood and Prediction Error Methods. Automatica 1980, 16, 551–574. [Google Scholar] [CrossRef]
  8. Gupta, N.K.; Mehra, R.K. Computational aspects of maximum likelihood estimation and reduction in sensitivity function calculations. IEEE Trans. Autom. Control 1974, AC-19, 774–783. [Google Scholar] [CrossRef]
  9. Walter, S.F. Structured Higher-Order Algorithmic Differentiation in the Forward and Reverse Mode with Application in Optimum Experimental Design. Ph.D. Thesis, Humboldt-Universität zu Berlin, Berlin, Germany, 2011. [Google Scholar]
  10. Bierman, G.J. Factorization Methods for Discrete Sequential Estimation; Academic Press: New York, NY, USA, 1977. [Google Scholar]
  11. Grewal, M.S.; Andrews, A.P. Kalman Filtering: Theory and Practice Using MATLAB, 4th ed.; John Wiley & Sons, Inc.: New York, NY, USA, 2015. [Google Scholar]
  12. Kailath, T.; Sayed, A.H.; Hassibi, B. Linear Estimation; Prentice Hall: Hoboken, NJ, USA, 2000. [Google Scholar]
  13. Gibbs, B.P. Advanced Kalman Filtering, Least-Squares and Modeling: A Practical Handbook; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2011. [Google Scholar]
  14. Grewal, M.S.; Weill, L.R.; Andrews, A.P. Global Positioning Systems, Inertial Navigation, and Integration; John Wiley & Sons, Inc.: New York, NY, USA, 2001. [Google Scholar]
  15. Tsyganova, J.V.; Kulikova, M.V. State sensitivity evaluation within UD based array covariance filters. IEEE Trans. Autom. Control 2013, 58, 2944–2950. [Google Scholar] [CrossRef] [Green Version]
  16. Tsyganova, Y.V.; Tsyganov, A.V. On the Computation of Derivatives within LD Factorization of Parametrized Matrices. Bull. Irkutsk. State Univ. Ser. Math. 2018, 23, 64–79. (In Russian) [Google Scholar] [CrossRef]
  17. Kaminski, P.G.; Bryson, A.E.; Schmidt, S.F. Discrete square-root filtering: A survey of current techniques. IEEE Trans. Autom. Control 1971, AC-16, 727–735. [Google Scholar] [CrossRef]
  18. Tsyganova, J.V.; Kulikova, M.V.; Tsyganov, A.V. A general approach for designing the MWGS-based information-form Kalman filtering methods. Eur. J. Control 2020, 56, 86–97. [Google Scholar] [CrossRef]
  19. Björck, A. Solving linear least squares problems by Gram-Schmidt orthogonalization. Bit Numer. Math. 1967, 7, 1–21. [Google Scholar] [CrossRef]
  20. Tsyganova, J.V.; Kulikova, M.V.; Tsyganov, A.V. Some New Array Information Formulations of the UD-based Kalman Filter. In Proceedings of the 18th European Control Conference (ECC), Napoli, Italy, 25–28 June 2019; pp. 1872–1877. [Google Scholar]
  21. Ljung, L. System Identification: Theory for the User, 2nd ed.; Prentice Hall PTR: Upper Saddle River, NJ, USA, 1999. [Google Scholar]
  22. Kulikova, M.V. Likelihood Gradient Evaluation Using Square-Root Covariance Filters. IEEE Trans. Autom. Control 2009, 54, 646–651. [Google Scholar] [CrossRef] [Green Version]
  23. Semushin, I.V.; Tsyganova, J.V.; Tsyganov, A.V. Numerically Efficient LD-computations for the Auxiliary Performance Index Based Control Optimization under Uncertainties. IFAC-PapersOnline 2018, 51, 568–573. [Google Scholar] [CrossRef]
  24. Boiroux, D.; Juhl, R.; Madsen, H.; Jørgensen, J.B. An Efficient UD-Based Algorithm for the Computation of Maximum Likelihood Sensitivity of Continuous-Discrete Systems. In Proceedings of the 2016 IEEE 55th Conference on Decision and Control (CDC), ARIA Resort & Casino, Las Vegas, NV, USA, 12–14 December 2016; pp. 3048–3053. [Google Scholar]
  25. Broxmeyer, C. Inertial Navigation Systems; McGraw-Hill Book Company: Boston, MA, USA, 1956. [Google Scholar]
  26. Semushin, I.V. Adaptation in Stochastic Dynamic Systems—Survey and New Results I. Int. J. Commun. Netw. Syst. Sci. 2011, 4, 17–23. [Google Scholar] [CrossRef] [Green Version]
  27. Semushin, I.V. Adaptation in Stochastic Dynamic Systems—Survey and new results II. Int. J. Commun. Netw. Syst. Sci. 2011, 4, 266–285. [Google Scholar] [CrossRef] [Green Version]
  28. Semushin, I.V.; Tsyganova, J.V. Adaptation in Stochastic Dynamic Systems—Survey and New Results IV: Seeking Minimum of API in Parameters of Data. Int. J. Commun. Netw. Syst. Sci. 2013, 6, 513–518. [Google Scholar] [CrossRef] [Green Version]
  29. Semushin, I.V. The APA based time-variant system identification. In Proceedings of the 53rd IEEE Conference on Decision and Control (CDC), Los Angeles, CA, USA, 15–17 December 2014; pp. 4137–4141. [Google Scholar]
  30. Tsyganova, Y.V. Computing the gradient of the auxiliary quality functional in the parametric identification problem for stochastic systems. Autom. Remote Control 2011, 72, 1925–1940. [Google Scholar] [CrossRef]
Figure 1. Implementation scheme of the measurement update step in Algorithm 4 based on the MWGS-UD transformation.
Figure 1. Implementation scheme of the measurement update step in Algorithm 4 based on the MWGS-UD transformation.
Mathematics 10 00126 g001
Figure 2. The values of the API identification criterion J ε ( θ , K ) depending on values of parameter θ , calculated using Algorithms 1 and 5.
Figure 2. The values of the API identification criterion J ε ( θ , K ) depending on values of parameter θ , calculated using Algorithms 1 and 5.
Mathematics 10 00126 g002
Figure 3. The values of the API gradient J ε ( θ , K ) θ depending on values of parameter θ , calculated using Algorithms 4 and 5.
Figure 3. The values of the API gradient J ε ( θ , K ) θ depending on values of parameter θ , calculated using Algorithms 4 and 5.
Mathematics 10 00126 g003
Table 1. Numerical values of parameters.
Table 1. Numerical values of parameters.
ParameterValue
τ 1 s
g 9.81 m / s 2
a 0.6378245 · 10 7 m
b 1 e γ τ 1 γ τ
a 1 H 1 1 b 1 2 H 1 2 γ τ
n G y 0.48 · 10 6 rad / s
H 1 0.10 · 10 3 m / s 2
γ 0.20 · 10 3 s 1
v k σ ξ k , ξ k N ( 0 , 1 )
σ 0.1 m / s
w k w k N ( 0 , 1 )
Table 2. Performance of the API-based identification of the model parameter γ .
Table 2. Performance of the API-based identification of the model parameter γ .
“True” value γ 0.0002
Resulting estimate θ ^ min 1.9999 · 10 4
J ε ( θ ^ min , K ) 0.0196
Relative estimation error | | γ θ ^ min | | / | | θ ^ min | | 7.8173 · 10 6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tsyganov, A.; Tsyganova, J. Algorithmic Differentiation of the MWGS-Based Arrays for Computing the Information Matrix Sensitivity Equations within the Problem of Parameter Identification. Mathematics 2022, 10, 126. https://doi.org/10.3390/math10010126

AMA Style

Tsyganov A, Tsyganova J. Algorithmic Differentiation of the MWGS-Based Arrays for Computing the Information Matrix Sensitivity Equations within the Problem of Parameter Identification. Mathematics. 2022; 10(1):126. https://doi.org/10.3390/math10010126

Chicago/Turabian Style

Tsyganov, Andrey, and Julia Tsyganova. 2022. "Algorithmic Differentiation of the MWGS-Based Arrays for Computing the Information Matrix Sensitivity Equations within the Problem of Parameter Identification" Mathematics 10, no. 1: 126. https://doi.org/10.3390/math10010126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop