Next Article in Journal
Experimental Investigation on the Effect of Coil Shape on Planar Eddy Current Sensor Characteristic for Blade Tip Clearance
Previous Article in Journal
Anomaly Detection Method for Industrial Control System Operation Data Based on Time–Frequency Fusion Feature Attention Encoding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Augmented Physics-Based Models for High-Order Markov Filtering

1
Electrical and Computer Engineering Department, Northeastern University, Boston, MA 02115, USA
2
Department of Cybernetics, University of West Bohemia, Univerzitní 8, 30100 Plzeň, Czech Republic
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 6132; https://doi.org/10.3390/s24186132
Submission received: 8 August 2024 / Revised: 13 September 2024 / Accepted: 20 September 2024 / Published: 23 September 2024
(This article belongs to the Section Navigation and Positioning)

Abstract

:
Hybrid physics-based data-driven models, namely, augmented physics-based models (APBMs), are capable of learning complex state dynamics while maintaining some level of model interpretability that can be controlled through appropriate regularizations of the data-driven component. In this article, we extend the APBM formulation for high-order Markov models, where the state space is further augmented with past states (AG-APBM). Typically, state augmentation is a powerful method for state estimation for a high-order Markov model, but it requires the exact knowledge of the system dynamics. The proposed approach, however, does not require full knowledge of dynamics, especially the Markovity order. To mitigate the extra computational burden of such augmentation we propose an approximated-state APBM (AP-APBM) implementation leveraging summaries from past time steps. We demonstrate the performance of AG- and AP-APBMs in an autoregressive model and a target-tracking scenario based on the trajectory of a controlled aircraft with delay-feedback control. The experiments showed that both proposed strategies outperformed the standard APBM approach in terms of estimation error and that the AP-APBM only degraded slightly when compared to AG-APBM. For example, the autoregressive (AR) model simulation in our settings showed that AG-APBM and AP-APBM reduced the estimate error by 31.1 % and 26.7 %. The time cost and memory usage were reduced by 37.5 % and 20% by AP-APBM compared to AG-APBM.

1. Introduction

State estimation and filtering with noisy measurements is an essential component in numerous information processing engineering applications [1]. The filtering process usually involves two steps: the prediction through the transition model and a correction based on the observed measurements and associated model. Based on the complexity of the models, the filters can be classified into two categories: (1) linear filters, e.g., Kalman filter (KF) for the linear models [2], and (2) nonlinear filters, e.g., extended Kalman filters (EKF) [3], sigma-point Kalman filters [4,5], and particle filters [6]. H filter is an alternative filtering technique other than Kalman filters, which works especially for unknown noise statistics and the worst-case estimation error [7]. H filters are also applied to varies applications, e.g., network systems [8], target tracking [9], battery charging [10], etc. In this article, we focus on Kalman filter techniques with given noise statistics.
When considering nonlinear dynamics, machine learning (ML) strategies are appealing when dealing with complex models due to their flexibility and effectiveness in constructing mappings and capturing intricate patterns. However, purely data-driven ML solutions, which do not leverage the knowledge brought by physics-based models (PBM), lack interpretability of the physical meaning of estimated quantities, which is especially relevant when one aims at recovering latent states [11,12]. To incorporate the information from physics knowledge, hybrid ML algorithms are usually used to provide corrections to the estimation [13,14] through neural networks (NNs) or estimate the state directly [15]. In parallel, it has been shown that the NN parameter learning process can be interpreted as a state estimation filtering problem [16,17], which can be leveraged to design sequential training schemes that jointly estimate the state process and NN parameters. This training approach is exploited by augmented physics-based models (APBMs) [18] where physics-based models are augmented by a data-driven component that complements or learns the behaviors that physics-based models cannot represent. A survey of augmented physics-based models for navigation systems is presented in [19].
High-order Markov chains are widely applied and are able to achieve better performance in many time-series data processing applications, especially when data latency or communication delay occurs, e.g., biological sequence analysis [20], speech recognition [21], classification and detection problem [22], and autoregressive process estimation [23,24]. Current solutions to this problem usually involve a more complex system model. In those situations the APBM has great potential, where the complexity that can be captured by the NN part while the physics-based component of the model is kept simple.
In this paper, we focus on adapting the APBM approach to high-order Markov models when the system dynamics and the order of the Markovity are not given. More precisely, we augment the state of the high-order Markov model to fit the APBM and train the parameters of the NN with longer-memory data. An approximation-based method is also put forward to reduce the computational cost brought by the higher-dimensional augmented state. Section 2 introduces these two proposed APBM-based methods. Finally, a tracking experiment based on a delayed-feedback control application is discussed in Section 3 to validate the two proposed approaches. The remainder of this section quickly reviews APBMs and how high-order Markovianity is typically handled in filtering contexts.

1.1. Augmented Physics-Based Model

Consider the Markovian dynamics and measurement models
x k = f ( x k 1 ) + w k 1 x y k = h ( x k ) + v k y ,
where x k R d x × 1 is the state vector at time k and y k R d y × 1 is the measurement vector at time k. f ( · ) and h ( · ) are, possibly nonlinear, vector functions describing the state dynamics and measurement equations, respectively. w k 1 x N ( 0 , Q x ) is the zero-mean Gaussian noise of dynamics model and v k y N ( 0 , R y ) is the zero-mean Gaussian noise of measurement model. The proposed hybrid NN framework is able to learn the dynamics (1) from both sampling data and physics knowledge. The state dynamics, as described by an APBM, can be expressed as
x k = g ˘ f ¯ ( x k 1 ) , x k 1 ; θ + w k 1 x ,
where f ¯ ( · ) : R d x R d x is the PBM, which can be obtained from the simplification, approximation, or partial knowledge of the true dynamics model f ( · ) . g ˘ ( · ) : R d x × R d x R d x is a vector-valued function, including a NN parameterized by θ R d θ to compensate the mismatch of the PBM f ¯ compared with true dynamics function f.
Furthermore, the regularization method involving a parameter value θ ¯ is introduced in the APBM framework [18] to prevent the NN augmentation from completely taking over the model dynamics and neglecting the physics. The value θ ¯ R d θ is defined such that the augmented model is equivalent to the PBM when θ = θ ¯ : g ˘ f ¯ ( x k 1 ) , x k 1 ; θ = θ ¯ = f ¯ ( x k 1 ) .
The learning process infers the state x k and the parameter vector θ estimates. Considering the Bayesian estimation training approach [16,17], based on augmentation of the state x k by the parameter vector θ k , the state-space dynamics and measurement model can be rewritten as
θ k x k = θ k 1 g ˘ f ¯ ( x k 1 ) , x k 1 ; θ k 1 + w k 1 θ w k 1 x y k θ ¯ = h ( x k ) θ k + v k y v k θ ,
where w k θ N ( 0 , Q θ ) denotes the zero-mean Gaussian noise for NN parameter dynamics. The equation θ ¯ = θ k + v k θ in (3) serves for regularization purposes and can be perceived as a soft constraint of the APBM with respect to the PBM. It has a form of the pseudo-measurement with noise v k θ N ( 0 , 1 λ I ) , where λ is the user-defined parameter that controls penalization of distance between θ k and θ ¯ .

1.2. Augmented State for High-Order Markov Models

The transition model shown in (1) actually follows a first-order Markov chain, which means the current state is independent of all previous states except the most recent one [25] (Ch. 13). In a probabilistic framework, the joint distribution for a sequence of K states [ x 1 , , x K ] under such a model (ignoring the measurements) is given by
p ( x 1 , , x K ) = p ( x 1 ) k = 2 K p ( x k | x k 1 ) .
The first-order Markov model is a general and useful assumption in many applications, but it is restrictive in others. In reality, the sequential observations usually indicate a trend, which means the past data could provide nontrivial information to the next prediction. In this case, we model a high-order Markov chain [25] (Ch. 13) and [26], which allows the prediction to depend on a sequence of previous states. For instance, the joint distribution of the states in a second-order Markov chain is given by
p ( x 1 , , x K ) = p ( x 1 ) p ( x 2 | x 1 ) k = 3 K p ( x k | x k 1 , x k 2 ) .
A common approach to deal with high-order Markov is to augment the state vector with the previous states [23,27,28] so that the model becomes first-order Markovian. The augmented-state example of a second-order linear Markov model is shown in [29] (Ch. 6). Considering a general pth-order Markov model for the dynamics, Equation (1) can be expressed as
x k = f ( x k 1 , , x k p ) + w k 1 x y k = h ( x k ) + v k y ,
where x k R d x × 1 is the state vector at time k, y k R d y × 1 is the measurement vector at the time k. To filter this model, the conventional technique is to augment the state as x ˜ k = [ x k , , x k p + 1 ] and the transition model becomes
x ˜ k = f ( x ˜ k 1 ) I d x p × d x p 0 d x p × 1 x ˜ k 1 + w k 1 x 0 d x p × 1 ,
where I and 0 denote the identity and zero matrices with the dimensions indicated in the sub-indices. This augmented state (AGS) transforms the high-order Markovity into a first-order Markov model. After the augmentation, standard Bayesian filtering techniques can be used on the transformed first-order Markov model. In this article, due to the nonlinear nature of the APBM and measurement models, we employ a Cubature Kalman filter (CKF) in the experiments, which propagates the cubature points through the transition and measurement model [5], although other filtering solutions could be considered without loss of generality.

1.3. Contributions

State estimation in high-order Markov models presents a significant challenge due to the dependence of system dynamics on previous states. A widely-used method to address this issue is the aforementioned AGS approach. However, AGS is not applicable when (1) the full knowledge of the complex system dynamics (associated with the high-order Markov model) is unavailable or (2) the order of Markovity is unknown. In this paper, we introduce the augmented-state APBM, which leverages the APBM technique to learn system dynamics via the AGS method, allowing for state estimation even with partially known system dynamics or an unknown Markovian order. To reduce the computational burden brought by the state augmentation (based on the order of Markovity), we proposed the approximated-state APBM by using the point estimate of the posterior distribution of the previous states as the training input, while this approximation introduces a slight degradation in estimation accuracy, it significantly reduces computational costs by lowering the dimension of the state space.

2. Augmented-State APBM for High-Order Markov Models

In this section, we extend APBMs to cope with high-order Markov models, aiming at learning the dynamics from data while constraining the augmented model around the PBM. Since high-order Markov models considered in this paper do not lead to changes in the measurement model (1), we focus next on the transition model and associated prediction process.

2.1. Augmented-State APBM

APBMs are appealing when the nonlinear dynamics f ( · ) of the pth-order Markov model in (6) are not accurately known, either due to their parametric representation and/or the knowledge of the Markov process order. In this case, we employ APBM to learn the transition model
x k = g f ¯ ( x k 1 ) , x k 1 , , x k l ; θ + w k 1 ,
where g ( · ) : R d x × × R d x l R d x includes the PBM and a NN parameterized by θ R d θ . We use g instead of g ˘ to denote the APBM function that accounts for longer state memory. l is a design parameter that is intuitively assumed to be chosen such that l p , although simulations show that reasonable results can be obtained otherwise.
For the pth-order Markov chain, the predictive distribution of the state can be computed as
p ( x k | y 1 : k 1 ) = p ( x k | x k 1 , , x k p , y 1 : k 1 ) · p ( x k 1 : k p | y 1 : k 1 ) d x k 1 d x k 2 d x k p .
However, the joint posterior distribution p ( x k 1 : k p | y 1 : k 1 ) is usually not accessible. As mentioned before, the typical approach to tackling this problem is through state augmentation. Considering a pth-order Markovian model (6) and the APBM (8), augmentation of the state vector with the l 1 previous states results in x ˜ k = [ x k , , x k l + 1 ] . The augmented-state APBM (AG-APBM) is defined by
x ˜ k = g f ¯ ( x k 1 ) , x k 1 , , x k l ; θ k 1 I d x l × d x l 0 d x l × 1 x ˜ k 1 + w k 1 x 0 d x l × 1 g ˜ ( x ˜ k 1 ; θ k 1 ) + w ˜ k 1 x θ k = θ k 1 + w k 1 θ ,
where w ˜ k x N 0 , diag [ Q x , 0 ( p 1 ) × ( p 1 ) ] denotes the augmented state processing noise. The first row in x ˜ k is the original pth-order Markov transition model and the other rows represent the dynamics of the augmentation. For the joint state-parameter estimation we assume the same augmented measurement model as in (3) since it regularizes the NN contribution to the dynamics, i.e., g ˜ ( x ˜ k 1 ; θ = θ ¯ ) = f ¯ ( x k 1 ) [18]. For this augmented state, the predictive distribution is given by
p ( x ˜ k | y 1 : k 1 ) = p ( x ˜ k | x ˜ k 1 , y 1 : k 1 ) · p ( x ˜ k 1 | y 1 : k 1 ) d x ˜ k 1 .
The above Equation (11) has the same quantities with Equation (9), but different meaning and feasibility. The posterior p ( x k 1 : k p | y 1 : k 1 ) in Equation (9) is not easy to compute because it is in the joint form. However, notice that by augmenting the vector, the posterior p ( x ˜ k 1 | y 1 : k 1 ) is not the joint distribution anymore, but a high-dimensional simpler posterior distribution, as it only involves a single variable x ˜ k 1 . This posterior distribution is accessible during the filtering process. For example, in a Kalman filter, we recursively compute the posterior for state estimation at each time step.

2.2. Approximated-State APBM

The augmentation approach discussed above can lead to very high-dimensional state spaces if l is too large; therefore, it increases the computational cost of such a solution. This cost is mostly due to the need to solve multi-dimensional integrals in the Bayesian filtering Equation (11). To circumvent this issue, using the point estimate from the previous steps—rather than using the entire distribution—reduces the computational complexity in evaluating the aforementioned integrals. This leads to the approximated-state APBM (AP-APBM):
x k = g f ¯ ( x k 1 ) , x k 1 , x ^ k 2 , , x ^ k l ; θ k 1 + w k 1 θ k = θ k 1 + w k 1 θ ,
where x ^ k j = E { x k j | y 1 : k j } , j = 2 , , l . It is worth noting that approximations have to be made to obtain the above model. Equation (13) provides a mathematical explanation of this intuition by approximating the posterior distribution previous to its use in the integral.
p ( x k 2 : k l | y 1 : k 1 ) i = 2 l p ( x k i | y 1 : k 1 ) i = 2 l p ( x k i | y 1 : k i ) i = 2 l δ x k i E { x k i | y 1 : k i } ,
where δ ( · ) is the Dirac delta function. Note that in the above approximation, we first assume that the posterior distributions at each time step are independent, and then we neglect the dependency between states and future measurements. In addition, we finally replace the posterior distribution of the states with their point estimates, which is equivalent to the model in (12).
Finally, we present the AP-APBM
θ k x k = θ k 1 g f ¯ ( x k 1 ) , x k 1 , x ^ k l : k 2 ; θ k 1 + w k 1 θ w k 1 x y k θ ¯ = h ( x k ) θ k + v k y v k θ .
where it can be observed that although the AP-APBM needs memory from past state estimates [ x ^ k 1 , , x ^ k l ] , its computational cost is much lower than the cost of the AG-APBM since the dimension of the state is reduced significantly.

3. Numerical Simulations

In this section, we implement the proposed AG- and AP-APBM approaches in the context of the high-order Markov models, particularly with the aim of performing state estimation in two common applications: the autoregressive (AR) model problem and the time-delayed control problem.

3.1. AR Model

The AR model is one of the most common high-order Markov models. Here, we implemented the proposed approaches to filter the states generated by an AR(3) model based on noisy measurements. Considering the 2-dimensional system state x = [ x 1 , x 2 ] and the 2-dimensional measurement y = [ y 1 , y 2 ] , the transition and measurement models are given by
x k = F 1 x k 1 + F 2 x k 2 + F 3 x k 3 + w k 1 x
y k = H x k + v k y ,
where w k 1 x N ( 0 , Q x ) and v k y N ( 0 , R y ) are additive Gaussian noise with Q x = σ x 2 I 2 × 2 , R y = σ y 2 I 2 × 2 and σ x = 0.1 , σ y = 0.1 . H = I 2 × 2 and
F 1 = 0.5 0.3 0.4 0.2 F 2 = 0.2 0.1 0.1 0.2 F 3 = 0.1 0.05 0.05 0.1 .
We then compared the performance of the estimation of the filtering process when using the true model above, the PBM-only model, AG- and AP-APBM in the 1st order and 3nd order. It is noted that the true model is constructed based on the AGS method as shown in Equation (7). In the experiments, the PBM refers to the below AR(1) model
x k = F 1 x k 1 + w k 1 x .
The APBM is a linear combination between the PBM and multilayer perceptrons (MLP) as shown below
g θ ( x k 1 , · ) = w 0 F 1 x k 1 + w 1 γ ϕ ( x k 1 , · ) ,
where θ = [ w 0 , w 1 , ϕ ] and the MLP consists of 1 hidden layer with 5 units and ReLu activation function and output layer with two output units and linear activation function. The second argument in the function (19) can be null for 1st-order APBM, [ x k 2 , x k 3 ] for 3rd-order AG-APBM, and [ x ^ k 2 , x ^ k 3 ] for 3rd-order AP-APBM. Figure 1 shows the root mean square errors (RMSEs) of the estimation using the measurements from the AR(3) model, which are computed based on 200 Monte Carlo (MC) simulations. The PBM gives the largest error because it employs limited knowledge with an AR(1) model, while the true model always achieves the best precision since it has full knowledge of the dynamics model. All three APBMs have better estimation precision compared to the PBM. In terms of the estimate error median, the 1st-order APBM reduces the error from PBM by 22.2 %, the 3rd-order AG-APBM reduces the error by 31.1 %, and the 3rd-order AP-APBM reduces the error by 26.7 %. Among the hybrid learning approaches, the 3rd-order AG-APBM has the best performance since it learns the dynamics through the augmented state with the last two steps. Instead of incorporating the last two steps into the state, the 3rd-order AP-APBM leverages the point estimate of the last two steps as the input of the NN. The median of the RMSE increases about 6.5% compared to 3rd-order AG-APBM, but still lower than the 1st-order APBM. More importantly, the computational cost is significantly reduced by the approximation. Figure 2 shows that the time cost of the 3rd-order AP-APBM decreases by about 37.5 % compared with the 3rd-order AG-APBM and the memory usage decreases by about 20% over 200 MC simulations. The simulations are implemented based on the following:
  • Processor (CPU): Intel Core i7-10700KF, 8 cores, 3.80 GHz. The multi-core CPU allowed for efficient parallel processing during data preprocessing.
  • Memory (RAM): 32 GB DDR4.
  • Storage: 1 TB NVMe SSD.
  • Operating System: Windows 11 Pro.
  • Software Platform: MATLAB R2023a.
It is shown that for AP-APBM, the time cost and memory usage are larger than the cost of the 1st-order APBM, since the extra memory is used to store the approximation { x ^ k 2 , x ^ k 3 } when estimating x ^ k . However, the computational cost of the AP-APBM has been greatly reduced compared to the AG-APBM excluding the extra augmented state.

3.2. A Delayed-Feedback Control Nonlinear Model

We test the different approaches based on a target tracking problem, where the target is moving on a two-dimensional horizontal plane. The target is operates with an internal feedback control loop, unobserved to the tracker, where communication delays usually occur inside [30,31]. The corresponding dynamics and measurement models are also common in target tracking problems [32]. Considering the state [ x , Ω ] , where x = [ p x , v x , p y , v y ] is the two-dimensional position and velocity, Ω is the turning rate of the target and the feedback control input u , the dynamics model is given by
x k = F x k 1 + B u k 1 + w k 1 x ,
Ω k = Ω k 1 + w k 1 Ω ,
where w k 1 x N 0 , diag ( 0.1 , 0.1 , 0.1 , 0.1 ) , w k 1 Ω N 0 , 10 4 , F defines the constant-velocity model
F = 1 T s 0 0 0 1 0 0 0 0 1 T s 0 0 0 1 ,
T s = 1 s denotes the sampling period, B = I 4 × 4 is the identity matrix, and u k 1 = s k c G k 1 ( x k 3 x ¯ k 1 ) is the input controlled by the difference between the previous state x k 3 and reference x ¯ k 1 . s ( A ) : R 4 × 4 R 4 × 4 denotes a saturation function which restricts all elements in the matrix A within [ 5 , 5 ] due to the possible limitation of the actuator and k c = 1 is the controller gain. Specifically, we introduce G k 1 , controlling the state only by the difference between the true velocity and the reference.
G k 1 = 0 sin Ω k 1 T s Ω k 1 0 1 cos Ω k 1 T s Ω k 1 0 cos Ω k 1 T s 0 sin Ω k 1 T s 0 1 cos Ω k 1 T s Ω k 1 0 sin Ω k 1 T s Ω k 1 0 sin Ω k 1 T s 0 cos Ω k 1 T s .
The error is usually defined between the current state x k 1 and the reference state x ¯ k 1 at the same moment k 1 . However, in this case, it is computed based on the state before two steps x k 3 , due to communication delay, which means the aircraft can access its state until two steps from the current time. (Note that the true state is known to the aircraft but it is unknown to the tracker and is thus estimated by the tracker.) Notice that G k 1 introduces the high-order Markovian property into the dynamics. The measurement model is given by the received signal strength and bearings from two collocated sensors
y k = 10 log 10 Ψ 0 p 0 p k q ( p 0 , p k ) + v k y ,
with p 0 = ( 0 , 0 ) being the position of the sensors, p k = ( p x , k , p y , k ) the unknown position of the target, 10 log 10 Ψ 0 = 30 dBm, q = 2.2 the path loss exponent, ( p 0 , p k ) denoting the angle between locations p 0 and p k in radians, and v k y N 0 , diag ( 1 , 0.1 ) the measurement noise. The above model can be rewritten as a third-order Markov form
x k = f ( x k 1 , , x k 3 ) + w k 1
y k = h ( x k ) + v k .

3.3. AG-APBM and AP-APBM Performance

We implement CKF to estimate the state of interest x = [ p x , v x , p y , v y ] in 200 MC experiments. We will compare the filtering performance of the AGS approach as described in (7), 1st-order Markov APBM in (3), AG-APBM in (10), AP-APBM in (14), pure NN, and pure PBM. It is noted that the AGS approach is considered as the benchmark in this tracking problem, since it has the full knowledge of the dynamics, including the Markovity order and the control input, while the other approaches are not aware of these facts. The constant-velocity model f ¯ ( x k 1 ) = F x k 1 is used as the PBM. Here, the APBM is similar to Equation (19), which consists of the PBM and MLP, denoted γ ϕ ( · ) and parameterized by ϕ , with appropriate inputs such that g θ ( x k 1 , · ) = w 0 F x k 1 + w 1 γ ϕ ( x k 1 , · ) , where θ = [ w 0 , w 1 , ϕ ] . All NNs have one hidden layer with five hidden units and ReLu activation functions, and output layer with dimension d x = 4 and linear activation functions. It is noted that the six approaches leverage different-level information of the transition model. The AGS has full knowledge of the system, while 1st-order Markov APBM is only given the assumed PBM. The AG-APBM and AP-APBM are aware of the existence of the high-order Markov process, but not the exact order, where they use l = 5 instead. Moreover, our reference is designed for the control problem, but the APBM approaches do not have the knowledge of the existence of the reference.
Figure 3 shows the box plot of the RMSE of each approach. Figure 4 shows the averaged experimental cumulative distribution function (CDF) over the 200 experiments. Both plots show statistically significant improvements of APBM approaches when compared with both PBM and NNs. In terms of the median, the error of 1st-order APBM is reduced by 41.2 %, the error of AG-APBM is reduced by 58.8 %, and the error of AP-APBM is reduced by 52.9 % compared to PBM. The performances of APBMs are only surpassed by the true model, AGS. When comparing AG- and AP-APBM results the difference is minimal with AP-APBM leading to slightly higher errors. We also highlight that the two proposed approaches led to RMSE values that are very close to the ones obtained with the AGS, despite the perfect knowledge of the order of the Markov process and the transition model assumed in the AGS implementation.

4. Conclusions and Future Work

In this study, we extended the APBM framework to deal with high-order nonlinear Markov processes. To this end, we proposed two different implementations with different levels of computational complexity. In the first, we proposed the AG-APBM, where we augmented the state space with past states, thus considerably increasing the number of computations required for the estimation process. In the second approach, AP-APBM, we mitigated the additional complexity by approximating past states by their point estimates, eliminating the state augmentation requirements. Simulated experiments demonstrated the performance of the proposed models in an AR(3) model state estimation and a nonlinear third-order Markov target tracking scenario. The estimation results from the AR model indicated that the high-order AG- and AP-APBM reduced the error by about 30% compared the PBM. Both of them outperformed the legacy 1st-order APBM since they leveraged more state information from the previous steps. Furthermore, the performance of AP-APBM only degrades slightly from the performance of AG-APBM, while significantly reducing the computational cost with 37.5 % reduction in time cost and 20% reduction in memory usage based on our simulation environment. The tracking results showed that the AG-APBM and AP-APBM reduced the estimate error by 58.8 % and 52.9 % compared with the error of PBM even without precise knowledge of the order of the underlying Markov process. Both of them result in a lower RMSE than the standard 1st-order APBM does.
As we mentioned earlier, this article focuses on the state estimation under unknown dynamics (especially the unknown order of Markovity) by learning the system dynamics with PBM. In relation to this, future works could focus on: (1) dealing with unknown noise parameters. The noise statistics is also an important factor in the dynamics system. The noise can be estimated by the maximum likelihood method [29] or the correlation measurement difference method [33]. It is also shown that the data-driven algorithm is effective when learning the noise statistics, e.g., Long short-term memory Kalman filters [34], EKFNet [35] and KalmanNet [36]; (2) considering more general state-space distributions beyond the widely used Gaussian assumption. That is, while APBMs focused on Gaussian models, the general concept can be applied to non-Gaussian setups; and (3) using a different NN structure to learn the high-Markov dynamics. We implemented an MLP to learn the system dynamics in the simulation. The recurrent neural network (RNN), long-short term memory (LSTM) [37], and Transformer [38] may be used due to their mechanisms of forgetting and memorizing the past information especially when the order of the Markovity is not given.

Author Contributions

Conceptualization, T.I.; Methodology, S.T., J.D. and O.S.; Software, S.T. and T.I.; Validation, S.T.; Writing—original draft, S.T.; Writing—review & editing, J.D., O.S. and P.C.; Supervision, P.C.; Funding acquisition, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been partially supported by the NSF under Awards ECCS-1845833 and CCF-2326559.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The article provides a definition for all the symbols and variables that are used in the derivations. For the convenience of the reader, a list of these notations is provided here.
x k State vector at time instance k
y k Measurement vector at time instance k
w k x Noise of the dynamics model
v k y Noise of the measurement model
h ( · ) Possibly nonlinear measurement model
f ( · ) Possibly nonlinear true dynamics model
f ¯ ( · ) Physics-based model (PBM)
g ˘ ( · ) Augmented physics-based model (APBM)
θ Neural network (NN) parameters
w k θ Noise of NN parameter dynamics model
θ ¯ Pseudo-measurement for NN parameter regularization
v k θ Noise of NN parameter pseudo-measurement model
x ˜ k Augmented state vector at time instance k
g ˜ ( · ) Augmented-state APBM (AG-APBM)
w ˜ k x Noise of AG-APBM
x ^ k Estimated state at time instance k
δ ( · ) Dirac delta function

References

  1. Duník, J.; Biswas, S.K.; Dempster, A.G.; Pany, T.; Closas, P. State Estimation Methods in Navigation: Overview and Application. IEEE Aerosp. Electron. Syst. Mag. 2020, 35, 16–31. [Google Scholar] [CrossRef]
  2. Särkkä, S.; Svensson, L. Bayesian Filtering and Smoothing; Cambridge University Press: Cambridge, UK, 2023; Volume 17. [Google Scholar]
  3. Kalman, R.E. Contributions to the theory of optimal control. Bol. Soc. Mat. Mex. 1960, 5, 102–119. [Google Scholar]
  4. Wan, E.A.; Van Der Merwe, R. The unscented Kalman filter. In Kalman Filtering and Neural Networks; Wiley: Hoboken, NJ, USA, 2001; pp. 221–280. [Google Scholar]
  5. Arasaratnam, I.; Haykin, S. Cubature Kalman filters. IEEE Trans. Autom. Control 2009, 54, 1254–1269. [Google Scholar] [CrossRef]
  6. Arulampalam, M.S.; Maskell, S.; Gordon, N.; Clapp, T. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process. 2002, 50, 174–188. [Google Scholar] [CrossRef]
  7. Simon, D. Optimal State Estimation: Kalman, H Infinity, and Nonlinear Approaches; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  8. Yan, S.; Gu, Z.; Park, J.H.; Shen, M. Fusion-Based Event-Triggered H-infinity State Estimation of Networked Autonomous Surface Vehicles With Measurement Outliers and Cyber-Attacks. IEEE Trans. Intell. Transp. Syst. 2024, 25, 7541–7551. [Google Scholar] [CrossRef]
  9. Xia, J.; Gao, S.; Qi, X.; Zhang, J.; Li, G. Distributed cubature H-infinity information filtering for target tracking against uncertain noise statistics. Signal Process. 2020, 177, 107725. [Google Scholar] [CrossRef]
  10. Chen, Z.; Zhou, J.; Zhou, F.; Xu, S. State-of-charge estimation of lithium-ion batteries based on improved H infinity filter algorithm and its novel equalization method. J. Clean. Prod. 2021, 290, 125180. [Google Scholar] [CrossRef]
  11. Haoqing, L.; Borsoi, R.A.; Imbiriba, T.; Closas, P.; Bermudez, J.C.M.; Erdoğmuş, D. Model-Based Deep Autoencoder Networks for Nonlinear Hyperspectral Unmixing. IEEE Geosci. Remote. Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  12. Imbiriba, T.; Wu, P.; LaMountain, G.; Erdoğmuş, D.; Closas, P. Recursive Gaussian Processes and Fingerprinting for Indoor Navigation. In Proceedings of the 2020 IEEE/ION Position, Location and Navigation Symposium (PLANS), Portland, ON, USA, 20–23 April 2020; pp. 933–940. [Google Scholar]
  13. Chin, L. Application of neural networks in target tracking data fusion. IEEE Trans. Aerosp. Electron. Syst. 1994, 30, 281–287. [Google Scholar] [CrossRef]
  14. Vaidehi, V.; Chitra, N.; Krishnan, C.; Chokkalingam, M. Neural network aided Kalman filtering for multitarget tracking applications. Comput. Electr. Eng. 2001, 27, 217–228. [Google Scholar]
  15. Gao, C.; Yan, J.; Zhou, S.; Varshney, P.; Liu, H. Long short-term memory-based deep recurrent neural networks for target tracking. Inf. Sci. 2019, 502, 279–296. [Google Scholar] [CrossRef]
  16. Singhal, S.; Wu, L. Training multilayer perceptrons with the extended Kalman algorithm. Adv. Neural Inf. Process. Syst. 1988, 1. Available online: https://proceedings.neurips.cc/paper/1988/file/38b3eff8baf56627478ec76a704e9b52-Paper.pdf (accessed on 7 August 2024).
  17. Haykin, S. Kalman Filtering and Neural Networks; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  18. Imbiriba, T.; Demirkaya, A.; Duník, J.; Straka, O.; Erdoğmuş, D.; Closas, P. Hybrid neural network augmented physics-based models for nonlinear filtering. In Proceedings of the 2022 25th International Conference on Information Fusion (FUSION), Linköping, Sweden, 4–7 July 2022; pp. 1–6. [Google Scholar]
  19. Imbiriba, T.; Straka, O.; Duník, J.; Closas, P. Augmented physics-based machine learning for navigation and tracking. IEEE Trans. Aerosp. Electron. Syst. 2024, 60, 2692–2704. [Google Scholar] [CrossRef]
  20. Krogh, A. An introduction to hidden Markov models for biological sequences. In New Comprehensive Biochemistry; Elsevier: Amsterdam, The Netherlands, 1998; Volume 32, pp. 45–63. [Google Scholar]
  21. Lee, L.M.; Lee, J.C. A study on high-order hidden Markov models and applications to speech recognition. In International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems; Springer: Berlin/Heidelberg, Germany, 2006; pp. 682–690. [Google Scholar]
  22. Salnikov, V.; Schaub, M.T.; Lambiotte, R. Using higher-order Markov models to reveal flow-based communities in networks. Sci. Rep. 2016, 6, 23194. [Google Scholar] [CrossRef]
  23. Urteaga, I.; Djurić, P.M. Sequential estimation of hidden ARMA processes by particle filtering—Part one. IEEE Trans. Signal Process. 2016, 65, 482–493. [Google Scholar] [CrossRef]
  24. Djuric, P.M.; Kay, S.M. Order selection of autoregressive models. IEEE Trans. Signal Process. 1992, 40, 2829–2833. [Google Scholar] [CrossRef]
  25. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4. [Google Scholar]
  26. Raftery, A.E. A model for high-order Markov chains. J. R. Stat. Soc. Ser. B Stat. Methodol. 1985, 47, 528–539. [Google Scholar] [CrossRef]
  27. Terwijn, B.; Porta, J.M.; Kröse, B.J. A particle filter to estimate non-Markovian states. In Proceedings of the International Conference on Intelligent Autonomous Systems, IAS, Singapore, 22–25 June 2004; Volume 4, pp. 1062–1069. [Google Scholar]
  28. Jin, B.; Guo, J.; He, D.; Guo, W. Adaptive Kalman filtering based on optimal autoregressive predictive model. GPS Solut. 2017, 21, 307–317. [Google Scholar] [CrossRef]
  29. Shumway, R.H.; Stoffer, D.S.; Stoffer, D.S. Time Series Analysis and Its Applications; Springer: Berlin/Heidelberg, Germany, 2000; Volume 3. [Google Scholar]
  30. Kerns, A.J.; Shepard, D.P.; Bhatti, J.A.; Humphreys, T.E. Unmanned aircraft capture and control via GPS spoofing. J. Field Robot. 2014, 31, 617–636. [Google Scholar] [CrossRef]
  31. Kling, M.T.; Lau, D.; Witham, K.L.; Closas, P.; LaMountain, G.M. System for Closed-Loop GNSS Simulation. US Patent Application No. 17/662,822, 10 May 2022. [Google Scholar]
  32. Bar-Shalom, Y.; Li, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation: Theory Algorithms and Software; John Wiley & Sons: Hoboken, NJ, USA, 2001. [Google Scholar]
  33. Kost, O.; Duník, J.; Straka, O. Measurement difference method: A universal tool for noise identification. IEEE Trans. Autom. Control 2022, 68, 1792–1799. [Google Scholar] [CrossRef]
  34. Coskun, H.; Achilles, F.; DiPietro, R.; Navab, N.; Tombari, F. Long short-term memory Kalman filters: Recurrent neural estimators for pose regularization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5524–5532. [Google Scholar]
  35. Xu, L.; Niu, R. EKFNet: Learning system noise statistics from measurement data. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 4560–4564. [Google Scholar]
  36. Revach, G.; Shlezinger, N.; Ni, X.; Escoriza, A.L.; Van Sloun, R.J.; Eldar, Y.C. KalmanNet: Neural network aided Kalman filtering for partially known dynamics. IEEE Trans. Signal Process. 2022, 70, 1532–1547. [Google Scholar] [CrossRef]
  37. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef]
  38. Vaswani, A. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. NIPS 2017. [Google Scholar]
Figure 1. Box plots for the root mean square error (RMSE) of state estimation for the autoregressive-3 (AR(3)) model computed over 200 Monte-Carlo (MC) simulations. The central line in the box indicates the median. The bottom and top edges of the box denote the 25th and 75th percentiles, and the circles denotes the outliers. The AGS and PBM represent the estimation based on the true model and AR(1) model.
Figure 1. Box plots for the root mean square error (RMSE) of state estimation for the autoregressive-3 (AR(3)) model computed over 200 Monte-Carlo (MC) simulations. The central line in the box indicates the median. The bottom and top edges of the box denote the 25th and 75th percentiles, and the circles denotes the outliers. The AGS and PBM represent the estimation based on the true model and AR(1) model.
Sensors 24 06132 g001
Figure 2. Computational cost over 200 MC simulations. Each MC repetition includes a filtering process of 600 time steps. The simulations were implemented on MATLAB. The cost of AGS and PBM model are not plotted here because the deficiency of training process makes their computational cost incomparable to the hybrid NN models.
Figure 2. Computational cost over 200 MC simulations. Each MC repetition includes a filtering process of 600 time steps. The simulations were implemented on MATLAB. The cost of AGS and PBM model are not plotted here because the deficiency of training process makes their computational cost incomparable to the hybrid NN models.
Sensors 24 06132 g002
Figure 3. Box plots for the RMSE of position estimations computed over 200 MC simulations. The central line in the box indicates the median. The bottom and top edges of the box denote the 25th and 75th percentiles, and the circles denotes the outliers. The AGS and PBM represents the true model and the constant-velocity model.
Figure 3. Box plots for the RMSE of position estimations computed over 200 MC simulations. The central line in the box indicates the median. The bottom and top edges of the box denote the 25th and 75th percentiles, and the circles denotes the outliers. The AGS and PBM represents the true model and the constant-velocity model.
Sensors 24 06132 g003
Figure 4. Each curve represents the average empirical cumulative distribution function (CDF) of squared error of position estimation. The average was computed over 200 MC simulations.
Figure 4. Each curve represents the average empirical cumulative distribution function (CDF) of squared error of position estimation. The average was computed over 200 MC simulations.
Sensors 24 06132 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, S.; Imbiriba, T.; Duník, J.; Straka, O.; Closas, P. Augmented Physics-Based Models for High-Order Markov Filtering. Sensors 2024, 24, 6132. https://doi.org/10.3390/s24186132

AMA Style

Tang S, Imbiriba T, Duník J, Straka O, Closas P. Augmented Physics-Based Models for High-Order Markov Filtering. Sensors. 2024; 24(18):6132. https://doi.org/10.3390/s24186132

Chicago/Turabian Style

Tang, Shuo, Tales Imbiriba, Jindřich Duník, Ondřej Straka, and Pau Closas. 2024. "Augmented Physics-Based Models for High-Order Markov Filtering" Sensors 24, no. 18: 6132. https://doi.org/10.3390/s24186132

APA Style

Tang, S., Imbiriba, T., Duník, J., Straka, O., & Closas, P. (2024). Augmented Physics-Based Models for High-Order Markov Filtering. Sensors, 24(18), 6132. https://doi.org/10.3390/s24186132

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop