An Improved Adaptive Iterative Extended Kalman Filter Based on Variational Bayesian

: The presence of unknown heavy-tailed noise can lead to inaccuracies in measurements and processes, resulting in instability in nonlinear systems. Various estimation methods for heavy-tailed noise exist. However, these methods often trade estimation accuracy for algorithm complexity and parameter sensitivity. To tackle this challenge, we introduced an improved variational Bayesian (VB)-based adaptive iterative extended Kalman filter. In this VB framework, the inverse Wishart distributionis used as the prior for the state prediction covariance matrix. The system state and noise parameter posterior distributions are then iteratively updated for adaptive estimation. Furthermore, we make adaptive adjustments to the IEKF filter parameters to enhance sensitivity and filtering accuracy, thus ensuring robust prediction estimation. A two-dimensional target tracking and nonlinear numerical UNGM simulation validated our algorithm. Compared to existing algorithms RKF-ML and GA-VB, our method showed significant improvements in RMSE pos and RMSE vel , with increases of 21.81% and 22.11% respectively, and a 49.04% faster convergence speed. These results highlight the method’s reliability and adaptability.

Numerous filters have been proposed to address the issue of heavy-tailed noise impacting system tracking.Notable among these are the robust variational Bayesian filter [18][19][20], Student's t-based filters (RSTFs) [21,22], Huber's M-estimation-based filters [23,24], and the maximum correlation entropy-based filters [25].While these filters demonstrate enhanced tracking performance and effectively mitigate the negative impact of heavy-tailed noise, they face challenges in precise parameter selection and computational complexity.The need for accurate parameter calibration, a potentially time-consuming and obstructive task, combined with the computational demands of these methods, may limit their practical applicability in real-time scenarios.Consequently, there is an urgent need to improve and streamline the parameter selection process, thereby reducing computational complexity.
As a key technique in nonlinear recursive filtering, the Extended Kalman Filter (EKF) is widely recognized and utilized [16,26,27].Its popularity stems from several advantages, such as obviating the need for calculating nominal trajectories, its straightforward methodology, and ease of implementation.However, enhancing the functionality and effectiveness of state estimation filters in real-world applications remains crucial.The Kalman filter algorithm based on Huber [28,29], for example, shows resilience to outliers and superior performance in non-Gaussian settings.Yet, it faces increased computational complexity, requires fine-tuning of parameters, and may be less efficient in Gaussian environments.The Interactive Multiple Model (IMM)-based Kalman filter [30,31] offers improved estimation accuracy and robustness amidst model uncertainties, but this benefit comes with increased computational and implementation complexities, as well as reliance on precise model selection and meticulous parameter tuning.A robust Gaussian approximate filter, integrating unknown scale matrices and degrees of freedom parameters, was introduced in the literature (GA-VB) [32].To enhance this algorithm, the variational Bayesian (VB) method was employed to modify the state and posterior probability density functions (PDFs) for each parameter individually, aiming to reduce the overall complexity while inadvertently increasing sensitivity to parameters.Concurrently, literature (RKF-ML) [33] presented a robust state estimation algorithm using the maximum likelihood (ML) distribution to model heavy-tailed noise.This method notably avoids the need for manually selecting the degrees of freedom (DOF) parameters, thus being less sensitive to free parameters.However, it requires a VB approach to adaptively estimate ML distribution parameters in a closed recursive state, adding to the overall complexity.These algorithms underscore the trade-offs between complexity and sensitivity in algorithm design, highlighting the necessity for a nuanced approach that carefully balances these factors to achieve optimal performance in state estimation algorithms under challenging conditions.
Addressing parameter sensitivity and algorithmic complexity, this paper introduces a new approach: an improved variational Bayesian (VB)-based adaptive iterative extended Kalman filter (VBAIEKF).This method aims to enhance the accuracy of filters handling imprecise heavy-tailed noise while simultaneously addressing parameter sensitivity and reducing algorithmic complexity.Initially, the inverse Wishart (IW) distribution is utilized to model inaccurate measurement noise, treating the system state and measurement noise covariance as parameters to be estimated.To estimate these parameters, the variational Bayesian (VB) method computes the joint posterior probability density functions (PDFs) for these parameters, thereby accomplishing the state estimation of the system.This approach effectively addresses variability in both noise and prediction error covariances.The algorithm exhibits the following highlights: (1) The approach is characterized by a notable insensitivity to its parameters, possessing the ability to withstand variations in parameter values.
(2) The approach exhibits lower computational complexity, rendering it suitable for real-time applications.
(3) This approach exhibits significant adaptability and reliability when confronted with an uncertain environment.
The algorithm proposed in this article adeptly addresses the challenge of concurrently managing noise fluctuations and prediction error covariance in the state estimation process.When compared with two optimized algorithms, GA-VB [32] and RKF-ML [33], the performance of the VBAIEKF is notably impressive.Characterized by low sensitivity to parameters, minimal complexity, excellent convergence, and adaptability, these attributes render it a particularly promising solution for achieving precise state estimation in scenarios with heavy-tailed noise.
The primary framework is outlined as follows.Section 2 delineates the procedural workflow of the VBAIEKF algorithm, where we deduce the filter and emphasize its unique characteristics.In Section 3, we conduct a thorough comparison between the proposed VBAIEKF and the GA-VB [32] and RKF-ML [33] algorithms.This comparison focuses on two-dimensional target tracking with nonlinear uncertain heavy-tailed noise and nonlinear numerical UNGM.The comparative analysis clearly demonstrates the superior performance of the VBAIEKF.Section 4 provides definitive remarks that summarize the findings and implications of the study.

Computational Procedure 2.1. System Model
Consider the following noisy nonlinear system shown by the state space model: where x t ∈ R n is an observable yet unmeasurable state vector, z t = R m is a vector of measurements that may be observed, f t−1 (•) and h t (•) are recognized nonlinear functions.
To simplify computations and ensure the convergence of the Taylor series expansion, assume that f t−1 (•) and h t (•) are, without loss of generality, analytic functions that are sufficiently smooth and differentiable.The covariance matrices of the random variables w t−1 and v t are denoted as Q t and R t , respectively.The process noise w t−1 is a Gaussian process with a mean of zero and an unknown covariance matrix Q t .The measurement noise, denoted as v t , is a type of Gaussian white noise with a mean of zero.The covariance matrix R t associated with this noise is unknown.First, a Taylor Series expansion (TSE) [34], which can be stated as the following, is used to linearize locally the nonlinear functions f t−1 (•) and h t (•): The requirement for bounded higher-order terms (H.O.T) is crucial to guarantee the convergence and accuracy of the Taylor series expansion.Following the application of the Taylor series expansion, the nonlinear functions f t−1 (•) and h t (•) are approximated by linear functions with error terms F t and H t , respectively.
When there is heavy-tailed noise, the covariance matrix of the process and measurement noise is usually imprecise.To resolve this problem, one can utilize a variational Bayesian (VB) method [19,35,36], and subsequently apply an iterative extended Kalman filter to estimate the state vector x t .This work presents an improved variational Bayesian iterative extended Kalman filtering (VBAIEKF) technique designed for tracking estimation.

Prior Distributions
Provide a detailed description of the updating process for prior distributions.In the Kalman filtering paradigm, the likelihood probability density function (PDF) p Zt | Xt and the one-step anticipated PDF p Xt | Z1:t−1 are Gaussian distributions: where the Gaussian distribution Ñ(A; α, σ) has the following mean and variance: α, σ, and its PDF is: The anticipated state vector Xt|t−1 and the related one-step predicted covariance matrix P t|t−1 can be represented as follows: where Xt−1|t−1 and P t−1|t−1 denote the state estimate and estimation error covariance matrix at time t − 1, respectively.Where Xt−1|t−1 and P t−1|t−1 represent the state estimation and the covariance matrix of the estimate error at time t − 1, respectively.The coefficient P t|t−1 in Equation ( 9) is imprecise, and the actual process noise covariance matrix Q t is not known.The inverse Wishart (IW) distribution is considered the conjugate prior distribution for a Gaussian distribution with a specified mean covariance matrix in Bayesian statistics.Variational Bayesian (VB) approaches offer a closed-form analytical solution to this intricate problem.Let us consider the Wishart distribution W φ −1 ; β, θ − 1 , where φ −1 is the inverse matrix of a positive definite matrix φ.The matrix φ is distributed according to the inverse Wishart (IW) distribution [37].
where β > d + where ût|t−1 and Ût|t−1 denote the degrees of freedom and scale matrices of p R t | Z1:t−1 , respectively, tt|t−1 and Tt|t−1 denote the degrees of freedom and scale matrices of p P t|t−1 | Z1:t−1 .Then, it is necessary to determine the values of tt|t−1 , Tt|t−1 , ût|t−1 and Ût|t−1 .
The nominal error covariance matrix P t|t−1 can be set equal to the mean of matrix Pt|t−1 .
Using E(φ) = θ β−d−1 , we arrive at the following: where n stands for the dimension of the state T t|t−1 , and Qt−1 stands for the nominal process noise covariance.Let: where T ≥ 0 is the correction factor.The outcome of entering Equation ( 14) into ( 13) is: The prior probability density function (PDF) distribution p R t−1 | Z1:t−1 of the measurement noise covariance matrix R t−1 is influenced by the inverse Wishart (IW) distribution.Similarly, the posterior PDF distribution p R t−1 | Z1:t−1 of this matrix should also follow an IW distribution, as stated in Equation ( 12): It is unknown what the dynamic model of the noise variance in Equation ( 17) appears like in real-world applications: p(R t | R t−1 ).To preserve and propagate the approximate posterior at the prior moment, we select the forgetting factor ρ and formulate the prior parameters as follows: The forgetting factor ρ reflects the degree of temporal variation, which ranges from 0 to 1, inclusive.A value of ρ = 1 indicates the variance at a steady state, with lower values of ρ corresponding to higher frequencies of variations throughout time.

Posterior PDFs
Here, we elaborate on the meticulous process of updating posterior PDFs.In order to calculate Xt , P t|t−1 , and R t , the joint posterior probability density function p Xt , P t|t−1 , R t | Z1:t needs to be determined.Since there is no analytical solution available for this joint posterior probability density function (PDF), the variational Bayesian (VB) approach is employed to derive an approximate PDF that is factorized and flexible for: where q(•) corresponds to the somewhat posterior PDF of p(•).Therefore, it is possible to produce the VB approximation by minimizing the Kullback-Leibler Divergence (KLD) [38] between the posterior PDFs of the genuine joint p Xt , P t|t−1 , R t | Z1:t and the approximation q Xt , q P t|t−1 , q(R t ): q Xt , P t|t−1 , q(R t ) = arg minKLD   q Xt , q P t|t−1 q(R t ) Therefore, it can also be written as: Additionally, Equations ( 22) and ( 27) can be obtained.q (i+1) (•) represents the approximate PDF of the i + 1th iteration of q(•), and Ã(i) t is defined as follows: where E (i) [ρ] represents the expected value of variable ρ at iteration i-th.
The IW distribution with the parameter of freedom t(i+1) t|t−1 and scale matrix T(i+1) t|t−1 can be regarded as q (i+1) P t|t−1 : where t(i+1) t|t−1 is the parameter of freedom and T(i+1) t| t−1 is the scale matrix: T(i+1) Its comprehensive form is: where k has the following definition: Equation ( 26) states that q (i+1) (R t ) may be thought of as the IW distribution, where û(i+1) t|t−1 stands in for the freedom parameter and Û(i+1) t|t−1 for the scale matrix.
When the scale matrix Û(i+1) t|t−1 and the freedom parameter û(i+1) t|t−1 are represented as: where can be represented as: After the i + 1th iteration update, the one-step predicted PDF and likelihood PDF may be described by the following equations: Following are the formulations for the matrix P(i+1) t|t−1 of the corrected predicted error covariance and matrix R(i+1) t of the measured noise covariance: Equations ( 33)-( 36) can be transformed into a Gaussian distribution with a mean of X(i+1) t|t and a variance of P (i+1) t|t by considering the following: The predicted value X(i+1) t|t and prediction error covariance P (i+1) t|t are then calculated using IEKF at the i + 1th iteration as follows: Equation (44).The observation quantity includes the target's distance and bearing, and the observation equation is given as Equation (45): where x t = [x t , y t , ẋt , ẏt ] T is the state of the target at moment j, (x t , y t ) is the position state, ( ẋt , ẏt ) is the velocity state in x and y directions.T = 1 s is the sampling time.w t−1 is the zero-mean Gaussian process noise with the covariance Q = The initial position and velocity of the state are as x 0 = 100 m, 100 m, 10 ms −1 , 10 ms −1 T , with covariance P 0 = diag 20 m 2 , 20 m 2 , 2 m 2 s −2 , 2 m 2 s −2 T .The typical heavy-tailed noise is selected d as the measurement noise, and the Gaussian mixture distribution is used: where R = diag (5 m) 2 , (0.05 rad) 2 is the measurement noise covariance, and set ε = 10 −6 .
In the simulation experiment, the improved robust GA-VB [32], RKF-ML [33], and the proposed VBAIEKF were used to test the filtering effect.In the proposed VBAIEKF and the improved robust GA-VB and RKF-ML, the individual parameters were set as follows: parameter b = 5, parameter a = 1, parameter u = 4, parameter U = τ R, forgetting factor ρ = 1 − exp(−4), adjustment factor τ = 6, and the number of iterations N = 1, 2, 3, 5, 10.All algorithms were coded with MATLAB R2020b on a computer.
To compare the state estimation accuracy of the aforementioned nonlinear filtering algorithms, the root mean square error (RMSE) [39] and the average root mean square error (ARMSE) of the state vector is chosen to represent the filtering accuracy, and is defined as follows: where (m n t , r n t ) is the true position, ( mn t , rn t ) is the estimated position, and N denotes the number of times the algorithm is run.
In Figure 1, the impact of iteration numbers on the VBAIEKF algorithm is examined.The iteration count influences the proposed VBAIEKF in a distinct manner, with values set at N = 1, 2, 3, 5, 10.The observed trend indicates that as the number of iterations increases, the convergence speed gradually decreases, while the computational load correspondingly increases.Analysis of the ARMSE in position and velocity reveals a clear pattern; the ARMSE values demonstrate that accuracy improves with an increase in iterations, ultimately converging.When the iteration count reaches N = 10, the ARMSE values tend to stabilize.This iterative improvement is significant, despite the anticipated increase in computational demands.In Figure 2, a detailed examination is conducted to measure the impact of the adjustment factor τ on the performance of the proposed VBAIEKF algorithm.The adjustment factor τ systematically varies within the range of τ = 1, 2, 3, 5 . . ., 10.The results demonstrate that throughout the entire random process, the algorithm is uniformly influenced within a predetermined stable range.Notably, the operation of the algorithm is subject to a slight but not significant influence by the adjustment factor and minimization and precision enhancement.The VBAIEKF exhibits low sensitivity to parameters, displaying minimal influence from varying parameters.Therefore, τ = 6 was selected.This influence consistently remains within a specific stable range throughout the random process.At τ = 6, the algorithm achieves an optimal balance between error algorithm comparison experiments.
In Figure 3, we conducted a comprehensive analysis to evaluate the impact of the forgetting factor ρ on the performance of the proposed VBAIEKF algorithm.Systematic variations of the variable ρ within a certain range can identify various factors affecting the algorithm's performance.Affecting the ρ algorithm's performance, ρ = 1 − exp(−4), the impact of the forgetting factor significantly diminishes, thereby accelerating the convergence speed, reducing the RMSE, and making the algorithm output trajectory more perfect.The current configuration appears to have achieved the optimal balance between convergence speed, accuracy, and overall stability.Therefore, we chose ρ = 1 − exp(−4) for the subsequent comparative analysis.
Table 1 provides a comprehensive summary of the relative performance of the proposed VBAIEKF, GA-VB [32], and RKF-ML [33] at various levels of noise.Table 1 lists several key metrics: average root mean square error (ARMSE) for position and velocity, and average execution time.The results show that VBAIEKF has better convergence properties and adaptability.Compared to GA-VB and RKF-ML, the VBAIEKF shows a decrease in both the average position ARMSE and average velocity ARMSE, indicating improved accuracy in position and velocity estimation.Furthermore, this algorithm reduces the average execution time, highlighting its computational efficiency.Under various noise levels, using example noise R = diag (5 m) 2 , (0.05 rad) 2 as a reference, the convergence ranges under noises R = diag (1 m) 2 , (0.01 rad) 2 and R = diag (10 m) 2 , (0.1 rad) 2 are computed.From Table 1, it is evident that the VBAIEKF algorithm under different levels of noise exhibits a position convergence range of +0.058%∼+0.58%and a velocity convergence range of −0.17%∼+0.21%,significantly smaller than RKF-ML and GA-VB.This indicates that the algorithm possesses excellent convergence and adaptability.Comparative analysis reveals that the proposed VBAIEKF is a highly reliable and effective solution, surpassing both GA-VB and RKF-ML methods across different noise levels.The results obtained in this study provide robust evidence of the reliability and effectiveness of the proposed VBAIEKF in various scenarios.The results of these studies confirm that this algorithm can be used as a reliable and efficient tool.Figure 4 presents a thorough comparative analysis, wherein the proposed VBAIEKF algorithm is juxtaposed with existing optimization algorithms GA-VB and RKF-ML in terms of the noise R = diag (5 m) 2 , (0.05 rad) 2 , a specific number of iterations (N = 10).The comparative analysis is based on the root mean square absolute error (ARMSE) of velocity and position estimates.The results underscore the efficacy of the VBAIEKF algorithm; compared to the existing optimization algorithms RKF-ML and GA-VB, the VBAIEKF exhibited significant improvements in RMSE pos and RMSE vel , with increases of 21.81% and 22.11%, respectively.Table 1 shows that under noise R = diag (5 m) 2 , (0.05 rad) 2 , the convergence speed was enhanced by 49.04%, demonstrating accelerated convergence and significantly enhanced estimation precision.This highlights the algorithm's capacity to provide reliable and accurate estimates.

Nonlinear Numerical UNGM
This paper employs a nonlinear numerical UNGM [40] to demonstrate the effectiveness of VBRAIEKF.The equation representing the state space of the UNGM model is as follows: x k = 0.5x k−1 + 25 where Q k−1 and r k represent measurement noise with a mean of zero and covariance, and v k−1 represents non-Gaussian process noise R k .
The measurement noise is selected to be the typical heavy-tailed noise [25].
v i ∼ 0.9N (0, R) + 0.1N (0, 100 R) (52) The symbol R represents the measurement noise covariance matrix, which is defined as diag (1 m) 2 , (0.01 rad) 2 .By adjusting the parameters to optimize the algorithm's performance, setting the forgetting factor ρ = 1 − exp(−4) significantly reduces its impact, resulting in a significant decrease in influence.By selecting ε = 10 −6 , the algorithm's convergence speed is greatly enhanced.By setting the tuning factor τ = 6, the algorithm attains an optimal equilibrium in error algorithm comparative experiments.Performing 10 iterations decreases the average root mean square error (ARMSE), resulting in a more precise trajectory for the algorithm's output.During the simulation experiment, GA-VB, RKF-ML, and VBAIEKF are employed to evaluate the filtering efficacy.
Figure 5 examines the comparability of RMSE pos and RMSE vel for the GA-VB, RKF-ML, and the VBAIEKF algorithms.The results indicate that the estimation state of VBAIEKF demonstrates higher precision, with RMSE values lower than those of GA-VB and RKF-ML.It can more accurately estimate the expected state covariance and the actual state covariance.The improvement in estimation accuracy for both position and velocity signifies its reliability.

Conclusions
We have successfully developed an improved adaptive iterative extended Kalman filter based on variational Bayesian for nonlinear situations (VBAIEKF).This innovative

Figure 2 .
Figure 2. RMSE pos and RMSE vel for different adjustment factor τ.

Figure 4 .
Figure 4. RMSE pos and RMSE vel for different algorithms.

Figure 5 .
Figure 5. RMSE pos and RMSE vel for different algorithms.
Due to the fact that P t|t−1 and R t are the normal PDF covariance matrices, p P t|t−1 | Z1:t−1 and p R t | Z1:t−1 can be written as IW distributions: 1 represents the degree of freedom, θ represents a d × d positive definite matrix, d represents the dimension of θ, Γ d (•) represents the multivariate gamma distribution, and trace (•) represents the trace function.The expected value of the matrix φ ∼ IW(φ; β, θ), if it follows the IW distribution, is E(φ) = θ β−d−1 .

Table 1 .
ARMSE pos , ARMSE vel and the average execution time for the RKF-ML, GA-VB, and VBAIEKF under different noise.