Next Article in Journal
A Century of Azulene Chemistry; A Brief Look at Azulenes Building
Previous Article in Journal
Some New Fibonacci Matrices and Error Detecting-Correcting Codes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Generalized Loss-Based Nonlinear Filtering with Generalized Conversion

College of Electronic and Information Engineering, Southwest University, Chongqing 400715, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(3), 334; https://doi.org/10.3390/sym17030334
Submission received: 10 January 2025 / Revised: 19 February 2025 / Accepted: 21 February 2025 / Published: 22 February 2025
(This article belongs to the Section Computer)

Abstract

:
To address the nonlinear state estimation problem, the generalized conversion filter (GCF) is proposed using a general conversion of the measurement under minimum mean square error (MMSE) criterion. However, the performance of the GCF significantly deteriorates in the presence of complex non-Gaussian noise as the symmetry of the MMSE is compromised, leading to performance degradation. To address this issue, this paper proposes a new GCF, named generalized loss-based GCF (GLGCF) by utilizing the generalized loss (GL) as the loss function instead of the MMSE criterion. In contrast to other robust loss functions, the GL adjusts the shape of the function through the shape parameter, allowing it to adapt to various complex noise environments. Meanwhile, a linear regression model is developed to obtain residual vectors, and the negative log-likelihood of GL is introduced to avoid the problem of manually selecting the shape parameter. The proposed GLGCF not only retains the advantage of GCF in handling strong measurement nonlinearity, but also exhibits robust performance against non-Gaussian noises. Finally, simulations on the target-tracking problem validate the strong robustness and high filtering accuracy of the proposed nonlinear state estimation algorithm in the presence of non-Gaussian noise.

1. Introduction

State estimation is a key problem in engineering and scientific fields, involving the estimation of a system’s internal state from noisy measurements [1,2]. The Kalman filter (KF) is known as the optimal linear estimation method, providing an optimal recursive solution for linear systems under Gaussian noises [3]. However, in real-world applications, nonlinear dynamics and non-Gaussian noises are common. Under such conditions, the standard KF may fail to maintain its optimality in minimizing the mean squared error (MSE) of the estimated state. Thus, it is crucial to expand KF to efficiently deal with state estimation problem of nonlinear systems in the presence of non-Gaussian noise.
For addressing nonlinear filtering problems, obtaining an exact or analytical solution is often impractical. Many approaches have been explored to address this problem, including extended Kalman filter (EKF) [4], unscented Kalman filter (UKF) [5], cubature Kalman filter (CKF) [6], and Gauss–Hermite Kalman filter (GHKF) [7]. EKF utilizes the first-order Taylor series expansion to linearize system equations. However, EKF suffers from poor approximation accuracy due to linearization errors, and if the system exhibits high nonlinearity, EKF may even diverge [8]. To address this issue, UKF, CKF, and GHKF have been proposed, which approximate the state conditional probability distribution using deterministic sampling (DS). Although UKF avoids the linearization step, offering more accurate estimates for nonlinear systems, it demands more computational resources and careful adjustment of parameters. Moreover, in high-dimensional nonlinear systems, UKF may encounter negative weights, leading to filter instability and potential divergence. While the CKF addresses the numerical instability issue in the UKF, it introduces the problem of nonlocal sampling. GHKF may encounter the curse of dimensionality in high-dimensional systems problems, resulting in significant computational burdens. Based on the Monte Carlo method, particle filter (PF) uses random sampling of particles to represent probability density [9]. Since PF does not assume prior or posterior probability density, it requires considerable computational demands to attain accurate state estimation. Thus, depending on the system and specific requirements, it is important to balance accuracy and computational demands to select the suitable filter.
Generally, the standard KF and its variants are based on the framework of linear minimum mean square error (LMMSE) estimation [10]. LMMSE estimation focuses on finding the optimal linear estimator in the original measurement space. When the measurement and the state are related to a lower degree of nonlinearity, the LMMSE estimation is expected to yield more accurate results [10]. Furthermore, by considering a wider or optimized set of measurements that are uncorrelated with the original ones but still contain relevant information about the system state, LMMSE estimation can achieve better estimation accuracy. The uncorrelated conversion-based filter (UCF) improves the LMMSE estimator in a similar manner, by incorporating new measurements through nonlinear transformations [11]. Moreover, the optimized conversion-sample filter (OCF) simplifies the optimization of the nonlinear transformation function and limits the errors introduced by the DS [12]. In contrast to the UCF and OCF, the generalized conversion filter (GCF) optimizes both the conversion’s dimension and its sample points, providing a generalized transformation of measurements using DS [13].
While the GCF designed based on the minimum mean square error (MMSE) criterion is effectively to handle Gaussian noise, it is highly sensitive to non-Gaussian noise [14]. It is noteworthy that the measurement noise often follows a non-Gaussian distribution and may contain outliers in many practical scenarios. Recently, numerous advancements have been made in enhancing filter performance when dealing with non-Gaussian noise. One such approach involves augmenting the system model with quadratic or polynomial functions to improve estimation accuracy in the presence of non-Gaussian noise [15,16]. The Student’s t filter is another technique, using Student’s t distribution to model measurement noise [17]. Moreover, information-theoretic learning (ITL) has been utilized to combat non-Gaussian noises in signal processing [18]. For instance, maximum correntropy Kalman filter [19], maximum correntropy GCF (MCGCF) [20], and minimum error entropy Kalman filter [21] have been proposed. Unlike the previously discussed techniques, the generalized loss (GL) [22] provides flexibility in adjusting the shape of the loss function. It integrates various loss functions, such as the squared loss, Charbonnier loss [23], Cauchy loss [24], Geman–McClure loss [25] and Welsch loss [26]. Therefore, by acting as a robust loss function that does not rely on the symmetry of Gaussian distributions, GL improves both filtering accuracy and robustness, effectively capturing higher-order statistical characteristics and mitigating the impact of symmetry disruption in non-Gaussian environments. Until now, several GL-based algorithms have been proposed to tackle different estimation problems. A variational Bayesian-based generalized loss cubature Kalman filter is proposed to handle unknown measurement noise and the presence of potential outliers simultaneously [27]. The iterative unscented Kalman filter with general robust loss function [28] and geometric unscented Kalman filter with GL function [29] are utilized to enhance state estimation in power systems, alleviating the impact of non-Gaussian noise.
In this paper, a new nonlinear filter named generalized loss-based generalized conversion filter (GLGCF) is proposed, which employs the GL to reformulate the measurement update process of GCF. By leveraging the GCF’s ability to utilize higher-order information from transformed measurements and the GL’s robustness in dealing with various types of noise, the GLGCF outperforms both the GCF and MCGCF in non-Gaussian noise environments. The main contributions of this paper are summarized as follows:
  • To combat non-Gaussian noises, the GLGCF employs a robust nonlinear regression based on GL, and the posterior estimate and its covariance matrix are updated using a fixed-point iteration.
  • To solve the problem of manually setting the shape parameter in the GL function, the residual error is integrated into negative log-likelihood (NLL) of GL, and the optimal shape parameter is determined through minimizing the NLL.
  • Simulations on the target-tracking models in non-Gaussian noise environments demonstrate the superiority of the GLGCF. Additionally, its recursive structure makes it suitable for online implementation.
The rest of this paper is organized as follows. Section 2 introduces the GL and GCF. Section 3 derives the proposed GLGCF algorithm. Section 4 demonstrates the effectiveness of the GLGCF by target-tracking simulations. Section 5 concludes this paper.

2. Preliminaries

2.1. Generalized Loss

The following GL function is proposed by [22]
λ ( x , μ , c ) = | μ 2 | μ ( x / c ) 2 | μ 2 | + 1 μ / 2 1 ,
where μ represents the shape parameter, and c > 0 denotes the scale parameter. By adjusting the value of μ in Equation (1), λ can be extended to apply to a range of loss functions, e.g., squared loss ( μ = 2 ) , Charbonnier loss [23] ( μ = 1 ) , Cauchy loss [24] ( μ = 0 ) , Geman–McClure loss [25] ( μ = 2 ) , and Welsch loss [26] ( μ = ) .
Let X and Y be two arbitrary scalar random variables, and GL can be defined as follows:
V GL ( X , Y ) = | μ 2 | μ G ( e ) + 1 μ / 2 1 d F X Y ( x , y ) ,
where F X Y ( x , y ) represents the joint distribution of X and Y, and G ( e ) = ( e / c ) 2 / | μ 2 | with e = x y . Given the constraint of limited data, F X Y ( x , y ) is commonly unknown. Thus, Equation (2) is calculated by
V ^ GL ( X , Y ) = 1 M i = 1 M | μ 2 | μ G ( e ( i ) ) + 1 μ / 2 1 ,
where e ( i ) = x ( i ) y ( i ) and { x ( i ) , y ( i ) } i = 1 M are M samples sampled from F X Y ( x , y ) . Since Equation (3) incorporates G ( e ) , GL provides information regarding higher-order moments. This feature offers greater resilience to outliers and noise, making it particularly effective in situations where the data are subjected to non-Gaussian noise or outliers.

2.2. Generalized Conversion Filter

By applying a DS technique, GCF effectively transforms the sample points, resulting in accurate estimations for nonlinear systems influenced by Gaussian noise. Consider a discrete-time nonlinear system as follows:
x k = f x k 1 + w k 1 ,
z k = h x k + v k ,
where x k R n represents the system state; z k R m stands for the measurement vector; the functions f (·) and h (·) signify the process model and measurement model, respectively; w k R n and v k R m denote the process noise and measurement noise, respectively. Generally, for the nonlinear dynamic system described by Equations (4) and (5), the GCF includes three steps, namely prediction, constraint generation and update [13].

2.2.1. Prediction

The prior state and its covariance matrix can be calculated by generating sample points with DS methods, such as the unscented transformation [5], Gauss–Hermite quadrature (GHQ) [7], and cubature rules [6]. First, we define a combined state vector ψ k 1 as follows:
ψ k 1 x k 1 T , w k 1 T T ,
and the estimate ψ ^ k 1 of ψ k 1 , MSE of ψ ^ k 1 given the measurements Z k 1 , denoted by P k 1 ψ , is expressed as
ψ ^ k 1 = x ^ k | k 1 w ¯ k 1 , P k 1 ψ = P k 1 0 0 Q k 1 w ,
where Q k 1 w is the covariance matrix of w k 1 , and E w k 1 = w ¯ k 1 . Using the previous estimate ψ ^ k 1 and covariance matrix P k 1 ψ , we generate a weight vector W ψ = w 1 ψ , w 2 ψ , , w n s ψ T and a sample set ψ k 1 l l = 1 n ψ of ψ k 1 by a DS method with n ψ being the number of sampling points. The sample ψ k 1 l consists of the state sample x k 1 l T and the white process noise sample w k 1 l T .
The transformed samples set can be obtained using the process model as
x k | k 1 l * = f x k 1 l , w k 1 l , l = 0 , , n ψ .
Thus, the predicted state and its covariance matrix are computed by
x ^ k k 1 = l w l ψ x k | k 1 l * ,
P k k 1 = l w l ψ x k | k 1 l * x ^ k k 1 x k | k 1 l * x ^ k k 1 T .

2.2.2. Constraint Generation

With respect to the measurement function Equation (5), the combined state vector ξ k is defined as
ξ k x k T , v k T T ,
and the estimate ξ ^ k 1 of ξ k 1 , MSE of ξ ^ k 1 given the measurements Z k 1 , represented by P k 1 ξ , can be written as
ξ ^ k = x ^ k k 1 v ¯ k ,
P k ξ = P k k 1 0 0 R k v ,
where x ^ k k 1 is the predicted state, E v k = v ¯ k , and R k v is the covariance matrix of v k .
Similar to the previous prediction step, based on ξ ^ k and P k ξ , we can generate a weight vector W = w 1 , , w n ξ T and a sample set ξ k s s = 1 n ξ of ξ k by a DS method with n ξ being the number of sampling points. The sample ξ k s consists of the state sample x k s T and the white measurement noise sample v k s T .
The transformed samples set can be derived using the measurement model
z k s = h x k s , v k s , s = 0 , , n ξ .
Thus, the samples of x k and z k can be obtained as follows:
X k x k 1 , , x k n ξ , Z k z k 1 , , z k n ξ .
Following the similar approach to that in Table I in [13], where Z i a is substituted by Z i , k a = e i , n z Z k , z k , the constraint matrices C i , k ( n ) can be derived for all i = 1 , , n z . By applying the constraint matrices C i , k ( n ) , the higher-order terms of the transformed samples can be neglected, which effectively restricts the higher-order errors and achieves the optimal transformation.

2.2.3. Update

To obtain the optimal sample points, we apply the QR decomposition to the constraint matrices C i , k ( n ) for all i = 1 , , n z .
Q 0 , i k , Q i , k R 0 , i k 0 = q r C i , k T ,
Q k Q 1 , k , , Q n z , k T ,
where C i , k is C i , k ( n ) without its last column.
Finally, the estimated state x ^ k and its covariance matrix P k can be obtained by
x ^ k = x ^ k k 1 K k W ,
P k = P k k 1 K k M x , k T ,
K k M x , k Q k T Q k M y Q k T 1 Q k ,
where
M x , k = X k x ^ k k 1 1 1 × n ξ d i a g ( W ) ,
M y = s w s e s , n ξ W e s , n ξ W T .

3. Generalized Loss-Based Generalized Conversion Filter

Since the MMSE depends only on the second-order statistics of the errors, the performance of the GCF deteriorates in non-Gaussian noise [30]. To improve the robustness of the GCF, we propose integrating the GL cost function into the GCF framework, resulting in a new variant of the GCF, namely GLGCF. This variant is expected to perform better in non-Gaussian noise environments, as the GL incorporates second- and higher-order moments of the errors.
To combine the GL with the GCF, we first define a linear model that combines the state estimate and the measurement as follows:
x ^ k k 1 z k = x k h x ^ k + ϵ k ,
where ϵ k = x k x ^ k k 1 v k and the covariance matrix of ϵ k can be obtained by
E ϵ k ϵ k T = P k k 1 0 0 R k v = U p , k k 1 U p , k k 1 T 0 0 U r , k U r , k T ,
with U k = U p , k k 1 0 0 U r , k . Multiplying both sides of Equation (23) by U k 1 yields
Γ k = d x k + ϕ k ,
where Γ k = U k 1 x ^ k z k , d x k = U k 1 x k h x ^ k , and ϕ k = U k 1 ϵ k .
The GL-based cost function is defined as
J ^ GL x k = 1 L i = 1 L | μ 2 | μ G ϕ i , k + 1 μ / 2 1 ,
where ϕ i , k = Γ i , k d i x k is the ith element of ϕ k , and Γ i , k is the ith element of Γ k , d i x k is the ith row of d x k .
Next, the optimal estimate of x k can be obtained by
x ^ k = arg min x k J ^ GL x k = arg min x k 1 L i = 1 L | μ 2 | μ G ϕ i , k + 1 μ / 2 1 .
The solution to Equation (27) can be obtained by solving the following equation:
J ^ GL x k x k = i = 1 L G ϕ i , k + 1 ( μ / 2 ) 1 ϕ i , k ϕ i , k x i , k = 0 .
Equation (28) can be further expressed in matrix form as
d x k x k T Θ k Γ k d x k = 0 ,
where
Θ k = Θ k x 0 0 Θ k y
with
Θ k x = diag κ ϕ 1 , k , , κ ϕ n , k ,
Θ k y = diag κ ϕ n + 1 , k , , κ ϕ n + m , k ,
and
κ ϕ i , k = G ϕ i , k + 1 ( μ / 2 ) 1 .
According to [19], updated Equation (25) can be applied with one iteration to yield similar results within the GCF framework by employing Θ k to modify the measurement data. As noted in [31], two methods can be used to achieve this: the first method modifies the residual error covariance using ϕ i , k = Γ i , k d i ( x k ) , and the second method reconstructs the ‘pseudo-observation’. Although both methods are equivalent in their final outcome. For simplicity, this paper presents the algorithm based on the first approach.
Let Λ k denote the modified covariance, which can be expressed as
Λ k = U k Θ k 1 U k T .
In the following analysis, we express Λ k in two parts such that
Λ k = Λ k x 0 0 Λ k y .
Given that the actual state x ( k ) is unknown, we set x ^ k | k 1 x k = 0 . Under this condition, it is straightforward to observe that
P k | k 1 = U p ( k | k 1 ) · I · U p T ( k | k 1 ) = Λ k x .
The modified measurement noise covariance matrix can be obtained as
R ¯ k v = Λ k y .
The GL-based cost function characterizes the error properties by weighting on the measurement uncertainty, which is reflected in the modified measurement noise covariance matrix R ¯ k v . This procedure allows for a more accurate representation of the error dynamics by incorporating the uncertainty in the measurements, thus refining the covariance matrix to accurately capture the true noise characteristics.
Next, we replace R k v in Equation (13) with R ¯ k v to obtain
P ¯ k ξ = P k k 1 0 v v 0 R ¯ k v .
Based on ξ ^ k and P ¯ k ξ , we can generate a weight vector W ¯ = w 1 , , w n ξ T and a sample set ξ k s s = 1 n ξ of ξ k by a DS method. Thus, the samples of x k and z k can be obtained as Equation (15).
By employing the approach from Table I in [13], where Z i a is substituted by Z i , k a = e i , n z Z k , z k , we obtain the constraint matrix C i , k ( n ) for i = 1 , , n z . Subsequently, Q k is calculated by C i , k ( n ) for i = 1 , , n z , as shown in Equations (16) and (17).
Thus, the filter estimated state x ^ k and its covariance matrix P k can be computed by
x ^ k = x ^ k k 1 K k W ¯ ,
P ¯ k = P k k 1 K ¯ k M ¯ x , k T ,
K ¯ k M ¯ x , k Q k T Q k , M ¯ y Q k T 1 Q k ,
where
M ¯ x , k = X k x ^ k k 1 1 1 × n ξ d i a g ( W ¯ ) ,
M ¯ y = s w s e s , n ξ W ¯ e s , n ξ W ¯ T .
   Algorithm 1: GLGCF
  • Step 1: Set the initial state estimate x ^ 0 | 0 and the error covariance matrix P 0 | 0 ,
  • choose a proper truncation limit value t and a minimum shape parameter μ m i n < 2.
  • Step 2: Compute the predicted sampling points x k | k 1 l * ; use x ^ k k 1 = l w l ψ x k | k 1 l * and
  • P k k 1 = l w l ψ x k | k 1 l * x ^ k k 1 x k | k 1 l * x ^ k k 1 T to calculate x ^ k k 1 and P k k 1 ; adopt
  • Cholesky decomposition to obtain U k .
  • Step 3: Utilize z k s = h x k l , v k l , l = 0 , , n ξ , to compute the propagated sampling points
  • z k s ; calculate the residual error ϕ k = Γ k d x k . Employ Equation (45) and residual error
  • samples ϕ i , k ( j ) i = 1 L to obtain μ * ; use Equations (34) and (37) to derive R ¯ k v .
  • Step 4: Obtain the samples of x k and z k as Equation (15). Compute the constraint matrix
  • C i , k ( n ) using Table I in [13] with Z i a substituted by Z i , k a = e i , n z Z k , z k .
  • Step 5: Compute Q k using Equations (16) and (17). Update the state estimate x ^ k and the
  • error covariance matrix P k by the following equations:
    x ^ k = x ^ k k 1 K k W ¯ ,
  • P ¯ k = P k k 1 K ¯ k M ¯ x , k T ,
  • K ¯ k M ¯ x , k Q k T Q k , M ¯ y Q k T 1 Q k ,
  • where M x , k = X k x ^ k k 1 1 1 × n ξ d i a g ( W ) and M y = s w s e s , n ξ W e s , n ξ W T .
The shape of the GL function is determined by the parameter μ , with its value influencing the level of outlier suppression. Given that μ directly affects filtering performance, finding the optimal μ is essential. To address this issue, we formulate the negative log-likelihood (NLL) of GL’s probability distribution function as follows:
Ξ NLL ( x , c , σ ) = log P ( x , μ , c )   = Ξ ( x , μ , c ) + log c T ( μ ) ,
where P ( x , μ , c ) = 1 c T ( μ ) exp ( Ξ ( x , μ , c ) ) ; T ( μ ) = t t exp ( Ξ ( x , μ , 1 ) ) d x is an approximate integral, with t representing the truncation limit [32]. Subsequently, we find the optimal μ by minimizing the NLL of the residual error ϕ k as follows:
μ * = arg min μ i = 1 L | μ 2 | μ G ϕ i , k + 1 μ / 2 1 + log c t t exp | μ 2 | μ G ( ϕ i , k ) + 1 μ / 2 1 d x .
Since obtaining an analytical solution for the partition function in Equation (45) is not feasible and μ is a scalar, a 1-D grid search within μ μ min , 2 can be used to minimize Equation (45). The choice of μ min < 2 ensures the stability of the loss function and reduces computational complexity. When the system is affected by measurement outliers, the symmetric loss function MMSE becomes sensitive to symmetry disruption, resulting in biased state estimation. By optimizing the shape parameter, the GL adapts more effectively to the characteristic of non-Gaussian noise, thereby enhancing the robustness of the GCF in complex noise environments.
Finally, the proposed GLGCF algorithm is summarized in Algorithm 1.

Computational Complexity

The computational complexity of the proposed GLGCF is presented. Note that n and m are dimensions of x k and z k , respectively. N i n t represents the number of integration subintervals. n ξ represents the number of sampling points, which is determined by the selected DS method. In this paper, we utilize the GHQ rule as the DS method. C h denotes the computational complexity of h ( · ) for calculation with the sampling points.
As we can see from Table 1, different DS methods exhibit distinct computational complexities. In scenarios with constrained computational resources, opting for a DS method that requires fewer sampling points can effectively balance the trade-off between computational efficiency and accuracy. Furthermore, a small truncation limit values t, and a small search interval [ μ min , 2 ] can achieve high accuracy with a finite complexity.

4. Simulation Results

In order to analyze the performance of the proposed algorithm, the results of the simulation are provided in this section. The proposed GLGCF algorithm is compared with several existing algorithms: CKF [33], UKF [8], PF [34], GCF [13], and MCGCF [20]. To evaluate the filtering accuracy, the root mean square error (RMSE) and average RMSE (ARMSE) are defined as follows:
RMSE ( k ) = 1 M j = 1 M [ ( x i , k j x ^ i , k j ) 2 ] , k = 1 , 2 , , n ,
ARMSE = 1 n k = 1 n 1 M j = 1 M [ ( x i , k j x ^ i , k j ) 2 ] ,
where x k j denotes the real state, and x ^ k j represents the estimate of x k j for jth Monte Carlo run at discrete time k, n is the number of iterations and M denotes the number of Monte Carlo runs.

4.1. Constant Velocity Tracking Model

In this simulation, a constant velocity (CV) surface-target-tracking model is used to evaluate the filtering performance of the proposed robust GLGCF algorithm in the presence of high maneuvering speed and non-Gaussian noise. The state and measurement equations for the CV model are defined as follows:
x k = 1 T 0 0 0 1 0 0 0 0 1 T 0 0 0 1 x k 1 + w k 1 ,
z k = α k 2 + β k 2 arctan β k α k + v k ,
where x k α k , α ˙ k , β k , β ˙ k T with position ( α k , α ˙ k ) and velocity ( β k , β ˙ k ) . T represents the sampling period. w k denotes the process noise with its covariance matrix Q k w , and v k stands for the measurement noise with its covariance matrix R k v .
We consider the two scenarios with different measurement noise model as follows.

4.1.1. Heavy-Tailed Noise

Scenario 1: Based on the non-Gaussian noise simulation method [21], assume that the presence of non-Gaussian measurement noise with heavy tails is modeled by a mixed Gaussian distribution as described below:
v k 0.9 N ( 0 , diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) + 0.1 N ( 0 , 400 diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) ,
The initial state is generated from x 0 N x ^ 0 , P 0 , where N ( x ^ 0 , P 0 ) is a Gaussian distribution with mean x ^ 0 = 100 m , 30 m / s , 100 m , 20 m / s T and with covariance P 0 = diag 10 m 2 , 1 m 2 / s 2 , 10 m 2 , 1 m 2 / s 2 T . The sampling period T = 1 s. The process noise covariance is Q k w = diag [ 0.01 M , 0.01 M ] with M = T 3 / 4 , T 3 / 2 ; T 2 / 2 , T . For GLGCF, the parameters t and c are set as 10 and 1, respectively. The minimum shape parameter is set as μ m i n = 15 . The results of state estimate are obtained after 50 Monte Carlo runs.
Figure 1 shows the RMSE of GLGCF against other nonlinear filters. As can be seen from Figure 1, GLGCF and MCGCF are superior to GCF. This finding shows that, under a heavy-tailed noise environment, replacing the MMSE criterion with the maximum correntropy criterion (MCC) or GL in the GCF improves both the accuracy and robustness of the filter. Furthermore, the comparison shows that the GLGCF outperforms the MCGCF, indicating that GL is more effective than MCC in handling heavy-tailed noise.

4.1.2. Mixed Noise

Scenario 2: In this scenario, the measurement noise is assumed to follow an additive combination of two Gaussian noise distributions.
v k 0.5 N ( 0 , 0.1 d i a g [ ( 20 m ) 2 , ( 2 π 180 rad ) 2 ] ) + 0.5 N ( 0 , 0.1 d i a g [ ( 30 m ) 2 , ( 4 π 180 rad ) 2 ] ) ,
The initial state is generated from x 0 N x ^ 0 , P 0 , where N ( x ^ 0 , P 0 ) is a Gaussian distribution with mean x ^ 0 = 100 m , 30 m / s , 100 m , 20 m / s T and with covariance P 0 = d i a g 10 m 2 , 1 m 2 / s 2 , 10 m 2 , 1 m 2 / s 2 T . The sampling period T = 1 s. The process noise covariance is Q k w = d i a g [ 0.01 M , 0.01 M ] with M = T 3 / 4 , T 3 / 2 ; T 2 / 2 , T . For GLGCF, the parameters t and c are set as 10 and 1, respectively. The minimum shape parameter is set as μ m i n = 15 . The results of state estimate are obtained after 50 Monte Carlo runs.
The comparative results of different filters in the presence of mixed measurement noise are shown in Figure 2, Table 2 and Table 3. In comparison to the UKF, CKF, GCF, and PF, both the MCGCF and the GLGCF exhibit better performance. This is due to the use of the generalized transformation, which limits the higher-order errors associated with the DS technique, resulting in improved accuracy. Futhermore, as can be seen in Figure 2, Table 2 and Table 3, the proposed GLGCF algorithm achieves the highest estimation accuracy, demonstrating its superior performance in predicting both position and velocity estimation in the mixed-noise environment.

4.2. Cooperative Turning Tracking Model

In this simulation, the cooperative turning (CT) target-tracking model is used to evaluate the filtering performance of the proposed GLGCF algorithm under heavy-tailed and mixed-noise environments. The CT model is well known as a fundamental maneuvering model in surface target tracking and is especially prevalent in describing the dynamics of maneuvering targets in unmanned surface vehicle tracking. The state and measurement equations for the CT model are presented below:
x k = 1 sin Ω T Ω 0 1 cos Ω T Ω 0 cos Ω T 0 sin Ω T 0 1 cos Ω T Ω 1 sin Ω T Ω 0 sin Ω T 0 cos Ω T x k 1 + w k ,
z k = α k 2 + β k 2 arctan β k α k + v k ,
where x k α k , α ˙ k , β k , β ˙ k T with position ( α k , α ˙ k ) and velocity ( β k , β ˙ k ) . T represents the sampling period, and Ω represents the tuning rate. w k denotes the process noise with its covariance matrix Q k w , and v k represents the measurement noise with covariance matrix R k v .
We consider the four scenarios with different measurement noise models below.

4.2.1. Heavy-Tailed Noise

Scenario 1: The robustness of the GLGCF algorithm against non-Gaussian noise is demonstrated by assuming the measurement noise in Equation (53) to be heavy-tailed, generated by a mixture of Gaussian distributions with different variances:
v k 0.9 N ( 0 , diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) + 0.1 N ( 0 , 400 diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) ,
Scenario 2: The probability of large outliers in the measurement noise is increased as follows:
v k 0.8 N ( 0 , diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) + 0.2 N ( 0 , 400 diag [ ( 10 m ) 2 , ( 10 mrad ) 2 ] ) .
The initial state is generated from x 0 N x ^ 0 , P 0 , where N ( x ^ 0 , P 0 ) is a Gaussian distribution with mean x ^ 0 = 100 m , 80 m / s , 100 m , 120 m / s T and with covariance P 0 = diag 10 m 2 , 1 m 2 / s 2 , 10 m 2 , 1 m 2 / s 2 T . The sampling period T = 1 s and tuning rate Ω = π 19 rad. The process noise covariance is Q k w = diag [ 0.04 M , 0.04 M ] with M = T 3 / 3 , T 3 / 2 ; T 2 / 2 , T . For GLGCF, the parameters t and c are set as 10 and 1, respectively. The minimum shape parameter is set as μ m i n = 15 . The results of the state estimate are obtained after 50 Monte Carlo runs.
The RMSEs for different algorithms under heavy-tailed noise in scenario 1 and scenario 2 are depicted in Figure 3 and Figure 4, respectively. The ARMSEs of position and velocity estimates in scenarios 1 and 2 are summarized in Table 4 and Table 5, respectively. As shown in Figure 3 and Figure 4, the filtering accuracy of GLGCF is superior to GCF, MCGCF, CKF, and UKF. This confirms that incorporating the GL into the GCF improves the estimation accuracy and robustness of the filter when dealing with heavy-tailed noise. In the presence of measurement noise that includes outliers, the GCF suffers from performance degradation. However, the use of MCC and GL function stabilizes the GCF by performing nonlinear iterations on the anomalous measurements. Both MCGCF and GLGCF maintain stable performance, with GLGCF achieving superior performance. The results indicate that, in the non-Gaussian noise environments, the optimal form of the loss function can be automatically determined by the error distribution, thanks to the use of the residual optimization parameter.

4.2.2. Mixed Noise

Scenario 3: The measurement noise is characterized by a mixture of Gaussian distributions, with different variances for the position and velocity:
v k 0.5 N ( 0 , 0.05 diag [ ( 25 m ) 2 , ( 3 π 180 rad ) 2 ] ) + 0.5 N ( 0 , 0.05 diag [ ( 30 m ) 2 , ( 9 π 180 rad ) 2 ] ) ,
Scenario 4: In this scenario, the measurement noise is also a mixture of two Gaussian distributions, with the following mixture weights:
v k 0.7 N ( 0 , 0.05 diag [ ( 25 m ) 2 , ( 3 π 180 rad ) 2 ] ) + 0.3 N ( 0 , 0.05 diag [ ( 30 m ) 2 , ( 9 π 180 rad ) 2 ] ) .
The initial state is generated from x 0 N x ^ 0 , P 0 , where N ( x ^ 0 , P 0 ) is a Gaussian distribution with mean x ^ 0 = 100 m , 80 m / s , 100 m , 120 m / s T and with covariance P 0 = text 10 m 2 , 1 m 2 / s 2 , 10 m 2 , 1 m 2 / s 2 T . The sampling period T = 1 s and tuning rate Ω = π 19 rad. The process noise covariance is Q k w = diag [ 0.04 M , 0.04 M ] with M = T 3 / 3 , T 3 / 2 ; T 2 / 2 , T . For GLGCF, the parameters t and c are set as 10 and 1, respectively. The minimum shape parameter is set as μ m i n = 15 . The results of state estimate are obtained after 50 Monte Carlo runs.
Figure 5 and Figure 6 show the RMSE curves of UKF, CKF, GCF, MCGCF and GLGCF in the mixed measurement noise. Table 4 and Table 5 summarize the ARMSEs of position and velocity estimates in scenario 3 and scenario 4. It can be seen in Figure 5 and Figure 6, Table 4 and Table 5, the RMSE and ARMSE achieved by GLGCF algorithm are smaller than those of other compared algorithms. In the mixed-noise environment, the estimation accuracy of MCGCF and GLGCF becomes closer, with GLGCF still providing the best filtering accuracy. Compared to the heavy-tailed environment, the advantage of GLGCF is less distinct in the mixed-noise environment. This can be attributed to the absence of large outliers in the measurement noise, where both GCF and MCGCF offer good filtering results. Therefore, GLGCF exhibits its robustness against the impact of outliers, providing superior performance in more complex environments.

5. Conclusions

In this paper, an extended version of the generalized conversion filter (GCF), namely generalized loss-based generalized conversion filter (GLGCF), is proposed. The GCF performs well in the Gaussian noise environment. However, the performance of GCF is significantly deteriorated when the system encounters non-Gaussian noise, primarily due to the the characteristics of the minimum mean square error (MMSE) criterion. To tackle this challenge, the proposed GLGCF algorithm is reformulated using a nonlinear regression model, employing the generalized loss (GL) instead of MMSE criterion to estimate the system state. Moreover, the residual vectors are incorporated into the negative log-likelihood of GL to optimize the parameters, thereby reducing the burden for manual tuning. Finally, two simulations on target-tracking models are carried out to verify the effectiveness and robustness of the proposed GLGCF algorithm. The results indicate that this proposed GLGCF algorithm exhibits excellent performance when the system is affected by non-Gaussian noise.

Author Contributions

Conceptualization, Z.K. and S.W.; methodology, S.W.; software, Z.K.; validation, S.W. and Y.Z.; formal analysis, S.W.; investigation, Y.Z.; resources, S.W.; data curation, Z.K. and Y.Z.; writing—original draft preparation, Z.K.; writing—review and editing, S.W. and Y.Z.; visualization, Z.K.; supervision, S.W. and Y.Z.; project administration, S.W.; funding acquisition, S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant: 62471406).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
KFKalman filter
GCFgeneralized conversion filter
GLgeneralized loss
MCCmaximum correntropy criterion
ITLinformation-theoretic learning
MCGCFmaximum correntropy generalized conversion filter
GLGCFgeneralized loss generalized conversion filter
DSdeterministic sampling
NLLnegative log-likelihood
UCFuncorrelated conversion-based filter
OCFoptimized conversion-sample filter
MMSEminimum mean square error
LMMSElinear minimum mean square error
EKFextended Kalman filter
UKFunscented Kalman filter
PFparticle filter
GHKFGauss–Hermite Kalman filter
GHQGauss–Hermite quadrature
CVconstant velocity
CTcooperative turning

References

  1. Xia, X.; Xiong, L.; Huang, Y.; Liu, Y.; Gao, L.; Xu, N.; Yu, Z. Estimation on IMU yaw misalignment by fusing information of automotive onboard sensors. Mech. Syst. Signal Process. 2022, 162, 107993. [Google Scholar] [CrossRef]
  2. Duník, J.; Biswas, S.K.; Dempster, A.G.; Pany, T.; Closas, P. State estimation methods in navigation: Overview and application. IEEE Aerosp. Electron. Syst. Mag. 2020, 35, 16–31. [Google Scholar] [CrossRef]
  3. Kalman, R.E. A new approach to linear filtering and prediction problems. Trans. ASME, Ser. D J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  4. Ding, L.; Wen, C. High-Order extended Kalman filter for state estimation of nonlinear systems. Symmetry 2024, 16, 617. [Google Scholar] [CrossRef]
  5. Garcia, R.V.; Pardal, P.C.P.M.; Kuga, H.K.; Zanardi, M.C. Nonlinear filtering for sequential spacecraft attitude estimation with real data: Cubature Kalman Filter, Unscented Kalman Filter and Extended Kalman Filter. Adv. Space Res. 2019, 63, 1038–1050. [Google Scholar] [CrossRef]
  6. Wan, J.; Ren, P.; Guo, Q. Application of interactive multiple model adaptive five-degree cubature Kalman algorithm based on fuzzy logic in target tracking. Symmetry 2019, 11, 767. [Google Scholar] [CrossRef]
  7. Klokov, A.; Kanouj, M.; Mironchev, A. A novel carrier tracking approach for GPS signals based on gauss–hermite Kalman filter. Electronics 2022, 11, 2215. [Google Scholar] [CrossRef]
  8. Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 401–422. [Google Scholar] [CrossRef]
  9. Jouin, M.; Gouriveau, R.; Hissel, D.; Péra, M.C.; Zerhouni, N. Particle filter-based prognostics: Review, discussion and perspectives. Mech. Syst. Signal Process. 2016, 72–73, 2–31. [Google Scholar] [CrossRef]
  10. Liu, Y.; Li, X.R.; Chen, H. Linear estimation with transformed measurement for nonlinear estimation. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 221–236. [Google Scholar] [CrossRef]
  11. Lan, J.; Li, X.R. Nonlinear estimation by LMMSE-based estimation with optimized uncorrelated augmentation. IEEE Trans. Signal Process. 2015, 63, 4270–4283. [Google Scholar] [CrossRef]
  12. Lan, J.; Li, X.R. Nonlinear estimation based on conversion-sample optimization. Automatica 2020, 121, 109160. [Google Scholar] [CrossRef]
  13. Lan, J. Generalized-conversion-based nonlinear filtering using deterministic sampling for target tracking. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 7295–7307. [Google Scholar] [CrossRef]
  14. Schick, I.C.; Mitter, S.K. Robust recursive estimation in the presence of heavy-tailed observation noise. Ann. Statist. 1994, 22, 1045–1080. [Google Scholar] [CrossRef]
  15. Wang, S.; Wang, Z.; Dong, H.; Chen, Y.; Alsaadi, F.E. Recursive quadratic filtering for linear discrete non-Gaussian systems over time-correlated fading channels. IEEE Trans. Signal Process. 2022, 70, 3343–3356. [Google Scholar] [CrossRef]
  16. Battilotti, S.; Cacace, F.; d’Angelo, M.; Germani, A. The polynomial approach to the LQ non-Gaussian regulator problem through output injection. IEEE Trans. Autom. Control 2019, 64, 538–552. [Google Scholar] [CrossRef]
  17. Huang, Y.; Zhang, Y.; Xu, B.; Wu, Z.; Chambers, J. A new outlier-robust student’s t based gaussian approximate filter for cooperative localization. IEEE/ASME Trans. Mechatronics 2017, 22, 2380–2386. [Google Scholar] [CrossRef]
  18. Chen, B.; Dang, L.; Zheng, N.; Príncipe, J.C. Kalman Filtering Under Information Theoretic Criteria; Springer: Cham, Switzerland, 2023. [Google Scholar]
  19. Chen, B.; Liu, X.; Zhao, H.; Príncipe, J.C. Maximum correntropy Kalman filter. Automatica 2017, 76, 70–77. [Google Scholar] [CrossRef]
  20. Dang, L.; Jin, S.; Ma, W.; Chen, B. Maximum correntropy generalized conversion-based nonlinear filtering. IEEE Sens. J. 2024, 24, 37300–37310. [Google Scholar] [CrossRef]
  21. Chen, B.; Dang, L.; Gu, Y.; Zheng, N.; Príncipe, J.C. Minimum errorentropy Kalman filter. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 5819–5829. [Google Scholar] [CrossRef]
  22. Barron, J.T. A general and adaptive robust loss function. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4331–4339. [Google Scholar]
  23. Charbonnier, P.; Blanc-Feraud, L.; Aubert, G.; Barlaud, M. Two deterministic half-quadratic regularization algorithms for computed imaging. In Proceedings of the 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 168–172. [Google Scholar]
  24. Qi, L.; Shen, M.; Wang, D.; Wang, S. Robust cauchy kernel conjugate gradient algorithm for non-Gaussian noises. IEEE Signal Process. Lett. 2021, 28, 1011–1015. [Google Scholar] [CrossRef]
  25. Lu, L.; Wang, W.; Yang, X.; Wu, W.; Zhu, G. Recursive GemanMcClure estimator for implementing second-order Volterra filter. IEEE Trans. Circuits Syst. II Exp. Briefs. 2019, 66, 1272–1276. [Google Scholar]
  26. Dennis, J.E.; Welsch, R.E. Techniques for nonlinear least squares and robust regression. Commun. Statist.-Simul. Comput. 1978, 7, 345–359. [Google Scholar] [CrossRef]
  27. Yan, W.; Chen, S.; Lin, D.; Wang, S. Variational Bayesian-based generalized loss cubature Kalman Filter. IEEE Trans. Circuits Syst. II Express Briefs. 2024, 71, 2874–2878. [Google Scholar] [CrossRef]
  28. Zhao, H.; Hu, J. Iterative unscented Kalman filter with general robust loss function for power system forecasting-aided state estimation. IEEE Trans. Instrum. Meas. 2023, 73, 9503809. [Google Scholar] [CrossRef]
  29. Chen, S.; Zhang, Q.; Lin, D.; Wang, S. Generalized loss based geometric unscented Kalman filter for robust power system forecasting-aided state estimation. IEEE Signal Process Lett. 2022, 29, 2353–2357. [Google Scholar] [CrossRef]
  30. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties, and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  31. Zhu, B.; Chang, L.; Xu, J.; Zha, F.; Li, J. Huber-based adaptive unscented Kalman filter with non-Gaussian measurement noise. Circuits, Syst. Signal Process. 2018, 37, 3842–3861. [Google Scholar] [CrossRef]
  32. Chebrolu, N.; Labe, T.; Vysotska, O.; Behley, J.; Stachniss, C. Adaptive robust kernels for non-linear least squares problems. IEEE Robot. Autom. Lett. 2021, 6, 2240–2247. [Google Scholar] [CrossRef]
  33. Arasaratnam, I.; Haykin, S. Cubature Kalman filters. IEEE Trans. Autom. Control. 2009, 54, 1254–1269. [Google Scholar] [CrossRef]
  34. Arulampalam, M.S.; Maskell, S.; Gordon, N.; Clapp, T. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process. 2002, 50, 174–188. [Google Scholar] [CrossRef]
Figure 1. RMSEs of different filters in estimating the CV model under the scenario 1. (a) Position (m); (b) velocity (m/s).
Figure 1. RMSEs of different filters in estimating the CV model under the scenario 1. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g001
Figure 2. RMSEs of different filters in estimating the CV model under the scenario 2. (a) Position (m); (b) velocity (m/s).
Figure 2. RMSEs of different filters in estimating the CV model under the scenario 2. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g002
Figure 3. RMSEs of different filters in estimating the CT model under the scenario 1. (a) Position (m); (b) velocity (m/s).
Figure 3. RMSEs of different filters in estimating the CT model under the scenario 1. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g003
Figure 4. RMSEs of different filters in estimating the CT model under the scenario 2. (a) Position (m); (b) velocity (m/s).
Figure 4. RMSEs of different filters in estimating the CT model under the scenario 2. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g004
Figure 5. RMSEs of different filters in estimating the CT model under the scenario 3. (a) Position (m); (b) velocity (m/s).
Figure 5. RMSEs of different filters in estimating the CT model under the scenario 3. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g005
Figure 6. RMSEs of different filters in estimating the CT model under the scenario 4. (a) Position (m); (b) velocity (m/s).
Figure 6. RMSEs of different filters in estimating the CT model under the scenario 4. (a) Position (m); (b) velocity (m/s).
Symmetry 17 00334 g006
Table 1. Computational complexity of GLGCF.
Table 1. Computational complexity of GLGCF.
EquationsAddition/Subtraction and MultiplicationDivision, Matrix Inversion, Matrix Decomposition and Nonlinear Function
(9)–(14) 2 n 3 + 2 n 2 n O ( n C h )
(45) ( 2 μ min + 1 ) ( n + m + N i n t ) 0
(34) ( n + m ) 3 + 2 ( n + m ) 2 O ( ( n + m ) 3 )
(16)–(17)0 O ( n n ξ ( n ξ 1 ) 2 )
(39) n ( n ξ + 1 ) 0
(40) n 2 ( n ξ + 1 ) O ( n ξ + n )
(41) n 2 n ξ + 2 n n ξ 2 + n ξ 3 O ( n n ξ ) + O ( n ξ 2 ) + O ( n 3 )
(42) 2 ( n + m ) n ξ O ( n ξ )
(43) n ξ 3 + 2 n ξ 2 + n ξ O ( n ξ )
Table 2. ARMSE of position (m) in the CV model under the different scenarios.
Table 2. ARMSE of position (m) in the CV model under the different scenarios.
ScenariosUKFCKFPFGCFMCGCFGLGCF
Scenario 1331.0570321.7577200.826519.922314.024713.6668
Scenario 2139.9913132.555089.135614.916213.125112.6863
Table 3. ARMSE of velocity (m/s) in the CV model under the different scenarios.
Table 3. ARMSE of velocity (m/s) in the CV model under the different scenarios.
ScenariosUKFCKFPFGCFMCGCFGLGCF
Scenario 16.56106.42215.69474.26212.27812.0044
Scenario 22.76872.64092.25592.06201.35401.2733
Table 4. ARMSE of position (m) in the CT model under the different scenarios.
Table 4. ARMSE of position (m) in the CT model under the different scenarios.
ScenariosUKFCKFPFGCFMCGCFGLGCF
Scenario 185.276594.727958.208623.292417.163416.2886
Scenario 294.5022101.377389.377949.474634.244830.4771
Scenario 323.889525.510523.000612.850511.11679.1751
Scenario 423.795123.061721.238312.92558.21387.8886
Table 5. ARMSE of velocity (m/s) in the CT model under the different scenarios.
Table 5. ARMSE of velocity (m/s) in the CT model under the different scenarios.
ScenariosUKFCKFPFGCFMCGCFGLGCF
Scenario 19.309810.25666.62124.89734.02053.5092
Scenario 210.734811.361011.13119.58346.99615.9547
Scenario 33.04113.20342.51921.89451.59661.3459
Scenario 42.95212.95452.48291.91091.32611.2425
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kuang, Z.; Wang, S.; Zheng, Y. Robust Generalized Loss-Based Nonlinear Filtering with Generalized Conversion. Symmetry 2025, 17, 334. https://doi.org/10.3390/sym17030334

AMA Style

Kuang Z, Wang S, Zheng Y. Robust Generalized Loss-Based Nonlinear Filtering with Generalized Conversion. Symmetry. 2025; 17(3):334. https://doi.org/10.3390/sym17030334

Chicago/Turabian Style

Kuang, Zhijian, Shiyuan Wang, and Yunfei Zheng. 2025. "Robust Generalized Loss-Based Nonlinear Filtering with Generalized Conversion" Symmetry 17, no. 3: 334. https://doi.org/10.3390/sym17030334

APA Style

Kuang, Z., Wang, S., & Zheng, Y. (2025). Robust Generalized Loss-Based Nonlinear Filtering with Generalized Conversion. Symmetry, 17(3), 334. https://doi.org/10.3390/sym17030334

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop