1. Introduction
Estimation and filtering techniques are essential to successfully modeling and controlling engineering systems. By having a well-estimated state, the noise in the measured signal can be reduced and the quality of the controller can be improved [
1]. Filters extract valuable information from noisy measurement signals and forward this information to a control system. Several works have studied filtering and estimation theory and can be summarized based on the principles that were used to develop the filter. Important features and methods that are considered for filters include optimality, robustness, and stability, combining techniques from the first two categories, and the utilization of artificial intelligence.
The most popular filter in the optimality category is the Kalman filter (KF), which was introduced by Rudolph Kalman in the 1960s [
2]. This work was formulated to solve a linear stochastic system that involves Gaussian-white signals by reducing the state error covariance matrix. Later, a significantly large number of works were published that modified the KF in order to improve robustness and make it applicable to nonlinear systems and measurements. These works include the extended KF (EKF) [
2,
3], where a first-order Taylor series expansion is used to linearize the system and measurement matrices. Other works used the unscented transformation, such as the unscented KF (UKF) [
3,
4], and the Cubature rule, such as the cubature KF (CKF) [
5], to approximate the nonlinear functions. These techniques used several points that are drawn from pre-defined density functions and propagated through the nonlinear functions. Next, the methods used statistical regression summation as a linearization technique.
Several works targeted the robustness of the KF by using simple algebraic techniques as in [
6,
7,
8,
9,
10,
11]. For example, some used QR decomposition and Cholesky factoring for the covariance matrices [
10,
11]. Other techniques used boundaries for the error to make sure it remained close to the state estimates [
12]. For example, given the upper bounds on the level of modeling uncertainty, the KF gain may be bounded to help improve the estimation stability [
12]. Other methodologies include the addition of fictitious system noise and adding a fading memory strategy [
3,
13]. Modifying the system noise matrix was performed when less confidence was placed in the system model used by the filter, or in other words when there was a great deal of system uncertainty [
3,
13,
14,
15]. Finally, the H∞ filter was introduced in [
16,
17]. This filter takes into consideration the worst-case uncertainties to derive the gain, which ensures that the estimates are bounded.
The second category targets the robustness and stability of the filter. The main filters in this area are sliding mode (SMO), variable structure (VSO) observers [
13,
18,
19,
20,
21], and the strong tracking filter (STF) [
22,
23]. These observers define hyperplanes, usually functions of the innovation, and apply a discontinuous switching force to keep the estimate within a region of the true trajectory. For VSO, these hyperplanes divide the trajectory space into sub-regions in which the system dynamics are continuous. After crossing one of these planes, the observer changes its structure [
20,
24]. On the other hand, the hyperplanes of SMO are used to keep the estimates on them while sliding to towards zero [
25,
26]. A special form of the VSO and SMO was introduced in [
1,
27,
28] in a predictor-corrector form, referred to as the smooth variable structure filter (SVSF). The SVSF gain is a function of the a priori and a posteriori innovation vectors and a memory function. Later, in 2020, a new filter referred to as the sliding innovation filter (SIF) joined the family [
29]. SIF shared the same principles as the SVS but was simple in structure. SIF was used in target tracking [
30,
31], linear systems [
32], unmanned aerial vehicles [
33], and fault detection and diagnosis [
34].
The third and fourth categories involve merging the first two categories to overcome limitations in each method, as per [
35,
36,
37,
38,
39]. For example, the use of sigma points with the KF, such as central difference and unscented KF and the cubature KF, were combined with the SVSF in order to improve the accuracy of the SVSF while maintaining robustness. Other works combined the H∞ filter with the KF as in [
40,
41]. Some other works combined AI techniques such as in [
42,
43,
44] where fuzzy logic was combined with the KF, and in [
45,
46,
47] where AI was combined with the SMO and SVSF. The SIF was combined with multiple model-based filters in [
48,
49] and particle filters in [
50].
In this paper, a novel algorithm referred to as the alpha sliding innovation filter (ASIF) is proposed. The proposed ASIF method gives an efficient yet simple mechanism that improves the performance of the standard SIF. By utilizing a forgetting factor coefficient, the level of measurement confidence can be optimized. Measurement noise affects the performance of the SIF; therefore, the SIF can be improved by minimizing the effects of noise using a forgetting factor. This paper is organized as follows. The SIF is introduced in
Section 2, followed by the proposed ASIF algorithm in
Section 3. The proof of stability for the proposed filter is derived in
Section 4. The algorithm is tested and verified in
Section 5 on three different systems. The paper is then concluded in
Section 6.
2. The Sliding Innovation Filter (SIF)
The SIF is a new estimation technique that was proposed in [
29] and later modified in [
51,
52]. The filter is a predictor-corrector filter that shares the same principles as the SVSF [
35,
53,
54,
55]. It utilizes the innovation as a switching hyperplane and forces the estimate to be within its neighborhood. The process starts by assuming that the system and measurements behave according to the following, respectively:
It then uses the knowledge of the system model to obtain a priori, predicted estimates and measurements as follows:
The a priori estimate is refined in the corrector stage of the a posteriori estimate, as follows:
where
is the correction gain of the SIF defined in Equation (6) and
refers to the innovation or measurement error.
where
is the saturation function written as diagonal elements in a matrix and is defined as:
The elements have values between 0 and +1. The boundary layer
is used to smooth the estimate and reduce the effect of the measurement noise [
29]. However, if the innovation becomes larger than the boundary layer, then
becomes the identity matrix. The SIF estimation process is illustrated in
Figure 1.
3. The Proposed Alpha Sliding Innovation Filter (ASIF) Strategy
This section describes the proposed modification to the standard SIF. A forgetting factor coefficient (FF) or
is introduced to create the ASIF. The SIF using (2.6) is considered robust [
29]. However, its performance in terms of optimality highly depends on the selection of the boundary layer. The selection mechanism proposed in [
29] to obtain the widths of these layers was based on trial and error. However, this is not an easy task for several reasons, summarized as follows:
The range of acceptable is infinitely wide, as it can be any positive value .
The number of boundary layers required is equal to the number of measurements, which means exhaustive trials for high-dimensional systems (e.g., AI output layers).
In most cases, is selected to be the maximum allowable error in the system. For small innovation amplitudes, the filter will have a very small convergence rate, and for large amplitudes, the estimates may chatter.
These limitations inspired the authors to develop a modified version of the SIF that improves the performance while overcoming these limitations. A fading memory is added to the SIF’s gain. By adding the coefficient , the term can be ignored and replaced with an identity matrix. In our case, is a measure of measurement confidence, as per the following:
If , ASIF collapses to the SIF with a very small boundary layer.
If , the filter depends more on the system and less on the measurement. This can be used to reduce the effect of the measurement noise. This helps significantly when is larger than .
If is large, the filter depends more on the measurement, which makes the ASIF sensitive to the measurement noise but less sensitive to modeling uncertainties. This helps when is larger than .
For the ASIF estimation strategy, the proposed gain has the following simplified structure:
Note that the ASIF estimation process is similar to the SIF, except that Equation (8) is used instead of Equation (6).
4. Proof of Stability
The SIF gain was developed to guarantee the stability and robustness of the filter. Adding the forgetting factor changes the filter structure. In this section, the stability of the new ASIF is examined. We consider the discrete Lyapunov function
defined as follows:
Then, according to Equation (9), the proposed filter is stable as long as the innovation is decreasing with time, as follows:
Referring to the equations of
Section 2 and
Section 3, taking the difference between Equations (1) and (5) yields:
Substituting Equation (2) in (4) results in:
Substituting Equation (12) in (11) and multiplying by
gives:
Subtracting Equation (3) from Equation (1) and using Equation (12) yields the following:
Substituting Equation (15) into Equation (13) yields:
Rewriting Equation (16) and simplifying yields:
where
is defined as an uncertainty vector as follows:
If the measurement and system noise vectors are white, then
. Therefore, taking the expectation of Equation (17) yields:
From Equation (19) and Equation (20), the filter is considered stable if
where
refers to the expectation or expected value. The system under study is stable if the eigenvalues of
are less than unity. Therefore, without losing generality,
is less than unity, which yields the following condition:
Solving Equation (21) further gives:
Therefore, for the ASIF estimation process to be stable, must be defined as per Equation (22).
5. Computer Experiments and Results
The proposed ASIF estimation strategy will be tested on first-, second-, and third-order systems. The ASIF is compared with the original SIF and the well-known KF. The study was conducted on MATLAB and was repeated 1000 times. The repetitions were conducted to obtain the Monte Carlo Simulation (MCS) in order to provide better insight into the different estimation methods. The parameters, , and are estimated for no modeling uncertainties case. Their values are obtained by trial and error. They are assumed to be valid for the molding uncertainties case.
The results are compared in terms of root mean squared error (
RMSE) and maximum absolute error (
MAE) under normal operating conditions, and modeling uncertainty or faulty conditions.
RMSE and
MAE are calculated using Equations (23) and (24), respectively, assuming
is the number of sampled data.
5.1. Mercury Thermometer
A thermometer has a simple first-order differential equation, which is derived from the conservation of energy. The fluid transfers the heat by convection and stores it in the fluid using:
The above relation can be simplified to:
where
is the time constant with a value of 1.7 sec [
56]. Rearranging Equation (26) yields:
Equation (25) can be discretized into the following:
or using the following state space form:
where
is the temperature of the surroundings and is treated as the system input. It is selected to change its value between
and
every second. The system and measurement noises (
and
) are normally distributed with zero mean and covariance’s
and
defined by Equations (31) and (32), respectively.
is equal to unity and
is selected as shown in Equation (33).
The initial values for the filters and their corresponding initial covariance values were set to zero and one, respectively. The boundary layer was set to 0.5 based on trial and error in order to minimize the estimation error. The simulations were repeated after injecting the system with an uncertainty of 20%, i.e.,
becomes 0.7995.
Figure 2 shows the averaged RMSE and MAE for the system with and without modeling uncertainties, respectively.
Figure 3 shows the actual estimates for the two respective cases.
Table 1 and
Table 2 include the calculated RMSE and MAE of the MCS without and with modeling uncertainties, respectively.
When no modeling uncertainties were present, the RMSE and MAE results for the ASIF illustrate a better performance compared with the KF and SIF strategies. The performance of the ASIF remains better than the KF when uncertainties were present. Compared to the standard SIF, the ASIF yielded better estimates. However, when uncertainties were present, both the SIF and ASIF estimation strategies performed similarly, with slight superiority to SIF.
5.2. Spring-Mass-Damper
The well-known spring-mass-damper system (S-M-D) is mathematically described using a second-order differential equation, according to the following formula:
Equation (34) can be represented as discretized state space equations (system and measurement) as follows:
where
is the force applied to the system, which is selected to change its value between 50 and 500 Newton every second. The system and measurement noises (
and
) are normally distributed with zero mean and covariance’s
and
defined by Equations (37) and (38), respectively. The boundary layer and
are chosen as shown in Equations (39) and (40), respectively.
is the identity matrix.
The initial values for the filters and their corresponding initial covariance values were set to zero vectors and identity matrices, respectively. The simulations were repeated after injecting the system with uncertainties as mentioned in
Table 3. The system parameters are summarized in
Table 3.
Figure 4 shows the RMSE and MAE for the first and second states for the two case scenarios: No modeling and modeling uncertainties present. The true and estimated state values for the first and second states with no modeling and with modeling uncertainties present are shown in
Figure 5.
Table 4 and
Table 5 summarize the averaged RMSE and MAE for the cases without and with modeling uncertainties. ASIF shows a similar performance to KF when no modeling uncertainties present, which is better than those obtained by SIF. Once modeling uncertainties were injected, ASIF has a similar performance to SIF, which is better than KF.
5.3. Electrohydrostatic Actuator
The third and final example in this study includes an aerospace flight surface actuator proposed and studied in [
6,
27,
29]. The linear system and measurements are formulated as:
where
is the controller input. In this case, it is defined as a multistep input that changes its value between −0.5 and 0.5
every second. The system and measurement noise covariance’s are
and
defined by Equations (43) and (44), respectively. As per the previous examples, modeling uncertainties are injected using Equation (45). The boundary layer and
are chosen as shown in Equations (46) and (47), respectively.
is the identity matrix.
The initial values for the filters, their corresponding initial covariance values, and the boundary layer’s width were set to be similar to the S-M-D example.
Figure 6 shows the RMSE and MAE for the three states without and with modeling uncertainties present.
Figure 7 shows the actual and estimated state values for the states in both scenarios.
Table 6 and
Table 7 summarize the RMSE and MAE for the cases without and with modeling uncertainties.
Under normal operating conditions, the KF provided the best results (the RMSE and MAE were the smallest). In the presence of uncertainties (e.g., a fault) and after tuning the KF and ASIF, they performed similarly. Note that the SIF and ASIF were not as sensitive to uncertainties compared with the KF.
To complete the comparison, the simulation time for the three previous examples was measured for the three filters, namely KF, SIF, and ASIF.
Figure 8 shows that the ASIF has the shortest time among the other filters. It requires less than half the time that is required for the other filters in the second and third examples. Adding this to the results obtained from RMSE and MAE tables, ASIF can be considered to have an overall superior performance.