Next Article in Journal
Wirelessly Powered Light and Temperature Sensors Facilitated by Electrically Small Omnidirectional and Huygens Dipole Antennas
Previous Article in Journal
FPGA-Based Degradation and Reliability Monitor for Underground Cables
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Criteria for the Identification of Inertial Sensor Error Models

CSRI Elektropribor, JSC, ITMO University, 190000 Saint Petersburg, Russia
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(9), 1997; https://doi.org/10.3390/s19091997
Submission received: 21 February 2019 / Revised: 22 April 2019 / Accepted: 25 April 2019 / Published: 29 April 2019
(This article belongs to the Section Physical Sensors)

Abstract

:
This paper considers performance criteria for the identification of sensor error models and the procedure for their calculation. These criteria are used to investigate the efficiency of the identification problem solution, depending on the initial data, and to carry out a comparative analysis of various suboptimal algorithms. The calculation procedure is based on an algorithm that solves the joint problem of hypothesis recognition and parameter estimation within the Bayesian approach. A performance analysis of the models traditionally used to describe errors of inertial sensors is given to illustrate the application of the procedure for the calculation of performance criteria.

1. Introduction

Currently, algorithms based on system description in state space are widely used in control and estimation problems. To estimate the state vector of such systems, linear and nonlinear stochastic Bayesian filtering algorithms, such as the Kalman filter, unscented Kalman filter, and particle filter [1,2,3,4], which require the derivation of system models in state space in the form of random processes, are often used. For example, such a problem arises during the processing of redundant measurements, in particular, navigation data fusion, which involves signals from multiple sensors [5,6,7]. To design such stochastic filtering algorithms, we need to describe sensor errors. This, in turn, generates the need to solve the identification problem of their models. The problem aims to reveal the components of sensor errors, i.e., to determine the model structure and estimate the unknown parameters of its components based on the data obtained during the tests. The data are usually obtained by comparing the readings of the sensors with reference values, and as a result only contain information about sensor errors. Common tests, possible structures of error models, and the corresponding parameters to be identified, as well as some methods for their derivation, are given, for example, in [8,9]. The review shows that methods based on the analysis of sample characteristics are widely used in applied problems concerned with the processing of measurement data. Such sample characteristics include power spectral density [10,11], Allan variance (AV) [12,13,14,15,16,17,18], etc. These methods are usually effective only for stationary ergodic processes; in addition, they require long-term measurements. Moreover, in some cases the obtained model cannot be applied directly to Bayesian stochastic filtering algorithms [1,2,3,4,5,6,7,19,20] that imply the description of errors in time domain using the shaping filter defined in state space. The wavelet analysis is becoming increasingly popular nowadays [21]. The approach proposed in [21,22,23] delivers a global selection criterion based on the wavelet variance that can be used to design an algorithm for automatic identification of a model structure.
Many methods for the identification of state space systems have been developed so far [24,25,26,27,28,29]. For example, [1,30,31,32] formulate the problem of state space model identification within the framework of the Bayesian approach as a joint problem of hypotheses about the model structure recognition and estimation of its parameters. The corresponding problem statement as well as its solution algorithms are inherently close to the statements and algorithms considered in tracking problems [33,34,35,36] or system diagnostics [37,38].
However, such methods have not been properly applied for the identification of sensor error models. In [39] and [40] the authors use the Bayesian approach to estimate sensor error parameters only. In [41,42] the authors show that the problem can be generalized for the identification of the sensor error model structure as a joint problem of hypothesis recognition and parameter estimation. The problem solution within the state space and Bayesian formulation allows us to calculate directly the probabilities of correct structure identification in order to choose the sensor error model. The obtained estimates of the model parameters are the optimal Bayesian estimates for any time interval. The corresponding root-mean-square (RMS) of estimation errors can also be calculated both for stationary and nonstationary models in the frames of this approach.
In this paper we propose a method for determining the estimation quality for the problem of sensor error model identification. The straightforward calculation of the structure identification probabilities allows us to compute the unconditional probability for correct model identification and unconditional relative standard deviations for the estimated parameters. These unconditional characteristics can be used as identification quality criteria. The criteria can be computed by simulation to answer the question of whether it is possible in principle to determine the model and its parameters under certain conditions. Thus, they provide a quantitative measure of possibility to identify a model under certain conditions and evaluate the quality of such problem solution within a certain suboptimal algorithm such as the AV.
The Bayesian approach-based algorithm for identification of a sensor error model was considered in [42], but in this paper we focus on obtaining performance criteria for solution of the identification problem, propose methods for their calculation and give an example of their practical application.
The paper is structured as follows: Part 2 considers the statement of the identification problem for the sensor error model and estimation of its parameters in the framework of the Bayesian approach. The general algorithm for its solution is described here as well. Part 3 introduces the quality criteria for the problem and the method for their calculation. Part 4 illustrates the proposed method by the example of the identification problem solution for models traditionally used to describe the errors of navigation sensors.

2. Problem Statement and Solution Algorithm

To state the problem within the Bayesian approach, let us introduce the following model. Assume that we have a set of hypotheses about sensor error models, one of which is true. Moreover, each model can be represented in state space by the shaping filter [39,41]:
x i k = Φ i k ( θ k ) x i 1 k + Γ i k ( θ k ) w i k , θ i k = θ i 1 k = θ k , }
y i = H i k ( θ k ) x i k + Ψ i k ( θ k ) v i k
where y i is a sensor error sample obtained in the tests; x i k is the error model state vector; k is the number of the hypothesis that specifies the structure of the sensor error model, k = 1 K ¯ is a natural number from 1 to K; Φ i k ( θ k ) , Γ i k ( θ k ) , H i k ( θ k ) , Ψ i k ( θ k ) are the matrices describing the shaping filter for this model, which are nonlinearly dependent on the constant vector of parameters θ k and time instance i; w i k is p k -dimensional system noise; v i k describes the m k -dimensional white-noise component of the sensor error. Both are discrete zero-mean Gaussian white noise with identity covariance matrices. Vector x i k is assumed to be Gaussian with known prior expectation and covariance. Prior probability density function (PDF) for θ k is also known. Note that for different hypotheses the dimensionality of vectors θ k and x i k can also be different and the model can be time-varying because matrices Φ i k ( θ k ) , Γ i k ( θ k ) , H i k ( θ k ) , Ψ i k ( θ k ) can depend directly on time instant i.
Under the assumptions made, the problem of model identification can be formulated as a joint problem of hypotheses recognition and parameter estimation, i.e., the problem of determining the hypothesis number k, which is best suited for measurements Y i = [ y 1 y i ] T , and estimating the corresponding vectors θ k and x i k . Thus, we can completely describe the sensor error model if we know the hypothesis number and the corresponding vector θ k estimate.
Such a problem can be solved with the algorithm described in [40,42]. It allows us to obtain the number of hypothesis that corresponds to the maximum posterior probability and the optimal Bayesian estimates corresponding to the hypothesis vectors θ ^ k and x i k . The algorithm starts with definition of initial conditions, which is followed by sequential calculation of the hypotheses probabilities and estimates of parameter and state space vectors for each time.
First, a set of K hypothesis models Equations (1) and (2) is defined, which is potentially suitable as a sensor error model. Prior PDFs for the vectors θ k and x 0 k are usually defined as uniform and Gaussian ones. Their parameters can be roughly estimated using sample characteristics. The PDFs for vectors θ ^ k are approximated using, for example, the point-mass method [1,2,43]:
f θ k ( θ k / Y i , H = h k ) j = 1 M k μ i k j δ ( θ k θ k j )
where θ k j , j = 1 M k ¯ is a grid for θ k derived on a domain of its PDF within a fixed hypothesis h k , μ i k j are PDF approximation weights. The initial weights μ 0 k j corresponding to the prior PDF are defined as equal and normed ones, so that j = 1 M k μ i k j = 1 ,     k . The Monte Carlo method can also be used for such approximation [1,44,45]. The feature of the algorithm is the subsequent sequential computation of the posterior PDF approximations weights μ i k j for vector θ k with k = 1 . K ¯ , which yields the required estimates and hypotheses probabilities. It should be noted that for a fixed value θ k j , the problem of vector x i k estimation in Equations (1) and (2) becomes linear and can be solved using a Kalman filter (KF). This method is also referred to as the Rao-Blackwellization procedure [1,44]. Furthermore, we describe this recursive procedure, which consists of five steps.
In the first step, matrices Φ i k ( θ k j ) , Γ i k ( θ k j ) , H i k ( θ k j ) , Ψ i k ( θ k j ) are calculated for the whole set of values θ k j within the set of grids of all hypotheses. For simplicity, hereinafter, these matrices are referred to as Φ i k j , Γ i k j , H i k j , Ψ i k j . Thus, the KF bank is formed in accordance with shaping filters described by Equations (1) and (2), and the estimates of vectors x i k j and their covariances P i k j are calculated using KF equations:
x ^ i / i 1 k j = Φ i k j x ^ i 1 k j , x ^ i k j = x ^ i / i 1 k j + K i k j ( y i H i k j x ^ i / i 1 k j ) ,
K i k j = P i / i 1 k j ( H i k j ) T ( H i k j P i / i 1 k j ( H i k j ) T + Ψ k j ( Ψ k j ) T ) 1
P i k j = ( E K i k j H i k j ) P i / i 1 k j , P i / i 1 k j = Φ i k j P i 1 k j ( Φ i k j ) T + Γ i k j Q i k ( Γ i k j ) T ,
Here, the upper indexes are used to denote that Equations (4)–(6) correspond to the linear filtering problem within the fixed grid values of vector θ k j and hypothesis h k . The innovations
λ i k j = y i H i k j x ^ i / i 1 k j
and the corresponding covariances
Λ i k j = H i k j P i / i 1 k j ( H i k j ) T + Ψ i k j R i k j ( Ψ i k j ) T
are also computed in this step using estimates x ^ i k j and the corresponding covariances P i k j .
In the second step, the values of likelihood functions
f y i ( y i / Y i 1 , H = h k , θ k j ) = N ( λ i k j ; 0 , Λ i k j ) , f x i k ( x i k / Y i , H = h k , θ k j ) = N ( x i k ; x ^ i k j , P i k j )
are calculated for the whole set of values θ k j by using estimates x ^ i k j , innovations λ i k j , and the corresponding covariances P i k j , Λ i k j . Here notation N ( x ; x ¯ , P ) means the Gaussian function of random vector x with expectation x ¯ and covariance matrix P . The PDFs in Equation (9) are Gaussian relative to λ i k j and x i k , respectively, because the problem considered in Equations (1) and (2) is linear for fixed values of θ k j .
In the third step, the posterior probabilities of all hypotheses h k are calculated according to
Pr ( H = h k / Y i ) [ j = 1 M k μ i 1 k j N ( ν i k j ; 0 , Λ i k j ) ] Pr ( H = h k / Y i 1 ) k = 1 K [ [ j = 1 M k μ i 1 k j N ( ν i k j ; 0 , Λ i k j ) ] Pr ( H = h k / Y i 1 ) ]
using the values of the likelihood functions (Equation (9)) computed in the previous step. Equation (10) can be obtained from Bayes’ rule [1,2], taking into account that the likelihood function for each hypothesis can be calculated as:
f y i ( y i / Y i 1 , H = h k ) j = 1 M k μ i 1 k j N ( ν i k j ; 0 , Λ i k j )
At the same step, hypothesis h k * with the highest computed probability Pr ( H = h k * / Y i ) is detected. However, the decision that the model corresponding to the hypothesis is true, is only made when Pr ( H = h k * / Y i ) γ , where γ is a threshold close to unity. As a rule, the value of γ is assumed to be about 0.8.
In the fourth step, the approximation coefficients μ i k j for posterior PDFs of vectors θ k are computed according to Bayes’ rule:
μ i k j = μ i 1 k j f y i ( y i / Y i 1 , H = h k , θ k j ) j = 1 L μ i 1 k j f y i ( y i / Y i 1 , H = h k , θ k j ) .
In the fifth step, the estimates of vectors θ k , x i k and their covariances are calculated according to the following equations:
θ ^ i k ( Y i ) = j = 1 M k μ i k j θ k j ,   x ^ i k ( Y i ) = j = 1 M k μ i k j x ^ i k j ,
P i θ k ( Y i ) = j = 1 M k μ i k j θ k j ( θ k j ) T θ ^ i k θ ^ i k T , P i x k ( Y i ) = j = 1 M k [ μ i k j ( x ^ i k j ( x ^ i k j ) T + P i x k j ) ] x ^ i k ( x ^ i k ) T
where the approximation coefficients μ i k j obtained in the previous step are used.
As follows from the presented expressions, the algorithm has a high computational complexity. In order to overcome this drawback, in future works it is supposed to design special procedures aimed at reducing the amount of computations.

3. Performance Criteria Calculation

The performance criteria for identification of stochastic model described by Equations (1) and (2) can be determined separately: firstly, for the probability of the true model structure identification, and secondly, for the accuracy of the model parameter estimates. For the model structure, the unconditional probability of the identification error can be used as a performance criterion. This probability can be computed directly using the method of statistical trials:
1 Pr i ¯ R ( H = h k ) 1 L * L ,
where L * is the number of samples by which the algorithm identified the models correctly, and L is the total number of the simulated samples. The accuracy of the model parameter estimates can be evaluated using the unconditional covariance matrix of estimation errors:
G i θ k = E { ( θ k θ ^ i k ) ( θ k θ ^ i k ) T } ,
where E { } is an expectation sign. Thus, the diagonal elements of this matrix will be used for parameter estimates as a performance criterion. The elements can also be computed using the method of statistical trials:
G i θ k = G i R θ k 1 L 1 l = 1 L ( θ k ( l ) θ ^ i k ( Y i l ) ) ( θ k ( l ) θ ^ i k ( Y i l ) ) T
where θ k ( l ) is a true value of the unknown parameters vector in sample l, θ ^ k ( Y i l ) is the estimate of vector θ k ( l ) computed using an algorithm and measurements Y i l . These elements determine the variance of the estimation errors for the components of the unknown parameters vector. In what follows, we call the performance criteria obtained using Equations (15) and (17) a real identification error probability and a real estimation error variance, correspondingly; they will be indexed with “R”.
The unconditional performance criteria in Equations (15) and (16) can also be computed using the probability (Equation (10)) and covariances (Equation (14)) from the above algorithm and the method of statistical trials:
1 Pr i ¯ C ( H = h k ) 1 L l = 1 L ( 1 Pr i l ( H = h k * / Y i l ) ) ,
G i θ k = G i C θ k = P i θ k ( Y i ) f ( Y i ) d Y i 1 L l = 1 L P i θ k ( Y i l )
where Pr i l ( H = h k * / Y i l ) is the probability of the true hypothesis h k * . From here on, the values of the performance criteria obtained from Equations (15) and (19) will be called the calculated identification error probability and the calculated estimation error variance, correspondingly, and they will be indexed with “C”. The coincidence of the real and calculated performance criteria is evidence in favor of the validity of the results obtained.
For a particular performance analysis, when the mean value of the unknown parameter is not zero, it is useful to compare not variances, but the relative standard deviation (RSD), which is defined as the ratio of the standard deviation to the mean [46]:
V i θ k ( j ) = G i θ k ( j , j ) / E { θ k ( j ) }
where j is the number of vector θ k component, G i θ k ( j , j ) is the element of Equation (16) with index ( j , j ) . Note that depending on the procedure by which Equation (16) was calculated, we will differentiate between the real RSD, corresponding to the real covariance matrix (Equation (17)), and the calculated RSD, corresponding to the calculated matrix (Equation (19)).

4. Examples of Performance Criteria Use

To illustrate the use of the performance criteria, we consider two hypotheses corresponding to the shaping filters:
{ x i 1 1 = x ( i 1 ) 1 1 + σ w w i 1 , x i 2 1 = x ( i 1 ) 2 1 ,
{ x i 1 2 = exp ( Δ t / τ m ) x ( i 1 ) 1 2 + σ m 1 exp ( 2 Δ t / τ m ) w i 2 , x i 2 2 = x ( i 1 ) 2 2 ,
and measurements, which represent a sample of a total sensor error, in the form:
y i = x i 1 k + x i 2 k + σ v k v i k ,    k = 1 , 2 .
In the above relations, w i k , v i k are independent of each other zero-mean Gaussian white noises with unit variances; σ w is the RMS deviation of the system noise for the random walk; τ , σ m are the correlation interval and the RMS deviation of the first-order Markovian process in Equation (22), respectively; σ v k is the RMS deviation of the white-noise component, Δ t is the discretization interval. These models are quite common to navigation sensor error models that usually contain constant bias, drift, and white-noise components [8,12].
The identification problem for this case can be formulated as follows. Find the number of the hypothesis corresponding to the maximum posterior probability Pr ( H = h k / Y i ) , optimal Bayesian estimates for the state vector x i k = [ x i 1 k x i 2 k ] T and the vector of parameters θ k for this hypothesis using the given measurements Y i = [ y 1 y i ] T . Thus, each hypothesis corresponds to the different sensor error models. The proposed algorithm chooses one of them using the posterior probability Pr ( H = h k / Y i ) and estimate the corresponding parameters θ ^ k . However, the aim of the simulation is to calculate the performance criteria (Equation (15)), and Equations (17)–(20) to show the possibility of identifying sensor error models in certain conditions.
To show how the estimation accuracy changes with the changes in the parameter values, we consider three different cases. In the first case, the RMS of the white-noise component is assumed to be zero ( σ v k = 0 ,   k = 1 , 2 ) . Note that we used the KF Equation (Equation (5)), in which the noise covariance matrix is not inverted. The vectors of unknown parameters θ k are defined as θ 1 = σ w for model (22) and as θ 2 = [ τ m σ m ] T for model (22). Both vectors are uniformly distributed with elements within the ranges of σ w [ 0.04 0.15 ] u, τ m [ 1 25 ] min, σ m [ 0.5 1.5 ] u. Here, “u” means some arbitrary units, for example, deg/s or deg/h in the case of gyro errors, or m/s2 in the case of accelerometer errors. In the second case, the white-noise component is assumed to be known with RMS deviation σ v k = 0.5   u . ,   k = 1 , 2 . In the third case, the RMS of the white-noise component is unknown and uniformly distributed in the domain σ v 1 , 2 [ 0.25 1 ]   u .

4.1. Results of Model Structure Identification Probability

The real (Equation (15)) and calculated (Equation (18)) identification error probabilities are computed using the method of statistical trials for all the three cases described above. This implies that the problem is solved using the proposed algorithm for 500 simulated samples for each case.
In the first case, the mean probability of the identification error is close to zero after a couple of measurements (Figure 1a). Thus, these processes can be easily identified. To illustrate that not all methods allow us to do so, we give the AV [9,12] of these processes in Figure 1b. The AV [9,12] are obtained using a complete set of measurements of length 2000 s. The model identification using AV usually consists in empirical finding of significant slopes in the plot. We can hardly find difference between the AV for different models in Figure 1b because we can see only +1/2 slope in this plot, but there is no significant slope −1/2 of the first-order Markovian process. Taking into consideration the correlation interval of the first-order Markovian process, the slope −1/2 starts after the averaging time of 200 s, where AV is unreliable because of the sample length of 2000 s. To distinguish between these processes using AV, we should take a sample exceeding 20,000 s in length.
In the second and third cases, the mean probabilities of the identification error after 2000 s of measurements are about 2% (Figure 2a) and 10% (Figure 3a), respectively. Here, we may conclude that the uncertainty of the white-noise component affects significantly the model identification probability for these two models. The AV for these cases have two significant slopes: −1/2 and +1/2 that correspond to the white noise and random walk (Figure 2b and Figure 3b).
Thus, we can separate white noise and correlated processes, but the random walk and the first-order Markovian processes can hardly be distinguished with the use of AV as in the first case. Note that for the second and third cases, we did not set apart the white noise as a separate hypothesis, which is why we did not calculate the probability of its detection directly. Hence, the possibility of white-noise component identification by using the proposed algorithm can be judged from the corresponding RMS error of the white-noise component, which is presented in the next subsection.
It is shown that the first-order Markovian process and random walk cannot be distinguished by AV for all three cases. These results show that the potential accuracy of model and parameter identification can be much higher than the accuracy provided with the use of AV. The criteria for the performance can be calculated as unconditional error probabilities (Equations (15) and (18)). Figure 1a, Figure 2a and Figure 3a also show the coincidence of the real and calculated identification probabilities, which implies that the calculated values are reliable for estimation of identification probability in a real-data experiment.

4.2. Results of Parameter Estimation Accuracy

The mean accuracy characteristics for parameter estimation can also be calculated using the proposed algorithm. This subsection is devoted to the study of accuracy in different conditions. However, before we proceed to the analysis of parameter estimation accuracy, let us show how the results of such analysis can be generalized using RSD. Consider Equations (22) and (23) with different uncertainty ranges of parameters. For example, let τ m be uniformly distributed within the range of [ 1 5 ] min in the first case, τ m [ 5 25 ] min in the second case, and τ m [ 25 125 ] min in the third case. For these cases we can obtain calculated and real RMS (Figure 4a).
It can be seen from the plots (Figure 4a) that the RMS of the estimation errors are substantially different due to different true values of the correlation interval and the level of initial uncertainty. However, the RSD (Figure 4b) for different sizes of the prior uncertainty domain of parameter τ m coincide. Thus, the result obtained for a parameter with the initial uncertainty domain θ [ θ 1 θ 2 ] can be generalized for all domains of the form θ [ r θ 1 r θ 2 ] , where r is an arbitrary real number different from zero. Thus, we use RSD as an accuracy characteristic in the following plots.
Figure 5 and Figure 6 show the RSD for the estimates of parameters for the both models and all three cases considered above. It may be inferred that in the presence of white noise, the accuracy of the parameter estimation does not vary much regardless of whether the RMS noise is known or not. The greatest influence that causes decrease in the estimation accuracy is explained by the presence of the white-noise component.
In the third case, the RMS of the white-noise component is estimated with a rather high accuracy, which is the same for the both alternative models (Figure 6b). It should be noted that if we try to estimate the RMS of the white noise in a sample without white noise, such as in the first simulation case, the white noise RMS estimates will converge to zero. The above facts allow us to conclude that the white-noise component is easy to detect both with the use of the proposed algorithm and AV. Note that in Figure 4, Figure 5 and Figure 6 the calculated and real RSD coincide, which is evidence of the validity of the results obtained and reliability of the calculated values.

5. Conclusions

The Bayesian approach to identification of sensor error models based on the solution of a joint hypotheses recognition and parameter estimation problem is considered. The following main advantages of this approach are emphasized: possibility of estimating an unknown structure and parameters of a model at any time interval both for stationary and nonstationary processes; capability of calculating characteristics for estimation accuracy; the possibility of obtaining a model in the form that is convenient for solution of the estimation problem using stochastic filtering algorithms.
In the framework of this approach, the paper proposes the performance criteria for the identification problem of sensor error models in the form of an unconditional probability for identification of a true model structure and a relative standard deviation (RSD) for unknown parameters of the model. The probability of choosing the correct structure of the model is defined as the probability of choosing a shaping filter corresponding to the correct model. The RSD of unknown parameters are determined using the unconditional covariance matrices calculated for the chosen filter. The introduced criteria create preconditions for an objective study of the performance of the identification algorithm depending on the initial data. This analysis should be carried out to evaluate the possibility of the identification problem solution under given conditions, including a given set of hypotheses, or to choose the conditions that provide a satisfactory solution.
An example of the identification problem solution is given for the models that are traditionally used to describe errors of navigation sensors. The models are represented as a sum of correlated processes specified by the random walk or the first-order Markovian process, white-noise and constant components. The dependencies in the probability of correct model identification and the accuracy of parameter estimation from the RMS of the white-noise component are analyzed, including the case when the RMS is not precisely known and should be identified during the problem solution. It has been shown that the probability of model identification largely depends on the presence of the white-noise component. At the same time, the fact of uncertainty in the knowledge of the white-noise component RMS has an insignificant effect on the probability of model identification and the accuracy of parameter estimation.
In our future work the proposed criteria will be used to evaluate the performance of suboptimal algorithms, such as the Allan variance, for the problems of estimating error models of navigation sensors. This study will involve the comparison of suboptimal and the proposed algorithms.

Author Contributions

Conceptualization, Methodology, Supervision, Project Administration, Funding Acquisition, Validation, O.S.; Investigation, Software, Data Curation, Writing-Original Draft Preparation, Visualization, A.M., Writing-Review & Editing O.S., A.M.

Funding

This work was financially supported by the Government of the Russian Federation (Grant 08-08) and by the Russian Foundation for Basic Research, project no. 18-08-01101А.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Sarkka, S. Bayesian Filtering and Smoothing; Cambridge University Press: Cambridge, UK, 2013; ISBN 9781139344203. [Google Scholar]
  2. Chen, Z.H.E. Bayesian Filtering: From Kalman Filters to Particle Filters, and Beyond. Statistics 2003, 182, 1–69. [Google Scholar] [CrossRef]
  3. Haykin, S. Kalman Filtering and Neural Networks; Haykin, S., Ed.; John Wiley & Sons, Inc.: New York, NY, USA, 2001; ISBN 9780471221548. [Google Scholar]
  4. Simon, D. Kalman Filtering. Embed. Syst. Program. 2001, 14, 72–79. [Google Scholar]
  5. Grewal, M.S.; Andrews, A.P.; Bartone, C.G. Global Navigation Satellite Systems, Inertial Navigation, and Integration, 3rd ed.; John Wiley & Sons, Inc.: New York, NY, USA, 2013. [Google Scholar]
  6. Markley, F.L.; Crassidis, J.L. Fundamentals of Spacecraft Attitude Determination and Control; Springer Verlag: New York, NY, USA, 2014; ISBN 978-1-4939-0801-1. [Google Scholar]
  7. Stepanov, O.A. Optimal and sub-optimal filtering in integrated navigation systems. In Aerospace Navigation Systems; Nebylov, A., Watson, J., Eds.; John Wiley & Sons, Inc.: New York, NY, USA, 2016; p. 392. [Google Scholar]
  8. IEEE Recommended Practice for Inertial Sensor Test Equipment, Instrumentation, Data Acquisition, and Analysis; IEEE Aerospace and Electronic Systems Society, 2005. Available online: https://ieeexplore.ieee.org/document/6673990 (accessed on 8 November 2018).
  9. IEEE Std 528-2001. IEEE Standard for Inertial Sensor Terminology; IEEE Aerospace and Electronic Systems Society, 2001. Available online: https://ieeexplore.ieee.org/document/972830 (accessed on 8 November 2018).
  10. Marple, S.L.J. Digital Spectral Analysis with Applications; Prentice-Hall: Upper Saddle River, NJ, USA, 1987. [Google Scholar]
  11. Quinchia, A.; Falco, G.; Falletti, E.; Dovis, F.; Ferrer, C. A Comparison between Different Error Modeling of MEMS Applied to GPS/INS Integrated Systems. Sensors 2013, 13, 9549–9588. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. El-Sheimy, N.; Hou, H.; Niu, X. Analysis and modeling of inertial sensors using Allan variance. IEEE Trans. Instrum. Meas. 2008, 57, 140–149. [Google Scholar] [CrossRef]
  13. Tehrani, M. Ring Laser Gyro Data-Analysis with Cluster Sampling Technique. Proc. SPIE Int. Soc. Opt. Eng. 1983, 412, 207–220. [Google Scholar]
  14. Allan, D.W. Statistics of Atomic Frequency Standards. Proc. IEEE 1966, 54, 221–230. [Google Scholar] [CrossRef]
  15. Allan, D.W. Historicity, strengths, and weaknesses of Allan variances and their general applications. Gyroscopy Navig. 2016, 7, 1–17. [Google Scholar] [CrossRef]
  16. Stepanov, O.A.; Chelpanov, I.B.; Motorin, A.V. Accuracy of sensor bias estimation and its relationship with Allan variance. Gyroscopy Navig. 2017, 8, 51–57. [Google Scholar] [CrossRef]
  17. Miao, Z.; Shen, F.; Xu, D.; He, K.; Tian, C. Online Estimation of Allan Variance Coefficients Based on a Neural-Extended Kalman Filter. Sensors 2015, 15, 2496–2524. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Yu, H.; Wu, W.; Wu, M.; Feng, G.; Hao, M. Systematic Angle Random Walk Estimation of the Constant Rate Biased Ring Laser Gyro. Sensors 2013, 13, 2750–2762. [Google Scholar] [PubMed] [Green Version]
  19. Niu, X.; Wu, J.; Zhang, Q. Research on Measurement Error Model of GNSS/INS Integration Based on Consistency Analysis. Gyroscopy Navig. 2018, 9, 243–254. [Google Scholar] [CrossRef]
  20. Chu, H.-J.; Tsai, G.-J.; Chiang, K.-W.; Duong, T.-T. GPS/MEMS INS Data Fusion and Map Matching in Urban Areas. Sensors 2013, 13, 11280–11288. [Google Scholar] [CrossRef]
  21. Guerrier, S.; Skaloud, J.; Stebler, Y.; Victoria-Feser, M.-P. Wavelet-Variance-Based Estimation for Composite Stochastic Processes. J. Am. Stat. Assoc. 2013, 108, 1021–1030. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Guerrier, S.; Molinari, R.; Skaloud, J. Automatic Identification and Calibration of Stochastic Parameters in Inertial Sensors. Navigation 2015, 62, 265–272. [Google Scholar] [CrossRef]
  23. Radi, A.; Bakalli, G.; El-Sheimy, N.; Guerrier, S.; Molinari, R. An Automatic Calibration Approach for the Stochastic Parameters of Inertial Sensors. In Proceedings of the 30th International Technical Meeting of The Satellite Division of the Institute of Navigation (ION GNSS+ 2017), Portland, OR, USA, 25–29 September 2017; pp. 3053–3060. [Google Scholar]
  24. Åström, K.J. Maximum likelihood and prediction error methods. Automatica 1980, 16, 551–574. [Google Scholar] [CrossRef]
  25. Ljung, L. System Identification: Theory for User, 2nd ed.; Prentice-Hall: Englewood Cliffs, NJ, USA, 1999; ISBN 0138816409. [Google Scholar]
  26. Stoica, P.; Jansson, M. MIMO system identification: State-space and subspace approximations versus transfer function and instrumental variables. IEEE Trans. Signal Process. 2000, 48, 3087–3099. [Google Scholar] [CrossRef]
  27. Zia, A.; Kirubarajan, T.; Reilly, J.P.; Yee, D.; Punithakumar, K.; Shirani, S. An EM Algorithm for Nonlinear State Estimation with Model Uncertainties. IEEE Trans. Signal Process. 2008, 56, 921–936. [Google Scholar] [CrossRef]
  28. Overschee, P.; Moor, B. Subspace Identification for Linear Systems: Theory-Implementation-Applications; Springer US: New York, NY, USA, 1996; ISBN 9781461304654. [Google Scholar]
  29. Simon, D. Optimal State Estimation: Kalman H∞ and Nonlinear Approaches; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  30. Li, X.R.; Bar-Shalom, Y. Multiple-model estimation with variable structure. IEEE Trans. Automat. Contr. 1996, 41, 478–493. [Google Scholar]
  31. Ozkan, E.; Lindsten, F.; Fritsche, C.; Gustafsson, F. Recursive Maximum Likelihood Identification of Jump Markov Nonlinear Systems. IEEE Trans. Signal Process. 2015, 63, 754–765. [Google Scholar] [CrossRef]
  32. Stepanov, O.A.; Koshaev, D.A.; Motorin, A.V. Designing models for signals and errors of sensors in airborne gravimetry using nonlinear filtering methods. In Proceedings of the 2015 International Technical Meeting of The Institute of Navigation, Dana Point, CA, USA, 26–28 January 2015; pp. 279–284. [Google Scholar]
  33. Bar-Shalom, Y.; LI, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation: Theory Algorithms and Software; John Wiley & Sons, Inc.: New York, NY, USA, 2001; ISBN 978-0-471-41655-5. [Google Scholar]
  34. Blom, H.A.P.; Bar-Shalom, Y. Interacting multiple model algorithm for systems with Markovian switching coefficients. IEEE Trans. Automat. Contr. 1988, 33, 780–783. [Google Scholar] [CrossRef]
  35. Maybeck, P.; Hentz, K. Investigation of moving-bank multiple model adaptive algorithms. In Proceedings of the 1985 24th IEEE Conference on Decision and Control, Fort Lauderdale, FL, USA, 11–13 December 1985; pp. 1874–1881. [Google Scholar]
  36. Mallick, M.; Krishnamurthy, V.; Vo, B.-N. (Eds.) Integrated Tracking, Classification, and Sensor Management; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2013; ISBN 9781118450550. [Google Scholar]
  37. Ru, J.; Li, X.R. Variable-structure multiple-model approach to fault detection, identification, and estimation. IEEE Trans. Control Syst. Technol. 2008, 16, 1029–1038. [Google Scholar] [CrossRef]
  38. Ji, Q.; Lan, J.; Li, X.R. Nested joint fault detection, identification, estimation, and state estimation. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–8. [Google Scholar]
  39. Nikolic, J.; Furgale, P.; Melzer, A.; Siegwart, R. Maximum Likelihood Identification of Inertial Sensor Noise Model Parameters. IEEE Sens. J. 2016, 16, 163–176. [Google Scholar] [CrossRef]
  40. Stepanov, O.A.; Motorin, A.V.; Vasilyev, V.A. Identification of Sensor Errors by Using of Nonlinear Filtering. IFAC-PapersOnLine 2015, 48, 808–813. [Google Scholar] [CrossRef]
  41. Motorin, A.V.; Stepanov, O.A. Designing an error model for navigation sensors using the Bayesian approach. In Proceedings of the 2015 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), San Diego, CA, USA, 14–16 September 2015; pp. 54–58. [Google Scholar]
  42. Stepanov, O.A.; Motorin, A.V. Problem-Oriented Approach to Identification of Sensor Error Models and its Application to Navigation Data Processing. IFAC-PapersOnLine 2017, 50, 2830–2835. [Google Scholar] [CrossRef]
  43. Bucy, R.S.; Senne, K.D. Digital synthesis of non-linear filters. Automatica 1971, 7, 287–298. [Google Scholar] [CrossRef]
  44. Doucet, A.; De Freitas, N.; Gordon, N. Sequential Monte Carlo Methods in Practice; Springer-Verlag: New York, NY, USA, 2001. [Google Scholar]
  45. Zaritsky, V.S.; Svetnik, V.B.; Shimelevich, L.I. The Monte-Carlo technique in problems of optimal data processing. Autom. Remote Control 1975, 36, 2015–2022. [Google Scholar]
  46. Everitt, B.S.; Skrondal, A. The Cambridge Dictionary of Statistics, 4th ed.; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
Figure 1. (a) Unconditional probability of the identification error without the white-noise component; (b) Allan variances for the random walk with σ w = 0.04 u (1—red), σ w = 0.15 u (2—green) and the first-order Markovian process with τ = 1 min, σ m = 0.5 u (3—black), τ = 5 min, σ m = 1 u (4—blue), τ = 25 min, σ m = 1.5 u (5—yellow).
Figure 1. (a) Unconditional probability of the identification error without the white-noise component; (b) Allan variances for the random walk with σ w = 0.04 u (1—red), σ w = 0.15 u (2—green) and the first-order Markovian process with τ = 1 min, σ m = 0.5 u (3—black), τ = 5 min, σ m = 1 u (4—blue), τ = 25 min, σ m = 1.5 u (5—yellow).
Sensors 19 01997 g001
Figure 2. (a) Unconditional probability of the identification error with white-noise component of known RMS; (b) Allan variances for the sums of white noises with σ v = 0.5 u and the random walk with σ w = 0.04 u (1—red), σ w = 0.15 u (2—blue) and with the first-order Markovian process with τ = 1 min, σ m = 0.5 u (3—black), τ = 5 min, σ m = 1 u (4—blue), and τ = 25 min, σ m = 1.5 u (5—yellow).
Figure 2. (a) Unconditional probability of the identification error with white-noise component of known RMS; (b) Allan variances for the sums of white noises with σ v = 0.5 u and the random walk with σ w = 0.04 u (1—red), σ w = 0.15 u (2—blue) and with the first-order Markovian process with τ = 1 min, σ m = 0.5 u (3—black), τ = 5 min, σ m = 1 u (4—blue), and τ = 25 min, σ m = 1.5 u (5—yellow).
Sensors 19 01997 g002
Figure 3. (a) Unconditional probability of the identification error with the white-noise component of RMS σ v 1 , 2 [ 0.25   1 ] u; (b) Allan variances for the sums of white noises and the random walk with σ w = 0.04 u, σ v = 0.5 u (1—red), σ w = 0.15 u, σ v = 1 u (2—green) and with the first-order Markovian process with τ = 1 min, σ m = 0.5 u, σ v = 0.5 u (3—black), τ = 25 min, σ m = 1.5 u, σ v = 1 u (4—blue).
Figure 3. (a) Unconditional probability of the identification error with the white-noise component of RMS σ v 1 , 2 [ 0.25   1 ] u; (b) Allan variances for the sums of white noises and the random walk with σ w = 0.04 u, σ v = 0.5 u (1—red), σ w = 0.15 u, σ v = 1 u (2—green) and with the first-order Markovian process with τ = 1 min, σ m = 0.5 u, σ v = 0.5 u (3—black), τ = 25 min, σ m = 1.5 u, σ v = 1 u (4—blue).
Sensors 19 01997 g003
Figure 4. (a) Estimation error RMS for the correlation interval of the Markovian process with different initial uncertainties: 1 (red)— τ m [ 1 5 ] min, 2 (green)— τ m [ 5 25 ] min, 3 (blue)— τ m [ 25 125 ] min; (b) estimation error RSD for the correlation interval of the Markovian process with the same parameters.
Figure 4. (a) Estimation error RMS for the correlation interval of the Markovian process with different initial uncertainties: 1 (red)— τ m [ 1 5 ] min, 2 (green)— τ m [ 5 25 ] min, 3 (blue)— τ m [ 25 125 ] min; (b) estimation error RSD for the correlation interval of the Markovian process with the same parameters.
Sensors 19 01997 g004
Figure 5. Real and calculated RSD for the parameters of the Markovian process: (a) Correlation interval; (b) RMS, 1 (red)—without white-noise component, 2 (green)—with white-noise component of known RMS, 3 (blue)—with white-noise component of unknown RMS (3).
Figure 5. Real and calculated RSD for the parameters of the Markovian process: (a) Correlation interval; (b) RMS, 1 (red)—without white-noise component, 2 (green)—with white-noise component of known RMS, 3 (blue)—with white-noise component of unknown RMS (3).
Sensors 19 01997 g005
Figure 6. (a) Real and calculated RSD for RMS deviation of system noise for the random walk: 1 (red)—without white-noise component, 2 (green)—with white-noise component of known RMS, 3 (blue)—with white-noise component of unknown RMS; (b) Real (1—red) and calculated (1—red dashed) RSD for white-noise component RMS estimates with model (21) and Real (2—green) and calculated (2—green dashed) RSD for white-noise component RMS estimates with Equation (22).
Figure 6. (a) Real and calculated RSD for RMS deviation of system noise for the random walk: 1 (red)—without white-noise component, 2 (green)—with white-noise component of known RMS, 3 (blue)—with white-noise component of unknown RMS; (b) Real (1—red) and calculated (1—red dashed) RSD for white-noise component RMS estimates with model (21) and Real (2—green) and calculated (2—green dashed) RSD for white-noise component RMS estimates with Equation (22).
Sensors 19 01997 g006

Share and Cite

MDPI and ACS Style

Stepanov, O.; Motorin, A. Performance Criteria for the Identification of Inertial Sensor Error Models. Sensors 2019, 19, 1997. https://doi.org/10.3390/s19091997

AMA Style

Stepanov O, Motorin A. Performance Criteria for the Identification of Inertial Sensor Error Models. Sensors. 2019; 19(9):1997. https://doi.org/10.3390/s19091997

Chicago/Turabian Style

Stepanov, Oleg, and Andrei Motorin. 2019. "Performance Criteria for the Identification of Inertial Sensor Error Models" Sensors 19, no. 9: 1997. https://doi.org/10.3390/s19091997

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop