Next Article in Journal
Bath-Ambience—A Mechatronic System for Assisting the Caregivers of Bedridden People
Next Article in Special Issue
Multisensor Parallel Largest Ellipsoid Distributed Data Fusion with Unknown Cross-Covariances
Previous Article in Journal
Infrared Thermography Sensor for Temperature and Speed Measurement of Moving Material
Previous Article in Special Issue
Development of Data Registration and Fusion Methods for Measurement of Ultra-Precision Freeform Surfaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Fusion Estimation with Multi-Step Random Delays and Losses in Transmission

by
Raquel Caballero-Águila
1,*,
Aurora Hermoso-Carazo
2 and
Josefa Linares-Pérez
2
1
Dpto. de Estadística, Universidad de Jaén, Paraje Las Lagunillas, 23071 Jaén, Spain
2
Dpto. de Estadística, Universidad de Granada, Avda. Fuentenueva, 18071 Granada, Spain
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(5), 1151; https://doi.org/10.3390/s17051151
Submission received: 7 April 2017 / Revised: 12 May 2017 / Accepted: 15 May 2017 / Published: 18 May 2017

Abstract

:
This paper is concerned with the optimal fusion estimation problem in networked stochastic systems with bounded random delays and packet dropouts, which unavoidably occur during the data transmission in the network. The measured outputs from each sensor are perturbed by random parameter matrices and white additive noises, which are cross-correlated between the different sensors. Least-squares fusion linear estimators including filter, predictor and fixed-point smoother, as well as the corresponding estimation error covariance matrices are designed via the innovation analysis approach. The proposed recursive algorithms depend on the delay probabilities at each sampling time, but do not to need to know if a particular measurement is delayed or not. Moreover, the knowledge of the signal evolution model is not required, as the algorithms need only the first and second order moments of the processes involved. Some of the practical situations covered by the proposed system model with random parameter matrices are analyzed and the influence of the delays in the estimation accuracy are examined in a numerical example.

1. Introduction

Over the last few decades, research on the estimation problem for networked stochastic systems has gained considerable attention, due to the undeniable advantages of networked systems, whose applicability is encouraged, among other causes, by the development and advances in communication technology and the growing use of wireless networks. As it is well known, the Kalman filter provides a recursive algorithm for the optimal least-squares estimator in stochastic linear systems, assuming that the system model is exactly known and all of the measurements are instantly updated. The development of sensor networks motivates the necessity of designing new estimation algorithms that integrate the information of all the sensors to achieve a satisfactory performance; thus, using different fusion techniques, the measurements from multiple sensors are combined to obtain more accurate estimators than those obtained when a single sensor is used. In this framework, important extensions of the Kalman filter have been proposed for conventional sensor networks in which the measured outputs of the sensors always contain the actual signal contaminated by additive noises, and the transmissions are carried through perfect connections (see, e.g., [1,2,3] and references therein).
However, in a network environment, usually the standard observation models are not suitable due to the existence of network-induced uncertainties that can occur in both of the sensor measured outputs, and during the data transmission through the network. Accordingly, the consideration of appropriate observation models is vitally important to address the estimation problem in networked systems. Random failures in the transmission of measured data, together with the inaccuracy of the measurement devices, cause often the degradation of the estimator performance in networked systems. In light of these concerns, the estimation problem with one or even several network-induced uncertainties is recently attracting considerable attention, and the design of new fusion estimation algorithms has become an active research topic (see, e.g., [4,5,6,7,8,9,10,11,12,13] and references therein). In addition, some recent advances on the estimation, filtering and fusion for networked systems with network-induced phenomena can be reviewed in [14,15], where a detailed overview of this field is presented.
One of the most common network-induced uncertainties in the measured outputs of the different sensors is the presence of multiplicative noise, due to different reasons, such as interferences or intermittent sensor failures. Specifically, in situations involving random observation losses (see [5]), sensor gain degradation (see [6]), missing or fading measurements (see [13,16], respectively), the sensor observation equations include multiplicative noises. A unified framework to model these random phenomena is provided by the use of random measurement matrices in the sensor observation model. For this reason, the estimation problem in networked systems with random measurement matrices has become a fertile research subject, since this class of systems allow for covering different networked-induced random uncertainties as those mentioned above (see, e.g., [17,18,19,20,21,22,23], and references therein).
In relation to the network-induced uncertainties during the data transmission, it must be indicated that sudden changes in the environment and the unreliability of the communication network, together with the limited bandwidths of communication channels, cause unavoidable random failures during the transmission process. Generally, random communication delays and/or transmission packet dropouts are two essential issues that must be taken into account to model the measurements, which, being available after transmission, will be used for the estimation. Several estimation algorithms have been proposed in multisensor systems considering either transmission delays or packet losses (see, e.g., [24,25]) and also taking into account random delays and packet dropouts simultaneously (see, e.g., [4,23,26]). By using the state augmentation method, systems with random delays and packet dropouts can be transformed into systems with random parameter matrices (see, e.g., [8,9,10,20]). Hence, systems with random parameter measurement matrices also provide an appropriate unified context for modelling these random phenomena in the transmission.
Nevertheless, it must be indicated that the state augmentation method leads to a rise of the computational burden, due to the increase of the state dimension. Actually, in models with more than one or two-step random delays, the computational cost can be excessive and alternative ways to model and address the estimation problem in this class of systems need to be investigated. Recently, a great variety of models have been used to describe the phenomena of multi-step random delays and packet losses during the data transmission in networked systems, and fusion estimation algorithms have been proposed based on different approaches—for example, the recursive matrix equation method in [6], the measurement reorganisation approach in [27], the innovation analysis approach in [28] and the state augmentation approach in [29,30,31]. It should be noted that, in the presence of multi-step random delays and packet losses during the data transmission, many difficulties can arise in the design of the optimal estimators when the state augmentation approach is not used.
In view of the above considerations, this paper is concerned with the optimal fusion estimation problem, in the least-squares linear sense, for sensor networks featuring stochastic uncertainties in the sensor measurements, together with multi-step random delays and packet dropouts during the data transmission. The derivation of the estimation algorithms will be carried out without using the evolution model generating the signal process. The uncertainties in the measured outputs of the different sensors are described by random measurement matrices. The multi-step random delays in the transmissions are modeled by using a collection of Bernoulli sequences with known distributions and different characteristics at each sensor; the exact value of these Bernoulli variables is not required, and only the information about the probability distribution is needed. To the best of the authors’ knowledge, the optimal estimation problem (including prediction, filtering and fixed-point smoothing) has not been investigated for systems involving random measurement matrices and transmission multi-step random delays simultaneously, and, therefore, it constitutes an interesting research challenge. The main contributions of this research can be highlighted as follows: (a) even though our approach, based on covariance information, does not require the signal evolution model, the proposed algorithms are also applicable in situations based on the state-space model (see Remark 1); (b) random measurement matrices are considered in the measured outputs, thus providing a unified framework to address different network-induced phenomena (see Remark 2); (c) besides the stochastic uncertainties in the sensor measurements, simultaneous multi-step random delays and losses with different rates are considered in the data transmission; (d) unlike most papers about multi-step random delays, in which only the filtering problem is considered, we propose recursive algorithms for the prediction, filtering an fixed-point smoothing estimators under the innovation approach, which are computationally very simple and suitable for online applications; and (e) optimal estimators are obtained without using the state augmentation approach, thus reducing the computational cost in comparison with the augmentation method.
The rest of the paper is organized as follows. In Section 2, we present the sensor network and the assumptions under which the optimal linear estimation problem will be addressed. In Section 3, the observation model is rewritten in a compact form and the innovation approach to the least-squares linear estimation problem is formulated. In Section 4, recursive algorithms for the prediction, filtering and fixed-point smoothing estimators are derived. A simulation example is given in Section 5 to show the performance of the proposed estimators. Finally, some conclusions are drawn in Section 6.
Notations. 
The notations used throughout the paper are standard. R n and R m × n denote the n-dimensional Euclidean space and the set of all m × n real matrices, respectively. For a matrix A, A T and A 1 denote its transpose and inverse, respectively. The shorthand D i a g ( A 1 , , A m ) stands for a block-diagonal matrix whose diagonal matrices are A 1 , , A m . 1 n = ( 1 , , 1 ) T denotes the all-ones n × 1 -vector and I n represent the n × n -identity matrix. If the dimensions of matrices are not explicitly stated, they are assumed to be compatible with algebraic operations. The Kronecker and Hadamard product of matrices will be denoted by ⊗ and ∘, respectively. δ k , s denotes the Kronecker delta function. For any a , b R , a b is used to mean the minimum of a and b.

2. Observation Model and Preliminaries

In this paper, the optimal fusion estimation problem of multidimensional discrete-time random signals from measurements obtained by a sensor network is addressed. At each sensor, the measured outputs are perturbed by random parameter matrices and white additive noises that are cross-correlated at the same sampling time between the different sensors. The estimation is performed in a processing centre connected to all sensors, where the complete set of sensor data is combined, but due to eventual communication failures, congestion or other causes, random delays and packet dropouts are unavoidable during the transmission process. To reduce the effect of such delays and packet dropouts without overloading the network traffic, each sensor measurement transmits a fixed number of consecutive sampling times and, when several packets arrive at the same time, the receiver discards the oldest ones, so that only one measured output is processed for the estimation at each sampling time.

2.1. Signal Process

The optimal estimators will be obtained using the least-squares (LS) criterion and without requiring the evolution model generating the signal process. Actually, the proposed estimation algorithms, based on covariance information, only need the mean vectors and covariance functions of the processes involved, and the only requirement will be that the signal covariance function must be factorizable according to the following assumption:
(A1)
The n x -dimensional signal process { x k ; k 1 } has zero mean and its autocovariance function is expressed in a separable form, E [ x k x s T ] = A k B s T , s k , where A k , B s R n x × n are known matrices.
Remark 1 (on assumption (A1)).
The estimation problems based on the state-space model require new estimation algorithms when the signal evolution model is modified; therefore, the algorithms designed for stationary signals driven by x k + 1 = Φ x k + ξ k cannot be applied for non-stationary signals generated by x k + 1 = Φ k x k + ξ k , and these, in turn, cannot be used in uncertain systems where x k + 1 = ( Φ k + ϵ k Φ ^ k ) x k + ξ k . A great advantage of assumption (A1) is that it covers situations in which the signal evolution model is known, for both stationary and non-stationary signals (see, e.g., [23]). In addition, in uncertain systems with state-dependent multiplicative noise, as those considered in [6,32], the signal covariance function is factorizable, as it is shown in Section 5. Hence, assumption (A1) on the signal autocovariance function provides a unified context to deal with different situations based on the state-space model, avoiding the derivation of specific algorithms for each of them.

2.2. Multisensor Observation Model

Assuming that there are m different sensors, the measured outputs before transmission, z k ( i ) R n z , are described by the following observation model:
z k ( i ) = H k ( i ) x k + v k ( i ) , k 1 ; i = 1 , , m ,
where the measurement matrices, H k ( i ) , and the noise vectors, v k ( i ) , satisfy the following assumptions:
(A2)
H k ( i ) ; k 1 , i = 1 , , m , are independent sequences of independent random parameter matrices, whose entries have known means and second-order moments; we will denote H ¯ k ( i ) E [ H k ( i ) ] , k 1 .
(A3)
v k ( i ) ; k 1 , i = 1 , , m , are white noise sequences with zero mean and known second-order moments, satisfying E v k ( i ) v s ( j ) T = R k ( i j ) δ k , s , i , j = 1 , , m .
Remark 2 (on assumption (A2)).
Usually, in network environments, the measurements are subject to different network-induced random phenomena and new estimation algorithms must be designed to incorporate the effects of these random uncertainties. For example, in systems with stochastic sensor gain degradation or missing measurements as those considered in [6,7], respectively, or in networked systems involving stochastic multiplicative noises in the state and measurement equations (see, e.g., [31,32]), new estimation algorithms are proposed since the conditions necessary to implement the conventional ones are not met. The aforementioned systems are particular cases of systems with random measurement matrices, and, hence, assumption (A2) allows for designing a unique estimation algorithm, which is suitable to address all of these situations involving random uncertainties. In addition, based on an augmentation approach, random measurement matrices can be used to model the measured outputs of sensor networks with random delays and packet dropouts (see, e.g., [8,9,10,20]). Therefore, assumption (A2) provides a unified framework to deal with a great variety of network-induced random phenomena as those mentioned above.

2.3. Measurement Model with Transmission Random Delays and Packet Losses

Assuming that the maximum time delay is D, the measured output of the i-th sensor at time r, z r ( i ) , is transmitted during the sampling times r , r + 1 , , r + D , but, at each sampling time k > D , only one of the measurements z k D ( i ) , , z k ( i ) is processed. Consequently, at any time k > D , the measurement processed can either arrive on time or be delayed by d = 1 , , D sampling periods, while at any time k D , the measurement processed can be delayed only by d = 1 , , k 1 sampling periods, since only z 1 ( i ) , , z k ( i ) are available. Assuming, moreover, that the transmissions are perturbed by additive noises, the measurements received at the processing centre, impaired by random delays and packet losses, can be described by the following model:
y k ( i ) = d = 0 ( k 1 ) D γ d , k ( i ) z k d ( i ) + w k ( i ) , k 1 ,
where the following assumptions on the random variables modelling the delays, γ d , k ( i ) , and the transmision noise, w k ( i ) , are required:
(A4)
For each d = 0 , 1 , , D , γ d , k ( i ) ; k > d , i = 1 , , m , are independent sequences of independent Bernoulli random variables with P [ γ d , k ( i ) = 1 ] = γ ¯ d , k ( i ) and d = 0 ( k 1 ) D γ d , k ( i ) 1 , k 1 .
(A5)
w k ( i ) ; k 1 , i = 1 , , m , are white noise sequences with zero mean and known second-order moments, satisfying E w k ( i ) w s ( j ) T = Q k ( i j ) δ k , s , i , j = 1 , , m .
Remark 3 (on assumption (A4)).
For i = 1 , , m , when γ 0 , k ( i ) = 1 , the transmission of the i-th sensor is perfect and neither delay nor loss occurs at time k; that is, with probability γ ¯ 0 , k ( i ) , the k-th measurement of the i-th sensor is received and processed on time. Since d = 0 ( k 1 ) D γ d , k ( i ) 1 , if γ 0 , k ( i ) = 0 , there then exists at most one d = 1 , , D , such that γ d , k ( i ) = 1 . If there exists d such that γ d , k ( i ) = 1 (which occurs with probability γ ¯ d , k ( i ) ), then the measurement is delayed by d sampling periods. Otherwise, γ d , k ( i ) = 0 for all d, meaning that the measurement gets lost during the transmission at time k with probability 1 d = 0 ( k 1 ) D γ ¯ d , k ( i ) .
Finally, the following independence hypothesis is assumed:
(A6)
For i = 1 , , m and d = 0 , 1 , , D , the processes { x k ; k 1 } , { H k ( i ) ; k 1 } , { v k ( i ) ; k 1 } , { w k ( i ) ; k 1 } and γ d , k ( i ) ; k > d are mutually independent.

3. Problem Statement

Given the observation Equations (1) and (2) with random measurement matrices and transmission random delays and packet dropouts, our purpose is to find the LS linear estimator, x ^ k / L , of the signal x k based on the observations from the different sensors { y 1 ( i ) , , y L ( i ) , i = 1 , , m } . Specifically, our aim is to obtain recursive algorithms for the predictor ( L < k ), filter ( L = k ) and fixed-point smoother (k fixed and L > k ).

3.1. Stacked Observation Model

Since the measurements coming from the different sensors are all gathered and jointly processed at each sampling time k, we will consider the vector constituted by the meaurements from all sensors, y k = y k ( 1 ) T , , y k ( m ) T T . More specifically, the observation Equations (1) and (2) of all sensors are combined, yielding the following stacked observation model:
z k = H k x k + v k , k 1 , y k = d = 0 ( k 1 ) D Γ d , k z k d + w k , k 1 ,
where z k = z k ( 1 ) T , , z k ( m ) T T , H k = H k ( 1 ) T , , H k ( m ) T T , v k = v k ( 1 ) T , , v k ( m ) T T , w k = w k ( 1 ) T , , w k ( m ) T T and Γ d , k = D i a g γ d , k ( 1 ) , , γ d , k ( m ) I n z .
Hence, the problem is to obtain the LS linear estimator of the signal, x k , based on the measurements y 1 , , y L , given in the observation Equation (3). Next, we present the statistical properties of the processes involved in Equation (3), from which the LS estimation algorithms of the signal will be derived; these properties are easily inferred from the assumptions (A1)–(A6).
(P1)
H k ; k 1 is a sequence of independent random parameter matrices with known means, H ¯ k E [ H k ] = H ¯ k ( 1 ) T , , H ¯ k ( m ) T T , and
E [ H k x k x s T H s T ] = E [ H k A k B s T H s T ] = E [ H k ( i ) A k B s T H s ( j ) T ] i , j = 1 , m , s k ,
where E H k ( i ) A k B s T H s ( j ) T = H ¯ k ( i ) A k B s T H ¯ s ( j ) T , for j i or s k , and the entries of E H k ( i ) A k B k T H k ( i ) T are computed as follows:
E H k ( i ) A k B k T H k ( i ) T p q = a = 1 n x b = 1 n x E [ h p a ( i ) ( k ) h q b ( i ) ( k ) ] ( A k B k T ) a b , p , q = 1 , , n z ,
where h p q ( i ) ( k ) denotes the ( p , q ) entry of the matrix H k ( i ) .
(P2)
The noises v k ; k 1 and w k ; k 1 are zero-mean sequences with known second-order moments given by the matrices R k R k ( i j ) i , j = 1 , , m and Q k Q k ( i j ) i , j = 1 , , m .
(P3)
{ Γ d , k ; k > d } , d = 0 , 1 , , D , are sequences of independent random matrices with known means, Γ ¯ d , k E [ Γ d , k ] = D i a g γ ¯ d , k ( 1 ) , , γ ¯ d , k ( m ) I n z , and if we denote γ d , k = γ d , k ( 1 ) , , γ d , k ( m ) T 1 n z and γ ¯ d , k = E [ γ d , k ] , the covariance matrices Σ d , d γ k E [ ( γ d , k γ ¯ d , k ) ( γ d , k γ ¯ d , k ) T ] , for d , d = 0 , 1 , , D , are also known matrices. Specifically,
Σ d , d γ k = D i a g C o v γ d , k ( 1 ) , γ d , k ( 1 ) , , C o v γ d , k ( m ) , γ d , k ( m ) 1 n z 1 n z T ,
with C o v γ d , k ( i ) , γ d , k ( i ) = γ ¯ d , k ( i ) ( 1 γ ¯ d , k ( i ) ) , d = d , γ ¯ d , k ( i ) γ ¯ d , k ( i ) , d d . Moreover, for any deterministic matrix S, the Hadamard product properties guarantee that
E Γ d , k Γ ¯ d , k S Γ d , k Γ ¯ d , k = Σ d , d γ k S .
(P4)
For d = 0 , 1 , , D , the signal, { x k ; k 1 } , and the processes { H k ; k 1 } , { v k ; k 1 } , { w k ; k 1 } and { Γ d , k ; k > d } are mutually independent.
Remark 4 (on the observation covariance matrices).
From the previous properties, it is clear that the observation process { z k ; k 1 } is a zero-mean sequence whose covariance function, Σ k , s z E [ z k z s T ] , is obtained by the following expression:
Σ k , s z = E H k A k B s T H s T + R k δ k , s , s k ,
where E H k A k B s T H s T and R k are calculated as it is indicated in properties (P1) and (P2), respectively.

3.2. Innovation Approach to the LS Linear Estimation Problem

The proposed covariance-based recursive algorithms for the LS linear prediction, filtering and fixed-point smoothing estimators will be derived by an innovation approach. This approach consists of transforming the observation process { y k ; k 1 } into an equivalent one of orthogonal vectors called an innovation process, which will be denoted { μ k ; k 1 } and defined by μ k = y k y ^ k / k 1 , where y ^ k / k 1 is the orthogonal projection of y k onto the linear space spanned by μ 1 , , μ k 1 . Since both processes span the same linear subspace, the LS linear estimator of any random vector α k based on the observations y 1 , , y N , denoted by α ^ k / N , is equal to that based on the innovations μ 1 , , μ N , and, denoting Π h = E [ μ h μ h T ] , the following general expression for the LS linear estimators of α k is obtained
α ^ k / N = h = 1 N E α k μ h T Π h 1 μ h .
Hence, to obtain the signal estimators, it is necessary to find an explicit formula beforehand for the innovations and their covariance matrices.
Innovation μ L and Covariance Matrix Π L . Applying orthogonal projections in Equation (3), it is clear that the innovation μ L is given by
μ L = y L d = 0 ( L 1 ) D Γ ¯ d , L z ^ L d / L 1 , L 2 ; μ 1 = y 1 ,
so it is necessary to obtain the one-stage predictor z ^ L / L 1 and the estimators z ^ L d / L 1 , for d = 1 , , ( L 1 ) D , of the observation process.
In order to obtain the covariance matrix Π L , we use Equation (3) to express the innovations as:
μ L = d = 0 ( L 1 ) D ( Γ d , L Γ ¯ d , L ) z L d + Γ ¯ d , L ( z L d z ^ L d / L 1 ) + w L , L 2
and, taking into account that
E ( Γ d , L Γ ¯ d , L ) z L d ( z L d z ^ L d / L 1 ) T Γ ¯ d , L = 0 , d , d ,
we have
Π L = d , d = 0 ( L 1 ) D Σ d , d γ L Σ L d , L d z + Γ ¯ d , L P L d , L d / L 1 z Γ ¯ d , L + Q L , L 2 ; Π 1 = Σ 0 , 0 γ 1 + γ ¯ 0 , 1 γ ¯ 0 , 1 T Σ 1 , 1 z + Q 1 ,
where the matrices Σ d , d γ L and Σ L d , L d z are given in the Equations (4) and (5), respectively, and P L d , L d / L 1 z E [ ( z L d z ^ L d / L 1 ) ( z L d z ^ L d / L 1 ) T ] .

4. Least-Squares Linear Signal Estimators

In this section, we derive recursive algorithms for the LS linear estimators, x ^ k / L , k 1 , of the signal x k based on the observations y 1 , , y L given in Equation (3); namely, a prediction and filtering algorithm ( L k ) and a fixed-point smoothing algorithm (k fixed and L > k ) are designed.

4.1. Signal Predictor and Filter x ^ k / L , L k

From the general expression given in Equation (6), to obtain the LS linear estimators x ^ k / L = h = 1 L E x k μ h T Π h 1 μ h , L k , it is necessary to calculate the coefficients
X k , h E x k μ h T = E x k y h T E x k y ^ h / h 1 T , h k .
  • On the one hand, using Equation (3) together with the independence hypotheses and assumption (A1) on the signal covariance factorization, it is clear that
    E x k y h T = d = 0 ( h 1 ) D E [ x k ( H h d x h d + v h d ) T ] Γ ¯ d , h = A k d = 0 ( h 1 ) D B h d T H ¯ h d T Γ ¯ d , h , h k .
  • On the other hand, since y ^ h / h 1 = d = 0 ( h 1 ) D Γ ¯ d , h z ^ h d / h 1 , h 2 , and taking into account that, from Equation (6), z ^ h d / h 1 = j = 1 h 1 Z h d , j Π j 1 μ j with Z h d , j = E [ z h d μ j T ] , the following identity holds
    E x k y ^ h / h 1 T = d = 0 ( h 1 ) D j = 1 h 1 X k , j Π j 1 Z h d , j T Γ ¯ d , h .
Therefore, it is easy to check that X k , h = A k E h , 1 h k , where E h is a function satisfying that
E h = d = 0 ( h 1 ) D B h d T H ¯ h d T Γ ¯ d , h d = 0 ( h 1 ) D j = 1 h 1 E j Π j 1 Z h d , j T Γ ¯ d , h , h 2 , E 1 = B 1 T H ¯ 1 T Γ ¯ 0 , 1 .
Hence, it is clear that the signal prediction and filtering estimators can be expressed as
x ^ k / L = A k e L , L k , k 1 ,
where the vectors e L are defined by e L = h = 1 L E h Π h 1 μ h , for L 1 , with e 0 = 0 , thus obeying the recursive relation
e L = e L 1 + E L Π L 1 μ L , L 1 ; e 0 = 0 .
Matrices E L . Taking into account the above relation, an expression for E L , L 1 , must be derived. For this purpose, Equation (10) is rewritten for h = L as
E L = d = 0 ( L 1 ) D B L d T H ¯ L d T Γ ¯ d , L d = 0 ( L 1 ) D j = 1 L 1 E j Π j 1 Z L d , j T Γ ¯ d , L , L 2 ,
and we examine the cases d = 0 and d 1 separately:
For d = 0 , using Equation (3), it holds that Z L , j = H ¯ L X L , j = H ¯ L A L E j , for j < L , and, by denoting Σ L e = h = 1 L E h Π h 1 E h , L 1 , we obtain that
j = 1 L 1 E j Π j 1 Z L , j T Γ ¯ 0 , L = j = 1 L 1 E j Π j 1 E j T A L T H ¯ L T Γ ¯ 0 , L = Σ L 1 e A L T H ¯ L T Γ ¯ 0 , L .
For d 1 , since Z L d , j = H ¯ L d A L d E j , for j < L d , we can see that
j = 1 L 1 E j Π j 1 Z L d , j T Γ ¯ d , L = Σ L d 1 e A L d T H ¯ L d T Γ ¯ d , L + j = L d L 1 E j Π j 1 Z L d , j T Γ ¯ d , L .
By substituting the above sums in E L , it is deduced that
E L = d = 0 ( L 1 ) D B L d A L d Σ L d 1 e T H ¯ L d T Γ ¯ d , L d = 1 ( L 1 ) D j = L d L 1 E j Π j 1 Z L d , j T Γ ¯ d , L , L 2 ; E 1 = B 1 T H ¯ 1 T Γ ¯ 0 , 1 ,
where the matrices Z L d , j , j L d , will be obtained in the next subsection, as they correspond to the observation smoothing estimators, and the matrices Σ L e are recursively obtained by
Σ L e = Σ L 1 e + E L Π L 1 E L T , L 1 ; Σ 0 e = 0 .
Finally, from assumption (A1) and since the estimation errors are orthogonal to the estimators, we have that the error covariance matrices, P k / L x E [ ( x k x ^ k / L ) ( x k x ^ k / L ) T ] , are given by
P k / L x = A k B k A k Σ L e T , L k , k 1 .

4.2. Estimators of the Observations z ^ k / L , k 1

As it has been already indicated, the Equation (7) require obtaining the observation estimators (predictor, filter and smoother). From the general expression for the estimators given in Equation (6), we have that z ^ k / L = j = 1 L Z k , j Π j 1 μ j , with Z k , j = E [ z k μ j T ] . Next, recursive expressions will be derived separately for L < k (predictors) and L k (filter and smoothers).
Observation Prediction Estimators. Since Z k , j = H ¯ k A k E j , for j < k , we have that the prediction estimators of the observations are given by
z ^ k / L = H ¯ k A L e L , L < k , k 1 .
Observation Filtering and Fixed-Point Smoothing Estimators. Clearly, the filter and fixed-point smoothers of the observations are obtained by the following recursive expression:
z ^ k / L = z ^ k / L 1 + Z k , L Π L 1 μ L , L k , k 1 ,
with initial condition given by the one-stage predictor z ^ k / k 1 = H ¯ k A k e k 1 .
Hence, the matrices Z k , L must be calculated for L k . Since the innovation is a white process, E z ^ k / L 1 μ L T = 0 and hence Z k , L = E [ z k μ L T ] = E ( z k z ^ k / L 1 ) μ L T . Now using Equation (8) for μ L and, taking into account that E ( z k z ^ k / L 1 ) z L d T ( Γ d , L Γ ¯ d , L ) = 0 , d , we have
Z k , L = d = 0 ( L 1 ) D P k , L d / L 1 z Γ ¯ d , L , L k ,
where P k , L d / L 1 z E [ ( z k z ^ k / L 1 ) ( z L d z ^ L d / L 1 ) T ] .
Consequently, the error covariance matrices P k , h / m z = E [ ( z k z ^ k / m ) ( z h z ^ h / m ) T ] must be derived, for which the following two cases are analyzed separately:
*
For m k h , using Equation (17) and taking into account that Z k , m = E ( z k z ^ k / m 1 ) μ m T , it is easy to see that
P k , h / m z = P k , h / m 1 z Z k , m Π m 1 Z h , m T , m k h .
*
For m < h k , using Equation (16), assumption (A1) and the orthogonality between the estimation errors and the estimators, we obtain
P k , h / m z = A k B h T H ¯ k A k Σ m e A h T H ¯ h T , m < h k .

4.3. Signal Fixed-Point Smoother x ^ k / L , L > k

Starting with the filter, x ^ k / k , and the filtering error covariance matrix, P k / k x , it is clear that the signal fixed-point smoother x ^ k / L , L > k , and the corresponding error covariance matrix, P k / L x E [ ( x k x ^ k / L ) ( x k x ^ k / L ) T ] , are obtained by
x ^ k / L = x ^ k / L 1 + X k , L Π L 1 μ L , L > k , k 1 , P k / L x = P k / L 1 x X k , L Π L 1 X k , L T , L > k , k 1 .
An analogous reasoning to that of Equation (18) leads to the following expression for the matrices X k , L :
X k , L = d = 0 ( L 1 ) D P k , L d / L 1 x z Γ ¯ d , L , L > k ,
where P k , L d / L 1 x z E [ ( x k x ^ k / L 1 ) ( z L d z ^ L d / L 1 ) T ] .
The derivation of the error cross-covariance matrices P k , h / m x z = E [ ( x k x ^ k / m ) ( z h z ^ h / m ) T ] is similar to that of the matrices P k , h / m z , and they are given by
P k , h / m x z = P k , h / m 1 x z X k , m Π m 1 Z h , m T , m k h , P k , h / m x z = A k ( B h A h Σ m e ) T H h T , m < h k , P k , h / m x z = ( B k A k Σ m e ) A h T H h T , m < k h ,
where X k , m = A k E m , for m k , and Z h , m = H ¯ h A m E m , for m < h , and otherwise these matrices are given by Equations (22) and (18), respectively.

4.4. Recursive Algorithms: Computational Procedure

The computational procedure of the proposed prediction, filtering and fixed-point smoothing algorithms can be summarized as follows:
(1)
Covariance Matrices. The covariance matrices Σ d , d γ k and Σ k z are obtained by Equations (4) and (5), respectively; these matrices only depend on the system model information, so they can be calculated offline before the observed data packets are available.
(2)
LS Linear Prediction and Filtering Recursive Algorithm. At the sampling time k, once the ( k 1 ) -th iteration is finished and E k 1 , Π k 1 , Σ k 1 e μ k 1 and e k 1 are known, the proposed prediction and filtering algorithm operates as follows:
(2a)
Compute Z k , k 1 = H ¯ k A k E k 1 and Z k d , k 1 , for d = 1 , , ( k 1 ) D , by Equation (18). From these matrices, we obtain the observation estimators z ^ k d / k 1 , for d = 0 , 1 , , ( k 1 ) D , by Equation (19) and (20), and the observation error covariance matrices P k d , k d / k 1 z , for d , d = 0 , 1 , , ( k 1 ) D , by Equation (19) and (20).
(2b)
Compute E k by Equation (13) and use P k d , k d / k 1 z to obtain the innovation covariance matrix Π k by Equation (9). Then, Σ k e is obtained by Equation (14) and, from it, the prediction and filtering error covariance matrices, P k / k s x and P k / k x , respectively, are obtained by Equation (15).
(2c)
When the new measurement y k is available, the innovation μ k is computed by Equation (7) using z ^ k d / k 1 , for d = 0 , 1 , , ( k 1 ) D , and, from the innovation, e k is obtained by Equation (12). Then, the predictors, x ^ k / k s and the filter, x ^ k / k are computed by Equation (11).
(3)
LS linear fixed-point smoothing recursive algorithm. Once the filter, x ^ k / k , and the filtering error covariance matrix, P k / k x are available, the proposed smoothing estimators and the corresponding error covariance matrix are obtained as follows:
For L = k + 1 , k + 2 , , compute the error cross-covariance matrices P k , L d / L 1 x z , for d = 0 , 1 , , ( k 1 ) D , using Equation (23) and, from these matrices, X k , L is derived by Equation (22); then, the smoothers x ^ k / L and their error covariance matrices P k / L x are obtained from Equation (21).

5. Computer Simulation Results

In this section, a numerical example is presented with the following purposes: (a) to show that, although the current covariance-based estimation algorithms do not require the evolution model generating the signal process, they are also applicable to the conventional formulation using the state-space model, even in the presence of state-dependent multiplicative noise; (b) to illustrate some kinds of uncertainties which can be covered by the current model with random measurement matrices; and (c) to analyze how the estimation accuracy of the proposed algorithms is influenced by the sensor uncertainties and the random delays in the transmissions.
Signal Evolution Model with State-Dependent Multiplicative Noise. Consider a scalar signal x k ; k 0 whose evolution is given by the following model with multiplicative and additive noises:
x k = ( 0 . 9 + 0 . 01 ϵ k 1 ) x k 1 + ξ k 1 , k 1 ,
where x 0 is a standard Gaussian variable and { ϵ k ; k 0 } , { ξ k ; k 0 } are zero-mean Gaussian white processes with unit variance. Assuming that x 0 , { ϵ k ; k 0 } and { ξ k ; k 0 } are mutually independent, the signal covariance function is given by E [ x k x s ] = 0.9 k s D s , s k , where D s = E [ x s 2 ] is recursively obtained by D s = 0.8101 D s 1 + 1 , for s 1 , with D 0 = 1 ; hence, the signal process satisfies assumption (A1) taking, for example, A k = 0.9 k and B s = 0.9 s D s .
Sensor Measured Outputs. As in [22], let us consider scalar measurements provided by four sensors with different types of uncertainty: continuous and discrete gain degradation in sensors 1 and 2, respectively, missing measurements in sensor 3, and both missing measurements and multiplicative noise in sensor 4. These uncertainties can be described in a unified way by the current model with random measurement matrices; specifically, the measured outputs are described according to Equation (1):
z k ( i ) = H k ( i ) x k + v k ( i ) , k 1 , i = 1 , 2 , 3 , 4 ,
with the following characteristics:
  • For i = 1 , 2 , 3 , H k ( i ) = C ( i ) θ k ( i ) and H k ( 4 ) = C ( 4 ) + C ( 4 ) ρ k θ k ( 4 ) , where C ( 1 ) = C ( 3 ) = 0.8 , C ( 2 ) = C ( 4 ) = 0.75 , C ( 4 ) = 0.95 , and { ρ k ; k 1 } is a zero-mean Gaussian white process with unit variance. The sequences { ρ k ; k 1 } and { θ k ( i ) ; k 1 } , i = 1 , 2 , 3 , 4 , are mutually independent, and { θ k ( i ) ; k 1 } , i = 1 , 2 , 3 , 4 , are white processes with the following time-invariant probability distributions:
    θ k ( 1 ) is uniformly distributed over the interval [ 0.1 , 0.9 ] ;
    P [ θ k ( 2 ) = 0 ] = 0.3 , P [ θ k ( 2 ) = 0.5 ] = 0.3 P [ θ k ( 2 ) = 1 ] = 0.4 ;
    For i = 3 , 4 , θ k ( i ) are Bernoulli random variables with the same time-invariant probabilities in both sensors P [ θ k ( i ) = 1 ] = p .
  • The additive noises are defined by v k ( i ) = c i η k v , i = 1 , 2 , 3 , 4 , where c 1 = 0.5 , c 2 = c 3 = 0.75 , c 4 = 1 , and { η k v ; k 1 } is a zero-mean Gaussian white process with unit variance. Clearly, the additive noises { v k ( i ) ; k 1 } , i = 1 , 2 , 3 , 4 , are correlated at any time, with R k ( i j ) = c i c j , i , j = 1 , 2 , 3 , 4 .
Observations with Bounded Random Delays and Packet Dropouts. Next, according to the theoretical observation model, let us suppose that bounded random measurement delays and packet dropouts, with different delay probabilities, exist in the data transmission. Specifically, assuming that the largest delay is D = 3 , let us consider the observation Equation (2):
y k ( i ) = d = 0 ( k 1 ) 3 γ d , k ( i ) z k d ( i ) + w k ( i ) , k 1 ,
where, for i = 1 , 2 , 3 , 4 and d = 0 , 1 , 2 , 3 , { γ d , k ( i ) ; k > d } , are sequences of independent Bernoulli variables with the same time-invariant delay probabilities for the four sensors γ ¯ d , k ( i ) = γ ¯ d , where d = 0 3 γ ¯ d 1 . Hence, the packet dropout probability is 1 d = 0 3 γ ¯ d . The transmission noise is defined by w k ( i ) = c i η k w , i = 1 , 2 , 3 , 4 , where { η k w ; k 1 } is a zero-mean Gaussian white process with unit variance.
Finally, in order to apply the proposed algorithms, and according to (A5), we will assume that all of the processes involved in the observation equations are mutually independent.
To illustrate the feasibility and effectiveness of the proposed algorithms, they were implemented in MATLAB (R2011b 7.13.0.564, The Mathworks, Natick, MA, USA) and one hundred iterations of the prediction, filtering and fixed-point smoothing algorithms have been performed. In order to analyze the effect of the network-induced uncertainties on the estimation accuracy, different values of the probabilities p of the Bernoulli random variables that model the uncertainties of the third and fourth sensors, and several values of the delay probabilities γ ¯ d , d = 0 , 1 , 2 , 3 , have been considered.
Performance of the Prediction, Filtering and Fixed-Point Smoothing Estimators. Considering the values p = 0.5 , γ ¯ 0 = 0.6 (packet arrival probability) and γ ¯ d = 1 γ ¯ 0 4 , d = 1 , 2 , 3 (delay probabilities), Figure 1 displays a simulated trajectory along with the prediction, filtering and smoothing estimations, showing a satisfactory and efficient tracking performance of the proposed estimators. Figure 2 shows the error variances of the predictors z ^ k / k 2 and z ^ k / k 1 , the filter z ^ k / k and the smoothers z ^ k / k + 1 and z ^ k / k + 2 . Analogously to what happens for non-delayed observations, the performance of the estimators becomes better as more observations are used; that is, the error variances of the smoothing estimators are less than the filtering ones which, in turn, are less than those of the predictors. Hence, the estimation accuracy of the smoothers is superior to that of the filter and predictors and improves as the number of iterations in the fixed-point smoothing algorithm increases. The performance of the proposed filter has also been evaluated in comparison with the standard Kalman filter; for this purpose, the filtering mean square error (MSE) at each sampling time was calculated by considering one thousand independent simulations and one hundred iterations of each filter. The results of this comparison are displayed in Figure 3, which shows that the proposed filter performs better than the Kalman filter, a fact that was expected since the latter does not take into account either the uncertainties in the measured outputs or the delays and losses during transmission.
Influence of the Missing Measurements. To analyze the sensitivity of the estimation performance on the effect of the missing measurements phenomenon in the third and fourth sensors, the error variances are calculated for different values of the probability p. Specifically, considering again γ ¯ 0 = 0.6 and γ ¯ d = 0.1 , d = 1 , 2 , 3 , Figure 4 displays the prediction, filtering and fixed-point smoothing error variances for the values p = 0.5 to p = 0.9. This Figure shows that, as p increases, the estimation error variances become smaller and, hence, as it was expected, the performance of the estimators improves as the probability of missing measurements 1 p decreases.
Influence of the Transmission Random Delays and Packet Dropouts. Considering a fixed value of the probability p, namely, p = 0.5 , in order to show how the estimation accuracy is influenced by the transmission random delays and packet dropouts, the prediction, filtering and smoothing error variances are displayed in Figure 5 when γ ¯ 0 is varied from 0.1 to 0.9, considering again γ ¯ d = 1 γ ¯ 0 4 , d = 1 , 2 , 3 . Since the behaviour of the error variances is analogous from a certain iteration on, only the results at the iteration k = 50 are shown in Figure 5. From this figure, we see that the performance of the estimators (predictor, filter and smoother) is indeed influenced by the transmission delay and packet dropout probabilities and, as it was expected, it is confirmed that the error variances become smaller, and hence the performance of the estimators improves, as the packet arrival probability γ ¯ 0 increases. Moreover, for the filter and the smoothers, this improvement is more significant than for the predictors. In addition, as it was deduced from Figure 2, it is observed that, for all the values of γ ¯ 0 , the performance of the estimators is better as more observations are used, and this improvement is more significant as γ ¯ 0 increases.
Finally, it must be remarked that analogous conclusions are deduced for other values of the probabilities p and γ ¯ d , d = 0 , 1 , 2 , 3 , and also when such probabilities are different at the different sensors.

6. Conclusions

This paper makes valuable contributions to the optimal fusion estimation problem in networked stochastic systems with random parameter matrices, when multi-step delays or even packet dropouts occur randomly during the data transmission. By an innovation approach, recursive prediction, filtering and fixed-point smoothing algorithms have been designed, which are easily implementable and do not require the signal evolution model, but only the mean and covariance functions of the system processes.
Unlike other estimation algorithms proposed in the literature, where the estimators are restricted to obey a particular structure, in this paper, recursive optimal estimation algorithms are designed without requiring a particular structure on the estimators, but just using the LS optimality criterion. Another advantage is that the current approach does not resort to the augmentation technique and, consequently, the dimension of the designed estimators is the same as that of the original signal, thus reducing the computational burden in the processing centre.

Acknowledgments

This research is supported by the “Ministerio de Economía y Competitividad” and “Fondo Europeo de Desarrollo Regional” FEDER (Grant No. MTM2014-52291-P).

Author Contributions

All of the authors contributed equally to this work. Raquel Caballero-Águila, Aurora Hermoso-Carazo and Josefa Linares-Pérez provided original ideas for the proposed model and collaborated in the derivation of the estimation algorithms; they participated equally in the design and analysis of the simulation results; and the paper was also written and reviewed cooperatively.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ran, C.; Deng, Z. Self-tuning weighted measurement fusion Kalman filtering algorithm. Comput. Stat. Data Anal. 2012, 56, 2112–2128. [Google Scholar] [CrossRef]
  2. Feng, J.; Zeng, M. Optimal distributed Kalman filtering fusion for a linear dynamic system with cross-correlated noises. Int. J. Syst. Sci. 2012, 43, 385–398. [Google Scholar] [CrossRef]
  3. Yan, L.; Li, X.; Xia, Y.; Fu, M. Optimal sequential and distributed fusion for state estimation in cross-correlated noise. Automatica 2013, 49, 3607–3612. [Google Scholar] [CrossRef]
  4. Ma, J.; Sun, S. Centralized fusion estimators for multisensor systems with random sensor delays, multiple packet dropouts and uncertain observations. IEEE Sens. J. 2013, 13, 1228–1235. [Google Scholar] [CrossRef]
  5. Gao, S.; Chen, P. Suboptimal filtering of networked discrete-time systems with random observation losses. Math. Probl. Eng. 2014, 2014, 151836. [Google Scholar] [CrossRef]
  6. Liu, Y.; He, X.; Wang, Z.; Zhou, D. Optimal filtering for networked systems with stochastic sensor gain degradation. Automatica 2014, 50, 1521–1525. [Google Scholar] [CrossRef]
  7. Chen, B.; Zhang, W.; Yu, L. Distributed fusion estimation with missing measurements, random transmission delays and packet dropouts. IEEE Trans. Autom. Control 2014, 59, 1961–1967. [Google Scholar] [CrossRef]
  8. Li, N.; Sun, S.; Ma, J. Multi-sensor distributed fusion filtering for networked systems with different delay and loss rates. Digit. Signal Process. 2014, 34, 29–38. [Google Scholar] [CrossRef]
  9. Wang, S.; Fang, H.; Tian, X. Recursive estimation for nonlinear stochastic systems with multi-step transmission delays, multiple packet dropouts and correlated noises. Signal Process. 2015, 115, 164–175. [Google Scholar] [CrossRef]
  10. Chen, D.; Yu, Y.; Xu, L.; Liu, X. Kalman filtering for discrete stochastic systems with multiplicative noises and random two-step sensor delays. Discret. Dyn. Nat. Soc. 2015, 2015, 809734. [Google Scholar] [CrossRef]
  11. García-Ligero, M.J.; Hermoso-Carazo, A.; Linares-Pérez, J. Distributed fusion estimation in networked systems with uncertain observations and markovian random delays. Signal Process. 2015, 106, 114–122. [Google Scholar] [CrossRef]
  12. Gao, S.; Chen, P.; Huang, D.; Niu, Q. Stability analysis of multi-sensor Kalman filtering over lossy networks. Sensors 2016, 16, 566. [Google Scholar] [CrossRef] [PubMed]
  13. Lin, H.; Sun, S. State estimation for a class of non-uniform sampling systems with missing measurements. Sensors 2016, 16, 1155. [Google Scholar] [CrossRef] [PubMed]
  14. Hu, J.; Wang, Z.; Chen, D.; Alsaadi, F.E. Estimation, filtering and fusion for networked systems with network-induced phenomena: New progress and prospects. Inf. Fusion 2016, 31, 65–75. [Google Scholar] [CrossRef]
  15. Sun, S.; Lin, H.; Ma, J.; Li, X. Multi-sensor distributed fusion estimation with applications in networked systems: A review paper. Inf. Fusion 2017, 38, 122–134. [Google Scholar] [CrossRef]
  16. Li, W.; Jia, Y.; Du, J. Distributed filtering for discrete-time linear systems with fading measurements and time-correlated noise. Digit. Signal Process. 2017, 60, 211–219. [Google Scholar] [CrossRef]
  17. Luo, Y.; Zhu, Y.; Luo, D.; Zhou, J.; Song, E.; Wang, D. Globally optimal multisensor distributed random parameter matrices Kalman filtering fusion with applications. Sensors 2008, 8, 8086–8103. [Google Scholar] [CrossRef] [PubMed]
  18. Shen, X.J.; Luo, Y.T.; Zhu, Y.M.; Song, E.B. Globally optimal distributed Kalman filtering fusion. Sci. China Inf. Sci. 2012, 55, 512–529. [Google Scholar] [CrossRef]
  19. Hu, J.; Wang, Z.; Gao, H. Recursive filtering with random parameter matrices, multiple fading measurements and correlated noises. Automatica 2013, 49, 3440–3448. [Google Scholar] [CrossRef]
  20. Linares-Pérez, J.; Caballero-Águila, R.; García-Garrido, I. Optimal linear filter design for systems with correlation in the measurement matrices and noises: Recursive algorithm and applications. Int. J. Syst. Sci. 2014, 45, 1548–1562. [Google Scholar] [CrossRef]
  21. Yang, Y.; Liang, Y.; Pan, Q.; Qin, Y.; Yang, F. Distributed fusion estimation with square-root array implementation for Markovian jump linear systems with random parameter matrices and cross-correlated noises. Inf. Sci. 2016, 370–371, 446–462. [Google Scholar] [CrossRef]
  22. Caballero-Águila, R.; Hermoso-Carazo, A.; Linares-Pérez, J. Networked fusion filtering from outputs with stochastic uncertainties and correlated random transmission delays. Sensors 2016, 16, 847. [Google Scholar] [CrossRef] [PubMed]
  23. Caballero-Águila, R.; Hermoso-Carazo, A.; Linares-Pérez, J. Fusion estimation using measured outputs with random parameter matrices subject to random delays and packet dropouts. Signal Process. 2016, 127, 12–23. [Google Scholar] [CrossRef]
  24. Feng, J.; Zeng, M. Descriptor recursive estimation for multiple sensors with different delay rates. Int. J. Control 2011, 84, 584–596. [Google Scholar] [CrossRef]
  25. Caballero-Águila, R.; Hermoso-Carazo, A.; Linares-Pérez, J. Least-squares linear-estimators using measurements transmitted by different sensors with packet dropouts. Digit. Signal Process. 2012, 22, 1118–1125. [Google Scholar] [CrossRef]
  26. Ma, J.; Sun, S. Distributed fusion filter for networked stochastic uncertain systems with transmission delays and packet dropouts. Signal Process. 2017, 130, 268–278. [Google Scholar] [CrossRef]
  27. Wang, S.; Fang, H.; Liu, X. Distributed state estimation for stochastic non-linear systems with random delays and packet dropouts. IET Control Theory Appl. 2015, 9, 2657–2665. [Google Scholar] [CrossRef]
  28. Sun, S. Linear minimum variance estimators for systems with bounded random measurement delays and packet dropouts. Signal Process. 2009, 89, 1457–1466. [Google Scholar] [CrossRef]
  29. Sun, S. Optimal linear filters for discrete-time systems with randomly delayed and lost measurements with/without time stamps. IEEE Trans. Autom. Control 2013, 58, 1551–1556. [Google Scholar] [CrossRef]
  30. Sun, S.; Ma, J. Linear estimation for networked control systems with random transmission delays and packet dropouts. Inf. Sci. 2014, 269, 349–365. [Google Scholar] [CrossRef]
  31. Wang, S.; Fang, H.; Tian, X. Minimum variance estimation for linear uncertain systems with one-step correlated noises and incomplete measurements. Digit. Signal Process. 2016, 49, 126–136. [Google Scholar] [CrossRef]
  32. Tian, T.; Sun, S.; Li, N. Multi-sensor information fusion estimators for stochastic uncertain systems with correlated noises. Inf. Fusion 2016, 27, 126–137. [Google Scholar] [CrossRef]
Figure 1. Simulated signal and proposed prediction, filtering and smoothing estimates when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Figure 1. Simulated signal and proposed prediction, filtering and smoothing estimates when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Sensors 17 01151 g001
Figure 2. Prediction, filtering and smoothing error variances when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1,2,3.
Figure 2. Prediction, filtering and smoothing error variances when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1,2,3.
Sensors 17 01151 g002
Figure 3. MSE of Kalman filter and proposed filter when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Figure 3. MSE of Kalman filter and proposed filter when p = 0.5 and γ ¯ 0 = 0.6 , γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Sensors 17 01151 g003
Figure 4. Prediction, filtering and fixed-point smoothing error variances for different values of the probability p, when γ ¯ 0 = 0.6 and γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Figure 4. Prediction, filtering and fixed-point smoothing error variances for different values of the probability p, when γ ¯ 0 = 0.6 and γ ¯ d = 0.1 , d = 1 , 2 , 3 .
Sensors 17 01151 g004
Figure 5. Estimation error variances for different values of the probability γ ¯ 0 , when p = 0.5 .
Figure 5. Estimation error variances for different values of the probability γ ¯ 0 , when p = 0.5 .
Sensors 17 01151 g005

Share and Cite

MDPI and ACS Style

Caballero-Águila, R.; Hermoso-Carazo, A.; Linares-Pérez, J. Optimal Fusion Estimation with Multi-Step Random Delays and Losses in Transmission. Sensors 2017, 17, 1151. https://doi.org/10.3390/s17051151

AMA Style

Caballero-Águila R, Hermoso-Carazo A, Linares-Pérez J. Optimal Fusion Estimation with Multi-Step Random Delays and Losses in Transmission. Sensors. 2017; 17(5):1151. https://doi.org/10.3390/s17051151

Chicago/Turabian Style

Caballero-Águila, Raquel, Aurora Hermoso-Carazo, and Josefa Linares-Pérez. 2017. "Optimal Fusion Estimation with Multi-Step Random Delays and Losses in Transmission" Sensors 17, no. 5: 1151. https://doi.org/10.3390/s17051151

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop