1. Introduction
The problem of distributed estimation for linear dynamic systems via the cooperation and coordination of multiple fixed/mobile agents has been extensively explored by numerous researchers [
1,
2,
3,
4,
5]. Distributed Kalman filtering has been broadly utilized as a powerful and efficient tool to solve the aforementioned scenario [
1,
6,
7,
8,
9,
10]. An essential problem involving cooperation of multi-agent systems that has attracted much attention is the
consensus problem. In order to reconstruct the states of the dynamic system, a network of sensing agents is adopted; these may form either a homogeneous or a heterogeneous network. Each agent transmits state estimates to its immediate neighbors based on the communication topology. Through the consensus strategy, all agents in use eventually agree with each other on a common estimation value regarding the states of the dynamic system.
The early formal study of consensus problems [
11] formed the basis of distributed computing [
12], which has found wide and popular utilization in sensor network applications [
13,
14]. The dynamic consensus problem appears frequently in the cooperation and coordination of multi-agent systems, including scenarios such as formation control [
5,
9,
15], self-alignment, flocking [
4,
16], and distributed filtering [
1,
17]. The typical consensus protocol and its performance analysis were first introduced by Olfati-Saber and Murray in the continuous-time model (see [
3,
7]). In [
7], the authors considered a cluster of first-order integrators working cooperatively under the average consensus control algorithm, in which each agent finally agrees on a common value that is the average of initial states of all agents as the individual ultimate state. In [
1], a distributed Kalman filtering (DKF) algorithm was proposed in which data fusion was achieved through dynamic consensus protocols [
18]. Later, in [
17], the same author extended the results of [
1] to use two identical consensus filters for sensor fusion with different observation matrices, then presented [
1] an alternative distributed Kalman filtering algorithm which applies consensus to the state estimates. This idea forms the foundation of the present paper, in which we propose an adaptive Kalman consensus filtering algorithm. In [
9], the authors presented a different view towards designing consensus protocols based on the Kalman filter structure. By adjusting the time-varying consensus gains, [
9] proved that consensus can be achieved asymptotically under a no-noise condition. In addition, graph theory [
7,
19,
20] has been adopted to construct the communication topology among distributed agents. In this paper, we assume a fixed topology; however, it is not necessarily relaxed to the point of all-to-all connection. This means that each agent is not constrained in communication with others, a consideration that would be more practical in real-world applications.
Recently, many extensions of consensus protocols have been explored to improve the convergence rate of dynamic systems among cooperative agents. This includes the study of communication topology design [
21,
22], optimal consensus-based estimation algorithms [
2,
23,
24], and adaptive consensus algorithms [
25,
26,
27] in both continuous-time and discrete-time scenarios. Other extensions for system control purposes have been studied to realize finite time consensus among agents, methods of which involves event-triggered and sliding mode control [
28,
29]. These research fruits have been considered and embedded in Kalman filtering algorithms [
30,
31], although the complexity of the algorithms makes practical implementation challenging, particularly when compared to the adaptive weight parameter method proposed in this paper.
The main contribution of this paper is to derive an adaptive Kalman consensus filtering (a-KCF) strategy in a continuous-time model and analyze its stability and convergence properties. Extensive simulation results aim to demonstrate better effectiveness of a-KCF compared with the previous work of Olfati-Saber[
17] along with a faster convergence rate of the estimation error when the consensus gains change adaptively based on the disagreement of these filters.
The remainder of this paper is organized as follows. In
Section 2, we provide preliminaries on algebraic graph theory [
20], which is the basis of the consensus strategy. In
Section 3, we provide a retrospective view of the previous work of Olfati-Saber on the Kalman consensus filtering algorithm, which our analysis relies upon. In
Section 4, we illustrate the main results of this paper, namely, derivation of the adaptive Kalman consensus filtering algorithm. The purpose is to adaptively adjust the consensus gain as the weight applied to the disagreement terms in order to improve the convergence of the estimation error. Simulation results are presented in
Section 5, then we conclude our work in
Section 6.
The following notations are be used throughout this paper: and denote the n dimensional square matrices and the set of all matrices, respectively. denotes the identity matrix with dimension . For a given vector or matrix, represents the transpose of the matrix A. For a given square matrix A, denotes its trace and the norm of A is defined by . If , this indicates a random variable f that complies with a Gaussian distribution with mean a and variance .
2. Problem Statement
Consider a continuous-time dynamic system that has the following form:
with
m states measured by
n distributed agents via sensing devices and local filters. Here,
denotes the states of this dynamic system,
represents the dynamical matrix,
represents the white Gaussian noise of the system, which is distributed by the matrix
B with zero mean and covariance
,
is the initial guess for the states of the dynamic system with error covariance
, and the sensing capability of each agent is determined by the following equation:
with the measurement matrix
. The sensing devices are not able to measure all the states of the system; thus, only partial information is available to the local filters for the state estimation. Here,
represents the white Gaussian noise of measurement for the
ith agent, with zero mean and covariance matrix
. Throughout this paper, we assume that there is no noise coupling effect among the agents; therefore, we can set both
and
as diagonal matrices. The
main purpose of this paper is to design an adaptive state estimation structure for each agent such that the state estimation of individuals can be exchanged among their immediate neighbors through the communication topology
.
2.1. Graph Theory Preliminaries
We consider n distributed agents working cooperatively through a communication network/topology characterized by a weighted graph with inconsistent information, where represents the set of agents, denotes the set of edges in , and represents the connected bridge of an ordered pair . For each , we mean that the ith agent can only receive information from the jth agent, not vice versa. In such cases, we call j the neighbor of i and denote as the set of neighbors of the ith agent. If , this means that there is no communication link between the jth and ith agents.
Let represent the adjacency matrix of , defined as and ; otherwise, .
We define the in-degree of the ith agent as and the out-degree of the ith agent as ; then, the graph Laplacian of can be defined as , where .
One important property of
L (see [
19] for more details) is that all eigenvalues of
L are non-negative, and at least one of them is zero. If we denote
as the
ith eigenvalue of
L, then we have the following valid relation for any graph
:
[
32].
In a special case, if the in-degree of the ith agent or , then we call the graph , as an all-to-all connected topology. In such cases, the graph Laplacian L is a symmetric and positive-semidefinite matrix with only one zero eigenvalue.
2.2. Consensus Protocols
Several types of consensus protocols have been explored, and these have been utilized in many different scenarios over the years. In this paper, we mainly adopt the “consensus in network” strategy mentioned in [
17].
Using the background information provided in
Section 2.1 and assuming that there are
n integrator agents working cooperatively with dynamics
, the “consensus in network” strategy forces those agents to reach an agreement on their states by
.
This distributed dynamic structure demonstrates that the
ith agent updates its state by penalizing the state disagreement between its immediate neighbor
and itself in order to ensure that all of those
n agents finally agree on a common value as their ultimate state. Via this protocol, which [
17] proved to be stable and convergent, we can say consensus is achieved through the communication and cooperation among the agents.
Furthermore, we can explore the collective dynamics of individual agents in the graph
, which can be expressed as
, where
L is the aforementioned graph Laplacian in
Section 2.1.
3. Kalman Consensus Filtering Algorithm
In this section, we provide a retrospective view of the previous work of Olfati-Saber [
17], in which the author presented the continuous-time distributed Kalman filtering strategy without adopting the consensus filtering algorithm, instead forcing consensus through state estimates. We treat this work as the fundamental background for designing our adaptive Kalman consensus filters.
In [
17], the author considered
n distributed agents working cooperatively to complete a common task for estimating the state of a linear system, defined in Equation (
1), with the sensing capability defined as in Equation (
2). Each agent shares its instant state estimates with its immediate neighbors
through the communication topology
. Olfati-Saber [
17] proposed individual agents applying the following distributed estimation algorithm:
with initial conditions
and
.
Through this distributed algorithm, it was claimed that the collective dynamics of the estimation errors are
(in the absence of noise, this assumption is made to ease the computation complexity of the proof; later work showed that even if process and measurement noise are present, the stability and convergence properties of the adaptive Kalman consensus filtering algorithm remain valid). The errors converge to zero through analysis of the Lyapunov function
. Furthermore, all agents with the estimator structure in Equation (
3) asymptotically agree with each other on their state estimates, which match the value of the true states of the linear system in Equation (
1), such as
. See the proof in [
17] for more detail.
4. Adaptive Kalman Consensus Filtering
In this section, we show our main contribution to this paper, in which we consider adding an adaptation mechanism to the Kalman consensus filtering algorithm shown in
Section 3. We name this adaptive Kalman consensus filtering, or a-KCF. The proposed
ith agent estimation structure with adaptive gain
on the consensus term is as follows:
where
represents the state estimates of the
ith agent and
is the scalar adaptive gain of the estimation differences between the
ith agent and its neighbors. The consensus weighting matrix
is designed based on Lyapunov stability analysis, as detailed in the following part.
If we consider a scenario in which no noise affects the error dynamics, as proposed in [
17], in a similar fashion, we can derive the local error dynamic of the
ith agent using the fact that
, as shown below:
The basic idea for seeking the adaptive gain involves consideration of two aspects:
Lemma 1. The proposed distributed estimator structure of the ith agent with adaptive gains for each disagreement term is provided by Equation (4). By means of analyzing a Lyapunov function , where is a symmetric positive definite matrix, the local adaptation law can be derived by satisfying the system stability conditions for both and , which turns out to be . Proof.
We consider the following local Lyapunov function for the
ith agent based on Lyapunov redesign methods
where
is a symmetric positive definite solution of the following Lyapunov equation:
with the assumption that
is a symmetric positive definite matrix and all eigenvalues of the matrix
have negative real parts.
Because , there exist many ways to design in order to satisfy the aforementioned assumption. One possible way is to use the pole placement method to design the value for such that generates an exponentially stable local system.
Now, the derivative of
is provided by
. By substituting
from Equation (
5) into Equation (
6), we obtain the following expression:
In order to further simplify Equation (
8), we can assume the weighted consensus matrix
. Thus, the expression of
is
Therefore, by setting the sum of the last two terms in Equation (
9) to equal zero, we can derive the adaptation law shown below:
Therefore, the local adaptation law is now provided by
□
Remark 1. The choice of the structure of is not necessarily unique; it is only necessary for it to be possible to cancel out the term in Equation (8), along with the arbitrary matrix , which has an appropriate dimension such that and are available signals for the estimation process. Here, we choose for simplification. Furthermore, if we examine the derivative of
, then we find that
However, we can only argue that
This means we need additional arguments to guarantee that the convergence of reaches zero, as is only negative semi-definite. Thus, in order to show that each individual error dynamic asymptotically converges to zero, we need to examine the collective error dynamics for all agents.
Let us denote
as the collective error vector; now, we can write the collective error dynamics in the following form (without noise):
where
and
Remark 2. The row sum of Γ is zero; furthermore, the sum of the off-diagonal entries in each row is equal to the negative value of the corresponding diagonal entry. The operator ⊗ denotes the Kronecker product, while denotes a identity matrix.
Theorem 1. Consider a wireless sensor network consisting of n
agents which form a communication topology . Each agent adaptively estimates the states of a linear dynamic system with the structure governed by Equation (4) using the local adaptation law in Equation (11). Each satisfies the Lyapunov equation as in Equation (7). Then, the collective dynamics of the estimation errors in Equation (14) (without noise) represent a stable system. Furthermore, the adaptive Kalman consensus filter (a-KCF) and the adaptation law law can be implemented through Algorithm 1, which generates a stable estimation system. Proof.
The error dynamics can now be written as
where
and
is defined as in Equation (
15).
We now can write the collective Lyapunov function as follows:
We denote
, where each entry in
is a symmetric positive definite solution of Equation (
7) with respect to a different index
i.
Now, we can use the collective error dynamics to represent
V in Equation (
18) as follows:
Then, we take the derivative of
V and combine it with the adaptation law in Equation (
11), resulting in the following expression:
where
and each entry of
must satisfy Equation (
7) for the corresponding
ith index.
Therefore, we can argue that the collective error dynamics
and
from Equation (
20) are bounded; thus, we can say that
. From Equation (
16), it can be concluded that
is bounded as well, that is,
. Therefore, using Barbalat’s Lemma, we can write
as
; in other words, the error dynamics of each agent asymptotically converge to zero. □
Algorithm 1 Adaptive Kalman Consensus Filter |
- 1:
Initialize: , , . - 2:
Compute the weighted consensus matrix: , where is a symmetric positive definite matrix that satisfies the Equation (7). - 3:
Compute the Kalman gain: = , and satisfies to generate an exponentially stable local system. - 4:
Compute the adaptive gain: - 5:
Compute the estimated state: - 6:
return
|
Special case: when the adaptive gain
If we assume that the adaptive gains
for all neighbors of the
ith agent are identical, such that
. Then, we can take the
term out of the summation in Equation (
4); in this manner, we can further simplify the structure of
in Equation (
15) to
. Therefore, the error dynamics in Equation (
16) change to
where
L is the graph Laplacian of the communication topology
.
Therefore, we can rewrite the adaptation law in this special case as follows:
Remark 3. By carefully examining the above analysis, we find that there is vagueness when observing the agreement evolution of our adaptive Kalman consensus filters. Thus, it is necessary to show that the convergence of individual a-KCF to a common value, which implies that each agent agrees with the others on their estimates. To this end, it is necessary to define a quantity to measure the disagreement level of the proposed a-KCF.
Lemma 2. In a similar manner as [
17],
we define the measure of disagreement of n a-KCFs as follows:where is the mean estimate of n a-KCFs. Therefore, represents the deviation of each agent’s estimate from the mean estimate. Furthermore, by examining the disagreement dynamics of Equation (23), it is proved to converge to zero in finite time. Proof.
The measurement of disagreement can be transformed into the following expression:
Then, we can write the collective dynamics of
as follows:
It can instantly be recognized that the convergence of
is related to the convergence of
. We have proved that
converges to zero in finite time; however, due to the zero eigenvalue of the matrix
, Equation (
24) does not simply imply that
converges to zero in finite time. Therefore, we need to find an alternate approach to prove this lemma.
Let us define
as the estimate difference between the
ith and the
jth agents and denote
as the estimation error difference between the
ith and the
jth agents. It is immediately clear that
. Now, we can rewrite Equation (
23) into the following form:
From Equation (
26), it can be argued that if we are able to prove the convergence of the dynamics of
or
to zero in finite time, we can conclude that
converges to zero, as
.
For simplification, consider the special case when
. According to Equation (
5), we can find the dynamics of
as follows:
In
Section 4, we showed that
generates an exponentially stable linear dynamic system for the
ith agent, while both
and
, along with
, are bounded quantities. In the proof of Theorem 1, we concluded that
; hence,
. To complete this proof, we need to prove that both
and
are
bounded.
We now proceed to study the
norm of
, which is denoted as
; then,
From above, it can be proved that
; therefore, we can conclude that
converges to zero, as
. Furthermore,
in Equation (
26) converges to zero in finite time. □
5. An Example
Consider a linear dynamic system in Equation (
1) with
; assume that there are three agents distributed in a 2D domain. Each agent embeds a sensing device with measurement capability
, such as
.
It is clear while that the system itself is not observable by individual agents, it is observable through all of them. In this simulation, we carry out a comparison between the Kalman consensus filtering algorithm proposed by Olfati-Saber in [
17] and our adaptive Kalman consensus filtering algorithm from Theorem 1 on their performance and effectiveness.
We set up our simulation with the following parameter configuration:
In both cases above, we assume the same initial conditions for the error dynamics, such as
;
and
. In order to construct a fair comparison between the KCF and a-KCF algorithms, we assume that the dynamics of the error covariance matrix in Equation (
3) comply with the algebraic Riccati equation, and use the same filter gains
in both cases.
The simulation results are shown in
Figure 1,
Figure 2,
Figure 3 and
Figure 4. From a comparison of
Figure 1a,b, we can conclude that a-KCF has better performance regarding the error convergence rate than the KCF in [
17].
Figure 2a,b illustrates the evolution of the disagreement term in both the KCF and the a-KCF algorithms. According to the simulation results in
Figure 3a,b, both the state estimation error and the disagreement term converge faster in the a-KCF case.
Figure 4 demonstrates the dynamics of the adaptive gain.