Next Article in Journal
Distance-Based Paper Device Combined with Headspace Extraction for Determination of Cyanide
Previous Article in Journal
An Integrative Framework for Online Prognostic and Health Management Using Internet of Things and Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Diffusion Minimum Kernel Risk-Sensitive Loss Algorithm over Multitask Sensor Networks

1
College of Artificial Intelligence, Southwest University, Chongqing 400715, China
2
Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, and College of Electronic and Information Engineering, Southwest University, and Chongqing Collaborative Innovation Center for Brain Science, Chongqing 400715, China
*
Author to whom correspondence should be addressed.
Current address: Chongqing Collaborative Innovation Center for Brain Science, Southwest University, Chongqing 400715, China.
Sensors 2019, 19(10), 2339; https://doi.org/10.3390/s19102339
Submission received: 28 April 2019 / Accepted: 17 May 2019 / Published: 21 May 2019
(This article belongs to the Section Sensor Networks)

Abstract

:
Distributed estimation over sensor networks has attracted much attention due to its various applications. The mean-square error (MSE) criterion is one of the most popular cost functions used in distributed estimation, which achieves its optimality only under Gaussian noise. However, impulsive noise also widely exists in real-world sensor networks. Thus, the distributed estimation algorithm based on the minimum kernel risk-sensitive loss (MKRSL) criterion is proposed in this paper to deal with non-Gaussian noise, particularly for impulsive noise. Furthermore, multiple tasks estimation problems in sensor networks are considered. Differing from a conventional single-task, the unknown parameters (tasks) can be different for different nodes in the multitask problem. Another important issue we focus on is the impact of the task similarity among nodes on multitask estimation performance. Besides, the performance of mean and mean square are analyzed theoretically. Simulation results verify a superior performance of the proposed algorithm compared with other related algorithms.

1. Introduction

Distributed data processing over sensor networks has emerged as an attractive and challenging research area for various applications such as industrial automation, cognitive radios and inference tasks [1,2,3,4]. Distributed estimation plays a significant role in distributed data processing, which estimates some parameters of interest from noise measurements by exchanging information with neighboring nodes. Most algorithms proposed for distributed estimation can be classified into a consensus strategy [5,6,7,8], incremental strategy [9,10,11] and diffusion strategy [12,13,14]. In our work, we center on a diffusion strategy, which is robust, fully distributed and flexible among these strategies [15,16,17,18,19].
Diffusion strategies are particularly attractive schemes in distributed estimation, such as diffusion Recursive Least Squares (RLS) [20,21], diffusion Least Mean Square (LMS) [13,14]. With the mean-square error (MSE) criterion, these algorithms can accomplish a satisfying performance in a Gaussian noise environment. However, their performance may deteriorate dramatically in the presence of impulsive noise [22,23]. Some algorithms have been proposed to solve the issue, including Diffusion least-mean power (D-LMP) and the Diffusion sign-error Least Mean Square (DSE-LMS) adaptive filtering algorithm [24,25]. To efficiently address the non-Gaussian noise, the correntropy [26,27] was proposed, which is a higher order statistic and widely used in adaptive filters. Moreover, the generalized maximum correntropy criterion (GMCC) algorithm and the minimum kernel risk-sensitive loss (MKRSL) were proposed [28,29], which provide more general frameworks and better performance. In this work, we consider the diffusion minimum kernel risk-sensitive loss (D-MKRSL) algorithm for distributed estimation over multitask networks.
In previous works, diffusion strategies mainly focus on the single-task estimation problem where an identical parameter vector is estimated by all the nodes [30]. On the contrary, many essential applications are multitask-oriented, such as regression, web page categorization and target location tracking. In the above situations, multiple optimum vectors are different but related, which are inferred synchronously over the networks by all nodes in a collaborative manner. This type of problem is known as a multitask problem. Generally, distributed estimation problems over multitask networks can be roughly classified into two fields. In the first case, there is no knowledge about the correlation of tasks. Meanwhile, which nodes share the same tasks is unknown and nodes cooperate according to network topology [31,32,33]. In the second situation, it is assumed that nodes know which cluster they belong to and the parameter vector in each cluster is the same. Exploiting the information about the similarity of tasks, diffusion strategies for distributed estimation over multitask are obtained [34,35,36,37]. In our work, we focus on the second case.
Inspired by the adapt-then-combine (ATC) DLMS algorithm, we propose the diffusion MKRSL algorithm over multitask networks. The algorithm can achieve desirable performance in both Gaussian and impulsive noise environments. Additionally, the impact of task relatedness on estimation performance is also studied. Moreover, the mean and mean square stability are analyzed theoretically. Effectiveness and advantages of the proposed algorithm are verified by simulation results.
The remaining parts of the article are organized as follows: In Section 2, we briefly introduce the data model of distributed estimation and propose the multitask Diffusion MKRSL algorithm. In Section 3, the mean and mean square performance of the multitask D-MKRSL algorithm are analyzed. Simulation results are demonstrated in Section 4. Finally, we draw conclusions in Section 5.
Notation: We use ( . ) T , E . and ⨂ to denote transposition, expectation and Kronecker product operators, respectively. I m denotes an m × m identity matrix. 1 is an N × 1 all-unity vector. . is the absolute value of a scalar.

2. Multitask Diffusion Estimation

2.1. Data Model

Let us consider a connected network with K nodes. Every node k 1 , 2 , , K has access to scalar random variables d k , i and a zero-mean M × 1 regression vector u k , i at every time instant i 0 . The data of node k is related via the linear regression model:
d k , i = u k , i T w k 0 + n k , i
where n k , i is the random measurement noise with zero-mean and variance σ n , k 2 , which is independent of regression vector u k , i . The goal of distributed estimation is to estimate an M × 1 deterministic but unknown vector w k 0 by exchanging and combining the data only from neighboring nodes, which is regarded as single-task problem with w k 0 = w 0 for k = 1 , 2 , , K , and multitask problem with w k 0 w l 0 for k l . It is assumed that there is no limit to how much information can be transmitted among neighbors.

2.2. Diffusion MKRSL Algorithm

In many previous works, the diffusion distributed estimation algorithms are based on the MSE criterion, which achieves desirable performance if the measurement noise is Gaussian, while their performance may deteriorate dramatically in an impulsive noise environment. To solve the parameter estimation problem over multitask sensors networks, it becomes a significant focus of our interest to design a novel algorithm that is robust to both Gaussian noises and impulsive noises.
The information theoretic learning (ITL) plays a significant role and provides a general framework in distributed parameter estimation for non-Gaussian cases. The correntropy is a local statistical similarity measure in ITL, which is defined by Reference [26]
V ( X , Y ) = E [ k σ ( X Y ) ] = k σ ( x y ) d F X Y ( x , y )
where X , Y are two random variables, k σ ( . ) is a shift-invariant Mercer kernel and σ > 0 denotes the kernel bandwidth. F X Y ( x , y ) is the joint distribution function of X , Y . In our work, we focus on the Gaussian kernel, which takes the following form:
k σ ( x y ) = exp ( x y ) 2 2 σ 2
The minimum kernel risk-sensitive loss (MKRSL) algorithm is derived by applying the KRSL to develop a new adaptive filtering algorithm, which shows better convex properties than correntropic loss on the error performance surface [29,38]. The KRSL between two random variables X and Y is defined by
L λ X , Y = 1 λ E [ exp ( λ ( 1 k σ ( X Y ) ) ) ] = 1 λ exp ( λ ( 1 k σ ( X Y ) ) ) d F X Y ( x , y )
where λ > 0 is the risk-sensitive parameter. Nevertheless, the exact joint distribution of X , Y is usually unavailable in application scenarios. On the contrary, only a limited number of sample values x i , y i i = 1 L are known. Therefore, the sample mean estimator of KRSL—called empirical KRSL—is calculated by an average over samples:
L ^ λ ( X , Y ) = 1 L λ i = 1 L exp ( λ ( 1 k σ ( x ( i ) y ( i ) ) ) )
Then, the KRSL cost function is derived as
J K R S L = 1 L λ i = 1 L exp ( λ ( 1 k σ ( e ( i ) ) ) )
with e ( i ) = d ( i ) u i T w . The time average of the KRSL cost function in the above equation can be replaced by the expectation
J K R S L = 1 λ E [ exp ( λ ( 1 k σ ( e ( i ) ) ) ) ]
Based on the KRSL cost function mention in the above Equation(7), the instantaneous cost function of the KRSL algorithm is approximated as
J ˜ K R S L = 1 λ exp ( λ ( 1 k σ ( e ( i ) ) ) )
For the distributed diffusion estimation problem, our goal is to seek the best w k 0 by minimizing the diffusion KRSL cost function at each node k by cooperating with all neighboring nodes. For each node k, N k is the one-hop neighbor set and c l , k are non-negative real cooperative according to Metropolis rule weights satisfying
c l , k = 1 max n k , n l , if l N k k , 1 l N k k c l , k , if l = k , 0 . if l N k ,
where n k is the degree of node k. The real, non-negative combining coefficients c l , k satisfy the following conditions: l N k k c l , k = 1 and c l , k = 0 i f l N k , C I = I , I T C = I T , where C is an N × N matrix. The KRSL local cost function at each node k can be formulated as
J k l o c ( w ) = l N k c l , k J ˜ K R S L ( e l , i ) = 1 λ l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) = 1 λ l N k c l , k exp ( λ ( 1 k σ ( d l , i u l , i T w ) ) )
Based on the KRSL local cost function, the derivative of (10) with respect to w can be derived as
J k l o c ( w ) = 1 λ l N k c l , k w exp ( λ ( 1 k σ ( e l , i ) ) ) = 1 σ 2 l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T
At node k, the weight vector update equation based on a stochastic gradient for w k 0 is obtained by
w k ( i ) = w k i 1 μ J k l o c ( w ) = w k i 1 + μ σ 2 l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T = w k i 1 + η l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T
where η = μ σ 2 is step-size and w k ( i ) is estimator for w k 0 at time index i. The above algorithm is a new expression of the MKRSL algorithm. Inspired by the general framework for a diffusion-based distributed estimation algorithm [13], an adapt-then-combine (ATC) strategy for a diffusion MKRSL algorithm is proposed. The ATC scheme first updates the value of the estimator for each node with the adaptive algorithm. Then, the intermediate estimates are fused from its neighbors for each node k. The intermediate estimate at each node k is defined as:
φ k i 1 = l N k β l , k w l i 1
The nodes update their intermediate estimates by
φ k i = φ k i 1 + η l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T
φ k i 1 is an intermediate estimate at time index i 1 for node k. The non-negative real value β l , k is a weight coefficient, which corresponds to the matrices B , especially B = I in the ATC scheme [12]. Therefore, we can obtain:
φ k i = w k i 1 + η l N k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T
In the above Equation (15), the task relatedness among nodes is ignored, which is called non-cooperative diffusion MKRSL in this article.
However, multitask estimation is an attracting filed in practical applications. Nodes are grouped into some clusters and each cluster has an identical task in clustered multi-task networks. Furthermore, utilizing the relatedness of tasks, the performance of distributed estimation can be improved. The Equation (15) is adjusted for multitask estimation:
φ k i = w k i 1 + η l N k c k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T + τ l N k c k ρ k l w l i w k i
c k is the cluster of node k, with the cluster of node k non-negative strength parameter τ , weights ρ k l and η ( i ) = exp ( λ ( 1 k σ ( e i ) ) ) k σ ( e i ) . The notation N k c ( k ) is the set of neighboring nodes k and in the same cluster as k. On the contrary, N k c ( k ) denotes the set of neighboring nodes of k that are not in the same cluster as k. The Equations (15) and (16) are defined as the increment step. The combination step can then be derived as
w k i = l N k c l , k φ l i
The step-size η ( i ) is a function of e ( i ) and the curves with different values of λ (where σ = η = 2.0 ) and σ (where λ = η = 2.0 ) is depicted in Figure 1.
It is shown that the step-size η ( i ) will approach zero as e ( i ) for different values of λ . Therefore, the MKRSL algorithm maintains the robustness to outliers, such as impulsive noise.
For a better understanding, the Multitask Diffusion MKRSL algorithm is summarized in Algorithm 1:
Algorithm 1: Multitask Diffusion MKRSL Algorithm
Input: d k , i , u k , i T , η , τ , and c l , k satisfying (10)
Initialization: Start with w l , 1 = 0 for all l.
for i = 1 : T
    for each node k:
    Adaptation
     φ k i = w k i 1 + η l N k c k c l , k exp ( λ ( 1 k σ ( e l , i ) ) ) k σ ( e l , i ) e l , i u l , i T
       + τ l N k c k ρ k l w l i w k i
    Communication
   Transmit the intermediate φ k i to all neighbors in N k
    Combination
             w k i = l N k c l , k φ l i
end for

3. Performance Analysis

The multitask D-MKRSL algorithms are evaluated theoretically under model (1) in this section. In the following, some common assumptions are adopted for tractable analysis [39,40].
(1) The regression vector u k , i is independently and identically distributed (i.i.d.) and E u k , i u k , i T = R u , k .
(2) For each node k at time index i, the input noise n k i is independent of u k , i and is a mixture signal of zero mean Gaussian, we have E [ n k , i ] = 0 .
(3) The step-size η is small enough, so the squared value can be negligible.
Then, the estimate-error vectors are defined as follows:
w ˜ k , i = w k 0 w k , i
and
φ ˜ k , i = w k 0 φ k , i
Furthermore, the global quantities are defined to covert the local variables to global ones:
K = b l o c k d i a g η I M , , η I M
X = b l o c k d i a g τ I M , , τ I M
w ˜ i = c o l w ˜ 1 , i , , w ˜ K , i
w i = c o l w 1 , i , , w K , i
w k 0 = c o l w 1 0 , , w K 0

3.1. Mean Performance

We consider the gradient error caused by replacing the cost function of KRSL with instantaneous values. The gradient error of the intermediate estimate at time i and each node k is defined as follows:
s k ( w k , i 1 ) = f ^ k ( w k , i 1 ) f k ( w k , i 1 )
where f ^ k ( w k , i 1 ) = 1 σ 2 exp ( λ ( 1 k σ ( e k , i 1 ) ) ) k σ ( e k , i 1 ) e k , i 1 u k , i 1 T and f k ( w k , i 1 ) = 1 σ 2 E [ exp ( λ ( 1 k σ ( e k , i 1 ) ) ) k σ ( e k , i 1 ) e k , i 1 u k , i 1 T ]
The update equation of the intermediate estimate can be rewritten as
φ k , i = w k , i 1 + μ s k ( w k , i 1 ) + f k ( w k , i 1 )
f k ( w k , i 1 ) is twice continuous differentiable in a neighborhood of a line segment between points w k 0 and w k 1 . Thus, based on the Theorem 1.2.1 in Reference [41], we have
f k ( w k , i 1 ) = f k ( w k 0 ) 0 1 H k w k 0 t w ˜ k , i 1 d t w ˜ k , i 1
where H k ( w ) is the Hessian matrix of f k ( w k , i 1 ) . w ˜ k , i 1 = w k 0 w k , i 1 is the weight error vector for node k. The unknown vector w k 0 is the real-value that we want to estimate, so f k ( w k 0 ) is equal to zero. The estimate of each node converges to the vicinity of the unknown vector w k 0 . Thus, w ˜ k , i is small enough such that it is negligible, yielding
f k ( w k , i 1 ) 0 1 H k w k 0 d t w ˜ k , i 1 = H k ( w k 0 ) w ˜ k , i 1 = β R u , k w ˜ k , i 1
where R u , k = E u k , i u k , i T and β is a constant.
So, the approximate value of the gradient error at the value of w k 0 is
s k ( w k , i 1 ) s k ( w k 0 ) = f ^ k ( w k 0 ) f k ( w k 0 ) = 1 σ 2 exp ( λ ( 1 k σ ( e k , i 1 ) ) ) k σ ( e k , i 1 ) e k , i 1 u k , i 1 T
Substituting (28) and (29) into (26) and adjusting for multitask estimation, we can get the intermediate estimate
φ k , i = w k , i 1 + μ ( s k ( w k 0 ) H k w k 0 w ˜ k , i 1 ) + τ Q w ˜ k , i 1 + w k 0
where
Q = I MN P I M
P is the matrix with k , l -th entry ρ k l . Substituting (30) into (17), we can get the update equation of w k i as follows
w k i = l N k c l , k [ w k , i 1 + μ ( s k ( w k 0 ) H k w k 0 w ˜ k , i 1 ) + τ Q w ˜ k , i 1 + w k 0 ]
Define global quantity H = d i a g H 1 w 1 0 , , H k w N 0 and rewrite (32) as
w i = C w i 1 + K s i KH w ˜ i 1 + XQ w ˜ i 1 + XQ w 0
Noting that C w 0 = w 0 , subtracting both sides of (33) from w 0 , the global vector is obtained:
w ˜ i 1 = C I MN KH + XQ w ˜ i 1 + CK s i + CXQ w 0
Calculating the expectation of (34) leads to
E w ˜ i 1 = C I MN KH + XQ E w ˜ i 1 + CK E s i + CXQ w 0
where E s i = c o l E [ s 1 ( w 1 0 ) , , s N ( w N 0 ) ] = 0 . Based on Lemma 1 of [13], the matrix I MN KH + X should be stable to guarantee mean stability. There it holds that
λ max I MN KH + XQ < 1
λ max is the largest eigenvalue of matrix. Thus, a sufficient condition for maintaining the stability of the algorithm is:
0 < η < 2 β λ max R u , k + 2 τ

3.2. Mean-Square Performance

In this section, we mainly focus on the mean-square performance of the proposed algorithm. Computing the weight norm of (34) and calculating the expectations, we can obtain
E w ˜ i Σ 2 = E w ˜ i 1 Γ 2 + E [ s i T K C T Σ CK s i ] + 2 XQ w 0 T Σ C I MN KH + XQ E [ w ˜ i 1 ] + CXQ w 0
where
Γ = I MN KH + XQ C T Σ C I MN KH + XQ
and Σ is an Hermitian non-negative-definite matrix. w ˜ i is dependent of Γ with Assumptions 1 and 2. Therefore, we have:
E w ˜ i 1 Γ 2 = E w ˜ i 1 E [ Γ ] 2
Let
γ = v e c E Γ
and
σ = v e c Σ
where v e c ( . ) is the transpose of the vectorization of a matrix. The Equation (40) can be rewritten to follow equation with (41), (42):
E w ˜ i σ 2 = E w ˜ i 1 γ 2 + E [ s i T K C T Σ CK s i ] + 2 XQ w 0 T Σ C I MN KH + XQ E [ w ˜ i 1 ] + CXQ w 0
The vectorization operator denoted by Reference [42] is:
v e c A B C = C T A v e c B
Taking expectation and vectorization operations with (38), (41), (42), we have
γ = δ σ
where
δ = E I MN KH + XQ I MN KH + XQ Z
Z = E C T C T
Based on the relationship of the matrix trace and the vectorization operator [42], we have
t r A T B = v e c T B v e c A
Σ is symmetric and deterministic, and we obtain
E [ s i T K C T Σ CK s i ] = v e c T V Z σ
where V = K E s i s i T K . According to A.1 and A.2, V can be evaluated as
V = b l o c k d i a g η 2 s 1 2 w 1 0 , , η 2 s K 2 w K 0
Substitution of (45) and (50) into (43) has
E w ˜ i σ 2 = E w ˜ i 1 δ σ 2 + v e c V Z σ + 2 XQ w 0 T Σ C I MN KH + XQ E [ w ˜ i 1 ] + CXQ w 0
the recursion of Equation (51) is stable and convergent if the matrix δ is stable. δ can be approximated as
δ I MN KH + XQ I MN KH + XQ Z
We know that all the entries of Z are non-negative and all its columns sum up to unity. From the above equation, the stability of δ is in accordance with the stability of I MN KH + XQ . Therefore, choosing the step-size lined with the Equation (37) can keep the proposed algorithm stable in the mean-square sense.

4. Simulation

In this section, we validate the performance of the proposed algorithm over multitask sensor networks in two scenarios: a Gaussian environment and an impulsive noise environment. The noise is assumed to be generated by a Gaussian mixture distribution, which is commonly used in signal processing [43,44]:
p n i = ( 1 v i ) N ( 0 , σ 1 2 ) + v i N ( 0 , σ 2 2 )
where N ( 0 , σ i 2 ) ( i = 1 , 2 ) is the Gaussian distribution with zero-mean and variance σ i 2 . And σ 2 2 is set to much larger than σ 1 2 , which can generate the impulsive noise.
More frequent impulses are achieved with an increase of v i , especially
if ν i = 0 G a u s s i a n if ν i 0 I m p u l s i v e .
Increasing ν i leads to more frequent impulses.
We consider a fully connected sensor network with 15 nodes. The network topology and cluster structures are demonstrated in Figure 2. From the network topology, we can easily find that nodes 1 to 6 belong to the first cluster. Meanwhile, nodes 7 to 10 compose the second cluster and nodes 11 to 15 are in the third cluster.
Input variances and noise variances based on Assumptions 1 and 2 are depicted in Figure 3.
Scenario 1 (Gaussian noises Environment): As shown in Figure 3, the desired signal is a random process with a zero-mean Gaussian (i.i.d.) noise signal. In the experiment, system parameters are set with λ = 2 , σ = 1.5 and the step-size is set with η = 0.02 . τ is a regularization parameter, which promotes similarities between the tasks of the neighboring cluster and is chosen τ = 0.1 . The learning curve of the mean square deviation(MSD) is defined as
M S D = 1 K k = 1 K w k 0 w k , i 2 2
which is adopted for performance comparison. d ( i ) is the average value of d k , i for all nodes k at time i in Figure 4a. We compare some related algorithms over multitask network, such as diffusion least mean p-power (D-LMP) [21], diffusion generalized maximum correntropy criterion algorithm (D-GMCC) [16], diffusion sign-error LMS (DSE-LMS) [22], D-LMS [12] and the proposed d-MKRSL algorithm in Figure 4b. The step-sizes of all algorithms are chosen after many experiments to ensure the same convergence speed, and other parameters for each algorithm are experimentally selected to achieve a desirable performance. From the above figure, we can conclude that the D-MKRSL algorithm outperforms other related algorithms in the Gaussian noise environment.
Scenario 2 (Impulsive noise Environment): The impulsive noise model (54) is adopted to depict the distribution of impulsive interference in the experiment. We now test the influence of the impulsive interference on the performance of the algorithms mentioned above. In Figure 5a and Figure 6a, the desired signals are plotted with v i = 0.05 , 0.03 impulsive noise. The corresponding performance of the algorithms in the impulsive noise environment is plotted in Figure 5b and Figure 6b. The value of the parameters α and λ for D-GMCC are selected to achieve the best performance in both the Gaussian and impulsive noise environments. We can observe that the proposed D-MKRSL algorithm is robust and also shows superior performance compared with other related algorithms in the impulsive noise environment.
Furthermore, we consider the performance of the algorithm in a nonstationary scenario and the unknown vector w k 0 is assumed to change at time 1000. From the convergence curves in Figure 7, it can be easily observed that the proposed algorithm maintains a desirable performance even in the presence of sudden changes of an unknown vector.
Another important aspect is how the correlation of tasks influence the estimation performance. First, we investigate whether the proposed algorithm can promote performance by utilizing the relatedness of tasks compared with non-cooperative strategy. Figure 8 compares the D-MKRSL algorithm with a non-cooperative strategy over a multitask network at identical relatedness of tasks. It is clear that utilizing the relatedness of tasks is beneficial to improve the performance of estimation. Next, the impact of the similarity of tasks on performance is studied. According to Reference [35], the optimum mean vector is assumed to uniformly distribute on a circle of radius r centered at w k 0 . The bigger the value of r is, the smaller the correlation of the tasks will be. Optimum parameter vectors over the multitask network will be different but related based on the model. The multitask estimation model can be expressed as:
w k 0 = w 0 + r cos θ k sin θ k θ k = 2 π k 1 / N + π / 8
Figure 9 demonstrates that the performance of the algorithms will be improved with the increasing similarity.

5. Conclusions

In this work, we consider the problem of distributed estimation over multitask sensor networks. Then, the D-MKRSL algorithm is proposed and can achieve a desirable performance. Through theoretical analysis, a sufficient condition for ensuring the stability of the D-MKRSL algorithm is obtained. Compared with related algorithms, the simulation results show that the D-MKRSL algorithm has better performance in both Gaussian and impulsive noise environments. Furthermore, we uncover the relationship between the relatedness of tasks and estimation performance. It is demonstrated that the performance is improved with a higher correlation among tasks by cooperation strategy.

Author Contributions

Data curation, F.C.; Funding acquisition, S.D.; Project administration, X.L.; Software, X.L. and Q.S.; Supervision, S.D. and F.C.; Writing–original draft, X.L.; Writing–review and editing, Q.S. and S.X.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant No.61875168) and Chongqing Research Program of Basic Research and Frontier Technology (No. cstc2017jcyjAX0265).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sayed, A.H. Adaptation, learning, and optimization over networks. Found. Trends Mach. Learn. 2014, 7, 311–801. [Google Scholar] [CrossRef]
  2. Lorenzo, P.D.; Barbarossa, S.; Sayed, A.H. Bio-inspired swarming for dynamic radio access based on diffusion adaptation. In Proceedings of the 2011 19th European Signal Processing Conference (EUSIPCO), Barcelona, Spain, 29 August–2 September 2011; pp. 402–406. [Google Scholar]
  3. Chen, J.; Cao, X.; Cheng, P.; Xiao, Y.; Sun, Y. Distributed collaborative control for industrial automation with wireless sensor and actuator networks. IEEE Trans. Ind. Electron. 2010, 57, 4219–4230. [Google Scholar] [CrossRef]
  4. Sayed, A.H.; Tu, S.; Chen, J.; Zhao, X.; Towfic, Z.J. Diffusion strategies for adaptation and learning over networks. IEEE Signal Process. Mag. 2013, 30, 155–171. [Google Scholar] [CrossRef]
  5. Olfati-Saber, R.; Fax, J.A.; Murray, R.M. Consensus and cooperation in networked multi-agent systems. Proc. IEEE 2007, 95, 215–233. [Google Scholar] [CrossRef]
  6. Kar, S.; Moura, J.M.F. Distributed consensus algorithms in sensor networks: Link failures and channel noise. IEEE Trans. Signal Process. 2009, 57, 355–369. [Google Scholar] [CrossRef]
  7. Wang, J.; Peng, D.; Jing, Z.; Chen, J. Consensus-Based Filter for Distributed Sensor Networks with Colored Measurement Noise. Sensors 2018, 18, 3678. [Google Scholar] [CrossRef] [PubMed]
  8. Nedic, A.; Ozdaglar, A. Distributed subgradient methods for multiagent optimization. IEEE Trans. Autom. Control 2009, 54, 48–61. [Google Scholar] [CrossRef]
  9. Nedic, A.; Bertsekas, D.P. Incremental subgradient methods for nondifferentiable optimization. SIAM J. Optim. 2001, 12, 109–138. [Google Scholar] [CrossRef]
  10. Rabbat, M.G.; Nowak, R.D. Quantized incremental algorithms for distributed optimization. IEEE J. Sel. Areas Commun. 2005, 23, 798–808. [Google Scholar] [CrossRef] [Green Version]
  11. Lopes, C.G.; Sayed, A.H. Incremental adaptive strategies over distributed networks. IEEE Trans. Signal Process. 2007, 48, 223–229. [Google Scholar] [CrossRef]
  12. Chen, J.; Sayed, A.H. Diffusion adaptation strategies for distributed optimization and learning over networks. IEEE Trans. Signal Process. 2012, 60, 4289–4305. [Google Scholar] [CrossRef]
  13. Cattivelli, F.S.; Sayed, A.H. Diffusion LMS strategies for distributed estimation. IEEE Trans. Signal Process. 2010, 58, 1035–1048. [Google Scholar] [CrossRef]
  14. Zhao, X.; Sayed, A.H. Performance limits for distributed estimation over LMS adaptive networks. IEEE Trans. Signal Process. 2012, 60, 5107–5124. [Google Scholar] [CrossRef]
  15. Tu, S.Y.; Sayed, A.H. Diffusion strategies outperform consensus strategies for distributed estimation over adaptive networks. IEEE Trans. Signal Process. 2012, 60, 6217–6234. [Google Scholar] [CrossRef]
  16. Chen, F.; Li, X.; Duan, S.; Wang, L.; Wu, J. Diffusion generalized maximum correntropy criterion algorithm for distributed estimation over multitask network. Digit. Signal Process. 2018, 81, 16–25. [Google Scholar] [CrossRef]
  17. Liu, Y.; Li, C.; Tang, W.K.S.; Zhang, Z. Distributed estimation over complex networks. Inf. Sci. 2012, 197, 91–104. [Google Scholar] [CrossRef]
  18. Chen, F.; Shao, X. Broken-motifs diffusion LMS algorithm for reducing communication load. Signal Process. 2017, 197, 91–104. [Google Scholar] [CrossRef]
  19. Chen, F.; Shao, X. Complementary performance analysis of general complex-valued diffusion LMS for noncircular signals. Signal Process. 2019, 160, 237–246. [Google Scholar]
  20. Cattivelli, F.S.; Lopes, C.G.; Sayed, A.H. A diffusion RLS scheme for distributed estimation over adaptive networks. In Proceedings of the 2007 IEEE 8th Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Helsinki, Finland, 17–20 June 2007; pp. 1–5. [Google Scholar]
  21. Cattivelli, F.S.; Lopes, C.G.; Sayed, A.H. Diffusion recursive leasts-quares for distributed estimation over adaptive networks. IEEE Trans. Signal Process. 2008, 56, 1865–1877. [Google Scholar] [CrossRef]
  22. Gao, W.; Chen, J. Kernel Least Mean p-Power algorithm. IEEE Signal Process. Lett. 2017, 24, 996–1000. [Google Scholar] [CrossRef]
  23. Shao, X.; Chen, F.; Ye, Q.; Duan, S. A Robust Diffusion Estimation Algorithm with Self-Adjusting Step-Size in WSNs. Sensors 2017, 17, 824. [Google Scholar] [CrossRef] [PubMed]
  24. Wen, F. Diffusion least-mean P-power algorithms for distributed estimation in alpha-stable noise environments. Electron. Lett. 2013, 49, 1355–1356. [Google Scholar] [CrossRef]
  25. Ni, J.; Chen, J.; Chen, X. Diffusion sign-error LMS algorithm: formulation and stochastic behavior analysis. Signal Process. 2016, 128, 142–149. [Google Scholar] [CrossRef]
  26. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  27. Chen, B.; Liu, X.; Zhao, H.; Principe, J.C. Maximum correntropy Kalman filter. Automatica 2017, 76, 70–77. [Google Scholar] [CrossRef] [Green Version]
  28. Chen, B.; Xing, L.; Zhao, H.; Zheng, N.; Principe, J.C. Generalized correntropy for robust adaptive filtering. IEEE Trans. Signal Process. 2016, 64, 3376–3387. [Google Scholar] [CrossRef]
  29. Chen, B.; Xing, L.; Xu, B.; Zhao, H.; Zheng, N.; Principe, J.C. Kernel Risk-Sensitive Loss: Definition, Properties and Application to Robust Adaptive Filtering. IEEE Trans. Signal Process. 2017, 65, 2888–2901. [Google Scholar] [CrossRef] [Green Version]
  30. Chen, J.; Richard, C.; Sayed, A.H. Diffusion LMS over multitask networks. IEEE Trans. Signal Process. 2015, 63, 2733–2748. [Google Scholar] [CrossRef]
  31. Chen, J.; Sayed, A.H. Distributed Pareto optimization via diffusion strategies. IEEE J. Sel. Top. Signal Process. 2013, 7, 205–220. [Google Scholar] [CrossRef]
  32. Zhao, X.; Sayed, A.H. Clustering via diffusion adaptation over networks. In Proceedings of the 2012 3rd International Workshop on Cognitive Information Processing (CIP), Parador de Baiona, Spain, 28–30 May 2012; pp. 1–6. [Google Scholar]
  33. Zhao, X.; Sayed, A.H. Distributed clustering and learning over networks. IEEE Trans. Signal Process. 2015, 63, 3285–3300. [Google Scholar] [CrossRef]
  34. Chen, J.; Richard, C.; Hero, A.O.; Sayed, A.H. Diffusion LMS for mu1titask problems with overlapping hypothesis subspaces. In Proceedings of the 2014 IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Reims, France, 21–24 September 2014; pp. 1–6. [Google Scholar]
  35. Bogdanovic, N.; Plata-Chaves, J.; Berberidis, K. Distributed diffusion-based LMS for node-specific parameter estimation over adaptive networks. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 7223–7227. [Google Scholar]
  36. Chen, J.; Richard, C.; Sayed, A.H. Multitask diffusion adaptation over networks. IEEE Trans. Signal Process. 2014, 62, 4129–4144. [Google Scholar] [CrossRef]
  37. Nassif, R.; Richard, C.; Ferrari, A. Proximal multitask learning over networks with sparsity-inducing coregularization. IEEE Trans. Signal Process. 2016, 64, 6329–6344. [Google Scholar] [CrossRef]
  38. Ma, W.; Chen, B.; Duan, J.; Zhao, H. Diffusion maximum correntropy criterion algorithms for robust distributed estimation. Digit. Signal Process. 2016, 58, 10–19. [Google Scholar] [CrossRef] [Green Version]
  39. Sayed, A.H. Adaptive Filters; Wiley: Hoboken, NJ, USA, 2008. [Google Scholar]
  40. Haykin, S. Adaptive Filter Theory; Prentice-Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  41. Kelley, C.T. Iterative Methods for Optimization; SIAM: Philadelphia, PA, USA, 1999. [Google Scholar]
  42. Abadir, K.M.; Magnus, J.R. Matrix Algebra; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  43. Chan, S.C.; Zou, Y.X. A recursive least M-estimate algorithm for robust adaptive filtering in impulsive noise: fast algorithm and convergence performance analysis. IEEE Trans. Signal Process. 2004, 52, 975–991. [Google Scholar] [CrossRef]
  44. Sayed, A.S.; Zoubir, A.M.; Sayed, A.H. Robust adaptation in impulsive noise. IEEE Trans. Signal Process. 2016, 64, 2851–2865. [Google Scholar] [CrossRef]
Figure 1. Curves of η ( i ) as a function of e ( i ) (a) different values of λ ( σ = η = 2.0 ) (b) different values of σ ( λ = η = 2.0 ).
Figure 1. Curves of η ( i ) as a function of e ( i ) (a) different values of λ ( σ = η = 2.0 ) (b) different values of σ ( λ = η = 2.0 ).
Sensors 19 02339 g001
Figure 2. Network topology.
Figure 2. Network topology.
Sensors 19 02339 g002
Figure 3. The variances of the input signal (a) and noise (b).
Figure 3. The variances of the input signal (a) and noise (b).
Sensors 19 02339 g003
Figure 4. Gaussian noise environment (a) desired signal (b) transient network MSD(dB).
Figure 4. Gaussian noise environment (a) desired signal (b) transient network MSD(dB).
Sensors 19 02339 g004
Figure 5. Impulsive interference environment of v i = 0.05 (a) desired signal (b) transient network MSD(dB).
Figure 5. Impulsive interference environment of v i = 0.05 (a) desired signal (b) transient network MSD(dB).
Sensors 19 02339 g005
Figure 6. Impulsive interference environment of v i = 0.03 (a) desired signal (b) transient network MSD(dB).
Figure 6. Impulsive interference environment of v i = 0.03 (a) desired signal (b) transient network MSD(dB).
Sensors 19 02339 g006
Figure 7. MSD learning curves in a non-stationary environment (a) Gaussian environment (b) Impulsive Interference.
Figure 7. MSD learning curves in a non-stationary environment (a) Gaussian environment (b) Impulsive Interference.
Sensors 19 02339 g007
Figure 8. Network MSD comparison over multitask environment.
Figure 8. Network MSD comparison over multitask environment.
Sensors 19 02339 g008
Figure 9. Network MSD comparison with different r value.
Figure 9. Network MSD comparison with different r value.
Sensors 19 02339 g009

Share and Cite

MDPI and ACS Style

Li, X.; Shi, Q.; Xiao, S.; Duan, S.; Chen, F. A Robust Diffusion Minimum Kernel Risk-Sensitive Loss Algorithm over Multitask Sensor Networks. Sensors 2019, 19, 2339. https://doi.org/10.3390/s19102339

AMA Style

Li X, Shi Q, Xiao S, Duan S, Chen F. A Robust Diffusion Minimum Kernel Risk-Sensitive Loss Algorithm over Multitask Sensor Networks. Sensors. 2019; 19(10):2339. https://doi.org/10.3390/s19102339

Chicago/Turabian Style

Li, Xinyu, Qing Shi, Shuangyi Xiao, Shukai Duan, and Feng Chen. 2019. "A Robust Diffusion Minimum Kernel Risk-Sensitive Loss Algorithm over Multitask Sensor Networks" Sensors 19, no. 10: 2339. https://doi.org/10.3390/s19102339

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop