Next Article in Journal
A Gloss Composition and Context Clustering Based Distributed Word Sense Representation Model
Previous Article in Journal
Numerical Investigation into Natural Convection and Entropy Generation in a Nanofluid-Filled U-Shaped Cavity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Letter

Proportionate Minimum Error Entropy Algorithm for Sparse System Identification

1
School of Electronic and Information Engineering, South China University of Technology, Guangzhou 510640, China
2
School of Electronic and Information Engineering, Xi'an Jiaotong University, Xi'an 710049, China
3
School of Electrical Engineering, Southwest Jiaotong University, Chengdu 610031, China
4
Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA
*
Author to whom correspondence should be addressed.
Entropy 2015, 17(9), 5995-6006; https://doi.org/10.3390/e17095995
Submission received: 22 May 2015 / Revised: 17 August 2015 / Accepted: 25 August 2015 / Published: 27 August 2015

Abstract

:
Sparse system identification has received a great deal of attention due to its broad applicability. The proportionate normalized least mean square (PNLMS) algorithm, as a popular tool, achieves excellent performance for sparse system identification. In previous studies, most of the cost functions used in proportionate-type sparse adaptive algorithms are based on the mean square error (MSE) criterion, which is optimal only when the measurement noise is Gaussian. However, this condition does not hold in most real-world environments. In this work, we use the minimum error entropy (MEE) criterion, an alternative to the conventional MSE criterion, to develop the proportionate minimum error entropy (PMEE) algorithm for sparse system identification, which may achieve much better performance than the MSE based methods especially in heavy-tailed non-Gaussian situations. Moreover, we analyze the convergence of the proposed algorithm and derive a sufficient condition that ensures the mean square convergence. Simulation results confirm the excellent performance of the new algorithm.
MSC Codes:
62B10

1. Introduction

Sparse system identification is an active research area at present, which finds various real-world applications in network echo cancelation, wireless multipath channels, underwater acoustic communications, and so on [1,2]. A system is qualitatively classified as a sparse system if only a small percentage of coefficients are active, while other coefficients are insignificant (i.e., equal or close to zero). It is worth noting that in compressive sensing and sparse coding, the term “sparse vector” usually means that most of the elements are exactly zero. In system identification and adaptive filtering, however, the term “sparse system (or filter)” in general means that most of the coefficients are equal or close to zero. For a sparse system, classic adaptive filtering algorithms like least mean square (LMS) and normalized LMS (NLMS) [3] may perform poorly in terms of steady state excess mean square error and convergence speed due to not using the a priori knowledge, especially in applications with long sparse systems. As a new scheme, the proportionate normalized least mean square (PNLMS) [4], which updates each filter coefficient in proportion to the magnitude of its estimate, has recently received a great deal of attention, and can perform much better than the conventional NLMS in the identification of sparse systems. Several improvements of the PNLMS algorithm have been proposed [5,6,7,8]. Moreover, several proportionate-type affine projection algorithms (APAs) were also developed [9,10,11,12].
Most of the existing proportionate-type adaptive algorithms (such as PNLMS) are developed based on the well-known mean square error (MSE) criterion. The MSE is computationally simple and mathematically tractable, and optimal when data are Gaussian. However, when the data are non-Gaussian (especially when data are disturbed by impulsive noises or containing large outliers), the MSE may be a poor descriptor of optimality. Man-made low frequency atmospheric noises and lighting spikes in natural phenomena can be described more accurately using non-Gaussian noise models [13,14]. From a statistical point of view, MSE only takes into account the second-order statistics, which is insufficient to capture all possible information from data. As a result, in non-Gaussian situations, the proportionate-type NLMS algorithms may perform poorly especially in the presence of impulsive noises.
Information theoretic learning (ITL) provides an appropriate framework for dealing with non-Gaussian signal processing [15,16]. In ITL, the quadratic Renyi’s entropy of the error was proposed as an alternative to MSE. With nonparametric Parzen window approach, the entropy can be easily estimated from the samples. Under the minimum error entropy (MEE) criterion, an adaptive system can be trained such that the error entropy between the model and unknown system is minimized [17,18,19,20,21,22,23,24,25]. Since entropy can capture higher-order statistics and information content of signals rather than simply their energy, the MEE based adaptive algorithms may achieve significant performance improvements in non-Gaussian situations. In this work, we propose a novel proportionate algorithm for sparse system identification, called the proportionate minimum error entropy (PMEE) algorithm. Instead of using the MSE criterion, the new algorithm is derived based on the MEE criterion. The PMEE algorithm may perform much better than the PNLMS when identifying a sparse system with non-Gaussian noises. In a recent paper [26], we proposed three sparse adaptive filtering algorithms under MEE criterion, namely ZAMEE, RZAMEE, and CIMMEE, which are derived by incorporating a sparsity penalty term into the MEE criterion. These algorithms also perform well for sparse system identification with non-Gaussian noises. However, simulation results in this work show that PMEE can outperform them.
The rest of the paper is organized as follows. In Section 2, after briefly introducing the MEE criterion, we derive the PMEE algorithm. In Section 3, we carry out the mean square convergence analysis. In Section 4, we present simulation results to confirm the excellent performance of the PMEE. Finally, In Section 5, we give the conclusion.

2. Proportionate Minimum Error Entropy Algorithm

2.1. Minimum Error Entropy Criterion

Figure 1 depicts an adaptive filtering scheme under MEE criterion. According to Figure 1, the adaptive filtering can be formulated as minimizing the error entropy between the filter output and the desired response. Since entropy quantifies the average uncertainty or dispersion of a random variable, its minimization makes the error concentrated. Consider a linear system where the desired signal is generated by:
d ( n ) = W * T X ( n ) + v ( n )
where X ( n ) = [ x n M + 1 , , x n 1 , x n ) ] T denotes the input vector at instant n , W * = [ w 1 * , w 2 * , , w M * ] T denotes the weight (parameter) vector of an finite impulse response (FIR) channel with M being the memory size, Τ denotes the transpose operator, and v ( n ) stands for the interference or measurement noise. Assume that the adaptive filter is also an FIR filter with weight vector W ( n ) = [ w 1 ( n ) , w 2 ( n ) , , w M ( n ) ] T . Then the filtering error is:
e ( n ) = d ( n ) y ( n ) = d ( n ) W T ( n ) X ( n )
where y ( n ) is the output of the adaptive filter at instant n . Let the filtering error be a random variable with probability density function (PDF) p e ( . ) . The quadratic Renyi’s entropy of error is:
H R 2 ( e ) = log p e 2 ( ξ ) d ξ = log V ( e )
where V ( e ) = p e 2 ( ξ ) d ξ is named the quadratic information potential (QIP) [17,18]. In practical applications, however, an analytical expression of the error entropy is not available in general; one has to estimate it directly from the error samples. By the Parzen window approach, the error PDF can be estimated as:
p e ( e ) = 1 N i = 1 N κ σ ( e e ( i ) )
where N is the samples number, κ σ denotes a kernel function with bandwidth σ , satisfying κ σ ( x ) 0 , and κ σ ( x ) d x = 1 . The most popular kernel function used in ITL is the Gaussian kernel:
κ σ ( x ) = 1 σ 2 π exp ( x 2 2 σ 2 )
Figure 1. Adaptive filtering under MEE criterion.
Figure 1. Adaptive filtering under MEE criterion.
Entropy 17 05995 g001
In the rest of the paper, unless mentioned otherwise, we will use the Gaussian kernel. Combining Equations (3) and (4) yields:
H R 2 ( e ) log ( 1 N i = 1 N κ σ ( e e ( i ) ) ) 2 d e             = log 1 N 2 i = 1 N j = 1 N κ σ ( e e ( i ) ) κ σ ( e e ( j ) ) d e             = log 1 N 2 i = 1 N j = 1 N κ σ 2 ( e ( i ) e ( j ) )
One can easily obtain:
V ( e ) 1 N 2 i = 1 N j = 1 N κ σ 2 ( e ( i ) e ( j ) )
Obviously, minimizing the quadratic Renyi entropy is equivalent to maximizing the QIP. Thus, the optimal weight vector under MEE can be formulated as:
W o p t = arg max W V ( e )

2.2. Proportionate Minimum Error Entropy

Before presenting the PMEE algorithm, a general form of the PNLMS-type algorithms is revisited. Generally, the weight update equation of the PNLMS-type algorithms can be expressed as [4,8]:
W ( n + 1 ) = W ( n ) + μ G ( n ) e ( n ) X ( n ) X T ( n ) G ( n ) X ( n ) + δ
where μ is a step size parameter, δ > 0 is a regularization parameter that prevents division by zero in Equation (9) and stabilizes the solution, G ( n ) is a diagonal matrix that modifies the step size of each tap according to a specific rule. In general, the matrix G ( n ) is given by:
G ( n ) = d i a g ( g 1 ( n ) , g 2 ( n ) , , g M ( n ) )
where:
g l ( n ) = ϖ l ( n ) i = 1 M ϖ i ( n )        1 l M
ϖ l ( n ) = max [ ε max { ϕ ,   | w 1 ( n ) | , | w M ( n ) | } ,   | w l ( n ) | ]
The parameter ε prevents the coefficients from stalling when they are much smaller than the largest one. The parameter ϕ is an initialization parameter that helps to prevent stalling of the weight updating at the initial stage when all the taps are initialized to zero.
To develop the PMEE algorithm for sparse system identification, we use the error entropy instead of the squared error as the adaptation cost. According to Equation (8), a steepest ascent algorithm for estimating the weight vector can be derived as:
W ( n + 1 ) = W ( n ) + μ V ( e ( n ) )
where V ( e ( n ) ) denotes the gradient of the QIP with respect to the weight vector, given by:
V ( e ( n ) ) = V ( e ( n ) ) W ( n ) = W ( n ) ( 1 L 2 i = n L + 1 n j = n L + 1 n κ σ 2 ( Δ e ( i , j ) ) )                 = 1 2 L 2 σ 2 i = n L + 1 n j = n L + 1 n [ κ σ 2 ( Δ e ( i , j ) ) ( Δ e ( i , j ) ) ( y ( i ) W ( n ) y ( j ) W ( n ) ) ]                 = 1 2 L 2 σ 2 i = n L + 1 n j = n L + 1 n [ κ σ 2 ( Δ e ( i , j ) ) ( Δ e ( i , j ) ) ( X ( i ) X ( j ) ) ]
where Δ e ( i , j ) = e ( i ) e ( j ) , and L denotes the sliding data length. Hence, inspired by the PNLMS-type algorithms, we propose the following weight update equation:
W ( n + 1 ) = W ( n ) + μ G ( n ) V ( e ( n ) ) = W ( n ) + G ( n ) μ 2 L 2 σ 2 i = n L + 1 n j = n L + 1 n [ κ σ 2 ( Δ e ( i , j ) ) ( Δ e ( i , j ) ) ( X ( i ) X ( j ) ) ]
where G ( n ) is determined by Equations (10)–(12). This algorithm is referred to as the PMEE algorithm.
Remark 1. 
Obviously, one can also propose a normalized version of PMEE by dividing the term X T ( n ) G ( n ) X ( n ) + δ , just like Equation (9). However, our simulation results indicate that the normalized PMEE performs well only when the underlying system is extremely sparse. Thus, in this work, we don’t consider the normalized PMEE. Maybe other updating laws may exist that can provide possibly better result, but this is beyond the scope of this work.

3. Mean Square Convergence Analysis

3.1. Energy Conservation Relation

We rewrite Equation (2) in a form of block data:
e ( n ) = d ( n ) y ( n ) = d ( n ) χ ( n ) W ( n )
where e ( n ) = [ e ( n L + 1 ) , , e ( n ) ] T , d ( n ) = [ d ( n L + 1 ) , , d ( n ) ] T , y ( n ) = [ y ( n L + 1 ) , , y ( n ) ] T , and χ ( n ) = [ X ( n L + 1 ) , X ( n L + 2 ) , , X ( n ) ] T stands for an L × M input matrix. Define the a priori error vector e a ( n ) and a posteriori error vector e p ( n ) as follows:
{       e a ( n ) = [ e a ( n L + 1 ) , , e a ( n ) ] T = χ ( n ) W ˜ ( n )                   e p ( n ) = [ e p ( n L + 1 ) , , e p ( n ) ] T = χ ( n ) W ˜ ( n + 1 )
where W ˜ ( n ) = W * W ( n ) is the weight error vector. Then, e a ( n ) and e p ( n ) satisfies:
e p ( n ) = e a ( n ) + χ ( n ) ( W ˜ ( n + 1 ) W ˜ ( n ) ) = e a ( n ) χ ( n ) ( W ( n + 1 ) W ( n ) )
Now Equation (15) can be rewritten as:
W ( n + 1 ) = W ( n ) + μ G ( n ) χ T ( n ) h ( e ( n ) )
where h ( e ( n ) ) = [ h 1 ( e ( n ) ) , h 2 ( e ( n ) ) , , h L ( e ( n ) ) ] T , in which:
h i ( e ( n ) ) = J M E E ( e ( n ) ) e ( n L + i )
To simplify the analysis, below we assume L = M . Combining Equations (18) and (19), one can derive:
     e p ( n ) = e a ( n ) μχ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) e p ( n ) e a ( n ) = μχ ( n ) G ( n ) ( χ ( n ) G ( n ) ) T h ( e ( n ) ) e p ( n ) e a ( n ) = μ Η ( n ) Η T ( n ) h ( e ( n ) ) R 1 ( n ) ( e p ( n ) e a ( n ) ) = μ h ( e ( n ) ) G ( n ) χ T ( n ) R 1 ( n ) ( e p ( n ) e a ( n ) ) = ( W ( n + 1 ) W ( n ) ) G ( n ) χ T ( n ) R 1 ( n ) ( e p ( n ) e a ( n ) ) = W ˜ ( n + 1 ) W ˜ ( n )
where G ( n ) = d i a g ( g 1 ( n ) , g 2 ( n ) , , g M ( n ) ) , Η ( n ) = χ ( n ) G ( n ) , and R ( n ) = Η ( n ) Η T ( n ) is an L × L -dimensional symmetric matrix. Assume that R ( n ) is invertible. Then we obtain:
W ˜ ( n + 1 ) = W ˜ ( n ) + G ( n ) χ T ( n ) R 1 ( n ) ( e p ( n ) e a ( n ) )
Squaring both sides of Equation (22), we have:
W ˜ T ( n + 1 ) W ˜ ( n + 1 ) = [ W ˜ ( n ) + G ( n ) χ T ( n ) R 1 ( n ) ( e p ( n ) e a ( n ) ) ] T                                    × [ W ˜ ( n ) + G ( n ) χ T ( n ) R 1 ( n ) ( e p ( n ) e a ( n ) ) ]
After some simple manipulations, we derive:
W ˜ ( n + 1 ) 2 + e a ( n ) 1 ( n ) 2 = W ˜ ( n ) 2 + e p ( n ) 1 ( n ) 2
where ( n ) = χ ( n ) χ T ( n ) , W ˜ ( n ) 2 = W ˜ T ( n ) W ˜ ( n ) , e a ( n ) 1 ( n ) 2 = e a T ( n ) 1 ( n ) e a ( n ) , and e p ( n ) 1 ( n ) 2 = e p T ( n ) 1 ( n ) e p ( n ) . Taking the expectations of the both sides of Equation (24) leads to the energy conservation relation [19,23,26]:
E [ W ˜ ( n + 1 ) 2 ] + E [ e a ( n ) 1 ( n ) 2 ] = E [ W ˜ ( n ) 2 ] + E [ e p ( n ) 1 ( n ) 2 ]
where E [ W ˜ ( n ) 2 ] is called the weight error power (WEP) at iteration n .

3.2. Sufficient Condition for Mean Square Convergence

Based on the energy conservation relation Equation (25), a sufficient condition that guarantees the mean square convergence (i.e., the monotonic decrease of the WEP) can be easily derived. Substituting e p ( n ) = e a ( n ) μχ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) into Equation (25), we have:
E [ W ˜ ( n + 1 ) 2 ]   = E [ W ˜ ( n ) 2 ] 2 μ E [ e a T ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) ]                               + μ 2 E [ h T ( e ( n ) ) χ T ( n ) G ( n ) χ ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) ]
It follows that:
         E [ W ˜ ( n + 1 ) 2 ] E [ W ˜ ( n ) 2 ]    μ 2 E [ h T ( e ( n ) ) χ T ( n ) G ( n ) χ ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) ]      2 μ E [ e a T ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) ]   μ 2 E [ ϒ ] E [ Θ ]
where
ϒ = e a T ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) ,
Θ = h T ( e ( n ) ) χ T ( n ) G ( n ) χ ( n ) 1 ( n ) χ ( n ) G ( n ) χ T ( n ) h ( e ( n ) ) .
Since μ 0 , a sufficient condition for the mean square convergence will be:
{ E [ ϒ ] > 0 0 < μ 2 E [ ϒ ] E [ Θ ]
The above sufficient condition ensures that the WEP will be monotonically decreasing (hence the algorithm will not diverge).

4. Simulation Results

Now we present simulation results about sparse system identification to demonstrate the performance of the PMEE, compared with ZAMEE, RZAMEE, CIMMEE [26] and PNLMS. The mean square deviation (MSD) is used as the performance index, calculated by:
MSD= E [ W * W ( n ) 2 ]
In order to show the performance of the algorithms in impulsive noise environments, we adopt the alpha-stable distribution [27,28] to generate the disturbance noise, whose characteristic function is:
f ( t ) = exp { t γ | t | α [ 1 + sgn ( t ) S ( t , α ) ] }
in which:
S ( t , α ) = { tan απ 2 , if α 1 2 π log | t | , if α = 1
where α ( 0 , 2 ] is the characteristic factor, < δ < + is the location parameter, β [ 1 , 1 ] ; is the symmetry parameter, and γ > 0 is the dispersion parameter. This distribution is called a symmetric alpha-stable ( S α S ) distribution when β = 0 . We define the parameter vector V = ( α , β , γ , δ ) . In addition, we consider four unknown channels with different parameter vectors:
(a) 
Channel 1:
W * = [ 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ]
(b) 
Channel 2:
W * = [ 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 1 ]
(c) 
Channel 3:
W * = [ 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 , 0 , 0 , 0 , 1 ]
(d) 
Channel 4:
W * = [ 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 ]
Clearly, the four channels (with memory size 40) have different sparsity ratio (1/40, 1/8, 1/4, 1). In all the simulations below, the input signal is a white Gaussian process with zero mean and unit variance. The parameters ε and ϕ are set at 1/40 and 0.001, respectively. The sliding data length is L = 20 . Simulation results are averaged over 200 independent Monte Carlo runs, and each run has 15,000 iterations.
In the first simulation, the parameter vector V is set at ( 1.2 , 0 , 0.4 , 0 ) , the kernel width σ is 1.0, and the step sizes are chosen such that all the algorithms have almost the same initial convergence rate. Figure 2 illustrates the convergence curves for different channels. One can observe: (i) when the system is very sparse (e.g., channel 1 and 2), the PMEE achieves much lower steady-state MSDs than other algorithms; (ii) when the channel is less sparse (e.g., channel 3), the steady-state MSDs of PMEE is still the lowest although there is a little loss in performance; (iii) when the system is completely non-sparse (e.g., channel 4), the performance of PMEE is comparable with CIMMEE. Note that in all the cases, the PNLMS algorithm cannot work since it is very sensitive to impulsive noises.
In the second simulation, we show the performance of the algorithms with different noise parameters (assume that the unknown system is the channel 1). The steady-state MSDs with different α (0.8,1.0,1.2,1.4,1.6,1.8,2.0) and different γ (0.2,0.4,0.6,0.8,1.0,1.2,1.4,1.6) are shown in Figure 3 and Figure 4, respectively. Evidently, the PMEE performs very well and achieves the lowest MSDs compared with ZAMEE, RZAMEE, and CIMMEE. The PNLMS performs well (even better than PMEE) only when α is very close to 2.0. The main reason for this is that, when α comes near to 2.0, the noise will be approximately Gaussian. Simulation results confirm that the PMEE can effectively identify a sparse system in a non-Gaussian impulsive noise environment.
Figure 2. Convergence curves for different channels: (a) channel 1; (b) channel 2; (c) channel 3; (d) channel 4.
Figure 2. Convergence curves for different channels: (a) channel 1; (b) channel 2; (c) channel 3; (d) channel 4.
Entropy 17 05995 g002
In the third simulation, we investigate how the kernel width affects the convergence performance. The noise parameters are set at V = ( 1.2 , 0 , 0.4 , 0 ) . Simulation results are shown in Figure 5, from which one can see that the kernel width has significant influence on the convergence performance. When the kernel width is too large or too small, the performance will become poor. In a practical application, the kernel width can be manually selected or optimized by trial and error.
Figure 3. Steady-state MSDs with different α ( γ = 0.4 ).
Figure 3. Steady-state MSDs with different α ( γ = 0.4 ).
Entropy 17 05995 g003
Figure 4. Steady-state MSDs with different γ ( α = 1.2 ).
Figure 4. Steady-state MSDs with different γ ( α = 1.2 ).
Entropy 17 05995 g004
Figure 5. Convergence curves with different kernel widths.
Figure 5. Convergence curves with different kernel widths.
Entropy 17 05995 g005

5. Conclusions

In this work, the proportionate minimum error entropy (PMEE) algorithm has been developed to identify a sparse system. Different from the existing proportionate-type adaptive filtering algorithms, such as the proportionate normalized least mean square (PNLMS), PMEE is derived by using the minimum error entropy (MEE) instead of the traditional mean square error (MSE) as the adaptation criterion. Convergence analysis based on energy conservation relation has been carried out, and a sufficient condition for ensuring the mean square stability is obtained. Simulation results have demonstrated the superior performance of the proposed algorithm especially in impulsive non-Gaussian situations.

Acknowledgments

This work was supported by 973 Program (No. 2015CB351703) and National Natural Science Foundation of China (No. 61271210, No. 61372152).

Author Contributions

Zongze Wu derived the algorithm and finished the draft; Siyuan Peng conducted the simulations; Badong Chen put forward the ideas and proved the main theorem; Haiquan Zhao and Jose C. Principe polished the language and were in charge of technical checking. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, Y.; Benesty, J.; Chen, J. Acoustic MIMO Signal Processing; Springer: New York, NY, USA, 2006. [Google Scholar]
  2. Paleologu, C.; Benesty, J.; Ciochina, S. Sparse Adaptive Filters for Echo Cancellation; Morgan and Claypool: San Rafael, CA, USA, 2010. [Google Scholar]
  3. Sayed, A.H. Fundamentals of Adaptive Filtering; Wiley: Hoboken, NJ, USA, 2003. [Google Scholar]
  4. Duttweiler, D. Proportionate normalized least-mean-squares adaptation in echo cancellers. IEEE Trans. Speech Audio Process. 2000, 8, 508–518. [Google Scholar] [CrossRef]
  5. Deng, H.; Doroslovacki, M. Improving convergence of the PNLMS algorithm for sparse impulse response identification. IEEE Signal Process. Lett. 2005, 12, 181–184. [Google Scholar] [CrossRef]
  6. Laska, B.N.M.; Goubran, R.A.; Bolic, M. Improved proportionate subband NLMS for acoustic echo cancellation in changing environments. IEEE Signal Process. Lett. 2008, 15, 337–340. [Google Scholar] [CrossRef]
  7. Das Chagas de Souza, F.; Seara, R.; Morgan, D.R. An enhanced IAF-PNLMS adaptive algorithm for sparse impulse response identification. IEEE Trans. Signal Process. 2012, 60, 3301–3307. [Google Scholar] [CrossRef]
  8. Deng, H.; Doroslovacki, M. Proportionate adaptive algorithms for network echo cancellation. IEEE Trans. Signal Process. 2006, 54, 1794–1803. [Google Scholar] [CrossRef]
  9. Paleologu, C.; Ciochină, S.; Benesty, J. An efficient proportionate affine projection algorithm for echo cancellation. IEEE Signal Process. Lett. 2010, 17, 165–168. [Google Scholar] [CrossRef]
  10. Yang, J.; Sobelman, G.E. Efficient μ-law improved proportionate affine projection algorithm for echo cancellation. Electron. Lett. 2010, 47, 73–74. [Google Scholar] [CrossRef]
  11. Yang, Z.; Zheng, Y.R.; Grant, S.L. Proportionate affine projection sign algorithms for network echo cancellation. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2273–2284. [Google Scholar] [CrossRef]
  12. Zhao, H.; Yu, Y.; Gao, S.; Zeng, X.; He, Z. Memory proportionate APA with individual activation factors for acoustic echo cancellation. IEEE Trans. Audio Speech Lang. Process. 2014, 22, 1047–1055. [Google Scholar] [CrossRef]
  13. Plataniotis, K.N.; Androutsos, D.; Venetsanopoulos, A.N. Nonlinear filtering of non-Gaussian noise. J. Intell. Robot. Syst. 1997, 19, 207–231. [Google Scholar] [CrossRef]
  14. Weng, B.; Barner, K.E. Nonlinear system identification in impulsive environments. IEEE Trans. Signal Process. 2005, 53, 2588–2594. [Google Scholar] [CrossRef]
  15. Principe, J.C. Information Theoretic Learning: Renyi’s Entropy and Kernel Perspectives; Springer: New York, NY, USA, 2010. [Google Scholar]
  16. Chen, B.; Zhu, Y.; Hu, J.C.; Jose Principe, J.C. System Parameter Identification: Information Criteria and Algorithms; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  17. Erdogmus, D.; Principe, J.C. An error-entropy minimization for supervised training of nonlinear adaptive systems. IEEE Trans. Signal Process. 2002, 50, 1780–1786. [Google Scholar] [CrossRef]
  18. Erdogmus, D.; Principe, J.C. Convergence properties and data efficiency of the minimum error entropy criterion in adaline training. IEEE Trans. Signal Process. 2003, 51, 1966–1978. [Google Scholar] [CrossRef]
  19. Chen, B.; Zhu, Y.; Hu, J. Mean-square convergence analysis of ADALINE training with minimum error entropy criterion. IEEE Trans. Neural Netw. 2010, 21, 1168–1179. [Google Scholar] [CrossRef] [PubMed]
  20. Chen, B.; Hu, J.; Pu, L.; Sun, Z. Stochastic gradient algorithm under (h, phi)-entropy criterion. Circuit Syst. Signal Process. 2007, 26, 941–960. [Google Scholar] [CrossRef]
  21. Li, C.; Shen, P.; Liu, Y.; Zhang, Z. Diffusion information theoretic learning for distributed estimation over network. IEEE Trans. Signal Process. 2013, 61, 4011–4024. [Google Scholar] [CrossRef]
  22. Wolsztynski, E.; Thierry, E.; Pronzato, L. Minimum-entropy estimation in semi-parametric models. Signal Process. 2005, 85, 937–949. [Google Scholar] [CrossRef]
  23. Al-Naffouri, T.Y.; Sayed, A.H. Adaptive filters with error nonlinearities: Mean-square analysis and optimum design. EURASIP J. Appl. Signal Process. 2001, 4, 192–205. [Google Scholar] [CrossRef]
  24. Chen, B.; Principe, J.C. Some further results on the minimum error entropy estimation. Entropy 2012, 14, 966–977. [Google Scholar] [CrossRef]
  25. Chen, B.; Principe, J.C. On the Smoothed Minimum Error Entropy Criterion. Entropy 2012, 14, 2311–2323. [Google Scholar] [CrossRef]
  26. Wu, Z.; Peng, S.; Ma, W.; Chen, B.; Principe, J.C. Minimum Error Entropy Algorithm with Sparsity Penalty Constraints. Entropy 2015, 17, 3419–3437. [Google Scholar] [CrossRef]
  27. Weng, B.; Barner, K.E. Nonlinear system identification in impulsive environments. IEEE Trans. Signal Process. 2005, 53, 2588–2594. [Google Scholar] [CrossRef]
  28. Wang, J.; Kuruoglu, E.E.; Zhou, T. Alpha-stable channel capacity. IEEE Commun. Lett. 2011, 15, 1107–1109. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Wu, Z.; Peng, S.; Chen, B.; Zhao, H.; Principe, J.C. Proportionate Minimum Error Entropy Algorithm for Sparse System Identification. Entropy 2015, 17, 5995-6006. https://doi.org/10.3390/e17095995

AMA Style

Wu Z, Peng S, Chen B, Zhao H, Principe JC. Proportionate Minimum Error Entropy Algorithm for Sparse System Identification. Entropy. 2015; 17(9):5995-6006. https://doi.org/10.3390/e17095995

Chicago/Turabian Style

Wu, Zongze, Siyuan Peng, Badong Chen, Haiquan Zhao, and Jose C. Principe. 2015. "Proportionate Minimum Error Entropy Algorithm for Sparse System Identification" Entropy 17, no. 9: 5995-6006. https://doi.org/10.3390/e17095995

Article Metrics

Back to TopTop