1. Introduction
In many real-world applications, adaptive filters of all sorts play a fundamental role [
1]. Basically, such an adaptive system consists of a digital filter with time-varying coefficients, which are updated based on an iterative method, also known as the adaptive algorithm [
2]. The choice of a particular adaptive algorithm for a specific application depends on several performance criteria, like the convergence rate, solution accuracy, robustness capability, and computational efficiency. While the least-mean-square (LMS) algorithm remains one of the most popular choices due to its simplicity, the improved performance features support using the recursive least-squares (RLS) algorithm for nowadays applications [
3].
The length of the adaptive filter (i.e., the number of coefficients) is a critical factor related to the practical implementation of such a system. The longer the filter, the more challenging its implementation/operation becomes. This aspect is not only associated with the computational complexity, which naturally increases with the number of coefficients, but also with the main performance criteria. A longer filter implies a slower convergence rate and tracking while also influencing the accuracy of the solution. Since there are many applications that require adaptive filters with hundreds of coefficients (e.g., echo cancelation, channel modeling, noise reduction, etc.), dealing with such long-length impulse responses represents an important challenge, especially in conjunction with the RLS-based algorithms [
4,
5].
Recent works have addressed this problem by focusing on impulse response decomposition techniques, which aim to reformulate a problem with a large parameter space (i.e., a large number of filter coefficients) as a combination of “smaller” problems, i.e., shorter filters [
6,
7,
8,
9,
10,
11,
12]. In this framework, one of the most appealing solutions relies on the nearest Kronecker product (NKP) decomposition of the impulse response, which exploits the low-rank features that characterize most of the real-world systems [
13,
14,
15,
16,
17,
18,
19,
20,
21].
The basis of the NKP approach was developed in the context of a second-order decomposition [
13,
14], where an impulse response of length
(with
) is approximated by a sum of
R Kronecker products (with
) between shorter impulse responses of lengths
and
. In fact,
R is the approximated rank of the
matrix that results from reshaping the impulse response of length
L. Apparently, a higher-order decomposition is desirable since it would involve even shorter impulse responses. Nevertheless, in this case, the impulse response would be reshaped as a higher-order tensor, while the rank of such a structure is not always straightforward to handle [
22,
23,
24]. The solution proposed in [
20] avoids this issue and controls a third-order tensor’s rank via a matrix rank. As a result, the RLS-based algorithm using such a third-order tensor (TOT) decomposition, namely RLS-TOT [
21], outperforms its previously developed counterpart that exploits the second-order NKP decomposition [
14].
Besides the previously mentioned performance criteria (e.g., convergence, tracking, and accuracy), there is another important aspect that should be taken into account when the system operates in noisy environments. In such scenarios, the robustness of the adaptive algorithm becomes a critical issue. For example, in echo cancelation, the microphone captures the background noise, which could be strong and nonstationary. Consequently, its presence could significantly bias the adaptive filter behavior if the algorithm is not equipped with strong robustness features. Usually, the robustness of the algorithm is considered in the design phase by including specific terms into the optimization cost function [
25,
26,
27,
28,
29,
30]. Nevertheless, the resulting regularized algorithms frequently include additional parameters that are not always easy to control in practice. In this context, the solution recently proposed in [
31] considers a cost function that includes the contribution of both the external noise and the system uncertainties, while the variances of these terms are estimated in a practical manner within the algorithm.
The RLS-TOT algorithm developed in [
21] exploits the TOT-based decomposition of the global impulse response of the filter (which could have a large number of coefficients). As a result, it solves the associated system identification problem by combining the estimates provided by three (usually much shorter) adaptive filters. Consequently, the resulting gain is twofold, in terms of performance assets (e.g., faster tracking and better accuracy) and lower computational complexity. However, the cost functions related to the three component filters follow the line of the conventional RLS algorithm and do not include any regularization component. Thus, the so-called forgetting factors [
21] remain the only control parameters of this TOT-based algorithm, so that its robustness is inherently limited in noisy environments.
Recently, a robust RLS-type algorithm was developed in [
31], following the line of the conventional RLS version (without decomposition), but including a proper regularization term within the cost function. This regularization technique considers the influence of two important factors that could significantly influence the overall performance in real-world environments. First, it includes the external noise power, which is further estimated in a practical manner within the regularized RLS-type algorithm (based on the error signal). Second, it considers the so-called model “uncertainties” [
31], which are related to the variability of the unknown system to be identified. The resulting parameter can also be estimated in a practical way by using a measure related to the Euclidean norm of the difference between the estimates provided by the adaptive filter from two consecutive moments of time. As a result, the algorithm proposed in [
31] has a significantly improved robustness in noisy conditions, as compared to the non-regularized counterparts. Nevertheless, it does not involve any decomposition-based technique; thus, it still faces the inherent challenges related to the identification of long impulse responses.
Motivated by these aspects, in this article, we follow the idea from [
31] in order to develop a regularized version of the RLS-TOT algorithm, with improved robustness features compared to its counterpart from [
21]. As a result, the proposed algorithm inherits the advantages of the decomposition-based approach and the robust behavior due to the regularization technique. Following this introduction, the TOT-based decomposition is presented in
Section 2. Then, the proposed regularized RLS-TOT algorithm is developed in
Section 3. The experimental scenario presented in
Section 4 considers an echo cancelation application, and the obtained results support the performance features of the proposed solution. A brief discussion related to the strengths and limitations of the regularized RLS-TOT algorithm, together with some potential extensions, is provided in
Section 5. The paper concludes in
Section 6, also outlining several perspectives for future research.
2. TOT Decomposition
In this section, the background behind the TOT decomposition is presented, together with the specific framework and notation. Let us consider a time-varying system characterized by the impulse response of length
L, denoted by the vector:
where
n represents the discrete-time index,
is the
lth coefficient (with
), and the superscript
⊤ stands for transposition. Next, it is assumed that the length can be factorized as
, with
. This may not seem like a trivial assumption, but in many finite-precision implementations, the length of a digital filter is usually set as a power of two (for practical/algorithmic reasons), i.e.,
, where
B is a positive integer. Consequently, it can be easily factorized as
, with
and
(where
and
are also positive integers).
Using the Kronecker product, the impulse response from (
1) can be equivalently expressed as [
13]
where
and
, with
, are shorter impulse responses of lengths
and
, respectively, while ⊗ denotes the Kronecker product [
32]. As shown in [
13], the coefficients of these shorter impulse responses can be evaluated using the singular value decomposition (SVD) of the matrix
, of size
, which is obtained by reshaping the elements of
. Nevertheless, in terms of the parameter space, there is no reason to use (
2) instead of (
1), since the former involves
coefficients, which exceed the length
L associated with (
1). However, for low-rank systems, i.e., the rank of
is smaller than
, there is a potential gain. Indeed, let us denote by
R the rank of
, with
. In this case, using the NKP decomposition [
13], (
2) becomes
where
and
, with
, are impulse responses of lengths
and
, respectively. Thus, the parameter space consists of
coefficients, a number that can be significantly smaller than
L. In fact, in many practical scenarios (e.g., echo cancelation, noise reduction, etc.), the setup is
[
13,
14], thus leading to important advantages.
Apparently, increasing the decomposition order would be a benefit in terms of reducing the parameter space. In this case, the length
L would be factorized in more than two terms, while more than two (even shorter) impulse responses would be combined via the Kronecker product in the decomposition of
, similar to (
2). Nevertheless, the number of terms in the sum would not be related to a matrix rank anymore, but to a tensor rank, which is a more sensitive issue [
22,
23,
24]. For example, while the rank of a matrix is limited by the number of its columns, a tensor rank could exceed the largest dimension of its decomposition.
In order to avoid such shortcomings, the TOT-based solution from [
20] starts with the factorization
, but considering that
, while also decomposing
, with
. Then, following the same low-rank framework, the decomposition of the impulse response results in [
20]
with
, and the shorter impulse responses
,
, and
having the lengths
,
, and
, respectively. Hence, according to (
4), the parameter space becomes
, which can be significantly smaller than
L, under the common settings
and
[
20,
21].
To facilitate the upcoming developments, the coefficients of the component impulse responses from (
4) are grouped into the following data structures, which are specific to the TOT decomposition:
with
. It should also be noticed that the terms that compose the sums in (
4) can be equivalently expressed as
where
,
, and
denote the identity matrices of size
,
, and
, respectively. These equivalent forms are useful to further extract the individual components of the global impulse response, i.e., the last terms that appear in the right-hand side of (
8), (
9), and (
10), respectively. In fact, these sets of coefficients compose the impulse responses from (
5)–(
7), i.e.,
,
, and
, which have the lengths
,
, and
, respectively. Concluding, the adaptive filters that exploit the TOT-based decomposition are designed to estimate these three component impulse responses, instead of directly modeling the global impulse response of length
L. Furthermore, the coefficients of the estimated (full-length) impulse response can be obtained similarly to (
4), as will be shown in the next section.
3. RLS-TOT Algorithm and Its Regularized Version
An adaptive filter consists of a digital filter controlled by an iterative algorithm, which updates its coefficients in order to minimize a certain cost function that depends on the error signal [
1,
2,
3]. This error is defined as the difference between a reference (or desired) signal,
, and the output of the adaptive filter. Let us consider that the reference signal is obtained at the output of an unknown system, which is defined by the impulse response
of length
L, as introduced at the beginning of
Section 2. Also, it is considered that a zero-mean additive noise,
, corrupts this output, so that the reference signal results in
where
contains the most recent
L samples of the zero-mean input signal
, which is independent of
. This signal also drives the input of the adaptive filter, with the objective of identifying the unknown impulse response and delivering its estimate, denoted by
. Consequently, the a priori error signal is evaluated using the estimated coefficients from time index
, i.e.,
while the common form of the filter update is
where
generally denotes the update term, which depends on the input vector and the error signal. This update term results in different forms, depending on the adaptive filtering algorithm.
In this framework, the least-squares (LS) optimization criterion [
1,
2,
3] represents one of the most reliable approaches, which targets the minimization of the cost function:
with respect to
, where
represents the forgetting factor (
). Following this approach, the normal equations to be solved result in
where
The solution can be formulated in a recursive manner, which leads to the well-known RLS algorithm that is defined by the update relation [
1,
2,
3]:
where
is the a priori error signal defined in (
13). In this context,
from (
17) represents an estimate of the input signal covariance matrix.
There are different versions of the RLS algorithm that avoid a direct matrix inversion in (
19), e.g., by using the matrix inversion lemma (also known as the Woodbury formula) [
1,
2,
3] or the line search methods [
33,
34,
35,
36,
37]. The first approach leads to the most popular version of the RLS algorithm, in which the inverse of
is recursively updated. The computational amount required by this version is proportional to
. Nevertheless, the second approach that exploits the line search methods is more advantageous in terms of computational complexity. In this case, following (
19), the filter update is rewritten as
, where the “increment” (i.e., the update term) results as the iterative solution of an auxiliary set of normal equations,
, where
is the so-called residual vector. Among these iterative techniques, the dichotomous coordinate descent (DCD) method proposed by Zakharov et al. [
34] stands out as the most representative, which leads to a computational complexity of the resulting RLS-DCD algorithm proportional to
. In this context, we should note that different “fast” RLS algorithms can be found in the literature, for various filter structures [
1,
2,
3,
5]. Their analysis is beyond the scope of this paper, which follows the basic form of the RLS-based update from (
19), for the sake of generality.
In the context of the TOT-based decomposition presented in
Section 2, the estimate provided by the adaptive filter can be decomposed into a form similar to (
4), i.e.,
where the impulse responses
,
, and
have the lengths
,
, and
, respectively. Furthermore, similar to (
5)–(
7), we can obtain the component filters
,
, and
, of lengths
,
, and
, respectively. As a result, based on the properties of the Kronecker product, like in (
8)–(
10), each of these three component filters can be processed separately [
20,
21].
First, focusing on
, which contains the taps of
, with
, the other two sets of coefficients can be grouped into the following matrices:
where
. Consequently, a cost function similar to (
15) can be formulated as
where
is a forgetting factor. The minimization of
with respect to
relies on a multilinear optimization strategy [
38], which considers that the other two component filters are fixed. Following similar steps that led to (
16)–(
19) and introducing the notation
, the relations that define the RLS-based algorithm associated with this component filter result in
In a similar manner, the optimization process is carried out for the second component filter,
, which groups the taps of
, with
and
. Starting with the notation:
the related cost function becomes
with the associated forgetting factor
. Using the same multilinear optimization strategy, the RLS-type algorithm for the second component filter is defined by the following relations:
with
.
The development of the third component filter follows the same path as the second one. The purpose is to update
, which contains the coefficients
, with
and
. Therefore, the associated notation is
which allows us to define the cost function
with the forgetting factor
. Finally, the minimization of
with respect to
, under the same multilinear optimization strategy, results in the associated RLS-based algorithm, which is defined by the relations:
where
.
It can be noted that
, so that the a priori error from (
13) is equivalent to the alternative expressions from (
23), (
27), and (
31). Nevertheless, for the fluency of the development, these three alternative forms are used. On the other hand, the conventional form of the RLS algorithm, defined by (
13), (
17), and (
19), is reformulated using the TOT-based decomposition as a combination of three shorter RLS-based filters, defined by (
22)–(
24), (
26)–(
28), and (
30)–(
32). This approach could lead to important advantages, in terms of both performance and complexity, as indicated in [
21], where a version of the RLS-TOT based on the matrix inversion lemma was analyzed. However, this version is inherently limited in terms of robustness, since the only control parameters are the forgetting factors associated with the three component filters, i.e.,
,
, and
, respectively. Even when using the maximum values of these parameters (i.e., one), they cannot cope in noisy environments, with high and/or nonstationary noises. This is the main motivation for the upcoming development, which targets the derivation of a robust version of the RLS-TOT algorithm, using a recently designed regularization technique [
31].
To this purpose, the cost functions from (
21), (
25), and (
29) need to be adjusted by considering the weighted LS criterion and also including appropriate regularization terms [
31]. Consequently, the proposed regularized RLS-TOT algorithm relies on the following modified cost functions:
where
is the power of the additive noise, according to the model from (
11), while
,
, and
are the corresponding regularization terms. Following the idea from [
31], which was developed in the framework of the conventional RLS algorithm, the regularization terms of the proposed RLS-TOT version are designed as
where
denotes the Euclidean norm and simplified first-order Markov models are considered for the component filters to be identified, shown in (
5)–(
7), i.e.,
with
,
, and
being zero-mean white Gaussian noises (uncorrelated with their corresponding impulse responses from time index
), having the variances
,
, and
, respectively.
The cost functions from (
33)–(
35), which include the regularization terms from (
36)–(
38), could look quite laborious at first glance. However, they contain important components that influence the overall behavior of the resulting algorithm. First, the noise power,
, is related to the influence of the external noise, which impacts the robustness behavior. Second, the parameters
,
, and
(which multiply the corresponding filter lengths) represent the “uncertainties” of the systems to be identified, as they relate to the tracking capabilities. Taking into account (and controlling) the influence of these parameters could lead to a better compromise in terms of the main performance criteria, i.e., robustness and tracking, which represents a critical issue in the context of adaptive filtering algorithms.
The minimization of the cost functions
,
, and
with respect to
,
, and
, respectively, is carried out under similar circumstances as in [
31]. Consequently, the updates of the component filters result in similar outputs to (
24), (
28), and (
32), respectively, as follows:
where
are the regularization parameters associated with the three component filters, while
,
, and
are the identity matrices of sizes
,
, and
, respectively. For the sake of fluency, the derivation of (
42) and (
45) is provided in
Appendix A; these are associated with the filter
. The update relations associated with the other two component filters, i.e.,
and
, can be obtained in a similar manner.
In the expressions of the regularization parameters from (
45)–(
47), we can find the lengths of the component filters (i.e.,
,
, and
) and the related forgetting factors, which are known and available. However, we also need the noise power and the model uncertainties, which should be estimated in a practical way. First, the parameter
can be related to the error signal, since in a system identification scenario based on the model from (
11), the goal is to recover
from the error of the adaptive filter after this one converges to the steady-state solution. Therefore, the recursive estimates of the noise power (for each component filter, with different lengths) can be obtained as
where
are the forgetting factors, which act now as weighting parameters, while the subscript notation generally indicates the component filter, i.e.,
; as mentioned after (
32), the error signals
are equivalent, while the notation is kept for the sake of fluency. The uncertainty parameters can be evaluated based on (
39)–(
41), but using the estimates provided by the component adaptive filters, together with the approximation
, where
generally denotes the length of the corresponding component filter and
is the associated variance, with
. As a result, we can recursively evaluate:
where
, with
, are the same weighting parameters (i.e., the forgetting factors) used in (
48). Since they use the recursive (time-dependent) estimates from (
48)–(
51), the resulting regularization parameters from (
45)–(
47) also become time-dependent, thus acting like a variable regularization technique. Summarizing, the proposed variable regularized RLS-TOT (VR-RLS-TOT) algorithm is shown in
Table 1, while the main symbols and notation are grouped in
Table 2. The positive constants
and
(see
Table 1) are introduced for practical reasons, in order to prevent any potential numerical issues, like divisions by zero. A brief convergence analysis in the mean value (for one of the component filters) is provided in
Appendix B.
The forgetting factors associated with the three component filters are set using a common rule of thumb for the RLS-based algorithms, taking into account the corresponding filter length [
21]. Since the forgetting factor is considered a memory-related parameter, its value is naturally associated with the filter length, according to the following relation:
with
, where the value of
K is considered to be an integer. Clearly, the larger this integer is set, the closer the value of the forgetting factor is to one. Also, operating with a longer filter requires a larger value of the forgetting factor. As a result, as also indicated in
Table 1, the only parameter to be set is the integer
K, while the forgetting factors
(with
) are determined based on (
52), since the lengths of the component filters are known. The influence of this parameter and its setting will be further discussed in
Section 4.
As compared to the non-regularized version of the RLS-TOT from [
21], which uses constant regularization parameters only for the initialization of the matrices
,
, and
, the extra computational amount required by the variable regularization parameters of the proposed VR-RLS-TOT is negligible. These specific terms, i.e.,
in
Table 1, with
, need the evaluation of the noise power,
, together with the uncertainty parameters,
. To recursively compute the estimate from (
48), only three multiplications and one addition are required for each component filter. In addition, each of the estimated parameters from (
49)–(
51) involves
multiplications and
additions, where
represents the length of the corresponding filter, i.e.,
,
, and
, respectively. Thus, the overall (extra) computational complexity is on the order of
, i.e., it is proportional to the sum of the corresponding filters’ lengths.
The computational complexity of the conventional RLS algorithm defined by (
13), (
17), and (
19) is proportional to
where
denotes the computational amount related to an
matrix inversion; this operation can be performed in different ways, as outlined after (
19). On the other hand, the proposed VR-RLS-TOT algorithm, which combines the solutions provided by the three shorter component filters, involves a computational complexity proportional to
with
,
, and
. This could be significantly lower compared to (
53) for the regular setup of the decomposition parameters, as will also be indicated in the next section.
4. Simulation Results
The experimental framework considers an echo cancelation scenario, where the main goal is to identify the impulse response of an echo path,
, in order to obtain a replica of the echo, which is further subtracted from the microphone signal [
4]. To this purpose, the first impulse response from ITU-T G168 Recommendation [
39] is selected, which consists of a cluster of 64 coefficients that is padded with zeros up to the length
. The input signal (also known as the far-end) is a speech sequence of a female voice, using a sampling rate of 8 kHz. The output of the echo path is corrupted by a background noise that is considered white and Gaussian, with a signal-to-noise ratio (SNR) of 20 dB. The SNR is defined in relation to (
11) as the ratio
, where
denotes the variance of the echo signal,
. Different types of noise also affect the microphone signal, as will be indicated in some of the upcoming experiments, in order to assess the robustness of the algorithms. In several experiments, an echo path change scenario is considered by changing the sign of the coefficients of
in the middle of the simulation, with the purpose of evaluating the tracking capabilities of the algorithms. The performance measure is the normalized misalignment, which evaluates the “difference” between the true impulse response,
, and its estimate,
. This is computed (in dB) as
A lower level of this performance measure indicates a better accuracy of the solution/estimate, while a steeper misalignment curve is associated with a faster convergence rate and/or tracking of the algorithm.
Besides the proposed VR-RLS-TOT, other comparison algorithms are involved in several experiments. Naturally, the RLS-TOT counterpart from [
21] is selected, which represents the non-regularized version of the VR-RLS-TOT. In addition, the regularized RLS-TOT (R-RLS-TOT) from [
40] is considered, which uses a different regularization technique that relies only on the SNR. All these TOT-based algorithms combine the three shorter adaptive filters (of lengths
,
, and
) and use the associated forgetting factors,
, with
. Since
, the most natural factorization is
and
. Finally, the variable-regularized RLS (VR-RLS) algorithm recently developed in [
31] is involved in comparisons. It uses a similar regularization technique as that of the VR-RLS-TOT, but for the conventional RLS algorithm, using a single long-length adaptive filter (and a single forgetting factor,
), with
L coefficients. The forgetting factors of all these RLS-based algorithms are set based on (
52). For example, in the case of the VR-RLS algorithm [
31], the length of the filter is
L, so that its (single) forgetting factor is evaluated as
. Clearly, this value is larger compared to the forgetting factors of the TOT-based algorithms, i.e.,
, with
.
Before the comparisons with its counterparts, in the first experiment, it is important to analyze the performance of the proposed VR-RLS-TOT algorithm for different values of the decomposition parameter
P. The value of
P is limited by
; however, a much lower value is required in practice. As indicated in [
20,
21], the value of
P is usually within the range
to
. Consequently, for the current decomposition setup, we evaluate the performance of the VR-RLS-TOT algorithm when using the values
and 4, while setting
for computing the three forgetting factors (i.e.,
, with
) according to (
52). The results are shown in
Figure 1, where an echo path change scenario is simulated at time 7.5 s. In terms of accuracy (i.e., misalignment level), we can notice a similar performance for all the values of
P. Only a slightly better tracking reaction is obtained when the value of
P increases. However, this also increases the computational complexity, as discussed at the end of
Section 3. Consequently, using
represents a proper compromise in terms of performance and complexity. This value will be further used for all the TOT-based algorithms analyzed in this section.
The advantage of the TOT-based decomposition lies in using a much smaller parameter space (i.e., a much smaller number of coefficients) compared to the conventional approach, which involves a single filter of length
L. This is supported in
Figure 2, where the filter indicated by “0” on the
x-axis stands for the regular solution using
coefficients. Here, the filters indicated as “1”, “2”, and “3” are associated with the TOT-based decomposition, with the lengths
and
, which correspond to the current decomposition setup using
. Clearly, when using the TOT-based decomposition, there is an important reduction in terms of the parameter space, which further results in a lower computational complexity and an improved performance compared to the conventional approach.
The forgetting factor is an important parameter that controls the performance of any exponentially weighted RLS-type filter. A larger value of the forgetting factor leads to a strong accuracy of the solution (i.e., low misalignment) but at the cost of a slow tracking behavior. On the other hand, reducing the value of this parameter affects the accuracy while improving the tracking capability. The proposed VR-RLS-TOT uses three forgetting factors, each one associated with one of the component filters. In
Figure 3, their values are set using different values of the integer
K, according to (
52). This is the only parameter that has to be set for evaluating the values of
(with
), since the lengths of the associated filters are known, i.e.,
,
, and
, respectively. Clearly, the forgetting factors are always positive subunitary constants, as indicated by their role within the cost functions, i.e., to “forget” the past errors, so that the adaptive filter could be able to track the potential changes/variations in the system to be identified. As expected, a smaller value of
K (i.e., a smaller forgetting factor) improves the tracking while also increasing the misalignment. For a better compromise in terms of these performance criteria, the “middle” value
represents a favorable choice, which will be used in the following experiments for all the RLS-based algorithms. Of course, other values of
K could be used, taking into account the previously discussed performance compromise.
Next, the proposed VR-RLS-TOT algorithm is compared to its TOT-based counterparts from [
21,
40], i.e., the RLS-TOT and R-RLS-TOT algorithms. Also, the VR-RLS algorithm from [
31] is included in comparisons. The forgetting factors for all the algorithms are set using (
52), with
. The decomposition parameter for the TOT-based algorithms is set to
, as outlined before. The results are presented in
Figure 4. First, it can be noted that the proposed VR-RLS-TOT is more robust in terms of its accuracy, reaching a slightly lower misalignment level compared to the other two TOT-based algorithms, which are more affected by the nonstationary nature of the input signal (i.e., speech). Also, the VR-RLS-TOT algorithm exhibits strong tracking behavior when the echo path changes (after 7.5 s). In this case, we can notice a significantly slower reaction of the VR-RLS algorithm, which operates with a single (long-length) adaptive filter, with
coefficients. On the other hand, the TOT-based algorithms combine the estimates provided by three (much) shorter adaptive filters, with 4, 64, and 64 coefficients, respectively (as indicated in
Figure 2), which lead to an improved tracking behavior.
In the following, the robustness features of the algorithms are tested in noisy conditions when the microphone signal at the near end captures different signals in addition to the echo signal and the background noise. First, in
Figure 5, two bursts of white Gaussian noise (WGN) are considered, with
dB and 0 dB, respectively. It can be noted that the proposed VR-RLS-TOT is very robust to such perturbations, while the other algorithms are significantly affected. The R-RLS-TOT [
40] performs slightly better compared to the RLS-TOT [
21], while the VR-RLS algorithm [
31] has a certain latency related to its reactions and transitions due to the long-length filter that needs to be updated.
Another scenario considers more challenging types of noise compared to the previous WGN bursts (which are stationary sequences). In
Figure 6, the perturbation source (and its nature) is changing by using two bursts of highway noise and engine noise, respectively, while the latter is stronger. All the algorithms are more affected under these conditions compared to the previous WGN scenario. Nevertheless, the proposed VR-RLS-TOT still excels in terms of its robustness, being less affected by the first burst of highway noise while recovering fairly well in the case of the second burst of engine noise.
Finally, the most challenging echo cancelation scenario is considered, which is the so-called double-talk case, when both speakers (far-end and near-end) talk at the same time [
4]. In this scenario, the near-end speech acts like a large level of nonstationary disturbance for the adaptive filter. In
Figure 7, two periods of double-talk appear, with different intensities, the second one being more significant. As we can notice, the behavior of both the RLS-TOT [
21] and R-RLS-TOT [
40] is seriously biased during double-talk. The VR-RLS algorithm [
31] has some robustness in this scenario, but it is still influenced by the inherent latency due to operating with a long-length filter (
). On the other hand, the proposed VR-RLS-TOT algorithm quickly recovers at the beginning of the first period of double-talk while being very robust during the second period of (stronger) double-talk. Concluding, it achieves a much better compromise in terms of the main performance criteria compared to the competitive algorithms.