Next Article in Journal
Classification of Heart Sounds Using Convolutional Neural Network
Next Article in Special Issue
Development of Classification Algorithms for the Detection of Postures Using Non-Marker-Based Motion Capture Systems
Previous Article in Journal
An Experimental Approach for the Direct Measurement of Temperatures in the Vicinity of the Keyhole Front Wall during Deep-Penetration Laser Welding
Previous Article in Special Issue
Three-Dimensional (3D) Model-Based Lower Limb Stump Automatic Orientation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiresolution Speech Enhancement Based on Proposed Circular Nested Microphone Array in Combination with Sub-Band Affine Projection Algorithm

by
Ali Dehghan Firoozabadi
1,*,
Pablo Irarrazaval
2,3,4,
Pablo Adasme
5,
David Zabala-Blanco
6,*,
Hugo Durney
1,
Miguel Sanhueza
1,
Pablo Palacios-Játiva
7 and
Cesar Azurdia-Meza
7
1
Department of Electricity, Universidad Tecnológica Metropolitana, Av. José Pedro Alessandri 1242, Santiago 7800002, Chile
2
Electrical Engineering Department, Pontificia Universidad Católica de Chile, Santiago 7820436, Chile
3
Biomedical Imaging Center, Pontificia Universidad Católica de Chile, Santiago 7820436, Chile
4
Institute for Biological and Medical Engineering, Pontificia Universidad Católica de Chile, Santiago 7820436, Chile
5
Electrical Engineering Department, Universidad de Santiago de Chile, Av. Ecuador 3519, Santiago 9170124, Chile
6
Department of Computing and Industries, Universidad Católica del Maule, Talca 3466706, Chile
7
Department of Electrical Engineering, Universidad de Chile, Santiago 8370451, Chile
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2020, 10(11), 3955; https://doi.org/10.3390/app10113955
Submission received: 6 April 2020 / Revised: 17 May 2020 / Accepted: 4 June 2020 / Published: 6 June 2020
(This article belongs to the Special Issue Mathematics and Digital Signal Processing)

Abstract

:
Speech enhancement is one of the most important fields in audio and speech signal processing. The speech enhancement methods are divided into the single and multi-channel algorithms. The multi-channel methods increase the speech enhancement performance by providing more information with the use of more microphones. In addition, spatial aliasing is one of the destructive factors in speech enhancement strategies. In this article, we first propose a uniform circular nested microphone array (CNMA) for data recording. The microphone array increases the accuracy of the speech processing methods by increasing the information. Moreover, the proposed nested structure eliminates the spatial aliasing between microphone signals. The circular shape in the proposed nested microphone array implements the speech enhancement algorithm with the same probability for the speakers in all directions. In addition, the speech signal information is different in frequency bands, where the sub-band processing is proposed by the use of the analysis filter bank. The frequency resolution is increased in low frequency components by implementing the proposed filter bank. Then, the affine projection algorithm (APA) is implemented as an adaptive filter on sub-bands that were obtained by the proposed nested microphone array and analysis filter bank. This algorithm adaptively enhances the noisy speech signal. Next, the synthesis filters are implemented for reconstructing the enhanced speech signal. The proposed circular nested microphone array in combination with the sub-band affine projection algorithm (CNMA-SBAPA) is compared with the least mean square (LMS), recursive least square (RLS), traditional APA, distributed multichannel Wiener filter (DB-MWF), and multichannel nonnegative matrix factorization-minimum variance distortionless response (MNMF-MVDR) in terms of the segmental signal-to-noise ratio (SegSNR), perceptual evaluation of speech quality (PESQ), mean opinion score (MOS), short-time objective intelligibility (STOI), and speed of convergence on real and simulated data for white and colored noises. In all scenarios, the proposed method has high accuracy at different levels and noise types by the lower distortion in comparison with other works and, furthermore, the speed of convergence is higher than the compared researches.

1. Introduction

In the current century, the smartphones and other communication devices have been an important part of human life, where it is impossible to have social communications without them [1,2]. One of the principal parts in these smartphones is the signal processing platform. This part has an important role in the telecommunication and audio signal processing. Denoising and dereverberation are two main sections in the signal processing and enhancement platforms, which is the aim of this article, to increase the performance of speech enhancement algorithms [3]. Increasing the number of sensors improves the accuracy of denoising algorithms due to the spatial spectrum extension by providing the proper information. The definition of accuracy in the enhancement algorithms is how the enhanced signal is closer to the original signal with a high level of noise elimination and less distortion. Therefore, the speech enhancement is the main part in such applications as: hearing aid systems, mobile communication, speaker localization and tracking, speech recognition, voice activity detection (VAD), speaker identification, etc. The denoising algorithms should be implemented in a way to keep the speech intelligibility in an acceptable range and to remove a high level of noise and reverberation. Then, the signal-to-noise ratio (SNR) cannot be the only specific factor for comparing the speech enhancement methods. The qualitative criteria such as: perceptual evaluation of speech quality (PESQ) [4], mean opinion score (MOS) [5], and short-time objective intelligibility (STOI) [6] are very useful to show the performance of denoising methods in comparison with other previous works along with quantitative criteria such as: overall SNR and segmental SNR (SegSNR) [7]. The performance of the denoising algorithms is calculated by considering the qualitative and quantitative criteria at the same time, which are the proper measurements for comparison with other previous works.
In recent years, many of the single and multi-channel methods have been proposed for speech enhancement. The single-channel methods are still challenging strategies for the speech enhancement due to the limited information. The traditional speech enhancement methods such as the Wiener filter (WF) and distributed multichannel WF (DB-MWF) [8,9], spectral subtraction [10,11], and statistical-model-based [12,13] have superior performances in stationary noisy environments but the stability and accuracy of these methods are strongly decreased in non-stationary noisy conditions. However, existing noise estimation methods such as minima-controlled recursive averaging [14,15] and minimum statistics [11,16] follow the stationary noise energy. However, they do not have the ability to follow the non-stationary noise energy. For example, the method proposed in [16] is presented to estimate the power spectral density (PSD) of a non-stationary noise signal. This method can be considered in combination with any speech enhancement algorithm, which requires the noise PSD estimation. The presented method follows the spectral minima in each frequency band by minimizing conditional mean square error (MSE) criteria in each time frame, which develops the optimal smoothing parameter for recursive smoothing of the PSD of the noisy speech signal. Therefore, an unbiased noise estimator is presented based on the optimally smoothed PSD estimation and the analysis of the statistics of spectral minima. Therefore, the noise estimation accuracy in some methods [15,16] is affected when the noise is non-stationary. A group of speech enhancement methods are proposed based on a priori information of speech signals such as the auto-regressive hidden Markov model (ARHMM) [17,18,19]. The noise and speech signals are modeled as an auto-regressive (AR) process in these methods. In addition, the hidden Markov model (HMM) is implemented for modeling the prior information of speech and noise features. For example, the methods in [18,19] are considered for modeling the speech and noise spectrum shape. Therefore, the spectrum gain is calculated instead of the whole spectrum for the speech and noise signals. The noise-spectrum gain estimation is adapted by the fast variations of the signal energy, which is known as non-stationary noise.
Masoud and Sina [20] proposed a novel method based on the normalized fractional of the two-channel least mean square (LMS) algorithm for enhancing the speech signal. The presented algorithm is known as fractional LMS, which is obtained by considering the fractional terms in the calculation of filter coefficients of the standard LMS algorithm. The normalization is a proper strategy to improve the performance of the LMS algorithm. Therefore, a normalization step is implemented on the fractional LMS in order to promote the performance of the enhancement method. The proposed two-channel method has a higher performance in terms of the MSE criteria in comparison with other works. Pagula and Kishore [21] proposed a recursive least square (RLS)-based adaptive filter for the application of speech enhancement. The segmentation step is considered for the microphone signals to provide a better stationary of the speech signals. In the following, the adaptive filter coefficients are calculated based on the modified version of the RLS method. The filter coefficients are calculated in a way to have the least distortion in the enhanced speech signals. The presented method has a high performance in the presence of white noise for a different range of SNRs. Qi et al. [22] proposed a method for estimation of the short-time linear prediction parameters of the Wiener filter. In the presented work, a speech signal spectrum modeling is proposed based on the prior information of the speech linear prediction in order to model the noise as same as the speech signal. The difference between the proposed method with other previous works is the use of multiplicative update rule for better estimation of the coefficients. Tavakoli et al. [23] introduced a framework for the speech enhancement based on an ad-hoc microphone array. A subarray is considered for coherence calculation in the speech signal. A coherence measurement is proposed based on the speech quality in the entrance of the array in order to select the subarrays in the local speech enhancements, when more than one subarray is used. The proposed method is evaluated based on quantitative and qualitative criteria such as: array gains, speech distortion ratio, PESQ, and STOI to show the superiority of the algorithm. Shimada et al. [24] proposed an unsupervised speech enhancement method based on the non-negative matrix factorization and sub-band beamforming for robust speech recognition against the noise. In the recent years, the minimum variance distortionless response (MVDR) beamforming is widely used to achieve the speech enhancement because this method properly works when there are steering vectors for the speech signal and spatial covariance matrix for the noise. In the presented algorithm, an unsupervised method decomposes each time-frequency bin to the sum of the noise and signal by implementing the multi-channel non-negative matrix factorization (MNMF). The presented method estimates the spatial covariance matrix (SCM) for the signal and noise by the use of spectral noise and speech features. In this paper, the online MVDR beamforming is proposed via an adaptive update for the MNMF parameters. Kavalekalam et al. [25] proposed a speech enhancement model-based method to increase the speech perception for auditory earphones applications. In the proposed method, a binaural speech enhancement framework is introduced, which is implemented by a speech production approach. The proposed speech enhancement framework is based on a Kalman filter, which is presented to use the speech production dynamic in the procedure of the speech enhancement. The Kalman filter needs to have an estimation from the short time predictor (STP) of clean speech, noise, and the pitch estimation of the clean speech. A binaural method for STP parameters estimation is proposed in this paper with a directional pitch predictor based on the harmonic model and maximum likelihood (ML) criteria for pitch features estimations. These parameters are calculated just based on 2-microphones signals equivalent to human ears. Botinhao et al. [26] proposed a simultaneous noise-reverberation enhancement method for text-to-speech (TTS) systems. The recorded voices in noisy-reverberant environments affects the quality of the TTS systems. A simple way is to increase the quality of the prerecorded speech signals for the TTS training system by speech enhancement methods such as: noise suppression and dereverberation algorithms. Then, a recurrent neural network is considered in this paper for the speech enhancement. The neural network is trained by parallel data of clean speech and recorded speech with low quality. The low quality speech signal is obtained by the addition of environmental noise and convolution between the room impulse response and the clean speech. The separated neural networks are trained by only-noise, only-reverberation, and noisy-reverberant data. The quality of the training data with a low quality speech signal is highly improved by the use of this neural network. Wang et al. [27] proposed a model-based method for speech enhancement in modulation domain by the use of a Kalman filter. The proposed predictor models the estimated amplitude spectral dynamically from the speech and noise to calculate the minimum mean square error (MMSE) of the speech amplitude spectrum taking into account that the noise and speech are additive in the complex plane. The stationary Gaussian model is proposed to consider the dynamic noise amplitude as same as the dynamic speech amplitude, which is a mixture of Gaussian models that the centers are located in a complex plane.
In our article, a multi-channel speech enhancement method is introduced based on the proposed circular nested microphone array in combination with the sub-band affine projection algorithm (CNMA-SBAPA). A nested microphone array increases the accuracy of the speech enhancement methods by increasing the information. Nevertheless, spatial aliasing is one of the challenges when microphone arrays are used. Firstly, a uniform circular nested microphone array (CNMA) is proposed for eliminating the spatial aliasing. Additionally, the array dimensions are designed in a way to be applicable in the real conditions. The speech components are variable in frequency bands. Therefore, a sub-band processing method is considered for speech signals. This method provides the high frequency resolution in low speech frequency components. Finally, the affine projection algorithm (APA), as an adaptive method for the speech enhancement, is implemented on sub-band signals from the circular nested microphone array (NMA). Since each APA block is implemented on a sub-band with specific information, the accuracy and speed of convergence are increased in this condition. In the last step, the synthesis filters are used to generate the enhanced speech signal. The proposed system with sub-band APA is compared by the quantitative (segmental SNR), qualitative (PESQ, MOS, and STOI) criteria, and speed of convergence with the least mean square (LMS), traditional APA, recursive least square (RLS), distributed multichannel Wiener filter (DB-MWF), and multichannel nonnegative matrix factorization-minimum variance distortionless response (MNMF-MVDR) algorithms on real and simulated data under white and colored noisy conditions. The results show the superiority of the proposed system in comparison with other previous works in all environmental conditions.
Section 2 shows the microphone signal model and the proposed uniform circular nested microphone array. Section 3 includes the proposed sub-band algorithm with analysis and synthesis filter banks in combination with the sub-band APA. The results on real and simulated data are discussed in Section 4. Section 5 includes some conclusions.

2. The Microphone Model and Proposed Nested Microphone Array

In this section, the microphone signal model was presented to produce the simulated data. In addition, the uniform CNMA was proposed for eliminating the spatial aliasing. Additionally, the nested subarrays and microphone combinations are introduced in this section.

2.1. Microphone Signal Model

The microphone signal modeling is an important part in the implementation of speech processing algorithms such as: speech enhancement, speaker tracking, speech recognition, etc. Two models are usually considered in this processing: ideal and real models [28]. In the ideal model, which is known as an open-space model, the received signal in a microphone place is a weakened and delayed version of the transmitted signal from the source location. The ideal model for microphone signals is expressed as:
x m [ n ] = 1 r m s [ n τ m ] + v m [ n ] ,
where xm[n] is the received signal in the m-th microphone, s [ n ] is the speech source signal (transmitted signal), r m is the distance between source and m-th microphone, τ m is the time delay between source and m-th microphone, and v m [ n ] is the additive noise in m-th microphone place. This model cannot show the real environments and close space conditions because the reverberation effect is discarded. Therefore, the real model is introduced for microphone signal simulations to provide the real environmental conditions for evaluating the speech enhancement algorithms. The real model simulates the microphone signal similar to the environmental conditions. The expression for real model is shown as:
x m [ n ] = s [ n ] * γ m [ r m , n ] + v m [ n ] ,
where the source signal is convolved to the room impulse response to model the real environments. In this equation, γ m [ r m , n ] is the impulse response between the source and m-th microphone, which contains the attenuation factor and whole reverberation effect in the real conditions, and * denotes to the convolution operator. The simulated signals are similar to real conditions by considering this mathematical real model.

2.2. The Proposed Uniform Circular Nested Microphone Array

The microphone array increases the accuracy of the speech enhancement algorithms due to increasing the information. However, the spatial aliasing based on the inter-microphone distances destroys the recorded speech signals, and in the following, the performance of the speech enhancement algorithms. Nested microphone array has the capability to eliminate the spatial aliasing [29]. In this section, a uniform CNMA is proposed where by having a symmetrical shape, provides the same probability for all speakers around the array, and the quality of the enhanced signals are not dependent on the position of the speakers. Additionally, its small structure helps to be applicable in most of the conditions in comparison with other big arrays. Figure 1 shows the block diagram of the proposed speech enhancement algorithm, where the NMA part with its analysis filters and down-sampler blocks are shown in the left side.
The speech signal has a frequency range of [0–8000] Hz with a sampling frequency F s = 16 , 000   Hz . The proposed CNMA is designed for the frequency range [50–7800] Hz, which covers the wideband speech spectrum. The CNMA is structured by four subarrays. The first subarray is designed for the range B1 = [3900–7800] Hz, of central frequency f c 1 = 5850   Hz . The inter-microphone distance ( d lim ) should be d lim < λ / 2 ( λ is the wavelength of the highest frequency component in the related sub-band) to avoid the spatial aliasing, this is d lim ( 1 ) < 2.2   cm for the first subarray. The second subarray covers the frequency range B2 = [1950–3900] Hz with a central frequency of f c 2 = 2925   Hz , therefore d lim ( 2 ) = 2 d 1 < 4.4   cm . The third subarray is defined for the frequency range B3 = [975–1950] Hz with a central frequency of f c 3 = 1462   Hz and d lim ( 3 ) = 4 d 1 < 8.8   cm . Finally, the forth subarray is designed for the frequency range B4 = [50–975] Hz with a central frequency of f c 4 = 512   Hz and the inter-microphone distance is d lim ( 4 ) = 8 d 1 < 17.6   cm . For a more complexity system, a higher number of microphones could be considered to design a larger nested microphone array. Table 1 shows the summarized information to design the uniform CNMA.
The microphone array was structured to have the closest microphone distances as d s i m ( 1 ) = 2.2   cm (for the simulated data) based on the designed CNMA. Therefore, the first subarray included the microphone pairs {1,2}, {2,3}, {3,4}, {4,5}, {5,6}, {6,7}, {7,8}, and {8,1}. The microphone pairs {1,3}, {3,5}, {5,7}, {7,1}, {2,4}, {4,6}, {6,8}, and {8,1} were selected for the second subarray with an inter-microphone distance of d s i m ( 2 ) = 4.2   cm . The third subarray has the inter-microphone distance of d s i m ( 3 ) = 5.6   cm . Then, the microphone pairs {1,4}, {2,5}, {3,6}, {4,7}, {5,8}, {6,1}, {7,2}, and {8,3} wereconsidered for this subarray. For the last subarray, the inter-microphone distance is d s i m ( 4 ) = 6   cm and the microphone pairs {1,5}, {2,6}, {3,7}, and {4,8} were selected for the implementation. Given our actual microphone array, the minimum inter-microphone distance that we could have was 2.7 cm (for the real data). For this reason, we did two evaluations, one for simulated data with d s i m ( 1 ) = 2.2   cm , as dictated by the theory, and one for real data with d r e a l ( 1 ) = 2.7   cm , to match our hardware. All subarrays are shown in Figure 2, which shows the designed CNMA with its small shape.
Each subarray needs an analysis filter bank to avoid the spatial aliasing and imaging. Figure 1 (left and right sides) shows the analysis and synthesis filter banks along with the up-sampler and down-sampler blocks. The multirate sampling by the use of up-samplers and down-samplers is implemented to provide the frequency bands. As shown in Figure 3a, the analysis filter bank H i ( z ) and down-sampler D i are realized as a multi-level tree structure. Each stage of the tree requires a high-pass filter (HPF) H P i ( z ) , a low-pass filter (LPF) L P i ( z ) , and a down-sampler D i (for the analysis filter bank) or up-sampler D i (for the synthesis filter bank). The relation between the analysis filter bank H i ( z ) , the LPFs, and HPFs in the tree structure is expressed as:
H 1 ( z ) = H P 1 ( z )   H 2 ( z ) = L P 1 ( z ) H P 2 ( z 2 ) H 3 ( z ) = L P 1 ( z ) L P 2 ( z 2 ) H P 3 ( z 4 ) H 4 ( z ) = L P 1 ( z ) L P 2 ( z 2 ) L P 3 ( z 4 ) .
The synthesis filters G i ( z ) are the mirror image of analysis filters H i ( z ) , which are implemented by the tree structure as seen in Figure 3b.
In each level of the tree, a 52-tap finite impulse response (FIR) LPF and HPF are implemented by the Remez method. The parallel filters have a stop-band attenuation of 50 dB and a transition band 0.0575. Figure 4 shows the frequency response for the analysis filter banks.

3. The Proposed Multiresolution Sub-band-APA for the Speech Enhancement

Speech is a wideband and non-stationary signal, where each frequency band has different information. This feature for the speech signal provides the conditions to evaluate the speech spectrum components by considering different frequency resolution. For example, speech information is condensed at the lower part of the spectrum. Therefore, the accuracy of the speech enhancement algorithm is increased by a focus to low frequency components. In this article, a specific sub-band processing along with a filter bank was proposed for paying more attention to lower frequencies by the use of filters with narrower bandwidths. Table 2 shows the information to design and implement this analysis filter bank. There is still not any certain rule for selecting the number of frequency bands. Of course, by having narrower band filters in low frequencies, we have more frequency resolution, but the concern is the computational complexity. In other hands, adding each more filter means entering more microphone pairs and more calculations. Based on the experiments, this number of frequency bands prepares enough performance and acceptable level of complexity.
As seen, the filter bandwidth is smaller in low frequencies in comparison with high frequencies. This property increases the frequency resolution for low frequencies. The most important benefit in sub-band processing is the noise estimation from the silent part of the speech signal in each sub-band. Since in the proposed denoising method, the noise estimation is required as an input for the enhancement algorithm. Therefore, the more accurate and stationary noise estimation is obtained by sub-band processing of the speech signal, which increases the denoising algorithm performance. If x m [ n ] is considered as an input signal for the m-th microphone, the analysis filter output for the CNMA is expressed as:
x m , i [ n ] = x m [ n ] * h i [ n ]   where   { m = 1 , , 8   and   i = 1 , , 4 } ,
where x m , i [ n ] is the analysis filter output and h i [ n ] is the impulse response for this filter. Therefore, the spatial aliasing is eliminated from each microphone pairs of CNMA by the use of analysis filters, which are designed specifically for each subarray. In the following, the microphone signals are entered to the proposed analysis filter bank for the sub-band processing. As shown in Table 2, each microphone signal is divided into 10 sub-bands. These numbers of sub-bands were selected based on our experiments in order to provide a proper efficiency and with low computational complexity, by preparing a high frequency resolution in low frequencies. Therefore, the output of the proposed analysis filter bank is expressed as:
y m , i , j [ n ] = x m , i [ n ] * F j [ n ]   where   { j = 1 , , 10   ,   i = 1 , , 4   ,   m = 1 , , 8 } ,
where F j [ n ] is the impulse response for each sub-band filter in the analysis filter bank and y m , i , j [ n ] is the output of the analysis filter bank for the j-th sub-bands and m-th microphone. The signals y m , i , j [ n ] are the sub-band microphone signals for the proposed sub-band-APA algorithm. In the following, the sub-band-APA (SBAPA) algorithm along with the circular nested microphone array (CNMA-SBAPA) is proposed for the speech enhancement. Adaptive filters as an important tool in digital signal processing have been utilized for many years in such application as: speech signal enhancement, system identification, localization and tracking, etc. In adaptive filters, the coefficients change periodically to be adapted based on the time varying features of the noise, and this property increases the performance of the denoising system in comparison with normal methods. In addition, these filters are non-linear and homogeneous since their features are dependent on the input signal. The adaptive filters have the following advantages: low delay and better tracking in non-stationary conditions [30]. These advantages are very important in dereverberation, denoising, time delay estimation, channel equalization, and speaker tracking applications. In these applications, low delay and robustness against of non-stationary noisy and reverberant conditions are important parameters to improve the performance of the proposed systems. The existence of the reference signal, which is hidden in the filter coefficient estimations, defines the system performance. Figure 5 shows the general structure of the adaptive filter in denoising applications.
We change the notation for input signal in adaptive filter ( y m , i , j [ n ] ) to y [ n ] for simplifying the mathematical expressions. An adaptive filter is expressed as follows [31]:
z [ n ] = w L [ n ] * y [ n ] ,
where n is the time index, z [ n ] is the adaptive filter output, and w L [ n ] is the adaptive filter coefficients with length L. The update algorithm in Figure 5 is considered as a principal part for an adaptive filter, which is the APA in this article. The main idea for an adaptive filter is to minimize the error signal e [ n ] to make the output of the filter as similar as the desired signal.
The input signal y [ n ] for the adaptive filter is considered as the summation of the noise (v[n]) and desired signal (d[n]), which is described as:
y [ n ] = d [ n ] + v [ n ] .
The adaptive filter has a FIR structure, namely the filter is designed based on the limited number of coefficients in the time domain. For a filter with order of L, the filter coefficients are defined as:
w L [ n ] = [ w [ 0 ] , w [ 1 ] , , w [ L 1 ] ] .
The error signal or cost function is defined as the difference between estimated and desired signal, namely:
e [ n ] = d [ n ] z [ n ] .
As shown in Equation (6), the output of the adaptive filter z [ n ] is defined as the convolution between the filter coefficients w L [ n ] and the input signal y [ n ] , where y [ n ] is considered as the input of the adaptive filter, namely:
y [ n ] = [ y [ n ] , y [ n 1 ] , , y [ n L ] ] .
In addition, the adaptive filter coefficients change during the time, which is written as:
w L [ n ] = w L [ n 1 ] + Δ w L [ n ] ,
where Δ w L [ n ] is defined as the correction factor for the filter coefficients. The adaptive filter produces the correction factor based on the input and error signal. In Figure 5, several algorithms can be considered for updating the filter coefficients. The APA is one of the fastest and most efficient methods for this purpose. The AP algorithms were introduced to improve the speed of convergence in the gradient-based algorithms, especially when the input signal has a non-stationary spectrum. It is because the speed of convergence is decreased in the case of non-stationary and constraint spectrums [30].
Filter update equation is one of the most important features in the AP algorithms, which uses N vectors of the input data to update the filter coefficients instead of using one vector of the input data, i.e., the normalized least mean square (NLMS). Therefore, more information was considered in the time for accurately updating the filter coefficients. Thus, the AP algorithm is known as an improved and extended version of the NLMS method or it can be expressed mathematically as a constraints minimization problem, which is expressed as follows.
The variation for L filter coefficients during the two consecutive times is given by:
Δ w L [ n ] = w L [ n ] w L [ n 1 ] .
We minimized Equation (13) under N constraints, which are shown in Equation (14) to extend the adaptive filter algorithm.
Δ w L [ n ] 2 = Δ w L T [ n ] Δ w L [ n ] ,
where N constraints are defined as follows:
w L T [ n ] y [ n k ] = d [ n k ]   f o r   k = 0 , , N 1 ,
where y [ n k ] is the vector of N last sample from the input signal and d [ n ] is the desired signal, see Figure 5. The proposed solution formulates the update algorithm for AP, which is expressed as:
w L [ n ] = w L [ n 1 ] + A T [ n ] ( A [ n ] A T [ n ] ) 1 e N [ n ] ,
where:
A [ n ] = ( y L [ n ] , y L [ n 1 ] , y L [ n 2 ] , , y L [ n N + 1 ] ) T ,
and e N [ n ] is a vector of size N × 1 , which is written as:
e N [ n ] = d N [ n ] A [ n ] w L [ n 1 ] .
The vector d N [ n ] is the desired signal with size N × 1 , namely:
d N [ n ] = ( d [ n ] , d [ n 1 ] , , d [ n N + 1 ] ) T .
The general format for AP algorithm is obtained by rewriting Equation (15) as:
w L [ n ] = w L [ n 1 α ( N 1 ) ] + μ A τ T [ n ] ( A τ [ n ] A τ T [ n ] + δ I ) 1 e N τ [ n ] .
If e N τ [ n ] is considered as e N τ [ n ] = d N τ [ n ] A τ w L [ n 1 α ( N 1 ) ] , then:
A τ [ n ] = ( y L [ n ] , y L [ n τ ] , , y L [ n ( N 1 ) τ ] ) T ,
and the signal d N τ [ n ] is expressed as:
d N τ T [ n ] = ( d [ n ] , d [ n τ ] , , d [ n ( N 1 ) τ ] ) .
As shown in Equation (19), the N required vectors to update the adaptive filter are not necessarily to be the last data vectors. Therefore, several versions of AP algorithms are defined based on the way to select the input data and parameters in Equation (19). There are some developed algorithms based on these parameters selections such as: the NLMS along with the orthogonal correction factor (OCF-NLMS) [32], the partial rank affine projection algorithm (PRAPA) [33], and the standard APA [34] whose parameters are α = 0 ,   δ = 0 ,   τ = 1 . If δ parameter differs to 0, the APA algorithm is extended to APA with regularization (R-APA) [35], where the update equation for the filter coefficients is a specific case of the Levenberg Marquardt regularized APA (LMR-APA) algorithm [36].
The introduced AP algorithm contains one input signal. Since pairs of microphones are used in the proposed CNMA, the AP algorithm is generalized to a two-microphone version [37]. Firstly, the generalization of a two-microphone structure is defined, where each microphone contains the mixing speech and noise signal, which is expressed as (see Figure 6a):
q m [ n ] = i = 1 2 r = 1 L 1 p i m [ r ] s i [ n r ]   ,   m = 1 , 2 ,
where s i [ n ] represents the source signals, q m [ n ] is the microphone signals, L is the impulse response length, and p i m [ r ] are the impulse responses between the microphone and sources. These impulse responses are considered as linear time-invariant (LTI) systems. Two source signals s i [ n ] are selected as the speech signal s [ n ] and noise signal b [ n ] . It is assumed that the speech and noise signals are independent, which means E { s [ n ] b [ n m ] } = 0 ,   m , where E denotes to expected value. Then, the noise and speech signals are uncorrelated. Based on the general structure, which is shown in Figure 6a, the microphone signals q 1 [ n ] and q 2 [ n ] are expressed as follows:
q 1 [ n ] = s [ n ] * p 11 + b [ n ] * p 21 ,
q 2 [ n ] = s [ n ] * p 12 + b [ n ] * p 22 .
In addition, p 11 and p 22 represent the impulse responses for direct path, and p 12 and p 21 are cross-coupling for the channels between the sources and microphones. The presented model is simplified by considering p 11 = p 22 = δ [ n ] , which is shown in Figure 6b as:
q 1 [ n ] = s [ n ] + b [ n ] * p 21 ,
q 2 [ n ] = s [ n ] * p 12 + b [ n ] .
Therefore, the microphone signals are generated based on the impulse responses between the source and microphones, noise, and speech signals. The structure in Figure 6c was proposed to retrieve the source signal from the received noisy signals q 1 [ n ] and q 2 [ n ] . The proposed structure provides the conditions to retrieve the original signal by the use of adaptive filters p 11 and p 22 . The signals u 1 [ n ] and u 2 [ n ] for the two-microphone structure are defined as follows:
u 1 [ n ] = q 1 [ n ] q 2 [ n ] * w 21 [ n ] ,
u 2 [ n ] = q 2 [ n ] q 1 [ n ] * w 12 [ n ] ,
where in Equations (27) and (28), w 12 [ n ] and w 21 [ n ] are the adaptive filters for eliminating the noise of microphone signal q 1 [ n ] and the speech of microphone signal q 2 [ n ] , respectively. Signals u 1 [ n ] and u 2 [ n ] are rewritten by replacing Equations (25) and (26) to Equations (27) and (28) as:
u 1 [ n ] = s [ n ] * [ δ [ n ] p 12 * p 21 ] + b [ n ] * [ p 21 w 21 [ n ] ] ,
u 2 [ n ] = b [ n ] * [ δ [ n ] p 21 * p 12 ] + s [ n ] * [ p 12 w 12 [ n ] ] .
Two adaptive filters w 12 [ n ] and w 21 [ n ] are required to retrieve the original speech signal from the noisy signals u 1 [ n ] and u 2 [ n ] . There is just a unique structure for adaptive filters w 12 [ n ] and w 21 [ n ] as w 12 [ n ] = p 12 and w 21 [ n ] = p 21 to retrieve the enhanced speech of noisy signals u 1 [ n ] and u 2 [ n ] . This structure requires a VAD for preparing the noise estimation from the silent part of the recorded signals.
The AP algorithm is generalized to a two-microphone structure based on the obtained Equation (19) for updating the filter coefficients. The AP algorithm is the generalized version of the two-microphone NLMS [38], which is shown in Figure 6c for adaptive speech enhancement algorithm. Therefore, the adaptive filter coefficients w 12 [ n ] and w 21 [ n ] for two-microphone APA are expressed as:
w 12 [ n ] = w 12 [ n 1 ] + μ 12 q 1 [ n ] q 1 [ n ] T + δ I q 1 [ n ] u 2 [ n ] ,
w 21 [ n ] = w 21 [ n 1 ] + μ 21 q 2 [ n ] q 2 [ n ] T + δ I q 2 [ n ] u 1 [ n ] ,
where q 1 [ n ] and q 2 [ n ] are defined as q 1 [ n ] = [ q 1 [ n ] , q 1 [ n 1 ] , , q 1 [ n N + 1 ] ] and q 2 [ n ] = [ q 2 [ n ] , q 2 [ n 1 ] , , q 2 [ n N + 1 ] ] . The matrices of the two-microphone signals q 1 [ n ] and q 2 [ n ] have dimensions L × N , where L is the adaptive filter length and N is the projection order. The two parameters μ 12 and μ 21 are the step sizes, which control the convergence of adaptive filters w 12 [ n ] and w 21 [ n ] . These parameters should be selected in the range [0,2] to assure the convergence of AP algorithm. If N is selected as 1, the AP algorithm is converted to the NLMS method.
The proposed sub-band APA not only increased the accuracy of the speech enhancement algorithm, but also the speed of convergence was improved (Table 6 in the results section) in the implementations because the noise was estimated separately for each sub-band and it was stationary on narrow bandwidths. Then, the SBAPA was implemented on generated sub-bands by the analysis filters in Figure 3. As shown in Figure 1, a symmetrical synthesis filter bank and synthesis filters related to the nested microphone array were implemented for the reconstruction the final enhanced signal. The synthesis filters as similar as the analysis filters were implemented based on the tree structure in Figure 3b. Finally, all sub-band signals were summed to generate the final enhanced signal. In the next section, the performance of the proposed CNMA-SBAPA was compared with other previous works.

4. Results and Discussion

The experiments in order to evaluate the performance of the proposed method were implemented on the real and simulated data. The TIMIT dataset was considered for the simulated data, where the data collection MDAB0 by four continuous sentences SX139, SX229, SX319, and SX409 were selected as a male speaker in the simulations [39]. This dataset includes short sentences for testing and training the algorithms. The tones and frequency components are two different parameters in the speech signal. There are pitch and speech spectrum components for the speakers. It is important to work with male or female signals for the algorithms, which works with the pitch parameter. Since this parameter changes highly based on the gender. Since we consider the speech spectrum, then the issue to use the male or female speakers does not change the results. Therefore, 12.5 s male-speech signal is used for implementations and experiments. A voice activity detector is implemented to detect the silence part of the speech signal [40], and the noise spectrum is estimated of these parts for the proposed SBAPA. Figure 7 shows the simulated room with the location of speakers and microphone array. The inter-microphone distances d s i m = 2.2   cm for the simulated data was selected based on the designed array. A speaker and a steered noise source were considered in the simulations. The room dimensions, speaker, and noise source locations were selected as 475,592,420cm, 374,146,110cm, and 362,412,120cm, respectively. These dimensions and locations were considered the same as the real room recording conditions. In addition, the proposed algorithm was implemented on real data to evaluate the real effect of the noise and reverberation on the performance. For this purpose, the real speech signal was recorded in the speech processing laboratory at Fondazione Bruno Kessler (FBK), Trento, Italy. Figure 8 shows a view of the recording room at FBK. Two electronic speakers were used instead of the human and noise source in the process of data recording. In addition, Figure 8 shows the position of the circular NMA in the center of the room. We were able to consider the minimum inter-microphone distance in the real conditions with our setup (see Figure 8) as d r e a l = 2.7   cm because of the microphone dimension, electronic board, and the microphone shield. Additionally, each microphone had a cross section, where in the real conditions it was about 0.7 cm. It means it is hard to measure the exact distance between two microphones and it has some errors. Since all cross sections in a microphone are areas for a sound recording, then, based on all limitations, we were forced to have this inter-microphone distance for real data implementation even with a few millimeters difference with the mathematical calculations. Therefore, the differences in the results of our proposed method for the real and simulated data were for this an issue. In the real condition, there are always some inaccuracy factors for the measurements. We found the center of the room and the microphones were located on the table based on the primary measurements. All microphones were connected to the sound recording system, which uses parallel acquisition for all microphone channels. All channel acquisitions were synchronized and there was not any delay between recorded signals in different microphones or channels. The phase error based on the recording condition was very low and was even close to zero based on the audio recording system. In the real room, the table did not make any direct reflection. All the reflected waves from the table will cross to the walls and ceiling firstly, and since all of them were covered with curtains and sound absorption panels, the indirect reflections to the microphones were very few. Both speakers were connected to the two computers for playing the speech and noise with a sampling frequency of F s = 16 , 000   Hz . The microphone, sound, and noise sources were selected in the simulations with exactly the same real conditions for the results to be comparable in these two conditions. Figure 9 shows the time-domain and spectrum of the male speech signal.
The reverberation effect was considered in the experiments to provide the simulation conditions similar to the real scenarios. The image model was implemented in the simulations to produce the reverberation effect similar to the real conditions [41]. The image model produced the room impulse response between the source and microphone by considering the speaker position, microphone location, sampling frequency, room dimension, room reflection coefficients, impulse response length, and reverberation time. The received signal to the microphone was simulated by the convolution between the generated impulse response by the image method and the source signal. The impulse response was generated for the noise and speech sources because both receive the same effect of the room reverberation. In addition, noise was additive with the speech signal in the microphone positions. The room reverberation time was selected as R T 60 = 350   ms , which was considered for a room with a low level of reverberation to be the same as the real conditions. To generate the noisy signal, five types of noise were considered for the simulated and real data such as white noise, babble noise, train noise, car noise, and restaurant noise. Figure 10 shows the time-domain and spectrum for these noisy signals according to a SNR = 0 dB. The noise signal duration was 12.5 s, the same as the speech signal.
The Hamming window with a length of 30 ms was selected for signal blocking to keep the stationarity of the signal in the short time. The projection order was considered as N = 4 to keep the computational complexity in an acceptable range in addition to a proper accuracy of the algorithm. Additionally, the step sizes were chosen as μ 12 = 1 and μ 21 = 1 to provide the fast convergence for the proposed SBAPA in the real-time implementations. The evaluations in this article were implemented by the use of MATLAB software version 2019b on a PC with processor Inter Core i7-7700k, 4.20 GHz, and with 32GB RAM to be able to implement the proposed algorithm in the real-time conditions.
The proposed SBAPA in combination with a proposed circular nested microphone array (CNMA-SBAPA) was compared with the LMS [20], traditional APA [31], RLS [21], DB-MWF [9], and MNMF-MVDR [24] algorithms. These methods were compared because all of them are based on the adaptive filters and multi-channel beamforming as a main category for comparison. There are many methods for comparison with the proposed algorithm but the comparison should be based on the common theme in implementations. Therefore, the adaptive filter-based algorithms were selected for this comparison. The qualitative and quantitative criteria were considered to show the superiority of the proposed method in comparison with other previous works. For this purpose, the SegSNR [7], PESQ [4], MOS [5], and STOI [6] criteria were selected for the comparison. The SegSNR is a quantitative criterion, which shows the improvement in the enhanced signal due to the percentage of the noise power elimination from the noisy signal, namely:
SegSNR ( d B ) = 1 R i = o R 10 log 10 ( j = 0 Q 1 | S j [ n ] | 2 j = 0 Q 1 | S j [ n ] Z j [ n ] | 2 V A D i )
where S [ n ] and Z [ n ] are the clean and enhanced speech signals, respectively. The variable Q is the mean averaging value of the SNR for the output signal. The variable R is the number of only-speech frames and V A D is a speech detector, which is 1 for only-speech frames and 0 for only-noise frames. Therefore, the SegSNR is appropriate to show the speech enhancement performance. Many of the speech enhancement algorithms eliminate some part of the speech signals in addition to the noise frames, which decreases the speech perception for the enhanced signals. Then, three well-known qualitative criteria are considered in the evaluations. The first one is the PESQ, which is defined based on the standard ITU-T P.862 for qualitative evaluations of speech signals in mobile stations [4,42]. In fact, the PESQ criteria is used in the numerical representation of qualitative evaluations for enhanced speech signals. The defined range for this criteria is [−0.5 4.5], where −0.5 and 4.5 show the lowest and highest quality of the enhanced speech, respectively. Additionally, the results were compared with the MOS score criteria. These are qualitative criteria in telecommunication systems that represent the clarity, perception, and intelligibility of the enhanced signal. The MOS criteria are defined based on the standard ITU-T P.800 [5,43] in telecommunication systems. The evaluation results based on the MOS criteria was implemented by the use of some volunteers, by listening to the enhanced signal, where 1 and 5 are the lowest and highest scores in this criteria, respectively. Table 3 shows the defined scores for the MOS criteria in the evaluations.
Finally, the last qualitative criteria for evaluations is the STOI. This criteria predicts the intelligibility of humans based on a series of cases. The speech intelligibility measurement is based on the existence of a series of pre-assumptions, but if the noisy signal is processed based on the time-frequency weighting, the final results are not trustable. The STOI is an objective intelligibility measurement, which represents the highest convolution value by the intelligibility of both noisy and weighted time-frequency noisy signals. In addition, the lowest and highest scores for the STOI criteria are 0 and 1, which represent the best and the worst enhancement performance, respectively.
Firstly, the proposed method was evaluated on the white noise and then, the other colored noise were considered in the experiments. The proposed CNMA-SBAPA was evaluated on real and simulated data in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR algorithms. Figure 11 shows the time-domain and spectrum for the noisy and enhanced signals in the presence of white noise for SNR = 0 dB. As seen in these figures, the proposed CNMA-SBAPA method decreased more level of the noise with less distortion in comparison with other works. However, the numerical values are necessary for comparison. In the following, the experiments were evaluated with quantitative and qualitative criteria.
In the following, the proposed method was compared by numerical criteria with other previous works. Figure 12 shows the SegSNR results in SNRs [−10, −5, 0, 5, 10, and 15] dB for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR for real and simulated data in the presence of white noise. As seen, the proposed method had a superior performance in different ranges of SNRs in comparison with the rest of the works, namely a better noise elimination was reached via the proposed algorithm. For example, the proposed method enhanced the noisy speech signal with SNR = −10 dB to SegSNR = 1.35 dB in comparison with SegSNR = −4.58 dB in LMS, SegSNR = −3.21 dB in APA, SegSNR = −1.57 dB in RLS, SegSNR = −1.68 dB in DB-MWF, and SegSNR = −0.94 dB in MNMF-MVDR. Nevertheless, the quantitative criteria are not enough to properly evaluate a method, and both quantitative and qualitative criteria should be considered in the evaluations.
In addition, the proposed method was compared with previous works by qualitative criteria such as the PESQ, MOS, and STOI. We used 20 volunteers, where they listened first to the clean signal by the headset to have an idea of an excellent signal with a rating of 5 in the MOS scale and a noisy signal (before enhancement), which is the worst option in the MOS scale with a rating of 1. Then, the enhanced signal in a different range of SNRs were played for them, and they were asked to select a rate between 1 and 5 based on the Table 3. Figure 13 shows the PESQ, STOI, and averaged MOS criteria for the enhanced signal by the proposed method in comparison with previous works on real and simulated data for different ranges of SNRs in the presence of white noise. As seen, the proposed method had the best performance in comparison with previous works. For example, the PESQ score was 3.41 in the proposed method in comparison to 1.82 in LMS, 2.51 in APA, 2.73 in RLS, 2.93 in DB-MWF, and 3.1 in MNMF-MVDR, for SNR = 15 dB for simulated data. In addition, the STOI criteria was 0.89 in the proposed method in comparison to 0.73 in LMS, 0.77 in APA, 0.81 in RLS, 0.83 in DB-MWF, and 0.85 in MNMF-MVDR, for SNR = 15 dB. The other criteria for comparison was the average MOS rate, which was 3.5 in the proposed method in comparison to 2.5 in LMS, 2.7 in APA, 2.9 in RLS, 3.0 in DB-MWF, and 3.0 in MNMF-MVDR, for SNR = 15 dB. Therefore, the proposed method was superior for enhancing the noisy signals by considering both quantitative (Figure 12) and qualitative (Figure 13) criteria in comparison to previous works in the presence of white noise. In addition, the proposed method was implemented on colored noises to show the reliability of the results. For this purpose, the proposed method was evaluated on babble, train, car, and restaurant noises for the real and simulated data and for SNR ranges [−10, −5, 0, 5, 10, and 15] dB. Table 4 and Table 5 show the results on the simulated and real data, respectively. As seen from the numbers in these tables, the proposed method had better results in most cases in comparison with traditional methods, which present the reliability of the proposed method in colored noisy conditions. Some of the methods had slightly better results in specific cases, for example in SNR = 15 dB, which cannot be generalized to all cases. In addition, the SegSNR values are shown in these tables to present better comparison with qualitative criteria.
Finally, Table 6 presents the speed of convergence for the proposed method in comparison with other previous works for all white and colored noises in seconds (the required time for convergence based on the configuration of the used PC) on the real data. As shown, the proposed method has a higher speed of convergence in comparison with other algorithms. The main reason for this high speed of convergence is the sub-band processing, because this multiresolution processing provides stationary noise in each frequency band, which is an important factor in the speed of convergence. When the noise is closer to stationary conditions, the speed of convergence is increased in adaptive filter-based algorithms. As clearly shown in this table, the speed of convergence in white noisy conditions was higher than the colored noisy scenarios. Therefore, the proposed CNMA-SBAPA method had superiority for the speech enhancement in comparison with LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR algorithms based on the quantitative SegSNR and qualitative PESQ, MOS, and STOI criteria, as well as the speed of convergence.

5. Conclusions

Speech enhancement is an important application in the signal processing for smart meeting rooms. The aim of speech enhancement is denoising, dereverberation, or denoising–dereverberation at the same time. The speech enhancement is implemented as a pre-processing step to produce the proper signal in such an application as speaker localization, tracking, speech recognition, text-to-speech, estimation the number of speakers, etc. The speech enhancement algorithms are divided into the single and multi-channels methods. The single-channel algorithms are challenging in the speech enhancement processes because of the lack of suitable information in the denoising procedure. In contrast, the multi-channel algorithms increase the enhancement accuracy due to having more information but the computational complexity is increased. In this article, a multi-channel speech enhancement method was proposed based on the microphone array. The microphone array increased the accuracy in the enhanced algorithms based on the increasing of information, but the spatial aliasing decreased the efficiency because of inter-microphone distances. In this article, a uniform circular nested microphone array was proposed for the speech enhancement algorithms. This nested array was designed in a way that the microphones were located at specific distances to eliminate the spatial aliasing, in combination with analysis filters to provide the proper information for the speech enhancement algorithms. In addition, the speech information is different in various frequency bands. Therefore, the specific sub-band processing was proposed to have especial attention to the speech spectrum components. The frequency bands were designed to have the maximum resolution in low frequency components. In the following, the APA was implemented on all frequency bands, which was obtained by the sub-band processing and circular nested microphone array. The projection factor (N=4) was considered for the CNMA-SBAPA in order to keep the computational complexity in an acceptable range along with the superior accuracy. Finally, the synthesis filter bank was implemented on the sub-band signals and the enhanced signal was generated by the summation through all sub-bands. The proposed algorithm was compared with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on the real and simulated data for white and colored noises under the SNRs range [−10,−5,0,5,10, and 15]dB. In all conditions the proposed method had a superior accuracy in comparison with previous works. In addition, the proposed method was compared based on the speed of convergence with previous works, which it was much faster among all the other algorithms. Since the proposed enhancement algorithm was implemented on stationary signals, where its benefit was increasing the speed of convergence in adaptive filters.
One of the future works is reducing the size of the array and decreasing the number of microphones (without having a high effect on the quality) to be applicable for smartphone applications. Even the type of the microphones is important. In this article, we used a high quality microphone, which provides the signals with proper amplitude from the environment. The use of normal microphones in smartphones is another challenge, which could be an area for future work. Another area for future work is to find the best numbers of sub-bands to provide the maximum performance and lowest computational complexity, where the numbers of sub-bands will not be fixed and it should be adaptive based on the speech components.

Author Contributions

Conceptualization, A.D.F. and P.A. and D.Z.-B.; methodology, A.D.F. and P.A.; software, A.D.F., P.I., P.A. and H.D.; validation, M.S., P.P., D.Z.-B. and C.A.-M.; formal analysis, A.D.F. and P.A.; investigation, A.D.F. and P.A.; resources, A.D.F., P.A., D.Z.-B. and P.I.; data curation, A.D.F.; writing—original draft preparation, A.D.F., P.A. and D.Z.-B.; writing—review and editing, P.P.-J., C.A. and D.Z.-B.; supervision, P.I.; project administration, P.A., H.D., M.S. and D.Z.-B.; funding acquisition, P.A. and A.D.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by FONDECYT Postdoctorado No. 3190147, FONDECYT No. 11180107 and ANID PFCHA/Beca de Doctorado Nacional/2019 21190489.

Acknowledgments

This work was supported by the Vicerrectoría de Investigación y Postgrado of the Universidad Tecnológica Metropolitana, the Vicerrectoría de Investigación y Postgrado, and Faculty of Engineering Science of the Universidad Católica del Maule.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APAffine projection
APAAffine projection algorithm
ARAuto-regressive
ARHMM Auto-regressive hidden Markov model
CNMA Circular nested microphone array
CNMA-SBAPACircular nested microphone array in combination with sub-band affine projection algorithm
DB-MWFDistributed multichannel Wiener filter
DNNDeep neural network
FBKFondazione Bruno Kessler
FIRFinite impulse response
HMM Hidden Markov model
HPFHigh-pass filter
IFDInstantaneous frequency deviation
LMR-APALevenberg Marquardt regularized-Affine projection algorithm
LMSLeast mean square
LPFLow-pass filter
MLMaximum likelihood
MSEMean square error
MMSEMinimum mean square error
MNMFMulti-channel non-negative matrix factorization
MNMF-MVDRMultichannel nonnegative matrix factorization-minimum variance distortionless response
MOSMean opinion score
MVDRMinimum variance distortionless response
NLMSNormalized least mean square
NMANested microphone array
OCF-NLMSOrthogonal correction factor-Normalized least mean square
PESQPerceptual evaluation of speech quality
PRAPAPartial rank affine projection algorithm
RLSRecursive least square
SBAPASub-band affine projection algorithm
SCMSpatial covariance matrix
SegSNRSegmental signal-to-noise ratio
SNRSignal-to-noise ratio
STOIShort-time objective intelligibility
STPShort time predictor
TTSText-to-speech
VADVoice activity detector
WFWiener filter

References

  1. Prasad, P.B.M.; Ganesh, M.S.; Gangashetty, S.V. Two microphone technique to improve the speech intelligibility under noisy environment. In Proceedings of the IEEE 14th International Colloquium on Signal Processing & Its Applications (CSPA), Penang, Malaysia, 9–10 March 2018; pp. 13–18. [Google Scholar]
  2. Fukui, M.; Shimauchi, S.; Hioka, Y.; Nakagawa, A.; Haneda, Y. Acoustic echo and noise canceller for personal hands-free video IP phone. IEEE Trans. Consum. Electron. 2016, 62, 454–462. [Google Scholar] [CrossRef]
  3. Ephraim, Y. Statistical-Model-Based Speech Enhancement Systems. Proc. IEEE. 1992, 80, 1526–1555. [Google Scholar] [CrossRef]
  4. Rix, A.W.; Beerends, J.G.; Hollier, M.P.; Hekstra, A.P. Perceptual evaluation of speech quality (PESQ)—A new method for speech quality assessment of telephone net works and codecs. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (Cat. No. 01CH37221), Salt Lake City, UT, USA, 7–11 May 2001; pp. 749–752. [Google Scholar]
  5. Streijl, R.C.; Winkler, S.; Hands, D.S. Mean opinion score (MOS) revisited: Methods and applications, limitations and alternatives. Multimed. Syst. 2016, 22, 213–227. [Google Scholar] [CrossRef]
  6. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. A short-time objective intelligibility measure for time-frequency weighted noisy speech. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 4214–4217. [Google Scholar]
  7. Pollak, P.; Vondrasek, M. Methods for Speech SNR Estimation: Evaluation Tool and Analysis of VAD Dependency. Radio Eng. 2005, 14, 6–11. [Google Scholar]
  8. Loizou, P.C. Speech Enhancement: Theory and Practice, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2007. [Google Scholar]
  9. Doclo, S.; Moonen, M.; Bogaert, T.V.; Wouters, J. Reduced-band width and distributed MWF-based noise reduction algorithms for binaural hearing aids. IEEE Trans. Audio Speech Lang. Process. 2009, 17, 38–51. [Google Scholar] [CrossRef]
  10. Boll, S.F. Suppression of Acoustic Noise in Speech Using Spectral Subtraction. IEEE Trans. Acoust. Speech Signal Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  11. Martin, R. Spectral subtraction based on minimum statistics. In Proceedings of the European Signal Processing Conference, Scotland, UK, 13–16 September 1994; pp. 1182–1185. [Google Scholar]
  12. Ephraim, Y.; Malah, D. Speech enhancement using a minimum mean square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1984, 32, 1109–1121. [Google Scholar] [CrossRef] [Green Version]
  13. Ephraim, Y.; Malah, D. Speech enhancement using a minimum mean-square error log-spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1985, 33, 443–445. [Google Scholar] [CrossRef]
  14. Cohen, I.; Berdugo, B. Noise estimation by minima controlled recursive averaging for robust speech enhancement. IEEE Signal Process. Lett. 2002, 9, 12–15. [Google Scholar] [CrossRef]
  15. Cohen, I. Noise spectrum estimation in adverse environments: Improved minima controlled recursive averaging. IEEE Trans. Speech Audio Process. 2003, 11, 466–475. [Google Scholar] [CrossRef] [Green Version]
  16. Martin, R. Noise power spectral density estimation based on optimal smoothing and minimum statistics. IEEE Trans. Speech Audio Process. 2001, 9, 504–512. [Google Scholar] [CrossRef] [Green Version]
  17. Sameti, H.; Sheikhzadeh, H.; Deng, L.; Brennan, R.L. HMM-based strategies for enhancement of speech signals embedded in non stationary noise. IEEE Trans. Speech Audio Process. 1998, 6, 445–455. [Google Scholar] [CrossRef]
  18. Zhao, D.Y.; Kleijn, W.B. HMM-based gain modeling for enhancement of speech in noise. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 882–892. [Google Scholar] [CrossRef]
  19. Deng, F.; Bao, C.C.; Kleijin, W.B. Sparse Hidden Markov Models for Speech Enhancement in Non-Stationary Noise Environments. IEEE Trans. Audio Speech Lang. Process. 2015, 23, 1973–1987. [Google Scholar] [CrossRef]
  20. Geravanchizadeh, M.; Osgouei, S.G. Dual-channel speech enhancement using normalized fractional least-mean-squares algorithm. In Proceedings of the 19th Iranian Conference on Electrical Engineering, Tehran, Iran, 17–19 May 2011; pp. 1–5. [Google Scholar]
  21. Rakesh, P.; Kumar, T.K. A novel RLS based adaptive filtering method for speech enhancement. Int. J. Electr. Comput. Electron. Commun. Eng. 2015, 9, 153–158. [Google Scholar]
  22. He, Q.; Bao, F.; Bao, C. Multiplicative Update of Auto-Regressive Gains for Codebook-Based Speech Enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 457–468. [Google Scholar] [CrossRef]
  23. Tavakoli, V.M.; Jensen, J.R.; Christensen, M.G.; Benesty, J. A Framework for Speech Enhancement with Ad Hoc Microphone Arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 1038–1051. [Google Scholar] [CrossRef]
  24. Shimada, K.; Bando, Y.; Mimura, M.; Itoyama, K.; Yoshii, K.; Kawahara, T. Unsupervised Speech Enhancement Based on Multi channel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 960–971. [Google Scholar] [CrossRef]
  25. Kavalekalam, M.S.; Nielsen, J.K.; Boldt, J.B.; Christensen, M.G. Model-Based Speech Enhancement for Intelligibility Improvement in Binaural Hearing Aids. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 99–113. [Google Scholar] [CrossRef] [Green Version]
  26. Valentini-Botinhao, C.; Yamagishi, J. Speech Enhancement of Noisy and Reverberant Speech for Text-to-Speech. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1420–1433. [Google Scholar] [CrossRef]
  27. Wang, Y.; Brookes, M. Model-Based Speech Enhancement in the Modulation Domain. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 580–594. [Google Scholar] [CrossRef] [Green Version]
  28. Koutrouvelis, A.I.; Hendriks, R.C.; Heusdens, R.; Jensen, J. Robust Joint Estimation of Multi microphone Signal Model Parameters. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 1136–1150. [Google Scholar] [CrossRef] [Green Version]
  29. Zheng, Y.R.; Goubran, R.A.; El-Tanany, M. Experimental evaluation of a nested microphone array with adaptive noise cancellers. IEEE Trans. Instrum. Meas. 2004, 53, 777–786. [Google Scholar] [CrossRef]
  30. Haykin, S. Adaptive Filter Theory, 4th ed.; Prentice-Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  31. Gonzalez, A.; Ferrer, M.; Albu, F.; Diego, M. Affine projection algorithms: Evolution to smart and fast algorithms and applications. In Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 1965–1969. [Google Scholar]
  32. Sankaran, S.G.; Beex, A.A.L. Normalized LMS algorithm with orthogonal correction factors. In Proceedings of the Conference Record of the Thirty-First Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 2–5 November 1997; pp. 1670–1673. [Google Scholar]
  33. Kratzer, S.G.; Morgan, D.R. The partial Rank Algorithm for adaptive beamforming. In Proceedings of the SPIE0564, Real-Time Signal Processing VIII, San Diego, CA, USA, 22–23 August 1985; pp. 9–14. [Google Scholar]
  34. Ozeki, K.; Umeda, T. An adaptive filtering algorithm using an orthogonal projection to an affine subspace and its properties. Electron. Commun. Jpn. 1984, 67-A, 19–27. [Google Scholar] [CrossRef]
  35. Gay, S.L.; Benesty, J. Acoustic Signal Processing for Telecommunication, 2nd ed.; Springer: Boston, MA, USA, 2000. [Google Scholar]
  36. Waterschoot, T.V.; Rombouts, G.; Moonen, M. Optimally regularized adaptive filtering algorithms for room acoustic signal enhancement. Signal Process. 2008, 88, 594–611. [Google Scholar] [CrossRef]
  37. Gabrea, M. Double affine projection algorithm-based speech enhancement algorithm. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Hong Kong, China, 6–10 April 2003; pp. I904–I907. [Google Scholar]
  38. Shin, H.C.; Sayed, A.H.; Song, W.J. Variable step-size NLMS and affine projection algorithms. IEEE Signal Process Lett. 2004, 11, 132–135. [Google Scholar] [CrossRef]
  39. Garofolo, J.S.; Lamel, L.F.; Fisher, W.M.; Fiscus, J.G.; Pallett, D.S.; Dahlgren, N.L.; Zue, V. TIMIT Acoustic-Phonetic Continuous Speech Corpus LDC93S1. Web Download. Philadelphia: Linguistic Data Consortium, 1993. Available online: https://catalog.ldc.upenn.edu/LDC93S1 (accessed on March 2019).
  40. Schwartz, O.; David, A.; Shahen-Tov, O.; Gannot, S. Multi-microphone voice activity and single-talk detectors based on steered-response power output entropy. In Proceedings of the IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE), Eilat, Israel, 12–14 December 2018; pp. 1–4. [Google Scholar]
  41. Allen, J.; Berkley, D. Image method for efficiently simulating small-room acoustics. J. Acoust. Soc. Am. 1979, 65, 943–950. [Google Scholar] [CrossRef]
  42. ITU-T: Methods for Subjective Determination of Transmission Quality; Recommendation P.862; International Telecommunications Union (ITU-T): Place des Nations, Geneva, Switzerland, 1996.
  43. ITU-T: Methods for Subjective Determination of Transmission Quality; Recommendation P.800; International Telecommunications Union (ITU-T): Place des Nations, Geneva, Switzerland, 1996.
Figure 1. The block diagram of the proposed circular nested microphone array in combination with the sub-band affine projection algorithm (CNMA-SBAPA) for the speech enhancement.
Figure 1. The block diagram of the proposed circular nested microphone array in combination with the sub-band affine projection algorithm (CNMA-SBAPA) for the speech enhancement.
Applsci 10 03955 g001
Figure 2. The proposed uniform CNMA and allocated microphones for each subarray.
Figure 2. The proposed uniform CNMA and allocated microphones for each subarray.
Applsci 10 03955 g002
Figure 3. The tree structure for (a) analysis and (b) synthesis filters in CNMA.
Figure 3. The tree structure for (a) analysis and (b) synthesis filters in CNMA.
Applsci 10 03955 g003
Figure 4. The frequency response for the analysis filter banks.
Figure 4. The frequency response for the analysis filter banks.
Applsci 10 03955 g004
Figure 5. The general structure of the adaptive filter for denoising applications.
Figure 5. The general structure of the adaptive filter for denoising applications.
Applsci 10 03955 g005
Figure 6. (a) The general structure of the proposed denoising system, (b) the simplified presented model for the two-microphone system, and(c) the affine projection algorithm (APA) structure for the two-microphone.
Figure 6. (a) The general structure of the proposed denoising system, (b) the simplified presented model for the two-microphone system, and(c) the affine projection algorithm (APA) structure for the two-microphone.
Applsci 10 03955 g006
Figure 7. A view of the simulated room with the positions of speakers, noise source, and microphones.
Figure 7. A view of the simulated room with the positions of speakers, noise source, and microphones.
Applsci 10 03955 g007
Figure 8. The real recording room in the speech processing laboratory, FBK, Trento, Italy.
Figure 8. The real recording room in the speech processing laboratory, FBK, Trento, Italy.
Applsci 10 03955 g008
Figure 9. The time-domain and spectrum for the male speech signal.
Figure 9. The time-domain and spectrum for the male speech signal.
Applsci 10 03955 g009
Figure 10. The time-domain and spectrum for a noisy speech signal with (a) white noise, (b) babble noise, (c) train noise, (d) car noise, and (e) restaurant noise according to a signal-to-noise ratio (SNR) = 0 dB.
Figure 10. The time-domain and spectrum for a noisy speech signal with (a) white noise, (b) babble noise, (c) train noise, (d) car noise, and (e) restaurant noise according to a signal-to-noise ratio (SNR) = 0 dB.
Applsci 10 03955 g010aApplsci 10 03955 g010b
Figure 11. The time-domain and spectrum representation for (a) white noisy signal and enhanced signal by the (b) least mean square (LMS), (c) APA, (d) recursive least square (RLS), (e)distributed multichannel Wiener filter (DB-MWF), (f) multichannel nonnegative matrix factorization-minimum variance distortionless response (MNMF-MVDR), and (g) proposed CNMA-SBAPA for SNR = 0 dB.
Figure 11. The time-domain and spectrum representation for (a) white noisy signal and enhanced signal by the (b) least mean square (LMS), (c) APA, (d) recursive least square (RLS), (e)distributed multichannel Wiener filter (DB-MWF), (f) multichannel nonnegative matrix factorization-minimum variance distortionless response (MNMF-MVDR), and (g) proposed CNMA-SBAPA for SNR = 0 dB.
Applsci 10 03955 g011aApplsci 10 03955 g011bApplsci 10 03955 g011c
Figure 12. The segmental signal-to-noise ratio (SegSNR) comparison between the proposed CNMA-SBAPA, LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on (a) simulated and (b) real data for white noise.
Figure 12. The segmental signal-to-noise ratio (SegSNR) comparison between the proposed CNMA-SBAPA, LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on (a) simulated and (b) real data for white noise.
Applsci 10 03955 g012
Figure 13. The perceptual evaluation of speech quality (PESQ), short-time objective intelligibility (STOI), and averaged mean opinion score (MOS) comparison between the proposed CNMA-SBAPA and the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods for (a) simulated and (b) real data by considering the white noise.
Figure 13. The perceptual evaluation of speech quality (PESQ), short-time objective intelligibility (STOI), and averaged mean opinion score (MOS) comparison between the proposed CNMA-SBAPA and the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods for (a) simulated and (b) real data by considering the white noise.
Applsci 10 03955 g013
Table 1. The information to design the proposed uniform CNMA.
Table 1. The information to design the proposed uniform CNMA.
BandBandwidthAnalysis Filter Bankfcdlim
1B1 = [3900–7800] HzB1,1 = [6825–7800] Hz
B1,2 = [5850–6825] Hz
B1,3 = [4875–5850] Hz
B1,4 = [3900–4875] Hz
5850 Hz<2.2 cm
2B2 = [1950–3900] HzB2,1 = [2925–3900] Hz
B2,2 = [1950–2925] Hz
2925 Hz<4.4 cm
3B3 = [975–1950] HzB3,1 = [1425–1950] Hz
B3,2 = [975–1425] Hz
1462 Hz<8.8 cm
4B4 = [50–975] HzB4,1 = [512–975] Hz
B4,2 = [50–512] Hz
512 Hz<17.6 cm
Table 2. The required information to design the analysis filter bank for sub-band processing in the proposed CNMA-SBAPA algorithm.
Table 2. The required information to design the analysis filter bank for sub-band processing in the proposed CNMA-SBAPA algorithm.
FiltersBandwidth (Hz)fmin (Hz)fmax (Hz)Filter Length (Samples)
F1[n]4625051293
F2[n]462512975115
F3[n]4509751425102
F4[n]52514251950124
F5[n]97519502925109
F6[n]97529253900118
F7[n]97539004875131
F8[n]97548755850140
F9[n]97558506825146
F10[n]97568257800151
Table 3. The numerical scores for the mean opinion score (MOS) criteria in the evaluation process.
Table 3. The numerical scores for the mean opinion score (MOS) criteria in the evaluation process.
RatingQuality (Standard ITU-T P.800)Impairment
5ExcellentImperceptible
4GoodPerceptible but not annoying
3FairSlightly annoying
2PoorAnnoying
1BadVery annoying
Table 4. The comparison between PESQ, MOS, STOI, and SegSNR for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on the simulated data for colored noises such as: train, babble, car, and restaurant noises in different range of SNRs (the bold numbers are the best results).
Table 4. The comparison between PESQ, MOS, STOI, and SegSNR for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on the simulated data for colored noises such as: train, babble, car, and restaurant noises in different range of SNRs (the bold numbers are the best results).
SNR (dB)MethodsBabble NoiseTrain NoiseCar NoiseRestaurant Noise
SegSNRPESQSTOIMOSSegSNRPESQSTOIMOSSegSNRPESQSTOIMOSSegSNRPESQSTOIMOS
LMS−5.230.340.441.10−5.740.290.411.05−6.250.250.361.05−6.870.180.371.00
−10APA−4.630.630.511.15−5.150.560.471.10−5.960.480.431.05−6.080.520.391.10
RLS−2.910.860.581.45−3.290.810.511.35−3.980.740.481.30−4.610.650.441.30
DB-MWF−2.480.920.571.55−3.080.890.531.45−3.460.780.501.45−4.770.790.471.40
MNMF-MVDR−2.231.040.601.60−2.830.960.561.50−3.290.850.531.55−4.290.840.481.45
CNMA-SBAPA−1.711.190.651.85−2.091.160.631.70−2.691.030.591.65−3.140.990.561.60
LMS−2.660.480.531.50−3.020.440.511.40−3.590.390.461.35−3.780.410.471.35
−5APA−1.890.760.591.55−2.130.680.561.45−2.410.610.511.40−3.260.650.541.4
RLS0.570.910.651.75−0.080.820.621.70−1.140.790.581.60−1.970.720.601.55
DB-MWF1.180.980.671.900.810.940.621.80−0.350.870.601.65−0.290.930.591.70
MNMF-MVDR1.971.130.681.951.361.060.651.950.871.030.611.750.960.980.611.70
CNMA-SBAPA3.631.430.732.303.061.360.702.202.731.320.682.152.261.200.642.10
LMS3.580.750.611.653.130.660.561.552.690.590.521.52.240.540.491.55
0APA4.071.140.631.83.751.050.611.703.210.990.571.653.560.930.551.65
RLS4.121.530.712.053.981.470.661.953.671.410.641.93.921.360.621.85
DB-MWF4.831.610.722.254.391.590.682.003.981.580.661.954.111.490.651.95
MNMF-MVDR5.071.720.752.354.561.670.72.254.311.690.692.154.391.610.652.10
CNMA-SBAPA5.402.040.782.855.131.980.752.704.951.910.732.654.781.840.752.55
LMS8.591.220.671.758.461.140.641.708.251.090.601.78.171.010.571.65
5APA8.961.630.692.208.741.570.662.158.391.50.632.058.221.460.642.00
RLS9.282.210.762.409.082.120.732.358.802.070.712.358.641.950.672.30
DB-MWF9.562.320.752.609.322.080.712.609.122.190.702.408.822.070.662.45
MNMF-MVDR10.122.440.782.759.662.340.752.709.542.310.722.609.252.240.682.50
CNMA-SBAPA10.632.590.823.2510.342.530.803.1010.212.480.782.9510.032.490.732.90
LMS12.521.530.692.112.371.470.682.0512.111.390.652.0011.951.410.631.95
10APA12.862.20.732.5512.622.130.702.4512.432.070.652.4012.272.080.662.45
RLS13.472.450.782.7013.282.390.752.612.862.330.712.5512.542.380.702.50
DB-MWF13.212.530.782.8013.312.480.772.7512.752.250.732.6512.562.490.722.55
MNMF-MVDR13.522.670.792.9513.432.610.782.8512.982.460.762.7012.712.650.752.65
CNMA-SBAPA14.023.080.833.2013.592.960.823.1013.282.870.793.0513.092.920.792.95
LMS15.121.760.722.4515.091.690.722.3515.131.640.692.3015.41.680.72.25
APA15.552.440.762.6515.482.370.752.5515.352.280.762.5015.322.250.742.45
15RLS15.742.610.812.8515.562.580.792.7515.352.510.802.6515.482.400.792.65
DB-MWF15.522.990.833.0015.382.830.762.9015.242.720.792.7515.522.710.802.80
MNMF-MVDR15.633.030.843.1015.422.960.783.0515.422.930.812.9015.452.800.832.85
CNMA-SBAPA15.713.310.893.3515.523.220.843.3015.693.240.803.2515.443.190.823.20
Table 5. The comparison between PESQ, MOS, STOI, and SegSNR for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on the real data for colored noises such as: train, babble, car, and restaurant noises in different range of SNRs (the bold numbers are the best results).
Table 5. The comparison between PESQ, MOS, STOI, and SegSNR for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods on the real data for colored noises such as: train, babble, car, and restaurant noises in different range of SNRs (the bold numbers are the best results).
SNR (dB)MethodsBabble NoiseTrain NoiseCar NoiseRestaurant Noise
SegSNRPESQSTOIMOSSegSNRPESQSTOIMOSSegSNRPESQSTOIMOSSegSNRPESQSTOIMOS
LMS−5.560.320.411.00−5.820.250.391.00−6.460.180.361.00−6.920.140.351.00
−10APA−4.820.610.471.10−5.420.550.461.15−5.830.510.421.10−6.210.480.391.05
RLS−3.040.820.521.35−3.870.740.511.30−4.520.710.491.20−4.860.640.451.15
DB-MWF−2.980.860.551.40−3.430.760.531.35−4.130.750.501.30−4.510.670.451.25
MNMF-MVDR−2.540.930.571.50−3.010.830.521.45−3.810.800.511.35−4.090.750.471.30
CNMA-SBAPA−2.211.080.581.65−2.671.010.561.50−3.140.960.531.45−3.460.920.531.40
LMS−2.830.460.501.45−3.20.440.481.40−3.540.410.471.30−3.920.380.431.25
−5APA−2.040.720.551.50−2.570.680.531.45−3.090.640.521.45−3.470.610.491.35
RLS0.140.880.621.60−0.520.820.611.55−1.450.740.581.50−1.880.690.541.50
DB-MWF0.930.910.641.750.120.850.641.65−0.780.790.601.60−1.570.730.561.65
MNMF-MVDR1.531.020.671.900.840.970.651.700.210.930.591.75−0.840.880.591.75
CNMA-SBAPA3.291.290.702.102.761.210.672.052.251.160.641.951.761.110.611.90
LMS3.250.720.571.52.840.650.521.452.430.590.511.352.070.530.461.30
0APA3.991.050.621.753.711.010.571.653.670.940.521.63.440.890.511.50
RLS4.151.490.691.954.031.440.651.853.951.380.621.83.791.330.591.70
DB-MWF4.321.540.72.054.261.510.671.954.041.440.641.93.931.400.601.75
MNMF-MVDR4.511.690.722.204.491.580.672.054.291.560.652.004.171.580.631.95
CNMA-SBAPA5.031.950.742.554.861.870.712.504.711.760.702.454.581.720.672.45
LMS8.411.180.661.68.221.090.621.558.141.020.591.508.010.960.561.50
5APA8.731.580.652.058.521.520.632.008.391.480.601.958.251.410.611.90
RLS9.042.150.742.258.962.040.712.28.711.960.702.158.521.890.662.05
DB-MWF9.322.230.732.59.382.050.72.48.832.060.702.258.861.920.652.10
MNMF-MVDR9.582.340.762.659.622.260.722.559.222.190.712.409.322.080.662.35
CNMA-SBAPA10.272.480.783.0010.082.410.752.909.952.360.722.859.762.320.692.75
LMS12.431.480.672.0512.261.450.661.9512.091.400.641.9011.871.370.651.95
10APA12.912.150.702.4012.622.110.692.3512.472.070.672.3512.312.010.662.25
RLS13.282.410.752.5513.012.390.742.5012.762.360.722.4012.622.310.732.35
DB-MWF13.362.480.762.7513.222.490.732.6012.892.470.712.5012.752.380.742.45
MNMF-MVDR13.512.560.762.9013.342.550.752.7512.962.510.742.6512.832.480.722.70
CNMA-SBAPA13.882.920.813.0513.512.880.793.0013.192.820.772.9512.992.790.753.00
LMS15.081.740.702.4015.11.690.672.3015.041.650.672.1515.251.600.662.10
APA15.412.380.732.5515.382.350.722.5015.862.280.702.4015.452.210.72.45
15RLS15.242.560.842.7015.292.510.772.7015.922.440.752.6015.762.370.742.55
DB-MWF15.392.810.852.8515.362.620.772.8015.812.500.782.7515.782.460.772.65
MNMF-MVDR15.482.940.823.0015.422.740.792.9015.802.680.802.8515.732.590.742.70
CNMA-SBAPA15.353.220.833.2015.493.150.823.2515.833.130.773.1515.613.090.733.10
Table 6. The speed of convergence, in seconds, for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods for white and colored noises on real data in different range of SNRs (the bold numbers are the best results).
Table 6. The speed of convergence, in seconds, for the proposed CNMA-SBAPA in comparison with the LMS, traditional APA, RLS, DB-MWF, and MNMF-MVDR methods for white and colored noises on real data in different range of SNRs (the bold numbers are the best results).
SNR(dB)MethodsSpeed of Convergence (Seconds)
White NoiseBabble NoiseTrain NoiseCar NoiseRestaurant Noise
LMS0.5410.5820.610.6540.668
−10APA0.5160.5510.5920.6110.627
RLS0.4220.4680.4960.5390.546
DB-MWF0.5860.6120.6520.6730.695
MNMF-MVDR0.510.5390.570.5920.637
CNMA-SBAPA0.3560.3670.3930.4190.427
LMS0.5370.5560.5790.6010.634
−5APA0.4970.5270.5410.560.572
RLS0.4030.4290.4470.4820.506
DB-MWF0.5450.5930.6150.6360.658
MNMF-MVDR0.4940.5090.5270.5530.587
CNMA-SBAPA0.3370.3560.3790.3910.411
LMS0.5160.5380.5620.5680.595
0APA0.4730.4820.5020.5360.539
RLS0.3960.4090.4270.4350.452
DB-MWF0.5310.5630.5790.6020.621
MNMF-MVDR0.4850.4980.5160.5310.546
CNMA-SBAPA0.3180.3290.350.3580.362
LMS0.4920.5050.5170.5250.529
5APA0.4640.4790.4920.4670.503
RLS0.3880.3950.4010.4120.418
DB-MWF0.5070.5120.5440.5650.586
MNMF-MVDR0.4590.4660.4870.5030.525
CNMA-SBAPA0.3270.3360.3470.3590.365
LMS0.4880.4980.5090.5210.538
10APA0.4510.4670.4830.4990.507
RLS0.3690.3810.3890.3950.411
DB-MWF0.4780.4960.5130.5430.559
MNMF-MVDR0.4370.4580.4790.4910.516
CNMA-SBAPA0.3050.3280.3390.3520.368
LMS0.4720.4850.4930.4980.506
APA0.4630.4740.4780.4850.49
15RLS0.3720.3760.3850.3960.408
DB-MWF0.4430.4550.4620.4810.494
MNMF-MVDR0.410.4270.4480.4630.48
CNMA-SBAPA0.2990.3190.3250.3490.374

Share and Cite

MDPI and ACS Style

Dehghan Firoozabadi, A.; Irarrazaval, P.; Adasme, P.; Zabala-Blanco, D.; Durney, H.; Sanhueza, M.; Palacios-Játiva, P.; Azurdia-Meza, C. Multiresolution Speech Enhancement Based on Proposed Circular Nested Microphone Array in Combination with Sub-Band Affine Projection Algorithm. Appl. Sci. 2020, 10, 3955. https://doi.org/10.3390/app10113955

AMA Style

Dehghan Firoozabadi A, Irarrazaval P, Adasme P, Zabala-Blanco D, Durney H, Sanhueza M, Palacios-Játiva P, Azurdia-Meza C. Multiresolution Speech Enhancement Based on Proposed Circular Nested Microphone Array in Combination with Sub-Band Affine Projection Algorithm. Applied Sciences. 2020; 10(11):3955. https://doi.org/10.3390/app10113955

Chicago/Turabian Style

Dehghan Firoozabadi, Ali, Pablo Irarrazaval, Pablo Adasme, David Zabala-Blanco, Hugo Durney, Miguel Sanhueza, Pablo Palacios-Játiva, and Cesar Azurdia-Meza. 2020. "Multiresolution Speech Enhancement Based on Proposed Circular Nested Microphone Array in Combination with Sub-Band Affine Projection Algorithm" Applied Sciences 10, no. 11: 3955. https://doi.org/10.3390/app10113955

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop