Next Article in Journal
Multitemporal SAR Image Despeckling Based on a Scattering Covariance Matrix of Image Patch
Next Article in Special Issue
Forensic Speaker Verification Using Ordinary Least Squares
Previous Article in Journal
An Efficient Hardware-Oriented Single-Pass Approach for Connected Component Analysis
Previous Article in Special Issue
Window-Based Constant Beamwidth Beamformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual Microphone Voice Activity Detection Based on Reliable Spatial Cues

1
School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology, 123 Cheomdan-gwagiro, Buk-gu, Gwangju 61005, Korea
2
AI Technology Unit, SK Telecom, 100 Eulji-ro, Jung-gu, Seoul 04551, Korea
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(14), 3056; https://doi.org/10.3390/s19143056
Submission received: 30 May 2019 / Revised: 5 July 2019 / Accepted: 10 July 2019 / Published: 11 July 2019

Abstract

:
Two main spatial cues that can be exploited for dual microphone voice activity detection (VAD) are the interchannel time difference (ITD) and the interchannel level difference (ILD). While both ITD and ILD provide information on the location of audio sources, they may be impaired in different manners by background noises and reverberation and therefore can have complementary information. Conventional approaches utilize the statistics from all frequencies with fixed weight, although the information from some time–frequency bins may degrade the performance of VAD. In this letter, we propose a dual microphone VAD scheme based on the spatial cues in reliable frequency bins only, considering the sparsity of the speech signal in the time–frequency domain. The reliability of each time–frequency bin is determined by three conditions on signal energy, ILD, and ITD. ITD-based and ILD-based VADs and statistics are evaluated using the information from selected frequency bins and then combined to produce the final VAD results. Experimental results show that the proposed frequency selective approach enhances the performances of VAD in realistic environments.

1. Introduction

Voice activity detection (VAD) which decides if the speech signal is present in the current frame of the input signal has become a crucial part of the speech enhancement, noise estimation, pitch extraction, and the variable rate speech codecs [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32]. Single microphone VAD usually utilizes energy-related features, statistical model-based statistics including signal-to-noise ratios (SNRs) and likelihood ratios [8,9,10,11], or speech-specific features such as linear predictive coefficients, formant shape, zero-crossing rate, and cepstral features [12,13,14,15,16,17,18,19,20]. Recently, single microphone VADs using deep neural networks are also proposed [22,23,24,25,26], although some of the application scenarios such as mobile devices may not accommodate much computation. These single channel approaches, however, cannot utilize spatial diversity of the sound sources and therefore the performance is limited when strong speech-like interference is present. Since the devices with multiple microphones have become popular these days, reliable multi-microphone VAD becomes more important.
Like the human binaural perception, two main cues that capture the spatial information may be the differences of the arrival times and signal levels in the microphones. These statistics, which are called interchannel time difference (ITD) and interchannel level difference (ILD), can be exploited for dual microphone VAD. When the range of the direction-of-arrival (DoA) of the desired signal is known in advance as in the case of the handset mode of the mobile phone, the VAD can be constructed based on the estimated DoAs. There have been several approaches to the DoA estimation and the VAD based on ITD. The generalized cross correlation with phase transform (GCC-PHAT) [33] may be the most popular approach for DoA estimation and it can be utilized for ITD-based VAD. The long term information of interchannel phase difference (LTIPD) was also proposed as a test statistics for ITD-based VAD, which measures how consistently the signal energy is concentrated in a small DoA range [27]. ILD can also be utilized for VAD when the target source is located close to one of the microphones or there are obstacles between the target source and one of the microphones. In [28], the normalized difference of power spectral density (NDPSD) was proposed as a test statistic for VAD. Choi and Chang [29] proposed two step power level difference ratio (PLDR) using two different smoothing factors as an alternative measure of ILD. There have also been approaches exploiting both ITD and ILD information, as they reflect different characteristics of spatial diversity. In [30], voice activity is decided using a support vector machine (SVM) for which the inputs include both ITD-based and ILD-based features. This approach can configure the relative importance of the ILD or ITD related features in different frequency bins based on the training data, but cannot modify it dynamically according to the input signal. Statistical model-based approaches were also proposed for multichannel VADs adopting complex Gaussian model [31] or the spherically invariant random process [32] for the distribution of each frequency component.
As speech signal is sparse in the time–frequency (TF) domain in nature, there are always TF bins with SNRs in which the ILD or ITD information are more reliable and those with low SNRs where the ILD or ITD information may not be useful at all in noisy environments. In this letter, we propose a VAD based on both ITD and ILD information from reliable frequency bins only. The reliability is determined for each TF bin by three conditions on signal energy, ILD, and ITD. ITD-based and ILD-based VADs or statistics evaluated using only the reliable frequency bins are combined together to construct final voice activity decision. The long term spectral divergence [17] and the subband order statistics filter [18] uses the energies in the specific percentiles among the neighboring frames for the given subband to evaluate the statistics for VAD. In [21], a subset of the frequency bins are utilized for the SNR-based VAD, but the selection criterion is only based on the energies unlike the proposed approach. Experimental results showed that the proposed approach can enhance the performance of the dual microphone VAD.

2. Dual Microphone VADs Based on Spatial Cues

The dual microphone VAD can perform better if prior knowledge on the location of the desired speaker is available. One example of the scenarios in which the range of the locations of the desired signal source is known in advance is the handset mode of the mobile phone, where user’s mouth is much closer to the primary microphone.
Let Y 1 ( l , k ) and Y 2 ( l , k ) be the K-point short-time Fourier transform (STFT) coefficients of the signals from the primary and secondary microphones for the kth frequency bin at the lth frame, respectively. The test statistics for the NDPSD-based VAD [28] are the difference of the powers normalized by the average power of two microphone signals, which are not dependent on the absolute signal level. The NDPSD for each frequency bin is given as
Δ Φ ( l , k ) = | Y 1 ( l , k ) | 2 | Y 2 ( l , k ) | 2 | Y 1 ( l , k ) | 2 + | Y 2 ( l , k ) | 2 ,
where | · | denotes the magnitude. The voice activity is decided by comparing the NDPSD averaged over all frequency bins with a threshold ξ N D P S D as
V N D P S D ( l ) = 1 , if 2 K k = 1 K / 2 Δ Φ ( l , k ) ξ N D P S D 0 , otherwise .
It is noted that each frequency bin contributes to the final test statistic equally, although the absolute values of Δ Φ ( l , k ) do not have much information on the presence of speech in the current frame if Δ Φ ( l , k ) is negative.
The two step PLDR method [29] also takes the difference of the powers in microphone signals as basic information. The power differences of the input signals and noises in the microphones for each frequency are recursively smoothed with two different smoothing factors to produce long-term and short-term smoothed input and noise power level difference, Δ P ^ Y L T ( l , k ) , Δ P ^ Y S T ( l , k ) , Δ P ^ N L T ( l , k ) , and Δ P ^ N S T ( l , k ) , with the help of estimated speech presence probability. Then, the l o g Δ P ^ Y j ( l , k ) Δ P ^ N j ( l , k ) are averaged over all frequency bins to produce two PLDRs. They are converted to two a posteriori probabilities of speech presence, and then thresholded to determine voice activity. Like NDPSD, PLDRs can also be strongly affected by the presence of the interfering source near the secondary microphone.
Another important source of information about the location of the sound sources is ITD. Ref. [27] proposed LTIPD which measures how much energy is concentrated in the frequency bins for which the DoA estimates for nearby frames fall into the same small DoA sector. The target DoA range is divided into U overlapped sectors of which the width are equal. LTIPD is then defined as
E ( l ) = max 1 i U k : C i ( l , k ) > κ i | Y 1 ( l , k ) | 2 ,
where i denotes the indices of the sectors, C i ( l , k ) is the number of frames in which DoA estimate for the k-th frequency bin indicates the ith sector among the last L frames, and κ i is the threshold of the concentration of DoAs. The VAD based on LTIPD is given as
V L T I P D ( l ) = 1 , if E ( l ) ξ L T I P D 0 , otherwise .
It is noted that the LTIPD-based VAD includes the reliable frequency selection, but the selection criterion is based only on the ITD information.
The performance of VAD may be improved by utilizing both ITD and ILD information simultaneously, as two statistics provide different information on the spatial diversity. In [30], the VADs based on both ITD and ILD information are proposed. The simplest way proposed in [30] is the logical combination of the ITD-based and ILD-based VADs. The VAD using the logical “and” operation of the voice activities from ITD and ILD is given as
V A N D ( l ) = 1 , if V I T D ( l ) = 1 and V I L D ( l ) = 1 0 , otherwise ,
which was found to be more effective than the “OR” operation from several experiments. Among the candidates for the V I T D and V I L D , the combination of the LTIPD-based VAD and the NDPSD-based VAD performed the best [30]. Another method proposed in [30] is to build a VAD using SVM for which the input includes both the ILD-based and ITD-based features. After training with the clean speech data mixed with various noises, the output of the SVM, y x ( l ) = w o T x ( l ) + b o where x ( l ) is the input of the SVM and w o and b 0 are the weight vector and the bias for the optimal hyperplane given the training set, respectively [34], is used to estimate the a posteriori probability p ( V S V M = 1 | x ( l ) ) in the test phase. Then, the decision rule becomes
V S V M ( l ) = 1 , if p ( V S V M = 1 | x ( l ) ) ξ S V M 0 , otherwise .
As for the input of the SVM, it is reported that | Y 1 ( l , k ) | 2 , | Y 2 ( l , k ) | 2 , and the phase difference between Y 1 ( l , k ) and Y 1 ( l , k ) , Δ ψ ( l , k ) , showed the best performance among a number of ILD- and ITD-related features [30]. However, in our experiments, including p L T ( l ) and p S T ( l ) from the two-step PLDR approach on top of | Y 1 ( l , k ) | 2 , | Y 2 ( l , k ) | 2 , and Δ ψ ( l , k ) slightly improved the performance of V S V M . Although the SVM can put different importance on the statistics from different frequency bins, the weights are fixed in the test phase and cannot be dynamically changed from frame to frame.
The performance of each VAD can be significantly enhanced by introducing the hangover scheme, which requires several consecutive frames with instantaneous VAD of 0 to make the final VAD to be 0. The number of hangover frames remains as a tunable parameter along with the thresholds for each VAD. For V A N D , the hangover scheme is applied when evaluating V I T D , V I L D , and V A N D with three separate hangover parameters.

3. Dual Microphone VAD Using Reliable Spatial Cues

As speech components are sparsely distributed in the TF domain, the spatial cues in some of the TF bins are useful while those from other TF bins are not reliable. In this paper, we propose a dual microphone VAD based on the spatial information from the selected frequency bins with high reliability, which is determined in each frame by signal energies, ILDs, and ITDs for the corresponding frequency bins. After we determine the reliability of information from each TF bin, the test statistics for conventional VAD approaches are modified to consider reliable TF bins only.
Let us denote the spectral mask to select reliable frequency bins for the k-th frequency bin at the l-th frame as m ( l , k ) , i.e., m ( l , k ) = 1 for the TF bins with reliable spatial information and m ( l , k ) = 0 for other TF bins. The first condition to determine the reliability of the spatial information in each TF bin is on the signal energy. If the energy of the input noisy signal in the primary microphone is not high enough, the probability of speech presence is low and the ILD and ITD information is vulnerable to measurement noises. Thus, the first sub-mask m 1 ( l , k ) is constructed based on the input signal energy as follows:
m 1 ( l , k ) = 1 , if | Y 1 ( l , k ) | 2 η 1 0 , otherwise ,
where η 1 is the threshold.
The second and third conditions determine the reliability in each bin with ILD and ITD information, respectively. The second sub-mask m 2 ( l , k ) becomes 1 if the instantaneous ILD for the frequency is high enough:
m 2 ( l , k ) = 1 , if log | Y 1 ( l , k ) | 2 | Y 2 ( l , k ) | 2 η 2 0 , otherwise ,
where η 2 is the threshold for the level difference. Since this sub-mask is only used to select reliable spectral bins from which the test statistics for the dual microphone VAD are computed, the threshold η 2 is not set to maximize the performance of the VAD, but is configured to discriminate the frequency bins that may contain desired speech signals and those in which there is definitely no speech signal. This sub-mask would enhance the robustness of the VAD to the noise sources that are close to the secondary microphone, which may have a huge adverse impact on the ILD statistics. The frequency bins with low enough ILD will not have the desired speech and should be excluded in the computation of the ITD-based test statistics. The third condition based on ITD also aims to eliminate the TF bins in which the speech is absent for sure. The third sub-mask m 3 ( l , k ) is 1 only if ITD in the TF bin is between τ 1 and τ 2 which correspond to the time differences of arrival (TDoAs) when the source is located at the boundaries of the target DoA range:
m 3 ( l , k ) = 1 , if τ 1 Δ ψ ( l , k ) K 2 π f s k τ 2 or k > K 2 or k < K 1 0 , otherwise ,
where ( K 1 , K 2 ) is the range of the frequency bin index. K 1 is set to exclude low frequency bins in which the DoA estimate is too sensitive to the small errors in phase measurement, whereas K 2 is set to avoid spatial aliasing with a certain margin. It essentially means that we will not consider ILD or ITD information from the TF bins for which the TDoA is out of ( τ 1 , τ 2 ) range. The combined spectral mask is given as m ( l , k ) = m 1 ( l , k ) m 2 ( l , k ) m 3 ( l , k ) , which passes through only the spectral information from TF bins satisfying all the three conditions to prevent the disturbance from the TF bins with interferences only. In addition, the frames with few valid bins are regarded as non-speech frames, i.e., the final voice activity is 0 if the number of valid frequency bins, K m ( l ) = k m ( l , k ) , is less than a certain threshold, K m i n . It is noted that η 1 , η 2 , τ 1 , and τ 2 are set to reject only TF bins with definitely no speech signals considering the following combination with other masks.
The proposed frequency selective approach using the spectral masks m ( l , k ) can be applied to any frequency domain VAD methods that summarize the information from each frequency. The frequency selective version of the NDPSD-based VAD becomes
V N D P S D F S ( l ) = 1 , if k = 1 K / 2 m ( l , k ) K m ( l ) Δ Φ ( l , k ) ξ N D P S D and K m ( l ) K m i n 0 , otherwise .
Similarly, the Equation (3) for the LTIPD-based VAD is modified to consider only reliable frequency bins as E F S ( l ) = max 1 i U k : C i ( l , k ) > κ i m ( l , k ) | Y 1 ( l , k ) | 2 , and then V L T I P D F S ( l ) can be constructed in a similar manner with the Equation (4). It is not straightforward to modify the two step PLDR method incorporating m ( l , k ) . The adaptation of short-term and long-term smoothing factors are originally governed by the speech presence probability, which is computed based on the ILD and the ILD for noise only period. As the smoothing factor control part can be viewed as a soft-decision version of the frequency selective approach although it relies solely on ILD information, we did not construct the frequency selective version of it. The frequency selective version of V A N D is simply obtained by “AND” operation of V N D P S D F S ( l ) and V L T I P D F S ( l ) . As for the VAD based on SVM, the input features corresponding to the TF bins with m ( l , k ) = 0 are set to 0 so that they do not contribute to the output of the SVM, y ( x ( l ) ) = w o T x ( l ) + b o . Then, V S V M F S ( l ) is constructed based on the a posteriori probability computed from the output of the SVM with masked input vectors.

4. Experimental Result

To demonstrate the performance of the proposed algorithm, we have recorded the desired speech, directional interferences, and diffuse noises with two microphones located on a commercial mobile phone, Samsung Galaxy S7. The placement of the user and the loudspeakers are illustrated in Figure 1. The size of the room was 3119 × 3232 × 2080 mm3 and the reverberation time of the room was approximately 120 ms. In the center of the room, one of the two male and two female speakers stood holding the mobile phone with the right hand in the handset mode. The desired near-end speech was spoken by those speakers in English. The distance between the microphones was about 140 mm. The diffuse noise field was generated by playing back white, babble, or car noises from NOISEX-92 database [35] with four loudspeakers facing the corners of the room to incur complex reflections as shown in Figure 1. Directional interferences were the speech utterances spoken by four male and four female speakers chosen from the TIMIT database, and were played from one of the four loudspeakers located 1000 mm away from the user at the directions {45°, 135°, 225°, 315°} facing the users as depicted in Figure 1. Two minutes of near-end speech spoken by one male and one female speakers was used for the training of the SVM and the parameter setting of other approaches, while another six minute long speech spoken by the other male and female speakers was used to generate the test data. The near-end speech was mixed with a directional interference coming from one of the four loudspeakers or one of the diffuse noises at the SNR level out of {−5, 0, …, 20} dB on the primary microphone, which makes the total length of the training and test data 84 and 252 min, respectively. The sampling rate was 8 kHz, and 256 point Hamming window was applied with 10 ms frame shift. The frequency index range ( K 1 , K 2 ) considered in m 3 ( l , k ) and LTIPD was (4, 31), which corresponds to (125 Hz, 968.75 Hz), considering that the microphone distance was 140 mm. The range of the target DoA for the ITD-based VAD statistics was set to be (10°, 70°) when 0° corresponds to the end-fire direction to the primary microphone side. The range of the target DoA for the reliable frequency selection that determines τ 1 and τ 2 in Equation (9) was (0°, 80°), which was a bit wider than that for VAD. These values were set to allow the deviation of the estimated DoA due to the noises and the individual variability in how to hold the phone. The thresholds for m 1 ( l , k ) and m 2 ( l , k ) , η 1 and η 2 , were set to minimize the equal error rate on training set. The parameters for the LTIPD-based VAD, L and U, were set to 12 and 10, respectively. K m i n was set to 3.
The performances of the various ILD- and ITD-based VAD methods with and without the proposed reliable frequency bin selection for the whole test data are shown in Figure 2 in the form of the receiver operating characteristics (ROC) curves. The original ROC curves are the collection of the (Hit rate, FAR) pairs with different threshold parameters, so one ROC curve can be drawn for each number of hangover frames for each VAD. The curves shown in Figure 2 are the collection of the leftmost points among all ROC curves with different hangover parameters given the hit rate for each VAD method. Adaptive multi-rate (AMR) VAD option 2 was also shown as a performance benchmark [7]. The performances of the original VAD methods using all frequency bins are shown as dashed lines, while the solid lines indicate the performances of the frequency selective versions of the corresponding methods. The VADs with the proposed frequency selective approach outperformed the original VAD methods. Among all the methods, V A N D F S , the frequency selective version of V A N D , performed the best when the false alarm rate (FAR) was less than 29.02%, while V S V M F S was the best when we need a hit rate higher than 98.06% in the expense of a higher FAR. Figure 3 shows the ROC curves for different SNRs averaged over all noise types. The tendency that the reliable frequency selection improves the performance was similar in all cases except some regions with very low hit rates. The performance difference between V A N D and V S V M was not as large as in [30], possibly due to the mismatch in the near-end speakers in the training and the test. V S V M F S could further enhance the performance of the V S V M with dynamic frequency selection, especially for the lower SNRs. Table 1 shows the performance of each VAD in terms of accuracy, precision, and recall at the operating point minimizing the overall error rate E O V R defined as E O V R = α × F R R + ( 1 α ) × F A R [30], in which FRR denotes the false rejection rate. α was set to 0.8 as the false rejection may be more critical for variable rate speech codecs or speech enhancement. Experimental results in a more reverberant environment are also available at the demo page (https://mspl.gist.ac.kr/vad_demo/dualchannelVAD.html).

5. Conclusions

In this letter, we have proposed a dual microphone VAD scheme based on the spatial information from reliable TF bins only. The frequency selection masks are constructed based on three criteria on signal energy, ILD, and ITD. Conventional ITD- and ILD-based VADs and the combination of them are modified to consider the spatial information from the reliable spectral bins only. Experimental results with a commercial smartphone showed that the frequency selective approach can enhance the performance of the ILD- and ITD-based VADs, the logical combination of them, and the SVM-based VAD in various noise environments.

Author Contributions

Conceptualization, S.H., Y.G.J., and J.W.S.; methodology, Y.G.J. and J.W.S.; software, S.H. and Y.G.J.; validation, J.W.S.; formal analysis, J.W.S.; investigation, Y.G.J.; resources, J.W.S.; data curation, S.H.; writing—original draft preparation, S.H.; writing—review and editing, J.W.S.; visualization, S.H.; supervision, J.W.S.; project administration, J.W.S.; funding acquisition, J.W.S.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (NRF-2016R1C1B1015291) and the Technology Innovation Program (10076583, Development of free-running speech recognition technologies for embedded robot system) funded by the Ministry of Trade, Industry and Energy, Korea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tavares, R.; Coelho, R. Speech Enhancement with Nonstationary Acoustic Noise Detection in Time Domain. IEEE Signal Process. Lett. 2016, 23, 6–10. [Google Scholar] [CrossRef]
  2. Reddy, C.K.A.; Shankar, N.; Bhat, G.S.; Charan, R.; Panahi, I. An Individualized Super-Gaussian Single Microphone Speech Enhancement for Hearing Aid Users With Smartphone as an Assistive Device. IEEE Signal Process. Lett. 2017, 24, 1601–1605. [Google Scholar] [CrossRef] [PubMed]
  3. Meyer, J.; Simmer, K.U.; Kammeyer, K.D. Comparison of one- and two-channel noise-estimation techniques. In Proceedings of the 5th International Workshop on Acoustic Echo Control Noise Reduction, London, UK, 3 March 1997; pp. 137–145. [Google Scholar]
  4. Reddy, M.K.; Rao, K.S. Robust Pitch Extraction Method for the HMM-Based Speech Synthesis System. IEEE Signal Process. Lett. 2017, 24, 1133–1137. [Google Scholar] [CrossRef]
  5. Drugman, T.; Huybrechts, G. Traditional Machine Learning for Pitch Detection. IEEE Signal Process. Lett. 2018, 25, 1745–1749. [Google Scholar] [CrossRef]
  6. TIA Document, PN-3292, Enhanced Variable Rate Codec, Speech Service Option 3 for Wide-Band Spectrum Digital Systems. Available online: https://www.3gpp2.org/Public_html/Specs/C.S0014-A_v1.0_040426.pdf (accessed on 11 July 2019).
  7. 3GPP TS 26.104. ANSI-C Code for the Floating-Point Adaptive Multi-Rate (AMR) Speech Codec; Rev. 12.0.0; 3GPP: Valbonne, France, 2014. [Google Scholar]
  8. Shin, J.W.; Chang, J.-H.; Kim, N.S. Voice activity detection based on a family of parametric distributions. Pattern Recognit. Lett. 2007, 28, 1295–1299. [Google Scholar] [CrossRef]
  9. Shin, J.W.; Kwon, H.J.; Jin, S.H.; Kim, N.S. Voice activity detection based on conditional MAP criterion. IEEE Signal Process. Lett. 2008, 15, 257–260. [Google Scholar] [CrossRef]
  10. Chang, J.-H.; Shin, J.W.; Kim, N.S. Voice activity detector employing generalized Gaussian distribution. Electron. Lett. 2004, 40, 1561–1562. [Google Scholar] [CrossRef]
  11. Shin, J.W.; Chang, J.-H.; Kim, N.S. Voice activity detection based on statistical models and machine learning approaches. Comput. Speech Lang. 2010, 24, 515–530. [Google Scholar] [CrossRef]
  12. Rabiner, L.R.; Sambur, M.R. Voiced-unvoiced-slience detection using Itakura LPC distance measure. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Hartford, CT, USA, 9–11 May 1977; pp. 323–326. [Google Scholar]
  13. Hoyt, J.D.; Wechsler, H. Detection of human speech in structured noise. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Adelaide, SA, Australia, 19–22 April 1994; pp. 237–240. [Google Scholar]
  14. Junqua, J.C.; Reaves, B.; Mark, B. A study of endpoint detection algorithms in adverse conditions: Incidence on a DTW and HMM recognize. In Proceedings of the EUROSPEECH ’91, Genova, Italy, 24–26 September 1991; pp. 1371–1374. [Google Scholar]
  15. Haigh, J.A.; Mason, J.S. Robust voice activity detection using cepstral feature. In Proceedings of the TENCON’93, Beijing, China, 19–21 October 1993; pp. 321–324. [Google Scholar]
  16. Drugman, T.; Stylianou, Y.; Kida, Y.; Akamine, M. Voice Activity Detection: Merging Source and Filter-based Information. IEEE Signal Process. Lett. 2016, 23, 252–256. [Google Scholar] [CrossRef]
  17. Ramírez, J.; Segura, J.C.; Benítez, C.; Torre, Á.; Rubio, A. Efficient voice activity detection algorithms using long-term speech information. Speech Commun. 2004, 42, 271–287. [Google Scholar] [CrossRef]
  18. Ramírez, J.; Segura, J.C.; Benítez, C.; Torre, Á.; Rubio, A. An effective subband OSF-based VAD with noise reduction for robust speech recognition. IEEE Trans. Speech Audio Process. 2005, 13, 1119–1129. [Google Scholar] [CrossRef]
  19. Babu, C.G.; Vanathi, P.T. Performance analysis of voice activity detection algorithms for robust speech recognition. TECHNIA Int. J. Comput. Sci. Commun. Technol. 2011, 2, 288–293. [Google Scholar]
  20. Graf, S.; Herbig, T.; Buck, M.; Schmidt, G. Features for voice activity detection: A comparative analysis. EURASIP J. Adv. Signal Process. 2015, 2015, 1–15. [Google Scholar] [CrossRef]
  21. Pencak, J.; Nelson, D. The NP speech activity detection algorithm. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Detroit, MI, USA, 9–12 May 1995; pp. 381–384. [Google Scholar]
  22. Zhang, X.-L.; Wu, J. Deep belief network based voice activity detection. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 697–710. [Google Scholar] [CrossRef]
  23. Zhang, X.-L.; Wang, D. Boosting Contextual Information for Deep Neural Network Based Voice Activity Detection. IEEE Trans. Audio Speech Lang. Process. 2016, 24, 252–264. [Google Scholar] [CrossRef]
  24. Zazo, R.; Sainath, T.N.; Simko, G.; Parada, C. Feature Learning with Raw-Waveform CLDNNs for Voice Activity Detection. Proc. Interspeech 2016, 3668–3672. [Google Scholar] [CrossRef]
  25. Kim, J.; Hahn, M. Voice Activity Detection Using an Adaptive Context Attention Model. IEEE Signal Process. Lett. 2018, 25, 1181–1185. [Google Scholar] [CrossRef]
  26. Kaushik, L.; Sangwan, A.; Hansen, J.H.L. Speech Activity Detection in Naturalistic Audio Environments: Fearless Steps Apollo Corpus. IEEE Signal Process. Lett. 2018, 25, 1290–1294. [Google Scholar] [CrossRef]
  27. Guo, Y.; Li, K.; Fu, Q.; Yan, Y. A two microphone based voice activity detection for distant talking speech in wide range of direction of arrival. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, 25–30 March 2012; pp. 4901–4904. [Google Scholar]
  28. Jeub, M.; Herglotz, C.; Nelke, C.; Beaugeant, C.; Vary, P. Noise reduction for dual-microphone mobile phones exploiting power level differences. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, 25–30 March 2012; pp. 1693–1696. [Google Scholar]
  29. Choi, J.H.; Chang, J.-H. Dual-microphone voice activity detection technique based on two-step power level difference ratio. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1069–1081. [Google Scholar] [CrossRef]
  30. Park, J.; Jin, Y.G.; Hwang, S.; Shin, J.W. Dual Microphone Voice Activity Detection Exploiting Interchannel Time and Level Difference. IEEE Signal Process. Lett. 2016, 23, 1335–1339. [Google Scholar] [CrossRef]
  31. Potamitis, I. Estimation of Speech Presence Probability in the Field of Microphone Array. IEEE Signal Process. Lett. 2004, 11, 956–959. [Google Scholar] [CrossRef]
  32. Lee, B.; Kalker, T. Multichannel voice activity detection with spherically invariant sparse distributions. In Proceedings of the 2009 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, New Paltz, NY, USA, 18–21 October 2009; pp. 225–228. [Google Scholar]
  33. Knapp, C.H.; Carter, G.C. The generalized correlation method for estimation of time delay. IEEE Trans. Acoust. Speech Signal Process. 1976, 24, 320–327. [Google Scholar] [CrossRef] [Green Version]
  34. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
  35. Varga, A.; Steeneken, H. Assessment for automatic speech recognition: II. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition system. Speech Commun. 1993, 12, 247–251. [Google Scholar] [CrossRef]
Figure 1. Geographical placement of the noise sources and the receiver.
Figure 1. Geographical placement of the noise sources and the receiver.
Sensors 19 03056 g001
Figure 2. Receiver operating characteristics curves for various voice activity detection (VAD) with and without the proposed frequency selective approach.
Figure 2. Receiver operating characteristics curves for various voice activity detection (VAD) with and without the proposed frequency selective approach.
Sensors 19 03056 g002
Figure 3. Receiver operating characteristics curves for various VADs with and without the proposed frequency selective approach for each signal-to-noise ratio (SNR).
Figure 3. Receiver operating characteristics curves for various VADs with and without the proposed frequency selective approach for each signal-to-noise ratio (SNR).
Sensors 19 03056 g003
Table 1. Performance of accuracy, precision, and recall at the operating point minimizing E O V R .
Table 1. Performance of accuracy, precision, and recall at the operating point minimizing E O V R .
METHODAMRPLDRNDPSDNDPSDFSLTIPDLTIPDFSSVMSVMFSANDANDFS
Accuracy83.2894.3587.7990.1993.3896.1794.4094.4696.4297.13
Precision76.9192.0984.1487.4892.3795.8392.1692.1196.9397.00
Recall99.6498.2095.9795.9795.9497.3198.2298.4096.5897.82

Share and Cite

MDPI and ACS Style

Hwang, S.; Jin, Y.G.; Shin, J.W. Dual Microphone Voice Activity Detection Based on Reliable Spatial Cues. Sensors 2019, 19, 3056. https://doi.org/10.3390/s19143056

AMA Style

Hwang S, Jin YG, Shin JW. Dual Microphone Voice Activity Detection Based on Reliable Spatial Cues. Sensors. 2019; 19(14):3056. https://doi.org/10.3390/s19143056

Chicago/Turabian Style

Hwang, Soojoong, Yu Gwang Jin, and Jong Won Shin. 2019. "Dual Microphone Voice Activity Detection Based on Reliable Spatial Cues" Sensors 19, no. 14: 3056. https://doi.org/10.3390/s19143056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop