Next Article in Journal
Square-Root Unscented Information Filter and Its Application in SINS/DVL Integrated Navigation
Previous Article in Journal
Design and Evaluation of FBG-Based Tension Sensor in Laparoscope Surgical Robots
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Threshold-Based Noise Detection and Reduction for Automatic Speech Recognition System in Human-Robot Interactions

Department of Electrical Engineering, National Cheng Kung University, No. 1, Dasyue Road, Tainan City 701, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(7), 2068; https://doi.org/10.3390/s18072068
Submission received: 29 April 2018 / Revised: 8 June 2018 / Accepted: 22 June 2018 / Published: 28 June 2018
(This article belongs to the Section Physical Sensors)

Abstract

:
This work develops a speech recognition system that uses two procedures of proposed noise detection and combined noise reduction. The system can be used in applications that require interactive robots to recognize the contents of speech that includes ambient noise. The system comprises two stages, which are the threshold-based noise detection and the noise reduction procedure. In the first stage, the proposed system automatically determines when to enhance the quality of speech based on the signal-to-noise ratio (SNR) values of the collected speech at all times. In the second stage, independent component analysis (ICA) and subspace speech enhancement (SSE) are employed for noise reduction. Experimental results reveal that the SNR values of the enhanced speech exceed those of the received noisy speech by approximately 20 dB to 25 dB. The noise reduction procedure improves the speech recognition rates by around 15% to 25%. The experimental results indicate that the proposed system can reduce the effect of noise in numerous noisy environments and improve the quality of speech for recognition purposes.

1. Introduction

Automatic speech recognition (ASR) provides a user-friendly means of efficiently convey commands or requests to devices of human-machine interface (HMI). These devices can automatically analyze the received data and behave toward humans in ways that are consistent with the recognition results. A substantial literature exists on data classification and incomplete data analysis to reduce imputation error [1,2]. In recent years, research into ASR has considered many scenarios and applications. Much of the literature involves ASR for intelligent human-robot interaction [3,4,5,6,7]. When an ASR system is used in a real environment, especially a noisy one, the environmental noise considerably influences the quality of speech. The ambient noise can affect the signal components of speech and worsen the representation of recognition result. To solve the problem of noise, many methods of mitigating the effect of noise on ASR development, have been developed [8,9,10,11,12,13,14,15,16,17,18].
Choi et al. [8] presented a speech enhancement and recognition method for service robots. Their proposed adaptive beamformer structure includes a circular microphone array that comprised eight microphones. Jung et al. [9] presented a speech acquisition and recognition system, which can be utilized in home-agent robots. They used a generalized sidelobe canceller-based (GSC-based) algorithm with a microphone array to compensate the effect of room reverberation. Betkowska et al. [10] studied a factorial hidden Markov model (FHMM), which can be combined with HMMs of clean speech and noise, to increase speech recognition accuracy in noisy environments. Gomez et al. [11] proposed a spectral subtraction-based method to eliminate room reverberation for human-machine interaction. Ohashi et al. [12] demonstrated the use of a spatial subtraction array for noise reduction and employed the noise superimposition to realize a hands-free device of speech recognition. Hong et al. [13] executed a multi-channel GSC-based speech enhancement algorithm and designed a masking-based Wiener filter to reduce the residual noise. A noisy environment-aware speech recognition system has been used in human-robot interaction [14]. The system can determine whether the speech should be enhanced from the initial SNR value of the speech in noisy environments. Mohammadiha et al. [15] presented a speech enhancement method with a Bayesian formulation of nonnegative matrix factorization (BNMF). They adopted a scheme to train the noise BNMF model; the training BNMF model can be used as an unsupervised speech enhancement system. A review of the aforementioned methods [8,9,10,11,12,13,14,15] indicates that they mainly improve speech recognition by reducing the effects of ambient noise or room reverberation. However, noise signals can have numerous properties in real-world situations. When noise signals are unknown, the noise reduction method, which is based on known noise information [10,12,15], cannot be used. In the previous work, [14], the system estimated only the initial SNR value in the noisy environment. The method may be affected when the noise is varied.
To make the ASR system more robust in noisy environments, the methodology of artificial neural networks (ANN), especially deep neural networks (DNN), has been widely utilized in speech enhancement for ASR in recent years [16,17,18]. The goal of DNN is to implement complex nonlinear numeric functions, which are used to directly map log-likelihood spectral features of noisy speech into corresponding clean speech. In DNN model training, several research studies develop the strategy of multi-style training on mixed speech and noise data. Although DNN-based methods can achieve high accuracy improvement in ASR, the DNN models require more training data to synthesize, the amount of training data even more than HMM-based systems.
Integrating the above-mentioned ASR methods, two aspects are considered in this work; the first one is the method of noise reduction, and the second one is the manner of training data. In noise reduction method, this work attempts to develop a blind source separation-based (BSS-based) method to remove the ambient noise. Since ambient noise is unknown and varied in numerous environments, the noise reduction method which does not require noise information is adequate to separate the noise from the noisy speech. To improve the quality of speech for recognition, another speech enhancement method is combined with BSS-based method. Closely investigating different noise situations, noise signals may not be obvious within several time intervals when noise signals are intermittent. In this case, noise reduction cannot be used because the over-filtering speech may cause speech distortion and reduce the speech recognition rate. To prevent the circumstance of over-filtering speech, a preprocessing scheme called threshold-based noise detection is proposed in this work. The proposed scheme can automatically determine that when noise should be eliminated according to the magnitude of the noise. With respect to training data, an HMM-based training system is used in this work due to the amount of training data and training time. This work implements the HMM-based training system by using hidden Markov model toolkit (HTK) [19]. The proposed system utilizes the HTK as a speech recognizer in speech recognition. Compared with the DNN-based system, the HMM-based system using HTK can usefully reduce the amount of training data because the recognizer is trained with only the clean speech.
According to the two aspects, this work presents an HMM-based speech recognition system for human-robot interactions in noisy environments. The system can be divided into two procedures, the first one is proposed threshold-based noise detection, and the second one is combined noise reduction. The system has following four properties: Training data requires only clean speech data, a proposed preprocessing scheme to prevent the over-filtering speech, noise reduction without predicted noise information, and valid effect on reducing ambient noise and improving speech quality. This work provides another feasible method for ASR system in noisy environments.
This paper is organized as follows. Section 2 provides an overview of the proposed recognition system. Section 3 describes in detail the proposed threshold-based noise detection and combined noise reduction procedure. Section 4 considers the experimental results. Section 5 briefly draws conclusions.

2. System Overview

Figure 1 shows an overview of the proposed system. First to obtain information about the initial noise power, a linear array of two microphones is used to receive the initial noise signal in a noisy environment. When the noise signal is recorded, its power can be estimated by using a noise power calculation. From a noisy speech recording, the noisy speech signal can be identified as a speech signal or a non-speech signal (noise signal), based on the results of voice activity detection (VAD). If the collected signal is identified as a noise signal, then the noise power data can be updated from the power estimation of the current noise signal. Subsequently, the signal is used to determine the SNR value. The noisy speech signal can be input to the noise reduction procedure or speech recognition procedure as determined by comparative SNR determination.
Following the noise reduction procedure, an ICA-based method [20] is adopted to separate the noise signal from the noisy speech signal. However, the signal that is separated in the ICA processing retains the residual noise signal. To reduce the effect of the residual noise on the noisy speech signal and to reduce the speech distortion, a method of subspace-based speech enhancement [21] is applied after the ICA processing.
In speech recognition, an HTK-based speech recognizer, which is trained with clean speech data, is used in speech recognition. The recognizer analyzes and takes the approximate content of speech, which is the recognition result of the recognition system.

3. Proposed Methods

3.1. Proposed Threshold-Based Noise Detection

Figure 2 displays the procedure of proposed threshold-based noise detection. A linear array is employed to collect the speech signal in a noisy environment. In the time domain, the observed signals x1(t) and x2(t) can be modeled as matrices and vectors in (1) and (2), where y(t) and n(t) denote the clean speech signal and the noise signal, respectively. Since the observed signal x1(t) is similar to x2(t), signal x1(t) is taken as the principal signal in the subsequent VAD, noise power calculation, and SNR determination.
X = Y + N ,
[ x 1 ( t ) x 2 ( t ) ] = [ y 1 ( t ) y 2 ( t ) ] + [ n 1 ( t ) n 2 ( t ) ]
The objective of VAD is to locate the speech signal component of the received signal. Two features, which are called short-time energy and zero-crossing rate (ZCR), are executed in VAD. The short-time energy is formulated as (3), where w(n) is the selected window function, and L is the length of the window. In the proposed system, the default window function is a Hamming window, which is defined in (4). The signal that has high amplitude can be found and treated as a speech signal. To detect the speech signal accurately, another feature, ZCR, is used in VAD.
E x = n = 0 L 1 [ x 1 ( n ) w ( n ) ] 2 ,
w ( n ) = 0.54 0.46 cos ( 2 π n L 1 ) , 0 n L 1
Equation (5) represents ZCR; z(t) equals one if the amplitude of the observed signal x1(t) is positive and zero otherwise. ZCR can be used to discover the voiced signal, which has a lower ZCR than an unvoiced signal or noise. In VAD, the non-speech signal, which has a lower short-time energy and higher ZCR, can be regarded as a noise signal and be used in the noise power calculation.
Z C R x = 1 L 1 t = 1 L 1 | z ( t ) z ( t 1 ) |
The purpose of the noise power calculation is to estimate and update the mean power of the noise signal, which is detected in VAD. Equation (6) presents the relationship between the average noise power Pn and the noise signal n(t). The mean noise power can be used in the following determination of the SNR threshold value.
P n = 1 L t = 1 L 1 | n ( t ) | 2
In the determination of the SNR threshold value, the SNR value of the collected speech can be estimated. Equation (7) represents SNR, where Py and Px are the mean power of clean speech and noisy speech. The proposed system sets a threshold value ε, which is compared with the SNR value. When the SNR value is less than or equal to ε, the received speech should be enhanced because the power of noise signal is obvious. If the SNR exceeds ε, it means the effect of the noise on the collected speech is not obvious, and the collected speech can be directly passed to the speech recognition process.
ε S N R = 10 log 10 P y P n = 10 log 10 P x P n P n

3.2. Combined Noise Reduction Procedure

The combined noise reduction procedure comprises ICA and SSE. The observed signals in (1) and (2) can be expressed using an unknown mixing matrix A, which is in (8). The speech signal y(t) and the noise signal n(t) are regarded as the original source signals.
X = [ x 1 ( t ) x 2 ( t ) ] = [ a 11 a 12 a 21 a 22 ] [ y ( t ) n ( t ) ]
Consistent with (8), to obtain the individual source signal from the received signals x1(t) and x2(t), a de-mixing matrix is estimated. Equation (9) represents the de-mixing matrix, where s1(t) and s2(t) are separated signals, and matrix W is the de-mixing matrix. The separated signals are similar to the original source signals.
S = [ s 1 ( t ) s 2 ( t ) ] = [ w 11 w 12 w 21 w 22 ] [ x 1 ( t ) x 2 ( t ) ] = W X
To calculate the de-mixing matrix, ICA exploits high-order statistics and information theory to measure the non-Gaussian characteristic of the property. The analysis of the non-Gaussian characteristic can be used to obtain the de-mixing matrix. In the ICA process, both source signals must be mutually independent. To solve the situation of mutually independent, two methods called signal centering and signal whitening are utilized in ICA. These methods ensure that the source signals can become uncorrelated. Signal centering is performed using (10), where X is the received signal, and E[X] is the mean of the received signal.
X ¯ = X E [ X ]
The purpose of signal whitening is to evaluate a “whitening matrix”. The signal data that is described using (10) can become uncorrelated when multiplied by the whitening matrix. Equations (11) and (12) represent the whitening process in which H is a whitening matrix; E [ X ^ X ^ T ] is the covariance matrix of the signal, and I is the identity matrix.
X ^ = H X ¯ ,
E [ X ^ X ^ T ] = I
ICA adopts negentropy maximization to analyze the non-Gaussianity property of the signal. Equation (13) describes the formula for negentropy calculation, where the Gaussian distribution of signal Y ^ Gauss has the identical covariance matrix as the estimated signal Y ^ , and the entropy is H ( ) .
J ( Y ^ ) = H ( Y ^ Gauss ) H ( Y ^ )
To accelerate the ICA, the proposed system uses an algorithm that is called Fast ICA [22]. The calculation of negentropy can be approximately written as (14), where α and β are constants; γ ^ is a zero-mean Gaussian variable with a standard deviation of unity, and G 1 ( ) and G 2 ( ) are contrast functions. Several functions, given by (15)–(17), can be taken as the contrast function in the negentropy calculation. The coefficient that is given by (15) is a constant with a value of between one and two.
J ( Y ^ ) α ( E [ G 1 ( Y ^ ) ] ) 2 + β ( E [ G 2 ( Y ^ ) ] E [ G 2 ( γ ^ ) ] ) 2 ,
G 1 ( x ) = 1 a log ( cosh ( a x ) ) ,
G 2 ( x ) = e x 2 2 ,
G 3 ( x ) = x 4
According to (14), when a single contrast function is used, the calculated negentropy can be proportional to a perfect square form, which is composed of the contrast function of the estimated signal Y ^ and the zero-mean Gaussian variable γ ^ . Equation (14) can be rewritten as (18) and (19).
J ( Y ^ ) ( E [ G ( Y ^ ) ] E [ G ( γ ^ ) ] ) 2 ,
J ( Y ^ ) ( E [ G ( W T X ^ ) ] E [ G ( γ ^ ) ] ) 2
Equation (19) indicates that the maximum negentropy can be obtained by determining the maximum E [ G ( W T X ^ ) ] . The de-mixing matrix can be derived from the Newton iteration. The de-mixing matrix is represented as (20), where g ( ) and g ( ) are the derivatives of contrast functions G ( ) and g ( ) .
W E [ X ^ g ( W T X ^ ) ] E [ g ( W T X ^ ) ] W
After the ICA processing, the two separated signals can be judged as the speech signal and the noise signal according to ZCR. Although ICA can separate noise from received signals, the separated signals retain the residual noise component. To remove the residual noise, the proposed system incorporates another technique, SSE, to design a filter and eliminate the effect of residual noise on the separated signals. The filter coefficients can be evaluated by subtracting the original speech signal from the filtered speech signal. Equation (21) describes this signal subtraction, where F denotes the SSE filter, and I is the identity matrix. Finally, the signal subtraction can be rewritten in terms of two parameters δy and δn. The former parameter specifies the speech distortion from the filter, and the latter specifies the residual noise after the filter process.
δ = F X Y = F ( Y + N ) Y = ( F I ) Y + F N = δ y + δ n
To optimize the filter process based on (21), the variances of speech distortion and residual noise, which are given by (22) and (23), are used to evaluate the filter coefficients.
δ ¯ y = E [ δ y T δ y ] ,
δ ¯ n = E [ δ n T δ n ]
In the filter coefficients evaluation, the recognition rate can be reduced when the speech distortion is obvious. Therefore, the speech distortion should be minimized as much as possible. The residual noise also can influence the recognition result. To prevent this situation, the residual noise should be suppressed. With respect to the speech distortion and the residual noise, the two aforementioned requests can be defined as follows:
arg min F   δ ¯ y ,
subject to
δ ¯ n γ σ n ,
where σ n is the variance of the noise signal, and γ is the adjustable parameter, whose value is between zero and one.
Consistent with (24) and (25), the optimal filter is obtained using the Lagrange multiplier method. The optimal filter is represented as (26), where RYY and RNN are the covariance matrices of the speech signal and the noise signal, respectively, and μ is a Lagrange multiplier.
F = R Y Y ( R Y Y + μ R N N ) 1
The covariance matrix RYY described in (26) can be given by (27) using eigenvalue decomposition (EVD), where Q is a square matrix whose ith column is the eigenvector qi, and ΛYY is a diagonal matrix.
R Y Y = Q Λ Y Y Q 1 ,
substituting (27) into (26) yields,
F = Q Λ Y Y ( Λ Y Y + μ Q 1 R N N Q ) 1 Q 1 ,
expressing RNN in EVD form enables (28) to be rewritten as (29).
F = Q Λ Y Y ( Λ Y Y + μ Λ N N ) 1 Q 1

3.3. Speech Recognition Process

The proposed system utilizes the HTK as a speech recognizer in speech recognition. About the selection of speech corpus, the system takes corpora of Mandarin speech data across Taiwan (MAT-400) to train acoustic models, numerous acoustic models have been trained in the HTK recognizer. For feature extraction of speeches, the HTK uses Mel-frequency cepstral coefficients (MFCCs) as the speech features in speech recognition. In recognition process, the HTK-based speech recognizer analyzes the speech features and selects the most appropriate content of speech as the recognition result.

4. Experimental Results

4.1. Experimental Setup

To realize the proposed system, a humanoid robot, called 16-DOF RobotinnoTM, is used herein. Figure 3 shows the illustration of the robot. With respect to the linear array, two omni-directional microphones are placed with a spacing of 0.1 m on the shoulder of the humanoid robot.
Figure 4 shows the layout of the test environment; the length and the width of the experimental chamber are 7.2 m and 6.1 m, respectively. The linear array collects the test speech signal with a sampling rate of 8 kHz. The distance from the robot to the speaker is 1.5 m, and the distance from the robot to the source of the noise is 2 m. The SNR threshold value ε is set to 10. In the experiments, three test directions (30°, 60°, and 90°) are utilized to collect the speech signal, and three directions (45°, 90°, and 135°) are used to record the noise signal.
The proposed system utilizes the noisex-92 database [23] to provide the test background noises. The noises in the database are of five types, which are babble noise, car noise, factory noise, pink noise, and white noise. During speech recording, each of the four speakers (three males and one female) utters 30 sentences, each taking approximately three to five seconds. For each type of noise, a total of 1080 sentences are examined with the speakers and the source of noise at varied directions. The linear array firstly records the environmental noise to obtain the initial power value of the noise signal in the test environment. The power value of the noise signal is estimated and updated during the subsequent speech recording.

4.2. Evaluation Results

In test speech recording, the system records noisy speech with SNR values of 0 dB, 5 dB, and 10 dB. To compare the quality of enhanced speech with noisy speech, two objective speech quality measures, SNR and segmental SNR, are estimated from the experimental results. Equations (30) and (31) represent the SNR and segmental SNR, where y(t), y′(t), N, M, and m are, respectively, the noisy speech, the enhanced speech, the length of the speech signal, the number of frames, and the frame index.
S N R = 10 log 10 t = 0 N 1 y 2 ( t ) t = 0 N 1 [ y ( t ) y ( t ) ] 2 ,
S e g S N R = 1 M m = 1 M S N R m
Table 1, Table 2 and Table 3 compare the average SNR and segmental SNR values of the noisy speech and enhanced speech using the proposed method. Speech with three SNR values (0 dB, 5 dB, and 10 dB) and five types of noise are used in the experiments. The average SNR values of the enhanced speech exceed the noisy speech by approximately 20 dB to 25 dB. The segmental SNR values of the enhanced speech are also superior to the noisy speech. Both experimental results reveal that the proposed system improves the quality of speech in varied noisy environments.
Figure 5, Figure 6 and Figure 7 present the speech recognition rates of the noisy speech, the related works [10,13], and the proposed method. Two related methods based on HMM system are compared with the proposed HMM-based system. Three SNR values of noisy speech with 0 dB, 5 dB, and 10 dB, are examined in experiments. The results indicate that the proposed method can increase the recognition rates than noisy speech by about 15% to 25%. Compared with the related works, the proposed method is superior to the related works; the recognition rates can be better than the related works by 0.94% to 5.52%. The experimental results demonstrate that the proposed system using combined noise separation and speech enhancement methods can effectively remove numerous types of noise and improve the speech quality for speech recognition process.

5. Conclusions

This work develops a speech recognition system that can be embedded in a device of interactive robot to recognize the content of speech in noisy environments. The system can be divided into two procedures; the first one is the proposed preprocessing called threshold-based noise detection, and the second one is combined noise reduction. The proposed preprocessing scheme can evaluate the magnitude of noise, to prevent the situation of over-filtering speech when the background noise is slight. In noise reduction, two methods called ICA and SSE are combined to eliminate the effect of noises on the speech signal. ICA is used to separate the noise from the noise-contaminated speech, and the SSE method improves the quality of speech by filtering out the residual noise.
Experimental results indicate that the proposed system can remove the ambient noise and increase the speech recognition rate. The proposed method yields higher SNR value and speech recognition rate than noisy speech. The speech recognition rate is also superior to the related works in experiments. In future work, the system can be combined with several research fields such as acoustic processing, technique of sound source localization, design of home-care service robot, and multimedia analysis [24,25,26], to provide more user-friendly services in application of human-robot interactions.

Author Contributions

Data curation, S.-C.L. and M.-H.C.; Formal analysis, S.-C.L., J.-F.W. and M.-H.C.; Methodology, S.-C.L. and J.-F.W.; Writing—original draft, S.-C.L.; Writing—review and editing, J.-F.W.

Funding

This work has been supported by own funding of cognitive multimedia integrated circuit system design (CMICSD) laboratory from the Department of Electrical Engineering, National Cheng Kung University, Taiwan.

Acknowledgments

This work was supported by own funding of CMICSD laboratory.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, B.-W.; Ji, W.; Rho, S.; Gu, Y. Supervised collaborative filtering based on ridge alternating least squares and iterative projection pursuit. IEEE Access 2017, 5, 6600–6607. [Google Scholar] [CrossRef]
  2. Chen, B.-W.; Rho, S.; Yang, L.T.; Gu, Y. Privacy-preserved big data analysis based on asymmetric imputation kernels and multiside similarities. Future Gener. Comput. Syst. 2018, 78, 859–866. [Google Scholar] [CrossRef] [Green Version]
  3. Huang, G.-S.; Lee, V.C.S.; Lin, H.-C.; Ju, M.-J. The ASR technique for meal service robot. In Proceedings of the 37th IEEE Annual Conference on Industrial Electronics Society, Melbourne, Australia, 7–10 November 2011; pp. 3317–3322. [Google Scholar]
  4. Takahashi, T.; Nakadai, K.; Komatani, K.; Ogata, T.; Okuno, H.G. An improvement in automatic speech recognition using soft missing feature masks for robot audition. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 964–969. [Google Scholar]
  5. Yoshida, T.; Nakadai, K.; Okuno, H.G. Two-layered audio-visual speech recognition for robots in noisy environments. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 988–993. [Google Scholar]
  6. Mohamad, S.N.A.; Jamaludin, A.A.; Isa, K. Speech semantic recognition system for an assistive robotic application. In Proceedings of the IEEE International Conference on Automatic Control and Intelligent Systems, Shah Alam, Malaysia, 22 October 2016; pp. 90–95. [Google Scholar]
  7. Marin, R.; Vila, P.; Sanz, P.J.; Marzal, A. Automatic speech recognition to teleoperate a robot via Web. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; pp. 1278–1283. [Google Scholar]
  8. Choi, C.; Kong, D.; Kim, J.; Bang, S. Speech enhancement and recognition using circular microphone array for service robots. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 27–31 October 2003; pp. 3516–3521. [Google Scholar]
  9. Jung, Y.-W.; Lee, J.; Kong, D.; Kim, J.; Lee, C. High-quality speech acquisition and recognition system for home-agent robot. In Proceedings of the IEEE International Conference on Consumer Electronics, Los Angeles, CA, USA, 17–19 June 2003; pp. 354–355. [Google Scholar]
  10. Betkowska, A.; Shinoda, K.; Furui, S. Speech recognition using FHMMs robust against nonstationary noise. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Honolulu, HI, USA, 15–20 April 2007; pp. IV-1029–IV-1032. [Google Scholar]
  11. Gomez, R.; Kawahara, T.; Nakadai, K. Robust hands-free automatic speech recognition for human-machine interaction. In Proceedings of the 10th IEEE International Conference on Humanoid Robots, Nashville, TN, USA, 6–8 December 2010; pp. 138–143. [Google Scholar]
  12. Ohashi, Y.; Nishikawa, T.; Saruwatari, H.; Lee, A.; Shikano, K. Noise-robust hands-free speech recognition based on spatial subtraction array and known noise superimposition. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Edmonton, AB, Canada, 2–6 August 2005; pp. 2328–2332. [Google Scholar]
  13. Hong, J.; Cho, K.; Hahn, M.; Kim, S.; Jeong, S. Multi-channel noise reduction with beamforming and masking-based wiener filtering for human-robot interface. In Proceedings of the 5th International Conference on Automation, Robotics and Applications, Wellington, New Zealand, 6–8 December 2011; pp. 260–264. [Google Scholar]
  14. Lee, S.-C.; Chen, B.-W.; Wang, J.-F. Noisy environment-aware speech enhancement for speech recognition in human-robot interaction. In Proceedings of the IEEE International Conference on System, Man, and Cybernetics, Istanbul, Turkey, 10–13 October 2010; pp. 3938–3941. [Google Scholar]
  15. Mohammadiha, N.; Smaragdis, P.; Leijon, A. Supervised and unsupervised speech enhancement using nonnegative matrix factorization. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 2140–2151. [Google Scholar] [CrossRef]
  16. Novoa, J.; Wuth, J.; Escudero, J.P.; Fredes, J.; Mahu, R.; Yoma, N.B. DNN-HMM based automatic speech recognition for HRI scenarios. In Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, Chicago, IL, USA, 5–8 March 2018; pp. 150–159. [Google Scholar]
  17. Vu, T.T.; Bigot, B.; Chng, E.S. Combining non-negative matrix factorization and deep neural networks for speech enhancement and automatic speech recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Shanghai, China, 20–25 March 2016; pp. 499–503. [Google Scholar]
  18. Weng, C.; Yu, D.; Seltzer, M.L.; Droppo, J. Single-channel mixed speech recognition using deep neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Florence, Italy, 4–9 May 2014; pp. 5632–5636. [Google Scholar]
  19. Young, S.; Evermann, G.; Gales, M.; Hain, T.; Kershaw, D.; Liu, X.; Moore, G.; Odell, J.; Ollason, D.; Povey, D.; et al. The HTK Book Version 3.4; Department of Engineering, University of Cambridge: Cambridge, UK, 2006; pp. 23–47. [Google Scholar]
  20. Pham, D.T. Blind separation of instantaneous mixture of sources via an independent component analysis. IEEE Trans. Signal Process. 1996, 44, 2768–2779. [Google Scholar] [CrossRef]
  21. Ephraim, Y.; Van Trees, H.L. A signal subspace approach for speech enhancement. IEEE Trans. Speech Audio Process. 1995, 3, 251–266. [Google Scholar] [CrossRef]
  22. Hyvarinen, A. Fast and robust fixed-point algorithms for independent component analysis. IEEE Trans. Neural Netw. 1999, 10, 626–634. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Varga, A.; Steeneken, H.J.M. Assessment for automatic speech recognition: II. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition systems. Speech Commun. 1993, 12, 247–251. [Google Scholar] [CrossRef]
  24. Chen, B.-W.; Ji, W.; Rho, S. Geo-conquesting based on graph analysis for crowdsourced metatrails from mobile sensing. IEEE Commun. Mag. 2017, 55, 92–97. [Google Scholar] [CrossRef]
  25. Togami, M.; Amano, A.; Sumiyoshi, T.; Obuchi, Y. DOA estimation method based on sparseness of speech sources for human symbiotic robots. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 3693–3696. [Google Scholar]
  26. Chen, B.-W.; Rho, S.; Imran, M.; Guizani, M.; Fan, W.-K. Cognitive sensors based on ridge phase-smoothing localization and multiregional histograms of oriented gradients. IEEE Trans. Emerg. Top. Comput. 2016. [Google Scholar] [CrossRef]
Figure 1. Overview of proposed system.
Figure 1. Overview of proposed system.
Sensors 18 02068 g001
Figure 2. Procedure of proposed threshold-based noise detection.
Figure 2. Procedure of proposed threshold-based noise detection.
Sensors 18 02068 g002
Figure 3. Humanoid robot 16-DOF RobotinnoTM.
Figure 3. Humanoid robot 16-DOF RobotinnoTM.
Sensors 18 02068 g003
Figure 4. Layout of chamber to the developed system.
Figure 4. Layout of chamber to the developed system.
Sensors 18 02068 g004
Figure 5. Speech recognition rates of noisy speech (0 dB), related works, and proposed method.
Figure 5. Speech recognition rates of noisy speech (0 dB), related works, and proposed method.
Sensors 18 02068 g005
Figure 6. Speech recognition rates of noisy speech (5 dB), related works, and proposed method.
Figure 6. Speech recognition rates of noisy speech (5 dB), related works, and proposed method.
Sensors 18 02068 g006
Figure 7. Speech recognition rates of noisy speech (10 dB), related works, and proposed method.
Figure 7. Speech recognition rates of noisy speech (10 dB), related works, and proposed method.
Sensors 18 02068 g007
Table 1. Average values of SNR and segmental SNR of noisy speech (0 dB) and enhanced speech.
Table 1. Average values of SNR and segmental SNR of noisy speech (0 dB) and enhanced speech.
Noise TypeNoisy SpeechAverage SNR of Enhanced SpeechSegmental SNR of Enhanced Speech
Babble0 dB24.12 dB32.67 dB
Car0 dB33.31 dB37.11 dB
Factory0 dB26.91 dB35.15 dB
Pink0 dB27.59 dB35.61 dB
White0 dB29.36 dB35.22 dB
Table 2. Average values of SNR and segmental SNR of noisy speech (5 dB) and enhanced speech.
Table 2. Average values of SNR and segmental SNR of noisy speech (5 dB) and enhanced speech.
Noise TypeNoisy SpeechAverage SNR of Enhanced SpeechSegmental SNR of Enhanced Speech
Babble5 dB25.41 dB32.95 dB
Car5 dB37.38 dB37.21 dB
Factory5 dB27.37 dB35.33 dB
Pink5 dB30.83 dB35.88 dB
White5 dB31.38 dB35.42 dB
Table 3. Average values of SNR and segmental SNR of noisy speech (10 dB) and enhanced speech.
Table 3. Average values of SNR and segmental SNR of noisy speech (10 dB) and enhanced speech.
Noise TypeNoisy SpeechAverage SNR of Enhanced SpeechSegmental SNR of Enhanced Speech
Babble10 dB27.67 dB35.36 dB
Car10 dB42.62 dB39.04 dB
Factory10 dB30.78 dB37.85 dB
Pink10 dB36.41 dB38.88 dB
White10 dB35.29 dB38.14 dB

Share and Cite

MDPI and ACS Style

Lee, S.-C.; Wang, J.-F.; Chen, M.-H. Threshold-Based Noise Detection and Reduction for Automatic Speech Recognition System in Human-Robot Interactions. Sensors 2018, 18, 2068. https://doi.org/10.3390/s18072068

AMA Style

Lee S-C, Wang J-F, Chen M-H. Threshold-Based Noise Detection and Reduction for Automatic Speech Recognition System in Human-Robot Interactions. Sensors. 2018; 18(7):2068. https://doi.org/10.3390/s18072068

Chicago/Turabian Style

Lee, Sheng-Chieh, Jhing-Fa Wang, and Miao-Hia Chen. 2018. "Threshold-Based Noise Detection and Reduction for Automatic Speech Recognition System in Human-Robot Interactions" Sensors 18, no. 7: 2068. https://doi.org/10.3390/s18072068

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop