Next Article in Journal
Analysis of Graphomotor Tests with Machine Learning Algorithms for an Early and Universal Pre-Diagnosis of Dysgraphia
Previous Article in Journal
Depth Data Denoising in Optical Laser Based Sensors for Metal Sheet Flatness Measurement: A Deep Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Speech Enhancement Algorithm for Voice Assistance Application

by
Jenifa Gnanamanickam
1,*,
Yuvaraj Natarajan
2 and
Sri Preethaa K. R.
1
1
Department of Artificial Intelligence and Data Science, KPR Institute of Engineering and Technology, Coimbatore 641407, India
2
Department of Computer Science and Engineering, KPR Institute of Engineering and Technology, Coimbatore 641407, India
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(21), 7025; https://doi.org/10.3390/s21217025
Submission received: 15 September 2021 / Revised: 17 October 2021 / Accepted: 18 October 2021 / Published: 23 October 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In recent years, speech recognition technology has become a more common notion. Speech quality and intelligibility are critical for the convenience and accuracy of information transmission in speech recognition. The speech processing systems used to converse or store speech are usually designed for an environment without any background noise. However, in a real-world atmosphere, background intervention in the form of background noise and channel noise drastically reduces the performance of speech recognition systems, resulting in imprecise information transfer and exhausting the listener. When communication systems’ input or output signals are affected by noise, speech enhancement techniques try to improve their performance. To ensure the correctness of the text produced from speech, it is necessary to reduce the external noises involved in the speech audio. Reducing the external noise in audio is difficult as the speech can be of single, continuous or spontaneous words. In automatic speech recognition, there are various typical speech enhancement algorithms available that have gained considerable attention. However, these enhancement algorithms work well in simple and continuous audio signals only. Thus, in this study, a hybridized speech recognition algorithm to enhance the speech recognition accuracy is proposed. Non-linear spectral subtraction, a well-known speech enhancement algorithm, is optimized with the Hidden Markov Model and tested with 6660 medical speech transcription audio files and 1440 Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) audio files. The performance of the proposed model is compared with those of various typical speech enhancement algorithms, such as iterative signal enhancement algorithm, subspace-based speech enhancement, and non-linear spectral subtraction. The proposed cascaded hybrid algorithm was found to achieve a minimum word error rate of 9.5% and 7.6% for medical speech and RAVDESS speech, respectively. The cascading of the speech enhancement and speech-to-text conversion architectures results in higher accuracy for enhanced speech recognition. The evaluation results confirm the incorporation of the proposed method with real-time automatic speech recognition medical applications where the complexity of terms involved is high.

1. Introduction

Speech-to-text transcription has gained importance in many applications and benefits in research, the military, medical sector, smart homes, transportation systems, automatic transcription on lectures, conversations, record-making [1]. Speech recognition technology (SRT) involves the identification of patterns in audio waves and matching them with phonetics of speech to convert them into text. The accuracy of SRT dramatically depends on the quality of audio. The presence of background noises, multiple speakers, or the speaker’s accent provides erroneous transcription. Speech enhancement is a significant problem in communications at airports, medical centers, and other familiar public places. The SRT requires enhancement of speech to improve the quality and intelligibility of the signal before translation. Various approaches have been proposed for improving the quality of speech, such as the spectral subtraction algorithm [2], the signal subspace system [3], and the adaptive wiener filtering approach [4].
The spectral subtraction algorithm extracts the speech from additive noise. Enhancement of speech was achieved by estimating the spectrum of the noise-free signal and subtracting the estimated noisy signal from an available observed signal. The spectral subtraction algorithm suffered from residual noise [2]. The signal subspace algorithm was used for enhancing uncorrelated additive noise. This approach decomposes the noisy signal’s vector space into signal plus subspace and orthogonal noise subspace using Karhunen–Loeve transforms (KLT) or eigenvalue decomposition. The signal plus slot is used for processing, and the noisy subspace is discarded [3]. The noisy speech frames were classified into speech-dominated frames and noise-dominated frames using a signal KLT-based technique [4]. A Weiner filter-based algorithm was proposed for enhancing the signals, and it had a drawback of fixed frequency that required estimation of the frequency spectrum of both clean signal and noise before filtering [5]. An Adaptive Wiener filter method was proposed to overcome the disadvantage of the traditional Weiner filer that used an adaptation of the filter transfer function on sample to sample based on speech signal statistics [6]. The adaptive Weiner Filter-based approach was found to provide the best improvement over the signal-to-noise ratio.
The neural networks that learn the statistical information automatically using non-linear processing units were introduced for noise reduction. The deeper networks are considered to be more efficient in learning than the shallow networks [7]. A deep auto-encoder (DAE) algorithm was proposed for training the deep network architectures [8]. The challenge with DAE is the difficulty in generalizing the algorithm for all types of speech signals. Over conventional minimum mean square error (MMSE) based statistical techniques, supervised methods using deep neural networks were proposed to enhance the large volume of speech data. These methods were found to handle non-stationary noises effectively [9]. A voice activity detector (VAD) was introduced to estimate the noise during the non-speech, but it failed on encrypted speech signals [10]. Recurrent neural network (RNN) based speech enhancement techniques were introduced. The RNNs are found to produce significant performance by taking advantage of the temporal information over the noisy spectrum. Long Short-Term Memory (LSTM) was implemented for optimal speech enhancement and produced optimal results [11].
Generative adversarial networks (GAN) were used to construct the clear speech signals from the noise signals over RNNs [12]. Multiple deep neural networks were recommended over the single neural network for speech enhancement with known and unknown sources of noise [13]. A de-reverberation method combining correlation-based blind de-convolution and modified spectral subtraction was presented for speech enhancement where inverse filtered reverberation was suppressed by the spectral subtraction [14]. Though the conventional subtraction method reduces the noise level from speech, it introduces distortion noise in considerable spectral variation. The multiband spectral subtraction algorithm was proposed to overcome the distortion, maintaining the quality of speech signal [15,16]. Recent studies focus on the non-linear spectral subtraction algorithm for speech enhancements due to the significant variation in signal-to-noise ratio. The spectrum of real-world noise is irregular, and they have been affected more adversely at some frequencies. Non-linear spectral subtraction approaches are recommended to handle cleaning the speech signal [17,18]. The sub-space-based signal estimation method was proposed based on modified singular value decomposition (SVD) of data matrices that recovers speech signals from noisy observations [19]. The technique focused on mapping the observed signal to a clean signal, suppressing the noise subspace. Though the subspace-based speech enhancement deals with noise distortion, it is also used to remove colored noise [20].
Generalized sidelobe canceller (GSC) has been used for non-stationary broadband signals. GSC separates the desired signals from the interference by exploiting spatial information about the source location [21]. Noise suppression techniques play a vital role in automatic voice recognition (AVR) strategies, aiming to provide a clear gap between the clean and noisy speech signals [22]. Wiener filter and spectral subtraction combined noise estimator was proposed to control the noise energy in current frames and estimate noise from preceding frames by minimizing over subtraction [23,24,25]. Iterative signal enhancement (ISE) algorithm based on truncated SVD was proposed to obtain improved selective frequency for filtering noises from speech signals. ISE performed better than other classical algorithms, especially with speech signals [26,27].
From the literature study, it is evident that Nonlinear Spectral Subtraction (NSS) and Iterative Signal Enhancement algorithm (ISE) are the most effective methods for speech enhancement [28]. This paper proposes hybridization of NSS and ISE methods for further enhancing the speech signals. The performance of the hybrid algorithm is compared with the implementation of individual practices.
Dynamic time warping (DTW) is a dynamic programming algorithm technique used for determining the correspondence between two sequences of speech that may differ in time or speed [29]. For example, resemblances in speaking a specific word would be identified, even if in one audio the person was speaking slowly and if in another the same person was talking more quickly, or even if there were hastening and slowing down during the course of one observation [30]. DTW can be applied to audio, video, or any data that can be represented using a linear representation. Table 1 gives the insights about the literature survey.

2. Proposed Methodology

This work aims to enhance the speech signal by suppressing the noise signals involved in voice control applications. Speech enhancement is an essential factor in speech recognition as it can be used as a pre-processor to enhance speech. Generally, the source of the noise signals can be background noise, electromagnetically induced noise. Reducing these noises will result in increasing the intelligibility of the speech and agreeable speech recognition. Figure 1 depicts the flow of the idea of enhancing the noisy speech signal.
In this article, a hybridized speech enhancement algorithm is proposed with experimental results. Section 2.1 and Section 2.2 present various speech enhancement and speech-to-text conversion algorithms used for various speech signals, respectively. Section 3 represents the results obtained for the speech enhancement algorithms and their comparisons by using the word error rate values by converting the enhanced speech to text.

2.1. Speech Enhancement Algorithms

Enhancement of speech is essential when the terms involved are complex and involve external noises. There are various speech enhancement (SE) algorithms available, which are discussed in the following section.

2.1.1. Iterative Signal Enhancement Algorithm (ISE)

The iterative signal enhancement algorithm is used for reducing the noise in speech. The algorithm is designed based on a trimmed singular value decomposition (SVD) procedure and can be used as a tool for enhancing the noisy speech signal by suppressing the additive noise present in the signal. Compared to the classical algorithms, the ISE algorithm escalates similar improvements in Signal-to-Noise Ratio (SNR).
The ISE algorithm executes in two phases. The first phase produces an enhanced signal s[i] from the noisy signal ns[i]. The enhanced signal s[i], which comes from phase I, still contains some noise. So, the noise removal phase is repeated a certain number of times depending on the level of noise present in the signal. The signal decomposition of the enhanced signal at each level is given a rank ranging from 1 to level l. The first rank signal decomposition sd[i] for the s[i] is obtained by averaging the anti-diagonals of the Hankel matrix to rank-1. Hankel matrices are square matrices with constant ascending skew-diagonals from left to right. Hankel matrix is constructed from the signal is helpful for decomposition of constant signals and time-frequency representation. The rank-1 signal decomposition covers most of the energetic spectral band of the noisy input signal in the frequency domain.
The signal decomposition sd[i] is subtracted from the input signal sn[i] to calculate the residual signal r[i], and the phase-I procedure is repeated by using the residual signal as the input for the upcoming iterations. The signal decomposition sd[i] is summed up throughout the iterations to obtain the enhanced signal s[i]. The iteration is stopped once the residual signal r[i] contains only the noise components. The phase-2 concentrates on the residual signal to maintain the number of iterations of phase-I. The working of the ISE algorithm is depicted in Figure 2.
The ISE algorithm is advantageous over the typical enhancement algorithms as ISE has a better frequency selectivity for filtering out the noise than the standard algorithms.

2.1.2. Subspace—Based Speech Enhancement

Subspace methods, also called high-resolution or super-resolution methods, use an eigenvalue analysis or eigendecomposition of the correlation matrix to derive frequency component values of a signal. The process for subspace-based speech enhancement can be stated as follows:
(1)
Isolating the subspace as signal and noise subspaces from the original subspace
(noise mixed speech)
(2)
Eliminating the noise-only subspace that has been isolated in step1.
Assume s(i) represent the pure speech signals and let n(i) indicates the zero-mean additive noises mixed in the pure speech. The observed noisy speech x(i) can be given by
x(i) = s(i) + n(i)
Allow Rx, Rs, and Rn to be (p × q) genuine autocorrelation matrices of x(i), s(i), and n(i), respectively, with q > p. Rx = Rs + Rn is obvious given the assumption of uncorrelated speech and noise. Regardless of the specific optimization criterion, speech enrichment is now obtained by
  • By nullifying the components in the noise subspace, the enhanced speech is constrained to inhabit only the signal subspace.
  • Changing (decreasing) the eigenvalues of the signal subspace.
The input speech signal with noise is split into unique spaces. Each and every space is individually processed. In Figure 3, the noisy input signal is fed as the input, whereas the individual subspaces are allocated with a vector value and rank, respectively. The filter bank output of each subspace is obtained by filtering the noisy signal n(i) with its corresponding eigenfilter vector (EVF) and its reversed eigenfilter vector (REFV). This filtered output is fed to the not gate (NG) and their summation results in the enhanced speech signal.

2.1.3. Nonlinear Spectral Subtraction (NSS)

Nonlinear Spectral subtraction is one of the most primitive and notably the most famous speech enhancement methods. NSS can be predominantly valid in cases where a boisterous environment contaminates an original speech signal with the same bandwidth as that of speech. To decrease external noise, the NSS algorithm is developed, which considers the change of signal-to-noise ratio across the speech spectrum and uses a distinct over-subtraction factor in each frequency band. Figure 4 describes the steps involved in extracting only the speech from the noisy speech signal. The speech signal with noise is given as input to the Fast Fourier Transform module. Let y(i) denote the noisy speech, i.e., the pure speech signal s(i) is polluted with the noise signal n(i). Then in the time domain, their relationship is described as:
y(i) = s(i) + n(i)
To obtain their relation in the spectral domain, take the Discrete Fourier Transform (DFT) and power magnitude for Equation (1) with the assumption that noise and speech signals are uncorrelated, then the relation is described as follows:
|Y(r,f)|2 = |S(r,f)|2 + |N(r,f)|2
where r and f denote the frame and frequency values respectively. Now with the assumption that |N(r,f)| and | N ^ (r,f)| can be estimated, the spectral subtraction can be formulated as follows:
| S ^ ( r , f ) | 2 = | Y ( r , f ) | 2 | N ^ ( r , f ) | 2
The spectral subtraction algorithm adapts the damaged speech signal’s short-term spectral magnitude appropriately. The signal is modified so that the synthesized signal feels as near to the unbroken voice signal as possible. A noise power estimate and a subtraction rule are used to calculate the appropriate weighting of spectral magnitudes.
The Word Error Rate represents the amount of word error occurring during the speech. It can be calculated using the formula
WER = Ns + Nd + (Ni/Nn)
where ‘Ns’ represents the number of substitutions, ‘Nd’ indicates the number of deletions, ‘Ni’ means the number of insertions, and ‘Nn’ stands for the number of words in a sentence.

2.2. Speech to Text Conversion

The primary purpose of speech to text (STT) or speech recognition is to enable a real-time dictation of audio signals into text. All the STT systems depend on the acoustic model and the language model. In the case of including the additional feature vocabulary systems, a pronunciation model can be used. It is challenging to construct a universal speech recognizer. To develop the best quality STT system, it needs to specialize in a specific language, idiom, application domain, speech type, and communication frequency.

Hidden Markov Model

Hidden Markov Models (HMM) are mainly used for general-purpose speech recognition systems. In general, the speech signals are observed as a stationary signal whose amplitude and frequency remains constant. It takes a very short time scale for a speech to be estimated as a static process.
In HMM, it is possible to train the data set automatically, which makes it easy computationally, and hence it is extensively used. The HMM would generate a sequence of n-dimensional real-valued vectors every ten milliseconds (with ‘n’ being a small integer value such as 5 or 10) in speech recognition. The first coefficient of a Fourier transform of a small part of the speech is extracted and decorrelated with the cosine transform to calculate the cepstral coefficient vectors.
In each state of the HMM, a statistical distribution of a mixture of diagonal covariance Gaussians is performed to present the probability of each identified vector. In the current speech recognition systems, every word or phoneme has its output distribution, whereas, in HMM, the sequence of the terms or phonemes is constructed by cascading the individually trained cepstral vectors for the specific words and phonemes, respectively.
Modern voice recognition systems use diverse combinations of various common strategies to improve outcomes beyond the fundamental approach. Context dependency of the phonemes is required for the traditional large-sized vocabulary system to enable the phonemes with different pre and post contexts to have unique HMM states. These HMM states can be used to normalize different recordings and the speaker conditions using the cepstral normalization method. Vocal tract length normalization (VLTN) can be used to further normalize male-female or other speaker criteria. To capture the speech dynamics, linking and linear discriminant analysis (LDA) based projects can be used, which is followed by either the heteroscedastic LDA step or global semi-tied covariance transform method. To optimize the classification-related measure of the training data, many systems employ discriminative training strategies, which can be used to avoid the purely statistical approach to estimate the HMM parameter.
The HMM is used to convert speech features into HMM parameters and calculate all speech samples’ likelihood. Recognition of this likelihood of speech samples is used to recognize the spoken words. Figure 5 represents the working model of the HMM that decodes the extracted features based on the parameters such as acoustic models, pronunciation dictionary, and the language model for which speech recognition is required.

3. Hybrid Speech Enhancement Algorithm (HSEA)

The speech signal can be a single, continuous, or spontaneous word. The spontaneous words are complicated as their speech is fast and the number of words is high. This spontaneousness creates complications in automatic speech recognition. The proposed work mainly concentrates on spontaneous speech in the medical domain. The words used in the medical environment are complicated, and the medication prescriptions need clarity. Hence, the idea is to develop a hybridized algorithm that can resolve the issue and assure accuracy.
The nonlinear spectral subtraction and the hidden Markov model are cascaded to form a hybrid architecture to enhance the accuracy of the speech recognition system. The NSS provides a better result for the spontaneous signal by suppressing its noise to a greater extent, and the HMM can give high accuracy in converting the speech to text. Figure 6 shows the proposed hybrid algorithm architecture to enhance the speech signal by cascading the NSS and the HMM algorithms. The steps involved in the HSEA is described in Algorithm 1.
Algorithm 1 Hybrid Speech Enhancement Algorithm
Input y(i) the noisy speech signal with n(i) the noise and s(i) the clear speech signal
  • Segmentation: the speech signal is segmented into frames.
  • Process: The noisy signal is processed frame by frame and the Fourier transform is calculated as
     y(r,f) = s(r,f) + n(r,f)
    where ‘r’ is the frame number.
  • Calculation: The short-term power spectrum of y(i) is calculated as
     |y(f)|2 = |s(f)|2 + |n(f)|2
  • Estimation: By removing the noise from the input signal, the speech is estimated as follows:
    | s ^ ( f ) | 2 = | y ( f ) | 2 | n ^ (r,f)|2
  • Obtain: The   noise   spectrum   | n ^ (f,t)|2 is obtained by averaging the recent pause frames
      | n ^ ( r , f ) | 2 = 1 N k = 0 N 1 | y k ( f ) | 2
    where N is the number of consecutive speech frames
  • Reconstruction:   Step   5   is   reconstructed   by   expressing   s ^ (t)2 as the product of noise and the spectral subtraction factor
      | s ^ ( f ) | 2 = | n ^ (f)|2/|y(f)|2
  • Formulation: The Hankel matrix of the enhanced speech signal is formulated from step 6.
  • Optimization: The enhanced speech is optimized using the least squares and the minimum variance optimal estimator as
    Hs = ΣVt
  • Conversion: The enhanced speech signal is converted to text.
Output: The text format of the noise removed speech signal with reduced WER.

4. Results and Discussion

The proposed model has been tested on a dataset that contains medical speech of about 8.5 h and RAVDESS emotional speech dataset. The speech includes both male and female speeches, with each audio time ranging from 0.2 s to 60 s.

4.1. Performance Analysis of Speech Enhancement Algorithms

Various speech enhancement algorithms are present to enhance the speech signal by reducing the noise in the surrounding. The proposed work tested the algorithms such as ISE, sunspace, and NSS for the medical audio dataset, and the findings are discussed in this section.

4.1.1. ISE for Spontaneous Signal

One of the numerous issues that automatic speech recognition systems face is processing spontaneous speech. Spontaneous speech is characterized as utterances that comprise well-formed phrases similar to those found in written texts. Disfluencies (complete pauses, repetitions, false starts, and so on) are the main characteristics of this type of speech, and numerous studies have concentrated on detecting and correcting them.
The waveform and the spectrogram results produced by the ISE algorithm are shown in Figure 7. Figure 7a,b show a speech signal’s time waveform and spectrogram, respectively. Figure 7a shows how the noisy waveform is getting transferred to noise reduced waveform signal using the ISE algorithm.
Figure 7b represents the spectrogram of a signal that has been enhanced using the ISE. It is visible that there is an enhancement in the noise corrupted speech while using the ISE speech enhancement algorithm. The WER is calculated for the speech before and after removing the noise. The comparison of the WER before and after reducing the noise is given in Table 2.
It is visible from Table 2 that the WER is reduced to a certain extent after using the speech enhancement algorithm ISE, and hence it increases the accuracy of the speech to a minimal extent.

4.1.2. Sub Space Method for Spontaneous Signal

The input noise contains the noise from the surroundings, such as the fan and vehicle noises. It is necessary to suppress the external noise to enhance the actual speech to improve its recognition irrespective of its length and complexity. Figure 8a shows the waveform and the spectrogram of the noisy speech signal of length 60 s. When this noisy speech is treated with the subspace method for enhancing the speech, the external noise is removed, and its corresponding spectrogram is shown in Figure 8b. The improved speech can be converted into text for further analysis.
Although the noise removal can be visibly seen in Figure 8b, the quantitative analysis is done by applying Equation (1) in the text of the enhanced speech signal to calculate the word error rate. Table 3 represents the WER calculated, and accuracy of the noise removed signal by using the subspace method.

4.1.3. NSS for Spontaneous Signal

Spectral subtraction is one of the most basic and also possibly the most famous speech enhancement methods. Figure 9a shows the time waveform and the spectrogram of the noisy speech signal, where the noise can be any domestic noise in the surroundings. The noise can be effectively removed using the non-linear spectral subtraction, and the corresponding waveform and the spectrogram are shown in Figure 9b. From Figure 9a,b, it is visible that the noise has been eliminated to a greater extent.
The enhanced speech signal is converted into text using the HMM or the DTW-based text conversion algorithm to find the accuracy level of the enhanced speech signal. Table 4 represents the WER calculated for the enhanced speech signal using the NSS.
Comparing Table 1, Table 2 and Table 3, it can be seen that NSS produces less WER, which makes NSS a more accurate speech enhancement algorithm compared to ISE and subspace methods.
The WER calculated for the spontaneous signal after enhancing the signal using the NSS, subspace, and ISE are compared and shown in Figure 10. As the number of words increases in the speech, the ISE and subspace produce more WER, whereas the NSS can maintain the accuracy of the speech.

4.2. Speech to Text Conversion of Enhanced Speech

The accuracy and efficiency of the speech can be visualized using the time waveform, log transformation, and spectrogram. The text analysis is used to represent the efficiency of the speech with quantitative measurement so that the comparison of speech enhancement becomes accurate.

Performance of HMM

The word error rate of a noisy signal and the enhanced speech signal is calculated using the HMM algorithm and is shown in Table 5. It can be seen that the word error rate of a noisy speech is high compared to that of an enhanced speech signal.

4.3. Performance Analysis of HSEA

There are various speech enhancement algorithms to reduce the domestic noises recorded during the speech and produce an enhanced signal. When the length of the speech (the number of words involved in the speech) increases, the performance of the algorithms does not seem to be effective as many of them work well with only small words. The enhancement also depends on the complexity of words involved in the speech. Specific applications such as medical speech need more accuracy as the terms involved are complex, and the noise removal should not affect the originality of the words uttered. Hence it is more important to select the best noise removal algorithm for complicated applications.
In the proposed method HSEA, the speech enhancement algorithm NSS and the speech to text conversion algorithm HMM are cascaded to further increase the accuracy of the speech. Table 6 shows the WER of the spontaneous signal with different lengths of words. The WER is calculated for the NSS algorithm without using the HMM speech-to-text conversion method, and the NSS is cascaded with HMM.
Figure 11 represents the graph for the WER values calculated for the spontaneous signal using the NSS speech enhancement signal with and without cascading it with HMM. Although the length of the spontaneous speech increases, the cascaded hybrid method produces less WER than the stand-alone NSS speech enhancement method.
Table 7 represents the WER calculation for the RAVDESS data with a spontaneous signal. RAVDESS contains audio of different length with both male and female speakers. The WER is calculated using the NSS algorithm and the proposed HSEA method. Compared to Table 6, it is visible that the WER of RAVDESS data is less than that of medical speech. This is because the complexity of medical speech is high compared to that of RAVDESS emotional speech data.
Figure 12 describes the performance comparison of HSEA and the NSS algorithm for the RAVDESS dataset with 1440 audio files. It can be seen that RAVDESS has less WER compared to that of the medical speech transcription data as the terms involved in medical speech are complex compared to that of emotional speech. Additionally, HSEA produces less WER compared to that of the NSS algorithm.

5. Conclusions

In this study, an HSEA is employed for medical audio data to reduce the word error rate. The entire procedure of the proposed method consists of the NSS and the HMM architecture with the medical data set. The typical NSS algorithm is optimized with least squares and minimum variance and then cascaded with the speech-to-text algorithm HMM to establish the proposed HSEA. In HSEA, the speech is enhanced using the NSS algorithm, and again the enhanced speech is optimized using the optimization criteria of the subspace method. As the double layer of enhancement is performed, HSEA can perform competitively with other typical speech enhancement models such as NSS, ISE, and Subspace. The proposed model has been validated with 6660 medical and 1440 RAVDESS audio data. The validation of the proposed HSEA architecture has proven to achieve maximum accuracy of 90.5% with a minimum word error rate of 9.5% for medical speech transcription and accuracy of 92.4%, the word error rate of 7.6% for the RAVDESS audio data. The validation of the proposed HSEA architecture has proven to achieve maximum accuracy of 90.5% with a minimum word error rate of 9.5%.
The proposed methodology can produce a clear speech signal as the output with reducing word error rate, making it efficient for all kinds of applications involving speech in open space. Despite the advantages, the methodology involves mathematical complexities as the layer of speech purification increases, which remains a limitation of the proposed method. Therefore, future work can concentrate on integrating the layers in such a way as to reduce the computational complexities and to achieve high efficiency.

Author Contributions

Conceptualization, J.G., Y.N.; methodology, J.G., Y.N.; software, J.G., Y.N.; validation, J.G., Y.N.; formal analysis, J.G., Y.N.; investigation, J.G., Y.N.; resources, J.G., S.P.K.R.; writing—original draft preparation, J.G., Y.N., S.P.K.R.; writing—review and editing, J.G., S.P.K.R.; supervision Y.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/paultimothymooney/medical-speech-transcription-and-intent.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Anasuya, M.A.; Katti, S.K. Speech recognition by machine: A review. Int. J. Comput. Sci. Inf. Secur. 2009, 6, 181–205. [Google Scholar]
  2. Boll, S. Suppression of acoustic noise in speech using spectral subtraction. IEEE Trans. Acoust. Speech Signal. Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  3. Ephraim, Y.; Van Trees, H. A signal subspace approach for speech enhancement. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing, Minneapolis, MN, USA, 27–30 April 1993; Volume 2, pp. 355–358. [Google Scholar]
  4. Rezayee, A.; Gazor, S. An adaptive KLT approach for speech enhancement. IEEE Trans. Speech Audio Process. 2001, 9, 87–95. [Google Scholar] [CrossRef]
  5. Abd El-Fattah, M.A.; Dessouky, M.I.; Diab, S.M.; Abd Elsamie, F.E. Adaptive wiener filtering approach for speech en-hancement. Ubiquitous Comput. Commun. J. 2010, 3, 23–31. [Google Scholar]
  6. Lu, X.; Unoki, M.; Matsuda, S.; Hori, C.; Kashioka, H. Controlling tradeoff between approximation accuracy and complexity of a smooth function in a reproducing kernel Hilbert space for noise reduction. IEEE Trans. Signal. Process. 2013, 61, 601–610. [Google Scholar] [CrossRef]
  7. Bengio, Y. Learning deep architectures for AI. Learn. Deep. Archit. AI 2009, 2, 1–127. [Google Scholar] [CrossRef]
  8. Lu, X.; Tsao, Y.; Matsuda, S.; Hori, C. Speech enhancement based on deep denoising autoencoder. In Proceedings of the In-terspeech, Annual Conference of the 3 International Speech Communication Association, Lyon, France, 25–29 August 2013; pp. 436–440. [Google Scholar]
  9. Xu, Y.; Du, J.; Dai, L.-R.; Lee, C.-H. A regression approach to speech enhancement based on deep neural networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 7–19. [Google Scholar] [CrossRef]
  10. Barnov, A.; Bar Bracha, V.; Markovich-Golan, S. QRD based MVDR beamforming for fast tracking of speech and noise dynamics. In Proceedings of the 2017 IEEE Workshop on Applications of Signal. Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 15–18 October 2017; pp. 369–373. [Google Scholar] [CrossRef]
  11. Weninger, F.; Erdogan, H.; Watanabe, S.; Vincent, E.; Le Roux, J.; Hershey, J.R.; Schuller, B. Speech enhancement with LSTM recurrent neural networks and its application to noise-robust ASR. In Proceedings of the International Conference on Latent Variable Analysis and Signal Separation, Liberec, Czech Republic, 25–28 August 2015; pp. 91–99. [Google Scholar]
  12. Pascual, S.; Bonafonte, A.; Serrà, J. SEGAN: Speech enhancement generative adversarial network. In Proceedings of the Interspeech, 18th Annual Conference of the International Speech Communication Association, Stockholm, Sweden, 20–24 August 2017; pp. 3642–3646. [Google Scholar]
  13. Kumar, A.; Florencio, D. Speech enhancement in multiple-noise conditions using deep neural networks. In Proceedings of the Interspeech, San Francisco, CA, USA, 8–12 September 2016; pp. 3738–3742. [Google Scholar]
  14. Furuya, K.; Kataoka, A. Robust speech dereverberation using multichannel blind deconvolution with spectral subtraction. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 1579–1591. [Google Scholar] [CrossRef]
  15. Kamath, S.; Loizou, P. A multi-band spectral subtraction method for enhancing speech corrupted by colored noise. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing, Orlando, FL, USA, 13–17 May 2002; Volume 4, p. 44164. [Google Scholar]
  16. Udrea, R.M.; Oprea, C.C.; Stanciu, C. Multi-microphone Noise reduction system integrating nonlinear multi-band spectral subtraction. In Pervasive Computing Paradigms for Mental Health; Oliver, N., Serino, S., Matic, A., Cipresso, P., Filipovic, N., Gavrilovska, L., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 133–138. [Google Scholar]
  17. Tufts, D.W.; Kumaresan, R.; Kirsteins, I. Data adaptive signal estimation by singular value decomposition of a data matrix. Proc. IEEE 1982, 70, 684–685. [Google Scholar] [CrossRef]
  18. Saliha, B.; Youssef, E.; Abdeslam, D. A Study on Automatic Speech Recognition. J. Inf. Technol. Rev. 2019, 10, 77–85. [Google Scholar]
  19. Hermus, K.; Wambacq, P.; Van Hamme, H. A review of signal subspace speech enhancement and its application to noise robust speech recognition. EURASIP J. Adv. Signal. Process. 2006, 2007, 045821. [Google Scholar] [CrossRef] [Green Version]
  20. Pardede, H.; Ramli, K.; Suryanto, Y.; Hayati, N.; Presekal, A. speech enhancement for secure communication using coupled spectral subtraction and wiener filter. Electronics 2019, 8, 897. [Google Scholar] [CrossRef] [Green Version]
  21. Jousse, V.; Petit-Renaud, S.; Meignier, S.; Esteve, Y.; Jacquin, C. Automatic named identification of speakers using diarization and ASR systems. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal, Processing, Taipei, Taiwan, 19–24 April 2009; pp. 4557–4560. [Google Scholar] [CrossRef]
  22. Herbordt, W.; Buchner, H.; Kellermann, W. An acoustic human-machine front-end for multimedia applications. EURASIP J. Appl. Signal. Process. 2003, 1, 21–31. [Google Scholar] [CrossRef] [Green Version]
  23. Doclo, S.; Dologlou, I.; Moonen, M. A Novel Iterative Signal Enhancement Algorithm for Noise Reduction in Speech. Available online: https://www.isca-speech.org/archive_v0/archive_papers/icslp_1998/i98_0131.pdf (accessed on 16 October 2021).
  24. Bin Amin, T.; Mahmood, I. Speech recognition using dynamic time warping. In Proceedings of the 2nd International Conference on Advances in Space Technologies, Islamabad, Pakistan, 29–30 November 2008; pp. 74–79. [Google Scholar]
  25. Jenifa, G.; Yuvaraj, N.; Karthikeyan, B.; Preethaa, K.R.S. Deep learning based voice assistance in hospitals using face recognition. J. Phys. Conf. Ser. 2021, 1916, 012159. [Google Scholar] [CrossRef]
  26. Yuvaraj, N.; Sanjeev, M.; Jenifa, G.; Preethaa, K.R.S. Voice activated face recognition based smart support system. J. Phys. Conf. Ser. 2021, 1916, 012158. [Google Scholar] [CrossRef]
  27. Markovich-Golan, S.; Gannot, S. Performance analysis of the covariance subtraction method for relative transfer function estimation and comparison to the covariance whitening method. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 544–548. [Google Scholar]
  28. Souden, M.; Chen, J.; Benesty, J.; Affes, S. Gaussian model-based multichannel speech presence probability. IEEE Trans. Audio Speech Lang. Process. 2010, 18, 1072–1077. [Google Scholar] [CrossRef]
  29. Serizel, R.; Moonen, M.; Van Dijk, B.; Wouters, J. Low-rank Approximation based multichannel wiener filter algorithms for noise reduction with application in cochlear implants. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 785–799. [Google Scholar] [CrossRef] [Green Version]
  30. Qazi, O.U.R.; van Dijk, B.; Moonen, M.; Wouters, J. Understanding the effect of noise on electrical stimulation sequences in cochlear implants and its impact on speech intelligibility. Hear. Res. 2013, 299, 79–87. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Proposed idea.
Figure 1. Proposed idea.
Sensors 21 07025 g001
Figure 2. Iterative Signal Enhancement Algorithm.
Figure 2. Iterative Signal Enhancement Algorithm.
Sensors 21 07025 g002
Figure 3. Subspace—based speech enhancement.
Figure 3. Subspace—based speech enhancement.
Sensors 21 07025 g003
Figure 4. Nonlinear spectral subtraction.
Figure 4. Nonlinear spectral subtraction.
Sensors 21 07025 g004
Figure 5. Working model of Hidden Markov Model.
Figure 5. Working model of Hidden Markov Model.
Sensors 21 07025 g005
Figure 6. Proposed hybridized algorithm for speech enhancement.
Figure 6. Proposed hybridized algorithm for speech enhancement.
Sensors 21 07025 g006
Figure 7. (a,b) Time waveform and Spectrogram of speech signal using ISE.
Figure 7. (a,b) Time waveform and Spectrogram of speech signal using ISE.
Sensors 21 07025 g007
Figure 8. (a,b) Time waveform and Spectrogram of speech signal using subspace.
Figure 8. (a,b) Time waveform and Spectrogram of speech signal using subspace.
Sensors 21 07025 g008
Figure 9. (a,b) Time waveform and Spectrogram of speech signal using NSS.
Figure 9. (a,b) Time waveform and Spectrogram of speech signal using NSS.
Sensors 21 07025 g009
Figure 10. WER comparison of typical SE algorithms.
Figure 10. WER comparison of typical SE algorithms.
Sensors 21 07025 g010
Figure 11. Performance of HSEA for medical speech.
Figure 11. Performance of HSEA for medical speech.
Sensors 21 07025 g011
Figure 12. Performance of HSEA for RAVDESS.
Figure 12. Performance of HSEA for RAVDESS.
Sensors 21 07025 g012
Table 1. Literature survey based on category of techniques.
Table 1. Literature survey based on category of techniques.
TechniquesPerformanceAdvantage(s)Disadvantage(s)
Spectral subtraction algorithm [2]Estimating the spectrum of the noise-free signal and subtracting the estimated noisy signal is done to enhance the speechCan be applied for both stationary and non-stationary noisesResultant speech contains residual noise
Signal subspace algorithm [3]Uses Karhunen-Loeve transforms (KLT) or eigenvalue decompositionDiscards the noisy space
Can directly use the state-space representation for the system
Realizations done by state-space are not unique
Weiner filter-based algorithm [5]Mainly used in real time applicationsBetter performance for noise cancellationNeeds more number of computations
Adaptive Weiner filter-based algorithm [6]Mainly used in real time applicationsReduced/moderate computational complexityMean square error is not always relevant
Deep Auto-Encoder (DAE) algorithm [8]Deep denoising autoencoders are used to enhance the speech featuresEfficient for resonant speech recognitionMainly used for clean/controlled speech only
Voice Activity Detector (VAD) [10]Works on the long pause between the wordsCan classify the noise even during the pause of the speechNot efficient for encrypted speech signals
Long Short-Term Memory (LSTM) [11]A type of RNN, and it can learn long period dependenciesProduces good result in speech recognitionConcentrates only on the size (length) of the speech
Generative Adversarial Networks (GAN) [12]It is a type of RNN and it constructs clear speech from the noisy speech Generate audio that looks similar to original audio by eliminating noiseHarder to train
Multiband spectral subtraction algorithm [15]Inverse filtered reverberation was suppressed by the spectral subtractionOvercome the distortion, maintaining the quality of speech signalNot suitable for highly random real-world noise
Table 2. WER and its percentage after using ISE.
Table 2. WER and its percentage after using ISE.
Length of the WordWER (%)Accuracy (%)
100 words41.658.4
150 words51.548.5
200 words55.544.5
250 words59.940.1
300 words69.530.5
350 words73.426.6
Table 3. WER and its percentage using subspace method.
Table 3. WER and its percentage using subspace method.
Length of the WordWER (%)Accuracy (%)
100 words29.860.2
150 words40.449.6
200 words43.146.9
250 words46.743.3
300 words52.837.2
350 words60.229.8
Table 4. WER and its percentage using NSS.
Table 4. WER and its percentage using NSS.
Length of the WordWER (%)Accuracy (%)
100 words11.988.1
150 words19.480.6
200 words20.379.7
250 words21.278.8
300 words23.176.9
350 words24.375.7
Table 5. WER count and percentage with and without noise.
Table 5. WER count and percentage with and without noise.
Length of the WordWER (%) with NoiseWER (%) without Noise
100 words26.221.4
150 words28.123.3
200 words31.628.9
250 words34.932.7
300 words41.539.8
350 words44.341.2
Table 6. WER, accuracy using the HSEA and NSS.
Table 6. WER, accuracy using the HSEA and NSS.
Length of the WordWER (%)
HSEANSS
100 words9.511.9
150 words11.419.4
200 words13.620.3
250 words16.721.2
300 words17.123.1
350 words19.924.3
Table 7. WER, accuracy using the HSEA and NSS for RAVDESS.
Table 7. WER, accuracy using the HSEA and NSS for RAVDESS.
Length of the WordWER (%)
HSEANSS
100 words7.68.1
150 words9.214.5
200 words11.917.9
250 words14.819.2
300 words15.321.1
350 words17.522.9
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gnanamanickam, J.; Natarajan, Y.; K. R., S.P. A Hybrid Speech Enhancement Algorithm for Voice Assistance Application. Sensors 2021, 21, 7025. https://doi.org/10.3390/s21217025

AMA Style

Gnanamanickam J, Natarajan Y, K. R. SP. A Hybrid Speech Enhancement Algorithm for Voice Assistance Application. Sensors. 2021; 21(21):7025. https://doi.org/10.3390/s21217025

Chicago/Turabian Style

Gnanamanickam, Jenifa, Yuvaraj Natarajan, and Sri Preethaa K. R. 2021. "A Hybrid Speech Enhancement Algorithm for Voice Assistance Application" Sensors 21, no. 21: 7025. https://doi.org/10.3390/s21217025

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop