Next Article in Journal
Phish Responder: A Hybrid Machine Learning Approach to Detect Phishing and Spam Emails
Previous Article in Journal
A Low-Cost Collaborative Robot for Science and Education Purposes to Foster the Industry 4.0 Implementation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Preliminary Study of Robust Speech Feature Extraction Based on Maximizing the Probability of States in Deep Acoustic Models

Department of Electrical Engineering, National Chi Nan University, Nantou 545, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Syst. Innov. 2022, 5(4), 71; https://doi.org/10.3390/asi5040071
Submission received: 30 June 2022 / Revised: 18 July 2022 / Accepted: 19 July 2022 / Published: 26 July 2022

Abstract

:
This study proposes a novel robust speech feature extraction technique to improve speech recognition performance in noisy environments. This novel method exploits the information provided by the original acoustic model in the automatic speech recognition (ASR) system to learn a deep neural network that converts the original speech features. This deep neural network is trained to maximize the posterior accuracy of the state sequences of acoustic models with respect to the speech feature sequences. Compared with the robustness methods that retrain or adapt acoustic models, the new method has the advantages of less computation load and faster training. In the experiments conducted on the medium-vocabulary TIMIT database and task, the presented method provides lower word error rates than the unprocessed baseline and speech-enhancement-based techniques. These results indicate that the presented method is promising and worth further developing.

1. Introduction

Electronic devices and services such as mobile phones, hearing aids, headphones, and teleconference systems play a significant role in our lives. In particular, voice-based functions and applications (voice interaction, voice communication, and speech recognition) are essential in these devices and services. However, various sources of interference exist that deteriorate speech signals during transmission, and thus they undermine the capability of the functions mentioned above and applications. These interference sources include additive noise, channel distortion, and reverberation. In order to alleviate the influence of interference, various methods have been presented in recent decades from different perspectives of speech processing systems, such as front-end signal processing, acoustic feature design, and back-end acoustic modeling. Roughly speaking, the methods regarding the front-end acoustic signal-domain processing belong to speech enhancement (SE). In contrast, the techniques for the acoustic features and models are related to robust speech recognition.
According to [1], speech enhancement (SE) algorithms can be divided into four categories: spectral-subtractive [1,2,3], statistical modeling [1,4,5,6,7], sub-space based [1,8,9], and masking-based [10,11,12]. In particular, due to the successful development of deep neural network (DNN) techniques in the recent decade, the modeling of speech enhancement algorithms and the respective procedures (such as noise estimation and sub-space basis extraction) are significantly upgraded to achieve even better performance. For example, a DNN can be used to learn the statistical relationship between noisy utterances and clean counterparts through a massive amount of noisy–clean utterance pairs in a prepared training set. In particular, since speech enhancement aims to transform noisy speech back to clean speech, which is a standard regression problem in machine/deep learning, the mean squared error (MSE) loss function is often used in the SE-wise DNN.
However, when it comes to the performance evaluation of a speech enhancement algorithm, some objective metrics, such as the perceptual evaluation of speech quality (PESQ) [13] and short-time objective intelligibility (STOI) [14], are often used. These metrics are not necessarily related to the mean squared distance between the enhanced and original clean speech. Therefore, in some recently developed deep-learning-based SE algorithms [15,16], PESQ and STOI are employed directly to be the objective functions for the training of DNN models.
PESQ is a standard objective metric to measure speech quality, recommended by ITU-T. It is developed to predict the mean opinion scores (MOS) in subjective listening tests, and according to [17], PESQ shows a high correlation ( r 0.9 ) with MOS. PESQ is widely used to evaluate speech separation and enhancement algorithms. It mainly quantizes speech quality by computing the disturbance between the clean and separated/enhanced speech using cognitive modeling. PESQ ranges within [−0.5, 4.5], with high values indicating better quality [13,18].
Here, we provide a little more discussion about the issue that using mean squared error (MSE) as the loss function in most deep-learning-based SE algorithms [10,19,20,21] does not necessarily give rise to higher PESQ/STOI scores in the processed utterances. According to [18,21], two fundamental assumptions of the MSE loss in these SE methods somewhat contradict human auditory perception, which are as follows:
  • Every estimated element is equally important [18]:
    The MSE loss function treats each single data point equally. Regardless of the locations of the processed data, the same difference between the elements of each processed output and the desired output corresponds to the equal MSE loss value. The underlying reason is that the MSE loss is derived from each sample point of the utterance separately and independently instead of the whole utterance trajectory. However, we evaluate the quality and intelligibility of an utterance in a different manner. For example, the speech-dominant portion is treated more crucially than the silence portion in the STOI metric.
    In contrast, the PESQ metric pays more attention to isolated samples because they are more involved with speech quality. In addition, when considered in the frequency domain, the MSE loss is usually defined at the linear frequency scale. At the same time, the speech quality/intelligibility metrics are measured at human auditory frequency scales (such as the mel scale and Bark scale). To sum up, minimizing the MSE loss is not necessary to optimize the PESQ or STOI scores for noisy utterances.
  • The optimal solution is determined by the whole utterance set [21]:
    To extend the idea of the previous point, using the MSE as the objective loss function tends to consider the portions in utterances with different characteristics in an averaged manner, which might alleviate the distinct and discriminative components that highlight the speech. For example, an ordinary utterance set possesses low- and high-pitch utterances. Minimizing the MSE at these utterances indiscriminately very likely causes these utterances to have median pitches, which is not our goal at all.
Inspired by the observations mentioned above and some other literature [15,18,22], in this study, we propose a novel deep-learning-based noise-robust speech feature extraction algorithm with an MSE-irrelevant loss function. The used loss function is directly associated with the performance of a speech recognition system in noisy environments. Briefly speaking, a deep neural network is trained to perform the speech feature mapping that maximizes the posterior state probability associated with the back-end acoustic models in the speech recognition system. The resulting new speech feature representation is expected to outperform the original feature in recognition accuracy and possess noise robustness.
It is noteworthy that many research efforts have employed deep neural networks to engage in an ASR system to improve its noise robustness. For example, an end-to-end (E2E) ASR framework with a single DNN directly maps an acoustic feature sequence to a word sequence [23,24,25]. This E2E DNN is learned to optimize criteria directly associated with the ASR performance, such as the word error rate (WER). Such an E2E model can behave better and be more compact than the conventional hybrid ASR model, consisting mainly of acoustic models, a lexicon, and language models. The authors in [26] claim that the improvements in speech quality metrics such as PESQ by speech enhancement techniques do not translate into better ASR performance. They propose a joint optimization of mask-estimating network-based speech enhancement and acoustic modeling to reduce the WER. In addition, a deep learning scheme for a priori SNR estimation, termed Deep Xi, is presented in [27,28] to facilitate the conventional minimum-mean-squared-error (MMSE)-based SE methods for a robust ASR. Compared with these complicated and fine-grained techniques, our newly presented method is a relatively lightweight network that is much easier to be learned, while it is likely less effective. However, our method is a DNN-wise transformation in the same feature domain and thus can be easily integrated with these advanced methods to boost the ASR performance.
In the following sections, we first introduce the original acoustic features and the associated acoustic models in a pre-processing stage and then present the novel feature extraction algorithm. Next, we compare our method with various speech features created by the utterances enhanced by some state-of-the-art enhancement methods. Finally, we give a concluding remark.

2. Introduction of Speech Feature Representation and Acoustic Model Configuration

In this section, we set forth the process of creating the widely used speech feature representation: mel-frequency filter-bank features (FBANK) and mel-frequency cepstral coefficients (MFCC). After that, we introduce the configuration of a deep-neural-network-based hidden Markov model (DNN-HMM), which is widely used as the acoustic model structure in the current automatic speech recognition (ASR) system. This section serves as the background knowledge for a novel noise-robust speech feature extraction algorithm, which is clarified in the next section.

2.1. Mel-Frequency Cepstral Coefficients and Mel-Frequency Filter-Bank Features

Mel-frequency filter-bank features (FBANK) and mel-frequency cepstral coefficients (MFCC) are two of the most widely used speech feature representations in the state-of-the-art ASR system. FBANK and MFCC approximate the response of the human auditory system to sounds better than the linearly scaled frequency-band energies (short-time Fourier transforms of sound segments), thereby corresponding to superior recognition accuracies in ASR.
The procedures of creating a series of FBANK and MFCC features from an utterance are depicted in Figure 1. From this figure, it is clear that computing FBANK is the same as computing MFCC except for the discrete cosine transform (DCT) in the final step. The DCT operation is believed to reduce the correlation of the input significantly, and thus MFCC features are much less correlated than FBANK features. It is noteworthy that in conventional ASR systems, MFCC is preferred over FBANK due to its nearly uncorrelated characteristic. However, FBANK usually behaves better than MFCC in some advanced ASR systems whose acoustic models are non-sensitive to feature correlation.
In the following, we briefly introduce each MFCC/FBANK creation procedure shown in Figure 1.
1.
Pre-emphasis:
The first step is to apply a pre-emphasis filter on the input signal to amplify the high-frequency components because utterances usually have smaller magnitudes in high frequencies compared to low frequencies. The pre-emphasis filter can be applied to a signal x [ n ] as the following equation:
y [ n ] = x [ n ] 0.97 x [ n 1 ] .
2.
Framing:
An utterance is non-stationary, and its inherent characteristics change over time. However, it is reasonable to assume that the characteristics remain unchanged in each short-time interval. Thus, the utterance is segmented into a series of partially overlapped frames. The frame duration and frame shift are usually set to 25 ms and 10 ms, respectively.
3.
Windowing:
After framing, each frame signal is multiplied with an appropriate window function to reduce the unwanted sidelobes in spectrum [29]. We usually use the Hamming window function here, which is:
w [ n ] = 0.53836 0.46164 cos ( 2 π n N 1 ) , n = 0 , 1 , , N 1 ,
where N is the frame size.
4.
Discrete Fourier transform (DFT):
The windowed time-domain frame signal is converted to the acoustic frequency domain via a discrete Fourier transform (DFT). In the frequency domain, the signal’s human hearing characteristics (such as formants and pitches) can be better revealed. The function of the DFT is as follows:
X [ k ] = n = 0 N 1 x [ n ] e j 2 π k n N , k = 0 , 1 , , N 1 .
5.
Mel-frequency wrapping:
The perceived frequency resolution in the human hearing mechanism decreases as the physical frequencies increase, making humans less sensitive to frequency change at high frequencies. The mel-filter-banks, which consist of triangular band-pass filters, are designed to simulate the above mechanism by locating more narrow-bandwidth filters at low frequencies and fewer wide-bandwidth filters at high frequencies. The lth mel-filter H l [ k ] of the filter-bank is defined by:
H l [ k ] = 0 , k < f [ l 1 ] k f [ l 1 ] f [ l ] f [ l 1 ] , f [ l 1 ] k f [ l ] f [ l 1 ] k f [ l 1 ] f [ l ] , f [ l ] k f [ l + 1 ] 0 , k > f [ l + 1 ]
in which f [ l ] denotes the center frequency of the lth mel-filter, calculated by:
f [ l ] = N f s m e l 1 [ m e l ( f l ) + l m e l ( f H ) m e l ( f L ) J + 1 ]
with f H and f L being, respectively, the highest and lowest frequencies and
m e l ( f ) = 2595 log 10 ( 1 + f 700 )
which transforms the acoustic frequency f (in Hz) to the perceptual frequency m e l .
With the set of mel-filters, we convert the magnitude spectrum | X [ k ] | for each frame to the mel-scaled (magnitude) spectrum:
S [ l ] = k = 0 N 2 1 H l [ k ] | X [ k ] | , l = 1 , 2 , , L
6.
Logarithmic operation:
The human auditory system tends to adjust the strength of the received sound to protect the ears. When the input sound has a high amplitude, it will be suppressed by the human ears. To portray such a function, a logarithmic operation is applied to the mel-scaled spectrum S [ l ] as follows:
S l [ l ] = log ( S [ l ] ) , l = 1 , 2 , , L
In particular, the features S ( l ) [ l ] obtained in Equation (8) are called logarithmic mel-scale filter-bank spectra, which are often referred to as FBANK features.
7.
Discrete Cosine Transform (DCT):
The FBANK features can be converted to the cepstral domain via the discrete cosine transform (DCT). As mentioned earlier, the DCT has the ability to produce nearly uncorrelated features, and here the resulting outputs are MFCC features. The DCT operation to obtain the MFCC, y [ p ] from FBANK S ( l ) [ l ] , is as follows:
y [ p ] = 1 L l = 1 L S ( l ) [ l ] cos ( π p ( l 0.5 ) L ) , p = 0 , 1 , . . . , J 1 < L

2.2. Hidden Markov Model with Deep Neural Networks

In an ASR system, while observing a sequence of acoustic features O = [ o 1 , o 2 , o 3 , , o T ] (e.g., FBANK or MFCC), the objective is to find the most likely word sequences W * among all possible word sequences W = ( w 1 , w 2 , w 3 , , w T ) . That is,
W * = arg max W P ( W | O ) = arg max W P ( O | W ) P ( W ) ,
where P ( W | O ) is the posterior probability of W, which is proportional to P ( O | W ) and P ( W ) computed by the acoustic models and language models, respectively.
Among the various acoustic models, a hidden Markov model (HMM) is a double stochastic process consisting of states, each describing the observations (speech features). An HMM is parameterized by state-transition probability and state-dependent observation probability. In particular, the state-dependent observation probability has been usually computed via a Gaussian mixture model (GMM) in the past. However, it is more frequently evaluated with a deep neural network (DNN) in the state-of-the-art ASR. In the following, we briefly describe the DNN-HMM [30].
  • State-dependent observation probability [31]:
    An HMM acoustic model consists of several states, and each state corresponds to a probability function that describes the random observation vector. To train a deep neural network (DNN) that exhibits the probability of observations for each state, a common procedure is to train a conventional GMM-HMM (using a GMM to describe each state) at first and then use the GMM-HMM to provide each observation (feature) in a training set with a state label. After that, supervised learning is conducted on the observations as the input together with the state labels as the desired output to obtain the deep neural networks (DNNs) for the states of the HMM. We often use the softmax activation for the output layer of the DNN and the log-likelihood as the loss function, and thus the resulting DNN can output the state-dependent observation probability, denoted by b j ( o t ) , where j is the state index and o t is the observation at an arbitrary time index t.
  • State-transition probability:
    Each DNN-HMM acoustic model usually corresponds to a single mono-phone or a tri-phone. To find the most likely state sequence in an HMM for an arbitrary input utterance, the Viterbi algorithm is often employed. Let λ refer to the model parameters of a specific HMM, and O = [ o 1 , o 2 , o 3 , , o T ] is the observation sequence corresponding to that HMM. The Viterbi algorithm tries to find the probability:
    δ t ( i ) = max q 1 q 2 q t 1 [ P ( q 1 q 2 q t 1 , q t = i , o 1 o 2 o t | λ ) ]
    for t = 1 , 2 , , T in series and finally find the best state sequence that produces the highest probability P * = max 1 i N [ δ T ( i ) ] . It can be found that:
    δ t + 1 ( j ) = max i [ δ t ( i ) a i j ] b j ( o t + 1 ) ,
    where a i j is the transition probability from the ith state to the jth state. According to Equation (12), δ t + 1 ( j ) considers both the observation probability b j ( o t + 1 ) and state-transition probability a i j of the HMM model.

3. The Presented New Method for Noise-Robust Feature Extraction

We propose a novel method to create noise-robust speech features in ASR. This method enhances the original acoustic features in noise robustness for an ASR system without alternating the acoustic models. In the following, we introduce the proposed method and then discuss its characteristics.

3.1. The Procedures of the Presented Method

In Section 1, we discuss the issue of using the mean squared error (MSE) as the loss function to train the speech enhancement model. The resulting model does not necessarily benefit speech quality and intelligibility in terms of the PESQ and STOI evaluation scores. The underlying reason is the objective mismatch, i.e., minimizing the MSE loss function versus maximizing the evaluation scores, PESQ and STOI. Inspired by this observation, in this study, we present a training algorithm for noise-robust speech feature extraction in which the loss function is directly related to the recognition accuracy. Briefly speaking, the main idea of our proposed method is to find the acoustic speech features o t in noisy environments that maximize the accuracy of state sequences q in HMMs with respect to the clean noise-free condition. The flowchart of the presented method is shown in Figure 2, and it consists of the following steps:
Step 1: 
We compute the FBANK and MFCC features for every utterance in the training set. Then, the features are further processed by mean and variance normalization (MVN), resulting in the final features o t , where t is the frame index.
Step 2: 
With the MFCC features in the training set, the GMM-HMM acoustic models (mono-phones and tri-phones) are trained following the standard recipes of Kaldi [32]. Moreover, linear discriminant analysis (LDA) [33], maximum likelihood linear transform (MLLT) [34], and speaker-adaptive training (SAT) [35] are applied to the speech features during the model training. After training, we obtain the GMM-HMM aligned state sequences q ¯ for each feature sequence o t in the training set, and the state sequence includes mono-phone labels and tri-phone labels.
Step 3: 
With the FBANK features and their state labels for the training set in hand, we train the corresponding DNN-HMM acoustic models [31]. It is a multi-classification task since we aim to find the transformation between the speech features o t and their corresponding state labels q t . The final layer of each DNN can produce the state-dependent observation probability of each o t . It is worth noting that here we adopt multi-condition speech training set in the DNN learning stage. As such, the resulting DNN-HMMs are expected to be more noise-robust than the original GMM-HMMs. With the more accurate state sequences for each frame, combined with the trained HMM model in Step 2, we could obtain the word error rates (WER, %) which are the evaluation scores in our experiment.
Step 4: 
This step is the core of our presented method. Here, we train a denoising deep neural network to transform the original speech feature o t to another representation o t as follows:
o t = f D N ( o t ) ,
where f D N ( . ) denotes the mapping function of the trained network. The objective of the trained network is to make the resulting new features o t predict a state sequence from the DNN-HMM created in Step 3 that is closest to the ground-truth state sequence. We express the underlying idea in the following:
f D N = arg max f ( A c c ( q , q = g ( f ( o t ) | λ ) ) )
where λ is the DNN-HMM acoustic model, g is the function to produce the highest-likelihood state sequence, q , of new features f ( o t ) given the model λ , q is the ground-truth state sequence, and A c c is the log-likelihood function to measure the accuracy of q relative to q.
Once the denoising deep neural network f D N is trained, in the recognition process we use it to map the speech features o t in a noise-corrupted utterance to new features o t . Then, we feed o t to the original (not re-trained) DNN-HMM acoustic models together with language models in order to produce the maximum-likelihood state sequence and the word sequence. Compared with the original features o t , the new features o t are supposed to be more noise-robust since they are created with the help of the multi-condition DNN-HMMs. In addition, the new features o t are supposed to produce lower word error rates in recognition since they correspond to the state sequence with a higher state posterior accuracy relative to the original o t .
Figure 2. The flowchart of proposed method.
Figure 2. The flowchart of proposed method.
Asi 05 00071 g002

3.2. The Characteristics of the Proposed Method

Compared with some DNN-based speech feature extraction methods using MSE as the loss function [36], the newly presented method has the following potential advantages:
  • Our method aims to maximize the state posterior accuracies of multi-condition acoustic models given the noise-corrupted data, which implicitly reduces the word error rates (WERs) for noisy speech recognition. In contrast, those methods minimizing the MSE among the clean–noisy speech feature pairs in the training set have little to do with the subsequent acoustic models and thus are not guaranteed to improve the recognition accuracy for noisy situations.
  • Our method employs the multi-condition training set to obtain GMM-HMM and DNN-HMM acoustic models in turn, which serve to train the denoising feature extraction network. However, we expect such a denoising framework to work well when using clean noise-free training data. The underlying reason is that the framework adapts the original speech features to fit the subsequent acoustic models. In the later section, we observe and discuss the corresponding results from the evaluation experiments.

4. The Evaluation Results and Discussions

This section provides a series of experiments to evaluate the presented new noise-robust speech feature extraction method. The experimental setup is given in Section 4.1, and the recognition results achieved by some well-known speech enhancement methods are given and discussed in Section 4.2. Finally, Section 4.3 and Section 4.4 present and analyze the presented method’s results as well as several comparative algorithms.

4.1. The Experimental Setup

In the following, we describe the used data corpus, the ASR system preparation, and the arrangement of the DNN for our presented method.

4.1.1. Data Corpus

We used the well-known TIMIT corpus [37] to experiment. TIMIT contains phonemically and lexically transcribed speech utterances from American English speakers of different genders and dialects. These utterances correspond to the broadband recordings of 630 speakers of eight major dialects of American English, each speaker reading ten phonetically rich sentences. Each utterance is a 16-bit, 16 kHz sampling rate waveform file accompanied by time-aligned orthographic, phonetic, and word transcriptions. The utterances in TIMIT were used as the training and test sets in our evaluation experiments. As for the multi-condition training set, 1000 utterances were randomly selected and corrupted with any of three noise types (babble, car, and street) at five signal-to-noise ratio (SNR) levels (−5 dB, 0 dB, 5 dB, 10 dB, and 15 dB). For the test set, 400 utterances different from the training set were selected and corrupted with any of three noise types (white, engine, and jackhammer) at six SNR levels (−6 dB, −3 dB, 0 dB, 3 dB, 6 dB, and 12 dB). In the experiments, we tested the utterances with respect to each noise type and each SNR level to see the detailed effects of the evaluated methods.

4.1.2. The ASR System Preparation

The speech features in the training set were used to learn context-dependent (CD) tri-phones as acoustic models, which were arranged to be two different structures, i.e., GMM-HMM and DNN-HMM, respectively. Stating in more detail, GMM-HMM and DNN-HMM used GMM and DNN, respectively, to represent each state of the HMM. As for GMM-HMM, each mono-phone of speech signals and the silence were respectively characterized by an HMM with three states, having 1000 Gaussians in total, and each tri-phone was characterized by an HMM with three states, having 2500 leaves and 1500 Gaussians in total. Furthermore, LDA, MLLT, SAT were applied to the speech features during the tri-phone model training. On the other hand, for the DNN structure, five hidden layers, each containing 1024 nodes with a dropout rate 15 % , were connected to two independent output layers for tri-phones and mono-phones, respectively, in the DNN-HMM. The training process of this DNN underwent 24 epochs with an SGD optimizer and used the log-likelihood loss function for tri-phones and mono-phones. The losses of tri-phones and mono-phones were added up for minimization in the model training. The Kaldi toolkit [32] was used to create the GMM-HMM and the Pytorch-Kaldi [38] toolkit was used to create the DNN-HMM.
In addition, a set of tri-gram language models for training utterances was created via following the standard recipes of Kaldi.

4.1.3. The DNN Structure for Denoising Feature Extraction

For each of the utterances in the training and test set, the 69-dimensional FBANK feature stream (23 static FBANKs together with their delta and delta-delta for each frame, with a 20 ms frame duration and a 10 ms frame shift) was created as the baseline feature representation. The presented denoising DNN framework took the FBANK as input to produce the updated feature following the procedures in Section 3.1 for the subsequent recognition. The denoising DNN model was a convolutional neural network (CNN) with four same-size one-dimensional convolutional layers, each following a configuration setting (30, 5, 2), where the setting representation was “(number of kernel, kernel size, number of padding)”. In addition, these four convolutional layers were followed by the two same fully connected layers with 759 nodes. The activation function for each layer output was the rectified linear unit (ReLU). The training process of this denoising framework underwent 30 epochs with the Adam optimizer and used the log-likelihood loss function.

4.2. The Effect of SE Methods in Speech Recognition for Noisy Speech

In this subsection, we would like to investigate whether the signals pre-enhanced by speech enhancement (SE) methods can result in speech features that bring about improved recognition accuracies relative to the baseline. We selected two SE methods: MMSE [39] (an unsupervised approach) and ideal ratio mask (IRM) [10] (a supervised approach) and conducted the evaluation experiments with a multi-condition mode in two different arrangements:
(1)
The SE process was conducted on the utterances in both the training and test sets.
(2)
The SE process was only conducted on the utterances in the test set, while the training set remained unchanged.
Table 1, Table 2 and Table 3 list the word error rates (WER in %) for the above two arrangements with respect to the test utterances of three noise types (white, engine, and jackhammer). In these tables, MMSE (TR, TS) and IRM (TR, TS) refer to the first arrangement, in which both the training and test sets were enhanced by MMSE/IRM, while MMSE (TS) and IRM (TS) refer to the second arrangement, in which MMSE/IRM enhanced only the test set. Furthermore, to see if the two SE methods can improve the quality and intelligibility of the test utterances as they promise, we also list the PESQ and STOI scores in these tables. From the results shown in these three tables, we have the following observations:
  • Both MMSE and IRM improve the PESQ scores, while MMSE behaves better than IRM in most cases. For example, MMSE obtains 1.647 and 1.967 in PESQ at the 12 dB SNR case of Table 1 and Table 2, while the corresponding values achieved by IRM are 1.275 and 1.493. This is possibly due to the mismatch between the training and test sets for IRM. However, neither MMSE nor IRM can bring a significant STOI upturn for almost all noise scenarios. In addition, IRM results in less STOI degradation than MMSE.
  • Comparing the two first arrangements (“TR, TS” and “TS”) as for MMSE and IRM, we find that conducting the SE methods on both the training and test sets shows inferior recognition performance (higher WERs) than conducting them only on the test set. Furthermore, the first arrangement “TR, TS” even causes worse results than the baseline. For example, the methods “MMSE (TR, TS)” and “MMSE (TS)” obtain 56.5 % and 55.0 % WERs at the 3 dB SNR case of Table 1, which are worse than the baseline 49.8 % . The underlying explanation is that MMSE and IRM, when conducted on the training set, result in an over-smooth trajectory for speech utterances in the time or spectral domains and therefore reduce the diversity and discriminant capability of the various phone realizations in the training set.
  • As for the second arrangement (i.e., MMSE and IRM conducted only on the test set), IRM appears to bring moderately lower WERs for the two noise types “white” and “engine” (shown in the lower part of Table 1 and Table 2), while it is not quite helpful for the “jackhammer” case (shown in the lower part of Table 3). Comparatively, MMSE worsens the recognition accuracy for all three noise situations.
Based on the above observations, we may conclude with the following points:
  • The two metrics, PESQ and WER, seem to have little correlation since the method MMSE, which causes better PESQ scores, results in higher WERs (except for some low SNR cases). That is, improving the speech quality does not necessarily reduce its recognition accuracy. Relative to PESQ, the STOI index has more to do with the recognition accuracy in more noisy situations. This can be observed in that the method IRM can improve STOI and recognition accuracy simultaneously for “white” and “engine” noises, even though the improvement is somewhat marginal. In addition, the increase in STOI does not necessarily reduce WER, while the reduction in WER always comes with the increase in STOI.
  • The two SE methods used here do not show obvious benefits of speech recognition in noisy situations. In our opinion, directly compensating speech features used for a noisy recognition system is probably more helpful in reducing WERs than enhancing speech signals by SE methods.
Table 1. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “white”. The results worse than the baseline are marked with a superscript “*”. In particular, we also give the same marking for the results worse than the baseline in the other tables.
Table 1. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “white”. The results worse than the baseline are marked with a superscript “*”. In particular, we also give the same marking for the results worse than the baseline in the other tables.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline1.0311.0361.0441.0621.0961.271
PESQMMSE1.0401.0541.0801.1301.2371.647
IRM1.0321.0361.0441.0631.0981.275
baseline0.5770.6480.7220.7900.8500.935
STOIMMSE0.5840.643 *0.697 *0.734 *0.766 *0.840 *
IRM0.5780.6490.7230.7920.8520.937
baseline66.162.157.049.844.834.4
MMSE (TR, TS)70.0 *66.7 *61.9 *56.5 *50.5 *40.5 *
WER (%)IRM (TR, TS)67.3 *63.1 *58.3 *52.5 *47.3 *37.5 *
MMSE (TS)70.3 *66.8 *60.4 *55.0 *49.7 *40.1 *
IRM (TS)65.861.956.550.444.434.3
Table 2. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “engine”.
Table 2. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “engine”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline1.0511.0641.0871.1281.1951.482
PESQMMSE1.1031.1031.1981.3501.5201.967
IRM1.0511.0641.0871.1281.1951.493
baseline0.5060.5750.6510.7260.7960.899
STOIMMSE0.5150.5790.638 *0.692 *0.743 *0.850 *
IRM0.5060.5750.6520.7260.7960.901
baseline65.361.755.148.241.531.1
MMSE (TR,TS)68.2 *63.8 *57.8 *51.4 *44.3 *32.6 *
WER (%)IRM (TR,TS)67.5 *64.7 *59.0 *52.9 *46.0 *35.3 *
MMSE (TS)68.3 *63.6 *57.3 *51.4 *44.9 *34.2 *
IRM (TS)65.9 *61.454.848.241.231.1
Table 3. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “jackhammer”.
Table 3. The various evaluation results of MMSE, IRM, and the baseline for the test set in noise environment “jackhammer”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline1.6821.9462.2422.5662.9193.533
PESQMMSE2.0872.4132.7393.0593.3233.812
IRM1.7732.0442.3502.6803.0243.624
baseline0.9240.9500.9680.9810.9890.997
STOIMMSE0.917 *0.944 *0.962 *0.976 *0.983 *0.991 *
IRM0.9320.9550.9700.9820.9900.997
baseline31.427.825.923.823.021.9
MMSE (TR,TS)32.3 *28.7 *26.4 *24.5 *23.9 *22.5 *
WER (%)IRM (TR,TS)33.6 *30.4 *27.4 *25.2 *24.1 *22.9 *
MMSE (TS)32.9 *29.0 *27.2 *25.2 *24.4 *23.2 *
IRM (TS)32.4 *29.8 *26.6 *25.2 *23.8 *22.8 *

4.3. Experimental Results and Discussions for the Proposed Method

In this subsection, we provide the experimental results of our presented method and some discussions. For ease of discussion, the presented method is termed “maximum state posterior probability” with the abbreviation “MSPP”. The experimental results for the multi-condition test utterances are further split into two parts: multi-condition training mode and clean-condition training mode. Note that the denoising framework in our method is with the help of the multi-condition training set, but we would like to see if the resulting enhanced speech features can behave well in both modes.
Furthermore, we employ a comparative method for evaluation which primarily uses a deep neural network (DNN) to update the input FBANK features. This DNN is learned by directly minimizing the mean squared error (MSE) between the noisy and clean FBANK feature pairs in the multi-condition training set. This comparative method is termed feature-based MSE and abbreviated by “FMSE” in the following discussions.
Before going into the experimental results, it is worth noting again that we used the original plain FBANK in the training set to prepare the acoustic models for both multi-condition and clean-condition training modes. Therefore, the various speech enhancement/robust feature extraction methods were just conducted on the test sets.

4.3.1. Multi-Condition Training Mode

Table 4, Table 5 and Table 6 list the WER (%) for our method, MSPP, and FMSE, together with the two SE methods (MMSE and IRM) in the preceding sub-section using the acoustic models learned from the multi-condition training set. The used acoustic models are common to the four comparative methods. Since the presented MSPP is directly related to state posteriors in the acoustic models, here we also list the state error rates with the various methods for the test utterances. From these three tables, we have the following observations:
  • On average, Table 6 shows that the WERs achieved by various methods in the jackhammer noise environment are significantly lower than those in the white and engine noise environments as in Table 4 and Table 5, indicating that jackhammer noise brings less distortion to speech signals compared with the white and engine noises. However, we find that all of the used methods here fail to outperform the baseline results for the jackhammer noise case, indicating that the speech enhancement/noise-robust feature methods might introduce further observable distortion to less contaminated utterances.
  • As for the white and engine noise situations as in Table 4 and Table 5, the newly proposed MSPP can achieve lower WERs in most SNR cases and outperform the other methods in comparison, which clearly validates the main thought of MSPP that increasing the state posteriors helps to improve the noise robustness and promote recognition accuracy. For example, observing the the lower part of Table 4 and Table 5, MSPP achieves 54.9 % and 54.6 % WER at 0 dB SNR for white and engine noise cases, better than the baseline results 57.0 % and 55.1 % . In particular, the denoising framework in the presented MSPP is created upon the noisy dataset in which noise types are neither white nor engine. Thus, MSPP is shown to have a generalization capability to conquer unseen noise.
  • When observing the state error rates achieved by the various methods as shown in the upper parts of Table 4, Table 5 and Table 6, we do not see they strongly correlate with the obtained WERs. For example, the presented MSPP does not always have a lower state error rate, while the method IRM, which has the lowest state error rate in some SNR cases, results in higher WERs. One possible explanation is that the state error rate is quite sensitive to distortion, making it not feasible to be a good evaluation metric. In addition, the presented MSPP attempts to choose the best possible state sequence and is a decision process, which does not necessarily coincide with the minimization of the state error, an estimation process.
  • The method FMSE, which is developed to minimize the mean squared error (MSE) between the clean–noisy FBANK pairs, behaves worse than the baseline in almost all noise situations. For example, in Table 4 and Table 5, FMSE achieves 57.2 % and 56.2 % WER at 0 dB SNR for white and engine noise cases, worse than the baseline results 57.0 % and 55.1 % . These results seem to agree with our earlier discussions in Section 1, i.e., the mismatch in evaluation (optimization) metrics makes the MMSE-wise transformed speech features give rise to worse recognition accuracy. Another reason is that the learned DNN in FMSE overfits the training data and thus fails to enhance the test data well.
Table 4. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “white”.
Table 4. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “white”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9510.9280.8950.8490.8070.717
MSPP0.9490.9250.8950.852 *0.809 *0.723 *
state error rateFMSE0.954 *0.930 *0.899 *0.856 *0.811 *0.729 *
MMSE0.968 *0.952 *0.929 *0.899 *0.867 *0.794 *
IRM0.9500.9250.8910.850 *0.8030.718 *
baseline66.162.157.049.844.834.4
MSPP65.561.054.948.943.234.7 *
WER (%)FMSE69.2 *63.3 *57.2 *50.4 *44.835.3 *
MMSE70.3 *66.8 *60.4 *55.0 *49.7 *40.1 *
IRM65.861.956.550.4 *44.434.3
Table 5. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “engine”.
Table 5. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “engine”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9580.9360.8980.8500.7960.689
MSPP0.9570.9330.899 *0.851 *0.799 *0.694 *
state error rateFMSE0.9560.9320.8920.8400.7820.676
MMSE0.964 *0.946 *0.917 *0.878 *0.832 *0.730 *
IRM0.9580.9360.8980.851 *0.7950.690 *
baseline65.361.755.148.241.531.1
MSPP65.5 *60.254.647.941.7 *32.3 *
WER (%)FMSE70.4 *64.5 *56.2 *48.241.031.2 *
MMSE68.3 *63.6 *57.3 *51.4 *44.9 *34.2 *
IRM65.9 *61.454.848.241.231.1
Table 6. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “jackhammer”.
Table 6. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with multi-condition training for the test set in noise environment “jackhammer”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.6930.6560.6220.6000.5850.568
MSPP0.704 *0.667 *0.638 *0.614 *0.598 *0.579 *
state error rateFMSE0.725 *0.692 *0.663 *0.644 *0.630 *0.618 *
MMSE0.712 *0.678 *0.647 *0.625 *0.610 *0.591 *
IRM0.707 *0.675 *0.645 *0.618 *0.604 *0.585 *
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline31.427.825.923.823.021.9
MSPP32.6 *29.4 *27.5 *25.7 *25.1 *23.9 *
WER (%)FMSE34.4 *31.3 *29.1 *27.6 *27.0 *26.1 *
MMSE32.9 *29.0 *27.2 *25.2 *24.4 *23.2 *
IRM32.4 *29.8 *26.6 *25.2 *23.8 *22.8 *

4.3.2. Clean-Condition Training Mode

In Table 7, Table 8 and Table 9, we present the WER results of the presented MSPP, FMSE, and two SE methods, MMSE and IRM, for the test set under the clean-condition training scenario. Please note that here we use the original FBANK features in the clean training set to train the acoustic models, which are common to those evaluation methods. Several observations can be found in these tables:
  • Compared with the multi-condition baseline results shown in the previous three tables (Table 4, Table 5 and Table 6), the clean-condition baseline results look worse. For example, at 0 dB SNR for the white noise case, the clean-condition baseline WER is 61.0 % (in Table 7), worse than the multi-condition baseline WER 57.0 % (in Table 4). This is probably because the clean-condition training data are more mismatched with the test set than the multi-condition training data.
  • The two SE methods, MMSE and IRM, give similar or higher WERs compared to the baseline results. For example, at 0 dB SNR for the engine noise case (in Table 8), the WERs for the baseline, MMSE, and IRM are 61.1 % , 62.5 % , and 60.6 % , respectively. It again shows that enhancing the utterances does not necessarily improve the respective recognition accuracy.
  • The presented MSPP is shown to provide significantly lower WERs than the baseline results for most noisy cases (except for the jackhammer noise at the SNRs higher than −3 dB). For example, at 0 dB SNR for the white noise case (in Table 7), the WERs for the baseline and MSPP are 61.0 % and 56.3 % , respectively. These results indicate that MSPP promises to promote the recognition accuracy of noisy speech features even when MSPP does not pre-process the training set for the acoustic model. It reconfirms our previous statement: MSPP has a generalization capability to conquer unseen noise.
Table 7. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “white”.
Table 7. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “white”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9680.9530.9280.8980.8640.787
MSPP0.9610.9420.9140.8810.8410.757
state error rateFMSE0.9620.9430.9190.8860.8510.767
MMSE0.977 *0.966 *0.948 *0.927 *0.899 *0.836 *
IRM0.9680.9500.9260.8960.8630.788 *
baseline67.664.661.055.650.941.2
MSPP64.360.656.350.845.436.6
WER (%)FMSE68.6 *64.358.953.047.738.9
MMSE69.9 *67.1 *63.7 *59.1 *54.4 *45.6 *
IRM67.464.360.655.050.540.7
Table 8. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “engine”.
Table 8. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “engine”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9730.9590.9380.9090.8700.790
MSPP0.9670.9480.9190.8810.8320.732
state error rateFMSE0.9660.9460.9190.8820.8320.734
MMSE0.977 *0.965 *0.947 *0.917 *0.882 *0.799 *
IRM0.974 *0.9580.9370.9060.8700.786
baseline67.364.761.156.350.439.4
MSPP64.460.955.249.843.834.7
WER (%)FMSE69.9 *65.5 *59.752.346.836.6
MMSE69.1 *65.9 *62.5 *57.3 *52.1 *40.7 *
IRM66.865.1 *60.655.749.739.4
Table 9. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “jackhammer”.
Table 9. The state error rate and WER (%) of the baseline, MSPP, FMSE, MMSE, and IRM with clean-condition training for the test set in noise environment “jackhammer”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.7390.7010.6770.6380.6120.576
MSPP0.7200.6820.6530.6250.6060.583 *
state error rateFMSE0.7330.6920.6600.6330.614 *0.587 *
MMSE0.746 *0.710 *0.6760.645 *0.621 *0.585 *
IRM0.7280.6920.6600.6310.6100.575
baseline35.331.528.526.724.923.1
MSPP33.830.628.7 *27.0 *26.3 *24.9 *
WER (%)FMSE35.032.2 *29.7 *28.1 *27.0 *25.4 *
MMSE36.5 *32.8 *29.6 *27.7 *25.3 *23.7 *
IRM34.831.8 *28.8 *26.624.923.3 *

4.4. Experimental Results and Discussions for the Proposed Method with Data Augmentation

In this subsection, we provide the experimental results of a variant of our presented method, which further adopted the data augmentation technique [40]. To realize it, the DNN-HMM acoustic models were retrained with the original FBANKs and MSPP-enhanced FBANKs in the multi-condition training set (the original acoustic models were trained with the original FBANKs only). In particular, the training set here was arranged such that 50% of the utterances were converted to plain FBANKs while the remaining 50% corresponded to MSPP-enhanced FBANKs. The resulting acoustic models were evaluated on MSPP-enhanced FBANKs in the noisy test set to see how the recognition accuracy would be influenced.
In Table 10, Table 11 and Table 12, we present the state error rates and WER (%) under the multi-condition training scenario with respect to the aforementioned two acoustic model configurations. Here, the configuration using data augmentation is termed “MSPP-Aug”. From the three tables, we see that with data augmentation, the resulting acoustic models can further reduce the WERs and the state error rates for the MSPP-enhanced test set compared with the original FBANK-wise acoustic models. The WER reduction can be as high as 1% for some SNR cases for the noise types “white” and “engine”. For example, at 0 dB SNR for the engine noise case (in Table 11), the WERs for the baseline, MSPP, and MSPP-Aug are 55.1 % , 54.6 % , and 53.0 % , respectively. These results clearly reveal the effectiveness of the data augmentation technique, which increases the diversity of training data and thus benefits the noise robustness of the resulting acoustic models.

5. Conclusions and Future Work

In this study, we primarily pay attention to the noise issue in automatic speech recognition (ASR). We analyze the possible effects of speech enhancement methods on acoustic feature robustness in ASR and then present a novel deep-learning-based framework to create noise-robust speech features. This framework exploits a deep neural network to maximize the state posteriors of the used acoustic models which provided the clean–noisy FBANK speech feature pairs in the training set. The preliminary experiments reveal that the newly presented method can improve the recognition accuracy of FBANK features, particularly under the moderate and severe noise situations. It can behave well regardless of the training set being mismatched multi-condition or clean noise-free condition that prepares the acoustic models. As for future avenues, we could further enhance the presented denoising network by adopting more versatile training data or increasing the amount of training data. We also plan to integrate this method with some other feature-based or model-based noise-robust algorithms to achieve even better performance.

Author Contributions

Conceptualization, L.-C.C. and J.-W.H.; methodology, L.-C.C. and J.-W.H.; software, L.-C.C.; validation, L.-C.C. and J.-W.H.; formal analysis, L.-C.C. and J.-W.H.; investigation, L.-C.C. and J.-W.H.; resources, J.-W.H.; data curation, L.-C.C. and J.-W.H.; writing—original draft preparation, L.-C.C. and J.-W.H.; writing—review and editing, J.-W.H.; visualization, L.-C.C. and J.-W.H.; supervision, J.-W.H.; project administration, J.-W.H.; funding acquisition, J.-W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Science Council (NSC) in Taiwan with grant number 110-2221-E-260-012.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Loizou, P.C. Speech Enhancement: Theory and Practice, 2nd ed.; CRC Press, Inc.: Boca Raton, FL, USA, 2013. [Google Scholar]
  2. Weiss, M.R.; Aschkenasy, E.; Parsons, T.W. Study and Development of the Intel Technique for Improving Speech Intelligibility; Final Technical Report; Nicolet Scientific Corp.: Northvale, NJ, USA, 1975. [Google Scholar]
  3. Boll, S. Suppression of acoustic noise in speech using spectral subtraction. IEEE Trans. Acoust. Speech Signal Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  4. Wiener, N. Extrapolation, Interpolation, and Smoothing of Stationary Time Series; The MIT Press: Cambridge, MA, USA, 1964. [Google Scholar]
  5. Kjems, U.; Jensen, J. Maximum likelihood based noise covariance matrix estimation for multi-microphone speech enhancement. In Proceedings of the 2012 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 295–299. [Google Scholar]
  6. Su, Y.; Tsao, Y.; Wu, J.; Jean, F. Speech enhancement using generalized maximum a posteriori spectral amplitude estimator. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 7467–7471. [Google Scholar]
  7. Jiang, H.; Hirose, K.; Huo, Q. Robust speech recognition based on a Bayesian prediction approach. IEEE Trans. Speech Audio Process. 1999, 7, 426–440. [Google Scholar] [CrossRef]
  8. Dendrinos, M.; Bakamidis, S.; Carayannis, G. Speech enhancement from noise: A regenerative approach. Speech Commun. 1991, 10, 45–57. [Google Scholar] [CrossRef]
  9. Ephraim, Y.; Van Trees, H.L. A signal subspace approach for speech enhancement. In Proceedings of the 1993 IEEE International Conference on Acoustics, Speech, and Signal Processing, Minneapolis, MN, USA, 27–30 April 1993; Volume 2, pp. 355–358. [Google Scholar]
  10. Narayanan, A.; Wang, D. Ideal ratio mask estimation using deep neural networks for robust speech recognition. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 7092–7096. [Google Scholar]
  11. Kjems, U.; Boldt, J.B.; Pedersen, M.S.; Lunner, T.; Wang, D. Role of mask pattern in intelligibility of ideal binary-masked noisy speech. J. Acoust. Soc. Am. 2009, 126, 1415–1426. [Google Scholar] [CrossRef] [PubMed]
  12. Kim, G.; Loizou, P.C. Improving Speech Intelligibility in Noise Using a Binary Mask That Is Based on Magnitude Spectrum Constraints. IEEE Signal Process. Lett. 2010, 17, 1010–1013. [Google Scholar] [CrossRef] [Green Version]
  13. Rix, A.; Beerends, J.; Hollier, M.; Hekstra, A. Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In Proceedings of the 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing, Salt Lake City, UT, USA, 7–11 May 2001; Volume 2, pp. 749–752. [Google Scholar] [CrossRef]
  14. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. An Algorithm for Intelligibility Prediction of Time–Frequency Weighted Noisy Speech. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2125–2136. [Google Scholar] [CrossRef]
  15. Fu, S.; Liao, C.; Tsao, Y.; Lin, S. MetricGAN: Generative Adversarial Networks based Black-box Metric Scores Optimization for Speech Enhancement. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  16. Fu, S.; Yu, C.; Hsieh, T.; Plantinga, P.; Ravanelli, M.; Lu, X.; Tsao, Y. MetricGAN+: An Improved Version of MetricGAN for Speech Enhancement. arXiv 2021, arXiv:2104.03538. [Google Scholar]
  17. Hu, Y.; Loizou, P.C. Evaluation of Objective Quality Measures for Speech Enhancement. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 229–238. [Google Scholar] [CrossRef]
  18. Zhang, H.; Zhang, X.; Gao, G. Training Supervised Speech Separation System to Improve STOI and PESQ Directly. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5374–5378. [Google Scholar]
  19. Xu, Y.; Du, J.; Dai, L.; Lee, C. An Experimental Study on Speech Enhancement Based on Deep Neural Networks. IEEE Signal Process. Lett. 2014, 21, 65–68. [Google Scholar] [CrossRef]
  20. Lu, X.; Tsao, Y.; Matsuda, S.; Hori, C. Speech enhancement based on deep denoising Auto-Encoder. Proc. Interspeech 2013, 2013, 436–440. [Google Scholar]
  21. Xu, Y.; Du, J.; Dai, L.; Lee, C. A Regression Approach to Speech Enhancement Based on Deep Neural Networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 7–19. [Google Scholar] [CrossRef]
  22. Xia, B.; Bao, C.C. Wiener filtering based speech enhancement with Weighted Denoising Auto-encoder and noise classification. Speech Commun. 2014, 60, 13–29. [Google Scholar] [CrossRef]
  23. Graves, A.; Jaitly, N. Towards End-To-End Speech Recognition with Recurrent Neural Networks. In Proceedings of the 31st International Conference on Machine Learning, Bejing, China, 22–24 June 2014; Xing, E.P., Jebara, T., Eds.; PMLR: Bejing, China, 2014; Volume 32, pp. 1764–1772. [Google Scholar]
  24. Hannun, A.; Case, C.; Casper, J.; Catanzaro, B.; Diamos, G.; Elsen, E.; Prenger, R.; Satheesh, S.; Sengupta, S.; Coates, A.; et al. Deep Speech: Scaling up end-to-end speech recognition. arXiv 2014, arXiv:1412.5567. [Google Scholar] [CrossRef]
  25. Chorowski, J.; Bahdanau, D.; Cho, K.; Bengio, Y. End-to-end Continuous Speech Recognition using Attention-based Recurrent NN: First Results. arXiv 2014, arXiv:1412.1602. [Google Scholar] [CrossRef]
  26. Menne, T.; Schlüter, R.; Ney, H. Investigation into Joint Optimization of Single Channel Speech Enhancement and Acoustic Modeling for Robust ASR. In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6660–6664. [Google Scholar] [CrossRef]
  27. Nicolson, A.; Paliwal, K. Deep Learning for Minimum Mean-Square Error Approaches to Speech Enhancement. Speech Commun. 2019, 111, 44–55. [Google Scholar] [CrossRef]
  28. Nicolson, A.; Paliwal, K.K. Deep Xi as a Front-End for Robust Automatic Speech Recognition. In Proceedings of the 2020 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE), Gold Coast, Australia, 16–18 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  29. Mitra, S. Digital Signal Processing: A Computer-Based Approach; McGraw-Hill International Edition Electrical Engineering Series; McGraw-Hill/Irwin: New York, NY, USA, 2001. [Google Scholar]
  30. Yu, D.; Deng, L. Automatic Speech Recognition: A Deep Learning Approach; Springer: London, UK, 2015. [Google Scholar]
  31. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.E.; Mohamed, A.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.N.; et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  32. Povey, D.; Ghoshal, A.; Boulianne, G.; Burget, L.; Glembek, O.; Goel, N.; Hannemann, M.; Motlicek, P.; Qian, Y.; Schwarz, P.; et al. The Kaldi Speech Recognition Toolkit. In Proceedings of the IEEE 2011 Workshop on Automatic Speech Recognition and Understanding, Big Island, HI, USA, 11–15 December 2011. [Google Scholar]
  33. Haeb-Umbach, R.; Ney, H. Linear discriminant analysis for improved large vocabulary continuous speech recognition. In Proceedings of the ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing, San Francisco, CA, USA, 23–26 March 1992; Volume 1, pp. 13–16. [Google Scholar]
  34. Gales, M. Maximum likelihood linear transformations for HMM-based speech recognition. Comput. Speech Lang. 1998, 12, 75–98. [Google Scholar] [CrossRef] [Green Version]
  35. Anastasakos, T.; McDonough, J.; Schwartz, R.; Makhoul, J. A compact model for speaker-adaptive training. In Proceedings of the Fourth International Conference on Spoken Language Processing, ICSLP ’96, Philadelphia, PA, USA, 3–6 October 1996; Volume 2, pp. 1137–1140. [Google Scholar]
  36. Han, K.; He, Y.; Bagchi, D.; Fosler-Lussier, E.; Wang, D. Deep neural network based spectral feature mapping for robust speech recognition. In Proceedings of the 16th Annual Conference of the International Speech Communication Association, Dresden, Germany, 6–10 September 2015. [Google Scholar]
  37. Garofolo, J.; Lamel, L.; Fisher, W.; Fiscus, J.; Pallett, D. DARPA TIMIT acoustic-phonetic continous speech corpus CD-ROM. NIST speech disc 1-1.1. NASA STI/Recon Tech. Rep. N 1993, 93, 27403. [Google Scholar]
  38. Ravanelli, M.; Parcollet, T.; Bengio, Y. The PyTorch-Kaldi Speech Recognition Toolkit. arXiv 2018, arXiv:eess.AS/1811.07453. [Google Scholar]
  39. Ephraim, Y.; Malah, D. Speech Enhancement Using a Minimum Mean-Square Error Log-Spectral Amplitude Estimator. IEEE Trans. Acoust. Speech Signal Process. 1985, 33, 443–445. [Google Scholar] [CrossRef]
  40. Van Dyk, D.A.; Meng, X.L. The art of data augmentation. J. Comput. Graph. Stat. 2001, 10, 1–50. [Google Scholar] [CrossRef]
Figure 1. Flow chart of MFCC calculation.
Figure 1. Flow chart of MFCC calculation.
Asi 05 00071 g001
Table 10. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “white”.
Table 10. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “white”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9510.9280.8950.8490.8070.717
state error rateMSPP0.9490.9250.8950.852 *0.809 *0.723 *
MSPP-Aug0.9480.9230.8920.851 *0.808 *0.720 *
baseline66.162.157.049.844.834.4
WER (%)MSPP65.561.054.948.943.234.7 *
MSPP-Aug64.359.954.048.442.734.0
Table 11. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “engine”.
Table 11. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “engine”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.9580.9360.8980.8500.7960.689
state error rateMSPP0.9570.9330.899 *0.851 *0.799 *0.694 *
MSPP-Aug0.9550.9300.8940.8460.7930.690 *
baseline65.361.755.148.241.531.1
WER (%)MSPP65.5 *60.254.647.941.7 *32.3 *
MSPP-Aug64.260.053.047.041.131.8 *
Table 12. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “jackhammer”.
Table 12. The state error rate and WER (%) of the baseline, MSPP, and MSPP-Aug with multi-condition training for the test set in noise environment “jackhammer”.
Signal-To-Noise Ratio (SNR)
−6 dB−3 dB0 dB3 dB6 dB12 dB
baseline0.6930.6560.6220.6000.5850.568
state error rateMSPP0.704 *0.667 *0.638 *0.614 *0.598 *0.579 *
MSPP-Aug0.694 *0.659 *0.627 *0.608 *0.595 *0.578 *
baseline31.427.825.923.823.021.9
WER (%)MSPP32.6 *29.4 *27.5 *25.7 *25.1 *23.9 *
MSPP-Aug31.9 *29.4 *26.9 *25.2 *24.2 *23.1 *
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chang, L.-C.; Hung, J.-W. A Preliminary Study of Robust Speech Feature Extraction Based on Maximizing the Probability of States in Deep Acoustic Models. Appl. Syst. Innov. 2022, 5, 71. https://doi.org/10.3390/asi5040071

AMA Style

Chang L-C, Hung J-W. A Preliminary Study of Robust Speech Feature Extraction Based on Maximizing the Probability of States in Deep Acoustic Models. Applied System Innovation. 2022; 5(4):71. https://doi.org/10.3390/asi5040071

Chicago/Turabian Style

Chang, Li-Chia, and Jeih-Weih Hung. 2022. "A Preliminary Study of Robust Speech Feature Extraction Based on Maximizing the Probability of States in Deep Acoustic Models" Applied System Innovation 5, no. 4: 71. https://doi.org/10.3390/asi5040071

Article Metrics

Back to TopTop