Next Article in Journal
The Influence of Air Pressure on the Dynamics of Flexural Ultrasonic Transducers
Next Article in Special Issue
Multimodal Speaker Diarization Using a Pre-Trained Audio-Visual Synchronization Model
Previous Article in Journal
Multiple Physiological Signals Fusion Techniques for Improving Heartbeat Detection: A Review
Previous Article in Special Issue
Forensic Speaker Verification Using Ordinary Least Squares
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarially Learned Total Variability Embedding for Speaker Recognition with Random Digit Strings

Department of Electrical and Computer Engineering and the Institute of New Media and Communications, Seoul National University, Seoul 08826, Korea
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(21), 4709; https://doi.org/10.3390/s19214709
Submission received: 30 September 2019 / Revised: 23 October 2019 / Accepted: 28 October 2019 / Published: 30 October 2019

Abstract

:
Over the recent years, various research has been conducted to investigate methods for verifying users with a short randomized pass-phrase due to the increasing demand for voice-based authentication systems. In this paper, we propose a novel technique for extracting an i-vector-like feature based on an adversarially learned inference (ALI) model which summarizes the variability within the Gaussian mixture model (GMM) distribution through a nonlinear process. Analogous to the previously proposed variational autoencoder (VAE)-based feature extractor, the proposed ALI-based model is trained to generate the GMM supervector according to the maximum likelihood criterion given the Baum–Welch statistics of the input utterance. However, to prevent the potential loss of information caused by the Kullback–Leibler divergence (KL divergence) regularization adopted in the VAE-based model training, the newly proposed ALI-based feature extractor exploits a joint discriminator to ensure that the generated latent variable and the GMM supervector are more realistic. The proposed framework is compared with the conventional i-vector and VAE-based methods using the TIDIGITS dataset. Experimental results show that the proposed method can represent the uncertainty caused by the short duration better than the VAE-based method. Furthermore, the proposed approach has shown great performance when applied in association with the standard i-vector framework.

1. Introduction

Over the last decade, the i-vector framework [1,2] has become one of the most dominant approaches for feature extraction in text-independent speaker recognition. The widespread popularity of the i-vector framework in the speaker recognition community can be attributed to its ability to map the distributive pattern of speech with various duration to a fixed dimensional vector. The i-vector framework, along with many other classical utterance-level speech representation methods (e.g., eigenvoice adaptation and joint factor analysis) [3,4], focuses on efficiently compressing the Gaussian mixture model (GMM) supervector, which is formed by concatenating the mean vector of each mixture component [5], via linear factorization. However, due to its linear nature, the i-vector framework may not describe the whole variability of the given speech utterance.
To represent the variability which could not be expressed through linear processing, various attempts have been made utilizing deep neural networks (DNNs) in learning the nonlinear relationship between the input and the output for feature extraction. Most recent works in deep learning-based feature extraction have focused on nonlinearly modeling the frame-level characteristics in a similar manner to the DNN-based acoustic modeling in automatic speech recognition (ASR) [6].
In [7], a DNN for frame-level speaker classification is trained and the activations of the last hidden layer are taken as a nonlinear speaker representation. In [8,9], a TDNN (time-delay neural network)-based utterance embedding technique is proposed, where the embedding is obtained by aggregating the frame-level activations of the TDNN. In [10,11], the speaker embedding neural networks are optimized to directly minimize the verification loss in an end-to-end fashion. Although the conventional deep learning-based feature extraction schemes have shown the potential of using nonlinearly extracted features in speaker recognition, they are usually trained in a supervised manner and it is almost impossible to apply them when no labeled data are available for training.
Recently, verifying users with randomized pass-phrase with constrained vocabulary has become an important task due to the increasing demand for voice-based authentication systems [12]. In this technique, the users are enrolled and verified by speaking random digit strings. One of the obstacles in random digit string task is the feature uncertainty caused by the short duration of the spoken utterance [13]. Since a critical amount of speaker relevant information is contained in the phonetic characteristics, the lack of phonetically informative frames can lead to degradation of the quality of the extracted utterance-level feature. It has been reported that the conventional i-vector framework suffers from severe performance deterioration when short duration speech is applied in the verification process [14,15,16]. In real-life applications, this problem should be considered more carefully as the speech samples for enrollment and trial are required to be short.
In our previous work [17], we successfully exploited the variational autoencoder (VAE) framework [18,19] to extract an utterance-level feature to capture the total variability and the uncertainty of the speech in a nonlinear fashion. The VAE has an autoencoder-like structure and assumes that the input data are generated through a directed graphical model induced by a latent variable. The latent variable of the VAE-based feature extraction model in [17] serves as an utterance-level representation and has shown better performance than the conventional i-vector in coping with the uncertainty caused by the short duration. Moreover, we could observe that the variance of the latent variable can be used as a proxy for the uncertainty caused by the short duration of the given speech samples. However, as mentioned in [20,21], since the Kullback–Leibler divergence (KL divergence) regularization term in the VAE objective function focuses on stretching the latent variable space over the entire training set to avoid assigning a small probability to any samples, the VAE output often lacks variety. This may lead the VAE-based feature extractor to yield utterance-level features which lack distinct speaker-dependent information, resulting in limited performance of the overall speaker recognition system.
In order to overcome this problem, we propose a novel approach to train a deep learning-based feature extraction network using the adversarial learning framework. The proposed method adopts an adversarially learned inference (ALI) model [22] to nonlinearly express the total variability and uncertainty of the given speech. Analogous to the VAE-based feature extractor in [17], the proposed model is trained according to the maximum likelihood criterion and the latent variable serves as the utterance-level feature representation. However, instead of simply regularizing the latent variable via KL divergence, the proposed method uses a discriminator network to make sure that the generated latent variable and GMM supervector are close to the latent prior distribution and the GMM obtained through maximum a posteriori (MAP) adaptation, respectively. While training the proposed network, the parameters of the feature extraction and the discriminator networks are updated competitively; the feature extraction network tries to generate a more realistic GMM supervector and latent variable while the discriminator network focuses on distinguishing the generated GMM and latent variable from the real ones. By training the feature extraction network in this adversarial fashion, the proposed system is expected to provide utterance-level features that can capture more prominent speaker relevant characteristics and the uncertainty within the given speech utterance. Similarly to the VAE-based feature extractor, the utterance-level features extracted from the proposed model can substitute the conventional i-vector extraction module without any difficulty.
To evaluate the performance of the proposed method, along with the VAE-based feature extraction scheme and the conventional i-vector framework in the random digits task, we conducted a set of speaker verification experiments using the TIDIGITS dataset. Experimental results show that the proposed method outperforms the standard i-vector framework and the VAE-based method in terms of equal error rate (EER). It is also interesting to see that impressive performance improvement is observed when the features extracted from the proposed method and the conventional i-vector are augmented together. This implies that the feature extracted using the proposed method is complementary to that extracted from the conventional i-vector framework.
The rest of this paper is organized as follows: We first briefly describe the basic ideas and important formulations of the conventional i-vector and the ALI frameworks in Section 2. In Section 3, the newly proposed ALI-based feature extractor and its training scheme are presented. The experiments and results are shown in Section 4. Finally, Section 5 concludes the paper.

2. Related Work

2.1. I-Vector Framework

Once a universal background model (UBM) is provided, which is a GMM representing the utterance-independent distribution of the frame-level features, an utterance-dependent GMM can be attained by adjusting the UBM parameters using the maximum a posteriori (MAP) adaptation algorithm [23]. By concatenating the mean vector of each Gaussian mixture component, a GMM supervector can be obtained, which summarizes the overall pattern of the frame-level feature distribution. However, using the GMM supervector as an utterance-level feature may limit the overall speaker recognition performance due to its high dimensionality.
To solve this problem, various methods for reducing the dimensionality of the GMM supervector have been proposed [1,3,4]. Specifically, the i-vector framework is now widely used to represent the idiosyncratic characteristics of the utterance in speaker and language recognition [24]. The extraction of an i-vector can be viewed as a factorization process decomposing the GMM supervector as
m ( X ) = u + T w ( X )
where m ( X ) , u , T , and w ( X ) indicate the ideal GMM supervector dependent on a given speech utterance X , UBM supervector, total variability matrix, and i-vector, respectively. The i-vector framework aims to find the optimal w ( X ) and T to adapt the UBM parameters to a given speech utterance X . Given X , the zeroth- and the first-order Baum–Welch statistics are obtained as
n c ( X ) = l = 1 L γ l ( c )
f ˜ c ( X ) = l = 1 L γ l ( c ) ( x l u c )
where, for each frame l within X with L frames, γ l ( c ) is the posterior probability that the lth frame feature x l is aligned to the cth mixture component of the UBM, u c is the mean vector of the cth mixture component of the UBM, and n c ( X ) and f ˜ c ( X ) are the zeroth- and the centralized first-order Baum–Welch statistics, respectively.
As shown in Figure 1, the i-vector framework can be considered as an adaptation process where the mean of each UBM mixture component is adjusted to maximize the likelihood with respect to a given utterance, and the estimated i-vector is served as the adaptation factor. Let Σ c denote the covariance matrix of the cth UBM mixture component and F be the dimensionality of the frame-level features. The posterior probability γ l ( c ) is computed as follows:
γ l ( c ) = π c N ( x l | u c , Σ c ) i = 1 C π i N ( x l | u i , Σ i )
where π c is the weight of the cth Gaussian mixture of the UBM. The log-likelihood given an utterance X conditioned on w ( X ) can be obtained as
log P ( X | T , w ( X ) ) = c = 1 C ( n c ( X ) log 1 ( 2 π ) F / 2 | Σ c | 1 / 2 1 2 l = 1 L γ l ( c ) ( x l m c ( X ) ) t Σ c 1 ( x l m c ( X ) ) )
where m c ( X ) is the mean of the cth mixture component of m ( X ) and the superscript t indicates matrix transpose. The log-likelihood given X is obtained by marginalizing (4) over w ( X ) as
log P ( X | T ) = log E w [ P ( X | T , w ) ] = log P ( X | T , w ) N ( w | 0 , I ) d w
The total variability matrix T is trained to maximize the log-likelihood (5) via the expectation– maximization (EM) algorithm. Interested readers are encouraged to refer to [1,2] for further details of the i-vector framework.

2.2. Adversarially Learned Inference

Like other variants of the generative adversarial network (GAN) [25], ALI aims to train a network for generating a realistic data sample with the help of a discriminator network, which tries to predict whether the input data are real or generated [22]. However, unlike the ordinary GAN framework that cannot analyze the data at an abstract level, ALI integrates an inference network for estimating the random latent variable z from the input data. As depicted in Figure 2, ALI is composed of three directed networks: encoder, decoder, and joint discriminator networks. Analogous to the VAE network, the encoder network outputs the mean and variance of the posterior distribution p ( z | x ) given an observation x . On the other hand, the decoder generates the data sample from a latent variable sampled from a prior distribution p ( z ) .
The encoder and decoder networks represent the joint probability distributions of the latent variable z and the observed data x as follows:
q ϕ ( x , z ) = q ϕ ( x ) q ϕ ( z | x ) , p θ ( x , z ) = p θ ( z ) p θ ( x | z )
In (6), ϕ , θ , q ϕ ( x , z ) , and p θ ( x , z ) denote the encoder parameters, decoder parameters, and the joint distributions represented by the encoder and decoder, respectively. The encoder marginal probability q ϕ ( x ) represents the real data distribution and the decoder marginal probability p θ ( z ) is the prior distribution of the latent variable, usually specified as a standard normal distribution. The conditional probability distributions q ϕ ( z | x ) and p θ ( x | z ) represent the inferred latent distribution output by the encoder and the distribution of the data generated by the decoder, respectively.
The joint discriminator takes a joint pair of the data and the latent variable as input and distinguishes between the samples from the encoder ( x , z ^ ( x ) ) q ϕ ( x , z ) and the ones from the decoder ( x ˜ ( z ) , z ) p θ ( x , z ) . The discriminator output D ( x , z ) is sigmoidal, ideally having a value close to 0 if the samples are drawn from p θ ( x , z ) and 1 if drawn from q ϕ ( x , z ) .
The encoder, decoder, and the joint discriminator networks of the ALI are trained adversarially by minimizing the following objective functions:
E D = E ( x , z ) q ϕ ( x , z ) [ log ( D ( x , z ) ) ] E ( x , z ) p θ ( x , z ) [ log ( 1 D ( x , z ) ) ]
E G = E ( x , z ) q ϕ ( x , z ) [ log ( 1 D ( x , z ) ) ] E ( x , z ) p θ ( x , z ) [ log ( D ( x , z ) ) ]
In Equations (7) and (8), E D denotes the discriminator loss function and E G is the generator loss function. The joint discriminator network is trained to minimize E D , which decreases as the network distinguishes between the samples from the encoder and the decoder better and the parameters θ , ϕ of the generator (i.e., encoder and decoder networks) are updated to minimize E G .

3. Adversarially Learned Feature Extraction

The proposed algorithm assumes that the variability of the utterance-dependent GMM supervector of the UBM can be represented by a nonlinear projection of a hidden variable onto the variability space as follows:
m ( X ) = u + g ( z ( X ) )
where g is a nonlinear function which transforms the hidden variable z ( X ) to the total variability of the GMM supervector m ( X ) . As shown in Figure 3, the proposed scheme employs an encoder network for inferring the hidden variable z ( X ) from the observed speech X and a decoder network for nonlinearly mapping z ( X ) onto the total variability space to generate an ideal utterance-dependent GMM supervector.
As with the i-vector adaptation framework, the main task of the proposed approach is to generate an utterance-dependent GMM that maximizes the likelihood given the Baum-Welch statistics of the utterance.The encoder network takes the zeroth- and first-order Baum-Welch statistics of the input utterance X and estimates the posterior distribution, namely the mean μ and the variance σ 2 of the latent variable through variational inference. The latent variable z is a random variable assumed to follow a Gaussian distribution and its components are uncorrelated with each other. The decoder network generates the GMM supervector given the latent variable according to the maximum likelihood criterion.

3.1. Maximum Likelihood Criterion

Analogous to (4), the log-likelihood of the GMM supervector m ^ ( X ) generated from the decoder network conditioned on the latent variable z ( X ) , which is sampled from the posterior distribution q ϕ ( z | X ) generated by the encoder network given input observation X , can be written as
log P ( X | ϕ , θ , z ( X ) ) = c = 1 C ( n c ( X ) log 1 ( 2 π ) F / 2 | Σ c | 1 / 2 1 2 l = 1 L γ l ( c ) ( x l m ^ c ( X ) ) t Σ c 1 ( x l m ^ c ( X ) ) )
where m ^ c ( X ) denotes the cth component of m ^ ( X ) . Instead of directly maximizing the marginal log-likelihood log P ( X | ϕ , θ ) = log E z ( X ) q ϕ ( z | X ) [ P ( X | ϕ , θ , z ( X ) ) ] , which is practically intractable, the proposed algorithm maximizes its lower bound which can be obtained via Jensen’s inequality as follows:
log P ( X | ϕ , θ ) = log E z ( X ) q ϕ ( z | X ) [ P ( X | ϕ , θ , z ( X ) ) ] E z ( X ) q ϕ ( z | X ) [ log P ( X | ϕ , θ , z ( X ) ) ]
Using the reparameterization trick [18], the Monte Carlo estimate of the marginal log-likelihood lower bound can be computed as
E z ( X ) q ϕ ( z | X ) [ log P ( X | ϕ , θ , z ( X ) ) ] 1 S s = 1 S log P ( X | ϕ , θ , z s ( X ) )
where S is the number of samples used for the estimation, and z s ( X ) is the reparameterized latent variable defined by
z s ( X ) = μ ( X ) + σ ( X ) ϵ s
In (13), μ ( X ) and σ ( X ) are respectively the mean and standard deviation of the latent variable z ( X ) generated from the encoder network, and ϵ s N ( 0 , I ) is an auxiliary noise variable.

3.2. Adversarially Learned Inference for Nonlinear I-Vector Extraction

In order to ensure that the generated latent variable z matches its prior distribution and the GMM supervector m ^ well preserves the distributive structure of the GMM driven from the UBM, the proposed scheme utilizes a joint discriminator for regularizing the encoder and decoder parameters. As shown in Figure 4, the joint discriminator of the proposed algorithm takes the GMM supervector and the latent variable and tries to determine whether the input pairs are generated from the encoder or the decoder networks.
However, since the decoder output m ^ ( z ) does not match the encoder inputs n ( X ) and f ˜ ( X ) , the joint discriminator cannot be applied directly. To alleviate this difficulty, the proposed scheme first estimates the GMM mean vectors via maximum a posteriori (MAP) adaptation [23] given as follows:
m c , M A P ( X ) = l = 1 L γ l ( c ) x l l = 1 L γ l ( c ) = f ˜ c ( X ) n c ( X ) + u c
where m c , M A P ( X ) is the estimated cth Gaussian mixture mean given the input speech utterance X . The mean vectors m c , M A P ( X ) for c = 1 , . . . , C are concatenated to form a GMM supervector m M A P ( X ) .
The joint discriminator takes the joint pair either from the encoder ( m M A P ( X ) , z ^ ) q ϕ ( m , z ) or from the decoder ( m ^ ( z s a m p ) , z s a m p ) p θ ( m , z ) as input, where z ^ and z s a m p are the latent variables sampled from N ( μ ( X ) , log σ 2 ( X ) ) and p θ ( z ) , respectively, and m ^ ( z s a m p ) is the GMM supervector generated by the decoder given z s a m p . We assume that the prior distribution for the latent variable p θ ( z ) to be N ( z | 0 , I ) , akin to the prior for w in the i-vector framework. As in (7), the discriminator network parameter is trained to minimize the following objective function:
E P r o p , D = E ( m , z ) q ϕ ( m , z ) [ log ( D ( m , z ) ) ] E ( m , z ) p θ ( m , z ) [ log ( 1 D ( m , z ) ) ]
By combining the generator loss function of ALI (8) and the marginal log-likelihood lower bound (12), the objective function of the encoder and decoder networks of the proposed framework can be written as
E P r o p , G = E ( m , z ) q ϕ ( m , z ) [ log ( 1 D ( m , z ) ) ] E ( m , z ) p θ ( m , z ) [ log ( D ( m , z ) ) ] 1 S s = 1 S log P ( X | ϕ , θ , z s ( X ) )
From (16), it is seen that the encoder and decoder networks are trained not only to generate latent variables and GMM supervectors from q ϕ ( m , z ) or p θ ( m , z ) that are identical to each other, but also to maximize the log-likelihood of the generated utterance-dependent GMM by minimizing E P r o p , G through error back-propagation [26].
The proposed adversarially learned feature extraction framework is trained by alternating between the optimization of the generator (i.e., encoder and decoder network) and the discriminator every epoch. Once the networks are trained, the encoder network is used as an feature extractor, where the latent variable mean and log-variance vectors are used as nonlinear utterance-level features.

3.3. Relationship to the VAE-Based Feature Extractor

The VAE-based feature extraction network [17] focuses on maximizing the log-likelihood of the generated GMM by minimizing the following objective function:
E V A E / F E ( X ) = D KL ( q ϕ ( z | X ) p θ ( z ) ) 1 S s = 1 S log P ( X | ϕ , θ , z s ( X ) )
The first term in the RHS of (17) is the KL divergence between the prior and the posterior distribution of the latent variable z , which can be viewed as the regularization term. The regularization term forces the encoder network to generate a latent variable distribution that is compatible with p θ ( z ) . However, the KL regularization term stretches the latent space over the entire training set to avoid assigning small probability to any training samples [20,21]. Due to such problem of the KL regularization term, the VAE tends to generate conservative outputs, which usually lack in variety. Especially in the image processing community, it has been reported that the VAE-based image generators result in blurry image samples [22]. In the same manner, the KL regularization term may lead the VAE-based feature extractor to produce utterance-level features with insufficient idiosyncratic representation for the speaker.
The proposed ALI-based feature extraction framework, in contrast, does not regularize the latent variable distribution with a KL divergence term. Instead, the proposed scheme employs a joint discriminator network, which encourages the encoder and decoder networks to generate realistic latent variables and GMM supervectors. Thus, the distinctive information within the latent variables generated by the ALI-based feature extractor is less likely to be tightly constrained by its prior distribution.

4. Experiments

4.1. Databases

In order to evaluate the performances of the baseline systems and the proposed scheme in a condition similar to real-life application where the speech data for training and enrolling are limited and usually have short durations, we performed experiments using the TIMIT dataset [27] as the development set and TIDIGITS dataset [28] as the enrollment and trial sets. The TIMIT dataset contains 6300 clean recorded utterances, 10 utterances spoken by each of 438 male and 192 female speakers. Each utterance in the TIMIT dataset has an average duration of three seconds. The TIMIT dataset was used for training the UBM and also used for training the total variability matrix. The TIDIGITS dataset contains 25,096 clean recorded utterances spoken by 111 male, 114 female, 50 boy, and 51 girl speakers. Each of the 326 speakers in the TIDIGITS dataset spoke a set of isolated digits and 2–7 digit sequences. The TIDIGITS dataset was split into two subsets, each containing 12,548 utterances from all 326 speakers, and they were separately used as the enrollment and trial data.

4.2. Experimental Setup

The acoustic features used in the experiments were 19-dimensional Mel-frequency cepstral coefficients (MFCCs) and the log-energy extracted at every 10 ms, using a 20 ms Hamming window via the SPro library [29]. Together with the delta and delta–delta of the 19-dimensional MFCCs and the log-energy, the frame-level acoustic feature used in our experiments was a 60-dimensional vector.
We trained a UBM containing 32 mixture components in a gender-independent manner, using all the speech utterances in the TIMIT dataset. Training the UBM, total variability matrix, and the i-vector extraction were done by using the MSR Identity Toolbox via MATLAB [30]. The encoder and decoder of the experimented VAE- and ALI-based networks were configured to have a single hidden layer with 4096 ReLU nodes, and the dimensionality of the latent variables was set to be 200. As depicted in Figure 5, the discriminator network of the ALI-based feature extraction model was configured to follow a similar structure to the multimodal network [31]. The first few hidden layers model the higher-level representation of the GMM supervector and latent variable, and the last hidden layer models the joint information between them. The implementation of the experimented networks was done using Tensorflow [32] and trained using the AdaGrad optimization technique [33]. In addition, dropout [34] with a fraction of 0.8 and L2 regularization with a weight of 0.01 were applied for training all the VAEs, and the Baum–Welch statistics extracted from the entire TIMIT dataset were used as training data. A total of 100 samples were used for reparameterization to approximate the expectations in (15).
For all the extracted utterance-level features, linear discriminant analysis (LDA) [35] was applied for feature compensation and the dimensionality was finally reduced to 200. Probabilistic linear discriminant analysis (PLDA) [36] was used for speaker verification, and the speaker subspace dimension was set to be 200.
Two performance measures were evaluated in our experiments: classification error (Class. err.) and EER. The classification error was measured while performing a speaker identification task where each trial utterance was compared with all the enrolled speakers via PLDA, and the enrolled speaker with the highest score was chosen as the identified speaker. Then, the ratio of the number of falsely classified samples to the total number of trial samples represents the classification error. The EER is a widely used measure for speaker verification which indicates the error when the false alarm rate (FAR) and the false reject rate (FRR) are the same [35].

4.3. Effect of the Duration on the Latent Variable

In order to evaluate the effectiveness of using the latent variable variance as a measure for uncertainty caused by short duration, the differential entropy, which measures the average uncertainty of a random variable, was computed. Since the latent variable z ( X ) is assumed to follow a Gaussian distribution, the differential entropy can be formulated as follows:
h ( z ( X ) ) = 1 2 log ( 2 π e ) K + 1 2 log k = 1 K σ k 2 ( X )
In (18), K represents the dimensionality of the latent variable, and σ k 2 ( X ) is the kth element of σ 2 ( X ) .
From each speech sample in the entire TIDIGITS dataset, the variance of the 200-dimensional latent variable was obtained using the encoder network and used for computing the differential entropy. As shown in Figure 6, we experimented with the latent variables extracted from three different feature extraction models:
  • VAE: the VAE-based feature extraction network trained to minimize (17)
  • ALI: the ALI-based feature extraction network trained to minimize the standard GAN objective function (8)
  • ALI/NLL: the proposed ALI-based feature extraction network trained to minimize the negative log-likelihood-based objective function (16)
The differential entropies are averaged in six different duration groups: less than one second (i.e., <1), 1–2 s, 2–3 s, 3–4 s, 4–5 s, and more than five seconds (i.e., >5).
As shown in the result, the differential entropies computed using the variances of the latent variables extracted from VAE and ALI/NLL gradually decrease as the duration increases. However, the differential entropy computed using the latent variable from ALI does not decrease dramatically compared to the other methods. This may be due to the fact that ALI are trained only to generate a GMM supervector similar to the one obtained via MAP adaptation (14), which is determined by the input Baum–Welch statistics. Therefore, the latent variable of ALI will only be able to preserve information needed for reconstructing a deterministic distribution. On the other hand, VAE and ALI/NLL is trained to generate a GMM distribution according to the maximum likelihood criterion (12), thus their latent variables may capture more information about the variability within the generated GMM distribution.
Another interesting observation from Figure 6 is that the relative decrement in entropy was much greater in ALI/NLL than VAE. While the change in entropy is rather conservative in the VAE case, where the relative decrement between the first duration group (i.e., less than one second) and the sixth duration group (i.e., more than five seconds) was 29.91%, the entropy in ALI/NLL changed dramatically with a relative decrement of 330.17%. This shows that regularizing the latent variable with a joint discriminator network is more effective than using the KL divergence-based regularization for capturing the uncertainty.

4.4. Speaker Verification and Identification with Different Utterance-Level Features

In this subsection, we evaluated the performance of the features extracted from various techniques. More specifically, we compared the performance of the conventional i-vector and the latent variable mean (LM) and log-variance (LV) extracted from the VAE- and ALI-based feature extractors (i.e., VAE, ALI and ALI/NLL). In addition, we conducted feature-level fusion between different features and evaluated their performance. For feature-level fusion, we simply concatenated the different features together to create a supervector. The i-vector features used in this experiment were:
  • i-vector(200): standard 200-dimensional i-vector
  • i-vector(400): standard 400-dimensional i-vector
  • i-vector(600): standard 600-dimensional i-vector
and the latent variable features were:
  • LM: 200-dimensional latent variable mean
  • LV: 200-dimensional latent variable log-variance
The fusion features used in this experiment were:
  • LM + LV: concatenation of the 200-dimensional latent variable mean and the log-variance, resulting in a 400-dimensional vector
  • i-vector + LM: concatenation of the 200-dimensional i-vector and the 200-dimensional latent variable mean, resulting in a 400-dimensional vector
  • i-vector + LV: concatenation of the 200-dimensional i-vector and the 200-dimensional latent variable log-variance, resulting in a 400-dimensional vector
  • i-vector + LM + LV: concatenation of the 200-dimensional i-vector and the 200-dimensional latent variable mean and log-variance, resulting in a 600-dimensional vector
Table 1 and Table 2 respectively give the EER and classification error results obtained by using these features. As depicted in Figure 7, the latent variable mean vector extracted from the VAE- and ALI-based feature extractors (i.e., VAE, ALI, ALI/NLL) shows promising performance. In particular, the performance yielded by the latent variable mean of VAE and ALI/NLL (i.e.,VAE-LM, ALI/NLL-LM) seems to be comparable to the conventional i-vector. On the other hand, as shown in Figure 8, the latent variable log-variance of VAE and ALI (i.e.,VAE-LV, ALI-LV) shows relatively poor performance compared to the latent variable mean features. However, as shown in Figure 8, the latent variable log-variance extracted from ALI/NLL (i.e., ALI/NLL-LV) outperformed the latent variable log-variance features extracted from all three networks (i.e., VAE, ALI, ALI/NLL) in terms of EER. This shows that the proposed network ALI/NLL is capable of generating latent variable variance which not only implies the uncertainty within the input speech but also encodes a sufficient amount of speaker-dependent information. Using the latent variable together (i.e., LM + LV) as a combined feature further improved the performance, and the best performing feature was obtained from ALI/NLL, which achieved a relative improvement of 28.73% in terms of EER compared to that of i-vector (400). Figure 9 shows the detection error trade-off (DET) curves obtained from the experiments using LM + LV.
As shown in Figure 10, augmenting the standard i-vector and the latent variable mean (i.e., i-vector + LM) further improved the speaker verification performance. This improvement may be attributed to the nonlinear feature extraction process of the VAE- and ALI-based methods. Since the latent variable mean is trained to capture the variability within the distribution of the input utterance via a nonlinear process, it is likely to encompass information not attainable in the standard i-vector. Especially the one augmented with the latent variable mean extracted from ALI/NLL (i.e., ALI/NLL-(i-vector + LM)) showed better performance than the ones extracted from VAE and ALI, achieving a relative improvement of 42.16% in terms of EER compared to i-vector (400). The reason behind this may be due to the fact that the latent variable of ALI/NLL can preserve the distinctive information much better by incorporating a joint discriminator instead of regularizing the latent variable distribution via KL divergence.
Likewise, using the i-vector in conjunction with the latent variable log-variance (i.e., i-vector + LV) also showed improvement in performance. Similar to the i-vector + LM experiments, the i-vector augmented with the latent variable log-variance extracted from ALI/NLL (i.e., ALI/NLL-(i-vector + LV)) outperformed the other methods (i.e., VAE-(i-vector + LV), ALI-(i-vector + LV)), achieving a relative improvement of 34.66% compared to i-vector (400) in terms of EER. This may be due to the capability of the latent variable variance extracted from ALI/NLL in capturing the amount of uncertainty, which has been discussed in the previous subsection. Figure 11 shows the DET curves obtained from the experiments using LV.
Further improvement was observed when augmented with both the latent variable mean and log-variance (i-vector + LM + LV), which can be seen in Figure 12. The standard i-vector used in conjunction with the latent variable mean and log-variance extracted from ALI/NLL (i.e., ALI/NLL-(i-vector + LM + LV)) showed a relative improvement of 56.68% in terms of EER compared to the conventional i-vector with the same dimension (i.e., i-vector(600)).

5. Conclusions

In this paper, a novel utterance-level feature extractor using an adversarial learning framework for speaker recognition is proposed. Analogous to the previously proposed VAE-based feature extractor, the architecture proposed in this paper is composed of an encoder and a decoder network where the former estimates the distribution of the latent variable given the speech and the latter generates the GMM from the latent variable. However, in order to prevent the potential loss of distinctive representation for the speaker within the extracted latent variable, the newly proposed feature extractor is trained according to the ALI framework where a joint discriminator network is exploited to ensure that the latent variable and the generated GMM are close to their prior distribution and the GMM obtained through MAP adaptation, respectively.
To evaluate the performance of the features extracted from the proposed system in a short duration scenario, we conducted a set of experiments using the TIDIGITS dataset. From the results, we observed that the variance of the latent variable extracted from the proposed ALI-based feature extractor is more useful to represent the level of uncertainty caused by the short duration of the given speech than the one extracted from the VAE-based feature extractor. Moreover, using the features extracted from the proposed ALI-based method in conjunction with the standard i-vector was shown to be far more effective than the VAE-based method.
Although the proposed method showed good performance in the short duration scenario, we believe that there are still some improvements that could be made. One limitation of the proposed ALI-based method is that it relies on a pretrained GMM model (i.e., UBM), which requires an extra training phase and yields representation of the data distribution limited by the GMM structure.
Therefore, in our future study, we will expand our proposed method to jointly train the background model (i.e., UBM) along with the feature extraction network. Moreover, we will explore the effect of the uncertainty caused by environmental contamination (e.g., noise, reverberation) on the latent variable extracted from the proposed model by evaluating using more adversarial datasets (e.g., VoxCeleb, NIST SRE).

Author Contributions

Conceptualization, W.H.K. and N.S.K.; Funding acquisition, N.S.K.; Investigation, W.H.K.; Methodology, W.H.K.; Supervision, N.S.K.; Writing—original draft, W.H.K.; Writing—review and editing, N.S.K.

Funding

This research was supported by Projects for Research and Development of Police science and Technology under Center for Research and Development of Police science and Technology and Korean National Police Agency funded by the Ministry of Science, ICT and Future Planning (PA-J000001-2017-101).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dehak, N.; Kenny, P.; Dehak, R.; Dumouchel, P.; Ouellet, P. Front-end factor analysis for speaker verification. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 788–798. [Google Scholar] [CrossRef]
  2. Kenny, P. A small footprint i-vector extractor. In Proceedings of the Odyssey 2012, Singapore, 25–28 June 2012; pp. 1–25. [Google Scholar]
  3. Kenny, P.; Boulianne, G.; Dumouchel, P. Eigenvoice modeling with sparse training data. IEEE Trans. Audio Speech Lang. Process. 2005, 13, 345–354. [Google Scholar] [CrossRef]
  4. Dehak, N.; Kenny, P.; Dehak, R.; Glembek, O.; Dumouchel, P.; Burget, L.; Hubeika, V.; Castaldo, F. Support vector machines and joint factor analysis for speaker verification. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 4237–4240. [Google Scholar]
  5. Campbell, W.M.; Sturim, D.; Reynolds, D. Support vector machines using GMM supervectors for speaker verification. IEEE Signal Process. Lett. 2006, 13, 308–311. [Google Scholar] [CrossRef]
  6. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.; Mohamed, A.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.; et al. Deep neural networks for acoustic modeling in speech recognition. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  7. Variani, E.; Lei, X.; McDermott, E.; Moreno, I.L.; G-Dominguez, J. Deep neural networks for small footprint text-dependent speaker verification. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 4080–4084. [Google Scholar]
  8. Snyder, D.; G-Romero, D.; Povey, D.; Khudanpur, S. Deep neural network embeddings for text-independent speaker verification. In Proceedings of the INTERSPEECH, Stockholm, Sweden, 20–24 August 2017; pp. 999–1003. [Google Scholar]
  9. Snyder, D.; G-Romero, D.; Sell, G.; Povey, D.; Khudanpur, S. X-vectors: robust DNN embeddings for speaker recognition. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5329–5333. [Google Scholar]
  10. Chowdhury, F.A.R.; Wang, Q.; Moreno, I.L.; Wan, L. Attention-based models for text-dependent speaker verification. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5359–5363. [Google Scholar]
  11. Wan, L.; Wang, Q.; Papir, A.; Moreno, I.L. Generalized end-to-end loss for speaker verification. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 4879–4883. [Google Scholar]
  12. Yao, S.; Zhou, R.; Zhang, P. Speaker-phonetic i-vector modeling for text-dependent speaker verification with random digit strings. IEICE Trans. Inf. Syst. 2019, E102D, 346–354. [Google Scholar] [CrossRef]
  13. Saeidi, R.; Alku, P. Accounting for uncertainty of i-vectors in speaker recognition using uncertainty propagation and modified imputation. In Proceedings of the INTERSPEECH, Dresden, Germany, 6–10 September 2015; pp. 3546–3550. [Google Scholar]
  14. Hasan, T.; Saeidi, R.; Hansen, J.H.L.; van Leeuwen, D.A. Duration mismatch compensation for i-vector based speaker recognition systems. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 7663–7667. [Google Scholar]
  15. Mandasari, M.I.; Saeidi, R.; McLaren, M.; van Leeuwen, D.A. Quality measure functions for calibration of speaker recognition systems in various duration conditions. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 2425–2438. [Google Scholar] [CrossRef]
  16. Mandasari, M.I.; McLaren, M.; van Leeuwen, D.A. Evaluation of i-vector speaker recognition systems for forensic application. In Proceedings of the INTERSPEECH, Florence, Italy, 27–31 August 2011; pp. 21–24. [Google Scholar]
  17. Kang, W.H.; Kim, N.S. Unsupervised learning of total variability embedding for speaker verification with random digit strings. Appl. Sci. 2019, 9, 1597. [Google Scholar] [CrossRef]
  18. Kingma, D.P.; Welling, M. Auto-encoding variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  19. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  20. Theis, L.; van den Oord, A.; Bethge, M. A note on the evaluation of generative models. In Proceedings of the ICLR, San Juan, PR, USA, 24 April 2016. [Google Scholar]
  21. Shang, W.; Sohn, K.; Akata, Z.; Tian, Y. Channel-recurrent variational autoencoder. arXiv 2017, arXiv:1706.03729v1. [Google Scholar]
  22. Dumoulin, V.; Belghazi, I.; Poole, B.; Mastropietro, O.; Lamb, A.; Arjovsky, M.; Courville, A. Adversarially learned inference. In Proceedings of the ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  23. Reynolds, D.; Quatieri, T.F.; Dunn, R.B. Speaker verification using adapted Gaussian mixture models. Digit. Signal Process. 2000, 10, 19–41. [Google Scholar] [CrossRef]
  24. Dehak, N.; T-Carrasquillo, P.A.; Reynolds, D.; Dehak, R. Language recognition via i-vectors and dimensionality reduction. In Proceedings of the INTERSPEECH, Florence, Italy, 27–31 August 2011; pp. 857–860. [Google Scholar]
  25. Goodfellow, I.J.; P-Abadie, J.; Mirza, M.; Xu, B.; W-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661v1. [Google Scholar]
  26. Salakhutdinov, R. Learning deep generative models. Annu. Rev. Stat. Its Appl. 2015, 2, 361–385. [Google Scholar] [CrossRef]
  27. Lopes, C.; Perdigao, F. Phone recognition on the TIMIT database. Speech Technol. 2011, 1, 285–302. [Google Scholar]
  28. Leonard, R.G. A database for speaker-independent digit recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, San Diego, CA, USA, 19–21 March 1984; pp. 328–331. [Google Scholar]
  29. Gravier, G. SPro: Speech Signal Processing Toolkit. Available online: http://gforge.inria.fr/projects/spro (accessed on 1 July 2019).
  30. Sadjadi, S.O.; Slaney, M.; Heck, L. MSR identity toolbox v1.0: A MATLAB toolbox for speaker recognition research. Speech Lang. Process. Tech. Comm. Newsl. 2013, 1, 1–32. [Google Scholar]
  31. Ngiam, J.; Khosla, A.; Kim, M.; Nam, J.; Lee, H.; Ng, A.Y. Multimodal deep learning. In Proceedings of the ICML, Bellevue, WA, USA, 28 June–2 July 2011; pp. 689–696. [Google Scholar]
  32. Abadi, M.; Agarwal, A.; Harham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Tensorflow: Large-Scale Machine Learning Heterogenous Systems. Available online: tensorflow.org (accessed on 1 July 2019).
  33. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradial methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  34. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  35. Hansen, J.; Hasan, T. Speaker recognition by machines and humans. IEEE Signal Process. Mag. 2015, 32, 74–99. [Google Scholar] [CrossRef]
  36. Garcia-Romero, D.; Espy-Wilson, C.Y. Analysis of i-vector length normalization in speaker recognition systems. In Proceedings of the INTERSPEECH, Florence, Italy, 27–31 August 2011; pp. 249–252. [Google Scholar]
Figure 1. Flow chart of the i-vector framework. Note that when training the i-vector extractor, the GMM supervector is not explicitly computed. The Baum–Welch statistics computation stage is described in Equations (2) and (3), and the log-likelihood formulation is described in Equation (5).
Figure 1. Flow chart of the i-vector framework. Note that when training the i-vector extractor, the GMM supervector is not explicitly computed. The Baum–Welch statistics computation stage is described in Equations (2) and (3), and the log-likelihood formulation is described in Equation (5).
Sensors 19 04709 g001
Figure 2. Flow chart of the ALI framework. The encoder, decoder and joint discriminators are feed-forward neural networks, each consisting of nonlinear hidden layers.
Figure 2. Flow chart of the ALI framework. The encoder, decoder and joint discriminators are feed-forward neural networks, each consisting of nonlinear hidden layers.
Sensors 19 04709 g002
Figure 3. Maximum likelihood training scheme for the proposed nonlinear feature extractor. Red shows the sampling operation. Blue shows the loss term. The log-likelihood computation and sampling operation are described in Equations (10) and (13), respectively.
Figure 3. Maximum likelihood training scheme for the proposed nonlinear feature extractor. Red shows the sampling operation. Blue shows the loss term. The log-likelihood computation and sampling operation are described in Equations (10) and (13), respectively.
Sensors 19 04709 g003
Figure 4. ALI-based training scheme for the proposed nonlinear feature extractor. Red shows the sampling operations. The GMM mean estimation is described in (14) and more detailed information on the joint discriminator structure is depicted in Figure 5.
Figure 4. ALI-based training scheme for the proposed nonlinear feature extractor. Red shows the sampling operations. The GMM mean estimation is described in (14) and more detailed information on the joint discriminator structure is depicted in Figure 5.
Sensors 19 04709 g004
Figure 5. Network structure for the joint discriminator of the ALI-based feature extractor. Each block represents a hidden layer with rectified linear unit (ReLU) activation function.
Figure 5. Network structure for the joint discriminator of the ALI-based feature extractor. Each block represents a hidden layer with rectified linear unit (ReLU) activation function.
Sensors 19 04709 g005
Figure 6. Average differential entropy computed using the latent variable variance extracted from the VAE- and ALI-based systems on different durations.
Figure 6. Average differential entropy computed using the latent variable variance extracted from the VAE- and ALI-based systems on different durations.
Sensors 19 04709 g006
Figure 7. DET curves of the speaker verification experiments using latent variable mean as features.
Figure 7. DET curves of the speaker verification experiments using latent variable mean as features.
Sensors 19 04709 g007
Figure 8. DET curves of the speaker verification experiments using latent variable log-variance as features.
Figure 8. DET curves of the speaker verification experiments using latent variable log-variance as features.
Sensors 19 04709 g008
Figure 9. DET curves of the speaker verification experiments using the concatenation of the latent variable mean and log-variance as features.
Figure 9. DET curves of the speaker verification experiments using the concatenation of the latent variable mean and log-variance as features.
Sensors 19 04709 g009
Figure 10. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable mean as features.
Figure 10. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable mean as features.
Sensors 19 04709 g010
Figure 11. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable log-variance as features.
Figure 11. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable log-variance as features.
Sensors 19 04709 g011
Figure 12. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable mean and log-variance as features.
Figure 12. DET curves of the speaker verification experiments using the i-vectors augmented with the latent variable mean and log-variance as features.
Sensors 19 04709 g012
Table 1. EER comparison between various feature-level fusions of the conventional i-vector, mean and log-variance of the latent variables extracted from the VAE- and ALI-based feature extractors [%].
Table 1. EER comparison between various feature-level fusions of the conventional i-vector, mean and log-variance of the latent variables extracted from the VAE- and ALI-based feature extractors [%].
LMLVLM+LVi-Vector + LMi-Vector + LVi-Vector + LM + LV
i-vector(200)3.36
i-vector(400) 2.68
i-vector(600) 2.17
VAE3.614.652.031.781.650.97
ALI4.394.562.321.591.551.03
ALI/NLL3.643.461.911.551.510.94
Table 2. Classification error comparison between various feature-level fusions of the conventional i-vector, mean and log-variance of the latent variables extracted from the VAE- and ALI-based feature extractors [%].
Table 2. Classification error comparison between various feature-level fusions of the conventional i-vector, mean and log-variance of the latent variables extracted from the VAE- and ALI-based feature extractors [%].
LMLVLM + LVi-Vector + LMi-Vector + LVi-Vector + LM + LV
i-vector(200)12.62
i-vector(400) 7.67
i-vector(600) 5.07
VAE11.8917.786.945.364.992.75
ALI16.6217.318.545.104.792.79
ALI/NLL12.5612.386.763.974.182.49

Share and Cite

MDPI and ACS Style

Kang, W.H.; Kim, N.S. Adversarially Learned Total Variability Embedding for Speaker Recognition with Random Digit Strings. Sensors 2019, 19, 4709. https://doi.org/10.3390/s19214709

AMA Style

Kang WH, Kim NS. Adversarially Learned Total Variability Embedding for Speaker Recognition with Random Digit Strings. Sensors. 2019; 19(21):4709. https://doi.org/10.3390/s19214709

Chicago/Turabian Style

Kang, Woo Hyun, and Nam Soo Kim. 2019. "Adversarially Learned Total Variability Embedding for Speaker Recognition with Random Digit Strings" Sensors 19, no. 21: 4709. https://doi.org/10.3390/s19214709

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop