Next Article in Journal
Adversarial Machine Learning for NextG Covert Communications Using Multiple Antennas
Previous Article in Journal
Cardio-Hypothalamic-Pituitary Coupling during Rest and in Response to Exercise
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition

1
Key Laboratory of Child Development and Learning Science (Ministry of Education), Southeast University, Nanjing 210096, China
2
School of Information Science and Engineering, Southeast University, Nanjing 210096, China
3
School of Biological Science and Medical Engineering, Southeast University, Nanjing 210096, China
4
School of Cyber Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2022, 24(8), 1046; https://doi.org/10.3390/e24081046
Submission received: 28 June 2022 / Revised: 26 July 2022 / Accepted: 26 July 2022 / Published: 29 July 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Cross-corpus speech emotion recognition (SER) is a challenging task, and its difficulty lies in the mismatch between the feature distributions of the training (source domain) and testing (target domain) data, leading to the performance degradation when the model deals with new domain data. Previous works explore utilizing domain adaptation (DA) to eliminate the domain shift between the source and target domains and have achieved the promising performance in SER. However, these methods mainly treat cross-corpus tasks simply as the DA problem, directly aligning the distributions across domains in a common feature space. In this case, excessively narrowing the domain distance will impair the emotion discrimination of speech features since it is difficult to maintain the completeness of the emotion space only by an emotion classifier. To overcome this issue, we propose a progressively discriminative transfer network (PDTN) for cross-corpus SER in this paper, which can enhance the emotion discrimination ability of speech features while eliminating the mismatch between the source and target corpora. In detail, we design two special losses in the feature layers of PDTN, i.e., emotion discriminant loss L d and distribution alignment loss L a . By incorporating prior knowledge of speech emotion into feature learning (i.e., high and low valence speech emotion features have their respective cluster centers), we integrate a valence-aware center loss L v and an emotion-aware center loss L c as the L d to guarantee the discriminative learning of speech emotions except an emotion classifier. Furthermore, a multi-layer distribution alignment loss L a is adopted to more precisely eliminate the discrepancy of feature distributions between the source and target domains. Finally, through the optimization of PDTN by combining three losses, i.e., cross-entropy loss L e , L d , and L a , we can gradually eliminate the domain mismatch between the source and target corpora while maintaining the emotion discrimination of speech features. Extensive experimental results of six cross-corpus tasks on three datasets, i.e., Emo-DB, eNTERFACE, and CASIA, reveal that our proposed PDTN outperforms the state-of-the-art methods.

1. Introduction

Emotions reflect the psychological state of human beings, which are usually manifested in physiological and psychological signals [1,2,3,4,5], e.g., facial expression, speech, and electroencephalogram (EEG). As a commonly used communication mean, speech contains rich emotional information. Therefore, making the machine recognize the emotional states of speech, known as the speech emotion recognition (SER) task, is crucial for human–computer interaction (HCI). Generally, the task setting of SER suggests that the training and testing data come from the same corpus, which will cause the trained model on the training data to perform poorly on a new corpus. In recent years, the SER task, which involves training one dataset or several datasets and testing others, i.e., cross-corpus SER, has attracted wide attention.
Since the training and testing data are collected from different datasets, cross-corpus SER faces the issue that the speech samples are quite different in terms of background noise, recording device, language, or speaker, which will lead to the “corpus bias” [6]. Therefore, the cross-corpus SER is more practical than traditional SER tasks. To deal with the bias issue, early works investigated the utilization of low-level descriptors (LLDs) to enhance the emotion discrimination of speech features [7,8]. For instance, Shami et al. [7] firstly implemented the cross-corpus task by using utterance-level acoustical parameters for the naive classifiers, e.g., K-nearest neighbors (KNNs) and support vector machines (SVMs). Further, Schuller et al. [8] defined the cross-corpus SER settings standardly and explored several normalization methods (e.g., speaker normalization, corpus normalization, and speaker-corpus normalization) to reduce the “corpus bias”.
With the rapid development of transfer learning (TL) [9], TL-based methods have shown to be promising for cross-corpus SER [6,10,11,12,13]. TL aims to eliminate the bias between the training and testing data, especially domain adaption (DA) [13,14,15,16], which focuses on the issue of only labeled training data (source domain) and unlabeled testing data (target domain). Motivated by TL, Hassan et al. [17] regarded the dataset difference as a covariance shift, and explored the application of three importance weights (IWs) methods, i.e., kernel mean matching (KMM), unconstrained least-squares importance fitting (uLSIF), and the Kullback–Leibler importance estimation procedure (KLIEP), into the support vector machine (SVM) classifier to reduce this shift and achieved the UAR of 42.7 % on FAU Aibo-Mont. Then, Zong et al. [10] proposed a domain-adaptive least-squares regression (DaLSR) method for the corpus shift in cross-corpus SER by projecting speech features to the emotion label space, in which the feature distributions of the training and testing data are as close as possible. The DaLSR obtained the best WAR ( 52.47 % ) and UAR ( 44.41 % ) for the task of eNTERFACE to Emo-DB at that time. Further, Zhang et al. [13] jointly performed the transfer subspace learning and regression in [10] to learn the corpus-invariant speech features, which achieved the UAR of 49.58 % on the task of CASIA to Emo-DB. These works mainly measure the distribution distance of two domains based on the maximum mean discrepancy (MMD). Instead of MMD, Song et al. [6,11,12] also adopted the nearest neighbor graph as the distribution distance metric to reduce domain shifts in the latent space of speech features.
In addition to the subspace learning methods, deep learning methods have also achieved dominating performance as a recent research hotspot. In [18], a domain classifier with a deep neural network (DNN) has been intergraded into the feature extractor and the emotion classifier to learn the emotion-discriminative and domain-invariant feature for cross-corpus SER. Similarly, Abdelwahab et al. [14] also adopted adversarial training to eliminate the domain discrepancy during the common feature learning by an additional domain classifier. Furthermore, Gideon et al. [15] introduced a “meet in the middle” method, i.e., adversarial discriminative domain generalization (ADDoG), to learn the feature of each dataset closer to one another, which can improve the generalization of the dataset representations and then extend it to multiclass ADDoG for the training data with more datasets. The ADDoG obtained the UAR of 0.4749 % on the task of MSP-Improv to IEMOCAP.
Although previous works have achieved promising progress in cross-corpus SER, these methods simply consider the SER task as a DA problem, which first learns the common space of speech emotion features for the source and target datasets, and then decreases the distribution distance between the two domains in the common space. In this case, therefore, excessively narrowing the domain distance will impair the emotion discriminativeness of speech features because it is difficult to maintain the completeness of the emotion space only by the emotion classifier (e.g., KNN, SVM, or DNN) [13]. Aiming at this issue, the subspace learning methods perform the sparse constraints (e.g., 1 norm, 2 norm, or 2 , 1 norm) on the projection matrix of speech features to avoid redundant information for the discriminative emotion feature space [6,10,13]. However, the linear mapping of the subspace learning limits the representation ability of features, which is one of its disadvantages. In addition, the deep learning methods on cross-corpus SER still only consider eliminating the distribution shifts across the source and target domains, while ignoring the preservation of emotion discrimination on speech features.
To address the above issues, we jointly consider the emotion discrimination preservation of speech features and the distribution elimination between the source and target domains, and integrate them into the deep feature extractor. A benefit of this approach is to enhance the emotion discriminativeness of speech while narrowing the distribution discrepancy between two domains such that the emotion-discriminative and domain-invariant speech features can be obtained through the training of a deep end-to-end network.
Therefore, in this paper, we propose a progressively discriminative transfer network (PDTN) for the cross-corpus SER. In the PDTN, we adopt two special losses i.e., emotion discriminant loss L d and distribution alignment loss L a , in the high-level feature layers (i.e., fc layers), where L d is combined with the emotion classification loss L c e to enhance the emotion discrimination of speech features and L a decreases the distribution distance of features between the source and target domains. Specifically, L d contains a valence-aware center loss L v and an emotion-aware center loss L c , which are inspired by the prior knowledge of speech emotions, i.e., speech emotion features of the high and low valences have their respective cluster centers. Further, we utilize the multi-layer MMD in L a to measure the domain shift of marginal distributions between two domains. The proposed PDTN integrates the three losses, i.e., L a , L d , and L c e to progressively eliminate the inter-domain discrepancy and improve the emotion discriminativeness of speech features through an end-to-end network training stage. Experimental results on three datasets, i.e., Emo-DB, eNTERFACE, and CASIA, demonstrate the superiority of our proposed PDTN over the comparison methods.
Overall, the contributions of this paper can be summarized as the following three points:
  • This paper proposes a novel progressively discriminative transfer network for cross-corpus SER, which jointly considers the two aspects of eliminating the distribution discrepancy across the source and target domains, and enhancing the emotion discrimination of speech features during deep feature learning. Thus, it can avoid the dilemma that previous methods only consider one of two above aspects.
  • As far as we know, it is the first work to introduce the prior knowledge of speech emotions, i.e., speech emotion features of high and low valences with their respective cluster centers, into the deep feature learning to enhance the emotion discrimination of speech representations.
  • We adopt high-level features of fc layers to perform a practical distribution discrepancy measures under multi-layer features between the source and target domains through a multi-layer MMD metric.
The rest of this paper is organized as follows: Section 2 illustrates the proposed method in detail. Then, we conduct our experiments and discuss the results in Section 3. Finally, Section 4 concludes the paper and gives some points for future research.

2. The Proposed Method

In this section, we will illustrate the framework of PDTN in detail, shown in Figure 1, which can be divided into three parts, i.e., deep feature extraction, emotion discrimination preservation, and distribution discrepancy elimination.

2.1. Deep Feature Extraction

Compared with traditional methods, deep learning has performed well in speech processing, e.g., SER, speech recognition, and speech enhancement. Especially in the SER, DNNs (e.g., CNN and RNN) can extract the high-level feature of speech with more discriminative emotion information. Therefore, we adopt the deep CNN (DCNN) as the backbone network of our proposed PDTN for the deep feature extraction of speech emotion, according to [18,19]. Moreover, as a time–frequency representation of speech, the spectrogram is commonly used for the input feature of DCNN instead of the hand-crafted features.
To illustrate the process of deep feature extraction clearly, firstly, we formalize the labeled training dataset as D s = { x i s , l i } i = 1 n s and the unlabeled testing dataset D t = { x j t } j = 1 n t , where x i s and x j t are donated as the spectrogram of the ith speech sample in source data and the jth sample in source data, l i represents the emotion label of the ith speech in source dataset, and n s and n t are the numbers of source and target samples. Notably, since our proposed method is based on unsupervised domain adaptation (UDA) in TL, the target speech samples have no labels.
Then, the spectrogram features of the source and target dataset are fed into the DCNNs to extract the high-level representations of speech emotions. In this paper, we select the AlexNet and VGGNet as the comparison backbones of the proposed PDTN to evaluate the method’s performance. Through the backbones, the spectrograms x are encoded in time and frequency domains by a series of stacked convolutional layers, and further pass through several fully connected (fc) layers to obtain high-dimensional emotional semantic features f = { f k s , f k t } k = 1 n l , where f k s and f k t represent the features of source and target datasets in the kth fc layer, and n l is the number of the fc layers. Eventually, the extraction process of the high-level emotion feature f k = [ f k s , f k t ] in the kth fc layer of backbone network G f ( · ) can be formalized as
f k = G f ( [ x s , x t ] ; θ f ) ,
where θ f is the parameters of the feature extraction network G f ( · ) . The numbers of fc layers in AlexNet and VGGNet are both set as 3 in this paper.

2.2. Emotion Discrimination Preservation

In cross-corpus SER, after extracting the deep speech emotion feature f k , the common practice is to either input these high-dimensional features into a fully connected network-based classifier for emotion recognition or to align the distribution of these features both in the source and target domains [14,15,18]. However, since speech emotion is easily disturbed by other factors, e.g., background noise, speaker identity, and language, the emotion features are always confused with the features of these factors [10,11,12]. Therefore, in cross-corpus SER, when the feature distributions between domains are aligned, only utilizing a single emotion classifier cannot effectively disentangle the emotion information from the confusing features in a sufficiently complete feature subspace, which will damage the emotion discrimination of speech features in the feature generalization learning. To address this issue, we introduce an emotion discrimination preservation learning of speech features in the distribution alignment process, which can decouple independent emotion features in the common feature space using the prior knowledge of emotions.
As we know, the emotions can be represented on the two-dimensional arousal–valence emotion wheel [20,21,22], shown in Figure 2. It is obvious that each of seven emotions (i.e., angry, disgust, fear, happy, neutral, sad, and surprise) have a specific position on the arousal and valence axes of the emotion wheel. According to these positions, the prior knowledge of emotion categories can be indicated, that is, the seven emotions are divided into the negative-valence group (i.e., angry, disgust, fear, sad, and surprise), the neutral valence group (i.e., neutral), and the positive-valence group (i.e., happy and surprise) on the valence axis. Under these groups, the emotions in the same group are naturally near each other on the valence, indicating that the centers of their classes are relatively close. On the contrary, the emotions in the different groups have distant centers of emotion classes. Therefore, we introduce the prior knowledge of emotion categories into deep feature learning to maintain the emotion discrimination of speech features.
Specifically, we design a valence-aware center loss L v to model the emotion similarity inside groups and dissimilarity outside groups, which can be donated as
L v = i = 1 n b m a x ( 0 , f k s , i v l , i 2 2 α 1 ) + m a x ( 0 , α 2 v n b v p b 2 2 ) ,
where n b is the mini-batch size; f k s , i represents the kth fc layer feature of the ith speech sample in the source dataset; v n b and v p b are the mini-batch feature centers of the negative-valence emotion group N = { a n g r y , d i s g u s t , f e a r , s a d , s u r p r i s e } and the positive-valence emotion group P = { n e u t r a l , h a p p y } , respectively; v l , i is the feature center of the emotion group where the lth class of the ith speech sample belongs; and α 1 and α 2 are the thresholds to adjust the feature distances within group and between groups, respectively. The feature center of v l , i can be obtained as follows
v l , i = v n , l i N , v p , l i P ,
where v n and v p are the global centers of the negative-valence and positive-valence emotion groups in the whole source data, which can be calculated during the parameter updating in Algorithm 1. Moreover, the negative-valence feature centers v n b and the positive-valence feature centers v p b in each mini-batch can be denoted as
v n b = 1 n b 1 i n b , l i N f k s , i ,
v p b = 1 n b 1 j n b , l j P f k s , j ,
where n b and n b are the numbers of speech samples belonging to N and P in a mini-batch, respectively, and n b = n b + n b .
In addition to the arousal-valance-based center loss, we also construct a fine-grained emotion discrimination preservation strategy by fully using the prior information of each emotion category to finely maintain the emotion discriminativeness. Specifically, we design a novel emotion-aware center loss L c , which can decrease the inter-class distance and increase the intra-class distance in the source data, represented as follows
L c = i = 1 n s m a x ( 0 , f k s , i c i 2 2 α 1 ) + p , q = 1 , p q c m a x ( 0 , α 2 c p b c q b 2 2 ) ,
where c i is the feature center of the emotion category corresponding to the ith speech sample in the whole source data, which is implemented for details in Algorithm 1. α 1 and α 2 are the thresholds to adjust the distances, respectively. c p b and c q b are the mini-batch feature centers of the pth and qth emotion category, where c q b can be formalized as
c q b = 1 n b q 1 i n b q f k s , i ,
where n b q is the number of speech samples in a mini-batch corresponding to the qth emotion category. The formalization of c p b is similar to c q b .
Consequently, we combine L v and L c in deep feature learning to ensure the discrimination of emotions from coarse to fine in the process of distribution discrepancy elimination. Therefore, the loss of emotion discrimination preservation can be represented as
L d = λ L v + γ L c ,
where λ and γ are the tradeoff parameters to balance the two losses.
Algorithm 1: Algorithm for the parameter optimization of PDTN.
Input: the input features of source and target data: { x i s } i = 1 n s , { x j t } j = 1 n t ;
   training labels of source data: { l i } i = 1 n s ; fc layers: [ f c 1 , f c 2 , f c 3 ] ;
   learning rate: l r and trade-off parameters λ , γ , and μ .
Initialize: θ f , θ c randomly.
Output: the optimized parameters: θ ^ f , θ ^ c .
while the total loss L t o t a l < ϵ or iter n < maxIter do
(1) Generate a mini-batch features of source and target data: { x i s } i = 1 n b , { x j t } j = 1 n b ;
(2) Extract the high-level features of source and target data: { f k s , f k t } k = 1 n l = G f ( [ x s , x t ] ; θ f ) ;
(3) Calculate the negative-valence and positive-valence feature centers v n b and v p b
 in each mini-batch by the Equations (4) and (5);
(4) Calculate the feature center of qth class c q b in each mini-batch by the Equation (7);
(5) if iter n = 1 :
  Initialize global centers v n , v p , and c q (or c p ) in whole source data using steps (4) and (5);
else:
v n = 1 1 + n b 1 i n b , l i N ( v n b f k s , i ) , v n v n η v n , v p = 1 1 + n b 1 j n b , l j P ( v p b f k s , i ) , v p v p η v p , c q = 1 1 + n s q 1 i n b q ( c q b f k s , i ) , c q c q η c q ;
(6) Calculate L v , L c , L a , L c e , and L t o t a l using Equations (2), (6), and (10)–(12), respectively;
(7) Update the parameter θ f and θ c :
θ c θ c μ L c e θ c , θ f θ f μ L t o t a l θ f ;
(8) n = n + 1 .
end while

2.3. Distribution Discrepancy Elimination

Besides the discriminative feature of emotional speech, another challenge in cross-corpus SER is how to eliminate the domain shift between the source and target data, caused by the factors such as background noise, speaker identity, language, etc. To address this challenge, the moment matching-based methods [12,13] and adversarial learning-based methods [14,15] have been widely investigated and achieved great success. Adversarial learning adopts a domain discriminator to confuse the domain information of features for the discriminative representation of emotional speech, which is prone to a lack of convergence [14]. Moment matching is used to find a suitable metric function to measure the discrepancy between domains, e.g., MMD [13], 2 distance [23], Deep Coral [24], which is a non-parameter method and easy to implement. Therefore, the previous works of cross-corpus SER mainly integrated MMD into the subspace learning. Nevertheless, the speech emotion features generated by subspace learning are low-level such that it cannot accurately represent the feature distribution of the source and target data, which brings errors to the distance measurement. Thus, in this paper, we utilize the high-level features in fc layers to measure the distribution distance precisely. In addition, since the features of each fc layer correspond to the specific discrimination, inspired by [25,26,27], we also extend the feature alignment of a single layer to a multi-layer adaptation to obtain a more accurate measurement for the domain shift.
Firstly, we implement the distribution discrepancy of the signal layer high-level feature in the kth fc layer, namely D k , which can be formalized as
D H k = 1 n s 2 i = 1 n s j = 1 n s K ( f k s , i , f k s , j ) + 1 n t 2 i = 1 n t j = 1 n t K ( f k t , i , f k t , j ) 2 n s n t i = 1 n s j = 1 n t K ( f k s , i , f k t , j ) ,
where k [ 1 , 2 , , n l ] , K ( f k s , i , f k s , j ) = ϕ ( f k s , i ) , ϕ ( f k s , j ) is the kernel function in the high-dimension reproducing kernel Hilbert space (RKHS) H , which is denoted as the inner product : , : of the source and target features’ mapping function ϕ .
Further, D H k can be extended to the multi-layer feature distribution distance measurement by integrating the MMD in the two domain features of several fc layers to match the discrepancy between the source and target domains more accurately. Therefore, we can obtain the multi-layer distribution discrepancy distance and take it as the distribution alignment loss L a to constrain the model to gradually eliminate the domain shift between domains during the feature learning process. So, L a can be represented as follows
L a = 1 n s 2 i = 1 n s j = 1 n s k = 1 n l K k ( f k s , i , f k s , j ) + 1 n t 2 i = 1 n t j = 1 n t k = 1 n l K k ( f k t , i , f k t , j ) 2 n s n t i = 1 n s j = 1 n t k = 1 n l K k ( f k s , i , f k t , j ) ,
where K k is the kernel function corresponding to the features in the kth fc layer.

2.4. PDTN for Cross-Corpus SER

In cross-corpus SER, the spectrograms x s and x t of the source and target data are fed into the backbone network (e.g., AlexNet, VGGNet) to extract the high-level emotion semantic features in the kth fc layer, i.e., f k s and f k t . After this step, the high-level features in the first fc layer of the source and target data are utilized to calculate the valence-aware center loss L v , and the features in the second fc layer are used to generate the emotion-aware center loss L c . The combination of L v and L c is regarded as the emotion discrimination preservation loss L d to maintain the emotion information of speech features from coarse to fine. Furthermore, the source feature f 3 s in the final fc layer is adopted to predict emotion labels for cross-entropy loss L c e by emotion classifier G c ( · ) , which can be represented as
L c e = i = 1 n s J ( G c ( f n l s , i ; θ c ) , l i ) ,
where θ c denotes the parameter of the emotion classifier G c ( · ) and J ( · ) is the cross-entropy function.
Then, the high-level features in three fc layers of the source and target data are adopted to produce the distribution alignment loss L a for eliminating domain shifts between the source and target domains. Consequently, we can obtain the corpus-invariant and discriminative emotion representation through the total loss L t o t a l , which can be denoted as
L t o t a l = L c e + L d + μ L a = L c e + λ L v + γ L c + μ L a ,
where L c e is the cross-entropy loss of the emotion classifier. λ , γ , and μ are all the tradeoff parameters used to balance the different losses.
According to the aforementioned pipeline, the proposed PDTN is optimized by the L t o t a l to update the parameters of the backbone network and classifier. The detailed optimization processing is illustrated in Algorithm 1. Thus, in this paper, we utilize three fc layers in both AlexNet and VGGNet backbones, i.e., f c 1 , f c 2 , and f c 3 . Specifically, the features in f c 1 and f c 2 are utilized to calculate the L v and L c , respectively. The L a is obtained by integrating the features in three fc layers into the alignment loss.

3. Experiments

In this section, several experiments are implemented to evaluate our proposed method, and the results are also discussed to illustrate its applicability for cross-corpus SER.

3.1. Dataset

  • eNTERFACE [28] is a public English multi-modal emotion dataset, which contains 1290 audio-visual samples with a sample rate of 48 kHz. In this dataset, six emotions, i.e., anger, disgust, fear, happiness, sadness, and surprise, are induced by the pre-prepared performance contents. Forty-three volunteers coming from different countries with males and females participated in the recording of the dataset.
  • CASIA [29] includes 7200 emotional speech sentences with the Chinese language. Each sample is recorded with six emotions, i.e., anger, fear, happiness, neutral, sadness, and surprise, through some acting contents from four actors containing two males and two females. We utilize 1200 public speech samples with the sample rate of 16 kHz for the experiments.
  • Emo-DB [30] is collected as a German emotional speech dataset with 535 speech samples by ten native speakers, including five males and five females. In Emo-DB, each sentence is recorded with 16 kHz under seven emotions, i.e., anger, boredom, disgust, fear, happiness, neutral, and sadness.
In this paper, to perform the cross-corpus SER conveniently, we pick common emotion categories inside two datasets which are adopted for the cross-corpus task. We also design six tasks according to three datasets and the detailed setting is shown in Table 1, in which e, c, and b represent the datasets of eNTERFACE, CASIA, and Emo-DB, respectively.

3.2. Experimental Setting

In order to obtain the input of the proposed PDTN, we transform the speech signals to spectrogram features through the short-time discrete Fourier transform (STFT) with the Hamming window, in which the frame length is set as 350, and the FFT points is 1024. It is noted that all speech samples are chosen as the signal channel data and resampled to the sample rate of 16 kHz.
In PDTN, we select the AlexNet [31] and VGGNet (i.e., VGGNet-11) [32] as the backbone networks to evaluate the PDTN’s performance on different networks. In the backbone networks, their three fc layers with the dimensions of 4096, 4096, and class number, i.e., f c 1 , f c 2 , and f c 3 , are adopted to calculate the emotion discrimination preservation loss L d and the distribution alignment loss L a . Moreover, to match the input size of backbone networks, the dimension of spectrogram features is resized as 224 × 224 . The implementation of our proposed PDTN is based on the deep learning framework Pytorch with NVIDIA GeForce RTX3090 GPUs and it is optimized by the Adam optimizer [33] with a batch size of 32. Its initialized learning rate is set as 0.0002 with a decay weight of 0.9 and the training epoch is set as 500.
We also describe other parameters for the detail as follows. For instance, we utilize the Gaussian kernel in the MMD of L a and its bandwidth is set according to [34]. For the trade-off parameters, we set the γ and λ by the grid search strategy in the parameter set [ 0.001 , 0.003 , 0.005 , 0.01 , 0.03 , 0.05 , 0.1 , 0.5 ] . μ is set by an adjusting strategy, which can be formalized as μ = 2 1 + e δ p 1 . Then, δ is fixed to 10 and p is defined as the ratio of the current number of iterations to the total number of iterations.
In addition, in this paper, we adopt the setting of the cross-corpus SER task by training the PDTN in one dataset (e.g., eNTERFACE) and testing the model in another dataset (e.g., CASIA). Therefore, the six cross-corpus tasks are generated by three datasets, which are summarized as the task section in Table 1. Furthermore, two widely used measure criteria for the recognition accuracy are adopted to evaluate the performance of our proposed PDTN, i.e., the weighted average recall (WAR) and the unweighted average recall (UAR). WAR is denoted as the ratio of the number of correctly predicted samples to the total number of samples, and UAR is the average of the correct rate of each class. UAR has an advantage on measuring the model’s performance on class imbalance databases over WAR. Therefore, combining WAR and UAR can more comprehensively evaluate the performance of PDTN with state-of-the-art methods.

3.3. Comparison Methods

To effectively estimate the performance of our proposed PDTN on the cross-corpus SER tasks, we choose several state-of-the-art methods for the comparison, which are described as follows:
  • Baseline methods: both backbone networks used to extract the high-level features for the experiments.
AlexNet [31]: includes five convolution blocks with the kernel of 5 × 5 or 3 × 3 and three fc layers with the dimensions of 4096, 4096, and class number.
VGGNet-11 [32]: consists of eight convolution blocks with the kernel of 3 × 3 and three fc layers with the dimensions of 4096, 4096, and class number.
  • DA-based methods: all domain adaptation-based methods for cross-corpus SER tasks by our own implementation.
DAN [27]: contains a deep feature extractor and a domain alignment layer with the MMD in multiple fc layers.
DANN [35]: utilizes the domain adversarial training strategy by a domain discriminator to obtain the task-specific and domain-invariant representation.
Deep CORAL [24]: integrates the CORAL loss based on the second-order statistics (i.e., covariances) into a deep neural network for the end-to-end unsupervised domain adaptation framework.
DSAN [34]: proposes a non-adversarial sub-domain adaptation to align the local distribution discrepancy using joint local MMD.
Note that our proposed PDTN is non-parameterized because the calculation of L c e , L d , and L a does not require the parameter updating. Therefore, the parameter number of PDTN depends on the backbone networks, i.e., PDTN (AlexNet) has a similar parameter number with AlexNet (60 millions) [31] and the parameter number of PDTN (VGGNet-11) is the same as VGGNet-11 (133 millions) [32]. Furthermore, the parameters of other comparison methods, e.g., DAN, Deep CORAL, and DSAN, also rest with backbone networks. However, DANN has larger parameters than others because of the additional domain discriminator [35]. In addition, compared with AlexNet, VGGNet-11, and DANN, the proposed PDTN, DAN, Deep CORAL, and DSAN all design novel lossless resulting in extra computational complexity. Specifically, PDTN, DAN, and DSAN are based on MMD ( O ( n 2 ) ) and Deep CORAL is based on the second-order covariance ( O ( n 4 ) ), where n is the larger one of source number n s and target number n t .

3.4. Results and Discussions

The experimental results of six cross-corpus SER tasks are reported in Table 2 with WAR and UAR. The comparison results reveal that our proposed PDTN based on the two backbone networks, i.e., AlexNet and VGGNet-11, can achieve the best performance over other state-of-the-art methods. In detail, the DA-based methods are superior to the baseline methods for all six tasks of cross-corpus SER on the average accuracies. For each task, the DA-based methods also surpassed the performance of most tasks. Significantly, the discrepancy-based methods, i.e., DAN and Deep CORAL, achieve the comparable recognition rate with the adversarial-based method, i.e., DANN, demonstrating that the distribution alignment strategy, either distance measurement or adversarial training, can promote the corpus-invariant emotion features. Furthermore, DSAN has better performance than these three DA-based methods due to the sub-domina alignment strategy taken into account in DSAN. Furthermore, our proposed method goes beyond the mentioned DA-based methods. This is because the proposed PDTN framework not only adapts the marginal distribution between multiple layers but also maintains the emotion discriminative of speech features.
From the results in the Table 2, we can also observe that the tasks of b e , e c , and c e have worse performances than other tasks (i.e., e b , b c , and c b ). This situation indicates that variations in training and test datasets may affect the generalization performance of all cross-corpus methods. In addition, it is also interesting to find that the actuaries of b e are less than e b , which may be because the database of Emo-DB is small such that it cannot sufficiently obtain robust speech emotion features. Furthermore, neither c e nor e c perform promisingly. This is very likely because the CASIA and eNTERFACE are based on different languages, as CASIA is a Chinese dataset and eNTERFACE is an English one. The disparities across languages lead to the emotion variations in speech, which is also a research hotspot in the field of SER. Nevertheless, our proposed PDTN outperforms both the average accuracies and the performance of each task, demonstrating the superiority of the PDTN.

3.5. Ablation Experiments

To verify the effects of different components in the proposed PDTN, we also conduct the ablation study to illustrate this point through extensive experiments. The results with WAR and UAR for ablation experiments are illustrated in Table 3, in which PDTN_S and PDTN_M represent the signal-layer and multi-layer distribution alignment strategy in the PDTN framework according to Section 2.3. Furthermore, we select several key components of PDTN to explore their functions for cross-corpus SER. For instance, PDTN_M w/o L c & L v , PDTN_M w/o L v denote the model under the PDTN framework without the L c and L v losses and the one without the L v loss, respectively. Thus, for convenient comparison purposes, we adopt VGGNet-11 as the backbone network of PDTN for the ablation study. Thus, the PDTM_M herein is the proposed PDTN (VGGNet-11) in Table 2.
From the ablation results in Table 3, firstly, it is clear that PDTN_M w/o L c & L v outperforms PDTN_S w/o L c & L v in terms of the average accuracies, which indicates that the multi-layer alignment can obtain more domain-invariant features of speech emotions. Secondly, the performances of PDTN_M w/o L c & L v and PDTN_M w/o L v demonstrate that the emotion-aware loss L c we designed in the PDTN framework could facilitate the speech emotion feature learning with more discrimination. Thirdly, the PDTN achieves the best performance compared to other ablation components in the average accuracies. Moreover, PDTN illustrates its superior recognition rates in most of cross-corpus SER tasks except c e . These comparison results all demonstrate our proposed emotion discrimination preservation loss L d , including the valence-aware loss L v and emotion-aware loss L c , and distribution alignment loss L a can obtain more discriminative and corpus-invariant representations of emotional speech.

3.6. Visualization for Feature Distribution

The key to copying with cross-corpus SER is to extract the discriminative speech emotion feature. Therefore, to demonstrate the superiority of the proposed method on emotion discriminative preservation, we choose the features under the task of e b for the visualization. The feature distributions of different emotions are visualized in Figure 3, in which the features are generated by the fc layers (i.e., f c 1 , f c 2 , and f c 3 ) in the PDTN based on VGGNet-11. The distributions are shown through t-SNE [36], and the points of different colors represent the corresponding emotions, i.e., anger, disgust, fear, happiness, and sadness.
The sub-figures from (a)–(c) of Figure 3(1)–(3) illustrate that the deeper the fc layer, the more compact the distribution margin of each emotion, indicating that the deeper fc layer features contain stronger emotion discrimination. In addition, from Figure 3(1)–(3), we can also observe that, with the integration of distribution alignment loss L a , emotion-aware loss L c , and valance-aware loss L v , the features in three fc layers become more dispersed among different emotions, and more compact among the same emotion. These visualizations all demonstrate that our proposed PDTN framework is adept at maintaining the emotion discrimination of speech features while eliminating the distribution shift between training and testing data.

4. Conclusions

In the paper, we propose a progressively discriminative transfer network (PDTN) for cross-corpus SER, aiming at preserving the emotion discrimination of speech emotion features and eliminating distribution discrepancy between the training and testing data. In PDTN, we design the special discriminative loss L d based on the prior knowledge of speech emotions, including the valence-aware loss L v and emotion-aware loss L c , to assist the emotion classifier in enhancing the discrimination of speech features in deep feature learning processing. Then, we also adopt the multi-layer distribution alignment based on MMD to reduce the domain shifts between the source and target data. The experimental results of six cross-corpus SER tasks on three public datasets (i.e., Emo-DB, eNTERFACE, and CASIA) show that our proposed PDTN can obtain the more discriminative and domain-invariant representation of emotional speech than the state-of-the-art methods. In fact, the distance metric we adopt is based on the marginal distribution. Therefore, we will explore integrating conditional distribution to obtain a finer-grained measure for the domain shift in the future.

Author Contributions

Conceptualization and methodology, C.L., C.T., J.Z. and Y.Z.; validation, C.L. and J.Z.; writing—original draft preparation, C.L. and C.T.; writing—review and editing, J.Z. and Y.Z.; funding acquisition, C.L. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural National Science Foundation of China (NSFC) under the Grants U2003207 and 61902064, in part by the Jiangsu Frontier Technology Basic Research Project under the Grant BK20192004, in part by the Zhishan Young Scholarship of Southeast University, and in part by the Scientific Research Foundation of Graduate School of Southeast University under Grant YBPY1955.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cowie, R.; Douglas-Cowie, E.; Tsapatsoulis, N.; Votsis, G.; Kollias, S.; Fellenz, W.; Taylor, J.G. Emotion recognition in human-computer interaction. IEEE Signal Process. Mag. 2001, 18, 32–80. [Google Scholar] [CrossRef]
  2. Schuller, B.W.; Picard, R.; André, E.; Gratch, J.; Tao, J. Intelligent signal processing for affective computing. IEEE Signal Process. Mag. 2021, 38, 9–11. [Google Scholar] [CrossRef]
  3. Lu, C.; Zheng, W.; Li, C.; Tang, C.; Liu, S.; Yan, S.; Zong, Y. Multiple spatio-temporal feature learning for video-based emotion recognition in the wild. In Proceedings of the 20th ACM International Conference on Multimodal Interaction, Boulder, CO, USA, 16–20 October 2018; pp. 646–652. [Google Scholar]
  4. Li, S.; Zheng, W.; Zong, Y.; Lu, C.; Tang, C.; Jiang, X.; Liu, J.; Xia, W. Bi-modality fusion for emotion recognition in the wild. In Proceedings of the 2019 International Conference on Multimodal Interaction, Suzhou, China, 14–18 October 2019; pp. 589–594. [Google Scholar]
  5. Song, T.; Zheng, W.; Song, P.; Cui, Z. EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans. Affect. Comput. 2018, 11, 532–541. [Google Scholar] [CrossRef]
  6. Song, P. Transfer Linear Subspace Learning for Cross-Corpus Speech Emotion Recognition. IEEE Trans. Affect. Comput. 2019, 10, 265–275. [Google Scholar] [CrossRef]
  7. Shami, M.; Verhelst, W. Automatic classification of expressiveness in speech: A multi-corpus study. In Speaker Classification II; Springer: Berlin/Heidelberg, Germany, 2007; pp. 43–56. [Google Scholar]
  8. Schuller, B.; Vlasenko, B.; Eyben, F.; Wöllmer, M.; Stuhlsatz, A.; Wendemuth, A.; Rigoll, G. Cross-corpus acoustic emotion recognition: Variances and strategies. IEEE Trans. Affect. Comput. 2010, 1, 119–131. [Google Scholar] [CrossRef]
  9. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
  10. Zong, Y.; Zheng, W.; Zhang, T.; Huang, X. Cross-corpus speech emotion recognition based on domain-adaptive least-squares regression. IEEE Signal Process. Lett. 2016, 23, 585–589. [Google Scholar] [CrossRef]
  11. Song, P.; Zheng, W.; Ou, S.; Zhang, X.; Jin, Y.; Liu, J.; Yu, Y. Cross-corpus speech emotion recognition based on transfer non-negative matrix factorization. Speech Commun. 2016, 83, 34–41. [Google Scholar] [CrossRef]
  12. Song, P.; Zheng, W. Feature selection based transfer subspace learning for speech emotion recognition. IEEE Trans. Affect. Comput. 2018, 11, 373–382. [Google Scholar] [CrossRef]
  13. Zhang, J.; Jiang, L.; Zong, Y.; Zheng, W.; Zhao, L. Cross-Corpus Speech Emotion Recognition Using Joint Distribution Adaptive Regression. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 3790–3794. [Google Scholar]
  14. Abdelwahab, M.; Busso, C. Domain adversarial for acoustic emotion recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 2423–2435. [Google Scholar] [CrossRef]
  15. Gideon, J.; McInnis, M.G.; Provost, E.M. Improving cross-corpus speech emotion recognition with adversarial discriminative domain generalization (ADDoG). IEEE Trans. Affect. Comput. 2019, 12, 1055–1068. [Google Scholar] [CrossRef]
  16. Lu, C.; Zong, Y.; Zheng, W.; Li, Y.; Tang, C.; Schuller, B.W. Domain Invariant Feature Learning for Speaker-Independent Speech Emotion Recognition. IEEE/ACM Trans. Audio, Speech, Lang. Process. 2022, 30, 2217–2230. [Google Scholar] [CrossRef]
  17. Hassan, A.; Damper, R.; Niranjan, M. On acoustic emotion recognition: Compensating for covariate shift. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 1458–1468. [Google Scholar] [CrossRef]
  18. Mao, Q.; Xu, G.; Xue, W.; Gou, J.; Zhan, Y. Learning emotion-discriminative and domain-invariant features for domain adaptation in speech emotion recognition. Speech Commun. 2017, 93, 1–10. [Google Scholar] [CrossRef]
  19. Zhang, S.; Zhang, S.; Huang, T.; Gao, W. Speech emotion recognition using deep convolutional neural network and discriminant temporal pyramid matching. IEEE Trans. Multimed. 2017, 20, 1576–1590. [Google Scholar] [CrossRef]
  20. Plutchik, R. A general psychoevolutionary theory of emotion. In Theories of Emotion; Elsevier: Amsterdam, The Netherlands, 1980; pp. 3–33. [Google Scholar]
  21. Russell, J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980, 39, 1161. [Google Scholar] [CrossRef]
  22. Yang, L.; Shen, Y.; Mao, Y.; Cai, L. Hybrid Curriculum Learning for Emotion Recognition in Conversation. arXiv 2021, arXiv:2112.11718. [Google Scholar] [CrossRef]
  23. Rakshit, S.; Banerjee, B.; Roig, G.; Chaudhuri, S. Unsupervised multi-source domain adaptation driven by deep adversarial ensemble learning. In German Conference on Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2019; pp. 485–498. [Google Scholar]
  24. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 443–450. [Google Scholar]
  25. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, USA, 8–13 December 2014; Volume 27. [Google Scholar]
  26. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics—JMLR Workshop and Conference Proceedings, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  27. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Deep transfer learning with joint adaptation networks. In Proceedings of the International Conference on Machine Learning (PMLR), Sydney, Australia, 6–11 August 2017; pp. 2208–2217. [Google Scholar]
  28. Martin, O.; Kotsia, I.; Macq, B.; Pitas, I. The eNTERFACE’05 audio-visual emotion database. In Proceedings of the 22nd IEEE International Conference on Data Engineering Workshops (ICDEW’06), Atlanta, GA, USA, 3–7 April 2006; p. 8. [Google Scholar]
  29. Zhang, J.T.F.L.M.; Jia, H. Design of speech corpus for mandarin text to speech. In The Blizzard Challenge 2008 Workshop; International Speech Communication Association: Baixas, France, 2008; pp. 1–4. [Google Scholar]
  30. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.F.; Weiss, B. A database of German emotional speech. In Proceedings of the Interspeech, Lisbon, Portugal, 4–8 September 2005; pp. 1517–1520. [Google Scholar]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25 (NIPS 2012), Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  34. Zhu, Y.; Zhuang, F.; Wang, J.; Ke, G.; Chen, J.; Bian, J.; Xiong, H.; He, Q. Deep subdomain adaptation network for image classification. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1713–1722. [Google Scholar] [CrossRef]
  35. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  36. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. The overview of the progressively discriminative transfer network (PDTN) for cross-corpus SER. PDTN firstly extracts the high-dimensional fc layer features (i.e., f 1 , f 2 , and f 3 ) of the source domain data x s and the target domain data x t through DNN (i.e., AlexNet and VGGNet). Then, it uses fc features to calculate the valence-aware center loss L v and emotion-aware center loss L c in emotion discriminant loss L d , and distribution alignment loss L a , respectively. Finally, it predicts the emotion label of source samples for the emotion classification cross-entropy loss L c e .
Figure 1. The overview of the progressively discriminative transfer network (PDTN) for cross-corpus SER. PDTN firstly extracts the high-dimensional fc layer features (i.e., f 1 , f 2 , and f 3 ) of the source domain data x s and the target domain data x t through DNN (i.e., AlexNet and VGGNet). Then, it uses fc features to calculate the valence-aware center loss L v and emotion-aware center loss L c in emotion discriminant loss L d , and distribution alignment loss L a , respectively. Finally, it predicts the emotion label of source samples for the emotion classification cross-entropy loss L c e .
Entropy 24 01046 g001
Figure 2. The representation of seven emotions on the 2-dimensional arousal–valence emotion wheel.
Figure 2. The representation of seven emotions on the 2-dimensional arousal–valence emotion wheel.
Entropy 24 01046 g002
Figure 3. The visualization of feature distributions under different emotions generated by three fc layers (i.e., f c 1 , f c 2 , and f c 3 ) of PDTN (VGGNet-11) for the task of e b .
Figure 3. The visualization of feature distributions under different emotions generated by three fc layers (i.e., f c 1 , f c 2 , and f c 3 ) of PDTN (VGGNet-11) for the task of e b .
Entropy 24 01046 g003
Table 1. Data statistics of six cross-corpus SER tasks on three public datasets, where e, c, and b represent eNTERFACE, CASIA, and Emo-DB, respectively.
Table 1. Data statistics of six cross-corpus SER tasks on three public datasets, where e, c, and b represent eNTERFACE, CASIA, and Emo-DB, respectively.
TaskDataset (# Total Number)Emotion Category (# Samples of Each Emotion)
b  e, e → bb  (375)anger (127),  disgust (46),  fear (69),  happiness (71),  sadness (62)
e  (1052)anger (211),  disgust (211),  fear (211),  happiness (208),  sadness (211)
b  c, c → bb  (408)anger (127),  fear (69),  happiness (71),  neutral (79),  sadness (62)
c  (1000)anger (200),  fear (200),  happiness (200),  neutral (200),  sadness (200)
c → e, e → cc  (1052)anger (200),  fear 200,  happiness (200),  sadness (200),  surprise (200)
e  (1000)anger (211),  fear (211),  happiness (208),  sadness (211),  surprise (211)
Table 2. The experimental results (WAR/UAR[%]) compared with the state-of-the-art methods on CASIA, eNTERFACE, and Emo-DB for cross-corpus SER tasks, where the best results are highlighted in bold.
Table 2. The experimental results (WAR/UAR[%]) compared with the state-of-the-art methods on CASIA, eNTERFACE, and Emo-DB for cross-corpus SER tasks, where the best results are highlighted in bold.
Methode → bb → eb → cc → be → cc → eAverage
AlexNet [31]42.40/31.0329.56/29.4932.90/32.9043.13/42.2327.60/27.6026.33/26.3033.65/31.59
VGGNet-11 [32]44.26/43.2330.70/30.7035.10/35.1044.36/38.9528.80/28.8029.65/29.6035.48/34.40
DAN [27]49.82/40.4136.12/36.1339.00/39.0050.98/49.8529.00/29.0031.46/31.4739.89/37.64
DANN [35]52.80/43.6833.27/33.3839.20/39.2054.16/53.7129.80/29.8029.24/29.2539.62/38.05
Deep CORAL [24]53.07/43.3835.07/35.0338.30/38.3050.73/48.2831.00/31.0030.89/30.8939.84/37.81
DSAN [34]52.16/46.9036.29/36.2540.30/40.3051.81/50.6929.70/29.7032.61/32.6140.47/39.41
PDTN (AlexNet)54.60/47.1238.30/38.3242.80/42.8057.59/57.2135.10/35.1035.50/35.5043.99/ 42.70
PDTN (VGGNet-11)56.80/54.4838.49/38.6044.70/44.6062.01/61.6535.20/35.2035.74/35.4345.49/44.99
Table 3. Ablation experiments of PDTN based on the VGGNet-11 backbone network, where the best results (WAR/UAR[%]) are highlighted in bold.
Table 3. Ablation experiments of PDTN based on the VGGNet-11 backbone network, where the best results (WAR/UAR[%]) are highlighted in bold.
Methode → bb → eb → cc → be → cc → eAverage
PDTN_S w/o L c & L v 52.80/50.4035.83/35.8140.20/40.2055.39/54.8534.10/34.1034.03/33.9742.05/41.54
PDTN_M w/o L c & L v 53.00/51.0736.31/36.3641.60/41.5058.33/54.5533.80/33.8034.88/34.7742.99/42.01
PDTN_M w/o L v 54.44/51.5638.02/37.9744.00/43.9059.06/58.6634.60/34.6035.93/35.6744.34/43.31
PDTN_M56.80/54.4838.49/38.6044.70/44.6062.01/61.6535.20/35.2035.74/35.4345.49/44.99
PDTN_S54.66/51.8736.43/36.3244.40/44.4057.84/ 56.5334.50/34.5035.45/35.1443.88/43.12
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, C.; Tang, C.; Zhang, J.; Zong, Y. Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition. Entropy 2022, 24, 1046. https://doi.org/10.3390/e24081046

AMA Style

Lu C, Tang C, Zhang J, Zong Y. Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition. Entropy. 2022; 24(8):1046. https://doi.org/10.3390/e24081046

Chicago/Turabian Style

Lu, Cheng, Chuangao Tang, Jiacheng Zhang, and Yuan Zong. 2022. "Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition" Entropy 24, no. 8: 1046. https://doi.org/10.3390/e24081046

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop