Next Article in Journal
Human Activity Recognition with an HMM-Based Generative Model
Next Article in Special Issue
Roman Urdu Hate Speech Detection Using Transformer-Based Model for Cyber Security Applications
Previous Article in Journal
Personalized LSTM Models for ECG Lead Transformations Led to Fewer Diagnostic Errors Than Generalized Models: Deriving 12-Lead ECG from Lead II, V2, and V6
Previous Article in Special Issue
Collaborative Learning Based Sybil Attack Detection in Vehicular AD-HOC Networks (VANETS)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human–Computer Interaction with a Real-Time Speech Emotion Recognition with Ensembling Techniques 1D Convolution Neural Network and Attention

College of Engineering, Al Faisal University, P.O. Box 50927, Riyadh 11533, Saudi Arabia
Sensors 2023, 23(3), 1386; https://doi.org/10.3390/s23031386
Submission received: 23 November 2022 / Revised: 19 January 2023 / Accepted: 20 January 2023 / Published: 26 January 2023

Abstract

:
Emotions have a crucial function in the mental existence of humans. They are vital for identifying a person’s behaviour and mental condition. Speech Emotion Recognition (SER) is extracting a speaker’s emotional state from their speech signal. SER is a growing discipline in human–computer interaction, and it has recently attracted more significant interest. This is because there are not so many universal emotions; therefore, any intelligent system with enough computational capacity can educate itself to recognise them. However, the issue is that human speech is immensely diverse, making it difficult to create a single, standardised recipe for detecting hidden emotions. This work attempted to solve this research difficulty by combining a multilingual emotional dataset with building a more generalised and effective model for recognising human emotions. A two-step process was used to develop the model. The first stage involved the extraction of features, and the second stage involved the classification of the features that were extracted. ZCR, RMSE, and the renowned MFC coefficients were retrieved as features. Two proposed models, 1D CNN combined with LSTM and attention and a proprietary 2D CNN architecture, were used for classification. The outcomes demonstrated that the suggested 1D CNN with LSTM and attention performed better than the 2D CNN. For the EMO-DB, SAVEE, ANAD, and BAVED datasets, the model’s accuracy was 96.72%, 97.13%, 96.72%, and 88.39%, respectively. The model beat several earlier efforts on the same datasets, demonstrating the generality and efficacy of recognising multiple emotions from various languages.

1. Introduction

Emotion is a complex phenomenon that is influenced by numerous circumstances. Charles Darwin [1], one of the earliest scientists to study emotions, saw emotional expression as the last of behavioural patterns that had become obsolete owing to evolutionary change. In [2] theory of emotion was only a partial assertion and that feeling still serves an essential purpose; it is simply the nature of this purpose that has developed. Emotions are occasionally felt when something unexpected occurs, something to which evolution has not yet adapted, and in the circumstances such as these, emotional consequences start to take control. It is well-recognised that people’s emotional states can cause physical changes in their bodies. Emotions, for instance, may affect voice alterations [1]. Therefore, speech signals, which account for 38% of all emotional communication, can recognise and convey emotions [3]. The sound signals contain some elements that represent the speaker’s emotional state and those that correspond to the speaker and the speech. As a result, the fundamental idea behind emotion detection is examining the acoustic difference that arises while pronouncing the same thing in various emotional contexts [4]. Voice signals can be picked up without any device connected to the individual, even if alterations cannot be recognised without a portable medical instrument. Due to this, most studies on the subject have concentrated on automatically identifying emotions from auditory cues [5].
In human–computer interaction (HCI) and its applications, automatic emotion recognition is crucial since it can be a powerful feedback mechanism [6,7]. The main tools used in traditional HCI include the keyboard, mouse, screen, etc. It cannot comprehend and adjust to people’s emotions or moods; it just seeks convenience and precision. It is not easy to expect the computer to have the same intellect as people if it cannot recognise and convey emotions. Furthermore, it is challenging to anticipate that HCI will be genuinely harmonious and natural. People naturally expect computers to have emotional skills in the process of HCI since human interactions and communication are natural and emotional. The goal of affective computing is to give computers the capacity to perceive, comprehend, and produce a variety of emotional qualities similar to those found in humans. This will eventually allow computers to engage with humans in a way that is natural, friendly, and vivid [8].
Applications include human-robot communication, where machines react to people based on the emotions they have been programmed to recognise [9], implementation in call centres to identify the caller’s emotional state in emergencies, determining the degree of a customer’s satisfaction, medical analysis, and education. A conversational chatbot is another suggestion for an emotion recognition application, where real-time SER applications can improve dialogue [10]. A real-time SER should determine the best compromise between little computer power, quick processing times, and excellent accuracy.
But one of SER’s most significant yet elusive tasks has been identifying the “best” or most distinctive acoustic qualities that describe certain emotions. Despite thorough investigation, the pace of improvement was modest, and there were some discrepancies between studies. Due to these factors, research has shifted toward techniques that do away with or drastically minimise the requirement for prior knowledge of “optimal features” in favour of the autonomous feature-generating processes provided by neural networks [11]. Both passes employ cutting-edge classifiers, notably convolutional neural networks (CNN) and DNN [12,13,14].
Additionally, emotion identification may distinguish between feelings conveyed in only one language or across several languages. However, multilingual emotion recognition is still a new research topic, although many papers on monolingual emotion recognition have been published [15,16]. Therefore, using English, German, and Arabic corpora, extensive experiments and analyses of multilingual emotion recognition based on speech are given in the current study.
The contributions of the study are as follows:
  • In this paper, two deep learning models for the SER are proposed: a 1D CNN with LSTM attention and a 2D CNN with multiple layers employing modified kernels and a pooling strategy to detect the sensitive cues-based input extracted features, which tend to be more discriminative and dependable in speech emotion recognition.
  • The models above were constructed using varied datasets from multiple languages to make the models more generalisable across language boundaries in analysing emotions based on the vibrational patterns in speech.
  • The performance of the proposed 1D CNN with the LSTM attention model is superior to that of the prior attempts on the various speech emotion datasets, suggesting that the proposed SER scheme will contribute to HCI.
The paper is organized as follows: the introduction is in Section 1, followed by the literature review in Section 2, then the dataset details, pre-processing, feature extraction, and methodology in Section 3, followed by the results in Section 4 and discussion in Section 5, and the conclusions is in Section 6, followed by the references.

2. Literature Review

Numerous studies have been done on extracting emotions from aural information. The SER system had two essential processes in standard Machine Learning techniques: manual feature extraction and emotion classification. This section discusses existing techniques that leverage various real-time speech emotion recognition datasets and feature extraction techniques available in the literature. Some of the existing models for real-time speech emotion recognition are also discussed in this section.
A flexible emotion recognition system built on the analysis of visual and aural inputs was proposed by [17] in 2017. Mel Frequency Cepstral Coefficients (MFCC) and Filter Bank Energies (FBEs) were two features used in his study’s feature extraction stage to lower the dimension of previously derived features. The trials used the SAVEE dataset and had a 65.28 percent accuracy rate.
In their suggested method [18], also used the spectral characteristics, or the 13 MFCC, to categorise the seven emotions with a 70% accuracy rate using the Logistic Model Tree (LMT) algorithm. Results from experiments using the Ryerson Audio-Visual Database of Emotional Speech and Song and the Berlin Database of Emotional Speech (EmoDB) (RAVDESS).
By integrating three models or experts, each focused on different feature extraction and classification strategy [19], achieved ensemble learning. The study used the IEMOCAP corpus, and a confidence calculation approach was suggested to overcome the IEMOCAP corpus’s data imbalance issue. The study found the key areas of crucial local characteristics work in conjunction with the attention mechanism to exploit each expert’s job from many aspects fully. Seventy-five percent accuracy was achieved [20].
The authors in [21] developed a new, lightweight SER model with a low computing overhead and good identification accuracy. The proposed methodology uses a straightforward rectangular filter with a modified pooling strategy that improves SER discriminative performance, as well as a Convolutional Neural Network (CNN) method to learn the deep frequency features. The proposed CNN model was trained using frequency features extracted from the voice data, and its ability to predict emotions was then assessed. The benchmark speech datasets for the proposed SER model were the interactive emotional dyadic motion capture (IEMOCAP) and the Berlin emotional speech database (EMO-DB). The evaluation results were 77.01 and 92.02 percent recognition, respectively.
A novel approach for SER based on the Bidirectional Long Short-Term Memory with Attention (BLSTMwA) model and Deep Convolutional Neural Network (DCNN) was developed by [8]. (DCNN-BLSTMwA). The speech samples are initially preprocessed using data improvement and dataset balancing. Second, as input for the DCNN, three-channel log Mel-spectrograms (static, delta, and delta-delta) are extracted. The segment-level features are then produced using the DCNN model that has already been trained on the ImageNet dataset. The unweighted average recall (UAR) for experiments using the EMO-DB and IEMOCAP databases was 87.86 and 68.50 percent, respectively [22,23,24].
In this work [25] used recurrent neural network (RNN) designs in their research. Their suggested model extracts relationships from 3D spectrograms across time steps and frequencies by combining parallel convolutional layers (PCN) with a squeeze-and-excitation network (SEnet), also known as PCNSE. Additionally, they used the Interactive Emotional Dyadic Motion Capture (IEMOCAP) and FAU-Aibo Emotion Corpus to demonstrate the viability of the suggested approach (FAU-AEC). On the IEMOCAP dataset, they successfully attained a weighted accuracy (WA) of 73.1 percent, an unweighted accuracy (UA) of 66.3 percent, and a UA of 41.1 percent on the FAU-AEC dataset [26,27,28].
Deep Convolutional Recurrent Neural Network was used by [29] to create an ensemble classifier for speech emotion recognition (SER). This paper introduces a novel method for SER tasks inspired by current work on speech emotion detection. They obtained 3-D log Mel-spectrograms of utterance-level log Mel-spectrograms along with their first and second derivatives (Static, Delta, and Delta-delta). They use deep convolutional neural networks to extract the deep features from 3-D log Mel-spectrograms and deep convolutional neural networks [26]. An utterance-level emotion is then produced by applying a bi-directional-gated recurrent unit network to express long-term temporal dependency among all features. An ensemble classifier employing Softmax and a Support Vector Machine classifier is used to increase the overall recognition rate [27]. On the RAVDESS (Eight emotional states) and Odia (Seven emotional states) datasets, the suggested framework is trained and tested. The experiment’s outcomes show that an ensemble classifier outperforms a single classifier in terms of performance. The degrees of accuracy attained are 77.54 percent and 85.31 percent [30,31].
A pre-trained audio-visual Transformer for understanding human behaviour was introduced by [32]. It was trained on more than 500k utterances from over 4000 celebrities in the VoxCeleb2 dataset. The model’s application in emotion recognition tries to capture and extract relevant information from the interactions between human facial and auditory activities. Two datasets, CREMAD-D (emotion classification) and MSP-IMPROV were used to test the model (continuous emotion regression). According to experimental findings, fine-tuning the pre-trained model increases the consistency correlation coefficients (CCC) in continuous emotion detection by 0.03–0.09 and the accuracy of emotion categorisation by 5–7 percent compared to the same model generated from scratch [33,34,35].
We recognize a pattern in the adoption of deep convolutional models that can teach from spectrogram descriptions of speech, per the literature study. There are numerous databases accessible, therefore choosing which databases to use for better training and validation is a difficult process. Although little research has been done on it, the attention mechanism can enhance the performance of SER systems. As a result, we suggested two models in this paper: the 1D CNN with LSTM attention model and the 2D CNN model. The results demonstrate that the proposed models outperformed the existing models in the literature. The brief description of the proposed model is present in the subsequent section.

3. Materials and Methods

This work aims to classify emotions based on the voice signal. The proposed methodology is illustrated in Figure 1. Based on the findings of individual studies. One cannot definitively state which classifier is superior for emotion recognition. The quality of the data has a direct bearing on the performance of the classifier. This is because the accuracy shifts depending on the characteristics of the data, such as the quantity, and density distribution of each class (emotions), and language as well [36]. In addition, the model can be trained using the features derived from the audio waves rather than feeding raw audio into a classifier. Since the features are more particular to the issue at hand, the model can be executed more effectively using the features that have been retrieved.
As shown in Figure 1, the first stage involved the extraction of features from the available labelled datasets, and the second stage involved the classification of the features that were extracted. ZCR, RMSE, and the renowned MFC Coefficients were retrieved as features. Then, two proposed models, 1D CNN combined with LSTM and attention and a proprietary 2D CNN architecture, were used for classification. After that, the evaluation of the performance of the proposed models is presented.

3.1. Datasets

This section briefly describes the datasets such as Arabic Natural Audio Dataset (ANAD) dataset, Basic Arabic Vocal Emotions Dataset (BAVED), The Surrey Audio-Visual Expressed Emotion (SAVEE) database, and The Berlin emotional speech database (EMO-DB).

3.1.1. Arabic Natural Audio Dataset

The (ANAD) [37] dataset is comprised of three distinct feelings, namely, angry (A), surprised (S), and happy (H). The audio recordings that correlate to these feelings were collected from eight videos of live calls between an anchor and a human located outside the studio. Additionally, the audio files were taken from online discussion programs in Arabic. Eighteen manual labourers contributed to the creation of the ground truth for the videos that were collected. First, they assigned emotions such as anger, surprise, and happiness to each film. The final label is determined by taking the average of the ratings given by all eighteen of the labellers for each video. After that, each film is cut into sections titled “callers” and “receivers”. Next, the areas with laughter, noise, and silence were eliminated. After that, each chunk was automatically broken down into speech units that lasted one second. Figure 2 presents the results of the quantitative analysis performed on the corpus. Figure 2a demonstrates the distribution of emotions in the ANAD dataset. This figure clearly indicates the count of emotions in three distinct feelings, namely, angry (A), surprised (S), and happy (H).

3.1.2. Basic Arabic Vocal Emotions Dataset

Basic Arabic Vocal Emotions Dataset (BAVED) [38] is an audio/wave dataset including Arabic words spelled with varying levels of conveyed emotions. The dataset consisted of the recorded forms of seven Arabic words (0—like, 1—unlike, 2—this, 3—file, 4—good, 5—neutral, and 6—bad) across three emotional degrees. Level 0 is when the speaker expresses a low level of emotion, comparable to feeling exhausted or depressed. Figure 2b demonstrates the distribution of emotions in the BAVED dataset. Level 1 is the average level, where the speaker expresses neutral feelings; level 2 is when the speaker expresses strong positive or negative emotions (happy, joy, grief, wrath, etc.). The collection consists of 1935 recordings made by 61 speakers (45 males and 16 females).

3.1.3. Surrey Audio-Visual Expressed Emotion Database

The Surrey Audio-Visual Expressed Emotion (SAVEE) database [39] has been recorded as a prerequisite for creating an automatic emotion identification system. Figure 2c demonstrates the distribution of emotions in the SAVEE dataset. The collection contains recordings of four male actors expressing seven distinct emotions: sadness, neutral, frustration, happiness, disgust, and anger. The dataset provided 480 British English utterances in total. The sentences were selected from the standard TIMIT corpus and balanced phonetically for each emotion. The data were recorded, analysed, and labelled using high-quality audio-visual equipment at a lab for visual media.

3.1.4. Berlin Database of Emotional Speech

The Berlin emotional speech database, also known as EMO-DB [9], is a public database that includes information on seven different emotional states. These states are anger, boredom, disgust, fear, happiness, neutral, and sadness. Verbal contents originate from 10 German (5 males and 5 females) pre-defined neutral utterances. It was decided to have ten professional actors read each utterance in each of the seven emotional states. Figure 2d demonstrates the distribution of emotions in the EMO-DB dataset which shows the count of emotions in the seven distinct types of feelings. The EMO-DB includes around 535 phrases representing each of the seven emotions. The sound files were captured at a sample rate of 16 kHz, a resolution of 16 bits, and a single channel. The length of time for each audio file is, on average, three seconds.

3.2. Feature Extraction

In the feature extraction process, the system extracts as much information as possible from each piece of audio data. For example, the current work extracts the three most relevant features ZCR, RMSE, and MFCC.

3.2.1. Zero Crossing Rate (ZCR)

The zero-crossing rate indicates the frequency at which the signal crosses the zero-amplitude level. Equation (1) provides the calculation of ZCR
ZCR = 1 T t = 1 T s t s t 1
where s(t) = 1 if the signal has a positive amplitude at t, otherwise 0.

3.2.2. Root Mean Square Error (RMSE)

It is the measure of the energy content in the signal. It is the measure of the energy content in the signal (Equation (2))
R M S t = 1 K k = t . K t + 1 . K 1 s k 2

3.2.3. Mel Frequency Cepstral Coefficients

Studies in the field of psychophysics have demonstrated that the way humans perceive the sound frequency contents of speech signals does not follow a linear scale. Therefore, a subjective pitch is assessed on a scale referred to as the ‘Mel’ scale. This scale is used for tones with an actual frequency, f, measured in hertz. The conversion of the actual frequency f to mel scale is given in Equation (3).
m e l = 1127   l o g   1 + f / 700
Mel frequency cepstral coefficients, often known as MFCC, are the individual coefficients that, when added up, form a MFC. They are a form of the cepstral portrayal of the audio sample, where they originated (a nonlinear “spectrum-of-a spectrum”). The typical method for deriving MFCCs is illustrated in Figure 3. Following the pre-processing stage, the speech frame will go through a hamming window, and then, a rapid Fourier transformation will be used to determine the energy distribution. To get rid of the influence that harmonics have, a Mel filter bank is utilised. The discrete cosine transformation is the last step in the process. In a basic MFCC, the first difference parameters, the second difference parameters, and energy, respectively, make up an N-dimensional MFCC feature.

3.3. Model Development

The methods of classification that were employed are discussed in the following subsections. To improve SER, the 1D CNN with LSTM and attention layers were implemented. In addition to that, a 2D CNN is also demonstrated here. The models were trained using the ZCR, RMSE, and MFCC features retrieved from the data.

3.3.1. One-Dimensional Convolution Neural Network with Long Short-Term Memory and Attention

As shown in Figure 4, the suggested model consists of five blocks of convolution operations and two completely connected layers. The input to the convolution was retrieved features from the unprocessed audio files that were scaled to fit into 1D convolutional layers. The amount of input varies depending on the utilised dataset. In addition, the weights of the convolution layer must be taught, whereas the pooling layers employ a fixed function to translate the activation. The Rectified Linear Unit (ReLU) is utilised to offer non-linearity to the entire network without affecting the receptive fields of the convolutional layer. Loops improve the effectiveness of the convolution layer. In addition, the model output is produced using a training dataset that includes the loss function, and the learning parameters (kernels and weights) are updated using backpropagation with the loss. Finally, the result is pooled, essentially a nonlinear spatial down-sampling procedure. Pooling reduces the representation’s spatial size, which helps decrease parameters and calculations, preventing overfitting.
In the second half, CNN’s output is sent into LSTM, determining the more profound temporal link between features. An additional attention layer was a way of acquiring additional information pertinent to each class. Lastly, the pooled feature map is remapped using two layers that are fully connected. Finally, the Softmax layer provided the likelihood of each class’s prediction.

3.3.2. Two Dimensional Convolution Neural Network (CNN)

The structure of the 2D CNN was similar to the common CNNs for the image classification tasks, with the convolution and pooling layers alternatively appearing. However, the number of the convolution and the pooling layers were less compared to the 1D CNN. This is because the primary audio signals were 1D, and the extracted features were 1D, but in the 2D CNN presented in Figure 5, the convolution kernel, feature map, and other network structures are 2D. Therefore, to process a 1D signal with 2D CNNs, the 1D signal is usually mapped to a 2D space. Then, these 2D features are input into the conventional 2D CNNs for further processing. Finally, an additional dropout layer is added for better training and to avoid overfitting the model.

3.4. Performance Evaluation

The most common measures that are used to evaluate model performance are presented in this section. Accuracy alone will not be enough; additional metrics such as precision, recall, and the F1-score evaluation criteria will be necessary. The F1- measure is the harmonic mean of the precision, as described by definition [40].

4. Results

The experiments conducted to validate the effectiveness of our suggested 1D CNN with the LSTM attention model are shown in this section. To avoid making observations based solely on a corpus examination, four different well-known databases were selected to demonstrate the effectiveness of the suggested approach. An ablation study was first conducted to clarify the advantages of adding LSTM and an attention mechanism to the 1D CNN architecture. A comparison of the created framework’s effectiveness with the outcomes of the designed 2D CNN architecture on the well-known databases ANAD, BAVED, SAVEE, and EMO-DB further emphasises its efficiency.
From the audio samples of each database, the Zero Crossing Rate (ZCR), Root Mean Square Error (RMSE), and Mel-Frequency Cepstral Coefficients (MFCC) features were extracted. The ZCR, which is measured as the number of zero crossings in the temporal domain during one second, is one of the most affordable and basic characteristics. One of the methods most frequently used to assess the accuracy of forecasts is RMSE, also known as the root mean square deviation. It illustrates the Euclidean distance between measured true values and forecasts. The windowing of the signal, application of the DFT, calculation of the magnitude’s log, warping of the frequencies on a Mel scale, and application of the inverse DCT are the main steps in the MFCC feature extraction approach.
Before feature extraction, the data were also enhanced with noise and pitch and provided for model training. The model per epoch evaluation of the model training reveals that the proposed models raise the degree of accuracy and decrease losses for the training and testing datasets, indicating the model’s importance and efficacy. However, some models were stopped early before reaching epoch 50 as the validation accuracy has not improved much. The suggested CNN model’s visual output is depicted in Figure 6, while the confusion matric acquired are displayed in Figure 7 followed and Figure 8 for Loss and accuracy plots. It displays the results of training the 1D CNN with LSTM attention and 2D CNN models for the four spoken emotion datasets. This indicates the training and validation accuracy and the loss for the BAVED, ANAD, SAVEE, and EMO-DB datasets. Figure 6 presents the loss and accuracy plots during the training process of the 1D CNN with LSTM attention for the different datasets of real-time speech emotion recognition.
The suggested model is analysed, and the analysis results are presented in Table 1, while the confusion matric acquired are displayed in Figure 9. The AUC-ROC curve, recall, precision, and accuracy of the proposed deep learning models can be measured with the use of a confusion matrix. Our model demonstrates accuracy, precision, and recall, as well as the F1-score, which reflects the model’s resilience straightforwardly and concisely. In addition, the performance of the 1D CNN model with LSTM attention layers is much better than that of the 2D CNN model.
Figure 8 presents the loss and accuracy plots during the training process of the 2D CNN for the different datasets of real-time speech emotion recognition.
The confusion matrix in Figure 7 and Figure 9 and Table 2, Table 3, Table 4 and Table 5 clearly explains the model performance in detecting individual emotions on the test dataset.
The 1D CNN with the LSTM attention model achieved the highest level of performance when applied to the ANAD dataset. It was clear from Figure 7a and Table 2 that the best individual performance was achieved for the feeling “Angry” also using the 2D CNN model (Figure 9a; Table 2), but the performance of the model 1D CNN with LSTM was quite acceptable on the “Surprised” emotion. In contrast, the performance of the 2D CNN was much worse for the “Surprised” emotion state, which greatly damaged the model’s overall accuracy.
Compared to the ANAD dataset, the performance of the BAVED dataset was significantly lower, even though it was also an Arabic dataset. The two distinct models each had a distinct impact on the performance of their respective particular classes. The “Neutral” emotion class achieved the highest performance when the 1D CNN model was evaluated. However, the performance of the 2D CNN is superior for the “Low” emotion class (Table 3; Figure 7b and Figure 9b).
For the SAVEE datasets (Table 4; Figure 7c and Figure 9c), the performance of predicting “Happy”, “Surprised”, and “Neutral” emotions was highest with a 1D CNN model. In contrast, the performance of predicting these emotions with a 2D CNN model was very much lower. The performance was only improved for the emotion “Neutral” option, and even then, it was only 77% recall. Compared to the inferior emotion (“Fear”, with a recall of 95 percent) that could be predicted using 1D CNN, this prediction was much less accurate.
The trend seen in SAVEE can also be seen in the German dataset -EMO-DB. The performance of the 1D CNN with LSTM attention demonstrated a significant and noticeable difference. The emotions that performed the best with 1D CNN were “Disgust” and “Anger”. “Anger” still predominates on 2D CNN, but “Disgust” is relatively low. “Sad” performed better (Table 5; Figure 7d and Figure 9d).

5. Discussion

Speech Emotion Recognition is one of the contributing factors to human–computer interaction. The human brain processes the multimodalities to extract the spatial and temporal semantic information that is contextually meaningful to perceive and understand an individual’s emotional state. However, for machines, even though human speech contains a wealth of emotional content, it is challenging to predict the emotions or the speaker’s intention hidden in their voice [41].

5.1. Dataset Variability

For the SER challenge, several of the earlier efforts only included a single language dataset [42] and nevertheless managed to obtain excellent performance on that dataset. Despite its outstanding performance, the model was constructed using only one language, limiting its ability to comprehend various feelings. Due to this, the evaluation of the potential for generalisation, and consequently the real-world impact of the different techniques, is effectively constrained [43]. Robust embeddings for speech emotion identification should perform well across various languages to act as a paralinguistic cue [44]. However, human speech is so diverse and dynamic that no model can be reserved to be used forever [45].
Additionally, the diversity of languages causes an imbalance of available datasets for emotion recognition for minority languages compared to well-established majority languages such as English. This is particularly true for languages that are spoken by a smaller population. Despite this, it is necessary to develop a generalised model for multilingual emotional data for us to use the currently available datasets. The proposed model was validated by testing it on four datasets spanning three distinct languages, each with a different dataset size and a different category of emotions. Even though the performance was not very good on some datasets, the model tried to achieve a generalisation across different sets of emotions expressed in different languages.

5.2. Feature Extraction

The current work was designed similarly to the traditional ML pipeline, with the manually extracted features contributing to the model classification. ZCR, RMSE, and the MFCCs were the features extracted. Although no Feature Selection has been performed, MFCCs have been selected because they are the most common representation in traditional audio signal processing, but the most recent works used spectrogram images as the input for developing deep learning models [46]. Some of them used raw audio waveforms also [47]. Raw waveforms and spectrograms avoid hand-designed features, which should allow better to exploit the improved modelling capability of deep learning models, learning representations optimised for a task. However, this incurs higher computational costs and data requirements, and benefits may be hard to realise in practice. Handcrafted features, even though incorporating huge manual tasks, understanding the features contributing to the classification task would be more efficient than manual feature extraction, especially for complex audio recognition tasks such as emotion recognition. The proposed work outperformed the work on image spectrograms, with an accuracy of 86.92% for EMO-DB and 75.00% for SAVEE datasets [46] and 95.10% for EMO-DB and 82.10% for SAVEE [48] (Table 6).

5.3. Classification

In this study, CNNs are used to construct the particular model with additional LSTM with attention. For example, the speech signal is a time-varying signal which needs particular processing to reflect time-fluctuating features [49]. Therefore, the LSTM layer is introduced to extract long-term contextual relationships. The results revealed that the 1D CNN model produced with LSTM had outperformed the model developed with 2D CNN architecture for all the datasets included in the study (Table 1). [50] followed the same pipeline as our study, where they retrieved manual features from the audio, such as MFCCs. They even included the feature selection method, but the Linear SVM classifier was used, but with a comparatively lightweight model, they obtained an accuracy of 96.02%, which is much closer to our findings. However, the proposed model was inferior to the model produced by [51] using the BAVED database. In another work [52] this technique is also applied. They used wav2vec2.0 and HuBERT as the feature extractors and the classifier head as the Multi-layer Perception Classifier coupled to a Bi-LSTM Layer with 50 hidden units. The performance is as shown in Table 7, Table 8 and Table 9.
The current effort sought to construct an SER capable of generalising across the language boundaries in speech emotion detection. The performance of the model was satisfactory. However, there is still an opportunity for further progress in some languages. In the future study, more extensive datasets from various disciplines will be focused on to get a better SER model for a more productive human–computer interaction. The field of human–computer interaction, especially voice-based interaction, is developing and changing virtually every day. The current work aims to construct artificially intelligent systems based on deep learning for recognising human emotions based on many modalities. In this paper, a unique 1D CNN architecture with LSTM and attention layers was created, as well as a 2D CNN, to recognise emotions expressed in Arabic, English, and German utterances. Before starting the model training process, the MFCC, ZCR, and RMSE features were retrieved. The learning models were implemented, and then, the outcomes of these models were exhibited using the audio dataset available in three languages. The proposed 1D CNN design with LSTM and attention layers recognised emotions with more precision than the 2D CNN architecture. The model’s accuracy was achieved at 96.72%, 97.13%, 96.72%, and 88.39% for the EMO-DB, SAVEE, ANAD, and BAVED datasets.

5.4. Limitations

The current work is performing well in terms of results and accuracy compared to other works, but it is observed that it is a little complex. The performance complexity evaluation is a bit heavy, and it can be resolved. Therefore, in the upcoming future, this can be taken into consideration for the smooth performance of the system.

6. Conclusions

The field of human–computer interaction, especially voice-based interaction, is developing and changing virtually every day. The current work aims to construct artificially intelligent systems based on deep learning for recognising human emotions based on many modalities. In this paper, a unique 1D CNN architecture with LSTM and attention layers was created, as well as a 2D CNN, to recognise emotions expressed in Arabic, English, and German utterances. Before starting the model training process, the MFCC, ZCR, and RMSE features were retrieved. The learning models were implemented, and then, the outcomes of these models were exhibited using the audio dataset available in three languages. The results revealed that the 1D CNN model produced with LSTM outperformed the model developed with 2D CNN architecture for all the datasets included in the study. The performance of the model was satisfactory; additionally, the performance of the 2D CNN was superior for the “Low” emotion class. The proposed model outperformed the work on image spectrograms, with an accuracy of 86.92% for EMO-DB, 75.00% for SAVEE [46], 95.10% for EMO-DB, and 82.10% for SAVEE. The proposed 1D CNN design with LSTM and attention layers recognised emotions with more precision than the 2D CNN architecture. The model’s accuracy was achieved at 96.72%, 97.13%, 96.72%, and 88.39% for the EMO-DB, SAVEE, ANAD, and BAVED datasets.

Funding

The author received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available publicly and cited in proper places in the text.

Acknowledgments

The author would like to thank Al Faisal University, College of Engineering for all the support.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Darwin, C.; Prodger, P. The Expression of the Emotions in Man and Animals; Oxford University Press: New York, NY, USA, 1998. [Google Scholar]
  2. Oatley, K. The importance of being emotional. New Sci. 1989, 123, 33–36. [Google Scholar]
  3. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  4. Davletcharova, A.; Sugathan, S.; Abraham, B.; James, A.P. Detection and analysis of emotion from speech signals. Procedia Comput. Sci. 2015, 58, 91–96. [Google Scholar] [CrossRef] [Green Version]
  5. Harár, P.; Burget, R.; Dutta, M.K. Speech emotion recognition with deep learning. In Proceedings of the 2017 4th International conference on signal processing and integrated networks (SPIN), Noida, India, 2–3 February 2017; pp. 137–140. [Google Scholar]
  6. Busso, C.; Bulut, M.; Narayanan, S.; Gratch, J.; Marsella, S. Toward Effective Automatic Recognition Systems of Emotion in Speech. In Social Emotions in Nature and Artifact: Emotions in Human and Human-Computer Interaction; Gratch, J., Marsella, S., Eds.; Oxford University Press: New York, NY, USA, 2013; pp. 110–127. [Google Scholar]
  7. Kim, T.; Vossen, P. Emoberta: Speaker-aware emotion recognition in conversation with roberta. arXiv 2021, arXiv:2108.12009. [Google Scholar]
  8. Zhang, S.; Liu, R.; Tao, X.; Zhao, X. Deep Cross-Corpus Speech Emotion Recognition: Recent Advances and Perspectives. Front. Neurorobot. 2021, 15, 162. [Google Scholar] [CrossRef] [PubMed]
  9. Burkhardt, F.; Ajmera, J.; Englert, R.; Stegmann, J.; Burleson, W. Detecting Anger in Automated Voice Portal Dialogs. 2006. Available online: https://www.researchgate.net/profile/Joachim-Stegmann/publication/221478646_Detecting_anger_in_automated_voice_portal_dialogs/links/09e4150d2338e8a639000000/Detecting-anger-in-automated-voice-portal-dialogs.pdf (accessed on 22 December 2022).
  10. Deriche, M.; Absa, A.H.A. A two-stage hierarchical bilingual emotion recognition system using a hidden Markov model and neural networks. Arab. J. Sci. Eng. 2017, 42, 5231–5249. [Google Scholar] [CrossRef]
  11. Lech, M.; Stolar, M.; Best, C.; Bolia, R. Real-Time Speech Emotion Recognition Using a Pre-trained Image Classification Network: Effects of Bandwidth Reduction and Companding. Front. Comput. Sci. 2020, 2, 14. [Google Scholar] [CrossRef]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  13. Abdel-Hamid, O.; Mohamed, A.; Jiang, H.; Deng, L.; Penn, G.; Yu, D. Convolutional neural networks for speech recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1533–1545. [Google Scholar] [CrossRef] [Green Version]
  14. Shaukat, K.; Luo, S.; Chen, S.; Liu, D. Cyber Threat Detection Using Machine Learning Techniques: A Performance Evaluation Perspective. In Proceedings of the 2020 International Conference on Cyber Warfare and Security (ICCWS), Islamabad, Pakistan, 20–21 October 2020; pp. 1–6. [Google Scholar]
  15. PHeracleous, P.; Yoneyama, A. A comprehensive study on bilingual and multilingual speech emotion recognition using a two-pass classification scheme. PLoS ONE 2019, 14, e0220386. [Google Scholar]
  16. Javed, U.; Shaukat, K.; Hameed, I.A.; Iqbal, F.; Alam, T.M.; Luo, S. A review of content-based and context-based recommendation systems. Int. J. Emerg. Technol. Learn. 2021, 16, 274–306. [Google Scholar] [CrossRef]
  17. Noroozi, F.; Marjanovic, M.; Njegus, A.; Escalera, S.; Anbarjafari, G. Audio-visual emotion recognition in video clips. IEEE Trans. Affect. Comput. 2017, 10, 60–75. [Google Scholar] [CrossRef]
  18. Zamil, A.A.A.; Hasan, S.; Baki, S.M.D.J.; Adam, J.M.D.; Zaman, I. Emotion detection from speech signals using voting mechanism on classified frames. In Proceedings of the 2019 International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10–12 January 2019; pp. 281–285. [Google Scholar]
  19. Zheng, C.; Wang, C.; Jia, N. An Ensemble Model for Multi-Level Speech Emotion Recognition. Appl. Sci. 2019, 10, 205. [Google Scholar] [CrossRef] [Green Version]
  20. Ibrar, M.; Hassan, M.A.; Shaukat, K.; Alam, T.M.; Khurshid, K.S.; Hameed, I.A.; Aljuaid, H.; Luo, S. A Machine Learning-Based Model for Stability Prediction of Decentralized Power Grid Linked with Renewable Energy Resources. Wirel. Commun. Mob. Comput. 2022, 2022, 2697303. [Google Scholar] [CrossRef]
  21. Anvarjon, T.; Mustaqeem; Kwon, S. Deep-Net: A Lightweight CNN-Based Speech Emotion Recognition System Using Deep Frequency Features. Sensors 2020, 20, 5212. [Google Scholar] [CrossRef]
  22. ZAli, Z.; Hayat, M.F.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Luo, S.; Basheer, S.; Ayadi, M.; Ksibi, A. A Proposed Framework for Early Prediction of Schistosomiasis. Diagnostics 2022, 12, 3138. [Google Scholar] [CrossRef]
  23. Kumar, M.R.; Vekkot, S.; Lalitha, S.; Gupta, D.; Govindraj, V.J.; Shaukat, K.; Alotaibi, Y.A.; Zakariah, M. Dementia Detection from Speech Using Machine Learning and Deep Learning Architectures. Sensors 2022, 22, 9311. [Google Scholar] [CrossRef]
  24. Luo, S.; Shaukat, K. Computational Methods for Medical and Cyber Security; MDPI: Basel, Switzerland, 2022. [Google Scholar] [CrossRef]
  25. Zhao, Z.; Li, Q.; Zhang, Z.; Cummins, N.; Wang, H.; Tao, J.; Schuller, B.W. Combining a parallel 2D CNN with a self-attention Dilated Residual Network for CTC-based discrete speech emotion recognition. Neural Netw. 2021, 141, 52–60. [Google Scholar] [CrossRef]
  26. Batool, D.; Shahbaz, M.; Asif, H.S.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Ramzan, Z.; Waheed, A.; Aljuaid, H.; Luo, S. A Hybrid Approach to Tea Crop Yield Prediction Using Simulation Models and Machine Learning. Plants 2022, 11, 1925. [Google Scholar] [CrossRef]
  27. Alam, T.M.; Shaukat, K.; Khelifi, A.; Aljuaid, H.; Shafqat, M.; Ahmed, U.; Nafees, S.A.; Luo, S. A Fuzzy Inference-Based Decision Support System for Disease Diagnosis. Comput. J. 2022, 1, bxac068. [Google Scholar] [CrossRef]
  28. Srinivas, C.; Prasad, K.S.; Zakariah, M.; Alothaibi, Y.A.; Shaukat, K.; Partibane, B.; Awal, H. Deep Transfer Learning Approaches in Performance Analysis of Brain Tumor Classification Using MRI Images. J. Healthc. Eng. 2022, 2022, 3264367. [Google Scholar] [CrossRef] [PubMed]
  29. Swain, M.; Maji, B.; Kabisatpathy, P.; Routray, A. A DCRNN-based ensemble classifier for speech emotion recognition in Odia language. Complex Intell. Syst. 2022, 8, 4237–4249. [Google Scholar] [CrossRef]
  30. Devnath, L.; Summons, P.; Luo, S.; Wang, D.; Shaukat, K.; Hameed, I.A.; Aljuaid, H. Computer-Aided Diagnosis of Coal Workers’ Pneumoconiosis in Chest X-ray Radiographs Using Machine Learning: A Systematic Literature Review. Int. J. Environ. Res. Public Health 2022, 19, 6439. [Google Scholar] [CrossRef] [PubMed]
  31. Yang, X.; Khushi, M.; Shaukat, K. Biomarker CA125 Feature Engineering and Class Imbalance Learning Improves Ovarian Cancer Prediction. In Proceedings of the 2020 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE), Gold Coast, Australia, 16–18 December 2020. [Google Scholar] [CrossRef]
  32. Tran, M.; Soleymani, M. A Pre-Trained Audio-Visual Transformer for Emotion Recognition. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 4698–4702. [Google Scholar]
  33. Shaukat, K.; Luo, S.; Varadharajan, V.; Hameed, I.A.; Xu, M. A survey on machine learning techniques for cyber security in the last decade. IEEE Access 2020, 8, 222310–222354. [Google Scholar] [CrossRef]
  34. Shaukat, K.; Luo, S.; Varadharajan, V.; Hameed, I.A.; Chen, S.; Liu, D.; Li, J. Performance Comparison and Current Challenges of Using Machine Learning Techniques in Cybersecurity. Energies 2020, 13, 2509. [Google Scholar] [CrossRef]
  35. Shaukat, K.; Luo, S.; Varadharajan, V. A novel method for improving the robustness of deep learning-based malware detectors against adversarial attacks. Eng. Appl. Artif. Intell. 2022, 116, 105461. [Google Scholar] [CrossRef]
  36. Partila, P.; Voznak, M.; Tovarek, J. Pattern Recognition Methods and Features Selection for Speech Emotion Recognition System. Sci. World J. 2015, 2015, 573068. [Google Scholar] [CrossRef] [Green Version]
  37. Klaylat, S.; Osman, Z.; Hamandi, L.; Zantout, R. Emotion recognition in Arabic speech. Analog Integr. Circuits Signal Process. 2018, 96, 337–351. [Google Scholar] [CrossRef]
  38. Aouf, A. Basic Arabic Vocal Emotions Dataset (Baved). 2019. Available online: https://github.com/40uf411/Basic-Arabic-Vocal-Emotions-Dataset (accessed on 22 December 2022).
  39. Jackson, P.; Haq, S. Surrey Audio-Visual Expressed Emotion (Savee) Database; University of Surrey: Guildford, UK, 2014. [Google Scholar]
  40. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [Green Version]
  41. Gupta, R.; Malandrakis, N.; Xiao, B.; Guha, T.; Van Segbroeck, M.; Black, M.; Potamianos, A.; Narayanan, S. Multimodal prediction of affective dimensions and depression in human-computer interactions. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge; Association for Computing Machinery: New York, NY, USA, 2014; pp. 33–40. [Google Scholar]
  42. Motamed, S.; Setayeshi, S.; Rabiee, A. Speech emotion recognition based on brain and mind emotional learning model. J. Integr. Neurosci. 2018, 17, 577–591. [Google Scholar] [CrossRef]
  43. Scheidwasser-Clow, N.; Kegler, M.; Beckmann, P.; Cernak, M. SERAB: A multi-lingual benchmark for speech emotion recognition. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 7697–7701. [Google Scholar]
  44. Schuller, B.; Zhang, Z.; Weninger, F.; Rigoll, G. Using Multiple Databases for Training in Emotion Recognition: To Unite or to Vote? 2011. Available online: https://d1wqtxts1xzle7.cloudfront.net/46654855/Using_Multiple_Databases_for_Training_in20160620-11120-drxpoc-libre.pdf?1466454195=&response-content-disposition=inline%3B+filename%3DUsing_Multiple_Databases_for_Training_in.pdf&Expires=1674226853&Signature=VZF3e5LoHlcOV4OluQy0uGZiOszIorGhfTLV8ln1bnCo18AXxFrovvSLpGs0KnnmISiClxXTavwfYk6ugDRhFk6noK8QQRA00f8qoBy67d~lXZU2a~fimpr4PN3EiDWd0wEntT588N6XLIx2OabQlIsd2xAvuZOJmkk22wuYIYzIZXTJTZvNEQljJWhwUPxIhMHiGIpl9qrB5h1Vg~8DqKmJAQxx~U1aI~li9C-gLxtMBHWYQIS90HeVKF3uOzhnxSBZcpYmKC6GK-bFtiF9i8TLFt-G1MP8RZ2gr6qGBXm5ndmCq4SnJzo3Bme6prLmfZKSTBs39eOUtNji6DsFPA__&Key-Pair-Id=APKAJLOHF5GGSLRBV4ZA (accessed on 22 December 2022).
  45. Wang, D.; Zheng, T.F. Transfer learning for speech and language processing. In Proceedings of the 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, China, 16–19 December 2015; pp. 1225–1237. [Google Scholar]
  46. Seo, M.; Kim, M. Fusing Visual Attention CNN and Bag of Visual Words for Cross-Corpus Speech Emotion Recognition. Sensors 2020, 20, 5559. [Google Scholar] [CrossRef] [PubMed]
  47. Orjesek, R.; Jarina, R.; Chmulik, M.; Kuba, M. DNN based music emotion recognition from raw audio signal. In Proceedings of the 2019 29th International Conference Radioelektronika (RADIOELEKTRONIKA), Pardubice, Czech Republic, 16–18 April 2019; pp. 1–4. [Google Scholar]
  48. Farooq, M.; Hussain, F.; Baloch, N.K.; Raja, F.R.; Yu, H.; Bin Zikria, Y. Impact of Feature Selection Algorithm on Speech Emotion Recognition Using Deep Convolutional Neural Network. Sensors 2020, 20, 6008. [Google Scholar] [CrossRef] [PubMed]
  49. Zhao, J.; Mao, X.; Chen, L. Speech emotion recognition using deep 1D\& 2D CNN LSTM networks. Biomed. Signal Process. Control 2019, 47, 312–323. [Google Scholar]
  50. Hamdi, S.; Moussaoui, A.; Oussalah, M.; Saidi, M. Gender identification from arabic speech using machine learning. In Proceedings of the International Symposium on Modelling and Implementation of Complex Systems, Batna, Algeria, 24–26 October 2020; pp. 149–162. [Google Scholar]
  51. Mohamed, O.; Aly, S.A. Arabic Speech Emotion Recognition Employing Wav2vec2.0 and HuBERT Based on BAVED Dataset. arXiv 2021, arXiv:2110.04425. [Google Scholar]
  52. Alam, T.M.; Shaukat, K.; Khan, W.A.; Hameed, I.A.; Almuqren, L.A.; Raza, M.A.; Aslam, M.; Luo, S. An Efficient Deep Learning-Based Skin Cancer Classifier for an Imbalanced Dataset. Diagnostics 2022, 12, 2115. [Google Scholar] [CrossRef]
Figure 1. Proposed methodology.
Figure 1. Proposed methodology.
Sensors 23 01386 g001
Figure 2. Distribution of emotions in the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Figure 2. Distribution of emotions in the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Sensors 23 01386 g002aSensors 23 01386 g002b
Figure 3. The MFCC production process.
Figure 3. The MFCC production process.
Sensors 23 01386 g003
Figure 4. The 1D CNN with LSTM attention.
Figure 4. The 1D CNN with LSTM attention.
Sensors 23 01386 g004
Figure 5. Two-Dimensional CNN.
Figure 5. Two-Dimensional CNN.
Sensors 23 01386 g005
Figure 6. Loss and accuracy plots of the (a) ANAD, (b) BAVED, (c) SAVEE, (d) EMO-DB datasets for 1D CNN with the LSTM attention model.
Figure 6. Loss and accuracy plots of the (a) ANAD, (b) BAVED, (c) SAVEE, (d) EMO-DB datasets for 1D CNN with the LSTM attention model.
Sensors 23 01386 g006aSensors 23 01386 g006b
Figure 7. Confusion matrix of 1D CNN with the LSTM attention model for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Figure 7. Confusion matrix of 1D CNN with the LSTM attention model for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Sensors 23 01386 g007
Figure 8. Loss and accuracy plots for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets for the 2D CNN model.
Figure 8. Loss and accuracy plots for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets for the 2D CNN model.
Sensors 23 01386 g008aSensors 23 01386 g008b
Figure 9. Confusion matrix of 2D CNN for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Figure 9. Confusion matrix of 2D CNN for the (a) ANAD, (b) BAVED, (c) SAVEE, and (d) EMO-DB datasets.
Sensors 23 01386 g009
Table 1. Performance evaluation metrics.
Table 1. Performance evaluation metrics.
ModelDatasetAccuracyRecallPrecisionF1-Score
2D CNNANAD90.16%85.79%88.48%86.94%
BAVED83.39%83.26%83.55%83.38%
SAVEE51.30%46.72%48.68%45.88%
EMO-DB50%45.26%48.86%43.11%
CNN+LSTM+AttentionANAD96.72%96.22%95.64%95.90%
BAVED88.39%88.46%88.97%88.52%
SAVEE97.13%97.12%96.89%96.91%
EMO-DB96.72%96.96%96.86%96.88%
Table 2. Individual class performance evaluation metrics—ANAD.
Table 2. Individual class performance evaluation metrics—ANAD.
DatasetModelEmotionsRecallPrecisionF1-Score
ANADCNN+LSTM+AttentionAngry99%97%98%
Happy94%98%96%
Surprised96%92%94%
2D CNNAngry93%95%94%
Happy92%86%88%
Surprised73%85%79%
Table 3. Individual class performance evaluation metrics—BAVED.
Table 3. Individual class performance evaluation metrics—BAVED.
Dataset ModelEmotionsRecallPrecisionF1-Score
BAVEDCNN+LSTM+AttentionLow86%97%91%
Neutral95%85%90%
High85%85%85%
2D CNNLow88%92%90%
Neutral81%81%81%
High81%78%79%
Table 4. Individual class performance evaluation metrics—SAVEE.
Table 4. Individual class performance evaluation metrics—SAVEE.
Dataset ModelEmotionsRecallPrecisionF1-Score
SAVEECNN+LSTM+AttentionAngry90%98%94%
Disgust98%94%96%
Fear95%98%96%
Happy100%100%100%
Neutral99%97%98%
Sad98%96%97%
Surprise100%100%100%
2D CNNAngry50%76%60%
Disgust24%36%29%
Fear29%54%38%
Happy38%23%29%
Neutral77%65%70%
Sad47%47%47%
Surprise62%40%49%
Table 5. Individual class performance evaluation metrics—EMO-DB.
Table 5. Individual class performance evaluation metrics—EMO-DB.
DatasetModelEmotionsRecallPrecisionF1-Score
EMO-DBCNN+LSTM+AttentionAngry100%96%98%
Bored95%95%95%
Disgust97%97%97%
Fear96%95%95%
Happy95%100%98%
Neutral97%96%97%
Sad93%100%96%
2D CNNAngry95%55%69%
Bored33%32%33%
Disgust7%50%12%
Fear47%43%45%
Happy14%40%23%
Neutral38%36%37%
Sad83%86%84%
Table 6. Performance comparison of the works in the EMO-DB dataset.
Table 6. Performance comparison of the works in the EMO-DB dataset.
WorkModelAccuracy
[8]DCNN-BLSTMwA87.86%
[46]VACNN+BOVW86.92%
[48]DCNN+CFS+SVM95.10%
Our work1D CNN with LSTM attention96.72%
Table 7. Performance comparison of the works in the SAVEE dataset.
Table 7. Performance comparison of the works in the SAVEE dataset.
WorkModelAccuracy
[46]VACNN+BOVW75
[48]DCNN+CFS+SVM82.1
Our work1D CNN with LSTM attention97.13%
Table 8. Performance comparison of the works in the ANAD dataset.
Table 8. Performance comparison of the works in the ANAD dataset.
WorkModelAccuracy
[50]Linear SVM96.02
Our work1D CNN with LSTM attention96.72%
Table 9. Performance comparison of the works in the BAVED dataset.
Table 9. Performance comparison of the works in the BAVED dataset.
WorkModelAccuracy
[51]wav2vec2.089
Our work1D CNN with LSTM attention88.39%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alsabhan, W. Human–Computer Interaction with a Real-Time Speech Emotion Recognition with Ensembling Techniques 1D Convolution Neural Network and Attention. Sensors 2023, 23, 1386. https://doi.org/10.3390/s23031386

AMA Style

Alsabhan W. Human–Computer Interaction with a Real-Time Speech Emotion Recognition with Ensembling Techniques 1D Convolution Neural Network and Attention. Sensors. 2023; 23(3):1386. https://doi.org/10.3390/s23031386

Chicago/Turabian Style

Alsabhan, Waleed. 2023. "Human–Computer Interaction with a Real-Time Speech Emotion Recognition with Ensembling Techniques 1D Convolution Neural Network and Attention" Sensors 23, no. 3: 1386. https://doi.org/10.3390/s23031386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop