Previous Article in Journal
Digital Image Processing and Convolutional Neural Network Applied to Detect Mitral Stenosis in Echocardiograms: Clinical Decision Support
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bangla Speech Emotion Recognition Using Deep Learning-Based Ensemble Learning and Feature Fusion

by
Md. Shahid Ahammed Shakil
1,
Fahmid Al Farid
2,
Nitun Kumar Podder
1,
S. M. Hasan Sazzad Iqbal
1,
Abu Saleh Musa Miah
3,*,
Md Abdur Rahim
1,* and
Hezerul Abdul Karim
2,*
1
Department of Computer Science and Engineering, Pabna University of Science and Technology, Pabna 6600, Bangladesh
2
Centre for Image and Vision Computing (CIVC), COE for Artificial Intelligence, Faculty of Artificial Intelligence and Engineering (FAIE), Multimedia University, Cyberjaya 63100, Selangor, Malaysia
3
Department of Computer Science and Engineering, Bangladesh Army University of Science and Technology (BAUST), Saidpur 5311, Bangladesh
*
Authors to whom correspondence should be addressed.
J. Imaging 2025, 11(8), 273; https://doi.org/10.3390/jimaging11080273
Submission received: 14 April 2025 / Revised: 23 July 2025 / Accepted: 30 July 2025 / Published: 14 August 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

Emotion recognition in speech is essential for enhancing human–computer interaction (HCI) systems. Despite progress in Bangla speech emotion recognition, challenges remain, including low accuracy, speaker dependency, and poor generalization across emotional expressions. Previous approaches often rely on traditional machine learning or basic deep learning models, struggling with robustness and accuracy in noisy or varied data. In this study, we propose a novel multi-stream deep learning feature fusion approach for Bangla speech emotion recognition, addressing the limitations of existing methods. Our approach begins with various data augmentation techniques applied to the training dataset, enhancing the model’s robustness and generalization. We then extract a comprehensive set of handcrafted features, including Zero-Crossing Rate (ZCR), chromagram, spectral centroid, spectral roll-off, spectral contrast, spectral flatness, Mel-Frequency Cepstral Coefficients (MFCCs), Root Mean Square (RMS) energy, and Mel-spectrogram. Although these features are used as 1D numerical vectors, some of them are computed from time–frequency representations (e.g., chromagram, Mel-spectrogram) that can themselves be depicted as images, which is conceptually close to imaging-based analysis. These features capture key characteristics of the speech signal, providing valuable insights into the emotional content. Sequentially, we utilize a multi-stream deep learning architecture to automatically learn complex, hierarchical representations of the speech signal. This architecture consists of three distinct streams: the first stream uses 1D convolutional neural networks (1D CNNs), the second integrates 1D CNN with Long Short-Term Memory (LSTM), and the third combines 1D CNNs with bidirectional LSTM (Bi-LSTM). These models capture intricate emotional nuances that handcrafted features alone may not fully represent. For each of these models, we generate predicted scores and then employ ensemble learning with a soft voting technique to produce the final prediction. This fusion of handcrafted features, deep learning-derived features, and ensemble voting enhances the accuracy and robustness of emotion identification across multiple datasets. Our method demonstrates the effectiveness of combining various learning models to improve emotion recognition in Bangla speech, providing a more comprehensive solution compared with existing methods. We utilize three primary datasets—SUBESCO, BanglaSER, and a merged version of both—as well as two external datasets, RAVDESS and EMODB, to assess the performance of our models. Our method achieves impressive results with accuracies of 92.90%, 85.20%, 90.63%, 67.71%, and 69.25% for the SUBESCO, BanglaSER, merged SUBESCO and BanglaSER, RAVDESS, and EMODB datasets, respectively. These results demonstrate the effectiveness of combining handcrafted features with deep learning-based features through ensemble learning for robust emotion recognition in Bangla speech.

1. Introduction

Speech-based emotion recognition (SER) has applications in natural language processing and human–computer interaction (HCI), among other areas [1]. It can contribute to the enhancement of HCI systems by enabling personalized and human-like interactions. SER is helpful in a number of areas, including marketing, education, mental health, speech synthesis, and customer satisfaction [1]. For instance, SER can enhance the user experience overall by recognizing disgruntled consumers and offering insights into user preferences and behavior. To evaluate speech emotions, a variety of methods and strategies are used, including machine learning algorithms and statistical and probabilistic models [2]. Deep learning techniques have recently been central in this field [2,3,4]. Speech emotion identification has shown promise for deep learning approaches, including CNNs [5], DBNs, RNNs, and LSTMs [6]. Again, not much study or development has been conducted in the Bangla language to identify emotions. Therefore, there is a need and potential to create a speech-emotion recognition system for Bangla. Our study’s main objective is to classify speech emotions using an ensemble learning approach that incorporates three types of deep learning algorithms: CNN, LSTM, BiLSTM, or combinations. Identifying emotions in spoken language, particularly in Bangla, is difficult because of varying linguistic usage, social and cultural factors, subjective experiences, and scant evidence [7,8]. Once more, individual and cultural variations, together with the variety of emotional displays in tones, dialects, and speech rates, present major challenges to algorithms trying to identify emotions in Bangla speech [7,8]. Recently, many researchers employed traditional handcrafted feature-based machine learning approaches such as MFCCs, Chroma, and Spectral Contrast to recognize emotions from Bangla speech [6,9]. However, their performance accuracy is not satisfactory. More recently, some researchers are employing deep learning approaches like CNN, LSTM, and BiLSTM models to improve the performance accuracy [6,10]. However, the mentioned research work still faces challenges in achieving good performance accuracy and generalization properties due to a lack of effective features. To overcome the challenges, we proposed integrating the hand-created and multi-stream deep learning features to develop a Bangla speech-based emotion recognition system. Major contributions of the proposed model are given below:
  • Fusion of Handcrafted and Deep Learning Features: We combine handcrafted features with deep learning-derived representations to capture both explicit speech characteristics and complex emotional patterns. This fusion enhances the model’s accuracy and robustness, improving generalization across different emotional expressions and speaker variations.
  • Handcrafted Features: We extract features such as Zero-Crossing Rate (ZCR), Mel-Frequency Cepstral Coefficients (MFCCs), spectral contrast, and Mel-spectrogram, which focus on key speech characteristics like pitch, tone, and energy fluctuations. Although our approach accepts 1D feature vectors as inputs to the model, some features such as MFCCs, Mel-spectrograms, and chromagrams are themselves generated from time–frequency representations of speech signals that can also be represented as images. This makes our approach sing with imaging-based pipelines of classification that process visual patterns as extracted from audio signals, putting it on the same conceptual level as image-based signal processing. These features provide valuable insights into emotional content, enhancing the model’s ability to distinguish subtle emotional variations.
  • Multi-Stream Deep Learning Architecture: Our model employs three streams: 1D CNNs, 1D CNNs with Long Short-Term Memory (LSTM), and 1D CNNs with bidirectional LSTM (Bi-LSTM), which capture both local and global patterns in speech, providing a robust understanding of emotional nuances. The LSTM and Bi-LSTM streams improve the model’s ability to recognize emotions in speech sequences.
  • Ensemble Learning with Soft Voting: We combine predictions from the three streams using an ensemble learning technique with soft voting, improving emotion classification by leveraging the strengths of each model.
  • Improved Performance and Generalization: Data augmentation techniques such as noise addition, pitch modification, and time stretching enhance the model’s robustness and generalization, addressing challenges like speaker dependency and variability in emotional expressions. Our approach achieves impressive performance, with accuracies of 92.90%, 85.20%, 90.63%, 67.71%, and 69.25% for the SUBESCO, BanglaSER, merged SUBESCO and BanglaSER, RAVDESS, and EMODB datasets, respectively, demonstrating its superiority over traditional models.

2. Related Works

Speech emotion recognition (SER) has been transformed by deep learning algorithms [11,12,13]; yet, research on SER in the Bangla language is scarce. A deep learning approach to speech emotion recognition was proposed in 2021 by Sadia Sultana, M. Zafar Iqbal, et al. [14]. Using the Bangla audio-only dataset SUBESCO, they used bidirectional LSTM networks and deep convolutional neural networks with a time-distributed flatten layer for their investigation. They used cross-lingual and multilingual training and testing sets in many tests, and their method with a TDF layer performed better than other cutting-edge CNN-based SER models. In order to conduct cross-lingual experiments in Bangla and English, they employed transfer learning and the SUBESCO and RAVDESS datasets for both cross- and multi-corpus training. The model achieved 86.9% (WA) for the SUBESCO dataset and 82.7% (WA) for the RAVDESS dataset. In 2018, Rahman, Md. Masudur, and associates presented a proposal for an automated voice recognition system for Bengali that included a support vector machine with dynamic time warping [15]. For the static features, they employed MFCCs, and for the dynamic features, MFCC derivatives. SVM with RBF was utilized for classification, and their modified DTW technique was utilized for feature matching. The system achieved 86.08% accuracy on 12 speakers. Using phase-based cepstral features, Chakraborty et al. suggested an automatic speech emotion recognition model in 2022 [8]. They used pre-processing methods with PBCC to extract phase-based information from speech data. The SUBESCO and BanglaSER datasets, which included a gradient-boosting machine-based classifier, were utilized to evaluate the model. In comparison with earlier methods, the findings demonstrated enhanced performance, with an average accuracy of 96% for both speaker-dependent and speaker-independent emotion detection tests. Using five characteristics taken from sound data and used as 1D CNN inputs, Dias Issa, M. Fatih Demirci, et al. created a unique approach for speech emotion identification in 2020 [16]. During testing on the RAVDESS, EMO-DB, and IEMOCAP databases, the model achieved good classification accuracy in speaker-independent audio. For instance, the model achieved 71.61% accuracy on the RAVDESS dataset with eight classes, 86.1% accuracy on EMO-DB (535 samples) in 7 classes, 95.71% accuracy on EMO-DB (520 samples) in 7 classes, and 64.3% accuracy on IEMOCAP in 4 classes. Without the requirement for visual aids, the suggested model performed better than the majority of existing models. Using 2D and 1D CNN LSTM networks, Jianfeng Zhao, Xia Mao, et al. (2018) presented a deep learning method for speech emotion identification [17]. From speech and log-mel spectrograms, their algorithms retrieved features associated with both local and global emotions. They achieved recognition accuracy of 95.33% and 95.89% in speaker-dependent and speaker-independent trials on EmoDB and 52.14% and 89.16% in speaker-independent and speaker-dependent testing on the IEMOCAP database, respectively, outperforming competing methods such as CNN and Deep Belief Network. A 1D dilated CNN with hierarchical features learner blocks (HFLBs) and a bi-directional gated recurrent unit (BiGRU) for SER was suggested by Mustaqeem and Soonil Kwon in 2021 [18]. Through a layered 1D dilated network (HFLBs), they employed spectral analysis to uncover unknown patterns from audio samples. After that, the characteristics were input into the BiGRU, which used a softmax layer to build the likelihood of speech emotions and learn temporal cues. On the IEMOCAP, EMO-DB, and RAVDESS datasets, their model yielded relative accuracy values of 72.75%, 91.14%, and 78.01%, respectively. In 2017, Badshah et al. proposed a model (CNN) with three convolutional and three FC layers [19]. The model was trained using the seven emotion classes found in the EMODB corpus. In order to train, they used spectrograms. The accuracy percentage of their study was 56%. In 2018, Etienne et al. developed the CNN-LSTM model for identifying speech emotions [20]. High-level features were extracted using convolutional layers, while long-term associations were gathered using recurrent layers. Their best-performing model, a convolutional (4 layers) and BLSTM combination (1 layer), produced 61.7% unweighted accuracy and 64.5% weighted accuracy for four emotions.
In 2020, Xusheng Ai, Victor S. Sheng, et al. proposed an ensemble learning approach for speech emotion identification using ACRNN [21]. Convolutional recurrent neural networks were utilized in conjunction with bagging and attention models to handle observation overlap problems. In order to solve further issues, they also employed redagging and augmentation strategies. Their research made use of the Emo-DB and IEMOCAP databases. In 2020, Mustaqeem and Kwon presented an SER model based on spectrogram features that classified emotions using a CNN with deep strides (DSCNN) [22]. The average accuracy for the RAVDESS dataset was 79.5%, whereas the average accuracy for the IEMOCAP dataset was 81.75%. In 2015, Zheng et al. used log-spectrograms as input for their proposed DCNN model for speech emotion recognition [23]. They discovered that their model performed better than conventional models that depended on manually created features after using PCA to minimize dimensionality. In 2021, an ensemble-based cross-lingual speech emotion recognition (SER) approach was proposed by Wisha Zehra et al. [24]. They focused on improving emotion recognition in robots interacting with speakers from multilingual and multicultural backgrounds. Unlike traditional SER systems that use the same corpus for both training and testing, their study trained classifiers on one corpus and tested them on another to simulate real-world multilingual environments. They applied an ensemble learning strategy using a majority voting technique, combining the predictions of multiple classifiers to enhance robustness across corpora. Experiments on Urdu, German, Italian, and English datasets showed improvements in accuracy by up to 13% in within-corpus tests and up to 15% in cross-corpus settings, outperforming several individual classifiers. This demonstrated the advantage of ensemble methods in multilingual SER tasks. A real-time, non-line-of-sight (NLOS) human–machine interaction framework for ensemble-based emotional communication was introduced in 2025 by Shaik Abdul Khalandar Basha, P. M. Durai Raj Vincent, et al. [25]. The authors introduced the EAS framework, which enables long-distance emotional realization through communication. Their system employs CNN-LSTM architectures with and without attention mechanisms for emotional classification and feature extraction. They also used Deep Canonical Correlation Analysis (DCCA) for feature correlation analysis. The CNN-LSTM model with attention design improved in classification accuracy to 87.08%, outperforming baseline CNN (81.11%) and LSTM (84.01%) models. The approach performed better than expected in real-world emotional communication scenarios, which proves the benefit of ensemble and attention-oriented solutions in real-world SER usage.
Although our research focused on emotion recognition from audio speech, researchers also used other modalities for recognition, including facial expressions, physiological signals such as EEG, body gestures, textual data, and video [26]. These modalities can provide rich biometric cues for emotion detection. While working with a biometric modality such as speech or body gestures, we can extract multiple features and combine them for better performance. Features from different biometric modalities can also be fused for more accurate emotion recognition. In 2023, Baole Fu et al. [27] introduced a novel multimodal emotion recognition approach in which a Multimodal Feature Fusion Neural Network (MFFNN) was employed to fuse EEG and eye movement signals. The model consists of a dual-branch feature extraction module and a multi-scale fusion mechanism with cross-channel soft attention to extract complementary spatial and temporal information. On four-class emotion classes of the SEED-IV dataset, the model was proven to attain an accuracy rate of 87.32%, demonstrating the effectiveness of feature fusion in improving emotion recognition performance.

3. Datasets

By using the SUBESCO and BanglaSER datasets, which are collections of Bengali speech samples with annotations of expressed emotions, our work seeks to develop an SER system for the language. The RAVDESS and EMODB datasets, each with seven emotion categories, were also used to assess our models. We conducted experiments using five or seven emotion classes, depending on the available labels in each dataset. For example, SUBESCO has seven emotion classes: neutral, angry, disgust, happy, sad, surprise, and fear. We did not select the emotion categories independently; rather, they are predefined within the datasets based on their original annotation protocol.

3.1. SUBESCO Dataset

SUBESCO, also known as the SUST Bangla Emotional Speech Corpus, is an audio-only dataset including about seven hours of emotional Bangla speech. It is made up of 7000 utterances that were recorded by 20 native speakers who are split equally between the sexes. Each speaker recorded 10 lines, which correspond to seven different moods. Each audio clip underwent four rounds of verification by male and female raters after the dataset was examined by fifty college students [28]. Table 1 shows emotion classes in the SUBESCO dataset with counts and percentages. It was developed to help in the development of Bangla SER systems and is publicly available.

3.2. BanglaSER Dataset

BanglaSER is a collection of speech-audio data from 34 speakers that reflect five fundamental emotional states, and it is used for SER in the Bangla language. Using computers and cell phones to capture spoken audio, 1467 recordings in total of three phrases per recording were made, with an equal number of recordings in each category and an equal distribution of male and female performers (17 male, 17 female) [29]. Table 2 shows emotion classes in the BanglaSER dataset with counts and percentages.

3.3. SUBESCO and BanglaSER Merged Dataset

We conducted an inquiry to develop a multi-corpus classification model for Bangla SER by merging the SUBESCO and BanglaSER datasets. Making a more broadly applicable SER model was the primary goal. BanglaSER offers five different emotion courses: fear, neutral, sad, disgusted, and pleased. SUBESCO offers seven different emotional lessons. Table 3 shows emotion classes in the merged dataset with counts and percentages.

3.4. RAVDESS Dataset

The RAVDESS collection is made up of audio and video clips of actors performing heartfelt songs and speeches. It is used for research and development in speech and audio processing, multimedia content analysis, and emotional analysis. It contains recordings of seven emotions sung by 24 performers. Three different forms of data are included in the dataset: audio-only, audio–video, and video-only [30]. We used audio-only data in our study. Table 4 shows emotion classes in the RAVDESS dataset with counts and percentages.

3.5. EMODB Dataset

The Technical University of Berlin developed the free-to-use German emotional database EMODB, which has 535 statements expressing seven different emotions (angry, bored, anxious, fear, happy, disgust, sadness, and neutral) from 10 expert speakers [31]. Table 5 shows emotion classes in the RAVDESS EMODB with counts and percentages.

4. Materials and Methods

We propose a novel multi-stream deep learning feature fusion approach for Bangla speech emotion recognition, addressing the limitations of existing methods, such as low accuracy, speaker dependency, and poor generalization across emotional expressions. Our approach begins with various data augmentation techniques applied to the training dataset, improving the model’s robustness and generalization across different speakers and emotional expressions. We extract a comprehensive set of handcrafted features, including ZCR, chromagram, spectral centroid, spectral roll-off, spectral contrast, spectral flatness, MFCCs, RMS energy, and Mel-spectrogram. These handcrafted features capture essential characteristics of the speech signal, providing valuable insights into the emotional content, particularly in terms of pitch, tone, and energy variations. Subsequently, we utilize a multi-stream deep learning architecture to automatically learn complex, hierarchical representations of emotional content in speech. Despite the fact that 2D spectrograms are frequently employed in speech emotion recognition, we concentrated on 1D feature vectors for modeling in this work. The 1D feature vector that we obtained after extracting the aforementioned characteristics naturally fits 1D models such as CNN, LSTM, and BiLSTM. Once more, we wanted to increase efficiency by combining different handcrafted features; in this way, some feature vectors would not need to be converted into 2D form while using 1D input. So, our architecture consists of three distinct streams: the first stream employs 1D CNN, the second integrates 1D CNN with LSTM, and the third stream combines 1D CNN with Bi-LSTM. These deep learning models capture intricate emotional nuances, offering a deeper understanding of speech emotion than handcrafted features alone. We fuse both handcrafted and deep learning-derived features, combining traditional signal characteristics with learned representations of emotional content. We combine predictions from the three streams using an ensemble learning technique with soft voting, improving emotion classification by leveraging the strengths of each model. We assess our model using three primary datasets—SUBESCO, BanglaSER, and a merged version of both—as well as two external datasets, RAVDESS and EMODB. An overview of the proposed methodology is provided in Figure 1.

4.1. Data Augmentation

To assess the efficacy of a speech-emotion classification algorithm, the dataset is typically split into training and testing sets. In this study, we split the dataset into training and testing sets in the ratios of 70/30 and 80/20 based on the dataset owner protocol. An overview of the train–test samples is provided in Table 6. Then we applied a data augmentation approach on the training dataset, which is described in Table 7. We used a data augmentation technique to produce fresh synthetic training samples by making small adjustments to the original training dataset [9,32]. In this case, we applied six different audio data augmentations to improve the speech emotion detection models. It is important to bear in mind that data augmentation needs to be limited to the training data once it has been split into train and test samples. On our training data, we applied polarity inversion, noise addition, time stretching, pitch change, sound shifting, and random gain [33,34,35]. To expand the number of test samples, we only added noise to our test data. Some examples of data augmentations are given in Figure 2.

4.2. Feature Extraction

We extracted nine audio features from the frequency, time, and time–frequency domains, including low- and mid-level features, to categorize emotions in speech. Although our feature extraction process utilizes 1D handcrafted features such as MFCCs, chromagram vectors, and spectral features, they are all derived from time–frequency representations of the speech signal. To visually indicate how emotions manifest in the frequency domain, spectrograms of audio samples from different emotion classes are presented in Figure 3. At first, we loaded the audio files using a fixed duration of 2.5 s with an offset of 0.6 s to avoid silence. We extracted the predefined acoustic features by computing their frame-wise values and then averaging them across all frames to obtain a single fixed-length vector per feature type. The features were concatenated into a single 1D vector using horizontal stacking (np.hstack). These features were extracted for both the original and augmented datasets. All resulting feature vectors are vertically stacked (np.vstack). These final feature vectors are directly used as inputs to the model’s dense layers in a conventional machine learning pipeline. Feature extraction was performed using the Python (v3.10.12) Librosa package (v0.10.1) [33]. The names of the nine features and other relevant information are shown in Table 8.

4.3. Deep Learning Model

After the data are prepared, the model has to be trained for classification. This is the process of giving the algorithm a large number of labeled samples to increase its accuracy. A loss function is used to minimize the discrepancy between the predicted and actual outputs. We were able to train three different models: a 1D CNN, a 1D CNN-LSTM, and a 1D CNN-bidirectional LSTM.

4.3.1. 1D-CNN Approach

First, we use a 1D-CNN, which has the same architecture across all five datasets, as our emotion classification model. The model includes several layers for one-dimensional convolution, pooling, batch normalization, dropout, activation, flattening, and fully connected layers [48,49]. The convolution layers extract deep features and produce feature maps [49]. Max-pooling is used by the pooling layers, which down-sample and optimize the spatial size of the feature maps [49,50]. The model with ReLU incorporates non-linearity through activation layers [49]. While batch normalization layers hasten deep neural network training, dropout layers prevent overfitting [51,52]. One dimension is extracted from the input by the Flatten layer [53]. The fully connected layer uses a Softmax function to create a probability distribution across the classes to produce predictions based on data from previous layers [49,54]. With a learning rate of 0.00001, Adam was employed as the optimizer [55]. Our proposed 1D CNN model comprises six convolutional layers (filters: 64–256, kernel sizes: 3 and 5), each followed by ReLU activation, batch normalization, and max pooling (pool sizes: 3 or 5, stride: 2). Dropout (rate: 0.3) is applied to reduce overfitting. The fully connected layer has 1024 units, followed by a softmax output layer (5 or 7 units depending on the dataset). The model is trained using a batch size of 32 for up to 100 epochs. The architecture of our 1D CNN model is given in Figure 4.

4.3.2. Integration of 1D-CNN and LSTM Approach

For our second classification model, we merged an LSTM with a 1D CNN in an attempt to use both models’ advantages. We use a CNN component for feature extraction, where the CNN structure remains unchanged from our 1D CNN model, and the model’s architecture is the same for all five types of datasets. After processing the CNN architecture’s output, a time-distributed flatten layer sends it to an LSTM layer for sequential analysis and a fully connected layer for predictions. The LSTM layer has 300 units, and dropout (rate: 0.3) is applied after the LSTM and dense layers. To make sure that the same weights and biases are given to each temporal timestep of the layer, a wrapper called the time-distributed flatten method was used [14]. A Softmax function is used for probability distribution over emotion categories [54]. The Adam optimizer was used with a 0.000001 learning rate [55]. The model is trained using a batch size of 32 for up to 100 epochs. An architecture of our 1D CNN-LSTM model is given in Figure 5.

4.3.3. Integration of 1D-CNN and Bi-LSTM Approach

Our third classification model extracts contextual and deep characteristics from the input data by combining two bidirectional LSTM layers (300 and 100 units) with a 1D CNN (same architecture as previous models). Similar to our second model, the CNN output is sent to the LSTM layers via a time-distributed flatten layer; however, this time, the layers are bidirectional, allowing for the acquisition of information from both past and future data points [14]. The input is categorized into one of seven or five emotion categories by the model using a Softmax layer [54]. Utilizing a 0.000001 learning rate, the Adam optimizer was used [55]. The dense layer has 1024 units. It also uses a batch size of 32 and is trained for up to 100 epochs. The architecture of our 1D CNN-BiLSTM model is given in Figure 6.

4.4. Soft Voting Ensemble Learning

We trained three classification models and then applied an ensemble learning technique. Using the sum rule ensemble technique, the three classification models (1D CNN, 1D CNN LSTM, and 1D CNN BiLSTM) trained in the prior stage were combined for better emotion categorization. Sum rule ensemble learning (a form of soft voting) aggregates predictions from various models for each class label and predicts the class label based on the highest summed probability, as shown in Figure 1 [56]. In this approach, the overall accuracy is improved by focusing the models on different data points and combining their predictions. The procedure involves generating predictions from each model and aggregating them to form a final prediction. The sum rule is given by Equation (1):
y ^ final = arg max i = 1 n P ( y | x , θ i )
where y ^ final is the predicted class label. P ( y | x , θ i ) is the probability prediction for class y from model i for input x, with parameters θ i . n is the total number of models in the ensemble. The prediction is based on the highest summed probability across all models. After generating these summed predictions, we apply the argmax function to convert continuous values into class labels. Finally, we compare the predicted class labels with the ground truth labels to assess the ensemble’s accuracy. An accuracy metric from scikit-learn, such as accuracy_score, is used for this evaluation.

5. Results

Three classification models and an ensemble learning strategy were tested on each of the five datasets that we used. Across the datasets, we saw notable variations in the model’s performance, and our ensemble technique continuously increased accuracy.

5.1. Ablation Study

Table 9 displays the ablation study where we show the performance for each of our suggested models. Our investigation revealed that the 1D CNN-LSTM model outperformed other base models on the SUBESCO and BanglaSER datasets. Conversely, the 1D CNN model outperformed the others for the SUBESCO and BanglaSER merged, RAVDESS, and EMODB datasets. Overall, our models performed better on the Bangla speech datasets than the other two. The observed accuracy differences can be attributed to variations in language, dataset size, and recording style. SUBESCO and BanglaSER, both Bengali speech corpora, yielded consistently higher performance, possibly indicating that our model generalizes better within the Bengali language domain. In contrast, RAVDESS (English) and EMODB (German) are acted and linguistically different, which may limit cross-lingual generalization.

5.2. Outcomes of the Models for SUBESCO Dataset

For the SUBESCO Bangla Speech Emotion Corpus, our proposed models performed well. Over 99% accuracy was achieved by each model in learning the training set of data. The three 1D CNN classification models, 1D CNN LSTM, 1D CNN BiLSTM, and 1D CNN, have testing accuracy scores of 90.93%, 90.98%, and 90.50%, respectively. Ultimately, the accuracy of the ensemble learning approach that combined the three models was 92.90%. This indicates a significant increase in accuracy. A summary of all these findings is given in Table 10. In Table 11, the accuracy of the Ensemble Learning approach for each emotion class of the SUBESCO dataset is given, and the performance metrics are given in Table 12.
In Figure 7, the learning curves (accuracy and loss) of our trained models for the SUBESCO dataset are given.

5.3. Outcomes of the Models for BanglaSER Dataset

On the BanglaSER dataset, the accuracy of the three classification models 1D CNN, 1D CNN LSTM, and 1D CNN BiLSTM was 83.67%, 84.52%, and 81.97%, in that order. Ultimately, the accuracy of the ensemble learning method that integrates the three models was 85.20%. This indicates an improvement in accuracy. A summary of all these findings is given in Table 13. In Table 14, the accuracy of the Ensemble Learning approach for each emotion class of the BanglaSER dataset is given, and the performance metrics are given in Table 15. The learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BILSTM, respectively) for the BanglaSER dataset are given in Figure 8.

5.4. Outcomes of the Models for SUBESCO and BanglaSER Merged Dataset

In order to evaluate the three classification models (1D CNN, 1D CNN LSTM, and 1D CNN BiLSTM), we combined two Bangla speech emotion datasets. On the combined dataset, the models’ accuracy was 88.92%, 88.61%, and 87.56%, in that order. Nonetheless, we obtained an accuracy of 90.63% with an ensemble learning approach that included all three models, suggesting an improvement in accuracy. A summary of model accuracies for the SUBESCO and BanglaSER merged dataset is given in Table 16. In Table 17, the accuracy of the Ensemble Learning approach for each emotion class of the SUBESCO and BanglaSER merged dataset is given, and the performance metrics are given in Table 18. The learning curves (accuracy and loss) of the models for the SUBESCO and BanglaSER merged dataset are given in Figure 9.
For all of our datasets except EMODB, the ensemble learning approach showed improved classification accuracy. Figure 10 shows the confusion matrices of the ensemble learning approach for the SUBESCO, BanglaSER, and SUBESCO-BanglaSER merged datasets, where each matrix provides a detailed view of the classification outcomes of the ensemble learning approach for each of the datasets. For the SUBESCO dataset, while overall accuracy was high, our model faced difficulty distinguishing between ‘disgust’ and ‘sad,’ and occasionally misclassified ‘surprise’ as ‘angry’ or ‘disgust’; ‘neutral’ expressions, however, were recognized very accurately. For the BanglaSER dataset, we noted good performance for ‘angry,’ ‘neutral’, ‘sad’, and ‘surprise’, but more errors with ‘happy’, which was sometimes confused with ‘surprise’ or ‘sad’, suggesting potential feature overlap in that dataset. Merging the BanglaSER dataset with SUBESCO led to an overall improvement in accuracy and a reduction in the confusion between ‘disgust’ and ‘sad.’ However, there was still some confusion for ‘surprise’. By comparing these matrices, we can evaluate the performance of our ensemble learning approach, which provided improved identification of emotion classes and handled class imbalance and misclassification across all the datasets.

5.5. Outcomes of the Models for RAVDESS and EMODB Datasets

The models on the RAVDESS and EMODB datasets were trained using the identical model setups as the previous datasets. The models’ accuracy on these datasets was mediocre; however, our ensemble learning model’s accuracy increased. The accuracy of our models for the RAVDESS and EMODB datasets is given in Table 19 and Table 20, respectively.

5.6. State of the Art Comparison

Table 21 presents a comparison of various state-of-the-art models in emotion recognition using speech datasets. Each study employs distinct features and models, with a variety of accuracy metrics for different datasets. For instance, Sultana et al. [14] achieved an accuracy of 86.9% on the SUBESCO dataset using Mel-spectrogram and Deep CNN combined with BLSTM, while Rahman et al. [15] used MFCCs and their derivatives with SVM and DTW, achieving 86.08% accuracy on Bangla 12 speakers. Chakraborty et al. [8] utilized PBCC features with a Gradient Boosting Machine, reaching an impressive 96% accuracy on their dataset. Notably, the proposed model demonstrates superior performance, with 92.90% accuracy on the SUBESCO dataset, and notable results on RAVDESS, EMO-DB, BanglaSER, and a combined dataset (SUBESCO + BanglaSER) with 90.63% accuracy. This highlights the robustness and effectiveness of the proposed ensemble model.

5.7. Discussion

Table 9 shows the ablation study, and Table 21 shows the state of the art comparison model. For every dataset except EMODB, the accuracy of the ensemble learning strategy was better than that of other base models. Establishing a method for improved speech emotion categorization from Bangla speech was the main goal of our research. Our method demonstrates how the classification accuracy of Bangla speech may be increased by combining the right features with data augmentations, followed by an ensemble of trained models. Although we did not directly use spectrogram or chromagram images as model inputs, the features we extracted (such as MFCCs and chromagrams) are processed from these representations. Their 2D structure and visualization potential establish a conceptual link with imaging methodologies, even though we process them as 1D feature vectors. In contrast to Sadia Sultana, M. Zafar Iqbal, et al. (2021), who used CNN and the BiLSTM model to obtain 86.9% accuracy on the SUBESCO dataset [14], our method achieved 92.90% accuracy on the SUBESCO dataset, 85.20% accuracy on the BanglaSER dataset, and 90.63% accuracy on the combined dataset of SUBESCO and BanglaSER. Regarding the classification of emotions from speech, the ensemble learning approach performs better than most of the models. One main limitation of our research is that our experiment has been conducted on acted datasets only. Our experiment was limited to the acted dataset, which is one of the key limitations of our study. In real-world situations, our model might not function well because of variances in speech data caused by environments, languages, and cultural differences. It is also important to take into account the various Bangla dialects. In a subsequent study, we may expand to incorporate natural speech data rather than just acted data to more accurately detect speech patterns and emotional expressions in real-world situations. Moreover, speech data from multiple regional dialects of the language or multilingual datasets may be integrated to create a more robust system that can accurately recognize emotions in a range of languages and cultural contexts. We may look into several sets of data augmentation and feature extraction strategies to find out which combinations perform best together for voice emotion recognition. We can expand upon our work by evaluating model robustness under adversarial or environmental audio perturbations. To improve the ensemble learning method, we may eventually try incorporating more categorization models and experimenting with other combinations. Our ensemble learning strategy demonstrated constant superiority over the other models on all datasets. Additionally, our method’s efficacy with Bengali datasets is demonstrated by the results obtained. With good accuracy levels on most datasets, our method demonstrated the overall effectiveness of ensemble learning for speech-emotion recognition.

6. Conclusions

In this study, we propose a novel multi-stream deep learning feature fusion approach for Bangla speech emotion recognition, which effectively addresses the challenges of low accuracy, speaker dependency, and poor generalization. By combining handcrafted features with deep learning-derived features and employing an ensemble learning technique, our method significantly enhances the robustness and accuracy of emotion recognition across diverse datasets. We combine predictions from the three streams using an ensemble learning technique with soft voting, improving emotion classification by leveraging the strengths of each model. The results demonstrate the ability to capture intricate emotional nuances in Bangla speech, achieving high performance in both primary and external datasets. This approach holds broader potential for expanding emotion recognition systems to other languages and emotional categories, paving the way for more universal and accurate speech-emotion recognition systems. Looking ahead, we plan to extend this research by applying the proposed approach to other underrepresented languages and exploring additional emotional categories. We aim to enhance the generalization capability of the model by incorporating more diverse datasets and refining the feature extraction techniques. Furthermore, future work will focus on integrating this emotion recognition model into real-time systems, such as virtual assistants and conversational AI platforms, to improve their responsiveness and empathy in human–machine interactions.

Author Contributions

Conceptualization, M.S.A.S. and S.M.H.S.I.; methodology, M.S.A.S., N.K.P., M.A.R. and A.S.M.M.; software, M.S.A.S.; validation, M.S.A.S., N.K.P. and S.M.H.S.I.; formal analysis, M.S.A.S.; investigation, M.S.A.S.; resources, A.S.M.M.; data curation, M.S.A.S.; writing—original draft preparation, M.S.A.S.; writing—review and editing, N.K.P., M.A.R. and A.S.M.M.; visualization, M.S.A.S.; supervision, S.M.H.S.I. and M.A.R.; project administration, M.A.R.; funding acquisition, H.A.K. and F.A.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Multimedia University, Cyberjaya, Selangor, Malaysia (Grant Number: PostDoc(MMUI/240029)).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. All data used are publicly available from the sources cited in the manuscript.

Acknowledgments

We would like to express our sincere gratitude to the Department of Computer Science and Engineering, Pabna University of Science and Technology, for providing necessary support and resources throughout this research.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

PBCCPhase-Based Cepstral Coefficients
DTWDynamic Time Warping
RMSRoot Mean Square
MFCCsMel-Frequency Cepstral Coefficients

References

  1. Bashari Rad, B.; Moradhaseli, M. Speech emotion recognition methods: A literature review. AIP Conf. Proc. 2017, 1891, 020105. [Google Scholar] [CrossRef]
  2. Ahlam Hashem, M.A.; Alghamdi, M. Speech emotion recognition approaches: A systematic review. Speech Commun. 2023, 154, 102974. [Google Scholar] [CrossRef]
  3. Muntaqim, M.Z.; Smrity, T.A.; Miah, A.S.M.; Kafi, H.M.; Tamanna, T.; Farid, F.A.; Rahim, M.A.; Karim, H.A.; Mansor, S. Eye Disease Detection Enhancement Using a Multi-Stage Deep Learning Approach. IEEE Access 2024, 12, 191393–191407. [Google Scholar] [CrossRef]
  4. Hossain, M.M.; Chowdhury, Z.R.; Akib, S.M.R.H.; Ahmed, M.S.; Hossain, M.M.; Miah, A.S.M. Crime Text Classification and Drug Modeling from Bengali News Articles: A Transformer Network-Based Deep Learning Approach. In Proceedings of the 2023 26th International Conference on Computer and Information Technology (ICCIT), Cox’s Bazar, Bangladesh, 13–15 December 2023; pp. 1–6. [Google Scholar]
  5. Rahim, M.A.; Farid, F.A.; Miah, A.S.M.; Puza, A.K.; Alam, M.N.; Hossain, M.N.; Karim, H.A. An Enhanced Hybrid Model Based on CNN and BiLSTM for Identifying Individuals via Handwriting Analysis. CMES-Comput. Model. Eng. Sci. 2024, 140, 1690–1710. [Google Scholar] [CrossRef]
  6. Khalil, R.A.; Jones, E.; Babar, M.I.; Jan, T.; Zafar, M.H.; Alhussain, T. Speech emotion recognition using deep learning techniques: A review. IEEE Access 2019, 7, 117327–117345. [Google Scholar] [CrossRef]
  7. Saad, F.; Mahmud, H.; Shaheen, M.; Hasan, M.K.; Farastu, P. Is Speech Emotion Recognition Language-Independent? Analysis of English and Bangla Languages using Language-Independent Vocal Features. arXiv 2021, arXiv:2111.10776. [Google Scholar]
  8. Chakraborty, C.; Dash, T.K.; Panda, G.; Solanki, S.S. Phase-based Cepstral features for Automatic Speech Emotion Recognition of Low Resource Indian languages. In Transactions on Asian and Low-Resource Language Information Processing; Association for Computing Machinery: New York, NY, USA, 2022. [Google Scholar]
  9. Ma, E. Data Augmentation for Audio. Medium. 2019. Available online: https://medium.com/@makcedward/data-augmentation-for-audio-76912b01fdf6 (accessed on 8 March 2023).
  10. Rintala, J. Speech Emotion Recognition from Raw Audio using Deep Learning; School of Electrical Engineering and Computer Science Royal Institute of Technology (KTH): Stockholm, Sweden, 2020. [Google Scholar]
  11. Tusher, M.M.R.; Al Farid, F.; Kafi, H.M.; Miah, A.S.M.; Rinky, S.R.; Islam, M.; Rahim, M.A.; Mansor, S.; Karim, H.A. BanTrafficNet: Bangladeshi Traffic Sign Recognition Using A Lightweight Deep Learning Approach. Comput. Vis. Pattern Recognit. 2024; preprint. [Google Scholar]
  12. Siddiqua, A.; Hasan, R.; Rahman, A.; Miah, A.S.M. Computer-Aided Osteoporosis Diagnosis Using Transfer Learning with Enhanced Features from Stacked Deep Learning Modules. arXiv 2024, arXiv:2412.09330. [Google Scholar] [CrossRef]
  13. Tusher, M.M.R.; Farid, F.A.; Al-Hasan, M.; Miah, A.S.M.; Rinky, S.R.; Jim, M.H.; Sarina, M.; Abdur, R.M.; Hezerul Abdul, K. Development of a Lightweight Model for Handwritten Dataset Recognition: Bangladeshi City Names in Bangla Script. Comput. Mater. Contin. 2024, 80, 2633–2656. [Google Scholar] [CrossRef]
  14. Sultana, S.; Iqbal, M.Z.; Selim, M.R.; Rashid, M.M.; Rahman, M.S. Bangla speech emotion recognition and cross-lingual study using deep CNN and BLSTM networks. IEEE Access 2021, 10, 564–578. [Google Scholar] [CrossRef]
  15. Rahman, M.M.; Dipta, D.R.; Hasan, M.M. Dynamic time warping assisted SVM classifier for Bangla speech recognition. In Proceedings of the 2018 International Conference on Computer, Communication, Chemical, Material and Electronic Engineering (IC4ME2), Rajshahi, Bangladesh, 8–9 February 2018; pp. 1–6. [Google Scholar]
  16. Issa, D.; Demirci, M.F.; Yazici, A. Speech emotion recognition with deep convolutional neural networks. Biomed. Signal Process. Control 2020, 59, 101894. [Google Scholar] [CrossRef]
  17. Zhao, J.; Mao, X.; Chen, L. Speech emotion recognition using deep 1D & 2D CNN LSTM networks. Biomed. Signal Process. Control 2019, 47, 312–323. [Google Scholar]
  18. Mustaqeem; Kwon, S. 1D-CNN: Speech emotion recognition system using a stacked network with dilated CNN features. CMC-Comput. Mater. Contin. 2021, 67, 4039–4059. [Google Scholar]
  19. Badshah, A.M.; Ahmad, J.; Rahim, N.; Baik, S.W. Speech emotion recognition from spectrograms with deep convolutional neural network. In Proceedings of the 2017 International Conference on Platform Technology and Service (PlatCon), Busan, Republic of Korea, 13–15 February 2017; pp. 1–5. [Google Scholar]
  20. Etienne, C.; Fidanza, G.; Petrovskii, A.; Devillers, L.; Schmauch, B. Cnn+ lstm architecture for speech emotion recognition with data augmentation. arXiv 2018, arXiv:1802.05630. [Google Scholar] [CrossRef]
  21. Ai, X.; Sheng, V.S.; Fang, W.; Ling, C.X.; Li, C. Ensemble learning with attention-integrated convolutional recurrent neural network for imbalanced speech emotion recognition. IEEE Access 2020, 8, 199909–199919. [Google Scholar] [CrossRef]
  22. Kwon, S. A CNN-assisted enhanced audio signal processing for speech emotion recognition. Sensors 2019, 20, 183. [Google Scholar] [CrossRef]
  23. Zheng, W.; Yu, J.; Zou, Y. An experimental study of speech emotion recognition based on deep convolutional neural networks. In Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), Xi’an, China, 21–24 September 2015; pp. 827–831. [Google Scholar]
  24. Zehra, W.; Javed, A.R.; Jalil, Z.; Khan, H.U.; Gadekallu, T.R. Cross corpus multi-lingual speech emotion recognition using ensemble learning. Complex Intell. Syst. 2021, 7, 1845–1854. [Google Scholar] [CrossRef]
  25. Basha, S.A.K.; Vincent, P.D.R.; Mohammad, S.I.; Vasudevan, A.; Soon, E.E.H.; Shambour, Q.; Alshurideh, M.T. Exploring Deep Learning Methods for Audio Speech Emotion Detection: An Ensemble MFCCs, CNNs and LSTM. Appl. Math 2025, 19, 75–85. [Google Scholar]
  26. Zhao, S.; Jia, G.; Yang, J.; Ding, G.; Keutzer, K. Emotion Recognition From Multiple Modalities: Fundamentals and methodologies. IEEE Signal Process. Mag. 2021, 38, 59–73. [Google Scholar] [CrossRef]
  27. Fu, B.; Gu, C.; Fu, M.; Xia, Y.; Liu, Y. A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals. Front. Neurosci. 2023, 17, 1234162. [Google Scholar] [CrossRef]
  28. Sultana, S.; Rahman, M.S.; Selim, M.R.; Iqbal, M.Z. SUST Bangla Emotional Speech Corpus (SUBESCO): An audio-only emotional speech corpus for Bangla. PLoS ONE 2021, 16, e0250173. [Google Scholar] [CrossRef]
  29. Das, R.K.; Islam, N.; Ahmed, M.R.; Islam, S.; Shatabda, S.; Islam, A.M. BanglaSER: A speech emotion recognition dataset for the Bangla language. Data Brief 2022, 42, 108091. [Google Scholar] [CrossRef] [PubMed]
  30. Livingstone, S.R.; Russo, F.A. The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 2018, 13, e0196391. [Google Scholar] [CrossRef] [PubMed]
  31. Burkhardt, F.; Paeschke, A.; Kienast, M.; Sendlmeier, W.F.; Weiss, B. Berlin EmoDB (1.3.0) [Data Set]. Zenodo. 2022. Available online: https://zenodo.org/records/7447302 (accessed on 5 December 2022).
  32. Paiva, L.F.; Alfaro-Espinoza, E.; Almeida, V.M.; Felix, L.B.; Neves, R.V.A. A Survey of Data Augmentation for Audio Classification. In Proceedings of the XXIV Congresso Brasileiro de Automática (CBA); 2022; Available online: https://sba.org.br/open_journal_systems/index.php/cba/article/view/3469 (accessed on 5 December 2022).
  33. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; Battenberg, E.; Nieto, O.; Dieleman, S.; Tokunaga, H.; McQuin, P.; NumPy; et al. librosa/librosa: 0.10.1. Available online: https://zenodo.org/records/8252662 (accessed on 4 January 2024).
  34. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 6–12 July 2015; Volume 8, pp. 18–25. [Google Scholar]
  35. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  36. Galic, J.; Grozdić, Đ. Exploring the Impact of Data Augmentation Techniques on Automatic Speech Recognition System Development: A Comparative Study. Adv. Electr. Comput. Eng. 2023, 23, 3–12. [Google Scholar] [CrossRef]
  37. Titeux, N. Everything You Need to Know About Pitch Shifting; Nicolas Titeux: Luxembourg, 2023; Available online: https://www.nicolastiteux.com/en/blog/everything-you-need-to-know-about-pitch-shifting/ (accessed on 5 March 2023).
  38. Jordal, I. Gain; Gain—Audiomentations Documentation. 2018. Available online: https://iver56.github.io/audiomentations/waveform_transforms/gain/ (accessed on 4 January 2023).
  39. Kedem, B. Spectral analysis and discrimination by zero-crossings. Proc. IEEE 1986, 74, 1477–1493. [Google Scholar] [CrossRef]
  40. Rezapour Mashhadi, M.M.; Osei-Bonsu, K. Speech Emotion Recognition Using Machine Learning Techniques: Feature Extraction and Comparison of Convolutional Neural Network and Random Forest. PLoS ONE 2023, 18, e0291500. [Google Scholar] [CrossRef]
  41. Shah, A.; Kattel, M.; Nepal, A.; Shrestha, D. Chroma Feature Extraction Using Fourier Transform. In Proceedings of the Conference at Kathmandu University, Nepal, January 2019. Available online: https://www.researchgate.net/publication/330796993_Chroma_Feature_Extraction (accessed on 5 January 2025).
  42. Sharma, G.; Umapathy, K.; Krishnan, S. Trends in Audio Signal Feature Extraction Methods. Appl. Acoust. 2020, 158, 107020. [Google Scholar] [CrossRef]
  43. Kumar, S.; Thiruvenkadam, S. An Analysis of the Impact of Spectral Contrast Feature in Speech Emotion Recognition. Int. J. Recent Contrib. Eng. Sci. IT (iJES) 2021, 9, 87–95. [Google Scholar] [CrossRef]
  44. West, K.; Cox, S. Finding An Optimal Segmentation for Audio Genre Classification. In Proceedings of the 6th International Conference on Music Information Retrieval, ISMIR 2005, London, UK, 11–15 September 2005; pp. 680–685. [Google Scholar]
  45. Peeters, G. A Large Set of Audio Features for Sound Description (Similarity and Classification) in the CUIDADO Project; Ircam: Paris, France, 2004. [Google Scholar]
  46. Panda, R.; Malheiro, R.; Paiva, R.P. Audio Features for Music Emotion Recognition: A Survey. IEEE Trans. Affect. Comput. 2023, 14, 68–88. [Google Scholar] [CrossRef]
  47. Roberts, L. Understanding the Mel Spectrogram. Medium. March 2020. Available online: https://medium.com/analytics-vidhya/understanding-the-mel-spectrogram-fca2afa2ce53 (accessed on 5 January 2023).
  48. Wu, J. Introduction to convolutional neural networks. Natl. Key Lab Nov. Softw. Technol. Nanjing Univ. China 2017, 5, 495. [Google Scholar]
  49. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  50. Zafar, A.; Aamir, M.; Mohd Nawi, N.; Arshad, A.; Riaz, S.; Alruban, A.; Dutta, A.; Alaybani, S. A Comparison of Pooling Methods for Convolutional Neural Networks. Appl. Sci. 2022, 12, 8643. [Google Scholar] [CrossRef]
  51. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar] [CrossRef]
  52. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  53. GeeksforGeeks. What Is a Neural Network Flatten Layer? GeeksforGeeks: Noida, India, 2024; Available online: https://www.geeksforgeeks.org/what-is-a-neural-network-flatten-layer/ (accessed on 10 August 2024).
  54. Nwankpa, C.; Ijomah, W.; Gachagan, A.; Marshall, S. Activation Functions: Comparison of trends in Practice and Research for Deep Learning. arXiv 2018, arXiv:cs.LG/1811.03378. [Google Scholar] [CrossRef]
  55. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar] [CrossRef]
  56. Mohammed, A.; Kora, R. A Comprehensive Review on Ensemble Deep Learning: Opportunities and Challenges. J. King Saud Univ. Comput. Inf. Sci. 2023, 35, 757–774. [Google Scholar] [CrossRef]
Figure 1. An overview of our proposed methodology.
Figure 1. An overview of our proposed methodology.
Jimaging 11 00273 g001
Figure 2. Examples of data augmentations: (a) Original. (b) Noise Added. (c) Time Stretched. (d) Sound Shifted. (e) Pitch Changed.
Figure 2. Examples of data augmentations: (a) Original. (b) Noise Added. (c) Time Stretched. (d) Sound Shifted. (e) Pitch Changed.
Jimaging 11 00273 g002
Figure 3. Spectrograms of audio signals from different emotional categories: (a) Happy. (b) Sad. (c) Angry. (d) Surprised. (e) Fearful. (f) Disgusted. (g) Neutral.
Figure 3. Spectrograms of audio signals from different emotional categories: (a) Happy. (b) Sad. (c) Angry. (d) Surprised. (e) Fearful. (f) Disgusted. (g) Neutral.
Jimaging 11 00273 g003
Figure 4. Architecture of our 1D CNN model.
Figure 4. Architecture of our 1D CNN model.
Jimaging 11 00273 g004
Figure 5. Architecture of our 1D CNN LSTM model.
Figure 5. Architecture of our 1D CNN LSTM model.
Jimaging 11 00273 g005
Figure 6. Architecture of our 1D CNN BiLSTM model.
Figure 6. Architecture of our 1D CNN BiLSTM model.
Jimaging 11 00273 g006
Figure 7. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BiLSTM, respectively) for the SUBESCO dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy curve; (f) 1D CNN-BiLSTM loss curve.
Figure 7. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BiLSTM, respectively) for the SUBESCO dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy curve; (f) 1D CNN-BiLSTM loss curve.
Jimaging 11 00273 g007
Figure 8. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BILSTM, respectively) for the BanglaSER dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy; (f) 1D CNN-BiLSTM loss curve.
Figure 8. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BILSTM, respectively) for the BanglaSER dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy; (f) 1D CNN-BiLSTM loss curve.
Jimaging 11 00273 g008
Figure 9. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BILSTM, respectively) for the SUBESCO and BanglaSER merged dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy curve; (f) 1D CNN-BiLSTM loss curve.
Figure 9. Learning curves (accuracy and loss) of the models (1D CNN, 1D CNN-LSTM, and 1D CNN-BILSTM, respectively) for the SUBESCO and BanglaSER merged dataset. (a) 1D CNN accuracy curve; (b) 1D CNN loss curve; (c) 1D CNN-LSTM accuracy curve; (d) 1D CNN-LSTM loss curve; (e) 1D CNN-BiLSTM accuracy curve; (f) 1D CNN-BiLSTM loss curve.
Jimaging 11 00273 g009
Figure 10. Confusion matrices of the ensemble learning approach for different datasets: (a) SUBESCO; (b) BanglaSER; (c) SUBESCO and BanglaSER merged.
Figure 10. Confusion matrices of the ensemble learning approach for different datasets: (a) SUBESCO; (b) BanglaSER; (c) SUBESCO and BanglaSER merged.
Jimaging 11 00273 g010
Table 1. Emotion classes in the SUBESCO dataset with counts and percentages.
Table 1. Emotion classes in the SUBESCO dataset with counts and percentages.
Class NameTotal CountPercentage
Happy100014.29
Sad100014.29
Angry100014.29
Surprise100014.29
Fear100014.29
Disgust100014.29
Neutral100014.29
Table 2. Emotion classes in the BanglaSER dataset with counts and percentages.
Table 2. Emotion classes in the BanglaSER dataset with counts and percentages.
Class NameTotal CountPercentage
Happy30620.87
Sad30620.87
Angry30620.87
Surprise30620.87
Neutral24316.53
Table 3. Emotion classes in the merged SUBESCO and BanglaSER dataset.
Table 3. Emotion classes in the merged SUBESCO and BanglaSER dataset.
Class NameTotal CountPercentage
Happy130615.41
Sad130615.41
Angry130615.41
Surprise130615.41
Fear100011.80
Disgust100011.80
Neutral124314.67
Table 4. Emotion classes in the RAVDESS dataset.
Table 4. Emotion classes in the RAVDESS dataset.
Class NameTotal CountPercentage
Happy19213.33
Sad19213.33
Angry19213.33
Surprise19213.33
Fear19213.33
Disgust19213.33
Neutral966.67
Table 5. Emotion classes in the EMODB dataset.
Table 5. Emotion classes in the EMODB dataset.
Class NameTotal CountPercentage
Happy7113.27
Sad6211.59
Angry12723.74
Boredom8115.14
Fear6912.90
Disgust468.60
Neutral7914.77
Table 6. Overview of our train–test samples.
Table 6. Overview of our train–test samples.
Dataset NameTotal SamplesTrain/Test RatioTrain SamplesTest Samples
SUBESCO700070/3049002100
BanglaSER146780/201173294
SUBESCO and BanglaSER merged846770/3059262541
RAVDESS (Audio-only)124870/301008432
EMODB53580/20428107
Table 7. Data augmentation techniques and their descriptions.
Table 7. Data augmentation techniques and their descriptions.
Augmentation NameDescription
Polarity InversionReverses the phase of the audio signal by multiplying it by 1 , effectively canceling the phase when combined with the original signal, resulting in silence [36]. It can simulate variations in recordings caused by microphone polarity errors, which may slightly enhance the generalization of the model. Polarity inversion was applied by multiplying the signal by 1 .
Noise AdditionAdds random white noise to the audio data to enhance its variability and robustness [9]. We have added white Gaussian noise at 10 dB and 20 dB SNR using numpy.random.normal() with a noise factor of 0.005.
Time StretchingAlters the speed of the audio by stretching or compressing time series data, increasing or decreasing sound speed [9]. We have applied time stretching on our audio data with a rate (stretch factor) of 0.8 using librosa.time_stretch() function.
Pitch ChangeChanges the pitch of the audio signal by adjusting the frequency of sound components, typically by resampling [37]. We have shifted ±2 semitones of our audio data with a pitch factor (steps to shift) of 0.8 using librosa.effects.pitch_shift() function.
Sound ShiftingRandomly shifts the audio by a predefined number of seconds, introducing silence at the shifted location if necessary [9]. We have shifted audio ±5 ms using numpy.roll() with shifts from random.uniform( 5 , 5).
Random GainAlters the loudness of the audio signal using a volume factor, making it louder or softer [38]. To apply random gain, we randomly adjusted volume between ×2 and ×4 using random.uniform().
Table 8. Feature extraction and their advantages.
Table 8. Feature extraction and their advantages.
Feature NameDescription and Advantage
Zero-Crossing Rate (ZCR)Counts the number of times the audio signal crosses the horizontal axis. It helps analyze signal smoothness and is effective for distinguishing voiced from unvoiced speech [39,40].
ChromagramRepresents energy distribution over frequency bands corresponding to pitch classes in music. It captures harmonic and melodic features of the signal, useful for tonal analysis [41].
Spectral CentroidIndicates the “center of mass” of a sound’s frequencies, providing insights into the brightness of the sound. It is useful for identifying timbral characteristics [42].
Spectral Roll-offMeasures the frequency below which a certain percentage of the spectral energy is contained. This feature helps in distinguishing harmonic from non-harmonic content [42].
Spectral ContrastMeasures the difference in energy between peaks and valleys in the spectrum, capturing timbral texture and distinguishing between different sound sources [43,44].
Spectral FlatnessQuantifies how noise-like a sound is. A high spectral flatness value indicates noise-like sounds, while a low value indicates tonal sounds, useful for identifying the type of sound [45].
Mel-Frequency Cepstral Coefficients (MFCCs)Captures spectral variations in speech, focusing on features most relevant to human hearing. It is widely used in speech recognition and enhances emotion recognition capabilities [42,45].
Root Mean Square (RMS) EnergyMeasures the loudness of the audio signal, offering insights into the energy of the sound, which is crucial for understanding the emotional intensity [46].
Mel-SpectrogramConverts the frequencies of a spectrogram to the mel scale, representing the energy distribution in a perceptually relevant way, commonly used in speech and audio processing [47].
Table 9. Summary of performances (accuracy %) of our proposed models.
Table 9. Summary of performances (accuracy %) of our proposed models.
Dataset1D CNN1D CNN LSTM1D CNN BiLSTMEnsemble Learning
SUBESCO90.93%90.98%90.50%92.90%
BanglaSER83.67%84.52%81.97%85.20%
SUBESCO + BanglaSER88.92%88.61%87.56%90.63%
RAVDESS65.63%64.93%60.76%67.71%
EMODB69.57%67.39%65.84%69.25%
Table 10. Model accuracy on the SUBESCO dataset.
Table 10. Model accuracy on the SUBESCO dataset.
ModelAccuracy
1D CNN90.93%
1D CNN LSTM90.98%
1D CNN BiLSTM90.50%
Ensemble Learning92.90%
Table 11. Accuracy of the Ensemble Learning approach for each emotion class of SUBESCO dataset.
Table 11. Accuracy of the Ensemble Learning approach for each emotion class of SUBESCO dataset.
EmotionAccuracy (%)
Angry93.93
Disgust84.98
Fear93.61
Happy94.98
Neutral98.69
Sad92.32
Surprise92.05
Table 12. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of SUBESCO dataset.
Table 12. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of SUBESCO dataset.
ClassPrecision (%)Recall (%)F1-Score (%)
Angry94.0094.0094.00
Disgust92.0085.0088.00
Fear93.0093.0093.00
Happy91.0095.0093.00
Neutral94.0099.0096.00
Sad95.0092.0094.00
Surprise91.0092.0092.00
Macro Average93.0093.0093.00
Weighted Average93.0093.0093.00
Accuracy (%) = 93.00
Table 13. Model accuracy on the BanglaSER dataset.
Table 13. Model accuracy on the BanglaSER dataset.
ModelAccuracy
1D CNN83.67%
1D CNN LSTM84.52%
1D CNN BiLSTM81.97%
Ensemble Learning85.20%
Table 14. Accuracy of the Ensemble Learning approach for each emotion class of BanglaSER dataset.
Table 14. Accuracy of the Ensemble Learning approach for each emotion class of BanglaSER dataset.
EmotionAccuracy (%)
Angry93.07
Happy76.61
Neutral93.02
Sad83.59
Surprise81.67
Table 15. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of BanglaSER dataset.
Table 15. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of BanglaSER dataset.
ClassPrecision (%)Recall (%)F1-Score (%)
Angry91.0093.0092.00
Happy83.0077.0080.00
Neutral89.0093.0091.00
Sad85.0084.0084.00
Surprise78.0082.0080.00
Macro Average85.0085.0085.00
Weighted Average85.0085.0085.00
Accuracy (%) = 85.00
Table 16. Model accuracy on the SUBESCO and BanglaSER merged dataset.
Table 16. Model accuracy on the SUBESCO and BanglaSER merged dataset.
ModelAccuracy
1D CNN88.92%
1D CNN LSTM88.61%
1D CNN BiLSTM87.56%
Ensemble Learning90.63%
Table 17. Accuracy of the Ensemble Learning approach for each emotion class of SUBESCO and BanglaSER merged dataset.
Table 17. Accuracy of the Ensemble Learning approach for each emotion class of SUBESCO and BanglaSER merged dataset.
EmotionAccuracy (%)
Angry94.84
Disgust86.19
Fear92.69
Happy89.03
Neutral96.74
Sad87.53
Surprise87.15
Table 18. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of SUBESCO and BanglaSER merged dataset.
Table 18. Performance metrics (Precision, Recall, and F1-Score) of the Ensemble Learning approach for each emotion class of SUBESCO and BanglaSER merged dataset.
ClassPrecision (%)Recall (%)F1-Score (%)
Angry93.0095.0094.00
Disgust91.0086.0088.00
Fear92.0093.0092.00
Happy89.0089.0089.00
Neutral90.0097.0093.00
Sad89.0088.0088.00
Surprise90.0087.0089.00
Macro Average91.0091.0091.00
Weighted Average91.0091.0091.00
Accuracy (%) = 91.00
Table 19. Model accuracy on the RAVDESS dataset.
Table 19. Model accuracy on the RAVDESS dataset.
ModelAccuracy
1D CNN65.63%
1D CNN LSTM64.93%
1D CNN BiLSTM60.76%
Ensemble Learning67.71%
Table 20. Model accuracy on the EMODB dataset.
Table 20. Model accuracy on the EMODB dataset.
ModelAccuracy
1D CNN69.57%
1D CNN LSTM67.39%
1D CNN BiLSTM65.84%
Ensemble Learning69.25%
Table 21. State of the art comparison of the proposed model.
Table 21. State of the art comparison of the proposed model.
ResearchFeatures UsedModelAccuracy for 5 Datasets (%)
SUBESCO (Bangla) RAVDESS (American English) EMO-DB BanglaSER (SUBESCO + BanglaSER)
Sultana et al. [14]Mel-spectrogramDeep CNN and BLSTM86.982.7
Rahman et al. (2018) [15]MFCCs, MFCC derivativesSVM with RBF, DTW86.08 (Bangla 12 speakers)-
Chakraborty  et al. (2022) [8]PBCCP Gradient Boosting Machine 96 96
Issa, et al. [16]MFCCs, Mel-spectrogram, Chroma-gram, Spectral contrast feature, Tonnetz representation1D CNN64.3% (IEMOCAP, 4 classes)71.61 (8 classes)95.71
Zhao, et al. 2019 [17]Log mel spectrogram1D CNN LSTM, 2D CNN LSTM89.16 (IEMOCAP dependent)52.14% (IEMOCAP independent) 95.33 (Emo-Db dependent)95.89% (independent)
Mustaqeem et al. [18]Spectral analysis1D Dilated CNN with BiGRU72.7578.0191.14--
Badshah et al. (2017) [19]SpectrogramsCNN (3 convolutional, 3 FC layers) 56%
Etienne, et al. (2018) [20]High-level features, log-mel SpectrogramCNN-LSTM (4 conv + 1 BLSTM layer)61.7% (Unweighted), 64.5% (Weighted)
ProposedZCR, chroma-gram, RMS, spectral centroid, spectral roll-off, spectral contrast, spectral flatness, mel spectrogram, and  MFCCsEnsemble of 1D CNN, 1D CNN LSTM, and 1D CNN BiLSTM92.90 % (SUBESCO),67.71% (RAVDESS),69.25% (EMODB)85.20% (BanglaSER),90.63% (SUBESCO + BanglaSER)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shakil, M.S.A.; Farid, F.A.; Podder, N.K.; Iqbal, S.M.H.S.; Miah, A.S.M.; Rahim, M.A.; Karim, H.A. Bangla Speech Emotion Recognition Using Deep Learning-Based Ensemble Learning and Feature Fusion. J. Imaging 2025, 11, 273. https://doi.org/10.3390/jimaging11080273

AMA Style

Shakil MSA, Farid FA, Podder NK, Iqbal SMHS, Miah ASM, Rahim MA, Karim HA. Bangla Speech Emotion Recognition Using Deep Learning-Based Ensemble Learning and Feature Fusion. Journal of Imaging. 2025; 11(8):273. https://doi.org/10.3390/jimaging11080273

Chicago/Turabian Style

Shakil, Md. Shahid Ahammed, Fahmid Al Farid, Nitun Kumar Podder, S. M. Hasan Sazzad Iqbal, Abu Saleh Musa Miah, Md Abdur Rahim, and Hezerul Abdul Karim. 2025. "Bangla Speech Emotion Recognition Using Deep Learning-Based Ensemble Learning and Feature Fusion" Journal of Imaging 11, no. 8: 273. https://doi.org/10.3390/jimaging11080273

APA Style

Shakil, M. S. A., Farid, F. A., Podder, N. K., Iqbal, S. M. H. S., Miah, A. S. M., Rahim, M. A., & Karim, H. A. (2025). Bangla Speech Emotion Recognition Using Deep Learning-Based Ensemble Learning and Feature Fusion. Journal of Imaging, 11(8), 273. https://doi.org/10.3390/jimaging11080273

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop