Next Article in Journal
EFDepth: A Monocular Depth Estimation Model for Multi-Scale Feature Optimization
Previous Article in Journal
From Static to Dynamic: Complementary Roles of FSR and Piezoelectric Sensors in Wearable Gait and Pressure Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MAGTF-Net: Dynamic Speech Emotion Recognition with Multi-Scale Graph Attention and LLD Feature Fusion

1
Key Laboratory of Information and Communication Systems, Ministry of Information Industry, Beijing Information Science and Technology University, Beijing 100192, China
2
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(23), 7378; https://doi.org/10.3390/s25237378
Submission received: 6 November 2025 / Revised: 30 November 2025 / Accepted: 2 December 2025 / Published: 4 December 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

In this paper, we propose a novel speech emotion recognition model, named MAGTF-Net (Multi-scale Attention Graph Transformer Fusion Network), which addresses the challenges faced by traditional hand-crafted feature-based approaches in modeling complex emotional nuances and dynamic contextual dependencies. Although existing state-of-the-art methods have achieved improvements in recognition performance, they often fail to simultaneously capture both local acoustic features and global temporal structures, and they lack adaptability to variable-length speech utterances, thereby limiting their accuracy and robustness in recognizing complex emotional expressions. To tackle these challenges, we design a log-Mel spectrogram feature extraction branch that combines a Multi-scale Attention Graph (MAG) structure with a Transformer encoder, where the Transformer module adaptively performs dynamic modeling of speech sequences with varying lengths. In addition, a low-level descriptor (LLD) feature branch is introduced, where a multilayer perceptron (MLP) is employed for complementary feature modeling. The two feature branches are fused and subsequently classified through a fully connected layer, further enhancing the expressive capability of emotional representations. Moreover, a label-smoothing-enhanced cross-entropy loss function is adopted to improve the model’s recognition performance on difficult-to-classify emotional categories. Experiments conducted on the IEMOCAP dataset demonstrate that MAGTF-Net achieves weighted accuracy (WA) and unweighted accuracy (UA) scores of 69.15% and 70.86%, respectively, outperforming several baseline models. Further ablation studies validate the significant contributions of each module in the Mel-spectrogram branch and the LLD feature branch to the overall performance improvement. The proposed method effectively integrates local, global, and multi-source feature information, significantly enhancing the recognition of complex emotional expressions and providing new theoretical and practical insights for the field of speech emotion recognition.

1. Introduction

Speech emotion recognition (SER) is a critical subfield of affective computing that aims to automatically identify the emotional states of speakers through the analysis of speech signals [1,2]. As early as 1872, Darwin highlighted the existence of emotional expressions in both humans and animals in his seminal work The Expression of the Emotions in Man and Animals [3]. Subsequently, Ekman proposed the six basic emotions theory, which has significantly advanced the study of emotions [4].
In recent years, the rapid development of artificial intelligence and human–computer interaction technologies has propelled SER from academic research into practical applications such as intelligent customer service, virtual assistants, and mental health monitoring [5,6,7], demonstrating its substantial research potential and practical value.
Traditional SER methods primarily relied on machine learning models, such as support vector machines (SVMs), Gaussian mixture models (GMMs), and hidden Markov models (HMMs), utilizing hand-crafted acoustic features like Mel-frequency cepstral coefficients (MFCCs), spectral features, and prosodic cues. Although these methods achieved some success, they struggled with dynamic emotional expressions and exhibited limited robustness under noisy real-world conditions [8,9,10].
The advent of deep learning has significantly advanced SER. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have enabled the automatic extraction of high-dimensional speech features, capturing long-range temporal dependencies more effectively and achieving superior recognition performance compared to traditional methods [11,12].
Despite these advancements, existing methods still face challenges in modeling complex emotional expressions. Specifically, there is a need for improved local time-frequency feature modeling, enhanced global emotional structure representation, and more efficient multi-source feature fusion strategies. Current SER approaches can be broadly categorized into three types:
  • Handcrafted Feature-Based Methods: These methods employ manually designed acoustic features such as MFCCs and low-level descriptors (LLDs) [13], classified using traditional machine learning models [14]. While simple to implement, these features have limited expressiveness and are sensitive to noise, leading to poor generalization.
  • End-to-End Deep Learning Methods: Deep models such as CNNs, RNNs, long short-term memory networks (LSTMs), and Transformers learn features directly from raw speech or spectrograms, improving emotion recognition by automatically extracting rich feature representations [15].
  • Multi-branch and Multi-scale Feature Fusion Methods: Recent research emphasizes the fusion of multi-scale, multi-source features through parallel network branches to better capture complex emotional characteristics [16,17].
However, there is still considerable room for improvement in feature fusion strategies, model efficiency, and generalization capability, as recently reported in several SER studies on IEMOCAP and Emo-DB [18,19]. To address these challenges, this paper proposes a dual-branch feature fusion network for SER. The framework includes the following:
  • A Mel-spectrogram Branch: Incorporating a Multi-scale Attention Graph (MAG) module and a Transformer encoder to effectively model both local and global emotional features;
  • An LLD Feature Branch: Utilizing multi-layer perceptrons (MLPs) to extract and complement fine-grained frame-level emotional cues.
This study leverages a fusion network to comprehensively integrate multi-scale features for holistic emotional representation. The core objectives are to capture speech’s local emotional details, model global long-range dependencies, and efficiently fuse cross-dimensional features—enhancing SER model accuracy, generalization, and robustness to advance affective computing research and applications. Key contributions are as follows:
  • We propose a novel dual-branch architecture. One branch integrates the MAG and Transformer structures, effectively combining the advantages of multi-scale local feature extraction and global temporal modeling, thereby significantly improving the performance of speech emotion recognition tasks. The other branch deeply explores the low-level descriptor (LLD) features, further enhancing the model’s ability to capture fine-grained emotional details and improving its robustness and generalization capability.
  • We introduce a dynamic natural-length modeling mechanism that better adapts to variable-length speech utterances. To address the varying lengths of Mel-spectrogram features in natural speech, this study employs a dynamic padding and masking mechanism for batch-wise alignment and effective frame modeling.
    The padding operation ensures that the time dimensions of different samples are unified, while the mask matrix selectively masks out the padded frames during the Transformer encoding and global pooling processes, enabling feature learning and aggregation to be performed only on valid frames.This design effectively enhances the accuracy and robustness of sequence modeling.
  • Extensive experiments conducted on benchmark datasets (IEMOCAP, EMO-DB, CASIA) demonstrate the effectiveness and generalization ability of the proposed method across different languages and environments.
The remainder of this paper is organized as follows: Section 2 reviews related work. Section 3 details the proposed method. Section 4 presents experimental results and analysis. Section 5 concludes the paper and discusses future directions.

2. Related Work

2.1. Traditional Machine Learning-Based Methods

Early approaches to speech emotion recognition (SER) primarily relied on traditional machine learning models, such as Gaussian mixture models (GMMs) [20], support vector machines (SVMs) [21], and hidden Markov models (HMMs) [8]. These methods typically utilized hand-crafted acoustic features, including Mel-frequency cepstral coefficients (MFCCs), perceptual linear predictive coefficients (PLPs), and low-level descriptors (LLDs) [22].
While these traditional methods demonstrated reasonable performance in controlled environments, they exhibited several limitations. Notably, they were highly sensitive to environmental noise and variability, struggled to capture the complex and dynamic nature of emotional expressions in real-world conditions, and often suffered from poor generalization to unseen data [23]. As a result, the need for more robust and adaptive SER systems became increasingly evident.

2.2. Deep Learning-Based Methods

To address the limitations of traditional methods, deep learning models represented by convolutional neural networks (CNNs) and recurrent neural networks (RNNs), particularly long short-term memory (LSTM) networks, have gradually attracted increasing attention from researchers.
These models are capable of automatically learning complex features from spectrograms or raw waveforms, effectively improving the accuracy of emotion recognition [24]. CNNs are advantageous for effectively capturing local patterns in the speech spectrogram and are particularly suitable for analyzing two-dimensional features.However, CNN models generally struggle to explicitly model the sequential and contextual relationships within speech signals.
To better capture the dynamic variations in speech signals, researchers have introduced recurrent neural networks (RNNs) and their variants, such as long short-term memory (LSTM) networks and gated recurrent units (GRUs). For instance, Tzinis et al. [25] proposed an end-to-end framework that integrates multi-scale CNNs with an attention mechanism to extract multi-granularity acoustic features and fuse key frame information, thereby enhancing temporal modeling. Similarly, Latif et al. [26] highlighted the advantages of LSTM and GRU in continuous emotion recognition and complex context understanding. However, these architectures still face challenges such as vanishing gradients and limited efficiency when modeling long-range dependencies [27].
In recent years, the Transformer model, with its self-attention mechanism, has offered new opportunities for speech emotion recognition (SER). Its ability to capture long-range dependencies has been validated in both speech and emotion recognition tasks [28,29]. For instance, ref. [30] proposed a Transformer-based framework that employs multi-head self-attention to effectively capture long-range contextual information; Li et al. [31] introduced DSTCNet with a triple attention mechanism across spectro-temporal-channel dimensions, significantly improving emotional information capture; Zhao [32] proposed Dual-TBNet, which combines Transformer and BiLSTM to jointly model short- and long-term dependencies, further enhancing robustness in complex emotion recognition.
Although Transformer-based architectures exhibit strong performance in capturing long-range information, there remains room for improvement in effectively modeling multi-scale local detail features. Recent transformer-based SER models also confirm that attention over Mel-spectrograms and sample-aware training can further improve robustness [33,34].

2.3. Multi-Branch and Multi-Scale Feature Fusion Methods

To more comprehensively model complex emotional features, there has been a growing interest in multi-scale and multi-branch feature fusion approaches in recent years. For example, ref. [17] proposed a multi-level acoustic information fusion method based on a co-attention mechanism, which effectively improves emotion recognition performance by integrating acoustic features across different scales. In addition, graph convolutional networks (GCNs) have also been successfully applied to speech emotion modeling [35]. Chandola et al. [36] proposed SERC-GCN, which constructs a dialogue-level emotional propagation graph structure, enabling the model to achieve good performance in capturing emotional dynamics within conversations. Beyond speech-related tasks, GNNs have also been investigated from broader methodological and system perspectives. Zhao et al. proposed a flexible diffusion convolution that generalizes message passing through node-adaptive continuous diffusion processes [37]. Ma et al. reviewed key GNN acceleration strategies, including sampling, simplification, quantization, pruning, and distillation [38]. Liu et al. modeled the micro-level mobility patterns of COVID-19 cases using a deep graph diffusion framework, showing that GNNs can effectively capture complex dependencies between geometric location graphs and diffusion trajectories in real-world mobility networks [39]. Complementarily, Hang et al. proposed MEGA, a machine learning–enhanced graph-analytics framework that combines automated feature engineering with GNN-based learning to detect spambots and influential spreaders in large-scale COVID-19 Twitter graphs, highlighting the effectiveness of GNNs in relational inference and structural pattern discovery [40]. These examples underscore the generality of graph architectures and motivate our adoption of a graph-based design in the proposed MAG module.
However, such multi-scale fusion methods often involve high computational complexity and intricate model structures, and there is still substantial room for improvement in the design of effective feature fusion strategies.
In addition, to further enhance the accuracy and robustness of emotion recognition, recent research has increasingly focused on multimodal approaches that integrate speech, text, and visual information. For example, in [41], a model combining speech and text modalities was proposed, leveraging a convolutional neural network (CNN) and a pre-trained language model (BERT) to effectively improve recognition performance. Building upon this, ref. [42] introduced visual information (such as facial expression features) and proposed a multimodal model that fuses speech, text, and visual features, significantly enhancing the recognition of complex emotional states.
However, despite the performance gains, multimodal fusion methods still face several critical challenges. Notably, significant modality discrepancies often exist between features from different modalities, and the extraction of high-quality text features heavily relies on accurate automatic speech recognition (ASR) systems. In real-world environments (e.g., under low signal-to-noise ratio conditions), ASR errors can lead to degraded emotion recognition accuracy [43]. Therefore, how to achieve improved model precision while simultaneously optimizing computational efficiency and generalization performance remains one of the key research directions in the field of multimodal emotion recognition.
In summary, although significant progress has been made in speech emotion recognition, there remains considerable room for further research, particularly in simultaneously capturing both local details and global temporal features, effectively fusing different types of speech features, and ensuring model generalization capability alongside computational efficiency.
To address these issues, our work proposes a dual-branch fusion framework that separately extracts Mel-spectrogram-based and LLD-based features. By combining multi-scale local feature modeling and global sequence modeling, we aim to achieve more comprehensive emotional representation with improved efficiency and robustness.

3. Methods

3.1. Overall Framework

The overall architecture of the proposed MAGTF-Net is illustrated in Figure 1. The framework integrates two parallel branches to extract and fuse multi-scale and multi-source emotional features from speech signals.
Specifically, the input raw audio signals are first preprocessed to obtain two types of features:
  • Log-Mel spectrogram features are extracted through short-time Fourier transform (STFT) followed by Mel-filterbank projection and logarithmic compression;
  • Low-level descriptors (LLDs) are extracted using standard acoustic feature extraction techniques, including short-time energy, zero-crossing rate, spectral centroid, bandwidth, roll-off, and spectral contrast.
Formally, given an input audio sample x, the extracted Mel-spectrogram feature matrix X m R T × F can be expressed as
X m = LogMel ( STFT ( x ) ) .
In the Mel-spectrogram branch, the input Mel-spectrogram features X m are first fed into the MAG encoder to perform multi-scale local feature extraction and modeling, resulting in the MAG feature representation X m . Subsequently, the output features from the MAG encoder are passed into the Transformer encoder, which captures the global temporal dependencies of the speech signal and produces the Transformer output feature representation X m . After the Transformer processing, the extracted features combine both local and global emotional information, forming the final feature representation of the Mel-spectrogram branch.
In the LLD feature branch, the LLD feature sequence X lld R T × D is extracted as
X lld = ExtractLLD ( x ) ,
where D represents the dimensionality of the LLD feature vector per frame. The LLD features are passed through a structure consisting of a multilayer perceptron (MLP) and batch normalization operations for feature extraction and mapping, resulting in an LLD feature representation X L that effectively captures detailed emotional information in speech. The purpose of this branch is to supplement the emotional features that may be overlooked by the Mel-spectrogram branch, thereby further enhancing the emotion recognition performance of the model.
The feature representations obtained from the two branches are fused through the feature fusion module. The fused feature representation can be expressed as
X = X m ; X L ,
and the concatenated feature X is then processed by the classifier. The final emotion category prediction is obtained using the softmax function:
y ^ = softmax W fc X + b fc ,
where W fc and b fc represent the weight matrix and bias vector of the classifier, respectively.

3.2. Mel-Spectrogram Branch

This branch first extracts the log-Mel spectrogram features from the raw audio signals. Subsequently, multi-scale convolution and channel-attention mechanisms are applied to capture local features, followed by a local graph convolution module to enhance inter-frame dependencies. Finally, a Transformer encoder is employed to model global temporal relationships, generating a fixed-dimensional feature vector. The model structure is shown in Figure 2.
The raw audio signals are loaded using the Librosa library at a sampling rate of 16 kHz. During the data loading stage, no additional static offset removal or normalization is performed; the original waveform data is preserved for subsequent feature extraction. The STFT is first applied to convert the signal into the frequency domain, followed by a Mel filterbank projection to map the frequencies onto a perceptual scale. A logarithmic compression is then performed to obtain the log-Mel spectrogram features. The extraction parameters are set as follows: FFT size ( n fft ): 512; window length ( win _ length ): 400 (corresponding to 25 ms); hop length ( hop _ length ): 160 (corresponding to 10 ms); and number of Mel filters ( n mels ): 40.
Due to the varying durations of different speech samples, the extracted Mel-spectrogram features have different frame lengths T across samples. To enable batch-wise parallel training, a sequence alignment and masking mechanism is adopted during the data preprocessing stage. Specifically, during data loading, padding is applied to zero-pad all samples along the time dimension to match the maximum frame length within each batch. Simultaneously, a mask matrix is constructed to indicate whether each frame is valid (mask = 1 ) or a padded frame (mask = 0 ). The mask matrix is later utilized during the Transformer encoder and the global pooling stages to selectively model and aggregate only the valid frames, effectively avoiding the influence of padded frames on the learning process. This selective masking mechanism also reduces data dispersion introduced by variable-length utterances. Since padded positions are explicitly excluded from both attention computation and global pooling, the Transformer encoder operates on a consistent distribution of valid frames across different samples, thereby improving the stability of temporal modeling under heterogeneous speech durations.
Building upon the initially extracted log-Mel spectrogram features, a multi-scale convolution module is employed to dynamically encode local features. This module consists of five parallel one-dimensional convolutional branches, each using a different kernel size ( 3 , 5 , 7 , 9 , 11 ) to capture local variations at different temporal scales. The computation process for each branch is expressed as
H i = σ Conv 1 D k i ( X m ) , k i { 3 , 5 , 7 , 9 , 11 } .
This approach, based on varying convolutional receptive fields, enables the network to dynamically capture local temporal features without manually fixing window segmentation. Subsequently, the output of each branch is further enhanced through a channel-attention module (SE block). The SE block first performs global average pooling along the temporal dimension to generate a channel-wise descriptor:
s = 1 T t = 1 T H i ( t ) .
The descriptor is then passed through two fully connected layers to generate the channel-attention weights a, which are multiplied with the original features along the channel dimension:
H ^ i = H i · a .
Each branch output is activated by a ReLU function and then fused to obtain the dynamic local feature representation.
To further enhance the dependencies between frames within local temporal segments, a local graph convolution module (GraphBlock) is introduced after the output. Before applying the convolution-based refinement procedure described below, we explicitly construct a learnable latent acoustic graph within the GraphBlock. Unlike conventional convolution-only approaches that implicitly capture inter-channel correlations through receptive fields, the proposed design defines a graph with c out learnable nodes along the channel dimension of the Mel-spectrogram representation. Two trainable node embedding matrices, E 1 R c out × r and E 2 R r × c out (with r = 10 ), are used to generate the adjacency matrix:
A = softmax ReLU ( E 1 E 2 ) ,
where each row of A R c out × c out is normalized to ensure valid outgoing edge weights. The adjacency matrix thus captures learnable correlations among latent acoustic subspaces. Graph propagation is then applied independently at each time step according to
X graph ( t ) = X m ( t ) A ,
allowing each channel to aggregate information from acoustically related nodes before entering the subsequent convolutional operations.
In the GraphBlock, a two-dimensional convolution (StartConv) with a kernel size of ( d model c out + 1 , 1 ) is first applied to integrate local region information. Subsequently, two 1 × 1 convolutions followed by GELU activation functions are used to complete the nonlinear transformations:
H = Conv 2 D ( d model c out + 1 , 1 ) ( H ) ,
H = σ Conv 2 D ( ( 1 , 1 ) ) ( H ) ,
Z graph = Conv 2 D ( ( 1 , 1 ) ) ( H ) .
Then, average pooling is performed on Z graph along the spatial dimension, and the result is transposed back to the sequence format. A linear projection is subsequently applied to adjust the channel dimension to 128, followed by a residual connection with the original input and a layer normalization operation:
Z local = LayerNorm Z graph + X m .
Subsequently, Z local is projected into a lower-dimensional embedding space through a linear layer, where the embedding dimension is set to 128. The embedded sequence is then fed into two stacked Transformer encoder layers, each consisting of a multi-head self-attention mechanism and a feedforward neural network. After residual connections and layer normalization, the global temporal feature representation Z trans R T × 128 is obtained:
Z trans = Z local W embed + b embed .
The self-attention operation is defined as
Attention ( Q , K , V ) = softmax Q K T d k V .
Due to the varying sequence lengths of input samples, a weighted global average pooling is applied to the Transformer outputs using the previously constructed mask matrix. Finally, a fully connected layer is applied to map the pooled vector to the desired dimensionality, resulting in the final output of the Mel-spectrogram branch:
z mel = 1 lengths t = 1 T X t · mask t ,
z mel , out = z mel W dense + b dense .

3.3. LLD Branch and Feature Fusion

To further enrich the emotional feature representation, an auxiliary feature extraction branch based on low-level descriptors (LLDs) is designed in addition to the Mel-spectrogram feature branch, as shown in Figure 3. At the final stage, the features extracted from the LLD branch are fused with those from the Mel-spectrogram branch to achieve multi-source information complementation.
The LLD features X lld are fed into a two-layer multilayer perceptron (MLP) for feature transformation and compression. Each MLP layer consists of a linear projection followed by a ReLU activation function:
H = σ X lld W + b ,
where W and b represent the weights and biases of the MLP, and σ denotes the ReLU activation function. To accommodate variable-length inputs, the LLD branch also adopts a mask-based weighted global average pooling mechanism, where only valid frames are pooled to generate a fixed-dimensional vector:
z lld , out = 1 lengths t = 1 T H t · mask t .
The feature vector output from the Mel-spectrogram branch and the feature vector output from the LLD branch are concatenated to obtain the fused feature:
z fused = z mel , out ; z lld , out .
The fused feature is then fed into the final classifier, where it is processed through a fully connected layer followed by a softmax function to output the emotion category distribution:
y ^ = softmax W cls z fused + b cls ,
where W cls and b cls represent the weights and biases of the classifier. Through the above multi-branch feature fusion design, the model is able to integrate both local dynamic variation features and fine-grained frame-level descriptive information, thereby enhancing the robustness and classification accuracy of speech emotion recognition.

3.4. Interpretability of MAGTF-Net

Although MAGTF-Net adopts a deep neural architecture, several components of the model inherently support interpretability due to their structural roles in the feature extraction process. These components provide transparent links between acoustic properties and the model’s learned representations.
First, the multi-scale convolutional block introduces interpretable temporal receptive fields. Each kernel size k { 3 , 5 , 7 , 9 , 11 } captures affective information at a different temporal resolution. Small kernels emphasize short-term acoustic fluctuations such as pitch jitter, micro-prosodic variations, and abrupt spectral changes, whereas larger kernels aggregate information over broader spans to capture long-range prosodic contours. This multi-resolution design naturally decomposes emotional cues across different timescales.
Second, the graph module provides relational interpretability enhanced by graph attention mechanisms. The learnable adjacency matrix explicitly models interactions among feature channels or temporal units. On top of this structure, the graph attention mechanism assigns importance weights to these relationships, enabling the model to selectively emphasize emotionally salient spectral bands or time regions. Higher attention coefficients indicate stronger influence in the graph propagation process. This mechanism also enables future visualization techniques—such as frequency-wise heatmaps or relevance maps over time—to highlight which parts of the spectro-temporal representation contribute most to emotional classification.
Third, the Transformer encoder contributes contextual interpretability. Self-attention weights reveal how distant frames within an utterance interact. Peaks in attention often correspond to emotionally informative regions such as stressed syllables, tonal transitions, or energy changes. This complements what is learned locally by multi-scale convolution and graph attention, providing a global temporal perspective.
Finally, the LLD branch is intrinsically interpretable because its descriptors have explicit physical meaning. Features such as RMS energy, spectral centroid, bandwidth, and spectral contrast are well-established acoustic indicators of emotional expression. Their role within the fusion process therefore provides a transparent acoustic interpretation that complements the learned Mel-based representations.
Together, these components form a layered interpretability framework: multi-scale convolution explains temporal granularity, graph attention highlights salient relational patterns, Transformer attention captures contextual relevance, and LLD descriptors provide physically grounded cues. This multi-perspective interpretability enhances the transparency and credibility of MAGTF-Net.

4. Results

4.1. Datasets

To verify the effectiveness of the proposed method, experiments were conducted on three widely used speech emotion recognition datasets: EMO-DB [44], IEMOCAP [45], and CASIA [46]. These datasets cover different languages, recording conditions, and emotion categories, making them suitable for evaluating the robustness and generalization capability of the model across multilingual and multi-context scenarios.
(1) IEMOCAP: Released by the University of Southern California, IEMOCAP is an English multimodal emotional database consisting of dialogue utterances from 10 speakers (5 males and 5 females) across five different sessions. It includes both improvised and scripted scenarios. Each speech sample is recorded at a sampling rate of 16 kHz and annotated with emotion labels by three annotators. In this work, only the improvised utterances are selected. Following prior studies, the excited and happiness categories are merged, and four emotional categories are retained: neutral, anger, sadness, and happiness.
(2) EMO-DB: Released by the Technical University of Berlin, EMO-DB is a German emotional speech database comprising utterances recorded by 10 professional actors (5 males and 5 females). The dataset contains a total of 535 speech samples covering seven emotion categories, including happiness, disgust, fear, boredom, neutral, sadness, and anger. All samples are recorded at a sampling rate of 16 kHz.
(3) CASIA: Released by the Institute of Acoustics, Chinese Academy of Sciences, CASIA is a Chinese emotional speech database consisting of utterances recorded by four professional announcers (2 males and 2 females). It covers six emotional states: anger, fear, happiness, neutral, sadness, and surprise. All samples are recorded at a sampling rate of 16 kHz in a professional recording environment.
In this experiment, the total numbers of speech samples in the three datasets (CASIA, IEMOCAP, and EMO-DB) are 960, 2937, and 535, respectively. Each dataset is divided into training, validation, and test sets following an 8:1:1 ratio [31].

4.2. Experimental Setup

To maintain a unified experimental protocol across the three emotional speech datasets (IEMOCAP, CASIA, and EMO-DB), all datasets are divided into training, validation, and test sets following an 8:1:1 utterance-level split. This configuration is consistent with recent SER studies such as Dual-TBNet [32], and ensures comparability across multilingual corpora with different speaker and session annotations. The utterance-level splitting strategy also exposes the model to diverse speaker characteristics and acoustic conditions during training, thereby enhancing the generalization capability of the proposed framework.
During preprocessing, raw audio signals were directly used for Mel-spectrogram extraction without fixed-length segmentation. To handle variable-length Mel-spectrograms in natural speech, dynamic padding and masking were applied to align batches and model valid frames. For the LLD branch, global statistical features were computed for each entire audio sample to obtain a fixed-length feature vector, which served as the model input.
Although no global normalization is applied to the raw waveform signals, the proposed preprocessing pipeline effectively mitigates data dispersion across speakers and recording sessions. In the Mel-spectrogram branch, dynamic padding and a binary mask are adopted to align variable-length utterances within each batch. The mask ensures that only valid frames participate in the Transformer encoding and the subsequent weighted global pooling, preventing padded frames from influencing the learned feature distribution. In the LLD branch, all low-level descriptors are processed through an MLP followed by Batch Normalization, which stabilizes the feature distribution during training and reduces scale variation across multilingual datasets. Furthermore, all datasets undergo identical STFT processing, Mel-filterbank projection, and logarithmic compression, ensuring consistent feature characteristics across corpora recorded under different languages and acoustic conditions.
To ensure stable training across different datasets, the hyperparameters were tuned using an automated grid search in the early stage of experimentation. The search focused on the main training-related hyperparameters, including the learning rate, batch size, dropout ratio, and number of training epochs (with early stopping). A compact discrete grid was defined based on commonly used ranges in speech emotion recognition, where the learning rate candidates were {0.01, 0.005, 0.001}, the batch size candidates were {16, 32, 64}, and the dropout candidates were {0.1, 0.3, 0.5}. Each configuration in the grid was automatically trained and evaluated on the validation split using weighted accuracy (WA) as the primary criterion.The configuration consisting of a learning rate of 0.001, a batch size of 32, and a dropout ratio of 0.3 achieved consistently strong validation performance and was therefore adopted as the final hyperparameter setting for all experiments. The model was trained for up to 50 epochs with early stopping (patience = 25), using label-smoothing-enhanced cross-entropy loss. All model training and testing were conducted using Python 3.7 on a computer equipped with Windows 10 and an NVIDIA GeForce GTX 1050 Ti GPU.

4.3. Evaluation Metrics

In this study, to evaluate the performance of the proposed speech emotion recognition model, two evaluation metrics are adopted: weighted accuracy (WA) and unweighted accuracy (UA).
Weighted accuracy (WA) measures the overall classification accuracy of the model across the entire test set, defined as the ratio of correctly classified samples to the total number of samples, as shown in Equation (22). WA reflects the overall classification performance of the model and is particularly suitable for datasets with imbalanced sample distributions:
WA = T P + T N T P + T N + F P + F N .
Unweighted accuracy (UA) is calculated by first computing the recall for each emotion category and then averaging the recall values across all categories, as defined in Equation (23). Recall is defined as the ratio of correctly predicted samples to the actual number of samples in each category. UA effectively evaluates the model’s balanced classification ability across different emotion categories and mitigates the bias that may occur in datasets with imbalanced class distributions:
UA = 1 C i = 1 C T P i T P i + F N i .

4.4. Comparative Analysis

To further demonstrate the advantages of each component in the proposed Mel-spectrogram branch model, a series of comparative experiments were conducted. The baseline model selected for comparison is CNN-BLSTM [47], which has been widely validated as a robust and efficient architecture in the field of speech processing. The CNN component can effectively capture local resonance patterns and spectral variation information at a lower level, while the BLSTM component comprehensively models emotional trends, intonation fluctuations, and temporal dependencies at a higher level within speech segments.
The experimental setup is as follows:
(a)
A clean CNN-BLSTM baseline model was trained and evaluated under the local experimental environment and parameter settings to analyze its emotion recognition performance;
(b)
The baseline CNN-BLSTM model was modified by replacing the fixed-length slicing of the audio source with the dynamic sequence modeling approach proposed in this study;
(c)
Based on the dynamic sequence modeling version of CNN-BLSTM, the CNN module was replaced with the proposed MAG module, forming the MAG-BLSTM model;
(d)
Based on the dynamic sequence modeling CNN-BLSTM model, the BLSTM module was replaced with the Transformer encoder, resulting in the CNN-Transformer model.
All the above experiments were conducted on the IEMOCAP and EMO-DB datasets, and their performances were compared with that of the final model proposed in this paper. The experimental results are summarized in Table 1 and Table 2. In the tables, models with the prefix “D” indicate that the dynamic sequence modeling method was adopted.
From the above experimental results (a–b), it can be observed that compared to the baseline CNN-BLSTM model, the dynamic sequence modeling method proposed in this study significantly improves emotion recognition performance. On the IEMOCAP dataset, WA increased from 53.64% to 62.79%, and UA improved from 61.04% to 63.82%. On the EMO-DB dataset, WA improved from 68.52% to 70.37%, and UA increased from 63.71% to 70.62%.
From the experimental results (b–d), it can be further observed that each module of the proposed model provides performance gains compared to the traditional CNN-BLSTM model. After replacing the CNN module with the proposed MAG module, WA on IEMOCAP improved from 62.79% to 69.83%, and UA from 63.82% to 67.57%. On EMO-DB, WA improved from 70.37% to 75.93%, and UA from 70.62% to 76.56%. Replacing BLSTM with the Transformer encoder further enhanced performance. Finally, the proposed MAG-Transformer achieved the best WA and UA, verifying the rationality and superiority of the model design.

4.5. Performance Analysis

To analyze the performance of our model on different emotions across the three datasets which is shown in Table 3, we conducted a statistical analysis of various emotion recognition metrics and plotted the corresponding confusion matrices.
Based on Table 3 and Figure 4, Figure 5 and Figure 6, it can be observed that on the CASIA dataset, the model achieves recognition accuracies as high as 98% for happiness, neutral, and sadness emotions, demonstrating overall stable performance. However, there is some confusion between anger and surprise, mainly due to the similarity in emotional polarity between these two categories.
On the IEMOCAP dataset, neutral emotion achieves the best recognition performance, while there are some misclassifications between anger and happiness, indicating that the model still has room for improvement in distinguishing high-arousal emotions under natural English conversational settings.
For the EMO-DB dataset, despite its smaller sample size, the model still achieves high recognition rates across most emotional categories, demonstrating good generalization capability. Overall, the model exhibits good stability and adaptability across the three datasets, showing particular strength in recognizing neutral and low-arousal emotions, and also performing well on small-sample datasets such as EMO-DB and CASIA.
We also collected recent studies in the field of speech emotion recognition for comparative analysis. As shown in Table 4, the recognition results are categorized based on different datasets.
It can be observed that the proposed MAGTF-Net achieves superior recognition performance on the IEMOCAP dataset compared with previous approaches. EMO-DB is a German speech emotion dataset and CASIA is a Chinese emotional dataset; both are relatively small in size. Our model maintains competitive performance across these multilingual datasets, confirming its robustness and generalization capability.
Finally, to further analyze the contributions of each component, ablation experiments were conducted on the EMO-DB dataset. (1) To verify the importance of the LLD feature branch, the branch was removed to evaluate its complementary effect. (2) To evaluate the graph-based local modeling, the GraphBlock layer within the MAG module was also removed for comparison.
As shown in Table 5, removing the LLD branch (-LLD) resulted in a decrease of 18.4% in UA and 14.3% in WA, indicating that low-level acoustic features play an important role in modeling emotional information. Similarly, removing the GraphBlock module (-GraphBlock) led to a 9.3% decrease in UA and a 6.9% decrease in WA, demonstrating that the graph structure positively contributes to capturing the internal relationships within Mel features.
In addition, to further validate the effectiveness of the proposed multi-scale design, we performed an extended kernel-size ablation study on the EMO-DB corpus using both single-kernel and reduced multi-kernel configurations. As shown in Table 6, certain single-scale settings—for example, k = 5 and k = 11 —achieve relatively high UA scores (80.58% and 80.23%), reflecting the fact that EMO-DB is a highly controlled studio-recorded dataset in which emotional cues are dominated by a small number of characteristic temporal scales.
However, the full multi-scale configuration { 3 , 5 , 7 , 9 , 11 } obtains the highest weighted accuracy (WA = 83.33%) among all tested variants, indicating that aggregating multiple temporal receptive fields provides more stable and balanced modeling of emotional expressions. Since WA is the most reliable metric under class-imbalance conditions, this result demonstrates that the full MAG configuration remains the most robust design on EMO-DB. Compared with reduced multi-scale subsets such as { 5 , 7 , 9 } or { 7 , 9 , 11 } , the full configuration preserves complementary short-, mid-, and long-range acoustic dynamics, thereby offering stronger generalization capability.
These results confirm that although individual temporal scales may achieve strong performance on certain controlled datasets, combining multiple temporal resolutions ensures stable behavior across diverse affective speech conditions.

5. Discussion and Conclusions

The proposed MAGTF-Net integrates multi-scale acoustic analysis, graph-based relational modeling, and complementary handcrafted descriptors to provide a unified framework for robust speech emotion recognition. By jointly capturing fine-grained local fluctuations and broader temporal–prosodic structures, MAGTF-Net achieves competitive or superior performance across English (IEMOCAP), German (EMO-DB), and Chinese (CASIA) emotional speech datasets. The fusion of Mel-based dynamic features and LLD-based global descriptors further enhances generalization under diverse speakers and recording conditions.
Despite these encouraging results, variations in recognition performance across emotional categories indicate that cross-linguistic and cross-cultural factors remain an intrinsic challenge. English and German primarily express affective cues through stress, intensity, and rhythm, whereas Mandarin Chinese encodes lexical meaning through pitch contours, which may overlap with or obscure emotional intonation. These differences partly explain why emotions such as happiness, surprise, or confusion exhibit inconsistent acoustic manifestations across languages. Future research into domain adaptation and cross-cultural representation learning may help mitigate such disparities.
Although the structural components of MAGTF-Net, such as multi-scale convolution, graph attention, and handcrafted LLD descriptors, provide inherent interpretability, post hoc explainability techniques also play an important role in analyzing model behavior. Applying methods such as SHAP or LIME to MAGTF-Net, however, is non-trivial because the model operates on variable-length acoustic sequences, incorporates dynamic padding and masking, and contains multiple interacting branches. These characteristics violate the fixed-dimensional input assumptions underlying most post hoc interpretability frameworks. Extending SHAP-based or attribution-based techniques to sequence-level emotional modeling therefore represents a promising direction for future work and may further enhance model transparency.
From a deployment perspective, the combination of multi-scale convolution, graph reasoning, and Transformer encoding introduces non-trivial computational overhead, which may limit real-time applicability in conversational assistants or embedded affective systems. While this study focuses on offline evaluation, exploring lightweight variants—such as reducing kernel-scale redundancy, compressing graph embeddings via low-rank or sparse formulations, pruning or narrowing the Transformer encoder, or adopting knowledge distillation—represents a promising direction for improving efficiency without sacrificing performance.
Overall, MAGTF-Net offers a flexible foundation for modeling emotional cues across languages and recording conditions. Continued progress on cross-cultural generalization, explainability extensions, and lightweight model design will further extend the applicability of the framework and deepen the scientific understanding of emotional expression in diverse linguistic and computational environments.

Author Contributions

Conceptualization, S.Z. and Y.X.; methodology, S.Z.; software, S.Z.; validation, S.Z.; formal analysis, S.Z.; investigation, S.Z.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z. and Y.X.; supervision, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Beijing Natural Science Foundation (Grant No. 4192023) and the Gold-Bridge Funds for Beijing (Grant No. 2021bjjq0017).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are publicly available at: IEMOCAP—https://sail.usc.edu/iemocap/ (accessed on 12 October 2025), EMO-DB—https://emodb.bilderbar.info/ (accessed on 12 October 2025), and CASIA—https://link.gitcode.com/i/b1a764ad5209d4d6a08667524ef00c43?uuid_tt_dd=10_19016666450-1746733703175-714152&isLogin=9&from_id=143074127 (accessed on 12 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ren, F. Affective information processing and recognizing human emotion. Electron. Notes Theor. Comput. Sci. 2009, 225, 39–50. [Google Scholar] [CrossRef]
  2. Huang, F.; Li, X.; Yuan, C.; Zhang, S.; Zhang, J.; Qiao, S. Attention-emotion-enhanced convolutional LSTM for sentiment analysis. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 4332–4345. [Google Scholar] [CrossRef]
  3. Darwin, C. The Expression of the Emotions in Man and Animals; John Murray: London, UK, 1872. [Google Scholar]
  4. Ekman, P.; Friesen, W.V. Constants across cultures in the face and emotion. J. Pers. Soc. Psychol. 1971, 17, 124. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, K.; Wei, J.; Zou, J.; Wang, P.; Yang, Y.; Shen, H.T. Improving Pre-trained Model-based Speech Emotion Recognition from a Low-level Speech Feature Perspective. IEEE Trans. Multimed. 2024, 26, 10623–10636. [Google Scholar] [CrossRef]
  6. Li, J.; Wang, X.; Lv, G.; Zeng, Z. GraphCFC: A Directed Graph Based Cross-Modal Feature Complementation Approach for Multimodal Conversational Emotion Recognition. IEEE Trans. Multimed. 2023, 26, 77–89. [Google Scholar] [CrossRef]
  7. Li, M.; Zhang, W.; Hu, B.; Kang, J.; Wang, Y.; Lu, S. Automatic assessment of depression and anxiety through encoding pupil-wave from HCI in VR scenes. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 20, 42. [Google Scholar] [CrossRef]
  8. Schuller, B.; Rigoll, G.; Lang, M. Hidden Markov model-based speech emotion recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Hong Kong, China, 6–10 April 2003; Volume 2, p. II-1. [Google Scholar]
  9. Wang, Y.; Guan, L. Recognizing Human Emotional State from Audiovisual Signals. IEEE Trans. Multimed. 2008, 10, 659–668. [Google Scholar] [CrossRef]
  10. Zeng, Z.; Tu, J.; Pianfetti, B.; Huang, T.S. Audio–Visual Affective Expression Recognition Through Multistream Fused HMM. IEEE Trans. Multimed. 2008, 10, 570–577. [Google Scholar] [CrossRef]
  11. Zhang, S.; Zhang, S.; Huang, T.; Gao, W. Speech Emotion Recognition Using Deep Convolutional Neural Network and Discriminant Temporal Pyramid Matching. IEEE Trans. Multimed. 2018, 20, 1576–1590. [Google Scholar] [CrossRef]
  12. Mustafa, M.B.; Yusoof, M.A.M.; Don, Z.M.; Malekzadeh, M. Speech emotion recognition research: An analysis of research focus. Int. J. Speech Technol. 2018, 21, 137–156. [Google Scholar] [CrossRef]
  13. Schuller, B.; Batliner, A.; Steidl, S.; Seppi, D. Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge. Speech Commun. 2011, 53, 1062–1087. [Google Scholar] [CrossRef]
  14. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  15. Ahmed, M.R.; Islam, S.; Islam, A.M.; Shatabda, S. An ensemble 1D-CNN-LSTM-GRU model with data augmentation for speech emotion recognition. Expert Syst. Appl. 2023, 218, 119633. [Google Scholar] [CrossRef]
  16. Mai, S.; Sun, Y.; Xiong, A.; Zeng, Y.; Hu, H. Multimodal Boosting: Addressing Noisy Modalities and Identifying Modality Contribution. IEEE Trans. Multimed. 2024, 26, 3018–3033. [Google Scholar] [CrossRef]
  17. Zou, H.; Si, Y.; Chen, C.; Rajan, D.; Chng, E.S. Speech emotion recognition with co-attention based multi-level acoustic information. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 7367–7371. [Google Scholar]
  18. Xu, C.; Liu, Y.; Song, W.; Liang, Z.; Chen, X. A New Network Structure for Speech Emotion Recognition Research. Sensors 2024, 24, 1429. [Google Scholar] [CrossRef] [PubMed]
  19. Ahn, Y.; Han, S.; Lee, S.; Shin, J.W. Speech Emotion Recognition Incorporating Relative Difficulty and Labeling Reliability. Sensors 2024, 24, 4111. [Google Scholar] [CrossRef]
  20. Sun, L.; Fu, S.; Wang, F. Decision tree SVM model with Fisher feature selection for speech emotion recognition. EURASIP J. Audio Speech Music Process. 2019, 2019, 1–14. [Google Scholar] [CrossRef]
  21. Al Dujaili, M.J.; Ebrahimi-Moghadam, A.; Fatlawi, A. Speech emotion recognition based on SVM and KNN classifications fusion. Int. J. Electr. Comput. Eng. 2021, 11, 1259. [Google Scholar] [CrossRef]
  22. Luengo, I.; Navas, E.; Hernáez, I. Feature Analysis and Evaluation for Automatic Emotion Identification in Speech. IEEE Trans. Multimed. 2010, 12, 490–501. [Google Scholar] [CrossRef]
  23. Fahad, M.S.; Ranjan, A.; Yadav, J.; Deepak, A. A survey of speech emotion recognition in natural environment. Digit. Signal Process. 2021, 110, 102951. [Google Scholar] [CrossRef]
  24. Mehrish, A.; Majumder, N.; Bharadwaj, R.; Mihalcea, R.; Poria, S. A review of deep learning techniques for speech processing. Inf. Fusion 2023, 99, 101869. [Google Scholar] [CrossRef]
  25. Sivanagaraja, T.; Ho, M.K.; Khong, A.W.H.; Wang, Y. End-to-End Speech Emotion Recognition Using Multi-Scale Convolution Networks. In Proceedings of the Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Kuala Lumpur, Malaysia, 12–15 December 2017; pp. 189–192. [Google Scholar]
  26. Latif, S.; Rana, R.; Khalifa, S.; Jurdak, R.; Qadir, J.; Schuller, B. Survey of deep representation learning for speech emotion recognition. IEEE Trans. Affect. Comput. 2021, 14, 1634–1654. [Google Scholar] [CrossRef]
  27. Guo, L.; Wang, L.; Xu, C.; Dang, J.; Chng, E.S.; Li, H. Representation learning with spectro-temporal-channel attention for speech emotion recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 6304–6308. [Google Scholar]
  28. Ma, H.; Wang, J.; Lin, H.; Zhang, B.; Zhang, Y.; Xu, B. A Transformer-Based Model With Self-Distillation for Multimodal Emotion Recognition in Conversations. IEEE Trans. Multimed. 2023, 26, 776–788. [Google Scholar] [CrossRef]
  29. Sun, T.; Chen, H.; Hu, G.; He, L.; Zhao, C. Explainability of Speech Recognition Transformers via Gradient-Based Attention Visualization. IEEE Trans. Multimed. 2023, 26, 1395–1406. [Google Scholar] [CrossRef]
  30. Liu, Y.; Sun, H.; Guan, W.; Xia, Y.; Zhao, Z. Multi-modal speech emotion recognition using self-attention mechanism and multi-scale fusion framework. Speech Commun. 2022, 139, 1–9. [Google Scholar] [CrossRef]
  31. Guo, L.; Ding, S.; Wang, L.; Dang, J. Dstcnet: Deep spectro-temporal-channel attention network for speech emotion recognition. IEEE Trans. Neural Netw. Learn. Syst. 2023, 36, 188–197. [Google Scholar] [CrossRef] [PubMed]
  32. Liu, Z.; Kang, X.; Ren, F. Dual-tbnet: Improving the robustness of speech features via dual-transformer-bilstm for speech emotion recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 2193–2203. [Google Scholar] [CrossRef]
  33. Li, H.; Li, J.; Liu, H.; Liu, T.; Chen, Q.; You, X. MelTrans: Mel-Spectrogram Relationship-Learning for Speech Emotion Recognition via Transformers. Sensors 2024, 24, 5506. [Google Scholar] [CrossRef]
  34. Li, F.; Zhang, D. Transformer-Driven Affective State Recognition from Wearable Physiological Data in Everyday Contexts. Sensors 2025, 25, 761. [Google Scholar] [CrossRef]
  35. Wang, H.; Kim, D.H. Graph Neural Network-Based Speech Emotion Recognition: A Fusion of Skip Graph Convolutional Networks and Graph Attention Networks. Electronics 2024, 13, 4208. [Google Scholar] [CrossRef]
  36. Chandola, D.; Altarawneh, E.; Jenkin, M.; Papagelis, M. SERC-GCN: Speech emotion recognition in conversation using graph convolutional networks. In Proceedings of the ICASSP 2024—IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 76–80. [Google Scholar]
  37. Zhao, S.; Yu, B.; Yang, K.; Zhang, S.; Hu, J.; Jiang, Y.; Yu, P.S.; Chen, H. A Flexible Diffusion Convolution for Graph Neural Networks. IEEE Trans. Knowl. Data Eng. 2025, 37, 3118–3131. [Google Scholar] [CrossRef]
  38. Ma, L.; Sheng, Z.; Li, X.; Gao, X.; Hao, Z.; Yang, L.; Nie, X.; Jiang, J.; Zhang, W.; Cui, B. Acceleration algorithms in GNNs: A survey. IEEE Trans. Knowl. Data Eng. 2025, 37, 3173–3192. [Google Scholar] [CrossRef]
  39. Liu, Y.; Rong, Y.; Guo, Z.; Chen, N.; Xu, T.; Tsung, F.; Li, J. Human Mobility Modeling during the COVID-19 Pandemic via Deep Graph Diffusion Infomax. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  40. Hang, C.N.; Yu, P.D.; Chen, S.; Tan, C.W.; Chen, G. MEGA: Machine Learning-Enhanced Graph Analytics for Infodemic Risk Management. IEEE J. Biomed. Health Inform. 2023, 27, 6100–6111. [Google Scholar] [CrossRef] [PubMed]
  41. Yoon, S.; Byun, S.; Jung, K. Multimodal speech emotion recognition using audio and text. In Proceedings of the IEEE Spoken Language Technology Workshop (SLT), Athens, Greece, 18–21 December 2018; pp. 112–118. [Google Scholar]
  42. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. M3ER: Multiplicative multimodal emotion recognition using facial, textual, and speech cues. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20), New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1359–1367. [Google Scholar]
  43. Khan, M.; Gueaieb, W.; El Saddik, A.; Kwon, S. MSER: Multimodal speech emotion recognition using cross-attention with deep fusion. Expert Syst. Appl. 2024, 245, 122946. [Google Scholar] [CrossRef]
  44. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.F.; Weiss, B. A database of German emotional speech. In Proceedings of the INTERSPEECH 2005—Eurospeech, 9th European Conference on Speech Communication and Technology, Lisbon, Portugal, 4–8 September 2005; Volume 5, pp. 1517–1520. [Google Scholar]
  45. Busso, C.; Bulut, M.; Lee, C.C.; Kazemzadeh, A.; Mower, E.; Kim, S.; Chang, J.N.; Lee, S.; Narayanan, S.S. IEMOCAP: Interactive emotional dyadic motion capture database. Lang. Resour. Eval. 2008, 42, 335–359. [Google Scholar] [CrossRef]
  46. Tao, J.; Liu, F.; Zhang, M.; Jia, H. Design of Speech Corpus for Mandarin Text-to-Speech. In Proceedings of the Blizzard Challenge 2008 Workshop, Brisbane, Australia, 21 September 2008. [Google Scholar]
  47. Satt, A.; Rozenberg, S.; Hoory, R. Efficient emotion recognition from speech using deep learning on spectrograms. In Proceedings of the INTERSPEECH 2017—18th Annual Conference of the International Speech Communication Association, Stockholm, Sweden, 20–24 August 2017; pp. 1089–1093. [Google Scholar] [CrossRef]
  48. Wang, Z.Q.; Tashev, I. Learning utterance-level representations for speech emotion and age/gender recognition using deep neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 5150–5154. [Google Scholar]
  49. Chen, M.; He, X.; Yang, J.; Zhang, H. 3-D convolutional recurrent neural networks with attention model for speech emotion recognition. IEEE Signal Process. Lett. 2018, 25, 1440–1444. [Google Scholar] [CrossRef]
  50. Cheng, S.; Zhang, D.; Yin, D. A DenseNet-GRU technology for Chinese speech emotion recognition. In Proceedings of the International Conference on Frontiers of Electronics, Information and Computation Technologies, Changsha, China, 21–23 May 2021; pp. 1–7. [Google Scholar]
  51. Sun, L.; Li, Q.; Fu, S.; Li, P. Speech emotion recognition based on genetic algorithm–decision tree fusion of deep and acoustic features. ETRI J. 2022, 44, 462–475. [Google Scholar] [CrossRef]
  52. Zhu, R.; Sun, C.; Wei, X.; Zhao, L. Speech emotion recognition using channel attention mechanism. In Proceedings of the 4th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 7–9 April 2023; pp. 680–684. [Google Scholar]
  53. Chen, L.W.; Rudnicky, A. Exploring wav2vec 2.0 fine tuning for improved speech emotion recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
Figure 1. The overall framework of MAGTF-Net.
Figure 1. The overall framework of MAGTF-Net.
Sensors 25 07378 g001
Figure 2. Architecture of the Mel-spectrogram feature branch.
Figure 2. Architecture of the Mel-spectrogram feature branch.
Sensors 25 07378 g002
Figure 3. Architecture of the LLD feature branch.
Figure 3. Architecture of the LLD feature branch.
Sensors 25 07378 g003
Figure 4. Confusion matrix of the model on the IEMOCAP dataset.
Figure 4. Confusion matrix of the model on the IEMOCAP dataset.
Sensors 25 07378 g004
Figure 5. Confusion matrix of the model on the EMO-DB dataset.
Figure 5. Confusion matrix of the model on the EMO-DB dataset.
Sensors 25 07378 g005
Figure 6. Confusion matrix of the model on the CASIA dataset.
Figure 6. Confusion matrix of the model on the CASIA dataset.
Sensors 25 07378 g006
Table 1. Performance comparison on the IEMOCAP dataset.
Table 1. Performance comparison on the IEMOCAP dataset.
ModelsWA (%)UA (%)
(a) CNN-BLSTM53.6461.04
(b) D_CNN-BLSTM62.7963.82
(c) D_MAG-BLSTM69.8367.57
(d) D_CNN-Transformer66.3864.56
(e) MAG-Transformer69.1570.86
Table 2. Performance comparison on the EMO-DB dataset.
Table 2. Performance comparison on the EMO-DB dataset.
ModelsWA (%)UA (%)
(a) CNN-BLSTM68.5263.71
(b) D_CNN-BLSTM70.3770.62
(c) D_MAG-BLSTM75.9376.56
(d) D_CNN-Transformer77.7874.71
(e) MAG-Transformer83.3379.11
Table 3. Detailed Performance of Our Model on Three Datasets.
Table 3. Detailed Performance of Our Model on Three Datasets.
IEMOCAPCASIAEMO-DB
P (%)R (%)F1 (%)P (%)R (%)F1 (%)P (%)R (%)F1 (%)
Angry66.6781.4873.3371.4393.7581.0885.7192.3188.89
Sad83.7859.6269.6667.8690.4877.55100.00100.00100.00
Neutral72.6568.5570.5496.1596.1596.1562.5071.4366.67
Happy59.2669.5764.00100.0066.6780.0075.0050.0060.00
Disgust100.00100.00100.00
Fear75.0047.3758.0676.92100.0086.96
Surprise85.7190.0087.80
Bored100.0040.0057.14
Table 4. Comparative results of speech emotion recognition on different datasets.
Table 4. Comparative results of speech emotion recognition on different datasets.
ModelsCASIAIEMOCAPEMO-DB
DNN-ELM [48]0.5800.809
3DRNN+Attention [49]0.6470.828
Dual-TBNet [32]0.9570.6480.841
DenseNet-GRU [50]0.8000.6310.820
SVM+Decision Tree [51]0.8530.858
CNN+ChannelAttention [52]0.8870.846
Wav2Vec2.0 [53]0.6310.820
Ours0.8120.7000.812
Table 5. Ablation study results on the EMO-DB dataset.
Table 5. Ablation study results on the EMO-DB dataset.
ModelsUAWA
-LLD0.6490.648
-GraphBlock0.7400.722
FusionNet0.8330.791
Table 6. Kernel-size ablation study on EMO-DB.
Table 6. Kernel-size ablation study on EMO-DB.
Kernel SetUA (%)WA (%)
{ 3 } 70.8166.67
{ 5 } 80.5879.63
{ 7 } 71.3768.52
{ 9 } 66.0261.11
{ 11 } 80.2377.78
{ 3 , 5 , 7 } 66.1668.52
{ 5 , 7 , 9 } 77.5275.93
{ 7 , 9 , 11 } 80.2778.33
{ 3 , 5 , 7 , 9 , 11 } (Full-MAG)79.1183.33
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, S.; Xie, Y.; Wang, Z. MAGTF-Net: Dynamic Speech Emotion Recognition with Multi-Scale Graph Attention and LLD Feature Fusion. Sensors 2025, 25, 7378. https://doi.org/10.3390/s25237378

AMA Style

Zhu S, Xie Y, Wang Z. MAGTF-Net: Dynamic Speech Emotion Recognition with Multi-Scale Graph Attention and LLD Feature Fusion. Sensors. 2025; 25(23):7378. https://doi.org/10.3390/s25237378

Chicago/Turabian Style

Zhu, Shiyin, Yinggang Xie, and Zhiliang Wang. 2025. "MAGTF-Net: Dynamic Speech Emotion Recognition with Multi-Scale Graph Attention and LLD Feature Fusion" Sensors 25, no. 23: 7378. https://doi.org/10.3390/s25237378

APA Style

Zhu, S., Xie, Y., & Wang, Z. (2025). MAGTF-Net: Dynamic Speech Emotion Recognition with Multi-Scale Graph Attention and LLD Feature Fusion. Sensors, 25(23), 7378. https://doi.org/10.3390/s25237378

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop