Next Article in Journal
Evaluation of Commercial Camera-Based Solutions for Tracking Hand Kinematics
Previous Article in Journal
Innovative Technologies to Improve Occupational Safety in Mining and Construction Industries—Part II
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition

1
Faculty of Applied Sciences, Macao Polytechnic University, Macao 999078, China
2
Guangxi Key Laboratory of Machine Vision and Intelligent Control, Wuzhou University, Wuzhou 543000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5718; https://doi.org/10.3390/s25185718 (registering DOI)
Submission received: 15 July 2025 / Revised: 22 August 2025 / Accepted: 2 September 2025 / Published: 13 September 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

A critical challenge in EEG-based emotion recognition is the poor generalization of models across different datasets due to significant domain shifts. Traditional methods struggle because they either overfit to source-domain characteristics or fail to bridge large discrepancies between datasets. To address this, we propose the Cross-corpus Attention-based Transfer Enhancement network (CATE), a novel two-stage framework. The core novelty of CATE lies in its dual-view self-supervised pre-training strategy, which learns robust, domain-invariant representations by approaching the problem from two complementary perspectives. Unlike single-view models that capture an incomplete picture, our framework synergistically combines: (1) Noise-Enhanced Representation Modeling (NERM), which builds resilience to domain-specific artifacts and noise, and (2) Wavelet Transform Representation Modeling (WTRM), which captures the essential, multi-scale spectral patterns fundamental to emotion. This dual approach moves beyond the brittle assumptions of traditional domain adaptation, which often fails when domains are too dissimilar. In the second stage, a supervised fine-tuning process adapts these powerful features for classification using attention-based mechanisms. Extensive experiments on six transfer tasks across the SEED, SEED-IV, and SEED-V datasets demonstrate that CATE establishes a new state-of-the-art, achieving accuracies from 68.01% to 81.65% and outperforming prior methods by up to 15.65 percentage points. By effectively learning transferable features from these distinct, synergistic views, CATE provides a robust framework that significantly advances the practical applicability of cross-corpus EEG emotion recognition.

1. Introduction

Emotion recognition has become a cornerstone of affective computing and human–computer interaction, with profound implications for mental health monitoring, adaptive user interfaces, and clinical diagnosis [1,2]. Among various physiological signals, electroencephalography (EEG) offers a direct window into the brain’s emotional processing. Its high temporal resolution and objective nature make it an ideal modality for capturing genuine emotional states, as EEG signals reflect neural activity that is difficult to consciously manipulate [3]. Advances in deep learning have enabled models to achieve impressive accuracies, often exceeding 95%, on single, standardized datasets like DEAP and SJTU Emotion EEG Dataset (SEED) [4,5].
However, this success in controlled settings masks a critical challenge that severely impedes real-world application: the poor generalization of models across different datasets, a problem known as the cross-corpus generalization challenge [6,7]. When a model trained on one dataset is tested on another, its performance often plummets by 20–30% [8,9]. This performance degradation arises from significant “domain shifts” caused by a combination of factors: inter-subject physiological variability (e.g., brain anatomy, skull thickness) and inter-dataset discrepancies in experimental protocols, emotion elicitation stimuli, recording equipment, and ambient noise [3,10]. This gap between laboratory performance and practical applicability remains the primary hurdle for deploying robust EEG-based emotion recognition systems.
To bridge this domain gap, researchers have explored a variety of computational strategies. Early attempts relied on conventional supervised learning, which failed to generalize due to overfitting to source-specific features. Subsequently, domain adaptation techniques [11], such as Domain Adversarial Neural Networks, were introduced to align feature distributions between source and target domains [5]. However, these methods often presuppose a closer proximity between domain distributions than typically exists in diverse EEG datasets, limiting their effectiveness. More recent efforts have incorporated advanced deep learning architectures. Attention mechanisms, particularly multi-head self-attention from Transformers, have shown promise in adaptively focusing on the most informative EEG channels and temporal segments [4,9]. Concurrently, self-supervised and contrastive learning paradigms have emerged as a powerful means to learn subject-invariant representations from unlabeled data, reducing the costly dependency on manual annotation [3,12]. These methods learn robust features by leveraging the inherent structure of EEG signals, for instance, by predicting temporal context or contrasting different augmentations of the same sample.
Despite these advancements, existing methods still exhibit significant limitations when faced with the complexities of cross-corpus EEG data. Each approach, while addressing a piece of the puzzle, fails to provide a comprehensive solution. The primary shortcomings of current state-of-the-art approaches are systematically summarized in Table 1.
Traditional domain adaptation methods like DANN assume that source and target distributions can be aligned through adversarial training. However, this assumption breaks down in cross-corpus EEG scenarios due to: (1) hardware variations—different EEG systems use varying sampling rates, electrode materials, and amplifier characteristics, creating systematic differences in signal properties; (2) protocol differences—SEED uses movie clips while other datasets may use images or music, resulting in fundamentally different neural response patterns; (3) subject heterogeneity—age, cultural background, and individual brain anatomy create irreducible inter-subject variability. When these differences accumulate, forcing alignment between such disparate distributions can cause negative transfer, where the model learns spurious correlations that harm performance. To systematically address these multifaceted challenges identified in Table 1, our proposed CATE framework introduces targeted solutions: (1) against overfitting in supervised learning, CATE employs two-stage training with self-supervised pre-training to learn generalizable features rather than memorizing source-specific patterns; (2) against domain adaptation’s distribution assumptions, CATE’s dual-view strategy learns inherently robust representations without forced alignment, where NERM models domain variations as noise while WTRM extracts domain-invariant multi-scale patterns; (3) against single-view limitations, CATE’s dual-view architecture simultaneously captures complementary information with NERM providing frequency-domain robustness and WTRM capturing temporal-frequency dynamics through wavelet decomposition; and (4) against augmentation dependency, CATE introduces EEG-specific strategies with NERM’s adaptive noise injection and WTRM’s wavelet-based augmentation that preserve emotional content while ensuring robust feature learning. This comprehensive design enables CATE to effectively overcome the limitations of existing approaches.
Several recent architectures have attempted to address these limitations through more sophisticated self-supervised and contrastive learning approaches. Generic contrastive methods adapted for EEG, including SimCLR [13] and Mixup variants [14], employ augmentation strategies unsuited for EEG’s non-stationary nature and operate on single-view representations, missing complementary multi-domain information. GMSS [15] applies SimCLR-based contrastive learning with meiosis augmentation but focuses solely on temporal features, neglecting crucial spatial channel interactions. MV-SSTMA [16] employs masked autoencoding across multiple domains but shows significant performance degradation in self-supervised settings. JCFA [17] uses dual-view contrastive learning from time-frequency domains and incorporates graph convolution during fine-tuning to model channel relationships; however, it suffers from prohibitively long training times due to the computational complexity of graph operations and lacks explicit mechanisms for handling domain-specific noise while assuming similar cross-domain distributions—an assumption that often fails in practice.
These limitations motivate our CATE framework, which introduces a fundamentally different approach to domain adaptation. Unlike existing methods that attempt to force alignment between disparate domains or rely on computationally expensive graph structures, CATE addresses domain shift through two synergistic mechanisms: (1) Noise-Enhanced Representation Modeling (NERM) actively learns to filter out domain-specific artifacts by training on adaptively noise-perturbed data, forcing the model to focus on robust, emotion-relevant patterns rather than superficial recording characteristics; and (2) Wavelet Transform Representation Modeling (WTRM) captures domain-invariant differential entropy patterns across frequency bands that persist across different recording conditions and protocols. This dual-view architecture, combined with efficient attention-based feature fusion instead of graph convolution, enables CATE to achieve superior cross-corpus performance with significantly reduced computational overhead compared to existing approaches. To address these multifaceted challenges, this paper introduces a novel Cross-corpus Attention-based Transfer Enhancement network (CATE). Our approach is built on a comprehensive two-stage framework that synergistically combines self-supervised pre-training with supervised fine-tuning. The core of CATE is a dual-view pre-training strategy that learns complementary representations from two perspectives: (1) NERM, which builds robustness to domain-specific noise and artifacts, and (2) Wavelet Transform Representation Modeling (WTRM), which captures essential frequency-domain differential entropy characteristics of EEG signals. During fine-tuning, attention mechanisms fuse these complementary features to perform robust emotion classification. A multi-component loss function, integrating alignment, contrastive, and style diversity objectives, ensures that the learned features are consistent, discriminative, and generalizable.
This work makes several key contributions to cross-corpus EEG-based emotion recognition:
  • We propose a novel dual-view self-supervised pre-training strategy that comprehensively captures complementary signal characteristics—noise robustness and multi-scale frequency patterns—via the NERM and WTRM views, respectively.
  • We design a multi-component optimization objective that effectively balances feature consistency, discriminability, and diversity, providing a solid foundation for cross-corpus generalization.
  • We develop an attention-based transfer enhancement mechanism that dynamically fuses multi-scale features while maintaining computational efficiency.
  • Through extensive experiments on the SEED series datasets, we demonstrate that CATE achieves state-of-the-art performance on six cross-corpus tasks, with accuracy improvements of 5.42% to 15.65% over existing methods, thereby setting a new benchmark for robust, practical EEG-based emotion recognition.
  • We pioneer the comprehensive evaluation of cross-corpus emotion recognition across multiple SEED datasets (SEED, SEED-IV, and SEED-V), particularly introducing SEED-V with its challenging five-class emotion taxonomy into cross-corpus research for the first time. Through extensive experimentation comparing multiple state-of-the-art models on these diverse emotion classification tasks (3-class, 4-class, and 5-class), we establish new benchmarks for multi-class cross-corpus EEG emotion recognition and demonstrate the scalability of our approach to increasingly complex emotion taxonomies.

2. The Proposed CATE Architecture

2.1. Overall Architecture

The overall architecture of CATE is shown in Figure 1. It consists of two stages: (1) a dual-path attention-driven self-supervised pre-training stage based on frequency band perturbation view and wavelet modulation view, and (2) a supervised fine-tuning stage.
In the pre-training stage, the model learns robust representations from two complementary views through a shared feature extractor consisting of linear layers with LayerNorm and GELU activation. This simplified architecture focuses on learning domain-invariant features through contrastive learning across the NERM and WTRM views.
In the fine-tuning stage, the outputs of the pre-trained dual-path attention encoders are fused through attention-weighted feature integration and then connected to a task-specific classifier, enabling the model to effectively focus on key emotion-related patterns in EEG signals and achieve more robust cross-corpus emotion recognition.
The pre-training architecture creates two distinct augmented views of EEG data through specialized enhancement modules, followed by direct feature extraction for robust representation learning. The framework generates two complementary views from the original EEG differential entropy features to capture diverse signal characteristics.

2.2. The Dual-Path Self-Supervised Pre-Training Stage

(1) the shared feature extractor ensures both views learn in the same feature space; (2) the alignment loss ensures cross-view consistency by minimizing the distance between representations of the same sample across views; (3) the style diversity loss prevents representation collapse and ensures the two views learn complementary rather than redundant features.

2.2.1. Noise-Enhanced Representation Modeling (NERM) View

Given the input EEG differential entropy features X R B × C × F , where B denotes the batch size, C represents the number of channels, and F = 5 indicates the five frequency bands (delta, theta, alpha, beta, gamma), the NERM view introduces a simple yet effective adaptive noise mechanism. This process enhances the robustness of the learned representations against domain-specific variations.
The core of this view is to add Gaussian noise whose magnitude is dynamically scaled based on the input features. The procedure is as follows:
First, an input scale is computed based on the mean absolute value of the features:
s input = mean ( | X | )
Next, this input scale is used to determine the noise scale, which is capped to prevent excessive perturbation:
s noise = min ( 0.05 , 0.1 · s input + ϵ )
where ϵ is a small constant (e.g., 1 × 10 8 ) to ensure numerical stability.
Finally, the noise-enhanced features X NERM are generated by adding scaled Gaussian noise to the original features:
X NERM = X + N ( 0 , I ) · s noise
where N ( 0 , I ) represents standard Gaussian noise.
This adaptive noise injection strategy forces the model to learn features that are less sensitive to minor variations and artifacts present in the input. By dynamically adjusting the noise level relative to the input’s magnitude, it ensures that the perturbation is meaningful without corrupting the underlying signal structure. This process effectively regularizes the model, discouraging it from overfitting to spurious, domain-specific characteristics and thereby improving its generalization capability for cross-corpus emotion recognition tasks.

2.2.2. Wavelet Transform Representation Modeling (WTRM) View

The WTRM view employs wavelet decomposition to capture multi-scale spectral energy distribution patterns across frequency bands:
For each sample x i R C × F where C = 62 channels and F = 5 frequency bands, the wavelet decomposition is applied to the spectral energy profile of each channel:
{ c A ( c ) , { c D j ( c ) } j = 1 L } = DWT ( x i ( c ) , ψ , L ) , c = 1 , , C
where x i ( c ) = [ D E d e l t a , D E t h e t a , D E a l p h a , D E b e t a , D E g a m m a ] ( c ) represents the differential entropy features across five frequency bands for channel c, c A ( c ) captures the overall spectral energy trend, { c D j ( c ) } captures the inter-band energy variations, ψ is the wavelet basis function (Haar or db1 due to signal length constraint), and L = 1 is the decomposition level (limited by F = 5 ).
Enhanced coefficients are generated through adaptive masking and perturbation:
c ˜ A ( c ) = c A ( c ) ( 1 + μ A · M A · ϵ A )
c ˜ D j ( c ) = c D j ( c ) ( 0.5 · M s u p p r e s s + M e n h a n c e · ( 1 + ϵ D j · σ D j ) )
where M A , M s u p p r e s s , M e n h a n c e are random binary masks with probabilities μ × 0.5 , μ , and μ × 0.5 , respectively, μ = 0.05 is the mask probability, ϵ A , ϵ D j are enhancement factors, and ⊙ denotes element-wise multiplication.
The WTRM view is reconstructed via an inverse wavelet transform:
X W T R M ( c ) = IDWT ( { c ˜ A ( c ) , { c ˜ D j ( c ) } j = 1 L } , ψ )
This decomposition captures the spectral energy distribution patterns that are characteristic of different emotional states, where the approximation coefficients represent the overall brain activation level and the detail coefficients capture the coordination between adjacent frequency bands, providing complementary information to the NERM view for robust emotion recognition.

2.2.3. Shared Feature Extraction

Both augmented views are processed through a shared feature extractor that directly maps the input to the latent representation space:
H b a s e = LayerNorm ( GELU ( W 1 X + b 1 ) ) W 2 + b 2
where W 1 R d m o d e l × d i n p u t and W 2 R d m o d e l × d m o d e l are learnable weight matrices, and b 1 , b 2 are bias vectors.
The features from both views are directly obtained through the shared extractor:
Z N E R M = H b a s e ( X N E R M )
Z W T R M = H b a s e ( X W T R M )
where Z N E R M , Z W T R M R B × d m o d e l represent the extracted features from NERM and WTRM views, respectively.

2.2.4. Multi-Component Loss Function

The self-supervised pre-training optimizes three complementary loss components operating on the directly extracted features:
(1)
Alignment Loss: This component ensures cross-view consistency by aligning representations from the same sample across different views in the feature space:
L a l i g n = 2 2 · 1 B i = 1 B z N E R M ( i ) · z W T R M ( i ) z N E R M ( i ) 2 z W T R M ( i ) 2
where z N E R M ( i ) and z W T R M ( i ) represent L2-normalized features from NERM and WTRM views, respectively. This loss function promotes semantic consistency between complementary views while preserving view-specific characteristics.
(2)
Style Diversity Loss: This component prevents representation collapse by enforcing orthogonality between different samples within the same view:
L s t y l e = G N E R M I F 2 + G W T R M I F 2
G N E R M = Z N E R M Z N E R M T , G W T R M = Z W T R M Z W T R M T
I is the identity matrix, and · F denotes the Frobenius norm. By constraining the Gram matrices to approximate the identity matrix, this loss effectively maintains feature diversity and prevents dimensional collapse during training.
(3)
Total Pre-training Loss: The complete pre-training objective combines the alignment and diversity components with appropriate weighting:
L t o t a l = λ a l i g n L a l i g n + λ s t y l e L s t y l e
where λ a l i g n and λ s t y l e are balancing coefficients that control the relative importance of alignment and diversity objectives. The combination of these two loss components forms the complete optimization objective for our self-supervised pre-training stage.
Through this multi-component loss framework, NERM and WTRM views form a synergistic interaction during pre-training: NERM focuses on learning representations robust to domain-specific noise and artifacts, while WTRM captures multi-scale spectral energy distribution patterns across frequency bands through wavelet decomposition. The alignment loss ensures both views produce consistent high-level semantic understanding of the same emotional state, while the style loss maintains feature diversity, preventing the model from simply learning identical representations. This dual-view approach enables the model to capture both noise-robust features and emotion-relevant spectral characteristics simultaneously.

2.3. Supervised Fine-Tuning Stage

During the supervised fine-tuning stage, the pre-trained Shared Feature Extractor serves as the backbone. Its learned weights are used to initialize two parallel encoders: the NERM View Encoder and the WTRM View Encoder. Each encoder is then augmented with its own multi-head attention mechanism, allowing the model to adapt the powerful, pre-trained representations for the specific downstream classification task.
The model input consists of standard Differential Entropy (DE) features, extracted from EEG signals that have been downsampled to 200 Hz and band-pass filtered (1–75 Hz). To adapt these sample-based DE features for our Transformer-based architecture, we treat each vector as a sequence of length one ( L = 1 ). An input batch of shape [ B , d input ] is reshaped to [ B , 1 , d input ] and processed through the initialized encoders. As illustrated in Figure 2, the refined features from the parallel NERM and WTRM encoders are then fused to generate a joint representation for classification. This strategy allows the model to effectively leverage the powerful pre-trained architecture with non-sequential input data.
This strategy allows the model to effectively leverage the powerful pre-trained architecture, even with input data that is not inherently sequential. The proposed dual-view encoder architecture is illustrated in Figure 2. This architecture employs two parallel encoders, NERM and WTRM, to process EEG feature representations from different domains, ultimately generating a joint representation through feature fusion for emotion classification.

2.3.1. Multi-Head Attention Mechanism

The core multi-head attention mechanism operates on the pre-trained features through learned query, key, and value projections:
Q = H W Q , K = H W K , V = H W V
where H R B × L × d model (with L = 1 in our case) represents the input feature sequence, and W Q , W K , W V R d model × d model are learnable projection matrices.
The scaled dot-product attention is computed as:
Attention ( Q , K , V ) = softmax Q K T d k V
where d k = d model / h and h denotes the number of attention heads.
Multi-head attention concatenates outputs from multiple attention heads:
MultiHead ( Q , K , V ) = Concat ( head 1 , , head h ) W O
where head i = Attention ( Q W Q i , K W K i , V W V i ) and W O R d model × d model is the output projection matrix.

2.3.2. NERM View Encoder

The NERM view adaptive encoder processes noise-enhanced representations through dual attention mechanisms to capture frequency-aware patterns:
H n e r m ( 1 ) = H n e r m + α · MultiHead n e r m ( H n e r m )
H n e r m ( 1 ) = LayerNorm ( H n e r m ( 1 ) )
The final NERM representation is obtained through global average pooling:
z N E R M f i n e = LayerNorm ( GlobalAvgPool ( H n e r m ( 1 ) ) )

2.3.3. WTRM View Encoder

The WTRM view encoder employs dual attention mechanisms to capture spectral energy distribution patterns from wavelet-enhanced representations.
H w t r m ( 1 ) = H w t r m + α · MultiHead w t r m ( H w t r m )
where α = 0.1 is a scaling factor for residual connections to ensure training stability.
The combined representation is normalized:
H w t r m ( 1 ) = LayerNorm ( H w t r m ( 1 ) )
The representation is further refined through a feed-forward network with scaled residual connection:
H f f = H w t r m ( 1 ) + α · FFN ( H w t r m ( 1 ) )
where FFN consists of:
FFN ( x ) = Dropout ( W 2 · GELU ( Dropout ( W 1 · x ) ) )
with W 1 R d m o d e l × 4 d m o d e l and W 2 R 4 d m o d e l × d m o d e l .
After layer normalization:
H n o r m = LayerNorm ( H f f )
The final WTRM representation for fine-tuning is obtained through adaptive pooling:
z W T R M f i n e = LayerNorm ( GlobalAvgPool ( H n o r m T ) )

2.3.4. Cross-Domain Feature Fusion

The enhanced features from the NERM and WTRM views, z N E R M f i n e and z W T R M f i n e , are combined to form a final representation. While complex fusion strategies like attention-weighting are possible, we opted for simple averaging:
z f u s e d = z N E R M f i n e + z W T R M f i n e 2
This choice was based on a combination of empirical evidence, stability, and simplicity. As detailed in Appendix B, we conducted experiments on the SEED- IV 4 SEED- V 4 task to evaluate different fixed weighting ratios (see Table A5).

2.3.5. Feature Fusion and Classification

Classification is performed by a multi-layer perceptron (MLP) head that takes the fused features z fused as input. The MLP projects the features into a hidden representation, applies a GELU activation and dropout for regularization, and finally produces the output logits for classification.
h cls = GELU ( LayerNorm ( W 1 z fused + b 1 ) )
h cls = Dropout ( h cls )
logits = W 2 h cls + b 2
where W 1 R d hidden × d model and W 2 R d output × d hidden are learnable weight matrices.

2.3.6. Fine-Tuning Objective

The fine-tuning stage adapts the model to the downstream classification task by optimizing a supervised objective. The loss function, therefore, consists of a standard cross-entropy classification loss:
L fine - tune = L cls
The classification loss is defined as the cross-entropy between the predicted logits and the one-hot encoded true labels:
L cls = CrossEntropy ( logits , y true )

2.3.7. Training Strategy

The fine-tuning process employs several standard optimization strategies to ensure stable and efficient training:
(1)
Mixed Precision Training: We employ automatic mixed precision (AMP) with gradient scaling to accelerate the fine-tuning process while maintaining numerical stability.
(2)
Gradient Clipping: To prevent the issue of exploding gradients, the norms of the gradients are clipped to a maximum threshold of 1.0, calculated as:
g clipped = min 1.0 , 1.0 g 2 · g
(3)
Learning Rate Scheduling: A ReduceLROnPlateau scheduler is utilized with a factor of 0.5 and a patience of 5 epochs. This scheduler monitors the total loss and adaptively adjusts the learning rate when improvements stagnate.
This fine-tuning strategy allows the model to effectively adapt the learned representations for the emotion classification task. The combination of a direct classification objective with robust optimization techniques ensures that the model is efficiently tailored to the downstream task.

3. Experimental Setting

3.1. Datasets

The SEED series of datasets is an emotional electroencephalogram (EEG) database established by the BCMI Laboratory at Shanghai Jiao Tong University. This series includes three main datasets, each targeting classification for a different number of emotional categories.
The SEED dataset (original version) focuses on three-category emotion recognition (negative, neutral, positive). It involved 15 subjects participating in 3 experimental sessions, each watching 15 emotion-inducing film clips. EEG signals were recorded using a 62-channel ESI Neuroscan system, and various EEG features such as differential entropy (DE) and differential asymmetry (DASM) were extracted.
The SEED-IV dataset expands to four-category classification (neutral, sad, fear, and happy emotions). It also involved 15 subjects participating in 3 experimental sessions, but each session included 24 film clips.
The SEED-V dataset further expands to five-category classification (disgust, fear, sad, neutral, happy). Each subject participated in 3 experimental sessions, watching 15 film clips per session, with both EEG and eye-tracking data recorded.

3.2. Pre-Processing

After collecting the raw EEG data, signal preprocessing and feature extraction performed. To increase the signal-to-noise ratio (SNR), the raw EEG signals are first downsampled to a 200 Hz sampling rate, followed by a band-pass filter from 1 Hz to 75 Hz . Subsequently, feature extraction is carried out. Differential Entropy (DE), as referenced in the literature [18], has the ability to distinguish patterns in different frequency bands. Therefore, we chose DE features as the input data for our model.
For one subject in one session in these three databases, a set of data is given in the form of channels (62) × trials (15 for SEED, 24 for SEED-IV, 15 for SEED-V) × frequency bands (5). We merge the channels and frequency bands, changing the form to trials × 310 ( 62 × 5 ).
Finally, all data are organized into 3394 × 310 (SEED dataset,3394 samples per subject per session), or 851 / 832 / 822 × 310 (SEED-IV dataset,851/832/822 samples per subject per session, respectively), or 681 / 541 / 601 × 310 (SEED-V dataset,681/541/ 601 samples per subject per session, respectively). The corresponding label vectors are generated in the form of 3394 × 1 , or 851 / 832 / 822 × 1 , or 681 / 541 / 601 × 1 .

3.3. Evaluation Metrics

Accuracy is used as a measure to evaluate the EEG classification, which is calculated as follows:
acc = A Num
where A refers to the number of samples correctly classified by the algorithm, and Num refers to the total number of samples to be classified.

4. Experiments

4.1. Experimental Setup

4.1.1. Datasets and Baselines

We evaluated our proposed model, CATE, on three publicly available EEG emotion recognition datasets: SEED [19], SEED-IV [20], and SEED-V [21]. We conducted a comprehensive comparison against five state-of-the-art (SOTA) methods: DANN [22], GECNN [23], E2STN [24], JCFA [17], and Mixup [14]. For fair comparison, all baselines were implemented using their officially reported hyperparameters.

4.1.2. Evaluation Protocol

We followed a strict subject-independent, cross-corpus protocol with supervised target-domain fine-tuning applied consistently across all methods. Importantly, all baseline models and our proposed CATE method were evaluated under the identical protocol, where target domain labels were used during fine-tuning. The experiments were conducted across six transfer tasks between the datasets. We use accuracy (ACC) and its standard deviation across all subjects in the test set as the primary evaluation metric. The specific trial splits for fine-tuning and testing in each task were as follows:
  • SEED ↔ SEED- IV 3 : For SEED, 9 trials were used for fine-tuning and 6 for testing. For SEED-IV (3-class), 12 trials were used for fine-tuning and 6 for testing.
  • SEED ↔ SEED- V 3 : For SEED, 9 trials for fine-tuning, 6 for testing. For SEED-V (3-class subset), 6 trials for fine-tuning, 3 for testing.
  • SEED- IV 4 ↔ SEED- V 4 : For SEED-IV (4-class), 16 trials for fine-tuning, 8 for testing. For SEED-V (4-class subset), 8 trials for fine-tuning, 4 for testing.

4.1.3. Implementation Details

Models were implemented in Python 3.8 using PyTorch 1.9.0 and trained on an NVIDIA Tesla A10 GPU. The self-supervised pre-training was conducted for 200 epochs with a batch size of 256. The subsequent fine-tuning was performed for 20 epochs with a batch size of 128.

4.2. Cross-Corpus Performance Analysis

Table 2 presents the performance comparison across all six cross-corpus tasks. Our proposed CATE model consistently and significantly outperforms all baseline methods, establishing a new state-of-the-art. Notably, CATE demonstrates not only higher mean accuracy but also lower standard deviation in most cases, indicating superior performance and stability.
For instance, in the SEED → SEED- IV 3 and SEED- IV 3 → SEED tasks, CATE achieves accuracies of 80.72% and 80.70%, surpassing the second-best method, JCFA, by a large margin of 15.20% and 14.37%, respectively. This substantial improvement highlights the effectiveness of our dual-view pre-training strategy. The Noise-Enhanced Representation Modeling (NERM) view learns to discard domain-specific noise stemming from different acquisition hardware, while the Wavelet Transform Representation Modeling (WTRM) view captures domain-invariant, multi-scale temporal-frequency patterns crucial for emotion recognition. The synergy between these views enables the model to learn robust and generalizable EEG representations.
The performance gains are also evident in the more challenging tasks involving SEED-V, which features a more complex emotion taxonomy. In the SEED- V 4 → SEED- IV 4 task, CATE achieves a 13.41% improvement over JCFA. This scenario, involving transfer between two distinct 4-class paradigms, underscores CATE’s ability to handle complex domain shifts and partially overlapping emotion categories.
Further analysis of the confusion matrices (Figure Figure 3, Figure 4 and Figure 5) reveals CATE’s discriminative power. For example, in the SEED → SEED- IV 3 task (Figure a), CATE significantly reduces the confusion between ‘positive’ and ‘neutral’ emotions compared to JCFA (Figure b). Similarly, in the challenging SEED- IV 4 → SEED- V 4 experiment (Figure 5c), CATE improves the recognition of ‘fear’ and reduces its confusion with ‘sad’, suggesting its capability to distinguish between nuanced emotional states. This confirms that our model learns more discriminative features, rather than just overfitting to the source domain’s characteristics.

4.3. Ablation Study

To validate the contributions of the core components of CATE, we conducted an extensive ablation study across all six tasks. We designed three variants: (1) w/o NERM, which removes the noise-enhanced view; (2) w/o WTRM, which removes the wavelet-based view; and (3) w/o pretrain, which trains the full model from scratch without self-supervised pre-training.
The results are summarized in Table 3. The full CATE model achieves the best performance across all settings, demonstrating the efficacy of its integrated design.
  • Removing the NERM component (‘w/o NERM’) causes the most significant performance degradation across all tasks (e.g., a 7.06% drop in the SEED → SEED- IV 3 task). This confirms that modeling and mitigating domain-specific noise is critical for cross-corpus generalization.
  • Removing the WTRM component (‘w/o WTRM’) also leads to a notable performance decline, highlighting the importance of capturing multi-scale temporal-frequency features, which serve as robust, emotion-related biomarkers.
  • Training the model from scratch (‘w/o pretrain’) results in inferior performance compared to the full model. This validates that the dual-view self-supervised pre-training provides a powerful initialization, enabling the model to learn more transferable representations before fine-tuning on limited labeled data.
Table 3. Ablation study results: average accuracy and standard deviation (%) of different model configurations across all cross-dataset tasks.
Table 3. Ablation study results: average accuracy and standard deviation (%) of different model configurations across all cross-dataset tasks.
MethodSEED → SEED- IV 3 SEED- IV 3 → SEEDSEED → SEED- V 3 SEED- V 3 → SEEDSEED- IV 4 → SEED- V 4 SEED- V 4 → SEED- IV 4
w/o NERM73.66 ± 18.8474.50 ± 17.8352.36 ± 23.5279.85 ± 12.6066.53 ± 20.7670.51 ± 16.47
w/o WTRM73.82 ± 19.1378.99 ± 18.3250.37 ± 20.4874.61 ± 14.2364.44 ± 15.5366.62 ± 14.83
w/o pretrain77.52 ± 18.6976.25 ± 16.4762.36 ± 19.8279.05 ± 10.2362.40 ± 19.5171.86 ± 15.94
Full (CATE)80.72 ± 14.2780.70 ± 10.0968.01 ± 19.3381.65 ± 09.3569.48 ± 16.7775.17 ± 14.01
The results demonstrate the contribution of each key component in the CATE model. NERM: Noise-Enhanced Representation Modeling; WTRM: Wavelet Transform Representation Modeling.
In summary, the ablation study empirically proves that the superior performance of CATE stems from the powerful synergy between its noise-robust modeling (NERM), multi-scale feature extraction (WTRM), and the effective feature initialization provided by the self-supervised pre-training paradigm.

5. Discussion

The empirical results robustly demonstrate that CATE sets a new state-of-the-art in cross-corpus EEG emotion recognition. Our discussion focuses on interpreting these performance gains by linking them to our model’s core architectural innovations, contextualizing its advantages over existing methods, and outlining the broader implications of this work.

5.1. Interpretation of CATE’s Performance

The consistent and substantial outperformance of CATE across all six transfer tasks validates our central hypothesis: a dual-view, self-supervised pre-training framework can effectively learn domain-invariant representations from heterogeneous EEG data. The most significant accuracy improvements (over 14 percentage points against the next-best method in SEED ↔ SEED-IV3 tasks) are not merely incremental gains but signify a fundamental advantage in generalization capability.
This advantage stems directly from the synergistic interplay of our two complementary views. The Noise-Enhanced Representation Modeling (NERM) view forces the model to become robust against superficial, domain-specific signal characteristics—akin to learning to recognize a voice despite different background noises or microphone qualities. This is crucial for cross-corpus tasks where recording equipment and environments inevitably vary. Concurrently, the Wavelet Transform Representation Modeling (WTRM) view captures the intrinsic, multi-scale temporal-frequency dynamics that are foundational to emotional expression in EEG signals. The ablation study (Table 3) provides clear evidence for this synergy: removing either view results in a significant performance drop, with the removal of NERM being particularly detrimental, confirming that tackling domain noise is a primary challenge.
Furthermore, the multi-head attention mechanism in the fine-tuning stage proves essential for adapting these robustly pre-trained features to specific classification tasks. As seen in the confusion matrices (e.g., Figure 5), CATE excels at distinguishing between nuanced and often-confused emotional states like ‘sad’ and ‘fear’. This suggests that the attention mechanism effectively learns to focus on the subtle yet discriminative neural signatures captured by the dual-view encoders, a capability that simpler feature extractors may lack.

5.2. Visualization Analysis of Feature Representations

To gain deeper insights into how each component of CATE contributes to effective emotion representation learning, we conducted t-SNE visualization analysis of the learned features across different model configurations. We randomly selected 200 samples from the SEED dataset and employed t-SNE to project the high-dimensional feature representations into a two-dimensional space. Figure 6 presents the t-SNE projections of feature representations from four model variants, providing intuitive evidence of each component’s contribution to discriminative feature learning.
From the perspective of an ablation study, the visualization results reveal several critical insights:
First, regarding the relative advantages of the full CATE model: Among the four configurations, the complete model demonstrates the relatively best feature organization structure. While overlaps between the three emotion categories still exist (which is common in EEG-based emotion recognition), compared to other configurations, the complete model exhibits more regular feature point distribution with relatively clearer inter-class boundaries. This indicates that the synergistic effect of the dual-view architecture and pre-training strategy indeed contributes to learning more discriminative feature representations.
Second, regarding the contribution of the NERM view encoder: When the NERM component is removed, feature distribution significantly deteriorates, manifesting as increased class overlap, particularly between neutral and negative emotions. This distribution degradation validates our quantitative findings that NERM view’s noise-enhanced modeling is crucial for extracting robust features that can distinguish subtle emotional differences in the presence of domain-specific noise. The absence of this component leads to the model’s difficulty in discriminating fine-grained emotional variations.
Third, regarding the effectiveness of the WTRM view encoder: Removing the WTRM component also compromises feature discriminability, though to a slightly lesser extent than removing NERM. The clusters show increased dispersion and partial overlap, particularly in the boundary regions between emotions. This suggests that wavelet-based multi-scale temporal-frequency features provide complementary information essential for capturing the full spectrum of emotion-related neural dynamics.
Most significantly, regarding the importance of the pre-training strategy: Compared to the full model, the model without pre-training exhibits the poorest feature organization, with all three emotion categories heavily intermingled. Despite achieving reasonable classification accuracy through supervised fine-tuning alone, the lack of self-supervised pre-training results in features that fail to capture the underlying structure of emotional states. This powerfully demonstrates that our dual-view pre-training strategy learns fundamental representations that supervised learning alone cannot achieve.
Core findings from the ablation study: The comprehensive ablation analysis reveals distinct contributions from each component across different transfer scenarios:
(1) Component-wise Analysis: Removing either the NERM or WTRM view consistently degrades performance, but their relative importance varies across tasks. NERM shows more significant impact in challenging cross-corpus scenarios (e.g., 7.06% drop in SEED → SEED- IV 3 ), while WTRM’s contribution is more pronounced in tasks involving complex emotion taxonomies (e.g., notable degradation in SEED-V related transfers).
(2) Pre-training vs. Architecture: The comparison between component removal and pre-training ablation reveals an interesting pattern. While removing individual components causes consistent performance drops, the impact of removing pre-training varies significantly across tasks. In some cases (e.g., SEED → SEED- IV 3 ), the architectural components (NERM/WTRM) contribute more than pre-training, while in others (e.g., SEED → SEED- V 3 ), pre-training shows greater importance. This suggests that the value of self-supervised pre-training is task-dependent and complements rather than replaces good architectural design.
(3) Synergistic Integration: The full model consistently outperforms all ablated versions, confirming that the optimal performance emerges from the synergistic combination of dual-view architecture and self-supervised pre-training, rather than from any single component dominance.
The t-SNE analysis provides compelling visual evidence supporting our architectural design choices. The synergistic combination of NERM and WTRM views, enhanced through self-supervised pre-training, creates a feature space where emotions are naturally organized along meaningful dimensions. This inherent structure in the learned representations explains CATE’s superior cross-corpus generalization capability—by learning features that capture the essential characteristics of emotional states rather than dataset-specific patterns, the model can effectively transfer knowledge across different experimental paradigms and recording conditions.

5.3. Comparison with Previous Cross-Corpus Methods

5.3.1. Comparison with Traditional Domain Adaptation Methods

DANN (2016), as a classical domain adversarial method, attempts to align feature distributions between source and target domains through adversarial training. However, our experimental results show that DANN performs poorly across all tasks (e.g., only 45.20% on SEED → SEED- IV 3 ), which is 35.52 percentage points lower than CATE. This substantial gap exposes DANN’s fundamental limitation: its assumption that source and target distributions can be directly matched is brittle for highly disparate EEG domains. In contrast, CATE does not force explicit alignment but rather learns core representations that are inherently robust to domain shifts, proving to be a more effective strategy.
GECNN (2021) introduced graph-embedded convolution to model spatial relationships between EEG channels. While achieving 57.25% accuracy on SEED → SEED- IV 3 , an improvement over DANN, it still falls 23.47 percentage points below CATE.GECNN’s primary limitation is its reliance on fixed spatial patterns that may not generalize across different recording setups and protocols. Without explicit mechanisms for handling domain-specific artifacts or capturing multi-scale spectral dynamics, GECNN struggles with the inherent variability in cross-corpus scenarios. Our dual-view architecture addresses these gaps through complementary noise-resilient (NERM) and wavelet-based spectral (WTRM) representations, achieving superior domain adaptation.

5.3.2. Comparison with Recent Self-Supervised Methods

E2STN (2025) employs an emotion style transfer network, achieving 61.24% on SEED → SEED- IV 3 . Despite attempting to mitigate domain differences through style transfer, it still lags behind CATE by 19.48 percentage points. E2STN’s main issue is that its style transfer mechanism assumes emotional “style” can be simply separated from content, but emotional expression in EEG signals is often tightly coupled with multiple aspects of the signal.
JCFA (2024) represents the closest competitor to CATE, employing dual-view contrastive learning with graph convolution. However, even this state-of-the-art method only achieves 65.52% on SEED → SEED- IV 3 , 15.20 percentage points lower than CATE. More importantly, JCFA’s training time is prohibitively long due to the computational complexity of graph operations. Another critical limitation of JCFA is its lack of explicit noise handling mechanisms, while our NERM view is specifically designed to filter domain-specific artifacts.

5.4. Limitations and Future Work

Despite its strong performance, this study has limitations that open avenues for future research. First, our validation was conducted on the SEED family of datasets. While these are standard benchmarks, testing CATE’s generalization capabilities on a wider array of datasets with even greater diversity (e.g., DEAP, AMIGOS) is a necessary next step. Second, the current CATE model operates in an offline setting. For many real-world applications, such as real-time adaptive interfaces or clinical monitoring, online adaptation capabilities are required.
Future work will, therefore, focus on three key directions: (1) Model Lightweighting: We will explore model compression and knowledge distillation techniques to create a more efficient version of CATE suitable for deployment on edge devices. (2) Broader-Scale Validation: We will extend our evaluation to more diverse and larger-scale EEG datasets to further probe the limits of CATE’s generalization. (3) Online and Multimodal Extension: We plan to develop an online adaptation mechanism for CATE and investigate the fusion of EEG data with other modalities (e.g., eye-tracking, ECG) to build even more robust and comprehensive emotion recognition systems.

6. Conclusions

In this paper, we introduced CATE, a novel two-stage deep learning framework designed to overcome the critical challenge of cross-corpus generalization in EEG-based emotion recognition. By synergistically combining a dual-view self-supervised pre-training strategy with an attention-based fine-tuning mechanism, CATE effectively learns robust, transferable representations from complex and noisy EEG signals. The core innovation lies in its complementary Noise-Enhanced (NERM) and Wavelet Transform (WTRM) modeling views, which equip the model to handle domain-specific noise while capturing essential temporal-frequency patterns of emotional states.
Extensive experiments across six challenging transfer tasks using the SEED, SEED-IV, and SEED-V datasets demonstrate the definitive superiority of our approach. CATE consistently set a new state-of-the-art, outperforming existing methods by substantial margins of up to 15.65 percentage points. These results, supported by a thorough ablation study, validate the architectural choices and confirm the efficacy of each component within the CATE framework. By significantly bridging the gap between single-dataset performance and real-world applicability, this work represents a meaningful advancement towards deploying practical and reliable EEG-based affective computing systems. CATE provides a robust and promising paradigm for future research in cross-domain physiological signal analysis.

Author Contributions

Conceptualization, Z.L. and K.-Y.W.; methodology, Z.L. and K.-Y.W.; software, Z.L.; validation, Z.L., K.-Y.W. and C.-T.L.; formal analysis, Z.L.; investigation, Z.L. and C.-T.L.; resources, K.-Y.W.; data curation, Z.L.; writing—original draft preparation, Z.L.; writing—review and editing, Z.L., K.-Y.W. and C.-T.L.; visualization, Z.L.; supervision, K.-Y.W.; project administration, K.-Y.W.; funding acquisition, K.-Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Guangxi Colleges and Universities Key Laboratory of Intelligent Software, and in part by the Guangxi Science and Technology Program under Grant No.AD22080026.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

This study does not involve humans.

Data Availability Statement

The database used in this study is publicly available at the following websites, SEED Dataset: https://bcmi.sjtu.edu.cn/home/seed/seed.html (accessed on 1 March 2021); SEED-IV Dataset: https://bcmi.sjtu.edu.cn/home/seed/seed-iv.html (accessed on 1 March 2021); SEED-V Dataset: https://bcmi.sjtu.edu.cn/home/seed/seed-v.html (accessed on 1 March 2022).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CATECross-corpus Attention-based Transfer Enhancement network
EEGElectroencephalography
NERMNoise-Enhanced Representation Modeling
WTRMWavelet Transform Representation Modeling

Appendix A. Implementation Details and Hyperparameter Settings

Appendix A.1. Training Configuration

Table A1. Hyperparameter settings for the pre-training stage.
Table A1. Hyperparameter settings for the pre-training stage.
ParameterValueDescription
OptimizerAdamWAdaptive moment estimation with weight decay
Base learning rate 4 × 10 4 Initial learning rate for pre-training
Weight decay 3 × 10 4 L2 regularization coefficient
Training epochs200Total number of pre-training epochs
Batch size256Number of samples per batch
Gradient clipping threshold1.0Maximum gradient norm
Scheduler decay factor0.5Learning rate reduction ratio
Scheduler patience5Epochs to wait for improvement
Table A2. Hyperparameter settings for the fine-tuning stage.
Table A2. Hyperparameter settings for the fine-tuning stage.
ParameterValueDescription
OptimizerAdamWAdaptive moment estimation with weight decay
Base learning rate 2 × 10 4 Initial learning rate for fine-tuning
Weight decay 3 × 10 4 L2 regularization coefficient
Training epochs20Total number of fine-tuning epochs
Batch size128Number of samples per batch

Appendix A.2. Loss Function Weights

Table A3. Multi-component loss function weight configuration.
Table A3. Multi-component loss function weight configuration.
Loss ComponentPre-TrainingFine-TuningDescription
λ a l i g n (Alignment loss)0.50.0Cross-view consistency weight
λ s t y l e (Style diversity loss)0.50.0Feature diversity weight
λ c l s (Classification loss)-1.0Cross-entropy classification weight

Appendix A.3. Model Architecture Parameters

Table A4. CATE model architecture configuration.
Table A4. CATE model architecture configuration.
ParameterValueDescription
Input dimension d i n p u t 31062 channels × 5 frequency bands
Model dimension d m o d e l 200Feature representation dimension
Hidden dimension d h i d d e n 256MLP hidden layer size
Output dimension d o u t p u t 3/4/5 *Number of emotion classes
Number of attention heads h8Multi-head attention heads
Number of Transformer layers2Encoder layers
FFN expansion factor4Feed-forward network expansion ratio
Residual scaling factor α 0.1Scaling coefficient for training stability
* Adjusted according to different dataset tasks.

Appendix B. Feature Fusion Ablation Study

To determine the optimal method for fusing features from the NERM and WTRM views, we conducted an ablation study on the SEED- IV 4 SEED- V 4 transfer task, experimenting with various fixed weighting ratios. The results, presented in Table A5, show that performance peaks around a 0.6:0.4 to 0.5:0.5 ratio. Specifically, simple averaging (a 0.5:0.5 ratio) achieved performance (e.g., Accuracy: 69.48%, F1-Score: 65.31%) that was statistically indistinguishable from the best-performing, slightly unbalanced ratio (0.6:0.4). Given the negligible performance difference, we selected simple averaging as it is a parameter-free, more stable, and less complex approach. This choice avoids introducing additional hyperparameters that might require separate tuning for each cross-corpus task, thereby enhancing the model’s generalizability and robustness.
Table A5. Performance comparison of different feature fusion weighting ratios for the NERM and WTRM views on the SEED- IV 4 SEED- V 4 task. The best performance for each metric is highlighted in bold.
Table A5. Performance comparison of different feature fusion weighting ratios for the NERM and WTRM views on the SEED- IV 4 SEED- V 4 task. The best performance for each metric is highlighted in bold.
Ratio NERM:WTRMAccuracyPrecisionF1-ScoreAUROCAUPRC
0.9 : 0.10.6813 ± 0.19600.6643 ± 0.22400.6333 ± 0.22090.8636 ± 0.15350.7960 ± 0.1894
0.8 : 0.20.6830 ± 0.17610.6807 ± 0.21030.6337 ± 0.19930.8648 ± 0.15110.7952 ± 0.1881
0.7 : 0.30.6937 ± 0.16510.6821 ± 0.21150.6458 ± 0.18890.8684 ± 0.15020.8012 ± 0.1879
0.6 : 0.40.6948 ± 0.16170.6827 ± 0.20950.6496 ± 0.18590.8708 ± 0.14650.8044 ± 0.1849
0.5 : 0.50.6948 ± 0.16770.6946 ± 0.19420.6531 ± 0.19100.8699 ± 0.14540.8019 ± 0.1839
0.4 : 0.60.6913 ± 0.17200.6933 ± 0.19420.6508 ± 0.19560.8682 ± 0.14230.7972 ± 0.1814
0.3 : 0.70.6868 ± 0.16680.6879 ± 0.19080.6456 ± 0.19220.8656 ± 0.14090.7910 ± 0.1803
0.2 : 0.80.6792 ± 0.16580.6827 ± 0.19100.6357 ± 0.19240.8629 ± 0.13970.7856 ± 0.1799
0.1 : 0.90.6708 ± 0.16070.6752 ± 0.18910.6282 ± 0.18690.8611 ± 0.13590.7782 ± 0.1785

References

  1. Torres, E.P.; Torres, E.A.; Hernández-Álvarez, M.; Yoo, S.G. EEG-based BCI emotion recognition: A survey. Sensors 2020, 20, 5083. [Google Scholar] [CrossRef] [PubMed]
  2. Pei, G.; Li, H.; Lu, Y.; Wang, Y.; Hua, S.; Li, T. Affective computing: Recent advances, challenges, and future trends. Intell. Comput. 2024, 1, 76. [Google Scholar] [CrossRef]
  3. Li, X.; Zhang, Y.; Tiwari, P.; Song, D.; Hu, B.; Yang, M. EEG based emotion recognition: A tutorial and review. ACM Comput. Surv. 2022, 55, 1–36. [Google Scholar] [CrossRef]
  4. Wang, X.; Ren, Y.; Luo, Z.; He, W.; Hong, J. Deep learning-based EEG emotion recognition: Current trends and future perspectives. Front. Psychol. 2023, 14, 1126994. [Google Scholar] [CrossRef] [PubMed]
  5. Ma, W.; Zheng, Y.; Li, T.; Li, Z.; Li, Y.; Wang, L. A comprehensive review of deep learning in EEG-based emotion recognition: Classifications, trends, and practical implications. PeerJ Comput. Sci. 2024, 10, e2065. [Google Scholar] [CrossRef] [PubMed]
  6. Roshdy, A.; Karar, A.; Kork, S.A.; Beyrouthy, T.; Nait-Ali, A. Advancements in EEG emotion recognition: Leveraging multi-modal database integration. Appl. Sci. 2024, 14, 2487. [Google Scholar] [CrossRef]
  7. Chen, Y.; Peng, Y.; Tang, J.; Camilleri, T.A.; Camilleri, K.P.; Kong, W.; Cichocki, A. EEG-based affective brain–computer interfaces: Recent advancements and future challenges. J. Neural Eng. 2025, 22, 031004. [Google Scholar] [CrossRef] [PubMed]
  8. Hu, X.; Chen, J.; Wang, F.; Zhang, D. Ten challenges for EEG-based affective computing. Brain Sci. Adv. 2019, 5, 1–20. [Google Scholar] [CrossRef]
  9. Spezialetti, M.; Placidi, G.; Rossi, S. Emotion recognition for human-robot interaction: Recent advances and future perspectives. Front. Robot. AI 2020, 7, 532279. [Google Scholar] [CrossRef] [PubMed]
  10. Pal, S.; Mukhopadhyay, S.; Suryadevara, N. Development and progress in sensors and technologies for human emotion recognition. Sensors 2021, 21, 5554. [Google Scholar] [CrossRef] [PubMed]
  11. Liang, S.; Li, L.; Zu, W.; Feng, W.; Hang, W. Adaptive Deep Feature Representation Learning for Cross-Subject EEG Decoding. BMC Bioinform. 2024, 25, 393. [Google Scholar] [CrossRef] [PubMed]
  12. Rafiei, M.H.; Gauthier, L.V.; Adeli, H.; Takabi, D. Self-Supervised Learning for Electroencephalography. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 1457–1471. [Google Scholar] [CrossRef] [PubMed]
  13. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  14. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  15. Kan, H.; Yu, J.; Huang, J.; Liu, Z.; Wang, H.; Zhou, H. Self-supervised group meiosis contrastive learning for EEG-based emotion recognition. Appl. Intell. 2023, 53, 27207–27225. [Google Scholar] [CrossRef]
  16. Li, R.; Wang, Y.; Zheng, W.L.; Lu, B.L. A Multi-View Spectral-Spatial-Temporal Masked Autoencoder for Decoding Emotions with Self-Supervised Learning. In Proceedings of the 30th ACM International Conference on Multimedia (MM ’22), Lisboa, Portugal, 10–14 October 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 3273–3281. [Google Scholar] [CrossRef]
  17. Liu, Q.; Zhou, Z.; Wang, J.; Liang, Z. Joint Contrastive Learning with Feature Alignment for Cross-Corpus EEG-based Emotion Recognition. In Proceedings of the 1st International Workshop on Brain-Computer Interfaces (BCI) for Multimedia Understanding, BCIMM ’24, Melbourne, Australia, 28 October–1 November 2024; Association for Computing Machinery: New York, NY, USA, 2024. [Google Scholar]
  18. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar]
  19. Zheng, W.L.; Lu, B.L. Investigating Critical Frequency Bands and Channels for EEG-based Emotion Recognition with Deep Neural Networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  20. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. EmotionMeter: A Multimodal Framework for Recognizing Human Emotions. IEEE Trans. Cybern. 2019, 49, 1110–1122. [Google Scholar] [CrossRef] [PubMed]
  21. Liu, W.; Qiu, J.L.; Zheng, W.L.; Lu, B.L. Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 715–729. [Google Scholar] [CrossRef]
  22. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  23. Song, T.; Zheng, W.; Liu, S.; Zong, Y.; Cui, Z.; Li, Y. Graph-embedded convolutional neural network for image-based EEG emotion recognition. IEEE Trans. Emerg. Top. Comput. 2021, 10, 1399–1413. [Google Scholar] [CrossRef]
  24. Zhou, Y.; Li, F.; Li, Y.; Ji, Y.; Zhang, L.; Chen, Y.; Wang, H. Enhancing Cross-Dataset EEG Emotion Recognition: A Novel Approach with Emotional EEG Style Transfer Network. IEEE Trans. Affect. Comput. 2025, 1–15. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the proposed CATE model for cross-corpus EEG-based emotion recognition. (a) Dual-Path Self-Supervised Pre-training Stage: The input EEG data undergoes dual-view data augmentation through Noise-Enhanced Representation Modeling (NERM) and Wavelet Transform Representation Modeling (WTRM). Both views are processed through a shared feature extractor with linear layers, LayerNorm, and GELU activation. The extracted NERM and WTRM features are then optimized using alignment loss ( L a l i g n ), and style loss ( L s t y l e ) to learn robust domain-invariant representations. (b) Supervised fine-tuning stage: The pre-trained model is adapted for emotion classification. Input data is normalized and processed through parallel NERM and WTRM view encoders, each containing attention mechanisms and feed-forward networks (FFN). The dual-view features are fused, pooled globally, and fed to a classifier for final emotion prediction. The inset shows the detailed internal structure of the encoders with multi-head attention layers.
Figure 1. The overall architecture of the proposed CATE model for cross-corpus EEG-based emotion recognition. (a) Dual-Path Self-Supervised Pre-training Stage: The input EEG data undergoes dual-view data augmentation through Noise-Enhanced Representation Modeling (NERM) and Wavelet Transform Representation Modeling (WTRM). Both views are processed through a shared feature extractor with linear layers, LayerNorm, and GELU activation. The extracted NERM and WTRM features are then optimized using alignment loss ( L a l i g n ), and style loss ( L s t y l e ) to learn robust domain-invariant representations. (b) Supervised fine-tuning stage: The pre-trained model is adapted for emotion classification. Input data is normalized and processed through parallel NERM and WTRM view encoders, each containing attention mechanisms and feed-forward networks (FFN). The dual-view features are fused, pooled globally, and fed to a classifier for final emotion prediction. The inset shows the detailed internal structure of the encoders with multi-head attention layers.
Sensors 25 05718 g001
Figure 2. NERM-WTRM dual-view feature fusion architecture. Two complementary encoders process EEG representations from different perspectives: (1) The NERM encoder builds robustness against domain-specific artifacts through noise-enhanced processing and MultiHead n e r m attention; (2) The WTRM encoder extracts emotion-relevant spectral energy distributions through wavelet decomposition and MultiHead w t r m attention. The complementary features z N E R M f i n e and z W T R M f i n e are fused to form the robust joint representation z f u s e d for emotion classification.
Figure 2. NERM-WTRM dual-view feature fusion architecture. Two complementary encoders process EEG representations from different perspectives: (1) The NERM encoder builds robustness against domain-specific artifacts through noise-enhanced processing and MultiHead n e r m attention; (2) The WTRM encoder extracts emotion-relevant spectral energy distributions through wavelet decomposition and MultiHead w t r m attention. The complementary features z N E R M f i n e and z W T R M f i n e are fused to form the robust joint representation z f u s e d for emotion classification.
Sensors 25 05718 g002
Figure 3. Confusion matrices for CATE vs. JCFA on the SEED ↔ SEED-IV3 tasks.
Figure 3. Confusion matrices for CATE vs. JCFA on the SEED ↔ SEED-IV3 tasks.
Sensors 25 05718 g003aSensors 25 05718 g003b
Figure 4. Confusion matrices for CATE vs. JCFA on the SEED ↔ SEED-V3 tasks.
Figure 4. Confusion matrices for CATE vs. JCFA on the SEED ↔ SEED-V3 tasks.
Sensors 25 05718 g004
Figure 5. Confusion matrices for CATE vs. JCFA on the SEED- IV 4 ↔ SEED- V 4 tasks.
Figure 5. Confusion matrices for CATE vs. JCFA on the SEED- IV 4 ↔ SEED- V 4 tasks.
Sensors 25 05718 g005
Figure 6. t-SNE visualization of learned feature representations from different CATE model configurations on the SEED dataset. Each point represents a sample projected into 2D space, colored by emotion category (red: positive, green: neutral, blue: negative).
Figure 6. t-SNE visualization of learned feature representations from different CATE model configurations on the SEED dataset. Each point represents a sample projected into 2D space, colored by emotion category (red: positive, green: neutral, blue: negative).
Sensors 25 05718 g006
Table 1. Limitations in Existing Cross-Corpus EEG Emotion Recognition Methods.
Table 1. Limitations in Existing Cross-Corpus EEG Emotion Recognition Methods.
Approach CategoryKey IdeaLimitations in Cross-Corpus Context
Conventional Supervised LearningTrain a model on a large, labeled source dataset.Fails to generalize to new target domains due to overfitting; performance drops significantly.
Traditional Domain AdaptationAlign the feature distributions of source and target domains, often adversarially.Assumes distributions are close, which is often not the case for EEG; susceptible to negative transfer if domains are too dissimilar.
Attention-Based ModelsAdaptively weight EEG channels, time steps, or frequency bands.Often operate on a single view of the data, failing to capture the interplay between spatial, temporal, and spectral patterns.
Self-Supervised & Contrastive LearningLearn representations from unlabeled data by solving pretext tasks or contrasting augmented views.Learning is highly dependent on the choice of data augmentation; may not sufficiently account for domain-specific noise and style variations.
Table 2. Comparison of average accuracy and standard deviation (%) of different methods across the six cross-corpus emotion recognition tasks.
Table 2. Comparison of average accuracy and standard deviation (%) of different methods across the six cross-corpus emotion recognition tasks.
MethodSEED → SEED- IV 3 SEED- IV 3 → SEEDSEED → SEED- V 3 SEED- V 3 → SEEDSEED- IV 4 → SEED- V 4 SEED- V 4 → SEED- IV 4
DANN(2016) *45.20 ± 24.3162.89 ± 08.6746.53 ± 23.5262.27 ± 12.6036.91 ± 14.8933.50 ± 17.35
GECNN(2021)57.25 ± 07.5358.02 ± 07.03----
MIXUP(2018) *60.23 ± 15.5663.24 ± 16.4751.11 ± 20.6672.21 ± 14.2353.25 ± 20.3459.78 ± 18.36
E2STN(2025)61.24 ± 15.1460.51 ± 05.41----
JCFA(2024) *65.52 ± 15.9266.33 ± 11.1554.07 ± 23.2679.73 ± 12.3264.06 ± 23.2261.76 ± 15.23
CATE (ours)80.72 ± 14.2780.70 ± 10.0968.01 ± 19.3381.65 ± 09.3569.48 ± 16.7775.17 ± 14.01
* Results obtained by our own implementation. “-” indicates results were not available in the original papers.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Wong, K.-Y.; Lam, C.-T. Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition. Sensors 2025, 25, 5718. https://doi.org/10.3390/s25185718

AMA Style

Li Z, Wong K-Y, Lam C-T. Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition. Sensors. 2025; 25(18):5718. https://doi.org/10.3390/s25185718

Chicago/Turabian Style

Li, Zongni, Kin-Yeung Wong, and Chan-Tong Lam. 2025. "Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition" Sensors 25, no. 18: 5718. https://doi.org/10.3390/s25185718

APA Style

Li, Z., Wong, K.-Y., & Lam, C.-T. (2025). Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition. Sensors, 25(18), 5718. https://doi.org/10.3390/s25185718

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop