Next Article in Journal
Tree–Hillclimb Search: An Efficient and Interpretable Threat Assessment Method for Uncertain Battlefield Environments
Previous Article in Journal
Entropy Methods on Finding Optimal Linear Combinations with an Application to Biomarkers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Subject EEG Emotion Recognition Using SSA-EMS Algorithm for Feature Extraction

1
Normal College, Jimei University, Xiamen 361021, China
2
National Engineering Research Center for E-Learning, Central China Normal University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(9), 986; https://doi.org/10.3390/e27090986
Submission received: 6 August 2025 / Revised: 17 September 2025 / Accepted: 19 September 2025 / Published: 21 September 2025

Abstract

This study proposes a novel SSA-EMS framework that integrates Singular Spectrum Analysis (SSA) with Effect-Matched Spatial Filtering (EMS), combining the noise-reduction capability of SSA with the dynamic feature extraction advantages of EMS to optimize cross-subject EEG-based emotion feature extraction. Experiments were conducted using the SEED dataset under two evaluation paradigms: “cross-subject sample combination” and “subject-independent” assessment. Random Forest (RF) and SVM classifiers were employed to perform pairwise classification of three emotional states—positive, neutral, and negative. Results demonstrate that the SSA-EMS framework achieves RF classification accuracies exceeding 98% across the full frequency band, significantly outperforming single frequency bands. Notably, in the subject-independent evaluation, model accuracy remains above 96%, confirming the algorithm’s strong cross-subject generalization capability. Experimental results validate that the SSA-EMS framework effectively captures dynamic neural differences associated with emotions. Nevertheless, limitations in binary classification and the potential for multimodal extension remain important directions for future research.

1. Introduction

The electroencephalogram (EEG) holds significant value for emotion recognition due to their high temporal resolution and non-invasive nature [1,2]. However, their practical application faces numerous challenges stemming from the inherent biophysical properties of EEG signals and the complexity of emotions themselves [3,4]. EEG signals are characterized by inherent non-stationarity and a low signal-to-noise ratio, making them highly susceptible to interference from physiological artifacts [5,6]. Although many analytical methods rely on the assumption of stationarity for practicality, this property is particularly prominent during transient events like seizures or in ERP analyses [7]. The challenges become even more pronounced in cross-subject EEG emotion recognition tasks [8,9]. Beyond the intrinsic complexity of the signals, individual differences in physiology and psychology further exacerbate the difficulty of EEG-based emotion recognition [10,11]. As a result, EEG emotion recognition models often perform well in within-subject scenarios but demonstrate significantly reduced generalization capability in cross-subject settings [12,13].
To address this challenge, researchers have proposed various innovative methods to enhance the cross-subject adaptability of models by improving signal processing techniques or machine learning frameworks [14,15]. Anuragi A et al. proposed an automated cross-subject emotion recognition framework based on EEG signals, employing the Fourier-Bessel Series Expansion-based Empirical Wavelet Transform (FBSE-EWT) method. They demonstrated the effectiveness of the proposed framework using machine learning models such as Artificial Neural Networks (ANN) and k-Nearest Neighbors (k-NN), showing superior performance in human emotion classification compared to other state-of-the-art emotion classification models [16]. Wenhui Guo and Yaxuan Li et al. introduced a Functional Connectivity-Enhanced Feature Group Attention Network (FC-FAN) for cross-subject emotion recognition. Comprehensive evaluations indicate that this framework achieves outstanding performance in cross-subject emotion recognition tasks, effectively capturing the spatiotemporal features of EEG signals while clearly analyzing the nonlinear relationships among electrode signals [17]. Yong Peng and Honggang Liu et al. proposed a joint EEG feature transfer and semi-supervised cross-subject emotion recognition model, which achieves global optimality by simultaneously optimizing the shared subspace projection matrix and target labels. The joint learning paradigm of this model significantly improves emotion recognition performance on both the SEED (The SJTU Emotion EEG Dataset) and SEED-IV datasets [18].
Schurger et al. proposed EMS algorithm [19]. Traditional spatial filtering methods rely on predefined spatial regions or fixed spatial filters, whose static nature limits their ability to dynamically adapt to the temporal evolution of neural activity. The core advantage of EMS lies in its “experiment-effect-driven” characteristic, which constructs time-varying spatial filters to capture the spatiotemporal dynamics of emotional differences [19]. Although the EMS algorithm can effectively track the dynamic changes in experimental effects, its direct construction of filters based on raw signal differences makes it susceptible to interference from high-frequency noise and artifacts. To address this issue, this study proposes a novel framework, SSA-EMS, which integrates EMS with SSA. The SSA-EMS framework leverages the strengths of SSA in time series denoising and reconstruction, as well as the advantages of EMS in spatial filtering and dynamic adaptation, aiming to achieve effective feature extraction from EEG signals.

2. Dataset and Methods

2.1. Dataset

The SEED dataset, provided by Shanghai Jiao Tong University, is a publicly available dataset primarily used for EEG-based affective computing research [20]. This dataset contains EEG signals from 15 Chinese participants (7 males and 8 females). EEG signals were recorded using a 62-channel electrode cap based on the international 10–20 system, covering major brain regions. Each participant completed 15 trials, during which they watched 15 film clips, each approximately 4 min long. These clips were carefully selected and edited to effectively elicit one of three distinct target emotions: positive, neutral, or negative. Prior to each clip, a 5 s cue was presented, followed by a 45 s period for self-assessment after viewing. A 15 s rest period was provided between consecutive clips. The presentation order of the films was specifically designed to prevent two clips targeting the same emotion from being shown consecutively [20]. The entire experiment was conducted over three sessions per participant, with sessions spaced approximately one week apart. Each session corresponds to one data file, resulting in a total of 45 files. Data in each file are stored as a matrix in the format “channels × data points” and saved in .mat format. To protect privacy, participant identities are anonymized using numerical identifiers [20].
This study utilized the “Preprocessed_EEG” EEG data files provided by the SEED dataset, as these files have already undergone downsampling and preprocessing. The data were downsampled to 200 Hz and filtered using a 0–75 Hz bandpass filter [21].

2.2. Methods

2.2.1. EMS Algorithm

Effect-Matched Spatial Filtering (EMS) is a data-driven, effect-oriented spatial filtering method designed to reduce high-dimensional EEG data into time series that reveal specific experimental effects. Grounded in matched filter theory and operating under the assumption of independent and identically distributed (i.i.d.) noise, EMS dynamically constructs spatial filters that align with the experimental effect of interest. At each time point, the method estimates, in a data-driven manner, a spatial filter that captures the spatial pattern of differences between experimental conditions. The original data are then projected onto this adaptively derived “functional axis” in sensor space [19]. Unlike methods based on anatomical landmarks or prior assumptions, EMS derives its filters directly from the empirically observed contrast patterns between experimental conditions. Furthermore, the filter weights are adjusted dynamically over time to maximize the expression of the targeted experimental effect in the sensor space, thereby enhancing the temporal resolution and sensitivity of effect detection.
The core mathematical principle of the EMS algorithm is to find the optimal projection direction ω that maximizes the difference between two classes of signals.
ω = arg max w w T ( μ 1 μ 0 ) w
Here, μ 1 and μ 0 represent the average response patterns under two different conditions, respectively. The analytical solution to this optimization problem is given by:
ω = μ 1 μ 0 μ 1 μ 0 2 .
The geometric interpretation is that by constructing the normal vector of a hyperplane, the signed distance from each data to the hyperplane is computed, where a positive value indicates that the data lies on the side pointed to by the normal vector, and a negative value indicates that it lies on the opposite side.

2.2.2. SSA Algorithm

Singular Spectrum Analysis (SSA) is a non-parametric time series analysis method based on linear algebra and dynamical systems theory [22,23,24,25]. Its core idea is to decompose a complex time series into several components with clear physical interpretations (such as trend, periodicity, and noise) through a “decomposition–reconstruction” procedure, thereby achieving objectives such as denoising, forecasting, and periodicity extraction [26,27]. Compared with traditional methods such as Fourier transform and wavelet transform, SSA does not rely on the assumption of data stationarity, thus exhibiting greater adaptability, especially for analyzing non-stationary and nonlinear time series [28,29,30]. The SSA algorithm consists of two main stages—decomposition and reconstruction [31,32,33]—and can be specifically divided into the following four key steps:
Step 1. Embedding
Given a time series sequence X = ( x 1 , x 2 , , x N ) of length N , select a window length L (typically L N / 2 ), and construct the trajectory matrix T . Here, K   =   N     L   +   1 , and the matrix T is an L × K Hankel matrix.
Step 2. Singular Value Decomposition (SVD)
The trajectory matrix T is subjected to SVD T = U Σ V T . The trajectory matrix can be represented as a sum of elementary components and
T = i = 1 L T i = i = 1 L σ i u i v i T ,
where u i and v i are the i -th columns of U and V , respectively.
Step 3: Grouping
Based on the magnitude of the singular values, select the first r components (typically r L   , and discard the remaining components (noise):
T reconstructed = i = 1 r T i .
For example, in the analysis of brain electrical signals, the first three components may correspond to the main neural oscillation modes, while higher-order components i 4 may represent noise.
Step 4. Reconstruction
The reconstructed matrix T reconstructed after dimensionality reduction is transformed back into a time series through diagonal averaging. The formula for this transformation is given by:
x ˜ n = 1 w n ( l , k ) D n T l , k ,
where D n = { ( l , k ) l + k 1 = n } represents the set of elements on the n -th diagonal line, and w n is the number of elements on the n -th diagonal line. Finally, the denoised signal X ˜ = ( x ˜ 1 , x ˜ 2 , , x ˜ N ) is obtained.
From the perspective of information theory, this process is essentially a re-distribution and filtering of the amount of information in time series [34]. Since singular values are discrete and larger singular values correspond to modes with greater energy contributions within the system, the information content of the system can be quantified as:
H = i = 1 n p i log p i ,
where p i represents the proportion of the total energy occupied by the i-th singular value, i.e.,
p i = σ i 2 j = 1 r σ j 2 ,
where σ i is the i-th singular value, and r is the rank of the trajectory matrix. This entropy value reflects the energy distribution of the signal in various modes. Therefore, the information-theoretic basis of SSA denoising lies in: retaining high p i modes (low entropy, corresponding to signals), and discarding low p i modes (high entropy, corresponding to noise), thus achieving signal-noise separation.

2.2.3. Feature Extraction Process Based on SSA-EMS Algorithm

The process of the EMS algorithm can be divided into two stages: spatial filter learning and data projection [19]. This paper combines the respective advantages of SSA and EMS, embedding SSA into the learning process of the EMS spatial filter, and proposes a SSA-EMS feature extraction algorithm based on a three-stage entropy control framework of “noise entropy suppression → mutual information optimization → entropy rate convergence”. The algorithm flow is as follows:
1. 
Spatial Filter Learning
Let the dataset be denoted as X N × C × T , where X i C × T represents the EEG of the i-th epoch (with N being the number of epochs, C the number of channels, and T the number of time points), and y i { 0 , 1 } is the binary classification label corresponding to X i C × T .
Step 1. Calculate the mean values of categories: Calculate the average response corresponding to two categories in the dataset.
X ¯ 0 = 1 N 0 i : y i = 0 X i , X ¯ 1 = 1 N 1 i : y i = 1 X i ,
where N 0 and N 1 are the number of epochs for category 0 and category 1, respectively.
Step 2. Apply SSA preprocessing: Perform singular spectrum analysis on the mean values of the two categories separately to remove noise and retain important components.
X ˜ 0 = SSA ( X ¯ 0 ) , X ˜ 1 = SSA ( X ¯ 1 )
Step 3. Construct the spatial filter: Calculate the differential spatial filter W = X ˜ 0 X ˜ 1 , where W C × T is the spatial filter matrix.
Step 4. Normalization processing: Normalize the spatial filter by time points column-wise.
W : , t = W : , t W : , t 2 + ϵ , t = 1 , 2 , , T ,
where ϵ = 10 10 is a small constant to prevent division by zero.
In SSA, a signal is typically decomposed into four components. In this study, the first three principal components were retained to achieve signal denoising and feature enhancement. These first three components generally correspond to the dominant trend and prominent oscillatory patterns in the signal, encompassing key neural rhythms related to the task, while the fourth component mostly contains high-frequency noise or non-specific fluctuations [35]. Directly computing the differential spatial filter based on class-averaged signals may amplify these noise components, compromising the stability of feature extraction. Therefore, by retaining the first three components, the core neural activity features can be effectively preserved, while more robustly extracting stable differential patterns between the two conditions, thereby enhancing the reliability of the analysis.
2. 
Data Projection—mutual information optimization
In the data projection stage, the learned spatial filter W C × T is applied to the dataset X N × C × T , where W C × T is the normalized spatial filter that can be calculated through Einstein summation and convention X transform = einsum ( n c t , c t n t , X , W ) to compute the projection feature vector X transform N × T .
SSA-EMS employs a time-varying spatial filter W C × T to achieve dynamic feature selection, which is essentially minimizing the conditional entropy H ( Y | X ) , where Y represents the target emotional states. Compared with traditional static filters, this algorithm reduces the conditional uncertainty between features and emotional labels by constructing a difference space W = X ^ 0 X ^ 1 . The optimization is achieved through:
min W H ( Y | X transform ) X transform = X , W ,
where the projected feature X transform is calculated using Einstein summation and dot product, optimizing the mutual information I ( Y ; X transform ) .
3. 
Construction of Stable Representations—Entropy rate convergence
To obtain stable neural dynamic representations and construct feature representations that can be used for classification, let X transform N × T , where X transform ( i ) T is the transformed feature for the i-th epoch, and y ( i ) { 0 , 1 } is its label. For each class k { 0 , 1 } , calculate the average feature vector corresponding to the two classes:
f k ¯ = 1 N k ¯ i : y ( i ) = k X transform ( i ) ,
where N k is the number of epochs in class k. This process, by averaging the same class of epochs in X transform , can obtain more stable neural dynamic representations.
The essence of this stage is to reduce entropy. The entropy H r ( X t r a n s f o r m ( i ) )   of the feature X t r a n s f o r m ( i ) within a single epoch reflects the temporal uncertainty. Under the assumption of independent and identically distributed (i.i.d.) noise, averaging over same class f k can lead to entropy reduction:
H r ( f k ) 1 N k H r ( X t r a n s f o r m ( i ) ) .
This process suppresses individual random variations and enhances the generalization ability across samples.
From the above process, it can be seen that the SSA-EMS algorithm can reduce the high-dimensional feature X N × C × T to a feature matrix F 2 × T , thus achieving the purpose of dimensionality reduction. F 2 × T is the time-domain feature vector extracted by the SSA-EMS algorithm, which can be used for subsequent classification training and recognition.
The aforementioned procedure can also employ leave-one-out or cross-validation methods for spatial filter learning and data projection to achieve better robustness. This study adopts the above procedure combined with a stratified 5-fold cross-validation method for spatial filter learning and data projection.

2.2.4. Cross-Subject Data Processing in Seed Emotional EEG

This study conducts cross-subject emotion recognition research based on the first experimental EEG data of 15 subjects from the SEED dataset. Each subject’s file contains 15 EEG segments (5 positive, 5 neutral, and 5 negative). A total of 225 EEG signal samples (75 per emotional category) were collected from the 15 subjects. All EEG data were preprocessed using the MNE library into standard Raw objects [36] and segmented into 1 s epochs X N × C × T .
1. 
Cross-subject sample combination data processing
In traditional cross-subject studies, “subject-independent” evaluation typically employs a simple strategy of splitting training and test sets, which is a coarse-grained approach that makes it difficult to precisely characterize model transfer performance under different subject combinations. To overcome this limitation, this study designs a more challenging cross-subject combination data fusion method: by computing the Cartesian product between emotional A data from each subject and emotional B data from each subject, thereby forming all pairwise combinations of samples across subjects. This method not only preserves the core advantages of “subject-independent” evaluation but also enables assessment of the model’s emotional transfer performance between subjects, suppressing interference caused by inter-subject variability. The specific implementation steps are as follows:
Step 1. Emotional pairing: Construct three emotional combinations: positive vs. negative, positive vs. neutral, and negative vs. neutral.
Step 2. Data combination: For each emotional combination (A vs. B), combine 75 samples of type A and 75 samples of type B using the Cartesian product, resulting in 5625 sample units.
Step 3. Data Concatenation: Concatenate the two types of emotions in each sample unit along the first dimension to form a three-dimensional matrix X ( M + N ) × C × T .
Step 4. Dimensionality Reduction: Apply the SSA-EMS method to reduce the three-dimensional matrix X ( M + N ) × C × T to a feature matrix F 2 × T , where each row represents the features of one emotion category.
Step 5. Feature Integration: After processing all 5625 sample units, construct a two-dimensional feature matrix F 11250 × T containing 11,250 (5625 × 2) feature vectors.
The specific steps are shown in Figure 1.
2. 
Subject-Independent Evaluation Method
This study also adopts the conventional subject-independent evaluation paradigm to verify the performance of the SSA-EMS algorithm. The specific implementation is as follows:
1. Data Partitioning:
Training set: EEG data from the first 7 subjects.
Test set: EEG data from the last 8 subjects.
2. Sample Construction Strategy:
For each pair of emotion combinations (A vs. B), the following procedure is adopted:
Step 1. For each subject, 15 samples of class A and 15 samples of class B are selected from their three experimental sessions (as noted, each subject participated in three sessions approximately one week apart).
Step 2. A Cartesian product operation is performed to generate 225 (15 × 15) sample pairs, all derived from the same subject. All sample pairs are strictly maintained as within-subject combinations, avoiding any cross-subject mixing.
3. Dataset Scale:
Total training set: 7 subjects × 225 units = 1575 sample units
Total test set: 8 subjects × 225 units = 1800 sample units
This scheme strictly adheres to the core requirements of subject-independent evaluation: (1) complete subject-level separation between training and test sets; (2) all data combinations maintain within-subject consistency; (3) ensuring that the model never accesses data from any test subject during training [37]. The specific steps are shown in Figure 2.

3. Results

3.1. Cross-Subject Sample Combination Classification

This experiment employs stratified 5-fold cross-validation to separately evaluate the influence of different frequency bands on the classification of emotional states for the three experimental groups: “positive vs. neutral”, “positive vs. negative”, and “neutral vs. negative”. A Random Forest classifier (n_estimators = 100) and a Radial Basis Function Support Vector Machine (RBF-SVM, kernel = ‘rbf’, C = 1, gamma = ‘scale’) are used to compare classification performance. The final experimental result is the average recognition rate from the stratified 5-fold cross-validation.
Figure 3 shows the recognition accuracy for pairwise classification of three emotional states (positive vs. neutral, negative vs. neutral, positive vs. negative) across different frequency bands (All band, Delta, Theta, Alpha, Beta, and Gamma). The x-axis represents different frequency bands, and the y-axis represents the recognition accuracy.
As shown in Figure 3, the RF outperformed SVM across all tasks and frequency bands. For instance, in the negative vs. neutral emotion classification task using the All Bands, RF achieved an accuracy of 98.61%, compared to 94.10% for SVM; in the Delta band for the negative vs. positive emotion classification task, RF achieved an accuracy of 95.11%, also higher than SVM’s 92.16%.
In the recognition of the three emotion state combinations, the performance using the full frequency band was superior to that of each individual frequency band. Notably, RF classifier achieved particularly high accuracy in the full frequency band, with classification accuracies of 98.61%, 98.44%, and 98.63% for the negative vs. neutral, negative vs. positive, and neutral vs. positive tasks, respectively.
In the single-band classification tasks, the Delta band consistently achieved the highest classification accuracy among the five individual bands (95.27% for negative vs. neutral, 95.11% for negative vs. positive, and 94.98% for neutral vs. positive). The Gamma band ranked second, with accuracies exceeding 90% across all tasks, demonstrating strong classification performance. The Theta, Alpha, and Beta bands generally yielded lower classification accuracies compared to the Delta and Gamma bands.
The confusion matrix is a fundamental tool for evaluating classification model performance, providing an intuitive visualization of the model’s discriminative ability across different classes [38,39]. Since RF demonstrated the best classification performance, Figure 4 presents the confusion matrices of the RF model on the test set using All-band features for the three binary emotion classification combinations (negative vs. neutral, negative vs. positive, neutral vs. positive), with the corresponding classification accuracies annotated. Overall, the model exhibited excellent performance across all three classification tasks, with accuracies exceeding 98%. The highest accuracy of 98.63% was achieved for the “neutral vs. positive” combination.
As evident from the confusion matrices, the diagonal elements (5544 vs. 5550, 5533 vs. 5542, 5547 vs. 5549) exhibit deep blue high-value regions, with proportions exceeding 98%, indicating the model’s exceptionally high accuracy in correctly classifying samples. The off-diagonal regions show lighter colors, corresponding to a small number of misclassified samples, specifically manifested as low false positive (FP) and false negative (FN) counts, with the values being close (81 vs. 75, 92 vs. 83, 78 vs. 76), suggesting the model does not exhibit a significant bias towards any particular class, demonstrating well-balanced classification performance.

3.2. Classification Results of “Subject-Independent” Evaluation

This experiment employed a Random Forest model (n_estimators = 100) to evaluate EEG-based emotion recognition in a subject-independent paradigm using a binary classification task. The final result is the average recognition rate obtained through stratified 5-fold cross-validation. The confusion matrix analysis results for the three test sets are shown in Figure 5. In Figure 5a (Neutral vs. Positive) and Figure 5b (Negative vs. Neutral), both false positive (FP) and false negative (FN) counts remain low and are similar in magnitude, indicating that the model exhibits no significant bias toward either class, demonstrating well-balanced classification performance. Notably, in Figure 5c (Negative vs. Positive), the misclassification rate for positive emotions is significantly higher than that for negative emotions (FN = 116 vs. FP = 18), indicating that positive emotion samples are more prone to misclassification.
Despite these differences, all three emotion binary classification tasks achieved excellent performance on the test set (accuracy > 96%), and comprehensive analysis indicates that the SSA-EMS algorithm possesses strong cross-subject generalization capability in cross-subject experiments.

4. Discussion

This study employed the SSA-EMS algorithm to process and analyze EEG data from three emotional unit combinations (neutral vs. positive, negative vs. neutral, negative vs. positive). The results demonstrate that the SSA-EMS algorithm exhibits excellent classification performance both in cross-subject data processing and in the “subject-independent” evaluation framework. This performance advantage stems from the complementary effects of SSA’s denoising capability and EMS’s strategy of maximizing inter-class differences. The physiological differences between emotional units are the fundamental cause of signal amplitude differentiation, while the mathematical principle of EMS lies in capturing the EEG pattern differences between two emotional states through spatial filters, thereby maximizing the divergence between the two classes. When brain activity differences are prominent, this also leads to larger mean differences between classes. The synergistic effect of SSA and EMS constitutes the key to the algorithm’s superior classification performance. Specifically, SSA enhances the signal-to-noise ratio by decomposing and reconstructing the trajectory matrix to remove high-frequency noise, which in turn enables the spatial filter weights computed by the subsequent EMS algorithm to achieve higher discriminability. Furthermore, the normalization mechanism of the filter amplifies the impact of intrinsic differences, essentially enhancing features by optimizing weight distribution across channels. The combined effect of these mechanisms enables the SSA-EMS algorithm to achieve significant advantages in feature extraction for emotional EEG signals.
From the perspective of information theory, the SSA-EMS algorithm constructs a three-stage entropy control framework: “noise entropy suppression → mutual information optimization → entropy rate convergence.”
1. SSA Layer: Reduces the noise entropy H noise of the input signal and improves the signal-to-noise ratio (SNR):
SNR = H signal H noise .
2. EMS Layer: Maximizes mutual information using a differential filter:
I ( W i ; Y ) .
3. Stable Representation Layer: Achieves entropy rate convergence:
H r ( f k ) const .
Figure 6 displays the average feature signals of three randomly selected emotional unit combinations from the full-frequency band, processed by the SSA-EMS algorithm. As shown, the average features of different emotional unit combinations exhibit significant distinguishability. It should be noted that the feature curve shapes presented in Figure 6 represent only one example among many possible configurations. Although diverse curve morphologies can be generated by different emotional unit combinations (not all shown due to space limitations), these features consistently maintain stable distinguishability. The results indicate that the SSA-EMS algorithm can effectively capture the intrinsic differential characteristics of distinct EEG emotional states, demonstrating significant universality in binary emotional classification.
This study is compared with the research conducted by Jie Quan, Yihang Dong, Longbin Jin, and their colleagues. Jie Quan et al. proposed a cross-subject emotion EEG classification algorithm based on multi-source domain selection and sub-domain adaptation. Experiments on two public datasets, SEED and SEED-IV, demonstrated that the proposed method achieved average classification accuracies of 92.83% and 79.30%, respectively [40]. Yihang Dong et al. introduced a multimodal cross-scale fusion model named “Mood Reader”. The related experiments were carried out on the SEED and SEED-V public datasets, showing that when the Mood Reader model combined DE features and employed a sliding window sequential combination (w SWC) strategy, the recognition rate was 93.12% ± 4.75% [41]. Longbin Jin and Eun Yi Kim innovatively proposed a cross-subject emotion prediction model integrating temporal analysis with spatial correlation characteristics. Experiments on the DEAP physiological signal emotion dataset and the Shanghai Jiao Tong University SEED dataset indicated significant advantages when channel features were used in conjunction with LSTM: achieving accuracies of 98.93% and 99.10% in the DEAP valence and arousal binary classification tasks, respectively, and reaching a recognition accuracy of 99.63% in the SEED three-classification task, surpassing the current state-of-the-art levels [42].
Compared with previous studies, the present research still has the following limitations: The SSA-EMS algorithm is restricted to binary classification tasks, which prevents it from handling multi-class problems simultaneously. Since emotion recognition typically involves multiple states (e.g., happiness, anger, sadness, surprise), the model cannot be directly applied to multi-class scenarios. Although three emotion categories are included in the study, they require multiple binary classifications to be combined, thereby reducing the model’s capability in distinguishing complex emotional states. Moreover, the performance of the method is also dependent on the number of available data samples, with larger datasets generally leading to improved results [43]. Therefore, extending the binary classification algorithm to a multi-class framework constitutes a direction for future research. Additionally, integrating the EEG features processed by SSA-EMS with other physiological signals to construct a multimodal fusion approach for emotion recognition could be explored.

5. Conclusions

The SSA-EMS algorithm proposed in this study significantly enhances cross-subject EEG-based emotion recognition performance by integrating the noise-reduction capability of singular spectrum analysis with the dynamic difference enhancement property of effect-matching spatial filtering. Experiments demonstrate that the algorithm achieves over 98% accuracy in binary classification tasks of positive, neutral, and negative emotions using All-band features, with high recognition rates also observed in the Delta and Gamma bands, and the Random Forest classifier outperforms the Support Vector Machine. Rigorous “subject-independent” evaluation further validates the model’s generalization ability on unseen subjects (accuracy > 96%), indicating that the SSA-EMS algorithm can stably extract dynamic features of emotional EEG. However, the algorithm supports only binary classification tasks, limiting its application in complex multi-emotion scenarios; future research should explore extensions to multi-class classification and strategies for multimodal physiological signal fusion.

Author Contributions

Conceptualization, Y.L.; methodology, Y.L.; validation, Y.L.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China under Grant 62377018, the National Key R\&D Program of China under Grant 2024YFC3308300, the selfdetermined Research Funds of CCNU from the Colleges’ Basic Research and Operation of MOE under Grants CCNU24ZZ154.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Muthunayagom, P.; Subhajini, A.C. Deciphering Feelings: Exploring EEG Data for Human Emotion Recognition. In Proceedings of the 2024 International Conference on Sustainable Communication Networks and Application (ICSCNA), Theni, India, 11–13 December 2024; pp. 1181–1187. [Google Scholar]
  2. Chang, J.; Wang, Z.; Jia, J. Enhancing Emotion Recognition in EEG Signals using Fractional Fourier Transform. In Proceedings of the 2024 IEEE 17th International Conference on Signal Processing (ICSP), Suzhou, China, 28–31 October 2024; pp. 110–115. [Google Scholar]
  3. Gkintoni, E.; Aroutzidis, A.; Antonopoulou, H.; Halkiopoulos, C. From Neural Networks to Emotional Networks: A Systematic Review of EEG-Based Emotion Recognition in Cognitive Neuroscience and Real-World Applications. Brain Sci. 2025, 15, 220. [Google Scholar] [CrossRef]
  4. Rahman, M.M.-E.-N.; Sarkar, A.K.; Hossain, M.A.; Hossain, M.S.; Islam, M.R.; Hossain, M.B.; Quinn, J.M.W.; Moni, M.A. Recognition of human emotions using EEG signals: A review. Comput. Biol. Med. 2021, 136, 104696. [Google Scholar] [CrossRef]
  5. Hbibi, B.; Khiari, C.; Wirsing, K.; Mili, L.; Baccar, K.; Mami, A. Identifying and Removing Interference and Artifacts in Multifractal Signals With Application to EEG Signals. IEEE Access 2023, 11, 119090–119105. [Google Scholar] [CrossRef]
  6. Sadiya, S.; Alhanai, T.; Ghassemi, M.M. Artifact Detection and Correction in EEG data: A Review. In Proceedings of the 2021 10th International IEEE/EMBS Conference on Neural Engineering (NER), Virtual Event, 4–6 May 2021; pp. 495–498. [Google Scholar]
  7. Jmail, N.; Gavaret, M.; Wendling, F.; Kachouri, A.; Hamadi, G.; Badier, J.M.; Bénar, C.G. A comparison of methods for separation of transient and oscillatory signals in EEG. J. Neurosci. Methods 2011, 199, 273–289. [Google Scholar] [CrossRef] [PubMed]
  8. Lu, W.; Zhang, X.; Xia, L.; Ma, H.; Tan, T.-P. Domain adaptation spatial feature perception neural network for cross-subject EEG emotion recognition. Front. Hum. Neurosci. 2024, 18, 1471634. [Google Scholar] [CrossRef] [PubMed]
  9. Liu, S.; Wang, Z.; An, Y.; Li, B.; Wang, X.; Zhang, Y. DA-CapsNet: A multi-branch capsule network based on adversarial domain adaption for cross-subject EEG emotion recognition. Knowl. Based Syst. 2023, 283, 111137. [Google Scholar] [CrossRef]
  10. Wang, W.; Yang, J.; Li, S.; Wang, B.; Yang, K.; Sang, S.; Zhang, Q.; Liu, B. A Random Forest Weights and 4-Dimensional Convolutional Recurrent Neural Network for EEG Based Emotion Recognition. IEEE Access 2024, 12, 39549–39563. [Google Scholar] [CrossRef]
  11. Jafari, M.; Shoeibi, A.; Khodatars, M.; Bagherzadeh, S.; Shalbaf, A.; López-García, D.; Górriz, J.M.; Acharya, U.R. Emotion recognition in EEG signals using deep learning methods: A review. Comput. Biol. Med. 2023, 165, 107450. [Google Scholar] [CrossRef]
  12. Chen, B.; Chen, C.L.P.; Zhang, T. GDDN: Graph Domain Disentanglement Network for Generalizable EEG Emotion Recognition. IEEE Trans. Affect. Comput. 2024, 15, 1739–1753. [Google Scholar] [CrossRef]
  13. Wang, Y.; Zhang, B.; Tang, Y. DMMR: Cross-Subject Domain Generalization for EEG-Based Emotion Recognition via Denoising Mixed Mutual Reconstruction. Proc. AAAI Conf. Artif. Intell. 2024, 38, 628–636. [Google Scholar] [CrossRef]
  14. Li, P.; Li, A.; Li, X.; Lv, Z. Cross-Subject Emotion Recognition with CT-ELCAN: Leveraging Cross-Modal Transformer and Enhanced Learning-Classify Adversarial Network. Bioengineering 2025, 12, 528. [Google Scholar] [CrossRef]
  15. Liang, S.; Li, L.; Zu, W.; Feng, W.; Hang, W. Adaptive deep feature representation learning for cross-subject EEG decoding. BMC Bioinform. 2024, 25, 1–19. [Google Scholar] [CrossRef]
  16. Anuragi, A.; Sisodia, D.S.; Pachori, R.B. EEG-based cross-subject emotion recognition using Fourier-Bessel series expansion based empirical wavelet transform and NCA feature selection method. Inf. Sci. 2022, 610, 508–524. [Google Scholar] [CrossRef]
  17. Guo, W.; Li, Y.; Liu, M.; Ma, R.; Wang, Y. Functional connectivity-enhanced feature-grouped attention network for cross-subject EEG emotion recognition. Knowl. Based Syst. 2024, 283, 111199. [Google Scholar] [CrossRef]
  18. Peng, Y.; Liu, H.; Kong, W.; Nie, F.; Lu, B.-L.; Cichocki, A. Joint EEG Feature Transfer and Semisupervised Cross-Subject Emotion Recognition. IEEE Trans. Ind. Inform. 2023, 19, 8104–8115. [Google Scholar] [CrossRef]
  19. Schurger, A.; Marti, S.; Dehaene, S. Reducing multi-sensor data to a single time course that reveals experimental effects. BMC Neurosci. 2013, 14, 122. [Google Scholar] [CrossRef]
  20. Zheng, W.-L.; Lu, B.-L. Investigating Critical Frequency Bands and Channels for EEG-based Emotion Recognition with Deep Neural Networks. IEEE Trans. Auton. Ment. Dev. (IEEE TAMD) 2015, 7, 162–175. [Google Scholar] [CrossRef]
  21. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential Entropy Feature for EEG-based Emotion Classification. In Proceedings of the 6th International IEEE EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar]
  22. Hassani, H. Singular Spectrum Analysis: Methodology and Comparison. J. Data Sci. 2007, 5, 239–257. [Google Scholar] [CrossRef]
  23. Rodrigues, P.C.; Tuy, P.G.S.E.; Mahmoudvand, R. Randomized singular spectrum analysis for long time series. J. Stat. Comput. Simul. 2018, 88, 1921–1935. [Google Scholar] [CrossRef]
  24. Kume, K.; Nose-Togawa, N. An Adaptive Orthogonal SSA Decomposition Algorithm for a Time Series. Adv. Data Sci. Adapt. Anal. 2018, 10, 1850002. [Google Scholar] [CrossRef]
  25. Alomar, A.; Dahleh, M.A.; Mann, S.; Shah, D. SAMoSSA: Multivariate Singular Spectrum Analysis with Stochastic Autoregressive Noise. arXiv 2023, arXiv:2305.16491. [Google Scholar] [CrossRef]
  26. Sanseverino, C.M.R. Singular spectrum analysis and forecasting of failure time series. Reliab. Eng. Syst. Saf. 2013, 114, 126–136. [Google Scholar] [CrossRef]
  27. Golyandina, N.; Zhigljavsky, A.A. SSA for Forecasting, Interpolation, Filtering and Estimation; Singular Spectrum Analysis for Time Series; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  28. Ghanbarzadeh, M.; Aminghafari, M. A New Hybrid-Multiscale SSA Prediction of Non-Stationary Time Series. Fluct. Noise Lett. 2016, 15, 1650005. [Google Scholar] [CrossRef]
  29. Harmouche, J.; Fourer, D.; Auger, F.; Borgnat, P.; Flandrin, P. The Sliding Singular Spectrum Analysis: A Data-Driven Nonstationary Signal Decomposition Tool. IEEE Trans. Signal Process. 2018, 66, 251–263. [Google Scholar] [CrossRef]
  30. Leles, M.C.R.; Sansão, J.P.H.; Mozelli, L.A.; Guimarães, H.N. Improving reconstruction of time-series based in Singular Spectrum Analysis: A segmentation approach. Digit. Signal Process. 2017, 77, 63–76. [Google Scholar] [CrossRef]
  31. Bonenberger, C.M.A.; Ertel, W.; Schwenker, F.; Schneider, M. Singular Spectrum Analysis and Circulant Maximum Variance Frames. Adv. Data Sci. Adapt. Anal. 2022, 14, 2250008. [Google Scholar] [CrossRef]
  32. Mahmoudvand, R.; Najari, N.; Zokaei, M. On the Optimal Parameters for Reconstruction and Forecasting in Singular Spectrum Analysis. Commun. Stat.—Simul. Comput. 2013, 42, 860–870. [Google Scholar] [CrossRef]
  33. Golyandina, N.; Nekrutkin, V.; Zhigljavsky, A.A. Analysis of Time Series Structure—SSA and Related Techniques. In Monographs on Statistics and Applied Probability; CRC Press: Boca Raton, FL, USA, 2001. [Google Scholar]
  34. Jiang, J.; Xie, H.-B. Denoising Nonlinear Time Series Using Singular Spectrum Analysis and Fuzzy Entropy. Chin. Phys. Lett. 2016, 33, 100501. [Google Scholar] [CrossRef]
  35. Bonizzi, P.; Karel, J.M.H.; Meste, O.; Peeters, R.L.M. Singular Spectrum Decomposition: A New Method for Time Series Decomposition. Adv. Data Sci. Adapt. Anal. 2014, 6, 1450011. [Google Scholar] [CrossRef]
  36. Gramfort, A.; Luessi, M.; Larson, E.; Engemann, D.A.; Strohmeier, D.; Brodbeck, C.; Goj, R.; Jas, M.; Brooks, T.; Parkkonen, L.; et al. MEG and EEG data analysis with MNE-Python. Front. Neurosci. 2013, 7, 267. [Google Scholar] [CrossRef]
  37. Hwang, S.; Ki, M.; Hong, K.; Byun, H. Subject-Independent EEG-based Emotion Recognition using Adversarial Learning. In Proceedings of the 2020 8th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Republic of Korea, 26–28 February 2020; pp. 1–4. [Google Scholar]
  38. Hamel, L. Model Assessment with ROC Curves. In Encyclopedia of Data Warehousing and Mining; 2009; Available online: https://homepage.cs.uri.edu/faculty/hamel/pubs/hamel-roc.pdf (accessed on 5 August 2025).
  39. Fahmy Amin, M. Confusion Matrix in Binary Classification Problems: A Step-by-Step Tutorial. J. Eng. Res. 2022, 6, 1. [Google Scholar] [CrossRef]
  40. Quan, J.; Li, Y.; Wang, L.; He, R.; Yang, S.; Guo, L. EEG-based cross-subject emotion recognition using multi-source domain transfer learning. Biomed. Signal Process. Control 2023, 84, 104741. [Google Scholar] [CrossRef]
  41. Dong, Y.; Chen, X.; Shen, Y.; Ng, M.K.-P.; Qian, T.; Wang, S. Multi-modal Mood Reader: Pre-trained Model Empowers Cross-Subject Emotion Recognition. arXiv 2024, arXiv:2405.19373. [Google Scholar]
  42. Jin, L.; Kim, E.Y. Interpretable Cross-Subject EEG-Based Emotion Recognition Using Channel-Wise Features. Sensors 2020, 20, 6719. [Google Scholar] [CrossRef] [PubMed]
  43. Antonacci, Y.; Toppi, J.; Pietrabissa, A.; Anzolin, A.; Astolfi, L. Measuring Connectivity in Linear Multivariate Processes with Penalized Regression Techniques. IEEE Access 2024, 12, 30638–30652. [Google Scholar] [CrossRef]
Figure 1. Cross-subject sample combination data processing.
Figure 1. Cross-subject sample combination data processing.
Entropy 27 00986 g001
Figure 2. Subject-Independent Evaluation Method.
Figure 2. Subject-Independent Evaluation Method.
Entropy 27 00986 g002
Figure 3. Recognition accuracy of emotional states across different frequency bands. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Figure 3. Recognition accuracy of emotional states across different frequency bands. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Entropy 27 00986 g003
Figure 4. Confusion matrix of RF under All-band feature classification. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Figure 4. Confusion matrix of RF under All-band feature classification. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Entropy 27 00986 g004
Figure 5. Confusion matrix of RF under “subject-independent” evaluation. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Figure 5. Confusion matrix of RF under “subject-independent” evaluation. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Entropy 27 00986 g005
Figure 6. Examples of average feature signals for three emotional unit combinations. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Figure 6. Examples of average feature signals for three emotional unit combinations. (a) negative vs. neutral; (b) negative vs. positive; (c) neutral vs. positive.
Entropy 27 00986 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, Y.; Chen, J. Cross-Subject EEG Emotion Recognition Using SSA-EMS Algorithm for Feature Extraction. Entropy 2025, 27, 986. https://doi.org/10.3390/e27090986

AMA Style

Lu Y, Chen J. Cross-Subject EEG Emotion Recognition Using SSA-EMS Algorithm for Feature Extraction. Entropy. 2025; 27(9):986. https://doi.org/10.3390/e27090986

Chicago/Turabian Style

Lu, Yuan, and Jingying Chen. 2025. "Cross-Subject EEG Emotion Recognition Using SSA-EMS Algorithm for Feature Extraction" Entropy 27, no. 9: 986. https://doi.org/10.3390/e27090986

APA Style

Lu, Y., & Chen, J. (2025). Cross-Subject EEG Emotion Recognition Using SSA-EMS Algorithm for Feature Extraction. Entropy, 27(9), 986. https://doi.org/10.3390/e27090986

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop