Next Article in Journal
The ‘Postural Rhythm’ of the Ground Reaction Force during Upright Stance and Its Conversion to Body Sway—The Effect of Vision, Support Surface and Adaptation to Repeated Trials
Previous Article in Journal
Alteration of Excitation/Inhibition Imbalance in the Hippocampus and Amygdala of Drug-Resistant Epilepsy Patients Treated with Acute Vagus Nerve Stimulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Subject-Independent EEG Emotion Recognition Based on Genetically Optimized Projection Dictionary Pair Learning

School of Information Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(7), 977; https://doi.org/10.3390/brainsci13070977
Submission received: 3 May 2023 / Revised: 18 June 2023 / Accepted: 19 June 2023 / Published: 21 June 2023
(This article belongs to the Section Neural Engineering, Neuroergonomics and Neurorobotics)

Abstract

:
One of the primary challenges in Electroencephalogram (EEG) emotion recognition lies in developing models that can effectively generalize to new unseen subjects, considering the significant variability in EEG signals across individuals. To address the issue of subject-specific features, a suitable approach is to employ projection dictionary learning, which enables the identification of emotion-relevant features across different subjects. To accomplish the objective of pattern representation and discrimination for subject-independent EEG emotion recognition, we utilized the fast and efficient projection dictionary pair learning (PDPL) technique. PDPL involves the joint use of a synthesis dictionary and an analysis dictionary to enhance the representation of features. Additionally, to optimize the parameters of PDPL, which depend on experience, we applied the genetic algorithm (GA) to obtain the optimal solution for the model. We validated the effectiveness of our algorithm using leave-one-subject-out cross validation on three EEG emotion databases: SEED, MPED, and GAMEEMO. Our approach outperformed traditional machine learning methods, achieving an average accuracy of 69.89% on the SEED database, 24.11% on the MPED database, 64.34% for the two-class GAMEEMO, and 49.01% for the four-class GAMEEMO. These results highlight the potential of subject-independent EEG emotion recognition algorithms in the development of intelligent systems capable of recognizing and responding to human emotions in real-world scenarios.

1. Introduction

EEG-based emotion recognition is a prominent research area in neuroscience and machine learning, as it holds the potential to enhance our understanding of the neural representation and objective measurement of emotions. Emotions play a crucial role in our daily lives, influencing decision making, behavior, and social interactions. However, traditional methods for measuring emotions, such as self-reporting or behavioral observation, possess limitations and are susceptible to bias. In contrast, an EEG offers a noninvasive, objective, and direct approach to measure emotions by detecting patterns of brain activity associated with specific emotional states [1]. This capability has the potential to advance the diagnosis and treatment of emotional disorders such as depression [2], anxiety [3], and post-traumatic stress disorder (PTSD) [4]. Moreover, the EEG holds promise for diverse applications, including human–computer interaction [5], affective computing [6,7], marketing research, and entertainment. Consequently, the development of reliable and accurate EEG emotion recognition systems bears great significance for the scientific community and society at large.
Numerous ongoing research efforts are dedicated to EEG emotion recognition, encompassing various research directions [8]. These include: (1) Time-Frequency Analysis, where the EEG signal is decomposed into different frequency bands using techniques such as wavelet transform or Fourier transform. Features extracted from each frequency band are employed to identify the emotional state [9,10]. (2) Independent Component Analysis (ICA) separates the EEG signal into independent components, with features extracted from each component used for emotion identification [11]. (3) Support Vector Machines (SVMs) are utilized as classifiers to categorize the EEG signal into distinct emotional states, employing features extracted from the signal [12]. (4) Fuzzy Logic models the uncertainty and imprecision in EEG signals for emotion recognition [13]. (5) Hidden Markov Models (HMMs) capture the temporal dynamics of EEG signals for emotion recognition [14]. (6) Multimodal emotion recognition combines EEG data with other modalities, such as facial expressions, speech, or physiological signals, to enhance the accuracy of EEG emotion recognition [15,16]. (7) Artificial Neural Networks (ANNs) are trained on the EEG signal to recognize different emotional states based on the extracted features [17]. (8) Deep learning techniques exhibit promising results in EEG-based emotion recognition, with ongoing exploration to develop models capable of capturing the intricate and dynamic patterns of brain activity associated with distinct emotional states [18,19,20,21]. (9) Transfer learning allows models trained on one dataset to adapt to another dataset with minimal retraining. Researchers are actively investigating the application of transfer learning to enhance the generalization performance of EEG emotion recognition models across different individuals and contexts [22,23].
Recent research has focused on developing algorithms capable of accurately classifying EEG signals associated with different emotional states. These algorithms employ machine learning techniques, such as support vector machines (SVMs) or deep learning, to assess the brain activity patterns within the EEG signal. However, a significant hurdle in EEG emotion recognition arises from the considerable variability in EEG signals across individuals and the challenges of controlling external factors that may influence brain activity, such as cognitive load or fatigue [22]. In order to address these challenges, we investigate methods to enhance the accuracy and resilience of classification algorithms.
The Projection Dictionary Pair Learning (PDPL) algorithm is a widely used machine learning technique employed in diverse applications, such as image processing [24], natural language processing [25], and Internet of Medical Things (IoMT) systems [26]. The PDPL serves as an effective approach for extracting significant features from high-dimensional data [27] and extends the standard Projection Dictionary Learning (PDL) algorithm. The fundamental concept of the PDPL revolves around identifying a pair of projection matrices that can map high-dimensional data to a lower-dimensional space while preserving the essential features. The algorithm comprises two main stages: dictionary learning and projection learning. In the dictionary learning stage, the algorithm acquires a set of basis functions or dictionary atoms that effectively and sparsely represent the input data by minimizing a cost function incorporating a data-fitting term and a sparsity-promoting term. In the projection learning stage, the algorithm determines a projection matrix that transforms the high-dimensional input data into a lower-dimensional space by minimizing a cost function containing a reconstruction error term and a regularization term. The PDPL process involves iterative updates of the dictionary and projection matrices until convergence. The PDPL offers several advantages over alternative machine learning algorithms, including improved sparsity, robustness to noise, enhanced performance, and interpretability, thereby rendering it suitable for a wide range of real-world applications.
The Projection Dictionary Pair Learning (PDPL) algorithm is a powerful and flexible method for learning dictionaries from high-dimensional data, offering advantages over traditional machine learning approaches. Its capacity to acquire interpretable and sparse dictionaries proves particularly valuable in applications that necessitate an understanding of the underlying features. The PDPL excels in selecting relevant features from datasets, enhancing the classification and clustering tasks. However, the effective tuning of its parameters relies on the researcher’s experience. In this study, we employ the Genetic Algorithm (GA), an optimization technique inspired by natural selection and evolution, to optimize the PDPL parameters for achieving optimal performance. The GA has demonstrated promising results in parameter optimization across various domains. For instance, the GA has been successfully employed to determine parameter values in the Deep Deterministic Policy Gradient, resulting in accelerated learning for the agent [28]. Additionally, the GA has been utilized for the multiobjective optimization of process parameters, employing a weighted objective sum method [29]. Moreover, the GA has been applied to SVM parameter optimization, effectively addressing grid search problems [30].
Genetic Algorithms (GA) simulate the natural selection process observed in biology, whereby individuals possessing favorable traits are more likely to survive, reproduce, and transmit their genes to subsequent generations [31]. In the context of optimization, these individuals represent potential solutions, while the genes correspond to the parameters or variables defining those solutions. GAs offer significant advantages in addressing intricate optimization problems, spanning domains such as engineering, finance, and artificial intelligence, where the search space is extensive, and alternative optimization algorithms prove ineffective [32].
The main contributions of this study can be summarized as follows:
  • In recognition of the variability in EEG-based emotion recognition among individuals, we applied the PDPL algorithm to perform cross-subject analysis, with a specific focus on feature selection.
  • The exploration of parameter space in the PDPL algorithm presents a substantial computational burden due to the wide range of parameter adjustments and the resulting extensive combinations. To address this challenge, we propose the utilization of the Genetic Algorithm (GA) for adaptive parameter optimization.
  • Our proposed method surpasses conventional machine learning approaches, demonstrating exceptional recognition performance. Specifically, it achieves an average accuracy of 69.89% on the SEED database, 24.11% on the MPED database, 64.34% for the two-class GAMEEMO dataset, and 49.01% for the four-class GAMEEMO dataset. These results shed light on the effectiveness of emotion recognition, particularly for females, providing valuable insights into their emotional susceptibility.
In this academic paper, we address the issue of individual differences in EEG emotion recognition. To overcome this challenge, we introduce a novel PDPL algorithm based on genetic optimization. The proposed algorithm demonstrates superior recognition performance across subjects. The subsequent sections of the paper are organized as follows: Section 2 provides an overview of the materials and our proposed methodology, Section 3 presents the experimental results and corresponding discussions, and Section 4 concludes the paper while outlining future research directions.

2. Materials and Methods

2.1. EEG Emotion Database

To validate the effectiveness of our proposed method, we conducted experiments on three well-established multicategory EEG emotion databases. Specifically, we classified the datasets using both discrete models, such as a two-class model for GAMEEMO, a three-class model for the SJTU Emotion EEG Dataset (SEED Database), and a seven-class model for the Multimodal Physiological Emotion Database (MPED Database), as well as the dimensional model, employing a four-class model for GAMEEMO.
The SEED Database (https://bcmi.sjtu.edu.cn/home/seed/seed.html accessed on 18 June 2023) comprises emotional EEG signals collected from 15 subjects. For our experiment, we carefully selected a set of 15 Chinese film clips from a larger pool of materials based on their emotional valence (positive, neutral, and negative). These clips were meticulously edited to ensure coherence in eliciting emotions and to maximize emotional impact. Each clip had an approximate duration of 4 min. The experiment consisted of a total of 15 trials. Prior to the start of each clip, participants were provided with a 5 s hint. After viewing each clip, participants were given 45 s for self-assessment, followed by a 15 s rest period before proceeding to the next clip in the session. To assess their emotional reactions to the stimulus, participants were required to complete a questionnaire immediately after each clip viewing [33]. In our study, we utilized the dominant features provided in the database, specifically the differential entropy (DE) features, as the input for our proposed method.
The MPED Database (https://github.com/Tengfei000/MPED accessed on 18 June 2023) is an emotion database that comprises an EEG, galvanic skin response, respiration, and electrocardiogram signals. This comprehensive database encompasses data from 23 subjects, capturing 64 EEG channel signals and brain activity, while presenting 28 emotional stimulation videos. The videos encompassed a range of emotions, including joy, funny, anger, fear, disgust, sadness, and neutrality. All data were meticulously collected within a controlled laboratory environment [16].
The GAMEEMO Database (https://data.mendeley.com/datasets/b3pn4kwpmn/3 accessed on 18 June 2023), a collection of EEG signals acquired during computer games, was obtained from 28 individuals using the portable and wearable 14-channel Emotiv Epoc+ EEG device. Each participant engaged in four different computer games (boring, calm, horror, and funny) for 5 min each, resulting in a total EEG data duration of 20 min per subject. To evaluate the emotional experience, participants rated each game using the Self-Assessment Manikin (SAM) form, which measures arousal and valence [34]. Based on the stimulus material, EEG emotion can be categorized into two types: Positive–Negative models and four types of Arousal–Valence models, as presented in Table 1.
The extraction of log spectral power features allows for a quantitative assessment of power distribution across different frequency bands, enabling the study of brain activity. In both the MPED and GAMEEMO databases, we extracted the log spectral power features from the EEG signals [35]. To perform this extraction, the EEG signals were filtered using an order-8 zero-phase IIR Butterworth filter in five frequency bands: delta + theta (1–8 Hz), alpha (8–12 Hz), beta (12–35 Hz), gamma-1 (35–70 Hz), and gamma-2 (70–100 Hz). The root mean square of the filtered signals within each frequency band was computed using nonoverlapping 1 s windows. Finally, the logarithm of the root mean square was calculated for each window and electrode.

2.2. GA-PDPL for EEG Emotion Recognition

The PDPL is a machine learning algorithm utilized for unsupervised feature learning and data representation. Its primary objective is to acquire a dictionary of basis vectors for representing features in a lower-dimensional space. Nevertheless, parameter adjustment in the PDPL relies on experience. To achieve the optimal parameter combination, we incorporated a genetic algorithm (GA) to optimize these parameters. This approach allowed us to enhance the effectiveness of the PDPL in EEG emotion recognition. Figure 1 illustrates the framework of the GA-PDPL algorithm. Upon inputting the training dataset into the system, the GA parameters were generated. Through several iterations of parameter optimization, the best parameter combination was determined and employed to train the PDPL model. Subsequently, the trained model was tested on the test set, ultimately yielding the emotional category of each test sample as the output.

2.2.1. Descriminative Dictionary Learning (DDL)

After the EEG signal was preprocessed, and the features were extracted, a sample could be expressed as f R ( b × c ) × 1 , where b and c are the number of frequency bands and the number of electrodes of the EEG signal, respectively. In addition, the label corresponding to the sample with a total of K emotional classes could be expressed as y { 1 , 2 , 3 , , k , , K } . We denote by F = { F 1 , , F k , , F K } and Y = { y 1 , , y k , , y K } a set of training samples and training labels from K classes, respectively, where F k = [ f 1 , f 2 , , f n ] R p × n is the training sample set of class k, p = b × c , y k = [ y 1 , y 2 , y 3 , , y n ] R 1 × n is the training label set of class k, and n is the number of samples of each class. DDL methods focus on acquiring a proficient data representation model from F to address classification tasks by leveraging the class label information of training data. This can be formulated within the framework presented below:
min D , A F DA F 2 + λ A p + Ψ ( D , A , Y ) .
In the training model (1), the scalar constant λ 0 , synthesis dictionary D , and coding coefficient matrix A of F over D are utilized. The data fidelity term F D A F 2 ensures the representation ability of D , while the p -norm regularizer A p is imposed on A . Additionally, a discrimination promotion function Ψ ( D , A , Y ) is used to ensure the discrimination power of D and A .

2.2.2. PDPL Model

Deep Learning (DL) methods exhibit variations in their approach to learning a dictionary and classifier for all classes. Some methods employ a shared dictionary, while others utilize a structured dictionary to enhance discrimination. However, these methods commonly rely on 0 or 1 -norm sparsity regularizers for the coding coefficients. Unfortunately, such reliance on sparsity regularization leads to inefficiencies during both the training and testing stages. To address this issue, we propose the PDPL model, which extends the conventional DL model presented in (1). The PDPL model introduces a pair of discriminative synthesis and analysis dictionaries. Unlike other DL methods, the PDPL model does not necessitate the use of costly 0 or 1 -norm sparsity regularizers. Instead, the coding coefficients can be explicitly obtained through linear projection.
The discriminative model in Equation (1) aims to train a synthesis dictionary D that can sparsely represent the signal F [24,27]. Unfortunately, obtaining the code A for this dictionary requires an expensive l rnomm sparse coding process. To improve the efficiency, we instead found an analysis dictionary P R m K × p that satisfied A = P F , enabling the highly efficient representation of F without the need for sparse coding. To accomplish this, we learned an analysis dictionary using the synthesis dictionary D , resulting in the following formulated model,
P * , D * = arg min P , D F D P F F 2 + Ψ ( D , P , F , Y ) .
In the DPL model, the analysis dictionary P was used for the analytical coding of F , while the synthesis dictionary D was utilized for the reconstruction of F , with discrimination function Ψ ( D , P , F , Y ) applied throughout. To improve the model’s efficiency, structured synthesis and analysis dictionaries D = D 1 , D 2 , , D k and P = P 1 , P 2 , , P K were learned. Each sub-dictionary pair for class k is produced by D k R p × m and P k R m × p . To ensure that samples from class i (where i k ) are projected towards a null space with the structured analysis dictionary P , P k was designed accordingly. This was achieved by leveraging sparse subspace clustering, which has demonstrated that under certain incoherence conditions, signals can be represented by their corresponding dictionary. The formulated equation for this process is shown below,
P k F i 0 , k i .
The structured synthesis dictionary D can also be utilized to reconstruct the data matrix F . Specifically, the sub-dictionary D k can efficiently reconstruct the data matrix F k from the projective code matrix P k F k . Therefore, the dictionary pair was utilized to minimize the reconstruction error,
min P , D k = 1 k F k D k P k F k F 2 .
Based on the preceding discussion, the formulation of the DPL model can be expressed as follows,
P * , D * = arg min P , D k = 1 k | | F k D k P k F k F 2 + λ | | P k F ¯ k | | F 2 , s . t . d i 2 2 1 .
The synthesis dictionary is represented by matrix D , which consists of atoms denoted as d i . To ensure stability in the Projection Dictionary Learning (PDL) process, the energy of each atom is constrained to prevent the trivial solution P k = 0 . Additionally, the complement of F k in the entire training set F is denoted as F ¯ k . While sparse coding is not necessarily crucial for classification, the DPL model offers faster computation and demonstrates highly competitive classification performance. Therefore, the following approach was adopted for classification purposes. To optimize the nonconvex objective function in Equation (5), a variable matrix A was introduced, and Equation (5) was relaxed to the following problem:
P * , A * , D * = arg min P , A , D k = 1 K F k D k A k F 2 = | | P k F k A k F 2 + λ | | P k F k | | F 2 , s . t . d i 2 2 1 .
The objective function in Equation (6) consists of terms involving the Frobenius norm, which is facilitated by a scalar constant τ , ensuring ease of solving. To initialize the analysis dictionary P and synthesis dictionary D , random matrices with a unit Frobenius norm were initially employed. Subsequently, the minimization process proceeded by iteratively updating A and D , P . The minimization procedure involved alternating between the following two steps:
(1)
Fix D and P , update A ,
A * = arg min A k = 1 K F k D k A k F 2 + τ P k F k A k F 2 .
We can obtain a closed-form solution for this standard least-squares problem:
A k * = D k T D k + τ I 1 τ P k F k + D k T F k .
(2)
Fix A , update D and P ,
P * = arg min P k = 1 K | | P k F k A k | | F 2 + λ | | P k F ¯ k | | F 2 D * = arg min D k = 1 K | | F k D k A k | | F 2 , s . t . d i 2 2 1 .
We can obtain closed-form solutions for P as follows:
P k * = τ A k F k T τ F k F i T + λ F ¯ k F ¯ k T + γ I ,
where γ is a small number, and I is the identity matrix. Introducing a variable S can optimize the D problem:
min D , S k = 1 K F k D k A k | | F 2 , s . t . D = S , S i 2 2 1 .
The optimal solution of (11) can be obtained by the Alternating Direction Method of Multipliers algorithm:
D ( r + 1 ) = arg min D k = 1 K F k D k A k F 2 + ρ | | D k S k ( r ) + T k ( r ) | | F 2 , S ( r + 1 ) = arg min S k = 1 K ρ | | D k ( r + 1 ) S k ( r ) + T k ( r ) | | F 2 , s . t . D = | | S i | | 2 2 1 , T ( r + 1 ) = T ( r ) + D k ( r + 1 ) S k ( r + 1 ) , update ρ if appropiate .
The proposed Dictionary Pair Learning (DPL) model exhibits a fast training process owing to the rapid convergence of closed-form solutions for variables A and P in each optimization step. The optimization of D is based on the Alternating Direction Method of Multipliers (ADMM), which also demonstrates quick convergence by stopping iterations when the energy difference between two consecutive iterations is below 0.01 . Upon convergence, the analysis dictionary P and synthesis dictionary D are obtained as outputs for classification. The objective functions presented in Equation (9) are designed to enhance the discriminative power of P while minimizing the reconstruction error. This balanced approach allowed the model to achieve both effective discrimination and strong representation capabilities.
During the classification stage, the residual values of samples within a class were utilized. The analysis sub-dictionary P k * produces small coefficients for samples that do not belong to class k, while the synthesis sub-dictionary D k ( * ) reconstructs the samples from class k. Consequently, the residual | | F k D k * P k * F k | | F 2 for a sample in class k is smaller than the residual | | F i D k * P k * F i | | F 2 for a sample not in class k. During the testing phase, the residual of an unknown query sample f t is computed for each class. The class associated with the minimum residual is assigned as the class for the testing sample. This testing process can be formulated as follows:
identity ( f t ) = arg min i f t D i P i f t 2 .
If the minimum residual in Equation (12) corresponds to class i, then sample f t is assigned to class i, where D i and P i represent the synthesis and analysis sub-dictionaries for that class, respectively.

2.2.3. The GA for the Parameter Optimization of the PDPL

The genetic algorithm (GA) is a widely used optimization technique inspired by the natural selection process in biology. It is commonly employed to find the optimal parameters for machine learning algorithms. Initially, the algorithm generates an initial population of potential parameter values, and each solution in the population is evaluated using a fitness function that assesses its performance on a given task. Solutions with higher fitness scores are given preference for selection in the next generation. Genetic operators, such as mutation and crossover, are then applied to the selected solutions to generate new offspring solutions. Mutation randomly alters some of the parameters in a solution, while crossover combines the parameters of two solutions to create a new one. The newly created offspring solutions undergo evaluation using the fitness function, and those with higher fitness scores are selected for the next generation. This process continues until a termination criterion is met, such as reaching a maximum number of iterations or convergence of the fitness function. Finally, the optimal parameter values for the given task are represented by the solution with the highest fitness score obtained at the end of the algorithm.
The parameter optimization algorithm flow of the GA-PDPL utilizes the Sheffield University genetic algorithm toolbox (gatbx).
  • Initialization: An initial population of solutions is generated by randomly assigning values to the parameters. The initialization parameters, which include the maximum genetic algebra, population size, crossover function, mutation probability, and t parameter of PDPL, are presented in Table 2. Furthermore, the GA optimization process involves tuning four PDPL parameters: m, τ , λ , and γ . The threshold ranges and coding methods for these parameters are provided in Table 3.
  • Evaluation: The fitness of each solution in the population is assessed using the projection dictionary pair learning (PDPL) algorithm and a fitness function. In this study, we defined the fitness function as the accuracy of the PDPL recognition on the test set. The calculation method for the fitness is illustrated in Figure 2 and can be found in Formula (14). To incorporate the research background, which was unrelated to the subjects, we introduced a leave-one-out subject cross test into the fitness calculation. The final fitness value was determined by averaging the accuracy across all subjects.
    F i t n e s s = 1 N i = 1 N Accuracy i .
  • Selection: Choose a subset of solutions to serve as parents for the next generation based on their fitness scores.
  • Crossover: Generate new solutions by combining the parameters of the selected parents through crossover.
  • Mutation: Introduce random changes to the parameters of some solutions to explore different areas of the search space.
  • Evaluation: Assess the fitness of the newly created solutions resulting from crossover and mutation.
  • Replacement: Select the top-performing solutions from both the previous and new generations to form the subsequent generation.
  • Termination: Stop the algorithm when a specified termination criterion is met, such as reaching the maximum number of generations or achieving the desired level of fitness.
  • Output: Provide the best solution obtained by the genetic algorithm (GA), which corresponds to the optimal parameter values for the projection dictionary pair learning algorithm.

3. Results and Discussion

The experimental test protocol followed the widely adopted leave-one-subject-out cross-validation (LOSOCV) procedure, which ensures robustness across subjects. We thoroughly analyze and discuss the obtained experimental results, focusing on aspects such as recognition accuracy, parameter variations, sex disparities, and a comparison with the state-of-the-art (SOTA) method.

3.1. Recognition Results of the GA-PDPL Method on Three Databases

Accuracy, an essential metric for evaluating the effectiveness of the proposed method, represents the ratio of the correctly classified samples to the total number of samples in a given test dataset. To demonstrate the efficacy of the proposed optimization method, we conducted a comparison of the effects of PDPL on three databases without parameter optimization. In this comparison, we utilized the default values for the four PDPL parameters, namely D i c t S i z e = 32 ,   τ = 0.03 ,   λ = 0.003 , and γ = 0.0001 . The recognition results are presented in Table 4. Across the four experimental conditions, the GA-PDPL consistently outperformed the PDPL in terms of accuracy. Particularly noteworthy was the significant improvement achieved by the GA parameter optimization on the SEED dataset, where the accuracy was 18.87% higher compared to the non-optimized version.
The 95% confidence interval is a valuable tool for assessing the reliability and uncertainty of recognition results in machine learning models. It provides a range within which we can be 95% confident that the actual result falls when conducting multiple experiments or sampling. In this study, we calculated the confidence intervals for all experimental accuracies, both with and without the GA for parameter adjustment. The results are presented in Table 5. Statistical analysis reveals that the upper and lower limits of the confidence interval for the proposed GA-PDPL method exceeded those of the PDPL method, thus demonstrating the effectiveness of our approach.
To enhance the visualization of the recognition results for each subject in the test set, we created bar graphs to display the outcomes for the three databases. Figure 3 represents the SEED database, Figure 4 corresponds to the MPED database, Figure 5 pertains to the GAMEEMO (two-class) dataset, and Figure 6 corresponds to the GAMEEMO (four-class) dataset. The GA-PDPL method demonstrates a substantial improvement in the recognition performance for most subjects across the four experiments.
The bar graph in Figure 3 illustrates that, with the exception of subjects 5 and 10, the accuracy rates for all other subjects surpassed 60%, with subject 11 achieving close to 100%. The optimization of the GA parameters resulted in a significant improvement in accuracy for all subjects except subject 14. These findings indicate that the GA enhanced the performance of the PDPL, enabling a better characterization of the EEG’s emotional characteristics.
Among the 23 subjects in the MPED database, as depicted in Figure 4, subject 14 achieved the highest accuracy, followed by subjects 18 and 23. The remaining subjects exhibited accuracy rates ranging from 20% to 30%. These considerable variations in individual performance can be attributed to several factors. Firstly, EEG signals are susceptible to various sources of noise interference, such as electromagnetic and power frequency disturbances. Additionally, there exist substantial individual differences among subjects, including disparities in emotional cognition and EEG responses. However, apart from subjects 8 and 22, the optimization of the GA parameters significantly enhanced the accuracy for all other subjects. This outcome demonstrates the efficacy of the GA in improving EEG emotion recognition through parameter optimization in a multi-classification setting.
The proposed method yielded a notable improvement in the accuracy rate of the latter subjects in both the two-class and four-class experiments of the GAMEEMO dataset. A comparison between the two-category and four-category experiments reveals that the two-category experiment achieved better recognition performance, while the four-category experiment demonstrated more substantial overall improvement. These findings are illustrated in Figure 5 and Figure 6.

3.2. Parameter Optimization Analysis of the GA-PDPL Method

The optimization procedure enhanced the performance of the Projection Dictionary Pair Learning (PDPL) algorithm through the utilization of Genetic Algorithms (GAs). Its objective was to iteratively refine the model’s parameters and feature selection process, thereby maximizing the classification performance. By employing the GA optimization algorithms and statistical techniques, the procedure aimed to find the optimal configuration that minimized the errors or maximized the performance metrics, resulting in improved classification outcomes. The optimization procedure offers several benefits. Firstly, it enables the fine-tuning of model parameters, optimizing their values to better align with the underlying data. This refinement process enhances the model’s ability to capture intricate patterns and relationships within the data, thereby improving the classification performance. Secondly, the optimization procedure facilitates feature selection or feature weighting, allowing for the identification of the most informative features for the classification task. By prioritizing the relevant features and reducing the influence of the irrelevant or redundant ones, the procedure enhances the discriminative power of the classifier. Additionally, the optimization procedure helps address the challenge of overfitting, which is common in classification tasks. By optimizing the regularization parameters or employing techniques such as cross validation, the procedure prevents the model from excessively memorizing the training data. Instead, it encourages the model to generalize well to unseen data, leading to an improved generalization performance and a reduced likelihood of erroneous classifications on new instances.
Figure 7, Figure 8, Figure 9 and Figure 10 illustrate the optimization results of the four parameters of the Genetic Algorithm (GA) utilized in the Projection Dictionary Pair Learning (PDPL) algorithm, as well as the corresponding change curves of the GA fitness function with increasing iterations on the three databases. In Figure 7, the variations in the number of iterations led to fluctuations in the m parameter, a decrease in the final value of τ , an increase in γ , a significant decrease in λ , and a gradual increase in the fitness function. Figure 8 exhibits a similar pattern to Figure 7 for each parameter and the fitness function, except for a substantial reduction in the τ parameter. The change curve of the MPED database demonstrated more pronounced changes compared to the SEED, potentially due to the larger amount of data available for MPED subjects, facilitating improved model learning.
The parameter changes in the SEED and MPED databases reveal that as the number of iterations increased, smaller values of τ and λ within their respective ranges and larger values of γ led to higher fitness levels. The m parameter underwent a process of feature selection, exhibiting a minimal significant increase or decrease. This adaptive process is beneficial for accommodating new test sets or subjects. Additionally, the growth trend of the fitness function can be observed through its change curve. With sufficient computing power, increasing the number of iterations allowed the fitness function to continue growing, resulting in a further improvement in the accuracy rates on each database.
Interestingly, for the same GAMEEMO dataset, different classification objectives yielded distinct trends and rules for the four parameters. Thus, increasing or decreasing specific parameters may have varying effects on different datasets. The change modes and combinations of the four parameters also differed across each dataset. Consequently, Genetic Algorithms (GA) provide an effective method for simultaneously optimizing these four parameters.

3.3. Emotion Recognition Performance of the GA-PDPL Method with Regard to Sex

It is widely acknowledged that emotional processing differs between men and women. Previous studies have provided evidence supporting this claim, indicating that women tend to display more authentic emotional expressions, while men exhibit greater control over their enthusiasm [36]. Furthermore, research has revealed that distinct brain networks are engaged by men and women when processing sad, depressed, and humorous audiovisual stimuli [37]. Women typically express their emotions through appearance and interpersonal interactions, whereas men tend to express their emotions through activities [38]. Although these studies contribute to our understanding of sex differences in emotional processing, they lack objective evidence and quantitative assessment [39]. Therefore, this study aimed to investigate sex differences in emotional processing using a more rigorous approach.
To address this objective, data were collected from two databases, and the recognition results of subjects of different sexes were analyzed. The mean and standard deviation values were calculated for accuracy, as depicted in Figure 11. The study’s findings revealed that the women exhibited higher average accuracy rates than the men in both databases, suggesting superior emotional cognition abilities. Additionally, the men displayed smaller standard deviations compared to the women, indicating greater emotional stability. These findings contribute objective evidence and provide a quantitative assessment of sex disparities in emotional processing. They support the notion that men and women process emotions differently, with women demonstrating higher emotional cognition abilities and men exhibiting greater emotional stability. These distinctions could be attributed to variations in brain networks and socialization processes.
The present study offers empirical support for the existence of sex differences in emotional processing. These findings emphasize the importance of considering sex when investigating emotional processing and have implications for enhancing communication and interpersonal relationships. Further research is warranted to explore the underlying mechanisms driving these sex disparities and their potential impact on mental health and wellbeing.

3.4. Training and Testing Time of the GA-PDPL

The genetic algorithm can be time-consuming when optimizing parameters, whereas the PDPL method offers the advantage of speed. In the context of the emotional brain–computer interface, the recognition time for the samples is a crucial factor. Thus, we recorded the training and testing time of the model on the test device (MATLAB 2019b, Intel(R) Core(TM) i5-9600KF CPU with 32.0 GB RAM). Considering the variations in sample size, stimuli, and categories across the three databases, we calculated the testing time for each sample, as presented in Table 6. Observing the results, although the time required for parameter optimization and model training was considerable, the testing time for each sample amounted to only a few thousandths of a second. This rapid calculation speed holds promising prospects for real-time detection of emotional changes in future real scenarios. In comparison to the PDPL method, the training time was longer, but the testing time was significantly reduced. Notably, the testing time for the GAMEEMO dataset was the shortest. This difference may stem from the dataset’s fewer signal channels, leading to a reduced number of features and consequently faster model recognition.

3.5. Comparison of the GA-PDPL Method and SOTA Method

We compared our proposed GA-PDPL method with state-of-the-art (SOTA) approaches in subject-independent EEG emotion recognition settings using the SEED, MPED, and GAMEEMO datasets. The results are presented in Table 7, Table 8, and Table 9, respectively. The compared algorithms were classic machine learning algorithms, which we applied to our experimental data and experimental results obtained under the same test protocol. Since the methods on the GAMEEMO dataset have not undergone subject validation, we included the classic method as a comparative approach in this experiment. From the tables, it can be concluded that our proposed method outperformed the current conventional methods in subject-independent protocols. Compared to the KLIEP [40], ULSIF [41], STM [42], SVM [43], KPCA [44], TCA [45], KNN [46], Random Forest [47], PDPL [27], and SA [43], our method utilized a comprehensive dictionary and an analysis dictionary to enhance the feature representation. Additionally, we employed a genetic algorithm (GA) for parameter optimization to select the optimal dictionary and parameters, thus achieving the best recognition performance. Furthermore, on the MPED database, our method outperformed two deep learning methods, DANN [48] and A-LSTM [16]. Deep learning methods exhibit great power, especially in parameter learning, but for small-sample datasets such as EEG emotions, deep learning models are prone to overfitting. To mitigate overfitting to some extent, our method reduced the dimensional space of the features to enhance the model training.

4. Conclusions

This paper presented a subject-independent EEG emotion recognition method employing genetically optimized projection dictionary pair learning. The experimental results demonstrated that the proposed method surpasses existing traditional machine learning algorithms. Moreover, our findings indicated superior recognition performance on female subjects. These outcomes hold implications for practical applications in emotional brain–computer interface devices. Nevertheless, this research had certain limitations. Due to computational constraints, we were unable to explore larger parameter settings within a limited timeframe, particularly in terms of the number of iterations, which resulted in a restricted optimization range. Future research should prioritize subject-independent and cross-database EEG emotional recognition. Such investigations have the potential to advance the development of emotional brain–computer interface systems applicable in real-world scenarios.

Author Contributions

Conceptualization, J.S. and H.C.; methodology, H.C.; formal analysis and investigation, T.S.; resources and data curation, J.Z.; writing, J.S. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the China Scholarship Council under Grant 202206090203.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xue, Y.; Zheng, W.; Zong, Y.; Chang, H.; Jiang, X. Adaptive Hierarchical Graph Convolutional Network for EEG Emotion Recognition. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–8. [Google Scholar]
  2. Chang, H.; Zong, Y.; Zheng, W.; Tang, C.; Zhu, J.; Li, X. Depression Assessment Method: An EEG Emotion Recognition Framework Based on Spatiotemporal Neural Network. Front. Psychiatry 2022, 12, 2620. [Google Scholar] [CrossRef] [PubMed]
  3. Al-Ezzi, A.; Kamel, N.; Faye, I.; Gunaseli, E. Review of EEG, ERP, and brain connectivity estimators as predictive biomarkers of social anxiety disorder. Front. Psychol. 2020, 11, 730. [Google Scholar] [CrossRef] [PubMed]
  4. Meyer, T.; Smeets, T.; Giesbrecht, T.; Quaedflieg, C.W.; Smulders, F.T.; Meijer, E.H.; Merckelbach, H.L. The role of frontal EEG asymmetry in post-traumatic stress disorder. Biol. Psychol. 2015, 108, 62–77. [Google Scholar] [CrossRef]
  5. Liu, B.; Chang, H.; Peng, K.; Wang, X. An End-to-End Depression Recognition Method Based on EEGNet. Front. Psychiatry 2022, 13, 864393. [Google Scholar] [CrossRef] [PubMed]
  6. Han, Z.; Chang, H.; Zhou, X.; Wang, J.; Wang, L.; Shao, Y. E2ENNet: An end-to-end neural network for emotional brain-computer interface. Front. Comput. Neurosci. 2022, 16, 942979. [Google Scholar] [CrossRef]
  7. Li, X.; Zheng, W.; Zong, Y.; Chang, H.; Lu, C. Attention-based Spatio-Temporal graphic LSTM for EEG emotion recognition. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
  8. Abdulrahman, A.; Baykara, M. A Comprehensive Review for Emotion Detection Based on EEG Signals: Challenges, Applications, and Open Issues. Trait. Signal 2021, 38, 1189–1200. [Google Scholar] [CrossRef]
  9. Murugappan, M.; Rizon, M.; Nagarajan, R.; Yaacob, S.; Hazry, D.; Zunaidi, I. Time-frequency analysis of EEG signals for human emotion detection. In Proceedings of the 4th Kuala Lumpur International Conference on Biomedical Engineering 2008: BIOMED 2008, Kuala Lumpur, Malaysia, 25–28 June 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 262–265. [Google Scholar]
  10. Murugappan, M.; Murugappan, S. Human emotion recognition through short time Electroencephalogram (EEG) signals using Fast Fourier Transform (FFT). In Proceedings of the 2013 IEEE 9th International Colloquium on Signal Processing and Its Applications, Kuala Lumpur, Malaysia, 8–10 March 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 289–294. [Google Scholar]
  11. Dongwei, C.; Fang, W.; Zhen, W.; Haifang, L.; Junjie, C. EEG-based emotion recognition with brain network using independent components analysis and granger causality. In Proceedings of the 2013 International Conference on Computer Medical Applications (ICCMA), Sousse, Tunisia, 20–22 January 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–6. [Google Scholar]
  12. Wang, X.W.; Nie, D.; Lu, B.L. EEG-based emotion recognition using frequency domain features and support vector machines. In Proceedings of the Neural Information Processing: 18th International Conference, ICONIP 2011, Shanghai, China, 13–17 November 2011; Springer: Berlin/Heidelberg, Germany, 2011. Part I 18. pp. 734–743. [Google Scholar]
  13. Matiko, J.W.; Beeby, S.P.; Tudor, J. Fuzzy logic based emotion classification. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 4389–4393. [Google Scholar]
  14. Fu, R.; Wang, H.; Zhao, W. Dynamic driver fatigue detection using hidden Markov model in real driving condition. Expert Syst. Appl. 2016, 63, 397–411. [Google Scholar] [CrossRef]
  15. Zhang, J.; Yin, Z.; Chen, P.; Nichele, S. Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review. Inf. Fusion 2020, 59, 103–126. [Google Scholar] [CrossRef]
  16. Song, T.; Zheng, W.; Lu, C.; Zong, Y.; Zhang, X.; Cui, Z. MPED: A multi-modal physiological emotion database for discrete emotion recognition. IEEE Access 2019, 7, 12177–12191. [Google Scholar] [CrossRef]
  17. Bazgir, O.; Mohammadi, Z.; Habibi, S.A.H. Emotion recognition with machine learning using EEG signals. In Proceedings of the 2018 25th National and 3rd International Iranian Conference on Biomedical Engineering (ICBME), Qom, Iran, 29–30 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–5. [Google Scholar]
  18. Islam, M.R.; Moni, M.A.; Islam, M.M.; Rashed-Al-Mahfuz, M.; Islam, M.S.; Hasan, M.K.; Hossain, M.S.; Ahmad, M.; Uddin, S.; Azad, A.; et al. Emotion recognition from EEG signal focusing on deep learning and shallow learning techniques. IEEE Access 2021, 9, 94601–94624. [Google Scholar] [CrossRef]
  19. Zhang, J.; Li, C.; Kosov, S.; Grzegorzek, M.; Shirahama, K.; Jiang, T.; Sun, C.; Li, Z.; Li, H. LCU-Net: A novel low-cost U-Net for environmental microorganism image segmentation. Pattern Recognit. 2021, 115, 107885. [Google Scholar] [CrossRef]
  20. Abdulrahman, A.; Baykara, M.; Alakus, T.B. A Novel Approach for Emotion Recognition Based on EEG Signal Using Deep Learning. Appl. Sci. 2022, 12, 10028. [Google Scholar] [CrossRef]
  21. Abdulrahman, A.; Baykara, M. Feature extraction approach based on statistical methods and wavelet packet decomposition for emotion recognition using EEG signals. In Proceedings of the 2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–7. [Google Scholar]
  22. Li, J.; Qiu, S.; Shen, Y.Y.; Liu, C.L.; He, H. Multisource transfer learning for cross-subject EEG emotion recognition. IEEE Trans. Cybern. 2019, 50, 3281–3293. [Google Scholar] [CrossRef]
  23. Li, W.; Huan, W.; Hou, B.; Tian, Y.; Zhang, Z.; Song, A. Can emotion be transferred?—A review on transfer learning for EEG-Based Emotion Recognition. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 833–846. [Google Scholar] [CrossRef]
  24. Das, A.; Mondal, P.; Pal, U.; Ferrer, M.A.; Blumenstein, M. Fast and efficent multimodal eye biometrics using projective dictionary pair learning. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1402–1408. [Google Scholar]
  25. Ameri, R.; Alameer, A.; Ferdowsi, S.; Nazarpour, K.; Abolghasemi, V. Labeled projective dictionary pair learning: Application to handwritten numbers recognition. Inf. Sci. 2022, 609, 489–506. [Google Scholar] [CrossRef]
  26. Cai, W.; Gao, M.; Jiang, Y.; Gu, X.; Ning, X.; Qian, P.; Ni, T. Hierarchical domain adaptation projective dictionary pair learning model for EEG classification in IoMT systems. IEEE Trans. Comput. Soc. Syst. 2022. [Google Scholar] [CrossRef]
  27. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Projective dictionary pair learning for pattern classification. Adv. Neural Inf. Process. Syst. 2014, 27, 1–9. [Google Scholar]
  28. Sehgal, A.; La, H.; Louis, S.; Nguyen, H. Deep reinforcement learning using genetic algorithm for parameter optimization. In Proceedings of the 2019 Third IEEE International Conference on Robotic Computing (IRC), Naples, Italy, 25–27 February 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 596–601. [Google Scholar]
  29. Dilip, D.G.; Panda, S.; Mathew, J. Characterization and parametric optimization of micro-hole surfaces in micro-EDM drilling on Inconel 718 superalloy using genetic algorithm. Arab. J. Sci. Eng. 2020, 45, 5057–5074. [Google Scholar] [CrossRef]
  30. Syarif, I.; Prugel-Bennett, A.; Wills, G. SVM parameter optimization using grid search and genetic algorithm to improve classification performance. TELKOMNIKA (Telecommun. Comput. Electron. Control) 2016, 14, 1502–1509. [Google Scholar] [CrossRef]
  31. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  32. Katoch, S.; Chauhan, S.S.; Kumar, V. A review on genetic algorithm: Past, present, and future. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef] [PubMed]
  33. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 81–84. [Google Scholar]
  34. Alakus, T.B.; Gonen, M.; Turkoglu, I. Database for an emotion recognition system based on EEG signals and various computer games–GAMEEMO. Biomed. Signal Process. Control 2020, 60, 101951. [Google Scholar] [CrossRef]
  35. Sani, O.G.; Yang, Y.; Lee, M.B.; Dawes, H.E.; Chang, E.F.; Shanechi, M.M. Mood variations decoded from multi-site intracranial human brain activity. Nat. Biotechnol. 2018, 36, 954–961. [Google Scholar] [CrossRef]
  36. Deng, Y.; Chang, L.; Yang, M.; Huo, M.; Zhou, R. Gender differences in emotional response: Inconsistency between experience and expressivity. PLoS ONE 2016, 11, e0158666. [Google Scholar] [CrossRef] [Green Version]
  37. Goshvarpour, A.; Goshvarpour, A. EEG spectral powers and source localization in depressing, sad, and fun music videos focusing on gender differences. Cogn. Neurodyn. 2019, 13, 161–173. [Google Scholar] [CrossRef] [PubMed]
  38. Kensinger, E.A. Remembering emotional experiences: The contribution of valence and arousal. Rev. Neurosci. 2004, 15, 241–252. [Google Scholar] [CrossRef] [PubMed]
  39. Bao, L.Q.; Qiu, J.L.; Tang, H.; Zheng, W.L.; Lu, B.L. Investigating sex differences in classification of five emotions from EEG and eye movement signals. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 6746–6749. [Google Scholar]
  40. Sugiyama, M.; Nakajima, S.; Kashima, H.; Von Buenau, P.; Kawanabe, M. Direct importance estimation with model selection and its application to covariate shift adaptation. In Advances in Neural Information Processing Systems 20 (NIPS 2007); Citeseer: Princeton, NJ, USA, 2007; Volume 7, pp. 1433–1440. [Google Scholar]
  41. Kanamori, T.; Hido, S.; Sugiyama, M. A least-squares approach to direct importance estimation. J. Mach. Learn. Res. 2009, 10, 1391–1445. [Google Scholar]
  42. Chu, W.S.; De la Torre, F.; Cohn, J.F. Selective transfer machine for personalized facial expression analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 529–545. [Google Scholar] [CrossRef]
  43. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  44. Schölkopf, B.; Smola, A.; Müller, K.R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998, 10, 1299–1319. [Google Scholar] [CrossRef] [Green Version]
  45. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2010, 22, 199–210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Cai, H.; Han, J.; Chen, Y.; Sha, X.; Wang, Z.; Hu, B.; Yang, J.; Feng, L.; Ding, Z.; Chen, Y.; et al. A pervasive approach to EEG-based depression detection. Complexity 2018, 2018, 5238028. [Google Scholar] [CrossRef] [Green Version]
  47. Rigatti, S.J. Random forest. J. Insur. Med. 2017, 47, 31–39. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2096-2030. [Google Scholar]
  49. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2066–2073. [Google Scholar]
Figure 1. Framework of the GA-PDPL algorithm for EEG emotion recognition.
Figure 1. Framework of the GA-PDPL algorithm for EEG emotion recognition.
Brainsci 13 00977 g001
Figure 2. Leave-one-subject-out cross validation and fitness.
Figure 2. Leave-one-subject-out cross validation and fitness.
Brainsci 13 00977 g002
Figure 3. Accuracy (%) of each subject in the SEED database.
Figure 3. Accuracy (%) of each subject in the SEED database.
Brainsci 13 00977 g003
Figure 4. Accuracy (%) of each subject in the MPED dataset.
Figure 4. Accuracy (%) of each subject in the MPED dataset.
Brainsci 13 00977 g004
Figure 5. Accuracy (%) of each subject in the GAMEEMO (two-class) dataset.
Figure 5. Accuracy (%) of each subject in the GAMEEMO (two-class) dataset.
Brainsci 13 00977 g005
Figure 6. Accuracy (%) of each subject in the GAMEEMO (four-class) dataset.
Figure 6. Accuracy (%) of each subject in the GAMEEMO (four-class) dataset.
Brainsci 13 00977 g006
Figure 7. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the SEED dataset. Here, ’Dict’ represents the m parameter. The first four line charts are the optimized values of the PDPL parameter optimization curve based on the GA at different iterations. The last line graph is the change curve of the fitness function as the number of iterations increases.
Figure 7. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the SEED dataset. Here, ’Dict’ represents the m parameter. The first four line charts are the optimized values of the PDPL parameter optimization curve based on the GA at different iterations. The last line graph is the change curve of the fitness function as the number of iterations increases.
Brainsci 13 00977 g007
Figure 8. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the MPED dataset. Here, ’Dict’ represents the m parameter.
Figure 8. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the MPED dataset. Here, ’Dict’ represents the m parameter.
Brainsci 13 00977 g008
Figure 9. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the GAMEEMO dataset (two—class). Here, ’Dict’ represents the m parameter.
Figure 9. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the GAMEEMO dataset (two—class). Here, ’Dict’ represents the m parameter.
Brainsci 13 00977 g009
Figure 10. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the GAMEEMO dataset (four—class). Here, ’Dict’ represents the m parameter.
Figure 10. Change curves of the four parameters (m, τ , λ , and γ ) and the fitness function on the GAMEEMO dataset (four—class). Here, ’Dict’ represents the m parameter.
Brainsci 13 00977 g010
Figure 11. The mean and standard deviation of the accuracy for subjects of different sexes on the two databases.
Figure 11. The mean and standard deviation of the accuracy for subjects of different sexes on the two databases.
Brainsci 13 00977 g011
Table 1. The stimuli used in the GAMEEMO dataset.
Table 1. The stimuli used in the GAMEEMO dataset.
Game NameStimuli TypePositive–NegativeArousal–Valence
G1BoringNegativeLANV
G2CalmPositiveLAPV
G3HorrorNegativeHANV
G4FunnyPositiveHAPV
HAPV: high arousal–positive valence; HANV: high arousal–negative valence; LANV: low arousal–negative valence; LAPV: low arousal–positive valence.
Table 2. Initialization parameter settings of the GA.
Table 2. Initialization parameter settings of the GA.
ParameterValue
Maximum generation50
Size of population20
Selection functionStochastic Universal Sampling
Rate of individuals to be selected0.9
Mutation probability0.7
Table 3. Matrix describing the length and how to decode each substring in the chromosome.
Table 3. Matrix describing the length and how to decode each substring in the chromosome.
FieldD  1 m τ λ γ
len  2 9999
lb  3 1000
ub  3 310/70  * 0.10.010.001
code  4 graygraygraygray
scale  5 arithmeticarithmeticarithmeticarithmetic
lbin  6 0011
ubin  6 1111
1 FieldD —Matrix describing the length and how to decode each substring in the chromosome. 2 len—row vector containing the length of each substring in the chromosome. sum(len) should equal the individual length. 3 lb, ub—Lower and upper bounds for each variable. 4 code—binary row vector indicating how each substring is to be decoded. 5 scale—binary row vector indicating where to use arithmetic and/or logarithmic scaling. 6 lbin, ubin—binary row vectors indicating whether or not to include each bound in the representation range. * 310 for SEED Dataset and MPED Dataset, 70 for GAMEEMO Dataset.
Table 4. Mean accuracies (ACC) and standard deviation (STD) of all experiments.
Table 4. Mean accuracies (ACC) and standard deviation (STD) of all experiments.
MethodPDPL ACC/STD (%)GA-PDPL ACC/STD (%)
SEED51.02/13.5769.89/14.39
MPED21.39/5.4124.87/5.83
GAMEEMO (two-class)61.76/3.9964.34/6.44
GAMEEMO (four-class)39.92/8.2849.01/8.46
Table 5. The 95% confidence interval for the accuracy of all the experiments.
Table 5. The 95% confidence interval for the accuracy of all the experiments.
MethodPDPLGA-PDPL
SEED[43.51, 58.54][59.52, 76.78]
MPED[19.05, 23.73][22.35, 27.39]
GAMEEMO (2-class)[60.21, 63.31][61.84, 66.84]
GAMEEMO (4-class)[36.71, 43.13][45.72, 52.29]
Table 6. The training and testing time of all the experiments.
Table 6. The training and testing time of all the experiments.
MethodPDPLGA-PDPL
Time (s)Training TimeTesting TimeTraining TimeTesting Time
SEED5.63300.020876,9020.001
MPED12.69850.0625335,3860.005
GAMEEMO (2-class)1.57340.001215,5750.0005
GAMEEMO (4-class)1.76600.003417,8920.0004
Table 7. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the SEED dataset (N = 15).
Table 7. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the SEED dataset (N = 15).
MethodM ± SEM (%)
KLIEP [40] *45.17 ± 4.59
PDPL [27] *51.02 ± 3.50
ULSIF [41] *51.18 ± 3.50
STM [42] *51.23 ± 3.83
SVM [43] *56.73 ± 4.21
KPCA [44] *61.28 ± 3.77
TCA [45] *63.64 ± 3.84
SA [43] *69.00 ± 2.81
GA-PDPL (ours )69.89 ± 3.72
* indicates the experiment results obtained by our own implementation. Bold indicates best result.
Table 8. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the MPED dataset (N = 23).
Table 8. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the MPED dataset (N = 23).
MethodM ± SEM (%)
KLIEP [40] *18.92 ± 0.95
ULSIF [41] *19.63 ± 0.79
TCA [45] *19.50 ± 0.75
SVM [43] *19.66 ± 0.83
GFK [49] *20.27 ± 0.91
SA [43] *20.74 ± 0.87
STM [42] *20.89 ± 0.75
PDPL [27] *21.39 ± 1.13
DANN [48]22.36 ± 0.91
A-LSTM [16]24.06 ± 0.96
GA-PDPL (ours )24.87 ± 1.22
* indicates the experiment results obtained by our own implementation. Bold indicates best result.
Table 9. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the GAMEEMO dataset (N = 28).
Table 9. Mean accuracies (M) and standard errors of the mean (SEM) of the subject-independent experiment on the GAMEEMO dataset (N = 28).
MethodTwo-Class M ± SEM (%)Four-Class M ± SEM (%)
KNN [46] *58.16 ± 1.4535.46 ± 2.06
Random Forest [47] *59.29 ± 2.1238.27 ± 2.95
PDPL [27] *61.76 ± 0.7539.92 ± 1.56
SVM [43] *63.17 ± 1.2746.62 ± 1.89
GA-PDPL(ours )64.34 ± 1.5649.01 ± 1.60
* indicates the experiment results obtained by our own implementation. Bold indicates best result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Su, J.; Zhu, J.; Song, T.; Chang, H. Subject-Independent EEG Emotion Recognition Based on Genetically Optimized Projection Dictionary Pair Learning. Brain Sci. 2023, 13, 977. https://doi.org/10.3390/brainsci13070977

AMA Style

Su J, Zhu J, Song T, Chang H. Subject-Independent EEG Emotion Recognition Based on Genetically Optimized Projection Dictionary Pair Learning. Brain Sciences. 2023; 13(7):977. https://doi.org/10.3390/brainsci13070977

Chicago/Turabian Style

Su, Jipu, Jie Zhu, Tiecheng Song, and Hongli Chang. 2023. "Subject-Independent EEG Emotion Recognition Based on Genetically Optimized Projection Dictionary Pair Learning" Brain Sciences 13, no. 7: 977. https://doi.org/10.3390/brainsci13070977

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop