Next Article in Journal
Diagnosis of Low-Grade Urothelial Neoplasm in the Era of the Second Edition of the Paris System for Reporting Urinary Cytology
Next Article in Special Issue
Brain–Computer Interface: The HOL–SSA Decomposition and Two-Phase Classification on the HGD EEG Data
Previous Article in Journal
Radiomics Applications in Spleen Imaging: A Systematic Review and Methodological Quality Assessment
Previous Article in Special Issue
PPG Signals-Based Blood-Pressure Estimation Using Grid Search in Hyperparameter Optimization of CNN–LSTM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing 1D-CNN-Based Emotion Recognition Process through Channel and Feature Selection from EEG Signals

1
Department of Computer Science, College of Arts and Science, Prince Sattam bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
2
Center of Excellence in Information Assurance (CoEIA), King Saud University, Riyadh 11543, Saudi Arabia
3
College of Computer & Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(16), 2624; https://doi.org/10.3390/diagnostics13162624
Submission received: 14 June 2023 / Revised: 28 July 2023 / Accepted: 29 July 2023 / Published: 8 August 2023
(This article belongs to the Special Issue Biomedical Signal Processing and Analysis)

Abstract

:
EEG-based emotion recognition has numerous real-world applications in fields such as affective computing, human-computer interaction, and mental health monitoring. This offers the potential for developing IOT-based, emotion-aware systems and personalized interventions using real-time EEG data. This study focused on unique EEG channel selection and feature selection methods to remove unnecessary data from high-quality features. This helped improve the overall efficiency of a deep learning model in terms of memory, time, and accuracy. Moreover, this work utilized a lightweight deep learning method, specifically one-dimensional convolutional neural networks (1D-CNN), to analyze EEG signals and classify emotional states. By capturing intricate patterns and relationships within the data, the 1D-CNN model accurately distinguished between emotional states (HV/LV and HA/LA). Moreover, an efficient method for data augmentation was used to increase the sample size and observe the performance deep learning model using additional data. The study conducted EEG-based emotion recognition tests on SEED, DEAP, and MAHNOB-HCI datasets. Consequently, this approach achieved mean accuracies of 97.6, 95.3, and 89.0 on MAHNOB-HCI, SEED, and DEAP datasets, respectively. The results have demonstrated significant potential for the implementation of a cost-effective IoT device to collect EEG signals, thereby enhancing the feasibility and applicability of the data.

1. Introduction

Emotion is defined as the reaction to or consciousness of external stimuli, playing an important role in daily life by affecting people’s routines. Basic emotions like happiness, anger, and sadness are constantly reflected by people voluntarily or involuntarily, significantly impacting their position in society. Negative emotions can lead to social exclusion, causing physiological and psychological effects [1]. Positive emotions relate to better living standards and a longer life [2]. Many studies have analyzed emotions, with recent ones focusing on understanding emotional behaviors [3,4]. However, the abstract and individualized nature of emotions makes this research challenging and limits progress [5]. Moreover, the multitude of data collection methods (facial, vocal, neural, bodily signals, etc.) and analysis techniques creates complex, time-consuming data processing [4]. Therefore, computer-aided AI-based analysis methods are needed. In summary, emotion analysis is critical but faces difficulties due to the inherent complexity of emotions. Advanced computational techniques are thus necessary to enable effective emotion research.
EEG-based emotion recognition has wide applications but can identify hidden internal states beyond external expressions. EEG directly measures brain signals using scalp electrodes, capturing real-time data on unspoken emotions. This enables intuitive responses to users’ psychological states in diverse systems. Therefore, this study has utilized EEG data to recognize emotions. The process of emotion recognition typically involves three main steps: preprocessing, feature extraction, and classification. Pre-processing procedures are essential for data preparation prior to further analysis. Feature extraction can be achieved using traditional machine learning or deep learning methods. Traditional methods often involve the use of handcrafted features [6]. However, owing to the high non-stationarity of EEG signals, extracting such features is challenging and requires expert knowledge. Recent studies have explored multivariate statistical analysis techniques in the frequency, time-frequency, and nonlinear domains that effectively represent EEG characteristics [7,8,9]. For instance, Zheng et al. [7] evaluated different feature extraction methods and achieved the best average accuracy using a discriminative graph-regularized Extreme Learning Machine with differential entropy features. Zhu et al. [8] constructed a graph structure based on differential entropy characteristics, learned channel relationships through dynamic simplifying graph convolutional networks, and recalibrated channel features using the style recalibration module. Through the fusion and classification of sub-band features, the method achieves improved classification accuracy compared to existing methods. Wang et al. [9] proposed an end-to-end convolutional neural network (CNN) model called SACNN to improve cross-subject emotion recognition accuracy. By restructuring the data dimensions and selecting the top 10 channels, the SACNN model achieved significant accuracy in detecting human emotions. Algarni et al. [10] proposed a stacked bi-directional Long Short-Term Memory (Bi-LSTM) Model for EEG-based emotion recognition that includes Binary Gray Wolf Optimizer for feature selection. By achieving high accuracy in classifying arousal, valence, and liking from the DEAP dataset, the proposed method demonstrated effective emotion recognition through EEG signals. Li et al. [11] proposed a C-RNN model combining CNN and RNN to identify emotions using multichannel EEG signals. While RNN-based methods have advantages for processing time series data and have achieved great results, they still have limitations when dealing with multichannel EEG data. Gao et al. [12] proposed a gradient particle swarm optimization (GPSO) model to automatically optimize CNN models. They implemented the GPSO algorithm to optimize the hyperparameters and architecture of CNNs. The experimental results demonstrated that CNN models optimized by their proposed GPSO approach achieved significantly higher classification accuracy compared to other CNNs. Hancer and Subasi [13] proposed a unique EEG emotion recognition framework including multi-scale PCA and wavelet filtering for preprocessing. For feature extraction, they used dual tree complex wavelet transform. For feature selection, they employed statistical criteria to reduce dimensions. Consequently, researchers have been utilizing various neural networks for EEG data analysis [14,15,16] with the aim of improving accuracy.
EEG-based emotion recognition tasks using deep learning models face a significant challenge due to the limited availability of EEG training datasets compared to visual and audio datasets. Currently, only a few public datasets, including MAHNOB-HCI [17], DEAP [18], and SEED [19], are accessible for EEG-based emotion recognition, and their scale is small to train a deep learning model efficiently [20]. To overcome this limitation, data augmentation techniques have been employed to generate additional training data. These techniques involve applying geometric modifications or adding Gaussian noise to the original EEG data [16]. However, a recent approach by Kalaganis et al. [21] executes a graph variant of the classical Empirical Mode Decomposition (EMD) to produce realistic EEG-like data. The generated data, combined with the original data, is then utilized to train a 1D-CNN model for classification and an Arousal-Valence model to identify emotions from complex and nonstationary EEG data. Experimental results have demonstrated that data augmentation significantly enhances the accuracy of classifiers in EEG-based emotion recognition.
The cerebral cortex has four lobes: frontal, parietal, temporal, and occipital. EEG devices vary in electrode number, categorized as low (1–32 electrodes), medium (33–128), or high (>128) resolution [22]. The international 10–20 system [23] standardizes electrode placement by mapping locations to scalp regions. Selecting optimal EEG channels for emotion recognition from key brain areas can reduce the electrodes needed, improving wearable headset comfort. Determining these optimal channels has been a focus for enabling real-world EEG applications. The frontal, parietal, temporal, and occipital lobes each contribute unique signals relevant for emotion recognition. By exploring the most informative regions and channels, this work aims to provide insights on channel selection to reduce computational demands while preserving emotion-relevant information from EEG signals. The goal is to improve the feasibility of EEG-based emotion recognition.
Previous research has recognized human emotions using a limited number of EEG channels. While this can increase computational speed, it decreases the accuracy of emotion recognition. In this paper, we introduced a differential entropy-based approach for channel selection that not only enhances the model’s performance but also reduces the reliance on using all EEG electrodes for signal extraction by selecting the most relevant channels. Moreover, challenges remain in computational efficiency, robust feature extraction, and optimal model architectures. Continued research is needed to develop end-to-end systems that leverage different deep networks’ complementary strengths with a fewer number of channel EEG analyses. Therefore, the paper utilizes a Decomposed Mutual Information Maximization (DMIM) method for sequential feature selection, which eliminates unwanted data from the selected channels to further improve the model’s performance. Furthermore, it proposes a lightweight 1D-CNN architecture based on deep learning for the classification of high-quality data obtained from selected EEG channels. Lastly, the paper incorporates a graph-EMD to generate realistic EEG-like data, enhancing the training dataset.
The remaining sections of the paper are organized as follows: Section 2 provides a detailed description of the research methodology, specifically the presented deep learning model. In Section 3, the results of the analysis are presented. Finally, Section 4 and 5 offer a discussion of the work and conclude the paper, respectively.

2. Materials and Methods

In this section, we provide a comprehensive overview of the MAHNOB-HCI, SEED, and DEAP datasets, along with the pre-processing procedures employed. Following that, we delve into the detailed introduction of our model. Lastly, we elucidate the training parameters utilized in our model for a comprehensive understanding of the training process. Figure 1 illustrates the process of proposed framework.

2.1. Dataset and Data Processing

The MAHNOB-HCI database [17] contains EEG and physiological signals recorded from 30 participants (data from 3 subjects was lost) watching 20 emotional video clips. The 17 women and 13 men healthy adult participants were 19–40 years old. These participants watched 20 emotional video clips chosen to elicit responses of disgust, amusement, fear, sadness, and joy. The clips were selected from movies and online videos. The clips ranged from 34 to 117 s, with a 30 s baseline before/after each clip. EEG data were recorded using caps with 32 electrode sensors and sampling rate of 512 Hz, which was then down-sampled to 256 Hz. After each video, participants self-reported their arousal and valence using the Self-Assessment Manikin (SAM) scale. Arousal measures the degree of excitement, rated from 1 (boring) to 9 (exciting), and valence indicates the polarity of emotion, rated from 1 (unpleasant) to 9 (pleasant) [24]. Therefore, the dataset includes subjective arousal and valence ratings from 1 to 9 for each video based on the participants’ emotional responses. These self-reported scores provide ground truth labels to relate the recorded EEG and physiological signals to emotional states defined along the dimensions of arousal and valence. In total, the database includes EEG and peripheral signals for emotion analysis from 27 subjects responding to 20 videos.
The DEAP dataset [18] contains EEG recordings from 32 participants (16 male, 16 female) as they watched 40 one-minute musical video clips. Each trial included a 3 s pre-trial baseline followed by 60 s of video stimulation. EEG data were recorded using caps with 32 electrode sensors and sampling rate of 512 Hz. The data were preprocessed by the DEAP team, including down-sampling the EEG signal to 128 Hz. After each video, participants self-reported their arousal valence, like/dislike, dominance, and familiarity levels using the SAM scale. Each subject underwent 40 trials, with various signals recorded as 40-channel data. The first 32 channels represented EEG signals, while the remaining 8 channels captured autonomous physiological signals. After watching each video, participants rated the levels of arousal, valence, like/dislike, dominance, and familiarity using SAM.
The SEED dataset [19] consists of EEG signal data obtained during an emotional experiment. It was collected from 15 subjects, comprising 7 males and 8 females. Sixty-two-channel EEG data were collected as participants viewed 15 four-minute Chinese film excerpts designed to elicit positive, neutral, and negative emotions (5 clips per emotion). During the experiments, subjects received a 5-s prompt before each clip, then completed a 45-s self-assessment of their emotions followed by 15 s of rest. The dataset focused on three specific emotions: positive, neutral, and negative, resulting in a total of 45 trials. Each trial involved recording the scalp EEG signals using a 62-channel standard 10–20 system at a sampling rate of 200 Hz, which was then down-sampled to 128 Hz. To eliminate physiological noises, a 3rd-order bandpass Butterworth filter with cutoff frequencies of 0 Hz to 75 Hz was applied. A comparison between the various characteristics of all the datasets under consideration has been depicted in Table 1.
For this work, the valence and arousal scales were used for classification. The SAM rating scale was divided into two parts, high (level > 5) and low (level ≤ 5), to establish low valence (LV), high valence (HV), low arousal (LA), and high arousal (HA) labels for MAHNOB-HCI and DEAP datasets. However, for SEED dataset, three classes (positive, neutral, and negative) were considered for classification.

2.2. Data Augmentation

In recent years, there has been growing scientific interest in applying data augmentation for EEG classification. George et al. [25] performed data augmentation for EEG signals using six different synthesis approaches. The results showed that the synthesized EEG data exhibited similar characteristics to real data. Using the augmented data for training increased classification accuracy by up to 3% and 12% on two public EEG datasets. Luo et al. [26] used conditional Wasserstein GAN (cWGAN) and selective VAE (sVAE) for data augmentation to improve emotion recognition from EEG signals. They transformed the training data into power spectral density and differential entropy representations before feeding them into the cWGAN and sVAE models. Their proposed augmentation technique helped in improving classification performance for EEG-based emotion recognition. Zhang et al. [27] proposed a new EEG data augmentation method using empirical mode decomposition (EMD). In their approach, EEG signals are decomposed into intrinsic mode functions (IMFs) using EMD. The IMFs are then recombined in different ways to artificially generate new training data. This augmented data are transformed into wavelet tensors and used to train a convolutional neural network (CNN) classifier. Experiments showed that their EMD-based data augmentation technique improved EEG classification accuracy compared to training without data augmentation. Abdelfattah et al. [28] proposed recurrent generative adversarial networks (RGANs) to generate synthetic EEG. The recurrent generator architecture allowed the RGANs to successfully capture the key temporal dynamics of EEG signals, giving them an advantage over standard GANs for EEG data generation. To address limited dataset sizes for EEG-based emotion recognition, this study explored using graph empirical mode decomposition (EMD) for augmenting EEG signals to improve classification.
The graph variant of the classical EMD method allows the generation of artificial EEG epochs using an arbitrary number of input epochs. To generate these artificial epochs, each individual EEG epoch is initially decomposed into a set of graph Intrinsic Mode Functions (IMFs) based on a graph structure G . By combining the graph IMFs from different epochs, an artificial EEG epoch can be created. This approach leverages the mono-component nature of the graph IMFs, resulting in artificial epochs that share similar characteristics with the original signals that contributed to their IMFs.
To enhance the capabilities of a classifier, the class information associated with each EEG epoch is considered during the generation of artificial epochs. Consequently, the corresponding IMFs of each epoch are also assigned the same class label, ensuring that each artificial EEG epoch is generated exclusively from graph IMFs belonging to a single class. This approach enables the creation of artificial EEG epochs that align with specific class characteristics, further enhancing the discriminative capabilities of the classifier.
The proposed data augmentation method as discussed by Kalaganis et al. [21] involves the following steps to generate artificial EEG epochs:
  • Random selection of class-specific EEG epochs: For each class, a random set of EEG epochs is chosen to contribute their Intrinsic Mode Functions (IMFs). The number of contributing EEG epochs is determined by the maximum number of IMFs present in a signal segment.
  • IMF selection for artificial EEG epoch generation: To create an artificial EEG epoch, the IMFs are selected in a sequential manner. The first IMF is taken from the first contributing EEG epoch, the second IMF from the second contributing EEG epoch, and so on. If a contributing EEG epoch has fewer IMFs than required, the additional IMFs are considered as zero graph signals. The required maximum number of IMFs is set to five.
By following this procedure, a substantial number of artificial EEG epochs can be generated, with the potential to reach the total number of EEG epochs raised to the power of the number of graph IMFs. This augmentation process effectively expanded the dataset, providing a greater variety of data samples for analysis and training purposes.

2.3. Proposed Architecture

2.3.1. Channel Selection

According to neuroscience, the brain waves are generated in specific regions of the brain. In particular, the pre-frontal, parietal, and temporal areas are associated with the detection of positive valence emotions [29]. Several recent studies have explored EEG channel selection for emotion recognition. Affes et al. [30] proposed a cascading deep learning model called CAtt-MLP for channel selection, consisting of convolutional blocks, an attention neural network, and a multi-layer perceptron (MLP). Ramadhani et al. [31] applied integrated selection to remove irrelevant channels, improving brain–computer interface performance. Dura and Wosiak [32] used a reversed correlation algorithm (RCA) to automatically select optimal EEG electrodes. RCA identified emotion-relevant band-channel combinations through inter-subject analysis. Daoud and Bayoumi [33] used transfer learning and semi-supervised approaches to identify relevant channels. According to previous research [34,35], the most commonly used electrodes for emotion recognition are O1, O2, FP1, FP2, F3, F4, C3, C4, AF3, AF4, FC5, FC6, T7, T8, F7, F8, P7, and P8. Moreover, the frontal lobe electrodes like F3, F4, and AF3/4 have been used in more than 70% of the studies, which shows that many researchers believe the frontal cortex plays an important role in emotion processing. Across these works, channel selection consistently improved accuracy compared to using all channels. By removing less informative channels, channel selection reduces complexity and noise while focusing on emotion-relevant EEG features.
In this work, we have employed the concept of differential entropy to select the channels. By exploiting this measure, we can identify and incorporate the relevant channels in this work. Formally, let (U, CD) be an informed decision, where U is the set of all EEG data samples under consideration, categorized into two classes C and D. For a particular EEG sample P belonging to class C (PC), the differential entropy E of P with respect to class C is defined as:
E P | U C = 1 U z U log 2 x C x P x P ,
We conducted a series of experiments using different threshold values between 1.0 and 2.5 to evaluate the performance of emotion recognition when using selected channels, with the help of differential entropy, as input to the proposed 1D-CNN model. Although the model performance was sufficiently accurate with several channel numbers, we tried to optimize channel selection across all datasets under consideration to find the best performing combination. Therefore, a threshold of 1.5 was chosen to select the most significant EEG channels that directly impacted class discrimination. This threshold value not only helped achieve the same number of channels across the SEED, DEAP, and MAHNOB-HCI datasets but also improved emotion recognition performance using a limited number of channels. Specifically, 10 common channels were selected from the pool of most significant channels for all three datasets, as shown in Figure 2. Using this data-driven approach to select the optimal channels enabled improved emotion recognition with a small subset of EEG channels that were most informative for discrimination. The threshold of 1.5 struck the right balance between channel reduction and preserving class-discriminative information.

2.3.2. Feature Selection

Feature selection is a well-established preprocessing technique that plays a crucial role in reducing dimensionality and enhancing the performance of regression and classification tasks. Among the various approaches, the sequential forward feature selection methods based on Mutual Information (MI) have gained popularity in practical applications due to their computational efficiency and classifier independence. In our study, we have specifically focused on the Decomposed Mutual Information Maximization (DMIM) method [36]. This method retains the desirable theoretical properties of the most effective existing methods while addressing the issue of complementarity penalization. By separately maximizing the inter-feature and class-relevant redundancies, DMIM overcomes this limitation and offers improved performance.
For this purpose, we first divided the entire signal of each selected channel into multiple samples. We achieved this by applying a 2-s window without overlapping, allowing for better evaluation. Additionally, we labeled each sample based on its position in the original signal before applying differential entropy. This labeling enables us to obtain the original signal of each sample after ranking them. We then applied differential entropy to each 2-s sample using the following Formula (2) to use it for DMIM-based feature ranking method:
h x = 1 2 log P + 1 2 log 2 π e N ,
where P represents the average energy value of the EEG signal by replacing the variance, and the N represents the length of fixed time window, which is 2 s in this case.
Afterwards, we utilized the DMIM based features ranking method described by [36], which maximizes inter-feature and class-relevance redundancy separately, enabling joint optimization of feature diversity and relevance for selection. The mutual information between two variables can be calculated as follows:
M I A , B = a = 𝒶 b = 𝒷 P A = a , B = b l n P A = a , B = b P A = a P B = b ,
Therefore, suppose X i F be the feature being used for prioritization and its relevance with the class C; the objective function DMIM using MI(.) can be defined as follows:
D M I M X i = M I X i ,   C max X s S M I X i ,   X s + max X s S M I ( X i ,   X s | C ) ,
Each term of the objective function DMIM states a unique type of contribution between the class (C) the set of selected features (S) and the features ( X i ), whereas X s S are the selected features. The first term MI(.) measures the relevance of feature X i to the class. The second term maxMI(.) measures redundancy between X i and X s . The third term maxMI(.) measures redundancy between X i and the class-specific information in selected features. Overall, the three terms quantify relevance, inter-feature redundancy, and class-conditional redundancy for feature selection.
Our complete evaluation process for feature selection is provided in Figure 3 in terms of flow chart. While the accuracy of recognition is increasing, the process of feature reduction continues until the accuracy does not improve further. In this way, the highly ranked features are selected and contain class-specific information while removing redundant features.
Once we obtained the feature ranking from DMIM, we selected only the top twenty features with the highest contribution to class discrimination with respect to classification accuracy and disregarded the remaining ones. On average, the first twenty features had a greater influence on the accuracy of emotion recognition compared to the rest of the features (see Figure 4). Therefore, we chose these twenty features as the high-quality features. Subsequently, we obtained the original EEG samples corresponding to these features for further use in the next stage of the method, which is the 1D-CNN model. The typical size of feature vectors obtained from each level are listed in Table 2.

2.3.3. 1D-CNN Model

The extracted high-quality features from the original EEG data are passed through the 1D-CNN model. The model consists of multiple pairs of convolutional, batch normalization and pooling layers, as well as an output layer. Prior to being input into the 1D-CNN model, the features are combined into a vector format, which is subsequently convolved with a set of one-dimensional filters within the convolutional layers. Following the pooling layer, the data undergo further down-sampling, resulting in vectors with reduced dimensions. The network weights and filters within the convolutional layers are trained using the back-propagation algorithm.
The CNN structure employed in this study is relatively straightforward. The input vector is a one-dimensional feature represented by X, with a shape of N × 1. At the convolutional layer, the input feature vector is convolved with filters Wf. The convolution operation results in the formation of an output map, and the feature map at the corresponding layer is obtained using the equation provided below:
f α = f W f X + b f ,
where, weight matrix W f R i × 1 , and the bias value b are utilized in the calculation, with f denoting the filter index ranging from 1 to n, where n represents the total number of filters in the convolutional layer. In this work, the activation function employed is the rectified linear unit (ReLU), denoted by the function f. ReLU is more efficient in preventing gradient disappearance compared to traditional neural network activation functions like sigmoid and tanh. The ReLU function is defined as follows:
f α = ReLU α = ln 1 + e α ,
where value of α has been defined in previous equation.
Batch normalization is a well-known technique in modern neural networks that serves two purposes: normalizing the output from the previous layer in a CNN model and regularizing the data to prevent overfitting. It plays a crucial role in ensuring stable training and enhancing the generalization capability of the network. At the max-pooling layer, the feature map undergoes down-sampling through the application of an average-pooling function. This function is employed because it has been observed that extracting the average value from the selected values within a given feature map is an effective approach. This significantly reduces the number of trainable parameters, leading to an accelerated training process. Also, the dropout layer is employed to prevent overfitting.
Following the last pooling layer, a fully connected layer is employed, where the output data from the pooling layer is flattened. Subsequently, a series of fully connected layers called DNN (Deep Neural Network) is utilized. In each layer of the DNN, the activation function employed is also ReLU.
For the output layer, two classification tasks are considered: binary classification and three-class classification. To address the binary classification task (for DEAP and MAHNOB-HCI dataset), the sigmoid function is utilized as an activation function, while for the three-class classification task (for SEED dataset), the softmax function is employed. Additionally, for the binary-classification task, the Adadelta optimizer is used, and the loss is calculated using the binary cross-entropy formula, as indicated by the equation below:
l o s s C E = n = 1 N y ^ i log y i + 1 y ^ i log 1 y ^ i ,
where N represents the number of samples, y i denotes the value in the encoded form, and y ^ i represents the output obtained from the output layer using sigmoid activation. In the case of three-class classification, Adam optimizer is employed, and the loss is calculated using categorical cross-entropy, which can be expressed as
l o s s C C E = n = 1 N y ^ i 1 log y i 1 + y ^ i 2 log y ^ i 2 + y ^ i 3 log y ^ i 3 ,
where N represents the number of samples, and y i 1 , y i 2 , and y i 3 correspond to the values of the label, which are encoded. The values y ^ i 1 , y ^ i 2 , and y ^ i 3 represent the three outputs obtained from the output layer, where softmax activation is applied. The model’s parameters are updated using the back-propagation algorithm. The error between the desired output and the actual output is computed, and the gradient descent method is utilized to update the parameters, aiming to minimize the error. The weight and bias update functions are illustrated as follows:
W f = W f η E W f ,     b f = b f η E b f ,
where Wf refers to the weight matrix, bf represents the bias, η denotes the learning rate, and E represents the error, which is the loss calculated in Equations (7) and (8). An illustration of the proposed 1D-CNN model has been provided in Figure 5.

2.3.4. Hyperparameters

In our study, we utilize a 1D-CNN to enhance the capture of EEG signal details linked to emotions. A grid search approach was considered to achieve the most suitable set of hyper-parameters of this 1D-CNN model. Table 3 presents the key hyperparameters, their corresponding range of values, and selected values for the proposed model in this study.
The model is implemented using Python 3.9. As depicted in Figure 5 and Table 2, the size of the input 1D vector comprises of 10 (EEG channels) × 2 (window size: 2 s) × 128 (sampling rate: 128 Hz (for DEAP and SEED datasets)/256 Hz (for MAHNOB-HCI dataset)). The batch size for the model is set to 64, resulting in input data shapes of (163,840) for the DEAP and SEED datasets and (327,680) for the MAHNOB-HCI dataset.
In the proposed model, the term “1DConvolution” represents the one-dimensional convolutional layer, “AveragePooling1D” denotes one-dimensional average pooling, “BatchNorm” refers to one-dimensional batch normalization, and “FC” represents the fully connected layer. Each convolutional layer in the model is followed by an activation layer, where ReLU is used as the activation function.
The proposed model consists of four convolutional layers, three batch normalizations, two dropout layers with probabilities of 0.25, one pooling layer, one flatten layer, and two fully connected layers. These layers are combined to form four convolution blocks. The first block contains a convolutional layer with a kernel size of 5 × 1, aimed at extracting emotion-relevant features. ReLU is used as the activation function for each convolutional layer. Each convolution block, except for one, includes a normalization layer. The third convolution block is connected to an average pooling layer at the end. The last convolution block is followed by a fully connected layer, which is then connected to the final output fully connected layer after a dropout layer with a probability of 0.5. The final output of the model provides the classification results for emotion recognition. The structure of the proposed model is illustrated in Figure 5.

3. Results

In this section, we analyze the implementation of the proposed feature selection method by utilizing the classification results obtained through our approach. We have conducted various evaluation parameters to validate the results and compared them with those obtained from the state-of-the-art models to verify the effectiveness of the proposed method. To mitigate the issue of overfitting, the dataset was typically partitioned into random subsets of equal size. Accordingly, we employed the leave-one-subject-out strategy to divide our dataset for training and testing sets.
We classified three (positive, neutral, negative) emotional states for the SEED dataset and two (HV, LV) and (HA, LA) emotional states for the DEAP and MAHNOB-HCI dataset using the designed 1D-CNN model. The accuracy of the selected features for each subject in the datasets was used to validate the results. To calculate the accuracy, precision, specificity, recall, F1-score, and p-score, we employed the confusion matrix and utilized the keywords True positive, False positive, True negative, and False negative. The performance metrics for the DEAP, MAHNOB-HCI, and SEED datasets are presented in Table 4, Table 5 and Table 6, respectively.
Moreover, a t-test was conducted to evaluate the p-score (probability score), aiming to assess the usefulness of the selected function. It determined whether there existed a significant difference between the selected feature vector and the original or actual feature vector. The p-value was derived by calculating the mean difference between the selected objects and the original object vector. A higher p-value indicated a higher quality factor for the selected item.
Additionally, we performed data augmentation technique to increase the number of samples. Usually, deep learning models require a large amount of datasets to produce better results. Therefore, we added the same number of generated samples of EEG data to the original datasets of DEAP, MAHNOB-HCI, and SEED, respectively. The results of adding fake data using graph-EMD have been presented in Figure 6 with class-specific accuracies. The average accuracies are valued on the left side of the Y-axis, while the values of the standard deviation (SD) associated with the respective average accuracies are represented on the right side of the Y-axis. As shown in the figure, the LA class of the MAHNOB-HCI dataset attained the highest average accuracy of 98.43% with 0.11 SD using augmented data. The least SD of 0.09 is attributed to the neutral class of the SEED dataset with an average accuracy of 94.65% with augmented data.
To illustrate the benefits of data augmentation using graph-EMD, we visualized the real and synthesized samples through t-SNE two-dimensional projections, as shown in Figure 7. From the results, we can observe that (1) the generated data matches the actual data distributions in each dataset and (2) the augmented data maintains class-specific properties as expected, without low-quality samples that could mislead the classifier. Overall, the t-SNE visualizations demonstrate that the synthesized data has similar distributions to the real data for each emotion class, validating the quality of the graph-EMD augmentation approach.

Performance Analysis

The computational time taken by the proposed model with and without the channels and the attributes selected through the proposed method has been calculated. The following equation has been used to calculate the time [37]:
C T = 1 F T c i = 1 C C P × ln f v + 1 C P × ln 1 f v ,
where the selected feature value is denoted as fv, the training features are depicted as FT, the features of a particular emotional category are specified as CP, the number of emotional categories is symbolized as C, and the categories index is symbolized as ci.
The processing time is reduced as a result of selecting a smaller number of features. By employing the differential entropy for channel selection and the feature reduction method using DMIM, emotions can be recognized within a shorter processing time while maintaining the overall classification performance. Table 7 presents the execution time of the proposed method, comparing the results with and without the channel/feature selection method.
The channel selection greatly enhances the overall rating performance, as depicted in Table 8, where the selected channel outperforms all other channels. This improvement can be attributed to the fact that not all channels provide high-quality stimuli for emotions in EEG signals. This study highlights the importance of channel selection and its impact on understanding the emotional behavior captured by the selected channel. Also, a comparison of average accuracies on augmentation has also been provided in Table 8. It shows the efficiency of the proposed method with and without augmentation. Moreover, we used the Kappa coefficient to evaluate emotion recognition accuracy. Kappa measures the agreement between predicted and true labels, normalized for chance and classifier count. Unlike basic accuracy, Kappa provides a standardized performance estimate adjusted for variations in classes and classifiers.

4. Discussion

Given the important role of emotion in human–machine interactions, this study aims to recognize emotions using EEG signals captured with a limited number of electrodes. While people may more easily identify specific emotions they feel, the dimensional perspective offers a more fundamental understanding of the emotion–behavior relationship [38]. Neuroscience studies also support this dimensional view. Evidence of neural activities shows that valence and arousal influence cognition and behavior through distinct brain mechanisms [39]. Consistent with this theoretical background, this study systematically manipulated and examined valence and arousal effects rather than focusing on discrete emotions. By taking a dimensional approach to emotion recognition from limited-channel EEG data, this work provides insight into the fundamental mechanisms linking EEG signals to emotional states.
We have introduced a deep learning-based 1D-CNN model that effectively utilizes differential entropy for channel selection and the DMIM method for feature selection. This model efficiently processes raw EEG signals to recognize emotions, enhancing the visibility and efficiency of class-specific features. The proposed architecture simultaneously extracts valuable EEG patterns, considering the recent literature emphasizing the significance of channel selection in emotion recognition. Consequently, we specifically employ differential entropy to identify the most relevant channels for emotion recognition. Furthermore, we employ the DMIM method to select high-quality features, dividing the EEG signal into samples of 2 s on a horizontal scale. This helped in assessing the signal in a better way on a temporal scale and reducing the feature vector for the deep learning model.
In Table 4, Table 5 and Table 6, we provide the evaluation metrics for each subject. It is important to highlight that the proposed 1D-CNN model demonstrated efficient performance across all datasets examined in this study. The average accuracy of the model is highest in the MAHNOB-HCI dataset, reaching 94.6%, and lowest in the DEAP dataset, with an accuracy of 84.2%. Additionally, the model successfully tackled the three-class problem in the SEED dataset, achieving an average accuracy of 91.7%.
Furthermore, the mean p-value (0.10) is highest in the MAHNOB-HCI dataset, indicating the significance of the selected feature vector compared to the original data. Although the average p-value (0.081) for the DEAP dataset is lower than that of MAHNOB-HCI, it is still satisfactory and demonstrates the importance of the selected features over the original ones.
In order to ensure a suitable training phase, we tested different numbers of augmented data and network designs. The experimental results determining the impact of augmented data are presented in Figure 6. As depicted in Figure 6, data augmentation proves to be more effective in the deep learning model. It considerably enhances the prediction accuracy of the 1D-CNN model. Specifically, doubling the data leads to improvements of up to 4.8%, 3.9%, and 3.5% for the DEAP, MAHNOB-HCI, and SEED datasets, respectively. This improvement can be attributed to the fact that deep neural network models require a larger amount of data compared to traditional machine learning models. In conclusion, data augmentation in EEG data is a complex task that requires careful consideration. In this experiment, a substantial amount of data was generated and added to the original dataset. However, augmented data did not uniformly improve results across all datasets. When increasing the original data amount threefold using our augmentation approach, we observed overfitting and degraded testing accuracy. This highlights the fact that that simply generating more data does not necessarily lead to better performance. There are optimal augmentation levels, beyond which additional synthetic data fails to provide new information and causes overfitting. The reliability of the classifier’s accuracy on data from different sources remains a significant concern in this task.
Table 8 presents the emotion recognition accuracy and Kappa coefficient using our proposed feature extraction method with and without data augmentation, compared to using all channels. Augmentation increased the maximum accuracy on MAHNOB-HCI to 97.61%, exceeding all-channel accuracy by 8.11%. For SEED, average accuracy was 95.27% with augmentation. DEAP accuracy improved by 4.81% to 89.02% with augmentation versus original data. The kappa scores were 0.841, 0.752, and 0.713 for MAHNOB-HCI, SEED, and DEAP, indicating substantial reductions in error versus chance levels. Overall, these results demonstrate that our proposed method of an emotion recognition and data augmentation approach considerably improves EEG-based emotion recognition accuracy and consistency.
Table 9 presents a concise overview and performance comparison between our proposed method and other existing emotion recognition methods. These methods are all deep learning models and have been considered for evaluation because they used the same dataset. Luo et al. [26] proposed three methods for augmenting EEG training data to improve the performance of emotion recognition models, utilizing deep generative models (VAE and GAN) and selective data augmentation strategies. These methods, namely cWGAN, sVAE, and sWGAN, were evaluated on SEED and DEAP datasets. Topic and Russo [40] presented a new model for emotion recognition using EEG signals, utilizing topographic and holographic representations of the signal characteristics. Deep learning was employed for feature extraction and fusion, resulting in improved emotion recognition performance on publicly available datasets. Zhang et al. [41] addressed the lack of interpretability in deep learning models for EEG-based tasks and explores the use of interpretable models to analyze CNN-based model SACNN in emotion recognition. By integrating brain science knowledge with interpretability analysis results, a new model was proposed, which demonstrated improved recognition accuracy on standard EEG datasets. Zhang et al. [42] presented a novel two-step spatial-temporal framework for emotion recognition based on EEG. The framework incorporated a hierarchical self-attention network to capture both local and global temporal information, reducing noise and localizing relevant segments. Additionally, the squeeze-and-excitation module and channel correlation loss were utilized to enhance spatial feature extraction and improve emotion recognition performance. Luo and Lu [43] addressed the challenge of limited EEG data for emotion recognition by proposing a Conditional Wasserstein GAN (CWGAN) framework. By leveraging generative adversarial networks, realistic EEG data in differential entropy form were generated and evaluated using quality indicators. Overall, the proposed method outperformed the methods from the literature and provided sufficient grounds for the idea that using a deep-learning-based lightweight 1D-CNN model with limited number of channels can produce successful results. Therefore, the major contributions of this study are listed below:
  • It has been observed that the proposed deep learning method performs better than methods in the literature, even without data augmentation.
  • Generally, large datasets are expected for deep learning studies. Therefore, the improved performance after augmentation further proves this point.
  • In deep learning, manual feature selection is not typically conducted. In this study, deep learning was performed after feature selection. Manually selecting features may have caused some information loss. However, the deep learning model still achieved successful classification.
  • Generally, all channels are used for emotion recognition. Despite the fact that this work utilized a limited number of channels, the deep learning model produced successful results.
  • Processing raw data can increase deep learning algorithm performance. Due to the small sample lengths in this study, we could process the raw data, which increased overall performance.
  • Although each dataset alone was small, the reliability of the proposed method was evaluated using different datasets and increasing the data size through augmentation on each dataset separately.
To summarize, Table 9 presents a comprehensive comparison between our proposed study and existing state-of-the-art methods in the literature, focusing on EEG-based techniques applied to the DEAP, MAHNOB-HCI, and SEED datasets. Our proposed approach outperforms other methods in terms of classification accuracy while also reducing computational overhead. By utilizing a 1D-CNN model and selecting high-quality features from specific channels, our method achieves simplicity and effectiveness. The results in Table 6 demonstrate that our model significantly improves precision compared to the previously employed model using the same datasets.

5. Conclusions

In this study, we tackled two key challenges in emotion identification from EEG signals: the high dimensionality problem and the scarcity of available data. To address these issues, we employed channel and feature selection methods to reduce the dimensionality of the EEG data and employed data augmentation using graph-EMD. The implemented approach exhibited superior accuracy on the proposed 1D-CNN model, highlighting the significance of addressing data scarcity in neural network models. The model was validated using DEAP, MAHNOB-HCI, and SEED datasets, demonstrating that our feature selection method effectively enhances overall classification performance while minimizing computational costs. Additionally, the proposed model offers notable improvements in feature extraction capabilities, while channel selection aids researchers in exploring the understanding of emotions through selective channels for precise emotion recognition. These results demonstrate the potential for a cost-effective IoT EEG device to enable accessible emotion recognition. By selectively extracting clean and relevant neural signals, our proposed methods can enhance the feasibility of real-world EEG acquisition. Our system provides an efficient and accurate pipeline from targeted EEG input to emotion state predictions. Moving forward, future research directions will explore further enhancements to the emotion analysis framework by integrating multiple neural networks. Considering the reduced testing and training time of the proposed model, it would be interesting to evaluate its performance across multiple emotion categories. Also, further analysis is needed to determine ideal augmentation factors for each dataset. While augmentation generally helped, more data did not linearly translate to accuracy gains. Additionally, this study was focused on small sample size for the sake of simplicity and reduced complexity; therefore, it is required to investigate the increased sample size for preprocessing to observe its impact on the performance.

Author Contributions

Conceptualization, H.A. and F.M.; methodology, H.A.; software, H.A. and F.M.; validation, S.A.-A., F.M. and H.A.; formal analysis, H.A.; investigation, H.A.; resources, F.M.; data curation, F.M.; writing—original draft preparation, H.A.; writing—review and editing, S.A.-A.; visualization, F.M.; supervision, S.A.-A.; project administration, S.A.-A.; funding acquisition, S.A.-A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank Deanship of Scientific Research in King Saud University for funding and supporting this research through the initiative of DSR Graduate Students Research Support (GSR).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. DEAP, MAHNOB and SEED can be found here respectively: https://www.eecs.qmul.ac.uk/mmv/datasets/deap/download.html (accessed on 10 July 2020), https://mahnob-db.eu/hci-tagging/ (accessed on 7 July 2020); https://bcmi.sjtu.edu.cn/home/seed/index.html (accessed on 13 July 2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Naji, M.; Firoozabadi, M.; Azadfallah, P. Emotion Classification during Music Listening from Forehead Biosignals. Signal Image Video Process 2015, 9, 1365–1375. [Google Scholar] [CrossRef]
  2. Gan, Y. Happy People Live Longer and Better: Advances in Research on Subjective Well-Being. Appl. Psychol. Health Well-Being 2020, 12, 3–6. [Google Scholar] [CrossRef] [PubMed]
  3. Sun, J.; Wang, X.; Zhao, K.; Hao, S.; Wang, T. Multi-Channel EEG Emotion Recognition Based on Parallel Transformer and 3D-Convolutional Neural Network. Mathematics 2022, 10, 3131. [Google Scholar] [CrossRef]
  4. Wang, X.; Ren, Y.; Luo, Z.; He, W.; Hong, J.; Huang, Y. Deep Learning-Based EEG Emotion Recognition: Current Trends and Future Perspectives. Front. Psychol. 2023, 14, 1126994. [Google Scholar] [CrossRef] [PubMed]
  5. Long, F.; Zhao, S.; Wei, X.; Ng, S.-C.; Ni, X.; Chi, A.; Fang, P.; Zeng, W.; Wei, B. Positive and Negative Emotion Classification Based on Multi-Channel. Front. Behav. Neurosci. 2021, 15, 720451. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, F.; Wu, S.; Zhang, W.; Xu, Z.; Zhang, Y.; Wu, C.; Coleman, S. Emotion Recognition with Convolutional Neural Network and EEG-Based EFDMs. Neuropsychologia 2020, 146, 107506. [Google Scholar] [CrossRef]
  7. Zheng, W.-L.; Zhu, J.-Y.; Lu, B.-L. Identifying Stable Patterns over Time for Emotion Recognition from EEG. IEEE Trans. Affect. Comput. 2019, 10, 417–429. [Google Scholar] [CrossRef] [Green Version]
  8. Zhu, X.; Liu, G.; Zhao, L.; Rong, W.; Sun, J.; Liu, R. Emotion Classification from Multi-Band Electroencephalogram Data Using Dynamic Simplifying Graph Convolutional Network and Channel Style Recalibration Module. Sensors 2023, 23, 1917. [Google Scholar] [CrossRef]
  9. Wang, Z.; Chen, M.; Feng, G. Study on Driver Cross-Subject Emotion Recognition Based on Raw Multi-Channels EEG Data. Electronics 2023, 12, 2359. [Google Scholar] [CrossRef]
  10. Algarni, M.; Saeed, F.; Al-Hadhrami, T.; Ghabban, F.; Al-Sarem, M. Deep Learning-Based Approach for Emotion Recognition Using Electroencephalography (Eeg) Signals Using Bi-Directional Long Short-Term Memory (Bi-Lstm). Sensors 2022, 22, 2976. [Google Scholar] [CrossRef]
  11. Li, R.; Ren, C.; Zhang, X.; Hu, B. A Novel Ensemble Learning Method Using Multiple Objective Particle Swarm Optimization for Subject-Independent EEG-Based Emotion Recognition. Comput. Biol. Med. 2022, 140, 105080. [Google Scholar] [CrossRef]
  12. Gao, Z.; Li, Y.; Yang, Y.; Wang, X.; Dong, N.; Chiang, H.-D. A GPSO-Optimized Convolutional Neural Networks for EEG-Based Emotion Recognition. Neurocomputing 2020, 380, 225–235. [Google Scholar] [CrossRef]
  13. Hancer, E.; Subasi, A. EEG-Based Emotion Recognition Using Dual Tree Complex Wavelet Transform and Random Subspace Ensemble Classifier. Comput. Methods Biomech. Biomed. Eng. 2022, 1–13. [Google Scholar] [CrossRef]
  14. Ari, B.; Siddique, K.; Alcin, O.F.; Aslan, M.; Sengur, A.; Mehmood, R.M. Wavelet ELM-AE Based Data Augmentation and Deep Learning for Efficient Emotion Recognition Using EEG Recordings. IEEE Access 2022, 10, 72171–72181. [Google Scholar] [CrossRef]
  15. Cheng, W.X.; Gao, R.; Suganthan, P.N.; Yuen, K.F. EEG-Based Emotion Recognition Using Random Convolutional Neural Networks. Eng. Appl. Artif. Intell. 2022, 116, 105349. [Google Scholar] [CrossRef]
  16. Salama, E.S.; El-Khoribi, R.A.; Shoman, M.E.; Wahby, M.A. EEG-Based Emotion Recognition Using 3D Convolutional Neural Networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 329–337. [Google Scholar] [CrossRef] [Green Version]
  17. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A Multimodal Database for Affect Recognition and Implicit Tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef] [Green Version]
  18. Koelstra, S.; Muhl, C.; Soleymani, M.; Jong-Seok, L.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  19. Zheng, W.-L.; Lu, B.-L. Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep Neural Networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  21. Kalaganis, F.P.; Laskaris, N.A.; Chatzilari, E.; Nikolopoulos, S.; Kompatsiaris, I. A Data Augmentation Scheme for Geometric Deep Learning in Personalized Brain–Computer Interfaces. IEEE Access 2020, 8, 162218–162229. [Google Scholar] [CrossRef]
  22. Gu, X.; Cao, Z.; Jolfaei, A.; Xu, P.; Wu, D.; Jung, T.-P.; Lin, C.-T. EEG-Based Brain-Computer Interfaces (BCIs): A Survey of Recent Studies on Signal Sensing Technologies and Computational Intelligence Approaches and Their Applications. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 1645–1666. [Google Scholar] [CrossRef] [PubMed]
  23. Towle, V.L.; Bolaños, J.; Suarez, D.; Tan, K.; Grzeszczuk, R.; Levin, D.N.; Cakmur, R.; Frank, S.A.; Spire, J.-P. The Spatial Location of EEG Electrodes: Locating the Best-Fitting Sphere Relative to Cortical Anatomy. Electroencephalogr. Clin. Neurophysiol. 1993, 86, 1–6. [Google Scholar] [CrossRef] [PubMed]
  24. Kim, H.N.; Sutharson, S.J. Emotional Valence and Arousal Induced by Auditory Stimuli among Individuals with Visual Impairment. Br. J. Vis. Impair. 2023, 41, 254–264. [Google Scholar] [CrossRef]
  25. George, O.; Smith, R.; Madiraju, P.; Yahyasoltani, N.; Ahamed, S.I. Data Augmentation Strategies for EEG-Based Motor Imagery Decoding. Heliyon 2022, 8, e10240. [Google Scholar] [CrossRef]
  26. Luo, Y.; Zhu, L.-Z.; Wan, Z.-Y.; Lu, B.-L. Data Augmentation for Enhancing EEG-Based Emotion Recognition with Deep Generative Models. J. Neural. Eng. 2020, 17, 056021. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Duan, F.; Sole-Casals, J.; Dinares-Ferran, J.; Cichocki, A.; Yang, Z.; Sun, Z. A Novel Deep Learning Approach With Data Augmentation to Classify Motor Imagery Signals. IEEE Access 2019, 7, 15945–15954. [Google Scholar] [CrossRef]
  28. Abdelfattah, S.M.; Abdelrahman, G.M.; Wang, M. Augmenting The Size of EEG Datasets Using Generative Adversarial Networks. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 4 July 2018; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar]
  29. Xie, C.; Shao, Y.; Li, X.; He, Y. Detection of Early Blight and Late Blight Diseases on Tomato Leaves Using Hyperspectral Imaging. Sci. Rep. 2015, 5, 16564. [Google Scholar] [CrossRef] [Green Version]
  30. Affes, A.; Mdhaffar, A.; Triki, C.; Jmaiel, M.; Freisleben, B. Personalized Attention-Based EEG Channel Selection for Epileptic Seizure Prediction. Expert. Syst. Appl. 2022, 206, 117733. [Google Scholar] [CrossRef]
  31. Ramadhani, A.; Fauzi, H.; Wijayanto, I.; Rizal, A.; Shapiai, M.I. The Implementation of EEG Transfer Learning Method Using Integrated Selection for Motor Imagery Signal. In Proceedings of the 1st International Conference on Electronics, Biomedical Engineering, and Health Informatics, Surabaya, Indonesia, 8–9 October 2021; Triwiyanto, N.H.A., Rizal, A., Caesarendra, W., Eds.; Lecture Notes in Electrical Engineering. Springer Singapore: Singapore, 2021; Volume 746, pp. 457–466. [Google Scholar]
  32. Dura, A.; Wosiak, A. EEG Channel Selection Strategy for Deep Learning in Emotion Recognition. Procedia Comput. Sci. 2021, 192, 2789–2796. [Google Scholar] [CrossRef]
  33. Daoud, H.; Bayoumi, M.A. Efficient Epileptic Seizure Prediction Based on Deep Learning. IEEE Trans. Biomed. Circuits Syst. 2019, 13, 804–813. [Google Scholar] [CrossRef] [PubMed]
  34. Apicella, A.; Arpaia, P.; Isgro, F.; Mastrati, G.; Moccaldi, N. A Survey on EEG-Based Solutions for Emotion Recognition With a Low Number of Channels. IEEE Access 2022, 10, 117411–117428. [Google Scholar] [CrossRef]
  35. Alarcao, S.M.; Fonseca, M.J. Emotions Recognition Using EEG Signals: A Survey. IEEE Trans. Affect. Comput. 2019, 10, 374–393. [Google Scholar] [CrossRef]
  36. Macedo, F.; Valadas, R.; Carrasquinha, E.; Oliveira, M.R.; Pacheco, A. Feature Selection Using Decomposed Mutual Information Maximization. Neurocomputing 2022, 513, 215–232. [Google Scholar] [CrossRef]
  37. Aurelio, Y.S.; De Almeida, G.M.; De Castro, C.L.; Braga, A.P. Learning from Imbalanced Data Sets with Weighted Cross-Entropy Function. Neural Process. Lett. 2019, 50, 1937–1949. [Google Scholar] [CrossRef]
  38. Barrett, L.F. Discrete Emotions or Dimensions? The Role of Valence Focus and Arousal Focus. Cogn. Emot. 1998, 12, 579–599. [Google Scholar] [CrossRef]
  39. Dolcos, F.; LaBar, K.S.; Cabeza, R. Dissociable Effects of Arousal and Valence on Prefrontal Activity Indexing Emotional Evaluation and Subsequent Memory: An Event-Related FMRI Study. NeuroImage 2004, 23, 64–74. [Google Scholar] [CrossRef] [Green Version]
  40. Topic, A.; Russo, M. Emotion Recognition Based on EEG Feature Maps through Deep Learning Network. Eng. Sci. Technol. Int. J. 2021, 24, 1442–1454. [Google Scholar] [CrossRef]
  41. Zhang, Y.; Cui, C.; Zhong, S. EEG-Based Emotion Recognition via Knowledge-Integrated Interpretable Method. Mathematics 2023, 11, 1424. [Google Scholar] [CrossRef]
  42. Zhang, Y.; Liu, H.; Zhang, D.; Chen, X.; Qin, T.; Zheng, Q. EEG-Based Emotion Recognition with Emotion Localization via Hierarchical Self-Attention. IEEE Trans. Affect. Comput. 2022, 1. [Google Scholar] [CrossRef]
  43. Luo, Y.; Lu, B.-L. EEG Data Augmentation for Emotion Recognition Using a Conditional Wasserstein GAN. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; IEEE: Honolulu, HI, USA, 2018; pp. 2535–2538. [Google Scholar]
Figure 1. Emotion recognition process using proposed framework.
Figure 1. Emotion recognition process using proposed framework.
Diagnostics 13 02624 g001
Figure 2. Selected most significant channels for class discrimination are the ones with filled circles.
Figure 2. Selected most significant channels for class discrimination are the ones with filled circles.
Diagnostics 13 02624 g002
Figure 3. Process of selection of features using DMIM based objective function.
Figure 3. Process of selection of features using DMIM based objective function.
Diagnostics 13 02624 g003
Figure 4. A plot between accuracy and selected features of a particular selected channel using DEAP, MAHNOB-HCI and SEED datasets.
Figure 4. A plot between accuracy and selected features of a particular selected channel using DEAP, MAHNOB-HCI and SEED datasets.
Diagnostics 13 02624 g004
Figure 5. Architecture of Proposed 1D-CNN model. Note. The different colors of input EEG samples at each row shows a signal from different channel.
Figure 5. Architecture of Proposed 1D-CNN model. Note. The different colors of input EEG samples at each row shows a signal from different channel.
Diagnostics 13 02624 g005
Figure 6. Performance (average accuracy/standard deviation (SD)) of the proposed model with/without augmented data using graph-EMD. The Y-axis on the left side shows the average accuracy associated with the Original and Original+Augmented data, while the Y-axis on the right side shows the values of the SD associated with the respective average accuracies.
Figure 6. Performance (average accuracy/standard deviation (SD)) of the proposed model with/without augmented data using graph-EMD. The Y-axis on the left side shows the average accuracy associated with the Original and Original+Augmented data, while the Y-axis on the right side shows the values of the SD associated with the respective average accuracies.
Diagnostics 13 02624 g006
Figure 7. The t-SNE visualization of the real and generated EEG data distributions for one subject from the SEED, DEAP, and MAHNOB-HCI datasets is shown in (ac), respectively. The lines represent the real EEG data, while the scattered points are the artificially generated augmented data. In (a), the data points colored red, blue, and yellow represent the negative, positive, and neutral classes, respectively. In (b,c), the data points colored red, orange, blue, and magenta represent the high valence, low valence, high arousal, and low arousal classes, respectively.
Figure 7. The t-SNE visualization of the real and generated EEG data distributions for one subject from the SEED, DEAP, and MAHNOB-HCI datasets is shown in (ac), respectively. The lines represent the real EEG data, while the scattered points are the artificially generated augmented data. In (a), the data points colored red, blue, and yellow represent the negative, positive, and neutral classes, respectively. In (b,c), the data points colored red, orange, blue, and magenta represent the high valence, low valence, high arousal, and low arousal classes, respectively.
Diagnostics 13 02624 g007
Table 1. Comparison of characteristics between all three datasets under discussion.
Table 1. Comparison of characteristics between all three datasets under discussion.
CharacteristicsDEAPMAHNOB-HCISEED
Participants322715
Trials402015
Channels323262
Sampling rate (Hz)128256128
Affective statesValence
Arousal
Dominance
Valence
Arousal
Positive
Negative
Neutral
Rating scale range1–91–9-
Table 2. Size of the feature vector obtained at step of proposed method for each subject.
Table 2. Size of the feature vector obtained at step of proposed method for each subject.
Raw EEGChannel SelectionFeature Selection
DEAP32 × 768010 × 768010 × 5120
MAHNOB-HCI32 × 15,36010 × 15,36010 × 10,240
SEED62 × 30,72010 × 30,72010 × 5120
Table 3. Hyperparameters of the proposed 1D-CNN model.
Table 3. Hyperparameters of the proposed 1D-CNN model.
Parameter TypeRange of ValuesSelected Value
Number of 1DConvolution Layers3–84
Number of filters16–12864, 32, 32, 16
Size of kernel1 × 1–7 × 15 × 1, 5 × 1, 3 × 1, 1 × 1
Batch size32–12864
Learning rate0.0005–0.010.005
Momentum0.1–0.90.9
Dropout0.25–0.50.25
Number of epochs50–200100
Table 4. Performance of the proposed method on DEAP dataset.
Table 4. Performance of the proposed method on DEAP dataset.
SubjectAccuracyPrecisionSpecificityRecallF1-Scorep-Score
Subject 196.17789900.890.091
Subject 289.99975760.830.098
Subject 385.97577830.810.0697
Subject 478.37673700.870.084
Subject 577.38275660.850.067
Subject 683.18171660.850.0748
Subject 786.49591920.870.083
Subject 883.29274770.840.0852
Subject 983.88978760.860.09
Subject 1082.98187810.800.08
Subject 1176.56789750.860.0849
Subject 1287.29179720.890.084
Subject 1383.87190860.80.0951
Subject 1487.67693690.940.075
Subject 1587.29192990.80.0732
Subject 16857374680.790.089
Subject 1783.88673700.840.085
Subject 1884.66991860.870.079
Subject 1981.87175900.90.08
Subject 2083.39466760.910.069
Subject 2184.18579820.870.09
Subject 2284.98978920.810.076
Subject 2386.036865860.820.072
Subject 2481.127677840.780.0921
Subject 2582.57170960.820.0695
Subject 2687.39477840.90.09
Subject 2782.528595740.830.088
Subject 2880.378581640.840.07
Subject 2985.619498980.870.087
Subject 3082.289488790.950.075
Subject 3183.849392780.890.076
Subject 3285.258278780.830.0785
Average84.28381800.850.081
Table 5. Performance of the proposed method on MAHNOB-HCI dataset.
Table 5. Performance of the proposed method on MAHNOB-HCI dataset.
SubjectAccuracyPrecisionSpecificityRecallF1-Scorep-Score
Subject 195.39389900.890.072
Subject 297.69296930.890.078
Subject 397.49489930.910.0839
Subject 491.28992900.990.073
Subject 592.79085870.990.072
Subject 696.58887860.970.0794
Subject 791.19693940.910.067
Subject 893.19789900.930.0802
Subject 995.89592920.970.08
Subject 1094.59394910.890.0812
Subject 1192.59498951.000.073
Subject 1296.79292981.000.0779
Subject 13958990960.970.0793
Subject 14969093990.990.79
Subject 15928894820.910.0684
Subject 1693.59089850.900.076
Subject 1796.69192920.990.08
Subject 1895.19286910.990.0858
Subject 1996.19590890.940.076
Subject 2094.29492940.990.0831
Subject 2193.99392960.890.0792
Subject 2295.79493930.960.087
Subject 2393.28289950.890.083
Subject 2496.28892900.940.086
Subject 25959889881.00.08
Subject 2695.69192920.950.082
Subject 2792.69292900.950.0717
Average94.69291910.950.10
Table 6. Performance of the proposed method on SEED dataset.
Table 6. Performance of the proposed method on SEED dataset.
SubjectAccuracyPrecisionSpecificityRecallF1-Scorep-Score
Subject 193.88792910.970.0975
Subject 291.49185860.970.0845
Subject 392.89489830.910.09
Subject 491.19784870.980.092
Subject 590.28982900.950.0945
Subject 692.59687890.950.0935
Subject 789.48996940.910.0965
Subject 894.68893890.930.094
Subject 990.49399890.980.095
Subject 1091.19192950.860.093
Subject 1189.69397980.980.091
Subject 1292.79791960.960.0885
Subject 1390.57880960.930.087
Subject 1493.59092980.970.095
Subject 1592.18185820.910.0806
Average91.79089910.940.092
Table 7. Computational time comparison of proposed method with and without channel/feature selection method.
Table 7. Computational time comparison of proposed method with and without channel/feature selection method.
DEAPMAHNOB-HCISEED
ChannelsFeature VectorTime (s)ChannelsFeature VectorTime (s)ChannelsFeature VectorTime (s)
CS+FS+1D-CNN1051,3002210102,400461051,30022
CS+1D-CNN10245,76018910491,52022110409,600249
FS+1D-CNN32524,288303321,048,576629622,539,5201066
1D-CNN322,621,4401100325,242,88012746210,158,0801675
Table 8. Comparison between all channels and selected channels in terms of accuracy with/without augmented data and Kappa coefficient for EEG classification.
Table 8. Comparison between all channels and selected channels in terms of accuracy with/without augmented data and Kappa coefficient for EEG classification.
Original +
Augmented
Original
(10 Channels)
All ChannelsKappa
DEAP89.0284.2181.060.713
MAHNOB-HCI97.6194.6589.500.841
SEED95.2791.7288.300.752
Table 9. Comparison between the results of the proposed method and the state of the art methods from literature according to various factors and the number of datasets used.
Table 9. Comparison between the results of the proposed method and the state of the art methods from literature according to various factors and the number of datasets used.
AugmentationDEAPMAHNOB-HCISEED
AccuracyChannelsClassesAccuracyChannelsClassesAccuracyChannelsClasses
ProposedYes8910497.6/1010495.3103
Luo et al. [26]Yes59.1324---93.5623
Topic and Russo [40]No77.72324---88.45623
Zhang et al. [41]No8332107110289102
Zhang et al. [42]No77.0332481.3732479.56623
Luo and Lu [43]Yes78.17324---86.96623
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aldawsari, H.; Al-Ahmadi, S.; Muhammad, F. Optimizing 1D-CNN-Based Emotion Recognition Process through Channel and Feature Selection from EEG Signals. Diagnostics 2023, 13, 2624. https://doi.org/10.3390/diagnostics13162624

AMA Style

Aldawsari H, Al-Ahmadi S, Muhammad F. Optimizing 1D-CNN-Based Emotion Recognition Process through Channel and Feature Selection from EEG Signals. Diagnostics. 2023; 13(16):2624. https://doi.org/10.3390/diagnostics13162624

Chicago/Turabian Style

Aldawsari, Haya, Saad Al-Ahmadi, and Farah Muhammad. 2023. "Optimizing 1D-CNN-Based Emotion Recognition Process through Channel and Feature Selection from EEG Signals" Diagnostics 13, no. 16: 2624. https://doi.org/10.3390/diagnostics13162624

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop