Next Article in Journal
Methods of Pre-Clustering and Generating Time Series Images for Detecting Anomalies in Electric Power Usage Data
Previous Article in Journal
Product Authentication Technology Integrating Blockchain and Traceability Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EEG Emotion Recognition Based on Federated Learning Framework

School of Information and Electrical Engineering, Zhejiang University City College, Hangzhou 310015, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(20), 3316; https://doi.org/10.3390/electronics11203316
Submission received: 16 September 2022 / Revised: 10 October 2022 / Accepted: 11 October 2022 / Published: 14 October 2022

Abstract

:
Emotion recognition based on the multi-channel electroencephalograph (EEG) is becoming increasingly attractive. However, the lack of large datasets and privacy concerns lead to models that often do not have enough data for training, limiting the research and application of Deep Learn (DL) methods in this direction. At present, the popular federated learning (FL) approach, which can collaborate with different clients to perform distributed machine learning without sending data to a central server, provides a solution to the above problem. In this paper, we extended the FL method to the field of emotion recognition based on EEG signals and evaluated its accuracy in the DEAP and SEED datasets, where the model accuracy reached 90.74% in our framework. We also divided the DEAP dataset into different clients. The accuracy of emotion recognition decreased by 29.31% compared to the FL method when the clients were trained using local data, which validates the necessity of the FL approach for emotion recognition tasks. In addition, we verified the impact of N-IID data on the accuracy of FL training. The experiment demonstrated that N-IID leads to a 14.89% decrease in accuracy compared to IID.

1. Introduction

Emotion recognition is a hot issue in human–machine interaction systems [1,2], and its accuracy directly affects the user’s interaction experience. At the same time, its application is beneficial for diagnosing diseases such as depression [3], tracking patients’ recovery effects and assisting psychology in studying emotional behavior. Emotion recognition can be based on behavioral signals such as actions and expressions [4,5] and physiological signals such as EEG and ECG [6,7]. Among them, EEG signals provide direct measurements of signals generated by the human nervous system, which is the most direct, reliable and accurate way to reflect human emotional behaviors, and therefore is widely used in emotion recognition.
In recent years, with the advancement of EEG acquisition technology, EEG signals’ temporal resolution and post-processing techniques have been significantly improved, which has laid a solid foundation for applying deep learning techniques to process EEG signals for emotion recognition [8,9]. Due to the temporal asymmetry and instability of EEG signals, poor signal-to-noise ratio and the existence of variability among different brain regions and individuals [10], it poses many challenges for EEG-based emotion recognition. Traditional machine learning methods categorize this task as a classification problem, using algorithms such as decision tree [11], multilayer perceptron [12] and support vector machine [13]. Due to the complexity of EEG signal recognition and categorization, traditional algorithms often have the problem of inaccurate or even unclassifiable classification. Deep learning networks have made significant progress in classification problems in medical image processing, computer vision and other fields due to their strong generalization ability and automatic learning of abstract features [14,15]. They have also achieved advantages far beyond traditional algorithms in the problem of emotion recognition based on EEG signals [16,17]. Deep learning algorithms represented by convolutional neural networks still have problems. Compared with image classification problems that can be processed locally, the classification of EEG signals has discrete characteristics in time and space, making it difficult for traditional convolutional neural networks to achieve higher accuracy rates. Acharya [18] achieved a recognition rate of 87.72 by increasing the number of layers of convolutional neural networks. Rudakov [19] proposed extracting features using differential entropy and power spectra density methods to fit EEG signals’ characteristics further and achieve 96.28% recognition accuracy, which has more value for EEG emotion recognition.
The increase in accuracy of the above emerging EEG emotion recognition algorithms comes at the cost of higher system parameter complexity and the number of training sets, i.e., a sufficient amount of training data is required for the models to be trained to achieve clinical-level accuracy. Since physiological signals such as EEG can reflect the data collector’s personality traits, emotional tendencies and other brain activity characteristics, the development of deep learning has heightened the privacy leakage risks that may arise from data misuse in terms of privacy protection [20]. Governments have started to pay extensive attention to data security and privacy protection issues, with the EU and China adopting the General Data Protection Regulation [21] in May 2018 and the Data Security Law of the People’s Republic of China in June 2021, respectively, requiring the use of personal data to be subject to the consent of the data owner.
Compared with traditional picture data in the field of deep learning, EEG data are subject to technical limitations leading to a more difficult acquisition of EEG data [22], and the high time domain signal-to-noise ratio of EEG data and other characteristics make clean EEG data that can be used for deep learning model training more challenging to obtain. As a result, the EEG data available to various medical institutions are insufficient to train emotion recognition models with high accuracy and robustness under increasingly stringent data security and privacy protection regulations. This makes the research and clinical application of deep learning methods in EEG more challenging and becomes one of the urgent problems to be solved.
In 2017, to address the issues of privacy preservation and data silos in devices, Google proposed federated learning (FL) [23], which provides us with a solution to the above problems. Federation learning is essentially a distributed machine learning method, which abandons the traditional way of training models with centralized data and allows data between devices to participate in training without going out of local by uploading training models for aggregation through local training, and this method protects data privacy.
FL has gained the attention and focus of many researchers in the healthcare field, where data privacy is highly valued because it has the potential to address trust and privacy issues arising from the sensitivity of patient data. Brisimi [24], in 18, proposed an FL model capable of predicting hospitalizations of cardiac patients using inter-institutional EHR data. In order to enable institutions with small amounts of Melanoma Detection data to collaborate in training the model and to address the problem of poor data availability, Agbley [25] extended the FL approach to Melanoma Detection for disease detection. Although the FL algorithm achieves to protect data privacy, it is still possible to attack the uploaded models to obtain data information. Therefore, Malekzadeh [26] proposed an FL system based on differential privacy stochastic gradient descent (DPSGD) and secure aggregation to improve the security of FL further. Moreover, with the explosion of COVID-19, the FL-based approach for COVID-19 detection also attracted the interest of many researchers, among which Feki [27] combined blockchain and FL approaches to achieve cross-institutional co-training of COVID-19 detection models using CTs of COVID-19 patients. At the same time, Zhang [28] designed a dynamic fusion of FL systems to improve the communication efficiency of the FL algorithm for COVID-19 detection accuracy. The application environment of the FL algorithm presents a key data challenge: the distribution of medical data held between different medical institutions is usually non-independently and identically distributed (N-IID). Such N-IID among data has been shown to cause substantial accuracy degradation in traditional image classification domains such as CIFAR-10 or MNIST. However, the N-IID problem of the FL algorithm in EEG data has not been taken seriously [29].
In this study, we expected to use the FL approach to address the problems of insufficient data volume and complexity faced by research and clinical applications in this field, enabling institutions with small datasets to train machine learning models collaboratively. Our contributions mainly include the following:
1.
This paper extends the FL method to EEG signal-based emotion recognition field and evaluates its accuracy in the DEAP and SEED datasets. Our validation shows that the FL method can lead to higher model accuracy;
2.
We constructed different DEAP datasets for evaluating the effect of the diversity of training data on emotion recognition models. It was verified that the accuracy of the emotion recognition model using EEG signals is highly dependent on subjects and that increasing the diversity of subjects can substantially improve the model’s generalization performance, demonstrating the need for the FL method to be applied in this domain;
3.
The impact of the FL method on the accuracy and convergence speed of the emotion recognition model when trained on EEG data with N-IID distribution was evaluated by simulating the N-IID distribution of the inter-client DEAP dataset. Compared with the IID distribution, there is a substantial decrease in the accuracy of the FL-trained emotion recognition model under the N-IID distribution.

2. Materials and Methods

The federated learning framework for training emotion recognition models is shown in Figure 1. The server distributes the emotion recognition models to the clients (individuals or medical institutions) with EEG data participating in the FL task. The client participating in the task processes the data locally into the data format required for training the FL task to train the sentiment recognition model, ends the training after the specified number of training rounds, and uploads the locally trained model to the server. The server aggregates the models uploaded by the client and determines whether to continue the FL training. The methods used in the framework are described in this section.

2.1. Electroencephalography—Emotion Recognition Dataset

The DEAP dataset is a multimodal dataset contributed by Koelstra [30]. This dataset recorded the EEG and peripheral physiological signals of 32 participants. In the experiment, 32 subjects watched music videos of 1 min in length as required, for a total of 40 videos per subject. The experimenters recorded 32 channels of EEG and 8 channels of peripheral physiological information simultaneously using a BioSemi EEG cap according to the international 10–20 standard. Subjects scored each video on four dimensions: validity, arousal, dominance, likeness and familiarity, as required. During pre-processing, the raw 512 hz EEG signal was downsampled and filtered to 128 hz to remove artifacts such as EOG and muscle movement.
The scoring and EEG data of 32 subjects were saved in two formats, MATLAB.mat and Python.dat. The data file corresponding to each subject contains two arrays: data and labels. Data has a data dimension of 40*40*8064, each video saves 40 channels of data, and each channel has 8064. The dimension of data is 40*40*8064, each video has 40 channels of data, and each channel has 8064 saved electrical signals; the dimension of labels is 40*4, and the scores of four aspects were recorded: valence, arousal, dominance, likeness and familiarity. In addition, frontal face clips of 22 subjects were recorded and saved in the face_video.zip file.
The SEED dataset is an emotion classification dataset provided by the BCMI lab of Shanghai Jiao Tong University [31]. The SEED dataset records EEG signals of 15 subjects through 62 channels of electrode caps with a sampling frequency of 1000 Hz according to the 10–20 standard. There were 5 movie clips for each emotion, for a total of 15 movie clips. The SEED dataset provides EEG data down sampled to 200 Hz with a 0.5–75 bandpass filter, which is stored in a .mat file in the Preprocessed_EEG folder, and the corresponding labels are stored in a label.mat file (−1 means negative, 0 means neutral, 1 means positive).

2.2. Signal Pre-Processing

We use fast Fourier transform (FFT) for feature extraction to transform the data relative to time variation into a spectrogram relative to frequency variation, making the model converge faster while having a higher accuracy rate. The sample dimension of the processed EEG signal data becomes (1, 70). These extracted features include five frequency bands: Theta-θ (4–8 Hz), Alpha-α (8–14 Hz), Beta-β (14–30 Hz) and Gamma-γ (31–50 Hz) [32].
FFT is a more efficient and faster computational method for the discrete Fourier transform (DFT) [33], which can be used to transform the signal domain from time to frequency. The use of this algorithm enables the computer to compute the discrete Fourier transform with a much-reduced number of multiplications, thus reducing the computational time as well as the computational complexity.
Since the fluctuations of emotional states are mainly concentrated in the 14 channels [34] in Table 1, we selected these channels for training, which can reduce the computational cost of this study and does not affect the accuracy of the model. The time window was set to 2 s, and the update step was updated every 0.125 s.

2.3. Emotion Recognition Model

Convolutional neural network (CNN) is a deep learning algorithm that can automatically learn spatial features in data samples without human extraction. Its advantages of high accuracy and high generalization ability in deep learning compared to traditional machine learning algorithms have made it widely used in the direction of image classification. Therefore, in this study, a standard CNN model was chosen as a machine learning model for evaluating the completion of the emotion classification task in FL methods. The network structure of this model is shown in Figure 2, which mainly consists of three one-dimensional convolutional neural networks, two pooling layers and three fully connected layers, and finally, the lined SoftMax layer was used for emotion classification.
In the experiments of this study, we input the data of shape (70*1) into this network with a convolutional kernel size of 5 and a step size of 1. After the convolutional network operation, the output was subjected to maximum pooling. The pooling layer has a pooling window of 2 and a step size of 1. Going through the pooling layer reduces the dimensionality of the information extracted from the convolutional layer and reduces the computational effort. After the third convolutional layer, the output was pulled into 1 dimension through the Flatten layer and sent to the fully connected layer for further training. All layers in this network use the ReLU function as the activation function, and the output data were normalized to a standard normal distribution with mean 0 and variance 1 by batch normalization to accelerate the model training.

2.4. Federated Learning Algorithm

This section introduces the basic composition structure, model training process and parameter updating methods of federated learning algorithms. Federated learning is a distributed machine learning algorithm that protects data privacy. The standard federated learning framework works together to train high-quality global models through the client, the server and the aggregation framework. In this process, the model obtained by the server through aggregation is usually called the central model, while the model distributed by the server to the client and trained locally is called the local model.
Usually, the FL training process consists of the following three steps:
Step 1: Initialize the FL task. The server determines the FL training task, i.e., the training target of FL task and the data requirements needed for training. Then the FL task is released to request clients who meet the conditions to join the FL task, and the global model and training parameters of the FL task are determined according to the computing power, communication and data volume constraints of the clients joining the training task. The server then distributes the initialized global model and training requirements to the clients participating in the task [35];
Step 2: Execute the FL task to achieve model training. The client uses the local data to train the issued global model W and update the model parameters according to the training requirements, which are usually minimized loss functions. After completing the training, the client uploads the updated local model to the server. The server waits for all clients to upload the local model and then obtains the new global model by weighting the parameters of the local model by the aggregation algorithm to obtain a weighted average;
Step 3: End of FL task. The updated global model is tested, and the FL task is ended when the global model performance meets the task goal. If the model performance does not meet the task target, the global model is resent to the client for parameter update waiting for a new round of model aggregation and testing by repeating step 2.
In the above FL training process, the client usually holds the training data and provides the computational resources needed for the model to be trained locally. The server is usually the commander of a node with reliable computational power, responsible for sending the global model to the client, receiving the local model returned by the client, and implementing the aggregation in the server, in addition to managing the communication between the client and itself. The aggregation algorithm is the core of the FL framework. The local model is updated into a new global model by aggregation algorithm, and the global model obtained by the aggregation algorithm should have good accuracy and generalization performance.
The aggregation algorithm used in this study is federated averaging (FedAvg), which performs a weighted average of the local model parameters for each client. The goal of Fedavg is usually to minimize the loss function of all samples, which can be expressed as Equation (1):
ω = a r g m i n L ω
where ω denotes the main model parameters and the L ω function denotes the global loss function. ω parameters are updated by the gradient uploaded by each client for optimization, which usually requires some training rounds. The server distributes the global model to client i. The client i calculates the local gradient L i ω i locally using the SGD algorithm and updates the local model with the learning rate λ. The expression of the model update is expressed as follows:
ω i t + 1 = ω i t λ L i ω i t .
The client can be trained locally several times before participating in aggregation. After the server gives the aggregation command and completes local training, the clients upload their model parameters to the server, and then the aggregator in the server aggregates the model. The aggregator uses the FedAvg method for aggregation as follows:
ω t + 1 = i = 1 N n i n ω i t + 1 .
Here, N is the number of clients participating in federated learning, n denotes the total amount of data from clients participating in aggregation, and n i denotes the amount of data owned by client i. The server aggregates the local models uploaded by each client based on the number of samples from each party with a weighted average of the model parameters to obtain the global model ω t + 1 for the next round. The above method can be used to achieve model aggregation at the server after local training of models by the clients in the process of step 2. The training stops when the global loss function, accuracy, or the number of training rounds reaches a threshold, and the global loss function and accuracy are two critical metrics for the same number of training rounds. Finally, this classical federation learning is applied in this study and experimentally proved to have better results.

3. Results

3.1. Experimental Setup

All our experiments were conducted on a Windows 10 computer with an AMD R7–4800H CPU and a GTX1650Ti GPU, and the experimental code was implemented in Python 3.7. We divided each subject’s data in the dataset into 80% training set and 20% test set, and the training data were equally distributed to all clients involved in the federation learning. In contrast, the test set was kept in the server measurement, and the accuracy of the aggregated model was tested and recorded after the completion of the aggregation. In our experiments, we set the optimizer for model training to Adam, the learning rate to 0.001 and the batch size to 2048, and each round of aggregation required the clients to train 1 epoch locally.

3.2. Experimental Results

In this study, we validated the accuracy of the FL framework on the DEAP and SEED datasets. In the experiments, the SEED dataset label is a tri-categorization task (negative, neutral, positive), and the DEAP data label is the subjects’ self-assessment of four feelings of arousal, valence, liking and dominance with scores ranging from 1 to 9. In order to make comparing experimental results from different datasets more intuitive, we divided the DEAP labels into four for the triple classification task by using the scores of 4 and 7 in the DEAP dataset labels as the threshold.
Table 2 records the tested accuracies of 1 client (without the FL method), the FL method with 5 clients, and the FL method with 10 clients when training 200 Epochs. The experiments show that the FL algorithm of five clients obtains higher accuracy when training 200 Epochs compared to the FL method without FL, where the classification accuracy of DEAP-Dominance reaches 90.74%
Figure 3 shows the variation curves of the loss function of the FL algorithm with the different number of clients when trained on the DEAP dataset with different labels and the SEED dataset. The total amount of training data is the same, but it is evident that the rate of loss decline of the FL algorithm receives the influence of the number of clients when training on the DEAP dataset. The convergence of the model leveled off at 60 epochs; the convergence rate gradually decreases with the increase in the number of clients involved in federal learning, and the FL model converges to level off only at 200 epochs of aggregation when 10 clients are used.
We simulated different training datasets for comparison experiments with the following experimental setup to verify the effect of training source diversity on model generalization performance.
Since real healthcare organizations usually have all the data of one or some subjects, the DEAP dataset is a sentiment classification dataset with 32 subjects. Therefore, we took the overall data of one subject (i.e., 1/32 = 3.125% of the data) as the base, reconstructed the DEAP dataset into the following four different training sets, and assumed that they are owned by four different medical institutions (A, B, C, D):
A. Data owned by two subjects from s01 to s02 (6.25% of DEAP);
B. Data owned by subjects from s01 to s16 (50% of DEAP);
C. Data owned by subjects from s017 to s32 (50% of DEAP);
D. Data owned by subjects from s01 to s32 (100% of DEAP).
In order to exclude the effect of sample size on the accuracy, we replicated the datasets of medical institution A 16 times and the datasets of medical institutions B and C twice, thus ensuring that each medical institution uses the same number of training samples. Finally, each medical institution uses the training data to train the emotion recognition model.
The models were all tested in a test set composed of the complete DEAP subject data, and the experimental data are presented in Figure 4. In the experiments with the label Arousal, the accuracy was 36.25% when training with s01 to s02 subjects and 58.63% when training with s01 to s16 subjects, which was much lower than the accuracy of 87.94% when using all the data. This trend is also shown in the other three labels.
Figure 5 shows the experiments show that the model is tested with an accuracy of up to 91.28% on the data sourced from the training set, but the same model is only 58.82% on the test set composed of all the subjects’ data. This highlights the specificity of EEG data, where each person has some variation in the EEG signal produced in response to external stimuli, and emotion recognition models trained using deep learning methods need to use a large amount of different data to ensure that the model meets the model robustness required for clinical care.
Since the sample collection for emotion recognition is based on the subjective thoughts of the subject or patient, the data distribution of the EEG dataset for emotion recognition available among medical institutions is mostly N-IID. Therefore, we validated the performance variation in the FL algorithm under the N-IID data distribution. In order to ensure intuitive experimental results, we classified the DEAP data according to 1~9 and sampled the DEAP dataset, retaining 10,000 data samples for each label, for a total of 90,000 data samples. Moreover, these data are assigned to different clients using the Dirichlet distribution to achieve the simulated N-IID distribution [36], and the client data distribution is shown in the following Table 3. As a comparison experiment, we equally distributed the resampled DEAP dataset to all clients to simulate the IID distribution.
Figure 6 shows the rising accuracy curves of the FL algorithm for N-IID and IID distributions, where all four labels produce a severe accuracy drop in the N-IID distribution, and an average accuracy decrease of 14.89% can be obtained from Table 4.
The convergence rate of the FL algorithm at the data distribution of N-IID, the model’s training, converges only after 450 rounds of aggregation, and the convergence curve at N-IID is more tortuous than the IID distribution when the model has converged after 200 rounds of aggregation.
The FL approach allows the model to achieve higher accuracy when using EEG for emotion classification. Table 5 summarizes the comparison of 600 Epoch trained with 10 clients with other state-of-the-art automatic emotion classification techniques, with our approach generally achieving higher accuracy.

4. Discussion

In recent years, the use of machine learning methods to extract and identify emotional features in EEG [41]. Since the training and validation of models in machine learning methods are highly data-dependent, it makes the research and clinical applications of machine learning methods in EEG fields such as emotion recognition, where there are few publicly available data sets, complex data collection and strict data privacy requirements, more difficult due to the lack of data quantity. For this reason, we investigated federated learning methods in EEG-based emotion recognition to address this problem and evaluate the effectiveness of federated learning methods and the new problems they pose. Our results support our expectation that by using a federated learning approach, owners of EEG sentiment data can use the data to jointly train a global model without sharing the data they have and that the global sentiment recognition models trained by this approach have high accuracy.
Unlike traditional machine learning models that are trained directly in the data to gain experience, federal learning aggregates multiple local models in such a way that the global model gains the experience learned by the local models during training, which makes the process of convergence of the global model affected by multiple local models at the same time. In order to evaluate the variability in the training emergence of emotion recognition models caused by the federal learning approach compared to the traditional machine learning approach, we first designed experiments under the data distribution of IID. We evaluated them on the DEAP and SEED datasets. From the experimental results presented in Section 3, it is clear that using the federation learning method with the same amount of data leads to higher accuracy but may affect the convergence rate of the model. We believe this is because the global model of federal learning learns the local model experience by averaging the gradients of the local model’s parameters, so the global model’s convergence rate is not affected by the number of local models involved in the aggregation. In contrast, when the number of global data is fixed for federation learning, as the number of clients increases, each client has fewer data, and the local model learns less experience after one round of epoch, so the convergence speed becomes slower as the number of clients increases, even though the number of data involved in federation learning is the same.
In order to verify the necessity of federation learning, we reconstructed four DEAP training and testing subsets with the same amount of data according to the source of data collection using the DEAP dataset and show the emotion recognition ability possessed by the model after training under different subsets in Section 3. The results show that the more complex the source of training data is when training a machine learning model, the stronger the classification ability and robustness of the model. This demonstrates that many healthcare organizations with EEG data can obtain more accurate and robust emotion classification models if they can share their data to form more complex datasets for training. Thus, federal learning methods that can accomplish sub-goals become very relevant in the face of data privacy constraints.
Although this study demonstrates that FL methods can solve the problem of insufficient data samples when studying machine learning algorithms in the EEG domain, our accuracy evaluation of FL tasks with N-IID data sample distribution among clients also demonstrates the shortcomings of the current FL algorithms. Experiments show that the accuracy of the global model when the FL algorithms are trained on N-IID data samples is substantially higher when compared to data. The accuracy of the global model decreases significantly when the FL algorithm is trained on a sample of N-IID data compared to the data distribution of IID. However, the distribution of EEG data among institutions with small EEG data must be N-IID, and there are still few studies to address the accuracy degradation of FL algorithms in medical data when facing N-IID data distribution. We believe this issue is one of the problems that must be solved for clinical applications of FL algorithms in EEG, so we plan to investigate this in our future work.

5. Conclusions

In response to the current deep learning research and clinical applications based on EEG signal data receive insufficient data size in public datasets, and EEG data cannot be used mutually between different institutions. This study introduces federated learning into the field of emotion recognition. The method allows for the aggregation of training models by having the client upload the training models so that data between devices can participate in the training without going out of the local area, avoiding the problem of data privacy leakage. The FL method decouples between emotion recognition models, and any emotion recognition model or algorithm can be trained together using the FL method for multiple devices. The effectiveness and superiority of the method for EEG signal emotion recognition are validated on two different tasks. Experiments show that the FL method exhibits higher accuracy, with a 2.75% improvement in accuracy compared to traditional centralized training on a single device. The need for diverse data sources and the necessity of federal learning for emotion recognition models based on EEG data is also demonstrated through experiments using subjects with different DEAP datasets. Finally, N-IID experiments with simulated data distributions show that training an emotion recognition model using the FL method produces a 14.89% decrease in accuracy when faced with N-IID data, a problem that needs to be addressed.

Author Contributions

Conceptualization, C.X. and H.L.; funding acquisition, H.L.; investigation, W.Q.; methodology, C.X.; software, C.X.; writing—original draft, C.X. and H.L.; writing—review and editing, C.X. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Scientific Research Foundation of Zhejiang University City College (No.X-202106); Zhejiang Provincial Natural Science Foundation of China under Grant No. LQ22F010002.

Data Availability Statement

In this paper, two EEG datasets, DEAP and SEED, are used for emotion recognition. You can find them at the following link. DEAP: http://www.eecs.qmul.ac.uk/mmv/datasets/deap/ (accessed on 10 October 2022). SEED: https://bcmi.sjtu.edu.cn/~seed/seed.html (accessed on 10 October 2022).

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Fragopanagos, N.; Taylor, J. Emotion recognition in human–computer interaction. Neural Netw. 2005, 18, 389–405. [Google Scholar] [CrossRef] [PubMed]
  2. Yin, Y.; Zheng, X.; Hu, B.; Zhang, Y.; Cui, X. EEG emotion recognition using fusion model of graph convolutional neural networks and LSTM. Appl. Soft Comput. 2020, 100, 106954. [Google Scholar] [CrossRef]
  3. Li, Y. A Survey of EEG Analysis based on Graph Neural Network. In Proceedings of the 2021 2nd International Conference on Electronics, Communications and Information Technology (CECIT), Sanya, China, 29 December 2021; pp. 151–155. [Google Scholar] [CrossRef]
  4. Huang, X.; Zhao, G.; Hong, X.; Zheng, W.; Pietikäinen, M. Spontaneous facial micro-expression analysis using Spatiotemporal Completed Local Quantized Patterns. Neurocomputing 2016, 175, 564–578. [Google Scholar] [CrossRef]
  5. Huang, X.; Wang, S.-J.; Liu, X.; Zhao, G.; Feng, X.; Pietikainen, M. Discriminative Spatiotemporal Local Binary Pattern with Revisited Integral Projection for Spontaneous Facial Micro-Expression Recognition. IEEE Trans. Affect. Comput. 2017, 10, 32–47. [Google Scholar] [CrossRef] [Green Version]
  6. Yin, Z.; Zhao, M.; Wang, Y.; Yang, J.; Zhang, J. Recognition of emotions using multimodal physiological signals and an ensemble deep learning model. Comput. Methods Programs Biomed. 2017, 140, 93–110. [Google Scholar] [CrossRef]
  7. Abadi, M.K.; Kia, M.; Subramanian, R.; Avesani, P.; Sebe, N. Decoding affect in videos employing the MEG brain signal. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  8. Jirayucharoensak, S.; Pan-Ngum, S.; Israsena, P. EEG-Based Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation. Sci. World J. 2014, 2014, 627892. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, B.; Han, X.; Tang, J. Three class emotions recognition based on deep learning using staked autoencoder. In Proceedings of the 2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; pp. 1–5. [Google Scholar] [CrossRef]
  10. Wang, X.-H.; Zhang, T.; Xu, X.-M.; Chen, L.; Xing, X.-F.; Chen, C.L.P. EEG Emotion Recognition Using Dynamical Graph Convolutional Neural Networks and Broad Learning System. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 1240–1244. [Google Scholar] [CrossRef]
  11. Cheng, J.; Chen, M.; Li, C.; Liu, Y.; Song, R.; Liu, A.; Chen, X. Emotion recognition from multi-channel eeg via deep forest. IEEE J. Biomed. Health. 2021, 25, 453–464. [Google Scholar] [CrossRef]
  12. George, F.P.; Shaikat, I.M.; Hossain, P.S.F.; Parvez, M.Z.; Uddin, J. Recognition of emotional states using EEG signals based on time-frequency analysis and SVM classifier. Int. J. Electr. Comput. Eng. (IJECE) 2019, 9, 1012–1020. [Google Scholar] [CrossRef]
  13. Fdez, J.; Guttenberg, N.; Witkowski, O.; Pasquali, A. Cross-Subject EEG-Based Emotion Recognition Through Neural Networks with Stratified Normalization. Front. Neurosci. 2021, 15, 626277. [Google Scholar] [CrossRef]
  14. Succetti, F.; Rosato, A.; Di Luzio, F.; Ceschini, A.; Panella, A.M. A FAST DEEP LEARNING TECHNIQUE FOR WI-FI-BASED HUMAN ACTIVITY RECOGNITION. Prog. Electromagn. Res. 2022, 174, 127–141. [Google Scholar] [CrossRef]
  15. Gong, D.; Ma, T.; Evans, J.; He, A.S. Deep Neural Networks for Image Super-Resolution in Optical Microscopy by Using Modified Hybrid Task Cascade U-Net. Prog. Electromagn. Res. 2021, 171, 185–199. [Google Scholar] [CrossRef]
  16. Islam, R.; Islam, M.; Rahman, M.; Mondal, C.; Singha, S.K.; Ahmad, M.; Awal, A.; Islam, S.; Moni, M.A. EEG Channel Correlation Based Model for Emotion Recognition. Comput. Biol. Med. 2021, 136, 104757. [Google Scholar] [CrossRef]
  17. Huang, D.; Chen, S.; Liu, C.; Zheng, L.; Tian, Z.; Jiang, D. Differences first in asymmetric brain: A bi-hemisphere discrepancy convolutional neural network for EEG emotion recognition. Neurocomputing 2021, 448, 140–151. [Google Scholar] [CrossRef]
  18. Acharya, D.; Jain, R.; Panigrahi, S.S.; Sahni, R.; Deshmukh, S.P.; Bhardwaj, A. Multi-class Emotion Classification Using EEG Signals. In International Advanced Computing Conference; Springer: Singapore, 2021; pp. 474–491. [Google Scholar] [CrossRef]
  19. Rudakov, E.; Laurent, L.; Cousin, V.; Roshdi, A.; Fournier, R.; Nait-Ali, A.; Beyrouthy, T.; Al Kork, S. Multi-Task CNN model for emotion recognition from EEG Brain maps. In Proceedings of the 2021 4th International Conference on Bio-Engineering for Smart Technologies (BioSMART), Online, 8–10 December 2021; pp. 1–4. [Google Scholar] [CrossRef]
  20. Solangi, Z.A.; Solangi, Y.A.; Chandio, S.; Aziz, M.B.S.A.; bin Hamzah, M.S.; Shah, A. The future of data privacy and security concerns in Internet of Things. In Proceedings of the 2018 IEEE International Conference on Innovative Research and Development (ICIRD), Bangkok, Thailand, 11–12 May 2018; pp. 1–4. [Google Scholar] [CrossRef]
  21. General Data Protection Regulation (GDPR). Available online: https://www.epsu.org/sites/default/files/article/files/GDPR_FINAL_EPSU.pdf (accessed on 10 October 2022).
  22. Valentin, O.; Ducharme, M.; Crétot-Richert, G.; Monsarrat-Chanon, H.; Viallet, G.; Delnavaz, A.; Voix, J. Validation and benchmarking of a wearable EEG acquisition platform for real-world applications. IEEE T. Biomed. Circ. S. 2019, 13, 103–111. [Google Scholar] [CrossRef]
  23. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data H. Artif. Intell. Stat. 2017, 54, 10. [Google Scholar]
  24. Theodora, S.B.; Chen, R.; Theofanie, M.; Olshevsky, A.; Paschalidis, I.C.; Shi, W. Federated learning of predictive models from federated electronic health records. Int. J. Med Inform. 2018, 112, 59–67. [Google Scholar] [CrossRef]
  25. Agbley, B.L.Y.; Li, J.; Haq, A.U.; Bankas, E.K.; Ahmad, S.; Agyemang, I.O.; Kulevome, D.; Ndiaye, W.D.; Cobbinah, B.; Latipova, S. Multimodal Melanoma Detection with Federated Learning. In Proceedings of the 2021 18th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 17–19 December 2021; pp. 238–244. [Google Scholar] [CrossRef]
  26. Malekzadeh, M.; Hasircioglu, B.; Mital, N.; Katarya, K.; Ozfatura, M.E.; Gündüz, D. Dopamine: Differentially Private Federated Learning on Medical Data. arXiv 2021, arXiv:2101.11693. [Google Scholar]
  27. Feki, I.; Ammar, S.; Kessentini, Y.; Muhammad, K. Federated learning for COVID-19 screening from Chest X-ray images. Appl. Soft Comput. 2021, 106, 107330. [Google Scholar] [CrossRef]
  28. Zhang, W.; Zhou, T.; Lu, Q.; Wang, X.; Zhu, C.; Sun, H.; Wang, Z.; Lo, S.K.; Wang, F.-Y. Dynamic-Fusion-Based Federated Learning for COVID-19 Detection. IEEE Internet Things J. 2021, 8, 15884–15891. [Google Scholar] [CrossRef]
  29. Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the Convergence of FedAvg on Non-IID Data. arXiv 2019, arXiv:1907.02189. [Google Scholar]
  30. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  31. Zheng, W.-L.; Lu, B.-L. Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep Neural Networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  32. Cimtay, Y.; Ekmekcioglu, E. Investigating the Use of Pretrained Convolutional Neural Network on Cross-Subject and Cross-Dataset EEG Emotion Recognition. Sensors 2020, 20, 2034. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Al-Fahoum, A.S.; Al-Fraihat, A.A. Methods of EEG Signal Features Extraction Using Linear Analysis in Frequency and Time-Frequency Domains. ISRN Neurosci. 2014, 2014, 730218. [Google Scholar] [CrossRef] [Green Version]
  34. Al-Qazzaz, N.K.; Sabir, M.K.; Ali, S.; Ahmad, S.A.; Grammer, K. Effective EEG Channels for Emotion Identification over the Brain Regions using Differential Evolution Algorithm. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; Volume 2019, pp. 4703–4706. [Google Scholar] [CrossRef]
  35. Wang, J.; Charles, Z.; Xu, Z.; Joshi, G.; McMahan, H.B.; Al-Shedivat, M.; Andrew, G.; Avestimehr, S.; Daly, K.; Data, D.; et al. A Field Guide to Federated Optimization. arXiv 2021, arXiv:2107.06917. [Google Scholar]
  36. Li, Q.; Diao, Y.; Chen, Q.; He, B. Federated Learning on Non-IID Data Silos: An Experimental Study. In Proceedings of the 2022 IEEE 38th International Conference on Data Engineering (ICDE), Online, 9–12 May 2022; pp. 965–978. [Google Scholar] [CrossRef]
  37. Luo, Y.; Fu, Q.; Xie, J.; Qin, Y.; Wu, G.; Liu, J.; Jiang, F.; Cao, Y.; Ding, X. EEG-Based Emotion Classification Using Spiking Neural Networks. IEEE Access 2020, 8, 46007–46016. [Google Scholar] [CrossRef]
  38. Nawaz, R.; Cheah, K.H.; Nisar, H.; Yap, V.V. Comparison of different feature extraction methods for EEG-based emotion recognition. Biocybern. Biomed. Eng. 2020, 40, 910–926. [Google Scholar] [CrossRef]
  39. Topic, A.; Russo, M. Emotion recognition based on EEG feature maps through deep learning network. Eng. Sci. Technol. Int. J. 2021, 24, 1442–1454. [Google Scholar] [CrossRef]
  40. Galvão, F.; Alarcão, S.; Fonseca, M. Predicting Exact Valence and Arousal Values from EEG. Sensors 2021, 21, 3414. [Google Scholar] [CrossRef]
  41. Li, X.; Zhang, Y.; Tiwari, P.; Song, D.; Hu, B.; Yang, M.; Zhao, Z.; Kumar, N.; Marttinen, P. EEG based Emotion Recognition: A Tutorial and Review. ACM Comput. Surv. 2022. [Google Scholar] [CrossRef]
Figure 1. Federated learning framework for training emotion recognition models.
Figure 1. Federated learning framework for training emotion recognition models.
Electronics 11 03316 g001
Figure 2. Emotion recognition model.
Figure 2. Emotion recognition model.
Electronics 11 03316 g002
Figure 3. The loss decreases curve of different numbers of clients during training on the four emotions of DEAP dataset (a) and SEED dataset (b). The ordinate is the loss value, and the abscissa is the aggregation rounds.
Figure 3. The loss decreases curve of different numbers of clients during training on the four emotions of DEAP dataset (a) and SEED dataset (b). The ordinate is the loss value, and the abscissa is the aggregation rounds.
Electronics 11 03316 g003
Figure 4. The accuracy of the models trained from the four DEAP data on the test set shows.
Figure 4. The accuracy of the models trained from the four DEAP data on the test set shows.
Electronics 11 03316 g004
Figure 5. The accuracy of training under s01~s16 DEAP data, using the test set composed of s01~s16 data and the test set composed of s01~s32.
Figure 5. The accuracy of training under s01~s16 DEAP data, using the test set composed of s01~s16 data and the test set composed of s01~s32.
Electronics 11 03316 g005
Figure 6. Federal learning of rising accuracy curves in the face of IID and N-IID data distributions, trained and validated in four labels of the DEAP dataset.
Figure 6. Federal learning of rising accuracy curves in the face of IID and N-IID data distributions, trained and validated in four labels of the DEAP dataset.
Electronics 11 03316 g006
Table 1. Channel Selection.
Table 1. Channel Selection.
ChannelDEAP IndexSeed
Index
ChannelDEAP IndexSeed
Index
AF313AF4174
F327F41911
F735F82013
FC5415FC62121
T4723T82531
P71141P82949
O01358O23160
Table 2. Accuracy performance of different training methods on the four emotions of DEAP dataset and SEED dataset.
Table 2. Accuracy performance of different training methods on the four emotions of DEAP dataset and SEED dataset.
MethodDEAP-ValenceDEAP-ArousalDEAP-DominanceDEAP-LikingSEED
1 client0.85690.86380.87930.85820.8225
FL-5 clients0.88430.89740.90740.87190.8572
FL-10 clients0.87940.88700.90390.86870.8463
Table 3. The data distribution of the clients after distributing the data using the Dirichlet distribution is shown with the columns indicating the data categories and the horizontal columns indicating the clients. The values in the table represent the data for each category in the different clients.
Table 3. The data distribution of the clients after distributing the data using the Dirichlet distribution is shown with the columns indicating the data categories and the horizontal columns indicating the clients. The values in the table represent the data for each category in the different clients.
ClassClass 1Class 2Class 3Class 4Class 5Class 6Class 7Class 8Class 9Distribution
Client
Client 1110289213458384755208249Electronics 11 03316 i001
Client 23474898527083581557650Electronics 11 03316 i002
Client 315430364135851521115046650Electronics 11 03316 i003
Client 41008542431072112027814278483Electronics 11 03316 i004
Client 50329142397936396548161587Electronics 11 03316 i005
Client 6210146688214151218536872770Electronics 11 03316 i006
Client 7242830696329000000Electronics 11 03316 i007
Client 839747163312134531193460731Electronics 11 03316 i008
Client 9338678535549672631535722550Electronics 11 03316 i009
Client 104911448214404855261260600Electronics 11 03316 i010
All Data10,00010,00010,00010,00010,00010,00010,00010,00010,000Electronics 11 03316 i011
Table 4. The accuracy of the models obtained by training on DEAP datasets with IID and N-IID distributions, respectively.
Table 4. The accuracy of the models obtained by training on DEAP datasets with IID and N-IID distributions, respectively.
Method.ValenceArousalDominanceLiking
IID0.92000.91020.92220.9286
N-IID0.76330.75900.79990.7633
Table 5. Comparison with state-of-the-art techniques using the same DEAP EEG dataset for the emotion recognition task.
Table 5. Comparison with state-of-the-art techniques using the same DEAP EEG dataset for the emotion recognition task.
MethodValenceArousalDominanceLiking
Luo et al. [37]0.74000.78000.80000.8627
Acharya et al. [18]0.85070.83830.81430.8574
Nawaz et al. [38]0.78960.77620.7760/
Topic et al. [39]0.77720.7661//
Galvao et al. [40]0.89840.8983//
FL-10 Client0.92000.91020.92220.9286
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, C.; Liu, H.; Qi, W. EEG Emotion Recognition Based on Federated Learning Framework. Electronics 2022, 11, 3316. https://doi.org/10.3390/electronics11203316

AMA Style

Xu C, Liu H, Qi W. EEG Emotion Recognition Based on Federated Learning Framework. Electronics. 2022; 11(20):3316. https://doi.org/10.3390/electronics11203316

Chicago/Turabian Style

Xu, Chang, Hong Liu, and Wei Qi. 2022. "EEG Emotion Recognition Based on Federated Learning Framework" Electronics 11, no. 20: 3316. https://doi.org/10.3390/electronics11203316

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop