EEG Data Augmentation for Emotion Recognition with a Task-Driven GAN
Abstract
:1. Introduction
2. Related Works
2.1. Emotion Recognition Based on EEGs
2.2. Data Augmentation Based on GAN
2.2.1. GAN
2.2.2. EEG Data Augmentation Based on GAN
- The multi-channel EEG differential entropy feature matrix, which retains the spatial distribution information of channels, was extracted as the data augmentation target.
- For emotion recognition research, this is conducive to the subsequent classification work, i.e., to retain as much spatial information of the data as possible in the feature extraction process. In this study, the augmentation target is the multi-channel EEG differential entropy feature matrix that retains the channel spatial distribution information rather than the original EEG signal.
- A task-driven data augmentation method for emotion recognition based on GAN is proposed.
- We introduce the classifier into the EEG data augmentation network based on GAN. Both the discriminator and the classifier will provide gradients for parameter optimization. The former promotes the generation of realistic EEG data, and the latter ensures that the generated data will help to classify the performance.
- We evaluate the performance of the proposed method from two aspects.
- The Wasserstein distance, MMD (maximum mean difference), and reduced dimension visualization (UMAP) methods are used to evaluate the data quality. Four different classifiers, including SVM and Lenet, are used to enhance the accuracy of downstream classification tasks by enhancing the emotional classification evaluation data.
3. The Proposed Method
3.1. The Basic Structure and Variants of GAN
- WGAN: A major problem of the original GAN is its instability during generative adversarial training resulting from gradient disappearance caused by the discontinuity of the Jensen–Shannon divergence. WGAN formalizes adversarial training by minimizing the Wasserstein distance rather than the Jensen–Shannon divergence between the distribution of generated data and real data so that it can continuously provide useful gradients for the parameter optimization of the generator. Formula (2) shows the objective function of the WGAN:
- CWGAN: CGAN specifies the types of generated data for the generation network by connecting labels and random noise at the input, so as to ensure the generation of data within a specific range. It can be simply combined with WGAN to form CWGAN. Formula (3) shows the objective function of the CWGAN:
3.2. EEG Data Augmentation for Emotion Recognition
3.2.1. Structure for Multi-Channel EEG Feature Map
3.2.2. The Task-Driven CWGAN
Algorithm 1: Training Process |
Training discriminator and classifier (update the parameters of the discriminator and classifier): |
|
|
|
|
|
|
Training generator (update the parameters of the generator): |
|
|
|
4. Experimental Works and Results
4.1. Experiment Works
4.1.1. Database
4.1.2. Data Processing
4.1.3. Experiment Details
4.2. Result
4.2.1. Quality of Generated Data
4.2.2. Improvement of Performance of Emotion Recognition Task
5. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Tong, X.; Chen, Y.; Zhou, S.; Yang, S. How background visual complexity influences purchase intention in live streaming: The mediating role of emotion and the moderating role of gender. J. Retail. Consum. Serv. 2022, 67, 103031. [Google Scholar] [CrossRef]
- Ferrer-Cairols, I.; Ferré-González, L.; García-Lluch, G.; Peña-Bautista, C.; Álvarez-Sánchez, L.; Baquero, M.; Cháfer-Pericás, C. Emotion recognition and baseline cortisol levels relationship in early Alz-heimer disease. Biol. Psychol. 2023, 177, 108511. [Google Scholar] [CrossRef] [PubMed]
- Jiang, Y.; Li, W.; Hossain, M.S.; Chen, M.; Alelaiwi, A.; Al-Hammadi, M. A Snapshot Research and Implementation of Multimodal Information Fusion for Da-ta-driven Emotion Recognition. Inf. Fusion 2019, 53, 209–221. [Google Scholar] [CrossRef]
- Pourebrahim, Y.; Razzazi, F.; Sameti, H. Semi-supervised parallel shared encoders for speech emotion recognition. Digit. Signal Process. 2021, 118, 103205. [Google Scholar] [CrossRef]
- Sima, Y.; Yi, J.; Chen, A.; Jin, Z. Automatic expression recognition of face image sequence based on key-frame generation and differential emotion feature. Appl. Soft Comput. 2021, 113, 108029. [Google Scholar] [CrossRef]
- Chowanda, A.; Sutoyo, R.; Tanachutiwat, S. Exploring Text-based Emotions Recognition Machine Learning Techniques on Social Media Conversation. Procedia Comput. Sci. 2021, 179, 821–828. [Google Scholar] [CrossRef]
- Yan, M.; Deng, Z.; He, B.; Zou, C.; Wu, J.; Zhu, Z. Emotion classification with multichannel physiological signals using hybrid feature and adaptive decision fusion. Biomed. Signal Process. Control. 2022, 71, 103235. [Google Scholar] [CrossRef]
- Lashgari, E.; Liang, D.; Maoz, U. Data augmentation for deep-learning-based electroencephalography. J. Neurosci. Methods 2020, 346, 108885. [Google Scholar] [CrossRef]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
- Hartmann, K.G.; Schirrmeister, R.T.; Ball, T. EEG-GAN: Generative adversarial networks for electroencephalograhic (EEG) brain signals. arXiv 2018, arXiv:1806.01875. [Google Scholar]
- Bouallegue, G.; Djemal, R. EEG data augmentation using Wasserstein GAN. In Proceedings of the 2020 20th International Conference on Sciences and Techniques of Automatic Control and Computer Engineering (STA), Monastir, Tunisia, 20–22 December 2020. [Google Scholar]
- Luo, Y.; Lu, B.L. EEG Data Augmentation for Emotion Recognition Using a Conditional Wasserstein GAN. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 2535–2538. [Google Scholar]
- Luo, Y.; Zhu, L.Z.; Wan, Z.Y.; Lu, B.L. Data Augmentation for Enhancing EEG-based Emotion Recognition with Deep Generative Models. J. Neural Eng. 2020, 17, 056021. [Google Scholar] [CrossRef] [PubMed]
- Zhang, A.; Su, L.; Zhang, Y.; Fu, Y.; Wu, L.; Liang, S. EEG data augmentation for emotion recognition with a multiple generator conditional Was-serstein GAN. Complex Intell. Syst. 2021, 8, 3059–3071. [Google Scholar] [CrossRef]
- Rahman, M.M.; Sarkar, A.K.; Hossain, M.A.; Hossain, M.S.; Islam, M.R.; Hossain, M.B.; Quinn, J.M.W.; Moni, M.A. Recognition of human emotions using EEG signals: A review. Comput. Biol. Med. 2021, 136, 104696. [Google Scholar] [CrossRef] [PubMed]
- Topic, A.; Russo, M. Emotion recognition based on EEG feature maps through deep learning network. Eng. Sci. Technol. Int. J. 2021, 24, 1442–1454. [Google Scholar] [CrossRef]
- An, Y.; Xu, N.; Qu, Z. Leveraging spatial-temporal convolutional features for EEG-based emotion recognition. Biomed. Signal Process. Control. 2021, 69, 102743. [Google Scholar] [CrossRef]
- Liu, Y.; Ding, Y.; Li, C.; Cheng, J.; Song, R.; Wan, F.; Chen, X. Multi-channel EEG-based emotion recognition via a multi-level features guided capsule network. Comput. Biol. Med. 2020, 123, 103927. [Google Scholar] [CrossRef] [PubMed]
- Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
- Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
- Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
- Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of Wasserstein GANs. Adv. Neural Inf. Process. Syst. 2017, 5769–5779. [Google Scholar]
- Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- Chaitanya, K.; Karani, N.; Baumgartner, C.F.; Erdil, E.; Becker, A.; Donati, O.; Konukoglu, E. Semi-supervised Task-driven Data Augmentation for Medical Image Seg-mentation. Med. Image Anal. 2021, 68, 101934. [Google Scholar] [CrossRef] [PubMed]
- Tripathi, S.; Chandra, S.; Agrawal, A.; Tyagi, A.; Rehg, J.M.; Chari, V. Learning to Generate Synthetic Data via Compositing. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 461–470. [Google Scholar]
- Yang, Y.; Wu, Q.; Fu, Y.; Chen, X. Continuous Convolutional Neural Network with 3D Input for EEG-Based Emotion Recognition. In Proceedings of the 25th International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018. [Google Scholar]
(a) | |
Layer (type) | Output Shape |
Linear-1 | [64, 324] |
ReLU-2 | [64, 4, 9, 9] |
Conv2d-3 | [64, 64, 9, 9] |
ReLU-4 | [64, 64, 9, 9] |
Conv2d-5 | [64, 32, 9, 9] |
ReLU-6 | [64, 32, 9, 9] |
Conv2d-7 | [64, 16, 9, 9] |
Conv2d-8 | [64, 4, 9, 9] |
(b) | |
Layer (type) | Output Shape |
Conv2d-1 | [64, 32, 9, 9] |
LeakReLU-2 | [64, 32, 9, 9] |
MaxPool2d-3 | [64, 32, 4, 4] |
Conv2d-4 | [64, 64, 4, 4] |
LeakReLU-5 | [64, 64, 4, 4] |
MaxPool2d-6 | [64, 64, 2, 2] |
Linear-7 | [64, 256] |
LeakReLU-8 | [64, 256] |
Linear-9 | [64, 1] |
(c) | |
Layer (type) | Output Shape |
Conv2d-1 | [64, 32, 9, 9] |
LeakReLU-2 | [64, 32, 9, 9] |
MaxPool2d-3 | [64, 32, 4, 4] |
Conv2d-4 | [64, 64, 4, 4] |
LeakReLU-5 | [64, 64, 4, 4] |
MaxPool2d-6 | [64, 64, 2, 2] |
Linear-7 | [64, 256] |
LeakReLU-8 | [64, 256] |
Linear-9 | [64, 1] |
Sigmoid-10 | [64, 1] |
Subjects | Videos | EEG Channels | Sampling Rate | Emotional Dimensions | Label Values |
32 | 40 | 32 | 128 Hz | Arousal and Valence | Continuous values in the range of 1–9 |
Data format of each subject (Array shape) | |||||
Videos × EEG Channels × Sampling rate × Segment length = 40 × 32 × 128 Hz × 63 s (3 s in rest and 60 s watching videos) |
Classifiers | Emotional Dimensions | Arousal | Valence | ||||
---|---|---|---|---|---|---|---|
Original | CWGAN | CWGAN-T | Original | CWGAN | CWGAN-T | ||
SVM | 0.8485 | 0.8589 | 0.9135 | 0.8330 | 0.8545 | 0.9064 | |
Tasknet | 0.8015 | 0.8876 | 0.9022 | 0.7880 | 0.8755 | 0.8956 | |
CNN-noPooling | 0.8499 | 0.9054 | 0.9225 | 0.8346 | 0.8977 | 0.9187 | |
Lenet-with-BN | 0.8748 | 0.9050 | 0.9352 | 0.8664 | 0.9021 | 0.9275 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, Q.; Hao, J.; Guo, Y. EEG Data Augmentation for Emotion Recognition with a Task-Driven GAN. Algorithms 2023, 16, 118. https://doi.org/10.3390/a16020118
Liu Q, Hao J, Guo Y. EEG Data Augmentation for Emotion Recognition with a Task-Driven GAN. Algorithms. 2023; 16(2):118. https://doi.org/10.3390/a16020118
Chicago/Turabian StyleLiu, Qing, Jianjun Hao, and Yijun Guo. 2023. "EEG Data Augmentation for Emotion Recognition with a Task-Driven GAN" Algorithms 16, no. 2: 118. https://doi.org/10.3390/a16020118
APA StyleLiu, Q., Hao, J., & Guo, Y. (2023). EEG Data Augmentation for Emotion Recognition with a Task-Driven GAN. Algorithms, 16(2), 118. https://doi.org/10.3390/a16020118