Next Article in Journal
Manufacturing and Functional Characterization of Bioengineered Liver Grafts for Extracorporeal Liver Assistance in Acute Liver Failure
Next Article in Special Issue
Classification of EEG Signals Based on Sparrow Search Algorithm-Deep Belief Network for Brain-Computer Interface
Previous Article in Journal
Extraction, Isolation, and Component Analysis of Turmeric-Derived Exosome-like Nanoparticles
Previous Article in Special Issue
A Novel Asynchronous Brain Signals-Based Driver–Vehicle Interface for Brain-Controlled Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Canonical Correlation Analysis for EEG Inter-Band Correlation Extraction

1
School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200030, China
2
Defense Innovation Institute, Academy of Military Sciences (AMS), Beijing 100000, China
3
Tianjin Artificial Intelligence Innovation Center (TAIIC), Tianjin 300450, China
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(10), 1200; https://doi.org/10.3390/bioengineering10101200
Submission received: 10 August 2023 / Revised: 29 September 2023 / Accepted: 5 October 2023 / Published: 16 October 2023
(This article belongs to the Special Issue Advances in Neurotechnology and Brain-Robot Interfaces)

Abstract

:
(1) Background: Emotion recognition based on EEG signals is a rapidly growing and promising research field in affective computing. However, traditional methods have focused on single-channel features that reflect time-domain or frequency-domain information of the EEG, as well as bi-channel features that reveal channel-wise relationships across brain regions. Despite these efforts, the mechanism of mutual interactions between EEG rhythms under different emotional expressions remains largely unexplored. Currently, the primary form of information interaction between EEG rhythms is phase–amplitude coupling (PAC), which results in computational complexity and high computational cost. (2) Methods: To address this issue, we proposed a method of extracting inter-bands correlation (IBC) features via canonical correlation analysis (CCA) based on differential entropy ( D E ) features. This approach eliminates the need for surrogate testing and reduces computational complexity. (3) Results: Our experiments verified the effectiveness of IBC features through several tests, demonstrating that the more correlated features between EEG frequency bands contribute more to emotion classification accuracy. We then fused IBC features and traditional DE features at the decision level, which significantly improved the accuracy of emotion recognition on the SEED dataset and the local CUMULATE dataset compared to using a single feature alone. (4) Conclusions: These findings suggest that IBC features are a promising approach to promoting emotion recognition accuracy. By exploring the mutual interactions between EEG rhythms under different emotional expressions, our method can provide valuable insights into the underlying mechanisms of emotion processing and improve the performance of emotion recognition systems.

Graphical Abstract

1. Introduction

Emotion is an essential medium of communication between people and the outside world, reflecting human physiological and psychological states, and influencing behavior and decision-making directly. In recent years, with the rapid development of human–computer interaction (HCI), affective computing has emerged as a research hotspot. Affective computing aims to empower computers with the ability to perceive, understand, and express human emotions, enhancing their intelligence. As the core of affective computing [1], emotion recognition has become increasingly critical. It holds significant importance in diagnosing emotion-related diseases and constructing harmonious and friendly human–computer interaction systems.
Current emotion recognition methods can be mainly divided into two categories. One is based on non-physiological signals, such as facial expressions [2,3], body movements [4] and speech signals [5,6]. The other type is based on physiological signals, such as Electroencephalography (EEG) [7,8], Electrocardiogram (ECG) [9,10], Electromyography (EMG) [11] and galvanic skin response (GSR) [12]. Compared with non-physiological signals, physiological signals have the advantage of being difficult to disguise, so they can reflect people’s emotional states more objectively. Among the numerous physiological signals, EEG is the signal most frequently used for emotion recognition because it can dynamically reflect changes in the central nervous system and contains rich emotional information [13].
In the emotion recognition task, the quality of feature extraction [14] affects its performance directly. For EEG signals, the features commonly used for emotion recognition mainly include time domain [15,16,17], frequency domain [15] and time–frequency domain [15,18] features. Duan proposed a feature called differential entropy ( D E ) [19]. The experimental results on subjects show that differential entropy is an effective feature for identifying emotions and outperforms traditional power spectral density (PSD) [20] features. Kamble [21] developed an integrated eigenvector centrality-variational nonlinear chirp mode decomposition-based EEG rhythm separation (EVNCERS) and analyzed the rhythms associated with several entropy-based features, which were designed to accurately detect affective emotions from various frequency bands of EEG rhythms. Vanitha [22] used the Hilbert–Huang transform (HHT) to detect EEG-related emotional patterns in the time–frequency domain and classified emotions using a multi-class support vector machine classifier (MC-SVM). The results demonstrate the effectiveness of the proposed time–frequency analysis method when it comes to detecting emotions using EEG signals. In addition, some studies suggest that the relationship between EEG channels is also a feature that needs to be considered. Moon [23] identified emotions by extracting the connectivity features between different brain channels, including Pearson correlation coefficient (PCC), phase lock value (PLV) and phase lag index (PLI). They constructed a connectivity matrix as the input of the convolutional neural network (CNN) by calculating the connectivity coefficient between any two EEG channels. The results showed that connectivity between the EEG channels was indeed an important feature of emotion, reflecting the synergy between different brain regions as emotions arise and change. Further, some senior features were extracted. Khare proposed a novel time-order representation (TOR) based on the S-transform for the identification of human emotions. This TOR was given as an input to CNN to automatically extract and classify the deep features [24], which proved that deep features extracted using this method were effective for emotion recognition.
However, the mechanism of mutual interactions between EEG frequency bands under different emotional expressions lacks full exploration. Theoretically, the generation of emotion is accompanied by the common activity of multiple frequency bands. Currently, the main way of information interaction between EEG frequency bands is phase–amplitude coupling (PAC) and it is indicated that PAC is a potential feature for emotion [25]. However, the high computational complexity of PAC is a huge challenge to the real-time performance of the model. Aru [26] indicated that a critical step of the typical approach to analyzing PAC was statistical evaluation. Because there is no direct way to quantify the significance of PAC scores, one needs to rely on statistical inference to reach a conclusion about the statistical significance of the measure. Presently, most studies of PAC rely on the frequentist approach of using surrogate data to estimate a p-value. Hülsemann [27] illustrated the method of surrogate testing. During surrogate testing, the observed coupling value is compared to a distribution of shuffled coupling values. Shuffling is usually repeated 200 to 1000 times, which leads to high computational complexity and is time consuming. In summary, the current research gap is that there is a lack of sufficient research on the cross-frequency coupling, and there is great room for improvement in the extraction of relevant emotional features between frequency bands, whether in terms of model accuracy or time cost.
Therefore, in this paper, we consider the synergy between EEG frequency bands. First, we extracted the DE features of each frequency band of the multi-channel EEG signals. Then, we extracted the correlating features between the EEG frequency band through canonical correlation analysis (CCA) [28,29] based on DE. Finally, we used the DE feature and the correlation feature between frequency bands to train a support vector machine (SVM) classifier, respectively, and integrated the two classifiers to identify emotions through decision-level fusion, which significantly improves the accuracy of emotion recognition compared with that of a single feature.
The other parts of this paper are arranged as follows: In Section 2, we describe in detail the dataset we used and the preprocessing, feature extraction, decision fusion, and cross-validation methods employed when analyzing the data. In Section 3, the experimental results are presented and some analysis and discussion are carried out. Section 4 summarizes the main conclusions and contributions of our study.

2. Materials and Methods

2.1. Datasets and Preprocessing

We used two datasets: the SJTU Emotion EEG Dataset (SEED) [30] and our local Continuous Upgrading Multi-Model Affect Elicitation (CUMULATE) dataset [31].
The SEED dataset consists of 15 participants (7 males and 8 females, mean age 23.27 years, variance 2.37). Each subject participated in three experiments, and in each experiment, they were required to watch 15 emotional film clips to elicit three emotions: positive, neutral, and negative. This dataset recorded EEG signals from 62 electrodes (FPZ, FP1, FP2, AF3, AF4, F7, F5, F3, F1, FZ, F2, F4, F6, F8, FT7, FC5, FC3, FC1, FCZ, FC2, FC4, FC6, FT8, T7, C5, C3, C1, CZ, C2, C4, C6, T8, TP7, CP5, CP3, CP1, CPZ, CP2, CP4, CP6, TP8, P7, P5, P3, P1, PZ, P2, P4, P6, P8, PO7, PO5, PO3, POZ, PO4, PO6, PO8, CB1, O1, OZ, O2, CB2) with ESI NeuroScan System. The acquired EEG signals were then preprocessed. In order to filter out noise and artifacts, the signals were processed using a bandpass filter from 0.3 to 50 Hz, followed by downsampling of the sampling rate from 1000 Hz to 200 Hz. Afterwards the EEG signal corresponding to the video clip was segmented and the data of each channel were divided into 5 s lengths without overlap.
The CUMULATE dataset is our local dataset, containing a three-classes sub-dataset of calm, fear and happy; a four-classes sub-dataset of calm, fear, happy and sad; and a five-classes sub-dataset of calm, fear, happy, sad and angry. Each sub-dataset contained 15 subjects, 5 of whom participated in both rounds of the experiment, so a total of 40 subjects participated in the experiment (30 males and 10 females, mean age 26.20 years, variance 4.74). In each type of sub-dataset, there were five video clips for each emotion. Thus 15, 20, and 25 video clips were viewed by subjects in the three-classes, four-classes, and five-classes sub-dataset, respectively. A total of 60 EEG channels (FPZ, FP1, FP2, AF3, AF4, F7, F5, F3, F1, FZ, F2, F4, F6, F8, FT7, FC5, FC3, FC1, FCZ, FC2, FC4, FC6, FT8, T7, C5, C3, C1, CZ, C2, C4, C6, T8, TP7, CP5, CP3, CP1, CPZ, CP2, CP4, CP6, TP8, P7, P5, P3, P1, PZ, P2, P4, P6, P8, PO7, PO5, PO3, POZ, PO4, PO6, PO8, O1, OZ, O2) were recorded with the ESI NeuroScan System. After that, the same bandpass filtering process of 0.3–50 Hz was performed on the original EEG signal to remove the high-frequency noise, and a notch filter of 50 Hz was used to remove power line interference. The signal was then downsampled to 200 Hz for subsequent processing. To avoid the interference of eye movement artifacts, the eye movement artifacts were removed via independent component analysis (ICA) [32]. Similarly, the pre-processed EEG signal was segmented. The data of each channel were divided into 5 s lengths without overlap.

2.2. Differential Entropy

For EEG signals, D E is a very effective feature, which is a generalized form of Shannon information entropy over continuous variables [19,30]. Assuming that X is a continuous random variable, the differential entropy is calculated using the formula shown in Equation (1):
D E = X p x log p x d x
where p   x is the probability density function of the random variable X . When the random variable X ~ N   ( μ , σ 2 ) , Equation (1) can be rewritten as Equation (2):
D E = 1 2 π σ 2 e x μ 2 2 σ 2 l o g 1 2 π σ 2 e x μ 2 2 σ 2 d x = 1 2 l o g 2 π e σ 2
Studies have shown that for a fixed-length EEG signal, the differential entropy of the EEG signal in a frequency band is equivalent to the logarithm of the energy spectrum in the frequency band [33]. Thus, we can calculate the differential entropy of five frequency bands ( δ : 1–4 Hz, θ : 4–8 Hz, α : 8–13 Hz, β : 13–30 Hz, γ : 30–50 Hz) per channel for each sample. For the SEED dataset, we can extract 310 dimensional features (62 channels × 5 bands) for each sample, and for the CUMULATE dataset, we can extract 300 dimensional features (60 channels × 5 bands) for each sample.

2.3. Canonical Correlation Analysis

CCA is usually used to study the correlation between two sets of multidimensional random variables. It is an extension of the correlation between two random variables under two sets of random variables [34]. For two random variables x and y , the correlation coefficients are defined as follows:
ρ = C o v x , y D x D y
where C o v x , y is the covariance of x , y and D x D y is the variance of x , y . For two sets of random variables X and Y , CCA discusses the correlation by finding a certain linear combination of the two sets of variables, respectively. As shown in Equations (4) and (5):
x = a T X
y = b T Y
where x and y is the linear combination of X and Y . The goal of CCA is to find the values of a and b that maximize the correlation coefficient between x and y . As follows:
argmax a , b ρ = argmax a , b C o v a T X , b T Y D a T X D b T Y
Because X and Y are preprocessed multi-dimensional variables with a mean of 0 and a variance of 1, Equation (6) can be transformed into Equation (7):
argmax a , b a T C o v X , Y b a T D X a b T D Y b
Since the numerator and denominator are increased by the same multiple, the optimization objective result remains unchanged, so our optimization objective can be transformed into the following:
argmax a , b a T C o v X , Y b s . t .     a T D X a = b T D Y b = 1
For the optimization objective of the above formula, it can be solved via SVD decomposition. Let S X Y = C o v X , Y , S X X = D X ,     S Y Y = D Y to simplify the writing. Firstly, Let a = S X X 1 / 2 u , b = S Y Y 1 / 2 v . So:
a T S X X a = u T S X X 1 / 2 S X X S X X 1 / 2 u = u T u = 1
b T S Y Y b = v T S Y Y 1 / 2 S Y Y S Y Y 1 / 2 v = v T v = 1
a T S X Y b = u T S X X 1 / 2 S X Y S Y Y 1 / 2 v
Therefore, our optimization objective transform into Equation (12):
argmax a , b u T S X X 1 / 2 S X Y S Y Y 1 / 2 v s . t .     u T u = v T v = 1
Let M = S X X 1 / 2 S X Y S Y Y 1 / 2 , and treat u and v as left and right singular vectors corresponding to a singular value of matrix M . Performing singular value decomposition on M gives us M = U Σ V T , where U and V are matrices formed by the left and right singular vectors of M , respectively, and Σ is a diagonal matrix composed of the singular values of M . According to the properties of singular value decomposition, all columns of U and V are mutually orthogonal. Thus, we can derive Equation (13):
u T S X X 1 / 2 S X Y S Y Y 1 / 2 v = u T U Σ V T v = σ u v
Therefore, the maximum value of the optimization objective of Equation (12) is the maximum singular value of the matrix M . u and v are the left and right singular vectors corresponding to the largest singular values, respectively. Then, a = S X X 1 / 2 u , b = S Y Y 1 / 2 v can be calculated.

2.4. Inter-Band Correlation Features Based on CCA

CCA is a multivariate statistical analysis method that aims to capture the overall correlation relationship between two sets of indicators. It does this by extracting two representative synthetic variables, U 1 and V , which are linear combinations of the variables in each set. These synthetic variables are then used to reflect the overall correlation between the two sets of indicators.
The CCA algorithm involves the following steps:
1.
Identify the linear combination of variables in each set that maximizes the correlation coefficient between the two sets.
2.
Select an uncorrelated linear combination from the remaining options and pair it with the previously selected linear combination, selecting the one with the highest correlation coefficient.
3.
Repeat steps 1 and 2 until all correlations between the two sets of indicators have been captured in the synthetic variables.
According to the CCA derivation formula in Section 2 and Section 3, when the top k singular values. singular values of the matrix M are selected, u and v are the matrices composed of the left and right singular vectors corresponding to the top k singular values, respectively. Then, a = S X X 1 / 2 u , b = S Y Y 1 / 2 v can be calculated. At this point, Equations (4) and (5) are equivalent to a linear transformation of X and Y , so that the correlation coefficients of the two sets of random variables after the transformation are the top k singular value of the matrix M , and the linearly transformed random variables x and y reflect the correlation characteristics of the two original random variables.
The generation of and changes in emotions usually involve multiple bands of brain activity, and there are synergies between these bands, so the inter-bands correlation (IBC) features can be extracted using CCA. The IBC calculation process is shown in Figure 1.
The DE features of five frequency bands ( δ : 1–4 Hz, θ : 4–8 Hz, α : 8–13 Hz, β : 13–30 Hz, γ : 30–50 Hz) for all channels are represented as X δ , X θ , X α ,   X β , X γ , respectively. DE features of any frequency band can be treated as a set of variables, and the CCA algorithm can be used to extract the correlation between any two frequency bands of DE features. For DE features of any two frequency bands X and Y , the correlation features X and Y between the two frequency bands can be obtained by performing a linear transformation using CCA. As shown in Equations (14) and (15):
X 1 = W 1 T X 1
X 2 = W 2 T X 2
According to [35], we set two parameters α 1 and α 2 , satisfying α 1 + α 2 = 1 , and perform a weighted summation of the correlation features between the two frequency bands obtained above, as shown in Equation (16):
X F = α 1 X + α 2 X 2
Thus, we obtain the IBC features X F between the two frequency bands. As depicted in Figure 1, for X δ and X θ , we can obtain the correlation features X δ and X θ by Equations (14) and (15). Then the IBC features X F δ θ between δ and θ frequency bands can be obtained by Equation (16). For the five frequency bands of the EEG, features similar to X F δ θ can be obtained between any two frequency bands, so we can obtain 10 sets of features like X F δ θ . Finally, we concatenate 10 sets of features including X F δ θ , X F δ α , X F δ β , X F δ γ , X F θ α , X F θ β , X F θ γ , X F α β , X F α γ , X F β γ , which is the final inter-band correlation feature X I B C , as shown in Equation (17):
X I B C = X F δ θ , X F δ α , X F δ β , , X F β γ

2.5. Decision-Level Fusion

Considering that some of the original information may be lost when the IBC features are obtained from the original DE features, a decision-level fusion method [15] is adopted for the original DE features and the IBC features to classify emotions. For each of these two features, a classifier is trained using a linear kernel SVM. The output of the linear kernel SVM is a vector, and each value of the vector means a probability estimate, representing the probability that the sample belongs to a certain category. The SVM-based decision fusion method can improve the accuracy and robustness of classifiers by weighting the results of multiple classifiers. Let ω k ( k = 1,2 , , c ) represent the category of emotions and y D E and y I B C represent the labels of emotions predicted by the classifier, respectively. The result of decision-level fusion is shown as follows:
  y F u s i o n = arg max w k p w k y D E + p w k y I B C     ( k = 1 ,   2 , ,   c )  
For each of the DE features and inter-band correlation (IBC) features, a classifier is trained using a linear kernel SVM. p w k y D E and p   w k y I B C ( k = 1,2 , , c ) represent the confidence of the classifiers trained using DE features and trained using IBC features to identify the k -th emotion, respectively. Then, we take a weighted summation for confidence of the two SVM classifiers outputs and take the labels corresponding to the highest sum of confidence of the two classifiers outputs as the final prediction label.

2.6. Cross-Validation

We use the block-wise cross-validation to evaluate the classification performance. Each block contains one video of each emotion type, and each data set has five blocks. The data of one block are used as the test set, and the data of the remaining blocks are used as the training set. This process is repeated five times to ensure that each block is used as a test set. Finally, the average classification accuracy of the block-wise cross-validation is taken as the evaluation metric.

3. Results

3.1. Validity of IBC Features

To ascertain the validity of the IBC features, we subjected the singular values of matrix M from the previous section to a thorough sorting in descending order. Subsequently, we selected the IBC features corresponding to the top half of these singular values and those corresponding to the bottom half, employing an SVM model for emotion classification. Specifically, for the CUMULATE dataset, we opted for the IBC features corresponding to the first 30 largest singular values and the last 30; whereas, on the SEED dataset, we selected the IBC features corresponding to the first 31 largest singular values and the last 31. In the weighted summation of correlation features across two frequency bands, we assumed equal importance for each band, hence α_1 = 0.5 and α_2 = 0.5. The outcomes of our classification are depicted in Figure 2. Specifically, CUMULATE3, CUMULATE4, and CUMULATE5 signify the three-class, four-class, and five-class datasets of CUMULATE, respectively.
Figure 2 illustrates a comparison of classification accuracy between the first 30/31 IBC features and the last 30/31 IBC features on CUMULATE and SEED datasets. A paired sample t-test was conducted to compare the classification accuracies of the first 30/31 IBC features and the last 30/31 IBC features for both datasets, respectively. The results demonstrate that the classification accuracy of the first 30/31 IBC features is significantly higher than that of the last 30/31 IBC features. For the three sub-datasets of CUMULATE, the average classification accuracy of the first 30 IBC features is 78.34%, 60.77%, and 53.23%, respectively, which is significantly higher than the corresponding average classification accuracies of the last 30 IBC features of 58.70% (p < 0.001), 46.61% (p < 0.001), and 41.27% (p < 0.001). On the SEED dataset, the average classification accuracy of the first 31 IBC features is 70.00%, which is significantly higher than the other half of the features corresponding to 58.83% (p < 0.001).
Indeed, a larger singular value corresponds to a more correlated feature between the frequency bands. Consequently, the first 30/31 IBC features are better suited to reflect the correlations between the two frequency bands compared to the last 30/31 IBC features. It can be inferred that the more correlated features between the frequency bands are more conducive to the recognition of emotions, hence the features that exhibit inter-band correlations may potentially encapsulate emotional information.
Furthermore, we investigate the effectiveness of training an SVM classifier with the IBC features corresponding to the top k largest singular values of M . The change curve of classification accuracy is plotted as the value of k gradually increases on both the local CUMULATE dataset (Figure 3a–c) and the public dataset SEED (Figure 3d). The dashed line and thick solid line indicate the variation in emotion recognition accuracy for each individual subject and the average classification accuracy fluctuation across all subjects, respectively.
As depicted in Figure 3, the classification accuracy demonstrates a pattern of sharp increase followed by a tendency towards flattening out across different datasets. This further underscores the notion that the correlation between frequency bands plays a crucial role in emotion recognition, with features corresponding to larger singular values being more influential. On the other hand, IBC features with smaller singular values may not be as significant in emotion recognition.
Therefore, we conducted an analysis to investigate the correlation between the IBC features corresponding to the k th singular values and the emotion labels. The mean value of the correlation coefficient between each feature and emotion label was calculated across subjects, as shown in Figure 4.
It can be observed that the IBC features corresponding to larger singular values, which implies stronger correlations between frequency bands, exhibit a stronger correlation with emotion labels. On the other hand, the features corresponding to smaller singular values, which indicate less correlation between frequency bands, display a weaker correlation with emotion labels. This finding substantiates the notion that more correlated features between frequency bands may encompass specific emotional characteristics.
Furthermore, we designate the correlation coefficient between IBC features and emotion labels as ICC, and the correlation coefficient among different frequency bands mentioned above as DCC. To verify the validity of IBC features, we conduct an analysis of the correlation between ICC and DCC.
Table 1 and Table 2 present the correlation coefficient (CC) between ICC and DCC on the CUMULATE dataset and SEED dataset, respectively. Our analysis reveals a strong positive correlation between DCC and ICC on most subjects across both datasets. This finding indicates that as the correlation coefficient among different frequency bands increases, the corresponding IBC features exhibit a stronger association with emotion labels. We posit that emotions are intricately linked to the synergy between frequency bands, and the correlation among different frequency bands is driven by emotional factors. The IBC features corresponding to higher correlation coefficients among different frequency bands are characterized by a rich abundance of emotional information.

3.2. Classification Results

Based on the findings from the previous subsection, we select the IBC features corresponding to the top 30/31 singular values to train an SVM classifier. Additionally, we train another SVM classifier using the original DE features. Finally, we apply the decision-level fusion sum rule to combine the outputs of both classifiers for emotion classification.
Figure 5 displays the emotion recognition accuracy utilizing DE features, IBC features, and their decision-level fusion on both CUMULATE and SEED datasets. Notably, there is no significant difference in the accuracy of emotion recognition between the weak classifiers trained on the original DE feature and IBC feature. However, upon fusing DE and IBC features at the decision level, the classification accuracy surpasses the results obtained from individual features. For the three sub-datasets of CUMULATE, the average classification accuracies after fusion are 81.36%, 63.88%, and 55.77%, which are significantly higher than the average classification accuracies of 80.02% (p < 0.05), 61.86% (p < 0.05), and 53.61% (p < 0.001) using DE features alone as well as significantly higher than the average classification accuracies of mean classification accuracy of 78.34% (p < 0.001), 60.77% (p < 0.001), and 53.23% (p < 0.001) using IBC features alone. The same results hold true for the SEED dataset. The average classification accuracy after fusion is 72.35%, which is significantly higher than the average classification accuracy of 71.49% (p < 0.01) using DE features alone and higher than the average classification accuracy of 70.00% (p < 0.001) using IBC features alone. This outcome demonstrates that IBC features provide supplementary emotional information via decision-level fusion, thereby enhancing the accuracy of emotion recognition.
In addition, we examine the classification accuracy of IBC features across different pairs of frequency bands. The results are presented in Figure 6, where subplots (a), (b), (c), and (d) correspond to the CUMULATE3, CUMULATE4, CUMULATE5, and SEED datasets, respectively. The diagonal elements of each subplot have been set to zero for clarity and are not considered in the analysis. Each cell in the subplot represents the classification accuracy of IBC features across different pairs of frequency bands.
The same trend is observed in all four subplots of Figure 6. The colors in the figure from light to dark means an increasing level of accuracy from low to high. The upper left part of each subplot exhibits low classification accuracy, while the lower right part demonstrates higher accuracy. This finding suggests that IBC features from higher frequency bands are more effective for emotion recognition. In particular, the correlation features between the β band and γ band yield the highest classification accuracy on each dataset, while the correlation features between the δ band and θ band provide the lowest accuracy. This result also reinforces the notion that the β and γ bands contain richer emotional information and are more effective for emotion recognition, which is consistent with the findings in [30].

4. Discussion

4.1. The Validity of IBC Features

We conducted four experiments to demonstrate the effectiveness of IBC feature extraction. Firstly, we compared the emotion classification results obtained from the IBC features corresponding to the largest half of the singular values and the IBC features corresponding to the other half of the singular values. The results showed that the emotion recognition accuracy for the first 30/31 IBC features was significantly higher than that of the last 30/31 IBC features. This indicates that most of the information related to emotions is concentrated in the first 30/31 IBC features after CCA. The emotional information contained in the last 30/31 IBC features is relatively insufficient. To further investigate this, we conducted a second experiment by selecting the IBC features corresponding to the top k largest singular values and observed the change in classification accuracy as the value of k increased gradually. The trend of change in classification accuracy was consistent across all datasets, with rapid growth at the beginning and then stabilization. This further confirms that the more correlated features between frequency bands have a greater impact on emotions, while features with weaker correlations between frequency bands contain less emotional information. In a third experiment, we directly explored the correlation between the k -th correlation feature and emotional labels. It was observed that features with a higher correlation between frequency bands had a greater correlation coefficient with emotional labels and may contain more emotional information. Finally, we performed an experiment to investigate the correlation between ICC and DCC. Our results suggest that the cross-frequency coupling features extracted using our method are potential features for emotion, and inter-bands coupling is driven by emotions.

4.2. Information Complementarity between IBC and DE Features

We compared the performance of emotion recognition using only DE features, IBC features, and their decision fusion on different datasets. Our results showed that the combination of DE and IBC features significantly improved the accuracy of emotion recognition compared to using either feature alone. This indicates that some information about emotions is lost when performing the IBC feature extraction via CCA based on original DE features. However, there is also information that is not included in the original DE features but is supplied by the IBC features. Therefore, the combination of these two types of features provides a complementary view of emotions, leading to better a performance in terms of emotion recognition.

4.3. Computational Complexity of IBC Features

Our method utilizes the SVM model and CCA for cross-frequency coupling feature extraction, avoiding surrogate testing. In terms of computational complexity, our method is more efficient than PAC coupling feature extraction, as it shortens the training time of the model. This not only improves the real-time performance of the model but also enhances its value in engineering applications. Additionally, our method contributes to higher accuracy in terms of emotion recognition.

4.4. Future Directions

Our proposed method has demonstrated superior results, indicating that the relevance of features across frequency bands plays a crucial role in emotion recognition. Furthermore, the underlying physiological mechanism is an important area for future research. Additionally, our feature extraction method reduces computational costs and enhances the real-time performance of the model. However, it should be noted that our approach is limited to emotion recognition research based on effective DE features. In the future, we aim to expand our algorithm ideas and apply them to other fields based on EEG signals, such as depression, anxiety [36], upper limb applications [37], and detection of epileptic seizures [38]. In doing so, we hope to contribute to a wider range of applications using EEG signals while advancing the field of emotion recognition.

5. Conclusions

In this paper, we present a novel approach for extracting emotional cross-band features based on differential entropy. Our method differs from traditional inter-band feature extraction methods such as phase–amplitude coupling (PAC) in that it relies on machine learning and avoids the need for surrogate testing, which can be time-consuming. This not only enhances the efficiency of our model but also improves its real-time performance. To validate the effectiveness of our proposed IBC features, we conducted experiments using two publicly available datasets: SEED and CUMULATE. The results demonstrate that combining cross-band DE features with traditional DE features outperforms methods using two single features alone, substantiating the effectiveness of our method for emotion recognition.

Author Contributions

Conceptualization, Z.W.; methodology, Z.W.; software, Z.W. and R.H.; validation, Z.W.; formal analysis, Z.W. and R.H.; investigation, R.H.; resources, Y.Y., Z.L. and S.Z.; data curation, Z.W.; writing—original draft preparation, Z.W.; writing—review and editing, Z.L. and S.Z.; visualization, Z.W.; supervision, Z.L., B.W. and J.J.; project administration, L.X. and E.Y.; funding acquisition, E.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the grants from the National Natural Science Foundation of China under Grant 62332019, 62076250 and 62176090; in part by STI 2030-major projects 2022ZD0208900; in part by Shanghai Municipal Science and Technology Major Project under Grant 2021SHZDZX; in part by the Program of Introducing Talents of Discipline to Universities through the 111 Project under Grant B17017. This research was also supported by National Government GuidedSpecial Funds for Local Science and Technology Development (Shenzhen, China) (No. 2021Szvup043) and by Project of Jiangsu Province Science and Technology Plan Special Fund in 2022 (Key research and development plan industry foresight and key core technologies) under Grant BE2022064-1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The local data CUMULATE that support the findings of this study are available on request from the corresponding author and the data are not publicly available due to privacy or ethical restrictions. The SEED dataset is available at https://bcmi.sjtu.edu.cn/home/seed/ (accessed on 1 March 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Picard, R.W. Affective Computing; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  2. Liu, Y.-J.; Zhang, J.-K.; Yan, W.-J.; Wang, S.-J.; Zhao, G.; Fu, X. A main directional mean optical flow feature for spontaneous micro-expression recognition. IEEE Trans. Affect. Comput. 2015, 7, 299–310. [Google Scholar] [CrossRef]
  3. Michel, P.; El Kaliouby, R. Real time facial expression recognition in video using support vector machines. In Proceedings of the 5th International Conference on Multimodal Interfaces, Vancouver, BC, Canada, 5–7 November 2003. [Google Scholar]
  4. Yan, J.; Zheng, W.; Xin, M.; Yan, W. Integrating facial expression and body gesture in videos for emotion recognition. IEICE Trans. Inf. Syst. 2014, 97, 610–613. [Google Scholar] [CrossRef]
  5. Mencattini, A.; Martinelli, E.; Ringeval, F.; Schuller, B.; Di Natale, C. Continuous estimation of emotions in speech by dynamic cooperative speaker models. IEEE Trans. Affect. Comput. 2016, 8, 314–327. [Google Scholar] [CrossRef]
  6. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  7. Wang, X.-W.; Nie, D.; Lu, B.-L. Emotional state classification from EEG data using machine learning approach. Neurocomputing 2014, 129, 94–106. [Google Scholar] [CrossRef]
  8. Chao, H.; Zhi, H.; Dong, L.; Liu, Y. Recognition of emotions using multichannel EEG data and DBN-GC-based ensemble deep learning framework. Comput. Intell. Neurosci. 2018, 2018, 9750904. [Google Scholar] [CrossRef] [PubMed]
  9. Agrafioti, F.; Hatzinakos, D.; Anderson, A.K. ECG pattern analysis for emotion detection. IEEE Trans. Affect. Comput. 2011, 3, 102–115. [Google Scholar] [CrossRef]
  10. Tivatansakul, S.; Ohkura, M. Emotion recognition using ECG signals with local pattern description methods. Int. J. Affect. Eng. 2016, 15, 51–61. [Google Scholar] [CrossRef]
  11. Jerritta, S.; Murugappan, M.; Wan, K.; Yaacob, S. Emotion recognition from facial EMG signals using higher order statistics and principal component analysis. J. Chin. Inst. Eng. 2014, 37, 385–394. [Google Scholar] [CrossRef]
  12. Wu, G.; Liu, G.; Hao, M. The analysis of emotion recognition from GSR based on PSO. In Proceedings of the 2010 International Symposium on Intelligence Information Processing and Trusted Computing 2010, Huanggang, China, 28–29 October 2010; pp. 360–363. [Google Scholar]
  13. Heller, W. Neuropsychological mechanisms of individual differences in emotion, personality, and arousal. Neuropsychology 1993, 7, 476–489. [Google Scholar] [CrossRef]
  14. Jenke, R.; Peer, A.; Buss, M. Feature extraction and selection for emotion recognition from EEG. IEEE Trans. Affect. Comput. 2014, 5, 327–339. [Google Scholar] [CrossRef]
  15. Kamble, K.; Sengupta, J. A comprehensive survey on emotion recognition based on electroencephalograph (EEG) signals. Multimed. Tools Appl. 2023, 82, 27269–27304. [Google Scholar] [CrossRef]
  16. Yuen, C.T.; San San, W.; Seong, T.C.; Rizon, M. Classification of human emotions from EEG signals using statistical features and neural network. Int. J. Integr. Eng. 2009, 1, 13. [Google Scholar]
  17. Liu, Y.; Sourina, O. Real-time fractal-based valence level recognition from EEG. In Transactions on Computational Science XVIII: Special Issue on Cyberworlds; Gavrilova, M.L., Tan, C.J.K., Kuijper, A., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2013; pp. 101–120. [Google Scholar]
  18. Hadjidimitriou, S.K.; Hadjileontiadis, L.J. Toward an EEG-based recognition of music liking using time-frequency analysis. IEEE Trans. Biomed. Eng. 2012, 59, 3498–3510. [Google Scholar] [CrossRef] [PubMed]
  19. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013. [Google Scholar]
  20. Wang, X.-W.; Nie, D.; Lu, B.-L. EEG-based emotion recognition using frequency domain features and support vector machines. In Proceedings of the International Conference on Neural Information Processing, Shanghai, China, 13–17 November 2011. [Google Scholar]
  21. Kamble, K.S.; Sengupta, J. EVNCERS: An integrated Eigenvector Centrality-Variational Nonlinear Chirp Mode Decomposition-Based EEG Rhythm Separation for Automatic Emotion Recognition. IEEE Sens. J. 2023, 23, 21661–21669. [Google Scholar] [CrossRef]
  22. Vanitha, V.; Krishnan, P. Time-frequency analysis of EEG for improved classification of emotion. Int. J. Biomed. Eng. Technol. 2017, 23, 191–212. [Google Scholar] [CrossRef]
  23. Moon, S.-E.; Jang, S.; Lee, J.-S. Convolutional neural network approach for EEG-based emotion recognition using brain connectivity and its spatial information. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018. [Google Scholar]
  24. Khare, S.K.; Nishad, A.; Upadhyay, A.; Bajaj, V. Classification of emotions from EEG signals using time-order representation based on the S-transform and convolutional neural network. Electron. Lett. 2020, 56, 1359–1361. [Google Scholar] [CrossRef]
  25. Zhang, C.; Yeh, C.-H.; Shi, W. Variational phase-amplitude coupling characterizes signatures of anterior cortex under emotional processing. IEEE J. Biomed. Health Inform. 2023, 27, 1935–1945. [Google Scholar] [CrossRef]
  26. Aru, J.; Aru, J.; Priesemann, V.; Wibral, M.; Lana, L.; Pipa, G.; Singer, W.; Vicente, R. Untangling cross-frequency coupling in neuroscience. Cur. Opin. Neurobiol. 2015, 31, 51–61. [Google Scholar] [CrossRef]
  27. Hülsemann, M.J.; Naumann, E.; Rasch, B. Quantification of phase-amplitude coupling in neuronal oscillations: Comparison of phase-locking value, mean vector length, modulation index, and generalized-linear-modeling-cross-frequency-coupling. Front. Neurosci. 2019, 13, 573. [Google Scholar] [CrossRef]
  28. Vergult, A.; De Clercq, W.; Palmini, A.; Vanrumste, B.; Dupont, P.; Van Huffel, S.; Van Paesschen, W. Improving the interpretation of ictal scalp EEG: BSS–CCA algorithm for muscle artifact removal. Epilepsia 2007, 48, 950–958. [Google Scholar] [CrossRef]
  29. Liu, Q.; Jiao, Y.; Miao, Y.; Zuo, C.; Wang, X.; Cichocki, A.; Jin, J. Efficient representations of EEG signals for SSVEP frequency recognition based on deep multiset CCA. Neurocomputing 2020, 378, 36–44. [Google Scholar] [CrossRef]
  30. Zheng, W.-L.; Lu, B.-L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  31. Wang, X.; Pei, Y.; Zhang, L.; Zhao, S.; Luo, Z.; Yan, Y.; Yin, E. The increasing types of negative emotion impose an extra pressure on the accuracy of emotion recognition. In Proceedings of the 2022 International Conference on Computer, Artificial Intelligence, and Control Engineering, Zhuhai, China, 25–27 February 2022; Volume 12288, pp. 259–266. [Google Scholar]
  32. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef] [PubMed]
  33. Shi, L.C.; Jiao, Y.Y.; Lu, B.L. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Osaka, Japan, 3–7 July 2013. [Google Scholar]
  34. Press, W.H. Canonical Correlation Clarified by Singular Value Decomposition; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  35. Liu, W.; Qiu, J.-L.; Zheng, W.-L.; Lu, B.-L. Comparing recognition performance and robustness of multimodal deep learning models for multimodal emotion recognition. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 715–729. [Google Scholar] [CrossRef]
  36. Khare, S.K.; March, S.; Barua, P.D.; Gadre, V.M.; Acharya, U.R. Application of data fusion for automated detection of children with developmental and mental disorders: A systematic review of the last decade. Inf. Fusion 2023, 99, 101898. [Google Scholar] [CrossRef]
  37. Khan, S.M.; Khan, A.A.; Farooq, O. Selection of features and classifiers for EMG-EEG-based upper limb assistive devices—A review. IEEE Rev. Biomed. Eng. 2019, 13, 248–260. [Google Scholar] [CrossRef]
  38. Rabcan, J.; Levashenko, V.; Zaitseva, E.; Kvassay, M. Review of methods for EEG signal classification and development of new fuzzy classification-based approach. IEEE Access 2020, 8, 189720–189734. [Google Scholar] [CrossRef]
Figure 1. IBC feature extraction based on original DE features.
Figure 1. IBC feature extraction based on original DE features.
Bioengineering 10 01200 g001
Figure 2. Comparison of classification accuracy between the first 30/31 IBC features and the last 30/31 IBC features on CUMULATE and SEED datasets, *** p < 0.001.
Figure 2. Comparison of classification accuracy between the first 30/31 IBC features and the last 30/31 IBC features on CUMULATE and SEED datasets, *** p < 0.001.
Bioengineering 10 01200 g002
Figure 3. Variation curve of classification accuracy with k value on CUMULATE and SEED datasets (a) CUMULATE3 dataset (b) CUMULATE4 dataset (c) CUMULATE5 dataset (d) SEED dataset.
Figure 3. Variation curve of classification accuracy with k value on CUMULATE and SEED datasets (a) CUMULATE3 dataset (b) CUMULATE4 dataset (c) CUMULATE5 dataset (d) SEED dataset.
Bioengineering 10 01200 g003
Figure 4. Correlation coefficient between the k -th IBC feature and emotion label.
Figure 4. Correlation coefficient between the k -th IBC feature and emotion label.
Bioengineering 10 01200 g004
Figure 5. Emotion recognition accuracy using DE features, IBC features and their decision-level fusion on local CUMULATE dataset and SEED dataset, * p < 0.05, ** p < 0.01, *** p < 0.001.
Figure 5. Emotion recognition accuracy using DE features, IBC features and their decision-level fusion on local CUMULATE dataset and SEED dataset, * p < 0.05, ** p < 0.01, *** p < 0.001.
Bioengineering 10 01200 g005
Figure 6. Classification accuracy of the IBC features between any two different frequency bands on local CUMULATE dataset and SEED dataset (a) CUMULATE3 dataset (b) CUMULATE4 dataset (c) CUMULATE5 dataset (d) SEED dataset.
Figure 6. Classification accuracy of the IBC features between any two different frequency bands on local CUMULATE dataset and SEED dataset (a) CUMULATE3 dataset (b) CUMULATE4 dataset (c) CUMULATE5 dataset (d) SEED dataset.
Bioengineering 10 01200 g006
Table 1. The correlation coefficient (CC) between ICC and DCC on CUMULATE dataset.
Table 1. The correlation coefficient (CC) between ICC and DCC on CUMULATE dataset.
CUMULATE3CUMULATE4CUMULATE5
SubjectCCSubjectCCSubjectCC
S10.71S10.70S10.76
S20.75S20.76S20.65
S30.74S30.72S30.80
S40.66S40.57S40.72
S50.76S50.72S50.81
S60.72S60.72S60.83
S70.76S70.67S70.77
S80.73S80.66S80.54
S90.74S90.61S90.59
S100.81S100.69S100.75
S110.72S110.75S110.68
S120.73S120.62S120.74
S130.84S130.84S130.83
S140.68S140.63S140.69
S150.85S150.54S150.77
AVG0.75AVG0.68AVG0.73
STD0.05STD0.07STD0.09
Table 2. The correlation coefficient (CC) between ICC and DCC on SEED dataset.
Table 2. The correlation coefficient (CC) between ICC and DCC on SEED dataset.
SEEDday1SEEDday2SEEDday3
SubjectCCSubjectCCSubjectCC
S10.75S10.66S10.51
S20.53S20.44S20.75
S30.34S30.66S30.59
S40.78S40.76S40.49
S50.75S50.82S50.83
S60.59S60.71S60.67
S70.75S70.81S70.67
S80.58S80.68S80.77
S90.82S90.81S90.79
S100.70S100.54S100.78
S110.77S110.76S110.75
S120.81S120.74S120.75
S130.72S130.76S130.72
S140.79S140.66S140.67
S150.71S150.66S150.71
AVG0.69AVG0.70AVG0.70
STD0.13STD0.10STD0.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Huang, R.; Yan, Y.; Luo, Z.; Zhao, S.; Wang, B.; Jin, J.; Xie, L.; Yin, E. An Improved Canonical Correlation Analysis for EEG Inter-Band Correlation Extraction. Bioengineering 2023, 10, 1200. https://doi.org/10.3390/bioengineering10101200

AMA Style

Wang Z, Huang R, Yan Y, Luo Z, Zhao S, Wang B, Jin J, Xie L, Yin E. An Improved Canonical Correlation Analysis for EEG Inter-Band Correlation Extraction. Bioengineering. 2023; 10(10):1200. https://doi.org/10.3390/bioengineering10101200

Chicago/Turabian Style

Wang, Zishan, Ruqiang Huang, Ye Yan, Zhiguo Luo, Shaokai Zhao, Bei Wang, Jing Jin, Liang Xie, and Erwei Yin. 2023. "An Improved Canonical Correlation Analysis for EEG Inter-Band Correlation Extraction" Bioengineering 10, no. 10: 1200. https://doi.org/10.3390/bioengineering10101200

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop