Next Article in Journal
Correction: Tong et al. Reduced Apoptotic Injury by Phenothiazine in Ischemic Stroke Through the NOX-Akt/PKC Pathway. Brain Sci. 2019, 9, 378
Previous Article in Journal
Assessment of Patterns of Infiltration and Relapse of Patients with Glioblastoma of the Occipital Lobe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ICTD: Combination of Improved CNN–Transformer and Enhanced Deep Canonical Correlation Analysis for Eye-Movement Emotion Classification

1
School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
Shunde Innovation School, University of Science and Technology Beijing, Foshan 528399, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2026, 16(3), 330; https://doi.org/10.3390/brainsci16030330
Submission received: 15 February 2026 / Revised: 14 March 2026 / Accepted: 17 March 2026 / Published: 19 March 2026
(This article belongs to the Section Cognitive, Social and Affective Neuroscience)

Highlights

What are the main findings?
  • This paper proposes a deep canonical correlation analysis method based on cosine similarity, non-linearly transforming feature vectors of different modalities into feature vectors with stronger correlation to improve the accuracy of emotion classification.
  • This paper proposes an incremental feature feedforward network (IFFN) to perform feature transformations in enhancement and simplification, replacing the FFN in the original transformer module.
What are the implications of the main findings?
  • Cosine similarity pays more attention to the direction of vectors than it does to their magnitude, is affected less by outliers, and does not require the data to satisfy specific distribution assumptions. The characteristics that are more suitable for eye-movement input data have provided a suitable processing method for eye-movement-based emotion classification.
  • Existing studies mostly rely on the statistical characteristics of the original data for emotion analysis. Moreover, the importance of each emotional feature in the calculation process is not primary or secondary, and the role of key features cannot be highlighted. By assigning higher weights to key features, the influence of indiscriminately input features is reduced, thereby enhancing the importance of key features. This design effectively addresses the lack of prioritization in feature importance and significantly improves the ability of eye-movement features to characterize emotional states.

Abstract

Background/Objectives: Emotion classification based on eye-movement features has become a widely adopted approach due to the simplicity of data acquisition and the strong association between ocular responses and emotional states. However, several challenges remain with regard to existing emotion recognition methods, including the relatively weak correlation between eye-movement features and emotional labels and the fact that the key features are not prominently presented. Methods: To address abovelimitations, this study proposes an improved CNN-transformer combined with enhanced deep canonical correlation analysis network (ICTD). The proposed method first performs preprocessing and reconstruction of raw eye-movement signals to extract informative features. Subsequently, convolutional neural networks (CNNs) and transformer architectures are employed to capture local and global feature, respectively. In addition, an incremental feature feedforward network is incorporated to enhance the transformer, enabling the model to assign higher importance to salient feature information. Finally, the extracted representations are processed through deep canonical correlation analysis based on cosine similarity in order to generate classification outcomes. Results: Experiments conducted on the SEED-IV, SEED-V, and eSEE-d datasets demonstrate that the proposed ICTD framework consistently outperforms baseline approaches and attains optimal classification results. (1) On the eSEE-d dataset, the results of three-category arousal and valence classification reach 81.8% and 85.2%, respectively; (2) on the SEED-IV dataset, the emotion four-category classification result reaches 91.2%; (3) finally, on the SEED-V dataset, the emotion five-category classification result reaches 85.1%. Conclusions: The proposed ICTD framework effectively improves feature representation and classification performance, showing strong potential for practical emotion recognition and physiological signal analysis.

1. Introduction

Emotion is a subjective physiological phenomenon of humans, and the expression of emotion is multifaceted. Accurately identifying emotional states plays a crucial role in cognitive science [1]. Picard [2] proposed the concept of affective computing, aiming to research and develop systems and devices capable of recognizing, interpreting, processing, and simulating human emotions [3]. Emotion recognition in humans has become an important area of research within affective computing, significantly contributing to both the management of emotional states and the identification of disorders linked to emotions [4].
In recent years, researchers have conducted a series of studies on the acquisition and classification of emotional states [5,6,7,8]. Currently, methods for acquiring emotional data are categorized into non-contact- and contact-based techniques. Contact methods typically involve physiological sensors such as electroencephalography (EEG), electrocardiography (ECG), and electrodermal activity (EDA). It is worth mentioning the functional near-infrared spectroscopy (fNIRS) technique for emotion detection, which has grown in popularity in recent years. Their main characteristics are the need to fix devices on the subject and the high requirements for the acquisition environment [9,10,11]. Non-contact methods primarily involve the analysis of voice, facial expressions, and eye-tracking signals. These approaches offer the benefits of straightforward signal collection while ensuring that participants experience no discomfort [12,13]. The study of eye-tracking signals has become an ideal research subject in affective computing due to the ease of their acquisition and their direct correlation with emotions. Acquiring eye-tracking signals does not require wearing cumbersome devices like EEG or ECG signals. Eye-tracking data primarily include fixation, saccade, and pupil-related information. The frequency and duration of fixations provide direct insights into the subject’s level of attention and focus on specific stimuli. Saccades, being the most frequent type of eye-movement, exhibit variations in speed and duration, which reflect different levels of attention directed toward distinct targets. Changes in pupil size are positively correlated with the current stimulus experienced by the subject. When individuals are in different emotional states, such as happiness, sadness or fear, their eye-tracking behavior shows systematic differences that are specifically reflected in fixation duration, saccade speed, and pupil diameter, among other aspects. For example, individuals in a happy state often exhibit longer fixation times on positive stimulus areas, while, in a fearful state, their pupils tend to dilate significantly, something which is accompanied by high-frequency saccadic behavior [14]. The combined influence of eye-movement characteristics and the inherent relationships between emotions and these features demonstrates that emotion recognition based on eye-tracking is both feasible and scientifically grounded. In recent years, a range of deep neural networks has been applied to affective computing, and their outstanding performance indicates that they surpass traditional shallow approaches in effectiveness [15]. Although eye-tracking features demonstrate numerous advantages in affective computing, there are still some major issues that limit the improvement of emotional computing accuracy.
This study proposes an improved CNN–transformer framework integrated with an enhanced deep canonical correlation analysis network for emotion recognition. The proposed approach comprehensively incorporates multiple eye-movement features into the emotion classification process, enabling more effective feature extraction and classification. These features provide richer representations of emotional characteristics and contribute to improved discrimination among different emotional states. The CNN–transformer architecture is employed to capture both local and global feature information. In addition, the conventional transformer structure is enhanced by introducing an incremental feature feedforward network, which allows important feature information to receive greater attention during feature learning. Finally, deep canonical correlation analysis based on cosine similarity is applied to the extracted features, and the final classification results are obtained by measuring the correlations between the learned feature representations.
The remaining part of this paper is structured as follows. Section 2 provides a summary of the common methods and network models for eye-movement emotion analysis. Section 3 details the overall structure of the ICTD algorithm and the implementation methods of each module. Section 4 introduces the dataset employed in the experiments, the experimental results, and the comparison results with baseline methods. Section 5 discusses the contributions of ICTD as well as its existing shortcomings. Finally, Section 6 introduces the conclusions and future work.

2. Related Work

Eye-tracking technology has been widely applied across numerous domains, including applied psychology, neuropsychology, and medical research [16,17,18,19], and has become an increasingly prominent analytical tool in recent years. Existing studies on emotion recognition primarily focus on improving classification performance through multimodal data integration [20], while comparatively limited attention has been devoted to exploring the intrinsic information embedded in eye-tracking signals themselves. As a result, although eye-tracking has been extensively utilized as a supporting modality in emotion analysis, research relying solely on eye-tracking parameters for emotion recognition remains relatively limited and is still an emerging methodological direction. Therefore, this section aims to systematically review the current research progress and identify the key challenges associated with emotion analysis based exclusively on eye-tracking parameters.

2.1. Eye-Movement Features for Emotion Recognition

Given the widespread interest in emotion recognition tasks using eye-tracking features, an increasing number of methods for assessing emotional states have been proposed. In the research on the affective analysis of eye-tracking features, Oliva et al. [21] investigated how variations in pupil diameter relate to emotion recognition using nonverbal vocal stimuli, and their findings indicated that pupil size changes are associated with cognitive processing [22]. Changes in baseline pupil size are related to task efficiency. The pupil diameter increases during task interruptions, while it decreases during task engagement. The study aimed to test stimuli with varying values, intensity, duration, and recognition difficulty. Aracena et al. [23] randomly selected 90 images from the IAPS dataset and categorized them into three emotional classes: positive, neutral, and negative, while also evaluating them according to three corresponding valence levels. Classification was subsequently conducted using neural networks [24] and binary decision trees. The best performance, obtained with DeepLearntools [25], reached a recognition rate of 82.8%, with a mean accuracy of 71.7%. Skaramagkas et al. [26] used eye-tracking features such as fixation duration, saccade amplitude, etc., and selected 48 participants to watch 10 emotion-inducing videos, recording their ocular responses. The authors classified emotions based on statistical features of the eye ocular responses, including fixation and blink frequency, saccade amplitude and duration, pupil diameter, fixation duration kurtosis, saccade duration variation and pupil diameter kurtosis. Participants evaluated four emotional states: tenderness, anger, disgust, and sadness, and a deep multilayer perceptron (DMLP) model achieved 92% accuracy in discriminating positive versus non-positive valence, as well as 84% accuracy in differentiating low from medium arousal.

2.2. Ocular Response-Based Emotional Computing for Differentiated Individuals

Emotional data collected from different subjects typically exhibits high individual variability, and current research focuses on model generalization studies to reduce inter-individual differences. Fu et al. [27] proposed a multimodal feature fusion neural network (MFFNN) designed to effectively extract complementary information from eye-movement signals and integrate it with EEG features. Their architecture employs a dual-branch feature extraction module to capture features from both modalities while maintaining temporal alignment. Additionally, a multi-scale feature fusion module is incorporated, which leverages cross-channel soft attention to selectively aggregate information from different spatial scales, thereby enabling effective multi-scale feature integration. Experiments conducted on the publicly available SEED-IV dataset demonstrate that the proposed model achieves an accuracy of 87.32% for four-class emotion recognition, encompassing happiness, sadness, fear, and neutral. Li et al. [28] employed a deep gradient convolutional neural network (DGCNN) by inputting eye-tracking signal images processed with FFT into the network, achieving an 87% accuracy in recognizing four emotional categories: nervous, calm, happy, and sad. Tarnowski et al. [29] utilized 18 eye-movement features to classify emotional states, including high arousal/negative valence, low arousal/medium valence, and high arousal/positive valence. The study employed three classifiers—support vector machines, linear discriminant analysis, and k-nearest neighbors—and evaluated performance using a leave-one-subject-out cross-validation approach. The results indicated a peak classification accuracy of 80%, with the strongest performance observed in distinguishing high and low arousal emotions. Wang et al. [30] collected and analyzed electrooculogram (EOG) signals and eye-tracking videos simultaneously, proposing an emotion perception algorithm based on eye-tracking information. They applied short-time Fourier transform (STFT) to the original multi-channel EOG signals to extract time–frequency eye-tracking features, integrated time–domain eye-tracking features using two feature fusion strategies, and conducted recognition experiments on positive, neutral, and negative emotional states, achieving average respective accuracies of 88.64% and 88.35%.

2.3. Multimodal Emotion Analysis Model with Eye-Tracking Data

In recent years, scholars have carried out research on emotion analysis by utilizing multimodal physiological signals including electroencephalogram (EEG) signals, electrocardiogram (ECG) signals, electrooculogram (EOG) signals, skin conductance, and eye-movement signals. Among these, the combination of EEG and eye-movement for emotion analysis is a relatively prevalent approach. Gong et al. [31] put forward a novel multimodal emotion recognition model, named integrity-induced adaptive broad learning (CiABL), to enhance the generalization performance for unseen target domain topics. This introduces a weighted representation distribution alignment mechanism that is capable of effectively aligning both marginal and conditional distributions, thereby mitigating the influence of individual differences. Experiments on the SJTU emotion EEG dataset (SEED) and SEED-FRA datasets verified its effectiveness and generalization ability, and it outperformed the current mainstream methods. Magdiel proposed a multimodal semi-supervised domain adaptation method, namely cross modal joint distribution alignment (CMJDA), which conducts cross modal learning and joint distribution alignment by leveraging the discriminative features of each modality through independent neural networks [32]. The model generates correlated features and consistent predictions between modalities and promotes the similarity of marginal and conditional distributions among labeled source data, limited labeled target data, and a large amount of unlabeled target data. Comparative experiments were conducted via “Leave-One-Subject -Out” cross-validation on two public emotion recognition datasets, SEED-IV and SEED-V. An average accuracy ranging from 92.50% to 96.13% was achieved in three available sessions of SEED-IV and SEED-V. Liu et al. [33] studied how people from China, Germany, and France recognize emotions by analyzing both EEG signals and eye-movement data. A systematic analysis was performed across four dimensions: emotion model performance, neural patterns, complementary information from multiple modalities, and cross-cultural emotion recognition to evaluate emotion recognition accuracy. The study concluded that, in terms of cross-cultural emotion recognition, German and French participants exhibit greater similarity in terrain patterns and attention weight distributions when compared with Chinese participants. Additionally, while Chinese data align well with test datasets, it is less suitable for training models for the other two cultural groups.
Based on the above research, the research methods and characteristics of different ocular response-based emotion classification studies are summarized as shown in Table 1.
From the above studies, several challenges remain in the field of eye-movement-based emotion classification.
(1) The correlation between eye-movement features and emotional states remains relatively weak. Emotional responses are complex and cannot be accurately represented by only a limited number of eye-movement features. In practice, the same eye-movement feature may exhibit different patterns across different emotional states [34]. Therefore, further investigation is required to better understand the relationships between emotional states and eye-movement features, as well as to enhance and identify discriminative feature representations to improve classification performance.
(2) Eye-movement features are often treated merely as auxiliary information in multimodal emotion recognition frameworks, which limits their independent analytical potential. For example, Zheng [35] combined EEG features and eye-movement features through a linear fusion strategy to classify four emotional states, achieving an average accuracy of 72%, whereas the accuracy obtained using only eye-movement features was 67%. This indicates that studies focusing solely on eye-movement-based emotion classification remain limited, and that methodological developments specifically designed for eye-movement signals are still insufficient.
(3) Most existing studies rely on statistical feature extraction from eye-movement data for emotion analysis, without differentiating the relative importance of individual features during the computational process. As a result, the contribution of key features cannot be sufficiently emphasized. When a large number of features are included as input, the weights of features with strong emotional representation capability may be diluted, thereby negatively affecting the overall emotion classification performance.

3. Proposed Method

3.1. Task Definition

In this research, the data samples in the dataset are represented as X R N × D , where N indicates the total number of samples and D denotes the feature dimensionality. The target set Y is defined as Y = Y i , Y i 0 , 1 , , C , which corresponds to the set of emotion categories, with C representing the number of classes. The primary objective of this work is to extract features from X and predict the corresponding emotion category Y.

3.2. The Structure of the Proposed Method

This paper presents an improved CNN–transformer combined with enhanced deep canonical correlation analysis module based on cosine similarity, as depicted in Figure 1.
In this framework, pupil diameter data and gaze data extracted from eye-tracking signals are treated as two independent yet correlated modalities. These modalities are first processed through separate CNN branches for hierarchical feature extraction, enabling the capture of local feature details. The extracted CNN features are subsequently fed into a transformer encoder. Within the transformer architecture, the conventional feedforward network (FFN) is replaced with an incremental feature feedforward network (IFFN), which is designed to enhance the weighting of salient features. Through the self-attention mechanism, the transformer models the internal dependencies among features and generates context-enhanced feature representations. The resulting feature representations from the two modalities are then input into a deep canonical correlation analysis (DCCA) module. Through projection networks, the features are mapped into a shared latent space and optimized using canonical correlation analysis to maximize the cosine similarity between the projected representations of the two modalities. Finally, the aligned dual-modal features are fused and fed into a classifier. By jointly optimizing classification loss and correlation loss, the proposed framework simultaneously enhances feature discriminability and inter-modal consistency, thereby achieving accurate multimodal emotion classification. The details of each component in the proposed architecture are described as follows.

3.2.1. CNN–Transformer Module

As the videos watched by the participants contain strong emotional incentives, this leads to sudden changes in the participants’ emotions, and the emotions generated gradually disappear as the video changes. Therefore, the data during the period from the generation to the disappearance of emotional incentives form an important feature in the emotional computing process and a decisive factor in ensuring the accuracy of emotional computing. In the experiments of this paper, the eye-tracking features used include both local features such as pupil size and fixation duration, as well as global statistical features such as saccade time kurtosis and fixation frequency. Thus, the model needs to accurately process local features while also having strong processing capabilities for global features.
The transformer architecture primarily relies on self-attention mechanisms to handle dependencies in the input sequence, exhibiting excellent performance in long sequence modeling. In contrast, CNN excel at handling local features and weight sharing, making them well suited for extracting local features. Therefore, embedding CNN into transformer models aims to combine the ability of CNN to capture local features with the advantages of transformers in modeling long-range dependencies, thereby improving the overall feature representation capability and prediction performance of the model.
To effectively model temporal dynamics at different resolutions, multiple average pooling layers with kernel sizes of 1 × 8, 1 × 16, 1 × 32, 1 × 64, 1 × 128, 1 × 256, and 1 × 512 are introduced in the CNN module. These pooling operations perform temporal aggregation over windows of different lengths, enabling the extraction of multi-scale temporal features. Specifically, smaller pooling kernels capture fine-grained short-term variations, while larger kernels encode long-range temporal dependencies and global trends. The integration of multi-scale pooled features provides richer temporal representations, which benefits subsequent sequence modeling in the transformer module.
Based on the characteristics of the aforementioned model, this study used two CNN-based modules to process the transformer encoder and decoder. The two CNN modules are employed to extract local features from the input data via one-dimensional convolutional layers. Within the encoder, a two-layer convolutional structure was implemented to increase the output feature channels to 128. The first convolutional layer utilizes a kernel size of 3, with 5 input channels and 64 output channels; the second layer also uses a kernel size of 3 and produces 128 output channels. Each convolution is followed by a ReLU activation, and the same padding is applied to preserve the time-step length, thereby maintaining the integrity of the temporal features. To prevent data redundancy from causing model overfitting, we set the number of heads of the transformer to 8 and the number of layers to 2.
Positional encoding plays a crucial role in transformer models, as the self-attention mechanism itself cannot sense the order of input sequences. To compensate for this shortcoming, position coding enables the model to obtain position information about the individual elements in the sequence, thereby helping to understand the sequential relationships of the input data. By combining element content and position information, position coding not only enhances the model’s expressive ability but also effectively avoids the common gradient disappearance problem in RNNs, making the transformer perform better when dealing with long-distance dependencies [36]. Position coding is typically calculated using fixed sine and cosine functions, whose values are closely related to position information in the sequence. For a given sequence position POS and encoding dimensions, the formula for calculating position coding is as follows:
P E ( p o s , 2 i ) = sin p o s 10,000 2 i d
P E ( p o s , 2 i + 1 ) = cos p o s 10,000 2 i d
where p o s is the position index of the element in the sequence, i is the dimension index, and d is the dimension of the input vector. This sine and cosine encoding method has smooth variation characteristics and can provide the model with relative position information between different positions.
In the architecture of the proposed model, an improved multi-layer transformer encoder module is used as the sequence model. The encoder module has two sub-layers: a multi-head self-attention layer as the first part, followed by a positionally fully connected feedforward network. Additionally, each encoder includes residual connections and layer normalization. For supervised classification, a softmax layer is added to the overall network to complete the classification task. During classification, let the labeled dataset be S, which consists of different eye-tracking feature vectors, namely S = [ S 1 ,   , S n ] . The input sequence is passed through a feature extraction layer and a multi-layer transformer encoder module to obtain the activation m l of the last encoder. Then, a feedforward layer with parameter W y and a SoftMax layer are added to predict y .
P y S 1 , , S n = softmax m l W y
In Equation (3), the conditional probability P is modeled using a neural network. Subsequent models will be described below.

3.2.2. Incremental Feature Feedforward Network

In classical transformer architectures, the standard feedforward feature network independently processes each input feature, significantly enhancing the representation of features by leveraging the self-attention mechanism. However, traditional models do not prioritize emotion features during computation, failing to highlight the importance of key features. Therefore, this paper proposes an incremental feature feedforward network (IFFN) to perform feature transformations in enhancement and simplification, replacing the FFN in the original transformer module. The IFFN is constructed by introducing partial convolution (PConv) operations to enhance the information elements in the features. Enhancing existing features strengthens the weights of key features, thereby improving the overall model accuracy. The specific process is shown in Algorithm 1.
Algorithm 1: Incremental Feature Feedforward Network.
Input:
   X ^ : Autoencoder encoding-decoding and reconstructing the input features
Output:
   X ^ o u t : Enhanced output features
  Step 1: Partial Convolution Projection
     ( X ^ ) GELU ( W 1 PConv ( X ^ ) )
  Step 2: Channel Splitting
     ( X ^ 1 , X ^ 2 ) Split ( X ^ )
  Step 3: Cross-Channel Interaction
     ( X ^ r ) X ^ 1 C o n v ( X ^ 2 )
  Step 4: Final Projection
     ( X ^ o u t ) GELU ( W 2 X ^ r )
  Step 5: Return Result
    Return X ^ o u t
W1 and W2 represent linear projection, the dim of W1 and W2 is 128 × 128. P C o n v ( · ) refers to the partial convolution [37] operation. Partial convolution introduces a mask on the basis of traditional convolution, and only performs convolution operation on the effective region of the input features. It can ignore missing or invalid regions, thereby realizing the “emphasis” operation of key features. The process is represented by splitting the input X ^ after encoding and decoding by the autoencoder, dividing the input into two parts. The purpose is to embed local spatial information into the features [38]. The features of the first branch X ^ 1 are multiplied element-wise with the features of the second branch X ^ 2 after “localization processing”, so that the local spatial importance weight extracted by the second branch can be added to the features of the first branch.
The purpose of this operation is to attach the key features of the second branch to the features of the first branch. By adjusting the features of the first branch with a dynamic, content-related spatial weight, the weight of the important features is increased, thereby improving the accuracy of the overall model.

3.2.3. Deep Canonical Correlation Analysis Model Based on Cosine Similarity

Deep learning-related analysis methods can learn two nonlinear projections with the help of deep neural networks, non-linearly transforming feature vectors of different modalities into feature vectors with stronger correlation [39]. The optimization objective is to maximize the total correlation coefficient, and its principle can be described as follows:
Suppose the features in the original input deep neural network are H 1 n × d 1 and H 2 n × d 2 , where n is the number of samples d 1 and d 2 are the input dimensions of the features. The corresponding nonlinear transformations of the deep neural networks are f 1 ( X 1 ; α 1 ) and f 2 ( X 2 ; α 2 ) , which are the parameters of the two deep neural networks. Let H 1 o × n and H 2 o × n be the new features obtained after undergoing the nonlinear transformations of their respective deep neural networks, where o is the output dimension, and n is the number of input data.
The goal of deep canonical correlation analysis (DCCA) is to jointly learn the parameters α 1 and α 2 of two neural networks so that the cosine similarity correlation between H 1 and H 2 is as high as possible, namely:
α 1 , α 2 = arg   max   Corr f 1 H 1 ; α 1 , f 2 H 2 ; α 2
In the process of calculating the correlation coefficient in DCCA, the feature vectors need to be centralized. This centralization operation may cause the data to lose their original distribution characteristics, thereby introducing bias. Additionally, in the presence of outliers or extreme values, the centralization effect may be poor, and gradient explosion may occur when using maximum correlation coefficient as a constraint during experiments. To address these issues, an alternative correlation measurement method is adopted: cosine similarity. Using cosine similarity allows one to largely avoid the problems encountered by the correlation coefficient and it has the following advantages when compared with the correlation coefficient:
(1) Cosine similarity pays more attention to the direction of vectors rather than their magnitude, making it more suitable for measuring directional similarity between two feature vectors. During the process of subjects watching videos, the eye-tracking features and gaze features may exhibit similar structural changes in their degree of variation, and the similarity in spatial direction of this information can be better reflected.
(2) Cosine similarity does not require the data to follow specific distributional assumptions. In practical scenarios, the data may be influenced by complex environments and varying objects, which can lead to deviations from strict Gaussian distributions. Therefore, cosine similarity is more suitable for measuring the similarity of such data.
(3) Cosine similarity is relatively robust to outliers because it mainly measures the directional similarity between vectors rather than relying on their magnitude. As a result, the influence of extreme values is reduced, leading to more accurate similarity estimation in the presence of outliers.
The back propagation algorithm is used to update, and the target equation and the calculation process are as follows. First, centralize the data as follows:
H 1 ¯ = H 1 1 n H 1
H 2 ¯ = H 2 1 n H 2
where H 1 ¯ = h 1 ( 1 ) , , h 1 ( n ) , H 2 ¯ = h 2 ( 1 ) , , h 2 ( n ) . Data centering ensures that the mean of the centered sample is 0, facilitating data alignment and subsequent processing. For the projection vector pair ( h 1 ( i ) , h 2 ( i ) ) of the i-th sample, the cosine similarity is as follows:
cos h 1 ( i ) , h 2 ( i ) = h 1 ( i ) T h 2 ( i ) h 1 ( i ) 2 h 2 ( i ) 2
The optimization objective is defined as maximizing the mean cosine similarity among all pairs of samples, which serves as the objective function:
max α 1 , α 2 J = 1 n i = 1 n h 1 ( i ) T h 2 ( i ) h 1 ( i ) 2 h 2 ( i ) 2 = tr H 1 T ¯ H 2 ¯ H 1 ¯ H 2 ¯ = cos H 1 ¯ , H 2 ¯
Here, t r ( ) represents the trace of the matrix. 2 represents the L2 norm of the vector set. The cosine similarity cos H 1 ¯ , H 2 ¯ is calculated using the Frobenius norm of the vector set, and its gradient expression with respect to the sum of eigenvectors is as follows:
cos H 1 ¯ , H 2 ¯ H 1 ¯ = 1 H 1 ¯ H 2 ¯ H 2 ¯
cos H 1 ¯ , H 2 ¯ H 2 ¯ = 1 H 1 ¯ H 2 ¯ H 1 ¯
After training and optimizing deep neural networks and through backpropagation of f 1 ( X 1 ; α 1 ) and f 2 ( X 2 ; α 2 ) , an approximate optimal solution for the parameters of the deep neural networks is obtained.

3.2.4. Loss Optimization Module

To enhance the effectiveness of features, this paper designs a dual-loss joint optimization function that improves the overall performance of the model and achieves synergistic optimization of feature representation and classification decision by adjusting the weights of different loss functions. The expression of the adopted loss function is as follows:
L = ω 1 L C E + ω 2 L D C C A
L C E = 1 n i = 1 n L i = 1 n i = 1 n c = 1 M y i c log ( p i c )
L D C C A = tr H 1 T ¯ H 2 ¯ H 1 ¯ H 2 ¯
Among these, L C E is the cross-entropy loss; L D C C A is the loss based on cosine similarity for deep canonical correlation analysis; ω 1 and ω 2 represent weights ( ω 1 + ω 2 = 1 ); n is the total number of samples; M is the total number of categories; y i c is the sign function, i.e., ( y i c { 0 , 1 } ); the true category c of sample i is 1, otherwise c is 0; p i c is the probability that sample i belongs to category c ; H 1 ¯ is the feature obtained by the gaze data after passing through the feature extractor; and H 2 ¯ is the feature obtained by the pupil data after passing through the feature extractor. During the experiment, we observed that the magnitude of L C E is approximately ten times that of L D C C A . To balance the influence of these two components, we adjusted the weighting parameters through multiple experiments. The ratio ω 1 : ω 2 = 1:9 (i.e., 0.1 and 0.9) was found to provide the most stable training process and the best performance. This setting helps compensate for the difference in scale between L C E and L D C C A , ensuring that both terms contribute appropriately during optimization. Consequently, in the experiment, ω 1 is set to 0.1 and ω 2 is set to 0.9.

4. Experimental Results and Analysis

4.1. Experimental Procedure

4.1.1. Experimental Design

The research objective of this experiment is to classify different types of emotions through the proposed algorithm model. Currently, there are two mainstream methods for emotion classification: one is to classify specific emotions; the other is based on the two-dimensional affective model of valence–arousal proposed by Russell for emotion classification. These emotions are highly systematically correlated through positive or negative correlations, where valence is used to describe the negative or positive degree of an event, and arousal is used to describe the degree of interest in the event. Russell’s model can represent the changes between different emotions and can also more clearly express certain emotions that are difficult to describe. In this experiment, a combined method of the two classification approaches is adopted based on the characteristics of different datasets for classification and, by comparing with existing baseline algorithms, the superiority of the proposed method is intended to be fully validated. The data splitting strategy plays an essential role in model evaluation, as it defines the division of the dataset into training and testing sets and directly influences the model’s performance. In this study, all experiments employ a subject-independent classification method for dataset partitioning.
Traditional machine learning and classic deep learning methods, including multiple approaches from both categories, are selected as comparison objects. Traditional machine learning methods include mainstream algorithms such as SVM, random forest, and decision trees, while deep learning includes models such as CNN and LSTM. All comparison models use completely consistent dataset partition ratios, evaluation metrics, and hyperparameter optimization processes to avoid result bias caused by experimental condition differences from the source.

4.1.2. Experimental Environment

The experiment’s software and hardware configurations are as follows: The hardware includes an NVIDIA GeForce RTX 4060 GPU with 12 GB of video random access memory. The software uses Python 3.12.3 and relies on libraries including Pandas, PyTorch 2.2.1, NumPy, and Scikit-learn 1.2.1.

4.1.3. Experiment Parameter Configuration

All models in the experiment were trained using the same parameter inputs, and the input features of all models remained consistent. The Adam optimizer was used for training. The specific parameter configuration is shown in Table 2.

4.1.4. Experiment Dataset

SEED-V Dataset
The SEED-V dataset [40] comprises five emotional categories: happy, sad, fear, disgust, and neutral. Sixteen healthy participants (6 males, 10 females) aged 19–28 were recruited. For each emotion, nine movie clips were selected based on their elicited emotional effects, with durations ranging from 2 to 4 min. The subjects in each experiment needed to watch 15 pieces of stimulating material and 3 pieces of each type of emotion. To study the stability of emotion recognition and ensure the effectiveness of the stimulation, each subject was asked to participate in the experiment three times, with at least three days between experiments.
In the self-assessment section, the subjects were asked to score according to the induction effect of the stimulus material. The scoring range was 0–5 points, of which 5 points represented the best induction effect and 0 points the worst. If the participant feels joy after watching the joy video, they should be scored 4–5 points, and if they do not feel anything, they should be given a score of 0. It is worthwhile to mention that if you are watching neutral stimulus material, if the subject’s mood fluctuates, the score should be 0, and the natural state is 5 points.
SEED-IV Dataset
The SEED-IV dataset [35] is similar to the SEED-V dataset in that it contains EEG and eye-movement signals recorded by 15 participants who watched a 2 min video clip to elicit four emotions: happy, sad, fear, and neutral. Each participant completed 6 trials for each emotion, resulting in 24 composite trials. The signals were down sampled to 200 Hz and segmented into non-overlapping 4 s intervals. Eye-movement signals were recorded using the SMI eye-tracking glasses.
eSEE-d Dataset
The eSEE-d dataset [26] is designed for the assessment of emotional states through eye-tracking data. Eye-movements of 48 participants were recorded while they viewed 10 emotionally evocative videos, each followed by a neutral clip. The collected signals encompassed various indicators of fixation, blinking, and pupil dynamics, allowing for the extraction of a comprehensive set of eye-tracking features, including fixations and saccades. Participants rated five emotions—tenderness, anger, disgust, sadness and neutral—on a scale from 0 to 10, which were subsequently transformed into emotional arousal and valence measures. Additionally, each participant completed three self-assessment questionnaires. A comprehensive analysis was conducted on the participants’ self-assessment scores on the questionnaires and their scores during the experiment. In this dataset, our study used the Russell model of the emotion coordinate system to convert the five emotions—anger, sadness, disgust, tenderness, and neutral—in the experiment into three categories of valence: positive valence (PV), medium valence (MV), and negative valence (NV); and three categories of arousal: high arousal (HA), medium arousal (MA), and low arousal (LA). Specifically: anger/disgust → high arousal—negative value (HANV), sadness → low arousal—negative value (LANV), tenderness → low arousal—positive value (LAPV), and neutral → medium arousal—medium value (MAMV). The statistical results after feature extraction are as follows: Classified by arousal, there were 1892 high arousal samples, 1605 low arousal samples, and 492 medium arousal samples.

4.1.5. Data Preprocessing and Feature Extraction

As the experimental data in this study came from three publicly available datasets containing complete raw data, some of the raw data might become unusable due to unexpected or unforeseen errors during the experiment; consequently, it was necessary to convert these raw signals into structured inputs suitable for the research. During the experiment, we performed data preprocessing and feature engineering on the data.
The SEED-IV and SEED-V datasets already have fully classified eye-tracking features. In order to compare with other methods that also use the same datasets, this paper has not changed the features in the datasets. The eye-movement features are based on the SEED-IV and SEED-V datasets and are shown in Table 3.
The eSEE-d dataset comprises raw eye-tracking data, which necessitates specific pre-processing steps such as filtering, data alignment, and interpolation-based smoothing. Feature extraction is conducted on the unprocessed data and computing metrics include the mean and standard deviation of pupil diameter, fixation duration, saccade duration, as well as the kurtosis and skewness of saccade velocity. All proposed features are derived directly from the raw dataset.
To ensure the statistical significance of the extracted features, a significance test was conducted. Specifically, one-way analysis of variance (ANOVA) was employed to evaluate whether the extracted features exhibit statistically significant differences across different emotion categories. Features with p-values lower than 0.05 were retained, indicating that they contribute significantly to distinguishing emotional states. This procedure helps eliminate redundant or non-informative features and improves the efficiency and generalization capability of the proposed model. The eye-tracking features derived from the eSEE-d dataset are summarized in Table 4.

4.2. Experimental Results and Comparison

To demonstrate the superiority of this method, the proposed ICTD is compared with other methods in some literature. The goal is to verify the effectiveness of the proposed model. For the evaluation metric, we mainly use accuracy for comparison. The calculation formula is as follows:
Accuracy = T P + T N T P + F N + T N + F P
In this context, true positive (TP) and false negative (FN) represent whether the target sample is classified correctly or incorrectly, while true negative (TN) and false positive (FP) represent whether the non-target sample is classified correctly or incorrectly.

4.2.1. Evaluation of ICTD’s Effectiveness on the eSEE-d Dataset

In the experiments on the eSEE-d dataset, this paper classifies the data separately according to the specific emotions, and 5-fold cross-validation is conducted in the experiments. In the experiment, 80% of the data were selected as the training set, and the remaining 20% were used as the test set. Table 5 shows the classification results on the eSEE-d dataset and the comparison results with existing methods.
The experimental results show that the multimodal improved deep canonical correlation CNN–transformer method proposed in this paper has 81.8% arousal accuracy and 85.2% valence accuracy on this dataset. The 95% arousal confidence interval is [80.9%, 82.7%] and valence confidence interval is [84.2%, 86.2%], indicating excellent performance. Compared with traditional machine learning, this method’s recognition accuracy exceeds SVM’s highest respective arousal and valence accuracies of 64.6% and 70.1%, and the accuracy is improved by a respective 17.2% and 15.1%. Compared with the deep learning method, the accuracy of arousal and valence is improved by 10.5% and 12%, compared with the popular CNN–transformer method. Compared with the deep typical correlation analysis commonly used in the field of emotion analysis in recent years, the accuracy is improved by 6.3% and 8.9%, respectively. Compared with the literature that proposed the eSEE-d dataset, the accuracy of arousal and valence was improved by 3.0% and 1.2%, respectively.
Overall, the results show that our model has a better effect on eye-movement emotion classification, and the accuracy is greatly improved compared with traditional machine learning methods, indicating that it can achieve high accuracy even with a small input sample size. Compared with the current commonly used deep learning methods, it can also achieve better classification results.
To provide a more detailed analysis of the proposed method’s performance on this dataset, Figure 2 presents the standardized confusion matrices obtained by aggregating the results from all cross-validation folds. Here, Figure 2a shows a confusion matrix of accuracy for a three-category arousal classification, with the corresponding arousal categories being 0-LA, 1-MA, and 2-HA. Figure 2b shows a confusion matrix of accuracy for a three-category valence classification, with the corresponding arousal categories being 0-NV, 1-MV, and 2-PV. The values on the diagonal of the confusion matrix represent the accuracy of each emotion when using a single modality. The confusion matrix results show that the model’s accuracy in identifying valence is higher than its accuracy in identifying arousal. This is because the dataset uses five emotional stimuli to stimulate the subjects. Among these five emotions, the ratio of positive valence, medium valence, and negative valence is 1:1:3, while the ratio of high arousal, medium arousal, and low arousal is 2:1:2. Positive valence and medium valence each correspond to only one emotion, resulting in higher accuracy after the model classifies emotions. Arousal, on the other hand, has a more even distribution and is more difficult to distinguish than valence, hence the above results. This indicates that the model’s results conform to the objective laws of the data. Furthermore, the results show that the model performs best in both arousal and valence classification, reaching 84.4% and 86.2% for medium arousal and neutral valence, respectively. The main confusion in arousal occurs when 14.9% of “low arousal” is incorrectly classified as “high arousal,” while 14.1% of negative valence is classified as positive valence. This also proves that eye-tracking parameters differ significantly from other categories in medium valence and medium arousal, resulting in the best classification results.

4.2.2. Evaluation of ICTD’s Effectiveness on the SEED-IV and SEED-V Datasets

As both SEED-IV and SEED-V datasets were proposed by Lu Baoliang’s team at Shanghai Jiao Tong University, and the experimental methods for both datasets are identical (differences only lie in the number of participants and the number of emotion categories), presenting the experimental results for both datasets simultaneously allows for a better comparison of the performance of the proposed method on both datasets. In the SEED-IV and SEED-V dataset experiments, this paper classifies the data separately according to the arousal and valence dimensions, and 5-fold cross-validation is conducted in the experiments. In the experiment, 80% of the data were selected as the training set, and the remaining 20% were used as the test set. Table 6 and Table 7 shows the classification results on the SEED-IV and SEED-V datasets and the comparison results with existing methods. Both datasets are multimodal datasets composed of EEG and eye-tracking data, only the eye-tracking data were used in this study.
Experimental results demonstrate that the proposed multimodal improved deep canonical correlation CNN–transformer method achieves accuracies of 91.2% and 85.1% on the SEED-IV and SEED-V datasets, respectively, and that the 95% confidence intervals are [89.1%, 93.3%] and [82.3%, 87.9%], exhibiting excellent performance. Compared with traditional machine learning, its recognition accuracy surpasses the highest arousal and valence accuracy of SVM by 72.3% and 68.6%, respectively, representing improvements of 19.9% and 16.5%. Compared with deep learning methods, and specifically the currently popular CNN–transformer method, it improves accuracy by 10.1% and 7.3% in arousal and valence, respectively. Compared with deep canonical correlation analysis, which has been commonly used in emotion analysis in recent years, it improves accuracy by 2.3% and 3.1%.
Figure 3 presents the confusion matrices illustrating the emotion classification performance on the SEED-IV and SEED-V datasets. Specifically, Figure 3a shows the confusion matrix for the SEED-IV dataset, where the emotion categories are defined as 0—neutral, 1—sad, 2—fear, and 3—happy. Figure 3b shows the confusion matrix for the SEED-V dataset, with the corresponding emotion categories being 0—disgust, 1—fear, 2—sad, 3—neutral, and 4—happy. Overall, the results show that our model outperforms both traditional machine learning methods and several commonly used deep learning methods listed. The experimental results on the SEED-V dataset are generally lower than those on SEED-IV because SEED-IV is a four-class emotion dataset, while SEED-V is a five-class emotion dataset, making classification more challenging.

4.2.3. Ablation Experiment

To verify the functional contribution of each module in ICTD, this section conducts ablation experiments on the CNN–transformer module, IFFN module, and cos-DCCA module. To align the ablation results more closely with the model’s performance in practical applications, the ablation dataset is used for four-classification training based on the SEED-IV subject-independent experiment, with 5-fold cross-validation implemented. The results of the ablation experiments are summarized in Table 8. The results show that removing the IFFN module causes a 1.8% decrease in model accuracy, which means that replacing the FFN module in the transformer with IFFN can enhance the performance of the model. After removing the cos-DCCA module, the model accuracy drops by 5.7% with the largest accuracy deviation, which means the DCCA module founded on cosine similarity has the greatest impact on classification accuracy. In addition, in the experiment where the cos-DCCA module and IFFN module are removed, the model accuracy decreases by 10.1%, which means that using only the CNN–transformer method results in a significant deviation from the ICTD model’s outcomes.

5. Discussion

In order to solve the problems mentioned in Section 2, this paper proposes an improved CNN–transformer combined with enhanced deep canonical correlation analysis network, which mainly contributes as follows:
(1) Traditional eye-tracking emotion analysis mostly uses feature splicing or double-branch alignment to process data, with which it is difficult to capture the deep association and hierarchical structure between eye-movement features, in turn often leading to fragmentation of emotion representation. In this paper, CNN–transformer is combined with an enhanced deep canonical correlation analysis model. After using CNN and transformer to capture local and global information, it learns high-order nonlinear features, and then effectively integrates eye-movement features, explores the correlation between features, and improves the practicability of eye-movement features.
(2) Existing studies mostly rely on the statistical characteristics of the original data for emotion analysis. Moreover, the importance of each emotional feature in the calculation process is not primary or secondary, and the role of key features cannot be highlighted. To solve this problem, an incremental feature feedforward network module is designed to improve the transformer module. Replacing features from indiscriminate inputs by adding weight to key features increases the importance of these key features. This design effectively addresses the lack of prioritization in feature importance, and significantly improves the ability of eye-movement features to characterize emotional states.
However, the ICTD has certain limitations. While the model performs well on the same dataset its performance may vary across different subjects, experimental settings, or eye-tracking devices, indicating a need for improved cross-subject generalization. In addition, the IFFN architecture, combined with transformer and local convolution operations, increases both training and inference time, potentially restricting real-time emotion analysis applications.

6. Conclusions

This study proposes an improved CNN–transformer framework integrated with an enhanced deep canonical correlation analysis network for emotion classification based on eye-movement data. The proposed method addresses the limitation that a small number of eye-tracking features cannot adequately represent complex emotional states. In the proposed architecture, convolutional neural networks (CNNs) and transformer modules are employed to capture local and global feature representations, respectively. Furthermore, the transformer is enhanced by introducing an incremental feature feedforward network, which enables the model to assign greater importance to salient feature information. The extracted features are subsequently processed using deep canonical correlation analysis based on cosine similarity so as to obtain the final classification results. Experimental evaluations conducted on the eSEE-d, SEED-IV, and SEED-V datasets demonstrate the effectiveness of the proposed approach.
In future work, we plan to incorporate additional datasets to further expand the data scale. Moreover, deep multimodal fusion strategies combining eye-tracking features with other physiological signals will be explored to identify features that better reflect emotional states, thereby improving the robustness, practicality, and application potential of the proposed method.

Author Contributions

Conceptualization, X.L. and J.C.; methodology, C.Z.; software, C.Z. and M.C.; validation, C.Z., J.C., M.C. and J.L.; formal analysis, C.Z.; investigation, C.Z., Q.G. and J.L.; writing—original draft preparation, C.Z. and M.C.; writing—review and editing, X.L. and J.C.; resources, X.L. and J.C.; visualization, M.C.; data curation, J.L. and Q.G.; supervision, X.L. and J.C.; project administration, J.L.; funding acquisition, J.C. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation for Young Scholars of China under grant 62206016; in part by the Guangdong Basic and Applied Basic Research Foundation under grants 2023A1515140086 and 2022A1515140016; in part by the Fundamental Research Funds for the Central Universities under grant FRF-TP-24-062A; in part by the Project “Vice President of Science and Technology” of Changping District, Beijing under grant 2025040040042; in part by the open project of the State Key Laboratory of Digital Manufacturing Equipment and Technology, Huazhong University of Science and Technology (DMETKF2021023); and in part by the Foundation of Engineering Research Center of Intelligence Perception and Autonomous Control, Ministry of Education, China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in reference number [eSEE-d] [https://doi.org/10.3390/brainsci13040589] [26]; [SEED-IV] [https://doi.org/10.1109/TCYB.2018.2797176] [35]; [SEED-V] [https://doi.org/10.1109/NER.2019.8716943] [40].

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IFFNIncremental feature feedforward network
LALow arousal
MAMedium arousal
HAHigh arousal
NVNegative valence
MVMedium valence
PVPositive valence
LSTMLong short-term memory
KNNk-nearest neighbors
SVMSupport vector machine
RNNRecurrent neural network
DGCNNDeep gradient convolutional neural network
DMLPDeep multi-layer perceptron
STFTShort-time Fourier transform
DCCADeep canonical correlation analysis
ANOVAAnalysis of variance
ICTDImproved CNN-transformer combined with enhanced deep canonical correlation Analysis network
SEEDThe SJTU emotion EEG dataset
SEED-FRASEED includes data of French subjects
SEED-VISEED includes 4 categories of emotional classification
SEED-VSEED includes 5 categories of emotional classification
fNIRSFunctional near-infrared spectroscopy

References

  1. Salovey, P.; Mayer, J.D. Emotional intelligence. Imagin. Cogn. Personal. 1990, 9, 185–211. [Google Scholar] [CrossRef]
  2. Picard, R.W. Affective Computing; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  3. Mühl, C.; Allison, B.; Nijholt, A.; Chanel, G. A survey of affective brain computer interfaces: Principles, state-of-the-art, and challenges. Brain-Comput. Interfaces 2014, 1, 66–84. [Google Scholar] [CrossRef]
  4. Wu, D.; Lu, B.L.; Hu, B.; Zeng, Z. Affective brain–computer interfaces (aBCIs): A tutorial. Proc. IEEE 2023, 111, 1314–1332. [Google Scholar] [CrossRef]
  5. Cowie, R.; Douglas-Cowie, E.; Tsapatsoulis, N.; Votsis, G.; Kollias, S.; Fellenz, W.; Taylor, J.G. Emotion recognition in human-computer interaction. IEEE Signal Process. Mag. 2001, 18, 32–80. [Google Scholar] [CrossRef]
  6. Shanechi, M.M. Brain–machine interfaces from motor to mood. Nat. Neurosci. 2019, 22, 1554–1564. [Google Scholar] [CrossRef] [PubMed]
  7. Widge, A.S.; Malone, D.A., Jr.; Dougherty, D.D. Closing the loop on deep brain stimulation for treatment-resistant depression. Front. Neurosci. 2018, 12, 175. [Google Scholar] [CrossRef] [PubMed]
  8. Liu, Y.H.; Wu, C.T.; Cheng, W.T.; Hsiao, Y.T.; Chen, P.M.; Teng, J.T. Emotion recognition from single-trial EEG based on kernel Fisher’s emotion pattern and imbalanced quasiconformal kernel support vector machine. Sensors 2014, 14, 13361–13388. [Google Scholar] [CrossRef] [PubMed]
  9. Alonso-Martin, F.; Malfaz, M.; Sequeira, J.; Gorostiza, J.F.; Salichs, M.A. A multimodal emotion detection system during human–robot interaction. Sensors 2013, 13, 15549–15581. [Google Scholar] [CrossRef]
  10. Peng, Y.; Lu, B.L. Discriminative extreme learning machine with supervised sparsity preserving for image classification. Neurocomputing 2017, 261, 242–252. [Google Scholar] [CrossRef]
  11. Si, X.; Huang, H.; Yu, J.; Ming, D. EEG microstates and fNIRS metrics reveal the spatiotemporal joint neural processing features of human emotions. IEEE Trans. Affect. Comput. 2024, 15, 2128–2138. [Google Scholar] [CrossRef]
  12. Liu, X.; Hu, B.; Si, Y.; Wang, Q. The role of eye movement signals in non-invasive brain-computer interface typing system. Med. Biol. Eng. Comput. 2024, 62, 1981–1990. [Google Scholar] [CrossRef]
  13. Startsev, M.; Zemblys, R. Evaluating eye movement event detection: A review of the state of the art. Behav. Res. Methods 2023, 55, 1653–1714. [Google Scholar] [CrossRef]
  14. Bradley, M.M.; Codispoti, M.; Cuthbert, B.N.; Lang, P.J. Emotion and motivation I: Defensive and appetitive reactions in picture processing. Emotion 2001, 1, 276. [Google Scholar] [CrossRef]
  15. Zheng, W.L.; Zhu, J.Y.; Peng, Y.; Lu, B.L. EEG-based emotion classification using deep belief networks. In 2014 IEEE International Conference on Multimedia and Expo (ICME); IEEE: New York, NY, USA, 2014; pp. 1–6. [Google Scholar]
  16. Holmqvist, K.; Nyström, M.; Andersson, R.; Dewhurst, R.; Jarodzka, H.; Van de Weijer, J. Eye Tracking: A Comprehensive Guide to Methods and Measures; Oxford University Press: Oxford, UK, 2011. [Google Scholar]
  17. Worthy, D.A.; Lahey, J.N.; Priestley, S.L.; Palma, M.A. An examination of the effects of eye-tracking on behavior in psychology experiments. Behav. Res. Methods 2024, 56, 6812–6825. [Google Scholar] [CrossRef]
  18. Ionescu, A.; Ștefănescu, E.; Strilciuc, Ș.; Rafila, A.; Mureșanu, D. Correlating eye-tracking fixation metrics and neuropsychological assessment after ischemic stroke. Medicina 2023, 59, 1361. [Google Scholar] [CrossRef]
  19. Ibragimov, B.; Mello-Thoms, C. The use of machine learning in eye tracking studies in medical imaging: A review. IEEE J. Biomed. Health Inform. 2024, 28, 3597–3612. [Google Scholar] [CrossRef] [PubMed]
  20. Lim, J.Z.; Mountstephens, J.; Teo, J. Emotion recognition using eye-tracking: Taxonomy, review and current challenges. Sensors 2020, 20, 2384. [Google Scholar] [CrossRef]
  21. Oliva, M.; Anikin, A. Pupil dilation reflects the time course of emotion recognition in human vocalizations. Sci. Rep. 2018, 8, 4871. [Google Scholar] [CrossRef] [PubMed]
  22. Gilzenrat, M.S.; Nieuwenhuis, S.; Jepma, M.; Cohen, J.D. Pupil diameter tracks changes in control state predicted by the adaptive gain theory of locus coeruleus function. Cogn. Affect. Behav. Neurosci. 2010, 10, 252–269. [Google Scholar] [CrossRef] [PubMed]
  23. Aracena, C.; Basterrech, S.; Snáel, V.; Velásquez, J. Neural networks for emotion recognition based on eye tracking data. In 2015 IEEE International Conference on Systems, Man, and Cybernetics; IEEE: New York, NY, USA, 2015; pp. 2632–2637. [Google Scholar]
  24. Cheng, B.; Titterington, D.M. Neural networks: A review from a statistical perspective. Stat. Sci. 1994, 9, 2–30. [Google Scholar]
  25. Palm, R.B. Prediction as a candidate for learning deep hierarchical models of data. Tech. Univ. Den. 2012, 5, 19–22. [Google Scholar]
  26. Skaramagkas, V.; Ktistakis, E.; Manousos, D.; Kazantzaki, E.; Tachos, N.S.; Tripoliti, E.; Fotiadis, D.I.; Tsiknakis, M. eSEE-d: Emotional state estimation based on eye-tracking dataset. Brain Sci. 2023, 13, 589. [Google Scholar] [CrossRef]
  27. Fu, B.; Gu, C.; Fu, M.; Xia, Y.; Liu, Y. A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals. Front. Neurosci. 2023, 17, 1234162. [Google Scholar]
  28. Li, Y.; Deng, J.; Wu, Q.; Wang, Y. Eye-tracking signals based affective classification employing deep gradient convolutional neural networks. Int. J. Interact. Multimedia Artif. Intell. 2021, 7, 34–43. [Google Scholar] [CrossRef]
  29. Tarnowski, P.; Kołodziej, M.; Majkowski, A.; Rak, R.J. Eye-tracking analysis for emotion recognition. Comput. Intell. Neurosci. 2020, 2020, 2909267. [Google Scholar] [CrossRef]
  30. Wang, Y.; Lv, Z.; Zheng, Y. Automatic emotion perception using eye movement information for E-healthcare systems. Sensors 2018, 18, 2826. [Google Scholar] [CrossRef]
  31. Gong, X.; Chen, C.P.; Hu, B.; Zhang, T. CiABL: Completeness-induced adaptative broad learning for cross-subject emotion recognition with EEG and eye movement signals. IEEE Trans. Affect. Comput. 2024, 15, 1970–1984. [Google Scholar] [CrossRef]
  32. Jiménez-Guarneros, M.; Fuentes-Pineda, G.; Grande-Barreto, J. Multimodal semi-supervised domain adaptation using cross-modal learning and joint distribution alignment for cross-subject emotion recognition. IEEE Trans. Instrum. Meas. 2025, 72, 2523911. [Google Scholar] [CrossRef]
  33. Liu, W.; Zheng, W.L.; Li, Z.; Wu, S.Y.; Gan, L.; Lu, B.L. Identifying similarities and differences in emotion recognition with EEG and eye movements among Chinese, German, and French People. J. Neural Eng. 2022, 19, 026012. [Google Scholar] [CrossRef] [PubMed]
  34. Skaramagkas, V.; Ktistakis, E.; Manousos, D.; Tachos, N.S.; Kazantzaki, E.; Tripoliti, E.E.; Fotiadis, D.I.; Tsiknakis, M. A machine learning approach to predict emotional arousal and valence from gaze extracted features. In 2021 IEEE 21st International Conference on Bioinformatics and Bioengineering (BIBE); IEEE: New York, NY, USA, 2021; pp. 1–5. [Google Scholar]
  35. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. Emotionmeter: A multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 2018, 49, 1110–1122. [Google Scholar] [CrossRef]
  36. Li, Y.; Zhang, K.; Cao, J.; Timofte, R.; Van Gool, L. Localvit: Bringing locality to vision transformers. arXiv 2021, arXiv:2104.05707. [Google Scholar]
  37. Chen, J.; Kao, S.H.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, don’t walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; IEEE: New York, NY, USA, 2023; pp. 12021–12031. [Google Scholar]
  38. Andrew, G.; Menglong, Z. Efficient convolutional neural networks for mobile vision applications. Mobilenets 2017, 10, 151. [Google Scholar]
  39. Qiu, J.L.; Liu, W.; Lu, B.L. Multi-view emotion recognition using deep canonical correlation analysis. In International Conference on Neural Information Processing; Springer International Publishing: Cham, Switzerland, 2018; pp. 221–231. [Google Scholar]
  40. Li, T.H.; Liu, W.; Zheng, W.L.; Lu, B.L. Classification of five emotions from EEG and eye movement signals: Discrimination ability and stability over time. In 2019 9th International IEEE/EMBS Conference on Neural Engineering (NER); IEEE: New York, NY, USA, 2019; pp. 607–610. [Google Scholar]
Figure 1. Structure diagram of ICTD.
Figure 1. Structure diagram of ICTD.
Brainsci 16 00330 g001
Figure 2. (a) Confusion matrix of accuracy for three-category arousal classification and (b) confusion matrix of accuracy for three-category valence classification.
Figure 2. (a) Confusion matrix of accuracy for three-category arousal classification and (b) confusion matrix of accuracy for three-category valence classification.
Brainsci 16 00330 g002
Figure 3. (a) Emotion classification accuracy confusion matrix of SEED-IV and (b) emotion classification accuracy confusion matrix of SEED-IV.
Figure 3. (a) Emotion classification accuracy confusion matrix of SEED-IV and (b) emotion classification accuracy confusion matrix of SEED-IV.
Brainsci 16 00330 g003
Table 1. Summary of methods and features in emotion classification.
Table 1. Summary of methods and features in emotion classification.
AuthorsObjectivesModalitiesFeaturesResults
Skaramagkas [26]To test new machine learning methods and attempt to develop the best model that uses eye-tracking functionality to predict emotional states. EyeFixation and blink frequency, saccade amplitude and duration, pupil diameter, fixation duration kurtosis, saccade duration variation and pupil diameter kurtosis.84%
Fu [27]To enhance the accuracy and stability of emotion recognition, a dual-branch feature extraction module is designed to extract features from both modalities simultaneously while maintaining temporal alignment between the modal signals.EEG, Eye33-dimension eye-movement features such as pupil diameter, gaze details, saccade details, blink details, and event details statistics, and a 62 channel EEG signal. 87%
Oliva [21]Investigate the relationship between pupil size fluctuations and the process of emotion recognition. Participants heard human nonverbal vocalizations and indicated the emotional state of the speakers as soon as they had identified it. EyePeak pupil dilation, the time of peak dilation, the rate of pre-peak dilation, and the rate of post-peak contraction.81%
Li [28]Develop a deep gradient convolutional neural network (DGCNN) using eye-movement signals for emotion classification.EyeImage features87%
Aracena [23]Using the evolution of the eye-tracking data during a window of time for recognizing the emotions which are provoked by visual stimuli of colored images labeled as negative, neutral and positive.EyePupil size average, pupil size change average72%
Gong [31]To overcome the individual differences caused by the non-stationary and low signal-to-noise properties that bring several challenges to cross-subject emotion recognition tasks, a novel multimodal emotion recognition model was proposed for improving the generalization performance to unseen target domain subjects.EEG, Eye33-dimension eye-movement features such as pupil diameter, gaze details, saccade details, blink details, and event details statistics, and a 62 channel EEG signal.96%
Liu [33]Identify the similarities and differences among Chinese, German, and French individuals in emotion recognition with electroencephalogram (EEG) and eye-movements from an affective computing perspective.EEG, Eye33-dimension eye-movement features such as pupil diameter, gaze details, saccade details, blink details, and event details statistics, and a 62 channel EEG signal.84%
Tarnowski [29]To classify three emotional states, the study employed support vector machines, linear discriminant analysis, and k-nearest neighbors, with performance evaluated using the leave-one-subject-out cross-validation approach.Eye18-dimension eye-movement features such as average duration of fixation, skewness of fixation durations, average amplitude of the saccades, variation of saccade amplitudes, average pupil diameter, pupil diameter variance, pupil diameter kurtosis.80%
Table 2. Experimental parameter Configuration Table.
Table 2. Experimental parameter Configuration Table.
Category/ParameterValue
Training and optimization 
OptimizerAdam
Learning rate0.0005
Momentum(0.9, 0.999)
Batch size128
Epochs100 (eSEE-d), 50 (SEED-IV, SEED-V)
Decay rate0.1
Weight decay0.01
Model architecture 
Hidden layer164
Hidden layer2128
Dropout0.3
Table 3. Eye-tracking features extracted from the SEED-IV and SEED-V datasets.
Table 3. Eye-tracking features extracted from the SEED-IV and SEED-V datasets.
Eye-Movement ParametersExtracted Features
Pupil diameters
(left and right)
Mean, standard deviation
Fixation duration (ms)
Dispersion (X and Y)
Saccade duration (ms)
Saccade amplitude (°)
Blink duration (ms)
Event statisticsFixation frequency, maximum fixation duration, maximum fixation dispersion, total fixation dispersion, saccade frequency, average saccade duration, average saccade amplitude, average saccade latency
Table 4. Eye-tracking features extracted from the eSEE-d dataset.
Table 4. Eye-tracking features extracted from the eSEE-d dataset.
Eye-Movement ParametersExtracted Features
Pupil diameterMean, Var, CV
Fixation duration (ms)Mean, Var, CV
Saccade duration (ms)Kurt, Skew, CV 
Saccade speedKurt, Skew
Saccade distanceKurt, Skew
Table 5. Classification accuracy comparison of various methods on the eSEE-d dataset.
Table 5. Classification accuracy comparison of various methods on the eSEE-d dataset.
MethodArousalValence
Accuracy (%) ± StdF1-Score (%) ± StdAccuracy (%) ± StdF1-Score (%) ± Std
SVM64.6 ± 1.663.5 ± 1.570.1 ± 1.168.3 ± 1.2
Random forest62.9 ± 1.361.8 ± 1.256.5 ± 0.854.8 ± 0.8
CNN69.5 ± 2.268.2 ± 2.071.2 ± 2.670.1 ± 2.5
CNN–transformer71.3 ± 1.870.0 ± 1.773.2 ± 2.373.5 ± 2.1
LSTM71.8 ± 2.570.5 ± 2.372.6 ± 2.475.6 ± 2.4
DCCA [39]75.5 ± 3.674.2 ± 3.376.3 ± 3.176.8 ± 3.0
DMLP [26]72.0 ± \\84.0 ± \\
DGCNN [28]78.8 ± 1.577.5 ± 1.382.1 ± 1.381.5 ± 1.2
ICTD (ours)81.8 ± 0.980.4 ± 0.785.2 ± 1.084.2 ± 0.9
Table 6. Accuracy comparison of various methods on the SEED-IV datasets.
Table 6. Accuracy comparison of various methods on the SEED-IV datasets.
MethodSEED-IV
Accuracy (%) ± StdF1-Score (%) ± Std
SVM72.3 ± 1.868.5 ± 2.1
Random forest66.2 ± 1.563.1 ± 1.8
CNN75.4 ± 2.971.7 ± 2.5
CNN–transformer81.1 ± 2.678.4 ± 2.2
LSTM80.3 ± 1.777.6 ± 1.4
DGCNN [28]87.8 ± 1.684.5 ± 1.7
DCCA [39]88.9 ± 2.885.7 ± 2.4
ICTD (Ours)91.2 ± 2.189.3 ± 2.0
Table 7. Accuracy comparison of various methods on the SEED-V datasets.
Table 7. Accuracy comparison of various methods on the SEED-V datasets.
MethodSEED-V
Accuracy (%) ± StdF1-Score (%) ± Std
SVM68.6 ± 2.367.5 ± 2.5
Random forest59.1 ± 1.758.2 ± 1.9
CNN71.2 ± 3.170.3 ± 3.3
CNN–transformer77.8 ± 3.476.8 ± 3.6
LSTM76.6 ± 1.676.0 ± 1.5
DGCNN [28]82.1 ± 2.381.5 ± 2.5
DCCA [39]82.0 ± 3.280.3 ± 3.4
ICTD (Ours)85.1 ± 2.884.5 ± 3.0
Table 8. Results of ablation experiments.
Table 8. Results of ablation experiments.
MethodAccuracy (%) ± StdF1-Score (%) ± Std
ICTD (ours)91.2 ± 2.189.3 ± 2.0
w/o IFFN89.4 ± 1.888.1 ± 1.7
w/o cos-DCCA85.5 ± 1.384.9 ± 1.2
w/o IFFN + cos-DCCA81.1 ± 2.678.4 ± 2.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, C.; Li, X.; Chi, J.; Cao, M.; Gu, Q.; Liu, J. ICTD: Combination of Improved CNN–Transformer and Enhanced Deep Canonical Correlation Analysis for Eye-Movement Emotion Classification. Brain Sci. 2026, 16, 330. https://doi.org/10.3390/brainsci16030330

AMA Style

Zhang C, Li X, Chi J, Cao M, Gu Q, Liu J. ICTD: Combination of Improved CNN–Transformer and Enhanced Deep Canonical Correlation Analysis for Eye-Movement Emotion Classification. Brain Sciences. 2026; 16(3):330. https://doi.org/10.3390/brainsci16030330

Chicago/Turabian Style

Zhang, Cong, Xisheng Li, Jiannan Chi, Ming Cao, Qingfeng Gu, and Jiahui Liu. 2026. "ICTD: Combination of Improved CNN–Transformer and Enhanced Deep Canonical Correlation Analysis for Eye-Movement Emotion Classification" Brain Sciences 16, no. 3: 330. https://doi.org/10.3390/brainsci16030330

APA Style

Zhang, C., Li, X., Chi, J., Cao, M., Gu, Q., & Liu, J. (2026). ICTD: Combination of Improved CNN–Transformer and Enhanced Deep Canonical Correlation Analysis for Eye-Movement Emotion Classification. Brain Sciences, 16(3), 330. https://doi.org/10.3390/brainsci16030330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop