Next Article in Journal
Comparison of Auditory Stream Segregation Abilities and Cerebral Asymmetry in Processing Speech in Noise in Carnatic Musicians, Bharatanatyam Dancers, and Non-Trained Individuals
Previous Article in Journal
Perceptual Decision Advantages in Open-Skill Athletes Emerge near the Threshold of Awareness: Behavioral, Computational, and Electrophysiological Evidence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Cross-Subject Band-Power Complexity Metric for Detecting Mental Fatigue Through EEG

by
Ang Li
1,2,3,
Zhenyu Wang
1,
Tianheng Xu
1,
Ting Zhou
4,
Xi Zhao
4,
Honglin Hu
1,2,* and
Marc M. Van Hulle
3,*
1
Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Laboratory for Neuro- & Psychophysiology, Department of Neurosciences, KU Leuven—University of Leuven, B-3000 Leuven, Belgium
4
School of Microelectronics, Shanghai University, Shanghai 200444, China
*
Authors to whom correspondence should be addressed.
Brain Sci. 2026, 16(2), 199; https://doi.org/10.3390/brainsci16020199
Submission received: 17 January 2026 / Revised: 4 February 2026 / Accepted: 5 February 2026 / Published: 7 February 2026

Highlights

What are the main findings?
  • The proposed Short-Term Second-Order Differential Entropy (ST-SODE) can capture fatigue from short-term band-power dynamics.
  • ST-SODE improves the robustness of cross-domain EEG fatigue detection.
What are the implications of the main findings?
  • ST-SODE reduces calibration burden for real-world fatigue monitoring.
  • ST-SODE enables lightweight cross-subject deployment.

Abstract

Background/Objectives: Electroencephalography (EEG) is a promising modality for fatigue detection because it directly reflects neural states; however, it is hindered by the need for subject-specific calibration and its reliance on unstable labeling. Moreover, classical EEG features are sensitive to intrinsic brain rhythm variations, causing pronounced domain shifts that degrade performance across sessions and subjects. Methods: Motivated by the biological fatigue rebound mechanism, we propose a robust cross-subject metric which we name Short-Term Second-Order Differential Entropy (ST-SODE). ST-SODE effectively suppresses the interference of background brain rhythms, enhancing robustness to cross-domain drift; consequently, its one-dimensional output can provide an indication of fatigue states without additional model training. Results: ST-SODE is validated on the public driving fatigue regression dataset SEED-VIG and on a private Vigilance classification dataset based on the N-Back task. ST-SODE achieves a correlation coefficient of 0.56 on SEED-VIG dataset (vs. 0.4 for differential entropy, DE) and a binary classification accuracy of 93.75% on the Vigilance dataset, outperforming other EEG-based fatigue metrics. Conclusions: ST-SODE offers a reliable solution for deployment in fields such as driving, manufacturing, and healthcare, where it could reduce safety incidents caused by fatigue.

1. Introduction

Statistics from the World Health Organization (WHO) show that 1.19 million people die from road traffic accidents every year [1]. Fatigue is one of the main culprits among the factors causing traffic accidents, predominantly in terms of rear-end car accidents; as such, fatigue is responsible for both human casualties and economic burden to society [2,3]. Fatigue reduces a driver’s vigilance, affecting the normal execution of driving tasks such as attention level, braking decisions, and reaction times [4,5]. Therefore, how to monitor and reduce driving fatigue has been a focus of research on traffic safety.
From the perspective of fatigue detection, three methodologies are used: physical feature identification [6], vehicle trajectory assessment [7], and biological signal assessment [8,9,10]. Biological signal assessment methods, among which Electroencephalography (EEG) stands out prominently, offer an effective and intuitive means of gauging a driver’s mental state. Consequently, signals such as EEG are frequently used to assess driver fatigue [11,12,13].
However, EEG signals are typically noisy and vary among subjects, leading researchers to develop a number of decoding methods. These methods primarily belong to two categories: the first involves traditional feature construction, in which time and frequency domain features such as Differential Entropy (DE) [14] and Power Spectral Density (PSD) [15,16] are constructed from EEG signals and supplied to machine learning classifiers; the second involves neural networks such as CNNs and RNNs [17,18,19,20], which can directly learn hidden features related to fatigue.
Although researchers have proposed various fatigue detection methods based on machine learning and deep learning, insufficient attention has been paid to the cross-session and cross-subject performance of these methods, which can lead to overfitting in practical applications because of lack of training data.
The brain exhibits rhythmic phenomena in which the power across different frequency bands of the brain varies over time and between subjects [21,22]. Classical fatigue features such as DE features use the magnitude of frequency-band power as an indicator of fatigue intensity; as such, they are susceptible to intrinsic fluctuations in brain rhythms, resulting in feature distribution shifts across subjects and across time periods. Deep learning models impose stricter requirements on feature distribution consistency, but these models typically need data recollection and model retraining for cross-subject and cross-session applications, which is a highly time-consuming process.
Therefore, this work aims to find a more robust metric to quantify mental fatigue. Fatigue is a protective response of the brain to prolonged high workloads. In realistic scenarios, the subject needs to maintain vigilance while driving or working. Through long-term experiments, we found that normally, when subjects feel tired, if they allow fatigue to develop, then they naturally fall into sleep. This is reflected by changes in the magnitude of the band’s power within the DE feature description [14]. When subjects fight against fatigue, such as during driving or working, this is manifested by fluctuation in the EEG frequency band power, known as the fatigue rebound effect [8,23,24]. In other words, as mental fatigue develops, EEG-band power does not exhibit a monotonic drift from vigilance to fatigue; instead, it typically progresses from a relatively stable alert state to a transient oscillatory phase, subsequently settling into a new stable state. Focusing on the fluctuations instead of on changes in the absolute magnitude helps to reduce the confounding effects of baseline cross-subject and cross-session variations.
Here, we use these fluctuations over a certain time window to describe the subject’s fatigue state during that period and propose the Short-Term Second-Order Differential Entropy (ST-SODE) metric to describe human fatigue. Because ST-SODE depends on fluctuation patterns rather than the absolute band-power magnitude, it is less sensitive to baseline shifts and scale differences across subjects and sessions, allowing for improved robustness under cross-subject and cross-session variability. We validated ST-SODE on a public fatigue dataset and a private binary vigilance–fatigue dataset. The results show that ST-SODE consistently outperforms existing metrics on both regression and cross-subject classification tasks. The contributions of this study are as follows:
  • This study proposes using the fatigue rebound effect to assess fatigue levels, then introduces ST-SODE to characterize this phenomenon.
  • ST-SODE describes the fluctuation of the short-term frequency-band power of the EEG signal without training; the one-dimensional output can directly represent fatigue states with cross-subject and cross-session robustness.
  • We provide theoretical evidence that short-time DE features computed from long EEG sequences approximately follow a Gaussian distribution, which supports the statistical modeling assumptions of ST-SODE.
  • This study includes a fatigue experiment based on the N-Back task for classification purposes. The proposed method was validated on both the N-Back dataset and the public SEED-VIG dataset. The results show that ST-SODE outperforms DE and PSD features in cross-subject scenarios using the Leave-One-Subject-Out (LOSO) strategy.
The rest of this paper is organized as follows: Section 2 introduces related works; Section 3 describes the proposed method; Section 4 introduces the datasets and experiment; Section 5 and Section 6 respectively present and discuss the results; finally, we conclude the paper in Section 7.

2. Related Works

An EEG-based fatigue detection system typically comprises four main components: signal acquisition, preprocessing, decoding, and postprocessing. Signal acquisition involves using an EEG cap to record the EEG signals associated with fatigue and vigilance states. Signal decoding forms the system’s core and includes two critical stages: feature construction or extraction, and decoding. Feature construction or extraction aims to derive key features that facilitate effective classification, employing methods such as the wavelet transform, Differential Entropy (DE) [25], Power Spectral Density (PSD) [16,26], etc. Extracted features such as entropy and time–frequency domain characteristics are utilized to differentiate between fatigue and vigilance states. Finally, features are decoded by training a classifier or regressor using machine learning methods. The results also require postprocessing (outlier detection, etc.) before application.

2.1. Feature Extraction

Methods for extracting EEG fatigue features primarily include time–frequency domain features and nonlinear features. Cui et al. [27] utilized the Fast Fourier Transform (FFT) to extract frequency-based features, integrating these with the Feature Weighted Episodic Training (FWET) classifier to detect fatigue. Zhang et al. [28] applied the Discrete Wavelet Transform (DWT) to extract entropy-based features, including the Wavelet Entropy (WE), Peak-to-Peak Approximate Entropy (PPApEn), and Peak-to-Peak Sample Entropy (PP-SampEn) to classify four driving fatigue stages. Duan et al. [25] were the first to employ DE features for driving fatigue detection. As a benchmark feature provided by the SEED-VIG dataset, DE has subsequently been widely cited and utilized in numerous studies [29,30]. When combining various features that have been proposed for EEG-based fatigue detection, the resulting high-dimensional feature space calls for dimensionality reduction or a feature selection process prior to or as part of the training process of the classifier or regressor. However, dimensionality reduction requires extensive labeled data, and challenges generalization to new subjects due to the cross-subject and cross-session variability of EEG signals.

2.2. Traditional Machine Learning Algorithms Applied in Classification

The classical machine learning methods of Support Vector Machine (SVM) and Support Vector Regression (SVR) for regression problems are extensively utilized in EEG-based fatigue detection. Shen et al. [31] proposed a probabilistic multiclass SVM detector for automatically classifying mental fatigue levels from EEG data. Li et al. [32] introduced the Auto-Correlation Function-based Sparse Support Matrix Machine (ACF-SSMM) algorithm, which incorporates location information into EEG signals and employs sparsity to reduce redundant features, resulting in enhanced fatigue detection performance. Wang et al. [33] extracted four graph-related features using Partial Directed Coherence (PDC)-based brain functional network construction. They utilized SVM for offline classification, achieving an accuracy of 87.16% for binary classification in a simulated driving experiment. For the SEED-VIG dataset labeled with continuous values, Zheng et al. [34] applied Differential Entropy (DE) features and Support Vector Regression (SVR) for regression-based prediction. In addition to SVM-based methods, other machine learning classifiers have also been applied to EEG-based fatigue detection, such as K-Nearest Neighbor (KNN) [35], Bayesian Neural Network (BNN) [36], and Feature Weighted Episodic Training (FWET) [27].

2.3. Deep Learning Models

Compared to traditional machine learning algorithms, deep learning methods require larger training datasets but offer significantly better prediction performance on test data within the same subject. Convolutional Neural Networks (CNN) are widely employed in image recognition and classification [17,18,37]. They possess the ability to automatically and adaptively learn the hierarchical feature space at each level, which effectively reduces the number of learned parameters [17]. Recent studies have highlighted the strong feature extraction capabilities of CNNs in EEG analysis. For instance, Wu et al. [18] proposed a deep sparse contractive autoencoder network combined with Finite Impulse Response (FIR) filters to extract local features and predict fatigue status, achieving 94.58% classification accuracy on a three-class private flight simulation fatigue dataset. Similarly, Hu et al. [19] introduced the Spatio-Temporal Fusion Network with Brain Region Partitioning Strategy (STFN-BRPS), a multi-branch deep learning network designed to enhance the accuracy and robustness of EEG-based driver fatigue detection. The model incorporates a recurrent multiscale convolution module for temporal features, a dynamic graph convolution module for spatial features, and a channel attention-based feature fusion module. This approach demonstrated superior performance compared to conventional methods on real-world driving tests.
While deep learning enhances feature extraction performance for specific domains, it is susceptible to overfitting [38]. EEG data exhibit significant variability across subjects [39,40], which means that new data must be collected from each subject in order to retrain the model or recalibrate the original model. This requirement poses a major challenge to the practicality of EEG-based fatigue detection methods.

3. Methods

3.1. EEG Frequency Bands

An Electroencephalograph (EEG) is a biosignal representing changes in surface potentials on the scalp, which reflect the electrophysiological activity of neural cells in the brain. EEG features are generally defined by their distinct frequency ranges and spatial patterns. Functionally, EEG frequency domain features are categorized into the five frequency bands shown in Table 1. Notably, the alpha and theta bands are strongly associated with fatigue [41,42,43,44]. Specifically, when subjects are mentally fatigued, their theta-band power will increase significantly. Their alpha-band power also increases; however some studies have observed weakening of the alpha band due to engagement in task-related areas [45]. The beta-band power may also weaken; however, the specific performance varies greatly between subjects.
To provide a more stable representation of fatigue, the frequency-band ratios θ / β , α / β , ( α + θ ) / β and ( α + θ ) / ( α + β ) are commonly used. This approach helps to mitigate baseline shifts that often occur when relying solely on a single frequency band over extended periods. However, the power within EEG frequency bands is sensitive to both long-term shifts and transient fluctuations caused by intrinsic physiological states or external environmental factors. These short-term variations can lead to instability in features from combined frequency bands, posing challenges for accurate fatigue prediction. Therefore, predicting the fatigue state of a subject solely based on the increase or decrease in frequency-band power is unreliable. In practical applications, frequency-band power features are extracted first, then machine learning methods are applied to learn their characteristics.

3.2. Short-Term Second-Order Differential Entropy

The framework of the proposed Short-Term Second-Order Differential Entropy (ST-SODE) method is illustrated in Figure 1. To quantify fatigue levels using EEG signals, we employ a dual-windowing strategy comprising the epoch window ( T e ) and the aggregation window ( T a ). The epoch window T e is a short-time window used to extract the instantaneous spectral features, while the aggregation window T a is a longer sliding window designed to capture the temporal variability of these features. Let N e denote the number of epoch windows contained within one aggregation window, f s the sampling rate, and s the fixed step size between adjacent epoch windows.
For each epoch window T e , the power spectrum is computed using Welch’s power spectral density estimate [46]:
P T e ( f ) = 1 N n = 0 N 1 w ( n ) · x ( n ) e j 2 π f n / f s 2
where w ( n ) is the Hamming window, represented as
w ( n ) = 0.54 0.46 cos 2 π n N 1 , n = 0 , 1 , 2 , , N 1 .
Then, the band power E i is calculated in four frequency bands (delta: 0.5–4 Hz theta: 4–8 Hz, alpha: 8–12 Hz, beta: 13–30 Hz). The Differential Entropy (DE) was initially defined as
h ( X ) = f X ( x ) log f X ( x ) d x = 1 2 π σ 2 exp ( x μ ) 2 2 σ 2 log 1 2 π σ 2 exp ( x μ ) 2 2 σ 2 d x = 1 2 log 2 π e σ 2 ,
with the time series X assumed to follow a Gaussian distribution N ( μ , σ 2 ) . It has since been proven that for a fixed-length EEG sequence, the DE is equivalent to the logarithm of the energy spectrum or the signal variance in a certain frequency band [14].
Subsequently, based on Section 3.1, the logarithm of the energy spectrum E i features is calculated. To capture the global state, the features are pooled by averaging across all channels and combined through frequency band i { δ , θ , α , β } to obtain the final DE feature D ϕ for each epoch window:
D ϕ = ϕ { D E δ , D E θ , D E α , D E β } ,
where ϕ represents the different band-power combinations discussed in Section 3.1.
Next, we prove that the distribution of DE features of windows from long-length EEG signals approximates a similar Gaussian distribution. Let signal segment X of length N follow a stationary Gaussian distribution, denoted as X N ( μ , σ 2 ) . The estimator for the Differential Entropy (DE), denoted as h ^ , is provided by
h ^ = 1 2 ln ( 2 π e S 2 ) ,
where S 2 represents the sample variance of the segment. According to the Central Limit Theorem [47], as the sample size N , the sample variance S 2 exhibits asymptotic normality:
N ( S 2 σ 2 ) d N ( 0 , 2 σ 4 ) .
This implies that for large N, S 2 is approximately distributed as N ( σ 2 , 2 σ 4 N ) .
To derive the distribution of h ^ , we apply the Delta method [48]. Let g ( u ) = 1 2 ln ( 2 π e u ) , such that h ^ = g ( S 2 ) . The first-order Taylor expansion of g ( S 2 ) around the true variance σ 2 is
h ^ g ( σ 2 ) + g ( σ 2 ) ( S 2 σ 2 ) .
The derivative of the function is g ( u ) = 1 2 u . Evaluated at σ 2 , we have g ( σ 2 ) = 1 2 σ 2 .
Because a linear transformation of a Gaussian random variable remains Gaussian, h ^ follows a normal distribution asymptotically. The variance of h ^ is derived as follows:
Var ( h ^ ) g ( σ 2 ) 2 · Var ( S 2 ) = 1 2 σ 2 2 · 2 σ 4 N = 1 4 σ 4 · 2 σ 4 N = 1 2 N .
Consequently, as N , the segmented DE values follow a similar Gaussian distribution:
h ^ N μ r e a l , 1 2 N .
Based on this proof, over the aggregation window T a containing N e epoch windows, we calculate the ST-SODE feature as the variance of the sequence { D ϕ ( T e 1 ) , , D ϕ ( T e n ) } . The ST-SODE is given by
S Φ ( T a ) = 1 N e 1 i = 1 N e D Φ ( T e i ) D ¯ Φ ( T a ) 2 ,
where D ¯ Φ ( T a ) is the mean DE value over the aggregation window and N e is the number of epoch window in one aggregation window. It has been proven in [14] that the variance estimation of signal sequence is just its average energy; therefore, the above operation can be regarded as a simplified version of the DE computation. We do not further apply a logarithmic operation here, since the potential negative values are undesirable for directly outputting the fatigue index. However, due to the monotonicity of the logarithm, its variation trend remains strictly consistent with that of the standard DE. A larger ST-SODE value indicates greater fluctuation and instability in the relative activity between the two frequency bands.

4. Datasets and Experiment Setup

In this study, we utilized two EEG-based fatigue datasets. The first public dataset, SEED-VIG, was collected by Shanghai Jiao Tong University and contains EEG signals recorded during simulated driving tasks. Eye-tracking devices were employed to monitor blink and closure, which served as the basis for fatigue regression labeling. The second dataset is a private dataset using the N-Back task to induce fatigue in subjects. In a typical N-Back paradigm, subjects continuously respond to a stream of stimuli by indicating whether the current stimulus matches the one presented N steps earlier, thereby demanding working memory and sustained attention. The N-Back task was initially introduced by Kirchner in [49] to impose varying cognitive workloads. Subsequently, Tanaka et al. [50] and Pergher et al. [51] demonstrated that the N-Back task can effectively induce fatigue when its design closely mirrors that of the control 1-Back task. The 1-Back task can induce fatigue, but less than the 2-Back task because of the higher task difficulty. Other fatigue classification studies have also employed N-Back experiments to induce varying degrees of fatigue [52,53]. Detailed descriptions of the experiment setup and data formats for both datasets are provided in the following subsections.

4.1. SEED-VIG

The first dataset was compiled and released by a research team at Shanghai Jiao Tong University [34]. They developed a simulated driving system to collect EEG data. The system was composed of a large LCD screen and a real vehicle without engine and other components. The vehicle was modified so the participants could operate the vehicle through the steering wheel and gas pedal using a screen in front of them. To induce fatigue, the LCD screen displayed a primarily straight and monotonous four-lane highway scene.
A total of 23 volunteers with an average age of 23.3 years participated in the fatigue driving experiments. To facilitate the induction of driving fatigue, subjects were required to participate in the experiment in the afternoon or late at night. The experiment lasted about 2 h, during which data were recorded. Both EEG and forehead EOG signals were recorded during the fatigue driving experiment; however, only EEG data are analyzed in this paper. EEG signals were recorded from eleven channels at the posterior site and six channels at the temporal site, following the international 10–20 electrode system. The sampling rate was 200 Hz. An annotation, called the PERCLOS measure, was adopted to quantify subjects’ fatigue. PERCLOS indicates the percentage of eye blink and closure, with the PERCLOS value between 0 (vigilance) and 1 (fatigue).

4.2. Vigilance Dataset

4.2.1. Experimental Protocol

The second dataset consists of a binary vigilance–fatigue dataset collected from eight subjects. To induce fatigue, we designed an experiment based on cognitive load experiments in psychology. In the N-Back task, participants are presented with a sequence of stimuli, which can be either visual or auditory, and must identify whether the current stimulus matches the one presented “N” steps earlier in the sequence. The value of “N” can be varied to adjust the difficulty of the task, with larger values requiring participants to hold more items in memory and track them over longer intervals.
In this study, we used different colors as stimuli and carried out two tasks for each subject, as illustrated in Figure 2. The 1-Back task was used to represent a relative vigilant working state, while the 2-Back task was used to induce the fatigue working state. To minimize interference between the two tasks, each subject first completed the 1-Back task; after a 10-min rest, the same subject then performed the 2-Back task. For the 1-Back task, the monitor displayed changing color blocks, each lasting 2 s with an interval of 1 s. The subject was instructed to concentrate on the screen and press the space bar when two consecutive colors were the same. For the 2-Back task, the subject was instructed to press the space bar when the current color was the same as two steps prior. Both tasks consisted of two rounds of 5 min each, with no interval between them.
Eight healthy subjects were recruited, seven males and one female (age: 23.75 ± 2.11 years, mean ± standard deviation). All subjects reported normal or corrected-to normal vision, with no history of substance addiction or mental disorders. The subjects were required to obtain a full night (>7 h) sleep before the day of the experiment. On the day of the experiment, they were required to avoid consuming caffeine or alcohol. The experiment was conducted in an electromagnetically shielded room. Before the start of the experiment, the subjects were trained to familiarize themselves with the N-Back task and to understand the necessity of completing the tasks as accurately as possible even when feeling tired. The study complied with the principles of the 2024 Declaration of Helsinki and was approved by the Research Ethics Committee of Shanghai University (no. ECSHU 2024-072). All subjects signed informed consent forms.

4.2.2. EEG Data Acquisition

Brain activity was recorded using EEG recording equipment (Compumedics Neuroscan, Charlotte, NC, USA) with 28 wet electrodes using the 10–20 system (without forehead electrodes, according to SEED-VIG). The selection and position of the channels are shown in Figure 3. The sampling rate was 1000 Hz. The impedances of all EEG channels were kept below 10 k Ω .

4.3. Experiment Setup

4.3.1. Effectiveness Experiment on SEED-VIG

The label of SEED-VIG is a floating value from 0 to 1; thus, we use the prediction curves and correlation analysis to demonstrate the performance of ST-SODE. The SEED-VIG dataset contains three parts: EEG (temporal lobe and posterior lobe), Forehead EEG, and EOG. Only the EEG data were utilized in this experiment, with the Forehead EEG and EOG data left out. Forehead electrodes can cause eye movement artifacts; after these data were removed, the raw EEG data were processed with a fourth-order Butterworth band-pass filter between 0.5 and 45 Hz to reduce artifacts and noise, referencing [34]. The filtering was completed before epoch window segmentation to reduce the edge effect. The epoch window size was set to 0.5 s and the aggregation window size to 8 s. The epoch window strategy was non-overlapping, with the step size the same as the epoch window length. After ST-SODE, the result was min–max normalized, then displayed in the same figure with PERCLOS labels.

4.3.2. Vigilance Dataset

The Vigilance dataset involves two tasks with different levels of workload, namely, the 1-Back task and 2-Back task. Each task consisted of two rounds, each of which began with the cue “Round 1/2 begins”. We selected the first round of the 1-Back task as the vigilant case and the second round of the 2-Back task as the fatigue case. The preprocessing was the same as SEED-VIG. The epoch window size was set at 0.5 s and the aggregation window size at 5, 10, and 20 s to assess the effect on the results. Each subject had 120, 60, and 30 trials for 5 s, 10 s, and 20 s, with fatigue and vigilance being equally divided. To classify fatigue without training on the test subject, we calculated the average value of the obtained segmented ST-SODE features from the other seven subjects, excepting the test subject. The training data length was [840/420/210, 5/10/20 * 1000 (Hz)] (trials, data points). Features from the test subject that were above the average value were predicted as fatigue, and vice versa as vigilance. The Differential Entropy (DE) and Power Spectral Density (PSD) were employed as a comparison. DE features were processed consistently with ST-SODE, and a Support Vector Machine with a Gaussian kernel was used to classify PSD features using the LOSO strategy, with the data of the other seven subjects excepting the test subject used for training. To compare the performance of ST-SODE with supervised subject-specific methods, we also conducted within-subject 5-fold cross-validation experiments using the DE and PSD features with supervised SVM classifiers.
In addition, we calculated the average ST-SODE results of fatigue and vigilance trials for each subject.

5. Results

5.1. Performance on SEED-VIG

A comparison of the PERCLOS labels and ST-SODE results among all subjects is depicted in Figure 4. ST-SODE (blue line) exhibits varying degrees of correlation with PERCLOS (red line), which are collected in Table 2. The performance is divided into three categories, as discussed below:
  • Most experiments show significant correlation between ST-SODE and PERCLOS changes: S01, S03, S04, S05, S06, S07, S10, S14, S15, S17, S18, S20, and S21. All their correlation coefficients are above average 0.56. Their peaks and troughs along the rising and falling edges remain largely coincident. Because ST-SODE and PERCLOS are both complete data modalities, the heights of the two curves may differ.
  • The second category consists of correlation coefficients between 0.2 to 0.56 but which still exhibit strong correlation in the image. This is specifically manifested as follows: when PERCLOS shows a peak in a short period of time, ST-SODE also demonstrates a similarly temporally coincident fluctuation. This category contains S11, S12, S13, S16, S19, S22, S23.
  • The third category comprises subjects with a correlation coefficient below 0.2. This group includes S02, S08, and S09. For S09, it can be observed that the two curves coincide almost exactly when there is a crest.
At the same time, we also calculated the performance of PSD and DE, which are widely used as features in fatigue prediction. In order to show the result without training, the processing method of PSD and DE was kept consistent with ST-SODE. The feature average of all channels was calculated to obtain a one-dimensional time series, then the correlation coefficient was calculated with the label. The results are shown in Table 2.
According to Table 2, when the PSD and DE features are not trained using classifiers with labels, their correlation with the labels is lower than ST-SODE. Among the four different frequency band combinations, the highest PSD is only 0.0325, the DE is 0.4001, and the chance level of the random signal between 0 and 1 is 0.005. The untrained ST-SODE method can achieve an average correlation coefficient of 0.56, significantly higher than DE or PSD (** p < 0.01 and *** p < 0.001, Wilcoxon signed-rank), which proves the effectiveness of ST-SODE in predicting fatigue without training.

5.2. Performance on Vigilance Dataset

The classification results of ST-SODE and the two compared methods (DE and PSD) are shown in Table 3 and Figure 5. Each method employs three different aggregation window lengths (5S, 10S, 20S).
It can be observed that the mean accuracy of ST-SODE across all three window lengths significantly exceeds those of the DE and PSD-SVM methods. ST-SODE achieves the highest average accuracy (93.75%) at a window length of 20 s, which is 6.46% higher than at 5 s and 1.46% higher than at 10 s. However, while increasing the sampling window length improves accuracy, it also leads to a reduction in system response speed. For the DE and PSD methods, accuracy significantly improves as the window length increases from 5 s to 10 s, with gains of 3.86% and 11.88%, respectively; however, when the sampling window length is further extended from 10 seconds to 20 s, DE and PSD show minimal accuracy gains or even slight declines, with changes of 0.62% and −1.05%, respectively. We believe that this is because keeping the total data length fixed while increasing the size of the sampling window results in a reduced number of training samples.
To quantify the magnitude of performance differences beyond statistical significance, we additionally report the rank-biserial correlation r as an effect-size measure for the paired Wilcoxon signed-rank test. The value of r ranges from [ 1 , 1 ] , with r > 0 indicating that ST-SODE achieves higher accuracy than the comparator and larger | r | implying a stronger and more consistent advantage across subjects. Based on the subject-wise paired accuracy results in Table 3, ST-SODE exhibits positive effect sizes over DE, with r = 0.72 / 0.50 / 0.69 for window lengths of 5 s/10 s/20 s. Compared to PSD-SVM, the effect sizes are r = 0.67 / 0.29 / 0.52 for 5 s/10 s/20 s, respectively. Overall, these results suggest that ST-SODE provides a consistent improvement over both DE and PSD-SVM across subjects.
In Table 4, we report the performance comparison between the cross-subject ST-SODE method and DE/PSD features with supervised SVM classifiers. The comparison indicates that the within-subject supervised method DE-SVM achieves the highest accuracy (97.50–98.44%), representing an accuracy improvement of 3.75–11.15% at the 20 s and 5 s window settings compared to the cross-subject ST-SODE. This performance gap reflects the advantage of subject-specific calibration in capturing individual EEG characteristics. Therefore, the choice of whether to sacrifice accuracy in order to avoid subject-specific calibration represents a tradeoff that users must evaluate based on their practical requirements.
In Table 5, it can be observed that the average of ST-SODE on the vigilance task is significantly lower than on the fatigue task (** p < 0.01). For example, when using an aggregation window length of 5 s, the average of ST-SODE on the vigilance task data is 0.000077, while on the fatigue task data it reaches 0.064361. Although ST-SODE scores higher for some subjects on the vigilance task, such as Subject 5 with 0.000257, this is still much lower than the lowest score on the fatigue task (Subject 1 with 0.000893). Meanwhile, the smallest difference between vigilance and fatigue is also from Subject 1, with 0.000079 and 0.000893, respectively, while the largest difference is from Subject 5, with 0.000257 and 0.219803.

5.3. Effectiveness Analysis

5.3.1. Performance When Using Other Band Power Ratios

ST-SODE uses α / β as the band-power combination due to its high performance compared to other combinations. We calculated the Spearman correlation between the predictions and PERCLOS on SEED-VIG using different band power ratios of θ / β ,   α / β , ( α + θ ) / β and ( α + θ ) / ( α + β ) , with the results shown in Table 2. The chance level is calculated using the shuffled label and the results of α / β .
From Table 2, it can be seen that α / β achieves the highest correlation among the four methods. The results obtained from the other three combinations are poorly correlated with the PERCLOS labels.

5.3.2. Feature Visualization

To validate the representation ability of the proposed method, we plot the t-SNE [54] transformed features in Figure 6.
The t-SNE is a nonlinear manifold learning technique that projects high-dimensional feature vectors into a low-dimensional space (typically 2D) for visualization by preserving local neighborhood relationships. Therefore, points that are closer together on the 2D map generally indicate more similar representations.
Specifically, we exhibit the DE, PSD, and ST-SODE features extracted from all subjects’ data in the Vigilance dataset. We used a learning rate of 500 and an iterative display frequency of 20 for the t-SNE calculation.
From the visualized features represented in the figure, it can be observed that the ST-SODE features are more class-discriminative than the other features. Additionally, the DE features are distributed in several regions which contain both vigilance and fatigue features. This issue illustrates why decoding methods are prone to significant deterioration on cross-subject tasks. Meanwhile, it is notable that some points of the PSD features are close to straight lines, which proves that features from the same subject in the same state are very similar compared to ST-SODE and DE while being very different from other subjects. This proves the difficulty of training a classifier that is universal for all subjects, as it is challenging to learn a hyperplane (or neural network) that can effectively distinguish the two types of states based on the confusing features.

6. Discussion

6.1. Topography of Channel-Level ST-SODE

The standard ST-SODE does not focus on the features between channels, but instead averages the values of all channels. To explore the spatial variation of ST-SODE across the scalp, Figure 7 shows the channel-level ST-SODE topography.
Notably, the fatigue condition shows significant increases in ST-SODE values within particular cortical regions for most subjects. Specifically, Subjects 1, 2, 4, 7, and 8 exhibit high ST-SODE activity in the left temporal electrodes and left parietal electrodes, while Subjects 5 and 6 show stronger ST-SODE features in the occipital electrodes.
This suggests that it is difficult to use individual channels for ST-SODE detection. The spatial differences in performance during fatigue and non-fatigue are different across subjects; therefore, attempting to use all of the channels for ST-SODE analysis can enhance generalizability to all subjects. However, when considering realistic application scenarios such as driver fatigue detection, it may not be possible to provide electrode caps that cover the whole brain, as in laboratory scenarios. Consequently, employing localized portable electrode setups centered on the left posterior region would optimize detection efficacy within these practical constraints.

6.2. Importance of Cross-Subject Method

Previous studies have shown that neuronal activities exhibit variability in the frequency and time domains over the states of vigilance and fatigue [11,12]. This variability is mainly exhibited by changes in the frequency domain features. Recent neurophysiological studies have suggested that this phenomenon captured by ST-SODE is likely attributable to Locus Coeruleus (LC) oscillations [55,56]. During fatigue, LC neurons exhibit functional fatigue and infra-oscillations. These fluctuations in noradrenergic gain cause cortical networks to oscillate between low-excitability and high-excitability states, generating the high spectral variance that ST-SODE measures. This mechanism aligns with observations by Doran et al. [57], who noted unstable fluctuations in attentional states during sleep deprivation. Most existing studies focus on utilizing a classifier trained on a single subject. This provides stronger fitting ability, but at the same time constrains realistic applications due to the need for subject-specific training data. In real-world situations, especially driving fatigue detection, there is urgent need for a white-box algorithm with clear internal workings and decision-making logic.
ST-SODE metrics effectively meet this requirement. Our findings reveal significant power fluctuation on EEG bands when subjects are in a fatigued state. The proposed ST-SODE method is designed based on this phenomenon to be white-box and training-free, and as such can be used in realistic fatigue detection scenarios.
Compared to PSD and DE features, ST-SODE features are less susceptible to shifts in frequency-band power across session and subject. Under normal conditions, frequency-band power varies with circadian rhythms [58] and arousal levels [59], making it challenging to evaluate fatigue changes using absolute power values across different periods. In contrast, ST-SODE features are unaffected by absolute value changes, enabling consistent fatigue standards to be applied for monitoring drivers across various sessions.
It is important to clarify the definition of “calibration-free” in this context. Although ST-SODE does not require any subject-specific calibration (where a new subject must provide labels to train a personal model), it still requires preparation for the classification threshold. As demonstrated by the LOSO setting, the threshold is derived from existing subjects. Therefore, the system preparation phase requires a database of labeled subjects in order to establish the threshold before the system can be applied to new subjects.

6.3. Labels of Fatigue Experiment

One of the major challenges in fatigue detection is labeling. The fatigue state of a subject is highly random and unstable, and is also affected by various factors. Sleeping state, circadian rhythms, task difficulty, and physical differences can all affect the fatigue level of the subject [60,61]. Most fatigue monitoring methods use supervised learning methods that require labeled data to fit a classifier [25,27,28,29,62], in which case the accuracy of the labeling affects the performance of the model for fatigue prediction.
Previously, fatigue was usually inferred indirectly by analyzing subject metrics such as eye blinks [63], frequency of yawns [64], and task errors [65]. However, there are many other factors associated with these signs. A classifier obtained by training on indirect labels still predicts indirect labels, leaving the correlation between fatigue level and indirect labels in need of further validation.
Our Vigilance dataset used cognitive load to induce the fatigue state. It was designed to produce a similar state to driving or performing other long-time tasks by means of a high-difficulty repetitive task over a short period of time, while utilizing a low-difficulty task to control for variables. We designed the metrics with reference to the properties of the brain when fatigued in order to avoid introducing other disturbances using supervised learning.
The test results on the Vigilance dataset demonstrate that ST-SODE metrics can effectively distinguish between two different work states. The process is a result-oriented one, and we consider that fatigue can be effectively detected if the two states can be distinguished.

6.4. Limitations

Despite its promising results, this study still has limitations that need to be addressed in future work. First, while ST-SODE achieves promising performance on the SEED-VIG dataset, poor performance on some subjects indicates that this metric alone is currently insufficient for operational deployment in safety critical scenarios, and should ideally be integrated into multimodal and distributed monitoring systems [66] to ensure reliability. Second, the Vigilance dataset employed an N-Back task to induce fatigue through high cognitive load, which differs from the vigilance decline typical of real-world driving; consequently, the generalizability of these findings to real driving conditions requires further validation in future on-road experiments.
Third, the frequency-band ratio ( α / β ) utilized in ST-SODE was selected based on its best performance on the SEED-VIG dataset. While this combination demonstrated effectiveness for the Vigilance dataset, we acknowledge that this selection is empirical. Therefore, when applying ST-SODE to new scenarios, users would need to obtain labeled data from their domain in order to determine the optimal frequency-band combination.
Last, although proportional drift across bands offers a reasonable hypothesis for the effectiveness of ST-SODE in cross-subject scenarios, the mechanistic explanation remains empirically demonstrated but theoretically underspecified in the current study. Conclusive validation of these specific drift patterns requires further targeted research into the neurophysiological mechanisms.

7. Conclusions

This study has proposed a cross-subject EEG-based fatigue metric called ST-SODE which can be applied to scenarios such as driving, manufacturing, and healthcare that require low fatigue levels. ST-SODE utilizes the fatigue rebound effect to represent fatigue levels; this approach reports fatigue levels by calculating the fluctuation of short-window DE features, equivalent to the second-order differential entropy. We validated the effectiveness of ST-SODE on the public SEED-VIG dataset as well as on a private dataset based on the N-Back task. ST-SODE achieved a correlation coefficient of 0.56 on the SEED-VIG regression dataset and 93.75% binary classification accuracy on the Vigilance dataset, outperforming other metrics under the same LOSO settings. ST-SODE can be deployed in existing environments while requiring no subject-specific calibration, thereby addressing the limitations of supervised fatigue detection methods in real-world applications.

Author Contributions

Conceptualization, A.L.; methodology, A.L.; software, A.L.; validation, A.L. and Z.W.; formal analysis, A.L., Z.W. and M.M.V.H.; investigation, A.L., Z.W. and X.Z.; resources, T.X., T.Z., H.H. and M.M.V.H.; data curation, A.L. and X.Z.; writing—original draft preparation, A.L.; writing—review and editing, A.L., Z.W., H.H. and M.M.V.H.; visualization, A.L.; supervision, H.H. and M.M.V.H.; project administration, T.X., H.H. and M.M.V.H.; funding acquisition, Z.W., T.X., T.Z., H.H. and M.M.V.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant received from Horizon Europe’s Marie Sklodowska-Curie Action (grant agreement No. 101118964), by the Horizon 2020 research and innovation programme under grant agreement No. 857375, by the special research fund of the KU Leuven (C24/18/098), by the Belgian Fund for Scientific Research—Flanders (G0A4321N, G0C1522N, G031426N), by the Hercules Foundation (AKUL 043), and by the National Natural Science Foundation of China under Grant No. U24A20209.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Shanghai University (no. ECSHU 2024-072, date of approval 4 September 2024).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The SEED-VIG dataset is available on request at: https://bcmi.sjtu.edu.cn/home/seed/seed-vig.html (accessed on 5 April 2024). The Vigilance dataset presented in this study is available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. Global Status Report on Road Safety 2023; World Health Organization: Geneva, Switzerland, 2024. [Google Scholar]
  2. Zhang, H.; Wu, C.; Yan, X.; Qiu, T.Z. The effect of fatigue driving on car following behavior. Transp. Res. Part F Traffic Psychol. Behav. 2016, 43, 80–89. [Google Scholar] [CrossRef]
  3. Zou, S.; Qiu, T.; Huang, P.; Bai, X.; Liu, C. Constructing multi-scale entropy based on the empirical mode decomposition (EMD) and its application in recognizing driving fatigue. J. Neurosci. Methods 2020, 341, 108691. [Google Scholar] [CrossRef]
  4. Fountas, G.; Pantangi, S.S.; Hulme, K.F.; Anastasopoulos, P.C. The effects of driver fatigue, gender, and distracted driving on perceived and observed aggressive driving behavior: A correlated grouped random parameters bivariate probit approach. Anal. Methods Accid. Res. 2019, 22, 100091. [Google Scholar] [CrossRef]
  5. Chen, J.; Wang, H.; Wang, Q.; Hua, C. Exploring the fatigue affecting electroencephalography based functional brain networks during real driving in young males. Neuropsychologia 2019, 129, 200–211. [Google Scholar] [CrossRef]
  6. Gu, H.; Ji, Q. An automated face reader for fatigue detection. In Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition; IEEE: New York, NY, USA, 2004; pp. 111–116. [Google Scholar]
  7. Rezaei, M.; Klette, R. Look at the driver, look at the road: No distraction! no accident! In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 129–136. [Google Scholar]
  8. Lal, S.K.; Craig, A. A critical review of the psychophysiology of driver fatigue. Biol. Psychol. 2001, 55, 173–194. [Google Scholar] [CrossRef]
  9. Bergasa, L.M.; Nuevo, J.; Sotelo, M.A.; Barea, R.; Lopez, M.E. Real-time system for monitoring driver vigilance. IEEE Trans. Intell. Transp. Syst. 2006, 7, 63–77. [Google Scholar] [CrossRef]
  10. Mårtensson, H.; Keelan, O.; Ahlström, C. Driver sleepiness classification based on physiological data and driving performance from real road driving. IEEE Trans. Intell. Transp. Syst. 2018, 20, 421–430. [Google Scholar] [CrossRef]
  11. Fan, C.; Peng, Y.; Peng, S.; Zhang, H.; Wu, Y.; Kwong, S. Detection of train driver fatigue and distraction based on forehead EEG: A time-series ensemble learning method. IEEE Trans. Intell. Transp. Syst. 2021, 23, 13559–13569. [Google Scholar] [CrossRef]
  12. Wang, H.; Xu, L.; Bezerianos, A.; Chen, C.; Zhang, Z. Linking attention-based multiscale CNN with dynamical GCN for driving fatigue detection. IEEE Trans. Instrum. Meas. 2020, 70, 2504811. [Google Scholar] [CrossRef]
  13. Alghanim, M.; Attar, H.; Rezaee, K.; Khosravi, M.; Solyman, A.; Kanan, M.A. A hybrid deep neural network approach to recognize driving fatigue based on EEG signals. Int. J. Intell. Syst. 2024, 2024, 9898333. [Google Scholar] [CrossRef]
  14. Shi, L.C.; Jiao, Y.Y.; Lu, B.L. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); IEEE: New York, NY, USA, 2013; pp. 6627–6630. [Google Scholar]
  15. Youngworth, R.N.; Gallagher, B.B.; Stamper, B.L. An overview of power spectral density (PSD) calculations. Opt. Manuf. Test. VI 2005, 5869, 206–216. [Google Scholar]
  16. Jap, B.T.; Lal, S.; Fischer, P.; Bekiaris, E. Using EEG spectral components to assess algorithms for detecting fatigue. Expert Syst. Appl. 2009, 36, 2352–2359. [Google Scholar] [CrossRef]
  17. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015); Computational and Biological Learning Society: Washington, DC, USA, 2015. [Google Scholar]
  18. Deng, P.Y.; Qiu, X.Y.; Tang, Z.; Zhang, W.M.; Zhu, L.M.; Ren, H.; Zhou, G.R.; Sheng, R.S. Detecting fatigue status of pilots based on deep learning network using EEG signals. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 575–585. [Google Scholar] [CrossRef]
  19. Hu, F.; Zhang, L.; Yang, X.; Zhang, W.A. EEG-Based Driver Fatigue Detection Using Spatio-Temporal Fusion Network With Brain Region Partitioning Strategy. IEEE Trans. Intell. Transp. Syst. 2024, 25, 9618–9630. [Google Scholar] [CrossRef]
  20. Du, X.; Meng, Y.; Qiu, S.; Lv, Y.; Liu, Q. EEG emotion recognition by fusion of multi-scale features. Brain Sci. 2023, 13, 1293. [Google Scholar] [CrossRef]
  21. Buzsaki, G.; Draguhn, A. Neuronal oscillations in cortical networks. Science 2004, 304, 1926–1929. [Google Scholar] [CrossRef] [PubMed]
  22. Klimesch, W. EEG alpha and theta oscillations reflect cognitive and memory performance: A review and analysis. Brain Res. Rev. 1999, 29, 169–195. [Google Scholar] [CrossRef]
  23. Borghini, G.; Astolfi, L.; Vecchiato, G.; Mattia, D.; Babiloni, F. Measuring neurophysiological signals in aircraft pilots and car drivers for the assessment of mental workload, fatigue and drowsiness. Neurosci. Biobehav. Rev. 2014, 44, 58–75. [Google Scholar] [CrossRef]
  24. Dinges, D.F.; Powell, J.W. Microcomputer analyses of performance on a portable, simple visual RT task during sustained operations. Behav. Res. Methods Instrum. Comput. 1985, 17, 652–655. [Google Scholar] [CrossRef]
  25. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER); IEEE: New York, NY, USA, 2013; pp. 81–84. [Google Scholar]
  26. Libert, A.; Van Hulle, M.M. Predicting premature video skipping and viewer interest from EEG recordings. Entropy 2019, 21, 1014. [Google Scholar] [CrossRef]
  27. Cui, Y.; Xu, Y.; Wu, D. EEG-based driver drowsiness estimation using feature weighted episodic training. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 2263–2273. [Google Scholar] [CrossRef]
  28. Zhang, C.; Wang, H.; Fu, R. Automated detection of driver fatigue based on entropy and complexity measures. IEEE Trans. Intell. Transp. Syst. 2013, 15, 168–177. [Google Scholar] [CrossRef]
  29. Wang, D.; Tong, J.; Yang, S.; Chang, Y.; Du, S. EEG Signal Driving Fatigue Detection based on Differential Entropy. In Proceedings of the 2024 IEEE International Conference on Mechatronics and Automation (ICMA); IEEE: New York, NY, USA, 2024; pp. 543–548. [Google Scholar]
  30. Peng, B.; Zhang, Y.; Wang, M.; Chen, J.; Gao, D. TA-MFFNet: Multi-feature fusion network for EEG analysis and driving fatigue detection based on time domain network and attention network. Comput. Biol. Chem. 2023, 104, 107863. [Google Scholar] [CrossRef] [PubMed]
  31. Shen, K.Q.; Li, X.P.; Ong, C.J.; Shao, S.Y.; Wilder Smith, E.P. EEG-based mental fatigue measurement using multi-class support vector machines with confidence estimate. Clin. Neurophysiol. 2008, 119, 1524–1533. [Google Scholar] [CrossRef]
  32. Li, Y.; Wang, D.; Liu, F. The auto-correlation function aided sparse support matrix machine for EEG-based fatigue detection. IEEE Trans. Circuits Syst. II Express Briefs 2022, 70, 836–840. [Google Scholar] [CrossRef]
  33. Wang, F.; Wu, S.; Ping, J.; Xu, Z.; Chu, H. EEG driving fatigue detection with PDC-based brain functional network. IEEE Sens. J. 2021, 21, 10811–10823. [Google Scholar] [CrossRef]
  34. Zheng, W.; Lu, B. A multimodal approach to estimating vigilance using EEG and forehead EOG. J. Neural Eng. 2017, 14, 026017. [Google Scholar] [CrossRef]
  35. Tuncer, T.; Dogan, S.; Ertam, F.; Subasi, A. A dynamic center and multi threshold point based stable feature extraction network for driver fatigue detection utilizing EEG signals. Cogn. Neurodyn. 2021, 15, 223–237. [Google Scholar] [CrossRef]
  36. Chai, R.; Naik, G.R.; Nguyen, T.N.; Ling, S.H.; Tran, Y.; Craig, A.; Nguyen, H.T. Driver fatigue classification with independent component by entropy rate bound minimization analysis in an EEG-based system. IEEE J. Biomed. Health Inform. 2016, 21, 715–724. [Google Scholar] [CrossRef]
  37. Pratticò, D.; Laganà, F. Infrared Thermographic Signal Analysis of Bioactive Edible Oils Using CNNs for Quality Assessment. Signals 2025, 6, 38. [Google Scholar] [CrossRef]
  38. Al-Saegh, A.; Dawwd, S.A.; Abdul-Jabbar, J.M. Deep learning for motor imagery EEG-based classification: A review. Biomed. Signal Process. Control 2021, 63, 102172. [Google Scholar] [CrossRef]
  39. Li, R.; Wang, L.; Sourina, O. Subject matching for cross-subject EEG-based recognition of driver states related to situation awareness. Methods 2022, 202, 136–143. [Google Scholar] [CrossRef] [PubMed]
  40. Wang, K.; Qiu, S.; Wei, W.; Yi, W.; He, H.; Xu, M.; Jung, T.; Ming, D. Investigating EEG-based cross-session and cross-task vigilance estimation in BCI systems. J. Neural Eng. 2023, 20, 056001. [Google Scholar] [CrossRef]
  41. Barwick, F.; Arnett, P.; Slobounov, S. EEG correlates of fatigue during administration of a neuropsychological test battery. Clin. Neurophysiol. 2012, 123, 278–284. [Google Scholar] [CrossRef]
  42. Borghini, G.; Vecchiato, G.; Toppi, J.; Astolfi, L.; Maglione, A.; Isabella, R.; Caltagirone, C.; Kong, W.; Wei, D.; Zhou, Z.; et al. Assessment of mental fatigue during car driving by using high resolution EEG activity and neurophysiologic indices. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society; IEEE: New York, NY, USA, 2012; pp. 6442–6445. [Google Scholar]
  43. Craig, A.; Tran, Y.; Wijesuriya, N.; Middleton, J. Fatigue and tiredness in people with spinal cord injury. J. Psychosom. Res. 2012, 73, 205–210. [Google Scholar] [CrossRef]
  44. Lal, S.K.; Craig, A.; Boord, P.; Kirkup, L.; Nguyen, H. Development of an algorithm for an EEG-based driver fatigue countermeasure. J. Saf. Res. 2003, 34, 321–328. [Google Scholar] [CrossRef]
  45. Li, G.; Huang, S.; Xu, W.; Jiao, W.; Jiang, Y.; Gao, Z.; Zhang, J. The impact of mental fatigue on brain activity: A comparative study both in resting state and task state using EEG. BMC Neurosci. 2020, 21, 20. [Google Scholar] [CrossRef]
  46. Hayes, M.H. Statistical Digital Signal Processing and Modeling; John Wiley & Sons: Hoboken, NJ, USA, 1996. [Google Scholar]
  47. Kwak, S.G.; Kim, J.H. Central limit theorem: The cornerstone of modern statistics. Korean J. Anesthesiol. 2017, 70, 144. [Google Scholar] [CrossRef]
  48. Van der Vaart, A.W. Asymptotic Statistics; Cambridge University Press: Cambridge, MA, USA, 2000; Volume 3. [Google Scholar]
  49. Kirchner, W.K. Age differences in short-term retention of rapidly changing information. J. Exp. Psychol. 1958, 55, 352. [Google Scholar] [CrossRef]
  50. Tanaka, M.; Shigihara, Y.; Ishii, A.; Funakura, M.; Kanai, E.; Watanabe, Y. Effect of mental fatigue on the central nervous system: An electroencephalography study. Behav. Brain Funct. 2012, 8, 48. [Google Scholar] [CrossRef] [PubMed]
  51. Pergher, V.; Wittevrongel, B.; Tournoy, J.; Schoenmakers, B.; Van Hulle, M.M. Mental workload of young and older adults gauged with ERPs and spectral power during N-Back task performance. Biol. Psychol. 2019, 146, 107726. [Google Scholar] [CrossRef]
  52. Chen, K.; Liu, Z.; Liu, Q.; Ai, Q.; Ma, L. EEG-based mental fatigue detection using linear prediction cepstral coefficients and Riemann spatial covariance matrix. J. Neural Eng. 2022, 19, 066021. [Google Scholar] [CrossRef]
  53. Karim, E.; Pavel, H.R.; Jaiswal, A.; Zadeh, M.Z.; Theofanidis, M.; Wylie, G.; Makedon, F. An EEG-based cognitive fatigue detection system. In Proceedings of the 16th International Conference on Pervasive Technologies Related to Assistive Environments, Corfu, Greece, 5–7 July 2023; pp. 131–136. [Google Scholar]
  54. van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  55. Silverman, D.; Chen, C.; Chang, S.; Bui, L.; Zhang, Y.; Raghavan, R.; Jiang, A.; Le, A.; Darmohray, D.; Sima, J.; et al. Activation of locus coeruleus noradrenergic neurons rapidly drives homeostatic sleep pressure. Sci. Adv. 2025, 11, eadq0651. [Google Scholar] [CrossRef]
  56. Osorio-Forero, A.; Foustoukos, G.; Cardis, R.; Cherrad, N.; Devenoges, C.; Fernandez, L.M.; Lüthi, A. Infraslow noradrenergic locus coeruleus activity fluctuations are gatekeepers of the NREM–REM sleep cycle. Nat. Neurosci. 2025, 28, 84–96. [Google Scholar] [CrossRef]
  57. Doran, S.M.; Van Dongen, H.; Dinges, D.F. Sustained attention performance during sleep deprivation: Evidence of state instability. Arch. Ital. Biol. 2001, 139, 253–267. [Google Scholar] [PubMed]
  58. Dijk, D.J.; Czeisler, C.A. Contribution of the circadian pacemaker and the sleep homeostat to sleep propensity, sleep structure, electroencephalographic slow waves, and sleep spindle activity in humans. J. Neurosci. 1995, 15, 3526–3538. [Google Scholar] [CrossRef]
  59. Torsvall, L. Sleepiness on the job: Continuously measured EEG changes in train drivers. Electroencephal. Clin. Neurophysiol. 1987, 66, 502–511. [Google Scholar] [CrossRef] [PubMed]
  60. Van Dongen, H.P.; Maislin, G.; Mullington, J.M.; Dinges, D.F. The cumulative cost of additional wakefulness: Dose-response effects on neurobehavioral functions and sleep physiology from chronic sleep restriction and total sleep deprivation. Sleep 2003, 26, 117–126. [Google Scholar] [CrossRef]
  61. Boksem, M.A.; Meijman, T.F.; Lorist, M.M. Effects of mental fatigue on attention: An ERP study. Cogn. Brain Res. 2005, 25, 107–116. [Google Scholar] [CrossRef] [PubMed]
  62. Liu, Y.; Xiang, Z.; Yan, Z.; Jin, J.; Shu, L.; Zhang, L.; Xu, X. CEEMDAN fuzzy entropy based fatigue driving detection using single-channel EEG. Biomed. Signal Process. Control 2024, 95, 106460. [Google Scholar] [CrossRef]
  63. Ojo, J.; Omilude, L.; Adeyemo, I. Fatigue detection in drivers using eye-blink and yawning analysis. Int. J. Comput. Trends Technol. 2017, 50, 87–90. [Google Scholar] [CrossRef]
  64. Xie, Y.; Chen, K.; Murphey, Y.L. Real-time and robust driver yawning detection with deep neural networks. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI); IEEE: New York, NY, USA, 2018; pp. 532–538. [Google Scholar]
  65. Gastaldi, M.; Rossi, R.; Gecchele, G. Effects of driver task-related fatigue on driving performance. Procedia-Soc. Behav. Sci. 2014, 111, 955–964. [Google Scholar] [CrossRef]
  66. Bibbò, L.; Angiulli, G.; Laganà, F.; Pratticò, D.; Cotroneo, F.; La Foresta, F.; Versaci, M. MEMS and IoT in HAR: Effective Monitoring for the Health of Older People. Appl. Sci. 2025, 15, 4306. [Google Scholar] [CrossRef]
Figure 1. Scheme of the proposed ST-SODE fatigue estimation framework; the multichannel scalp EEG is segmented into aggregation windows, then band-pass filtered into different epoch-level band powers. The extracted DE features and the equivalent second-order differential entropy of the aggregation windows are used to describe the intensity of the fatigue rebound effect, which indicates the mental fatigue level.
Figure 1. Scheme of the proposed ST-SODE fatigue estimation framework; the multichannel scalp EEG is segmented into aggregation windows, then band-pass filtered into different epoch-level band powers. The extracted DE features and the equivalent second-order differential entropy of the aggregation windows are used to describe the intensity of the fatigue rebound effect, which indicates the mental fatigue level.
Brainsci 16 00199 g001
Figure 2. The N-Back experiment used to construct the Vigilance dataset. (a) Experimental setup: The experimental equipment included a NeuroScan SynAmps2 amplifier system, an electrode cap, and a computer with a monitor. (b) Experimental procedure: Two rounds of 1-Back tasks and two rounds of 2-Back tasks were conducted, with a 10-min rest period between 1-Back and 2-Back tasks. The 1-Back task was used to represent a vigilant working state, while the 2-Back task was used to induce the fatigued working state. (c) Explanation of N-Back experiments: Participants were told to press the spacebar when the current color block matched the previous one (1-Back) or the one before that (2-Back). The 2-Back task required sustained recall of historical color blocks within a short timeframe, resulting in significantly higher workload than 1-Back and rapidly inducing fatigue.
Figure 2. The N-Back experiment used to construct the Vigilance dataset. (a) Experimental setup: The experimental equipment included a NeuroScan SynAmps2 amplifier system, an electrode cap, and a computer with a monitor. (b) Experimental procedure: Two rounds of 1-Back tasks and two rounds of 2-Back tasks were conducted, with a 10-min rest period between 1-Back and 2-Back tasks. The 1-Back task was used to represent a vigilant working state, while the 2-Back task was used to induce the fatigued working state. (c) Explanation of N-Back experiments: Participants were told to press the spacebar when the current color block matched the previous one (1-Back) or the one before that (2-Back). The 2-Back task required sustained recall of historical color blocks within a short timeframe, resulting in significantly higher workload than 1-Back and rapidly inducing fatigue.
Brainsci 16 00199 g002
Figure 3. Position of the 28 electrodes during construction of the Vigilance dataset. Electrodes labeled in green were used to model fatigue. The reference electrode (default REF) was located between CZ and CPZ.
Figure 3. Position of the 28 electrodes during construction of the Vigilance dataset. Electrodes labeled in green were used to model fatigue. The reference electrode (default REF) was located between CZ and CPZ.
Brainsci 16 00199 g003
Figure 4. Visualization results of ST-SODE on the SEED-VIG dataset. The red line is the fatigue Level given by PERCLOS, with a high value indicating fatigue, while the blue line is the predicted result. Each figure is labeled with the correlation coefficient ρ . The sampling length is 8 s. Because the PERCLOS label is a discrete value derived by the eye tracker in another modality and the prediction is made without training or any use of label information, the two curves do not align precisely. Nonetheless, it is still possible to observe temporal synchronization by examining the rising and falling edges of the two curves.
Figure 4. Visualization results of ST-SODE on the SEED-VIG dataset. The red line is the fatigue Level given by PERCLOS, with a high value indicating fatigue, while the blue line is the predicted result. Each figure is labeled with the correlation coefficient ρ . The sampling length is 8 s. Because the PERCLOS label is a discrete value derived by the eye tracker in another modality and the prediction is made without training or any use of label information, the two curves do not align precisely. Nonetheless, it is still possible to observe temporal synchronization by examining the rising and falling edges of the two curves.
Brainsci 16 00199 g004
Figure 5. Comparison of DE, PSD-SVM, and ST-SODE on the Vigilance dataset. Three window settings (5 s, 10 s, and 20 s) are applied. PSD-SVM uses cross-subject setting (data from the seven subjects other than the test subject were used for training) in order to remain consistent with the untrained ST-SODE method. Here, ** indicates p < 0.01 on the Wilcoxon signed-rank test.
Figure 5. Comparison of DE, PSD-SVM, and ST-SODE on the Vigilance dataset. Three window settings (5 s, 10 s, and 20 s) are applied. PSD-SVM uses cross-subject setting (data from the seven subjects other than the test subject were used for training) in order to remain consistent with the untrained ST-SODE method. Here, ** indicates p < 0.01 on the Wilcoxon signed-rank test.
Brainsci 16 00199 g005
Figure 6. t-SNE visualization of DE, PSD, and ST-SODE features. All subject data from the Vigilance dataset are combined.
Figure 6. t-SNE visualization of DE, PSD, and ST-SODE features. All subject data from the Vigilance dataset are combined.
Brainsci 16 00199 g006
Figure 7. Channel-level ST-SODE brain topography. We calculated ST-SODE separately for each channel, with the aggregation window set to 5 s. The experiment did not include frontal electrodes. The color-bar range is fixed for each subject in terms of vigilance and fatigue; red indicates higher band-power fluctuations, signifying a more severe state of fatigue, while blue represents vigilance.
Figure 7. Channel-level ST-SODE brain topography. We calculated ST-SODE separately for each channel, with the aggregation window set to 5 s. The experiment did not include frontal electrodes. The color-bar range is fixed for each subject in terms of vigilance and fatigue; red indicates higher band-power fluctuations, signifying a more severe state of fatigue, while blue represents vigilance.
Brainsci 16 00199 g007
Table 1. Frequency bands of EEG signals.
Table 1. Frequency bands of EEG signals.
NumberFrequency BandFrequency Range/HzAmplitude/μV
1Delta ( δ )0.5–410–20
2Theta ( θ )4–820–40
3Alpha ( α )8–1210–100
4Beta ( β )13–305–30
5Gamma ( γ )30–100≈40
Table 2. Correlation between prediction and labels using different frequency-band settings without training.
Table 2. Correlation between prediction and labels using different frequency-band settings without training.
MethodDEPSDST-SODE
Sub θ / β α / β ( α + θ ) / β ( α + θ )
/ ( α + β )
θ / β α / β ( α + θ ) / β ( α + θ )
/ ( α + β )
θ / β α / β ( α + θ ) / β ( α + θ )
/ ( α + β )
Chance Level
1−0.21940.72050.8198−0.2191−0.2599−0.2295−0.2678−0.23860.38010.84160.83070.1819−0.0021
20.16850.13680.16580.16150.16590.18190.18730.0074−0.02620.32710.2836−0.02190.0119
30.12310.54530.46420.16530.0882−0.3300−0.23550.41460.24530.63310.57280.2079−0.0076
4−0.05140.53480.6043−0.0677−0.01810.15850.0737-0.14920.40860.74640.68810.26130.0134
50.23430.56850.64420.22990.20570.03570.08780.1525−0.14100.66570.6531−0.26950.0152
6−0.37410.27020.2678−0.3821−0.41770.0113−0.1646−0.3937−0.06340.71610.4527−0.20180.0098
70.2609−0.0211−0.00470.25850.24630.09960.22520.23380.26960.79840.29500.2004−0.0080
80.05960.11450.03280.09200.1413−0.1028−0.13360.22360.11600.28010.20500.1050−0.0040
90.29830.51250.36600.31030.23080.07940.20780.28260.1933−0.0178−0.04580.2003−0.0278
100.06560.73350.63050.03170.20220.41070.3902−0.13740.12430.68500.6650−0.0077−0.0026
11−0.27030.3677−0.0631−0.2700−0.2449−0.1866−0.2472−0.2368−0.05110.21280.0357−0.05750.0130
120.32990.11540.10800.34020.27050.07940.18130.24410.14560.34690.28430.08770.0082
130.26760.33050.30870.26200.21020.13100.18280.1468−0.09670.38080.3579−0.08160.0206
14−0.18110.62560.4644−0.1820−0.2874−0.1062−0.2041−0.2825−0.00660.70650.5697−0.1356−0.0098
150.28840.36070.46780.28690.22780.22580.23150.2144−0.03200.73590.5811−0.2392−0.0040
160.06650.27090.37360.06560.03790.04450.04210.0609−0.03900.55150.4489−0.1012−0.0042
170.20160.24590.39070.19720.15470.10620.13760.05650.24440.72300.54740.20020.0062
18−0.04000.48860.6983−0.0539−0.03770.05800.0028−0.12550.54710.80210.75640.43210.0014
190.04290.80650.66130.05010.0281−0.0608−0.03330.0737−0.08690.43730.4143−0.15000.0086
20−0.01920.64100.6480−0.0147−0.0272−0.1442−0.0373−0.00390.32860.71350.67860.17140.0283
21−0.38510.31080.1864-0.3883−0.2926−0.2538−0.2876−0.2957−0.00420.63130.4510−0.08310.0186
22−0.0359−0.0943−0.0130−0.0279−0.0336−0.2058−0.08410.04110.12280.49150.19170.04720.0038
230.26010.61800.42110.25600.15770.00990.08650.13210.15770.47140.44460.05870.0251
Mean0.04740.40010.37580.04790.03250.00050.01480.01830.11900.5600 **†††0.45050.03190.0050
Significance: Wilcoxon signed-rank test across subjects. ** denotes ST-SODE vs. DE with p < 0.01 . ††† denotes ST-SODE vs. PSD with p < 0.001 . The bold number indicates the highest correlation.
Table 3. Accuracy comparison of DE, PSD-SVM and ST-SODE on the Vigilance dataset.
Table 3. Accuracy comparison of DE, PSD-SVM and ST-SODE on the Vigilance dataset.
SubjectDEPSD-SVMST-SODE
5 s 10 s 20 s 5 s 10 s 20 s 5 s 10 s 20 s
144.1733.3353.3371.6770.0080.0082.5088.3390.00
260.0065.0073.3382.5080.0086.6773.3370.8366.67
385.0093.3393.3384.1785.0083.3377.5081.6783.33
458.3370.0073.3360.00100.0086.6795.83100.00100.00
577.5088.3380.0059.1790.0083.3381.6781.6783.33
667.5075.0073.3365.0081.6766.6789.1793.3393.33
750.0050.0050.0059.1776.6793.3395.8396.67100.00
888.3386.6770.0088.3381.6776.6778.3376.6780.00
Mean (%)66.3570.2170.8371.2583.1382.0887.29 **92.29 **††93.75 **††
Significance: Wilcoxon signed-rank test across subjects. ** denotes ST-SODE vs. DE with p < 0.01; /†† denote ST-SODE vs. PSD-SVM with p < 0.05/0.01. The bold numbers indicate the highest accuracy for each setting.
Table 4. Accuracy comparison of DE-SVM (within-subject), PSD-SVM (within-subject), and ST-SODE (cross-subject) on the Vigilance dataset.
Table 4. Accuracy comparison of DE-SVM (within-subject), PSD-SVM (within-subject), and ST-SODE (cross-subject) on the Vigilance dataset.
SubjectDE-SVM (Within-Subject)PSD-SVM (Within-Subject)ST-SODE
5 s 10 s 20 s 5 s 10 s 20 s 5 s 10 s 20 s
198.3398.3396.6795.0098.3396.6782.5088.3390.00
299.17100.00100.0086.6793.3394.1773.3370.8366.67
3100.0098.3393.3390.0093.3388.3377.5081.6783.33
498.3395.00100.0095.0095.00100.0095.83100.00100.00
599.1796.67100.0088.3394.1795.0081.6781.6783.33
698.33100.0096.6796.67100.0096.6789.1793.3393.33
795.0098.3396.6795.0098.3396.6795.8396.67100.00
899.1796.6796.6789.1790.0091.6778.3376.6780.00
Mean (%)98.4497.9297.5092.0895.8395.8387.2992.2993.75
Table 5. Average results of ST-SODE for all trials on the vigilance task and fatigue task using different aggregation window lengths (5 s, 10 s, 20 s).
Table 5. Average results of ST-SODE for all trials on the vigilance task and fatigue task using different aggregation window lengths (5 s, 10 s, 20 s).
SubjectVigilance
(1-Back)
Fatigue
(2-Back)
VigilanceFatigueVigilanceFatigue
10.0000790.0008930.0000830.0008680.0000820.000993
20.0000910.0021790.0000930.0021970.0000920.002282
30.0002540.1217830.0002510.1407960.0002430.165035
40.0000310.1175060.0000320.1175100.0000320.115746
50.0002570.2198030.0002570.2328600.0002610.328026
60.0001680.0224310.0001590.0227370.0001690.023496
70.0000660.0821600.0000700.0816500.0000710.088062
80.0000390.0008620.0000370.0009260.0000380.000937
Mean0.0000770.0643610.0000750.0682610.0000760.084273
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, A.; Wang, Z.; Xu, T.; Zhou, T.; Zhao, X.; Hu, H.; Van Hulle, M.M. A Cross-Subject Band-Power Complexity Metric for Detecting Mental Fatigue Through EEG. Brain Sci. 2026, 16, 199. https://doi.org/10.3390/brainsci16020199

AMA Style

Li A, Wang Z, Xu T, Zhou T, Zhao X, Hu H, Van Hulle MM. A Cross-Subject Band-Power Complexity Metric for Detecting Mental Fatigue Through EEG. Brain Sciences. 2026; 16(2):199. https://doi.org/10.3390/brainsci16020199

Chicago/Turabian Style

Li, Ang, Zhenyu Wang, Tianheng Xu, Ting Zhou, Xi Zhao, Honglin Hu, and Marc M. Van Hulle. 2026. "A Cross-Subject Band-Power Complexity Metric for Detecting Mental Fatigue Through EEG" Brain Sciences 16, no. 2: 199. https://doi.org/10.3390/brainsci16020199

APA Style

Li, A., Wang, Z., Xu, T., Zhou, T., Zhao, X., Hu, H., & Van Hulle, M. M. (2026). A Cross-Subject Band-Power Complexity Metric for Detecting Mental Fatigue Through EEG. Brain Sciences, 16(2), 199. https://doi.org/10.3390/brainsci16020199

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop