Next Article in Journal
Approximating the Performance of a Time-Domain Pulsed Induction EMI Sensor with Multiple Frequency-Domain FEM Simulations for Improved Modelling of Arctic Sea-Ice Thickness
Previous Article in Journal
Robust Line-Scan Image Registration via Disparity Estimation for Train Fault Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on the Prediction of Driver Fatigue Degree Based on EEG Signals

School of Electrical Engineering, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(23), 7316; https://doi.org/10.3390/s25237316
Submission received: 20 October 2025 / Revised: 24 November 2025 / Accepted: 26 November 2025 / Published: 1 December 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

Objective: Predicting driver fatigue degree is crucial for traffic safety. This study proposes a deep learning model utilizing electroencephalography (EEG) signals and multi-step temporal data to predict the next time-step fatigue degree indicator percentage of eyelid closure (PERCLOS) while exploring the impact of different EEG features on prediction performance. Approach: A CTL-ResFNet model integrating CNN, Transformer Encoder, LSTM, and residual connections is proposed. Its effectiveness is validated through two experimental paradigms, Leave-One-Out Cross-Validation (LOOCV) and pretraining–finetuning, with comparisons against baseline models. Additionally, the performance of four EEG features—differential entropy, α / β band power ratio, wavelet entropy, and Hurst exponent—is evaluated, using RMSE and MAE as metrics. Main Results: The combined input of EEG and PERCLOS significantly outperforms using PERCLOS alone validated by LSTM, and CTL-ResFNet surpasses baseline models under both experimental paradigms. In LOOCV experiments, the α / β band power ratio performs best, whereas differential entropy excels in pretraining–finetuning. Significance: This study presents a high-performance hybrid deep learning framework for predicting driver fatigue degree and reveals the applicability differences in EEG features across experimental paradigms, offering guidance for feature selection and model deployment in practical applications.

1. Introduction

Fatigued driving is a condition in which a driver’s driving skills deteriorate after driving a vehicle for an extended period of time [1]. Driver fatigue can be caused by poor sleep quality, prolonged driving of vehicles, single road conditions, and taking medications prohibited for driving vehicles. Continuing to drive while fatigued is highly dangerous and may lead to traffic accidents. Timely detection of the driver’s drowsiness can help prevent accidents caused by fatigue. Fatigue-related traffic accidents often result in catastrophic outcomes, including multiple fatalities and injuries [2]. An analysis of rail accidents in the UK shows that about 21% of high-risk rail accidents are caused by driver fatigue [3]. The above figures show that fatigue driving is not only common but also has serious consequences, thus demonstrating the importance of preventing fatigue driving.
Currently, research on fatigue driving detection primarily focuses on binary classification (fatigued or alert) or ternary classification (alert, fatigued, and drowsy) of the driver’s fatigue state [4,5,6,7,8,9]. However, these approaches have notable limitations. First, classification models can only passively identify fatigue states that have already occurred, lacking the capability to dynamically predict the evolution of fatigue. Second, traditional methods often rely on single-modal signals, such as eye movements or facial features, making them inadequate to address individual differences and environmental interference in complex driving scenarios. To address these issues, this paper proposes a multimodal temporal prediction framework, CTL-ResFNet, which integrates multidimensional dynamic features such as eye state and EEG signals. This framework achieves a paradigm shift from state recognition to fatigue trend prediction, providing a novel technical approach for proactive fatigue intervention.
The main contributions of this paper include:
(1)
The CTL-ResFNet hybrid neural network is proposed, which for the first time organically integrates the local feature extraction capability of CNN, the global dependency modeling of Transformer, and the temporal dynamic capturing of LSTM through residual connections, addressing the representational limitations of traditional single models in cross-modal EEG-fatigue temporal prediction. Experiments demonstrate that this architecture significantly outperforms baseline models in both leave-one-out cross-validation (LOOCV) and transfer learning scenarios, providing a new paradigm for physiological signal temporal prediction.
(2)
Through LOOCV, it was discovered that the spatiotemporal coupling of EEG differential entropy features and PERCLOS can improve prediction accuracy, revealing the complementary enhancement effect of physiological signals on subjective fatigue labels. Further research revealed that in LOOCV, the predictive performance of the α / β band energy ratio significantly outperforms comparative features such as differential entropy and wavelet entropy, demonstrating its superior zero-shot cross-subject generalization ability and stronger robustness to individual differences.
(3)
To validate the small-sample individual adaptation capability of CTL-ResFNet, this study established a pretraining–finetuning experimental framework. The results demonstrate that differential entropy features exhibit optimal performance in transfer learning scenarios. This finding provides methodological guidance for optimizing feature selection in practical fatigue monitoring system applications.
This study focuses on the prediction of driver fatigue degree based on EEG signals. The paper is divided into five sections: Section 1 elaborates on the hazards of fatigued driving and the significance of the research, presenting the research questions. Section 2 systematically reviews existing fatigue detection methods and summarizes related work. Section 3 introduces the proposed CTL-ResFNet hybrid neural network architecture, which integrates the advantages of CNN, Transformer, and LSTM while incorporating residual connections. Section 4 presents the experimental validation, including a detailed description of the SEED-VIG dataset, experimental setup, evaluation metrics, and comparative experiments. Section 5 analyzes the results and discusses the effectiveness of different EEG features in predicting fatigue degree.

2. Related Work

In the field of driver fatigue detection, numerous scholars have conducted extensive research based on EEG signals. Early studies primarily focused on the correlation between EEG rhythmic waves and fatigue states. For instance, researchers observed that as fatigue levels in-creased, the power of α and θ waves significantly rose [10,11], while the power of β waves exhibited a declining trend [12]. Subsequently, Jap et al. reported that the ( θ + α ) / β ratio showed the most pronounced increase at the end of driving tasks [13], which aligns with the conclusions of earlier studies. These findings laid a crucial foundation for subsequent research, prompting scholars to systematically explore the quantitative relationship between EEG activities in different frequency bands and fatigue levels.
With the advancement of machine learning techniques, researchers gradually shifted from traditional spectral analysis to more sophisticated feature extraction and pattern recognition methods. For example, Khushaba et al. employed wavelet packet transform based on fuzzy mutual information for feature extraction, achieving an average classification accuracy of 95–97% for driver fatigue states [14]. Zhao et al. utilized a multivariate autoregressive model to extract features from multi-channel EEG signals, combined with kernel principal component analysis and support vector machines, attaining an identification accuracy of 81.64% for three driving-related mental fatigue states [15]. In addition, some scholars have explored fatigue detection by integrating multimodal physiological signals. For instance, Huo et al. fused EEG and frontal electrooculogram (EOG) signals and applied a discriminant graph regularized extreme learning machine to detect driver fatigue levels. Their results demonstrated that the fused modality outperformed single-modality signals in fatigue detection performance [16].
In recent years, the rise of deep learning methods has brought new breakthroughs in fatigue detection. For example, Chaabene et al. achieved a high accuracy of 90.42% in fatigue state classification based on EEG signals and a convolutional neural network architecture [17]. Cui et al. utilized separable convolutions to process spatiotemporal sequences of EEG signals, reporting an average accuracy of 78.35% in LOOCV across 11 participants [18]. In terms of feature engineering, some researchers have begun exploring the application of nonlinear dynamic features. Shi et al. proposed a novel EEG feature called differential entropy in their study [19], with experimental results demonstrating that differential entropy (DE) is the most accurate and stable EEG feature for reflecting vigilance changes. Notably, most existing fatigue detection methods focus on identifying fatigue after it has already occurred, lacking the capability for dynamic prediction of fatigue degree, which provides a critical research opportunity for this study.

3. Methods

3.1. Overall Architecture

The proposed EEG-based fatigue degree prediction model consists of two core components: a feature extraction backbone network and a prediction regression head. The backbone network integrates CNN, Transformer encoder, LSTM, and residual networks, while the regression head comprises a fully connected layer with an output dimension of 1. The overall architecture is illustrated in Figure 1.
The model employs a CNN for dimensionality reduction in input data, utilizes the Transformer encoder architecture to capture diverse features and relationships within sequences, and subsequently feeds the extracted feature sequences into an LSTM network to extract temporal dependencies. Furthermore, residual connections are incorporated to enhance information flow and mitigate gradient vanishing issues. Finally, a fully connected layer performs regression prediction of fatigue degree. The model parameters are shown in the following table (Table 1).
To provide a theoretical rationale, the CTL-ResFNet is designed according to a “Spatial-Global-Temporal” processing hierarchy tailored to EEG characteristics. First, the CNN employs 1 × 1 convolutions as learnable spatial filters. Unlike temporal convolutions, this design deliberately fuses cross-channel information to simulate neural integration across cortical regions while strictly preserving the temporal resolution of the sequence. Second, the Transformer Encoder models global functional connectivity, capturing long-range dependencies between distant electrodes that local operations miss. A single-layer configuration was adopted to prevent overfitting, balancing model capacity with the limited sample size of the EEG dataset. Third, the LSTM layer captures the cumulative temporal evolution of fatigue, treating drowsiness as a dynamic process accumulated over time. Finally, residual connections ensure training stability against the high variability inherent in physiological signals.

3.2. Position Encoding and Activation Function

Since the Transformer encoder does not inherently capture positional information, sinusoidal positional encoding is employed:
P E ( p o s , 2 i ) = sin p o s 10,000 2 i / d m o d e l
P E ( p o s , 2 i + 1 ) = cos p o s 10,000 2 i / d m o d e l
where p o s denotes the input position, d m o d e l is the feature dimension, and i the feature index.
The model uses two nonlinear activation functions: ReLU and Sigmoid. ReLU is defined as
f ( x ) = max ( 0 , x )
which maintains a gradient of 1 for positive inputs, mitigating gradient vanishing in deep networks.
The Sigmoid function,
f ( x ) = 1 1 + e x
outputs values in ( 0 , 1 ) , making it suitable for probability prediction.

3.3. CNN Module

CNN is adopted due to its strong capability for extracting local spatial correlations in multichannel EEG data, which helps to capture frequency-specific spatial patterns related to fatigue. Compared with traditional feature engineering or simple MLP, CNN can automatically learn discriminative spatial filters from EEG inputs [20]. Convolutional Neural Networks were first proposed by Yann LeCun et al. [21] in 1998, primarily for image processing. CNNs have since been extended to handle one-dimensional data, such as time series. In this study, we employ a one-dimensional CNN (1D CNN) architecture, where convolution kernels slide along the input sequence to compute dot products of local regions.
Figure 2 illustrates the 1 × 1 convolutional kernel used in this study. It operates on individual elements of the input data, producing output with the same spatial dimensions but cannot capture relationships between adjacent elements. This operation is applied to multi-channel input for dimensionality reduction: as shown in Figure 3, a 1D convolution with size 1 transforms a 3-channel input into a 2-channel output without altering sequence length.

3.4. Transformer Encoder Module

In 2017, Vaswani et al. [22] introduced the Transformer architecture, consisting of an encoder and a decoder. In this study, only the Transformer encoder is employed, as shown in Figure 4.
The Transformer encoder is composed of n stacked layers ( n = 1 in this study), each containing two sub-layers: a multi-head self-attention mechanism followed by layer normalization and residual connection, and a position-wise feed-forward network with similar normalization and residual structure.
The scaled dot-product attention is defined as:
A t t e n t i o n ( Q , K , V ) = s o f t m a x Q K T d k V
where Q, K, and V are query, key, and value vectors, and d k is the key dimension.
Multi-head attention enables capturing diverse feature representations:
Q i = X W i Q , K i = X W i K , V i = X W i V
h e a d i = Attention ( Q i , K i , V i )
MultiHead ( Q , K , V ) = Concat ( h e a d 1 , , h e a d h ) W O
where W i Q , W i K , W i V , W O are learnable weights.
The feed-forward network enhances model capacity:
f ( x ) = max ( 0 , x W 1 + b 1 ) W 2 + b 2
Residual connections [23] mitigate vanishing/exploding gradients:
H ( x ) = F ( x ) + x
Layer normalization stabilizes training by reducing internal covariate shift, normalizing inputs across channels.

3.5. LSTM Module

Long Short-Term Memory (LSTM), proposed by Hochreiter and Schmidhuber in 1997 [24], is a variant of traditional RNNs with superior ability to capture long-range dependencies in sequential data. LSTM architectures have been successfully applied to fatigue detection tasks [25].
An LSTM unit relies on four key components: the forget gate, input gate, cell state, and output gate. Figure 5 illustrates the internal structure of a single LSTM unit. At timestep t, the unit receives the input x t and the previous hidden state h t 1 , producing an updated cell state C t and hidden state h t , which are then propagated to the next timestep.
The forget gate decides which historical information to discard:
f t = σ W f · h t 1 , x t + b f
The input gate determines how much new information to store:
i t = σ W i · h t 1 , x t + b i
C ˜ t = tanh W C · h t 1 , x t + b C
The cell state is updated as:
C t = f t C t 1 + i t C ˜ t
The output gate controls the hidden state output:
o t = σ W o · h t 1 , x t + b o
h t = o t tanh C t
This simplified structure illustrates how LSTM selectively retains or discards information, enabling effective modeling of long-range dependencies in sequential data.

3.6. Regression Prediction Head

The hidden state output from the final timestep of the LSTM layer in the feature extraction backbone network is connected to the regression prediction head. The regression head consists of a fully connected (FC) layer with an input dimension of 32 and an output dimension of 1. Since the output values are constrained within the range (0,1), a sigmoid activation function is employed.

4. Experiments

4.1. SEED-VIG Dataset

EEG signals represent the scalp-recorded electrical potentials generated by synchronous discharges of neuronal populations, primarily originating from the synchronized synaptic activities of pyramidal cells in the cerebral cortex [26]. Specifically, the human brain contains tens of thousands of interconnected neurons. These neurons receive signals from other neurons and generate EEG signals when the received signals exceed a certain threshold. Essentially, brain waves represent the electrical signals produced by the collective activity of these neurons. Single-channel EEG signals provide limited information with poor determinacy, leading to random research outcomes, whereas multi-channel EEG signals can capture more comprehensive information and better reflect global brain activities. EEG signals possess inherent non-replicability as physiological characteristics. Relevant studies have demonstrated significant differences in EEG patterns between fatigue and non-fatigue states [27]. Among various physiological indicators for fatigue assessment, EEG signals are recognized as one of the most reliable biomarkers [28]. Researchers have made notable breakthroughs in monitoring fatigue levels by leveraging EEG signal analysis [29,30,31].
The SEED-VIG database [32], a benchmark for driver fatigue studies, was adopted in this work. Developed by Shanghai Jiao Tong University’s Brain-Inspired Computing team, it includes data from 23 subjects (mean age: 23.3 years). Experiments were conducted in a virtual driving simulator with a projected road scene. The 118-min experimental sessions were primarily scheduled in the afternoon to facilitate fatigue induction. Participants wore EEG caps for electrophysiological signal acquisition and SMI gaze-tracking apparatus to measure eye movements, with PERCLOS values calculated for fatigue-level annotation. EEG recordings in the SEED-VIG dataset were obtained at 1000 Hz sampling rate from 17 standard channels referenced to CPZ, covering frontal to occipital areas (FT7, FT8, T7, T8, TP7, TP8, CP1, CP2, P1, PZ, P2, PO3, POZ, PO4, O1, OZ, O2). Figure 6 illustrates the schematic configuration of these 17 electrode positions.
The amplitude of EEG signals is at the microvolt level, making them highly susceptible to interference from various noise sources [33]. Consequently, the acquired data consists of both EEG signals and diverse noise components, with all non-EEG signals collectively referred to as artifacts [34]. The EEG signals were initially processed using a 1–75 Hz bandpass filter for noise suppression, with the sampling rate subsequently de-creased to 200 Hz to optimize computational efficiency. During the feature extraction stage, power spectral density (PSD) and DE features were extracted from EEG sig-nals across five frequency bands ( δ : 1–4 Hz, θ : 4–8 Hz, α : 8–14 Hz, β : 14–31 Hz, and γ : 31–50 Hz). These features were computed using short-time Fourier transform (STFT) with non-overlapping 8-s windows. Additionally, features were extracted across the entire 1–50 Hz frequency band with a 2 Hz resolution. PSD reflects the energy distribution of EEG signals across different frequency bands, which has been demonstrated to be highly correlated with fatigue and drowsiness [35,36,37]. DE, characterizing the complexity and uncertainty of EEG signals, can be regarded as a complexity feature. Studies have shown that various entropy-based and complexity measurement methods can effectively identify fatigue states [38]. Notably, empirical evidence suggests that differential entropy features outperform power spectral density in fatigue detection [16].
One of the most significant manifestations of fatigue is reflected in the eyes [39], where the PERCLOS demonstrates a robust association with fatigue levels. PERCLOS is measured as the duration during which the eyelids cover the pupils within a given period, providing an objective metric for fatigue assessment with its computational formula given in Equation (17).
PERCLOS = T closure T total × 100 %
where T closure is the eyelid closure time within the given time window, and T total is the total observation time window. In the experiment, PERCLOS was computed over non-overlapping 8-s windows ( T total = 8 s), generating 885 time steps in total. The PERCLOS metric is bounded by 0 and 1, where elevated scores correspond to increased fatigue degree. Based on reference [32], the fatigue states were classified as: alert [0, 0.35], fatigued [0.35, 0.7], and drowsy [0.7, 1]. Although PERCLOS is often discretized into two or three levels in prior studies [40,41,42], this study treats it as a continuous variable to preserve subtle temporal dynamics and enable fine-grained prediction. The regression approach facilitates proactive fatigue degree forecasting rather than post hoc state recognition.
In this study, we further analyzed the data distribution of PERCLOS values in the SEED-VIG dataset. The samples are distributed as follows: 36.32% in the alert range (0–0.35), 43.76% in the fatigued range (0.35–0.7), and 19.91% in the drowsy range (0.7–1.0). This distribution reveals a mild imbalance across fatigue levels, reflecting the natural evolution of fatigue during prolonged simulated driving. To mitigate potential bias caused by this imbalance, we conducted LOOCV experiments at the subject level, ensuring subject independence and preventing data leakage. Moreover, regression-based metrics (RMSE and MAE) were adopted, which are less sensitive to uneven class proportions compared with categorical accuracy, thereby maintaining fair evaluation across different fatigue levels.

4.2. Evaluation Metrics

Model performance was evaluated using both RMSE and MAE metrics, calculated as follows:
RMSE = 1 m i = 1 m y i y ^ i 2
MAE = 1 m i = 1 m | y i y ^ i |
In the above, y i is the true value, y ^ i is the predicted value, and m is the number of samples.
The RMSE measures the square root of the average squared differences between predicted and actual values. It penalizes larger errors more heavily, making it particularly sensitive to outliers. Therefore, RMSE reflects the overall deviation magnitude of the model predictions.
The MAE represents the average of the absolute differences between predicted and actual values. It provides a more intuitive interpretation of prediction accuracy, as it directly reflects the average prediction deviation without disproportionately amplifying large errors.
In summary, MAE indicates the average level of prediction error, while RMSE emphasizes the influence of large errors. Using both metrics together provides a comprehensive evaluation of the prediction performance.

4.3. Implementation Details

4.3.1. LOOCV Experiment

The LOOCV strategy was adopted to evaluate the model’s generalization performance across subjects. The core concept is that each time one subject is selected as the test set, while all remaining subjects form the training set. This process is repeated until every subject has served as the test set once, and finally, all test results are aggregated for comprehensive evaluation. Subject order is shuffled across participants during training, with within-subject temporal continuity strictly maintained. The data partitioning scheme is illustrated in Figure 7.

4.3.2. Cross-Subject Pre-Training with Within-Subject Fine-Tuning Experiment

The experiment adopts a two-stage training strategy, consisting of a pre-training stage and a fine-tuning stage. This study falls within the scope of transfer learning, which has demonstrated exceptional cross-task and cross-domain adaptation capabilities across multiple fields including computer vision, natural language processing, and biomedical signal analysis, with its effectiveness being extensively validated [43,44,45,46,47,48].
During the pre-training stage, the training sets of all 23 samples are combined into a unified pre-training dataset, and the order of the training sets is randomly shuffled each epoch while preserving the internal temporal sequence of each sample. Early stopping is monitored simultaneously using the validation sets of all 23 samples during pre-training. In the fine-tuning stage, the pre-trained weights obtained from the pre-training stage are used to fine-tune the parameters for individual samples, with each sample’s training set utilized for fine-tuning. Early stopping is monitored using the respective sample’s validation set, and the final evaluation is performed on the sample’s test set. This cross-subject to within-subject hybrid design enables simultaneous learning of population-level common features and adaptation to individual variations.
This study constructs an EEG dataset by concatenating DE and eyelid closure degree at the feature level. Let the differential entropy matrix be X D E R 885 × 85 , the eyelid closure vector be y R 885 × 1 , and the concatenated dataset be D = X D E , y R 885 × 86 . Subsequently, a sliding window is constructed for the dataset D R 885 × 86 , using the data from the previous three time steps to predict the PERCLOS value at the next time step. Time step 3 is the optimal solution obtained through comparative experiments (Section 4.5.1). The 23 time-series data samples are split into training, validation, and test sets in an 8:1:1 ratio, without shuffling the temporal order. The training set is used for model training, the validation set for monitoring the training process in combination with an early-stopping strategy, and the test set for the final evaluation of model performance. For each i in the training, validation, and test sets:
X i = D [ i ] D [ i + 1 ] D [ i + 2 ] R 3 × 86 , y i = D [ i + 3 , 86 ] R
To eliminate dimensional and scale differences while accelerating model convergence, this study processes the data using Z-score standardization, scaling the data to follow a standard normal distribution. The calculation formula is as follows:
z = x μ σ
μ corresponds to the mean, and σ to the standard deviation.

4.3.3. Additional Feature Extraction

To further explore the discriminative patterns in EEG signals, we extracted three additional features from the raw EEG data in the SEED-VIG dataset: the α / β band power ratio, wavelet entropy, and Hurst exponent. Prior to feature extraction, the raw EEG data underwent preprocessing involving 1–75 Hz bandpass filtering and artifact removal through Independent Component Analysis (ICA). These procedures were implemented using the EEGLAB toolbox, a MATLAB (version R2021a) toolkit specifically designed for EEG signal processing.
Filtering effectively isolates fatigue-related frequency bands and suppresses signal interference. As shown in Figure 8, the filtered power spectral density exhibits significant attenuation beyond 75 Hz.
ICA is an effective method for blind source separation [49], which enables the isolation of interference signals unrelated to fatigue, such as EOG and electromyographic (EMG) signals. As illus-trated in Figure 9, the ICA-derived results comprise 17 components, among which some represent genuine EEG signals while others constitute artifacts. For subsequent analysis, it is essential to remove these artifactual components while retaining the purified EEG signals.
A representative example of artifacts is shown in Figure 10, where EMG signals are illustrated-with the left panel displaying the topographic brain map and the right panel presenting the time-domain representation of EMG artifacts.
A comparison between the raw EEG signals and preprocessed EEG signals is shown in Figure 11, where the blue waveforms represent the original EEG data and the red waveforms denote the preprocessed EEG signals.
After preprocessing the raw EEG signals to obtain clean electrophysiological data, we subsequently extracted three key EEG features for analysis: the α / β band power ratio, wavelet entropy, and Hurst exponent. Below we provide a brief introduction to each of these features.
Band power ratio, a commonly used quantitative metric in EEG signal analysis, reflects the activity intensity of different frequency bands. Typically, EEG signals are divided into five frequency bands. The α band is associated with relaxation and eye-closed states, while the β band correlates with concentration and cognitive activities. A high α / β ratio is commonly observed during relaxed or meditative states. An increased α / β ratio indicates mental fatigue. This ratio is calculated as the power ratio between the α and β bands. In this study, the α / β ratio was computed using STFT with non-overlapping 8-s windows. The formula for calculating the α / β band power ratio is as follows:
α / β R a t i o = P α P β
P α = f = 8 Hz 14 Hz PSD ( f )
P β = f = 14 Hz 31 Hz PSD ( f )
Wavelet Entropy [50] is a signal analysis method that combines wavelet transform and information entropy to quantify signal complexity. The wavelet entropy value decreases when the EEG signal exhibits regular and ordered patterns, whereas it increases when the signal becomes complex and disordered. To compute wavelet entropy, the signal is first decomposed using wavelet transform to obtain wavelet coefficients W j ( t ) at different scales. The energy E j at each scale j is then calculated, followed by the determination of relative energy p j . Finally, wavelet entropy is derived using Shannon’s entropy formula as follows:
E j = t W j t 2
p j = E j k E k
W E = j p j · log p j
This study employs non-overlapping 8-s windows for wavelet transformation using db4 (Daubechies 4) as the mother wavelet with a 5-level decomposition. The δ , θ , α , β , and γ frequency bands were extracted, and wavelet entropy was subsequently calculated based on these five frequency bands.
The Hurst exponent was originally proposed by hydrologist Harold Hurst [51] and has since been widely applied across various disciplines. This metric quantifies the long-range dependence or self-similarity of EEG signals, thereby reflecting the complexity of brain activity. During fatigue states, the complexity of neural activity decreases while the regularity of EEG signals increases—changes that can be effectively captured by the Hurst exponent. In this study, we performed 5-level decomposition using db4 wavelets for each 8-second non-overlapping window per channel. The reconstructed signals from each frequency band were subsequently analyzed using the rescaled range (R/S) analysis to compute their respective Hurst exponents. The calculation formula for the Hurst exponent is as follows:
H = log R / S log T
The formula parameters are defined as follows: R represents the rescaled range, S denotes the standard deviation of the time series, T indicates the time span (number of data points) of the series, and H is the Hurst exponent whose value typically ranges between 0 H 1 .

4.4. Training Settings

4.4.1. LOOCV Experiment

The experiment was conducted using the PyTorch (1.11.0) framework on an NVIDIA RTX 3050 GPU. During the training process, the MSE loss function was used, with the Adam [52] optimizer, an initial learning rate of 0.001, a batch size of 64, and a fixed number of 20 training epochs.
To ensure reproducibility, all experiments were conducted with fixed random seeds (refer to the hyperparameter settings in Table A1). All model weights were initialized using PyTorch’s default schemes (Kaiming uniform for convolutional layers and Xavier uniform for linear layers). The Z-score normalization described in Equation (21) was applied, where the mean and standard deviation were computed from the training set only to avoid data leakage.

4.4.2. Cross-Subject Pre-Training with Within-Subject Fine-Tuning Experiment

In the pre-training phase, the Adam optimizer was used with an initial learning rate of 0.001, and the learning rate was dynamically adjusted—halved if the validation loss did not decrease for 5 epochs. The total number of training epochs was 150, and an early stopping mechanism was employed, terminating the training if the validation loss showed no improvement for 10 epochs. The batch size was set to 64. Using the MSE loss function. In the fine-tuning phase, the model was trained for 50 epochs with a learning rate of 0.0001, while all other training configurations remained consistent with those used in the pre-training phase.

4.5. Main Results

4.5.1. Window Length Experiment

redAs shown in Table 2, from the average results (Avg), a window length of 3 performed the best (RMSE = 0.0598, MAE = 0.0509), with the lowest error. This indicates that for fatigue detection based on EEG (DE features) + PERCLOS, selecting an appropriate time window can significantly improve the model’s prediction accuracy.

4.5.2. Comparison of Univariate and Multimodal Fatigue Prediction Based on LSTM

This study employs an LSTM model for fatigue degree prediction and designs two comparative experiments based on a LOOCV paradigm: the first group uses univariate modeling based solely on PERCLOS (PERCLOS-only), while the second group employs multimodal fusion modeling combining EEG signals (DE features) and PERCLOS (EEG + PERCLOS). As shown in Table 3, the prediction performance of the multimodal fusion model (Avg RMSE = 0.0598) is significantly better than that of the univariate model (Avg RMSE = 0.0683), with a relative error reduction of 12.4%. These results confirm the effectiveness of multimodal physiological signal fusion in fatigue degree prediction, demonstrating that EEG features and PERCLOS indicators exhibit complementarity, and their joint modeling can improve prediction accuracy.

4.5.3. Ablation Study

To rigorously evaluate the contribution of each architectural component in CTL-ResFNet, we conducted a systematic ablation study based on a module-removal paradigm. The experiments were performed using Leave-One-Out Cross-Validation (LOOCV) on the dataset, utilizing Differential Entropy (DE) as the input feature. We compared the complete CTL-ResFNet against four incomplete variants:
  • -CNN: Removing the Convolutional Neural Network module to test the importance of local feature extraction.
  • -Transformer: Removing the Transformer module to evaluate the impact of the self-attention mechanism.
  • -LSTM: Removing the Long Short-Term Memory network to assess the necessity of temporal dynamic modeling.
  • -Residual: Removing the residual connections to verify their role in feature fusion and training stability.
The quantitative results of these experiments are summarized in Table 4.
As evidenced by the average metrics in Table 4, the complete CTL-ResFNet architecture achieves the best performance with the lowest RMSE (0.0266) and MAE (0.0215). The removal of any single component leads to varying degrees of performance degradation, confirming that all designed modules contribute positively to the model’s predictive capability.
Impact of Residual Connections. The most significant performance drop occurs in the -Residual variant, where the average RMSE surges to 0.0692 and MAE to 0.0549. This stark contrast suggests that residual connections are critical for the network. They likely facilitate efficient gradient flow and prevent information loss during the integration of features from different layers, ensuring the model can effectively learn complex mappings without degradation.
Impact of Temporal Modules (LSTM). The -LSTM variant shows the second-highest error rates (RMSE: 0.0531). Since EEG signals are inherently time-series data, this result underscores the importance of the LSTM module in capturing long-term temporal dependencies. Without LSTM, the model fails to adequately account for the temporal dynamics and historical context of the brain signals.
Impact of Feature Extraction (CNN and Transformer). Removing the feature extraction modules also negatively impacts performance, though to a lesser extent than the structural components above. The -CNN variant yields an RMSE of 0.0449, indicating that the CNN is essential for extracting local spatial-spectral features from the DE entropy maps.Meanwhile, the -Transformer variant results in an RMSE of 0.0331. While this is the smallest drop among the variants, it is still notably higher than the full model. This demonstrates that the Transformer’s self-attention mechanism provides a valuable refinement by capturing global dependencies and weighting critical feature segments, thereby enhancing the final prediction accuracy.
In conclusion, the ablation study validates the rationale behind CTL-ResFNet: the CNN extracts local features, the Transformer refines global attention, the LSTM captures temporal evolution, and the residual connections ensure stable integration, all working synergistically to achieve optimal performance.

4.5.4. Cross-Subject Generalization and Individual Adaptation Analysis

This study systematically validated the superior performance of the CTL-ResFNet model through LOOCV experiments and pretraining–finetuning comparison experiments. The experiments selected CNN, Transformer, LSTM, and CNN-Transformer as baseline models for comparative analysis. As shown in Table 5, in the LOOCV, CTL-ResFNet achieved an average RMSE of 0.0266 and an average MAE of 0.0215, significantly outperforming all comparative models. This result fully demonstrates the model’s exceptional generalization capability in zero-shot cross-subject scenarios, highlighting its unique architectural advantages. Further fine-tuning experimental results (Table 6) revealed that CTL-ResFNet (Avg RMSE = 0.0935) also exhibited significant advantages in individual adaptation ability, further validating the model’s strong transferability and adaptability.
The training curves of the experiments are shown in Figure 12 and Figure 13.

4.5.5. Comparison of Feature Performance Between LOOCV and Fine-Tuning Experiments

This study further extracted three feature indicators from EEG signals: the α / β band power ratio, wavelet entropy, and Hurst exponent. Two experimental paradigms—LOOCV and pretraining–finetuning (based on the CTL-ResFNet framework)—were employed for validation. As shown in Table 7, under the LOOCV setting, the α / β band power ratio demonstrated the best predictive performance, achieving an average RMSE of 0.0190. In contrast, in the pretraining–finetuning experiments presented in Table 8, the DE feature performed the best, with an average RMSE of 0.0935. This comparative result indicates that the optimal feature selection varies significantly under different validation paradigms, suggesting that the most suitable feature indicators should be chosen based on the specific experimental design in practical applications.
To evaluate the feasibility of deploying CTL-ResFNet in real-time fatigue monitoring scenarios, we further measured the inference speed on different hardware platforms. On a standard laptop equipped with an AMD R7-6800H CPU (Advanced Micro Devices, Inc., Santa Clara, CA, USA), the average inference time per sample was 2.022 ms, corresponding to a throughput of approximately 495 samples per second. On an embedded NVIDIA Jetson Nano B01 device (NVIDIA Corporation, Santa Clara, CA, USA), the average inference time per sample was 8.578 ms (≈117 samples/s). These results demonstrate that CTL-ResFNet can operate in real-time even under limited computational resources. The lightweight design effectively balances temporal modeling capacity and computational efficiency, making the model suitable for integration into embedded automotive fatigue-monitoring systems.

5. Conclusions

This paper introduces CTL-ResFNet, a novel deep learning model that integrates CNN, TransformerEncoder, LSTM, and residual connections to predict driver fatigue degree using historical EEG signals and PERCLOS. By effectively capturing spatiotemporal dependencies in EEG data, the proposed model outperforms baseline approaches (CNN, Transformer, LSTM, and CNN-Transformer) in both LOOCV and pretraining–finetuning experiments, as reflected by lower RMSE and MAE values.
Results demonstrate that combining EEG-based DE features with PERCLOS substantially improves fatigue degree prediction performance compared to using PERCLOS alone, as confirmed through LSTM validation experiments. Comparative analysis of alternative EEG features—including the α / β band power ratio, wavelet entropy, and Hurst exponent—reveals that while the α / β ratio performs best in cross-validation, DE features exhibit superior effectiveness in pretraining–finetuning scenarios, highlighting the need for context-aware feature selection.
These findings emphasize the promise of multimodal deep learning for fatigue monitoring, with CTL-ResFNet providing a robust framework for real-world deployment. Future research directions could include enhancing model interpretability through attention-based visualization or explainable AI methods such as SHAP or Grad-CAM to analyze which EEG channels and features contribute most to fatigue prediction. In addition, since the SEED-VIG dataset used in this study involves participants with a relatively narrow age range and unspecified gender distribution, the effects of demographic factors such as age and gender on fatigue prediction were not explicitly examined. Future work will therefore expand the dataset to include participants of diverse ages and genders to explore demographic influences on fatigue modeling and improve the generalizability of the proposed method.

Author Contributions

Conceptualization, Z.W. and X.D.; methodology, Z.W.; software, Z.W.; validation, Z.W., C.J. and J.S.; formal analysis, Z.W.; investigation, C.J. and J.S.; resources, X.D.; data curation, Z.W.; writing—original draft preparation, Z.W.; writing—review and editing, X.D.; visualization, Z.W.; supervision, X.D.; project administration, X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study is based on the publicly available SEED-VIG dataset. No new experiments involving human participants were conducted by the authors. The present study only involves secondary analysis of publicly available and anonymized data. Under these circumstances, ethical approval was not required for the present study.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data analyzed in this study were sourced from the SEED-VIG dataset provided by Shanghai Jiao Tong University, accessible at: https://bcmi.sjtu.edu.cn/home/seed/seed-vig.html (accessed on 25 November 2025). This dataset is subject to access restrictions and requires submitting an application to the data provider for approval prior to use.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A.1

Figure A1. Prediction performance of each model in the LOOCV experiment, with the horizontal axis representing the time steps (a total of 885 time steps, each separated by 8 s) and the vertical axis representing PERCLOS.
Figure A1. Prediction performance of each model in the LOOCV experiment, with the horizontal axis representing the time steps (a total of 885 time steps, each separated by 8 s) and the vertical axis representing PERCLOS.
Sensors 25 07316 g0a1
Figure A2. Prediction results of each model in the pre-training and fine-tuning experiment, with the x-axis representing time in seconds (total 7080 s) and the y-axis representing PERCLOS (Subject 7 (left) and Subject 19 (right)).
Figure A2. Prediction results of each model in the pre-training and fine-tuning experiment, with the x-axis representing time in seconds (total 7080 s) and the y-axis representing PERCLOS (Subject 7 (left) and Subject 19 (right)).
Sensors 25 07316 g0a2

Appendix A.2

Table A1. Summary of hyperparameter settings under two experimental configurations.
Table A1. Summary of hyperparameter settings under two experimental configurations.
HyperparameterLOOCV Default ValuePretraining–Finetuning Default Value
Data Processing Related
Sequence length33
Validation set ratio0.1
Test set ratio0.1
Batch size6464
Normalization methodZ-Score normalizationZ-Score normalization
Model Architecture Related
Batch size6464
Embedding dimension6464
LSTM hidden size3232
Number of LSTM layers22
Number of Transformer heads44
Transformer feedforward dimension6464
Transformer dropout rate0.10.1
Convolution kernel size11
Input feature dimension8686
Maximum positional encoding length10001000
Positional encoding dropout rate00
Training Strategy Related
Optimizer typeAdamAdam
Loss functionMSELossMSELoss
Learning rate0.001Pretraining: 0.001/Finetuning: 0.0001
Number of epochs20Pretraining: 150/Finetuning: 50
Learning rate schedulerNoneReduceLROnPlateau
Scheduler decay factor0.5
Scheduler patience5
Early stoppingNoneEnabled
Early stopping patience10
Minimum improvement threshold0.001
Randomness and Device Settings
Random seed4242
DeviceRTX3050RTX3050
Environment Configuration
Python version3.93.9
CUDA version11.311.3
PyTorch version1.11.01.11.0
Algorithm A1 Leave-One-Out Cross-Validation (LOOCV) Workflow for CTL-ResFNet
Require: 
Full Dataset D = { S 1 , S 2 , , S 23 } , where S i is the time-series data for subject i.
Require: 
Hyperparameters: Window size w = 3 , Batch size B = 64 , Epochs E = 20 , Learning rate η = 0.001 .
Ensure: 
Average performance metrics (RMSE, MAE) across all subjects.
  1:
for  i 1 to 23 do                                                                                                                          ▹ Iterate through each subject as test set
  2:
       D t e s t S i
  3:
       D t r a i n D { S i }
  4:
      Step 1: Sequence Generation (Sliding Window)
  5:
       X t r a i n , Y t r a i n SlidingWindow ( D t r a i n , w )                                                                        ▹ Input X: 3 steps, Target Y: 4th step PERCLOS
  6:
       X t e s t , Y t e s t SlidingWindow ( D t e s t , w )
  7:
      Step 2: Z-score Normalization (Crucial: Prevent Data Leakage)
  8:
                                             ▹ Compute statistics solely on the aggregated training data
  9:
       μ mean ( X t r a i n , axis = 0 )
10:
       σ std ( X t r a i n , axis = 0 )
11:
       X t r a i n ( X t r a i n μ ) / ( σ + ϵ )
12:
       X t e s t ( X t e s t μ ) / ( σ + ϵ )                                                                                                             ▹ Apply training stats to the test set
13:
      Step 3: Data Loader Initialization
14:
       L t r a i n GroupLoadersBySubject ( D t r a i n , B , shuffle = False )  ▹ Each subject gets a separate DataLoader, preserving temporal order
15:
       L t e s t DataLoader ( D t e s t , B )
16:
      Step 4: Model Training
17:
      Initialize Model M θ and Optimizer (Adam, η )
18:
      for  e p o c h 1 to E do
19:
             M θ . train ( )
20:
            Shuffle order of subject loaders in L t r a i n
21:
            for each subject_loader L s in L t r a i n  do
22:
                  for batch ( x b , y b ) in L s  do
23:
                         y ^ b M θ ( x b )
24:
                         L MSELoss ( y ^ b , y b )
25:
                         Backpropagate ( L )
26:
                         Optimizer . step ( )
27:
                  end for
28:
            end for
29:
      end for
30:
      Step 5: Evaluation
31:
       M θ . eval ( )
32:
       Y ^ t e s t EvaluateModel ( M θ , L t e s t )
33:
      Record RMSE and MAE between Y ^ t e s t and Y t e s t
34:
  end for
35:
  return Average RMSE and MAE across all LOOCV folds.
Algorithm A2 Cross-Subject Pre-training with Within-Subject Fine-tuning
Require: 
All Subjects Data D a l l = { S 1 , S 2 , , S 23 } .
Require: 
Hyperparameters: Pre-train η p r e = 0.001 , Fine-tune η f t = 0.0001 .
Require: 
Early Stopping Patience P = 10 , LR Scheduler Patience P l r = 5 .
Ensure: 
Performance metrics on the test set of the target subject.
  1:
Function ChronologicalSplitAndNormalize( S s u b )                 ▹ Splits data S s u b strictly by time and normalizes using training stats
  2:
D t r , D v a l , D t e ChronologicalSplit ( S s u b , ratios = 0.8 , 0.1 , 0.1 )
  3:
μ , σ Stats ( D t r )
  4:
D t r , D v a l , D t e Normalize ( D t r , D v a l , D t e , μ , σ )
  5:
return  L t r , L v a l , L t e (DataLoaders for each set)
  6:
End Function
  7:
Phase 1: Cross-Subject Pre-training
  8:
Initialize Model M θ , Optimizer ( η p r e ) , and Scheduler (ReduceLROnPlateau)
  9:
L t r a i n _ a l l [ ] , L v a l _ a l l [ ]
10:
for  i 1 to 23 do
11:
       L t r i , L v a l i , L t e i ChronologicalSplitAndNormalize ( S i )                                                     ▹ Normalize each subject’s data independently
12:
       L t r a i n _ a l l . append ( L t r i )
13:
       L v a l _ a l l . append ( L v a l i )
14:
end for
15:
ES _ counter 0 , L m i n
16:
while  ES _ counter < P  and  e p o c h < 150  do
17:
       Shuffle ( L t r a i n _ a l l ) , Shuffle ( L v a l _ a l l )                                                                ▹ Shuffle subject order, not sequence order within subject
18:
       L t r a i n AggregateTrainLoss ( M θ , L t r a i n _ a l l )
19:
       L v a l AggregateValidationLoss ( M θ , L v a l _ a l l )
20:
       Scheduler . step ( L v a l )
21:
      if  L v a l < L m i n  then
22:
             L m i n L v a l , Save best θ p r e θ , ES _ counter 0
23:
      else
24:
             ES _ counter ES _ counter + 1
25:
      end if
26:
end while
27:
Load θ θ p r e
28:
Phase 2: Within-Subject Fine-tuning (Example: Target Subject k)
29:
k 1                                                                                                                                                                                        ▹ Target subject
30:
L t r k , L v a l k , L t e k ChronologicalSplitAndNormalize ( S k )                                                                    ▹ Re-split and Re-normalize for subject k
31:
Initialize Optimizer ( η f t = 0.0001 ) and Scheduler (ReduceLROnPlateau)
32:
ES _ counter 0 , L m i n
33:
while  ES _ counter < P  and  e p o c h < 50  do
34:
       L t r a i n k TrainLoss ( M θ , L t r k )
35:
       L v a l k ValidationLoss ( M θ , L v a l k )
36:
       Scheduler . step ( L v a l k )
37:
      if  L v a l k < L m i n  then
38:
             L m i n L v a l k , Save best θ f t θ , ES _ counter 0
39:
      else
40:
             ES _ counter ES _ counter + 1
41:
      end if
42:
end while
43:
Load θ θ f t                                                                                                                                                  ▹ Load best fine-tuned weights
44:
Y ^ t e s t k EvaluateModel ( M θ , L t e k )
45:
return RMSE and MAE on S k ’s test set.

References

  1. Sprajcer, M.; Dawson, D.; Kosmadopoulos, A.; Sach, E.J.; E Crowther, M.; Sargent, C.; Roach, G.D. How tired is too tired to drive? A systematic review assessing the use of prior sleep duration to detect driving impairment. Nat. Sci. Sleep 2023, 15, 175–206. [Google Scholar] [CrossRef]
  2. Khunpisuth, O.; Chotchinasri, T.; Koschakosai, V.; Hnoohom, N. Driver drowsiness detection using eye-closeness detection. In Proceedings of the 2016 12th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Naples, Italy, 28 November–1 December 2016; pp. 661–668. [Google Scholar]
  3. Fan, J.; Smith, A.P. A preliminary review of fatigue among rail staff. Front. Psychol. 2018, 9, 634. [Google Scholar] [CrossRef]
  4. Ahmed, M.; Masood, S.; Ahmad, M.; El-Latif, A.A.A. Intelligent driver drowsiness detection for traffic safety based on multi CNN deep model and facial subsampling. IEEE Trans. Intell. Transp. Syst. 2021, 23, 19743–19752. [Google Scholar] [CrossRef]
  5. Yogarajan, G.; Singh, R.N.; Nandhu, S.A.; Rudhran, R.M. Drowsiness detection system using deep learning based data fusion approach. Multimed. Tools Appl. 2024, 83, 36081–36095. [Google Scholar]
  6. Gao, Z.; Chen, X.; Xu, J.; Yu, R.; Zhang, H.; Yang, J. Semantically-Enhanced Feature Extraction with CLIP and Transformer Networks for Driver Fatigue Detection. Sensors 2024, 24, 7948. [Google Scholar] [CrossRef] [PubMed]
  7. Ganguly, B.; Dey, D.; Munshi, S. An Attention Deep Learning Framework-Based Drowsiness Detection Model for Intelligent Transportation System. IEEE Trans. Intell. Transp. Syst. 2025, 26, 4517–4527. [Google Scholar] [CrossRef]
  8. Lyu, X.; Akbar, M.A.; Manimurugan, S.; Jiang, H. Driver Fatigue Warning Based on Medical Physiological Signal Monitoring for Transportation Cyber-Physical Systems. IEEE Trans. Intell. Transp. Syst. 2025, 26, 14237–14249. [Google Scholar] [CrossRef]
  9. Wang, K.; Mao, X.; Song, Y.; Chen, Q. EEG-based fatigue state evaluation by combining complex network and frequency-spatial features. J. Neurosci. Methods 2025, 416, 110385. [Google Scholar]
  10. Åkerstedt, T.; Torsvall, L. Continuous Electrophysiological Recording. In Breakdown in Human Adaptation to ‘Stress’ Towards a Multidisciplinary Approach Volume I; Springer: Dordrecht, The Netherlands, 1984; pp. 567–583. [Google Scholar]
  11. Åkerstedt, T.; Kecklund, G.; Knutsson, A. Manifest sleepiness and the spectral content of the EEG during shift work. Sleep 1991, 14, 221–225. [Google Scholar] [CrossRef]
  12. Eoh, H.J.; Chung, M.K.; Kim, S.-H. Electroencephalographic study of drowsiness in simulated driving with sleep deprivation. Int. J. Ind. Ergon. 2005, 35, 307–320. [Google Scholar] [CrossRef]
  13. Jap, B.T.; Lal, S.; Fischer, P.; Bekiaris, E. Using EEG spectral components to assess algorithms for detecting fatigue. Expert Syst. Appl. 2009, 36, 2352–2359. [Google Scholar] [CrossRef]
  14. Khushaba, R.N.; Kodagoda, S.; Lal, S.; Dissanayake, G. Driver drowsiness classification using fuzzy wavelet-packet-based feature-extraction algorithm. IEEE Trans. Biomed. Eng. 2010, 58, 121–131. [Google Scholar] [CrossRef] [PubMed]
  15. Zhao, C.; Zheng, C.; Zhao, M.; Tu, Y.; Liu, J. Multivariate autoregressive models and kernel learning algorithms for classifying driving mental fatigue based on electroencephalographic. Expert Syst. Appl. 2011, 38, 1859–1865. [Google Scholar] [CrossRef]
  16. Huo, X.-Q.; Zheng, W.-L.; Lu, B.-L. Driving fatigue detection with fusion of EEG and forehead EOG. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 897–904. [Google Scholar]
  17. Chaabene, S.; Bouaziz, B.; Boudaya, A.; Hökelmann, A.; Ammar, A.; Chaari, L. Convolutional neural network for drowsiness detection using EEG signals. Sensors 2021, 21, 1734. [Google Scholar] [CrossRef] [PubMed]
  18. Cui, J.; Lan, Z.; Sourina, O.; Müller-Wittig, W. EEG-based cross-subject driver drowsiness recognition with an interpretable convolutional neural network. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 7921–7933. [Google Scholar]
  19. Shi, L.-C.; Jiao, Y.-Y.; Lu, B.-L. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 6627–6630. [Google Scholar]
  20. Bashivan, P.; Rish, I.; Yeasin, M.; Codella, N. Learning representations from EEG with deep recurrent-convolutional neural networks. arXiv 2015, arXiv:1511.06448. [Google Scholar]
  21. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 2002, 86, 2278–2324. [Google Scholar]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  25. Jiao, Y.; Deng, Y.; Luo, Y.; Lu, B.-L. Driver sleepiness detection from EEG and EOG signals using GAN and LSTM networks. Neurocomputing 2020, 408, 100–111. [Google Scholar] [CrossRef]
  26. Avitan, L.; Teicher, M.; Abeles, M. EEG generator—A model of potentials in a volume conductor. J. Neurophysiol. 2009, 102, 3046–3059. [Google Scholar] [CrossRef]
  27. Gharagozlou, F.; Saraji, G.N.; Mazloumi, A.; Nahvi, A.; Nasrabadi, A.M.; Foroushani, A.R.; Kheradmand, A.A.; Ashouri, M.; Samavati, M. Detecting driver mental fatigue based on EEG alpha power changes during simulated driving. Iran. J. Public Health 2015, 44, 1693. [Google Scholar]
  28. Panicker, R.C.; Puthusserypady, S.; Sun, Y. An asynchronous P300 BCI with SSVEP-based control state detection. IEEE Trans. Biomed. Eng. 2011, 58, 1781–1788. [Google Scholar] [CrossRef]
  29. Cui, Y.; Xu, Y.; Wu, D. EEG-based driver drowsiness estimation using feature weighted episodic training. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 2263–2273. [Google Scholar] [CrossRef]
  30. Gao, Z.; Wang, X.; Yang, Y.; Mu, C.; Cai, Q.; Dang, W.; Zuo, S. EEG-based spatio–temporal convolutional neural network for driver fatigue evaluation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2755–2763. [Google Scholar] [CrossRef]
  31. Akin, M.; Kurt, M.B.; Sezgin, N.; Bayram, M. Estimating vigilance level by using EEG and EMG signals. Neural Comput. Appl. 2008, 17, 227–236. [Google Scholar] [CrossRef]
  32. Zheng, W.-L.; Lu, B.-L. A multimodal approach to estimating vigilance using EEG and forehead EOG. J. Neural Eng. 2017, 14, 026017. [Google Scholar] [CrossRef] [PubMed]
  33. Hu, L.; Zhang, Z. EEG Signal Processing and Feature Extraction; Springer: Singapore, 2019. [Google Scholar]
  34. Luck, S.J. An Introduction to the Event-Related Potential Technique; MIT Press: Cambridge, MA, USA, 2014. [Google Scholar]
  35. Wei, C.-S.; Lin, Y.-P.; Wang, Y.-T.; Lin, C.-T.; Jung, T.-P. A subject-transfer framework for obviating inter-and intra-subject variability in EEG-based drowsiness detection. NeuroImage 2018, 174, 407–419. [Google Scholar] [CrossRef] [PubMed]
  36. Wei, C.-S.; Wang, Y.-T.; Lin, C.-T.; Jung, T.-P. Toward drowsiness detection using non-hair-bearing EEG-based brain-computer interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 400–406. [Google Scholar] [CrossRef]
  37. Liu, Y.; Lan, Z.; Cui, J.; Sourina, O.; Müller-Wittig, W. Inter-subject transfer learning for EEG-based mental fatigue recognition. Adv. Eng. Inform. 2020, 46, 101157. [Google Scholar] [CrossRef]
  38. Zhang, C.; Wang, H.; Fu, R. Automated detection of driver fatigue based on entropy and complexity measures. IEEE Trans. Intell. Transp. Syst. 2013, 15, 168–177. [Google Scholar] [CrossRef]
  39. Sigari, M.-H.; Fathy, M.; Soryani, M. A driver face monitoring system for fatigue and distraction detection. Int. J. Veh. Technol. 2013, 2013, 263983. [Google Scholar] [CrossRef]
  40. Zhang, Z.; Zhang, J. A new real-time eye tracking based on nonlinear unscented Kalman filter for monitoring driver fatigue. J. Control Theory Appl. 2010, 8, 181–188. [Google Scholar] [CrossRef]
  41. Du, G.; Zhang, L.; Su, K.; Wang, X.; Teng, S.; Liu, P.X. A multimodal fusion fatigue driving detection method based on heart rate and PERCLOS. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21810–21820. [Google Scholar] [CrossRef]
  42. Dai, L.; Li, Y.; Zhang, M. Detection of operator fatigue in the main control room of a nuclear power plant based on eye blink rate, PERCLOS and mouse velocity. Appl. Sci. 2023, 13, 2718. [Google Scholar] [CrossRef]
  43. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  44. Panigrahi, S.; Nanda, A.; Swarnkar, T. A survey on transfer learning. In Intelligent and Cloud Computing: Proceedings of ICICC 2019, Volume 1; Springer: Singapore, 2020; pp. 781–789. [Google Scholar]
  45. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  46. Ruder, S.; Peters, M.E.; Swayamdipta, S.; Wolf, T. Transfer learning in natural language processing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Tutorials, Minneapolis, MN, USA, 2–7 June 2019; pp. 15–18. [Google Scholar]
  47. Maqsood, M.; Nazir, F.; Khan, U.; Aadil, F.; Jamal, H.; Mehmood, I.; Song, O.-Y. Transfer learning assisted classification and detection of Alzheimer’s disease stages using 3D MRI scans. Sensors 2019, 19, 2645. [Google Scholar] [CrossRef] [PubMed]
  48. Schweikert, G.; Rätsch, G.; Widmer, C.; Schölkopf, B. An empirical analysis of domain adaptation algorithms for genomic sequence analysis. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–11 December 2008; Volume 21. [Google Scholar]
  49. Hyvärinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar]
  50. Särkelä, M.O.K.; Ermes, M.J.; van Gils, M.J.; Yli-Hankala, A.M.; Jäntti, V.H.; Vakkuri, A.P. Quantification of epileptiform electroencephalographic activity during sevoflurane mask induction. Anesthesiology 2007, 107, 928–938. [Google Scholar] [CrossRef]
  51. Hurst, H. Long-term storage of reservoirs: An experimental study. Trans. Am. Soc. Civ. Eng. 1951, 116, 770–799. [Google Scholar]
  52. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. CTL-ResFNet Model overall structure.
Figure 1. CTL-ResFNet Model overall structure.
Sensors 25 07316 g001
Figure 2. 1D CNN (kernel_size = 1).
Figure 2. 1D CNN (kernel_size = 1).
Sensors 25 07316 g002
Figure 3. 1D CNN (kernel_size = 1, in_channels = 3, out_channels = 2).
Figure 3. 1D CNN (kernel_size = 1, in_channels = 3, out_channels = 2).
Sensors 25 07316 g003
Figure 4. Transformer Encoder.
Figure 4. Transformer Encoder.
Sensors 25 07316 g004
Figure 5. LSTM structure.
Figure 5. LSTM structure.
Sensors 25 07316 g005
Figure 6. Schematic diagram of electrode channel and EEG signal acquisition.
Figure 6. Schematic diagram of electrode channel and EEG signal acquisition.
Sensors 25 07316 g006
Figure 7. Schematic diagram of dataset partitioning in the LOOCV experiment (using sub2 as the test set example).
Figure 7. Schematic diagram of dataset partitioning in the LOOCV experiment (using sub2 as the test set example).
Sensors 25 07316 g007
Figure 8. Power spectral density after 1–75 Hz filtering.
Figure 8. Power spectral density after 1–75 Hz filtering.
Sensors 25 07316 g008
Figure 9. 17 ICA components.
Figure 9. 17 ICA components.
Sensors 25 07316 g009
Figure 10. EMG artifact-related components.
Figure 10. EMG artifact-related components.
Sensors 25 07316 g010
Figure 11. Comparison of EEG signals before and after preprocessing After preprocessing the raw EEG signals to obtain.
Figure 11. Comparison of EEG signals before and after preprocessing After preprocessing the raw EEG signals to obtain.
Sensors 25 07316 g011
Figure 12. Training curve, taking sub2 as an example.
Figure 12. Training curve, taking sub2 as an example.
Sensors 25 07316 g012
Figure 13. The pre-training process training curve, where the point with the lowest validation set loss is the moment when the early stopping mechanism saves the model.
Figure 13. The pre-training process training curve, where the point with the lowest validation set loss is the moment when the early stopping mechanism saves the model.
Sensors 25 07316 g013
Table 1. CTL-ResFNet Parameter Information.
Table 1. CTL-ResFNet Parameter Information.
ComponentParameters
PositionalEncodingd_model = 86
Conv1dinput = 86, output = 64, kernel = 1
TransformerEncoderLayerd_model = 64, nhead = 4, dim_feedforward = 64
TransformerEncoder1 layer
LSTMinput = 64, hidden = 32, layers = 2
FCinput = 32, output = 1
Table 2. Experimental Study on Window Length Selection Based on LSTM (LOOCV Experiment, EEG + PERCLOS).
Table 2. Experimental Study on Window Length Selection Based on LSTM (LOOCV Experiment, EEG + PERCLOS).
Sub.Window Length 1Window Length 2Window Length 3Window Length 4Window Length 5
RMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
10.11540.10040.11310.09740.07680.06520.07930.07110.07330.0630
20.04930.03930.04810.04140.03560.03020.08300.07230.02830.0213
30.08610.07670.07780.07280.04690.04070.05230.04890.09350.0870
40.05310.04330.10000.08440.04430.03710.05080.04170.07540.0664
50.05720.04680.06360.05640.07300.06510.07100.06190.08170.0724
60.08880.06590.06150.04880.05820.04290.06250.04670.07410.0611
70.16500.11610.17640.11700.05260.04450.13520.10100.09100.0698
80.05800.04790.03050.02490.06720.06220.04630.04360.05240.0492
90.12020.10910.18090.16550.05660.05380.03780.03270.05490.0518
100.03430.02760.04130.03190.05450.04110.03510.02690.03770.0295
110.02350.01830.04180.03230.03650.02810.04740.04390.04980.0467
120.04860.04400.04380.03930.04510.03940.04740.04350.04820.0448
130.03210.02830.03840.03070.07070.06490.06220.05780.08220.0750
140.08280.06750.08430.06890.06920.05660.06350.05120.06370.0520
150.10750.09900.05750.05110.07060.06070.04990.04250.06840.0604
160.06770.05670.07190.06000.04970.04330.04640.03720.05910.0510
170.05550.05020.04420.03770.04830.04380.05470.04930.04660.0416
180.06000.04900.10090.08760.07890.07040.06870.06020.06980.0621
190.05730.04460.05730.04950.07090.05960.06270.05010.06680.0559
200.09180.07190.09270.07130.08270.06880.06190.05310.05690.0482
210.03650.03360.04360.03390.05810.05220.10080.07910.03890.0363
220.07590.06280.07190.04910.05280.04250.05530.04140.07460.0573
230.08210.07110.06050.05060.07640.05740.04950.04120.05590.0435
Avg.0.07170.05960.07400.06090.05980.05090.06190.05210.06270.0542
Table 3. Experimental Comparison of LSTM-based EEG + PERCLOS and PERCLOS-only Approaches (LOOCV Experiment).
Table 3. Experimental Comparison of LSTM-based EEG + PERCLOS and PERCLOS-only Approaches (LOOCV Experiment).
Sub.EEG + PERCLOSPERCLOS-Only
RMSEMAERMSEMAE
10.07680.06520.08420.0767
20.03560.03020.03990.0335
30.04690.04070.05950.0557
40.04430.03710.04650.0390
50.07300.06510.06720.0575
60.05820.04290.08520.0644
70.05260.04450.05640.0462
80.06720.06220.04910.0458
90.05660.05380.05200.0496
100.05450.04110.05920.0437
110.03650.02810.05370.0502
120.04510.03940.05420.0496
130.07070.06490.06980.0656
140.06920.05660.08660.0703
150.07060.06070.08610.0813
160.04970.04330.06400.0541
170.04830.04380.04640.0416
180.07890.07040.08020.0700
190.07090.05960.08530.0776
200.08270.06880.10280.0843
210.05810.05220.07060.0677
220.05280.04250.07860.0492
230.07640.05740.09410.0581
Avg.0.05980.05090.06830.0579
Table 4. Results of ablation experiments (LOOCV experiment).
Table 4. Results of ablation experiments (LOOCV experiment).
Sub.CTL-ResFNet-CNN-Transformer-LSTM-Residual
RMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
10.04110.03620.04210.03440.03690.03220.07790.06040.02510.0207
20.02570.02220.04030.03580.05390.04820.03500.03020.08070.0694
30.01360.01100.02520.02260.05510.04890.08250.07510.06120.0426
40.04010.03020.03140.02540.02880.02360.03350.02540.04460.0337
50.02560.02090.03810.03140.01930.01550.03600.03040.14610.1168
60.03320.02550.03500.02800.03150.02260.03740.02800.07430.0555
70.02610.01530.04870.03830.12900.09340.03470.02750.06810.0537
80.02310.02040.05900.05230.04700.04000.03150.02400.07310.0562
90.02150.01770.05890.05170.05410.04820.08120.07330.07760.0675
100.01940.01650.04010.03330.03560.02690.03320.02520.11210.0959
110.04150.03800.04960.04370.01710.01480.04780.03970.04820.0389
120.01580.01280.10850.09250.02160.01800.12450.10220.04460.0341
130.01600.01280.02960.02490.01370.01070.03980.03470.07720.0657
140.02830.02630.05840.04930.02160.01910.07980.07340.09660.0738
150.02240.01800.02880.02270.01880.01550.06630.04840.08010.0537
160.03550.02390.03090.02210.01910.01520.04110.03070.06780.0470
170.02240.01710.01940.01710.01860.01530.02060.01410.03950.0311
180.02040.01710.02430.01880.02490.01840.07060.06030.04350.0357
190.01910.01540.01700.01370.01690.01460.04440.02570.05120.0360
200.04260.03110.11390.06620.01900.01600.05930.04850.11700.0902
210.01130.00820.02850.01720.01670.01390.01630.01310.03310.0290
220.01390.01150.02380.01720.02410.01940.05540.05120.02680.0215
230.05490.04710.08070.07350.03710.03200.07210.06270.10410.0935
Avg.0.02660.02150.04490.03620.03310.02710.05310.04370.06920.0549
Table 5. LOOCV Evaluation Results for Each Model.
Table 5. LOOCV Evaluation Results for Each Model.
Sub.CTL-ResFNetCNNTransformerLSTMCNN-Transformer
RMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
10.04110.03620.04050.03150.05890.04890.07680.06520.04530.0400
20.02570.02220.08300.07150.07890.07060.03560.03020.03310.0289
30.01360.01100.08730.05430.07930.05190.04690.04070.02390.0203
40.04010.03020.15990.07860.07920.05660.04430.03710.02670.0209
50.02560.02090.07700.06490.12580.10450.07300.06510.04200.0338
60.03320.02550.14260.11950.07520.05730.05820.04290.03370.0272
70.02610.01530.45330.36040.17190.12790.05260.04450.04600.0390
80.02310.02040.04440.03310.05700.04660.06720.06220.03370.0275
90.02150.01770.02600.01860.06130.04780.05660.05380.05590.0494
100.01940.01650.05940.05150.08460.06400.05450.04110.03110.0232
110.04150.03800.06580.05710.06630.05050.03650.02810.03340.0293
120.01580.01280.06090.04570.11050.07630.04510.03940.03270.0274
130.01600.01280.09330.02690.09740.08150.07070.06490.04020.0340
140.02830.02630.11640.09050.11130.09300.06920.05660.04140.0341
150.02240.01800.11210.09450.10850.08470.07060.06070.04210.0356
160.03550.02390.11000.08040.06560.05250.04970.04330.03980.0299
170.02240.01710.05160.03990.07170.06000.04830.04380.02340.0182
180.02040.01710.10770.08890.13520.11650.07890.07040.03850.0313
190.01910.01540.04760.03900.14300.13010.07090.05960.03390.0267
200.04260.03110.21370.16660.11860.10100.08270.06880.04290.0329
210.01130.00820.13040.07020.07020.04600.05810.05220.03540.0282
220.01390.01150.10970.09040.11430.10120.05280.04250.03860.0326
230.05490.04710.10520.09100.07720.06500.07640.05740.06310.0569
Avg.0.02660.02150.10620.08110.09400.07540.05980.05090.03810.0316
Table 6. Fine-Tuning Results of Each Model.
Table 6. Fine-Tuning Results of Each Model.
Sub.CTL-ResFNetCNNTransformerLSTMCNN-Transformer
RMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
10.06140.05750.26100.25350.13100.11030.25820.25680.02220.0203
20.05620.04630.08110.07070.18690.18220.04420.03570.12100.1128
30.02730.02340.04320.03620.06840.05710.09700.09250.05060.0470
40.02480.02070.07020.06180.15810.14480.03720.03000.07830.0692
50.17460.15700.13980.11900.12300.10800.20560.17970.17790.1571
60.14570.13820.30410.27580.10210.09180.32900.29180.13250.1248
70.02140.01700.06950.06040.16330.12390.05310.03180.08090.0638
80.06650.05290.03530.02840.03940.03140.03890.03250.07040.0617
90.05190.03520.04620.03810.06560.04250.04290.03540.07630.0566
100.26500.24060.23570.22720.23910.22280.20030.18920.29020.2312
110.03370.03020.03890.03080.10150.09450.05300.04660.06410.0552
120.09200.08270.09680.07390.09390.07870.04950.04260.14880.1311
130.04680.03570.06600.05520.06570.05790.15040.14510.06500.0516
140.28760.26010.40930.36950.20250.17830.24200.21820.43970.3945
150.15450.15450.16200.14940.07550.05210.16590.16320.01240.0118
160.14010.10990.13970.11820.12430.09960.15840.12100.19690.1514
170.04650.04050.05730.04540.09750.08500.05710.04580.07940.0728
180.11680.11570.32790.32120.16270.14250.37490.37140.22200.1951
190.01910.01650.04450.03690.04190.03200.05720.05050.05770.0466
200.04610.04130.14810.13640.09320.07960.05990.04750.12900.1198
210.05280.04860.03950.03400.04840.04140.08080.07830.05770.0504
220.06240.05890.09400.08550.06290.04940.03980.03080.04930.0451
230.15740.15190.30420.28090.11720.09710.23660.22010.18710.1056
Avg.0.09350.08410.13980.12650.11150.09580.13100.11980.12210.1041
Table 7. Experimental Results for Different Dimensional Features (CTL-ResFNet, LOOCV).
Table 7. Experimental Results for Different Dimensional Features (CTL-ResFNet, LOOCV).
Sub. α / β RatioWavelet EntropyHurst ExponentDE
RMSEMAERMSEMAERMSEMAERMSEMAE
10.01560.01160.01670.01400.02180.01770.04110.0362
20.01180.01020.01740.01480.02220.01940.02570.0222
30.01910.01750.01610.01300.01710.01450.01360.0110
40.01250.00880.01160.00930.01280.01000.04010.0302
50.03070.02300.02800.02230.03370.02830.02560.0209
60.01950.01330.02410.01890.03120.02480.03320.0255
70.01240.00910.01690.01340.01090.00830.02610.0153
80.01010.00840.04190.03350.01720.01450.02310.0204
90.00840.00740.07710.07070.04000.03820.02150.0177
100.01520.01180.01630.01150.01770.01420.01940.0165
110.02340.02210.02030.01540.01280.00990.04150.0380
120.01540.01060.02390.01870.01990.01510.01580.0128
130.01090.00760.01490.01200.01260.00990.01600.0128
140.02120.01640.02400.01990.02140.01760.02830.0263
150.02740.02320.01440.01110.01660.01370.02240.0180
160.01710.01180.01660.01280.01360.01080.03550.0239
170.02090.01310.02700.01950.00990.00720.02240.0171
180.02100.01860.01480.01190.02770.02300.02040.0171
190.01630.01230.01820.01270.02470.02180.01910.0154
200.01960.01410.01960.01720.02440.01880.04260.0311
210.00800.00540.01060.00860.01230.00970.01130.0082
220.04290.02820.01560.01360.01340.01110.01390.0115
230.03790.03260.04000.03550.03360.02940.05490.0471
Avg.0.01900.01470.02290.01870.02030.01680.02660.0215
Table 8. Experimental results for different dimensional features (based on CTL-ResFNet, pretraining–finetuning).
Table 8. Experimental results for different dimensional features (based on CTL-ResFNet, pretraining–finetuning).
Sub. α / β Band Power RatioWavelet EntropyHurst ExponentDE
RMSEMAERMSEMAERMSEMAERMSEMAE
10.08470.08460.06450.06450.08160.08110.06140.0575
20.05130.04340.05460.04650.04850.04010.05620.0463
30.05360.04800.05950.04650.06550.06860.02730.0234
40.04940.04140.07280.06180.04870.03880.02480.0207
50.15190.12870.21200.18890.19950.17110.17460.1570
60.10930.10140.10680.09820.13790.12340.14570.1382
70.04020.03280.06140.05140.06170.04920.02140.0170
80.07060.06040.10200.08320.06920.05670.06650.0529
90.10240.06710.14540.10190.09740.07400.05190.0352
100.24220.22150.30990.26720.21810.19420.26500.2406
110.03480.02900.01780.01440.06620.05950.03370.0302
120.07320.06470.08560.07120.10690.09200.09200.0827
130.04870.04010.06700.05620.07600.06570.04680.0357
140.33350.31120.36060.30350.28920.24230.28760.2601
150.13130.13130.11380.11270.08980.08780.15450.1545
160.16140.12860.17170.13940.17860.13700.14010.1099
170.05360.04820.02230.01780.04620.04200.04650.0405
180.09220.09080.08330.08200.06630.06430.11680.1157
190.02450.02010.06020.05470.04090.03350.01910.0165
200.04930.04570.05060.04720.06170.05600.04610.0413
210.04850.04060.06390.05300.05390.04350.05280.0486
220.06030.05610.06490.06120.17200.16610.06240.0589
230.14590.13320.17780.16180.16490.13450.15740.1519
Avg.0.09710.08560.10990.09500.10610.09290.09350.0841
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Du, X.; Jiang, C.; Sun, J. Research on the Prediction of Driver Fatigue Degree Based on EEG Signals. Sensors 2025, 25, 7316. https://doi.org/10.3390/s25237316

AMA Style

Wang Z, Du X, Jiang C, Sun J. Research on the Prediction of Driver Fatigue Degree Based on EEG Signals. Sensors. 2025; 25(23):7316. https://doi.org/10.3390/s25237316

Chicago/Turabian Style

Wang, Zhanyang, Xin Du, Chengbin Jiang, and Junyang Sun. 2025. "Research on the Prediction of Driver Fatigue Degree Based on EEG Signals" Sensors 25, no. 23: 7316. https://doi.org/10.3390/s25237316

APA Style

Wang, Z., Du, X., Jiang, C., & Sun, J. (2025). Research on the Prediction of Driver Fatigue Degree Based on EEG Signals. Sensors, 25(23), 7316. https://doi.org/10.3390/s25237316

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop