Next Article in Journal
Evaluating Machine Learning-Based Classification of Human Locomotor Activities for Exoskeleton Control Using Inertial Measurement Unit and Pressure Insole Data
Next Article in Special Issue
Intelligent Compaction System for Soil-Rock Mixture Subgrades: Real-Time Moisture-CMV Fusion Control and Embedded Edge Computing
Previous Article in Journal
Cross-Domain Object Detection with Hierarchical Multi-Scale Domain Adaptive YOLO
Previous Article in Special Issue
Intelligent Detection and Description of Foreign Object Debris on Airport Pavements via Enhanced YOLOv7 and GPT-Based Prompt Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TMU-Net: A Transformer-Based Multimodal Framework with Uncertainty Quantification for Driver Fatigue Detection

School of Electronic Information Engineering, Xi’an Technological University, Xi’an 710021, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(17), 5364; https://doi.org/10.3390/s25175364
Submission received: 31 July 2025 / Revised: 22 August 2025 / Accepted: 28 August 2025 / Published: 29 August 2025
(This article belongs to the Special Issue AI and Smart Sensors for Intelligent Transportation Systems)

Abstract

Driving fatigued is a prevalent issue frequently contributing to traffic accidents, prompting the development of automated fatigue detection methods based on various data sources, particularly reliable physiological signals. However, challenges in accuracy, robustness, and practicality persist, especially for cross-subject detection. Multimodal data fusion can enhance the effective estimation of driver fatigue. In this work, we leverage the advantages of multimodal signals to propose a novel Multimodal Attention Network (TMU-Net) for driver fatigue detection, achieving precise fatigue assessment by integrating electroencephalogram (EEG) and electrooculogram (EOG) signals. The core innovation of TMU-Net lies in its unimodal feature extraction module, which combines causal convolution, ConvSparseAttention, and Transformer encoders to effectively capture spatiotemporal features, and a multimodal fusion module that employs cross-modal attention and uncertainty-weighted gating to dynamically integrate complementary information. By incorporating uncertainty quantification, TMU-Net significantly enhances robustness to noise and individual variability. Experimental validation on the SEED-VIG dataset demonstrates TMU-Net’s superior performance stability across 23 subjects in cross-subject testing, effectively leveraging the complementary strengths of EEG (2 Hz full-band and five-band features) and EOG signals for high-precision fatigue detection. Furthermore, attention heatmap visualization reveals the dynamic interaction mechanisms between EEG and EOG signals, confirming the physiological rationality of TMU-Net’s feature fusion strategy. Practical challenges and future research directions for fatigue detection methods are also discussed.

1. Introduction

Fatigue driving poses a significant challenge to global road safety. According to statistics from the World Health Organization, approximately 12–20% of road traffic accidents are directly attributed to driver fatigue [1]. As is well-documented, fatigue impairs perception, memory, decision-making, and judgment, thereby increasing the risk of traffic accidents [2,3]. A recent report by the National Highway Traffic Safety Administration (NHTSA) revealed that in the United States alone, fatigue-related driving accidents exceed 100,000 annually, resulting in severe casualties and substantial economic losses [4]. This critical situation has spurred continuous efforts in both academia and industry to explore more effective driver fatigue detection technologies.
Recent studies have primarily pursued three technical approaches for driver fatigue detection. (1) Vehicle motion-based methods analyze vehicular dynamic parameters such as steering wheel angle, speed, and acceleration. Research indicates that these methods achieve high detection accuracy in structured road environments [5]; however, their performance is significantly influenced by road complexity, with accuracy potentially declining markedly in urban settings [6]. (2) Driver behavior-based methods employ computer vision to monitor facial features (e.g., eye closure duration, yawning frequency) [7]. The hybrid LSTM-CNN model developed by Quddus et al. [8] attained 89.7% accuracy under standardized test conditions. Nevertheless, such methods face limitations in low-light conditions or when drivers wear glasses [9], and struggle to distinguish fatigue from transient eye closures [10]. (3) Physiological signal-based methods leverage biomarkers, among which PERCLOS (Percentage of Eyelid Closure Over the Pupil) is widely regarded as the gold standard for fatigue assessment [11]. Initially proposed by Dinges et al. in NHTSA-sponsored research [11], PERCLOS was later validated by Wierwille et al. through large-scale on-road experiments, demonstrating strong correlation with actual driver fatigue states [12].
Further studies demonstrate that when PERCLOS values exceed 0.15, drivers exhibit a significantly elevated risk of accidents [13]. However, computer vision-based PERCLOS detection faces substantial challenges in real-world driving conditions. Bergasa et al. [13] reported that detection accuracy may decrease by over 40% in low-light environments (e.g., nighttime or tunnels). Ji et al. [14] found that head rotations exceeding 15° markedly increase measurement errors, while Garcia et al. [15] noted that obstructions like eyeglasses or sunglasses can lead to a 35% failure rate in PERCLOS detection. Additionally, interethnic variations in palpebral fissure height (2–3 mm) further compromise measurement consistency [15].
These limitations have driven researchers toward more reliable direct physiological signal detection methods. Multimodal physiological signal fusion approaches, which integrate information from diverse modalities (e.g., electroencephalography (EEG) for brain activity and electrooculography (EOG) for eye movements), effectively overcome the constraints of single-signal systems [16]. Zheng and Lu [17] demonstrated that such multimodal methods offer distinct advantages over traditional PERCLOS detection, including the following: Robustness to environmental interference; earlier fatigue state detection; and direct measurement of neurological activity.
Current multimodal frameworks for driver fatigue detection predominantly adopt three fusion strategies: (1) Early fusion methods, exemplified by Zheng et al.’s EEG-EOG feature-level concatenation model [18], demonstrate computational efficiency but suffer from information redundancy and feature interference, with studies showing 15–20% accuracy degradation when handling modalities with significant sampling rate disparities [19]; (2) late fusion approaches, like Luo et al.’s decision-level system [20], enable modular asynchronous processing but require stringent temporal alignment and may disregard cross-modal correlations; and (3) hybrid methods, such as Cheng et al.’s VigilanceNet [21], achieve optimal laboratory performance, yet their computational complexity hinders real-time implementation in actual driving scenarios. Recent advances in multimodal information processing have provided insights for optimizing fusion strategies. For instance, Cai et al. [22] proposed a visual grounding-based reinforcement-learning framework that enhances model robustness in dynamic environments through object-level feature integration. This approach offers valuable reference for the multimodal fusion of EEG and EOG signals, particularly in addressing signal interference issues in complex driving scenarios.
Through a systematic analysis of the existing literature, we identified several critical limitations in current research. (i) Absence of dynamic reliability assessment—signal quality fluctuates with motion artifacts (EEG contamination in 23–41% of recordings [16]) and environmental noise (EOG variance during head movements [23]), while current static-weighting schemes fail to adapt; (ii) inadequate uncertainty quantification—despite evidence that probabilistic modeling enhances noise robustness by 18–22% [24], most systems neglect this capability; (iii) limited cross-subject generalization—model performance typically degrades by 30–40% for new subjects due to physiological variability, where adversarial domain adaptation [25] reduces this gap by 12% at the expense of tripled computational overhead.
This study aims to develop a robust and accurate driver fatigue detection framework (TMU-Net) to address the limitations of existing methods. To this end, we propose key design choices involving multimodal fusion (EEG + EOG) and the simultaneous use of 2 Hz full-band and five-band EEG features. Single physiological modalities are susceptible to noise and provide limited information. Electroencephalogram (EEG) signals directly reflect changes in central neural activity, serving as an early internal indicator of fatigue, but are prone to contamination from electromyographic and motion artifacts. In contrast, electrooculogram (EOG) signals capture prominent external behavioral manifestations of fatigue, exhibit high correlation with PERCLOS, and offer greater stability, though their reliability may be affected by visual obstructions or lighting variations. Therefore, fusing EEG and EOG leverages the complementary strengths of the “internal and early” nature of EEG and the “external and direct” nature of EOG, enhancing system robustness across diverse scenarios. Concurrently, 2 Hz full-band features capture subtle, frequency-specific variations in brain activity, providing a rich data foundation and improving the model’s sensitivity to minor fatigue changes. Conversely, the classical five-band features are linked to established physiological mechanisms, offering strong interpretability and smoothing noise to some extent through spectral consolidation, thus yielding more stable feature representations. By integrating both feature sets, the model benefits from the sensitivity of high-resolution data and the robustness and interpretability of standard frequency bands.
The main contributions of this paper are summarized as follows:
  • We propose TMU-Net, a novel framework that for the first time integrates EEG signals (both 2 Hz full-band and five-band features) with EOG signals, incorporating uncertainty modeling to enhance the accuracy and robustness of driver fatigue detection.
  • We design a unimodal feature extraction module that employs causal convolution, ConvSparseAttention, and Transformer encoders to capture spatiotemporal features, while integrating uncertainty modeling to improve prediction reliability.
  • We dynamically integrate features through a weighted gating mechanism, thereby achieving adaptive optimization of fatigue detection performance.

2. Methodology

In this section, we present our model, TMU-Net. As illustrated in Figure 1, the model architecture primarily consists of two core modules: the unimodal feature extraction module and the multimodal fusion module. Raw EEG and EOG signals are first processed independently by the unimodal module, which extracts spatiotemporal features and estimates uncertainty for each modality. These features, along with their uncertainty estimates, are then fed into the multimodal fusion module, where cross-modal attention and uncertainty-weighted gating mechanisms dynamically integrate the information to produce the final fatigue prediction. These two modules will be described in detail in Section 2.1 and Section 2.2.

2.1. Unimodal Feature Extraction Module

The unimodal feature extraction module is a core component of TMU-Net, designed to extract time-frequency features from each modality while quantifying prediction reliability through uncertainty modeling. This is achieved by integrating causal convolution, a convolutional sparse attention mechanism (ConvSparseAttention), and a Transformer encoder. Taking the EEG-2Hz modality as an example, the detailed architecture of the feature extractor is illustrated in Figure 2. The feature extractors for other modalities (e.g., EOG and EEG-5Bands) follow a similar structure, with only parameter configurations differing.
Let the input data of EEG-2Hz be denoted as x R B   ×   S   ×   D , where   B   represents the batch size, S denotes the time series length, and D indicates the feature dimension. The input data undergoes a dimension permutation, transforming from R B   ×   S   ×   D to R B   ×   D   ×   S , to accommodate the 1D convolution operation.
The input data undergoes local feature extraction through a two-layer 1D causal convolutional neural network (CNN). Each convolutional layer is followed by a ReLU activation function, and causal padding is applied to preserve the temporal causality of the sequence. After each convolutional layer, a ConvSparseAttention module is introduced to enhance the spatial and temporal representation capabilities of the features.
Compress the input channels into a single channel using a one-dimensional convolution (kernel_size = 1) to generate a feature map of shape [ B × 1 × S ] , which is then repeatedly expanded to [ B × C o u t × S ] .
Apply depthwise convolution ( g r o u p s = C o u t ) followed by pointwise convolution to enhance temporal interactions while preserving channel independence, with ReLU activation introducing nonlinearity.
The features are transposed to B × S × C o u t dimensions, followed by computation of attention weights through a multi-head attention mechanism. These weights are normalized via Softmax to generate attention-weighted features.
Features are extracted through the other convolutional pathway consisting of 1D convolution, batch normalization, and ReLU activation. These features undergo element-wise multiplication with the attention-weighted features, followed by integration with the original input features via residual connection.
Subsequently, a Transformer encoder processes the sequence, with a CLS token appended to aggregate global information. The CLS token output is split into two branches:
One branch generates the mean prediction via linear transformation:
μ 1 = W μ 1 x c l s + b μ 1
The other branch produces the logarithm of the variance through a linear transformation, followed by an exponential function to ensure non-negativity:
σ 1 2 = e W σ 1 x c l s + b σ 1
Under the Gaussian distribution assumption, the predicted value is generated as follows:
y ^ 1 = μ 1 + σ 1 2 · ε
where ε ~ N ( 0,1 ) is random noise introduced during training, following a standard normal distribution.
The unimodal feature extraction module of TMU-Net extracts time-frequency features from EEG and EOG signals while quantifying prediction reliability via uncertainty modeling. Causal convolution ensures temporal causality for real-time driver fatigue detection. ConvSparseAttention merges CNNs’ local pattern capture with attention mechanisms’ long-range dependency modeling for efficient feature extraction. The Transformer encoder captures global context. Uncertainty estimation enables dynamic weighting in the fusion module, enhancing robustness over static averaging.

2.2. Multimodal Feature Fusion Module

The multimodal fusion module integrates the predicted values, feature representations, and uncertainty estimates derived from EEG-2Hz, EOG, and EEG-5Bands modalities to generate the final fused prediction y ^ f u s i o n R B   ×   1 . This module achieves dynamic fusion through multi-scale feature extraction, cross-modal attention mechanisms, gating operations, and residual connections. The detailed structure of the module is illustrated in Figure 3.
The multimodal fusion module receives the outputs from the unimodal feature extractors, including the following:
The unimodal prediction values y ^ 1 ,   y ^ 2 ,   y ^ 3 , corresponding to the predictions from EEG-2Hz, EOG, and EEG-5Bands modalities, respectively.
The predictive uncertainty variances σ 1 2 , σ 2 2 , σ 3 2 , representing the uncertainty associated with each modality’s prediction.
The feature representations extracted from the CLS token outputs, denoted as f e a t u r e 1 ,   f e a t u r e 2 ,   f e a t u r e 3 .
To model the interactions between modalities, the CLS token features from each modality to a unified dimensional space of size d m o d e l through a linear transformation. The projected features are then stacked to form a tensor ( f e a t u r e s R B   ×   3   ×   d m o d e l ).
As shown in the Figure 4, for each modality-specific feature, query (Q), key (K), and value (V) vectors are generated as follows:
Q = f e a t u r e s · W Q + b Q ,   K = f e a t u r e s · W K + b K ,   V = f e a t u r e s · W V + b V
where W Q , W K , W V R d m o d e l   ×   d m o d e l and b Q , b K , b V R m o d e l , and the resulting tensors are Q , K , V R B   ×   3   ×   d m o d e l .
The Q , K , V   tensors are split into n h = 4 attention heads, with each head having dimensionality   d h = d m o d e l n h , Accordingly, the head-specific tensors are denoted as follows: Q h , K h , V h R B   ×   n h   ×   3   ×   d h .
For each attention head, the cross-modal attention score is computed as follows:
S h = Q h · K h T d h
where d h serves as a scaling factor to mitigate the numerical instability of high-dimensional dot-product operations.
The uncertainty of each modality is represented by its variance. Precision is defined as the inverse of the variance:
π i = 1 σ i 2 + ε ,     i = 1,2 , 3
where ε = 10 8 is a constant added to avoid division by zero, π i R B   ×   1 serves as the precision of modality i, with higher values indicating lower uncertainty.
The relative confidence weights are computed for each modality based on their precision and normalized as follows:
w i = π i j 3 π j ,     i = 1,2 , 3
where the normalization ensures i 3 w i = 1 , enhancing both interpretability and stability of the weights, with w i R B   ×   1 .
An initial confidence weight vector is constructed as follows:
W r a w = w 1 , w 2 , w 3 R B   ×   3
W r a w is mapped to the model dimension d m o d e l through a linear projection layer.
To align with the dimensionality requirements of the multi-head attention mechanism, we perform expansion and reshaping operations:
W = r e s h a p e r e p e a t W r a w , n h R B   ×   n h   ×   3   ×   d h
The projected weights are adapted for multi-head attention through dimensional transformations, modulating attention scores to prioritize high-confidence modalities:
S h = S h · s o f t m a x W p r o j · β , d i m = 1
where β = 0.5 serves as a scaling factor to control the intensity of confidence weights, and the softmax operation normalizes along the modality dimension.
The attention output is generated as follows:
O h = S h V h
where O h R B   ×   n h   ×   3   ×   d h incorporates both cross-modal interaction information and uncertainty weights:
O = C o n c a t O 1 , , O n h W O + b O
with learnable parameters W O R n h   ×   d h   ×   d m o d e l and   b O R d m o d e l to restore the model dimension d m o d e l .
Then, we perform modality-wise averaging to produce the fused feature representation:
f e a t u r e f u s e d = 1 3 i = 1 3 O : , i , :
The resulting f e a t u r e f u s e d R B   ×   d m o d e l integrates feature representations from all three modalities (EEG-2Hz, EOG, EEG-5Bands), combining cross-modal interactions with uncertainty-weighted information, which serves as input for subsequent gating mechanisms.
The multimodal fusion module simultaneously receives both the predicted values y ^ 1 , y ^ 2 , y ^ 3 from single-modality feature extractors and their uncertainty variances σ 1 2 , σ 2 2 , σ 3 2 , concatenating to form the input tensor: x f u s i o n = y ^ 1 , y ^ 2 , y ^ 3 , σ 1 2 , σ 2 2 , σ 3 2 R B   ×   6 .
This tensor is further processed by a multi-scale feature extractor to capture the complex relationships between the predicted values and their associated uncertainties.
As illustrated in Figure 5, the multi-scale feature extractor first projects the dimensionality from 6 to 32. LayerNorm is then applied to stabilize training and accelerate convergence. A second linear layer further projects the dimensionality to 64, followed by dropout with probability 0.3 to prevent overfitting:
This module extracts deep features of x f u s i o n through multi-scale transformations, providing rich representations x f u s i o n e n h a n c e d for subsequent gated fusion.
As illustrated in Figure 6, the gating mechanism dynamically adjusts the contributions of cross-modal attention outputs and multi-scale features. The input is formed by concatenating the two representations: G i n = [ x f u s i o n e n h a n c e d ,   f e a t u r e f u s e d ] R B   ×   2 d m o d e l
The gating weights are generated via a fully connected network, which first projects the dimensionality from 2 d m o d e l to d m o d e l and then applies a Sigmoid activation to produce normalized weights:
G = S i g m o i d W G ( R e L U W G G i n + b G ) + b G
With W G R 2 d m o d e l   ×   d m o d e l , b G R d m o d e l ,   W G R d m o d e l   ×   d m o d e l , b G R d m o d e l .
The final gated output is fused via element-wise multiplication:
G f i n a l = G f e a t u r e f u s e d
Here, G f i n a l R B   ×   d m o d e l further dynamically modulates the weights of the fused features through the gating mechanism, ensuring that higher-confidence modalities dominate the final representation.
The final prediction is generated by integrating two branches, which combine the fused features and the original input information: the fusion connection branch and the residual connection branch.
The fusion connection branch processes the output from the gating mechanism, denoted as G f i n a l .
First, the first linear layer maps the input dimension from d_model to 32 dimensions, followed by a ReLU activation function and a dropout layer with a dropout rate of 0.3 to obtain   h 1 .
The second linear layer then maps the dimension from 32 to a single dimension, generating the fusion prediction y ^ f u s e d
y ^ f u s e d R B   ×   1 denotes the prediction result obtained via the integration of cross-modal attention and gating mechanisms, which synthesizes features and uncertainty information from EEG-2Hz, EOG, and EEG-5Bands modalities.
This branch processes the initial input x f u s i o n of the multimodal fusion module, where x f u s i o n = [ y ^ 1 , y ^ 2 , y ^ 3 , σ 1 2 , σ 2 2 , σ 3 2 ] R B   ×   6 contains the single-modality prediction values and uncertainty variances.
The first linear layer maps the input dimension from 6 to 32, followed by a ReLU activation function to obtain h r . A normalization layer applies LayerNorm to stabilize training and enhance numerical stability.
The second linear layer maps the dimension from 32 to a single dimension, generating the residual prediction:
y ^ r e s i d u a l = W r 2 h r + b r 2
where W r 2 R 32   ×   1 ,   b r 2 R 1 are learnable parameters, and y ^ r e s i d u a l R B   ×   1   represents the direct contribution based on single-modality predictions and variances, retaining the original input information.
The outputs of the fusion connection branch and the residual connection branch are combined through addition to generate the final prediction:
y ^ f u s i o n = y ^ f u s e d + y ^ r e s i d u a l
The residual connection fuses the predictions from both branches by addition, preventing information loss and enhancing the model’s ability to model complex modality relationships.
The multimodal fusion module achieves dynamic fusion through multi-scale feature extraction, cross-modal attention, gating operations, and residual connections. Multi-scale feature extraction captures complex relationships between predictions and uncertainties, stabilizing training. Cross-modal attention models complementary inter-modal interactions, prioritizing high-confidence modalities. Gating operations dynamically adjust feature contributions, enhancing adaptability. Residual connections integrate unimodal and cross-modal information, preventing information loss. These designs ensure TMU-Net’s robust, high-accuracy fatigue prediction in driving scenarios.

2.3. Design of the Model Loss Function

To ensure that TMU-Net effectively integrates multimodal signals and enhances the accuracy and robustness of fatigue detection, we have designed a comprehensive loss function that combines the single-modality uncertainty loss, multimodal fusion loss, and auxiliary branch loss, aiming to optimize the feature extraction and dynamic fusion processes.
Single-Modality Uncertainty Loss: For each modality (EEG-2Hz, EOG, and EEG-5Bands), we use the negative log-likelihood (NLL) loss, combining the predicted mean and variance, while introducing a regularization term to prevent excessively large uncertainty:
L u n i m o d a l , i = 1 B b = 1 B 1 2 log σ b , i 2 + 1 2 μ b , i y b 2 σ b , i 2 + λ r e g σ b , i 2
where i { 1,2,3 } represents the modality, y ^ b , i is the predicted value for modality i , σ b , i 2 is the uncertainty variance, y b is the true label, and λ r e g = 0.001 is the regularization coefficient.
The total single-modality loss is the weighted sum of the three modalities, with weights dynamically computed based on uncertainty:
L u n i m o d a l = i = 1 3 w i L u n i m o d a l , i   ,     w i = π i j = 1 3 π j   ,     π i = 1 σ b , i 2 + ε   ,       ε = 10 8
The dynamic weight w i ensures that modalities with higher confidence (lower variance) contribute more to the loss.
Multimodal Fusion Loss: The final fused prediction y ^ f u s i o n is supervised using mean squared error (MSE):
L f u s i o n = 1 B b = 1 B y ^ b , f u s i o n y b 2
Auxiliary Branch Loss: The fusion branch ( y ^ f u s e d ) and the residual branch ( y ^ r e s i d u a l ), are supervised separately to ensure their effective collaboration:
L a u x = 1 B b = 1 B y ^ b , f u s e d y b 2 + y ^ b , r e s i d u a l y b 2
Total Loss Function: The total loss function is defined as a weighted sum of three components:
L t o t a l = α L u n i m o d a l + β L f u s i o n + γ L a u x
Here, α, β, and γ are hyperparameters set to 1.0, 1.0, and 0.5, respectively, and are tuned empirically through experiments.
This loss function is designed to collaboratively optimize model performance under multimodal uncertainty conditions. The unimodal uncertainty loss simultaneously optimizes predicted values and variances to prevent overconfidence, while dynamically weighting modalities to emphasize those with higher confidence. The multimodal fusion loss directly supervises the accuracy of the final output, and the auxiliary loss promotes branch collaboration via intermediate supervision. The total loss balances multiple optimization objectives through weight allocation, enhancing model accuracy and robustness.

3. Experiments

3.1. Dataset and Preprocessing

In this study, we employ the publicly available SEED-VIG dataset [17], developed by the research team led by Baoliang Lü at Shanghai Jiao Tong University. The dataset was collected using a simulated driving system designed by the team to synchronously acquire electroencephalography (EEG) and electrooculography (EOG) signals with high-precision labeling. A total of 23 subjects participated in the experiment, during which EEG signals, EOG signals, and PERCLOS values were recorded simultaneously under fatigue-induced driving conditions.
As shown in Figure 7, data acquisition was conducted in a realistic vehicle cockpit equipped with a large liquid crystal display. The screen presented a simplified four-lane highway scenario, with unnecessary elements such as engine components removed to minimize distraction. Subjects controlled the virtual vehicle using a steering wheel and accelerator, with the driving scene updated in real time based on their operations. The road was designed to be predominantly straight and monotonous to effectively induce driver fatigue. Each subject performed a continuous driving task for approximately 2 h to ensure the collection of high-quality physiological signals sufficient to induce fatigue.
As shown in Figure 8, the dataset includes EEG signals sampled at 200 Hz through 17 channels (following the international 10–20 system), as well as EOG signals captured from the forehead region. The EEG data provide two types of frequency-domain representations: full-frequency band features (1–50 Hz with 2 Hz resolution) and five sub-band features (Delta, Theta, Alpha, Beta, and Gamma bands), both including power spectral density (PSD) and differential entropy (DE). For the EOG signals, vertical and horizontal eye movement features were extracted using independent component analysis (ICA) and subtraction separation techniques. All data were annotated with fatigue levels based on the PERCLOS index (percentage of eyelid closure over the pupil per minute), which ranges from [0,1] and effectively reflects continuous changes in the subjects’ fatigue state.
In the data preprocessing stage, this study first performed systematic dimension adjustment and standardization on the raw physiological signals. Considering the characteristics of multimodal data, we placed the time-step dimension of EEG and EOG signals at the front and flattened the EEG frequency-domain features.
Eventually, three standardized data formats were formed: full-band EEG features (dimension: 20,355 × 4 × 425), EOG features (dimension: 20,355 × 3 × 36), and five-band EEG features (dimension: 20,355 × 4 × 85). To eliminate dimensional differences between channels and frequency bands, a channel-level Z-score standardization method was adopted, with the calculation formula given as follows: X ^ B , C , F = X B , C , F μ C , F σ C , F + ϵ , where μ C , F and σ C , F represent the mean and standard deviation of each channel-frequency band, and ϵ = 10 8 is used to ensure numerical stability.
In terms of experimental design, this study employs leave-one-out cross-validation. In each experiment, 885 samples from one subject are used as the test set, and the remaining 20,355 samples from the other 22 subjects serve as the training set. This approach is used to assess the model’s generalization ability to new subjects.

3.2. Parameter

In terms of experimental environment configuration, the hardware platform utilized an NVIDIA GeForce RTX 4090 GPU (equipped with 16,384 CUDA cores, 24 GB GDDR6X memory, and a 384-bit memory interface width), paired with a 20-core CPU and 80 GB RAM. The implementation was conducted on a Windows 10 Professional operating system using Python 3.6.13 and the PyTorch 1.10.2 framework. To assess the practical applicability of TMU-Net for real-time deployment, we evaluated its computational efficiency on this platform. The model demonstrated a low complexity of 0.003 GFLOPs. The average inference time for a single forward pass was 6.053 ms. When performing uncertainty quantification with 10 Monte Carlo samples, the average inference time was 9.408 ms, which is well below the 33.33 ms requirement for real-time processing at 30 FPS, confirming its strong potential for real-world deployment.
The unimodal feature extraction module adopts a hierarchical convolutional architecture to process multi-source physiological signals. For the three input modalities—EEG_2Hz (4 channels × 425 time-series points), EEG_5Bands (4 channels × 85 frequency-band features), and EOG (3 channels × 36 eye-movement features), they are first processed through a two-layer one-dimensional causal convolutional network. The convolution kernel size is set to 3, and a replication padding strategy is adopted (padding = 1). The first convolutional layer uniformly maps each modality to 16 channels, while the second layer performs differentiated processing: EEG signals are expanded to 64 channels, and EOG signals to 32 channels. To enhance the feature expression capability, a 4-head sparse convolutional attention mechanism (ConvSparseAttention) is introduced into the network. For temporal modeling, a 2-layer Transformer encoder is employed. The Transformer architecture for all modalities used 4 attention heads ( n h = 4). For EEG signals, the model dimension ( d m o d e l ) was set to 64, resulting in a key/query/value dimension of 16. For EOG signals, d m o d e l was set to 32, resulting in a key/query/value dimension of 8. The dimension of the feed-forward network within the Transformer was 128. All residual connections integrate LayerNorm and Dropout regularization (dropout rate p = 0.3).
In the multimodal fusion stage, features from each modality are first linearly projected into a 64-dimensional shared space for alignment. A three-head cross-modal attention mechanism (with a dropout rate of 0.1) is employed to establish dynamic inter-modal relationships. To achieve adaptive feature fusion, a two-layer gated MLP structure is designed, with both hidden layer dimensions of 64 and employing the Sigmoid activation function, combined with a variance-based weighting strategy (with a regularization coefficient λ = 0.001), incorporated to balance the contribution of each modality. Throughout the fusion process, residual connections are maintained, and LayerNorm is applied after each sub-module to stabilize the training procedure.
For model training, we employ the AdamW optimizer with differentiated learning rates for distinct modules: 4 × 10−4 for the EEG_2Hz module, 7 × 10−4 for the EEG_5Bands module, 3 × 10−4 for the EOG module, and 5 × 10−4 for the fusion module. The training batch size is fixed at 128, and training is conducted for 50 epochs. During inference, to evaluate model uncertainty, we performed 10 Monte Carlo sampling iterations (mc_samples = 10) while maintaining the dropout rate at 0.3. The final prediction was obtained by averaging the outputs of these 10 stochastic forward passes. The predictive uncertainty for each sample was quantified as the variance of these 10 outputs.

3.3. Experimental Results

3.3.1. Fatigue Detection Performance

As shown in Table 1, the model achieved an average RMSE of 0.150599 and an average MAE of 0.123802 on the SEED-VIG dataset, demonstrating high prediction accuracy. Subject 10 exhibited the lowest RMSE, reaching 0.046472, and the lowest MAE, reaching 0.033936, indicating that the model predicted fatigue states for this individual with exceptional accuracy. In contrast, Subject 13 recorded the highest RMSE at 0.361604 and the highest MAE at 0.330456, suggesting that the model’s generalization capability is limited for certain individuals. The relatively large prediction errors for these two subjects are closely related to the quality of the participants and their physiological characteristics. Some physiological signals contained artifacts, and the noise patterns varied across individuals, ultimately leading to increased prediction errors. The standard deviations of RMSE and MAE, which are 0.067982 and 0.063950, respectively, reflect the model’s stability across different subjects and further highlight the variability in performance between individuals.

3.3.2. Performance Comparison

As shown in Table 2, the proposed TMU-Net outperforms all baseline methods, achieving the best average RMSE of 0.1506 with a standard deviation of 0.0680. Compared to the best baseline DResNet, TMU-Net reduces RMSE by 4.01 percent and standard deviation by 7.48 percent, indicating improved accuracy and stability. This performance gain is attributed to its novel multimodal fusion architecture. TMU-Net leverages cross-modal attention to integrate complementary features from EEG_2Hz, EOG, and EEG_5Bands, enhancing sensitivity to fatigue changes. The uncertainty-based dynamic weighting improves robustness, while the ConvSparseAttention module significantly reduces computational cost without compromising key spatiotemporal feature extraction. Additionally, the Transformer-based unimodal encoder enables efficient global feature aggregation. These components work together to make TMU-Net highly effective for PERCLOS prediction, offering both high precision and strong robustness to individual variability and noise—critical for real-time driver fatigue monitoring.

3.3.3. Comparison Between Unimodal and Multimodal Approaches

As shown in Table 3, the single-modal methods for fatigue detection are based on EEG_2Hz, EOG, and EEG_5Bands. Among the three single-modal experiments, the EOG modality achieved the best performance, with an average RMSE of 0.158179 and MAE of 0.122523, and standard deviations of 0.082630 and 0.070814, respectively, indicating higher stability and accuracy in predicting PERCLOS values. This superior performance may be attributed to EOG data being directly derived from eye movement tracking, which is highly correlated with PERCLOS labels based on eyelid closure duration, thus effectively capturing fatigue-related ocular movement features. However, EOG performed poorly for certain subjects, such as subject 13 (RMSE: 0.445883, MAE: 0.385439), possibly due to the influence of lighting conditions or device calibration issues. The EEG_2Hz modality yielded an average RMSE of 0.226376 and MAE of 0.184766, with standard deviations of 0.106922 and 0.098614, respectively. EEG_2Hz provides high-resolution frequency features that can capture fine-grained variations in brain electrical activity. Nevertheless, its performance exhibited substantial inter-subject variability, such as subject 12 with a high RMSE of 0.593778 and MAE of 0.532934, which may be related to the susceptibility of EEG signals to noise or individual differences in brain activity patterns. The EEG_5Bands modality recorded the highest average RMSE (0.267976) and MAE (0.224698), with standard deviations of 0.131726 and 0.119818, respectively, indicating the lowest prediction accuracy and stability among single-modal methods. Although EEG_5Bands reduces feature dimensionality by utilizing five frequency bands, this may lead to the loss of critical frequency details, impairing the model’s ability to capture complex fatigue-related EEG patterns.
Dual-modal fusion methods, by combining features from two modalities, outperform their respective single-modal counterparts, demonstrating the advantage of complementary feature integration. According to Table 3, compared to single-modal EEG_2Hz (RMSE: 0.226376, MAE: 0.184766), the EEG_2Hz+EOG combination reduces RMSE and MAE by 31.73% and 32.52%, respectively, while EEG_2Hz+EEG_5Bands reduces RMSE and MAE by 1.96% and 1.84%. Compared to single-modal EEG_5Bands (RMSE: 0.267976, MAE: 0.224698), the EOG+EEG_5Bands combination reduces RMSE and MAE by 41.73% and 43.88%, respectively, while EEG_2Hz+EEG_5Bands reduces RMSE and MAE by 17.17% and 16.26%. Compared to single-modal EOG (RMSE: 0.158179, MAE: 0.122523), EEG_2Hz+EOG and EOG+EEG_5Bands reduce RMSE by 2.27% and 1.26%, respectively, while MAE increases slightly by 1.75% and 2.95%. However, the MAE standard deviations for EEG_2Hz+EOG (0.066955) and EOG+EEG_5Bands (0.068813) are lower than that of single-modal EOG (0.070814), indicating significantly improved prediction consistency. The advantage of dual-modal methods lies in integrating complementary information from brain electrical activity and eye movement signals. For instance, EEG_2Hz+EOG effectively mitigates EEG_2Hz’s sensitivity to noise and EOG’s dependence on external conditions, while EOG+EEG_5Bands leverages EOG’s high correlation to enhance EEG_5Bands’ predictive capability. Although EEG_2Hz+EEG_5Bands shows smaller improvements, it still enhances performance by combining two EEG feature sets. The lower standard deviations of dual-modal methods compared to single-modal methods indicate reduced prediction variability, which is critical for fatigue detection applications across diverse subjects.
The multimodal fusion approach, integrating features from EEG_2Hz, EOG, and EEG_5Bands via the MultiModalFusion module with a cross-modal attention mechanism and uncertainty modeling through Monte Carlo Dropout and epistemic and aleatoric uncertainty estimation, achieves the best overall performance, with an average RMSE of 0.150599, MAE of 0.123802, and standard deviations of 0.067982 and 0.063950. Compared to the best single-modal EOG, RMSE is reduced by 4.78%, while MAE increases slightly by 1.05%, but the MAE standard deviation decreases from 0.070814 to 0.063950. Compared to the best dual-modal EEG_2Hz+EOG, RMSE and MAE are reduced by 2.58% and 0.74%, respectively, with the MAE standard deviation further reduced to 0.063950. Compared to EEG_2Hz and EEG_5Bands, RMSE is reduced by 33.46% and 43.79%, and MAE by 33.00% and 44.92%, respectively. The low standard deviations of the multimodal approach indicate the highest prediction stability, effectively mitigating limitations of single- and dual-modal methods through comprehensive feature integration, demonstrating superior adaptability to diverse physiological and behavioral characteristics across subjects.

3.3.4. Temporal Prediction Results

Figure 9 presents a comparative analysis of multimodal and three unimodal prediction curves against the ground truth curve for the PERCLOS values of Subject 23. In Figure 9a, the multimodal fusion prediction curve exhibits a high degree of alignment with the ground truth in both trend and amplitude, particularly during fatigue peaks (sample points 200–600, PERCLOS > 0.6) and low-fatigue phases (sample points 400–800, PERCLOS < 0.2). This advantage stems from the MultiModalFusion module, which effectively integrates complementary information from EEG_2Hz, EOG, and EEG_5Bands through a cross-modal attention mechanism. Notably, the multimodal approach demonstrates greater sensitivity in responding to rapid changes in fatigue states, attributed to the robustness of uncertainty modeling in handling noise and dynamic variations. Although some deviations persist in certain peak and trough regions (e.g., an evident underestimation of the peak near sample point 450), the error remains relatively small compared to other model predictions. Figure 9b compares the prediction results of three unimodal models (EEG_2Hz: blue dashed line; EOG: green dashed line; EEG_5Bands: red dashed line). The EEG_2Hz prediction curve exhibits significant fluctuations and a tendency to overfit, closely aligning with some high-value points but deviating considerably elsewhere. The EOG curve is relatively stable but lacks sensitivity, underestimating many peaks. The EEG_5Bands curve, due to insufficient frequency band feature resolution, shows the largest fluctuations and severe oscillations with substantial noise, yielding poor overall fit despite occasional proximity to the ground truth. This experiment further confirms that the multimodal approach, with an RMSE of 0.150599 and an MAE of 0.123802, outperforms all unimodal models (EEG_2Hz: RMSE 0.226376; EOG: RMSE 0.158179; EEG_5Bands: RMSE 0.267976), highlighting its superior temporal resolution and prediction stability.

3.3.5. Comparative Study on PERCLOS Prediction

As shown in Table 4, the ablation study results demonstrate that the complete TMU-Net model achieves optimal performance in PERCLOS prediction, with an RMSE of 0.151 ± 0.068 and an MAE of 0.124 ± 0.064, surpassing all ablated configurations and confirming the significant contribution of synergistic interactions among modules to enhanced prediction accuracy. The convolutional sparse attention module, by integrating local convolution and global attention mechanisms during unimodal feature extraction, effectively captures complex spatiotemporal patterns in EEG signals at 2 Hz and across five bands, as well as EOG signals, which are critical for PERCLOS prediction. Removal of this module results in a substantial performance decline, with an RMSE of 0.179 ± 0.126 and an MAE of 0.145 ± 0.112, exhibiting the highest mean and standard deviation, underscoring its pivotal role in both prediction accuracy and stability. The absence of this mechanism impairs the model’s ability to model complex signals, leading to significantly increased prediction errors and variability, particularly under noisy conditions. The cross-modal attention module enhances multimodal interactions, such as synergy between EEG brain activity and EOG eye movement features, improving fusion effectiveness. Its removal causes a modest performance drop, with an RMSE of 0.157 ± 0.073 and an MAE of 0.129 ± 0.069, but the decline is mitigated by other fusion mechanisms, with a slightly elevated standard deviation indicating its secondary contribution to stability. The multi-scale feature extraction module, by integrating multimodal predictions and uncertainty information, strengthens feature representation during the fusion stage. Its removal yields performance comparable to cross-modal attention, with an RMSE of 0.154 ± 0.067 and an MAE of 0.126 ± 0.062, with a lower standard deviation suggesting improved stability through smoother feature distributions. The uncertainty weight calculation module, via dynamic weighting based on modal uncertainty, prioritizes reliable modalities, such as high-quality EOG signals, enhancing fusion robustness. Its removal leads to a performance decline, with an RMSE of 0.155 ± 0.075 and an MAE of 0.127 ± 0.070, with a slightly higher standard deviation reflecting its auxiliary role in stability, particularly when signal quality varies. The complete model, integrating all modules, extracts high-quality features in the unimodal stage and achieves efficient fusion in the multimodal stage, fully leveraging the spatiotemporal information and complementarity of multimodal signals to attain the lowest error and moderate stability. The results highlight the convolutional sparse attention module as the most critical, followed by cross-modal attention, multi-scale feature extraction, and uncertainty weight calculation. The complete model’s superior RMSE and MAE demonstrate robust predictive performance, providing theoretical support for the TMU-Net design and guidance for model optimization in applications such as driver fatigue monitoring.

3.3.6. Attention Mechanism Analysis

As shown in Figure 10, the model exhibits distinct variations in the allocation of attention weights across different physiological signal submodalities. Taking Subject 1 (Sample 115) as an example, the PERCLOS value is approximately 0.29, approaching or reaching the threshold for moderate fatigue [31], indicating a significant fatigue state during this period. In this state, the model assigns higher attention weights to EOG and EEG_2Hz, while the attention to EEG_5Bands is notably lower. This distribution pattern is highly consistent with the physiological mechanisms of fatigue development: in the early stages of fatigue, behavioral changes (such as increased blink frequency and slowed eye movements) typically precede central nervous system responses, making eye movement signals more effective in capturing critical cues of initial fatigue. The higher attention weight assigned to EOG during this stage demonstrates that the model effectively recognizes its diagnostic value in fatigue monitoring. Meanwhile, EEG_2Hz, corresponding to low-frequency brain activity (primarily in the theta band), is associated with the brainstem–thalamus-cortex network and serves as an electrophysiological marker of early central nervous fatigue. The increase in its attention weight reflects fluctuations in brain regulatory capacity. In contrast, EEG_5Bands, which integrates features across multiple frequency bands, does not yet exhibit sufficient discriminative power in the early stages of fatigue progression, resulting in lower attention allocation by the model. This attention mechanism not only demonstrates the model’s strong task-focused capability in its architecture but also physiologically reflects the fatigue evolution process in a scientifically interpretable manner, providing a robust and explainable basis for fatigue detection based on multimodal signals.

4. Discussion

The experimental results demonstrate that TMU-Net achieves superior driver fatigue detection performance on the SEED-VIG dataset, with an average RMSE of 0.1506 ± 0.0680 and MAE of 0.1238 ± 0.0640, outperforming baseline methods such as DResNet (RMSE: 0.1569 ± 0.0735) by reducing RMSE by 4.01% and standard deviation by 7.48%. These improvements stem from TMU-Net’s multimodal fusion architecture, which integrates EEG (2Hz full-band and five-band features) and EOG signals through cross-modal attention, uncertainty-weighted gating mechanisms, and ConvSparseAttention modules. By leveraging uncertainty estimation to dynamically adjust the contribution of each modality, the model enhances robustness to noise and inter-subject variability, addressing key limitations in signal quality fluctuations and cross-subject generalization identified in prior studies. Shi et al. [32] proposed a multimodal fusion approach based on time-space-frequency features, utilizing a convolutional autoencoder (CAE) to integrate EEG and EOG signals, achieving an RMSE of 0.08 and a correlation coefficient of 0.96 on the SEED-VIG dataset, and thereby validating the potential of multimodal fusion to enhance detection accuracy.
The comparison between unimodal and multimodal approaches highlights the synergistic advantages of multimodal fusion. Using EOG alone yields a competitive RMSE of 0.1582 ± 0.0826, but its performance degrades under conditions such as low lighting or when subjects wear glasses. EEG-based modalities (2Hz and five-band) provide complementary neurological information but are limited by higher variability (RMSE of 0.2264 ± 0.1069 for EEG_2Hz and 0.2680 ± 0.1317 for EEG_5Bands), likely due to motion artifacts and inter-subject physiological differences [16,25]. Consistent with findings by Alghanim et al. [33], TMU-Net’s multimodal fusion captures interactions between eye movement (EOG) and brain activity (EEG) through cross-modal attention mechanisms, achieving reductions in RMSE and MAE of 4.82% and 0.55%, respectively, compared to the best unimodal approach (EOG), and reductions of 33.46% and 43.79% in RMSE and 33.00% and 44.92% in MAE compared to EEG_2Hz and EEG_5Bands, respectively.
The ablation study further elucidates the contributions of TMU-Net’s components. The ConvSparseAttention module is critical for unimodal feature extraction, and its removal leads to a significant performance decline, underscoring its pivotal role in capturing spatiotemporal dependencies in EEG and EOG signals. The cross-modal attention and uncertainty-weighted gating mechanisms enhance fusion by dynamically prioritizing reliable modalities, as confirmed by attention heatmap analysis. These mechanisms address the limitations of static-weighting schemes noted in prior studies, improving robustness to signal noise. TMU-Net’s prediction standard deviation (0.0680) is lower than that of baseline methods such as SVR (0.2076) and MLP (0.1822), indicating improved cross-subject generalization and partially mitigating the 30–40% performance degradation for new subjects reported in earlier research. Similarly, Nemcová et al. [34] demonstrated that a multimodal neural networks integrating EEG, EOG, and facial image features via a coupling mechanism achieved over 95% classification accuracy on the DROZY dataset, further validating the efficacy of dynamic feature fusion in enhancing generalization.
Despite these advancements, TMU-Net exhibits certain limitations. The model shows considerable performance variability across subjects, with Subject 13 yielding an RMSE of 0.3616 and MAE of 0.3305, likely due to signal artifacts or physiological differences. Although uncertainty modeling mitigates some noise effects, its efficacy is limited in cases of severe signal degradation. Furthermore, the use of the SEED-VIG dataset, collected in a virtual reality environment, introduces constraints. The potential for simulator sickness among participants [35] could contaminate physiological signals (e.g., causing atypical eye movements or brain activity), thereby confounding fatigue assessment, and this factor was not monitored or controlled for in the original dataset. Several practical barriers remain for real-world deployment. Sensor discomfort from prolonged wear of EEG/EOG devices may hinder user adoption, necessitating research into dry-electrode or non-contact alternatives. Finally, the high cost and complexity of integrating multi-modal physiological sensors present scalability challenges.
Future research can explore the following directions: First, incorporating adversarial domain adaptation or transfer learning could further reduce the 30–40% performance gap across subjects by leveraging graph convolutional networks to enhance the cross-subject generalization of EEG signals. Second, optimizing TMU-Net’s computational efficiency through model pruning or lightweight attention mechanisms would facilitate real-time applications. Third, validating the model on diverse real-world driving datasets, such as the DROZY dataset used by Nemcová et al., which includes multimodal data under real sleep deprivation conditions, would strengthen its practical applicability. Future data collection efforts should also incorporate screening tools like the Simulator Sickness Questionnaire (SSQ) to identify and mitigate the impact of simulator sickness on data quality. Finally, integrating additional modalities, such as heart rate variability or vehicle dynamics parameters, could further improve detection accuracy by capturing a broader range of fatigue indicators. In conclusion, TMU-Net significantly advances driver fatigue detection through multimodal fusion, uncertainty quantification, and Transformer-based feature extraction. Its superior accuracy and robustness highlight its substantial potential for enhancing road safety via reliable real-time fatigue monitoring. Addressing the aforementioned limitations will further enhance its applicability in real-world driving scenarios.

5. Conclusions

This study proposes a novel driver fatigue detection framework, TMU-Net, which integrates EEG (2Hz full-band and five-band features) and EOG signals, combining cross-modal attention, uncertainty-weighted gating mechanisms, and a ConvSparseAttention module to significantly enhance the accuracy and robustness of fatigue detection. Compared to single-modal approaches, multi-modal fusion captures complementary information from EEG and EOG signals, substantially reducing prediction errors and improving robustness to noise and inter-subject variability through uncertainty modeling.
The core contribution of TMU-Net lies in its innovative architectural design. The ConvSparseAttention module effectively captures spatiotemporal features of physiological signals, the cross-modal attention mechanism integrates multi-modal information, and the uncertainty-weighted gating mechanism dynamically adjusts modal contributions to enhance adaptability to complex driving scenarios. Attention heatmap analysis further reveals the model’s prioritized focus on EOG and low-frequency EEG signals, which aligns closely with the physiological mechanisms of early fatigue. These features enable TMU-Net to demonstrate superior performance and stability in PERCLOS prediction tasks, offering a reliable solution for real-time driver fatigue monitoring with significant potential for road safety applications.
However, TMU-Net has room for improvement. Inter-subject performance variability (e.g., an RMSE of 0.3616 for Subject 13) indicates that the model’s robustness to extreme signal artifacts requires further enhancement, and its computational complexity limits real-time deployment in resource-constrained environments. Future research could improve cross-subject generalization by incorporating adversarial domain adaptation, optimizing the model architecture to reduce computational costs, and validating performance on real-world driving datasets. Additionally, integrating supplementary modalities, such as heart rate variability or vehicle dynamics parameters, could further enhance detection accuracy. Subsequent studies will focus on addressing these limitations to improve temporal coherence and classification accuracy.

Author Contributions

Conceptualization, X.X.; methodology, Y.Z. and X.X.; validation, Y.Z., X.X., Y.D. and N.Z.; investigation, Y.Z. and X.X.; data curation, Y.D. and N.Z.; writing—original draft, Y.Z.; writing—review and editing, Y.D. and N.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available at https://bcmi.sjtu.edu.cn/home/seed/index.html, accessed on 27 August 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. Global Status Report on Road Safety 2018; World Health Organization: Geneva, Switzerland, 2019. [Google Scholar]
  2. Jia, Z.; Xiao, H.; Ji, P. End-to-end fatigue driving EEG signal detection model based on improved temporal-graph convolution network. Comput. Biol. Med. 2023, 152, 106431. [Google Scholar] [CrossRef]
  3. Lv, C.; Nian, J.; Xu, Y.; Song, B. Compact vehicle driver fatigue recognition technology based on EEG signal. IEEE Trans. Intell. Transp.Syst. 2022, 23, 19753–19759. [Google Scholar] [CrossRef]
  4. Gao, D.; Wang, K.; Wang, M.; Zhou, J.; Zhang, Y. SFT-Net: A network for detecting fatigue from EEG signals by combining 4D feature flow and attention mechanism. IEEE J. Biomed. Health Inform. 2024, 28, 4444–4455. [Google Scholar] [CrossRef] [PubMed]
  5. Yang, L.; Guan, W.; Ma, R.; Li, X. Comparison among driving state prediction models for car-following condition based on EEG and driving features. Accident Anal. Prev. 2019, 133, 105296. [Google Scholar] [CrossRef] [PubMed]
  6. Qu, Y.; Hu, H.; Liu, J.; Zhang, Z.; Li, Y.; Ge, X. Driver state monitoring technology for conditionally automated vehicles: Review and future prospects. IEEE Trans. Instrum. Meas. 2023, 72, 3000920. [Google Scholar] [CrossRef]
  7. Duchowski, A.T. Eye Tracking Methodology: Theory and Practice, 2nd ed.; Springer: London, UK, 2017. [Google Scholar]
  8. Quddus, A.; Zandi, A.S.; Prest, L.; Comeau, F.J.E. Using long short term memory and convolutional neural networks for driver drowsiness detection. Accident Anal. Prev. 2021, 156, 106107. [Google Scholar] [CrossRef]
  9. Knapik, M.; Cyganek, B. Driver’s fatigue recognition based on yawn detection in thermal images. Neurocomputing 2019, 338, 274–292. [Google Scholar] [CrossRef]
  10. Min, J.; Cai, M.; Gou, C.; Xiong, C.; Yao, X. Fusion of forehead EEG with machine vision for real-time fatigue detection in an automatic processing pipeline. Neural Comput. Appl. 2023, 35, 8859–8872. [Google Scholar] [CrossRef]
  11. Dinges, D.F.; Grac, R. PERCLOS: A Valid Psychophysiological Measure of Alertness as Assessed by Psychomotor Vigilance. 1998. Available online: https://rosap.ntl.bts.gov/view/dot/113 (accessed on 30 July 2025).
  12. Wierwille, W.W.; Ellsworth, L.A. Evaluation of driver drowsiness by trained raters. Accident Anal. Prev. 1994, 26, 571–581. [Google Scholar] [CrossRef]
  13. Bergasa, L.M.; Nuevo, J.; Sotelo, M.A.; Barea, R.; Lopez, M.E. Real-time system for monitoring driver vigilance. IEEE Trans. Intell. Transp. Syst. 2006, 7, 63–77. [Google Scholar] [CrossRef]
  14. Ji, Q.; Zhu, Z.; Lan, P. Real-time nonintrusive monitoring and prediction of driver fatigue. IEEE Trans. Veh. Technol. 2004, 53, 1052–1068. [Google Scholar] [CrossRef]
  15. Garcia, I.; Bronte, S.; Bergasa, L.M.; Almazán, J.; Yebes, J. Vision-based drowsiness detection for real driving conditions. In Proceedings of the 2012 IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 618–623. [Google Scholar]
  16. Shahbakhti, M.; Beiramvand, M.; Rejer, I.; Augustyniak, P.; Broniec-Wójcik, A.; Wierzchon, M.; Marozas, V. Simultaneous eye blink characterization and elimination from low-channel prefrontal EEG signals enhances driver drowsiness detection. IEEE J. Biomed. Health Inform. 2022, 26, 1001–1012. [Google Scholar] [CrossRef] [PubMed]
  17. Zheng, W.L.; Lu, B.L. A multimodal approach to estimating vigilance using EEG and forehead EOG. J. Neural Eng. 2017, 14, 026017. [Google Scholar] [CrossRef] [PubMed]
  18. Zheng, W.L.; Zhu, J.W.; Lu, B.L. Multimodal vigilance estimation using deep learning. IEEE Trans. Cybern. 2022, 52, 5296–5308. [Google Scholar]
  19. Jiang, W.B.; Liu, X.H.; Zheng, W.L.; Lu, B.L. Multimodal adaptive emotion transformer with flexible modality inputs on a novel dataset with continuous labels. In Proceedings of the 31st ACM International Conference on Multimedia, New York, NY, USA, 29 October–3 November 2023; pp. 5975–5984. [Google Scholar]
  20. Luo, Y.; Lu, B.L. Wasserstein-distance-based multi-source adversarial domain adaptation for emotion recognition and vigilance estimation. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021; pp. 1424–1428. [Google Scholar]
  21. Cheng, X.; Wei, W.; Du, C.; Qiu, S.; Tian, S.; Ma, X.; He, H. VigilanceNet: Decouple intra- and inter-modality learning for multimodal vigilance estimation. In Proceedings of the 30th ACM International Conference on Multimedia, New York, NY, USA, 10–14 October 2022; pp. 209–217. [Google Scholar]
  22. Jiang, H.; Lu, Z. Visual Grounding for Object-Level Generalization in Reinforcement Learning. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 55–72. [Google Scholar]
  23. Huo, X.; Zheng, W.L.; Lu, B.L. Real-time driver fatigue detection using multi-modal physiological signals. IEEE Trans. Intell. Transp. Syst. 2020, 21, 3356–3368. [Google Scholar]
  24. Kendall, A.; Gal, Y. What uncertainties do we need in Bayesian deep learning for computer vision? Adv. Neural Inf. Process. Syst. 2017, 30, 5574–5584. [Google Scholar]
  25. Ma, B.Q.; Li, H.; Luo, Y.; Lu, B.L. Depersonalized cross-subject vigilance estimation with adversarial domain generalization. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  26. Ko, W.; Oh, K.; Jeon, E.; Suk, H. Vignet: A deep convolutional neural network for EEG-based driver vigilance estimation. In Proceedings of the 2020 8th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Republic of Korea, 26–28 February 2020; pp. 1–3. [Google Scholar]
  27. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  28. Zhu, Y.; Zhuang, F.; Wang, D. Aligning domain-specific distribution and classifier for cross-domain classification from multiple sources. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; pp. 5989–5996. [Google Scholar]
  29. Gou, M.; Yin, H.L.; Lu, B.L.; Zheng, W.L. Multi-modal adversarial regressive transformer for cross-subject fatigue detection. In Proceedings of the 2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 15–19 July 2024; pp. 1–4. [Google Scholar]
  30. Boqun, M. Cross-Subject Emotion Recognition and Fatigue Detection Based on Adversarial Domain Generalization. Master’s Thesis, Shanghai Jiao Tong University, Shanghai, China, 2019. [Google Scholar]
  31. Dinges, D.F.; Mallis, M.M.; Maislin, G.; Powell, J.W. Evaluation of Techniques for Ocular Measurement as an Index of Fatigue and as the Basis for Alertness Management. 1998. Available online: https://rosap.ntl.bts.gov/view/dot/2518 (accessed on 30 July 2025).
  32. Shi, J.X.; Wang, K. Fatigue driving detection method based on time-space-frequency features of multimodal signals. Biomed. Signal Process. Control 2023, 84, 85–93. [Google Scholar] [CrossRef]
  33. Alghanim, M.; Alotaibi, B.; Ahmad, I.; Dias, S.B.; Althobaiti, T.; Alzahrani, M.; Almutairi, B. A hybrid deep neural network approach to recognize driving fatigue based on EEG signals. Int. J. Intell. Syst. 2024, 2024, 4721863. [Google Scholar] [CrossRef]
  34. Nemcová, A.; Slaninová, K.; Justrová, A.; Šimková, K.; Smíšek, R.; Vítek, M. Optimized driver fatigue detection method using multimodal neural networks. Sci. Rep. 2025, 15, 55723. [Google Scholar] [CrossRef]
  35. Brooks, J.O.; Gangadharaiah, R.; Rosopa, E.B.; Pool, R.; Jenkins, C.; Rosopa, P.J.; Mims, L.; Schwambach, B.; Melnrick, K. Using the functional object detection—Advanced driving simulator scenario to examine task combinations and age-based performance differences: A case study. Appl. Sci. 2024, 14, 11892. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of TMU-Net.
Figure 1. The overall architecture of TMU-Net.
Sensors 25 05364 g001
Figure 2. Structural Diagram of 2 Hz EEG Feature Extractor.
Figure 2. Structural Diagram of 2 Hz EEG Feature Extractor.
Sensors 25 05364 g002
Figure 3. Structural Diagram of MultiModalFusion.
Figure 3. Structural Diagram of MultiModalFusion.
Sensors 25 05364 g003
Figure 4. Architecture of multi-head attention between two channels.
Figure 4. Architecture of multi-head attention between two channels.
Sensors 25 05364 g004
Figure 5. Architecture of Multi-Scale Feature Extraction.
Figure 5. Architecture of Multi-Scale Feature Extraction.
Sensors 25 05364 g005
Figure 6. Illustration of Gating Mechanism.
Figure 6. Illustration of Gating Mechanism.
Sensors 25 05364 g006
Figure 7. SEED-VIG dataset acquisition equipment and environment schematic: (a) Urban road scene constructed in virtual reality; (b) Neuroscan system for EEG and EOG signal acquisition with SMI eye-tracking glasses for gaze monitoring; and (c) Real vehicle equipped with basic driving controls.
Figure 7. SEED-VIG dataset acquisition equipment and environment schematic: (a) Urban road scene constructed in virtual reality; (b) Neuroscan system for EEG and EOG signal acquisition with SMI eye-tracking glasses for gaze monitoring; and (c) Real vehicle equipped with basic driving controls.
Sensors 25 05364 g007
Figure 8. The electrode placement for EEG and EOG signals in the SEED-VIG dataset follows standard biosignal acquisition protocols.
Figure 8. The electrode placement for EEG and EOG signals in the SEED-VIG dataset follows standard biosignal acquisition protocols.
Sensors 25 05364 g008
Figure 9. Comparison Between Predicted and True PERCLOS Over Time (Fold 23): (a) Multimodal fusion prediction; and (b) Individual Predictions of Unimodal EEG_2Hz, EOG, and EEG_5Bands.
Figure 9. Comparison Between Predicted and True PERCLOS Over Time (Fold 23): (a) Multimodal fusion prediction; and (b) Individual Predictions of Unimodal EEG_2Hz, EOG, and EEG_5Bands.
Sensors 25 05364 g009
Figure 10. Fusion attention heatmap of Subject 1 (Sample 115).
Figure 10. Fusion attention heatmap of Subject 1 (Sample 115).
Sensors 25 05364 g010
Table 1. Performance evaluation using LOOCV.
Table 1. Performance evaluation using LOOCV.
IndexRMSEMAE
10.1189390.096812
20.1100410.089023
30.2303040.191571
40.1698180.143578
50.1012540.076963
60.1282470.098166
70.1167460.087638
80.1611940.11413
90.0977680.076226
100.0464720.033936
110.1943150.165127
120.2553150.222368
130.3616040.330456
140.1573390.103876
150.0664190.048681
160.1499740.126567
170.1985230.168296
180.1381140.121762
190.0736810.05776
200.1406260.112973
210.1355390.121048
220.180230.159634
230.1313080.100847
Avg0.1505990.123802
Std0.0679820.06395
Table 2. Performance comparison with previous methods on SEED_VIG.
Table 2. Performance comparison with previous methods on SEED_VIG.
MethodsRMSE
SVR0.2071 ± 0.2076
MLP0.1942 ± 0.1822
VIGNet [26]0.1876 ± 0.1903
DG-DANN [27]0.1543 ± 0.1490
MFSAN [28]0.1780 ± 0.1559
MART (w/o Domain Generalization) [29]0.1556 ± 0.1311
DResNet [30]0.1569 ± 0.0735
TMU-Net0.1506 ± 0.0680
Table 3. Performance comparison of unimodal versus multimodal fusion approaches on SEED_VIG.
Table 3. Performance comparison of unimodal versus multimodal fusion approaches on SEED_VIG.
ModalityRMSEMAE
EEG_2HZ0.226376 ± 0.1069220.184766 ± 0.098614
EOG0.158179 ± 0.082630.122523 ± 0.070814
EEG_5BANDS0.267976 ± 0.1317260.224698 ± 0.119818
EEG_2HZ+EOG0.154578 ± 0.0729350.124668 ± 0.066955
EEG_2HZ+ EEG_5BANDS0.221911 ± 0.0898780.188159 ± 0.083861
EOG+ EEG_5BANDS0.156201 ± 0.0706160.126112 ± 0.068813
All Modalities0.150599 ± 0.0679820.123802 ± 0.063950
Table 4. Performance impact of removing key components in TMU-Net.
Table 4. Performance impact of removing key components in TMU-Net.
Model ConfigurationRMSEMAE
No_ Cross-Modal Attention0.157 ± 0.0730.129 ± 0.069
No_ ConvSparseAttention0.179 ± 0.1260.145 ± 0.112
No_ Multi-Scale Feature Extraction0.154 ± 0.0670.126 ± 0.062
No_ Uncertainty Weight Calculation0.155 ± 0.0750.127 ± 0.070
Full Mode0.151 ± 0.0680.124 ± 0.064
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Xu, X.; Du, Y.; Zhang, N. TMU-Net: A Transformer-Based Multimodal Framework with Uncertainty Quantification for Driver Fatigue Detection. Sensors 2025, 25, 5364. https://doi.org/10.3390/s25175364

AMA Style

Zhang Y, Xu X, Du Y, Zhang N. TMU-Net: A Transformer-Based Multimodal Framework with Uncertainty Quantification for Driver Fatigue Detection. Sensors. 2025; 25(17):5364. https://doi.org/10.3390/s25175364

Chicago/Turabian Style

Zhang, Yaxin, Xuegang Xu, Yuetao Du, and Ningchao Zhang. 2025. "TMU-Net: A Transformer-Based Multimodal Framework with Uncertainty Quantification for Driver Fatigue Detection" Sensors 25, no. 17: 5364. https://doi.org/10.3390/s25175364

APA Style

Zhang, Y., Xu, X., Du, Y., & Zhang, N. (2025). TMU-Net: A Transformer-Based Multimodal Framework with Uncertainty Quantification for Driver Fatigue Detection. Sensors, 25(17), 5364. https://doi.org/10.3390/s25175364

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop