You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

5 September 2025

A Novel Hybrid Approach for Drowsiness Detection Using EEG Scalograms to Overcome Inter-Subject Variability

,
,
,
and
1
Service d’électronique et de Microélectronique, University of Mons, 7000 Mons, Belgium
2
Research Laboratory of Technology and Medical Imaging-LTIM-LR12ES06, Faculty of Medicine of Monastir, Monastir 5000, Tunisia
3
National Engineering School of Sousse, University of Sousse, Sousse 4054, Tunisia
4
Laboratory of Electronics and Microelectronics, Faculty of Sciences of Monastir, Monastir 5019, Tunisia
This article belongs to the Section Biomedical Sensors

Abstract

Drowsiness constitutes a significant risk factor in diverse occupational settings, including healthcare, industry, construction, and transportation, contributing to accidents, injuries, and fatalities. Electroencephalography (EEG) signals, offering direct measurements of brain activity, have emerged as a promising modality for drowsiness detection. However, the inherent non-stationary nature of EEG signals, coupled with substantial inter-subject variability, presents considerable challenges for reliable drowsiness detection. To address these challenges, this paper proposes a hybrid approach combining convolutional neural networks (CNNs), which excel at feature extraction, and support vector machines (SVMs) for drowsiness detection. The framework consists of two modules: a CNN for feature extraction from EEG scalograms generated by the Continuous Wavelet Transform (CWT), and an SVM for classification. The proposed approach is compared with 1D CNNs (using raw EEG signals) and transfer learning models such as VGG16 and ResNet50 to identify the most effective method for minimizing inter-subject variability and improving detection accuracy. Experimental evaluations, conducted on the publicly available DROZY EEG dataset, show that the CNN-SVM model, utilizing 2D scalograms, achieves an accuracy of 98.33%, outperforming both 1D CNNs and transfer learning models. These findings highlight the effectiveness of the hybrid CNN-SVM approach for robust and accurate drowsiness detection using EEG, offering significant potential for enhancing safety in high-risk work environments.

1. Introduction

Vigilance is defined as the state of optimal reactivity of the organism to external or internal stimuli, encompassing the ability to maintain attention and respond appropriately to various situations over an extended period [1]. This alertness is essential for cognitive performance, safety, and productivity, and it is influenced by internal factors such as fatigue and external factors such as the environment. Reduced alertness, often caused by insufficient sleep, poor sleep quality, circadian rhythm disruptions, or medical conditions [2], impairs attention and increases the risk of errors, slower reactions, and accidents [3].
Human vigilance is primarily divided into four states: deep sleep [4]; light sleep [5]; active wakefulness [6]; and passive wakefulness, also known as drowsiness [7]. Drowsiness is characterized by a decrease in alertness and an increased tendency to relax. Although this state typically lasts only a few minutes, its consequences can be disastrous in many professional sectors, such as medicine, construction, industry, and transportation. Drowsiness can lead to serious accidents, which is why it is so important to detect and prevent it effectively.
Drowsiness detection [8] relies on various methods, each with its advantages and limitations. Traditional methods include self-reported questionnaires [9], behavioral observation, and the use of cameras to analyze facial expressions or eye movements [10]. While these methods are useful, they often lack precision and objectivity. For example, facial expression-based methods analyze facial features such as eyelid opening, yawning, and eye movements to detect drowsiness. These methods can be affected by lighting conditions, visual obstructions, or distractions, and may not detect early signs of drowsiness before they become apparent.
Analysis of EEG signals [11], on the other hand, offers an advanced and effective method of detecting drowsiness. EEG measures the brain’s electrical activity via electrodes placed on the scalp, providing a direct perspective on the brain’s state. This method enables the identification of subtle drowsiness-related changes in brain activity through EEG, often before visible signs emerge [12,13]. EEG is particularly effective for early drowsiness detection as it captures neural changes that typically precede observable symptoms. Moreover, EEG provides objective and quantifiable data, unlike observation-based or self-reported methods. EEG-based systems can then monitor alertness in real time, enabling rapid intervention, which is crucial in critical environments such as driving or operating machinery.
EEG signals, while highly effective for detecting drowsiness, are extremely sensitive to artifacts caused by muscle movements and power line interferences. This sensitivity necessitates a rigorous artifact elimination process to ensure data quality. The size of EEG segments is also a critical parameter for accurate and reliable detection, as it directly impacts the effectiveness of feature extraction.
The frequency characteristics of EEG signals are valuable markers for the detection of drowsiness [14]. However, the large variability between subjects poses problems, leading to substantial differences in indicators of drowsiness between individuals. Temporal characteristics further contribute to understanding vigilance states, making the integration of time-frequency features particularly advantageous. Time-frequency representations have the potential to mitigate the impact of inter-subject variability, thereby enhancing detection accuracy. In this context, the Continuous Wavelet Transform represents a highly effective method for generating time-frequency representations of EEG signals, producing scalograms that are particularly useful for advanced analysis and classification [15].
With the advent of artificial intelligence (AI) [16], numerous AI-based drowsiness detection approaches using EEG signals have emerged. Several studies focus on classical machine learning (ML) tools [17], such as SVMs, due to their ability to handle the non-linearity of EEG data and their classification performance. However, classical ML tools are limited in their capacity to manage high-dimensional data and extract features.
A growing body of research is exploring the possibility of using deep learning (DL) tools for drowsiness detection from EEG signals [18], particularly convolutional neural networks [19]. Originally designed to process images, CNNs have demonstrated exceptional performance due to their ability to extract complex and relevant features from visual data.
SVMs excel in classification and handling EEG data non-linearity, while CNNs are superior in feature extraction and processing high-dimensional data like EEG scalograms [20]. A hybrid CNN-SVM approach, combining both strengths, would make the detection of drowsiness from EEG signals particularly effective.
This study proposes a generalized approach for EEG-based drowsiness detection that addresses inter-subject variability thanks to the rich time-frequency information provided by scalograms and the effectiveness of the CNN-SVM combination. By taking advantage of the richness of the data provided by scalograms, the computational efficiency of CNN-based feature extraction, and the classification capability of SVMs, this approach aims to significantly improve the accuracy and reliability of sleepiness detection, even in the presence of significant individual differences between subjects.
The key contributions of this work are as follows:
  • We propose a hybrid CNN-SVM model that combines a CNN for automatic feature extraction with an SVM for robust classification, improving drowsiness detection accuracy.
  • We utilize EEG scalograms to extract meaningful time-frequency features, preserving both temporal and frequency information for enhanced detection performance.
  • We mitigate inter-subject variability by designing a model that learns generalized patterns across different individuals, ensuring consistent and reliable drowsiness detection.
This paper is structured as follows: Section 2 provides an overview of the latest developments related to this research. In Section 3, we develop the proposed methodology, providing a detailed explanation. Section 4 presents the experimental results on which we build a detailed discussion. Finally, Section 5 provides a concise conclusion and explores future orientations for this study.

3. Methodology

This study focuses on the detection of drowsiness using EEG signals. To achieve this goal, our approach begins by applying a rigorous pipeline to remove unwanted artifacts from EEG segments. Next, these segments are converted into scalograms using CWT. The resulting scalograms, which are rich in time-frequency information, are used as input to a 2D CNN. The CNN is responsible for automatically extracting distinctive features from the scalograms. Finally, an SVM classifier is used to detect drowsiness states. Figure 1 and Figure 2 illustrate in detail the various stages of the proposed method for detecting drowsiness.
Figure 1. The suggested drowsiness detection workflow.
Figure 2. Comprehensive overview of features extraction and classification of the proposed method.

3.1. DROZY Database

The selection of an appropriate database is a crucial factor in the development of robust drowsiness detection systems. While numerous publicly available datasets predominantly focus on sleep onset, this study targets the detection of drowsiness as a distinct state. Accordingly, the ULg Multimodality Drowsiness Database (DROZY) [32] was utilized. DROZY includes high-quality recordings from five EEG channels (Fz, Cz, C3, C4, and Pz) stored in EDF format and sampled at 512 Hz. This dataset is specifically designed to investigate vigilance states in 14 healthy participants, all screened to exclude drug use, alcohol consumption, and sleep disorders.
The data collection protocol required participants to perform a Psychomotor Vigilance Task (PVT) three times across two consecutive days, amounting to 28.30 hours of sustained wakefulness. For each PVT session, 10 minutes of EEG signal recordings are collected per subject. These sessions were strategically distributed across morning, afternoon, and nighttime periods to capture variations in vigilance levels. The experimental protocol is comprehensively outlined in Figure 3. Electrodes placement adhered to the internationally standardized 10–20 system [33], as depicted in Figure 4, ensuring accurate and reliable EEG recordings. Following each PVT session, participants were asked to report their subjective vigilance levels using the Karolinska Sleepiness Scale (KSS) [34]. The KSS, a nine-point scale widely used to assess sleepiness, provides a granular representation of vigilance states.
Figure 3. Protocol for collecting data per subject over 2 days, organized by activity and PVT (Psychomotor Vigilance Task) sessions [32].
Figure 4. Distribution and Labeling of EEG Electrodes (10–20 Standard) with Highlighted Positions for the DROZY Database (Fz, Cz, C3, C4, Pz in Blue Circles) [33].
In this study, we just aim to detect drowsiness, so we do not need to differentiate specific intermediate vigilance levels. Therefore, levels 0, 1, 2, and 3 are grouped as level 0 (alert), while levels 4, 5, 6, 7, 8, and 9 are classified as level 1 (drowsy). Figure 5 displays the EEG signals for both, an alert and a drowsy state individual.
Figure 5. EEG signals for an alert (a) and a drowsy (b) individual [33].

3.2. Preprocessing

For drowsiness detection using EEG signals, the preprocessing phase is crucial [35], as it significantly affects both the performance and the accuracy of detection. Indeed, EEG signals are particularly sensitive to artifacts [36], which can be either physiological or non-physiological. To minimize the effect of these artifacts, several removal techniques have been developed. Many studies have demonstrated that effective drowsiness detection can be achieved with band-pass filters operating within the [0.1–30] Hz range. We, therefore, applied a band pass FIR filter with the same frequency range to clean up the EEG signals [37,38,39].
The segmentation phase of EEG signals plays a crucial role in the accuracy of drowsiness detection systems [40]. The segment length must be carefully chosen to balance temporal resolution with the ability to capture meaningful physiological patterns associated with the transition from alertness to drowsiness. Since the shift between wakefulness and drowsiness is typically gradual rather than abrupt, researchers commonly use 30-s epochs, as this duration effectively captures the evolving brain activity during the transition phase. A segment that is too short may fail to reflect these slow physiological changes, while longer segments risk reducing temporal responsiveness and increasing computational demands. Prior studies have demonstrated that 30-s segments provide an optimal trade-off between information richness and detection reliability [41]. Accordingly, in this study, the filtered EEG signals were segmented into 30-s windows to ensure accurate and efficient analysis [42].

3.3. Continuous Wavelet Transform

The time-frequency domain [43] offers an integrated approach for analyzing EEG signals in drowsiness detection by combining the temporal and spectral characteristics of the data. Unlike separate time-domain or frequency-domain analyses, it provides a comprehensive exploration of signal variations. The CWT [44], a robust time-frequency analysis method, represents a signal (t) as contributions at different time and frequency scales. The CWT is defined as Equation (1):
W ( l , n ) = x ( t ) 1 | l | ψ * t n l d t
where  W ( l , n ) represents the wavelet coefficients, l controls the scale (frequency), n defines the temporal position relative to time t, and  ψ * is the complex conjugate of the mother wavelet  ψ ( t ) . This approach enables a multi-resolution analysis, critical for detecting rapid transitions between wakefulness and drowsiness [45]. The resulting scalograms [46], representing graphically the intensity of the transform, facilitate the visualization of frequency variations associated with vigilance states.
Compared to the Short-Time Fourier Transform (STFT) [47], which uses fixed-size windows and uniform time-frequency resolution, the Wavelet Transform (WT) employs scalable mother wavelets, providing long windows for low frequencies and short windows for high frequencies. This adaptability makes CWT particularly effective for non-stationary EEG signals, allowing the precise localization of events in both the time and frequency domains. In this study, the Morlet wavelet [48] is applied to generate scalogram images, as illustrated in Figure 6 for alert and drowsy subjects. A total of 1400 EEG scalogram images were generated across the five channels for all 14 subjects.
Figure 6. Scalogram generation with continuous Morlet wavelet.

3.4. Image Resizing

The scalogram images generated by the CWT applied to EEG data initially have a resolution of 662 × 536 pixels. To adapt these images for analysis with a CNN, they are resized to 256 × 256 pixels using cubic interpolation [49].
This resizing choice was determined after experimenting with various dimensions, with 256 × 256 offering an optimum balance between accuracy and computational efficiency. First of all, CNNs need uniform input dimensions for consistent processing, and resizing to 256 × 256 pixels guarantees this uniformity. Moreover, this resolution optimizes computational complexity, enabling efficient model training while maintaining a reasonable demand on resources. Cubic interpolation is used to effectively preserve essential scalogram details and features, ensuring high-quality input data. Lastly, the resized dimensions facilitate data normalization and significantly improve the convergence and overall performance of the CNN, demonstrating its suitability for this application.

3.5. CNN

Convolutional layers [50] extract image features by applying filters (kernels) that detect patterns like edges, shapes, and textures through convolution operations. Each filter generates a feature map, highlighting specific image attributes. Key parameters include filter size (commonly 3 × 3 or 5 × 5 pixels) and the number of filters, which determine the diversity and richness of extracted features. Networks typically use dozens, if not hundreds, of filters per layer, enabling both simple and complex patterns to be efficiently detected.
In this work, we compare a basic 2D CNN with a 1D CNN to highlight the effectiveness of the proposed approach. Additionally, the proposed method is benchmarked against the most commonly used transfer learning models in drowsiness detection, namely, VGG16 and ResNet50, to validate its performance and reliability.
The proposed CNN architecture for feature extraction is implemented using the Keras library, designed with a focus on simplicity and efficiency. The model begins with a convolutional layer [47] comprising 16 filters, a kernel size of (3, 3), and a ReLU activation function [51], applied to input images with dimensions of (64, 64, 3). This layer is followed by a MaxPooling2D operation with a pool size of (2, 2) [52], which reduces the spatial dimensions, and a Dropout layer with a rate of 0.25 to prevent overfitting by randomly deactivating 25% of the neurons during training [53]. The second convolutional layer expands the feature extraction capacity, utilizing 64 filters with the same kernel size and activation function as the first layer. It is similarly followed by a MaxPooling2D layer with a pool size of (2, 2) and another Dropout layer with a 0.25 dropout rate. The extracted feature maps from the convolutional layers are then flattened into a one-dimensional vector using a Flatten layer [54], which serves as input to a fully connected dense layer containing 128 neurons [55] and a ReLU activation function. To further improve generalization, a second Dropout layer with a rate of 0.5 is applied, deactivating 50% of the neurons.

3.5.1. General Hyperparameters

Before detailing each model’s architecture, we first summarize the common hyperparameters applied to all implemented methods. These shared settings ensured a fair training procedure and a consistent evaluation across different models, as shown in Table 1.
Table 1. General hyperparameters used across all models.

3.5.2. Model-Specific Hyperparameters

In addition to the general training settings, each architecture required specific configurations aligned with its design. These details ensured optimal performance and a fair comparison between generic models (i.e., VGG16, ResNet50) and EEG-oriented architectures (i.e., DeepConvNet, EEGNet, ShallowConvNet), as shown in Table 2.
Table 2. Architecture-specific hyperparameters for each model.

3.6. Classification

For the classification phase, we use support vector machines because of their high performance in binary classification and their ability to handle non-linearity.
SVMs are a widely used supervised learning method for classification [56]. They work by identifying the optimal hyperplane that maximizes the separation between classes in the feature space. The equation for the hyperplane can be expressed as shown in Equation (2):
ω · x + b = 0
The goal of the SVM is to maximize the margin between the classes. Mathematically, this is translated into minimizing Equation (3):
1 2 ω 2
subject to the constraint that (4)
y i ( ω · x i + b ) 1
for all training data  { ( x i , y i ) } i = 1 N , where  y i represents the class labels.
To handle non-linear data, SVMs use kernel functions such as the linear, polynomial, Gaussian, and Radial basis function (RBF) kernels [57]. These kernels map the data into a higher-dimensional feature space where linear separation becomes feasible. In this context, the optimization problem is formulated as minimizing Equation (5):
1 2 ω 2 + C i = 1 N ξ i
subject to the constraints in Equation (6):
y i ( ω T φ ( x i ) + b ) 1 ξ i , ξ i 0
where  φ ( x i ) represents the transformation of the data via the kernel function, and  ξ i are the slack variables that allow for margin violations.
The parameter C balances the trade-off between maximizing the margin and minimizing classification error. A higher C emphasizes minimizing classification errors, while a lower C prioritizes a larger margin, potentially allowing some misclassifications.
SVMs are particularly well-suited for the classification of EEG data because they can handle the complexity and non-linearity of EEG signals using appropriate kernels. EEG signals, by nature, exhibit complex and non-linear patterns, making SVMs ideal for this task. By using kernels such as the RBF kernel, SVMs can project EEG data into a feature space where differences between alert and drowsy states become more apparent and linearly separable. This enables SVMs to identify subtle and complex patterns in EEG signals, offering high performance in binary classification.

3.7. Performance Evaluation

To evaluate the classification performance of the detection approach in this work, a binary confusion matrix metric is used [58]. Figure 7 illustrates this confusion matrix. The matrix provides a visual representation of the model’s performance by presenting the number of true positives, false positives, true negatives, and false negatives:
Figure 7. Confusion matrix of binary classification.
  • True positive (TP): prediction of drowsiness when the actual state is drowsiness;
  • False positive (FP): prediction of drowsiness when the actual state is alertness;
  • True negative (TN): prediction of alertness when the actual state is alertness;
  • False negative (FN): prediction of alertness when the actual state is drowsiness.
These elements are essential for calculating various performance measures such as accuracy, precision, specificity, sensitivity, and F-score, thereby offering a comprehensive assessment of the effectiveness of the proposed detection model. The different performance metrics are presented in Table 3.
Table 3. Mathematical formulations of performance metrics.
The DROZY database, as previously mentioned, consists mainly of EEG signals recorded from five electrodes: Fz, Cz, C3, C4, and Pz. In this study, our goal was to develop an optimized approach suitable for integration into an embedded system for drowsiness detection. Based on our recent experiments and previous work, we evaluated the classification accuracy obtained from each individual electrode in the DROZY dataset. We found that the C4 electrode provided the highest accuracy in detecting drowsiness, followed closely by C3 [59]. However, to avoid relying on a single EEG derivation, which could lead to system malfunction in the case of electrode failure or poor contact with the scalp, we decided to test the combination of C3 and C4 electrodes. Our results demonstrated that the use of both electrodes improved classification performance compared to using either one alone [60]. Therefore, we chose to continue with these two electrodes (C3 and C4), as they offer a good trade-off between performance and system simplicity. This selection is also consistent with our objective of designing an approach that is adaptable to the constraints of embedded system implementation, where minimizing the number of electrodes is essential for practicality and user comfort. By combining data from these two electrodes across all subjects, a total of 3360 scalograms were obtained for the wakefulness state and 1680 scalograms for the drowsiness state. To rebalance the dataset before training, a data augmentation technique was applied to the scalograms associated with the drowsiness state. This step resulted in a final balanced dataset composed of 6720 scalograms, evenly distributed between the two classes: 3360 for the wakefulness state and 3360 for the drowsiness state. The objective of this experiment is to evaluate the robustness of the approach with respect to inter-subject variability and to test its generalization capability. To this end, a combined data split was adopted, with 70% used for training and 30% for testing.

4. Results and Discussion

In this section, we present the results of the evaluation of drowsiness detection in the inter-subject mode using the proposed approach. This evaluation aims to verify the generalizability of the model by testing its performance on subjects outside the training set. Results are analyzed in terms of performance metrics, such as accuracy, sensitivity, specificity, precision, and F-score. These metrics will allow us to assess the robustness and effectiveness of the approach under inter-subject conditions, where physiological variability is a critical factor. These results will help to validate the approach’s feasibility in practical applications to detect drowsiness in a real-life environment.

4.1. Influence of CNN Depth on Drowsiness Detection Accuracy

In this part, we will explore how convolutional depth influences the accuracy of drowsiness detection models. The objective is to assess how varying the number of layers (either increasing or decreasing) affects the model’s ability to extract relevant features from images and, consequently, its overall performance. We will evaluate the results as a function of the number of layers, focusing on the strengths and weaknesses of the different CNN setups. Table 4 presents the performance outcomes for the various CNN configurations, where  C N N 1 consists of one convolutional layer,  C N N 2 has two layers,  C N N 3 includes three layers,  C N N 5 is composed of five convolutional layers, and  C N N 10 features a deeper architecture with 10 convolutional layers.
Table 4. Performance metrics for each CNN model.
The results indicated that the  C N N 3 model, which consists of three convolutional layers, demonstrates the highest efficiency in extracting features relevant to drowsiness detection.

4.2. Combined CNN-SVM Model

In order to improve the classification of our drowsiness detection system, we apply the SVM technique to the features extracted by the CNN. To determine the optimal SVM kernel, we evaluate multiple options, including linear, polynomial, RBF, and sigmoid kernels. Table 5 summarizes the classification results for each of these SVM kernels.
Table 5. Performance metrics for each SVM kernel.
The results indicate that the RBF kernel was the most effective among those tested (more than 2% improvement). While the linear, polynomial, and sigmoid kernels also performed well, the RBF kernel achieved an impressive accuracy of 98.33%, distinguishing itself through its ability to model non-linear decision boundaries. This is particularly important for detecting drowsiness, where EEG signal relationships tend to be complex and non-linear. Furthermore, the RBF kernel demonstrates superior generalization on test data, reducing the likelihood of overfitting. Its flexibility enables it to pick up subtle variations in the signal, improving the accuracy of drowsiness detection. The SVM-RBF hyperparameters, penalty factor C = 1 and kernel factor  γ = 0.4 , were selected using the GridSearch optimization tool.
Moreover, replacing the final softmax layer of the CNN3 model with an SVM classifier led to an increase in accuracy of 0.53%. This improvement can be explained by the intrinsic differences between the two classifiers: While the softmax layer functions as a linear classifier optimized for binary classification through cross-entropy loss, it may struggle with complex, high-dimensional, and non-linearly separable EEG feature spaces. In contrast, the SVM with an RBF kernel is designed to find an optimal separating hyperplane in a transformed feature space, maximizing the margin between the two classes and effectively handling non-linear relationships. This allows the SVM to better capture subtle non-linear EEG patterns associated with drowsiness, especially in the high-dimensional feature space extracted by CNN layers, thus enhancing classification performance. The accuracy and loss curves (during training and test) for the CNN SVM (RBF) are shown in Figure 8. Figure 9 illustrates the confusion matrix for the classification operation.
Figure 8. Loss and accuracy curves for training (blue) and test (red) of the proposed using CNN SVM (RBF) method.
Figure 9. The confusion matrix of the proposed approach.
The evolution of the loss and accuracy curves highlights the efficient convergence of the model. The loss curve shows a rapid decrease during the initial epochs, stabilizing near zero after approximately 10 iterations. The validation loss follows a similar trend, with minor fluctuations in later epochs, suggesting good generalization capabilities.
Regarding accuracy, a sharp increase is observed during the first few epochs, reaching a plateau close to 99% by the tenth iteration. The absence of a significant gap between the training and validation curves indicates that the model neither underfits nor overfits. These results suggest optimal convergence and a strong ability to generalize to unseen data.
The analysis of the confusion matrix reveals high classification performance. The model correctly identified 116 samples from class 0 and 120 samples from class 1. In this context, class 0 refers to the alert state, while class 1 refers to the drowsy state. However, four samples from class 0 were misclassified as class 1, introducing false positives. Notably, no false negatives were observed, indicating perfect sensitivity in detecting class 1.
These results demonstrate high precision and recall, confirming the robustness of the model. The absence of false negatives is particularly beneficial for critical applications where detecting class 1 is essential. Although the presence of a few false positives may lead to unnecessary alerts, their low occurrence suggests that the model maintains reliable and robust classification performance.
Figure 10 presents two complementary evaluation curves for the model: the precision–recall curve (left) and the ROC curve (right). The precision–recall curve, with an area under the curve (PR AUC) of 0.9986, lies almost entirely in the upper-right region, reflecting very high precision and recall across all decision thresholds. This shape indicates that the model maintains excellent precision even when recall is maximized, which is particularly relevant for potentially imbalanced datasets. The ROC curve, with an area under the curve (ROC AUC) of 0.9986, shows a nearly vertical rise followed by a plateau close to 1, illustrating an almost perfect discriminative ability between positive and negative classes. These results confirm that the model effectively detects the positive class while minimizing false positives, thereby demonstrating outstanding overall performance.
Figure 10. The PR and ROC ACU for the proposed method.

4.3. Comparison Between the Proposed Method and 1D CNN

This experiment aims to evaluate the effectiveness of the feature extraction method for drowsiness detection by comparing a 2D CNN + SVM model, which processes time-frequency representations (scalograms), with a 1D CNN that directly analyzes raw EEG signals.
The results in Table 6 demonstrate that the 2D CNN + SVM approach outperforms the 1D CNN, highlighting the advantage of leveraging time-frequency features for capturing relevant patterns in drowsiness detection. This comparison highlights the importance of using advanced feature extraction techniques, such as scalograms, to improve the performance of EEG-based drowsiness detection systems.
Table 6. Superior performance of 2D CNN + SVM in drowsiness detection over 1D CNN. A: accuracy, S: sensitivity, F1: F1-score, Sp: specificity, P: precision.
The results clearly show that the proposed method (2D CNN + SVM) is much more effective at detecting drowsiness and overcoming inter-subject variability compared to CNN 1D. Specifically, the performance of the proposed approach achieved an accuracy of 98.33%, while CNN 1D reached 72.79%, highlighting a significant improvement of approximately 35.09%.

4.4. Comparison with Transfer Learning Models

In this section, we compare the proposed 2D CNN + SVM approach with transfer learning models, specifically, VGG16, ResNet50, DeepConvNet, EEGNet, and ShallowConvNet, to evaluate its effectiveness for drowsiness detection in an inter-subject context. The goal is to demonstrate the advantages of the proposed method, particularly the integration of SVM in the classification phase, which enhances the system’s performance and robustness.
Transfer learning leverages pre-trained models trained on large datasets, enabling them to capture general features that can be fine-tuned for specific tasks like drowsiness detection. Transfer learning models provide strong initial feature extraction, which can be adapted to EEG data characteristics, offering a robust starting point for domain-specific tasks. By contrasting the results with established transfer learning models, we aim to highlight the superior accuracy, sensitivity, and F1-score of the 2D CNN + SVM approach, as illustrated in Table 7. This comparison underscores the importance of tailored feature extraction and classification techniques in improving drowsiness detection systems.
Table 7. Comparison between the proposed approach and transfer learning models.
For the comparison of our proposed approach with state-of-the-art models including 1D-CNN, VGG16, ResNet50, DeepConvNet, EEGNet, and ShallowConvNet, we employed commonly accepted standard hyperparameters as reported in the EEG signal processing literature. All models were trained for 50 epochs using the Adam optimizer with learning rates adapted to each architecture to ensure fair and effective convergence. The selection of hyperparameters was conducted through a grid search procedure to optimize model performance systematically.
Specifically, the 1D-CNN was configured with a learning rate of 0.001, a batch size of 32, a dropout rate of 0.5, and kernel sizes of 64 and 32 in its convolutional layers with ReLU activations. VGG16 and ResNet50, pretrained on ImageNet, were fine-tuned with a learning rate of 1 × 10−4, batch size of 32, and dropout of 0.5 applied on fully connected layers; these models use 3 × 3 kernels for VGG16 and a combination of 7 × 7 initial kernels followed by bottleneck blocks for ResNet50, all with ReLU activations. The EEG-specific architectures DeepConvNet, EEGNet, and ShallowConvNet were trained with a learning rate of 0.001, batch size of 64, and dropout rates ranging from 0.25 to 0.5. DeepConvNet and EEGNet utilize ELU activation functions, while ShallowConvNet applies square and logarithmic non-linearities suited for oscillatory EEG features. Kernel sizes varied per model according to their original designs: DeepConvNet uses temporal and spatial kernels of size [1 × 10], EEGNet employs depthwise and separable convolutions with kernels of sizes [1 × 64] and [1 × 16], and ShallowConvNet uses a kernel size of [1 × 13]. Batch normalization was applied after convolution layers in all models to stabilize training. This standardized configuration ensured consistent and fair evaluation of all models on the same datasets, allowing for a rigorous comparison of performance.

4.5. Discussion

The findings of this study highlight the potential of the proposed EEG-based method for drowsiness detection, achieving notable accuracy even when applied across different subjects. However, while the CWT successfully generated scalogram images that captured detailed time-frequency features, the method’s real-world practicality remains questionable. The reliance on CWT to extract meaningful information, although beneficial in revealing the evolution of frequency content over time, may introduce complexity and computational overhead. Furthermore, while this approach is an improvement on traditional methods of representing non-stationary signals such as EEG, its scalability and performance in more varied and less controlled environments still need to be proven, leaving room for further refinement and validation. Although the use of CNNs for automatic feature extraction has shown great promise for detecting patterns associated with drowsiness, it is essential to critically evaluate the wider implications of this approach, particularly in terms of its impact on detection accuracy and system performance. The implementation of an SVM allows efficient differentiation between drowsy and alert states in a cross-subjects framework, improving the robustness of the model to individual variability. However, it is worth asking whether this combination of CWT, CNN, and SVM effectively offers a superior solution to existing methods. Although the preliminary results are encouraging, further validation with diverse populations and under real-world conditions is needed to confirm that it systematically outperforms previous approaches in real-world applications.
In Table 8, we compare our proposed method with existing literature on drowsiness detection in an inter-subject setting. In [23], the authors employed a simple 1D CNN for drowsiness detection using 10-s EEG segments as input for classification. The final dense softmax layer of the CNN was responsible for decision-making, achieving an accuracy of 73.22%. In [24], the same approach was followed, except for a reduction in EEG segment length to 4 seconds. This modification resulted in a significant accuracy improvement to 95%, highlighting the crucial impact of segment length on detection performance.
In another study [25], the authors utilized EEG spectrograms for drowsiness detection, employing an RCNN model that achieved an accuracy of 88.39%. Similarly, in [26], a CNN-LSTM architecture was used to predict drowsiness states based on EEG spectrograms, reaching an accuracy of 82.73%. In [27], EEG epochs of 13 seconds were used as input features for a CNN-based model (CpNet), which achieved an accuracy of 86.44%. Finally, in [28], 5-s EEG epochs were directly fed into a CNN, resulting in an accuracy of 93%. In this work [61], the authors propose a deep learning model called AMD-GCN, which utilizes the power spectral density (PSD) of EEG signals filtered with a band-pass filter ([1–50] Hz) from 17 electrodes for drowsiness detection. The model was validated on the SEED-VIG database and achieved an overall accuracy of 89.94%. While this approach demonstrates strong performance in the field of drowsiness detection, it remains limited in terms of implementation on embedded systems due to its algorithmic complexity. Furthermore, the authors did not address inter-subject variability, as the evaluation was performed only in the intra-subject mode with a relatively high number of required electrodes.
In the work [62], the authors proposed a drowsiness detection approach focusing on minimizing inter-subject variability. In this context, they employed a Random Forest (RF) classifier, achieving an accuracy of 86%. The method was validated using the SEED-VIG database, extracting power spectral density (PSD) features from 8-s EEG epochs. While this approach is optimized from an implementation perspective, its accuracy remains relatively low, particularly in the context of subject-independent detection. Furthermore, the method focuses exclusively on a single type of feature (PSD), which may limit its generalization capability.
Table 8. Comparison of different EEG-based drowsiness detection methods.
Table 8. Comparison of different EEG-based drowsiness detection methods.
RefMethodSegment SizeClassifierDatabaseA (%)
[24]1D EEG + CNN4 sFully connected layer with Softmax activationSEED-VIG95.00
[25]Spectrogram + RCNNFully connected layer with Softmax activationSEED-VIG88.39
[26]Spectrogram + CNN-LSTMFully connected layer with Softmax activationSEED-VIG82.73
[63]AGL-NetFully connected layer with Softmax activationSEED-VIG87.30
[64]VIG-Net8 sFully connected layer with Softmax activationSEED-VIG95.00
[61]AMD-GCN model10 sFully connected layer with Softmax activationSEED-VIG89.94
[62]ML+PSD8 sRFSEED-VIG86
Proposed2D CNN + SVM30sSVMSEED-VIG95.8
Note: A is accuracy.
The literature analysis clearly demonstrates that the proposed method outperforms existing approaches, achieving an accuracy of 98.33%. This result indicates that our approach effectively mitigates inter-subject variability due to several key factors. One crucial factor is the length of EEG epochs, which plays a significant role in drowsiness detection. The use of 30-s epochs provides a more comprehensive representation of vigilance states compared to shorter segments of 1 s, 4 s, 5 s, or 13 s.
Another major contributor to the improved accuracy is the use of scalograms, which offer a time-frequency representation of EEG signals derived from the Continuous Wavelet Transform (CWT). Unlike raw EEG signals, which are challenging to interpret due to their non-stationary nature, or spectrograms generated via the Fast Fourier Transform (FFT), which provide only a frequency-domain representation, scalograms capture both temporal and spectral information. Additionally, CNNs are highly effective in processing images, making scalogram-based feature extraction particularly relevant. CNNs can automatically learn meaningful features associated with drowsiness, further enhancing classification performance.
Finally, the choice of a classification model that excels in binary classification, while effectively handling non-linearity and high-dimensional data, significantly improves detection accuracy compared to a simple dense softmax layer. Through the analysis of previous studies, we observe that one of the most widely used datasets for drowsiness detection research is the SEED-VIG database.
To assess the ability of our approach to minimize inter-subject variability and ensure robust generalization, we evaluated its performance on the SEED-VIG database. This dataset contains EEG recordings from 21 subjects, collected using 12 electrodes (CP1, CPz, CP2, P1, Pz, P2, PO3, POz, PO4, O1, Oz, and O2). Each participant underwent a two-hour driving simulation during two key periods of the day—the afternoon and the evening—when drowsiness levels are most likely to vary. From these recordings, 5040 scalograms were generated for analysis. The results show a slight performance degradation compared to the original DROZY dataset, with an accuracy of 95.8%, a sensitivity of 94.7%, and an F1-score of 95.4%, against 98.33% accuracy on DROZY. This performance drop can be primarily attributed to the difference in electrode configurations between the two datasets, particularly the absence of the C3 and C4 derivations in SEED-VIG. These central electrodes are known to capture cortical activity highly relevant to vigilance regulation and early drowsiness onset, and their absence reduces the discriminative information available to the classifier.
To further investigate the robustness and generalizability of our approach across different experimental conditions and subject populations, an additional validation was conducted using an independent dataset collected at the Sahloul University Hospital (Monastir, Tunisia). This dataset comprises 45 hours of EEG recordings from eight healthy subjects aged between 21 and 25 years, all without a history of alcoholism or drug use. Data collection was performed at the Vigilance and Sleep Center of the Faculty of Medicine in Monastir, following an experimental protocol approved by the faculty’s Ethics Committee. All participants signed an informed consent form prior to the experiment, in compliance with ethical research standards. EEG signals were recorded from 19 channels (Fp1, Fp2, F2, F3, Fz, F4, F8, T3, C3, Cz, C4, T4, T5, P3, Pz, P4, T6, O1, and O2).
For validation purposes, only the C3 and C4 electrodes were retained, ensuring consistency with our minimal-electrode configuration. On this dataset, the proposed method achieved an accuracy of 97.83%, which is only marginally lower than the performance obtained on the DROZY database. These results demonstrate that the approach is capable of maintaining high detection performance even when applied to EEG data collected from different populations, under varying recording setups, and in diverse environments. The ability to achieve comparable performance with minimal electrodes across heterogeneous datasets strongly supports our claim that the proposed method effectively mitigates inter-subject variability and remains suitable for practical drowsiness monitoring applications, where electrode configurations and acquisition protocols may differ from one scenario to another. To assess whether the observed differences in performance across datasets were statistically significant, we conducted independent t-tests with corresponding p-value analysis. The comparison between the DROZY (98.33%) and SEED-VIG (95.8%) datasets revealed a highly significant difference (t = 23.72, p < 0.001), confirming that the absence of C3 and C4 electrodes in SEED-VIG had a measurable impact on performance. A smaller but significant difference was also found between DROZY (98.33%) and the Sahloul dataset (97.83%) (t = 5.56, p < 0.001). Finally, SEED-VIG and Sahloul also showed a statistically significant difference (t = −18.33, p < 0.001), highlighting the role of electrode configurations and acquisition protocols in shaping classification outcomes. Despite these statistical differences, all datasets consistently achieved accuracies above 95%, demonstrating that the proposed method remains robust and generalizable across heterogeneous populations and experimental conditions.

5. Conclusions

Inter-subject variability represents a major challenge in EEG-based drowsiness detection, often limiting the reliability and generalizability of existing systems. In this study, we addressed this issue by proposing a hybrid detection approach that leverages the Continuous Wavelet Transform (CWT) with the Morlet wavelet to generate scalogram images, thereby capturing both temporal and spectral information associated with vigilance states. These scalograms were resized and processed by a lightweight CNN, specifically designed to maintain computational efficiency while ensuring high discriminative power. The extracted features were then classified using an SVM with an RBF kernel, which excels in handling the non-linear and high-dimensional nature of EEG data. This combination achieved an accuracy of 98.33% in inter-subject scenarios, confirming its ability to mitigate variability between subjects and provide a highly generalized and accurate detection system.
The novelty of the approach lies not only in the CNN–SVM fusion, but also in its integration of several complementary strategies: (i) the use of 30-s epochs to capture gradual vigilance transitions, (ii) reliance on a minimal-electrode configuration (C3 and C4) to enhance practicality for real-world use, (iii) feature standardization to further reduce inter-subject variability, and (iv) the adoption of a lightweight CNN architecture optimized for FPGA-friendly implementation. These design choices collectively ensure a strong balance between accuracy, robustness, and computational efficiency, with results consistently above 95% across heterogeneous datasets.
Looking ahead, we plan to extend this work in two directions. First, the integration of complementary physiological signals (i.e., EOG, ECG) and facial expressions will be explored to further improve robustness. Second, as suggested by recent advances, we acknowledge the potential of super-resolution techniques for time–frequency mapping and more advanced classifiers such as Vision Transformers (ViT) and large vision models like CLIP. While these models currently demand high computational resources and are less suited to embedded contexts, future investigations will explore lightweight adaptations of such architectures. This direction aims to combine the strength of transformer-based models with our embedded implementation strategy, ensuring that drowsiness detection systems remain both accurate and deployable in real-time portable platforms such as FPGAs.

Author Contributions

Conceptualization, A.Z. and K.B.K.; methodology, A.Z.; software, A.Z.; validation, N.B., K.B.K. and C.V.; formal analysis, A.Z.; investigation, A.Z.; resources, A.Z.; data curation, A.Z.; writing—original draft preparation, A.Z. and K.B.K.; writing—review and editing, A.Z. and M.H.B.; visualization, K.B.K. and N.B.; supervision, K.B.K., N.B. and C.V.; project administration, K.B.K. and C.V.; funding acquisition, K.B.K. and C.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available DOI:10.1109/WACV.2016.7477715.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. van Schie, M.K.; Lammers, G.J.; Fronczek, R.; Middelkoop, H.A.; van Dijk, J.G. Vigilance: Discussion of related concepts and proposal for a definition. Sleep Med. 2021, 83, 175–181. [Google Scholar] [CrossRef] [PubMed]
  2. Nir, Y.; de Lecea, L. Sleep and vigilance states: Embracing spatiotemporal dynamics. Neuron 2023, 111, 1998–2011. [Google Scholar] [CrossRef]
  3. Murray, S.; Amaya, S. The strategic allocation theory of vigilance. Wiley Interdiscip. Rev. Cogn. Sci. 2024, 15, e1693. [Google Scholar] [CrossRef]
  4. Park, J.H.; Lee, J.D. A Customized Deep Sleep Recommender System Using Hybrid Deep Learning. Sensors 2023, 23, 6670. [Google Scholar] [CrossRef]
  5. Hassinger, A.B.; Afzal, S.; Rauth, M.; Breuer, R.K. Pediatric Intensive Care Unit related Sleep and Circadian Dysregulation: A focused review. Semin. Pediatr. Neurol. 2023, 48, 101077. [Google Scholar] [CrossRef]
  6. Ibe, T.; Fujiwara, K.; Hiraoka, T.; Abe, E.; Yamakawa, T. Development of game-like system using active behavior input for wakefulness-keeping support in driving. IEEE Trans. Intell. Veh. 2020, 6, 323–332. [Google Scholar] [CrossRef]
  7. Slater, J.D. A definition of drowsiness: One purpose for sleep? Med. Hypotheses 2008, 71, 641–644. [Google Scholar] [CrossRef] [PubMed]
  8. Saleem, A.A.; Siddiqui, H.U.R.; Raza, M.A.; Rustam, F.; Dudley, S.; Ashraf, I. A systematic review of physiological signals based driver drowsiness detection systems. Cogn. Neurodynamics 2023, 17, 1229–1259. [Google Scholar] [CrossRef] [PubMed]
  9. Albadawi, Y.; Takruri, M.; Awad, M. A review of recent developments in driver drowsiness detection systems. Sensors 2022, 22, 2069. [Google Scholar] [CrossRef] [PubMed]
  10. Safarov, F.; Akhmedov, F.; Abdusalomov, A.B.; Nasimov, R.; Cho, Y.I. Real-time deep learning-based drowsiness detection: Leveraging computer-vision and eye-blink analyses for enhanced road safety. Sensors 2023, 23, 6459. [Google Scholar] [CrossRef] [PubMed]
  11. Das, R.K.; Martin, A.; Zurales, T.; Dowling, D.; Khan, A. A survey on EEG data analysis software. Sci 2023, 5, 23. [Google Scholar] [CrossRef]
  12. Othmani, A.; Sabri, A.Q.M.; Aslan, S.; Chaieb, F.; Rameh, H.; Alfred, R.; Cohen, D. EEG-based neural networks approaches for fatigue and drowsiness detection: A survey. Neurocomputing 2023, 557, 126709. [Google Scholar] [CrossRef]
  13. Luu, L.; Pham, P.; Le, T.Q. Feature Extraction and Electrophysiological Modeling in Personalized Deep Brain Structure Using Electroencephalography Signal. In Proceedings of the International Conference on the Development of Biomedical Engineering in Vietnam, Ho Chi Minh, Vietnam, 27–29 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 551–556. [Google Scholar]
  14. Minhas, R.; Peker, N.Y.; Hakkoz, M.A.; Arbatli, S.; Celik, Y.; Erdem, C.E.; Peker, Y.; Semiz, B. Improved drowsiness detection in drivers through optimum pairing of EEG features using an optimal EEG channel comparable to a multichannel EEG system. Med. Biol. Eng. Comput. 2025, 1–18. [Google Scholar] [CrossRef] [PubMed]
  15. Antoine, J.P.; Murenzi, R. The continuous wavelet transform, from 1 to 3 dimensions. In Subband and Wavelet Transforms: Design and Applications; Springer: Berlin/Heidelberg, Germany, 1996; pp. 149–187. [Google Scholar]
  16. Murdick, D.; Dunham, J.; Melot, J. AI Definitions Affect Policymaking; CSET Issue Brief; Center for Security and Emerging Technology: Washington, DC, USA, 2020. [Google Scholar]
  17. Mobarak, M.H.; Mimona, M.A.; Islam, M.A.; Hossain, N.; Zohura, F.T.; Imtiaz, I.; Rimon, M.I.H. Scope of machine learning in materials research—A review. Appl. Surf. Sci. Adv. 2023, 18, 100523. [Google Scholar] [CrossRef]
  18. Soori, M.; Arezoo, B.; Dastres, R. Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cogn. Robot. 2023, 3, 54–70. [Google Scholar] [CrossRef]
  19. Raj, R.; Kos, A. An extensive study of convolutional neural networks: Applications in computer vision for improved robotics perceptions. Sensors 2025, 25, 1033. [Google Scholar] [CrossRef]
  20. Pham, T.T.A.; Nguyen, T.D.H.; Le, Q.K.; Huynh, Q.L. Application of portable EEG device in detection and classification drowsiness by support vector machine. In Proceedings of the International Conference on the Development of Biomedical Engineering in Vietnam, Ho Chi Minh, Vietnam, 27–29 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 521–526. [Google Scholar]
  21. Abidi, A.; Ben Khalifa, K.; Ben Cheikh, R.; Valderrama Sakuyama, C.A.; Bedoui, M.H. Automatic detection of drowsiness in EEG records based on machine learning approaches. Neural Process. Lett. 2022, 54, 5225–5249. [Google Scholar] [CrossRef]
  22. Wang, H.; Zhang, L.; Yao, L. Application of genetic algorithm based support vector machine in selection of new EEG rhythms for drowsiness detection. Expert Syst. Appl. 2021, 171, 114634. [Google Scholar] [CrossRef]
  23. Cui, J.; Lan, Z.; Liu, Y.; Li, R.; Li, F.; Sourina, O.; Müller-Wittig, W. A compact and interpretable convolutional neural network for cross-subject driver drowsiness detection from single-channel EEG. Methods 2022, 202, 173–184. [Google Scholar] [CrossRef]
  24. Obaidan, H.B.; Hussain, M.; AlMajed, R. EEG_DMNet: A Deep Multi-Scale Convolutional Neural Network for Electroencephalography-Based Driver Drowsiness Detection. Electronics 2024, 13, 2084. [Google Scholar] [CrossRef]
  25. Gao, D.; Tang, X.; Wan, M.; Huang, G.; Zhang, Y. EEG driving fatigue detection based on log-Mel spectrogram and convolutional recurrent neural networks. Front. Neurosci. 2023, 17, 1136609. [Google Scholar] [CrossRef]
  26. Alghanim, M.; Attar, H.; Rezaee, K.; Khosravi, M.; Solyman, A.; Kanan, M.A. A hybrid deep neural network approach to recognize driving fatigue based on EEG signals. Int. J. Intell. Syst. 2024, 2024, 9898333. [Google Scholar] [CrossRef]
  27. Guarda, L.; Tapia, J.E.; Droguett, E.L.; Ramos, M. A novel Capsule Neural Network based model for drowsiness detection using electroencephalography signals. Expert Syst. Appl. 2022, 201, 116977. [Google Scholar] [CrossRef]
  28. Boudaya, A.; Bouaziz, B.; Chaabene, S.; Chaari, L.; Ammar, A.; Hökelmann, A. EEG-based hypo-vigilance detection using convolutional neural network. In Proceedings of the The Impact of Digital Technologies on Public Health in Developed and Developing Countries: 18th International Conference, ICOST 2020, Hammamet, Tunisia, 24–26 June 2020; Proceedings 18. Springer: Berlin/Heidelberg, Germany, 2020; pp. 69–78. [Google Scholar]
  29. Pérez-Velasco, S.; Santamaria-Vazquez, E.; Martinez-Cagigal, V.; Marcos-Martinez, D.; Hornero, R. EEGSym: Overcoming inter-subject variability in motor imagery based BCIs with deep learning. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 1766–1775. [Google Scholar] [CrossRef]
  30. Vo, K.; Vishwanath, M.; Srinivasan, R.; Dutt, N.; Cao, H. Composing graphical models with generative adversarial networks for EEG signal modeling. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 1231–1235. [Google Scholar]
  31. Kim, J.H.; Nam, H.; Won, D.; Im, C.H. Domain-generalized Deep Learning for Improved Subject-independent Emotion Recognition Based on Electroencephalography. Exp. Neurobiol. 2025, 34, 119–130. [Google Scholar] [CrossRef]
  32. Massoz, Q.; Langohr, T.; François, C.; Verly, J.G. The ULg multimodality drowsiness database (called DROZY) and examples of use. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; IEEE: New York, NY, USA, 2016; pp. 1–7. [Google Scholar]
  33. Homan, R.W.; Herman, J.; Purdy, P. Cerebral location of international 10–20 system electrode placement. Electroencephalogr. Clin. Neurophysiol. 1987, 66, 376–382. [Google Scholar] [CrossRef] [PubMed]
  34. Miley, A.Å.; Kecklund, G.; Åkerstedt, T. Comparing two versions of the Karolinska Sleepiness Scale (KSS). Sleep Biol. Rhythm. 2016, 14, 257–260. [Google Scholar] [CrossRef] [PubMed]
  35. Sharma, R.; Meena, H.K. Emerging Trends in EEG Signal Processing: A Systematic Review. SN Comput. Sci. 2024, 5, 415. [Google Scholar] [CrossRef]
  36. Mumtaz, W.; Rasheed, S.; Irfan, A. Review of challenges associated with the EEG artifact removal methods. Biomed. Signal Process. Control 2021, 68, 102741. [Google Scholar] [CrossRef]
  37. Sen, D.; Mishra, B.B.; Pattnaik, P.K. A review of the filtering techniques used in EEG signal processing. In Proceedings of the 2023 7th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 11–13 April 2023; IEEE: New York, NY, USA, 2023; pp. 270–277. [Google Scholar]
  38. Buzsaki, G. Rhythms of the Brain; Oxford University Press: Oxford, UK, 2006. [Google Scholar]
  39. Kaur, G.; Aggarwal, H.; Goel, N. An Insightful Analysis of Preprocessing Methods Used in EEG Signals for Computer-Assisted Cognitive Domain. In Proceedings of the Congress on Intelligent Systems, Bengaluru, India, 4–5 September 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 53–65. [Google Scholar]
  40. Li, Y.; Liu, A.; Yin, J.; Li, C.; Chen, X. A segmentation-denoising network for artifact removal from single-channel EEG. IEEE Sens. J. 2023, 23, 15115–15127. [Google Scholar] [CrossRef]
  41. Tawhid, M.N.A.; Siuly, S.; Wang, K.; Wang, H. Automatic and efficient framework for identifying multiple neurological disorders from EEG signals. IEEE Trans. Technol. Soc. 2023, 4, 76–86. [Google Scholar] [CrossRef]
  42. Asl, N.S.; Baghdadi, G.; Ebrahimian, S.; Haghighi, S.J. Toward applicable EEG-based drowsiness detection systems: A review. Front. Biomed. Technol. 2022, 9, 323–350. [Google Scholar]
  43. Pachori, R.B. Time-Frequency Analysis Techniques and Their Applications; CRC Press: Boca Raton, FL, USA, 2023. [Google Scholar]
  44. Bajaj, N. Wavelets for EEG analysis. In Wavelet Theory; IntechOpen: London, UK, 2020; pp. 1–16. [Google Scholar]
  45. Djaballah, S.; Meftah, K.; Khelil, K.; Sayadi, M. Deep transfer learning for bearing fault diagnosis using CWT time–frequency images and convolutional neural networks. J. Fail. Anal. Prev. 2023, 23, 1046–1058. [Google Scholar] [CrossRef]
  46. Almanza-Conejo, O.; Almanza-Ojeda, D.L.; Contreras-Hernandez, J.L.; Ibarra-Manzano, M.A. Emotion recognition in EEG signals using the continuous wavelet transform and CNNs. Neural Comput. Appl. 2023, 35, 1409–1422. [Google Scholar] [CrossRef]
  47. Latreche, I.; Slatnia, S.; Kazar, O.; Barka, E.; Harous, S. A Review on Deep Learning Techniques for EEG-Based Driver Drowsiness detection systems. Informatica 2024, 48, 359–378. [Google Scholar] [CrossRef]
  48. Abdubrani, R.; Mustafa, M.; Zahari, Z.L. Enhancement of Morlet Mother Wavelet in Time–Frequency Domain in Electroencephalogram (EEG) Signals for Driver Fatigue Classification. In Advances in Intelligent Manufacturing and Mechatronics: Selected Articles from the Innovative Manufacturing, Mechatronics & Materials Forum (iM3F 2022), Pahang, Malaysia; Springer: Berlin/Heidelberg, Germany, 2023; pp. 151–161. [Google Scholar]
  49. Said, A. New Filters for Image Interpolation and Resizing. arXiv 2023, arXiv:2312.00926. [Google Scholar] [CrossRef]
  50. Rossi, A.; Hagenbuchner, M.; Scarselli, F.; Tsoi, A.C. A Study on the effects of recursive convolutional layers in convolutional neural networks. Neurocomputing 2021, 460, 59–70. [Google Scholar] [CrossRef]
  51. Apicella, A.; Donnarumma, F.; Isgrò, F.; Prevete, R. A survey on modern trainable activation functions. Neural Netw. 2021, 138, 14–32. [Google Scholar] [CrossRef]
  52. Vargas, J.A.C.; Esquivel, J.Z.; Tickoo, O. Introducing region pooling learning. In Pattern Recognition, Proceedings of the ICPR International Workshops and Challenges (ICPR 2021), Virtual Event, 10–15 January 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 714–724. [Google Scholar]
  53. Wei, C.; Kakade, S.; Ma, T. The implicit and explicit regularization effects of dropout. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 10181–10192. [Google Scholar]
  54. Krichen, M. Convolutional Neural Networks: A Survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  55. Cong, S.; Zhou, Y. A review of convolutional neural network architectures and their optimizations. Artif. Intell. Rev. 2023, 56, 1905–1969. [Google Scholar] [CrossRef]
  56. Abdullah, D.M.; Abdulazeez, A.M. Machine learning applications based on SVM classification a review. Qubahan Acad. J. 2021, 1, 81–90. [Google Scholar] [CrossRef]
  57. Wang, F.; Wan, Y.; Li, M.; Huang, H.; Li, L.; Hou, X.; Pan, J.; Wen, Z.; Li, J. Recent Advances in Fatigue Detection Algorithm Based on EEG. Intell. Autom. Soft Comput. 2023, 35, 3573–3586. [Google Scholar] [CrossRef]
  58. Balam, V.P. Systematic Review of Single-Channel EEG-Based Drowsiness Detection Methods. IEEE Trans. Intell. Transp. Syst. 2024, 25, 15210–15228. [Google Scholar] [CrossRef]
  59. Zayed, A.; Khalifa, K.B.; Belhadj, N.; Bedoui, M.H.; Sakuyama, C.A.V. Discrete wavelet transform coefficients for drowsiness detection from eeg signals. In Proceedings of the 2023 IEEE International Conference on Design, Test and Technology of Integrated Systems (DTTIS), Gammarth, Tunisia, 1–4 November 2023; IEEE: New York, NY, USA, 2023; pp. 1–6. [Google Scholar]
  60. Zayed, A.; Belhadj, N.; Ben Khalifa, K.; Bedoui, M.H.; Valderrama, C. Efficient Generalized Electroencephalography-Based Drowsiness Detection Approach with Minimal Electrodes. Sensors 2024, 24, 4256. [Google Scholar] [CrossRef] [PubMed]
  61. Liu, H.; Liu, Q.; Cai, M.; Chen, K.; Ma, L.; Meng, W.; Zhou, Z.; Ai, Q. Attention-based multi-semantic dynamical graph convolutional network for eeg-based fatigue detection. Front. Neurosci. 2023, 17, 1275065. [Google Scholar] [CrossRef] [PubMed]
  62. Rogel, J.M.H.; Beltrán, E.T.M.; Pérez, M.Q.; Bernal, S.L.; Pérez, G.M.; Celdrán, A.H. Studying drowsiness detection performance while driving through scalable machine learning models using electroencephalography. arXiv 2022, arXiv:2209.04048. [Google Scholar]
  63. Fang, W.; Tang, L.; Pan, J. AGL-Net: An efficient neural network for EEG-based driver fatigue detection. J. Integr. Neurosci. 2023, 22, 146. [Google Scholar] [CrossRef]
  64. Ko, W.; Oh, K.; Jeon, E.; Suk, H.I. Vignet: A deep convolutional neural network for eeg-based driver vigilance estimation. In Proceedings of the 2020 8th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Republic of Korea, 26–28 February 2020; IEEE: New York, NY, USA, 2020; pp. 1–3. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.