Efficient Feature Extraction for EEG-Based Classification: A Comparative Review of Deep Learning Models
Abstract
1. Introduction
- A comprehensive review of 114 DL-based EEG classification papers, including systematic reviews, CNN-based models, Transformer-based models, CNN–Transformer hybrids, and other recurrent-based hybrids.
- An evaluation and discussion of 88 DL-based EEG models, covering the most common network architectures, along with an analysis of efficiency and performance challenges.
- An in-depth trade-off analysis using different approaches of evaluation to cover a wider spectrum of possible trade-offs.
- The identification of current challenges in DL-based EEG classification and potential directions to inform future research.
2. Methods
2.1. Identification
2.2. Screening

2.3. Inclusion
2.4. Analysis
3. Related Work
3.1. Background
3.2. Recent Reviews
3.3. Study Scope
4. Feature Extraction for EEG-Based Classification
4.1. Efficiency in CNN-Based Models
4.1.1. Applications to Raw EEG
4.1.2. Frequency- and Spectral-Domain Approaches
4.1.3. EEG Representation and Topology Strategies
4.1.4. ERP- and VEP-Based CNN Models
4.1.5. Multiscale and Temporal Modeling Strategies
4.1.6. Compact and Lightweight Architectures
4.1.7. Benchmarking and Next-Generation CNN Designs
4.1.8. Fusion and Hybrid Feature Extraction Strategies
4.1.9. Cross-Integration Overlaps
| Ref. | Author/Year | Model | Protocol | Samples | Channels | Inspiration Basis | Acc. |
|---|---|---|---|---|---|---|---|
| [70] | Ma et al. 2015 | CNN-based | Resting | 10 | 64 | CNN | 88.00 |
| [71] | Mao et al. 2017 | CNN-based | VAT | 100 | 64 | CNN | 97.00 |
| [75] | Gonzalez et al. 2017 | ES1D | AVEPs | 23 | 16 | 1D CNN [Inception] | 94.01 |
| [82] | Das et al. 2017 | CNN-based | MI, VEPs | 40 | 17 | CNN | 98.80 |
| [83] | Cecotti et al. 2017 | CNN1-6 | RSVP | 16 | 64 | CNN | 83.10–90.50 |
| [11] | Schirrmeister et al. 2017 | ConvNets | MI | 9/54 | 22/54/3 | ResNet, Deep/Shallow ConvNet | 81.00–85.20 |
| [87] | Bai et al. 2018 | TCN | CNN | 97.20–99.00 | |||
| [12] | Lawhern et al. 2018 | EEGNet | ERPs, ERN, SMR, MRCP | 15/26/13/9 | 64/56/22 | CNN | 0.91 [AUC] |
| [104] | Wu et al. 2018 | CNN-based | RSVP | 10/15 | 16 | CNN | 97.60 |
| [72] | Schons et al. 2018 | CNN-based | Resting | 109 | 64 | CNN | 99.00 |
| [73] | Di et al. 2018 | CNN-based | ERPs | 33 | 64 | CNN | 99.30–99.90 |
| [74] | Zhang et al. 2018 | HDNN/CNN4EEG | RSVP | 15 | 64 | CNN | 89.00 |
| [76] | Waytowich et al. 2018 | Compact EEGNet | SSVEP | 10 | 8 | EEGNet | 80.00 |
| [78] | Lai et al. 2019 | CNN-based | Resting | 10 | 64 | CNN | 83.21/79.08 |
| [85] | Chen et al. 2019 | GSLT-CNN | ERPs, RSVP | 10/32/157 | 28/64 | CNN | 97.06 |
| [79] | Wang et al. 2019 | CNN-based | SSVEP | 10 | 8 | CNN | 99.73 |
| [77] | Yu et al. 2019 | M-Shallow ConvNet | SSVEP | 8 | 9 | CNN | 96.78 |
| [84] | Cecotti et al. 2019 | 1/2/3D CNN | RSVP | 16 | 64 | CNN | 92.80 |
| [80] | Wang et al. 2019 | CNN-based | Resting | 109/59 | 64/46 | Graph CNN | 99.98/98.96 |
| [105] | Özdenizci et al. 2019 | Adversarial CNN | RSVP | 3/10 | 16 | CNN [Adversarial] | 98.60 |
| [93] | Salimi et al. 2020 | N-Back-EEGNet | N-back | 26 | 28 | EEGNet | 95.00 |
| [94] | Ingolfsson et al. 2020 | EEG-TCNet | MI | 9 | 22 | EEGNet-TCN | 77.35–97.44 |
| [88] | Riyad et al. 2020 | Incep-EEGNet | MI | 9 | 22 | Inception-EEGNet | 74.08 |
| [89] | Liu et al. 2020 | PSTSA-CNN | MI | 9/14 | 22/44 | CNN-S Attention | 74.07–97.68 |
| [95] | Kasim et al. 2021 | 1DCNN | Photic stimuli | 16 | 16 | CNN | 97.17 |
| [90] | Zhu et al. 2021 | RAMST-CNN | MI | 109 | 64 | CNN [ResNet] | 96.49 |
| [106] | Musallam et al., 2021 | TCNet-Fusion | MI | 9/14 | 22/44 | EEG-TCNet | 83.73–94.41 |
| [107] | Mane et al. 2021 | FBCNet | MI | 9/54/37/34 | 22/20/27 | CNN | 74.70–81.11 |
| [86] | Salami et al. 2022 | EEG-ITNet | MI | 54/9 | 20/22 | Inception-TCN | 76.19/78.74 |
| [81] | Zhang et al. 2022 | 3D CNN | VEPs | 70 | 16 | CNN | 82.33 |
| [99] | Bidgoly et al. 2022 | CNN-based | Resting | 109 | 64/32/3 | CNN | 98.04 |
| [96] | Wu et al. 2022 | Mixed-FBCNet | MI | 109/9/10 | 64/22/10 | FBCNet | 98.89–99.48 |
| [97] | Altuwaijri et al. 2022 | MBEEG-SE | MI | 9 | 22 | EEGNet-S Attention | 82.87–96.15 |
| [98] | Autthasan et al. 2022 | MIN2Net | MI | 9/14/54 | 20/15 | AE [CNN] | 72.03/68.81 |
| [92] | Ding et al. 2023 | TSception | EMO | 32/27 | 32/32 | GoogleNet | 61.27/63.75 |
| [100] | Alsumari et al. 2023 | CNN-based | Resting | 109 | 3 | CNN | 99.05 |
| [101] | Yap et al. 2023 | GoogleNet, ResNet, EfficientNet, DenseNet, Inception | ERPs | 30 | 14 | CNN | 80:00 |
| [102] | Chen et al. 2024 | EEGNeX | ERPs, MI, SMR, ERN | 1/54/6/26 | 14/20/22/56 | EEGNet | 78.81–93.81 |
| [103] | Shakir et al. 2024 | STFE/MTFE-R-CNN | MI | 109 | 64 | CNN | 89.00/95.00 |
| [91] | Lakhan et al. 2025 | EEG-BBNet | MI, ERPs, SSVEP | 54 | 62/14/8 | CNN–Graph CNN | 99.26 |
4.2. Efficiency in Transformer-Based Models
4.2.1. Applications to Raw EEG
4.2.2. Generative and Self-Supervised Foundation Models
4.2.3. Modular and Dual-Branch Spatiotemporal Transformers
4.2.4. Ensemble and Multidomain Transformers
4.2.5. Specialized Attention Mechanisms and Dual Architectures
4.2.6. Cross-Integration Overlaps
| Ref. | Author/Year | Model | Protocol | Samples | Channels | Inspiration Basis | Acc. |
|---|---|---|---|---|---|---|---|
| [38] | Arjun et al. 2021 | ViT-CWT, ViT-Raw EEG | EMO-VAT | 32 | 32 | T Encoder (ViT) | 97.00/95.75 99.40/99.10 |
| [34] | Dosovitskiy et al. 2021 | ViT (Base/Large/Huge) | T Encoder (ViT) | 77.63–94.55 | |||
| [39] | Song et al. 2021 | S3T | MI | 9/9 | 22/3 | T Encoder [CNN] | 82.59/84.26 |
| [40] | Tao et al. 2021 | Gated Transformer | MI, VAT | 109/6 | 64/128 | T Encoder [GRU] | 61.11/55.4 |
| [42] | Du et al. 2022 | ETST | Resting | 109 | 64 | T Encoder | 97.29–97.90 |
| [43] | Zeynali et al. 2023 | Ensemble Transformer | VEPs | 8 | 64 | T Encoder | 96.10 |
| [112] | Wei et al. 2023 | TC-Net | EMO, AVP | 32/23 | 48/15 | T Encoder [CapsNet, ViT] | 98.59–98.82 |
| [41] | Siddhad et al. 2024 | Transformer-based | Resting | 60/48 | 14 | T Encoder | 95.28 |
| [44] | Omair et al. 2024 | GET | MI/Alpha EEGs | 9/20 | 3/16 | Transformer | 85.00 |
| [46] | Wang et al. 2024 | EEGPT | ERPs, MI, SSVEP, EMO | 9-2383 | 58/3-128 | Transformer [BERT, ViT] | 58.46–80.59 |
| [108] | Hu et al. 2024 | HASTF | EMO | 32/15 | 32/62 | Transformer [BERT] | 98.93/99.12 |
| [110] | Wei et al. 2025 | Fusion Transformer | EMO | 62 | T Encoder | 87.38/95.73 | |
| [45] | Lim et al. 2025 | EEGTrans | MI | 1/7/9/14 | 3/22/59/128 | Transformer | 80.69–90.84 |
| [111] | Ghous et al. 2025 | AE-BMD, CD-FTA | EMO | 15/20/23 | 62 | T Encoder [RNN, MLP] | 79.00–95.00 |
| [109] | Muna et al. 2025 | SSTAF | MI | 103/9 | 64/22 | Transformer | 68.30/76.83 |
4.3. Efficiency in CNN–Transformer-Based Hybrids
4.3.1. Sequential Pipelines
4.3.2. Parallel and Multibranch Blocks
4.3.3. Integrated and Hierarchical Attention
4.3.4. TCN-Enhanced Models for Lightweight Temporal Modeling
4.3.5. Pretrained and Self-Supervised Transformers
4.3.6. Cross-Integration Overlaps
| Ref. | Author/Year | Model | Protocol | Samples | Channels | Inspiration Basis | Acc. |
|---|---|---|---|---|---|---|---|
| [113] | Sun et al. 2021 | Fusion-CNN-Trans | MI | 109 | 64 | CNN–T Encoder | 87.80 |
| [131] | Kostas et al. 2021 | BENDR | Raw EEG | >10,000 | 20 | CNN–T Encoder | 86.70 |
| [124] | Bagchi et al. 2022 | EEG-ConvTranformer | VEPs | 10 | 128 | CNN–T Encoder (MHA) | 89.64 |
| [120] | Xie et al. 2022 | CTrans | MI | 109 | 64 | CNN–T Encoder | 83.31 |
| [128] | Altaheri et al. 2023 | ATCNet | MI | 9 | 22 | EEGNet–T Encoder (MHA)–TCN | 70.97–85.38 |
| [132] | Yang et al. 2023 | ViT2EEG | Raw EEG | 27 | 128 | EEGNet–T Encoder (ViT) | 55.40–61.70 |
| [119] | Li et al. 2023 | Dual-TSST | MI, EMO | 9/15 | 22/3/62 | CNN–T Encoder | 96.65 |
| [125] | Song et al. 2023 | EEG Conformer | MI | 9/9/15 | 22/3/62 | CNN–T Encoder (MHA) | 78.66/95.30 |
| [115] | Wan et al. 2023 | EEGFormer | SSVEP | 70/15/12 | 64/62/6 | CNN–T Encoder-CNN | 92.75 |
| [121] | Si et al. 2023 | TBEM | EMO | 80/6 | 30/30 | CNN–T Encoder-CNN (HybridNet–PureConvNet) | 42.50 |
| [127] | Gong et al. 2023 | ACTNN | EMO | 15 | 62 | CNN–T Encoder | 95.30 |
| [116] | Ma et al. 2023 | CNN-Transformer | MI | 9 | 22 | CNN–Transformer | 83.91 |
| [114] | Omair et al. 2024 | ConTraNet | MI | 9/105 | 3/64 | CNN–T Encoder | 86.98 |
| [126] | Si et al. 2024 | MACTN | EMO | 80/32 | 30/28 | CNN–T Encoder (MHA) | 67.80 |
| [117] | Zhao et al. 2024 | CTNet | MI | 9 | 22/3 | CNN–T Encoder | 83.11–97.81 |
| [133] | Jiang et al. 2024 | LaBraM | Resting, MI, Raw EEG | >140 | 19-64 | CNN–T Encoder | 82.58 |
| [118] | Liu et al. 2024 | ERTNet | EMO-AAT | 32/16 | 32/62 | CNN–T Encoder | 74.23 |
| [122] | Yao et al. 2024 | EEG ST-TCNN | EMO | 15/32 | 62/32 | T Encoder–CNN | 95.73–96.95 |
| [134] | Li et al. 2024 | MTL-Transformer1-2 | EEG, Eye tracking | 356 | 128 | ViT2EEG–CNN | |
| [123] | Lu et al. 2024 | CIT-EmotionNet | EMO | 15 | 62 | ResNet II–T Encoder | 92.09/98.57 |
| [129] | Nguyen et al. 2024 | EEG-TCNTransformer | MI | 9 | 22 | EEG-TCNet–T Encoder (MHA) | 83.41 |
| [130] | Cheng et al. 2024 | MSDCGTNet | EMO | 32/15 | 32/62 | CNN–T Encoder–TCN | 98.85/99.67 |
4.4. Efficiency in Recurrent Deep Learning Models
4.4.1. Attention-Based Architectures
4.4.2. CNN–Recurrent Hybrids
4.4.3. Stimulus-Locked Models
4.4.4. Multimodal and Parameter-Efficient Hybrids
4.4.5. Cross-Integration Overlaps
| Ref. | Author/Year | Model | Protocol | Samples | Channels | Inspiration Basis | Acc. |
|---|---|---|---|---|---|---|---|
| Pure Architectures | |||||||
| [144] | Kumar et al. 2019 | BLSTM-NN | VEPs | 33/58 | 14/16 | LSTM | 97.57 |
| [141] | Puengdang et al. 2019 | LSTM-based | SSVEP, ERPs | 20 | 6 | LSTM | 91.44 |
| [142] | Zheng et al. 2020 | ERP-LSTM | VEP | 10 | 128 | LSTM | 66.81 |
| [145] | Chakladar et al. 2021 | mSNN | MI | 70 | 14 | SNN | 98.57 |
| Other Hybrids | |||||||
| [138] | Wilaiprasitporn et al. 2015 | CNN-LSTM, CNN-GRU | ERPs, EMO | 32/40 | 5/32 | CNN-LSTM/GRU | 99.17–99.90 |
| [135] | Zhang et al. 2018 | MindID | Resting | 8 | 14/64 | Attention RNN | 98.20–99.89 |
| [139] | Sun et al. 2019 | 1DCNN-LSTM | Resting | 109 | 64/32/4 | 1D CNN-LSTM | 94.34–99.58 |
| [136] | Zhang et al. 2020 | DeepKey | Relaxing | 7 | 14 | Attention RNN | 99.00 |
| [143] | Jin et al. 2021 | CTNN | Resting, MI, EEG | 105/20/32 | 64/32/7 | CNN-TN | 99.50 |
| [140] | Chakravarthi et al. 2022 | ResNet152-LSTM | Resting | 20 | 4 | ResNet-LSTM | 98.00 |
| [137] | Balci et al. 2023 | DM-EEGID | Resting | 109 | 48 | Attention LSTM | 99.97–99.70 |
5. Comparative Analysis of Efficiency Trade-Offs
5.1. Proxy Metric Development
- We collected any metrics reported by the authors based on four axes: (1) accuracy (Acc.) to represent the overall performance of the model—we kept the highest reported value; (2) computational resources to represent system costs, such as the architectural cost (parameters), compute cost (FLOPs/MACs), memory footprint, and inference latency; (3) operational costs for acquisition costs like the epoch length (EEG segment) and channel count—for the channels and sample size, we kept the minimum reported values; (4) training costs such as training time and the GPU/TPU/cloud environment used for training. Thus, we were able to approximate the total cost to train and deploy these systems.
- Due to the inconsistent and incomplete reporting of system metrics, we employed a mixed approach to create four proxy metrics (each scored on a scale from 1 (Best/Lowest Cost) to 5 (worst/highest cost)). This calculation was based on evaluating a set of quantitative metrics (white background in tables) and the authors’ qualitative claims (orange background in tables) to define a unified cost dimension across all models.
- Complex. Proxy (Complex.): This metric reflects the size and depth of the architecture using the number of parameters (k/M) and EEG channels used. It quantifies the memory cost and architectural burden. Low Complex. is ideal for edge devices (Table 6).
| Scale | Rationale |
|---|---|
| 1 (Very Low) | Very simple or lightweight design. Lowest parameter counts. Designed for few channels. Minimal depth. |
| 2 (Low) | Small or simplified models. Low parameter counts. Low channel count. Uses efficient blocks. |
| 3 (Medium) | Standard deep learning models. Moderate parameter count. Low channel count. |
| 4 (High) | Deep, complex, or specialized models. High parameter count. Applied to high-channel datasets. |
| 5 (Very High) | Very complex models. Very high parameter count. Requires full channel count. |
- Computational Cost Proxy (Comp. Cost): This metric reflects the hardware resources required for a single prediction (inference) after training is complete. It is measured by MACs/FLOPs and the model parameters. It quantifies the processing cost. A low cost is crucial for real-time BCIs (Table 7).
| Scale | Rationale |
|---|---|
| 1 (Minimal) | Very low FLOPs/MACs. Designed for mobile/embedded focus. |
| 2 (Low) | Low parameter count. Claims to be more efficient/faster than standard models. |
| 3 (Medium) | Standard operational load (for most desktop/GPU systems). Claims of optimization but lacks high efficiency. |
| 4 (High) | High operational load (a powerful discrete GPU is required). High resource use optimized for Acc. |
| 5 (Very High) | Very high FLOPS/MACs. Model’s Complex. indicates high computational demand. |
- Operational Cost Proxy (Oper. Cost): This metric reflects the time required for a model to generate a prediction (latency) and the data requirements. It quantifies the real-world usability and user burden needed for the system. This is critical for systems needing seamless and immediate user interaction (Table 8).
| Scale | Rationale |
|---|---|
| 1 (Real-Time) | Reported latency is in sub-seconds. Uses few channels, short segments, and raw signals. No calibration. |
| 2 (Near Real-Time) | Latency is slightly longer (suitable for basic BCI control). Minimal preprocessing. |
| 3 (Acceptable) | Latency is half to a few seconds (adequate for user-paced, non-immediate tasks). Uses data transformation. |
| 4 (Slow) | Latency exceeds seconds (unsuitable for immediate interaction). Extensive pretraining and fine-tuning. |
| 5 (Very Slow) | Relies on a high number of channels or long segments. Large-scale pretraining or high input dimensions. |
- Training Cost Proxy (Train. Cost): This metric reflects the training time and the number of training epochs required to train the model from scratch. It quantifies the initial development and time investment before deployment. A low cost is desirable for research and rapid development (Table 9).
| Scale | Rationale |
|---|---|
| 1 (Instant) | Extremely low epoch count. Training time reported in minutes. Efficient transfer learning for new users. |
| 2 (Fast) | Low to moderate epoch counts (meaning quick convergence). No high-cost hardware. |
| 3 (Standard) | Typical training time for most deep learning models on one GPU. Common epoch count for the domain. |
| 4 (Long) | Extended training time on one GPU (suggesting a larger dataset or deeper model). High epoch count. |
| 5 (Very Long) | Extensive training time (hours/days) or complex iterative process on high GPUs. Very high epoch count. |
5.1.1. Efficiency Insights: CNNs
- Acc. vs. Complex. and Comp. Cost: Compact architectures can maintain strong performance with a limited Comp. Cost. Ref. [12] achieved 91% Acc. using only 1.066 K parameters. In [94], the results showed suitability for low-power deployment, with 4.27 K parameters, 6.8 M MACs, and 77.35% Acc. Ref. [106] used 17.58 K parameters and 20.69 M MACs to achieve 83.73% Acc., although the authors noted that the increased MACs might limit use on lightweight devices for a modest Acc. gain.
- Acc. vs. Oper. Cost: Reducing the number of channels lowers the setup complexity and hardware requirements without necessarily reducing Acc. With only three channels, Refs. [99,100] achieved 98.04% and 99.05% Acc., respectively. Ref. [100] also used 95 s epochs, balancing Acc. and recording efficiency. The epoch length is another important factor for real-time operation. Ref. [70] achieved only 76% Acc. with 62.5 ms segments, showing a speed–Acc. trade-off. In contrast, Refs. [79,80] reached around 99.9% Acc. using 1 s segments, trading speed for better performance. Low latency is critical for online BCI and authentication systems. Ref. [97] reported 1.79 ms latency with 96.15% Acc., suitable for real-time use. Ref. [94] showed 197 ms latency, representing a slower but computationally efficient design.
- Acc. vs. Train. Cost: Training requirements affect deployment feasibility. Ref. [11] reported 24.77 min training time, compared to 33 s for the FBCSP baseline, reflecting the high Comp. Costs of deep models for a moderate Acc. gain (85.20%). Refs. [79,80] required only 2–4 fine-tuning epochs (<1 min) to adapt to new users, reducing the calibration time. Ref. [96] reported 8–18 min enrollment time per subject, showing the time cost of subject-specific model adaptation.
5.1.2. Efficiency Insights: Transformers
- Acc. vs. Complex. and Comp. Cost: Model size influences both Acc. and computational efficiency. The ViT and EEGPT models [34,46] lack inductive biases useful for small datasets, which resulted in lower Acc. compared to ResNet. Ref. [34] used large-scale pretraining with 14–300 M images to overcome this trade-off. In [46], performance was scaled with up to 101 M parameters, but this increased the model size and memory usage. In contrast, Ref. [39] achieved 84.26% Acc. with only 6.50–8.68 K, which means that compact models can maintain competitive performance. Architectural choices also affect computational efficiency. Sequential RNNs/LSTMs have long training times for long sequences. Transformer-based models [40,42] address this by using faster parallel attention mechanisms. Expanding the input signal window provides a richer context and better performance but increases the computational demands and memory usage. Ref. [44] mitigated this by projecting the input into a lower-dimensional latent space (100). Ref. [45] had to limit its codebook to fit GPU constraints during training.
- Acc. vs. Oper. Cost: Transformer-based models suppress heavy FE, which reduces the operational complexity. Using raw signals [38] led to 99.4% Acc. with lower Oper. Costs compared to the CWT method (97% Acc.), meaning that reduced preprocessing can maximize performance. However, the raw data approach [41] yielded lower performance compared to traditional feature engineering. In [43], the simple temporal Transformer had a low Oper. Cost, while the spectrotemporal ensemble model had a high Oper. Cost but provided 96.1% Acc.
| Ref. | Author/Year | Parameters (K/M) | MACs/FLOPs (M/G) | Latency (s) | Training Time Epochs (s/m/h) | Memory Footprint | Epoch Length (Segment) (ms/s) | GPU TPU Cloud | Acc. (%) | Sample Size | Channels | Complex. | Comp. Cost | Oper. Cost | Train. Cost |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [70] | Ma et al. 2015 | - | - | - | - | - | 62.5 ms | - | 88.00 | 10 | 64 | 3 | 3 | 2 | 3 |
| [71] | Mao et al. 2017 | - | - | - | 0.3 h | - | - | - | 97.00 | 100 | 64 | 3 | 3 | 3 | 2 |
| [75] | Gonzalez et al. 2017 | - | - | - | 1 M iterations | - | 1 s | - | 94.01 | 23 | 16 | 2 | 3 | 3 | 4 |
| [82] | Das et al. 2017 | - | - | - | 50 epochs | - | 600 ms | - | 98.80 | 40 | 17 | 2 | 2 | 3 | 2 |
| [83] | Cecotti et al. 2017 | - | - | - | - | - | 800 ms | NVIDIA GTX 1080 | 90.50 | 16 | 64 | 2 | 3 | 3 | 5 |
| [11] | Schirrmeister et al. 2017 | - | - | - | 24 m 46 s | - | 4 s | NVIDIA GeForce GTX 980 | 85.20 | 9 | 3 | 3 | 4 | 3 | 4 |
| [87] | Bai et al. 2018 | 70 K | - | - | Fast convergence | Low | 1 s | - | 99.00 | - | - | 3 | 2 | 3 | 2 |
| [12] | Lawhern et al. 2018 | 1.066 K | - | - | 500 epochs | - | 4 s | NVIDIA Quadro M6000 GPU | 91.00 | 9 | 22 | 1 | 2 | 3 | 3 |
| [104] | Wu et al. 2018 | - | - | 7s | 500 epochs | - | 2 s | - | 97.60 | 10 | 16 | 2 | 4 | 4 | 3 |
| [72] | Schons et al. 2018 | - | - | - | - | - | 1 s | - | 99.00 | 109 | 64 | 3 | 3 | 3 | 3 |
| [73] | Di et al. 2018 | - | - | - | - | - | 1 s | GPU | 99.90 | 33 | 64 | 4 | 3 | 3 | 3 |
| [74] | Zhang et al. 2018 | - | - | - | Long training | 10x | 1.25 s | - | 89.00 | 15 | 64 | 2 | 3 | 3 | 4 |
| [76] | Waytowish et al. 2018 | - | - | - | - | - | 1s | - | 80.00 | 10 | 8 | 2 | 2 | 2 | 3 |
| [78] | Lai et al. 2019 | - | - | - | 30 repetitions | - | - | - | 83.21 | 10 | 64 | 2 | 2 | 3 | 2 |
| [85] | Chen et al. 2019 | - | - | - | 0.5 h | - | - | NVIDIA GeForce GTX TITAN X | 97.06 | 10 | 28 | 4 | 2 | 3 | 2 |
| [79] | Wang et al. 2019 | - | - | - | 2–4 epochs (fine-tune) | - | 1 s | - | 99.73 | 10 | 8 | 2 | 2 | 2 | 1 |
| [77] | Yu et al. 2019 | - | - | - | 50 iterations | - | - | - | 96.78 | 8 | 9 | 2 | 2 | 3 | 2 |
| [84] | Cecotti et al. 2019 | - | - | - | - | - | 800 ms | - | 92.80 | 16 | 64 | 4 | 3 | 3 | 3 |
| [80] | Wang et al. 2019 | - | - | - | <1 min, 0 epochs | - | 1s | - | 99.98 | 59 | 46 | 2 | 2 | 3 | 1 |
| [105] | Özdenizci et al. 2019 | - | - | - | 100 epochs | - | 0.5 s | - | 98.60 | 3 | 16 | 3 | 3 | 3 | 2 |
| [93] | Salimi et al. 2020 | - | - | - | 100 epochs | - | 1.1 s | NVIDIA Tesla K80 | 95.00 | 26 | 28 | 1 | 2 | 3 | 2 |
| [94] | Ingolfsson et al. 2020 | 4.27 K | 6.8 M | 197 ms | 750 epochs | 396 kB | 4 s | NVIDIA GTX 1080 Ti GPU | 97.44 | 9 | 22 | 1 | 1 | 4 | 4 |
| [88] | Riyad et al. 2020 | - | - | - | 180 epochs | - | 4 s | NVIDIA P100 GPU | 74.08 | 9 | 22 | 3 | 3 | 4 | 2 |
| [89] | Liu et al. 2020 | - | - | - | - | - | 4 s | NVIDIA RTX 2080Ti GPU | 97.68 | 9 | 22 | 4 | 4 | 4 | 3 |
| [95] | Kasim et al. 2021 | - | - | - | 1200 epochs | - | 3 s | - | 97.17 | 16 | 16 | 3 | 3 | 3 | 5 |
| [90] | Zhu et al. 2021 | - | - | - | - | - | 1 s | - | 96.49 | 109 | 64 | 4 | 3 | 3 | 3 |
| [106] | Musallam et al. 2021 | 17.58 K | 20.69 M | - | 1000 epochs | 1188 kB | 4.5 s | TensorFlow | 94.41 | 9 | 22 | 3 | 4 | 4 | 4 |
| [107] | Mane et al. 2021 | - | - | - | 600/1500 epochs | - | 4 s | - | 81.11 | 9 | 20 | 2 | 3 | 4 | 5 |
| [86] | Salami et al. 2022 | 3 K | - | - | 500 epochs | - | 4 s | - | 78.74 | 9 | 20 | 1 | 2 | 4 | 3 |
| [81] | Zhang et al. 2022 | - | - | - | Hour level | - | 4 s | - | 82.33 | 70 | 16 | 4 | 5 | 4 | 5 |
| [99] | Bidgoly et al. 2022 | - | - | - | 30 epochs | - | 1 s | - | 98.04 | 109 | 3 | 2 | 2 | 1 | 2 |
| [96] | Wu et al. 2022 | 450.626 K | - | 4 s | 8–10 m (enrollment) | - | 4 s | - | 99.48 | 9 | 10 | 4 | 4 | 2 | 2 |
| [97] | Altuwaijri et al. 2022 | 10.17 K | - | 1.79 ms | 1000 epochs | - | 4.5 s | Google Colab | 96.15 | 9 | 22 | 2 | 1 | 4 | 4 |
| [98] | Autthasan et al. 2022 | 55.232 K | - | 0.1–0.3 s | 0.47–1.36 s/epoch | - | 2 s | NVIDIA Tesla V100 GPU | 72.03 | 9 | 15 | 3 | 2 | 3 | 2 |
| [92] | Ding et al. 2023 | 12.56 K | - | - | 500 epochs | - | 2–4 s | - | 63.75 | 27 | 32 | 2 | 3 | 3 | 3 |
| [100] | Alsumari et al. 2023 | 74.071 K | - | - | 20 epochs | - | 5 s | Google Colab | 99.05 | 109 | 3 | 2 | 3 | 1 | 1 |
| [101] | Yap et al. 2023 | 5–45 M | - | 2 s | 30 epochs | - | 4.5 s | GTX 1080 Ti | 80.00 | 30 | 14 | 2 | 1 | 2 | 4 |
| [102] | Chen et al. 2024 | - | - | - | - | - | 2 s | - | 93.81 | 1 | 56 | 4 | 3 | 3 | 3 |
| [103] | Shakir et al. 2024 | - | - | - | - | - | 1 s | - | 95.00 | 109 | 3 | 2 | 2 | 1 | 3 |
| [91] | Lakhan et al. 2025 | - | - | - | 20 epochs | - | - | NVIDIA Tesla V100GPU | 99.26 | 54 | 8 | 4 | 3 | 2 | 1 |
| Ref. | Author/Year | Parameters (K/M) | MACs/FLOPs (M/G) | Latency (s) | Training Time Epochs (s/m/h) | Memory Footprint | Epoch Length (Segment) (ms/s) | GPU TPU Cloud | Acc. (%) | Sample Size | Channels | Complex. | Comp. Cost | Oper. Cost | Train. Cost |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [38] | Arjun et al. 2021 | - | - | - | - | - | 6 s | - | 99.40 | 32 | 32 | 1 | 1 | 1 | 2 |
| [34] | Dosovitskiy et al. 2021 | - | - | - | - | - | 14–300 M image patches | TPU v3 core days | 94.55 | - | - | 3 | 2 | 3 | 4 |
| [39] | Song et al. 2021 | 6.50–8.68 K | - | - | - | - | Small segments | - | 84.26 | 9 | 3 | 1 | 1 | 1 | 1 |
| [40] | Tao et al. 2021 | - | - | - | - | - | 20–460 ms | - | 61.11 | 6 | 64 | 3 | 2 | 2 | 2 |
| [41] | Siddhad et al. 2024 | - | - | - | - | - | - | - | 95.28 | 48 | 14 | 2 | 2 | 1 | 2 |
| [42] | Du et al. 2022 | - | - | - | - | - | 1s | - | 97.90 | 109 | 64 | 2 | 2 | 1 | 2 |
| [43] | Zeynali et al. 2023 | - | - | - | 1000 epochs | - | - | - | 96.10 | 8 | 64 | 4 | 3 | 3 | 5 |
| [44] | Omair et al. 2024 | - | - | - | - | Latent dim. 100 | 150 time stamps | - | 85.00 | 9 | 3 | 3 | 3 | 2 | 2 |
| [45] | Lim et al. 2025 | (Embed. size 256) | - | - | <1 day | >24 GB | 2 s | RTX 4090 GPU | 90.84 | 1 | 3 | 5 | 5 | 4 | 4 |
| [46] | Wang et al. 2024 | 10–101 M | - | - | 200 epochs | - | 4 s | 8 NVIDIA 3090s GPUs | 80.59 | 9 | 3 | 5 | 5 | 5 | 5 |
| [108] | Hu et al. 2024 | - | - | - | 100 epochs | - | - | NVIDA TESLA T4 Tensor Core GPU | 99.12 | 15 | 32 | 4 | 3 | 2 | 3 |
| [109] | Muna et al. 2025 | - | - | - | 20 epochs | - | - | CUDA Cloud | 76.83 | 9 | 22 | 3 | 3 | 3 | 1 |
| [111] | Ghous et al. 2025 | - | - | - | 50 epochs | - | - | - | 95.00 | 15 | 62 | 4 | 4 | 4 | 3 |
5.1.3. Efficiency Insights: CNN–Transformer Hybrids
- Acc. vs. Complex. and Comp. Cost: Models with high parameter counts deliver higher accuracy at the expense of computational efficiency. For instance, architectures with up to 23.55 M [131] or even 369 M [133] parameters achieved strong performance but required more computation. In contrast, Ref. [117] achieved very high Acc. (97.81%) with a very low parameter count (24.9–25.7 K), indicating high architectural efficiency. Similarly, Ref. [128] is noted for its small memory footprint and low parameter count (115.2 K), making it suitable for resource-constrained applications or embedded BCI applications. The models in [120,126] show that longer EEG data segments or window lengths generally increase Acc. but also raise the computational complexity.
- Oper. Cost vs. Train. Cost: Some models optimize real-time usability at the expense of training overhead. The model in [130] showed a low latency of 0.0043 s with high Acc. (99.67%), showing a strong design for real-time operation (low Oper. Cost). Others [129] had up to 5000 epochs, indicating a high Train. Cost.
5.1.4. Efficiency Insights: Recurrent Hybrids
- Acc. vs. Complex. and Comp. Cost: Multimodal EEG systems increase user complexity yet yield high security. Ref. [144] boosted the Acc. from 97.57% (unimodal) to 98.78% (EEG and signature). Similarly, Ref. [136] achieved 99.57% overall Acc. by combining EEG and gait. Advanced models can reduce the computational burden without compromising Acc. Ref. [143] used tensor-train decomposition for computational efficiency gains. It required only 1.6 K parameters for classification, compared to a traditional model at 1.28 M parameters. This led to a reduced memory footprint with 99.50% Acc.
- Acc. vs. Oper. Costs: Channels’ dimensionality reduction increases efficiency and user practicality while maintaining high Acc. Refs. [139,140] demonstrated high Acc. (99.58% and 98.00%, respectively) using a minimal number of four channels. Ref. [138] achieved a 100% CRR with 32 channels and still maintained a 99.17% CRR when reduced to five, making the system practical and efficient. Ref. [137] found the optimal efficiency–Acc. balance at 48 channels out of 64. An operational burden during data collection is sometimes accepted for FE gains to improve both the signal-to-noise ratio and Acc. Ref. [142] accepted a very high Oper. Cost, requiring over 50,000 trials, for an enhanced feature space. This resulted in a 30.09% improvement in classification Acc. over comparable methods.
- Oper. Cost vs. Train. Cost: A few authors have accepted high Train. Costs in exchange for very low Oper. Costs. Ref. [135] had a long training time in exchange for less than 1 s latency for a better authentication decision and practical deployment. Ref. [139] achieved lower batch testing latency of 0.065 s. Ref. [145] used one-shot learning training on as few as six pairs. It resulted in a reduced initial Oper. Cost for user enrollment despite a long total training time (870 min).
| Ref. | Author/Year | Parameters (K/M) | MACs/FLOPs (M/G) | Latency (s) | Training Time Epochs (s/m/h) | Memory Footprint | Epoch Length (Segment) (ms/s) | GPU TPU Cloud | Acc. (%) | Sample Size | Channels | Complex. | Comp. Cost | Oper. Cost | Train. Cost |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [113] | Sun et al. 2021 | - | - | - | - | - | - | - | 87.80 | 109 | 64 | 3 | 3 | 2 | 3 |
| [131] | Kostas et al. 2021 | - | Quadratic | - | - | - | - | - | 86.70 | >10,000 | 20 | 5 | 5 | 4 | 3 |
| [124] | Bagchi et al. 2022 | 4.56–23.55 M | Quadratic | - | 35–80 epochs | - | - | - | 89.64 | 10 | 128 | 5 | 5 | 4 | 3 |
| [120] | Xie et al. 2022 | - | - | - | - | - | - | - | 83.31 | 109 | 64 | 3 | 3 | 3 | 3 |
| [128] | Altaheri et al. 2023 | 115.2 K | - | - | - | Small | - | - | 85.38 | 9 | 22 | 1 | 1 | 1 | 2 |
| [132] | Yang et al. 2023 | 86 M | - | - | 15 epochs | - | - | - | 61.70 | 27 | 128 | 4 | 4 | 3 | 2 |
| [119] | Li et al. 2023 | - | - | - | - | - | - | - | 96.65 | 9 | 3 | 4 | 3 | 3 | 3 |
| [125] | Song et al. 2023 | - | - | 0.27 | - | - | - | GPU | 95.30 | 9 | 3 | 3 | 3 | 2 | 2 |
| [115] | Wan et al. 2023 | Avoids huge complexity | - | - | - | - | - | - | 92.75 | 12 | 6 | 3 | 2 | 2 | 3 |
| [121] | Si et al. 2023 | Hybrid: slightly lower | - | - | - | - | 14 s | - | 42.50 | 6 | 30 | 4 | 3 | 3 | 3 |
| [127] | Gong et al. 2023 | - | - | - | - | - | - | - | 95.30 | 15 | 62 | 3 | 3 | 3 | 3 |
| [116] | Ma et al. 2023 | - | - | - | 200 epochs | - | - | - | 83.91 | 9 | 22 | 3 | 3 | 3 | 4 |
| [114] | Omair et al. 2024 | - | - | - | 100 epochs | - | - | - | 86.98 | 9 | 3 | 3 | 3 | 3 | 3 |
| [126] | Si et al. 2024 | - | - | - | - | - | 14 s Optimal | - | 67.80 | 32 | 28 | 3 | 4 | 3 | 3 |
| [117] | Zhao et al. 2024 | 24.9–25.7 K | - | - | - | - | - | RTX3090 | 97.81 | 9 | 3 | 2 | 2 | 2 | 3 |
| [133] | Jiang et al. 2024 | 5.8–369 M | - | - | Fine tuning costly | Costly | 1 s | - | 82.58 | >140 | 19 | 5 | 5 | 5 | 5 |
| [118] | Liu et al. 2024 | - | - | - | - | - | - | - | 74.23 | 16 | 32 | 2 | 2 | 2 | 3 |
| [122] | Yao et al. 2024 | - | - | - | - | - | 3 s | - | 96.95 | 15 | 32 | 3 | 3 | 2 | 3 |
| [134] | Li et al. 2024 | - | - | - | 15 epochs | - | - | RTX4090 | - | 356 | 128 | 3 | 2 | 2 | 2 |
| [123] | Lu et al. 2024 | - | - | - | - | - | - | - | 98.57 | 15 | 62 | 4 | 3 | 3 | 3 |
| [129] | Nguyen et al. 2024 | - | - | - | Up to 5000 epochs | - | - | - | 83.41 | 9 | 22 | 3 | 3 | 3 | 5 |
| [130] | Cheng et al. 2024 | Linear complexity | - | 0.0043 | 200–300 epochs | - | 2–17 s | 2080Ti | 99.67 | 15 | 32 | 2 | 1 | 1 | 4 |
| Ref. | Author/Year | Parameters (K/M) | MACs/FLOPs (M/G) | Latency (s) | Training Time Epochs (s/m/h) | Memory Footprint | Epoch Length (Segment) (ms/s) | GPU/TPU/Cloud | Acc. (%) | Sample Size | Channels | Complex. | Comp. Cost | Oper. Cost | Train. Cost |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [144] | Kumar et al. 2019 | - | - | - | - | - | - | - | 97.57 | 33 | 14 | 4 | 2 | 4 | 2 |
| [141] | Puengdang et al. 2019 | - | - | - | 28.5 m 30–50 epochs | - | - | - | 91.44 | 20 | 6 | 3 | 2 | 3 | 1 |
| [142] | Zheng et al. 2020 | - | - | - | - | - | - | - | 66.81 | 10 | 128 | 3 | 4 | 5 | 3 |
| [145] | Chakladar et al. 2021 | - | - | - | 870 m 150 epochs | - | - | - | 98.57 | 70 | 14 | 4 | 3 | 3 | 5 |
| [138] | Wilaiprasitporn et al. 2015 | - | - | - | Fast | - | - | - | 99.90 | 32 | 5 | 4 | 2 | 2 | 2 |
| [135] | Zhang et al. 2018 | - | - | <1 | Increased | - | - | - | 99.89 | 8 | 14 | 4 | 1 | 3 | 4 |
| [139] | Sun et al. 2019 | - | - | 0.065 | Long | - | - | GPU | 99.58 | 109 | 4 | 4 | 1 | 1 | 4 |
| [136] | Zhang et al. 2020 | - | - | 0.39 | - | - | - | - | 99.00 | 7 | 14 | 5 | 2 | 5 | 3 |
| [143] | Jin et al. 2021 | 1.6 K (vs. 1.28 M) | - | - | - | Reduced | - | - | 99.50 | 20 | 7 | 5 | 1 | 2 | 2 |
| [140] | Chakravarthi et al. 2022 | - | - | - | - | - | - | - | 98.00 | 20 | 4 | 4 | 3 | 1 | 3 |
| [137] | Balci et al. 2023 | - | - | - | Long | - | - | - | 99.97 | 109 | 48 | 4 | 3 | 4 | 4 |
5.2. Comprehensive Weighted Sum Model
5.2.1. CNN Heatmap Analysis
- Acc. vs. Oper. Cost: In S1, a high weight on Oper. Cost means accepting constraints like low channel counts or very short epoch lengths. The most balanced models [79,80] achieved very high Acc. with very short epoch lengths (1 s) and efficient fine tuning. Ref. [104] scores highly in S1, but its 7 s latency shows a poor speed/operational trade-off, reducing its score despite its good Acc.
- Acc. vs. Complex.: In S2, a high weight for Complex. means sacrificing peak Acc. or longer training. The model in Ref. [12] is designed to be ultra-compact (1.066 k parameters), resulting in a high score for S2 but with modest Acc. of 91.00% and a high number of training epochs (500).
- Acc. vs. Training Cost: A heavy Train. Cost is necessary for a well-performing model. Studies like [81,83] have the lowest scores due to long training, which means that a high Train. Cost does not guarantee high operational efficiency. However, models like that in [97], with a high number of training epochs (1000), achieved excellent latency, showing a trade-off between training effort and post-deployment speed.
5.2.2. Transformer Heatmap Analysis
- Acc. vs. Oper. Cost: This trade-off is critical in S1, as it determines whether an Acc. gain justifies added preprocessing Complex. and slowness. Some studies [38,41,42] use raw EEG signals and achieve the best overall performance with a minimal Oper. Cost. Ref. [38] best embodies an efficient lightweight solution that balances Complex., Oper., and Comp. Costs. In contrast, Ref. [43] combined raw temporal and spectral (PSD) features, which resulted in a slight increase in Acc. but incurred a higher Oper. Cost due to the added computation.
- Acc. vs. Complex.: This trade-off is central for S2, where the model size and computational load are minimized. The tiny model in [39] had few parameters and obtained the best Complex. This shows that attention mechanisms can be effective at low parameter counts. This model trades Acc. (84.26%) for very high efficiency. On the other hand, Ref. [46] demonstrates a large-model penalty, with millions of parameters and high Complex. While its size boosts its performance via scalability, the resulting high Comp. Cost cancels out any Acc. benefits.
- Acc. vs. Training Cost: This trade-off reflects the computational and time resources required to train a model, heavily influencing the all-rounder efficiency. The studies in [43,45] represent the resource-heavy end of the spectrum, both requiring 1000 training epochs. This high demand for training time indicates that these models need a high development cost to reach high Acc. In contrast, Ref. [109] achieved fast convergence, completing training in only 20 epochs. This reflects a highly efficient training process, balancing a low development cost with moderate Acc. (76.83%).
5.2.3. CNN–Transformer Heatmap Analysis
- Acc. vs. Total Cost Efficiency: The model in [118] demonstrates the maximum trade-off in favor of efficiency. It achieves good but not the best raw Acc. of 85.38%, which is compensated for by a zero-cost profile across Complex., Comp. Cost, and Oper. Cost. The model’s minimal resource footprint makes it the ideal edge solution.
- Acc. vs. Oper. Cost: The model in [130] represents the most desirable trade-off. It provides peak performance across the dataset with the highest raw Acc. of 99.67%, while simultaneously demonstrating minimal latency due to its zero-cost Oper. Cost profile. This secures its status as the optimal real-time solution.
- Acc. vs. Complex. and Oper. Cost: The model in [129] clearly prioritizes performance, as shown by its near-best raw Acc. of 98.08%. However, this high performance comes at the cost of its resource profile, with only average performance in Complex. and Oper. Cost. It can be chosen only when high Acc. is mandatory and the system can tolerate its high resource utilization. Overall, it is less efficient than the two top models [118,130].
- Acc. vs. Complex. and Comp. Cost: The model in [133] demonstrates a poor trade-off. It yields modest raw Acc. of 70.11% with the maximum resources across the Complex. and Comp. Costs. It proves the diminishing returns of architectural scaling that do not translate into superior performance. This is the least practical choice in constrained environments.
5.2.4. Recurrent Hybrid Heatmap Analysis
- Acc. vs. Oper./Comp. Cost: In S1, the model in [139] ranks first and achieves 99.58% Acc. by minimizing resource use during operation. It secures efficiency by using only four channels and has the fastest testing latency at 1 s. This performance comes at the cost of a long training time due to its complex recurrent architecture. Ref. [140] ranks second and achieves 98.00% Acc. and uses four channels, but its Comp. Cost score is lower because of its latency of 1.7 s.
- Acc. vs. Complex./Train. Cost: In S2, the model in [141] ranks first due to having the lowest Acc. among the top tier (91.44%) in exchange for being the most resource-efficient system. It requires only minutes of training time. In contrast, Ref. [139] ranks second and maintains 99.58% Acc. but its recurrent architecture results in high Complex., increasing its memory footprint.
- Acc. vs. Total System Cost: In S3, Ref. [139] ranks first, achieving 99.58% Acc. and combining this with superior operational efficiency (four channels and 1 s latency). Its score confirms that its fast low-channel operation outweighs the penalty of its long initial training time. Ref. [138] ranks second and achieves high Acc. (99.17%) and reliable performance across all cost metrics, with six channels and 1.25 s latency. This allows it to avoid the extreme cost trade-offs of the rank-1 models.
5.3. The Efficiency Frontier
5.3.1. Trade-off Analysis
5.3.2. Architectural Insights
5.3.3. Scenario Performance Insights
5.4. The Pareto Frontier

5.5. Performance vs. Total Real Cost Analysis
6. Analysis of Design Trends
6.1. CNN Design Patterns
6.1.1. Depthwise Separable Convolutions and Compact CNNs
6.1.2. Outperforming Traditional Methods
6.1.3. Multiscale and Multibranch Architectures
6.1.4. Residual Connections and Feature Fusion
6.1.5. Attention Mechanisms
6.1.6. Adversarial and Transfer Learning
6.1.7. Real-Time, Embedded Focus and Cloud/Edge Deployment
6.1.8. Lightweight 1D/2D/3D CNNs
6.2. Transformer Design Patterns
6.2.1. Raw Signal Use with Minimal Preprocessing
6.2.2. Time–Frequency Transform Integration
6.2.3. Modular Attention Architectures (Spatiotemporal Separation)
6.2.4. Attention-Enhanced Gating and Graph Mechanisms
6.2.5. Self-Supervised and Generative Pretraining
6.2.6. Specialized Tokenization and Representation Learning
6.3. CNN–Transformer Architecture Design Patterns
6.3.1. Dual-Branch Fusion Strategies
6.3.2. CNN Precompression for Efficient Attention
6.3.3. Attention Optimization
6.3.4. Smart Feature Enhancement
6.3.5. Interactive Learning Modules
6.4. Recurrent Hybrid Design Patterns
6.4.1. Frequency-Focused Fusion
6.4.2. Multimodal Fusion with Recurrent Backbones
6.4.3. Two-Stage Pipelines
6.4.4. Lightweight and Portable Personalization
6.4.5. Efficiency-Driven Architectures
6.4.6. Multitask and Feature Generalization
7. Findings and Discussion
7.1. Domains of Application
7.1.1. Trends over Time—Architectures
7.1.2. Architecture–Domain Mapping
7.1.3. Performance–Efficiency Trade-Offs
7.1.4. Notable Trends in the Latest Period
7.2. Future Directions
7.2.1. Bridging the Performance–Efficiency Gap
- Developing novel Transformer and CNN–Transformer architectures that maintain very high performance and improve efficiency beyond media. This could be achieved by (1) exploring knowledge distillation from complex and less efficient models to smaller and faster ones; (2) implementing sparsity techniques in Transformer layers; and (3) researching hardware-aware network designs specific to BCI/EEG applications.
- Given the stability of foundational CNNs, revising and optimizing lightweight and high-performing CNN variants that are deployable on low-power devices.
7.2.2. Expanding Domain Specialization and Generalization
- Increase research attention to the ER domain, which has emerged in the most recent period. The latest models show promising high performance/high efficiency in this area, revealing that it is an impactful research domain.
- Invest in the GF domain, using Transformer and CNN–Transformer architectures to reduce the need for domain-specific models. The goal should be to build models that can achieve high performance across multiple domains (Bio, MI, EP) without extensive retraining.
- Explore novel and niche BCI domains outside personalized medicine with newer architectures to see if the performance gains translate to these areas.
7.2.3. Deeper Analysis of Architecture Components
- Perform ablation studies on hybrid CNN–Transformers by isolating the contributions of the CNN part vs. the Transformer part across different domains. This will determine the optimal split to maximize performance and efficiency.
- Standardizing performance and efficiency metrics since the use of low to high scales is relative. Future research should adopt quantitative standardized metrics for reporting to allow for the rigorous and fair comparison of architectures.
7.2.4. Longitudinal Studies and Reproducibility
- It is necessary to conduct studies to track architectural lifecycles (how long a design remains relevant)—for instance, investigate whether the efficiency gains seen in early CNNs can be replicated with modern training techniques on new architectures.
- Since the initial focus has been on peak performance, future work should prioritize measuring robustness and generalization. Very high performance is less valuable if it is not reproducible by other researchers.
- It is important to develop foundational models that are pretrained on low-density EEG to support robust edge deployment.
7.2.5. Generative Models as the Next Frontier
7.3. Limitations
- Performance metrics—Acc., EER, AUC, CRR, etc.;
- Model size—parameters in K or M;
- Computational complexity—MACs or FLOPs per inference window;
- Memory footprint—memory usage at inference, including weights and activations;
- Inference latency—measured on embedded, mobile, desktop, or server systems;
- Training details—training time, number of epochs, batch size, and hardware used;
- Validation protocol—within/cross-subject or cross-session and number of subjects;
- Operational setup—number of EEG channels, epoch duration, and calibration/enrollment time per subject/session.
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| Acc. | Accuracy |
| ANN | Artificial Neural Network |
| AR | Autoregressive |
| AUC | Area Under the Curve |
| BCI | Brain–Computer Interface |
| BLSTM-NN | Bidirectional Long Short-Term Memory Neural Network. |
| Borda Count | Rank-Based Aggregation Method |
| CAR | Common Average Referencing |
| CCA | Canonical Correlation Analysis |
| CNN | Convolutional Neural Network |
| CRR | Correct Recognition Rate |
| CS | Cosine Similarity |
| CSP | Common Spatial Patterns |
| CWT | Continuous Wavelet Transform |
| DBN | Deep Belief Network |
| DE | Differential Entropy |
| DL | Deep Learning |
| DML | Deep Metric Learning |
| ECG | Electrocardiography |
| EEG | Electroencephalography |
| EER | Equal Error Rate |
| EOG | Electrooculography |
| FAR | False Acceptance Rate |
| FBCSP | Filter Bank Common Spatial Patterns |
| FC | Fully Connected layer |
| FE | Feature Extraction |
| FLOPs | Floating-Point Operations Per Second |
| FRR | False Rejection Rate |
| FuzzyEn | Fuzzy Entropy |
| GAN | Generative Adversarial Network |
| GAT | Graph Attention Network |
| GET | Generative EEG Transformer |
| GFCC | Gammatone Frequency Cepstral Coefficient |
| GNN | Graph Neural Network |
| GRU | Gated Recurrent Unit |
| GSO | Gram–Schmidt Orthogonalization |
| GSR | Galvanic Skin Response |
| GT | Gated Transformer |
| HTER | Half Total Error Rate |
| ITR | Information Transfer Rate |
| LSTM | Long Short-Term Memory |
| MACs | Millions of Multiply–Accumulate Operations |
| Max Rule | A Decision Fusion Strategy Selecting the Maximum Score |
| MFCC | Mel-Frequency Cepstral Coefficient |
| MI | Motor Imagery |
| ML | Machine Learning |
| MLP | Multilayer Perceptron |
| MSE | Mean Squared Error |
| NN | Nearest Neighbor |
| PLV | Phase Locking Value |
| PSD | Power Spectral Density |
| RF | Random Forest |
| RMSprop | Root Mean Square Propagation |
| RNN | Recurrent Neural Network |
| R/S Analysis | Rescaled Range Analysis |
| SA | Self-Attention |
| SAFE | Spatial Attention Feature Extractor |
| SE | Squeeze-and-Excitation |
| SMOTE | Synthetic Minority Oversampling Technique |
| SNN | Siamese Neural Network |
| SPA | Spectral Power Analysis |
| STE | Spatial Transformer Encoder |
| STFT | Short-Time Fourier Transform |
| TTNN | Tensor-Train Neural Network |
| TAFE | Temporal Attention Feature Extractor |
| TTE | Temporal Transformer Encoder |
| TCN | Temporal Convolutional Network |
| ViT | Vision Transformer |
| WT | Wavelet Transform |
| XGB | XGBoost |
| BCI Competition IV-2a | Motor Imagery EEG Benchmark Dataset (22 channels, 9 subjects). |
| CD FTA | Cross-Dataset Fine-Tuning Adaptation. Benchmarking for adapting EEG models across datasets. |
| DEAP | Dataset for Emotion Analysis using EEG and peripheral Physiological signals during video watching. |
| DREAMER | Dataset for emotion analysis using EEG and ECG while subjects watched affective videos. |
| EEGMMIDB | Large-scale PhysioNet dataset for EEG Motor Movement/Imagery Database. |
| HGD | Gamma dataset. |
| ImageNet | Large-scale visual dataset (1.2 M images, 1000 categories), widely used for pretraining. |
| ImageNet 21k | Extended ImageNet with 21,000 categories for large-scale vision pretraining. |
| JFT 300 M | Google’s large-scale proprietary dataset of 300 M images, used for pretraining. |
| MOABB | Mother of All BCI Benchmarks. A standardized benchmarking framework for EEG datasets. |
| MPED | Multimodal Physiological Emotion Database. EEG and physiological modalities for emotion recognition. |
| PhysioNet EEG | Collection of EEG datasets on PhysioNet for various clinical and cognitive studies. |
| SEED-IV | SJTU Emotion EEG Dataset (four-class emotion recognition for 15 participants). |
| SEED-V | SJTU Emotion EEG Dataset (five-class emotion recognition for 16 participants). |
| SMR BCI | Sensorimotor Rhythm Brain–Computer Interface Dataset. Longitudinal MI dataset (600 h, 62 participants). |
References
- Teplan, M. Fundamentals of EEG Measurement. Meas. Sci. Rev. 2002, 2, 1–11. Available online: https://www.measurement.sk/2002/S2/Teplan.pdf (accessed on 15 September 2025).
- Craik, A.; He, Y.; Contreras-Vidal, J.L. Deep Learning for Electroencephalogram (EEG) Classification Tasks: A Review. J. Neural Eng. 2019, 16, 031001. [Google Scholar] [CrossRef] [PubMed]
- Lin, F.; Cho, K.W.; Song, C.; Xu, W.; Jin, Z. Brain Password: A Secure and Truly Cancelable Brain Biometrics for Smart Headwear. In Proceedings of the 16th Annual International Conference on Mobile Systems, Applications, and Services, Munich, Germany, 10–15 June 2018; ACM: New York, NY, USA, 2018; pp. 296–309. [Google Scholar] [CrossRef]
- Yu, Y.-C.; Wang, S.; Gabel, L.A. A Feasibility Study of Using Event-Related Potential as a Biometrics. In Proceedings of 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; IEEE: New York, NY, USA, 2016; pp. 4547–4550. [Google Scholar] [CrossRef]
- Singh, A.K.; Krishnan, S. Trends in EEG Signal Feature Extraction Applications. Front. Artif. Intell. 2023, 5, 1072801. [Google Scholar] [CrossRef] [PubMed]
- Ramoser, H.; Muller-Gerking, J.; Pfurtscheller, G. Optimal Spatial Filtering of Single Trial EEG during Imagined Hand Movement. IEEE Trans. Rehab. Eng. 2000, 8, 441–446. [Google Scholar] [CrossRef]
- Subasi, A. EEG Signal Classification Using Wavelet Feature Extraction and a Mixture of Expert Model. Expert Syst. Appl. 2007, 32, 1084–1093. [Google Scholar] [CrossRef]
- Roy, Y.; Banville, H.; Albuquerque, I.; Gramfort, A.; Falk, T.H.; Faubert, J. Deep Learning-Based Electroencephalography Analysis: A Systematic Review. J. Neural Eng. 2019, 16, 051001. [Google Scholar] [CrossRef] [PubMed]
- Takahashi, S.; Sakaguchi, Y.; Kouno, N.; Takasawa, K.; Ishizu, K.; Akagi, Y.; Aoyama, R.; Teraya, N.; Bolatkan, A.; Shinkai, N.; et al. Comparison of Vision Transformers and Convolutional Neural Networks in Medical Image Analysis: A Systematic Review. J. Med. Syst. 2024, 48, 84. [Google Scholar] [CrossRef]
- Sun, C.; Mou, C. Survey on the Research Direction of EEG-Based Signal Processing. Front. Neurosci. 2023, 17, 1203059. [Google Scholar] [CrossRef]
- Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep Learning with Convolutional Neural Networks for EEG Decoding and Visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef]
- Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A Compact Convolutional Neural Network for EEG-Based Brain–Computer Interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
- Vafaei, E.; Hosseini, M. Transformers in EEG Analysis: A Review of Architectures and Applications in Motor Imagery, Seizure, and Emotion Classification. Sensors 2025, 25, 1293. [Google Scholar] [CrossRef]
- Zhang, X.; Yao, L.; Wang, X.; Monaghan, J.; McAlpine, D.; Zhang, Y. A Survey on Deep Learning-Based Non-Invasive Brain Signals: Recent Advances and New Frontiers. J. Neural Eng. 2021, 18, 031002. [Google Scholar] [CrossRef]
- Fu, J. A Comparison of CNN and Transformer in Continual Learning. Master’s Thesis, KTH Royal Institute of Technology, Stockholm, Sweden, 2023. Available online: https://kth.diva-portal.org/smash/get/diva2:1820229/FULLTEXT01.pdf (accessed on 16 September 2025).
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015; Available online: https://ora.ox.ac.uk/objects/uuid:60713f18-a6d1-4d97-8f45-b60ad8aebbce (accessed on 27 September 2025).
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: New York, NY, USA, 2015; pp. 1–9. [Google Scholar] [CrossRef]
- Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: New York, NY, USA, 2017; pp. 5987–5995. [Google Scholar] [CrossRef]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: New York, NY, USA, 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
- Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. Available online: http://proceedings.mlr.press/v97/tan19a.html (accessed on 16 September 2025).
- Radosavovic, I.; Kosaraju, R.P.; Girshick, R.; He, K.; Dollar, P. Designing Network Design Spaces. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; IEEE: New York, NY, USA, 2020; pp. 10425–10433. [Google Scholar] [CrossRef]
- Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; IEEE: New York, NY, USA, 2022; pp. 11966–11976. [Google Scholar] [CrossRef]
- Chen, G.; Zhang, X.; Zhang, J.; Li, F.; Duan, S. A Novel Brain-Computer Interface Based on Audio-Assisted Visual Evoked EEG and Spatial-Temporal Attention CNN. Front. Neurorobot. 2022, 16, 995552. [Google Scholar] [CrossRef] [PubMed]
- Mai, N.-D.; Hoang Long, N.M.; Chung, W.-Y. 1D-CNN-Based BCI System for Detecting Emotional States Using a Wireless and Wearable 8-Channel Custom-Designed EEG Headset. In Proceedings of the 2021 IEEE International Conference on Flexible and Printable Sensors and Systems (FLEPS), Manchester, UK, 20–23 June 2021; IEEE: New York, NY, USA, 2021; pp. 1–4. [Google Scholar] [CrossRef]
- Huang, J.; Wang, C.; Zhao, W.; Grau, A.; Xue, X.; Zhang, F. LTDNet-EEG: A Lightweight Network of Portable/Wearable Devices for Real-Time EEG Signal Denoising. IEEE Trans. Consum. Electron. 2024, 70, 5561–5575. [Google Scholar] [CrossRef]
- Borra, D.; Magosso, E. Deep Learning-Based EEG Analysis: Investigating P3 ERP Components. J. Integr. Neurosci. 2021, 20, 791–811. [Google Scholar] [CrossRef]
- Ramakrishnan, K.; Groen, I.I.A.; Smeulders, A.W.M.; Scholte, H.S.; Ghebreab, S. Deep Learning for EEG-Based Brain Mapping. bioRxiv 2017, 178541. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Available online: https://papers.nips.cc/paper/7181-attention-is-all-you-need (accessed on 23 September 2025).
- Lin, T.; Wang, Y.; Liu, X.; Qiu, X. A Survey of Transformers. arXiv 2021, arXiv:2106.04554. [Google Scholar] [CrossRef]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. OpenAI. 2018. Available online: https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 17 September 2025).
- Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
- Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 1–67. Available online: https://jmlr.org/papers/v21/20-074.html (accessed on 22 September 2025).
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16×16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 3–7 May 2021; Available online: https://openreview.net/forum?id=YicbFdNTTy (accessed on 19 September 2025).
- Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. PaLM: Scaling Language Modeling with Pathways. J. Mach. Learn. Res. 2023, 24, 1–113. Available online: https://jmlr.org/papers/volume24/22-1144/22-1144.pdf (accessed on 21 September 2025).
- Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. Meta AI. 2023. Available online: https://ai.meta.com/research/publications/llama-open-and-efficient-foundation-language-models (accessed on 25 September 2025).
- Google Developers/Google DeepMind. Introducing Gemini: Google’s Most Capable AI Model Yet. Google Blog 6 December 2023. Available online: https://blog.google/technology/ai/google-gemini-ai/ (accessed on 20 September 2025).
- Arjun, A.; Rajpoot, A.S.; Raveendranatha, P.M. Introducing Attention Mechanism for EEG Signals: Emotion Recognition with Vision Transformers. In Proceedings of 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 1–5 November 2021; IEEE: New York, NY, USA, 2021; pp. 5723–5726. [Google Scholar] [CrossRef]
- Song, Y.; Jia, X.; Yang, L.; Xie, L. Transformer-Based Spatial-Temporal Feature Learning for EEG Decoding. arXiv 2021, arXiv:2106.11170. [Google Scholar] [CrossRef]
- Tao, Y.; Sun, T.; Muhamed, A.; Genc, S.; Jackson, D.; Arsanjani, A.; Yaddanapudi, S.; Li, L.; Kumar, P. Gated Transformer for Decoding Human Brain EEG Signals. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 1–5 November 2021; IEEE: New York, NY, USA, 2021; pp. 125–130. [Google Scholar] [CrossRef]
- Siddhad, G.; Gupta, A.; Dogra, D.P.; Roy, P.P. Efficacy of Transformer Networks for Classification of EEG Data. Biomed. Signal Process. Control 2024, 87, 105488. [Google Scholar] [CrossRef]
- Du, Y.; Xu, Y.; Wang, X.; Liu, L.; Ma, P. EEG Temporal–Spatial Transformer for Person Identification. Sci. Rep. 2022, 12, 14378. [Google Scholar] [CrossRef]
- Zeynali, M.; Seyedarabi, H.; Afrouzian, R. Classification of EEG Signals Using Transformer Based Deep Learning and Ensemble Models. Biomed. Signal Process. Control 2023, 86, 105130. [Google Scholar] [CrossRef]
- Omair, A.; Saif-ur-Rehman, M.; Metzler, M.; Glasmachers, T.; Iossifidis, I.; Klaes, C. GET: A Generative EEG Transformer for Continuous Context-Based Neural Signals. arXiv 2024, arXiv:2406.03115. [Google Scholar] [CrossRef]
- Lim, J.-H.; Kuo, P.-C. EEGTrans: Transformer-Driven Generative Models for EEG Synthesis. In Proceedings of the 13th International Conference on Learning Representations (ICLR 2025), Singapore, 24–28 April 2025; Available online: https://openreview.net/forum?id=ydw2l8zgUB (accessed on 18 September 2025).
- Wang, G.; Liu, W.; He, Y.; Xu, C.; Ma, L.; Li, H. EEGPT: Pretrained Transformer for Universal and Reliable EEG Representation Learning. In Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 10–15 December 2024; pp. 1–12. [Google Scholar]
- Saeidi, M.; Karwowski, W.; Farahani, F.V.; Fiok, K.; Taiar, R.; Hancock, P.A.; Al-Juaid, A. Neural Decoding of EEG Signals with Machine Learning: A Systematic Review. Brain Sci. 2021, 11, 1525. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Zhang, Y.; Tiwari, P.; Song, D.; Hu, B.; Yang, M.; Zhao, Z.; Kumar, N.; Marttinen, P. EEG Based Emotion Recognition: A Tutorial and Review. ACM Comput. Surv. 2023, 55, 1–57. [Google Scholar] [CrossRef]
- Prabowo, D.W.; Nugroho, H.A.; Setiawan, N.A.; Debayle, J. A Systematic Literature Review of Emotion Recognition Using EEG Signals. Cogn. Syst. Res. 2023, 82, 101152. [Google Scholar] [CrossRef]
- Vempati, R.; Sharma, L.D. A Systematic Review on Automated Human Emotion Recognition Using Electroencephalogram Signals and Artificial Intelligence. Results Eng. 2023, 18, 101027. [Google Scholar] [CrossRef]
- Mohammed, S.A.; Jasim, S.S.; Thamir, B.A.; Alabdel Abass, A. A Survey on EEG Signal Analysis Using Machine Learning. UTJES 2025, 15, 89–106. [Google Scholar] [CrossRef]
- Gatfan, K.S. A Review on Deep Learning for Electroencephalogram Signal Classification. J. AI-Qadisiyah Comp. Sci. Math. 2024, 16, 137–151. [Google Scholar] [CrossRef]
- Suhaimi, N.S.; Mountstephens, J.; Teo, J. EEG-Based Emotion Recognition: A State-of-the-Art Review of Current Trends and Opportunities. Comput. Intell. Neurosci. 2020, 2020, 1–19. [Google Scholar] [CrossRef] [PubMed]
- Rahman, M.M.; Sarkar, A.K.; Hossain, M.A.; Hossain, M.S.; Islam, M.R.; Hossain, M.B.; Quinn, J.M.W.; Moni, M.A. Recognition of Human Emotions Using EEG Signals: A Review. Comput. Biol. Med. 2021, 136, 104696. [Google Scholar] [CrossRef] [PubMed]
- Khare, S.K.; Blanes-Vidal, V.; Nadimi, E.S.; Acharya, U.R. Emotion Recognition and Artificial Intelligence: A Systematic Review (2014–2023) and Research Recommendations. Inf. Fusion 2024, 102, 102019. [Google Scholar] [CrossRef]
- Jafari, M.; Shoeibi, A.; Khodatars, M.; Bagherzadeh, S.; Shalbaf, A.; García, D.L.; Gorriz, J.M.; Acharya, U.R. Emotion Recognition in EEG Signals Using Deep Learning Methods: A Review. Comput. Biol. Med. 2023, 165, 107450. [Google Scholar] [CrossRef] [PubMed]
- Ma, W.; Zheng, Y.; Li, T.; Li, Z.; Li, Y.; Wang, L. A Comprehensive Review of Deep Learning in EEG-Based Emotion Recognition: Classifications, Trends, and Practical Implications. PeerJ Comput. Sci. 2024, 10, e2065. [Google Scholar] [CrossRef]
- Gkintoni, E.; Aroutzidis, A.; Antonopoulou, H.; Halkiopoulos, C. From Neural Networks to Emotional Networks: A Systematic Review of EEG-Based Emotion Recognition in Cognitive Neuroscience and Real-World Applications. Brain Sci. 2025, 15, 220. [Google Scholar] [CrossRef] [PubMed]
- Al-Saegh, A.; Dawwd, S.A.; Abdul-Jabbar, J.M. Deep Learning for Motor Imagery EEG-Based Classification: A Review. Biomed. Signal Process. Control 2021, 63, 102172. [Google Scholar] [CrossRef]
- Ko, W.; Jeon, E.; Jeong, S.; Phyo, J.; Suk, H.-I. A Survey on Deep Learning-Based Short/Zero-Calibration Approaches for EEG-Based Brain–Computer Interfaces. Front. Hum. Neurosci. 2021, 15, 643386. [Google Scholar] [CrossRef]
- Pawan; Dhiman, R. Machine Learning Techniques for Electroencephalogram Based Brain-Computer Interface: A Systematic Literature Review. Meas. Sens. 2023, 28, 100823. [Google Scholar] [CrossRef]
- Saibene, A.; Ghaemi, H.; Dagdevir, E. Deep Learning in Motor Imagery EEG Signal Decoding: A Systematic Review. Neurocomputing 2024, 610, 128577. [Google Scholar] [CrossRef]
- Moreno-Castelblanco, S.R.; Vélez-Guerrero, M.A.; Callejas-Cuervo, M. Artificial Intelligence Approaches for EEG Signal Acquisition and Processing in Lower-Limb Motor Imagery: A Systematic Review. Sensors 2025, 25, 5030. [Google Scholar] [CrossRef]
- Wang, X.; Liesaputra, V.; Liu, Z.; Wang, Y.; Huang, Z. An In-Depth Survey on Deep Learning-Based Motor Imagery Electroencephalogram (EEG) Classification. Artif. Intell. Med. 2024, 147, 102738. [Google Scholar] [CrossRef]
- Hassan, J.; Reza, S.; Ahmed, S.U.; Anik, N.H.; Khan, M.O. EEG Workload Estimation and Classification: A Systematic Review. J. Neural Eng. 2025, 22, 051003. [Google Scholar] [CrossRef] [PubMed]
- de Bardeci, M.; Ip, C.T.; Olbrich, S. Deep Learning Applied to Electroencephalogram Data in Mental Disorders: A Systematic Review. Biol. Psychol. 2021, 162, 108117. [Google Scholar] [CrossRef]
- Nwagu, C.; AlSlaity, A.; Orji, R. EEG-Based Brain-Computer Interactions in Immersive Virtual and Augmented Reality: A Systematic Review. Proc. ACM Hum.-Comput. Interact. 2023, 7, 1–33. [Google Scholar] [CrossRef]
- Dadebayev, D.; Goh, W.W.; Tan, E.X. EEG-Based Emotion Recognition: Review of Commercial EEG Devices and Machine Learning Techniques. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 4385–4401. [Google Scholar] [CrossRef]
- Klepl, D.; Wu, M.; He, F. Graph Neural Network-Based EEG Classification: A Survey. arXiv 2023, arXiv:2310.02152. [Google Scholar] [CrossRef]
- Ma, L.; Minett, J.W.; Blu, T.; Wang, W.S.-Y. Resting State EEG-Based Biometrics for Individual Identification Using Convolutional Neural Networks. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; IEEE: New York, NY, USA, 2015; pp. 2848–2851. [Google Scholar] [CrossRef]
- Mao, Z.; Yao, W.X.; Huang, Y. EEG-Based Biometric Identification with Deep Learning. In Proceedings of the 2017 8th International IEEE/EMBS Conference on Neural Engineering (NER), Shanghai, China, 25–28 May 2017; IEEE: New York, NY, USA, 2017; pp. 609–612. [Google Scholar] [CrossRef]
- Schons, T.; Moreira, G.J.P.; Silva, P.H.L.; Coelho, V.N.; Luz, E.J.S. Convolutional Network for EEG-Based Biometric. In Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications; Mendoza, M., Velastín, S., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 10657, pp. 601–608. [Google Scholar] [CrossRef]
- Di, Y.; An, X.; Liu, S.; He, F.; Ming, D. Using Convolutional Neural Networks for Identification Based on EEG Signals. In Proceedings of the 2018 10th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China, 25–26 August 2018; IEEE: New York, NY, USA, 2018; pp. 119–122. [Google Scholar] [CrossRef]
- Zhang, F.Q.; Mao, Z.J.; Huang, Y.F.; Xu, L.; Ding, G.Y. Deep Learning Models for EEG-Based Rapid Serial Visual Presentation Event Classification. J. Inf. Hiding Multimed. Signal Process. 2018, 9, 177–187. [Google Scholar]
- Gonzalez, P.A.; Katsigiannis, S.; Ramzan, N.; Tolson, D.; Arevalillo-Herrez, M. ES1D: A Deep Network for EEG-Based Subject Identification. In Proceedings of the 2017 IEEE 17th International Conference on Bioinformatics and Bioengineering (BIBE), Washington, DC, USA, 23–25 October 2017; IEEE: New York, NY, USA, 2017; pp. 81–85. [Google Scholar] [CrossRef]
- Waytowich, N.; Lawhern, V.J.; Garcia, O.; Cummings, J.; Faller, J.; Sajda, P.; Vettel, J.M. Compact Convolutional Neural Networks for Classification of Asynchronous Steady-State Visual Evoked Potentials. J. Neural Eng. 2018, 15, 066031. [Google Scholar] [CrossRef]
- Yu, T.; Wei, C.-S.; Chiang, K.-J.; Nakanishi, M.; Jung, T.-P. EEG-Based User Authentication Using a Convolutional Neural Network. In Proceedings of the 2019 9th International IEEE/EMBS Conference on Neural Engineering (NER), San Francisco, CA, USA, 20–23 March 2019; IEEE: New York, NY, USA, 2019; pp. 1011–1014. [Google Scholar] [CrossRef]
- Lai, C.Q.; Ibrahim, H.; Abdullah, M.Z.; Abdullah, J.M.; Suandi, S.A.; Azman, A. Arrangements of Resting State Electroencephalography as the Input to Convolutional Neural Network for Biometric Identification. Comput. Intell. Neurosci. 2019, 2019, 1–10. [Google Scholar] [CrossRef] [PubMed]
- Wang, M.; El-Fiqi, H.; Hu, J.; Abbass, H.A. Convolutional Neural Networks Using Dynamic Functional Connectivity for EEG-Based Person Identification in Diverse Human States. IEEE Trans. Inform. Forensic Secur. 2019, 14, 3259–3272. [Google Scholar] [CrossRef]
- Wang, M.; Hu, J.; Abbass, H. Stable EEG Biometrics Using Convolutional Neural Networks and Functional Connectivity. Aust. J. Intell. Inf. Process. Syst. 2019, 15, 19–26. [Google Scholar]
- Zhang, R.; Zeng, Y.; Tong, L.; Shu, J.; Lu, R.; Li, Z.; Yang, K.; Yan, B. EEG Identity Authentication in Multi-Domain Features: A Multi-Scale 3D-CNN Approach. Front. Neurorobot. 2022, 16, 901765. [Google Scholar] [CrossRef]
- Das, R.; Maiorana, E.; Campisi, P. Visually Evoked Potential for EEG Biometrics Using Convolutional Neural Network. In Proceedings of the 2017 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; IEEE: New York, NY, USA, 2017; pp. 951–955. [Google Scholar] [CrossRef]
- Cecotti, H. Convolutional Neural Networks for Event-Related Potential Detection: Impact of the Architecture. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Seogwipo, Republic of Korea, 11–15 July 2017; IEEE: New York, NY, USA, 2017; pp. 2031–2034. [Google Scholar] [CrossRef]
- Cecotti, H.; Jha, G. 3D Convolutional Neural Networks for Event-Related Potential Detection. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: New York, NY, USA, 2019; pp. 4160–4163. [Google Scholar] [CrossRef]
- Chen, J.X.; Mao, Z.J.; Yao, W.X.; Huang, Y.F. EEG-Based Biometric Identification with Convolutional Neural Network. Multimed. Tools Appl. 2020, 79, 10655–10675. [Google Scholar] [CrossRef]
- Salami, A.; Andreu-Perez, J.; Gillmeister, H. EEG-ITNet: An Explainable Inception Temporal Convolutional Network for Motor Imagery Classification. IEEE Access 2022, 10, 36672–36685. [Google Scholar] [CrossRef]
- Bai, S.; Kolter, J.Z.; Koltun, V. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar] [CrossRef]
- Riyad, M.; Khalil, M.; Adib, A. Incep-EEGNet: A ConvNet for Motor Imagery Decoding. In Image and Signal Processing; El Moataz, A., Mammass, D., Mansouri, A., Nouboud, F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 12119, pp. 103–111. [Google Scholar] [CrossRef]
- Liu, X.; Shen, Y.; Liu, J.; Yang, J.; Xiong, P.; Lin, F. Parallel Spatial–Temporal Self-Attention CNN-Based Motor Imagery Classification for BCI. Front. Neurosci. 2020, 14, 587520. [Google Scholar] [CrossRef]
- Zhu, Y.; Peng, Y.; Song, Y.; Ozawa, K.; Kong, W. RAMST-CNN: A Residual and Multiscale Spatio-Temporal Convolution Neural Network for Personal Identification with EEG. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2021, E104.A, 563–571. [Google Scholar] [CrossRef]
- Lakhan, P.; Banluesombatkul, N.; Sricom, N.; Sawangjai, P.; Sangnark, S.; Yagi, T.; Wilaiprasitporn, T.; Saengmolee, W.; Limpiti, T. EEG-BBNet: A Hybrid Framework for Brain Biometric Using Graph Connectivity. IEEE Sens. Lett. 2025, 9, 1–4. [Google Scholar] [CrossRef]
- Ding, Y.; Robinson, N.; Zhang, S.; Zeng, Q.; Guan, C. TSception: Capturing Temporal Dynamics and Spatial Asymmetry from EEG for Emotion Recognition. IEEE Trans. Affect. Comput. 2023, 14, 2238–2250. [Google Scholar] [CrossRef]
- Salimi, N.; Barlow, M.; Lakshika, E. Towards Potential of N-back Task as Protocol and EEGNet for the EEG-based Biometric. In Proceedings of the 2020 IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, ACT, Australia, 1–4 December 2020; pp. 1718–1724. [Google Scholar] [CrossRef]
- Ingolfsson, T.M.; Hersche, M.; Wang, X.; Kobayashi, N.; Cavigelli, L.; Benini, L. EEG-TCNet: An Accurate Temporal Convolutional Network for Embedded Motor-Imagery Brain–Machine Interfaces. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; IEEE: New York, NY, USA, 2020; pp. 2958–2965. [Google Scholar] [CrossRef]
- Kasim, Ö.; Tosun, M. Biometric Authentication from Photic Stimulated EEG Records. Appl. Artif. Intell. 2021, 35, 1407–1419. [Google Scholar] [CrossRef]
- Wu, B.; Meng, W.; Chiu, W.-Y. Towards Enhanced EEG-Based Authentication with Motor Imagery Brain-Computer Interface. In Proceedings of the 38th Annual Computer Security Applications Conference, Austin, TX, USA, 5–9 December 2022; ACM: New York, NY, USA, 2022; pp. 799–812. [Google Scholar] [CrossRef]
- Altuwaijri, G.A.; Muhammad, G.; Altaheri, H.; Alsulaiman, M. A Multi-Branch Convolutional Neural Network with Squeeze-and-Excitation Attention Blocks for EEG-Based Motor Imagery Signals Classification. Diagnostics 2022, 12, 995. [Google Scholar] [CrossRef]
- Autthasan, P.; Chaisaen, R.; Sudhawiyangkul, T.; Rangpong, P.; Kiatthaveephong, S.; Dilokthanakul, N.; Bhakdisongkhram, G.; Phan, H.; Guan, C.; Wilaiprasitporn, T. MIN2Net: End-to-End Multi-Task Learning for Subject-Independent Motor Imagery EEG Classification. IEEE Trans. Biomed. Eng. 2022, 69, 2105–2118. [Google Scholar] [CrossRef] [PubMed]
- Bidgoly, A.J.; Bidgoly, H.J.; Arezoumand, Z. Towards a Universal and Privacy Preserving EEG-Based Authentication System. Sci. Rep. 2022, 12, 2531. [Google Scholar] [CrossRef] [PubMed]
- Alsumari, W.; Hussain, M.; Alshehri, L.; Aboalsamh, H.A. EEG-Based Person Identification and Authentication Using Deep Convolutional Neural Network. Axioms 2023, 12, 74. [Google Scholar] [CrossRef]
- Yap, H.Y.; Choo, Y.-H.; Mohd Yusoh, Z.I.; Khoh, W.H. An Evaluation of Transfer Learning Models in EEG-Based Authentication. Brain Inf. 2023, 10, 19. [Google Scholar] [CrossRef]
- Chen, X.; Teng, X.; Chen, H.; Pan, Y.; Geyer, P. Toward Reliable Signals Decoding for Electroencephalogram: A Benchmark Study to EEGNeX. Biomed. Signal Process. Control 2024, 87, 105475. [Google Scholar] [CrossRef]
- Shakir, A.M.; Bidgoly, A.J. Task-Independent EEG-Based Authentication. J. Tianjin Univ. Sci. Technol. 2024, 57, 1–14. [Google Scholar] [CrossRef]
- Wu, Q.; Zeng, Y.; Zhang, C.; Tong, L.; Yan, B. An EEG-Based Person Authentication System with Open-Set Capability Combining Eye Blinking Signals. Sensors 2018, 18, 335. [Google Scholar] [CrossRef]
- Ozdenizci, O.; Wang, Y.; Koike-Akino, T.; Erdogmus, D. Adversarial Deep Learning in EEG Biometrics. IEEE Signal Process. Lett. 2019, 26, 710–714. [Google Scholar] [CrossRef]
- Musallam, Y.K.; AlFassam, N.I.; Muhammad, G.; Amin, S.U.; Alsulaiman, M.; Abdul, W.; Altaheri, H.; Bencherif, M.A.; Algabri, M. Electroencephalography-Based Motor Imagery Classification Using Temporal Convolutional Network Fusion. Biomed. Signal Process. Control 2021, 69, 102826. [Google Scholar] [CrossRef]
- Mane, R.; Chew, E.; Chua, K.; Ang, K.K.; Robinson, N.; Vinod, A.P.; Lee, S.-W.; Guan, C. FBCNet: A Multi-View Convolutional Neural Network for Brain-Computer Interface. arXiv 2021, arXiv:2104.01233. [Google Scholar] [CrossRef]
- Hu, F.; Wang, F.; Bi, J.; An, Z.; Chen, C.; Qu, G.; Han, S. HASTF: A Hybrid Attention Spatio-Temporal Feature Fusion Network for EEG Emotion Recognition. Front. Neurosci. 2024, 18, 1479570. [Google Scholar] [CrossRef] [PubMed]
- Muna, U.M.; Shawon, M.M.H.; Jobayer, M.; Akter, S.; Sabuj, S.R. SSTAF: Spatial-Spectral-Temporal Attention Fusion Transformer for Motor Imagery Classification. arXiv 2025, arXiv:2504.13220. [Google Scholar] [CrossRef]
- Wei, C.; Zhou, G. EEG Emotion Recognition Based on Attention Mechanism Fusion Transformer Network. In Proceedings of the 2024 11th International Conference on Biomedical and Bioinformatics Engineering, Osaka, Japan, 8–11 November 2024; ACM: New York, NY, USA, 2024; pp. 146–150. [Google Scholar] [CrossRef]
- Ghous, G.; Najam, S.; Alshehri, M.; Alshahrani, A.; AlQahtani, Y.; Jalal, A.; Liu, H. Attention-Driven Emotion Recognition in EEG: A Transformer-Based Approach with Cross-Dataset Fine-Tuning. IEEE Access 2025, 13, 69369–69394. [Google Scholar] [CrossRef]
- Wei, Y.; Liu, Y.; Li, C.; Cheng, J.; Song, R.; Chen, X. TC-Net: A Transformer Capsule Network for EEG-Based Emotion Recognition. Comput. Biol. Med. 2023, 152, 106463. [Google Scholar] [CrossRef]
- Sun, J.; Xie, J.; Zhou, H. EEG Classification with Transformer-Based Models. In Proceedings of the 2021 IEEE 3rd Global Conference on Life Sciences and Technologies (LifeTech), Nara, Japan, 9–11 March 2021; IEEE: New York, NY, USA, 2021; pp. 92–93. [Google Scholar] [CrossRef]
- Omair, A.; Saif-ur-Rehman, M.; Glasmachers, T.; Iossifidis, I.; Klaes, C. ConTraNet: A Hybrid Network for Improving the Classification of EEG and EMG Signals with Limited Training Data. Comput. Biol. Med. 2024, 168, 107649. [Google Scholar] [CrossRef]
- Wan, Z.; Li, M.; Liu, S.; Huang, J.; Tan, H.; Duan, W. EEGformer: A Transformer–Based Brain Activity Classification Method Using EEG Signal. Front. Neurosci. 2023, 17, 1148855. [Google Scholar] [CrossRef]
- Ma, Y.; Song, Y.; Gao, F. A Novel Hybrid CNN-Transformer Model for EEG Motor Imagery Classification. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; IEEE: New York, NY, USA, 2022; pp. 1–8. [Google Scholar] [CrossRef]
- Zhao, W.; Jiang, X.; Zhang, B.; Xiao, S.; Weng, S. CTNet: A Convolutional Transformer Network for EEG-Based Motor Imagery Classification. Sci. Rep. 2024, 14, 20237. [Google Scholar] [CrossRef]
- Liu, R.; Chao, Y.; Ma, X.; Sha, X.; Sun, L.; Li, S.; Chang, S. ERTNet: An Interpretable Transformer-Based Framework for EEG Emotion Recognition. Front. Neurosci. 2024, 18, 1320645. [Google Scholar] [CrossRef] [PubMed]
- Li, H.; Zhang, H.; Chen, Y. Dual-TSST: A Dual-Branch Temporal-Spectral-Spatial Transformer Model for EEG Decoding. arXiv 2024, arXiv:2409.03251. [Google Scholar] [CrossRef]
- Xie, J.; Zhang, J.; Sun, J.; Ma, Z.; Qin, L.; Li, G.; Zhou, H.; Zhan, Y. A Transformer-Based Approach Combining Deep Learning Network and Spatial-Temporal Information for Raw EEG Classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 2126–2136. [Google Scholar] [CrossRef]
- Si, X.; Huang, D.; Sun, Y.; Huang, S.; Huang, H.; Ming, D. Transformer-Based Ensemble Deep Learning Model for EEG-Based Emotion Recognition. Brain Sci. Adv. 2023, 9, 210–223. [Google Scholar] [CrossRef]
- Yao, X.; Li, T.; Ding, P.; Wang, F.; Zhao, L.; Gong, A.; Nan, W.; Fu, Y. Emotion Classification Based on Transformer and CNN for EEG Spatial–Temporal Feature Learning. Brain Sci. 2024, 14, 268. [Google Scholar] [CrossRef] [PubMed]
- Lu, W.; Xia, L.; Tan, T.P.; Ma, H. CIT-EmotionNet: Convolution Interactive Transformer Network for EEG Emotion Recognition. PeerJ Comput. Sci. 2024, 10, e2610. [Google Scholar] [CrossRef] [PubMed]
- Bagchi, S.; Bathula, D.R. EEG-ConvTransformer for Single-Trial EEG-Based Visual Stimulus Classification. Pattern Recognit. 2022, 129, 108757. [Google Scholar] [CrossRef]
- Song, Y.; Zheng, Q.; Liu, B.; Gao, X. EEG Conformer: Convolutional Transformer for EEG Decoding and Visualization. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 710–719. [Google Scholar] [CrossRef]
- Si, X.; Huang, D.; Liang, Z.; Sun, Y.; Huang, H.; Liu, Q.; Yang, Z.; Ming, D. Temporal Aware Mixed Attention-Based Convolution and Transformer Network for Cross-Subject EEG Emotion Recognition. Comput. Biol. Med. 2024, 181, 108973. [Google Scholar] [CrossRef]
- Gong, L.; Li, M.; Zhang, T.; Chen, W. EEG Emotion Recognition Using Attention-Based Convolutional Transformer Neural Network. Biomed. Signal Process. Control 2023, 84, 104835. [Google Scholar] [CrossRef]
- Altaheri, H.; Muhammad, G.; Alsulaiman, M. Physics-Informed Attention Temporal Convolutional Network for EEG-Based Motor Imagery Classification. IEEE Trans. Ind. Inf. 2023, 19, 2249–2258. [Google Scholar] [CrossRef]
- Nguyen, A.H.P.; Oyefisayo, O.; Pfeffer, M.A.; Ling, S.H. EEG-TCNTransformer: A Temporal Convolutional Transformer for Motor Imagery Brain–Computer Interfaces. Signals 2024, 5, 605–632. [Google Scholar] [CrossRef]
- Cheng, Z.; Bu, X.; Wang, Q.; Yang, T.; Tu, J. EEG-Based Emotion Recognition Using Multi-Scale Dynamic CNN and Gated Transformer. Sci. Rep. 2024, 14, 31319. [Google Scholar] [CrossRef] [PubMed]
- Kostas, D.; Aroca-Ouellette, S.; Rudzicz, F. BENDR: Using Transformers and a Contrastive Self-Supervised Learning Task to Learn from Massive Amounts of EEG Data. Front. Hum. Neurosci. 2021, 15, 653659. [Google Scholar] [CrossRef]
- Yang, R.; Modesitt, E. ViT2EEG: Leveraging Hybrid Pretrained Vision Transformers for EEG Data. arXiv 2023, arXiv:2308.00454. [Google Scholar] [CrossRef]
- Jiang, W.; Zhao, L.; Lu, B. LaBraM: Large Brain Model for Learning Generic Representations with Tremendous EEG Data. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 7–11 May 2024; pp. 1–22. Available online: https://proceedings.iclr.cc/paper_files/paper/2024/file/47393e8594c82ce8fd83adc672cf9872-Paper-Conference.pdf (accessed on 23 September 2025).
- Li, W.; Zhou, N.; Qu, X. Enhancing Eye-Tracking Performance Through Multi-Task Learning Transformer. In Augmented Cognition; Schmorrow, D.D., Fidopiastis, C.M., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2024; Volume 14695, pp. 31–46. [Google Scholar] [CrossRef]
- Zhang, X.; Yao, L.; Kanhere, S.S.; Liu, Y.; Gu, T.; Chen, K. MindID: Person Identification from Brain Waves through Attention-Based Recurrent Neural Network. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–23. [Google Scholar] [CrossRef]
- Zhang, X.; Yao, L.; Huang, C.; Gu, T.; Yang, Z.; Liu, Y. DeepKey: A Multimodal Biometric Authentication System via Deep Decoding Gaits and Brainwaves. ACM Trans. Intell. Syst. Technol. 2020, 11, 1–24. [Google Scholar] [CrossRef]
- Balcı, F. DM-EEGID: EEG-Based Biometric Authentication System Using Hybrid Attention-Based LSTM and MLP Algorithm. Trait. Du Signal 2023, 40, 65–79. [Google Scholar] [CrossRef]
- Wilaiprasitporn, T.; Ditthapron, A.; Matchaparn, K.; Tongbuasirilai, T.; Banluesombatkul, N.; Chuangsuwanich, E. Affective EEG-Based Person Identification Using the Deep Learning Approach. IEEE Trans. Cogn. Dev. Syst. 2020, 12, 486–496. [Google Scholar] [CrossRef]
- Sun, Y.; Lo, F.P.-W.; Lo, B. EEG-Based User Identification System Using 1D-Convolutional Long Short-Term Memory Neural Networks. Expert Syst. Appl. 2019, 125, 259–267. [Google Scholar] [CrossRef]
- Chakravarthi, B.; Ng, S.-C.; Ezilarasan, M.R.; Leung, M.-F. EEG-Based Emotion Recognition Using Hybrid CNN and LSTM Classification. Front. Comput. Neurosci. 2022, 16, 1019776. [Google Scholar] [CrossRef] [PubMed]
- Puengdang, S.; Tuarob, S.; Sattabongkot, T.; Sakboonyarat, B. EEG-Based Person Authentication Method Using Deep Learning with Visual Stimulation. In Proceedings of the 2019 11th International Conference on Knowledge and Smart Technology (KST), Phuket, Thailand, 23–26 January 2019; IEEE: New York, NY, USA, 2019; pp. 6–10. [Google Scholar] [CrossRef]
- Zheng, X.; Cao, Z.; Bai, Q. An Evoked Potential-Guided Deep Learning Brain Representation for Visual Classification. In Neural Information Processing; Yang, H., Pasupa, K., Leung, A.C.-S., Kwok, J.T., Chan, J.H., King, I., Eds.; Communications in Computer and Information Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 1333, pp. 54–61. [Google Scholar] [CrossRef]
- Jin, X.; Tang, J.; Kong, X.; Peng, Y.; Cao, J.; Zhao, Q.; Kong, W. CTNN: A Convolutional Tensor-Train Neural Network for Multi-Task Brainprint Recognition. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 103–112. [Google Scholar] [CrossRef] [PubMed]
- Kumar, P.; Saini, R.; Kaur, B.; Roy, P.P.; Scheme, E. Fusion of Neuro-Signals and Dynamic Signatures for Person Authentication. Sensors 2019, 19, 4641. [Google Scholar] [CrossRef] [PubMed]
- Chakladar, D.D.; Kumar, P.; Roy, P.P.; Dogra, D.P.; Scheme, E.; Chang, V. A Multimodal-Siamese Neural Network (mSNN) for Person Verification Using Signatures and EEG. Inf. Fusion 2021, 71, 17–27. [Google Scholar] [CrossRef]










| Ref. | Author/Year | Focus | Contribution |
|---|---|---|---|
| Broad Surveys and Comparative Reviews | |||
| [8] | Roy et al. 2019 | 156 DL EEG studies (2010–2018) | -Concluded CNNs most common (40%), RNNs come next (13%) -DL gave 5.4% Acc. gain with extensive trends |
| [2] | Craik et al. 2019 | 90 DL EEG studies | -Categorized tasks (ER, MI, workload, seizure, ERP/sleep) -Noted CNNs/RNNs/DBNs strongest |
| [47] | Saeidi et al. 2021 | Supervised EEG decoding in both ML and DL | -Mapped full pipeline from preprocessing to classification -Highlighted SVMs’ continued use and growing DL adoption |
| [48] | Li et al. 2022 | Conceptual EEG classification | -EEG research within psychological and physiological contexts |
| [49] [50] | Prabowo et al. 2023 Vempati and Sharma 2023 | DL foundations and EEG classification | -Summarized theoretical motivations and methods |
| [52] | Gatfan 2024 | ML vs. DL for EEG classification | -Showed CNNs consistently outperform ML |
| [51] | Mohammed et al. 2025 | ML and DL across seizure detection, ER, MI, workload | -Surveyed CNNs, RNNs, Transformers -Noted growing hybrid adoption |
| Application-Specific Reviews: ER | |||
| [53] | Suhaimi et al. 2020 | EEG-based ER (2016–2019) | -Compared emotion elicitation methods (hardware, ML use, and VR as emerging tool) |
| [54] | Rahman et al. 2021 | EEG and emotion theories | -Combined theoretical emotion models with EEG techniques |
| [55] | Khare et al. 2023 | Multimodal ER (EEG + ECG/GSR) | -Reviewed datasets and challenges (signal variability, trust, real-time issues) -proposed federated meta-learning |
| [56] | Jafari et al. 2023 | DL for EEG ER | -Discussed hardware acceleration (SoC, FPGA, ASIC) for real-time EEG |
| [57] | Ma et al. 2024 | DL for EEG ER | -Framework: subject-dependent vs. -independent models |
| [58] | Gkintoni et al. 2025 | 64 DL EEG studies on ER | -Found multimodal EEG + physiology led to >90% Acc. -Recommended adaptive models and ethical practices |
| Application-Specific Reviews: MI and BCIs | |||
| [59] | Al-Saegh et al. 2021 | DL for MI EEG | -Summarized input representations (raw, TF) -Noted CNN/hybrid dominance |
| [60] | Ko et al. 2021 | Calibration-free DL-based BCIs | -Categorized DA (generative, geometric) and TL (explicit, implicit) -Recommended generative DA and explicit TL |
| [61] | Pawan et al. 2023 | 220 ML-based BCI studies | -Organized classification pipelines, common features, and classifiers -Noted DL emergence |
| [62] | Saibene et al. 2024 | DL for MI EEG (public datasets) | -Highlighted preprocessing, benchmarks, wearable integration |
| [64] | Wang et al. 2024 | 67 DL-based studies for MI EEG (PRISMA) | -Benchmarked 13 models with public code -Their ablations resulted in design recommendations -Confirmed multistream CNN + LSTM best—FC layers are costly |
| [63] | Moreno-Castelblanco et al. 2025 | 35 ML- and DL-based studies on lower limb MI in neurorehabilitation | -Identified multimodal fusion, low-channel, portable BCIs as trends |
| Application-Specific Reviews: Cognitive Workload and Neuropsychology | |||
| [66] | Bardeci et al. 2021 | DL for EEG-based psychiatric diagnosis and prediction | -Evaluated methodological rigor in clinical context of EEG -Highlighted flaws in reporting and validation |
| [65] | Hassan et al. 2024 | EEG-based cognitive workload (PRISMA) | -Showed SVM + DL dominance (CNNs, RNNs, and hybrids) -Emphasized multimodal integration and real-world validation |
| Application-Specific Reviews: Special Contexts (Devices, VR/AR) | |||
| [68] | Dadebayev et al. 2022 | Consumer vs. research-grade EEG | -Found limited data quality and ML performance with commercial headsets |
| [67] | Nwagu et al. 2023 | EEG-based BCIs in VR/AR | -First review for VR/AR BCIs -Defined trends: SSVEP in AR, MI in VR -Noted discomfort and ITR issues |
| Architecture-Focused Reviews | |||
| [69] | Klepl et al. 2023 | GNNs for EEG | -Reviewed graph node/edge features -Highlighted spectral graph convolutions’ dominance -Common node features (raw EEG signals and DE) |
| [13] | Vafaei and Hosseini 2025 | Transformers for EEG | -Categorized (TS, vision, GAT, hybrid) Transformers -Provided a roadmap with data augmentation and TL |
| Scenario Focus | Primary Weights | Insights |
|---|---|---|
| S1: Real-Time efficiency | Oper. Cost (Speed) | Reveals best models for synchronous, low-latency applications |
| S2: Edge Deployment Efficiency | Complex. (Size) | Shows ideal models for portable, low-memory devices |
| S3: Comprehensive All-Rounder | Neutral (Balanced) | Ranks models with the highest overall utility and efficiency |
| Metric | S1: Real-Time | S2: Edge Deployment | S3: All-Rounder |
|---|---|---|---|
| Accuracy (Acc.) | 35% | 35% | 40% |
| Oper. Cost | 30% | 15% | 20% |
| Complex. | 15% | 30% | 15% |
| Comp. Cost | 10% | 10% | 10% |
| Train. Cost | 10% | 10% | 15% |
| Domain/Task | CNNs | Transformers | CNN–Transformers | Recurrent Hybrids |
|---|---|---|---|---|
| Biometrics (Bio) | [70,71,72,73,75,77,78,79,80,81,82,83,85,90,91,93,96,99,100,101,103,105] | [41,42] | - | [135,136,137,138,139,143,144,145] |
| Motor Imagery (MI/BCI) | [11,12,86,88,89,94,96,97,98,99,102,103,104,106,107] | [39,40,44,45,109] | [113,114,116,117,119,120,125,128,129] | - |
| Emotion Recognition (ER) | [92] | [38,108,111] | [118,119,121,122,123,125,127,130] | [140] |
| Evoked Potentials (EPs) | [12,74,76,77,80,81,82,83,84,85,91,95,102,104,105,107] | [43] | [115,120,124,129] | [141,142] |
| General Foundation (GF) | [87] | [34,44,45,46] | [131,132,133] | - |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Hallal, L.; Rhinelander, J.; Venkat, R.; Newman, A. Efficient Feature Extraction for EEG-Based Classification: A Comparative Review of Deep Learning Models. AI 2026, 7, 50. https://doi.org/10.3390/ai7020050
Hallal L, Rhinelander J, Venkat R, Newman A. Efficient Feature Extraction for EEG-Based Classification: A Comparative Review of Deep Learning Models. AI. 2026; 7(2):50. https://doi.org/10.3390/ai7020050
Chicago/Turabian StyleHallal, Louisa, Jason Rhinelander, Ramesh Venkat, and Aaron Newman. 2026. "Efficient Feature Extraction for EEG-Based Classification: A Comparative Review of Deep Learning Models" AI 7, no. 2: 50. https://doi.org/10.3390/ai7020050
APA StyleHallal, L., Rhinelander, J., Venkat, R., & Newman, A. (2026). Efficient Feature Extraction for EEG-Based Classification: A Comparative Review of Deep Learning Models. AI, 7(2), 50. https://doi.org/10.3390/ai7020050

