Author Contributions
Conceptualization, T.F.; methodology, T.F. and Z.Z.; software, T.F.; validation, Z.Z. and W.Y.; formal analysis, T.F.; investigation, Z.Z.; resources, T.F.; data curation, T.F.; writing—original draft preparation, T.F.; writing—review and editing, Z.Z. and W.Y.; visualization, T.F.; supervision, Z.Z. and W.Y. All authors have read and agreed to the published version of the manuscript.
Figure 1.
Overall architecture of the proposed SCGTNet. The general architecture of the proposed SCGTNet. The input EEG or sEMG signals are adjusted to the significance of different channels through the ASA module. The MSC module connects the GRUs to extract local features. Transformer Encoder is utilized to obtain timing dependencies. B, T, C, F, and G are the batch size, time_steps, number of channels, filters in Conv2D and gru units.
Figure 1.
Overall architecture of the proposed SCGTNet. The general architecture of the proposed SCGTNet. The input EEG or sEMG signals are adjusted to the significance of different channels through the ASA module. The MSC module connects the GRUs to extract local features. Transformer Encoder is utilized to obtain timing dependencies. B, T, C, F, and G are the batch size, time_steps, number of channels, filters in Conv2D and gru units.
Figure 2.
Architecture of the Adaptive Spatial Attention Module. B, T, C, and W are the batch size, time_steps, number of channels and channel weights.
Figure 2.
Architecture of the Adaptive Spatial Attention Module. B, T, C, and W are the batch size, time_steps, number of channels and channel weights.
Figure 3.
Architecture of the Multi-Scale Convolution Module. B, T, C, and F are the batch size, time_steps, number of channels and filters in Conv2D.
Figure 3.
Architecture of the Multi-Scale Convolution Module. B, T, C, and F are the batch size, time_steps, number of channels and filters in Conv2D.
Figure 4.
Training loss and accuracy curves during the ablation experiments on the BCI IV-2a dataset. These plots demonstrate the contributions of each module to convergence speed and final performance on motor intent decoding.
Figure 4.
Training loss and accuracy curves during the ablation experiments on the BCI IV-2a dataset. These plots demonstrate the contributions of each module to convergence speed and final performance on motor intent decoding.
Figure 5.
t-SNE visualization for SCGTNet under different ablation settings on BCI IV 2a: (a) full model; (b) without ASA; (c) without MSC; (d) MSC replaced by convolution. Points represent trial feature vectors in 2D space, color-coded by class (LEFT is blue, RIGHT is orange, FEET is green, TONGUE is red). The degree of cluster overlap indicates the model’s class separation capability under each ablation.
Figure 5.
t-SNE visualization for SCGTNet under different ablation settings on BCI IV 2a: (a) full model; (b) without ASA; (c) without MSC; (d) MSC replaced by convolution. Points represent trial feature vectors in 2D space, color-coded by class (LEFT is blue, RIGHT is orange, FEET is green, TONGUE is red). The degree of cluster overlap indicates the model’s class separation capability under each ablation.
Figure 6.
Training loss and accuracy curves during the ablation experiments on the NinaPro DB2 dataset. These plots demonstrate the contributions of each module to convergence speed and final performance on gesture recognition.
Figure 6.
Training loss and accuracy curves during the ablation experiments on the NinaPro DB2 dataset. These plots demonstrate the contributions of each module to convergence speed and final performance on gesture recognition.
Figure 7.
Confusion matrix for SCGTNet on the BCI IV-2a dataset. Columns denote the ground-truth motor imagery classes (LEFT, RIGHT, FEET, TONGUE) and rows columns denote predicted classes. The color intensity indicates the proportion of trials assigned to each predicted class.
Figure 7.
Confusion matrix for SCGTNet on the BCI IV-2a dataset. Columns denote the ground-truth motor imagery classes (LEFT, RIGHT, FEET, TONGUE) and rows columns denote predicted classes. The color intensity indicates the proportion of trials assigned to each predicted class.
Figure 8.
Leave-one-subject-out (LOSO) performance of SCGTNet and baseline methods on the BCI IV 2a dataset. The x-axis denotes subjects (Sub1–Sub9), and the y-axis denotes classification accuracy.
Figure 8.
Leave-one-subject-out (LOSO) performance of SCGTNet and baseline methods on the BCI IV 2a dataset. The x-axis denotes subjects (Sub1–Sub9), and the y-axis denotes classification accuracy.
Figure 9.
Confusion matrix for SCGTNet on the NinaPro DB2 dataset (49 classes). Columns denote ground-truth gesture classes and rows denote predicted classes. Color intensity corresponds to the proportion of correctly classified trials per gesture.
Figure 9.
Confusion matrix for SCGTNet on the NinaPro DB2 dataset (49 classes). Columns denote ground-truth gesture classes and rows denote predicted classes. Color intensity corresponds to the proportion of correctly classified trials per gesture.
Table 1.
Ablation study results for SCGTNet on the BCI IV-2a dataset. These results highlight the contribution of each component to overall performance.
Table 1.
Ablation study results for SCGTNet on the BCI IV-2a dataset. These results highlight the contribution of each component to overall performance.
Configuration | Accuracy | Recall | Precision | F1-Score |
---|
Without ASA | 73.33% | 72.45% | 74.31% | 73.36% |
Without MSC | 74.12% | 76.47% | 71.26% | 73.77% |
CNN | 76.04% | 73.26% | 74.63% | 73.94% |
Full model | 79.47% | 79.33% | 80.96% | 78.13% |
Table 2.
Ablation study results for SCGTNet on the NinaPro DB2 dataset. These results highlight the contribution of each component to overall performance.
Table 2.
Ablation study results for SCGTNet on the NinaPro DB2 dataset. These results highlight the contribution of each component to overall performance.
Configuration | Accuracy | Recall | Precision | F1-Score |
---|
Without ASA | 83.28% | 81.02% | 84.90% | 82.88% |
Without MSC | 81.72% | 87.52% | 85.78% | 81.26% |
CNN | 81.15% | 85.43% | 86.46% | 80.60% |
Full model | 85.87% | 87.42% | 83.19% | 85.42% |
Table 3.
Impact of Morlet Wavelet Bandwidth (fb) on BCI IV 2a classification performance. Accuracies, recalls, precisions, and F1-scores were measured under different bandwidth settings (fb = 1.5, 2.0, 2.5).
Table 3.
Impact of Morlet Wavelet Bandwidth (fb) on BCI IV 2a classification performance. Accuracies, recalls, precisions, and F1-scores were measured under different bandwidth settings (fb = 1.5, 2.0, 2.5).
Number of fb | Accuracy | Recall | Precision | F1-Score |
---|
1.5 | 68.79% | 67.04% | 70.61% | 68.74% |
2 | 79.47% | 79.33% | 80.96% | 78.13% |
2.5 | 69.42% | 68.04% | 71.08% | 69.48% |
Table 4.
Impact of multi-head attention head number on BCI IV 2a and Ninapro DB2 classification performance.
Table 4.
Impact of multi-head attention head number on BCI IV 2a and Ninapro DB2 classification performance.
Dataset | num_heads | Accuracy | Recall | Precision | F1-Score |
---|
BCI IV 2a | 1 | 74.41% | 73.11% | 75.16% | 74.12% |
BCI IV 2a | 2 | 77.47% | 78.24% | 76.27% | 77.24% |
BCI IV 2a | 4 | 79.47% | 79.33% | 76.96% | 78.13% |
BCI IV 2a | 8 | 79.51% | 80.09% | 77.62% | 78.84% |
Ninapro DB2 | 1 | 80.23% | 82.14% | 79.16% | 80.62% |
Ninapro DB2 | 2 | 81.88% | 83.37% | 80.21% | 81.80% |
Ninapro DB2 | 4 | 85.87% | 87.42% | 83.19% | 85.42% |
Ninapro DB2 | 8 | 84.95% | 86.02% | 82.17% | 84.05% |
Table 5.
Performance comparison of SCGTNet with other models on the BCI IV 2a dataset.
Table 5.
Performance comparison of SCGTNet with other models on the BCI IV 2a dataset.
Model | Accuracy | Recall | Precision | F1-Score |
---|
CNN [44] | 70.65% | 70.31% | 71.26% | 70.78% |
LSTM [45] | 71.24% | 74.33% | 74.21% | 74.27% |
EEG-Inception [25] | 74.76% | 72.47% | 71.69% | 72.08% |
MIN2Net [46] | 75.87% | 75.09% | 74.16% | 74.62% |
EEG-TCNet [47] | 75.53% | 76.63% | 75.31% | 75.96% |
TFT [30] | 74.49% | 75.54% | 73.36% | 74.43% |
SCGT (Ours) | 79.47% | 79.33% | 76.96% | 78.13% |
Table 6.
Performance comparison of SCGTNet with other models on the NinaPro DB2 dataset. Recall/Precision/F1 is marked ‘-’ for baselines that cannot be reproduced, indicating that the original article was not reported or that we were unable to reproduce it with the same settings.
Table 6.
Performance comparison of SCGTNet with other models on the NinaPro DB2 dataset. Recall/Precision/F1 is marked ‘-’ for baselines that cannot be reproduced, indicating that the original article was not reported or that we were unable to reproduce it with the same settings.
Model | Accuracy | Recall | Precision | F1-Score |
---|
CNN [44] | 70.28% | 68.58% | 72.03% | 70.21% |
LSTM [45] | 73.13% | 72.22% | 74.57% | 73.40% |
Wei et al. [48] | 83.7% | - | - | - |
Ding et al. [26] | 78.86% | - | - | - |
Zhang et al. [49] | 83.29% | - | - | - |
Hu et al. [50] | 84.84% | - | - | - |
TFT [30] | 83.64% | 83.85% | 80.06% | 82.40% |
SCGT (Ours) | 85.87% | 87.42% | 83.19% | 85.42% |
Table 7.
Total and trainable parameter counts and floating-point operations (FLOPs) of SCGTNet and comparison models on the BCI IV 2a dataset.
Table 7.
Total and trainable parameter counts and floating-point operations (FLOPs) of SCGTNet and comparison models on the BCI IV 2a dataset.
Model | Total Params | Trainable Params | No-Trainable Params | FLOPs |
---|
CNN [44] | 11.5k | 11.3k | 0.2k | 0.83M |
LSTM [45] | 18k | 16.6k | 1.4k | 0.56M |
TFT [30] | 220k | 217.6k | 2.4k | 3.50M |
SCGT (Ours) | 56k | 55.4k | 0.6k | 1.12M |
Table 8.
Total and trainable parameter counts and floating-point operations (FLOPs) of SCGTNet and comparison models on the NinaPro DB2 Dataset.
Table 8.
Total and trainable parameter counts and floating-point operations (FLOPs) of SCGTNet and comparison models on the NinaPro DB2 Dataset.
Model | Total Params | Trainable Params | No-Trainable Params | FLOPs |
---|
CNN [44] | 9.8k | 9.6k | 0.2k | 1.25M |
LSTM [45] | 14.5k | 14.1k | 0.4k | 2.87M |
TFT [30] | 180k | 180k | 1.8k | 52M |
SCGT (Ours) | 48k | 47.4k | 0.6k | 12M |