1. Introduction
Every year, spinal cord injuries (SCIs) affect approximately 250,000 to 500,000 individuals globally, with an estimated two to three million people living with SCI-related disabilities [
1]. SCI arises from damage to the spinal cord or surrounding structures, disrupting communication between the brain and body [
2]. Causes include traumatic incidents such as vehicular accidents, falls, and sports injuries, as well as non-traumatic factors. Clinical manifestations vary depending on the injury’s severity and location, commonly resulting in sensory and motor impairments, muscular weakness, and complications in physiological functions [
3]. While complete injuries typically lead to permanent deficits, partial injuries may permit some functional recovery.
Technological advancements have significantly improved rehabilitation approaches and patient quality of life. Among these, brain-computer interfaces (BCIs) that leverage electroencephalography (EEG) have emerged as promising tools. EEG-based BCIs enable direct communication between the brain and external devices, offering a non-invasive, portable, and cost-effective solution for individuals with limited motor control [
4,
5]. EEG signals, which capture oscillatory neural activity, are acquired via electrodes placed on the scalp. These signals can be decoded in real time using machine learning algorithms to infer user intentions [
6].
Despite their potential, EEG signals present analytical challenges due to their high dimensionality, low signal-to-noise ratio, and variability across sessions and subjects [
7]. Effective dimensionality reduction is thus essential for improving signal decoding accuracy and computational efficiency. Traditional techniques like Principal Component Analysis (PCA) have been widely used, but recent research focuses on nonlinear methods better suited to the intrinsic geometry of EEG data [
8].
Manifold learning is a powerful class of nonlinear dimensionality reduction methods that seeks low-dimensional representations while preserving local or global data structures [
9,
10,
11]. These methods are particularly advantageous in processing EEG data due to their ability to retain discriminative features critical for classification [
12]. Prominent manifold learning algorithms include ISOMAP [
11], Locally Linear Embedding (LLE) [
10], t-Distributed Stochastic Neighbor Embedding (t-SNE) [
13], Spectral Embedding [
14], and Multidimensional Scaling (MDS) [
15].
In recent years, the integration of manifold learning techniques with shallow classifiers such as k-nearest neighbors (
k-NN), Support Vector Machines (SVMs), and Naive Bayes has shown promise in decoding motor imagery (MI) tasks from EEG [
16,
17]. These combinations enable efficient real-time EEG decoding with reduced computational burden. Moreover, comparative studies suggest that manifold learning can improve classification accuracy in EEG-based BCIs, particularly for applications in neurorehabilitation and assistive technology [
18].
This study aims to explore the effectiveness of manifold learning techniques paired with shallow classifiers for classifying EEG data collected from six healthy participants performing five wrist and hand motor imagery tasks. The performance of various dimensionality reduction-classifier pairs is evaluated across binary, ternary, and five-class scenarios to identify robust, low-complexity pipelines suitable for real-time BCI applications.
1.1. State of the Art
Recent studies have highlighted the potential of manifold learning and advanced feature extraction techniques in enhancing the classification performance of EEG-based BCIs, particularly in motor imagery tasks.
Li et al. [
19] introduced an adaptive feature extraction framework combining wavelet packet decomposition (WPD) and semidefinite embedding ISOMAP (SE-ISOMAP). This approach utilized subject-specific optimal wavelet packets to extract time-frequency and manifold features, achieving 100% accuracy in binary classification tasks and significantly outperforming conventional dimensionality reduction methods.
Yamamoto et al. [
20] proposed a novel method called Riemann Spectral Clustering (RiSC), which maps EEG covariance matrices as graphs on the Riemannian manifold using a geodesic-based similarity measure. They further extended this framework with odenRiSC for outlier detection and mcRiSC for multimodal classification, where mcRiSC reached 73.1% accuracy and outperformed standard single-modal classifiers in heterogeneous datasets.
Krivov and Belyaev [
21] incorporated Riemannian geometry and Isomap to reveal the manifold structure of EEG covariance matrices in a low-dimensional space. Their method, evaluated with Linear Discriminant Analysis (LDA), reported classification accuracies of 0.58 (CSP), 0.61 (PGA), and 0.58 (Isomap) in a four-class task, underlining the potential of manifold methods in representing EEG data structures.
Tyagi and Nehra [
22] compared LDA, PCA, FA, MDS, and ISOMAP for motor imagery feature extraction using BCI Competition IV datasets. A feedforward artificial neural network (ANN) trained with the Levenberg-Marquardt algorithm yielded the lowest mean square error (MSE) with LDA (0.1143), followed by ISOMAP (0.2156), while other linear methods showed relatively higher errors.
Xu et al. [
23] designed an EEG-based attention classification method utilizing Riemannian manifold representation of symmetric positive definite (SPD) matrices. By integrating amplitude and phase information using a filter bank and applying SVM, their approach reached a classification accuracy of 88.06% in a binary scenario without requiring spatial filters.
Lee et al. [
24] assessed the efficacy of PCA, LLE, and ISOMAP in binary EEG classification using LDA. The classification errors were reported as 28.4% for PCA, 25.8% for LLE, and 27.7% for ISOMAP, suggesting LLE’s slight edge in capturing intrinsic EEG data structures.
Li, Luo, and Yang [
25] further evaluated the performance of linear and nonlinear dimensionality reduction techniques in motor imagery EEG classification. Nonlinear methods such as LLE (91.4%) and parametric t-SNE (94.1%) outperformed PCA (70.7%) and MDS (75.7%), demonstrating the importance of preserving local neighborhood structures for robust feature representation.
Sayılgan [
26] investigated EEG-based classification of imagined hand movements in spinal cord injury patients using Independent Component Analysis (ICA) for feature extraction and machine learning classifiers including SVM,
k-NN, AdaBoost, and Decision Trees. The highest accuracy was achieved with SVM (90.24%), while
k-NN demonstrated the lowest processing time, with the lateral grasp showing the highest classification accuracy among motor tasks.
These studies collectively underline the critical role of dimensionality reduction, particularly manifold learning, in effectively decoding motor intentions from EEG data, thereby improving the performance and applicability of BCI systems in neurorehabilitation contexts.
1.2. Contributions
EEG recordings, collected from multiple scalp locations, are inherently high-dimensional, often containing redundant information and being susceptible to various noise sources and artifacts. Such properties can hinder the accuracy and robustness of motor intention decoding in brain-computer interface (BCI) systems. While conventional linear dimensionality reduction approaches, such as Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA), are widely used, they often fail to capture the complex, nonlinear temporal-spatial relationships embedded in EEG activity patterns. In contrast, manifold learning techniques offer a promising alternative by projecting data into a lower-dimensional space while preserving its intrinsic local geometry.
In this work, we present a comprehensive and adaptable manifold learning-based processing framework for EEG analysis, designed to support the development of rehabilitation-oriented BCIs. The proposed pipeline integrates multiple nonlinear dimensionality reduction algorithms with shallow classifiers to alleviate overfitting, enhance inter-class separability, and improve overall decoding performance.
The main contributions of this study can be outlined as follows:
Introduction of a unified manifold learning framework for the classification of motor imagery EEG signals into binary (two-class), ternary (three-class), and multi-class (five-class) categories, using real-time data acquired from healthy participants.
Systematic comparison of five widely recognized manifold learning algorithms: Spectral Embedding, Locally Linear Embedding (LLE), Multidimensional Scaling (MDS), Isometric Mapping (ISOMAP), and t-distributed Stochastic Neighbor Embedding (t-SNE) for their effectiveness as feature transformation tools in motor intent recognition.
Alignment of the proposed methodology with practical rehabilitation needs, specifically for integration into a cost-efficient, two-degree-of-freedom robotic platform employing a straightforward control strategy.
Emphasis on building a sustainable machine learning model capable of accurately detecting motor intentions in healthy users while ensuring high classification performance, thereby enabling scalability to clinical scenarios.
Addressing a notable gap in the literature by exploring high-accuracy three-class and five-class EEG-based BCI paradigms for spinal cord injury (SCI) rehabilitation, and benchmarking binary classification results against state-of-the-art systems.
Analysis of task combination compatibility across different classification schemes, with performance metrics aggregated over all participants to support model generalizability.
Comprehensive evaluation using multiple performance indicators—accuracy, precision, recall, and F1-score-to assess the robustness of each manifold-classifier pairing under varying task complexities.
3. Classifiers
3.1. Support Vector Machine (SVM)
Support Vector Machine (SVM) is a supervised learning algorithm introduced by Vapnik [
48], and it has demonstrated strong performance in various real-world classification problems, including brain-computer interface (BCI) applications [
16,
36]. SVM aims to find an optimal hyperplane that maximally separates data points of different classes in a high-dimensional space, thereby minimizing generalization error.
3.1.1. Step 1: Fundamental Concepts of SVM
SVM identifies a hyperplane
to separate two classes (
and
)
where
To ensure correct classification, the condition below must hold:
3.1.2. Step 2: Optimization Problem
The margin is maximized by minimizing
, leading to the following convex optimization problem:
3.1.3. Step 3: Lagrange Multipliers and Dual Form
Using Lagrange multipliers
, the dual problem becomes
The optimal weight vector is
3.1.4. Step 4: Kernel and Soft Margin
This study employs the Radial Basis Function (RBF) kernel:
To handle overlapping classes, slack variables
are introduced:
Here,
C is the regularization parameter balancing margin maximization and classification error [
49].
3.2. k-Nearest Neighbors (k-NN)
The
k-nearest neighbor (
k-NN) is a non-parametric, supervised learning algorithm that classifies data points based on the majority vote of their
k closest neighbors [
50]. It is widely used due to its simplicity and effectiveness, particularly in multi-class problems.
The core idea is to assign a class to a new observation by evaluating the distance to
k training samples. The most common metric is Euclidean distance, defined as
The choice of
k significantly influences performance. A small
k may be sensitive to noise, while a large
k may blur class boundaries. Thus, an odd
k is often used to avoid ties in binary classification [
51,
52].
3.3. Naive Bayes
Naive Bayes is a probabilistic classifier grounded in Bayes’ theorem. It assumes strong (Naive) independence between features given the class label. Despite this often unrealistic assumption, Naive Bayes classifiers have demonstrated competitive performance in various practical applications due to their simplicity, efficiency, and robustness.
3.3.1. Step 1: Bayes’ Theorem
The core of the Naive Bayes classifier is Bayes’ theorem, which expresses the posterior probability
of a class
y given a feature vector
:
where
is the posterior probability of class y given features x,
is the prior probability of class y,
is the likelihood of features x given class y,
is the evidence or marginal likelihood of observing x (often omitted in classification as it is constant across classes).
3.3.2. Step 2: Conditional Independence Assumption
The fundamental assumption of Naive Bayes is that the features
are conditionally independent given the class label
y:
This simplifies the computation of the posterior as
3.3.3. Step 3: Classification Decision
The final prediction is made by selecting the class
y that maximizes the posterior probability:
3.4. Hyperparameter Tuning and Cross-Validation
To ensure reproducibility and fair comparison, all hyperparameters for manifold learning and classification algorithms were optimized using a systematic grid search approach. For t-SNE, perplexity values were tested in the range of 5–50, with a step size of 5, and learning rates between 100 and 1000. For ISOMAP and LLE, the number of neighbors was varied between 5 and 20. The SVM classifier was tuned over and . For k-NN, k values from 1 to 15 were tested. Nested cross-validation (inner loop for hyperparameter optimization, outer loop for performance estimation) was employed to avoid overfitting. The final selected parameters were those maximizing mean accuracy on the validation sets.
Dimensionality reduction and classification steps were executed with predefined hyperparameters, as outlined below.
Manifold Learning Algorithms:
All manifold learning algorithms were configured to reduce the original EEG signal features into a three-dimensional subspace, facilitating visualization and efficient classification.
Classification Algorithms:
k-Nearest Neighbors (k-NN): Configured with , using the Euclidean distance metric and uniform weighting.
Support Vector Machine (SVM): Configured with a cost parameter C = 1.00, epsilon = 0.10, and radial basis function (RBF) kernel; numerical tolerance was set to , with a maximum of 100 iterations.
Naïve Bayes: Implemented based on Bayes’ Theorem, assuming conditional independence among features.
This comprehensive pipeline enabled the systematic evaluation of different manifold learning and classification methods for EEG-based motor intention decoding tasks in healthy individuals.
3.5. Evaluation of Manifold Learning Algorithms Performance
In this study, the performance evaluation of manifold learning algorithms was conducted using stratified k-fold cross-validation with . In this procedure, the dataset is partitioned into five equal subsets while preserving the class distribution. During each iteration, one subset is reserved as the test set, and the remaining four subsets are used for training. This process is repeated five times, and the final performance is computed as the average of the individual results. The use of stratified sampling ensures that the class balance is maintained in each fold, thereby yielding a more realistic and robust estimation of model performance.
To assess the efficacy of the manifold learning algorithms, standard evaluation metrics were employed, including Area Under the Curve (AUC), classification accuracy (CA), F1-score, and precision. These metrics offer complementary insights into the classification model’s effectiveness, especially in the context of binary classification problems.
In binary classification, each instance is assigned to one of two classes: positive or negative. The outcomes of a classifier can be summarized in a confusion matrix, which categorizes predictions as follows:
True Positives (TPs): Correctly predicted positive instances.
False Positives (FPs): Incorrectly predicted as positive when they are negative.
True Negatives (TNs): Correctly predicted negative instances.
False Negatives (FNs): Incorrectly predicted as negative when they are positive.
This framework facilitates a comprehensive understanding of the model’s predictive capabilities [
53,
54].
3.5.1. Area Under the Curve (AUC)
The Area Under the Receiver Operating Characteristic (ROC) Curve, abbreviated as AUC, is a widely used metric that quantifies the classifier’s ability to distinguish between classes. The ROC curve plots the True Positive Rate (TPR) against the False Positive Rate (FPR) at various threshold settings. A higher AUC value indicates better overall classification performance.
The TPR (or sensitivity) and the True Negative Rate (TNR or specificity) are calculated as
AUC provides an aggregate measure of performance across all classification thresholds, making it particularly valuable in imbalanced datasets [
55,
56].
3.5.2. Classification Accuracy (CA)
Classification accuracy (CA) measures the proportion of correctly predicted instances (both positive and negative) over the total number of samples:
Although widely used, accuracy alone may be misleading in imbalanced datasets. Therefore, additional metrics such as F1-score and precision are necessary for a more nuanced evaluation [
4,
57].
3.5.3. F1 Score
The F1-score is the harmonic mean of precision and recall and is especially useful when dealing with class imbalance. It is defined as
A higher F1-score indicates a better trade-off between precision and recall, especially when false positives and false negatives carry different costs [
58,
59].
3.5.4. Precision
Precision is defined as the proportion of true positives among all instances classified as positive:
It evaluates the classifier’s exactness and is crucial when the cost of false positives is high [
53].
3.6. Data Partitioning and Leakage Prevention
To prevent data leakage, raw EEG trials were first segmented and assigned at the subject level before any feature extraction or dimensionality reduction steps. All preprocessing, including filtering, dimensionality reduction (ISOMAP, LLE, Spectral Embedding, t-SNE, MDS), and classifier training was performed strictly within training folds.
The Orange Data Mining software’s Pipeline Builder was used only to construct modular workflows; training and validation data were kept strictly separated, and nested 5 × 5-fold cross-validation was applied. This ensured no information from validation/test folds leaked into the training pipeline.
4. Results
This section reports the performance of each manifold learning method (ISOMAP, LLE, Spectral Embedding, t-SNE, MDS) combined with three classifiers (SVM,
k-NN, Naïve Bayes) across two-, three-, and five-class settings. We summarize metrics as AUC, CA, F1, and precision without interpretation; detailed comparisons and implications are discussed in
Section 5.
According to the results shown in
Table 2, the most effective classification method in the ISOMAP method is
k-NN. The values of AUC, CA, F1 score, and precision are between 99.5% and 99.3%, indicating that the combination of ISOMAP and
k-NN is quite successful. As the number of classes increases, the performances of all classifiers decrease, but this is expected. Naive Bayes shows the best performance after
k-NN. However, a serious decrease is seen in the five-class case. SVM is the model with the lowest performance in all classes.
Table 3 compares the performance of the classification algorithms (SVM,
k-NN, Naive Bayes) applied after dimensionality reduction with the Local Linear Embedding (LLE) method in two-, three-, and five-class classifications through AUC, CA, F1, and precision metrics.
k-NN is the model with the highest metric value in all classes (two, three, five). According to the metrics,
k-NN is between 96.3–81.5% SVM is the model with the lowest performance. According to the SVM results, some values remained around or below 0.5. This shows that the model has weak discrimination power between classes. The Naive Bayes method is the method with the best results after
k-NN.
Table 4 shows the performances of the classification algorithms (SVM,
k-NN, Naive Bayes) with the Spectral Embedding method in two-, three- and five-class classifications.
k-NN showed the highest performance in all metrics, staying between 96–72.1%. Although SVM gave a very high AUC (96%) value, especially in the three-class case, it showed low performance in other metrics. Naive Bayes gave worse results than
k-NN but better than SVM. Although SVM reached the highest value in AUC with 96% in three-class values, it failed in other metrics (CA = 39%, F1 = 36.5%). This shows that the model cannot provide balance while trying to distinguish the classes.
An overview of
Table 5 shows that the t-SNE +
k-NN pipeline achieved the highest overall performance across all classification scenarios, while SVM provided competitive results only in binary and ternary cases and degraded significantly in five-class classification. Naive Bayes exhibited moderate success, with performance notably dropping as class cardinality increased.
An overview of
Table 6 shows that MDS combined with
k-NN achieved the highest classification performance across binary, ternary, and five-class scenarios. While SVM provided moderate accuracy in simpler binary tasks and Naïve Bayes yielded more balanced but limited performance,
k-NN consistently outperformed both methods in all class settings.
When
Figure 7,
Figure 8 and
Figure 9, which are comparatively presented after the dimensionality reduction methods (ISOMAP, LLE, Spectral Embedding, t-SNE, and MDS) applied on the dataset, are examined, it is seen that the highest performance belongs to the
k-NN algorithm in all classification scenarios. Especially in binary classification (
Figure 7),
k-NN achieved 99.6% accuracy with t-SNE, 98.4% with ISOMAP and 97.1% with MDS, demonstrating a significantly superior success compared to the other two models (SVM and Naive Bayes). SVM gave a partially competitive result with 94.3% accuracy with t-SNE in this scenario, while Naive Bayes generally remained in the range of 72–76%.
In
Figure 8, which includes three-class classification results,
k-NN again stands out as the most successful model in all dimensionality reduction methods. Reaching 99.3% accuracy with t-SNE, 96.6% with ISOMAP and 94.3% with MDS,
k-NN largely maintained its performance despite the increase in the number of classes. SVM achieved a competitive result of 91.2% only with t-SNE, while its accuracy remained below 50% for other methods. Naive Bayes provided moderate results in the range of 56–61% with methods such as LLE and MDS, but the difference with
k-NN remained significant.
When the classification accuracies obtained after the ISOMAP dimensionality reduction method are examined, it is observed that the k-NN algorithm achieves the highest accuracy rates in all motor task pairs. Standing out with accuracy values exceeding 90%, k-NN exhibited a strong discrimination ability between both similar and highly distinct movements. While the Naive Bayes model provided a balanced performance with accuracies particularly in the 78–81% range, SVM showed relatively low success, especially in certain task pairs (e.g., “Palmar G.-Pronation” and “Lateral G.-Supination”). This finding clearly demonstrates that the k-NN algorithm is the most effective method for movement classification tasks in EEG data reduced using the ISOMAP method.
Table 7 summarizes the classification results of the ISOMAP method for different two-class combinations. Similar to the previous spectral embedding results,
k-NN consistently achieved the highest accuracy across all class pairs, with values ranging between 90% and 98.1%. In contrast, SVM generally showed lower accuracies, while Naive Bayes performed moderately, yielding results better than SVM but still below
k-NN.
Table 8 presents the two-class classification performances using the Local Linear Embedding (LLE) method. Consistent with previous findings,
k-NN achieved the highest accuracy across all class pairs, ranging from 80.7% to 93.8%. While Naive Bayes generally provided moderate results and outperformed SVM, the SVM classifier showed comparatively lower accuracies in almost all cases.
In the classification analyses performed on dimensionally reduced data with the Local Linear Embedding (LLE) method, the k-NN algorithm stood out as the most successful model by reaching the highest accuracy rates in all motor task pairs. The fact that k-NN provided accuracy, especially in the “Palmar G.-Pronation” (93.2%), “Hand O.-Palmar G.” (93.8%), and “Hand O.-Lateral G.” (93.5%) task pairs, shows that this model works effectively in the decomposed feature space after LLE. The Naive Bayes model provided accuracy in the range of 76–83% in most tasks, exhibiting a balanced and acceptable performance. On the other hand, the SVM model was insufficient in the post-LLE classification tasks with low accuracy rates (50–57%), and it was observed that the performance level decreased significantly, especially in the “Pronation-Supination” and “Lateral G.-Pronation” task pairs. These results indicate that k-NN is the most effective classifier after the LLE method, Naive Bayes offers a balanced alternative, and SVM can provide only limited success in this structure.
Classification analyses performed after the Spectral Embedding dimensionality reduction method revealed that the k-NN algorithm achieved the highest accuracy rates in distinguishing motor task pairs. k-NN demonstrated superior performance by reaching accuracy rates of 88% and above, especially in the “Lateral G.-Supination” (90.8%), “Lateral G.-Pronation” (89.2%), and “Lateral G.-Palmar G.” (88.1%) task pairs. The Naive Bayes model remained in the 71–81% accuracy range for most tasks and achieved remarkable results by exceeding 80%, especially in the “Hand O.-Pronation” and “Palmar G.-Supination” task pairs. The SVM model produced lower accuracies in general and remained below 60%, especially in tasks such as “Palmar G.-Pronation” and “Lateral G.-Pronation”. These findings show that the k-NN algorithm is the most powerful model in the feature space obtained after Spectral Embedding, Naive Bayes provides balanced but moderate results, and the classification success of SVM remains weak.
Table 9 reports the classification accuracies obtained with the Spectral Embedding method for two-class combinations. As in the previous methods,
k-NN achieved the highest performance across all pairs, with accuracies ranging between 83.9% and 90.8%. Naive Bayes yielded moderate results, consistently outperforming SVM, which again showed the lowest accuracy values among the three classifiers.
Table 10 presents the two-class classification results obtained with the t-SNE method. In this case,
k-NN achieved near-perfect accuracies for all class pairs (99.4–99.7%), clearly outperforming both SVM and Naive Bayes. While SVM also provided relatively high and stable results around 87–94%, Naive Bayes yielded lower accuracies compared to the other classifiers.
Classification analyses performed after the t-SNE dimensionality reduction method revealed that the k-NN algorithm showed the highest performance with accuracy rates close to 99% in all motor task pairs. Especially in the “Hand O.-Supination”, “Lateral G.-Supination”, and “Palmar G.-Supination” task pairs, accuracy values exceeding 99.6% show that k-NN can perform a near-perfect separation between classes in the feature space obtained with t-SNE. In this scenario, the SVM model exhibited a competitive performance with high accuracy values (89–94%), unlike other dimensionality reduction methods. The Naive Bayes model, on the other hand, showed stable success in the range of 72–83%, but fell behind k-NN and SVM. These findings show that the classification algorithm that best fits the feature representation after t-SNE is k-NN, SVM demonstrates a significant performance increase in this method, while Naive Bayes offers relatively stable but limited performance.
Classification analyses performed after the MDS dimensionality reduction method revealed that the k-NN algorithm was the most effective model by achieving the highest accuracy rates in all motor task pairs. Especially in task pairs such as “Lateral G.-Supination” (96.2%) and “Pronation-Supination” (94.4%), k-NN provided over 94% accuracy and demonstrated a high discrimination capacity between classes. The Naive Bayes model generally provided stable results in the 74–85% accuracy range and became a competitive alternative by surpassing SVM in some tasks. Although SVM exhibited a more balanced performance with the MDS method, it achieved low accuracy in some task pairs, especially in “Palmar G.-Supination” (61.5%). These findings show that k-NN is the most reliable classification algorithm in data reduced by the MDS method, Naive Bayes provided balanced but limited success, and SVM lagged behind k-NN with partial improvements.
Table 11 shows the classification accuracies obtained with the MDS method for two-class combinations. As in the other dimensionality reduction techniques,
k-NN outperformed the other classifiers, reaching the highest accuracy values between 88.1% and 96.2%. SVM achieved moderate results, while Naive Bayes performed slightly better than SVM in several cases but still remained below the accuracy levels of
k-NN (in
Figure 10).
In
Table 12, the
k-NN algorithm showed superior performance by achieving the highest accuracy rates in all tasks. The highest accuracy value of 87.7% was obtained in the “Hand O.-Palmar G.-Lateral G.” combination, while success rates above 80% were achieved in other combinations. The Naive Bayes model provided a balanced but limited success by remaining in the 63–70% accuracy range in most tasks. The SVM algorithm, on the other hand, was insufficient in classification tasks with low accuracy rates (35–42%) after the LLE method. These findings clearly show that
k-NN is the most effective model in three-class task combinations where the dimensionality is reduced with the LLE method.
Table 13 shows that the
k-NN algorithm achieved the highest accuracy rates in all tasks. Especially achieving high success in the “Hand O.-Pronation-Lateral G.” (78.9%) and “Pronation-Palmar G.-Lateral G.” (77.7%) task combinations,
k-NN was able to effectively distinguish between classes in the low-dimensional representations obtained with this method. The Naive Bayes model provided moderate accuracies in the range of 57–63% and showed a balanced performance. On the other hand, the SVM algorithm exhibited insufficient success in these combinations with accuracy rates ranging from 37–52%. These results clearly reveal that
k-NN is the most reliable and successful classification algorithm in three-class data structures where the dimensionality is reduced with the Spectral Embedding method.
Classification analyses performed on triple motor task combinations created using the t-SNE dimensionality reduction method revealed that the
k-NN algorithm showed the highest performance by achieving over 99% accuracy in each task set. Reaching accuracies of 99.2% and above in many combinations such as “Hand O.-Supination-Lateral G.”, “Hand O.-Pronation-Palmar G.” and “Supination-Palmar G.-Lateral G.”,
k-NN showed an almost error-free classification success in this method. Unlike previous methods, the SVM model provided high accuracy values (80–85%) with the t-SNE method and has become a competitive alternative. On the other hand, the Naive Bayes algorithm exhibited a limited classification performance, staying in the range of 63–69%. These results show that
k-NN is the most successful model in multi-class structures where dimensionality is reduced with the t-SNE method (
Table 14).; SVM stands out only in this method; and Naive Bayes generally performs lower.
The results given in
Table 15 show that the
k-NN algorithm exhibited the highest performance by achieving over 85% accuracy in all tasks. Achieving 87.9% accuracy in the “Pronation-Supination-Lateral G.” task trio and 86.6–86.8% accuracy in tasks such as “Hand O.-Supination-Palmar G.” and “Hand O.-Supination-Lateral G.”,
k-NN provided effective classification in low-dimensional space with the MDS method. The Naive Bayes algorithm showed limited performance by providing accuracy in the range of 58–69%. Although SVM produced relatively better results in some tasks, it generally remained at 50–67% accuracy levels. These findings reveal that the
k-NN algorithm is the most suitable classifier that provides consistent and high success for three-class combinations in which the dimensionality is reduced with the MDS method (
Figure 11).
5. Discussion
This section interprets the empirical results, synthesizing trends across manifold methods, classifiers, and class cardinalities, and relates them to prior work and application constraints.
5.1. Overall Patterns Across Manifold Methods and Classifiers
Across all class settings,
k-NN coupled with manifold learning consistently achieved the strongest performance, while SVM and Naïve Bayes trailed with method-dependent variability (
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6). In particular, t-SNE and ISOMAP frequently yielded the highest discriminability, with MDS and LLE following, and Spectral Embedding generally underperforming. These trends persisted in pairwise and triple-combination analyses (
Table 7,
Table 8,
Table 9,
Table 10,
Table 11,
Table 12,
Table 13,
Table 14 and
Table 15) and extended to the five-class summary (
Table 16).
The detailed results in
Table 5 highlight that t-SNE combined with
k-NN consistently achieved the best performance across binary, ternary, and five-class scenarios, confirming its robustness in capturing discriminative structures in EEG data. Although SVM exhibited strong performance in binary and ternary tasks (AUC∼0.95), its classification accuracy and other metrics dropped substantially in the five-class setting, suggesting that its discriminative capability diminishes under higher complexity following nonlinear embedding. Naïve Bayes showed moderate success in simpler tasks but suffered from a notable decline in metrics such as F1 score and precision in the three- and five-class scenarios, consistent with its assumption of feature independence, which is difficult to satisfy in EEG data. While the AUC for Naïve Bayes remained relatively high (up to 0.73) even in the five-class case, this did not necessarily translate to balanced predictive performance, emphasizing the need to interpret AUC in conjunction with other metrics, particularly for imbalanced datasets.
The MDS-based dimensionality reduction method combined with k-NN consistently demonstrated superior classification performance across all tasks, confirming its robustness in capturing global data structure. In binary classification, k-NN achieved near-perfect accuracy (97.1%), while SVM performed moderately well, outperforming Naïve Bayes in simpler tasks. However, SVM’s performance sharply declined in three- and five-class scenarios, highlighting its sensitivity to class overlap and increased complexity.
Naïve Bayes, although underperforming compared to k-NN, exhibited relatively balanced predictions, particularly in the five-class case, where its performance surpassed SVM in some instances. These observations suggest that MDS effectively preserves separable global features, enabling k-NN to capitalize on neighborhood information, while SVM and Naïve Bayes require careful tuning to remain competitive in complex EEG classification tasks.
Figure 7,
Figure 8 and
Figure 9 further confirm these findings, illustrating
k-NN’s consistently superior accuracy across manifold methods. For example, in binary classification,
k-NN achieved 99.6% accuracy with t-SNE, 98.4% with ISOMAP, and 97.1% with MDS, outperforming both SVM and Naïve Bayes, which generally remained in the 72–76% range. These results highlight that
k-NN is the most stable and powerful choice for EEG-based motor imagery classification across manifold learning techniques.
5.2. Binary, Ternary, and Five-Class Behavior
As class cardinality increased from 2 to 5, performance declined across methods and classifiers, which is expected due to higher decision complexity. Nevertheless, t-SNE + k-NN preserved comparatively high metrics in ternary tasks and remained competitive in the five-class setting. ISOMAP + k-NN also maintained robust performance across class counts, underscoring the benefit of geometry-preserving embeddings for EEG representations.
In the most complex classification structure with five classes (
Figure 9), although a general decrease in model performance was observed,
k-NN maintained consistently high accuracy rates. Providing 93.3% accuracy with ISOMAP, 81.6% with LLE and 89.0% with t-SNE,
k-NN produced quite effective results compared to other models despite the difficulty brought by multi-class structures. On the other hand, SVM was inadequate in the classification task with low accuracy rates (23–40%) in all methods, while Naive Bayes produced partially more balanced results (34–46%) but still lagged behind
k-NN.
All these findings reveal that the k-NN algorithm, especially when used with t-SNE and ISOMAP dimensionality reduction methods, exhibited superior performance by providing the highest accuracy rates at both low and high class numbers. While SVM produced effective results with t-SNE only in binary classification, it experienced a serious decreases in its performance as the number of classes increased. Naive Bayes, on the other hand, achieved more balanced but generally moderate accuracies and although it behaved more stably in multi-class scenarios, it could not provide sufficient performance for applications requiring high accuracy.
In the classification analyses conducted for dual motor task pairs, when the accuracy rates obtained with different dimensionality reduction methods (ISOMAP, LLE, Spectral Embedding, t-SNE, MDS) were examined comparatively, it was observed that the highest accuracy was provided by the k-NN algorithm in all task pairs. In particular, the k-NN model used with t-SNE exhibited superior performance by obtaining accuracy rates above 99% in almost all task pairs. Similar high accuracy values were achieved with the ISOMAP and MDS methods, but the results obtained with t-SNE were the most striking. Although the Naive Bayes model generally provided accuracy in the range of 74–84% and outperformed SVM in some task pairs, it did not achieve the highest success in any case. Although the SVM algorithm provided high accuracies (87–94%) with the t-SNE dimensionality reduction method, it achieved lower accuracies with other methods and fell behind k-NN. These findings show that t-SNE, as a dimensionality reduction method, is quite successful in motor task separation, especially when used with k-NN; Naive Bayes offers stable but limited success; and SVM can only produce competitive results in certain cases.
Classification analyses performed on triple motor task combinations created using the ISOMAP dimensionality reduction method revealed that the k-NN algorithm performed significantly better than other models. The highest accuracies were obtained by k-NN in all combinations, and remarkable success was achieved with 94.3% accuracy, especially in the “Pronation-Supination-Lateral G.” task trio. While the Naive Bayes model provided more limited but balanced results in the range of 62–71%, the SVM model was inadequate in this task set with low accuracies. Especially in the “Hand O.-Palmar G.-Lateral G.” combination, SVM provided only 40.5% accuracy. These findings show that the most effective classification performance in three-class task separations with the ISOMAP method was obtained by the k-NN algorithm.
As a result of the classification analyses performed on triple motor task combinations, it was observed that the k-NN algorithm achieved the highest accuracy rates by far in all methods, regardless of the dimensionality reduction method. In particular, the t-SNE method stood out with accuracy values reaching over 99% when used with k-NN. Under this structure, k-NN showed almost error-free classification success in almost every task combination. While the ISOMAP and MDS methods provided very successful results in the 85–94% accuracy range when used with k-NN, the LLE and Spectral Embedding methods produced slightly lower but still high accuracies (between 73 and 88%).
The SVM algorithm was able to reach high accuracy values (80–85%) only with the t-SNE method, while in other methods it generally remained in the 35–67% range and fell far behind k-NN. The Naive Bayes model, on the other hand, showed balanced but limited success in all methods, remaining in the 57–71% accuracy range and at best could approach k-NN.
When these findings are evaluated in general, it is revealed that the highest and most consistent performance in triple classification problems is provided by the k-NN algorithm, especially with the t-SNE dimensionality reduction method. Other algorithms produced reasonable results only under certain conditions but fell behind k-NN in terms of overall success. Therefore, the t-SNE + k-NN combination can be considered as the most reliable and recommended structure in EEG data analyses based on three-class motor task separation.
The findings obtained in the five-classification scenario clearly show that the
k-NN algorithm achieves the highest classification accuracies among all dimensionality reduction methods. Especially when used with ISOMAP (79.3%) and t-SNE (79.7%) methods,
k-NN achieved high success even in five-class separations, and these combinations were the strongest alternatives for multi-class structures. The MDS method closely follows these two methods with 79.1% accuracy. On the other hand, the LLE (67.9%) and Spectral Embedding (64.5%) methods remained more limited in class separation with relatively lower accuracy rates, even when used with
k-NN (
Figure 12).
The Naive Bayes model produced accuracy values between 49.8% and 54.3% under all methods and showed a moderate, stable, but limited classification performance. Naive Bayes, which gave the best result with 54.3% accuracy using ISOMAP, generally fell far behind
k-NN (orange bars in
Figure 12).
On the other hand, the SVM algorithm stood out as the weakest model for five-class structures; accuracy remained below 30% in most methods. Especially when used with t-SNE, it achieved only 9.8% accuracy, indicating that this model is not compatible with high class numbers and low-dimensional representations (see
Figure 12). The highest accuracy for SVM was achieved with the MDS method at 37.2%.
5.3. Detailed View of the Best-Performing Configuration
The best-performing pipeline (t-SNE +
k-NN, five-class) achieved accuracy of 99.7% and an AUC of 0.995 (95% CI: 0.992–0.998), with sensitivity 0.98 and specificity 0.97 (
Table 17). These ROC-based metrics contextualize the single-point accuracy and indicate excellent separability even under multi-class conditions.
To complement the accuracy results, we provide a comprehensive statistical evaluation for the best-performing configuration (t-SNE +
k-NN, five-class). In addition to accuracy, we report AUC with 95% CI, sensitivity, and specificity at a fixed threshold of 0.5 (
Table 17).
The average five-class classification accuracy across all six participants was 89.0% ± 4.2%, highlighting inter-subject variability even in this limited cohort.
5.4. Method-Specific Observations
SVM benefited substantially from t-SNE in binary and ternary settings yet degraded in the five-class case, suggesting sensitivity to class overlap after non-linear embeddings. Naïve Bayes provided stable but modest performance, likely reflecting independence assumptions that are challenged by EEG feature dependencies. Spectral Embedding’s comparatively low scores align with its sensitivity to graph construction in noisy settings.
5.5. Relation to Prior Work and Application Constraints
The cross-study perspective (healthy real-time vs. SCI offline [
60]) indicates that manifold learning + shallow classifiers transfer across acquisition contexts, with t-SNE offering top-tier accuracy and ISOMAP providing a favorable accuracy/latency balance for real-time feasibility. While t-SNE remained highly accurate, ISOMAP’s lower processing time can be more pragmatic for patient-oriented deployments.
To further support claims of real-time feasibility,
Table 18 summarizes approximate per-trial computation times for each manifold learning method combined with the
k-NN classifier. These measurements, obtained on a standard workstation (Intel i7 CPU, 16 GB RAM, MATLAB R2024a, Orange 3.36), indicate that while t-SNE is the most computationally expensive (∼350 ms per trial), ISOMAP and MDS maintain processing times below 150 ms, making them practical for near real-time BCI systems. These results complement accuracy-based evaluations by demonstrating that the proposed pipelines are not only accurate but also computationally feasible for online operation, subject to further optimization on embedded or GPU-accelerated platforms.
5.6. Limitations and Future Directions
While the reported classification accuracies, particularly those obtained with the t-SNE + k-NN pipeline, are exceptionally high, these results must be interpreted with caution. The controlled experimental setup, relatively small sample size, and the use of subject-dependent validation may have contributed to the elevated performance metrics. Although these findings demonstrate the feasibility and promise of manifold learning techniques for real-time EEG decoding, they should not be taken as a direct indication of generalizable performance in large-scale or clinical environments. Future research will focus on validating these approaches with subject-independent cross-validation schemes and larger, more heterogeneous datasets to ensure robustness and reproducibility. This explicit acknowledgment is intended to mitigate the risk of overestimation and to provide a realistic perspective on the proposed methodology relative to the current state-of-the-art.
This study included six healthy subjects, which limits generalizability. While averaged metrics (e.g., mean accuracy ± standard deviation) provide an overview, statistical power remains low. Future work will increase the sample size, incorporate statistical significance testing across participants, and include neurologically impaired populations to strengthen conclusions. Despite strong metrics, performance inevitably decreases with higher class counts and potential inter-subject variability. Future work should report per-subject variability, calibrate thresholds for sensitivity/specificity in imbalanced settings, and extend latency profiling to embedded/edge deployments. Robust hyperparameter selection (e.g., nested CV, grid ranges) and explicit artifact handling remain critical to guard against overestimation and support generalization.