Next Article in Journal
Clinical Outcomes of Passive Sensors in Remote Monitoring: A Systematic Review
Previous Article in Journal
A Wearable Open-Source Neuroprosthesis/Neuro-Orthosis for Restoring Hand Function
Previous Article in Special Issue
Estimation of Pressure Pain in the Lower Limbs Using Electrodermal Activity, Tissue Oxygen Saturation, and Heart Rate Variability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Electroencephalography Motor Execution Signals Using a Hybrid Neural Network Based on Instantaneous Frequency and Amplitude Obtained via Empirical Wavelet Transform

1
Institute of Communication and Computer Networks, Poznan University of Technology, 60-965 Poznan, Poland
2
Faculty of Mechanical Engineering, Poznan University of Technology, 60-965 Poznan, Poland
3
Institute of Material Technology, Poznan University of Technology, 60-965 Poznan, Poland
4
Institute of Electrical Engineering and Electronics, Poznan University of Technology, 60-965 Poznan, Poland
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3284; https://doi.org/10.3390/s25113284
Submission received: 24 April 2025 / Revised: 20 May 2025 / Accepted: 22 May 2025 / Published: 23 May 2025

Abstract

:
Brain–computer interfaces (BCIs) have garnered significant interest due to their potential to enable communication and control for individuals with limited or no ability to interact with technologies in a conventional way. By applying electrical signals generated by brain cells, BCIs eliminate the need for physical interaction with external devices. This study investigates the performance of traditional classifiers—specifically, linear discriminant analysis (LDA) and support vector machines (SVMs)—in comparison with a hybrid neural network model for EEG-based gesture classification. The dataset comprised EEG recordings of seven distinct gestures performed by 33 participants. Binary classification tasks were conducted using both raw windowed EEG signals and features extracted via bandpower and the empirical wavelet transform (EWT). The hybrid neural network architecture demonstrated higher classification accuracy compared to the standard classifiers. These findings suggest that combining featuring extraction with deep learning models offers a promising approach for improving EEG gesture recognition in BCI systems.

1. Introduction

Over the years, brain–computer interfaces (BCIs) have gained increasing interest. This communication link supports individuals who are unable to use everyday technologies in a conventional way, providing them with essential accessibility and interaction capabilities. It assists individuals who, as a result of various incidents, have experienced spinal cord injuries, brain trauma, or strokes, as well as those affected by diseases that impair the nervous and muscular systems, such as amyotrophic lateral sclerosis (ALS), multiple sclerosis, or muscular dystrophies [1]. The need for physical, manual interaction with external devices is eliminated by utilizing the user’s brain activity [2,3]. This allows brain–computer interfaces to be applied in various medical and rehabilitation applications, with their functionality limited only by the patient’s needs and the ability to generate the appropriate EEG signal pattern. BCIs are used in rehabilitation systems [2,4,5,6,7] and they enable communication with the external world [8,9,10], control of prosthetic limbs [11,12,13,14], wheelchairs [15,16,17], and hearing aids [18], and they provide people with support in pursuing hobbies and passions, such as painting [19]. Their primary classification is based on invasiveness, distinguishing between invasive and non-invasive BCIs. The first type allows for signal measurement inside the skull, requiring surgical intervention. Implanted electrodes measure brain activity beneath the skull and within the cerebral cortex. A less invasive method, electrocorticography (ECoG), records neural signals from between the cortical surface and the skull. Invasive brain–computer interfaces provide higher spatial resolution compared to non-invasive ones, which rely on devices that do not directly interfere with the patient’s body. In the case of non-invasive BCIs, brain signal measurements are primarily conducted using magnetoencephalography (MEG), electroencephalography (EEG), functional magnetic resonance imaging (fMRI), positron emission tomography (PET), and functional near-infrared spectroscopy (fNIRS). Each of these techniques is based on different physical principles and varies in terms of equipment size, cost, measurement method, and spatial and temporal resolution [2,3,17,20]. Among these, the most commonly used method for data acquisition in non-invasive brain–computer interfaces is EEG. Modern BCIs are based on the principle that each brain region is associated with a specific function of the body. Popular signals include the following: Steady-State Visual Evoked Potentials (SSVEPs), which can be detected in the occipital lobe; P300 potential (also known as the P300 wave), which appears in the parietal lobe; Event-Related Desynchronization (ERD) and Event-Related Synchronization (ERS), which occur primarily in sensorimotor areas [3,21,22,23,24]. ERD/ERS are therefore related to the motor activity of the extremities. Movement of any limb first leads to desynchronization, meaning a decrease in mu and beta oscillations, which is observed during the body’s preparation for movement. After the actual movement is performed, these oscillations increase again (synchronization). In the case of upper limbs, this relationship is visible in the contralateral hemisphere of the brain relative to the moving extremity (i.e., for the right hand, it is the left hemisphere, and for the left hand, the right hemisphere). For the lower limbs, however, the brain region where these changes occur is located in the central area. Interestingly, this phenomenon also occurs during motor imagery (MI) [24,25,26,27]. Brain–computer interfaces based on motor execution (ME) or imagery for detecting ERD and ERS mainly use bandpower as a feature [28,29,30]. However, in recent years, researchers have focused their attention on identifying additional features that can enhance the effectiveness of classification. Among the most prominent approaches are the use of the Spatial Covariance Matrix (SCM), statistical and spatial features, wavelet coefficients at various decomposition levels, higher-order statistics (HOSs), Hjorth parameters, fractal dimensions, and their combinations [31,32,33]. The research that utilized instantaneous amplitude (IA) and instantaneous frequency (IF) , along with HOS derived from them, is worth noting; this research yielded satisfactory results in motor imagery classification [34].
The key components of BCIs are preprocessing, feature extraction, and classification [34]. Proper signal preparation is essential for eliminating noise or artifacts and isolating the desired signal bands. This can be achieved through appropriate filtering or other methods such as Principal Component Analysis (PCA) and independent component analysis (ICA) [35,36,37]. It is also worth noting that PCA and ICA are also used as a feature reduction techniques [38,39]. In addition, independent components (ICs) corresponding to electrical sources identified by ICA can be used as a features to support ERD/ERS classification. In general, ICs representing time–frequency domain activity contribute more significantly to classification accuracy than other components [40,41]. Effective approaches also include combining ICA with Powers Spectral Density (PSD) for feature extraction and employing supervised independent component analysis algorithms for MI tasks [42,43]. Feature extraction and classification enable the automated and desired operation of a device based on expected actions. Among the available feature extraction methods, notable ones include Fourier Transform (FT), autoregressive (AR) models, common spatial patterns (CSPs), statistical clustering, iterative spatio-spectral patterns learning (ISSPL), and optimal allocation [44,45,46,47,48,49]. Despite their extensive use in motor-related applications, these methods also come with various drawbacks, such as the lack of temporal information for EEG signals, susceptibility to noise, and a low success rate in classification [35,50]. In recent years, signal decomposition (SD) methods have been increasingly used for the same issue. These methods operate by generating multiple sub-band signals. Among them are wavelet transform (WT), wavelet packet decomposition (WPD), discrete wavelet transform (DWT), continuous wavelet transform (CWT), empirical mode decomposition (EMD), and empirical wavelet transform (EWT) [34,35,45,50,51,52]. Unlike traditional wavelet-based approaches such as DWT or CWT, the EWT method enables efficient extraction of time-varying components by using empirically determined wavelet filter parameters. This avoids unnecessary data redundancy and improves the precision of separating nonstationary and multi-component signals, thereby minimizing band mixing artifacts. In contrast, classical wavelet methods apply fixed filter parameters over time, which can lead to erroneous segmentation, such as splitting useful signal components or separating channels that contain no meaningful information. Findings across various disciplines indicate that EWT offers significant advantages over classical DWT, especially in the context of nonstationary signals such as EEG data [53]. Prior studies have shown that EWT achieves improved performance in signal decomposition and feature extraction compared to traditional wavelet-based approaches. Moreover, the existing literature suggests that EMD offers improved segmentation capabilities relative to DWT [54], while other work has demonstrated that EWT outperforms EMD in specific applications [55], further reinforcing its utility as a robust tool for analyzing complex, time-varying signals. EWT has been successfully applied in the medical domain, including signal processing (such as EEG and ECG), medical image analysis, and anatomical structures segmentation [55,56,57,58]. One noteworthy application is its use in motor imagery-related research [34,35,50,51]. Therefore, an area worthy of further investigation is its application in the motor execution paradigm, which could be useful in BCI systems supporting rehabilitation. EWT is an extended version of empirical mode decomposition used for analyzing non-linear and nonstationary signals. It enables multiresolution analysis (MRA) by employing an adaptive wavelet partitioning scheme, decomposing the signal into distinct sub-bands based on the information it carries [55]. Due to the nature of its operation, the EWT can produce inappropriate segmentations in the presence of high noise, which may negatively impact the performance of the machine learning method applied. One approach to mitigate this issue is the Enhanced Empirical Wavelet Transform (EEWT), which improves signal decomposition (particularly in non-stationary signals) through spectrum pre-regularization and segmentation based on the signal’s spectral envelope. However, EEWT is computationally more complex and requires careful hyperparameter optimization, making it challenging to implement on hardware with limited processing capabilities.
Gesture recognition using EEG signals requires advanced time–series analysis methods. Traditional approaches rely on feature engineering-based algorithms and classical statistical models [59]. Among the classifiers commonly used in ME and MI tasks that yield satisfactory results, linear discriminant analysis (LDA) [6,60,61,62] and support vector machine (SVM) [6,35,51,61,62,63] are notable. Furthermore, their extensions, such as least squares support vector machine (LS-SVM) [34,34] and time-varying linear discriminant analysis (TVLDA) [29], are also used. However, advancements in deep learning have enabled the automatic extraction of relevant information without the need for manual feature extraction [64]. In neural networks designed for EEG signals, the use of convolutional layers is particularly important to analyze for the spatial structure of the data. Convolutional neural networks (CNNs) enable the automatic detection of local patterns in brain signals, which is crucial for identifying the distinctive features of a given gesture [65]. The application of long short-term memory (LSTM) networks enables the capture of long-term dependencies over time, which is essential for the sequential nature of EEG signals [66,67,68]. To ensure effective gesture classification, the appropriate selection of dense layers and the output layer is crucial [69]. Despite significant progress being made in applying neural networks to EEG analysis, several challenges require further research. One of the main issues is the limited availability of data, which can lead to model overfitting [70]. Another challenge is model personalization, as brain signal patterns can vary significantly between users, making classification more difficult [71,72]. The need to develop adaptive methods that allow effective adjustment of the model to individual users is becoming increasingly important in the context of real-world BCI applications [73,74].
The effectiveness of gesture recognition can be enhanced by incorporating other electrophysiological signals, such as electromyography (EMG) [75]. However, this approach complicates the measurement infrastructure [76,77]. A simpler and more convenient alternative is to process brain signals directly, which can be recorded using widely available EEG headsets. When detecting movement in specific limbs (e.g., left arm vs. right arm or arm vs. leg), motion recognition may prove easier for traditional classifiers. This is due to brain physiology and the observable dominance of potentials in specific electrode placements (e.g., for right-hand movement, ERD/ERS can be observed in the left hemisphere, leading to detectable changes in channels such as C3). In contrast, the detection of more complex gestures has not yet been sufficiently explored in the context of non-invasive signal acquisition methods, necessitating extensive further research.
In this article, we compare the use of artificial neural networks (ANNs) with standard classifiers for detecting specific gestures. In addition, we evaluate the effectiveness of ANNs using both filtered signals and extracted features as input. The aim of this work is to develop an algorithm capable of detecting specific gestures, which will contribute to the advancement of non-invasive BCI systems designed to support the rehabilitation of individuals experiencing neuromuscular dysfunction, such as those recovering from a stroke.

2. Materials and Methods

The article presents a comparison of classification using the two most common classifiers in motor imagery applications (LDA and SVM) and the results achieved with a hybrid neural network (HNN) that integrates CNN and LSTM, allowing efficient processing of the spatial–temporal aspects of EEG signals [75,78]. Furthermore, the influence of signal processing and feature extraction on the results will be analyzed. All classification tools will be trained based on the preprocessed signal, along with feature extraction. For the neural network, training will also be conducted on data without extracted features to enable a comparison.

2.1. Dataset Description

In the presented study, the available online dataset [79] was utilized. The experiment involved 33 non-disabled subjects without any neuromuscular dysfunction. This group consisted of 17 right-handed men, 10 right-handed women, 3 left-handed men, and 3 left-handed women, with an average age of 25 ± 5.5 years. Among all participants, 24 were familiar with and had experience using BCIs. The measurement setup is an electroencephalograph. Brain signals acquisition was performed using the OpenBCI Ultracortex “MARK IV” headset, which included the following measurement electrodes: F3, F4, C3, Cz, C4, P3, Pz, and P4. This configuration enabled coverage of the motor cortex, thus maximizing the registration of motor-related potentials. One of two reference electrodes serving as the reference and bias (ground with common-mode noise rejection) for EEG system was placed on each earlobe. All electrodes were placed according to the international 10-20 system. The sampling rate was 250 Hz. The experimental protocol was explained to the subjects beforehand. It consisted of performing six repetitions (the last two repetitions were performed while standing) of each of the seven gestures: large diameter, cut, medium diameter, open, power, prismatic pinch, and three-finger sphere grasp. Each trial was preceded by the display of the target gesture on a screen located 1.5 m from the participant, along with a simultaneous acoustic tone signaling the need to begin performing the indicated gesture for 5 s. After each trial, a 3 s rest interval followed. In total, 42 trials were conducted per subject. To minimize environmental interference and enhance signal quality, all experimental sessions were conducted in a noise-reduced room.

2.2. Preprocessing

The EEG data in the dataset are represented in microvolts (μV). Initially, the signal was filtered using a 5th order Butterworth bandstop filter [80] to remove the 60 Hz component to prevent interference from power line noise. Then, two processing variants were considered (Figure 1), additionally involving the use of either all measurement channels or only C3, Cz, and C4. In the first variant, a frequency band from 8 to 25 Hz was extracted using a 5th order Butterworth bandpass filter, corresponding to the mu and beta rhythm ranges [78]. To increase the number of training samples, the signals were segmented using a sliding window with 60% overlap [75]. Finally, band power was computed for each window, serving as the feature for classifier training. It is important to emphasize that the quality of input data is crucial for the correct performance of both standard classifiers and neural network models, making preprocessing steps like filtering and segmentation essential for reliable model training and evaluation. In the second variant, after removing power line interference, the empirical wavelet transform method was applied to decompose the nonstationary signal into different modes, each adjusted around an explicit frequency to fulfill the properties of intrinsic mode functions (IMFs). This was achieved using filter banks obtained through a process that first computed the signal spectrum in the range of 0 to π using Fast Fourier Transform (FFT). Then, EWT boundary detection was performed (hyperparameters were optimally selected using brute force optimization method), followed by the application of empirical wavelets to all segmented spectra as bandpass filters [55]. Using the MATLAB 2024b library created by the author of EWT [81], the EEG signal was decomposed for each channel, empirically setting the number of modes to 10 per channel. For each mode, power spectrum analysis was performed using the Welch PSD method excluding frequencies up to 2 Hz, to detect the mode with the highest power concentration, indicating its highest sensitivity to the ME task. Based on the findings of [34], feature extraction focused on HOS derived from instantaneous frequency (IF) and instantaneous amplitude (IA). To extract IF and IA, the Hilbert–Huang Transform (HHT) approach [82] was applied to each sliding window with 60% overlap from the most sensitive mode. For each IMF in the first step, the Hilbert–Huang procedure computes the analytic signal:
z i ( t ) = x i ( t ) + j H { x i ( t ) } ,
where H x i is the Hilbert Transform of x i (data values in the set). Then, it expresses z i as
z i ( t ) = a i ( t ) e j θ i ( t ) ,
where a i (t) is the instantaneous amplitude and θ i (t) is the instantaneous phase. Finally, it computes the instantaneous frequency
ω i ( t ) d θ i ( t ) d t .
For classification, four time domain feature vectors were extracted from the instantaneous components:
  • Mean absolute deviation (MAD) measures statistical data distribution (4).
  • Interquartile range (IQR) determines dataset distribution by excluding lower and upper quartiles (5) [83,84].
  • Skewness (Sk) describes the asymmetry of data distribution around the mean (6).
  • Kurtosis (Kt) quantifies data flatness compared to a Gaussian distribution; combined with skewness, this helps identify linear, stationary, and Gaussian anomalies in signals (7) [45].
MAD = 1 n i = 1 n | x i x ¯ | ,
IQR = Q 3 Q 1 ,
Sk = 1 n i = 1 n ( x i x ¯ ) 3 ( 1 n i = 1 n ( x i x ¯ ) 2 ) 3
Kt = 1 n i = 1 n ( x i x ¯ ) 4 ( 1 n i = 1 n ( x i x ¯ ) 2 ) 2
where n is the number of data values, x ¯ is average value of the dataset, Q1 (first quartile) is the median of the lower half of the dataset, and Q3 (third quartile) is median of the upper half of the dataset. The choice of these four features is supported by their increasing popularity in the analysis of biomedical signal, as well as by findings from MI detection studies, where their use has been shown to improve classification accuracy and enhance the stability of detection results across different MI tasks [34,45,83,84,85]. Therefore, it is worth investigating whether these features can yield similar benefits in the context of ME tasks.

2.3. Deep Learning Models

In this study, deep learning was employed to perform binary classification of EEG gesture data using both windowed signals and features derived from the computed band power and the empirical wavelet transform, specifically the instantaneous amplitude and frequency calculated from the same windows. While patient-oriented modeling proves more efficient in this case [86], our dataset was not sufficiently large to support personalized models. Therefore, we focused on developing a generalized model capable of capturing common neural features between subjects. For each classification task, a one-vs-rest (OvR) strategy was used, in which a dedicated model was trained to distinguish a single gesture class from all others. In order to increase reliability and generalizability, training and testing was carried out in a mixed dataset that represents the trials of a given gesture made by individual subjects. The deep learning architecture was designed to take advantage of the temporal and spatial structure of the input signal. The architecture in general consists of the following components:
  • Input layer—input data in form of time–series matrices.
  • Convolutional layers—a stack of two 1D convolutional layers (32 filters of size 5 and 64 filters of size 3, respectively) capture local temporal features from the multichannel signal.
  • Pooling layers—max-pooling layers with a stride of 2 reduce the temporal dimension and help avoid overfitting [87].
  • Bidirectional LSTM layer—a bidirectional long short-term memory (BiLSTM) layer with 64 units enables the model to learn long-range temporal dependencies in both forward and backward directions, which is crucial for modeling the dynamics of gestures [88].
  • Dense layers—a fully connected layer with 64 neurons (ReLU activation) and an output layer with a single neuron (sigmoid activation) produce the binary prediction.
The model was trained with the 80/20 split dataset, utilizing the Adam optimizer and the binary cross-entropy loss function. Batch sizes ranging from 6 to 20 and the number of epochs per gesture were adjusted to optimize performance. Due to the inherent class imbalance in the OvR setting, the training set was balanced using RandomOverSampler from the "imbalanced-learn" package to avoid model bias toward the majority class [89]. Each model was evaluated using unseen trials of the target gesture as positive samples and gestures of all other classes as negative samples.

2.4. Performance Metrics

To evaluate the accuracy of the classification, the dataset was divided into training and test sets using the k-fold cross-validation method (in the case of LDA and SVM). The parameter k determines the number of equal-sized partitions into which the dataset is divided. A higher value of k results in more iterations and more partitions (which also means significantly more feature vectors in the training set compared to the test set) [90,91]. In this study, the parameter k was set to 5, which means that the dataset was divided into five equal partitions, with 80% used for training and 20% for testing. This split ensures a sufficiently large test partition for obtaining reliable results while maintaining a reasonable number of experiments, reducing computational time, and minimizing estimator variance [34]. Cross-validation was performed using a dataset of 1386 samples, each sample representing an individual’s execution of a particular gesture and comprising features derived from 61 time windows per EEG channel. Accuracy, representing the ratio of the number of correct assessments to the number of all assessments (8), and F1 score, calculated as the harmonic mean of the precision and recall scores (9), were used as statistical parameters to measure algorithm performance:
accuracy = TP + TN TP + TN + FP + FN ,
F 1 score = TP TP + 1 2 ( FP + FN ) ,
where TP (true positive) represents the number of correctly classified executions of the desired gesture. TN (true negative) refers to the number of correctly classified executions of any other gesture. FP (false positive) indicates the number of incorrectly classified executions of other gestures that were misclassified as the desired gesture. Conversely, FN (false negative) represents the number of incorrectly classified executions of the desired gesture, misclassified as one of the other gestures.

3. Results

The classifiers were trained using the OvR strategy based on the dataset studied in different configurations. This includes extracting channels C3, Cz, and C4, as well as evaluating the set of characteristics that considers IF and IA together and separately in the second variant. The best accuracy for the given variants and their combinations, which also achieved the highest F1 score in the studied dataset, is presented in Table 1 (the result in bold text indicates the highest score obtained in the entire study).
In addition, for comparison, the results of the remaining classifiers for the configuration that achieved the best performance are also included. During the study, it was observed that the best results were obtained when using signals from all EEG measurement channels. Therefore, information on the channels utilized was omitted from the variant description in Table 1. To compare the proposed hybrid model with traditional classifiers, Wilcoxon signed rank tests were performed using both the F1 score and the accuracy metrics in all gesture classes. The results indicate that the improvements achieved by the hybrid model are statistically significant (p < 0.05).

4. Discussion

The collected EEG dataset comprised seven gestures performed by 33 participants, each of whom completed six tests of each movement, with the last two repetitions performed while standing. However, this dataset presented significant limitations. In particular, only the signal associated with the execution of the gesture was recorded, without any temporal segments captured before or after the gesture. This required gesture detection to rely solely on the data recorded during its execution, thus necessitating a one-vs-rest strategy. Such an approach limits the ability to thoroughly analyze the dynamics of brain activity (e.g., it may omit the preparatory phase characterized by neural desynchronization or the post-movement physiological phenomenon of neural synchronization) and significantly complicates the ability of models to learn and recognize patterns. It is also important to note that the nature of EEG signals does not allow for typical interpolation or augmentation techniques, which forces deep learning models to rely solely on real data; this is problematic in this case due to the relatively small dataset. In addition, the spatial resolution of EEG is relatively low compared to other invasive methods. Significantly improved detection accuracy of individual finger movements is possible through the use of electrocorticography [29]. This is equivalent to making the entire task more challenging and making individual gestures more difficult. Furthermore, the dominant hand of the patients can also influence all of the results, potentially affecting the power of the generated signals, as well as the pattern of signal occurrence on specific measurement electrodes. The analysis included the use of basic classifiers such as LDA and SVM, as well as a complex CNN model with a bidirectional LSTM layer, trained on sequential data. An approach used the full sequence of 61 time windows without additional feature extraction. Another used a dataset consisting of features extracted from the data within each window. The tested approaches used band power (in the first variant) and feature extraction based on higher-order statistics derived from IA and IF parameters (in the second variant), which also reduced data dimensionality while preserving their characteristics. The highest accuracy and F1 score were achieved using the second variant, which applied higher-order statistics derived from both instantaneous frequency and instantaneous amplitude.
The highest obtained accuracy was equaled to 81.9%. The baseline for comparison was a study [75] using the identical dataset. In [75], an accuracy of 49.3% and an F1 score of 49.0% were reported. However, it should be noted that the approach presented in that article differs from the one compared here. In the present study, the focus was on creating a generalized algorithm to distinguish gesture data, while the other article involved tailoring a neural network model to each individual participant, which could ultimately have led to a higher average F1 score.
In the case of the authors of [34], who studied the motor imagery paradigm in five patients individually, their best results (accuracy 95.19%, F1 score was not provided) were achieved using the same HOS derived from IA. Their study was limited to only two MI tasks for the right hand and leg and did not cover as many gestures performed by the same limb. Given the physiological basis of ERD/ERS, detecting these events might have been easier for the classifier, at least because of the dominance of this phenomenon in different EEG channels.
Another example of using EEG signals to gesture recognition was research presented in [92]. The algorithm applied is NeuroGraps—a dual-stage deep learning framework. Two scenarios were considered—learning to distinguish between two gestures and learning to distinguish between four gestures. The number of gestures significantly influenced the accuracy obtained (68% for four gestures, 86% for two gestures). Learning was based not only on EMG, but was supported by EMG signals. In contrast, only the EEG signal was used during network testing. The accuracy obtained during the tests performed (89.1%) is comparable to a scenario in which two gestures were considered. Nevertheless, adding more gestures can significantly reduce the accuracy of the classification—therefore, it may be necessary to extend the signal set with EMG data. This phenomenon was also observed in a study [93] where a classification accuracy of 60.8–62.4% was achieved applying single-modal EEG. In the case of recognition based on the EEG signal, but with a pre-trained model using both the EEG and the EMG signals, an accuracy of 65.1–66.5% was obtained. Using both EEG and EMG signals for identification, the accuracy increased to a range of 85.93–87.49%. A comparable accuracy for classification based on both EEG and EMG signals (80.5%) was obtained in a study [94] in which dynamic elbow flexion–extension movements were performed under varying load conditions. However, research for optimal gesture recognition tools based on the EEG signal alone is important to minimize the impact of signal strength-dependent variability of the EMG signal, injury, or surgical interventions performed.
The variability of the accuracy obtained in the population was subject-specific. Within one model, the difference between the patient with the highest F1-score and the lowest was even 51%. F1-scores vary significantly between patients, suggesting that the model does not generalize equally effectively for all patients. The results suggest that individual differences between patients have a significant impact on the performance of the classifier. It may be recommended to personalize the model. For all the cases analyzed, the model was significantly more effective in classifying open, power, prismatic pinch, and three-finger sphere grasp gestures. The results obtained indicate that the resolution of EEG signal may not be sufficient to classify the other three hand gestures.
Despite not accounting for the variability introduced by the displacement of the electrode or the motion artifacts, our model demonstrated satisfactory performance under various measurement conditions. This suggests that stable temporal and spatial patterns in EEG signals may offer a degree of resilience to common sources of signal degradation [95].
Due to the limited dataset, balanced for each gesture with oversampling, the models used were carefully observed to determine the possibility of the occurrence of overfitting by comparing the accuracies achieved for the different training and testing sets. However, based on the large differences between the performance of individuals and gestures, it may be necessary to personalize the model.
In the study presented in this article, the detection of dependencies that may occur across all channels is not yet fully understood. Standard classifiers consistently performed worse than ANN, indicating the superiority of the latter and the strong potential for its use in BCI applications. The low F1 score observed in the results presented may be due to the limited number of training examples, algorithmic and data instability, and the high variability in EEG signals influenced by the characteristics of individual patients. It should be noted that the CNN + LSTM model worked correctly from a technical standpoint, but was unable to generalize to new cases given the limited dataset. It is also worth noting that the neural network trained on extracted features achieved better results than when using the time-windowed signal after filtering.
During the experiments, it was observed that, from a user’s perspective—particularly individuals with limited motor abilities for whom BCI could have real-world applications—the presented approaches could achieve satisfactory performance. However, its personalized mode capabilities should be considered. In other words, the system would need to be trained separately for each individual, which could prolong the calibration process and limit user convenience.

5. Conclusions

Applied ANNs architecture demonstrate higher effectiveness compared to standard classifiers. This may be due to the complexity of brain-derived signals. Therefore, future research should place strong emphasis on the development of neural network models capable of detecting the desired phenomena from brain-generated signals, while also offering adaptability to individual patients. It would be worth considering solutions that allow the classifier to be continuously fine-tuned during use, which would positively impact its performance by enabling adaptation to individual traits and the environment in which the patient operates.
To enhance the practical applicability of the presented methods, it would be necessary in the future to collect a significantly larger and more diverse dataset, including resting-state periods as well as pre- and post-gesture phases. This approach would not only improve the accuracy of movement initiation detection, but also leverage dynamic changes in brain activity to enhance model performance. Furthermore, techniques such as transfer learning or domain adaptation could enable initial training on a large general dataset, followed by rapid user-specific adaptation using minimal samples. Higher-order statistics derived from instantaneous frequency and amplitude may yield promising results in the case of motor execution, which should be further investigated on a larger dataset. It would also be advisable to evaluate the model’s behavior by comparing gesture execution signals with resting state (baseline) signals. In addition, performance should be assessed on a dominant hand-only dataset to isolate motor-specific neural patterns. Finally, a comprehensive comparative study is needed to validate these methods for individual applications.
Despite these limitations, the experimental results demonstrate that high classification accuracy can be achieved using EEG signals elicited by hand gestures. This represents a significant advance toward the development of simple and responsive brain–computer interfaces for patients who require assistive technologies. However, creating a universal model that performs well in diverse users remains a challenge due to the limited number of trial runs and the inherent variability of EEG signals between individuals.

Author Contributions

Conceptualization, P.Z., K.F., A.M.-C. and P.K.; methodology, P.Z., K.F., A.M.-C. and P.K.; software, P.Z. and K.F.; validation, A.M.-C. and P.K.; formal analysis, P.K.; investigation, P.Z. and K.F.; resources, A.M.-C.; data curation, P.Z. and K.F.; writing—original draft preparation, P.Z. and K.F.; writing—review and editing, A.M.-C. and P.K.; visualization, P.Z.; supervision, P.K.; project administration, P.K.; funding acquisition, P.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by statutory funds from the Faculty of Control, Robotics and Electrical Engineering of the Poznan University of Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wolpaw, J.R.; Birbaumer, N.; McFarland, D.J.; Pfurtscheller, G.; Vaughan, T.M. Brain–computer interfaces for communication and control. Clin. Neurophysiol. 2002, 113, 767–791. [Google Scholar] [CrossRef] [PubMed]
  2. Chaudhary, U.; Birbaumer, N.; Ramos-Murguialday, A. Brain–computer interfaces for communication and rehabilitation. Nat. Rev. Neurol. 2016, 12, 513–525. [Google Scholar] [CrossRef] [PubMed]
  3. Rao, R.P.N. Brain-Computer Interfacing: An Introduction; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
  4. Sieghartsleitner, S.; Sebastián-Romagosa, M.; Ortner, R.; Cho, W.; Guger, C. BCIs for stroke rehabilitation. In Brain-Computer Interfaces; Elsevier: Amsterdam, The Netherlands, 2025; pp. 131–150. [Google Scholar]
  5. Ang, K.K.; Guan, C. Brain–computer interface for neurorehabilitation of upper limb after stroke. Proc. IEEE 2015, 103, 944–953. [Google Scholar] [CrossRef]
  6. Sieghartsleitner, S.; Sebastián-Romagosa, M.; Cho, W.; Grünwald, J.; Ortner, R.; Scharinger, J.; Kamada, K.; Guger, C. Upper extremity training followed by lower extremity training with a brain–computer interface rehabilitation system. Front. Neurosci. 2024, 18, 1346607. [Google Scholar] [CrossRef]
  7. Sebastián-Romagosa, M.; Cho, W.; Ortner, R.; Sieghartsleitner, S.; Von Oertzen, T.J.; Kamada, K.; Laureys, S.; Allison, B.Z.; Guger, C. Brain–computer interface treatment for gait rehabilitation in stroke patients. Front. Neurosci. 2023, 17, 1256077. [Google Scholar] [CrossRef]
  8. Farwell, L.A.; Donchin, E. Talking off the top of your head: Toward a mental prosthesis utilizing event-related brain potentials. Electroencephalogr. Clin. Neurophysiol. 1988, 70, 510–523. [Google Scholar] [CrossRef]
  9. Kohler, J.; Ottenhoff, M.C.; Goulis, S.; Angrick, M.; Colon, A.J.; Wagner, L.; Tousseyn, S.; Kubben, P.L.; Herff, C. Synthesizing speech from intracranial depth electrodes using an encoder-decoder framework. arXiv 2021, arXiv:2111.01457. [Google Scholar] [CrossRef]
  10. Wairagkar, M.; Card, N.S.; Singer-Clark, T.; Hou, X.; Iacobacci, C.; Hochberg, L.R.; Brandman, D.M.; Stavisky, S.D. An instantaneous voice synthesis neuroprosthesis. bioRxiv 2024. [Google Scholar] [CrossRef]
  11. Collinger, J.L.; Wodlinger, B.; Downey, J.E.; Wang, W.; Tyler-Kabara, E.C.; Weber, D.J.; McMorland, A.J.; Velliste, M.; Boninger, M.L.; Schwartz, A.B. High-performance neuroprosthetic control by an individual with tetraplegia. Lancet 2013, 381, 557–564. [Google Scholar] [CrossRef]
  12. Hochberg, L.R.; Bacher, D.; Jarosiewicz, B.; Masse, N.Y.; Simeral, J.D.; Vogel, J.; Haddadin, S.; Liu, J.; Cash, S.S.; Van Der Smagt, P.; et al. Reach and grasp by people with tetraplegia using a neurally controlled robotic arm. Nature 2012, 485, 372–375. [Google Scholar] [CrossRef]
  13. Müller-Putz, G.R.; Scherer, R.; Pfurtscheller, G.; Rupp, R. EEG-based neuroprosthesis control: A step towards clinical practice. Neurosci. Lett. 2005, 382, 169–174. [Google Scholar] [CrossRef] [PubMed]
  14. Murphy, D.P.; Bai, O.; Gorgey, A.S.; Fox, J.; Lovegreen, W.T.; Burkhardt, B.W.; Atri, R.; Marquez, J.S.; Li, Q.; Fei, D.Y. Electroencephalogram-based brain–computer interface and lower-limb prosthesis control: A case study. Front. Neurol. 2017, 8, 696. [Google Scholar] [CrossRef] [PubMed]
  15. Rebsamen, B.; Burdet, E.; Guan, C.; Zhang, H.; Teo, C.L.; Zeng, Q.; Ang, M.; Laugier, C. A brain-controlled wheelchair based on P300 and path guidance. In Proceedings of the First IEEE/RAS-EMBS International Conference on Biomedical Robotics and Biomechatronics, Pisa, Italy, 20–22 February 2006; IEEE: Piscataway Township, NJ, USA, 2006; pp. 1101–1106. [Google Scholar]
  16. Na, R.; Hu, C.; Sun, Y.; Wang, S.; Zhang, S.; Han, M.; Yin, W.; Zhang, J.; Chen, X.; Zheng, D. An embedded lightweight SSVEP-BCI electric wheelchair with hybrid stimulator. Digit. Signal Process. 2021, 116, 103101. [Google Scholar] [CrossRef]
  17. Choi, K. Control of a vehicle with EEG signals in real-time and system evaluation. Eur. J. Appl. Physiol. 2012, 112, 755–766. [Google Scholar] [CrossRef]
  18. Choudhari, V.; Han, C.; Bickel, S.; Mehta, A.D.; Schevon, C.; McKhann, G.M.; Mesgarani, N. Brain-Controlled Augmented Hearing for Spatially Moving Conversations in Multi-Talker Environments. Adv. Sci. 2024, 11, 2401379. [Google Scholar] [CrossRef]
  19. Münßinger, J.I.; Halder, S.; Kleih, S.C.; Furdea, A.; Raco, V.; Hösle, A.; Kübler, A. Brain painting: First evaluation of a new brain–computer interface application with ALS-patients and healthy volunteers. Front. Neurosci. 2010, 4, 182. [Google Scholar] [CrossRef]
  20. Zhu, H.Y.; Hieu, N.Q.; Hoang, D.T.; Nguyen, D.N.; Lin, C.T. A human-centric metaverse enabled by brain–computer interface: A survey. IEEE Commun. Surv. Tutor. 2024, 26, 2120–2145. [Google Scholar] [CrossRef]
  21. Baseler, H.; Sutter, E.; Klein, S.; Carney, T. The topography of visual evoked response properties across the visual field. Electroencephalogr. Clin. Neurophysiol. 1994, 90, 65–81. [Google Scholar] [CrossRef]
  22. Li, J.; Pu, J.; Cui, H.; Xie, X.; Xu, S.; Li, T.; Hu, Y. An online P300 brain–computer interface based on tactile selective attention of somatosensory electrical stimulation. J. Med Biol. Eng. 2019, 39, 732–738. [Google Scholar] [CrossRef]
  23. Huang, R.S.; Jung, T.P.; Makeig, S. Event-related brain dynamics in continuous sustained-attention tasks. In Proceedings of the International Conference on Foundations of Augmented Cognition, Beijing, China, 22–27 July 2007; Springer: Berlin/Heidelberg, Germany, 2007; pp. 65–74. [Google Scholar]
  24. Pfurtscheller, G.; Da Silva, F.L. Event-related EEG/MEG synchronization and desynchronization: Basic principles. Clin. Neurophysiol. 1999, 110, 1842–1857. [Google Scholar] [CrossRef]
  25. Pfurtscheller, G.; Neuper, C. Motor imagery and direct brain–computer communication. Proc. IEEE 2001, 89, 1123–1134. [Google Scholar] [CrossRef]
  26. Tang, Z.; Sun, S.; Zhang, S.; Chen, Y.; Li, C.; Chen, S. A brain-machine interface based on ERD/ERS for an upper-limb exoskeleton control. Sensors 2016, 16, 2050. [Google Scholar] [CrossRef] [PubMed]
  27. Dekleva, B.M.; Chowdhury, R.H.; Batista, A.P.; Chase, S.M.; Yu, B.M.; Boninger, M.L.; Collinger, J.L. Motor cortex retains and reorients neural dynamics during motor imagery. Nat. Hum. Behav. 2024, 8, 729–742. [Google Scholar] [CrossRef] [PubMed]
  28. Neuper, C.; Wörtz, M.; Pfurtscheller, G. ERD/ERS patterns reflecting sensorimotor activation and deactivation. Prog. Brain Res. 2006, 159, 211–222. [Google Scholar]
  29. Gruenwald, J.; Znobishchev, A.; Kapeller, C.; Kamada, K.; Scharinger, J.; Guger, C. Time-variant linear discriminant analysis improves hand gesture and finger movement decoding for invasive brain–computer interfaces. Front. Neurosci. 2019, 13, 901. [Google Scholar] [CrossRef]
  30. Schober, T.; Wenzel, K.; Feichtinger, M.; Schwingenschuh, P.; Strebel, A.; Krausz, G.; Pfurtscheller, G. Restless legs syndrome: Changes of induced electroencephalographic beta oscillations—An ERD/ERS study. Sleep 2004, 27, 147–150. [Google Scholar] [CrossRef]
  31. Shuqfa, Z.; Belkacem, A.N.; Lakas, A. Decoding multi-class motor imagery and motor execution tasks using Riemannian geometry algorithms on large EEG datasets. Sensors 2023, 23, 5051. [Google Scholar] [CrossRef]
  32. Mohamed, A.F.; Jusas, V. Developing Innovative Feature Extraction Techniques from the Emotion Recognition Field on Motor Imagery Using Brain–Computer Interface EEG Signals. Appl. Sci. 2024, 14, 11323. [Google Scholar] [CrossRef]
  33. Molla, M.K.I.; Ahamed, S.; Almassri, A.M.; Wagatsuma, H. Classification of Motor Imagery Using Trial Extension in Spatial Domain with Rhythmic Components of EEG. Mathematics 2023, 11, 3801. [Google Scholar] [CrossRef]
  34. Sadiq, M.T.; Yu, X.; Yuan, Z.; Fan, Z.; Rehman, A.U.; Li, G.; Xiao, G. Motor imagery EEG signals classification based on mode amplitude and frequency components using empirical wavelet transform. IEEE Access 2019, 7, 127678–127692. [Google Scholar] [CrossRef]
  35. Siviero, I.; Brusini, L.; Menegaz, G.; Storti, S.F. Motor-imagery EEG signal decoding using multichannel-empirical wavelet transform for brain computer interfaces. In Proceedings of the 2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), Ioannina, Greece, 27–30 September 2022; IEEE: Piscataway Township, NJ, USA, 2022; pp. 1–4. [Google Scholar]
  36. Jolliffe, I.T. Principal Component Analysis for Special Types of Data; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  37. Xu, N.; Gao, X.; Hong, B.; Miao, X.; Gao, S.; Yang, F. BCI competition 2003-data set IIb: Enhancing P300 wave detection using ICA-based subspace projections for BCI applications. IEEE Trans. Biomed. Eng. 2004, 51, 1067–1072. [Google Scholar] [CrossRef] [PubMed]
  38. Alomari, M.H.; Samaha, A.; AlKamha, K. Automated classification of L/R hand movement EEG signals using advanced feature extraction and machine learning. arXiv 2013, arXiv:1312.2877. [Google Scholar]
  39. Yu, X.; Chum, P.; Sim, K.B. Analysis the effect of PCA for feature reduction in non-stationary EEG based motor imagery of BCI system. Optik 2014, 125, 1498–1502. [Google Scholar] [CrossRef]
  40. Kim, H.; Yoshimura, N.; Koike, Y. Characteristics of kinematic parameters in decoding intended reaching movements using electroencephalography (EEG). Front. Neurosci. 2019, 13, 1148. [Google Scholar] [CrossRef]
  41. Mohamed, A.K.; Marwala, T.; John, L. Single-trial EEG discrimination between wrist and finger movement imagery and execution in a sensorimotor BCI. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; IEEE: Piscataway Township, NJ, USA, 2011; pp. 6289–6293. [Google Scholar]
  42. Wan, B.; Zhou, Z.; Ming, D.; Qi, H. Detection of ERD/ERS signals based on ICA and PSD. J. Tianjin Univ. 2008, 41, 1383–1390. [Google Scholar]
  43. Zou, Y.; Zhao, X.; Chu, Y.; Xu, W.; Han, J.; Li, W. A supervised independent component analysis algorithm for motion imagery-based brain computer interface. Biomed. Signal Process. Control 2022, 75, 103576. [Google Scholar] [CrossRef]
  44. Lotte, F.; Bougrain, L.; Cichocki, A.; Clerc, M.; Congedo, M.; Rakotomamonjy, A.; Yger, F. A review of classification algorithms for EEG-based brain–computer interfaces: A 10 year update. J. Neural Eng. 2018, 15, 031005. [Google Scholar] [CrossRef]
  45. Kevric, J.; Subasi, A. Comparison of signal decomposition methods in classification of EEG signals for motor-imagery BCI system. Biomed. Signal Process. Control 2017, 31, 398–406. [Google Scholar] [CrossRef]
  46. Li, Y.; Wen, P.P. Clustering technique-based least square support vector machine for EEG signal classification. Comput. Methods Programs Biomed. 2011, 104, 358–372. [Google Scholar]
  47. Wu, W.; Gao, X.; Hong, B.; Gao, S. Classifying single-trial EEG during motor imagery by iterative spatio-spectral patterns learning (ISSPL). IEEE Trans. Biomed. Eng. 2008, 55, 1733–1743. [Google Scholar] [CrossRef]
  48. Wang, H.; Zhang, Y. Detection of motor imagery EEG signals employing Naïve Bayes based learning process. Measurement 2016, 86, 148–158. [Google Scholar]
  49. Siuly, S.; Li, Y. Improving the separability of motor imagery EEG signals using a cross correlation-based least square support vector machine for brain–computer interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2012, 20, 526–538. [Google Scholar] [CrossRef] [PubMed]
  50. Sadiq, M.T.; Yu, X.; Yuan, Z.; Zeming, F.; Rehman, A.U.; Ullah, I.; Li, G.; Xiao, G. Motor imagery EEG signals decoding by multivariate empirical wavelet transform-based framework for robust brain–computer interfaces. IEEE Access 2019, 7, 171431–171451. [Google Scholar] [CrossRef]
  51. Shan, W.; Wang, Y.; He, Q.; Xie, P. EEG Recognition of Motor Imagery Based on EWT in Driving Assistance. In Proceedings of the 2017 7th International Conference on Mechatronics, Computer and Education Informationization (MCEI 2017), Shenyang, China, 16–18 June 2017; Atlantis Press: Dordrecht, The Netherlands, 2017; pp. 11–15. [Google Scholar]
  52. Lee, H.K.; Choi, Y.S. Application of continuous wavelet transform and convolutional neural network in decoding motor imagery brain–computer interface. Entropy 2019, 21, 1199. [Google Scholar] [CrossRef]
  53. Karbasi, M.; Jamei, M.; Malik, A.; Kisi, O.; Yaseen, Z.M. Multi-steps drought forecasting in arid and humid climate environments: Development of integrative machine learning model. Agric. Water Manag. 2023, 281, 108210. [Google Scholar] [CrossRef]
  54. Yücelbaş, C.; Yücelbaş, Ş.; Özşen, S.; Tezel, G.; Küççüktürk, S.; Yosunkaya, Ş. Automatic detection of sleep spindles with the use of STFT, EMD and DWT methods. Neural Comput. Appl. 2018, 29, 17–33. [Google Scholar] [CrossRef]
  55. Gilles, J. Empirical wavelet transform. IEEE Trans. Signal Process. 2013, 61, 3999–4010. [Google Scholar] [CrossRef]
  56. Elouaham, S.; Dliou, A.; Jenkal, W.; Louzazni, M.; Zougagh, H.; Dlimi, S. Empirical wavelet transform based ecg signal filtering method. J. Electr. Comput. Eng. 2024, 2024, 9050909. [Google Scholar] [CrossRef]
  57. Nayak, A.B.; Shah, A.; Maheshwari, S.; Anand, V.; Chakraborty, S.; Kumar, T.S. An empirical wavelet transform-based approach for motion artifact removal in electroencephalogram signals. Decis. Anal. J. 2024, 10, 100420. [Google Scholar] [CrossRef]
  58. Anuragi, A.; Sisodia, D.S. Empirical wavelet transform based automated alcoholism detecting using EEG signal features. Biomed. Signal Process. Control 2020, 57, 101777. [Google Scholar] [CrossRef]
  59. Al-Nashash, H.A.; Zalzala, A.M.; Thakor, N.V. A neural networks approach to EEG signals modeling. In Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (IEEE Cat. No. 03CH37439), Cancun, Mexico, 17–21 September 2003; IEEE: Piscataway Township, NJ, USA, 2003; Volume 3, pp. 2451–2454. [Google Scholar]
  60. Tangwiriyasakul, C.; Mocioiu, V.; van Putten, M.J.; Rutten, W.L. Classification of motor imagery performance in acute stroke. J. Neural Eng. 2014, 11, 036001. [Google Scholar] [CrossRef] [PubMed]
  61. Chang, T.; Liao, J.; Wu, Y.; Quan, Y.; Chen, S.; Zhao, W.; Li, Y.; Yu, P.; Fang, Y.; Zong, Y.; et al. Improving Performance of Electroencephalography-Based Malignant Brain Tumors Screening from Neural Modulation of Motor Execution. SSRN. p. 5044979. Available online: https://ssrn.com/abstract=5044979 (accessed on 21 May 2025).
  62. Ahmed, S.S.; Khan, M.; Bukhari, S.M.; Khan, R.A. Robust Feature Engineering Techniques for Designing Motor Imagery-Based Bci Systems. SSRN. p. 5081953. Available online: https://ssrn.com/abstract=5081953 (accessed on 21 May 2025).
  63. Chen, C.; Chen, P.; Belkacem, A.N.; Lu, L.; Xu, R.; Tan, W.; Li, P.; Gao, Q.; Shin, D.; Wang, C.; et al. Neural activities classification of left and right finger gestures during motor execution and motor imagery. Brain-Comput. Interfaces 2021, 8, 117–127. [Google Scholar] [CrossRef]
  64. Rakhmatulin, I.; Dao, M.S.; Nassibi, A.; Mandic, D. Exploring convolutional neural network architectures for EEG feature extraction. Sensors 2024, 24, 877. [Google Scholar] [CrossRef]
  65. Lun, X.; Yu, Z.; Chen, T.; Wang, F.; Hou, Y. A simplified CNN classification method for MI-EEG via the electrode pairs signals. Front. Hum. Neurosci. 2020, 14, 338. [Google Scholar] [CrossRef]
  66. Shanmugam, S.; Dharmar, S. A CNN-LSTM hybrid network for automatic seizure detection in EEG signals. Neural Comput. Appl. 2023, 35, 20605–20617. [Google Scholar] [CrossRef]
  67. de Benito-Gorron, D.; Lozano-Diez, A.; Toledano, D.T.; Gonzalez-Rodriguez, J. Exploring convolutional, recurrent, and hybrid deep neural networks for speech and music detection in a large audio dataset. EURASIP J. Audio Speech Music Process. 2019, 2019, 9. [Google Scholar] [CrossRef]
  68. Rajagukguk, R.A.; Ramadhan, R.A.; Lee, H.J. A review on deep learning models for forecasting time series data of solar irradiance and photovoltaic power. Energies 2020, 13, 6623. [Google Scholar] [CrossRef]
  69. Haghighi, E.B.; Palm, G.; Rahmati, M.; Yazdanpanah, M.J. A new class of multi-stable neural networks: Stability analysis and learning process. Neural Netw. 2015, 65, 53–64. [Google Scholar] [CrossRef]
  70. Simard, P.Y.; Steinkraus, D.; Platt, J.C. Best practices for convolutional neural networks applied to visual document analysis. In Proceedings of the Icdar. Edinburgh, Edinburgh, UK, 6 August 2003; Volume 3. [Google Scholar]
  71. Smit, D.J.; Boomsma, D.I.; Schnack, H.G.; Pol, H.E.H.; de Geus, E.J. Individual differences in EEG spectral power reflect genetic variance in gray and white matter volumes. Twin Res. Hum. Genet. 2012, 15, 384–392. [Google Scholar] [CrossRef] [PubMed]
  72. Chen, A.C.; Feng, W.; Zhao, H.; Yin, Y.; Wang, P. EEG default mode network in the human brain: Spectral regional field powers. Neuroimage 2008, 41, 561–574. [Google Scholar] [CrossRef] [PubMed]
  73. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. {TensorFlow}: A system for {Large-Scale} machine learning. In Proceedings of the 12th USENIX symposium on operating systems design and implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  74. Venugopal, A.; Resende Faria, D. Boosting EEG and ECG Classification with Synthetic Biophysical Data Generated via Generative Adversarial Networks. Appl. Sci. 2024, 14, 10818. [Google Scholar] [CrossRef]
  75. Dere, M.D.; Jo, J.H.; Lee, B. Event-driven edge deep learning decoder for real-time gesture classification and neuro-inspired rehabilitation device control. IEEE Trans. Instrum. Meas. 2023, 72, 4011612. [Google Scholar] [CrossRef]
  76. Chen, Y.; Akutagawa, M.; Emoto, T.; Kinouchi, Y. The removal of EMG in EEG by neural networks. Physiol. Meas. 2010, 31, 1567. [Google Scholar] [CrossRef]
  77. Chen, Y.; Akutagawa, M.; Katayama, M.; Zhang, Q.; Kinouchi, Y. Neural network based EEG denoising. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Virtual Event, 20–25 August 2008; IEEE: Piscataway Township, NJ, USA, 2008; pp. 262–265. [Google Scholar]
  78. McFarland, D.J.; Miner, L.A.; Vaughan, T.M.; Wolpaw, J.R. Mu and beta rhythm topographies during motor imagery and actual movements. Brain Topogr. 2000, 12, 177–186. [Google Scholar] [CrossRef]
  79. LEE, B. EMG-EEG Dataset for Upper-Limb Gesture Classification; IEEE: Piscataway Township, NJ, USA, 2023. [Google Scholar] [CrossRef]
  80. Center, M.H. Butterworth Filter Design. 2004. Available online: https://www.mathworks.com/help/signal/ref/butter.html (accessed on 14 May 2025).
  81. MATLAB Central File Exchange. Empirical Wavelets. 2025. Available online: https://ww2.mathworks.cn/help/wavelet/ug/empirical-wavelet-transform.html (accessed on 23 April 2025).
  82. Huang, N.E. Hilbert-Huang Transform and Its Applications; World Scientific: Singapore, 2014; Volume 16. [Google Scholar]
  83. Biju, K.; Hakkim, H.A.; Jibukumar, M. Ictal EEG classification based on amplitude and frequency contours of IMFs. Biocybern. Biomed. Eng. 2017, 37, 172–183. [Google Scholar]
  84. Al Ghayab, H.R.; Li, Y.; Siuly, S.; Abdulla, S. A feature extraction technique based on tunable Q-factor wavelet transform for brain signal classification. J. Neurosci. Methods 2019, 312, 43–52. [Google Scholar] [CrossRef]
  85. Kutlu, Y.; Kuntalp, D. Feature extraction for ECG heartbeats using higher order statistics of WPD coefficients. Comput. Methods Programs Biomed. 2012, 105, 257–267. [Google Scholar] [CrossRef]
  86. Barrowclough, J.; Nnamoko, N.; Korkontzelos, I. Personalised Affective Classification Through Enhanced EEG Signal Analysis. Appl. Artif. Intell. 2025, 39, 2450568. [Google Scholar] [CrossRef]
  87. Sharma, S.; Mehra, R. Implications of pooling strategies in convolutional neural networks: A deep insight. Found. Comput. Decis. Sci. 2019, 44, 303–330. [Google Scholar] [CrossRef]
  88. Yang, J.; Huang, X.; Wu, H.; Yang, X. EEG-based emotion classification based on bidirectional long short-term memory network. Procedia Comput. Sci. 2020, 174, 491–504. [Google Scholar] [CrossRef]
  89. Hayaty, M.; Muthmainah, S.; Ghufran, S.M. Random and synthetic over-sampling approach to resolve data imbalance in classification. Int. J. Artif. Intell. Res. 2020, 4, 86–94. [Google Scholar] [CrossRef]
  90. Berrar, D. Cross-validation. In Reference Module in Life Sciences Encyclopedia of Bioinformatics and Computational Biology; Elsevier: Amsterdam, The Netherlands, 2019. [Google Scholar]
  91. Nti, I.K.; Nyarko-Boateng, O.; Aning, J. Performance of machine learning algorithms with different K values in K-fold cross-validation. Int. J. Inf. Technol. Comput. Sci. 2021, 13, 61–71. [Google Scholar]
  92. Cho, J.H.; Jeong, J.H.; Lee, S.W. NeuroGrasp: Real-time EEG classification of high-level motor imagery tasks using a dual-stage deep learning framework. IEEE Trans. Cybern. 2021, 52, 13279–13292. [Google Scholar] [CrossRef]
  93. Kim, S.; Shin, D.Y.; Kim, T.; Lee, S.; Hyun, J.K.; Park, S.M. Enhanced recognition of amputated wrist and hand movements by deep learning method using multimodal fusion of electromyography and electroencephalography. Sensors 2022, 22, 680. [Google Scholar] [CrossRef]
  94. Tryon, J.; Trejos, A.L. Evaluating convolutional neural networks as a method of EEG–EMG fusion. Front. Neurorobot. 2021, 15, 692183. [Google Scholar] [CrossRef]
  95. Du, Y.; Xu, Y.; Wang, X.; Liu, L.; Ma, P. EEG temporal–spatial transformer for person identification. Sci. Rep. 2022, 12, 14378. [Google Scholar] [CrossRef]
Figure 1. The individual preprocessing steps for both variants (first on the left, second on the right). Blocks with a dashed border indicate the feature extraction stage.
Figure 1. The individual preprocessing steps for both variants (first on the left, second on the right). Blocks with a dashed border indicate the feature extraction stage.
Sensors 25 03284 g001
Table 1. Assessment of classification.
Table 1. Assessment of classification.
ClassifierVariantAccuracy [%]F1 Score [%]
HNN1st with feature extraction
(bandpower) for all channels
74.5622.77
LDA50.1222.04
SVM67.8225.33
HNN1st without feature extraction for all channels74.7714.08
HNN2nd with feature extraction
(IF + IA) for all channels
81.9136.89
LDA73.0818.35
SVM75.7517.78
HNN2nd without feature extraction for all channels76.3616.80
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zych, P.; Filipek, K.; Mrozek-Czajkowska, A.; Kuwałek, P. Classification of Electroencephalography Motor Execution Signals Using a Hybrid Neural Network Based on Instantaneous Frequency and Amplitude Obtained via Empirical Wavelet Transform. Sensors 2025, 25, 3284. https://doi.org/10.3390/s25113284

AMA Style

Zych P, Filipek K, Mrozek-Czajkowska A, Kuwałek P. Classification of Electroencephalography Motor Execution Signals Using a Hybrid Neural Network Based on Instantaneous Frequency and Amplitude Obtained via Empirical Wavelet Transform. Sensors. 2025; 25(11):3284. https://doi.org/10.3390/s25113284

Chicago/Turabian Style

Zych, Patryk, Kacper Filipek, Agata Mrozek-Czajkowska, and Piotr Kuwałek. 2025. "Classification of Electroencephalography Motor Execution Signals Using a Hybrid Neural Network Based on Instantaneous Frequency and Amplitude Obtained via Empirical Wavelet Transform" Sensors 25, no. 11: 3284. https://doi.org/10.3390/s25113284

APA Style

Zych, P., Filipek, K., Mrozek-Czajkowska, A., & Kuwałek, P. (2025). Classification of Electroencephalography Motor Execution Signals Using a Hybrid Neural Network Based on Instantaneous Frequency and Amplitude Obtained via Empirical Wavelet Transform. Sensors, 25(11), 3284. https://doi.org/10.3390/s25113284

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop