Next Article in Journal
Petroleum Sector-Driven Roadmap for Future Hydrogen Economy
Previous Article in Journal
Progress and Summary of Photodarkening in Rare Earth Doped Fiber
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating the Motor Imagery Classification Performance of a Double-Layered Feature Selection on Two Different-Sized Datasets

by
Minh Tran Duc Nguyen
1,2,*,
Nhi Yen Phan Xuan
1,2,
Bao Minh Pham
1,2,
Trung-Hau Nguyen
1,2,
Quang-Linh Huynh
1,2 and
Quoc Khai Le
1,2,*
1
Department of Biomedical Engineering, Faculty of Applied Science, Ho Chi Minh City University of Technology (HCMUT), 268 Ly Thuong Kiet, District 10, Ho Chi Minh City 700000, Vietnam
2
Vietnam National University Ho Chi Minh City, Linh Trung Ward, Thu Duc District, Ho Chi Minh City 700000, Vietnam
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2021, 11(21), 10388; https://doi.org/10.3390/app112110388
Submission received: 30 September 2021 / Revised: 29 October 2021 / Accepted: 1 November 2021 / Published: 5 November 2021
(This article belongs to the Section Biomedical Engineering)

Abstract

:
Numerous investigations have been conducted to enhance the motor imagery-based brain–computer interface (BCI) classification performance on various aspects. However, there are limited studies comparing their proposed feature selection framework performance on both objective and subjective datasets. Therefore, this study aims to provide a novel framework that combines spatial filters at various frequency bands with double-layered feature selection and evaluates it on published and self-acquired datasets. Electroencephalography (EEG) data are preprocessed and decomposed into multiple frequency sub-bands, whose features are then extracted, calculated, and ranked based on Fisher’s ratio and minimum-redundancy-maximum-relevance (mRmR) algorithm. Informative filter banks are chosen for optimal classification by linear discriminative analysis (LDA). The results of the study, firstly, show that the proposed method is comparable to other conventional methods through accuracy and F1-score. The study also found that hand vs. feet classification is more discriminable than left vs. right hand (4–10% difference). Lastly, the performance of the filter banks common spatial pattern (FBCSP, without feature selection) algorithm is found to be significantly lower (p = 0.0029, p = 0.0015, and p = 0.0008) compared to that of the proposed method when applied to small-sized data.

1. Introduction

To help patients with functional motor disabilities, such as amyotrophic lateral sclerosis (ALS) or spinal cord injury (SCI), Brain-Computer Interface (BCI) has been researched for over decades. BCI is a system that establishes an artificial communication pathway between the central nervous system and external devices, acting as a peripheral motor, when a biological neuron path is weak, or no longer works properly [1]. Various applications applying BCI to real-world problems have been carried out: brain-actuated speller where a patient could type words on a computer screen with their thought [2,3]; controlling computer cursor by gazing [4]; controlling incoming calls [5]; computer control interface [6]. Owing to the emerging advances in novel signal processing and computational algorithms, BCI has an important role in facilitating the recovery of neural functions, restoring functional corticospinal, and corticomuscular connections and neurorehabilitation.
The non-invasive electroencephalography (EEG) technique has been widely used in BCI research because of its clinical simplicity, cost efficiency, and safety [7]. Event-related desynchronization/synchronization (ERD/ERS), in which the amplitude of mu and beta rhythms is suppressed or increased depending on the event [8,9,10], is one of the main types of non-invasive BCI. Specifically, motor imagery (MI) elicits mu and beta ERD/ERS patterns in both healthy subjects as well as patients with severe motor disabilities [11,12]. MI has the advantage over other methods as it only requires mental simulation of movements without actual actions or external stimuli.
In recent years, many studies put their focus on investigating and reinforcing feature extraction and classification algorithms for MI-based BCI, thus enhancing BCI performance. Traditional feature extractors have been widely used and claimed to be effective up to now, however, they still have drawbacks. For example, the temporal resolution and spectral resolution of the short-time Fourier transform (STFT) algorithm are determined by the width of the time window [13]. That is to say, there is a trade-off between temporal and spectral resolution: the better the frequency resolution as the window gets wider, the worse the time resolution and vice versa. Another example is wavelet transform, which can perform multi-scale time-frequency analysis. The main disadvantage of this algorithm is that it requires manual screening for statistical features such as maximum, minimum, mean, and standard deviation of wavelet coefficients, whereas EEG pattern, motor imagery to be specific, is not stationary and it varies among subjects [14]. Meanwhile, common spatial pattern (CSP) is one of the promising candidates for the shortcoming of EEG low spatial resolution characteristic [15]. Moreover, the CSP algorithm, one of the most effective filters optimizing spatial information in motor tasks discrimination [16,17], has not been fully investigated.
Since CSP needs to identify discriminative frequency bands for effective classification, there is a substantial amount of studies on finding optimal frequency bands. The first approach is a combination of CSP and time-frequency analysis [18,19]. The simultaneous optimization of the spatial spectrum filter is used in the second type [20,21]. The third type of approach involves filtering the original EEG signals into multiple frequency sub-bands, then CSP is applied to extract features from these sub-bands, and classification is performed based on the optimum frequency bands [22,23,24,25]. The intelligent optimization method is the fourth kind of technique utilized to determine the optimum frequency bands. This approach is capable of selecting informative frequency bands of any length and is independent of subjective experience [26,27,28]. In the first method, because frequency information is obtained via a time-frequency analysis, the EEG signals of each channel must be decomposed, which involves numerous calculations and takes considerable computation time. The second method is challenging to implement and effortlessly leads to local solutions [29]. The drawback of the fourth method is the requirement of a lengthy training phase of models, which is insufficient for practical BCI. Although sub-band filtering in the third method requires high computation cost, it seems to have surpassed the other methods in terms of performance and robustness.
With regard to feature selection, there are three main present methods: filter, wrapper, and embedded [30]. Filter method initiates with all of the features and these features are chosen based on criteria such as information gain, consistency, dependency, correlation, and distance measures [31,32]. The wrapper method evaluates feature selection by comparing it to the classifier’s performance [33,34,35]. However, when assessing each applicant feature subset, the wrapper method must train and test the classifier, and that is prone to overfitting and computationally costly. The embedded technique incorporates feature selection into the classifier’s training phase, performing both tasks concurrently [36,37]. Although proven to be efficient in most studies, the embedded technique has the highest computational cost in all three feature selection methods. In contrast, Fisher was used in a unified framework for feature extraction, selection, and classification [38]. Fisher’s ratio was also used in [39] as a channel selector along with the correlation coefficient. In [40], the author performed manual screening for optimal frequency bands and suggested building a more sophisticated algorithm for auto selection. Yet his following study [26] implemented temporal filter parameter optimization combined with optimization algorithms (particle swarm optimization, artificial bee colony, and genetic algorithm) and the drawback was long training time as the framework was so sophisticated. Hence, Fisher’s ratio is the right tool for a simpler, lower computational cost and more effective method. Furthermore, in [41], the author also suggested using mutual information-based feature selection (i.e., mRmR) as it is well-known for selecting the most significant features. Consequently, filter methods, Fisher’s ratio, and mRmR, are utilized in this study due to their low computational cost.
With regard to classifiers, the three most popular algorithms because of their robustness are linear discriminant analysis (LDA), support vector machine (SVM), and Bayesian [42,43,44]. Deep learning has emerged and is combined with traditional algorithms to create reliable and powerful techniques in MI tasks recognition problems in many studies [14,45,46,47,48]. However, deep learning not only poses a high computational cost but also extracts incomprehensible features [49]. Furthermore, deep learning has no clear advantages over conventional machine learning methods [50]. Hence, in this study, LDA is utilized for classification as it has been successfully implemented in various applications [51,52,53].
Lastly, most of the previous studies evaluated only on BCI competition datasets with the absence of comparing with their own data. Li et al. [38] established their own datasets of 16 subjects to evaluate the study’s proposed framework, which used Fisher’s objective function for feature extraction and feature selection. The study’s dataset size was small as it only contained 80 trials per class per subject and was not compared to other published datasets, such as BCI competition. Peterson et al. [18] also established a dataset of 7 subjects, 80 trials per class per subject, to evaluate his proposed combined CSP and time-frequency analysis framework. The study’s proposed framework was also evaluated on other BCI competition datasets. Consequently, carrying out both subjective and objective datasets would be more reliable.
To this end, the aim of this paper is to: (i) propose a novel framework for motor imagery tasks based on spatial filters, automatic feature selection, and traditional classifier; (ii) evaluate the proposed methods on both a self-acquired dataset and a BCI Competition dataset in terms of accuracy, F1-score, and ROC. The proposed framework (called Fisher—mRmR discriminative filter bank common spatial pattern, DFBCSP—FmRmR) would use filter bank for sub-band filtering of EEG signal, common spatial pattern as feature extractor, Fisher’s ratio and mutual information-based mRmR for automatic feature selection, and LDA as the classifier.
The content of this paper is organized as follows:
  • Data acquisition, signal processing, conventional feature extraction, and feature selection algorithms as well as the proposed method are introduced in Section 2.
  • The data visualizations and results are shown in Section 3.
  • Discussion and further explanation are in Section 4.
  • The conclusion is provided in Section 5.

2. Materials and Methods

2.1. Data Acquisition of Self-Acquired Dataset (Dataset 1)

Subjects were 16 healthy, both undergraduate and graduate students (6 women, 10 men; age range 20–23 years, mean 21 ± 0.89). Surveys were given to subjects and reported that no subjects had disorders relating to neurology. None of them had any prior experience with BCI. Two to three preparatory runs were provided to all subjects before data acquisition and were paid USD 2 per measurement for taking part in. Before and after each session, subjects were asked to fill out some information relating to their conditions to ensure the consistency of awareness during the experiment. Throughout the whole procedure, subjects had to comply with some regulations: (i) no caffeine or any stimulants consumption a few hours before experiments; (ii) restrict negative feelings; (iii) avoid sleep deprivation. During the experiment, participants were asked to sit on a chair in a Faraday-shielded room that blocks electromagnetic radiation and observe a computer screen. Minimizing eye movements, body adjustments, and avoiding any undesired movements during a cue onset. Subjects were allowed to give signs to pause or stop the experiment should any unexpected conditions show up (cough, scratch, sleepy, tired, etc.).
Data were recorded by NicoletOne V32 (Natus Medical Inc., Pleasanton, CA, USA) with monopolar montage with electrodes C3, Cz, C4, P3, Pz, and P4 according to the International 10–20 electrode placement system [54] (Figure 1). A reference and a ground electrode were attached to the left and right earlobes accordingly. All electrodes’ impedance was kept below 5 kΩ and the sampling frequency was 500 Hz. Nuprep was used to clean the scalp to reduce skin impedance, and conductive gel Ten20 was used for electrode attachment (Figure 1). Six electrode positions were chosen based on the fact that the most noticeable mu and beta ERD/ERS can be recognized when performing imagery tasks [55,56].
There were five measurements that occurred on five different days for each participant. There were five 8 min sessions in each measurement and a 10 min break was provided between each session. In each session, subjects were asked to pay attention to a MATLAB simulation program for 30 segments. Each segment lasted for 16 s (Figure 2): the first 6 s, the fixation cross appeared on the screen denoted relaxation state (no physical movements or mental activities); the next 6 s were for kinesthetic motor imagery tasks accompanied with cue denoted by a red arrow pointing left, right, and downward (for left hand, right hand and feet respectively); the last 4 s were for cue offset and subjects returned to relaxation. A cue onset and offset were indicated by a beep sound. In total, each subject has around 220–250 trials for each class. However, some were unable to complete the full experiment due to an inappropriate schedule. Subjects “ID03”, “ID06”, “ID10”, “ID12” have exceptionally fewer trials, approximately 180, 195, 140, and 134 trials for each class, respectively.
Most MI-BCI research conducted worldwide follows the same measuring procedure: when the subject is at resting state and observing the screen, the arrow appears in a random order after a cue, and the subject performs the imagining; lastly, the subject rests for random durations. The time required for arrows to appear, the duration required to complete the task, and the resting period are the variations between the measured procedures in these studies. The procedure for measuring in [57] began with the subject observing a cross in the center of the screen, and then the subject observed an arrow from the 2nd to the 6th second. This was followed by a two-second blank screen and a random rest period between 0.5 and 1.5 s. The measuring method described in [56] begins with a cue at the 2nd second and the arrow at the 3rd to 5th seconds, followed by a blank screen for the next second and a random interval of 0–1 s. The research by [58] used a similar paradigm as [56], except that the arrow only showed up for 1.25 s instead of 2 s. Study [59] had an inter-trial transitional period between 0.5 and 2 s. The interval between time points is not significant and is altered to suit the data analysis procedure. As a result, the research will follow the procedure outlined above, with minor modifications to the timeline to accommodate the measuring object, the data analysis, and processing.

2.2. BCI Competition Datasset (Dataset 2)

Dataset IIa of BCI competition IV (2008) [60]. This dataset contains 22 channels at a 250 Hz sampling rate. Nine participants (A01–A09) performed left hand, right hand, foot, and tongue motor imagery. This dataset has two sessions recorded on two different days, each with six runs of 48 trials (12 trials per class), yielding 288 trials per session (72 trials per class). The first session was used as the training set and the second as the test set. In this study, only the first session of each subject is used to train and evaluate for the number of trials as close to that of one measurement of the self-acquired dataset as possible, hence it would be easier for comparison between datasets. Furthermore, only three tasks (left hand, right hand, and feet) are performed in this study. Consequently, each subject in this dataset has 216 trials in total, before any artifacts removal, equally divided into three classes. Please visit the following website for further information: http://www.bbci.de/competition/IV/ (accessed on 10 June 2021).
There are a few key differences between the author’s self-acquired and BCI Competition datasets:
  • Firstly, the data acquisition paradigms are slightly different in terms of timing.
  • Secondly, the number of measurements of the self-acquired dataset are more and occurred on different days while there is only one measurement for each subject in BCI Competition. That results in a different number of trials for each subject and possibly affects the outcomes.

2.3. Signal Pre-Processing

For the self-acquired dataset: Firstly, artifact removal is done to raw EEG data by eliminating any trial if its amplitude exceeds 100 µV. A 50 Hz notch filter is then applied to remove electrical interference. Thirdly, a 6th order Butterworth filter is applied to the EEG data for 4–40 Hz band-pass filtering with a time window of 6 s, starting from the beginning of the MI cue until the end.
For the BCI Competition IV dataset IIa: Firstly, trials affected by artifacts were marked previously by experts and removed subsequently in this study. Next, a 50 Hz notch filter is then applied to remove electrical interference. Lastly, clean trials with a time window of 3 s, from the beginning of the MI cue until the end, traverse through a 6th order Butterworth filter for 4–40 Hz band-pass filtering.

2.4. Common Spatial Pattern

The common spatial pattern (CSP) [21,61] method is increasingly used for feature extraction and classification in two-class MI-BCI, by linearly mapping the feature into a more discriminative space. It maximizes the variance of the filtered data in one class while reducing the variance of the filtered data in the other. As a function, the resulting feature vectors enhance discriminating between the two classes by lowering intra-class variance and increasing class difference. The CSP attempts to convert the multi-channel EEG data into a low-dimensional area via linear transformation by using a projection matrix.
Z = W E ,
where E is an N × T matrix containing the raw EEG measurement data from a single trial, N is the number of channels, T is the number of samples, and W is the CSP projection matrix. This is accomplished by designing the rows of W or spatial filters so that the variances of the first and last rows of Z provide the greatest discrimination between two classes of MI tasks. As a result, the feature vector Fp is generated by (2), where Zp is the first and last m rows of Z, p ∈ {1 … 2m}.
F p = log [ ( v a r ( Z p ) ) / ( i = 1 2 m v a r ( Z i ) ) ]

2.5. Fisher’s Ratio

The DFBCSP method uses a series of filters called parent FB in the sequel [23] to ex-tract subject-specific discriminative frequency bands. The band selection block leverages the parent FB filters EEG and the Fisher’s ratio of the filtered EEG to identify subject-specific discriminative frequency bands. The parent filter bank analyzes signals and for the purposes of acquiring subject-specific DFB, the equation below is utilized to compute an approximation of spectral power associated with each sub-band:
P ( f i , t ) = 1 T n = 1 T x t , f ( n ) 2 ,  
for each t trial, P ( f i , t ) is the spectral power estimated in the ith band output, and the sample count in the filtered EEG signal x f ( n ) is denoted by T. Nf is the total number of bands, and Nt is the total number of trials, thus an Nf × Nt matrix matching to spectral power is produced. Across all frequency bands, each trial is paired with an estimated p-value calculated.
For the purpose of selecting the most useful filter banks, the Fisher’s ratio, or FR, is computed from all filter outputs from the parent filter bank. The following equation is used to get the Fisher’s ratio at each band output:
F R ( f ) = S B S W ,  
where SW and SB are the within-class variance and between-class variance, respectively. SW and SB are computers in Equations (5) and (6):
S W = k = 1 C t = 1 n k ( P t m k ) 2 ,  
S B = t = 1 C n k ( m m k ) 2 ,
where m is the average power across classes, m k is the average power within class k ,   ( k = 1 ,   2 ) , C denotes the number of classes, and nk denotes the number of trials for class k . A filter bank with a higher FR index permits for better classification and after selecting out the subject-specific frequency bands, these discriminative bands are used to filter in the range of all data channels.

2.6. Minimum-Redundancy-Maximum-Relevance (mRmR)

Minimum-redundancy-maximum-relevance (mRmR) [62,63] function selection works by means of selecting features that have an excessive correlation with the class and low correlation with different features. After CSP, mRmR is adopted in automatic feature selection by substituting the correlation between variables with the statistical dependence, which is defined through mutual information. It is possible to assess non-linear relationships between two random variables by measuring their mutual information with each other. Furthermore, it reveals how much information can be gleaned from a random variable by monitoring another random variable.
I ( x ; y ) = i , j log p ( x i , y j ) p ( x i ) p ( y j )
The variables x and y are feature vectors in Equation (7), p ( x , y ) is the joint probability function, and p ( x ) , p ( y ) are the marginal probability functions of x and y, respectively. Assume that S is a predefined set of features and h is a variable belonging to a specific class. The redundancy of S is quantified as follows:
W I = 1 | S | 2 i , j S I ( i , j )
The relevance of S is determined by the following criteria:
V I = 1 | S | i S I ( h , j )
Two methods are used to evaluate S:
M I D :   V I W I
M I Q :   V I / W I
In the vast majority of cases, MIQ is proven to be superior to MID. Due to the fact that we cannot test all possible subsets of features S for a given dataset, the mRMR algorithm employs a forward search approach in its implementation.

2.7. Proposed Framework

After the preprocessing block: notch filter is for electrical grid interference elimination, low-pass and high-pass filters are for cut-off frequency from 4–40 Hz; peak removal for each trial has voltage exceeding 100 µV, the preprocessed data is then split into 3 pairs, namely: Left hand vs. right hand (pair 1), Right hand vs. feet (pair 2) and feet vs. left hand (pair 3), for binary classification. Train and test data are created based on 10-fold cross-validation (Figure 3).
In the training phase, a filter bank is applied to each EEG trial with a frequency range of 4–40 Hz, a bandwidth of 4 Hz, and 2 Hz overlapping; thus creating 17 filter banks in total. The proposed frequency configuration is based on study [18,23,64]. Next, the Fisher’s ratio and mRmR algorithms are applied to data in each filter bank to select the most discriminative filter banks. By calculating the power of data samples, Fisher’s ratio of each channel can be calculated through the criterion “maximize between-class variance and minimize within-class variance” and sorted in descending order. In this study, the most discriminative filter bank in each channel that has the highest Fisher’s ratio is selected for further selection. The maximum filter banks we can have in this case are 6 as dataset 1 has 6 channels. For dataset 2, the maximum filter banks would be 17 filter banks as it has a number of channels exceeding the number of filter banks.
Meanwhile, prior to using the mRmR feature selection algorithm, features of data of each filter bank have to be extracted by CSP. mRmR works on the two independent variables, which are “features” and “labels” in this study. mRmR finds the relationship, or how much information, the two independent variables have in common. As a result, the features in dataset 1, which are 6 channels multiplied by 17 filter banks, are sorted in descending order based on the amount of mutual information the algorithm found. For dataset 2, the number of features would be 22 channels multiplied by 17 filter banks. The top 6 filter banks are selected for further selection, regardless of which channel they are generated from. The maximum number of filter banks we can have in this case is 6 to match with Fisher’s case.
The selection criteria of this study are as follows: we take the intersection of the two cases, which means the common filter banks are selected as the final selection. If the intersection is not possible, the filter banks selected in the mRmR case will be chosen as it is claimed to be more effective than Fisher’s [40,65,66]. In the testing phase, the test data goes through the filter bank which was previously chosen in the training phase and the features are extracted with the corresponding WCSP. In the last step, LDA is utilized to perform binary classification.
To visualize how informative features are extracted and selected by the proposed method, ERD/ERS responses, discriminative frequency bands, and features dimensions of a single measurement of a random subject of the self-acquired dataset are plotted in Figure 4. Figure 4A–C shows the responses in the time-frequency map of a random subject when performing left hand, right hand, and feet imagery respectively. Only C3, Cz, and C4 were plotted to represent the sides of the cortex. The imagination of limb movement causes ERD in mu and beta waves on the contralateral hemisphere [8,9,10,11,12]. The more dominant mu (10–14 Hz) and beta (22–26 Hz) suppressions of left-hand imagery can be seen on channel C4, while those of right-hand imagery are on channel C3. As feet imagery involves imagination of moving both feet and its somatosensory area is close to the mid-central of the cortex, the feet imagery causes mu suppression in C3, Cz, and C4. These features (frequency bands and channels) are informative for discriminating the tasks from one another. The proposed method successfully selected the most informative features (Figure 4D) for left hand vs. right hand. By simultaneously performing Fisher’s ratio and mRmR algorithms, the proposed method selected features by choosing the common bands between the two algorithms. Comparison between pre-selection of features (Figure 4E) and post-selection of features (Figure 4F) is performed by mapping the feature vectors on a 3D plane using t-distributed stochastic neighbor embedding (t-SNE) [67]. t-SNE is a non-linear dimensionality reduction algorithm that visualizes high dimensional data in two or three dimensions but preserves relative distances between objects. The features in Figure 4F are more discriminable than those in Figure 4E after the bands were selected by the proposed method.

2.8. Evaluation

The proposed framework is compared with three other conventional algorithms:
  • FBCSP (without feature selection): this method has 17 filter banks, ranging from 4–40 Hz, a 4 Hz bandwidth, 2 Hz overlapping, and the number of spatial filters m is three.
  • DFBCSP—Fisher: This method has 17 filter banks, ranging from 4–40 Hz, a 4 Hz bandwidth, 2 Hz overlapping and the number of spatial filters m is three Fisher’s ratio is used for band selection.
  • DFBCSP—mRmR: This method has 17 filter banks, ranging from 4–40 Hz, a 4 Hz bandwidth, 2 Hz overlapping, and the number of spatial filters m is three. mRmR is used for band selection, which could have a maximum of six filter banks.
  • Proposed method (DFBCSP—FmRmR): t\This method has 17 filter banks, ranging from 4–40 Hz, a 4 Hz bandwidth, 2 Hz overlapping, and the number of spatial filters m is three. Both Fisher’s ratio and mRmR are used for band selection, which can be carried out without manual screening. The selected bands are varied after each loop and can be up to six filter banks.
The results are compared in terms of accuracy and F1-score. While accuracy is a metric indicating the proportion of properly recognized instances, F1-score is primarily concerned with the disagreement between the true positive rate and the positive predictive value. Consequently, this study employs both metrics for a more objective evaluation. The models are evaluated on both the self-acquired dataset and the public BCI competition dataset. The results are visualized through bar charts, confusion matrixes, tables, and ROC curves.

3. Results

3.1. ROC and AUC

To examine how well LDA adapts to the self-acquired dataset and evaluate its performance, we compare three binary classifiers (pair 1—“Left hand vs. Right hand”, pair 2—“Right hand vs. Feet”, and pair 3 “Feet vs. Left hand”) of four methods. Receiver operator character (ROC) curves and area under the curve (AUC) of three LDA models combined with four feature extraction are shown in Figure 5. It is found that the ROC curves of all three classifiers are close to the top left corner and the AUC values are over 0.93. Noticeably, the proposed method follows a similar pattern compared to other methods, and even surpasses the FBCSP (0.99 compared to 0.98 in pair 1), although it is not significant. Additionally, the “Left hand vs. Right hand” pair performance is surprisingly higher than the other two (0.99 and 0.98 compared to 0.93 and 0.94). Besides ROC curves and AUC values, the confusion matrix is also an indication of the suitability of the models.
To further examine the performance of the four frameworks, heat maps are generated based on a confusion matrix (Figure 6). Actual labels are compared with predicted labels generated by 4 methods with a heat bar indicating the number of trials. We find that the performance of the proposed method surpasses conventional FBCSP in the “Left hand vs. Right hand” pair. Furthermore, the 1st pair classifications (left column) of the four methods are better than those of the 2nd and 3rd pair (middle and right column), around 20–30 trials. The accuracy of each pair for each method is computed by dividing the number of correctly classified samples (the darker blue diagonal) by the total within one confusion matrix. The result above is well supported by the ROC curve and AUC in Figure 5. Taken together, we can conclude that these methods are highly effective for the subject. The accuracy and F1-score of both datasets are compiled in the following subsections for a more transparent verification.
The averaged ROC curves and AUC values are depicted in Figure 7. There are no significant discrepancies between curves as well as the AUC values between methods. However, pair 2 and pair 3 have higher AUC values than pair 1.

3.2. Accuracy and F1-Score of BCI Competition’s Dataset

To evaluate the classification capacity of conventional and proposed frameworks on a reliable published dataset, both parameters for evaluation are composed in Table 1, for pair 1, pair 2, and pair 3, accordingly. The variation between F1-score and accuracy in Table 1 is not considerable, mostly under 4%, except for the pair 1 discrimination by DFBCSP—Fisher in subject 4, which is nearly 7%. Furthermore, the mean values of the two metrics are not different from each other, indicating the stability of the method as well as the data. A noteworthy point here is that DFBCSP frameworks achieve higher performance than the FBCSP framework in nine subjects. The mean values show that DFBCSP is 7–15% better than FBCSP in terms of accuracy and F1-score. This result is consistent with previous studies [40,65]. Subjects “01”, “02”, “03”, “07”, and “08” have good performance based on DFBCSP frameworks in general. Moreover, the proposed DFBCSP outweighs other methods in subjects “03” and “06 for pair 1, subjects “03”, “05”, “06”, “07”, and “08” for pair 2, and subject “04”, “07”, “08”, and “09” for pair 3 (bolded values). It can be concluded that DFBCSP—FmRmR is comparable to other methods.
Another key point is that the classification of pair 2 and pair 3 is better than that of pair 1 in most subjects, and the mean values also reassure that. In the FBCSP approach, approximately 60% from pair 1 compared to 65% from pair 2 and 64% from pair 3. In DFBCSP (3 approaches) around 68% from pair 1 compared to 78% from both pair 2 and pair 3. Only subject “08” shows an opposite trend. It can be concluded that MI patterns between hands and feet are easier to be recognized than those between the left and right hand.
However, the variations in performance between subjects are considerable as they elicit high standard deviation, which indicates the methods are not suitable for every participant. Further discussion will be given to the matter in the next section.

3.3. Accuracy and F1-Score of Self-Acquired Dataset

To both evaluate the performance of the proposed framework and validity of the self-acquired dataset, accuracy, and F1-score are compiled in Table 2, Table 3 and Table 4 for pair 1, pair 2, and pair 3 accordingly. Similar to Table 1, the variation between F1-score and accuracy in Table 2, Table 3 and Table 4 is not considerable, mostly under 4%. Moreover, the mean values of the two metrics are not different from each other, indicating the stability of the method as well as the data. It can be observed that there are subjects (“ID01”, “ID02”, “ID06”, “ID08”, “ID11”, “ID13” and “ID16”) achieving high classification rates, except for pair 1 of subjects “ID06” and “ID08”, approximately 85 to over 90%. The DFBCSP—FmRmR has an equal efficiency to its variants. Although it does outperform other methods, we can conclude that the proposed framework is effective in both datasets. Furthermore, the aforementioned subjects in this dataset have the potential to use a practical BCI system, while others are affected by the “BCI illiteracy” phenomenon.
Likewise, subjects in the self-acquired dataset, except for subjects “ID 01”, “ID 04”, “ID 07”, and “ID 10”, elicit a similar pattern of having discrimination for pair 2 and pair 3 better than pair 1. For the four methods, the difference in the mean values is from 2–5%, and it is not as significant as in dataset IIa. Surprisingly, the mean and SD values of this dataset (Table 2, Table 3 and Table 4) are almost equivalent to those of dataset IIa (Table 1), which ranges from 70–80% (mean) ± 10–15% (SD). The results above further verify the validity and similarity between the two datasets as well as the consistency of the proposed framework.
However, it is of great importance to notice that FBCSP has an equal efficiency to the other three models in most subjects, unlike in dataset 2. This outcome is inconsistent with studies [40,65], in which only BCI Competition was used. We hypothesize that this inconsistency might be due to the differences between the two datasets. As mentioned in the previous section, the number of trials in the BCI competition dataset is only 72 per class while that of the self-acquired dataset is over 200 trials per class for each subject. Therefore, we decided to analyze models in each measurement for each subject for a more balanced comparison (72 trials compared to 50 trials). That is also why we chose only one session in the BCI competition dataset (Figure 8).
To test our hypothesis of FBCSP performance affected by the amount of data, the evaluation on one subject is broken down to each measurement (Figure 8). Three pairs and four models are also evaluated and the maximum number of measurements for each subject is five. The last column is the data for all measurements. The most obvious recognition is that FBCSP in each measurement (or each day) is outweighed by other algorithms, approximately 20–30%. The DFBCSP frameworks still perform well compared to each other. However, all algorithms, including FBCSP, have almost equal efficiency when all measurements of one subject are computed. This can be inferred that FBCSP works better when the number of data is large enough, while DFBCSP algorithms only require a lower amount to be sufficient. Consequently, BCI utilizing DFBCSP would require less data and time for both training and testing. This also implies that DFBCSP algorithms are more suitable for a robust and practical BCI system.
To further reassure our hypothesis, the difference in the accuracy of “all measurement” and “single measurement” of every subject is shown in Figure 9. “All” is defined as data of all recordings taken into account, and “single” is defined as data of each day taken into account and then accuracy is averaged across each measurement. A positive value indicates the accuracy of “all measurement” is better than that of “single measurement”, and vice versa. Most subjects have an accuracy difference between “all measurement” and “single measurement” to be positive. This indicates that a larger data size improves classification performance. The differences in all three DFBCSP frameworks are mainly below 8%, which is not remarkable as there were variations in performance among measurements of each subject. However, there is a great fall of the FBCSP framework in three pairs in most subjects when the evaluation is performed in each measurement (smaller data size). The difference ranges from 10–30%, which means the performance of FBCSP in “single measurement” is significantly lower than in “all measurement” (p = 0.0029, p = 0.0015 and p = 0.0008 for pair 1, pair 2 and pair 3 accordingly). The mean values of the difference of three DFBCSP frameworks are no more than 3% for three pairs, while that of FBCSP is 12.30%, 14.22%, and 15.37% for pair 1, pair 2, and pair 3, respectively (Table 5). Taken together, FBCSP performance is negatively affected by limited data size while that of the proposed DFBCSP is not. There are no differences between pairs.

4. Discussion

Intelligent optimization methods [26] along with deep learning networks [14,45,46,47,48] have widely been investigated in recent years. Although their performances are slightly superior to traditional machine learning, the burdens in terms of computational complexity and massive data requirement are far from trading off. Secondly, traditional algorithms such as CSP have been proved in countless studies to be effective for practical BCI, yet its potential needs further exploration. Finally, most studies evaluated their proposed framework on the BCI competition dataset but in the absence of actual data of local population, which is one of the criteria both to evaluate the validity of the proposed method and examine the suitability of various populations. Consequently, the novel approach of this study make contributions as follows:
  • Propose a novel and reliable framework for motor imagery tasks based on spatial filters, automatic feature selection, and traditional classifier
  • Evaluate the proposed methods on both self-acquired dataset and a BCI Competition dataset in terms of accuracy, F1-score, and ROC
From the results, there are numerous key points that have to be taken into account. Firstly, the validity of the self-acquired dataset has been proven through similarities with BCI Competition dataset IIa. The pattern of better classification of hands vs. feet over that of left vs. right hand is seen in both datasets, Table 1, Table 2, Table 3 and Table 4, and also consistent with previous studies [68]. Moreover, DFBCSP methods outweigh the FBCSP method in both datasets (single measurement cases for self-acquired data) and it has been clearly proven in several studies [18,39,40,65,69,70]. This can be explained because FBCSP (without feature selection) generates an enormous amount of features, i.e., 22 × 17 features in BCI competition datasets IIa, and obviously also contains redundant information that would degrade the performance. Molla et al. [70] also stated that the mean accuracy of a framework without feature selection is far worse than that of a framework integrated with feature selection as it does not use irrelevant features. That is why several algorithms have been built to select informative features to mainly improve the performance and reduce computational costs. The third similarity between the two datasets is in the mean values and their standard deviation. Both datasets imply that the BCI performance varied among users (moderate mean value and high standard deviation) and not every subject is suitable for a certain framework, or decoder, or even type of BCI. The phenomenon is known as “BCI illiteracy” which will be explained later.
Secondly, the robustness and reliability of the proposed method are clearly demonstrated by the results. ROC curve, accuracy, and F1-score of the proposed method are consistent throughout the study and they are comparative to other methods. Although the framework might not be better than other conventional ones, there are some cases that have proven its superiority. Likewise, other methods are superior to one another depending on subjects and pairs. It can be inferred that the subjects from both datasets suffer from “BCI illiteracy”. BCI illiteracy is caused by a variety of factors that may be classified into two categories [71]: physiological and non-physiological causes. Regarding physiological reasons, variations in the brain’s anatomy, such as the dispersion of neurons in the brain folds, may make recording brain signals through the electrodes problematic. Additionally, transitory variables should be considered. These could include changes in attention level, exhaustion, or impediment, as well as social aspects such as emotional reactions, interactions, and social cues encountered during BCI training, which should be taken into consideration. Non-physiological reasons include a lack of proficiency in operating the system or inadequate training procedures, which lead the participant to misunderstand the instructions and execute the experiment erroneously [71]. That is why it leads to high standard deviation values in both datasets. Consequently, the proposed method could only handle a few cases in “BCI illiteracy” and further research is required to tackle the phenomenon.
Thirdly, the classification of feet vs. hand imagery is more discriminative than left vs. right-hand imagery as shown in both datasets, although it is more noticeable in BCI competition dataset IIa (4–10% difference) than in self-acquired dataset (2–5% difference). This might be due to more discriminable EEG patterns generated by feet MI than those induced by the left or right hand. Pfurtscheller et al. [68] explained that feet MI would elicit both mu and/or beta ERD in the mid-central cortex and mu ERS on two hemispheres, specifically in the hand presentation area [72]. The so-called “focal ERD/surround ERS” phenomenon was found in [8]. Hence, the authors in the aforementioned studies reported a more significant variation in 10–12 Hz band power when comparing different limbs (hand vs. feet) than homologous ones.
Finally, the FBCSP (without feature selection) could only work with data that is large enough, while DFBCSP could be effectively functional with small-sized data. It can clearly be seen that FBCSP efficiency is comparative to other methods (Table 2, Table 3 and Table 4) when it is applied to the self-acquired dataset, whose number of trials ranges from 220–250 per class, but downgrades to 10–30% if applied to a BCI competition IIa (72 trials per class) or single measurement of the self-acquired dataset (50 trials per class) (Figure 9). To the best of the author’s knowledge, the bottom line of this outcome lies in the number of features and physiological differences between days of data recording. In a single measurement, FBCSP extracts all possible features, including redundant ones, while DFBCSP selectively chooses them; thus, it has a better performance than FBCSP. However, due to daily physiological changes of human nature, the EEG patterns when performing MI tasks may have changed as well leading to changes in required features for task discrimination. When including all data of each subject into one evaluation, redundant features may become informative ones, resulting in better FBCSP performance. Further research is needed to confirm this hypothesis. Yet, the FBCSP is obviously unable to have an optimal computation cost as DFBCSP. Hence, DFBCSP is more suitable as a practical BCI system in terms of performance and computation cost.

5. Conclusions

Brain–computer interface has attracted a lot of interest in the hope of improving human life. Motor imagery-based BCI is expected to have potential in the medical field, especially in neurorehabilitation. However, this type of system has been encountered several challenges in terms of the limited number of commands, insufficient performance and training time, robust online analysis, and so forth. Various algorithms and approaches have been proposed to deal with the challenges, and CSP is one of the candidates for a practical BCI. Hence, numerous researchers put their effort in harnessing the full potential of CSP and its variants through methods such as the combination of CSP and time-frequency analysis, simultaneous optimization of the spatial spectrum filter, filtering the EEG signals into several frequency bands, and intelligent optimization. In addition, filter, wrapper, and embedded methods have also been established to further improve the performance. Yet, there are still limitations requiring further investigations to shed light on. To this end, the authors hope to make contributions to this crusade. The study successfully proposed a novel DFBCSP framework in combination with Fisher’s ratio and the mRmR algorithm. The framework was evaluated on both datasets and produced noteworthy findings mentioned in the previous section. The proposed method could be applied in any classification problems, mainly in electrophysiological signal processing but not limited to other researches requiring pattern recognition, biometrics identification, and machine learning-related problems. Furthermore, the proposed method could be utilized in small-sized data, which is essential in real-world applications. Lastly, the reassurance of the discrimination between hand vs. feet better than left vs. right hand provides more proof about responses on motor cortex when performing motor imagery in the field of neuroscience.
Nevertheless, there are some drawbacks to the study. Firstly, the performance of the proposed method is yet satisfactory for practical BCI, especially in a medical context. One way to address this issue is utilizing SVM and tuning its parameters to obtain optimal results. In addition, trial selection could be performed in the pre-processing step to eliminate undesired trials. Secondly, other metrics such as information transfer rate (ITR) and computational time, which are critical for practical BCI, are not considered in the paper. Thirdly, the study is carried out only in offline analysis. The authors hope to conduct further investigations into these matters. Segmenting the time window into 1 s makes it suitable to establish a model to perform online analysis. Hence the authors hope to establish a robust decoder for real-time motor imagery training with feedbacks. In future investigations, comparison with other feature selection methods or combination with optimization methods [26,27,29] while minimizing computation cost should be carried out.

Author Contributions

Conceptualization, M.T.D.N.; methodology, M.T.D.N.; validation, M.T.D.N., N.Y.P.X. and B.M.P.; formal analysis, M.T.D.N.; investigation, M.T.D.N.; resources, M.T.D.N.; data curation, M.T.D.N. and Q.K.L.; writing—original draft preparation, M.T.D.N. and B.M.P.; writing—review and editing, M.T.D.N. and B.M.P.; visualization, N.Y.P.X.; supervision, T.-H.N. and Q.-L.H.; project administration, T.-H.N., Q.K.L. and Q.-L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

We acknowledge the support of time and facilities from Ho Chi Minh City University of Technology (HCMUT), VNU-HCM for this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wolpaw, J.R.; Wolpaw, E.W. Brain–Computer Interfaces: Something New under the Sun. In Brain–Computer Interfaces: Principles and Practice; Oxford University Press (OUP): Oxford, UK, 2012; pp. 3–12. [Google Scholar]
  2. Donchin, E.; Spencer, K.; Wijesinghe, R. The mental prosthesis: Assessing the speed of a P300-based brain-computer interface. IEEE Trans. Rehabil. Eng. 2000, 8, 174–179. [Google Scholar] [CrossRef] [Green Version]
  3. Li, Y.; Nam, C.S.; Shadden, B.B.; Johnson, S.L. A P300-based brain–computer interface: Effects of interface type and screen size. Int. J. Hum. Comput. Interact. 2010, 27, 52–68. [Google Scholar] [CrossRef]
  4. Middendorf, M.; McMillan, G.; Calhoun, G.; Jones, K. Brain-computer interfaces based on the steady-state visual-evoked response. IEEE Trans. Rehabil. Eng. 2000, 8, 211–214. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Katona, J.; Peter, D.; Ujbanyi, T.; Kovari, A. Control of incoming calls by a windows phone based brain computer interface. In Proceedings of the 2014 IEEE 15th International Symposium on Computational Intelligence and Informatics (CINTI), Budapest, Hungary, 19–21 November 2014; pp. 121–125. [Google Scholar]
  6. Katona, J.; Kovari, A. EEG-based computer control interface for brain-machine interaction. Int. J. Online Biomed. Eng. 2015, 11, 43–48. [Google Scholar] [CrossRef] [Green Version]
  7. Saker, M.; Rihana, S. Platform for EEG signal processing for motor imagery—Application Brain Computer Interface. In Proceedings of the 2013 2nd International Conference on Advances in Biomedical Engineering, Tripoli, Lebanon, 11–13 September 2013; pp. 30–33. [Google Scholar]
  8. Pfurtscheller, G.; da Silva, F.L. Event-related EEG/MEG synchronization and desynchronization: Basic principles. Clin. Neurophysiol. 1999, 110, 1842–1857. [Google Scholar] [CrossRef]
  9. Neuper, C.; Müller-Putz, G.R.; Scherer, R.; Pfurtscheller, G. Motor imagery and EEG-based control of spelling devices and neuroprostheses. Prog. Brain Res. 2006, 159, 393–409. [Google Scholar] [PubMed]
  10. Neuper, C.; Scherer, R.; Wriessnegger, S.; Pfurtscheller, G. Motor imagery and action observation: Modulation of sensorimotor brain rhythms during mental control of a brain–computer interface. Clin. Neurophysiol. 2009, 120, 239–247. [Google Scholar] [CrossRef]
  11. Pfurtscheller, G.; Neuper, C.; Flotzinger, D.; Pregenzer, M. EEG-based discrimination between imagination of right and left hand movement. Electroencephalogr. Clin. Neurophysiol. 1997, 103, 642–651. [Google Scholar] [CrossRef]
  12. Neuper, C.; Pfurtscheller, G. 134 ERD/ERS based brain computer interface (BCI): Effects of motor imagery on sensorimotor rhythms. Int. J. Psychophysiol. 1998, 30, 53–54. [Google Scholar] [CrossRef]
  13. Lee, H.K.; Choi, Y.-S. Application of Continuous Wavelet Transform and Convolutional Neural Network in Decoding Motor Imagery Brain-Computer Interface. Entropy 2019, 21, 1199. [Google Scholar] [CrossRef] [Green Version]
  14. Xu, B.; Zhang, L.; Song, A.; Wu, C.; Li, W.; Zhang, D.; Xu, G.; Li, H.; Zeng, H. Wavelet Transform Time-Frequency Image and Convolutional Network-Based Motor Imagery EEG Classification. IEEE Access 2018, 7, 6084–6093. [Google Scholar] [CrossRef]
  15. Srinivasan, R. Methods to improve the spatial resolution of EEG. Int. J. Bioelectromagn. 1999, 1, 102–111. [Google Scholar]
  16. Ramoser, H.; Muller-Gerking, J.; Pfurtscheller, G. Optimal spatial filtering of single trial EEG during imagined hand movement. IEEE Trans. Rehabil. Eng. 2000, 8, 441–446. [Google Scholar] [CrossRef] [Green Version]
  17. Arvaneh, M.; Guan, C.; Ang, K.K.; Quek, C. Optimizing Spatial Filters by Minimizing Within-Class Dissimilarities in Electroencephalogram-Based Brain–Computer Interface. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 610–619. [Google Scholar] [CrossRef]
  18. Peterson, V.; Wyser, D.; Lambercy, O.; Spies, R.; Gassert, R. A penalized time-frequency band feature selection and classification procedure for improved motor intention decoding in multichannel EEG. J. Neural Eng. 2019, 16, 016019. [Google Scholar] [CrossRef] [Green Version]
  19. Lin, J.; Liu, S.; Huang, G.; Zhang, Z.; Huang, K. The Recognition of Driving Action Based on EEG Signals Using Wavelet-CSP Algorithm. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018. [Google Scholar]
  20. Aghaei, A.S.; Mahanta, M.S.; Plataniotis, K. Separable Common Spatio-Spectral Patterns for Motor Imagery BCI Systems. IEEE Trans. Biomed. Eng. 2015, 63, 15–29. [Google Scholar] [CrossRef]
  21. Yuksel, A.; Olmez, T. Filter Bank Common Spatio-Spectral Patterns for Motor Imagery Classification. In Proceedings of the Lecture Notes in Computer Science, Porto, Portugal, 5–8 September 2016; Springer Science and Business Media: Berlin/Heidelberg, Germany, 2016; pp. 69–84. [Google Scholar]
  22. Ang, K.K.; Chin, Z.Y.; Zhang, H.; Guan, C. Filter Bank Common Spatial Pattern (FBCSP) in Brain-Computer Interface. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Institute of Electrical and Electronics Engineers (IEEE), Hong Kong, China, 1–8 June 2008. [Google Scholar]
  23. Thomas, K.P.; Guan, C.; Lau, C.T.; Vinod, A.P.; Ang, K.K. A New Discriminative Common Spatial Pattern Method for Motor Imagery Brain–Computer Interfaces. IEEE Trans. Biomed. Eng. 2009, 56, 2730–2733. [Google Scholar] [CrossRef] [PubMed]
  24. Park, G.-H.; Lee, Y.-R.; Kim, H.-N. Improved Filter Selection Method for Filter Bank Common Spatial Pattern for EEG-Based BCI Systems. Int. J. Electron. Electr. Eng. 2014, 2, 101–105. [Google Scholar] [CrossRef]
  25. Wu, F.; Gong, A.; Li, H.; Zhao, L.; Zhang, W.; Fu, Y. A New Subject-Specific Discriminative and Multi-Scale Filter Bank Tangent Space Mapping Method for Recognition of Multiclass Motor Imagery. Front. Hum. Neurosci. 2021, 15, 104. [Google Scholar] [CrossRef] [PubMed]
  26. Kumar, S.; Sharma, A. A new parameter tuning approach for enhanced motor imagery EEG signal classification. Med. Biol. Eng. Comput. 2018, 56, 1861–1874. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Jiao, Y.; Zhou, T.; Yao, L.; Zhou, G.; Wang, X.; Zhang, Y. Multi-View Multi-scale optimization of feature representation for EEG classification improvement. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 2589–2597. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Zhou, T.; Wu, W.; Xie, H.; Zhu, H.; Zhou, G.; Cichocki, A. Improving EEG decoding via clustering-based multitask feature learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–11. [Google Scholar] [CrossRef]
  29. Zhang, S.; Zhu, Z.; Zhang, B.; Feng, B.; Yu, T.; Li, Z. The CSP-based new features plus non-convex log sparse feature selection for motor imagery EEG classification. Sensors 2020, 20, 4749. [Google Scholar] [CrossRef] [PubMed]
  30. Singh, A.; Hussain, A.; Lal, S.; Guesgen, H. A comprehensive review on critical issues and possible solutions of motor imagery based electroencephalography brain-computer interface. Sensors 2021, 21, 2173. [Google Scholar] [CrossRef] [PubMed]
  31. Venkatesh, B.; Anuradha, J. A Review of Feature Selection and Its Methods. Cybern. Inf. Technol. 2019, 19, 3–26. [Google Scholar] [CrossRef] [Green Version]
  32. Ang, K.K.; Chin, Z.Y.; Zhang, H.; Guan, C. Mutual information-based selection of optimal spatial–temporal patterns for single-trial EEG-based BCIs. Pattern Recognit. 2012, 45, 2137–2144. [Google Scholar] [CrossRef]
  33. Rakshit, P.; Bhattacharyya, S.; Konar, A.; Khasnobish, A.; Tibarewala, D.N.; Janarthanan, R. Artificial Bee Colony Based Feature Selection for Motor Imagery EEG Data. In Proceedings of the Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2013; pp. 127–138. [Google Scholar]
  34. Baig, M.Z.; Aslam, N.; Shum, H.P.; Zhang, L. Differential evolution algorithm as a tool for optimal feature subset selection in motor imagery EEG. Expert Syst. Appl. 2017, 90, 184–195. [Google Scholar] [CrossRef]
  35. Atyabi, A.; Shic, F.; Naples, A. Mixture of autoregressive modeling orders and its implication on single trial EEG classification. Expert Syst. Appl. 2016, 65, 164–180. [Google Scholar] [CrossRef] [Green Version]
  36. Sreeja, S.; Samanta, D.; Mitra, P.; Sarma, M. Motor Imagery EEG Signal Processing and Classification using Machine Learning Approach. Jordanian J. Comput. Inf. Technol. 2018, 4, 80. [Google Scholar] [CrossRef]
  37. Wang, J.; Xue, F.; Li, H. Simultaneous channel and feature selection of fused EEG features based on sparse group lasso. BioMed Res. Int. 2015, 2015, 703768. [Google Scholar] [CrossRef] [Green Version]
  38. Li, X.; Guan, C.; Zhang, H.; Ang, K.K. A Unified fisher’s ratio learning method for spatial filter optimization. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2727–2737. [Google Scholar] [CrossRef]
  39. Park, Y.; Chung, W. Selective Feature generation method based on time domain parameters and correlation coefficients for filter-bank-CSP BCI systems. Sensors 2019, 19, 3769. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Kumar, S.; Sharma, A.; Tsunoda, T. An improved discriminative filter bank selection approach for motor imagery EEG signal classification using mutual information. BMC Bioinform. 2017, 18, 125–137. [Google Scholar] [CrossRef] [Green Version]
  41. Kumar, S.; Mamun, K.; Sharma, A. CSP-TSM: Optimizing the performance of Riemannian tangent space mapping using common spatial pattern for MI-BCI. Comput. Biol. Med. 2017, 91, 231–242. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Bhattacharyya, S.; Khasnobish, A.; Chatterjee, S.; Konar, A.; Tibarewala, D. Performance analysis of LDA, QDA and KNN algorithms in left-right limb movement classification from EEG data. In Proceedings of the 2010 International Conference on Systems in Medicine and Biology, Kharagpur, India, 16–18 December 2010. [Google Scholar]
  43. Singla, S.; Garsha, S.N.; Chatterjee, S. Characterization of classifier performance on left and right limb motor imagery using support vector machine classification of EEG signal for left and right limb movement. In Proceedings of the 2016 5th International Conference on Wireless Networks and Embedded Systems (WECON), Rajpura, India, 14–16 October 2016. [Google Scholar]
  44. He, L.; Hu, D.; Wan, M.; Wen, Y.; Von Deneen, K.M.; Zhou, M. Common Bayesian Network for Classification of EEG-Based Multiclass Motor Imagery BCI. IEEE Trans. Syst. Man Cybern. Syst. 2015, 46, 843–854. [Google Scholar] [CrossRef]
  45. Wu, H.; Niu, Y.; Li, F.; Li, Y.; Fu, B.; Shi, G.; Dong, M. A parallel multiscale filter bank convolutional neural networks for motor imagery EEG classification. Front. Neurosci. 2019, 13, 1275. [Google Scholar] [CrossRef] [Green Version]
  46. Chaudhary, S.; Taran, S.; Bajaj, V.; Sengur, A. Convolutional neural network based approach towards motor imagery tasks EEG signals classification. IEEE Sens. J. 2019, 19, 4494–4500. [Google Scholar] [CrossRef]
  47. Wang, Z.; Cao, L.; Zhang, Z.; Gong, X.; Sun, Y.; Wang, H. Short time Fourier transformation and deep neural networks for motor imagery brain computer interface recognition. Concurr. Comput. Pract. Exp. 2018, 30, e4413. [Google Scholar] [CrossRef]
  48. Zhu, X.; Li, P.; Li, C.; Yao, D.; Zhang, R.; Xu, P. Separated channel convolutional neural network to realize the training free motor imagery BCI systems. Biomed. Signal Process. Control. 2019, 49, 396–403. [Google Scholar] [CrossRef]
  49. Padfield, N.; Zabalza, J.; Zhao, H.; Masero, V.; Ren, J. EEG-Based brain-computer interfaces using motor-imagery: Techniques and challenges. Sensors 2019, 19, 1423. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Lotte, F.; Bougrain, L.; Cichocki, A.; Clerc, M.; Congedo, M.; Rakotomamonjy, A.; Yger, F. A review of classification algorithms for EEG-based brain–computer interfaces: A 10 year update. J. Neural Eng. 2018, 15, 031005. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Sharma, A.; Paliwal, K.K. Rotational Linear Discriminant Analysis Technique for Dimensionality Reduction. IEEE Trans. Knowl. Data Eng. 2008, 20, 1336–1347. [Google Scholar] [CrossRef] [Green Version]
  52. Sharma, A.; Paliwal, K.K. A two-stage linear discriminant analysis for face-recognition. Pattern Recognit. Lett. 2012, 33, 1157–1162. [Google Scholar] [CrossRef]
  53. Sharma, A.; Paliwal, K.K.; Imoto, S.; Miyano, S. A feature selection method using improved regularized linear discriminant analysis. Mach. Vis. Appl. 2014, 25, 775–786. [Google Scholar] [CrossRef] [Green Version]
  54. Klem, G.H.; Lüders, H.O.; Jasper, H.H.; Elger, C. The ten-twenty electrode system of the International Federation. The International Federation of Clinical Neurophysiology. Electroencephalogr. Clin. Neurophysiol. Suppl. 1999, 52, 370–375. [Google Scholar]
  55. Cinar, E.; Sahin, F. New classification techniques for electroencephalogram (EEG) signals and a real-time EEG control of a robot. Neural Comput. Appl. 2011, 22, 29–39. [Google Scholar] [CrossRef]
  56. Kaiser, V.; Kreilinger, A.; Müller-Putz, G.R.; Neuper, C. First steps toward a motor imagery based stroke BCI: New strategy to set up a classifier. Front. Neurosci. 2011, 5, 86. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Allison, B.Z.; Brunner, C.; Kaiser, V.; Müller-Putz, G.R.; Neuper, C.; Pfurtscheller, G. Toward a hybrid brain–computer interface based on imagined movement and visual attention. J. Neural Eng. 2010, 7, 026007. [Google Scholar] [CrossRef] [PubMed]
  58. Horki, P.; Solis-Escalante, T.; Neuper, C.; Müller-Putz, G. Combined motor imagery and SSVEP based BCI control of a 2 DoF artificial upper limb. Med. Biol. Eng. Comput. 2011, 49, 567–577. [Google Scholar] [CrossRef]
  59. Pfurtscheller, G.; Leeb, R.; Slater, M. Cardiac responses induced during thought-based control of a virtual environment. Int. J. Psychophysiol. 2006, 62, 134–140. [Google Scholar] [CrossRef] [PubMed]
  60. Brunner, C.; Leeb, R.; Müller-Putz, G.; Schlögl, A.; Pfurtscheller, G. BCI Competition 2008–Graz Data Set A. Institute for Knowledge Discovery (Laboratory of Brain-Computer Interfaces); Graz University of Technology: Graz, Austria, 2008. [Google Scholar]
  61. Raza, H.; Cecotti, H.; Prasad, G. Optimising frequency band selection with forward-addition and backward-elimination algorithms in EEG-based brain-computer interfaces. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015. [Google Scholar]
  62. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  63. Jo, I.; Lee, S.; Oh, S. Improved Measures of Redundancy and Relevance for mRMR Feature Selection. Computers 2019, 8, 42. [Google Scholar] [CrossRef] [Green Version]
  64. Zhang, Y.; Zhou, G.; Jin, J.; Wang, X.; Cichocki, A. Optimizing spatial patterns with sparse filter bands for motor-imagery based brain–computer interface. J. Neurosci. Methods 2015, 255, 85–91. [Google Scholar] [CrossRef] [PubMed]
  65. Radman, M.; Chaibakhsh, A.; Nariman-Zadeh, N.; He, H. Feature fusion for improving performance of motor imagery brain-computer interface system. Biomed. Signal Process. Control. 2021, 68, 102763. [Google Scholar] [CrossRef]
  66. Gupta, A.; Agrawal, R.K.; Kirar, J.S.; Andreu-Perez, J.; Ding, W.-P.; Lin, C.-T.; Prasad, M. On the utility of power spectral techniques with feature selection techniques for effective mental task classification in noninvasive BCI. IEEE Trans. Syst. Man, Cybern. Syst. 2021, 51, 3080–3092. [Google Scholar] [CrossRef] [Green Version]
  67. Van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  68. Pfurtscheller, G.; Linortner, P.; Winkler, R.; Korisek, G.; Müller-Putz, G. Discrimination of Motor Imagery-Induced EEG Patterns in Patients with Complete Spinal Cord Injury. Comput. Intell. Neurosci. 2009, 2009, 104180. [Google Scholar] [CrossRef] [Green Version]
  69. Luo, J.; Wang, J.; Xu, R.; Xu, K. Class discrepancy-guided sub-band filter-based common spatial pattern for motor imagery classification. J. Neurosci. Methods 2019, 323, 98–107. [Google Scholar] [CrossRef]
  70. Molla, M.K.I.; Al Shiam, A.; Islam, M.R.; Tanaka, T. Discriminative feature selection-based motor imagery classification using EEG Signal. IEEE Access 2020, 8, 98255–98265. [Google Scholar] [CrossRef]
  71. Thompson, M.C. Critiquing the Concept of BCI Illiteracy. Sci. Eng. Ethic 2018, 25, 1217–1233. [Google Scholar] [CrossRef]
  72. Pfurtscheller, G.; Brunner, C.; Schlögl, A.; Da Silva, F.L. Mu rhythm (de)synchronization and EEG single-trial classification of different motor imagery tasks. NeuroImage 2006, 31, 153–159. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (A) Experimental setup with EEG acquired by Natus NicoletOne V32; (B) electrode placement of C3, Cz, C4, P3, Pz, and P4 according to the 10–20 International Electrode system.
Figure 1. (A) Experimental setup with EEG acquired by Natus NicoletOne V32; (B) electrode placement of C3, Cz, C4, P3, Pz, and P4 according to the 10–20 International Electrode system.
Applsci 11 10388 g001
Figure 2. Self-acquired data paradigm. A full trial consists of a 6 s pre-task where the subject was asked to relax, 6 s of cue appearance and performing an indicated motor imagery task, and 6 s of post-task where the subject returned to the relaxing state. The cue appeared and disappeared by two beeps.
Figure 2. Self-acquired data paradigm. A full trial consists of a 6 s pre-task where the subject was asked to relax, 6 s of cue appearance and performing an indicated motor imagery task, and 6 s of post-task where the subject returned to the relaxing state. The cue appeared and disappeared by two beeps.
Applsci 11 10388 g002
Figure 3. The proposed DFBCSP framework.
Figure 3. The proposed DFBCSP framework.
Applsci 11 10388 g003
Figure 4. (AC) ERDS map of left hand, right hand, and feet imagery respectively with channels C3, Cz and C4 from top to bottom. (D) Band selection criteria of the proposed method for discriminating one pair (i.e, Left vs. Right hand). (E) Visualization of features in 3D t-SNE plot before features selection (17 sub-bands). (F) Visualization of features in 3D t-SNE plot after features selection by the proposed method (i.e., 5 sub-bands). Magenta dot is for left hand and blue dot for right hand.
Figure 4. (AC) ERDS map of left hand, right hand, and feet imagery respectively with channels C3, Cz and C4 from top to bottom. (D) Band selection criteria of the proposed method for discriminating one pair (i.e, Left vs. Right hand). (E) Visualization of features in 3D t-SNE plot before features selection (17 sub-bands). (F) Visualization of features in 3D t-SNE plot after features selection by the proposed method (i.e., 5 sub-bands). Magenta dot is for left hand and blue dot for right hand.
Applsci 11 10388 g004
Figure 5. ROC curve and AUC of three binary classifiers for three pairs, “Left hand vs. Right hand”, “Right hand vs. Feet” and “Feet vs. Left hand” respectively left to right. Four feature extraction algorithms are in different colors.
Figure 5. ROC curve and AUC of three binary classifiers for three pairs, “Left hand vs. Right hand”, “Right hand vs. Feet” and “Feet vs. Left hand” respectively left to right. Four feature extraction algorithms are in different colors.
Applsci 11 10388 g005
Figure 6. Performance of four feature extraction algorithms for three pairs in terms of confusion matrix of a random subject. Each row depicts the classification of each method and each column depicts each pair.
Figure 6. Performance of four feature extraction algorithms for three pairs in terms of confusion matrix of a random subject. Each row depicts the classification of each method and each column depicts each pair.
Applsci 11 10388 g006
Figure 7. Averaged ROC curve and AUC of the self-acquired dataset. Four feature extraction algorithms are in different colors.
Figure 7. Averaged ROC curve and AUC of the self-acquired dataset. Four feature extraction algorithms are in different colors.
Applsci 11 10388 g007
Figure 8. Accuracy calculated for each day of a random subject from the author’s dataset. Four methods are in different colors.
Figure 8. Accuracy calculated for each day of a random subject from the author’s dataset. Four methods are in different colors.
Applsci 11 10388 g008
Figure 9. The “all measurement” and “single measurement” accuracy differences are distributed according to the four frameworks. Pair 1, pair 2, and pair 3 are denoted by green, blue, and magenta, respectively.
Figure 9. The “all measurement” and “single measurement” accuracy differences are distributed according to the four frameworks. Pair 1, pair 2, and pair 3 are denoted by green, blue, and magenta, respectively.
Applsci 11 10388 g009
Table 1. Accuracy and F1-score of four methods on BCI Competition’s dataset for three pairs.
Table 1. Accuracy and F1-score of four methods on BCI Competition’s dataset for three pairs.
SubjectLH-RH
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0171.010.776.090.7677.540.7873.910.74
0256.620.5461.760.6254.410.4961.030.57
0372.260.7492.70.9388.320.8990.510.91
0450.590.4551.760.4463.530.6162.350.57
0557.360.5752.710.5161.240.5851.940.5
0659.290.5753.980.550.440.4960.180.59
0762.410.6369.920.7176.690.7768.420.69
0871.970.7388.640.8985.610.8587.880.88
0939.660.3962.930.5770.690.6965.520.61
Mean ± SD60.13 ± 10.850.59 ± 0.1267.83 ± 15.260.66 ± 0.1769.83 ± 13.350.68 ± 0.1569.08 ± 12.90.67 ± 0.14
SubjectRH-F
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0183.210.8497.080.9797.810.9897.080.97
0269.630.6983.70.8481.480.83800.81
0385.070.8694.780.9596.270.9697.010.97
0455.170.5456.320.5957.470.5555.170.57
0544.70.4753.030.5450.760.5163.640.64
0655.660.5861.320.6462.260.6765.090.69
0774.440.7487.970.8887.970.8789.470.89
0858.460.5886.920.8883.850.8490.770.91
0958.870.5872.580.7377.420.7769.350.72
Mean ± SD65.02 ± 13.790.65 ± 0.1477.08 ± 16.770.78 ± 0.1677.25 ± 16.880.78 ± 0.1778.62 ± 15.770.8 ± 0.15
SubjectF-LH
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0175.180.7696.350.9694.890.9594.160.94
0272.180.7183.460.8489.470.8988.720.89
0378.520.7995.560.9694.810.9593.330.93
0457.140.667.860.763.10.6276.190.76
0555.040.5155.810.5651.940.4751.940.48
0657.140.5857.140.5962.860.6554.290.57
0776.120.7793.280.9388.060.8893.280.93
0863.080.6482.310.8282.310.8483.080.84
0948.250.4472.810.772.810.6876.320.73
Mean ± SD64.74 ± 110.64 ± 0.1278.29 ± 15.750.78 ± 0.1577.81 ± 15.720.77 ± 0.1779.03 ± 16.230.79 ± 0.17
Table 2. Accuracy and F1-score of four methods on the author’s dataset for “Left hand vs. Right hand” pair.
Table 2. Accuracy and F1-score of four methods on the author’s dataset for “Left hand vs. Right hand” pair.
SubjectLH-RH
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0192.370.9295.130.9595.760.9695.970.96
0280.860.8179.230.7981.060.8181.060.81
0349.30.4952.960.5554.930.5853.520.56
0478.360.7878.990.7977.310.7776.260.77
0571.560.6973.890.7176.920.7573.190.69
0667.190.6665.630.6670.310.764.580.65
0769.30.6973.810.7574.720.7676.30.77
0870.70.7171.340.7172.190.7264.970.65
0944.940.4449.580.5145.780.4547.260.45
1069.230.6969.230.761.920.6168.850.7
1196.990.9797.190.9796.990.9796.790.97
1268.010.6868.750.6968.010.6970.590.7
1389.460.8989.890.990.750.9190.540.91
1472.710.7375.050.7574.630.7570.580.71
1561.010.6153.960.5458.370.5954.190.55
1681.690.8284.980.8584.160.8481.480.81
Mean ± SD72.73 ± 14.080.72 ± 0.1473.73 ± 14.090.74 ± 0.1473.99 ± 14.280.74 ± 0.1472.88 ± 14.420.73 ± 0.14
Table 3. Accuracy and F1-score of four methods on the author’s dataset for “Right hand vs. Feet” pair.
Table 3. Accuracy and F1-score of four methods on the author’s dataset for “Right hand vs. Feet” pair.
SubjectRH-F
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0187.210.8783.80.8486.570.8685.070.85
0285.430.8680.360.8187.450.8883.40.84
0349.720.4949.720.4650.280.4649.440.47
0469.830.771.280.7272.520.7270.660.71
0576.360.7673.850.7378.450.7877.20.77
0682.640.8287.050.8786.790.8785.490.86
0767.820.6867.820.6867.590.6767.590.69
0893.510.9391.420.9293.510.9488.910.89
0951.270.545.990.4748.950.4947.890.47
1067.180.6766.410.6469.50.764.860.64
1192.380.9289.780.988.780.8986.970.87
1274.160.7481.650.8280.520.8178.280.78
1399.79198.940.9998.720.9999.791
1472.860.7375.850.7671.580.7170.940.71
1565.390.6665.170.6662.470.6459.780.62
1689.530.991.790.9290.760.9184.190.85
Mean ± SD76.57 ± 14.580.76 ± 0.1576.31 ± 14.90.76 ± 0.1577.15 ± 14.810.77 ± 0.1575.03 ± 14.480.75 ± 0.15
Table 4. Accuracy and F1-score of four methods on the author’s dataset for “Feet vs. Left hand” pair.
Table 4. Accuracy and F1-score of four methods on the author’s dataset for “Feet vs. Left hand” pair.
SubjectF-LH
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
AccF1AccF1AccF1AccF1
0186.940.8786.510.8786.510.8686.30.86
0289.610.989.210.8988.190.8887.580.88
0354.290.5455.680.5553.460.5353.190.51
0477.620.7779.710.7979.080.7978.450.78
0575.620.7375.620.7276.30.7276.520.73
0686.790.8687.310.8787.310.8788.60.88
0763.820.6461.290.6362.440.6463.820.65
0893.530.9392.90.9393.110.9391.650.92
0949.580.4946.850.4846.640.4847.270.48
1063.710.6459.850.667.950.6954.440.54
1194.180.9494.980.9593.780.9491.970.92
1272.320.7374.170.7475.650.77690.7
1398.920.9999.350.9999.57198.280.98
1484.710.8581.530.8282.380.8272.820.73
1565.10.6563.530.6566.220.6860.850.63
1684.80.8580.080.882.550.8371.660.72
Mean ± SD77.6 ± 14.830.77 ± 0.1576.79 ± 15.390.77 ± 0.1577.57 ± 14.880.78 ± 0.1574.53 ± 15.560.74 ± 0.15
Table 5. Mean difference in accuracy of “all measurement” and “single measurement” of self-acquired dataset.
Table 5. Mean difference in accuracy of “all measurement” and “single measurement” of self-acquired dataset.
LH-RHRH-FF-LH
FBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmRFBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmRFBCSPDFBCSP—FisherDFBCSP—mRmRDFBCSP—FmRmR
12.301.911.450.1214.220.140.45−0.3715.37−0.14−0.65−2.62
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nguyen, M.T.D.; Phan Xuan, N.Y.; Pham, B.M.; Nguyen, T.-H.; Huynh, Q.-L.; Le, Q.K. Evaluating the Motor Imagery Classification Performance of a Double-Layered Feature Selection on Two Different-Sized Datasets. Appl. Sci. 2021, 11, 10388. https://doi.org/10.3390/app112110388

AMA Style

Nguyen MTD, Phan Xuan NY, Pham BM, Nguyen T-H, Huynh Q-L, Le QK. Evaluating the Motor Imagery Classification Performance of a Double-Layered Feature Selection on Two Different-Sized Datasets. Applied Sciences. 2021; 11(21):10388. https://doi.org/10.3390/app112110388

Chicago/Turabian Style

Nguyen, Minh Tran Duc, Nhi Yen Phan Xuan, Bao Minh Pham, Trung-Hau Nguyen, Quang-Linh Huynh, and Quoc Khai Le. 2021. "Evaluating the Motor Imagery Classification Performance of a Double-Layered Feature Selection on Two Different-Sized Datasets" Applied Sciences 11, no. 21: 10388. https://doi.org/10.3390/app112110388

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop