Next Article in Journal
Concurrent fNIRS and EEG for Brain Function Investigation: A Systematic, Methodology-Focused Review
Next Article in Special Issue
Low-Light Image Enhancement Using Hybrid Deep-Learning and Mixed-Norm Loss Functions
Previous Article in Journal
Event-Related Potential-Based Brain–Computer Interface Using the Thai Vowels’ and Numerals’ Auditory Stimulus Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Quick-Response Eigenface Analysis Scheme for Brain–Computer Interfaces

1
Department of Electronic Engineering, Gachon University, 1342 Seongnam-daero, Seongnam 13306, Korea
2
School of Electronic Engineering, Kumoh National Institute of Technology, 61 Daehak-ro, Gumi 39177, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5860; https://doi.org/10.3390/s22155860
Submission received: 1 July 2022 / Revised: 29 July 2022 / Accepted: 1 August 2022 / Published: 5 August 2022
(This article belongs to the Special Issue Deep Learning Technology and Image Sensing)

Abstract

:
The brain–computer interface (BCI) is used to understand brain activities and external bodies with the help of the motor imagery (MI). As of today, the classification results for EEG 4 class BCI competition dataset have been improved to provide better classification accuracy of the brain computer interface systems (BCIs). Based on this observation, a novel quick-response eigenface analysis (QR-EFA) scheme for motor imagery is proposed to improve the classification accuracy for BCIs. Thus, we considered BCI signals in standardized and sharable quick response (QR) image domain; then, we systematically combined EFA and a convolution neural network (CNN) to classify the neuro images. To overcome a non-stationary BCI dataset available and non-ergodic characteristics, we utilized an effective neuro data augmentation in the training phase. For the ultimate improvements in classification performance, QR-EFA maximizes the similarities existing in the domain-, trial-, and subject-wise directions. To validate and verify the proposed scheme, we performed an experiment on the BCI dataset. Specifically, the scheme is intended to provide a higher classification output in classification accuracy performance for the BCI competition 4 dataset 2a (C4D2a_4C) and BCI competition 3 dataset 3a (C3D3a_4C). The experimental results confirm that the newly proposed QR-EFA method outperforms the previous the published results, specifically from 85.4% to 97.87% ± 0.75 for C4D2a_4C and 88.21% ± 6.02 for C3D3a_4C. Therefore, the proposed QR-EFA could be a highly reliable and constructive framework for one of the MI classification solutions for BCI applications.

1. Introduction

The brain–computer interface (BCI) is a continuously developing technology that implement the brain’s thought to manipulate external bodies without thinking through the human nerves to the hands or feet [1,2]. Thus, BCI systems are intended to generate new communication channels for other human parts [3,4]. There are two types of justifications available. One is uni-directional communication from a brain to a computer and the other is bi-direction communication between a brain and computer, as well as brain-to-brain. The current state of art technology for the BCI is bi-direction BCI technology or brain to brain interface (B2BI) [5]. The B2BI functions via on one hand as a brain–computer interface (BCI) to retrieve messages and, on the other hand, a computer-brain interface (CBI) [6]. The communication paths that do not go through human’s nerves could help people with physical disabilities by controlling objects, such as a wireless mouse [7]. The signals that are generated from the human brain are used to implement the systems, and they are extremely complex; such signals are recorded in electronic system, i.e., an electroencephalogram (EEG) [8]. In the beginning of EEG research, EEG measurement used invasive and painful methods, in which several electrodes were inserted directly into the human skin [9]. The non-invasive methods of attaching the electrodes to the human scalps have been used because they are non-painful, user-friendly, and economical ways to obtain EEG signals [10]. In particular, the classification of motor imagery (MI) is researched because EEG signals are shown when the right and left hands are related, respectively [11].
In the BCI systems, raw EEG signals need to be processed with filtering, extraction, and classification techniques [12]. We describe the whole BCI workflow based on the OSI 7-layer model; the OSI 7-layer model is a kind of the communication model. Therefore, it could be useful for readers to understand the feature extraction and classification algorithms in the signal processing workflow. In OSI 7-layer model, there are two types of disciplines. One is a datagram for a layered network hierarchy, and the other is a conceptual model for multistage processing. In the layered hierarchy model, there are the physical, medium access control, network, transport, or routing, etc. In the conceptual model for multistage processing, the focus is on multistage processing, such as the input, data acquisition, preprocessing, decision/detection, and hypothesis test. In this article, based on the previous literature of IEEE [11], we compared the similarity between the BCI dataflow and OSI 7-layer in the conceptual model for better understanding of the signal progressing in BCIs. According to N. Khadijah et al., in the previous related work [13], we used the open system interconnection (OSI) layer, as shown in Figure 1. In the BCI model, we used a layer architecture staring from data acquisition (layer 1), going through pre-processing (layer 2), approaching at feature extraction (layer 3), and then ending at classification (layer 4).
In the signal processing steps, feature extraction was used; generally, a filter block is applied before feature extraction [14,15,16]. There are several feature extraction methods, such as wavelet transform, short-time Fourier transform (STFT), common spatial pattern (CSP), regularized CSP (RCSP), common spatio-spectral pattern (CSSP), common sparse spectral spatial pattern (CSSSP), etc. [8]. As a representative visualization method, short time Fourier transform (STFT), spectrogram, wavelet, etc., were used [17]. Among those methods, the CSP is one of the most widely accepted feature extraction methods [18]. However, the CSP method has overfitting problems when data are sparse in the domain. The CSP method is also noise sensitive, so it is difficult to classify multi-class EEG signals. Blankertz et al. focused how a CSP filter affected the EEG signals [19]. The combination of CSP has been used by many researchers, including Lotte and Guan [20]. Their idea was used to regularize CSP and substitute the normal CSP. The limitation is that the researchers were required to use the full set of data for the training and testing sections. The training data should have a larger number of data than the test data. The full set of data that were used, especially the test data, may lead to an overestimation in performance because all the information is being used. To classify the MI for a subject, the algorithms required the use of other subjects, as well. Therefore, the algorithms are not suitable for classifying a single subject dataset. The extended algorithm of the CSP was used by Lemm et al. [10]. This algorithm is known as CSSP. The CSSP method introduced a delay to the system, so that the spectral filter was included in the system. The simulation results, except for six datasets, showed better accuracy than CSP.
In current EEG classification studies, linear discriminant analysis (LDA), support vector machine (SVM), deep learning, etc., were used [21]. Deep learning (DL) is a machine learning (ML) method used in various fields, such as the voice and video processing fields [22]. The DL works successfully on non-linear and non-stationary data, and it works efficiently even in the fields that are difficult for humans to distinguish [23]. The convolutional neural network (CNN) is one of the DL algorithms that is widely used for data classification [24]. Due to CNN characteristics, there are attempts to apply CNN for EEG signal classification. To use the CNN method specialized for image classification, various methods to visualize EEG signal are being studied [25]. Visualization of these EEG signals was used to help improve the classification performance of DL models using CNN algorithms. Compared to those traditional classification algorithms, such as LDA and SVM, the deep learning model requires the use of large numbers of the dataset [25]. Therefore, the limited numbers of the dataset need to be increased using a data augmentation technique. Considering the finite training data in BCIs, Huang et at.al proposed a data augmentation scheme via a CNN [26]. The data augmentation has accomplished by mixing and recombining the images. Lee et al. convers the lingering problem of zero-training by utilizing a proposed CNN model connected to P300 information [27].
In the EFA method, the calculated eigenface coefficient was used as a feature for data classification. In this case, it is possible to reconstruct the pictures using the eigenface coefficients. We proposed a novel EEG signals classification method, called quick response EFA (QR-EFA) utilizing deep knowledge in standardized and sharable QR image formulations. The QR-EFA is subjected in the EEG signal preprocessing stage. Image reconstruction or data augmentation using QR-EFA has been shown to generate QR image features suitable for CNN algorithms. To overcome the constraints, such as the trial numbers and limited BCI competition dataset in EEG, we proposed an innovative classification technique that was organically designed for EFA and CNN based on data augmentation. After learning the impulse response filter for the BCI competition IV dataset 2a (C4D2a_4c) [28] and BCI competition III dataset 3a (D3D3a_4c) [29], using a seven-layer simple CNN model (one input and output, one convolution, one pooling, and three fully connected), the test EEG data classification performance was measured according to the transfer function. As of today, the highest classification results for EEG 4 classes on BCI competition IV 2a is 85.4% [1]. Based on this observation, our proposed QR-EFA method could provide higher accuracy performance of MI classification for competition dataset in BCIs.

2. Materials and Principles

The proposed EFA algorithm is a feature extraction method from EEG data that builds up neuro images, emphasizing the discriminability of classes; the feature is a kind of tool.
The QR-EFA method was formulated based on the previous EFA result. From egienfaces derived from the previous EFA, we will provide the description. First, the eigenface is asymmetric in horizontal and vertical directions, and the image size is excessively bigger than the size that CNN can process. Second, in BCI competition dataset, there are many discrepancies in data sizes among number of classes, trials, and subjects. Based on this fact, we proposed QR eigenface to provide a symmetrical, standardized, and sharable sizes in images. In the framework level, QR-EFA consists of the parts, such as EFA, eigenface restructuring, data augmentation, and CNN. Figure 2 shows the relationship between the previous proposed EFA and proposed QR-EFA.
The EFA is different PCA image recognition type for dimension reduction [30]. The fundamental EFA method is depicted in Figure 3. The EEG data were preprocessed. As shown in steps 1–3, EEG data were converted to image data to build up eigenface. From eigenface, coefficients, which we called features, were extracted for training and testing procedures. Afterwards, the classes were classified for the next step.
Let us define the next step of the data interpretation after EFA procedure. We also used the same data interpretation techniques for classification.
Step 1: The three-dimensional (3D) converted EEG image data could be separated as time, channels, and trials. The data were recognized in each separated 3D converted EEG image data, and the generated data could be differentiated, depending on the viewpoint for each 3D direction.
Step 2: From the differentiated 3D EEG image data, the covariance matrix must be obtained. Afterwards, we could determine and build up the eigenfaces.
Step 3: The training data were projected to obtain the requested features. The dataset was composed of, or represented with, ‘time (S)’, ‘channels (C)’, and ‘trials (N)’, as described in Equation (1).
Dataset = time (S) × channels(C) × trials (N)
EFA is a method to obtain the differentiated EEG data with different directions. The EEG image data with the time, channels, and trials were combined into the dataset, which is an infinite number (N) of the image data with the same manner as that illustrated in Equation (2). In other words, we consider 3D data in two-dimensional (2D) images by combining or concatenating ‘time’ and ‘channel’ data together as in T = SC (time × channel). In fact, the derived tentative dataset, T is composed of ‘time’ and ‘channel’ components in series. The desired dataset or image matrix was obtained by rearranging the component T in horizontal direction and component N in vertical direction as shown in Figure 4. Consequently, Equation (2) shows the final dataset or 2D images.
Dataset = TN where T = SC
The eigenface was then obtained from differentiated 3D EEG image, and newly obtained image data Φ and value Ψ need to be calculated.
Φk =Datasetk − Ψk, k = 1, 2, …, N
The covariance matrix extracted from the differentiated 3D EEG image data without the mean value is obtained in Equation (4).
C = 1 L   k = 1 L Φ k Φ k T
Let us define the eigenvectors of X with eigenvalues of L of the covariance matrix C after solving the following equation CX = kX. Among the vectors extracted from this matrix, the k vectors were chosen. The eigenfaces must be extracted with only training eigenface Гtrain. Subsequently, the training features were extracted from the training eigenface and data. The extracted eigenface coefficients were projected in Equation (5).
Ωtrain= Φtrain Гtrain
The weight coefficient Ωtrain was obtained to be a training feature. Equation (6) shows how to obtain the feature coefficient Ωtest. The eigenspace was trained, and the EEG test data can be classified as shown in Equation (6).
Ωtest= Гtrain Φtest

3. QR-EFA

3.1. Idea Formulation for QR-EFA

The description of the procedure and flowchart for QR-EFA is shown. The proposed QR-EFA procedure is given below.
  • Perform the process with channel direction.
  • Original image eigenface formulation.
  • Confine the number of eigenfaces, up to the number of classes required and trials in common among subjects.
  • Adjust eigenface size horizontally with number of classes and vertically with number of trials.
  • Modify image eigenface formulation by multiplying a brightness factor with steps 3 and 4.
  • Obtaining training coefficients by projecting train images to the eigenface.
  • Implement QR neuro train images with the coefficients using the result of step 6.
  • Obtaining testing coefficients by projecting test images to the eigenface.
  • Implement QR neuro test images with the coefficients using the result of step 8.
  • Neuro image augmentation to diversify the limited neuro train images.
  • CNN training process with the results produced, using the result of step 10.
Figure 5 shows the flowchart of the procedure for QR-EFA, described above.

3.2. Data Augmentation for the Limited Training Data

After we finished the EFA method for the C4D2a_4C and C3D3a_4C datasets, the features were obtained. Based on the features, QR code-like brain neuro images will be formulated. There are two different image types. One is for the training process, and the other is for the testing process. The amount of training data set is definitely limited in BCI competitions, so we need a data augmentation process to multiply the limited images to the desired number of training images. To augment the limited training images, we adapted a brightness control for the input training images, according to the following probability model. For the given normal probability density distribution (PDF), with mean (μ) and standard deviation (σ) x ~ N (μ, σ2), the statistical random data x[n] will be obtained as follows in Equation (7) [31].
PDF :   f ( x ) = 1 2 π σ e ( x μ ) 2 / 2 σ 2 , < x < x = μ + σ
The MATLAB code for training image data augmentation by Gaussian blur (brightness and contrast implication or image bacteria culturing per pixel) is as follows (Box 1).
Box 1. The MATLAB code for training image data augmentation by Gaussian brightness control.
rand_img = 1 + (0.7).* randn(is1,is2);
% 1/0.7 (example of is1,is2,is3: 18,16,30)
timg = trimages(:,:,j).* rand_img;
% neuro image culture by multiplication
The data augmentation per pixel was demonstrated to enhance the performance of deep learning approaches by reducing overfitting problems. The overfitting or high variance in ML models are produced if the dataset used to “teach” the model is greater than the testing accuracy. The overfitting problems can be generated when the model has very high errors in the testing dataset. According to the data science literature, CNN needs a huge number of neuro images for the model to be trained effectively [32]. This proposed technique could be helpful for increasing the performance of the model, thereby reducing overfitting problems from the EEG data. Most BCI competition datasets for classification and object detection datasets have a few hundred to thousands of neuro images. Considering the rule of larger numbers or the invariance property of CNN, the objects can even be classified when they are visible in different sizes, orientations, or potentials on channels. Hence, we can take the limited neuro competition dataset of images and transform the objects into different sizes by controlling for brightness, scaling in size, rotating in orientation, or zooming in and out. Through this type data augmentation, we can create intense and diverse neuro image datasets with the extent of variations.

3.3. Covolution Neural Network for QR-EFA

Compared to conventional classifiers, such as LDA, SVM, and multilayer perceptron (MLP), the convolution neural network (CNN) is the most widely used multi-layered neural network ML method for image classification [19]. A general CNN method learns an input image through convolution, pooling, and fully connected layers, so it acts as a classifier [33]. The convolutional layer is composed of feature maps, with several different weight vectors; each feature map is calculated from the input image. The CNN-based image classification models, including ResNet and DarkNet, are widely known [33]. Using an appropriate ML model and neural network structure, according to the target images to be classified, is very important for obtaining high classification accuracy [34]. To better show that EEG images are easily classified using the QR-EFA method, a simple CNN model using only one convolutional layer, a pooling layer, and a fully connected layer was used.
In Figure 6, a simple one convolution and mean pooling are shown to reduce a computation time and minimize the task burden. The size of input images is 12 × 10 and then it goes through a convolution layer with kernel or filter size of 7 × 1. After the layer process is done, it would be 6 × 10 output; then, it goes through 2 × 2 mean and max pooling layers. After processing a fully connected layer, we finally obtained the output with the class classification output for BCI dataset.
Figure 6 shows a CNN model that classifies the EEG images converted by the QR-EFA method. The convolution layer of the CNN model, which consists of sixteen filters of size 9 × 3 and a pooling layer of size 2 × 2, was used. Using softmax and cross entropy methods, two class classification results are obtained. As hyper parameters, the learning rate was set to 0.0001, and a batch size of 512 and momentum optimizer were used. The CNN model was trained for sufficient learning, and training was set to finish early to prevent overfitting problems.

4. Experimental Results

The QR codes are widely used for quick response as matrix barcodes. Just as with QR codes, we also obtained the images, so we called them QR images. After performing QR image data, we need further steps. In obtaining QR images for training and testing images, there are three factors to consider for better classification, in order to have accurate, robust, and reliable results. First, let us define some terminologies, such as the domain-wise similarity among domains, trial-wise similarity among trials, and subject-wise similarity among subjects. Domain-wise similarity is the degree of similarity that indicates some common characteristics on the eigenfaces between the training and testing domains. Trial- and subject-wise similarity were degrees of similarity among trials and subjects, respectively. We could recognize the similarity from QR images. Our proposed QR-EFA algorithm maximizes the similarities, with the respect to domains, trials, and subjects. As result, it could generate unique features per domain, so it is possible to discriminate the classes efficiently.
The background regarding EEG datasets from BCI competitions needs to be explained. To validate the proposed method, we used two EEG datasets, such as the BCI competition III dataset 3a (C3D3a_4C) and BCI competition IV dataset 2a (C4D2a_4C), for four classes. Specifically, C3D3a_4C is dataset from three subjects or participants, and C4D2a_4C is dataset from nine subjects, which are the off-line, publicly available, and open accessible dataset of the BCI competition database. Hence, this article focuses on the C3D3a_4c and C4D2a_4C. Table 1 shows the detailed number of trials per subject for the C3D4a_4C used in this article.
The detailed information of the property for C3D3a_4C is given as follows (Box 2).
Box 2. The detailed information of the property for C3D3a_4C.
comment1: ‘dataset: C3D3a_4C’
date: ‘2021.12.28’
madeby: ‘4C’
affiliation: ‘KNIT’
window: ‘offset: 3.5, length: 2’
subject: ‘subject #: 1’
prefiltering: ‘off’
s: 250 (sample/sec)
c: [1 × 60 cell]
x: [500 × 60 × 180 double]
y: [1 × 180 double]
The MI classification EEG images extracted with QR-EFA were classified for C3D3a_4C. The data augmentation using Gaussian distribution was used for sufficient training of the ML model and prevention of overfitting. From 10,800 augmented training images, which come from the data augmentation of the original QR neuro images, in brightness and without data augmentation, a total of 2592 non-augmented test images were used. As a result of the final classification experiment, the accuracy was obtained for the test dataset of 2592 sheets, as shown in Table 2.
Figure 7 shows the examples of eigenfaces, formulated based on the QR code. Although there are lots of QR-eigenfaces are available, we will confine only a small number of eigenfaces, considering the number of classes and trials among the subjects. We chose the eigefacecs reflecting the number of classes. We designed the first to left, second, right, the third to feet, and the last to tongue classes. However, the order is subject to change as the supervisory learning constrictions. This type of decision is related to the choice regarding the approach between the supervised and unsupervised learning.
Thus far, we obtained the training coefficients by projecting the training images to the QR-eigenfaces and testing the coefficients by projecting the testing images to the QR-eigenfaces. The next step is to conceive of new QR images for training and testing the dataset. Subsequently, linear combinations of the QR eigenfaces, weighted by the relevant coefficients, were used as QR images for training and testing data.
During the data augmentation process for overcoming the limited dataset problem in the BCI competition, we modified the brightness of the recovered images based on the Gaussian noise contamination equation. Although the grayscale and quantized images from the first to last appear similar, the real binary values of the images are different and unique, and they can be used for classification. Because of ML or AI paradigm considerations, a sufficiently large number of training images are needed to train CNNs and fine-tune the logic. However, as the BCI training dataset and its QR images are finite and limited, we need to amplify or diversify them by utilizing data augmentation in the brightness direction. Figure 8 shows one of the EFA QR code implementation results after the brightness data augmentation. Among four classes, considering its dominance and the activation region of the brain, we only selected and data augmented four QR training images for the left and tongue classes, as shown.
Because the BCI data are random signals, they must be treated in statistical signal-processing domains. Figure 9 shows the typical neuro raw images considered for the QR-EFA. The left image is for the first trial, and the right image is for the last trial.
Based on the QR-eigenfaces for left, right, feet, and tongue, we proposed considering domain-, trial-, and subject-wise similarities. The detailed considerations are as follows. First, subject-wide similarities among the subjects were checked. Second, domain-wise similarity between the training and testing domains was considered. Finally, a trial-wise similarity was investigated among the trials.
Figure 10 shows the domain-wise similarity between the training and testing domains for subject 1. Considering their importance and the degree of information, only the QR images for subject 1 and trial 1 are shown among training or testing data and several trials. There are many similarities between training and testing QR images. We observed the trial-wise similarity of QR images between the first and last trials for subject 1. We also examined subject-wise similarity of QR images between subjects in the same first trial. It was confirmed that QR-EFA maximizes the three proposed similarities in domain, trials, and subjects.
Using QR-EFA, a sample output of the CNN testing results for BCI competition III dataset 3a (C3D3a_4C) is shown in Figure 11. As the number of epochs increases during simulations, the cost or loss function decreases and reaches a limit.
The MATLAB code for a sample mean of C3D3a_4C data and its confidence interval, when the sample size n = 10 is, as follows (Box 3).
Box 3. The MATLAB code for a sample mean of C3D3a_4C.
% result set #1
x1(1) = 0.907143; x1(2) = 0.795238; x1(3) = 0.802381; x1(4) = 0.980952;
x1(5) = 0.919048; x1(6) = 0.859524; x1(7) = 0.921429; x1(8) = 0.904762;
x1(9) = 0.909524; x1(10) = 0.821429;
% result set #2
x1(1) = 0.857143; x1(2) = 0.930952; x1(3) = 0.928571; x1(4) = 0.926190;
x1(5) = 0.945238; x1(6) = 0.864286; x1(7) = 0.871429; x1(8) = 0.783333;
x1(9) = 0.928571; x1(10) = 0.952381;
SD1 = std(x1); % Standard deviation (SD)
SE1 = SD1/sqrt(length(x1)); % Standard error(SE)
ts1 = tinv([0.025 0.975],length(x1)−1);
% T-value/score for 95% CI (2.26)
CI1 = mean(x1) + ts1 * SE1;
% Confidence Intervals
>>mean(x1)
Ans = 0.8821
>>std(x1)
ans = 0.0602
>>CI1 % Confidence_Interval
CI1 = 0.8390 0.9252
>>CI1(2)-mean(x1) %Confidence Interval
Ans = 0.0431
The computation, simulation, verification, and validation on the proposed scheme requires considerations of the confidence interval for the given number of trials (n) and variance or standard deviations. Based on the nature of random seeds in statistical signal processing, we consider a 95% confidence interval, due its variance and spread for data augmentation, CNN layer initiations for kernel filters, and QR-eigenface selections for training images.
From the above results, with a 95% confidence interval, the sample mean was estimated to be 0.88, with a confidence interval of 0.0431. Note that, in this case, the Student’s t-distribution is applicable for calculating the confidence interval and sample variance (s2), because we do not have information regarding the variance or standard deviation (SD) of the population distribution in BCI datasets (C3D3a_4C or C4D2a_4C both). Table 3 shows the 10 simulation results, as the random seeds changes uniformly for statistically justification for confidence interval.
To calculate a 95% confidence interval for the sample mean μ, an estimated coefficient value 2.26 was used [35]. Thus, the mean and confidence interval of accuracy were 88.21% ± 6.02%, with the confidence interval of 4.31, i.e., (83.90, 92.52). Consequently, the probability for the same mean under the 95% confidence interval is given by Equation (8). For a comparison, the latest and best classification accuracy reported thus far for EEG four classes on BCI competition IV 2a was 85.4 [31].
P ( | X ¯ μ | < 2.26 σ / n ) = P ( | X ¯ μ s / n | < 2.26 ) = 0.95
Then, we compute the confidence interval using Equation (9).
[ X ¯ 2.26 s n , X ¯ + 2.26 s n ] = [ 83.90 , 92.52 ]
To validate and verify QR-EFA in a real dataset, with a comparison to C3D3a_4C, the next section is for the result of C4D2a_4C. Table 3 shows the number of trials per subjects for C4D2a_4C. The C4D2a_4C dataset is composed of nine subjects and the predefined number of experimental trials. The number of trials for left, right, foot, and tongue classes in the C4D2a_4C dataset, composed of nine subjects and the predefined number of experimental trials, was 72.
The detailed information of the property for C4D2a_4C is given as follows (Box 4).
Box 4. The detailed information of the property for C3D2a_4C on MATLAB.
comment1: ‘dataset: C4D2a_4C’
date: ‘2021.02.12’
madeby: ‘4C’
affiliation: ‘KNIT’
window: ‘offset: 3.5, length: 2’
subject: ‘subject #: 1’
prefiltering: ‘off’
s: 250 (sample/sec)
y: [1 × 288 double]
x: [500 × 22 × 288 double]
c: [22 × 1 cell]
Using the designated CNN model, which is depicted in Figure 7, MI classification EEG images extracted with QR-EFA were classified for C4D2a_4C. The accuracy was obtained for the test dataset of 2592 sheets, as shown in Table 4.
In Figure 12, the eigenfaces has been formulated. The eigenfaces reflecting the number of classes were selected. Considering a supervisory learning limitation, we designed the first to left, second, right, third to feet, and last to tongue classes.
The training coefficients by projecting the training images to the QR-eigenfaces and testing coefficients by projecting the testing images to the QR-eigenfaces were obtained. The next step is to conceive of new QR images for the training and testing dataset. Afterwards, we will perform a liner combination of the QR-eigenface, depending on the relevant coefficients. The results of the linear combination were the QR images for the training and testing data. Figure 13 shows the results after brightness data augmentation in the brightness direction. Among four classes, the selected and data augmented images for left and tongue classes are shown.
Figure 14 shows the typical neuro raw images. The left image was of the first trial and the right was of the last trial for subject. First, we need to check the subject-wide similarity among subjects. Secondly, we check the domain-wise similarity between training and testing domain. Finally, we check the trial-wise similarity among trials.
Figure 15 shows the domain-wise similarity between training and testing domain for subject 1 and trial 1. There were also similarities between the training and testing images. Hence, we also maximized the similarities in the directions of domain, trials, and subjects.
In summary, it is clear that QR-EFA guarantees a distinguishable and discriminative feature per class, as well as a higher similarity, with respect to domain-, trial-, and subject-wise similarities. This unique novel scheme accelerates the improved accuracy in CNNs. With QR-EFA, a sample output of CNN testing results for in C4D2a_4C is provided in Figure 16. It also shows that the cost function reached the limit.
The MATLAB code for a sample mean of in C4D2a_4C data and its confidence interval, when the sample size n = 10 is, as follows (Box 5).
Box 5. The MATLAB code for a sample mean of in C4D2a_4C data.
% result set #1
x1(1) = 0.966435; x1(2) = 0.976852; x1(3) = 0.984954; x1(4) = 0.971065;
x1(5) = 0.969136; x1(6) = 0.983025; x1(7) = 0.986883; x1(8) = 0.985725;
x1(9) = 0.984182; x1(10) = 0.978395;
SD1 = std(x1); % Standard deviation (SD)
SE1 = SD1/sqrt(length(x1)); % Standard error(SE)
ts1 = tinv([0.025 0.975],length(x1)−1); % T-value/score for 95% CI (2.26)
CI1 = mean(x1) + ts1 * SE1; % Confidence Intervals
>>mean(x1)
Ans = 0.9787
>>std(x1)
ans = 0.0075
>>CI1 % Confidence_Interval
CI1 = 0.9733 0.9840
>>CI1(2)-mean(x1) %Confidence Interval
Ans = 0.0054
Table 5 shows the 10 simulation results, as the random seeds changes uniformly for statistically justification for confidence interval.
From the above results, with a 95% confidence interval, we obtained a range of estimates for a sample mean (0.98), and the confidence interval was 0.0054. The mean and confidence interval of the accuracy was 97.87 ± 0.0075%, with a confidence interval of 0.0054, i.e., (97.33, 98.40). The latest and best classification result is shown below. Then, we compute the confidence interval using Equation (10).
[ X ¯ 2.26 s n , X ¯ + 2.26 s n ] = [ 97.33 , 98.40 ]

5. Conclusions

In this study, we proposed the QR-EFA method for efficient motor-imaginary (MI) EEG classification and showed that the EEG signal, when converted into a standardized and sharable QR image, can be classified when using a simple CNN. To obtain a unique feature per class, EEG signal data can be utilized with the concepts of domain-, trial-, and subject- similarities. The EEG data measured from several subjects were used using the BCI competition 3 dataset 3a (C3D3a_4C) and BCI competition 4 dataset 2a (C4D2a_4C). Through QR-EFA method, the EEG signal was converted into EEG QR image, as formed with the EFA method, and a data augmentation technique was applied to solve the limited EEG image problem.
For optimum and best classification performance, QR-EFA maximizes the proposed domain-, trial-, and subject-wise similarities. As far as we are concerned, none of the literature in BCIs has considered the domain-, trial-, and subject-wise similarities so far. Based on this observation, our proposed QR-EFA method was used to maximize the three similarities in the directions of domains, trials, and subjects. Using a simple seven-layer CNN model, data classification results showed an exceptional and remarkable accuracy of 97.87% ± 0.75, with a confidence interval of 0.54, i.e., (97.33, 98.40) for the C4D2a_4C and 88.21% ± 6.02, with a confidence interval of 4.31, i.e., (87.90, 92.52) for the C3D3a_4C. Unlike the CSP method, which is only robust in the original two-class classification, the proposed QR-EFA method is applicable to the classification of two multi-classes of EEG signal; hence, it definitely advantages from the use of the QR-EFA algorithm. Because the QR-EFA method extracts classification features and generates EEG QR images that are well-differentiated between classes, it is suitable as training and testing input images for further ML process and can greatly contribute to classification performance improvement.
Because the ML technique for image classification is becoming stronger with the development of better performance models and hardware performance, QR-EFA’s flexible EEG conversion method or frameworks, which should be applied to non-motor imaginary (MI), such as word thinking, emotion detections, arithmetic calculation operations, and multi-class classification, will be developed in the future. The application to a larger number of classes, such as five or six categories, will be applied, with slight extensions in time.

Author Contributions

Conceptualization, H.C., J.P. and Y.-M.Y.; methodology, J.P. and Y.-M.Y.; formal analysis, H.C. and Y.-M.Y.; writing—original draft preparation, H.C. and Y.-M.Y.; supervision, Y.-M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are included within the article.

Acknowledgments

The work reported in this paper was conducted during the sabbatical year of Kumoh National Institute of Technology in 2021. This work was supported by the Gachon University research fund of 2021 (GCU-202110180001).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
EFAEigenface analysis
QR-EFAQuick eigenface analysis
BCIBrain computer interface
EEGElectroencephalogram
CSPCommon spatial pattern
C3D3a_4CBCI competition III dataset IIIa for four classes
C4D2a_4CBCI competition IV dataset IIa for four classes
PCAPrincipal component analysis
ICAIndependent component analysis

References

  1. Classification Ranking for EEG 4 Classes on BCI Competition IV 2a. Available online: https://paperswithcode.com/sota/eeg-4-classes-on-bci-competition-iv-2a (accessed on 22 April 2022).
  2. Pfurtscheller, G.; Neuper, C.; Guger, C.; Harkam, W.; Ramoser, H.; Schlogl, A.; Obermaier, B.; Pregenzer, M. Current trends in Graz brain-computer interface (BCI) research. IEEE Trans. Neural Syst. Rehabil. Eng. 2000, 8, 216–219. [Google Scholar] [CrossRef] [PubMed]
  3. Schalk, G.; McFarland, D.J.; Hinterberger, T.; Birbaumer, N.; Wolpaw, J.R. BCI2000: A general-purpose brain-computer interface (BCI) system. IEEE Trans. Biomed. Eng. 2004, 51, 1034–1043. [Google Scholar] [CrossRef] [PubMed]
  4. Belwafi, K.; Gannouni, S.; Aboalsamh, H. An Effective Zeros-Time Windowing Strategy to Detect Sensorimotor Rhythms Related to Motor Imagery EEG Signals. IEEE Access 2020, 8, 152669–152679. [Google Scholar] [CrossRef]
  5. Nam, C.S.; Nijholt, A.; Lotte, F. Brain–Computer Interfaces Handbook: Technological and Theoretical Advances; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  6. Boi, F.; Moraitis, T.; De Feo, V.; Diotalevi, F.; Bartolozzi, C.; Indiveri, G.; Vato, A. A bidirectional brain-machine interface featuring a neuromorphic hardware decoder. Front. Neurosci. 2016, 10, 563. [Google Scholar] [CrossRef] [Green Version]
  7. Chin-Teng, L.; Yu-Chieh, C.; Teng-Yi, H.; Tien-Ting, C.; Li-Wei, K.; Sheng-Fu, L.; Hung-Yi, H.; Shang-Hwa, H.; Jeng-Ren, D. Development of Wireless Brain Computer Interface with Embedded Multitask Scheduling and its Application on Real-Time Driver’s Drowsiness Detection and Warning. IEEE Trans. Biomed. Eng. 2008, 55, 1582–1591. [Google Scholar]
  8. Nicolas-Alonso, L.F.; Gomez-Gil, J. Brain Computer Interfaces, a Review. Sensors 2012, 12, 1211–1279. [Google Scholar] [CrossRef]
  9. Jinyi, L.; Yuanqing, L.; Hongtao, W.; Tianyou, Y.; Jiahui, P.; Feng, L. A Hybrid Brain Computer Interface to Control the Direction and Speed of a Simulated or Real Wheelchair. IEEE Trans. Neural Syst. Rehabil. Eng. 2012, 20, 720–729. [Google Scholar]
  10. Lemm, S.; Blankertz, B.; Curio, G.; Muller, K. Spatio-spectral filters for improving the classification of single trial EEG. IEEE Trans. Biomed. Eng. 2005, 52, 1541–1548. [Google Scholar] [CrossRef]
  11. Aznan, N.K.N.; Yeon-Mo, Y. Applying Kalman filter in EEG-Based Brain Computer Interface for Motor Imagery classification. In Proceedings of the 2013 International Conference on ICT Convergence (ICTC), Jeju, Korea, 14–16 October 2013; pp. 688–690. [Google Scholar]
  12. Lotte, F.; Congedo, M.; Lécuyer, A.; Lamarche, F.; Arnaldi, B. A review of classification algorithms for EEG-based brain–computer interfaces. J. Neural Eng. 2007, 4, R1. [Google Scholar] [CrossRef]
  13. Aznan, N.K.N.; Huh, K.-M.; Yang, Y.-M. EEG-based motor imagery classification in BCI system by using unscented Kalman filter. Int. J. Inf. Commun. Technol. 2016, 9, 492–508. [Google Scholar] [CrossRef]
  14. Roijendijk, L.; Gielen, S.; Farquhar, J. Classifying Regularized Sensor Covariance Matrices: An Alternative to CSP. IEEE Trans. Neural Syst. Rehabil. Eng. 2016, 24, 893–900. [Google Scholar] [CrossRef] [PubMed]
  15. Robinson, N.; Vinod, A.P.; Kai Keng, A.; Keng Peng, T.; Guan, C.T. EEG-Based Classification of Fast and Slow Hand Movements Using Wavelet-CSP Algorithm. IEEE Trans. Biomed. Eng. 2013, 60, 2123–2132. [Google Scholar] [CrossRef] [PubMed]
  16. Husain, A.M.; Sinha, S.R. Continuous EEG Monitoring: Principles and Practice; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  17. Jin, J.; Xiao, R.; Daly, I.; Miao, Y.; Wang, X.; Cichocki, A. Internal feature selection method of CSP based on L1-norm and Dempster–Shafer theory. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4814–4825. [Google Scholar] [CrossRef] [PubMed]
  18. Blankertz, B.; Muller, K.; Curio, G.; Vaughan, T.M.; Schalk, G.; Wolpaw, J.R.; Schlogl, A.; Neuper, C.; Pfurtscheller, G.; Hinterberger, T.; et al. The BCI competition 2003: Progress and perspectives in detection and discrimination of EEG single trials. IEEE Trans. Biomed. Eng. 2004, 51, 1044–1051. [Google Scholar] [CrossRef] [PubMed]
  19. Vidaurre, C.; Krämer, N.; Blankertz, B.; Schlögl, A. Time Domain Parameters as a feature for EEG-based Brain–Computer Interfaces. Neural Netw. 2009, 22, 1313–1319. [Google Scholar] [CrossRef]
  20. Lotte, F.; Guan, C. Regularizing common spatial patterns to improve BCI designs: Unified theory and new algorithms. IEEE Trans. Biomed. Eng. 2010, 58, 355–362. [Google Scholar] [CrossRef] [Green Version]
  21. Choi, H.; Park, J.; Lim, W.; Yang, Y.-M. Active-beacon-based driver sound separation system for autonomous vehicle applications. Appl. Acoust. 2021, 171, 107549. [Google Scholar] [CrossRef]
  22. Yu, X.; Chum, P.; Sim, K.-B. Analysis the effect of PCA for feature reduction in non-stationary EEG based motor imagery of BCI system. Optik 2014, 125, 1498–1502. [Google Scholar] [CrossRef]
  23. Pearson, K. LIII. On lines and planes of closest fit to systems of points in space. Lond. Edinb. Dublin Philos. Mag. J. Sci. 1901, 2, 559–572. [Google Scholar] [CrossRef] [Green Version]
  24. Lu, H.; Eng, H.-L.; Guan, C.; Plataniotis, K.N.; Venetsanopoulos, A.N. Regularized common spatial pattern with aggregation for EEG classification in small-sample setting. IEEE Trans. Biomed. Eng. 2010, 57, 2936–2946. [Google Scholar]
  25. Bengio, Y.; Goodfellow, I.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
  26. Wang, Y.; Huang, G.; Song, S.; Pan, X.; Xia, Y.; Wu, C. Regularizing deep networks with semantic data augmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3733–3748. [Google Scholar] [CrossRef] [PubMed]
  27. Lee, J.; Won, K.; Kwon, M.; Jun, S.C.; Ahn, M. CNN with large data achieves true zero-training in online P300 brain-computer interface. IEEE Access 2020, 8, 74385–74400. [Google Scholar] [CrossRef]
  28. The BCI Competition IV Dataset 2a for Four Classes (C4D2a_4C). BCI-Competition III (2008). Available online: https://www.bbci.de/competition/iv (accessed on 25 July 2022).
  29. The BCI Competition III Dataset 3a for Four Classes (C3D3a_4C). BCI-Competition III (2005). Available online: https://www.bbci.de/competition/iii (accessed on 25 July 2020).
  30. He, L.; Hu, D.; Wan, M.; Wen, Y.; von Deneen, K.M.; Zhou, M. Common Bayesian network for classification of EEG-based multiclass motor imagery BCI. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2016, 46, 843–854. [Google Scholar] [CrossRef]
  31. Kim, K.M.; Choe, S.-H.; Ryu, J.-M.; Choi, H. Computation of Analytical Zoom Locus Using Padé Approximation. Mathematics 2020, 8, 581. [Google Scholar] [CrossRef] [Green Version]
  32. Skansi, S. Introduction to Deep Learning: From Logical Calculus to Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  33. Kim, J.; Ko, J.; Choi, H.; Kim, H. Printed Circuit Board Defect Detection Using Deep Learning via A Skip-Connected Convolutional Autoencoder. Sensors 2021, 21, 4968. [Google Scholar] [CrossRef] [PubMed]
  34. Riaz, H.; Park, J.; Choi, H.; Kim, H.; Kim, J. Deep and Densely Connected Networks for Classification of Diabetic Retinopathy. Diagnostics 2020, 10, 24. [Google Scholar] [CrossRef] [Green Version]
  35. Jung, U.; Choi, H. Active echo signals and image optimization techniques via software filter correction of ultrasound system. Appl. Acoust. 2022, 188, 108519. [Google Scholar] [CrossRef]
Figure 1. The BCI layer model compared to OSI network model.
Figure 1. The BCI layer model compared to OSI network model.
Sensors 22 05860 g001
Figure 2. Overall collaboration between QR-EFA and EFA.
Figure 2. Overall collaboration between QR-EFA and EFA.
Sensors 22 05860 g002
Figure 3. The EFA algorithm procedure.
Figure 3. The EFA algorithm procedure.
Sensors 22 05860 g003
Figure 4. Data analysis on the viewpoint direction.
Figure 4. Data analysis on the viewpoint direction.
Sensors 22 05860 g004
Figure 5. Overall flowchart for QR-EFA (solid line) including EFA (dashed line). Cross reference in the above data flowchart for step 1~11.
Figure 5. Overall flowchart for QR-EFA (solid line) including EFA (dashed line). Cross reference in the above data flowchart for step 1~11.
Sensors 22 05860 g005
Figure 6. The CNN for neuro images in QR-EFA.
Figure 6. The CNN for neuro images in QR-EFA.
Sensors 22 05860 g006
Figure 7. An original eigenface images of subjects for training data for 4 classes: left, right, feet, and tongue after whitening (12 × 10: = 30 × 4) in C3D3a_4C.
Figure 7. An original eigenface images of subjects for training data for 4 classes: left, right, feet, and tongue after whitening (12 × 10: = 30 × 4) in C3D3a_4C.
Sensors 22 05860 g007
Figure 8. Some selected sample QR training images of data augmentation for the class of (a) left and (b) tongue (12 × 10) in C3D3a_4C.
Figure 8. Some selected sample QR training images of data augmentation for the class of (a) left and (b) tongue (12 × 10) in C3D3a_4C.
Sensors 22 05860 g008
Figure 9. Primitive raw images for EEG data signal for the selected first and last trials (80 × 60) in C3D3a_4C.
Figure 9. Primitive raw images for EEG data signal for the selected first and last trials (80 × 60) in C3D3a_4C.
Sensors 22 05860 g009
Figure 10. Training QR code realization of subject 1 at trial 1 for 4 classes: left, right, feet, and tongue (12 × 10) in C3D3a_4C.
Figure 10. Training QR code realization of subject 1 at trial 1 for 4 classes: left, right, feet, and tongue (12 × 10) in C3D3a_4C.
Sensors 22 05860 g010
Figure 11. Data graph of the CNN cost function evolution vs. number of epochs in C3D3a_4C.
Figure 11. Data graph of the CNN cost function evolution vs. number of epochs in C3D3a_4C.
Sensors 22 05860 g011
Figure 12. An original eigenface images of subjects for training data for 4 classes: left, right, feet, and tongue after whitening (18 × 16: = 72 × 4) in C4D2a_4C.
Figure 12. An original eigenface images of subjects for training data for 4 classes: left, right, feet, and tongue after whitening (18 × 16: = 72 × 4) in C4D2a_4C.
Sensors 22 05860 g012
Figure 13. Some selected sample QR training images of data augmentation for the classes of (a) left and (b) tongue (18 × 16 in C4D2a_4C).
Figure 13. Some selected sample QR training images of data augmentation for the classes of (a) left and (b) tongue (18 × 16 in C4D2a_4C).
Sensors 22 05860 g013
Figure 14. Primitive raw images for EEG data signal for the selected first and last trials (50 × 50) in C4D2a_4C.
Figure 14. Primitive raw images for EEG data signal for the selected first and last trials (50 × 50) in C4D2a_4C.
Sensors 22 05860 g014
Figure 15. Training QR code realization of subject 1 at trial 1 for 4 classes: left, right, feet, and tongue (18 × 16) in C4D2a_4C.
Figure 15. Training QR code realization of subject 1 at trial 1 for 4 classes: left, right, feet, and tongue (18 × 16) in C4D2a_4C.
Sensors 22 05860 g015
Figure 16. Data graph of the CNN cost function evolution vs. number of epochs in C4D2a_4C.
Figure 16. Data graph of the CNN cost function evolution vs. number of epochs in C4D2a_4C.
Sensors 22 05860 g016
Table 1. The C3D3a_4C dataset composed of three subjects and predefined number of experimental trials.
Table 1. The C3D3a_4C dataset composed of three subjects and predefined number of experimental trials.
SubjectClass (# of Trials)
Left (L)Right (R)Foot (F)Tongue (T)
145454545
230303030
330303030
Table 2. EEG analysis accuracy for C3D3a_4C using QR-EFA method.
Table 2. EEG analysis accuracy for C3D3a_4C using QR-EFA method.
Subjects
A1A2A3Average
Accuracy
(%)
EFA_LDA,
two classes
52.2246.6763.3354.07
QR-EFA90.0093.3390.8391.11
Table 3. Accuracy results of classification for 10 simulations, with changing random seeds uniformly in C3D3a_2C.
Table 3. Accuracy results of classification for 10 simulations, with changing random seeds uniformly in C3D3a_2C.
# of SimulationsSubject (Failed/Trials)Success Rate
123
126/18015/12019/1200.857143
216/18010/1203/1200.930952
311/18013/1206/1200.928571
49/18017/1205/1200.926190
55/18012/1206/1200.945238
630/18014/12013/1200.864286
727/1809/12018/1200.871429
842/18032/12017/1200.783333
910/18015/1205/1200.928571
1018/1800/1202/1200.952381
Table 4. EEG analysis accuracy for C4D2a_4C, using QR-EFA method.
Table 4. EEG analysis accuracy for C4D2a_4C, using QR-EFA method.
Subjects
123456789Average
Accuracy
(%)
EFA_LDA,
two classes
53.4752.0855.5555.5554.1645.1358.3347.7251.3852.55
QR-EFA97.1597.9997.0598.3398.4797.8598.7897.3697.8197.87
Table 5. Accuracy results of classification for 10 simulations, with changing random seeds uniformly in C4D2a_2C.
Table 5. Accuracy results of classification for 10 simulations, with changing random seeds uniformly in C4D2a_2C.
# of SimulationsSubject (Failed/Trials)Success Rate
123456789
118/2886/28819/2889/2887/2886/2882/28813/2887/2880.966435
27/2889/2883/2885/2884/2888/2886/2886/28812/2880.976852
32/2882/28810/2881/2886/2885/2880/2884/2889/2880.984954
419/2888/28810/2887/2881/2889/2887/2888/2886/2880.971065
511/2885/28810/28813/28810/28812/2885/28810/2884/2880.969136
66/2884/2887/2885/2883/2882/2882/2888/2887/2880.983025
73/2885/2883/2881/2883/2887/2884/2886/2882/2880.986883
83/2884/2885/2880/2883/2884/2887/2885/2886/2880.985725
91/2888/2889/2883/2883/2884/2880/2889/2884/2880.984182
1012/2887/2889/2884/2884/2885/2882/2887/2886/2880.978395
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Choi, H.; Park, J.; Yang, Y.-M. A Novel Quick-Response Eigenface Analysis Scheme for Brain–Computer Interfaces. Sensors 2022, 22, 5860. https://doi.org/10.3390/s22155860

AMA Style

Choi H, Park J, Yang Y-M. A Novel Quick-Response Eigenface Analysis Scheme for Brain–Computer Interfaces. Sensors. 2022; 22(15):5860. https://doi.org/10.3390/s22155860

Chicago/Turabian Style

Choi, Hojong, Junghun Park, and Yeon-Mo Yang. 2022. "A Novel Quick-Response Eigenface Analysis Scheme for Brain–Computer Interfaces" Sensors 22, no. 15: 5860. https://doi.org/10.3390/s22155860

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop