Next Article in Journal
Application of an Ultra-Low-Cost Passive Sampler for Light-Absorbing Carbon in Mongolia
Previous Article in Journal
Walking Bout Detection for People Living in Long Residential Care: A Computationally Efficient Algorithm for a 3-Axis Accelerometer on the Lower Back
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Multiple Channels EEG Signal Classification Based on Hierarchical Extreme Learning Machine

Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8976; https://doi.org/10.3390/s23218976
Submission received: 17 October 2023 / Revised: 29 October 2023 / Accepted: 3 November 2023 / Published: 4 November 2023
(This article belongs to the Section Biosensors)

Abstract

:
The human brain can be seen as one of the most powerful processors in the world, and it has a very complex structure with different kinds of signals for monitoring organics, communicating to neurons, and reacting to different information, which allows large developments in observing human sleeping, revealing diseases, reflecting certain motivations of limbs, and other applications. Relative theory, algorithms, and applications also help us to build brain-computer interface (BCI) systems for different powerful functions. Therefore, we present a fast-reaction framework based on an extreme learning machine (ELM) with multiple layers for the ElectroEncephaloGram (EEG) signals classification in motor imagery, showing the advantages in both accuracy of classification and training speed compared with conventional machine learning methods. The experiments are performed on software with the dataset of BCI Competition II with fast training time and high accuracy. The final average results show an accuracy of 93.90% as well as a reduction of 75% of the training time as compared to conventional deep learning and machine learning algorithms for EEG signal classification, also showing its prospects of the improvement of the performance of the BCI system.

1. Introduction

The human brain can be seen as a powerful and comprehensive system responsible for processing, monitoring all behaviors, and reacting to the outside through neural activities, and it has various research aspects in the fields of biology and neuroscience. To bridge the gaps between the human brain and electronic devices, including computers, the brain–computer interface (BCI) system has shown its performance by decoding and processing the signals in the human brain. Therefore, one of the most essential tasks of developing a BCI system is to discover and decipher the various signals generated by the brain and establish their relationship with the corresponding behaviors. However, the collection of the signals in the human brain may be different. With the development of this area, among the signal collection methods, the ElectroEncephaloGram (EEG) signal is widely utilized due to its safety in collection and high resolution in recording, which are its advantages over previous methods. It mainly involves detecting electrical activity in the brain using small metal discs (work as electrodes) attached to the scalp, measuring voltage fluctuations resulting from ionic current within the brain’s neurons [1]. Basically, EEG signals are often divided into different types by frequency, and researchers have already identified and explored various applications and uses for these EEG frequency bands [2]. As summarized in Table 1, five waves correspond to five ranges of frequencies in EEG signals [3].
As for the functions and applications, EEG signals are very powerful tools as a standard in medical and diagnostic research [6], building communication system [7], hand movement direction [8], and driver fatigue detection [9]. To analyze certain EEG signals, big data is growing in the field of healthcare [10], and relative machine learning algorithms improve the efficiency in analysis. Among the applications of EEG signal, motor imagery is the mental execution of a movement without any overt movement or without any peripheral activation [11]. These activities can be reflected by the EEG signal after certain processing.
To realize these functions, BCI systems serve as a connection between the human brain and devices in brain science studies [12]. This technology enables users to control certain electronic devices through the bio-signals generated from brain activities, where the EEG signals or other bio-signals act as the media. About 100 years ago, the relationship between signals and behaviors was first identified by Hans Berger in 1924 when he recorded human brain activity by using the EEG signal. Firstly, Berger could identify the oscillatory activities, such as the alpha wave. Over the years, BCIs have primarily focused on different aspects, such as researching, assisting, or repairing human cognitive or sensory-motor functions [13]. Presently, BCI systems have found broad applications in areas such as sleep monitoring [14], disease detection [15], motor imagery [16], visual realization control, and rehabilitation [17]. Among these applications, motor imagery EEG signals are usually collected from the brain motor cortex area when a person imagines performing specific movements [18]. These EEG signals are typically found in the alpha and central beta frequency bands and can be used to control electronic devices, such as moving a computer’s cursor, operating a wheelchair, or manipulating a robotic arm. Various acquisition techniques can capture motor imagery EEG signals.
To collect EEG signals, this has become easy in recent days, and general EEG devices usually use electrodes, detecting the brain activities by allocating them on the subjects’ scalps consisting of amplifiers, filters, and an analog-to-digital converter. There are portable EEG machines, wireless EEG machines, noninvasive scalp EEG, and other types of machines for specific applications. However, among the BCI applications using EEG signals, it is more difficult to connect the processing and recognition of motor imagery signals and control commands for the computer system than collecting the signals. It may take a lot of time to detect and decode with experiments. Generally, in machine learning algorithms for signal processing, the whole analysis procedure can be separated into two main parts: feature extraction of the signals and the final classifying of the signals. As for feature extraction methods, currently, there are many different algorithms and methods that have been applied including band power (BP) analysis [19], power spectral density (PSD) values [20], and Principal Component Analysis (PCA) [21]. Commonly used classification techniques for interpreting the results include support vector machines (SVM) [22], neural networks-based machine learning methods [23], K-Nearest Neighbors (K-NN) [24], and Naive Bayes [25]. This paper proposes a method that utilizes independent component analysis (ICA) and hierarchical extreme learning machine (H-ELM), combined for motor imagery recognition with the classification of EEG signals. The following are the main contributions of this work:
  • With the powerful learning ability of H-ELM, all-channel signals are processed rather than manually select specific channels for particular purposes, which may be more suitable for different brain–computer interface (BCI) systems of multi-function applications.
  • Optimize and apply hierarchical-ELM (H-ELM) on EEG signal classification and this method can be extracted to all 1-D signals.
  • The results show an accuracy of 93.9% for the classification in motor imagery and an improvement in reducing 75% of the training time of conventional machine learning algorithms.
The following content is proposed as follows: Section 2 illustrates the materials and dataset in this work; Section 3 shows the whole methodology, including the theory, algorithms, and designs in detail; Section 4 presents and compares the results with analysis; and Section 5 concludes the work with future works.

2. Materials

The dataset applied and processed in the experiments of this work is offered by the BCI Competition with its Dataset Ia [26] for motor imagery classification based on a 6-channel EEG signal. For applications in the real world, there are different types of EEG signal-collecting devices with different accuracy, channels, and functions. For example, in this work, the data are collected and saved in a 6-channel EEG signal as shown in Figure 1, and for others, it may be saved in a 128-channel EEG. For different applications, some channels may need to be abandoned by the professional for higher efficiency and accuracy. However, in this work, we design a system processing on all channel signals for further applications to make it more general to be used on the BCI system. This can help it be used by different people to avoid further selections. All channels of signals are extended into one sequence for comprehensive analysis. The data can be easily transformed into EEGLab or other methods.
According to the description of the dataset, the data were collected by PsyLab EEG8 amplifier. The A/D-converter utilized was the Computer Boards PCIM-DAS1602/16 bit, with the amplitude of +/−1000 μ V for the recorded data. The sampling rate employed during the recording was 256 samples per second (S/s). For the positions of electrodes, the EEG data were collected by 6-channel electrodes with the location set as Channel 1, represented by the A1-Cz position in the 10/20 system, with A1 indicating the left mastoid. Channel 2 corresponded to the A2-Cz position. Channel 3 was located 2 cm frontal to C3, while Channel 4 was positioned 2 cm parietal to C3. Similarly, Channel 5 was situated 2 cm frontal to C4, and Channel 6 was placed 2 cm parietal to C4 [26]. As for details, the total location is shown in shown in Figure 1.
For the details of the experiments of the dataset, according to the demands, each subject intended to manipulate a computer’s cursor on the screen by performing upward and downward movements, during which the machines recorded all of their cortical potentials. Throughout the recording session, the participant received visual feedback on their potentials. Finally, the positive activities of cortical resulted in moving down the cursor on the screen, while negative activities led to moving down. A duration of about 6 s was recorded on each trail. The sampling rate of this experiment is 256 Hz, and the length of the recording is 3.5 s. Finally, 896 samples were recorded for each channel of each trial. After the experiments, 268 trials were recorded as the training set with the labels, and 293 trials were recorded for testing [26].

3. Methodology

An overview of the design flow is shown in Figure 2. To make the signals more general in the processing, the signals are sampled under a certain frequency and recorded into digital signals. Then, ICA is utilized for feature extraction noise reduction and separates the data of 6 channels for the independent analysis. To improve the accuracy in certain applications, another ELM of a single layer is used for channel selections of relative and relative channels in different applications under supervision. Then, the selected channels of EEG signals are sent to the training module of H-ELM, in which the hidden nodes are randomly set and are not changed throughout the whole process. Finally, another single-layer ELM works for final decision making to decide whether there is motor imagery. This work can be divided into 2 parts: pre-processing and training. As shown in Figure 2, the two subfigures in the first column show the pre-processing, and the others show the training and final decisions. The details of the input and output in each step are shown as follows.
  • Input: Collected EEG signal
    Transform collected EEG signals into digital formats.
    Extend all channels of signals into a sequence
    Output: EEG Signal sequence in digital formats
  • Input: EEG signal sequence
    Set the parameters in EEGLab.
    Apply ICA on EEG signals.
    Reduce the noise of signals.
    Output: Processed EEG signal after ICA
  • Input: EEG signal after processed by ICA
    Set the parameters of single layer ELM.
    Decide whether channels are relative or not.
    Output: Processed EEG signal after selection
  • Input: Relative EEG signal
    Set certain parameters in training such as iterative times and calculation accuracy.
    Train the H-ELM with training data layer by layer
    Calculate the final parameter β of each layer
    Calculate the final training results.
    Record the accuracy and time in both the training and testing stages.
    Output: Parameters of the last layer
In this work, we can achieve classification and decision making, using the theory and features of H-ELM to process the EEG signals. As for the core component, H-ELM instead of single-layer ELM represents a multi-layer neural network, with enhanced performance in terms of accuracy due to its tighter and more flexible structure, multiple layers, and more number of neurons. At the same time, other relative steps also help improve the performance of this work to make it more feasible for the BCI system. So, the selection and parameters of EEG signals, the usage of ICA, and the theory of ELM are introduced in this section.

3.1. EEG Signal Pre-Processing

Firstly, the data used in this work are reshaped into a certain format for further processing by the feature extraction module and channel selection module, to improve the accuracy in training. As for the raw data from the dataset, all of the reactions of the subjects were recorded in a matrix as a generalized EEG signal in voltages. As an example, the size of this matrix of the training set is 268 × 5377 , where the first entry of each column records the class of each trial in 0 or 1, illustrating whether there is motor imagery or not, and the rest of the columns of this matrix contain all 6 channel EEG signals in a row. Then, the label of each subject is eliminated, and the remaining 5376 entries are reshaped into 6 channels according to the description of the dataset for the further selection of relative channels. After the pre-processing of EEG signals, all the data are sent to the feature extraction module of ICA for noise reduction.

3.2. ICA for Feature Extraction

ICA is a very important tool in statistics and machine learning. It is widely used to separate the mixed signals into sub-components or independent ones. Sometimes, the signals collected from different electrodes may be mixed, making it hard for further processing. The signal may sometimes be allocated into one channel in the time sequence, leading to computing errors. It has been widely used for optical imaging of neurons, face recognition, communications, and removing artifacts from EEG signals.
Generally, the definition of ICA in the mathematical model, considering a set of observed signals x = [ x 1 , x 2 , , x n ] , aims to find a linear transformation represented by a matrix
A such that the source signals of estimation s = [ s 1 , s 2 , , s n ] can be calculated by [27]:
s = Ax
where s and x represent n-dimensional column vectors, and the matrix A represents the mixing matrix.
Therefore, the objective of ICA is to calculate the inversion of the mixing matrix W . Then, we can obtain the estimated source signals by:
s = Wx
where W = A 1 .
Among the machine learning algorithms, many results are generated based on statistics analysis for the estimation. Similarly, the ICA algorithm aims to maximize the statistical independence of the estimated source signals by updating the elements of the mixing matrix through iterative calculations.
Based on the theory of ICA, it is employed to extract the main features related to raw EEG signals’ motor imagery to figure out each trial’s features and reduce the noise of raw data as the feature extraction method. This will help simplify the problems in the training process and improve the training speed.

3.3. ELM Theory

As for the core algorithm and its theory in this work, Huang et al. first presented the concept and theory of ELM in 2004. This algorithm and network largely help reduce the training time of networks and improve the performance [28]. Previously, to obtain the results from networks, back-propagation (BP) [19] was widely used by randomly setting the parameters in hidden layers and using the iterative algorithm to update, costing a large number of resources and time. Certain applications may require the network to finish the training process quickly and put it into use in a short time. And the robustness of the network and the frequency of upgrading are also very important. So, ELM shows its advantages in this area because it randomly generates the bias and weights, and once fixed, they do not need to be changed. Then, the results are calculated by the least-square algorithm, making the process more general for all kinds of platforms and coding languages. With less training time, it improves the flexibility for specific applications in realizing frequent training with the latest dataset. The general structure of ELM is presented in Figure 3. For its general calculation process, X is defined as the training set. Here, several N labeled pairs ( x i , y i ) are set. x i R denotes the ith input vector, and y i R represents the target. Overall, to describe the ELM function, f ( x ) is:
f ( x ) = i = 1 M w j a ( r j x + b j )
For each layer of ELM, there are M neurons of the input layer connecting to H neurons in the hidden layer with the weights represented by r j R , j = 1 , , H . Each hidden neuron, denoted by j, incorporates a bias term b j and a nonlinear activation function a ( ) . The hidden layer is further connected to the output neurons through a vector defined as w R . As for the core feature of ELM, the pairs r j , b j in Equation (1) are randomly defined and not to be tuned in training. Then, let H also represent the matrix of activation, where h i j H ( i = 1 , , N ; j = 1 , , H ) represents the jth hidden neuron for the ith input pattern activation value. The relationship is h i j = a ( r j x i + b j ) , and the detailed results are shown in Equation (2):
H = H ( w 1 , , w n ˜ , b 1 , , b n ˜ , x 1 , , x n ˜ )
H = a ( w i x ˙ j + b i ) a ( w N ˜ x 1 ˙ + b N ˜ ) a ( w i x ˙ N + b i ) a ( w N ˜ x N ˙ + b N ˜ )
Therefore, ELM training can be simplified in minimizing:
min ( w , b ) H w Y 2
Overall, training an ELM involves a straightforward procedure that can be divided into the following steps:
  • Set the parameters pairs r j , b j for the hidden layer randomly;
  • Calculate H as the activation matrix;
  • Calculate the output weights by Equation (3);
  • Finally, the core computation can be expressed as shown in Equation (5).

3.4. Basic Structure of H-ELM

For the disadvantages of conventional ELMs with a single layer, the learning ability may be limited, even if the layer’s scale is very large. The reason is that ELM involves matrix inversion, and the value of the entries of hidden layer varies from −1 to 1. When a certain entry has a very small value, the inversion calculation may lead to a very large output as an outlier, and this will cause inaccurate results in training. To keep the same total number of hidden nodes, separating a large layer into more small layers can not only improve the accuracy but it can also improve the flexibility in setting the layers. Therefore, multi-layer ELM is required for higher accuracy and stable structures.
H-ELM [29] is a framework with multiple layers designed as an effective and efficient Multilayer Perceptron (MLP); with a multi-layer structure, certain requirements including accuracy can be fulfilled compared to conventional ELM. It has shown advantages in the accuracy of 99.13% in the conventional MNIST dataset, where some of the deep learning methods, such as DBM and DBN, only realize accuracies of 99.05% and 98.87%, but deep learning methods cost nearly 200 times to 600 times the consumption in training time [29]. H-ELM separates the machine learning process into two steps: unsupervised multi-layer training, and supervised feature classification and final decision making. To improve the performance in training accuracy and speed, ELM sparse auto-encoder is also employed in H-ELM [30]. In the ELM sparse auto-encoder approach, each layer in the stack architecture operates as an independent subsystem or submodule. Additionally, a random matrix is employed to disperse the high-level features of the networks during feature classification. Finally, the original single-layer ELM is applied for the ultimate decision-making process. The general structure of H-ELM is shown in Figure 4.
In order to establish connections between each single-layer ELM, similar to conventional multi-layer models, the output and weights of the last layer are directly transmitted as input to the subsequent layer. Consequently, an N-layer unsupervised learning process is conducted to obtain high-level sparse features. The calculation of the output for each hidden layer can be expressed as:
H i = g ( H i 1 β ˙ )
where H i is the result of the ith layer ( i [ 1 , K ] ), and H i 1 represents the result of the ( i 1 ) th layer respectively. Then g ( · ) represents the activation function, and β denotes the output weights. Each hidden layer is seen as an independent single-layer ELM in the analysis. The core feature of the ELM structure is that once the hidden layer from the previous layer is randomly set, the weights or parameters of the current hidden layer remain fixed and do not necessitate tuning. This distinguishes H-ELM from deep learning algorithms that utilize backpropagation (BP) calculations and give the high training speed of ELM. Therefore, it reduces the process of setting and revising the hidden layers and improves the overall training speed or time consumption in applications.
As shown in Figure 4, the calculation results of the ith layer can be regarded as the input features. In the subsequent step, when these features are utilized in applications, they are randomly set and used to compute the results. This exemplifies the high capability of each layer of ELM in various applications, making it a versatile approach. Similarly, H-ELM can also be summarized in the following steps:
  • Given the training set N = ( x i , t i ) | x i R n , t i R m , i = 1 , , N with the activation function g ( x ) , and hidden neuron number N ^ .
  • Calculate the hidden layer output matrix H by
    min ( w , b ) H w Y 2 + λ w 2
  • Obtain β by β = H T as the output.
  • Follow the principle of multi-layer connection to connect each layer through H i = g ( H i 1 β ˙ ) .

3.5. ELM Sparse Auto-Encoder

For a further introduction of the ELM sparse auto-encoder, it transfers certain calculations of input data into approximation. In theory, it aims to achieve the equation:
h θ ( x ) x
where θ = A , b and A represents the hidden weight, and b denotes the bias of the framework. Moreover, the auto-encoder helps reduce errors. Generally, L 1 optimization is used for establishing the ELM auto-encoder, and the singular values of L 2 n o r m are calculated for high-level features optimization. The optimized mathematics model of this auto-encoder can be written as follows:
O β = argmin β { | | H β X | | 2 + | | β | | 1 }
Similarly, X denotes the input, H represents the randomly mapped output, and β represents the weight. Furthermore, X represents the original data for ELM, and H is randomly generalized, and once fixed does not need to be tuned. The algorithm for solving H-ELM through a sparse autoencoder is shown as follows:
  • Input: data X = x i where i = 1 , , N , output matrix of hidden nodes H = h i where i = 1 , , L .
  • Determine the Lipschitz constant ( γ ) of the gradient of a smooth convex function p .
  • Calculate β iteratively, y 1 = β 0 , t 1 = 1 , i 1 .
  • β j = s γ ( y j ) , where s γ is given by
    s γ = argmin β γ 2 | | β ( β i 1 1 γ p ( β ( i 1 ) ) | | 2 + q ( β ) γ 2 | | β ( β i 1 1 γ p ( β i 1 ) | | 2 .
  • t i + 1 = 1 + 1 + 4 t i 2 2 .
  • y i + 1 = β i + ( t i 1 t i + 1 ) ( β i β i 1 ) .
  • Output weight β .
Using the auto-encoder enhances the computational efficiency and application performance. In the H-ELM framework, the learning process in previous layers is designed for training, while the final layer is responsible for decision making. In this study, after the experiments under various conditions, a three-layer H-ELM is employed in applications with the sparse ELM autoencoder to enhance efficiency. Details of the experiment result with analysis are shown in Section 4.

4. Experiment Results and Discussion

In this work, all experiments are performed on a PC with the CPU of Intel(R) Core(TM) i5-6400 CPU @2.70 GHz and 24 GB 2133 MHz RAM. The programs are performed by Matlab 2020a. The data from each subject are labeled as the description of the dataset from [31]. Among the datasets used in our experiments, 268 trials from the training set with labels are used for training, and the other 293 trials are also labeled for testing. For the data from each subject, −1 or 1 are labeled to represent the motor imagery in the test. The results of the experiments are shown in detail as follows.
Firstly, the number of layers should be selected according to this dataset. After primary experiments, under the same conditions, 3-layer H-ELM can realize a balance between the accuracy of classification and time consumption in training as shown in Table 2. To explain the results, the H-ELM may have overfittings in training, and the 3-layer framework can realize the classification. Different numbers of neurons are selected for the EEG signal classification to test and find the most suitable results in balancing the accuracy and training time. In total, the trends in the training process and the testing process are shown in Figure 5 and Figure 6. From the figures, the training accuracy arises 100% if more than 700 neurons are set for training. Therefore, only more than 700 hidden nodes should be considered in this application; otherwise, the wrong training may also lead to wrong testing and final decisions. The training time increases nearly linearly in the training part, so selecting the number of hidden nodes should consider the results in both aspects.
For the testing part, the testing time is also increasing linearly, and it may affect the applications. Therefore, the time in testing may also be influenced by the dataset. However, the trend of testing accuracy is highly related to the number of hidden nodes, which increases when the number ranges from 300 to 1000, and then decreases after the number of hidden nodes is larger than 1000, but it will be nearly unchanged. Besides the overfitting, which is mainly because when the ELM has a very large scale in the hidden matrix, it may lead to unstable results in calculations since inversion calculations are involved in ELM as mentioned in Section 3.4, although the multi-layer structure has separated the errors to more layers. Therefore, the number 1000 is selected as the best performance in this application of EEG signal classification, although it varies depending on the scale of the selected dataset.
Figure 5 and Figure 6 show that H-ELM in EEG signal classification can finish the training within only 2.5 s for this dataset, and the testing can also be finished with 0.04 s. At the same time, the framework shows an accuracy of 93.9% in EEG signal classification based on all channel signals. The results demonstrate the advantages of the proposed approach in terms of both accuracy and training time when compared to conventional machine learning methods and single-layer ELM. The comparison results mainly focus on time consumption in training and testing accuracy, and they are shown in Table 3. From the comparisons with other methods such as K-NN, it has relatively high accuracy, especially in applications with large datasets, but it requires large time consumption. For certain applications that may require fast reactions, these methods may not be able to realize real-time reactions on certain occasions. For certain application areas, such as clinical, the frameworks may need frequent updates to maintain high accuracy. So, the proposed method balances the advantages in training time and accuracy, and it has a very small structure and is easy to be used and applied on hardware [32]. In addition, these experiments do not require the manual selection of the channels, which is more general in application, especially implemented on large systems, such as BCI systems.

5. Conclusions

This paper shows the application of H-ELM on the classification of all channel EEG signals with very high training and testing speed compared with other methods, such as improving nearly four times with k-NN for the same application and dataset. The results from this work also show the application prospects in other 1-D signal processing and classifications. Additionally, it can realize an average accuracy of 93.9%. To further improve the accuracy, fast feature extraction methods can be applied to this network, and hidden nodes can also be set for certain applications for proper performance, under the condition that the device can react to the change of the EEG signal from the electrodes in a short time. From these results, H-ELM shows potential for hardware architecture due to its simple structure and low resource occupation. However, ELM may lead to unstable results in the training process because when the scale of the hidden matrix is large, the inversion calculation may have an outlier value in the matrix and cause wrong results. Therefore, the training and testing need to be optimized to obtain the average results. Currently, it shows advantages in some work, such as direction-of-arrival (DOA) measurement of mice EEG [36], emotion detection [37], and building a more flexible BCI system [38]. Therefore, for future research directions, other relative methods, including feature extractions and noise reductions, will be applied, and it can be applied on hardware platforms to realize low-power and real-time measurements and analyses with high and efficient performance.

Author Contributions

Conceptualization, S.L. and R.C.C.C.; methodology, S.L.; software, S.L.; validation, S.L. and R.C.C.C.; formal analysis, S.L.; investigation, S.L.; resources, S.L. and R.C.C.C.; data S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L. and R.C.C.C.; visualization, S.L.; supervision, R.C.C.C.; project administration, R.C.C.C.; funding acquisition, R.C.C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable. The authors declare that the work described has not involved experimentation on humans or animals.

Data Availability Statement

This paper uses the public dataset of BCI Competition II [26].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Niedermeyer, E.; da Silva, F.L. Electroencephalography: Basic Principles, Clinical Applications, and Related Fields; Lippincott Williams & Wilkins: Philadelphia, PA, USA, 2005. [Google Scholar]
  2. Babiloni, C.; Barry, R.J.; Başar, E.; Blinowska, K.J.; Cichocki, A.; Drinkenburg, W.H.; Klimesch, W.; Knight, R.T.; Lopes da Silva, F.; Nunez, P.; et al. International Federation of Clinical Neurophysiology (IFCN)—EEG research workgroup: Recommendations on frequency and topographic analysis of resting state EEG rhythms. Part 1: Applications in clinical research studies. Clin. Neurophysiol. 2020, 131, 285–307. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, J.; Wang, M. Review of the emotional feature extraction and classification using EEG signals. Cogn. Robot. 2021, 1, 29–40. [Google Scholar] [CrossRef]
  4. Kirmizi-Alsan, E.; Bayraktaroglu, Z.; Gurvit, H.; Keskin, Y.H.; Emre, M.; Demiralp, T. Comparative analysis of event-related potentials during Go/NoGo and CPT: Decomposition of electrophysiological markers of response inhibition and sustained attention. Brain Res. 2006, 1104, 114–128. [Google Scholar] [CrossRef] [PubMed]
  5. Chaddad, A.; Wu, Y.; Kateb, R.; Bouridane, A. Electroencephalography Signal Processing: A Comprehensive Review and Analysis of Methods and Techniques. Sensors 2023, 23, 6434. [Google Scholar] [CrossRef] [PubMed]
  6. Mamitsuka, H. Selecting features in microarray classification using ROC curves. Pattern Recognit. 2006, 39, 2393–2404. [Google Scholar] [CrossRef]
  7. Roman-Gonzalez, A. EEG signal processing for BCI applications. In Human–Computer Systems Interaction: Backgrounds and Applications 2; Springer: Berlin/Heidelberg, Germany, 2012; pp. 571–591. [Google Scholar]
  8. Waldert, S.; Preissl, H.; Demandt, E.; Braun, C.; Birbaumer, N.; Aertsen, A.; Mehring, C. Hand movement direction decoded from MEG and EEG. J. Neurosci. 2008, 28, 1000–1008. [Google Scholar] [CrossRef] [PubMed]
  9. Sheykhivand, S.; Rezaii, T.Y.; Meshgini, S.; Makoui, S.; Farzamnia, A. Developing a deep neural network for driver fatigue detection using EEG signals based on compressed sensing. Sustainability 2022, 14, 2941. [Google Scholar] [CrossRef]
  10. Gurve, D.; Delisle-Rodriguez, D.; Bastos-Filho, T.; Krishnan, S. Trends in compressive sensing for EEG signal processing applications. Sensors 2020, 20, 3703. [Google Scholar] [CrossRef]
  11. Mulder, T. Motor imagery and action observation: Cognitive tools for rehabilitation. J. Neural Transm. 2007, 114, 1265–1278. [Google Scholar] [CrossRef]
  12. Guest Editorial The Third International Meeting on Brain-Computer Interface Technology: Making a Difference. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 126–127. [CrossRef]
  13. Krucoff, M.O.; Rahimpour, S.; Slutzky, M.; Edgerton, V.R.; Turner, D. Enhancing Nervous System Recovery through Neurobiologics, Neural Interface Training, and Neurorehabilitation. Front. Neurosci. 2016, 10, 584. [Google Scholar] [CrossRef] [PubMed]
  14. Nakamura, T.; Goverdovsky, V.; Morrell, M.J.; Mandic, D.P. Automatic Sleep Monitoring Using Ear-EEG. IEEE J. Transl. Eng. Health Med. 2017, 5, 2800108. [Google Scholar] [CrossRef] [PubMed]
  15. Kang, Y.; Escudero, J.; Shin, D.; Ifeachor, E.; Marmarelis, V. Principal Dynamic Mode Analysis of EEG Data for Assisting the Diagnosis of Alzheimer’s Disease. IEEE J. Transl. Eng. Health Med. 2015, 3, 1800110. [Google Scholar] [CrossRef] [PubMed]
  16. Razzak, I.; Hameed, I.A.; Xu, G. Robust Sparse Representation and Multiclass Support Matrix Machines for the Classification of Motor Imagery EEG Signals. IEEE J. Transl. Eng. Health Med. 2019, 7, 2000508. [Google Scholar] [CrossRef]
  17. Vilela, M.; Hochberg, L.R. Chapter 8—Applications of brain-computer interfaces to the control of robotic and prosthetic arms. In Handbook of Clinical Neurology; Ramsey, N.F., del R. Millán, J., Eds.; Elsevier: Amsterdam, The Netherlands, 2020; Volume 168, pp. 87–99. [Google Scholar] [CrossRef]
  18. Pfurtscheller, G.; Neuper, C.; Schlogl, A.; Lugger, K. Separability of EEG signals recorded during right and left motor imagery using adaptive autoregressive parameters. IEEE Trans. Rehabil. Eng. 1998, 6, 316–325. [Google Scholar] [CrossRef] [PubMed]
  19. Pfurtscheller, G.; Neuper, C.; Flotzinger, D.; Pregenzer, M. EEG-based discrimination between imagination of right and left hand movement. Electroencephalogr. Clin. Neurophysiol. 1997, 103, 642–651. [Google Scholar] [CrossRef]
  20. Chiappa, S.; Bengio, S. HMM and IOHMM Modeling of EEG Rhythms for Asynchronous BCI Systems; Technical Report; IDIAP: Martigny, Switzerland, 2003. [Google Scholar]
  21. Lekamalage, L.; Kasun, C.; Zhou, H.; Huang, G.B.; Cambria, E. Representational learning with ELMs for big data. IEEE Intell. Syst. 2013, 11, 31–34. [Google Scholar]
  22. Camps-Valls, G.; Bruzzone, L. Kernel-based methods for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1351–1362. [Google Scholar] [CrossRef]
  23. Subasi, A.; Ercelebi, E. Classification of EEG signals using neural network and logistic regression. Comput. Methods Programs Biomed. 2005, 78, 87–99. [Google Scholar] [CrossRef]
  24. Li, X.; Hu, B.; Shen, J.; Xu, T.; Retcliffe, M. Mild depression detection of college students: An EEG-based solution with free viewing tasks. J. Med. Syst. 2015, 39, 187. [Google Scholar] [CrossRef]
  25. LeVan, P.; Urrestarazu, E.; Gotman, J. A system for automatic artifact removal in ictal scalp EEG based on independent component analysis and Bayesian classification. Clin. Neurophysiol. 2006, 117, 912–927. [Google Scholar] [CrossRef]
  26. Birbaumer, N.; Flor, H.; Ghanayim, N.; Hinterberger, T.; Iverson, I.; Taub, E.; Kotchoubey, B.; Kbler, A.; Perelmouter, J. A brain-controlled spelling device for the completely paralyzed. Natureä398 2001, 297–298. [Google Scholar]
  27. Hyvärinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar] [CrossRef] [PubMed]
  28. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  29. Tang, J.; Deng, C.; Huang, G. Extreme Learning Machine for Multilayer Perceptron. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 809–821. [Google Scholar] [CrossRef] [PubMed]
  30. Kasun, L.L.C.; Yang, Y.; Huang, G.B.; Zhang, Z. Dimension reduction with extreme learning machine. IEEE Trans. Image Process. 2016, 25, 3906–3918. [Google Scholar] [CrossRef]
  31. Dornhege, G.; Blankertz, B.; Müller, K.R. BCI Competition II, Dataset Ia. 2003. Available online: https://www.bbci.de/competition/ii/results/dornhege_ia_desc.pdf (accessed on 31 August 2019).
  32. Decherchi, S.; Gastaldo, P.; Leoncini, A.; Zunino, R. Efficient digital implementation of extreme learning machines for classification. IEEE Trans. Circuits Syst. II Express Briefs 2012, 59, 496–500. [Google Scholar] [CrossRef]
  33. Wu, T.; Yan, G.-Z.; Yang, B.-Y.; Hong, S. EEG feature extraction based on wavelet packet decomposition for brain computer interface. Measurement 2008, 41, 618–625. [Google Scholar]
  34. Wang, B.; Jun, L.; Bai, J.; Peng, L.; Li, G.; Li, Y. EEG recognition based on multiple types of information by using wavelet packet transform and neural networks. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 17–18 January 2006; pp. 5377–5380. [Google Scholar]
  35. Sun, S.; Zhang, C. Assessing features for electroencephalographic signal categorization. In Proceedings of the Proceedings. (ICASSP’05). IEEE International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, PA, USA, 23 March 2005; Volume 5, pp. v/417–v/420. [Google Scholar]
  36. Chowdhury, M.H.; Eldaly, A.B.; Agadagba, S.; Cheung, R.C.C.; Chan, L.L.H. Machine Learning Based Hardware Architecture for DOA Measurement from Mice EEG. IEEE Trans. Biomed. Eng. 2022, 69, 314–324. [Google Scholar] [CrossRef]
  37. Gonzalez, H.A.; Muzaffar, S.; Yoo, J.; Elfadel, I.M. BioCNN: A Hardware Inference Engine for EEG-Based Emotion Detection. IEEE Access 2020, 8, 140896–140914. [Google Scholar] [CrossRef]
  38. Chiesi, M.; Guermandi, M.; Placati, S.; Scarselli, E.F.; Guerrieri, R. Creamino: A Cost-Effective, Open-Source EEG-Based BCI System. IEEE Trans. Biomed. Eng. 2019, 66, 900–909. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The 6-channel electrodes for EEG signals.
Figure 1. The 6-channel electrodes for EEG signals.
Sensors 23 08976 g001
Figure 2. Overview of the design of the whole process.
Figure 2. Overview of the design of the whole process.
Sensors 23 08976 g002
Figure 3. General structure of ELM.
Figure 3. General structure of ELM.
Sensors 23 08976 g003
Figure 4. General structure of H-ELM.
Figure 4. General structure of H-ELM.
Sensors 23 08976 g004
Figure 5. Trend of training accuracy and training time under different numbers of neurons.
Figure 5. Trend of training accuracy and training time under different numbers of neurons.
Sensors 23 08976 g005
Figure 6. Trend of testing accuracy and testing time under different numbers of neurons.
Figure 6. Trend of testing accuracy and testing time under different numbers of neurons.
Sensors 23 08976 g006
Table 1. Different kinds of EEG signals.
Table 1. Different kinds of EEG signals.
BandFrequencyOccasions
Delta0.5–4 HzMainly found in babies, adult slow-wave sleep, and some continuous-attention tasks [4]
Theta4–8 HzMainly found in young people and children, drowsiness in adults and teens with inhibition of elicited responses [4]
Alpha8–13 HzFound in relaxed, seemingly with the purpose of timing inhibitory activity in different locations across the brain.
Beta13–30 HzFound in active thinking, focus, and anxious
Gammalarger than 30 HzFound in advanced cognitive processing and the merging of sensory information [5].
Table 2. Relative results under different numbers of layers.
Table 2. Relative results under different numbers of layers.
LayersHidden Nodes of Each LayerTest Accuracy (Average)Training Time (Average)
370093.90%1.59 s
470093.25%2.51 s
Table 3. Accuracy and training time of different methods on BCI competition datasets.
Table 3. Accuracy and training time of different methods on BCI competition datasets.
MethodAccuracy in PercentTraining Time in Average
ELM87.500.90 s
DBM93.99around 100 s
k-NN [33]92.15around 10 s
Neural Network [34]91.47NA
Bayes [35]90.44NA
Proposed H-ELM93.902.60 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lyu, S.; Cheung, R.C.C. Efficient Multiple Channels EEG Signal Classification Based on Hierarchical Extreme Learning Machine. Sensors 2023, 23, 8976. https://doi.org/10.3390/s23218976

AMA Style

Lyu S, Cheung RCC. Efficient Multiple Channels EEG Signal Classification Based on Hierarchical Extreme Learning Machine. Sensors. 2023; 23(21):8976. https://doi.org/10.3390/s23218976

Chicago/Turabian Style

Lyu, Songyang, and Ray C. C. Cheung. 2023. "Efficient Multiple Channels EEG Signal Classification Based on Hierarchical Extreme Learning Machine" Sensors 23, no. 21: 8976. https://doi.org/10.3390/s23218976

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop