Next Article in Journal
Development of Impurity-Detection System for Tracked Rice Combine Harvester Based on DEM and Mask R-CNN
Next Article in Special Issue
Automatic Identification of Involuntary Muscle Activity in Subacute Patients with Upper Motor Neuron Lesion at Rest—A Validation Study
Previous Article in Journal
Digital-Twin-Assisted Edge-Computing Resource Allocation Based on the Whale Optimization Algorithm
Previous Article in Special Issue
Embodying Language through Gestures: Residuals of Motor Memories Modulate Motor Cortex Excitability during Abstract Words Comprehension
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Investigation of EEG-Based Biometric Identification Using State-of-the-Art Neural Architectures on a Real-Time Raspberry Pi-Based System

1
Department of Electrical Engineering and Computer Science, University of California, Irvine, CA 92697, USA
2
Northwood High School, Irvine, CA 92620, USA
3
Department of Computer Science, University of California, Irvine, CA 92697, USA
4
Department of Biomedical Engineering, University of California, Irvine, CA 92697, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(23), 9547; https://doi.org/10.3390/s22239547
Submission received: 17 October 2022 / Revised: 1 December 2022 / Accepted: 2 December 2022 / Published: 6 December 2022
(This article belongs to the Special Issue Biosignal Sensing Analysis (EEG, MEG, ECG, PPG))

Abstract

:
Despite the growing interest in the use of electroencephalogram (EEG) signals as a potential biometric for subject identification and the recent advances in the use of deep learning (DL) models to study neurological signals, such as electrocardiogram (ECG), electroencephalogram (EEG), electroretinogram (ERG), and electromyogram (EMG), there has been a lack of exploration in the use of state-of-the-art DL models for EEG-based subject identification tasks owing to the high variability in EEG features across sessions for an individual subject. In this paper, we explore the use of state-of-the-art DL models such as ResNet, Inception, and EEGNet to realize EEG-based biometrics on the BED dataset, which contains EEG recordings from 21 individuals. We obtain promising results with an accuracy of 63.21%, 70.18%, and 86.74% for Resnet, Inception, and EEGNet, respectively, while the previous best effort reported accuracy of 83.51%. We also demonstrate the capabilities of these models to perform EEG biometric tasks in real-time by developing a portable, low-cost, real-time Raspberry Pi-based system that integrates all the necessary steps of subject identification from the acquisition of the EEG signals to the prediction of identity while other existing systems incorporate only parts of the whole system.

1. Introduction

Biometrics is a well-valued approach to subject identification for protecting access to a source of information and providing identification without the need for a password or other physical methods [1]. It captures the physiological characteristics of a subject for identification and classification. Biometrics requires measurable physical or behavioral characteristics capable of satisfying the following considerations: universality, distinctiveness, collectability, circumvention, permanence, acceptability, and performance. There are different types of biometrics depending on the physiological characteristic, such as fingerprint, facial structure, iris, and voice, to the less popular ones such as EEG, ECG, deoxyribonucleic acid (DNA), gait, and ear shape [2,3,4]. EEG is an electrophysiological signal that reflects unique cognitive and neurological information of the person; thus, EEG possesses the potential to be used for robust biometrics [1,5,6]. EEG provides a high level of security since it is very difficult to reproduce the EEG patterns of a specific person. In addition, in the event that the person is subjected to a forced EEG reading against his/her will, the EEG biometric system may detect the person’s stress and deny access.
Deep learning-based models have achieved state-of-the-art results in a wide range of clinical applications [7] and biometric identification problems, ranging from phone authentication to bank security systems [8]. The use of DL models for biometric identification has been increasing in recent years. Within the field of EEG biometrics, DL models have been leveraged to improve the accuracy of EEG biometric identification systems.
Most Machine Learning (ML) projects are performed on ordinary computers or Graphics Processing Units (GPUs) because of the computing power they offer. However, since the purpose of this project is to develop a portable Internet of Things (IoT) system, it is not feasible to use common computers. In this work, we use Raspberry Pi as the core, which is compact, portable, and Wi-Fi capable. Raspberry Pi is widely used for portable IoT applications because of its versatility. The Raspberry Pi has been used in a number of projects for a variety of purposes, including monitoring the health status of patients [9], a security system [10], and a testing system device [11] to name a few. The Raspberry Pi has also been used for projects based on EEG signals. For example, in [12], EEG is used to control a car, and, in [13], it is used to monitor the depth of anesthesia. The purpose of this work is to develop a first approximation to a fully functional portable system for subject identification that uses trained DL models to process the signals.
This paper is organized as follows: The data description is presented in Section 2. In Section 3, we describe the software implementation for the subject identification using EEG signals. Section 4 shows the hardware implementation of the system. Section 5 reports the results. Finally, the discussion and conclusion will be drawn in Section 6.

2. Dataset Description

The BED (Biometric EEG Dataset) [14] is a dataset specifically designed to test EEG-based biometric approaches that use relatively inexpensive consumer-grade devices. The dataset includes EEG responses from 21 subjects to 12 different stimuli, which were broadly divided into four different types, namely affective stimuli, cognitive stimuli, visual evoked potentials, and resting-state. Each stimulus contains data across three different chronologically disjointed sessions. Fourteen-channel EEG signals containing the channels—AF3, F7, F3, FC5, T7, P7, O1, O2, P8, T8, FC6, F4, F8, and AF4 as shown in Figure 1—were collected at a sampling rate of 256 Hz. The BED dataset includes the raw EEG recordings with no preprocessing and the log files of the experimental procedure, in text format. The EEG recordings were segmented, structured, and annotated according to the presented stimuli, in Matlab format. The BED dataset also includes Autoregression Reflection Coefficients (ARRC), Mel-Frequency Cepstral Coefficients (MFCC), and Spectral (SPEC) features that were extracted from each EEG segment. In this work, however, only raw EEG recordings without any manually extracted features were used. Out of the 12 different stimuli, we only used EEG signals recorded during the Rest-Closed (RC) Stimulus since the action of “closing your eyes” is an easy and natural action and could be best replicated in a real-world scenario for biometric applications with ease. Having stimuli in the experimental procedure will require additional devices to present them to the individuals, which in turn will increase the complexity of the whole setup and practical use in the real world.
The BED dataset was simulated using our real-time Raspberry Pi-based system as a proof of concept to test the feasibility of this work for any real independent practical use. The simulated data acquisition in our Raspberry Pi-based system is performed by an analog-to-digital converter (ADC) that reads the analog output from the digital-to-analog converter (DAC) that receives the input from the original saved BED dataset. More details will be discussed in Section 4.

3. Software Implementations

In this section, we describe the data preprocessing steps and the deep learning models used. We also discuss how these models will be evaluated.

3.1. Data Processing

3.1.1. Preprocessing of EEG Signals

The preprocessing step of EEG signals is a crucial step in the DL pipeline because of its impact on the EEG analysis process. Without the preprocessing step, there may be noisy data and artifacts that can mask distinct features in the EEG signals. This can cause the model to have a harder time distinguishing between relevant EEG features, resulting in poorer performance of the model. In addition, we must pay attention to the quality of the preprocessing step as it can introduce unwanted artifacts if the early stages of the pipeline are not properly addressed. For example, although ordinary average referencing improves the signal-to-noise ratio, noisy channels which depend on the reference can contaminate the results.
Figure 2 shows the different steps of EEG signal preprocessing. The well-known preprocessing technique—PREP pipeline [15]—introduces specific important functionality for referencing the data, removing line noise, and detecting bad channels in order to deal with noisy channel-reference interactions. The PREP pipeline also removes artifacts such as muscle movement, jaw clenching, and eye blinking. The pipeline consists of various steps. First, the signal is filtered using a 1 Hz high pass filter followed by line noise removal using a notch filter at 60 Hz. In addition, finally, the signal is robustly referenced with respect to an estimate of the true mean reference, thereby enabling the detection of faulty channels. These channels are then interpolated relative to the same reference. We then use a lowpass filter of 50 Hz and divide the EEG data into overlapping epochs with an overlap rate of 90 percent. Finally, we standardize the EEG signals for each channel using StandardScaler. Figure 3 depicts an EEG epoch before and after preprocessing.

3.1.2. Deep Learning Implementation

The deep learning models used for subject identification are based on ResNet [16], an extension of the neural network into internal structures that add direct connections to the internal residual blocks to allow the gradient to flow directly through the lower layers; Inception [17], a convolutional neural network architecture that executes multiple operations with multiple filter sizes in parallel to avoid facing any compensation and allows the network to automatically extract relevant features from the time series; and EEGNet [18], a compact convolutional neural network that has been designed to build an EEG-specific model as it includes concepts and tools specific to EEG signals such as feature extraction and optimal spatial filtering to reduce the number of trainable parameters. The utilization of these three DL models was based on the fact that they are state-of-the-art models that have achieved good results for various other applications. For example, ResNet and InceptionTime were designed for Time Series Classification. EEGNet is a DL model that was designed for EEG-based Brain-Computer Interfaces. As a result, we wanted to examine the use of these models for the EEG Biometrics application.
Modifications for ResNet include adding additional residual blocks, to understand whether a more complex model that could extract complex features would perform better. For the Inception model, additional dropout layers and inception blocks were added. The activation function was changed from Rectified Linear Unit (ReLU) to Exponential Linear Unit (ELU) as the model was overfitting. For the EEGNet model, we fine-tuned the length of temporal convolution in the first layer and the number of channels after trial and error. Other modifications, such as adding GlobalAveragePooling2D layer, varying the dropout rate from 40 percent to 60 percent, and rearranging the order of layers, did not significantly improve the model performance.
In addition, within all our models, we added a callback function that reduces the learning rate based on the training loss. Specifically, we added the hyperparameter called patience, which is the number of epochs of non-decreasing loss values that the model runs before reducing the learning rate by half.
Because we want to check the feasibility of EEG-based Biometric identification over long periods of time, these presented models were trained using the first two weeks of the three chronologically disjoint sessions in the BED dataset. The third week of data was used for testing the trained model.
Once the models were trained, we had their hyperparameters, including learning rate, batch size, number of filters, kernel size, and number of epochs fine-tuned. The hyperparameter learning rate was set to 0.003 for the ResNet model and 0.009 for the Inception and EEGNet models. In addition, the number of epochs has been modified, 150 for Inception and 400 for ResNet. Figure 4, Figure 5 and Figure 6 briefly show the overall architectures of our modified Resnet-based, Inception-based, and EEGNet-based framework used in this work.

3.2. Model Evaluation

The model was evaluated using the third session of the three chronologically disjoint sessions provided in the dataset. The trained model would take one EEG segment from the testing dataset at a time, and output the prediction of which person it belongs to. Once all the EEG segments in the testing dataset have been predicted, we would use the confusion matrix to evaluate the model performance. The evaluation of the machine learning algorithms was carried out by comparing the most relevant indices for the prediction of the subjects [19]. The accuracy of the models calculates the ratio of correct predictions over the total number of instances evaluated. Precision is used to measure the positive patterns that are correctly predicted from the total predicted patterns in a positive class. Recall is used to measure the fraction of positive patterns that are correctly classified. F1 Score is calculated to represent the harmonic mean between recall and precision values. Finally, the Precison vs. Recall (P–R) curves are obtained to provide a graphical representation of the DL models’ performance.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 S c o r e = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
T P : True Positive F P : False Positive T N : True Negative F N : False Negative

4. Hardware Implementation

In this section, we describe the different parts of the system which are responsible for the generation and acquisition of EEG data and its processing for subject identification. In addition, finally, we present the system validation and its performance.

4.1. Data Acquisition

In this experimental work, data acquisition is simulated by using data from an existing database as mentioned in Section 2. Real-time EEG acquisition is in the scope of the future where we intend to implement the best deep learning model obtained from this study along with real-time signal acquisition for biometric application. Keeping in mind our goals for future work and existing time constraints, we decided to simulate analog EEG input in this work as mentioned later in this section.
The system consists of a Digital-to-Analog Converter (DAC) in charge of converting the stored data to analog signals mimicking real EEG acquisition scenarios [20]. The DAC is the 12-bit MCP4725 chip with an Inter-Integrated Circuit (I2C) communication bus [21]. Since the converter is a 12-bit converter, the range of digital values that can be converted is from 0 to 4095. Therefore, before converting the EEG signals stored in the memory of the Raspberry Pi to analog signals, the data were transformed by scaling each value to the range from 0 to 4095. The design includes an electronic loop with an Analog-to-Digital Converter (ADC) that transforms the analog EEG signal from the subjects—in this case, the DAC, to a digital signal for its processing by machine learning algorithms. The ADC is the 10-bit MCP3008 chip with Serial Peripheral Interface (SPI) [22]. The analog data are converted to digital signals by 10-bit architecture, so the digital data range is from 0 to 1023. As a result, we simulated the acquisition of EEG signals by means of an electronic loop between an ADC and a DAC using the data from the BED dataset. Figure 7 illustrates the complete hardware of the system and its connections.

4.2. Data Processing

The acquired signals are processed by the system controller, which is based on the Raspberry Pi 4 Model B 4 GB RAM, which performs the tasks of capturing the EEG signals, pre-processing them using a pipeline, and processing them through machine learning algorithms for the identification of the subject. The Raspberry Pi is a single-board computer based on Linux [23]. The programming language used is Python [24] as it is the most suitable for machine learning applications owing to a large number of available libraries. Task management is carried out by threads that run in parallel. The first thread consists of constant EEG data acquisition by the ADC, which is stored in a buffer for their processing, and the second thread is in charge of the pre-processing technique and the classification in real time of the samples stored in the buffer.
The communication between the user and the Raspberry is established using the SSH protocol [25]. Therefore, the result of the subject identification can be obtained in other edge devices, such as a PC or a server. Here, the result is displayed on the terminal indicating the identification of the subject.

4.3. System Validation

For test management, the Raspberry Pi handles another thread responsible for converting the dataset into EEG signals that are acquired by the ADC. To avoid data overlapping in the process of writing—DAC—and reading—ADC—the samples, a synchronization method is used between the two threads. Synchronization involves multiple threads efficiently waiting for each other to finish tasks. Specifically, the ADC thread waits to read the data until the DAC finishes writing the data, and the DAC waits to write new data until the ADC finishes reading the previous one. The ADC thread stores the read samples in a buffer, from which the classification thread reads them. Figure 8 illustrates the block diagram of the algorithm of the system software.

5. Results

5.1. Dataset Generator Efficiency

Generating analog EEG signals from a previously saved dataset by means of a DAC may generate an error between the original and the generated data. This error may be due to the mismatch of the bit architecture between the two converters or due to the error generated by each converter when generating the analog or digital signal in each case. Therefore, to analyze the efficiency of the dataset regeneration, the Mean Squared Error (MSE) between the two datasets was calculated using (4):
M S E = 1 N · n = 1 N ( y i y ^ i ) 2
where N denotes the total number of samples in the dataset, y i denotes the original dataset sample, and y ^ i denotes the sample generated by the DAC. The MSE between the original dataset and the dataset generated by the DAC is 27.57 , which represents 0.67 % of the dataset range.

5.2. Model Efficiency

Table 1 shows the performance of the three DL models used. We evaluated the models using four metrics: accuracy, F1 score, precision, and recall. Using the preprocessing PREP pipeline, the EEGNet-based model obtains the best accuracy of 86.47% followed by the Inception-based model with an accuracy of 70.18% and the ResNet-based model with an accuracy of 63.21%. In addition, we obtained for each DL model the processing time of one sample.
Figure 9 illustrates the precision vs. recall curves for different DL models to provide a graphical representation of each model’s efficiency. It can be seen that the P–R curve of the EEGNet model indicates the best results with higher precision and recall values.

6. Discussion and Conclusions

In this work, we investigated EEG-based subject identification using a consumer-grade EEG acquisition device (Emotiv EPOC+ headset) using several deep learning models that are based on ResNet, Inception, and EEGnet. We used our deep learning models on the BED dataset, an existing online EEG dataset and came out with a good deep learning EEGNet model at an accuracy of 86.74%. The high accuracy shows a promising future for EEG-based subject identification over long periods of time.
The high performance of the EEGNet-based model compared to the Inception-based model and ResNet-based model may be due to its use of Depthwise and Separable Convolutions allowing the construction of an EEG-specific model incorporating well-known EEG features.
We plan to validate these deep learning models on more training datasets once we collect them ourselves.
To better examine the performance of our deep learning models, we compared our results with previous works that use the same BED dataset [14] on subject identification tasks. The comparison is shown in Table 2. Compared to their best accuracy of 47.70%, 83.51%, and 29.69% respectively, our work on the EEGNet-based model demonstrates a higher accuracy of 86.74%. Here, ACC refers to accuracy and CCR refers to Correct Classification Rate.
As for the hardware design of this work, a proof of concept EEG-based real-time identification system using Raspberry Pi is demonstrated that integrates all the components needed for the practical use of a real subject identification system. The execution of the thread parallelism has been performed without any problem, and all the threads are synchronized to achieve their tasks without any errors. The MSE is low with respect to the range of the data conveying high similarity between the original dataset and the data generated by the DAC, which demonstrates the reliability of the results obtained during the model evaluation stage.
Finally, Raspberry Pi provides portability and allows the system to be integrated into a network; hence, the information is accessible from anywhere. We plan to develop a system for EEG acquisition and integrate it with the current Raspberry Pi-based system. The development of an acquisition system will eliminate the use of converter DAC and the need for a pre-stored dataset. The motivation for developing our device comes from the comparison of the different commercially available products. The devices found were developed only for acquisition purposes or for more complex applications and most of them are very costly. For this reason, we are motivated to develop our own end-to-end, low-cost, real-time device involving EEG acquisition from minimum necessary channels for EEG biometrics, preprocessing, feature extraction, and subject identification. In the future, to improve our device, we will perform a deeper analysis of the hardware by studying the power and time consumption and alternative technology for minimization of cost and size of the device.

Author Contributions

Conceptualization, M.V. and K.V.; Methodology, M.B., S.C., M.V., K.V. and H.C.; Formal analysis, M.B. and S.C.; Investigation, H.C.; Data curation, M.B. and S.C.; Writing—original draft, M.B. and S.C.; Writing—review & editing, M.V. and H.C.; Supervision, M.V., K.V. and H.C.; Funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSF NCS grant number #1926818 to Hung Cao.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pozo-Banos, M.D.; Alonso, J.B.; Ticay-Rivas, J.R.; Travieso, C.M. Electroencephalogram subject identification: A review. Expert Syst. Appl. 2014, 41, 6537–6554. [Google Scholar] [CrossRef]
  2. Sun, Y.; Lo, B. An Artificial Neural Network Framework for Gait-Based Biometrics. IEEE J. Biomed. Health Inform. 2019, 23, 987–998. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Jain, A.; Ross, A.; Pankanti, S. Biometrics: A Tool for Information Security. IEEE Trans. Inf. Forensics Secur. 2006, 1, 125–143. [Google Scholar] [CrossRef] [Green Version]
  4. Moolla, Y.; De Kock, A.; Mabuza-Hocquet, G.; Ntshangase, C.S.; Nelufule, N.; Khanyile, P. Biometric Recognition of Infants using Fingerprint, Iris, and Ear Biometrics. IEEE Access 2021, 9, 38269–38286. [Google Scholar] [CrossRef]
  5. Chan, H.L.; Kuo, P.C.; Cheng, C.Y.; Chen, Y.S. Challenges and Future Perspectives on Electroencephalogram-Based Biometrics in Person Recognition. Front. Neuroinform. 2018, 12, 66. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Maiorana, E.; Rocca, D.L.; Campisi, P. Eigenbrains and Eigentensorbrains: Parsimonious bases for EEG biometrics. Neurocomputing 2016, 171, 638–648. [Google Scholar] [CrossRef]
  7. Vishwanath, M. Detection of Traumatic Brain Injury Using a Standard Machine Learning Pipeline in Mouse and Human Sleep Electroencephalogram. Ph.D. Thesis, University of California, Irvine, CA, USA, 2021. [Google Scholar]
  8. Kodituwakku, S. Biometric authentication: A review. Int. J. Trend Res. Dev. 2015, 2, 113–123. [Google Scholar]
  9. Kamarozaman, N.B.; Awang, A.H. IOT COVID-19 Portable Health Monitoring System using Raspberry Pi, Node-Red and ThingSpeak. In Proceedings of the 2021 IEEE Symposium on Wireless Technology & Applications (ISWTA), Shah Alam, Malaysia, 17 August 2021; pp. 107–112. [Google Scholar] [CrossRef]
  10. Abhilash, D.; Chandrashekar, C.; Shalini, S. Economical, energy efficient and portable home security system based on Raspberry Pi 3 using the concepts of OpenCV and MIME. In Proceedings of the 2017 International Conference on Circuits, Controls, and Communications (CCUBE), Bangalore, India, 15–16 December 2017; pp. 60–64. [Google Scholar] [CrossRef]
  11. Sreejithlal, A.; Syam, M.N.; Letha, T.M.; Madhusoodanan, K.P.M.; Shooja, A. Pressure Sensor Test System Using Raspberry Pi. In Proceedings of the 2018 IEEE Recent Advances in Intelligent Computational Systems (RAICS), Thiruvananthapuram, India, 6–8 December 2018; pp. 182–185. [Google Scholar] [CrossRef]
  12. Amareswar, E.; Naik, M.; Prasad, S.; Chandra, N.V.; Karthik, R. Design of Brain Controlled Robotic Car using Raspberry Pi. In Proceedings of the 2021 5th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 3–5 June 2021; pp. 185–189. [Google Scholar] [CrossRef]
  13. Park, Y.; Han, S.H.; Byun, W.; Kim, J.H.; Lee, H.C.; Kim, S.J. A Real-Time Depth of Anesthesia Monitoring System Based on Deep Neural Network With Large EDO Tolerant EEG Analog Front-End. IEEE Trans. Biomed. Circuits Syst. 2020, 14, 825–837. [Google Scholar] [CrossRef] [PubMed]
  14. Arnau-González, P.; Katsigiannis, S.; Arevalillo-Herráez, M.; Ramzan, N. BED: A New Data Set for EEG-Based Biometrics. IEEE Internet Things J. 2021, 8, 12219–12230. [Google Scholar] [CrossRef]
  15. Bigdely-Shamlo, N.; Mullen, T.; Kothe, C.; Su, K.M.; Robbins, K.A. The PREP pipeline: Standardized preprocessing for large-scale EEG analysis. Front. Neuroinform. 2015, 9, 16. [Google Scholar] [CrossRef] [PubMed]
  16. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Fawaz, H.I.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.A.; Petitjean, F. InceptionTime: Finding AlexNet for Time Series Classification. arXiv 2019, arXiv:1909.04939. [Google Scholar]
  18. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Olson, D.L.; Delen, D. Advanced Data Mining Techniques, 2008th ed.; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  20. Dhillon, N.; Sutandi, A.; Vishwanath, M.; Lim, M.; Cao, H.; Si, D. A Raspberry Pi-Based Traumatic Brain Injury Detection System for Single-Channel Electroencephalogram. Sensors 2021, 21, 2779. [Google Scholar] [CrossRef] [PubMed]
  21. MCP4725 12-Bit DAC Datasheet; Technical Report; Microchip Technology Inc.: Chandler, AZ, USA, 2009.
  22. MCP3004/3008 10-Bit ADC Datasheet; Technical Report; Microchip Technology Inc.: Chandler, AZ, USA, 2008.
  23. Raspberry Pi 4 Model B Product Brief; Technical Report; Raspberry Pi Trading Ltd.: Cambridge, UK, 2021.
  24. van Rossum, G. Python Tutorial; Technical Report CS-R9526; Centrum voor Wiskunde en Informatica (CWI): Amsterdam, The Netherlands, 1995. [Google Scholar]
  25. Ylonen, T. SSH—Secure Login Connections over the Internet. In Proceedings of the 6th USENIX Security Symposium, San Jose, CA, USA, 22–25 July 1996. [Google Scholar]
  26. Bakırcıoglu, K.; Tanimu, M.B.; Özkurt, N.; Seçmen, M.; Güzeliş, C.; Yıldız, O. Multi Channel EEG Based Biometric System with a Custom Designed Convolutional Neural Network. In Proceedings of the International Congress of Electrical and Computer Engineering, Virtual Event, 9–12 February 2022; Springer: Berlin, Germany, 2022; pp. 124–134. [Google Scholar]
  27. Katsigiannis, S.; Arnau-González, P.; Arevalillo-Herráez, M.; Ramzan, N. Single-channel EEG-based subject identification using visual stimuli. In Proceedings of the 2021 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI), Athens, Greece, 27–30 July 2021; pp. 1–4. [Google Scholar] [CrossRef]
Figure 1. Electrode montage.
Figure 1. Electrode montage.
Sensors 22 09547 g001
Figure 2. EEG preprocessing steps.
Figure 2. EEG preprocessing steps.
Sensors 22 09547 g002
Figure 3. EEG epoch (a) raw data; and (b) preprocessed epoch.
Figure 3. EEG epoch (a) raw data; and (b) preprocessed epoch.
Sensors 22 09547 g003
Figure 4. ResNet model architecture.
Figure 4. ResNet model architecture.
Sensors 22 09547 g004
Figure 5. Inception model architecture.
Figure 5. Inception model architecture.
Sensors 22 09547 g005
Figure 6. EEGNet model architecture.
Figure 6. EEGNet model architecture.
Sensors 22 09547 g006
Figure 7. System hardware setup.
Figure 7. System hardware setup.
Sensors 22 09547 g007
Figure 8. Real-time acquisition algorithm.
Figure 8. Real-time acquisition algorithm.
Sensors 22 09547 g008
Figure 9. P–R Curves for DL models (a) ResNet model; (b) Inception model; (c) EEGNet model.
Figure 9. P–R Curves for DL models (a) ResNet model; (b) Inception model; (c) EEGNet model.
Sensors 22 09547 g009aSensors 22 09547 g009b
Table 1. Comparison of performance between Deep Learning Models.
Table 1. Comparison of performance between Deep Learning Models.
ModelResNet-BasedInception-BasedEEGNet-Based
Accuracy (%)63.2170.1886.74
F1 Score (%)61.5569.8786.69
Precision (%)63.9074.2289.13
Recall (%)63.2170.1486.68
Processing Time (ms)9.3510.510.28
Table 2. Comparison with existing works.
Table 2. Comparison with existing works.
PaperData UsedMetrics
BED: A New Data Set for EEG-Based Biometrics [14]MFCC, ARRC, and SPEC features for all 12 stimulusACC: 47.79%
Multi Channel EEG Based Biometric System with a Custom Designed Convolutional Neural Network [26]Raw EEG using 4 channels of RC StimulusCCR: 83.51%
Single-channel EEG-based subject identification using visual stimuli [27]MFCC, ARRC, and SPEC features of all stimulus for each channelACC: 29.69%
Our proposed best model based on EEGNetRaw EEG of RC StimulusACC: 86.74%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Benomar, M.; Cao, S.; Vishwanath, M.; Vo, K.; Cao, H. Investigation of EEG-Based Biometric Identification Using State-of-the-Art Neural Architectures on a Real-Time Raspberry Pi-Based System. Sensors 2022, 22, 9547. https://doi.org/10.3390/s22239547

AMA Style

Benomar M, Cao S, Vishwanath M, Vo K, Cao H. Investigation of EEG-Based Biometric Identification Using State-of-the-Art Neural Architectures on a Real-Time Raspberry Pi-Based System. Sensors. 2022; 22(23):9547. https://doi.org/10.3390/s22239547

Chicago/Turabian Style

Benomar, Mohamed, Steven Cao, Manoj Vishwanath, Khuong Vo, and Hung Cao. 2022. "Investigation of EEG-Based Biometric Identification Using State-of-the-Art Neural Architectures on a Real-Time Raspberry Pi-Based System" Sensors 22, no. 23: 9547. https://doi.org/10.3390/s22239547

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop