Electroencephalogram Dataset of Visually Imagined Arabic Alphabet for Brain–Computer Interface Design and Evaluation
Abstract
:1. Summary
- This dataset is the first publicly available EEG dataset focusing on visually imagined Arabic letters, contributing to the understanding of cognitive processes in language recognition for non-Latin characters and complementing previous Latin alphabet studies.
- The EEG dataset was collected from 30 healthy native Arabic-speaking participants, providing a robust sample size for reliable analysis and research.
- The dataset, comprising 28 distinct classes, offers significant versatility for classification and decoding experiments and applications. Moreover, researchers developing EEG-based BCI systems can utilize this variety in the available classes to design control mechanisms for assistive devices and evaluate their performance.
- The dataset can be analyzed to investigate the time, frequency, and spatial characteristics of EEG signals during VI tasks. Consequently, this dataset can be used to design new signal processing techniques and machine learning models that can advance the development of EEG-based BCI systems.
2. Methods
2.1. Subjects
2.2. EEG Data Acquisition System
- Sampling Rate: The EEG signals are internally sampled at 2048 Hz and downsampled to either 128 Hz or 256 Hz (user-configurable). For this study, we set the downsampling rate to 256 Hz.
- Bandwidth: A bandpass filter with a bandwidth of 0.16–43 Hz is applied, along with built-in digital notch filters at 50 Hz and 60 Hz.
- Filtering: An integrated 5th-order Sinc digital filter is applied for noise reduction.
2.3. Experimental Procedure
- Relaxation interval (5 s): A blank white screen was displayed, during which subjects were instructed to relax while keeping their eyes open.
- Visual observation interval (5 s): A black Arabic letter appeared at the center of the screen against a white background, prompting subjects to carefully observe the displayed letter.
- Visual imagination interval (8 s): A black screen was shown, and subjects were instructed to close their eyes and mentally visualize the letter they had observed in the previous interval. During this interval, subjects were explicitly instructed to envision the visual pattern of the letter as if it were still present in front of them, rather than thinking of its sound or imagining the act of writing it.
2.4. Quality Control and Artifact Inspection
3. Data Description
- Sx denotes the subject, with x being an integer between 1 and 30, reflecting the 30 subjects in the study;
- Ly denotes the Arabic letter, with y being an integer from 1 to 28;
- Tk denotes the trial number, with k being an integer from 1 to 10.
4. Validation of the Dataset’s Utility
5. Conclusions, Limitations, and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Knauff, M.; Kassubek, J.; Mulack, T.; Greenlee, M.W. Cortical activation evoked by visual mental imagery as measured by fMRI. NeuroReport Rapid Commun. Neurosci. Res. 2000, 11, 3957–3962. [Google Scholar] [CrossRef] [PubMed]
- Rademaker, R.L.; Pearson, J. Training visual imagery: Improvements of metacognition, but not imagery strength. Front. Psychol. 2012, 10, 3–224. [Google Scholar] [CrossRef] [PubMed]
- Wilson, H.; Chen, X.; Golbabaee, M.; Proulx, M.J.; O’Neill, E. Feasibility of decoding visual information from EEG. Brain-Comput. Interfaces 2024, 11, 33–60. [Google Scholar] [CrossRef]
- Zhang, X.; Yao, L.; Sheng, Q.Z.; Kanhere, S.S.; Gu, T.; Zhang, D. Converting your thoughts to texts: Enabling brain typing via deep feature learning of eeg signals. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications (PerCom), Athens, Greece, 19–23 March 2018. [Google Scholar]
- Yang, J.; Awais, M.; Hossain, M.A.; Yee, L.; Haowei, M.; Mehedi, I.M.; Iskanderani, A.I.M. Thoughts of brain EEG signal-to-text conversion using weighted feature fusion-based multiscale dilated adaptive DenseNet with attention mechanism. Biomed. Signal Process. Control 2023, 86, 105120. [Google Scholar] [CrossRef]
- Liu, M.; Wu, W.; Gu, Z.; Yu, Z.; Qi, F.; Li, Y. Deep learning based on batch normalization for P300 signal detection. Neurocomputing 2018, 275, 288–297. [Google Scholar] [CrossRef]
- Bose, R.; Goh, S.K.; Wong, K.F.; Thakor, N.; Bezerianos, A.; Li, J. Classification of brain signal (EEG) induced by shape-analogous letter perception. Adv. Eng. Inform. 2019, 42, 100992. [Google Scholar] [CrossRef]
- Alazrai, R.; Al-Saqqaf, A.; Al-Hawari, F.; Alwanni, H.; Daoud, M.I. A time-frequency distribution-based approach for decoding visually imagined objects using EEG signals. IEEE Access 2020, 8, 138955–138972. [Google Scholar] [CrossRef]
- Ullah, S.; Halim, Z. Imagined character recognition through EEG signals using deep convolutional neural network. Med. Biol. Eng. Comput. 2021, 59, 1167–1183. [Google Scholar] [CrossRef] [PubMed]
- Alazrai, R.; Abuhijleh, M.; Ali, M.Z.; Daoud, M.I. A deep learning approach for decoding visually imagined digits and letters using time–frequency–spatial representation of EEG signals. Expert Syst. Appl. 2022, 203, 117417. [Google Scholar] [CrossRef]
- Ramirez-Quintana, J.A.; Macias-Macias, J.M.; Ramirez-Alonso, G.; Chacon-Murguia, M.I.; Corral-Martinez, L.F. A novel deep capsule neural network for vowel imagery patterns from EEG signals. Biomed. Signal Process. Control. 2023, 81, 104500. [Google Scholar] [CrossRef]
- Ahmadieh, H.; Gassemi, F.; Moradi, M.H. A hybrid deep learning framework for automated visual image classification using EEG signals. Neural Comput. Appl. 2023, 35, 20989–21005. [Google Scholar] [CrossRef]
- Emotiv Epoc X. Available online: https://www.emotiv.com/products/epoc-x (accessed on 22 February 2025).
- EmotivPro Software. Available online: https://www.emotiv.com/products/emotivpro (accessed on 22 February 2025).
- Gómez-Herrero, G.; De Clercq, W.; Anwar, H.; Kara, O.; Egiazarian, K.; Van Huffel, S.; Van Paesschen, W. Automatic removal of ocular artifacts in the EEG without an EOG reference channel. In Proceedings of the 7th Nordic Signal Processing Symposium-NORSIG, Reykjavik, Iceland, 7–9 June 2006; pp. 130–133. [Google Scholar]
- Alazrai, R.; Alwanni, H.; Daoud, M.I. EEG-based BCI system for decoding finger movements within the same hand. Neurosci. Lett. 2019, 698, 113–120. [Google Scholar] [CrossRef] [PubMed]
- Alazrai, R.; Homoud, R.; Alwanni, H.; Daoud, M.I. EEG-based emotion recognition using quadratic time-frequency distribution. Sensors 2018, 18, 2739. [Google Scholar] [CrossRef] [PubMed]
- Alazrai, R.; Alwanni, H.; Baslan, Y.; Alnuman, N.; Daoud, M.I. Eeg-based brain-computer interface for decoding motor imagery tasks within the same hand using choi-williams time-frequency distribution. Sensors 2017, 17, 1937. [Google Scholar] [CrossRef] [PubMed]
Subject | Gender | Age | Handedness | Subject | Gender | Age | Handedness |
---|---|---|---|---|---|---|---|
S01 | Male | 21 | Right-handed | S16 | Male | 20 | Right-handed |
S02 | Female | 21 | Right-handed | S17 | Female | 21 | Left-handed |
S03 | Male | 20 | Right-handed | S18 | Male | 21 | Right-handed |
S04 | Male | 20 | Left-handed | S19 | Male | 21 | Right-handed |
S05 | Male | 21 | Right-handed | S20 | Female | 22 | Right-handed |
S06 | Female | 21 | Right-handed | S21 | Male | 22 | Right-handed |
S07 | Female | 21 | Right-handed | S22 | Female | 20 | Right-handed |
S08 | Female | 20 | Right-handed | S23 | Female | 21 | Right-handed |
S09 | Female | 21 | Left-handed | S24 | Male | 21 | Right-handed |
S10 | Male | 20 | Right-handed | S25 | Male | 21 | Right-handed |
S11 | Female | 21 | Right-handed | S26 | Male | 21 | Right-handed |
S12 | Male | 22 | Right-handed | S27 | Male | 19 | Right-handed |
S13 | Male | 20 | Right-handed | S28 | Male | 20 | Right-handed |
S14 | Male | 21 | Right-handed | S29 | Male | 23 | Right-handed |
S15 | Male | 21 | Left-handed | S30 | Female | 19 | Right-handed |
Letter ID | Arabic Letter | Letter ID | Arabic Letter | Letter ID | Arabic Letter |
---|---|---|---|---|---|
L01 | أ | L11 | ز | L21 | ق |
L02 | ب | L12 | س | L22 | ك |
L03 | ت | L13 | ش | L23 | ل |
L04 | ث | L14 | ص | L24 | م |
L05 | ج | L15 | ض | L25 | ن |
L06 | ح | L16 | ط | L26 | ه |
L07 | خ | L17 | ظ | L27 | و |
L08 | د | L18 | ع | L28 | ي |
L09 | ذ | L19 | غ | ||
L10 | ر | L20 | ف |
Field | Description | |||
---|---|---|---|---|
timestamp | The recorded timestamps for each sample point in the EEG data. These timestamps are organized as a row vector with dimensions Nsp × 1, where Nsp denotes the number of sample points recorded within a specific trial. | |||
data | The recorded EEG data for each trial are organized as a matrix with dimensions Nch × Nsp, where Nch represents the number of EEG channels (14 channels in this study). The order and description of these EEG channels are detailed in the channels field. | |||
channels | A cell array with dimensions of 14 × 1, where each cell contains a string representing the name of an EEG channel, arranged in the order shown below. | |||
Channel index | Channel name | Channel index | Channel name | |
1 | AF3 | 8 | O2 | |
2 | F7 | 9 | P8 | |
3 | F3 | 10 | T8 | |
4 | FC5 | 11 | FC6 | |
5 | T7 | 12 | F4 | |
6 | P7 | 13 | F8 | |
7 | O1 | 14 | AF4 | |
samplingrate | The number of samples measured in one second. | |||
events | A row vector consisting of four values, each representing the sample index marking the start or end of a specific event within recorded trial. The descriptions of these four events are as follows: | |||
Event index | Event description | |||
1 | The sample index marking the beginning of the trial. | |||
2 | The sample index marking the end of the relaxation interval and the start of the letter observation interval. | |||
3 | The sample index marking the start of the letter imagination interval and the end of the letter observation interval. | |||
4 | The sample index marking the end of the letter imagination interval, which also marks the end of the trial. | |||
subjectname | A string representing the subject ID, for example, “S03” for subject 3. | |||
objectname | A string representing the letter ID, for example, “L03” for letter 3. | |||
trialnumber | A string representing the trial number, for example, “T03” for trial 3. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alazrai, R.; Naqi, K.; Elkouni, A.; Hamza, A.; Hammam, F.; Qaadan, S.; Daoud, M.I.; Ali, M.Z.; Al-Nashash, H. Electroencephalogram Dataset of Visually Imagined Arabic Alphabet for Brain–Computer Interface Design and Evaluation. Data 2025, 10, 81. https://doi.org/10.3390/data10060081
Alazrai R, Naqi K, Elkouni A, Hamza A, Hammam F, Qaadan S, Daoud MI, Ali MZ, Al-Nashash H. Electroencephalogram Dataset of Visually Imagined Arabic Alphabet for Brain–Computer Interface Design and Evaluation. Data. 2025; 10(6):81. https://doi.org/10.3390/data10060081
Chicago/Turabian StyleAlazrai, Rami, Khalid Naqi, Alaa Elkouni, Amr Hamza, Farah Hammam, Sahar Qaadan, Mohammad I. Daoud, Mostafa Z. Ali, and Hasan Al-Nashash. 2025. "Electroencephalogram Dataset of Visually Imagined Arabic Alphabet for Brain–Computer Interface Design and Evaluation" Data 10, no. 6: 81. https://doi.org/10.3390/data10060081
APA StyleAlazrai, R., Naqi, K., Elkouni, A., Hamza, A., Hammam, F., Qaadan, S., Daoud, M. I., Ali, M. Z., & Al-Nashash, H. (2025). Electroencephalogram Dataset of Visually Imagined Arabic Alphabet for Brain–Computer Interface Design and Evaluation. Data, 10(6), 81. https://doi.org/10.3390/data10060081