Next Article in Journal
Geometric Analysis of a System with Chemical Interactions
Previous Article in Journal
Winsorization for Robust Bayesian Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Classification of Event-Related Potentials for the Recognition of Gender Differences in the Attention Task

by
Karina Maciejewska
1,* and
Wojciech Froelich
2
1
Institute of Biomedical Engineering, Faculty of Science and Technology, University of Silesia in Katowice, 75 Pulku Piechoty 1a Street, 41-500 Chorzow, Poland
2
Institute of Computer Science, Faculty of Science and Technology, University of Silesia in Katowice, Bedzinska 39 Street, 41-205 Sosnowiec, Poland
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(11), 1547; https://doi.org/10.3390/e23111547
Submission received: 17 October 2021 / Revised: 15 November 2021 / Accepted: 17 November 2021 / Published: 20 November 2021

Abstract

:
Research on the functioning of human cognition has been a crucial problem studied for years. Electroencephalography (EEG) classification methods may serve as a precious tool for understanding the temporal dynamics of human brain activity, and the purpose of such an approach is to increase the statistical power of the differences between conditions that are too weak to be detected using standard EEG methods. Following that line of research, in this paper, we focus on recognizing gender differences in the functioning of the human brain in the attention task. For that purpose, we gathered, analyzed, and finally classified event-related potentials (ERPs). We propose a hierarchical approach, in which the electrophysiological signal preprocessing is combined with the classification method, enriched with a segmentation step, which creates a full line of electrophysiological signal classification during an attention task. This approach allowed us to detect differences between men and women in the P3 waveform, an ERP component related to attention, which were not observed using standard ERP analysis. The results provide evidence for the high effectiveness of the proposed method, which outperformed a traditional statistical analysis approach. This is a step towards understanding neuronal differences between men’s and women’s brains during cognition, aiming to reduce the misdiagnosis and adverse side effects in underrepresented women groups in health and biomedical research.

1. Introduction

Recently, the importance of recognizing gender differences in health and biomedical research, including neuroscience, has become a significant matter of concern, prompting new policies to be implemented at the funding agencies (Canadian Institutes of Health Research, European Commission, US National Institutes of Health or German Research Foundation) to supervise gender analysis [1,2,3,4].
In recent decades, experiments in biomedical research were carried out without considering gender, resulting in more significant health risks for women due to their under-representation in the preclinical, clinical studies, and drug trials. Higher rates of misdiagnosis and adverse side effects from drug treatment were also more common for women. Acknowledging gender-specific functional differences is of great importance in the clinical research of brain disorders such as autism, conduct disorder, attention-deficit hyperactivity disorder (ADHD), schizophrenia, dyslexia, stuttering, Tourette’s syndrome, major depression, anxiety, panic disorders, obsessive-compulsive disorder (OCD), posttraumatic stress disorder (PTSD), bulimia, migraines, multiple sclerosis (MS), myasthenia gravis, and Alzheimer’s disease.
In this paper, we address this crucial research problem of cognitive neuroscience. We examine whether gender differences in human cognition can be recognized while performing the attention task. We used event-related potentials (ERPs) measured while performing attention tasks by a cohort of women and men to accomplish that. We used that data as an input to the data mining method we propose in this paper. The method proposed here combines practical and theoretical efforts to solve the underlying stated problem. As interdisciplinary research, this work contributes to both neuroscience and computer science. The contribution of this paper should be considered from empirical and theoretical perspectives:
  • We propose a carefully designed original experimentation procedure enabling the acquisition of electroencephalography (EEG) signals in the human attention task. In contrast to many EEG-based classification studies, where the EEG datasets were acquired from open databases, we designed and ran a neurophysiological experiment. Thus we ensured complete control of the possible confounds that are unknown when using public databases, allowing us to detect gender differences using the proposed data mining technique;
  • The theoretical contribution is the data mining method, which relies on the hierarchical segmentation and classification of ERPs. In particular, we transform the preprocessed ERPs to a multivariate time series. The time-series underlies segmentation, enabling the construction of a chain of classifiers yielding the targeted gender classification;
To the best of our knowledge, this paper is the first study proposing data mining recognition of gender differences using the ERP signal from an attention task.
The remainder of this paper is organized as follows. In Section 2, by presenting a literature review on the addressed problem, we motivate and prove the originality of the research undertaken in this paper. To deal with the addressed problem, we formalize it in Section 3.1. Then, in Section 3.2, we go to the detailed presentation of the proposed method. The results proving the high effectiveness of the proposed method are provided in Section 4. Section 5 concludes the paper.

2. Literature Review

ERPs are time-locked brain responses to stimuli, measured using EEG. They are scalp-recorded signatures of neural processes generated by groups of neurons activated during cognitive functions. The ERP component is defined as the scalp-recorded activity generated by a specific neural or psychological process, which produces defined polarity, latency, and scalp distribution, and is sensitive to experimental manipulations [5]. Traditionally, in ERP studies, amplitude and latency (i.e., the time interval between stimulus onset and the waveform) are measured, and these measures are compared between conditions. However, averaging the ERP trials to the stimuli onset across all the trials, i.e., the presentations of the stimuli during an examination, introduces a jitter that lowers the effect size and standard methods may not be sensitive enough [5].
Due to large individual differences and the small effect size of gender differences, there is still no consensus in the currently available literature regarding gender-related differences in the ERP waveforms [6,7,8,9,10,11,12,13,14,15]. Therefore, there is a need to search for more sensitive tools that are able to capture gender-related differences among healthy participants. It is necessary to explore computational approaches in neuroscience to present results that are both accurate and more robust. Consequently, neuroscientists have recently started searching for different machine learning classification algorithms as a tool for EEG-based decoding for studying the neural coding of human cognition, as they may increase both the stability of ERP data and classification accuracy.
ERP classification can be used to determine whether studied groups of subjects differ from each other. The purpose of using classification techniques for EEG signal analysis is to increase the statistical power of the differences between conditions that are too weak to be detected using standard ERP analytical methods. A classifier is then trained to distinguish between ERP time series from one group and the ERP time series from another group, and it then separates the two classes. Therefore, EEG classification methods may serve as a valuable tool for studying the temporal dynamics of human brain activity, and this study was designed to that end. That means that the classification was performed on averaged ERPs to understand the nature of underlying signals. That is why in such applications, obtained accuracies are informative, even though they are not as high as observed in brain-computer interface (BCI) applications, where high accuracy is essential due to single-trial decoding for prediction.
However, classification techniques to analyze EEG signals for understanding are still a novelty. Even less has been done to classify ERPs related to attention, one of the most important cognitive processes. While most EEG-based classification studies rely on resting-state EEG recording, in this work we focused on ERPs in reaction to visual stimuli as a result of neuronal activity related to the cognitive process, allowing for the isolation of specific cognitive functions of the human brain. In an oddball attention task, a P3 component is observed, which is related to the number of attentional resources engaged during task performance and reflects stimulus classification speed, thus being a sensitive temporal measure of the neural activity underlying the processes of attention allocation and immediate memory [16,17].
Among all gender classification methods, EEG-based gender recognition is precious, since EEG captures cognitive processes, revealing the estimated gender-related differences in the brain. It has multiple advantages, such as: high accuracy, high permanence, inherent uniqueness, universality, and resistance to deception [18]. However, the number of studies focused on gender classification in cognition is very limited due to the small effect size of gender differences and the impossibility to use a within-subject study design. Some of the previous studies showed the reliability and potential trustworthiness for gender classification using EEG signal [19,20,21,22,23,24,25,26,27].
Hu et al. compared a combination of four entropy feature sets, six single classifiers, and three ensemble algorithms as a method to identify gender based on resting-state EEG signal [20]. The authors found their results promising in providing a more efficient method for recognizing gender. Van Putten et al. explored deep learning for gender classification based on frequency features of the resting-state EEG [26]. Nguyen et al. [24] proposed a framework of automatic age and gender classification using EEG data from a person. The features were sent to a machine learning module, for example, support vector machine (SVM), to build age and gender models for that person. The experiments suggested that the paralinguistic features were very promising for this task. In another work of these authors [25], they proposed a framework based on parallel factors (PARAFAC), multilinear partial least squares (N-PLS) and SVM, which that can automatically classify age and gender using frequency features of the resting-state EEG data with eyes open and eyes closed. Li et al. [23] investigated the potential gender differences in resting-state EEG signals. Ghani [19] investigated gender classification of normal subjects based on their frontal resting-state EEG signals. Kaushik [22] used deep bi-directional long short term memory (BLSTM-LSTM) neural network to construct a hybrid learning framework for age and gender prediction based on resting-state EEG recording of subjects with closed eyes. Kaur et al. [21] presented an automatic age and gender prediction framework of users based on their neural signals captured during eyes-closed resting-state EEG. Wang [27] proposed a hybrid model for EEG-based gender recognition in the resting-state, which showed the potential applicability of the proposed approach and its ability to identify personal gender in an EEG-based biological recognition system.
However, these studies were based on resting-state EEG recordings, which do not capture gender-related differences in cognitive processes, such as attention, memory, language comprehension, or decision making. Specific cognitive tasks may achieve better performance due to more pronounced gender differences. There have been studies that applied EEG-based classification methods to different cognitive processes: anxiety [28], motor imaginary [29], working memory and spatial attention [30,31], emotions [32,33,34,35], or preference recognition in neuromarketing [36]. However, we found only two studies that investigated the gender-related difference in EEG during a cognitive task. De La Pava et al. [37] studied the gender differences present in an EEG-based emotion level classification system by means of simple K-nearest neighbor classifiers. The obtained results showed a gender-related difference for the valence dimension of the emotion scale in terms of classification performance. However, this work did not use the ERPs, which capture the dynamics of brain activity. Bilalpur et al. [38] studied gender and emotion recognition with ERPs and eye-tracking during emotional face processing. The authors concluded that gender differences were encoded best for anger and disgust.
In this paper, we use a careful and well-thought-out preprocessing pipeline, which is used in conventional ERP studies but not in the EEG-based classification studies (see [39] for a review showing that over 60% of the studies reviewed did not systematically remove EEG artifacts). In addition, using Target-Standard difference waveforms, i.e., subtracting ERPs to standard stimuli from ERPs to target stimuli, minimized potential confounds related to non-neural sources (e.g., physical properties of the stimuli or non-neural between-subject differences). This is a recommended approach in ERP studies that uses traditional methods but is not commonly used, especially not in EEG-based classification studies.
In our previous work, we examined ERPs in a standard visual P3 paradigm among healthy participants by means of an empirical and statistical approach to evaluate gender-related differences in ERPs [40]. A nonparametric cluster-based test showed significant differences in ERPs between men and women. In this paper, we present a new, hierarchical approach for more efficient gender recognition.
The novelty of our current approach lies in combining the neuronal signal preprocessing with the classification method, with a contribution of segmentation step, into a full line of electrophysiological signal analysis during the attention task to a better differentiation of the attention-related ERPs in men from ERPs in women.

3. Materials and Methods

Before going to the presentation of the proposed method, we specify the addressed problem.

3.1. Problem Specification

Let p = [ 1 , 2 , , p m a x ] be an index of a person participating in an experiment, where p m a x Z is a parameter, the number of persons considered. From each pth person, we gather continuous measurements from d electrodes that are indexed by e = [ 1 , 2 , , d ] . We assume z e p ( τ ) R is a single, real-valued measurement, gathered during the experiment from the pth person and eth electrode. By τ [ τ m i n , τ m a x ] we denote real-time measured from the beginning of the measurement at which the measurement was taken, where τ m i n = 0 . The period of the experiment is determined by an interval [ τ m i n , τ m a x ] , where τ m i n , τ m a x are the real-valued parameters.
By putting together all the measurements gathered at time τ from the pth person and d electrodes, we get a d-dimensional vector Z p ( τ ) R d . A sequence { Z p ( τ ) } of those vectors is gathered over the time of each recording.
Let us consider now a mapping between { Z p ( τ ) } and the gender of the considered pth person. We denote that mapping as M ( { Z p ( τ ) } ) G , where G = { f e m a l e , m a l e } is the set of class labels g G that are to be assigned to the sequence { Z p ( τ ) } . Note that under the above assumptions, the mapping M is a classifier identifying the gender of a given person.
For the sake of completeness of the introduced notation, we denote by 2 Z ( τ ) a power-set of { Z ( τ ) } , i.e., a huge space of data sequences that can be potentially gathered from the experiments independent of the considered person.
Let us assume the gender of a given person is unknown, i.e., we do not have a classifier M available. Thus, we face a problem of discovering M and then making an assignment g = M ( { Z p ( τ ) } ) , if possible, for each { Z p ( τ ) } 2 Z ( τ ) . This is a problem we address.

3.2. The Proposed Method

The method proposed here covers the data acquisition process, data preprocessing, and classification. Let us first take a look at the proposed approach from a general perspective.
  • Acquisition of ERP signals. To cope with the stated problem, we discover M using a supervised type of learning, i.e., we induce M from sample data { S 2 Z ( τ ) } with the known classifications. Those sample data S are gathered through experiments. They are also used for the validation of the proposed approach;
  • Data preprocessing. We preprocess the gathered data to form a d-dimensional multi-variate time series. This step leads to a reduction of the considered data. Note that by the data preprocessing we replace the problem of discovering M, by looking for a classifier M 1 that deals with the produced multivariate time series instead of the raw data gathered from the experiments. Therefore, the quality of data pre-processing is pivotal for the reliability of the proposed approach;
  • Time series segmentation. We segment the previously produced time series in the time domain. Thus, we replace the problem of discovering M 1 by a more straightforward problem of constructing M 2 a classifier that, instead of dealing with the entire time series, classifies only a much shorter part (segment) of them. The issue that arises here and that we solve through computational experiments is the selection of the time segment that best suits the classification when combined with the classifier M 2 ;
  • Bottom-level classification. Finally, we classify the selected segment of the time series by combining the classifications of each vector contained in that segment. It means we construct M 2 by combining the classifications delivered by a standard, state-of-the-art classifier denoted here as M 3 . The obstacle that we come across here is the selection of M 3 .
In the above-described way, we create a hierarchy of data processing steps. In particular, we form a chain of classifiers M M 1 M 2 M 3 . Thanks to that, we transform the addressed problem in a way that on its bottom level (classifier M 3 ) we are able to use a standard, state-of-the-art classifier known from the literature.

3.2.1. Acquisition of EEG Signal

Twenty students participated in the experiment (23.1 ± 1.1 years, 10 women). All participants were right-handed, had normal color perception, normal (or corrected to normal) visual acuity, and normal blood pressure and body temperature at the time of the study. They were healthy, non-smokers, and had no history of neurological or psychological disorders. Information about their health conditions and lifestyles was collected via questionnaire. None of the participants had consumed alcohol, coffee, intoxicants, or energizing beverages (or other such substances) within 12 h prior to the study (based on the questionnaire), which could have had an impact on subjects’ cognition. Participants were also asked to get adequate rest, not to attend parties or other tiring events, and not to consume large amounts of alcohol the day before the examination.
The participants were seated in a comfortable chair in front of the computer screen at a distance of 1 m, in a dimmed room. A two-stimulus oddball paradigm, which included standard (frequent) and target (rare) stimuli, was used. The visual stimuli consisted of images of white and black geometric figures, presented in a randomized order in the center of a 19 inch LCD monitor. A black square on a white background was the target stimulus, and a white circle on a black background was the standard stimulus. The length of the square side and the diameter of the circle were 9 cm each. Participants were instructed to press a button when they saw the target stimulus, and to gaze at the center of the black screen during the inter-stimulus interval. The visual P3b potential was elicited in response to task-relevant target stimuli. The parameters of the stimuli were: 150 ms duration, 1000 ms inter-stimulus interval, 20% target and 80% standard stimulus probabilities. The total number of stimulus presentations was 300, including 240 standard and 60 target stimuli. The scenario of the experiment was created using Eevoke software (ANT Neuro, Hengelo, The Netherlands).
Continuous EEG was recorded from 32 Ag/AgCl electrodes embedded in an elastic Waveguard TM EEG cap (ANT Neuro, Hengelo, The Netherlands), using extended 10/20 EEG montage system with the AFz electrode as the ground electrode. Before the signal acquisition, participants’ skin was prepared for the examination. Everi (Spes Medica s.r.l., Genova, Italy) abrasive and conductive paste was used to clean the skin on the hairless areas of the scalp to remove dead skin and skin impurities before putting on the EEG cap. Then, the EEG cap was put on and OneStep Clear Gel (H + H Medizinprodukte GbR, Münster, Germany) was inserted into all the electrodes in order to provide contact between skin and electrodes. The impedance at each electrode site was kept below 5 k Ω . The signals were recorded using a common average reference. The EEG signal was collected by way of ANT Neuro (Hengelo, The Netherlands) amplifier (AMP-TRF40AB model) in DC with 20,000 amplification gain and 256 Hz sampling rate. No high-pass filter was applied during data acquisition. The acquisition and data pre-processing of EEG signals were performed according to the International Federation of Clinical Neurophysiology (IFCN) Guidelines for eliciting, recording, and quantifying mismatch negativity, P300, and N400 [41]. Advanced Source Analysis system ASA-Lab (ANT Neuro) with ASA v.4.8 software was used for the acquisition and offline data pre-processing.
In this study, an EEG signal was measured from each participant’s head during the attention task. During acquisition, the biological analog signal is converted into a discrete, digitized signal, where each data point (i.e., potential value in μV) is measured and saved every time step. This time step is the result of the sampling frequency F s [Hz = 1/s], which is the number of samples obtained in one second. Our data have been measured with 256 Hz, so the time step is 1/256 Hz = 0.00390625 s = 3.90625 ms. Therefore, the raw, continuous signal contains all the data points collected within the experiment, which is approximately 5 min. However, since our goal is to classify the EEG signal in the attention task, we have to perform pre-processing of the raw data. Aside from dealing with the artifacts, pre-processing contains epoching and averaging the epochs, which allows obtaining ERPs time-locked to the presented stimuli.
This way, we produce the sequence { Z p ( τ ) } . Note that the length of that sequence differs across the participants since there was an approximately 30 s additional time interval before the start and after the stop of the stimulation presentation of the attention task. This was done in order to allow the EEG signal to stabilize and for the subsequent data pre-processing. However, this difference does not influence the classification process because the classification was performed on the data epoched time-locked to the stimuli onset.

3.2.2. Data Pre-Processing

To obtain ERP from the raw EEG { Z p ( τ ) } , we introduce a function Φ representing data pre-processing. Using that function, we perform a transformation Φ ( { Z p ( τ } ) { X p ( t ) } . This transformation is performed due to the fact that during the EEG measurement time τ [ τ m i n , τ max ] , a set of visual stimuli were presented to the participants, divided into two categories: standard stimuli and target stimuli. Participants were asked to focus their attention on the target stimuli. Therefore, the task-related ERPs come from the transformed X p ( t ) dataset. Figure 1 shows the workflow of EEG pre-processing, which included data filtering, blink correction using PCA, baseline correction, epoching, artifact detection, ERP averaging, and ERP target-standard difference waveforms calculation.
In the data pre-processing step, the recorded EEG signal was first filtered using a non-causal Butterworth band-pass filter (with 0.01–30 Hz half-amplitude cutoff and 24 dB/oct slope). In order to correct for eye blinks artifacts, principal component analysis (PCA) was conducted. An alternative way to reduce artifacts is independent component analysis (ICA) [42,43,44], however, in order to maintain a full agreement to the previously presented results, we performed the classification on the same pre-processed data as the previous experiment. A signal with amplitude over ±75 μV was detected and removed from the analysis. After signal filtering and correcting for eye artifacts (see Figure 1a–c), for each participant, the continuous data has been epoched, i.e., epochs (time intervals) from 100 ms to 1000 ms time-locked to each of the 300 stimuli have been extracted (see Figure 1d). This means each epoch started 100 ms before the stimulus onset (baseline) and ended 1000 ms after the stimulus onset, resulting in 281 time points in each epoch. After baseline correction and detrending, all the epochs were averaged across all standard and target stimuli separately (see Figure 1d,e). The mean (± standard deviation) number of standard and target stimuli after data pre-processing were: 160 ± 32 and 40 ± 8 in women and 164 ± 51 and 40 ± 13 in men, respectively. Finally, in order to isolate the ERPs related to the task, the ERP difference waveforms were calculated by subtracting the averaged ERPs from the standard stimulus category from the averaged ERPs from the target stimulus category (target-standard difference waveforms, see Figure 1e,f).
Note that while performing the transformation Φ ( { Z p ( τ } ) { X p ( t ) } , the initially considered continuous time scale is epoched, and the epochs are averaged across the presented stimuli. After that, it is represented by t T , where: T = { 1 , 2 , , t max } . The period of the experiment measured in real-time [ τ m i n , τ max ] is thus mapped to [ 1 , t max ] , where t m a x is a parameter common for all persons and experiments. Thus instead of dealing with time in a continuous scale τ , we consider time steps t within the epoch.
Thanks to the data pre-processing and the time discretization, which we describe in detail in the following section, the gathered data can be considered as multivariate time series. Specifically, data collected at time t from the eth the electrode of a particular pth person form a univariate time series { X e p ( t ) } . By collecting data from all electrodes, we get a multivariate time series. i.e., { X p ( t ) } = [ { X 1 p ( t ) } , { X 2 p ( t ) } , , { X d p ( t ) } ] , where t T . Thus, the time series { X p ( t ) } is composed of univariate time series { X e p ( t ) } .

3.2.3. Data Segmentation

After collecting the time series { X p ( t ) } we partition them into segments { W k p ( t ) } { X p ( t ) } , where k = 1 , 2 , k max is an index of a segment, and k m a x is a number of those segments. Figure 2 presents the epoched data specification (panel a) and a segment specification (panel b).
We measure the length | { W k p ( t ) } | of each segment as a number of vectors X p ( t ) { W k p ( t ) } contained in the kth time interval, corresponding to that segment.
Thanks to the performed segmentation, instead of classifying the entire time series { X p ( t ) } , we use for that purpose a single, much shorter part of it, namely a segment { W k p ( t ) } . This way, we again reduce the amount of data that needs to be classified.
The result of the segmentation process is illustrated in Figure 3, which presents the grand averaged ERPs for men and women, i.e., the mean ERPs from each participant averaged across all the men and all the women, for a representative Pz electrode site (midline parietal).
The bottom time scale represents an epoch used for data averaging (i.e., −100 to 1000 ms time-locked to the stimulus presentation) with −100–0 ms serving as the baseline. The top time scale represents consecutive numbers of time segments (in the range 1- k m a x ), k max = 22 , which correspond to 50-ms time bins in the original time scale. For example, the segment k = 10 covers time series related to the epoched data gathered in period t [ 350 400 ms ] measured between 350 ms and 400 ms after the stimuli onset.
Thanks to the segmentation, instead of using M 1 ( { X p ( t ) } ) , we are able to use a classifier M 2 ( { W k p ( t ) } ) , for kth segment. The selection of that segment, i.e., the value of k, for which the classification of { W k p ( t ) } leads to the best classification accuracy, is the goal of the computational experiments we perform at the experimental stage.

3.2.4. Bottom-Level Classification of ERP Waveforms

Note that each segment { W k p ( t ) } is a multivariate time-series, i.e., it contains a sequence of d-dimensional vectors X p ( t ) that have to be classified. To classify each of those vectors separately, we are able to use one of the standard, state-of-the-art classifiers. Here, we denote it as M 3 ( X p ( t ) ) .
More specifically, we employ M 3 to generate g 3 ( t ) = M 3 ( X p ( t ) ) , where g 3 ( t ) denotes the classification delivered by the classifier M 3 for each X p ( t ) { W k p ( t ) } } . Note that this classification may not always be correct, i.e., g 3 ( t ) g .
Going backward in our hierarchy of classifiers, we need to produce the classification for the entire considered segment, i.e., calculate g 2 ( t ) = M 2 ( { W k p ( t ) } ) . For that purpose, we count classifications g 3 ( t ) = ’female’ and g 3 ( t ) = ’male’ delivered by M 3 for all vectors X p ( t ) { W k p ( t ) } . The greater value, i.e., the value that better supports a particular gender, is selected as the classification of the entire segment, i.e., g 2 ( t ) = arg m a x g α ( { W k p ( t ) } , g ) , where α ( { W n p ( t ) , g } is calculated as:
α ( { W n p ( t ) } , g ) = | M 3 ( X p ( t ) ) = g | | { W k p ( t ) } | .
In Formula (1), we denote by | M 3 ( X p ( t ) ) = g | the number of correctly classified vectors X p ( t ) belonging to segment { W n p ( t ) } . The number of all vectors that are classified, i.e., the length of segment { W k p ( t ) } is denoted by | { W k p ( t ) } | . Thus the value of α is the rate of correct classification yielded by M 3 for a particular gender g in particular kth segment { W k p ( t ) } .
Following backward our hierarchy of classifiers, we calculate the classification of the entire time series and, i.e., produce g 1 = M 1 ( { X p ( t ) ) } . The issue is that g 2 ( t ) produced previously by M 2 is still time-dependent, i.e., it depends on the segment W k p ( t ) for which it was obtained. For that reason, considering the entire { X p ( t ) } we get a series of g 2 ( t ) , where some of them can be correct, i.e., g 2 ( t ) = g , and some not, i.e., g 2 ( t ) g .
To cope with that issue, we must select the segment (the value of k) that will be used for gender recognition. For that purpose, we perform a series of computational experiments using the gathered data sample S. The goal of those experiments is to select that segment for which the classification g 2 ( t ) = M 2 ( { ( W k p ( t ) } ) is the best, i.e., it delivers the highest number of correctly identified genders. For each segment { W k p ( t ) } , k = 1 , 2 , , k max we calculate:
γ ( { W k ( t ) } ) = p = 1 p m a x | M 2 ( { ( W k p ( t ) } ) = g | p m a x ,
where p is an index of persons and p m a x is the number of them.
The segment with the highest γ , i.e., W best = a r g m a x k γ ( { W k ( t ) } ) is selected as that carrying the most useful information regarding the final classification. Therefore, we assume the classification of the chosen segment as that for the entire time series, i.e., M 1 ( { X p ( t ) } ) = M 2 ( { W best ( t ) } ) .
Going to the uppermost level of the classification chain, we assume the data pre-processing Φ was made in the best possible way and enables us to conclude that M = M 1 . This way, we solve the initially stated problem of gender recognition.

4. Results and Discussion

4.1. Experimental Setup

At each level of the proposed approach, we deal with a number of parameters. The values of those parameters are provided in Table 1.
Due to the low number of participants (20) we opted to use the leave-one-out cross-validation, meaning we performed 20 validation trials. In each of them, we used data gathered from 19 participants for learning classifier M. A single person left in each trial was used for testing. This way, we tested our approach for each of the 20 participating subjects. We also avoided randomness, and its consequences occurring, by using standard k-fold cross-validation.
Averaging classification results over 20 learn-and-test trials we calculated α and γ using Formulas (1) and (2), respectively. All the obtained results were rounded to two decimal places.
As the selection of the best classifier (among hundreds of those that are available) is a complex research problem that has been addressed for years [45], we relied in that case on the recommendations provided in the literature [45]. For the comparative experiments, we assumed Near-Neighbor (k-NN), Naive Bayes (NB), Random Forrest (RF), and Support Vector Machine (SVM) classifiers as those frequently used for the classification of EEG signals. Note that the chosen classifiers rely on diverse theoretical grounds. Thus, we validate different approaches to the classification task.
The k-NN classifier assumes that similar data instances can be classified to the same class [46]. The classifier is called lazy because it does not learn any data model. It only stores training data. The classifier calculates the distance between the query example and each of the stored data instances from the training data set. The collection of the calculated distances is sorted from smallest to largest. Then, the class labels are picked from the first k entries of that list, where k is the parameter of the classifier. The mode of those labels is returned as the classification of the query example. For this study, we used the k-NN with the most popular Euclidean distance method.
A completely different approach represents the NB classifier that relies on counting data instances while ignoring similarities between them. The classifier uses the Bayesian formula to calculate the most probable class label for the classified data instance [47]. For that purpose, conditional probabilities of each possible class y | x i and for each feature x i are calculated. In the case of numerical attributes, it is assumed that those values are sampled from a Gaussian distribution. As the NB also assumes that the features are independent, i.e., a feature does not affect the other, the calculation of P ( y | x 1 , , x n ) can be easily performed using the Bayes formula and the multiplication of the obtained probabilities. The class y with the maximum P ( y | x 1 , , x n ) is assigned to the query data instance.
The RF classifier is an ensemble learning method that uses a chosen number of decision trees [48]. As the RF relies on decision trees, it takes into account both the similarity and statistical properties of data. For this paper, we checked various tree split criteria and selected the information gain ratio as the most promising in terms of the finally obtained classification accuracy. Each of the decision trees is based on a different set of data instances and a different random set of features of size p , where p is the total number of attributes. The output of the RF is the class selected by most trees.
The SVM classifier also relies on the calculation of distances between data instances [49]. In that sense, it is similar to the k-NN. However, in the case of the SVM, the goal is to maximize the margin distance between data instances. The margin is understood as a gap between the hyperplanes separating data instances belonging to different classes, and the positions and orientations of those hyperplanes are determined by support vectors. By their proper adjustment, it is possible to maximize the margins of the SVM. As a result, data instances falling on either side of the hyperplanes are attributed to different classes. The issue here is that the separation of data using linear hyperplanes is frequently challenging. Therefore, the SVM uses the so-called “kernel trick” to transform the original data to a higher-dimensional space. A kernel is a function that maps the data to a higher dimension. After performing numerous trials, for the purpose of this paper, we decided to use a polynomial kernel.
The statistical analysis was performed using R (Version 3.6.1) and RStudio (Version 1.1.463, RStudio, Inc., Boston, MA, USA) to determine whether the obtained accuracies of the classifiers were significantly above chance. Since the data did not meet the criterion of normal distribution, a non-parametric one-sample Wilcoxon signed-rank test with Bonferroni correction for multiple comparisons was used. p < 0.05 was regarded as significant.

4.2. Experimental Results

In a standard oddball experiment, several ERPs are generated, of which the most important is the P3b waveform, related to attentional processes. It is a broad positive peak, from around 300 to 750 ms, with maximal amplitude around 350–500 ms after the stimulus onset. Other waveforms (N1, P2, and N2) are negative (N1 and N2) and positive (P2) peaks, reflecting sensory and early attention processes, and are observed between 100 and 250 ms post-stimulus (Figure 3).
At the first step of the proposed approach, we performed a series of experiments to select the time segment for which the obtained classification accuracy is the best. Table 2 shows accuracies for each time segment for four analyzed classifiers: 3-NN, NB, RF, and SVM.
Since this study used a binary classification, we were interested in those accuracies that were above chance, i.e., 0.5. Moreover, due to their high temporal resolution, ERPs are the neural correlates of the dynamical cognitive processes occurring in the brain. Therefore, we can specify the time window for which the ERP signal is meaningful, i.e., it maps true brain electrical activity. Since the information in the visual modality in this paradigm is processed not earlier than after the first 100 ms, and in this experiment, we used an oddball paradigm to evoke the P3 waveform, which extends until around 750 ms, and we should expect meaningful results within 100–800 ms. All classifiers achieved high accuracy at the time segments for which statistical significance by means of non-parametric permutation cluster-based analysis was obtained in our previous experiment, i.e., approximately 540–710 ms [40], which corresponds to the segment numbers 14–16. Moreover, two of these classifiers achieved their maximal accuracies among all segments within this time window: γ = 0.8 and 0.85 for NB and RF, respectively. Both of these accuracies were observed for segment 16 (650–700 ms). This shows an agreement of the classification performed in this study with the statistical analysis performed in our previous work. Interestingly, however, 3-NN, NB and RF classifiers also achieved high accuracies for segment 9 (300–350 ms), which corresponds to the time window when the P3 waveform rises and reaches the maximal amplitude. NB and RF classifiers were especially good because the performance of NB classifier reached its highest accuracy ( γ = 0.8 ) for this time segment among all segments, and the RF classifier achieved accuracy γ = 0.75 . To better visualize the high accuracies achieved within the neurobiologically interesting segments, the calculated accuracies are presented in Figure 4.
To summarize, we achieved successful gender classification based on the ERP waveform from the attention task. We obtained high accuracies in the neurobiologically meaningful time windows. In particular, the highest accuracies (0.8 and 0.85) were observed for segments, where a previous analysis revealed statistical differences between men and women, and for the additional segment, for which statistical analysis failed to find differences. This shows that our hierarchical classification approach is capable of recognizing the subject’s gender based on neurophysiological signal.
The next point was to compare the performance of the four studies classifier and to indicate the best one. The discussion of our results from the previous paragraph suggests 3-NN, NB, and RF might be good candidates. Therefore, we next compared the number of segments within the meaningful 100–800 ms time window, for which each classifier achieved accuracy above chance. As it can be seen in Figure 4, among 14 segments within this time window, the number of segments, for which γ > 0.5 was: 9, 10, 9, and 6 for 3-NN, NB, RF, and SVM classifier, respectively.
Moreover, the fact that classification was highly accurate, especially for two segments discussed above (9 and 16) is interesting, because it shows an agreement with previous work (later time segment), but in addition, it also suggests there are gender-related differences in the earlier time window of the P3 waveform, which were not observed by statistical analysis, even using an advanced calculation method. Therefore, to further explore the classification performance within those two segments (given in columns of the table), we present in Table 3 and Table 4 the detailed results for each experimentation trial of the leave-one-out cross validation, for each used classifier. The number of trial corresponds to the person p for who the testing was performed.
Having detailed information about the performance of each classifier for each participant, we checked whether the accuracies of the analyzed classifiers are above chance (Figure 5).
The statistical analysis revealed that an average accuracy for both segments: 9 (300–350 ms) and segment 16 (600–650 ms) was above chance only for NB ( V 19 = 187 , Pcorr = 0.009 and V 19 = 187 , Pcorr = 0.006 for segments 9 and 16, respectively) and RF (( V 19 = 183 , Pcorr = 0.014 and V 19 = 198 , Pcorr = 0.001 for segments 9 and 16, respectively) classifiers, while not significantly different from the chance level for 3-NN (( V 19 = 148 , Pcorr = 0.422 and V 19 = 152 , Pcorr = 0.298 for segments 9 and 16, respectively) and SVM (( V 19 = 103 , Pcorr = 1.000 and V 19 = 151 , Pcorr = 0.353 for segments 9 and 16, respectively) classifiers.
We also calculated the proportion of the participants, for which the accuracy of each classifier was above chance (>0.5) for segment 9 (300–350 ms) and 16 (650–700 ms). These proportions were: 0.6, 0.85, 0.8, and 0.5 for 3-NN, NB, RF, and SVM, respectively within segment 9, and: 0.8, 0.8, 0.8, and 0.6 for 3-NN, NB, RF, and SVM, respectively within segment 16. From these analyses and the comparison of the highest accuracies reached by these classifiers, discussed earlier, we can see that 3-NN, NB, and RF seem to perform much better than SVM and that NB and RF seem to be the best ones.
As it can be seen in Table 3 and Table 4, the classification rate ( γ ) for a few participants was 0.0, i.e., the signal was not classified correctly for these segments. This means that individual ERP differences for these participants were more similar to the ERP differences in the opposite gender class. One of the reasons may be the nature of the ERP signal, i.e., the fact that it is an average of all trials measured time-locked to all of the presented stimuli. Therefore, variation from trial to trial of the single-trial waveform introduces distortion to the averaged waveform. The variability may also reflect stable individual differences among the subjects, e.g., due to differences in the pattern of individual folding of the cortex or neural processes. As an example, we present individual ERP of subject s4, who had the largest number of γ = 0.0 (three out of eight) among both segments and all four classifiers. Figure 6 shows an individual ERP difference waveform from subject s4, compared with grand averaged ERP differences for men (blue) and women (orange).
The amplitude of the ERP trace for this subject is higher and noisier than both the grand-averages, for a simple reason that the grand average is the average of the voltage values of all participants, which are higher for some of them and lower for others. Therefore the grand averages are usually smaller than the most single-subject ERPs. These individual differences may cause the fact that an ERP falls closer to the opposite gender class for some individuals. Large inter- and intra- variability due to individual differences is a common issue in all studies performed on humans. However, our results show that even with a sample size as small as 20, we were able to obtain high gender classification accuracies.
To summarize, researchers recently have focused on developing new classification methods to improve neurophysiological signal analysis during rest and cognitive processes in normal and pathological conditions. However, decoding physiological signals is not trivial, especially among healthy participants, because differences in normal cognitive processes are less distinctive than differences between normal and pathological signals. One of the most intriguing attempts is gender recognition. Most gender recognition methods found in the literature focused on the external physiological features, where the recognition accuracy was poor, and classification efficiency was affected by feature extractors and classification algorithms [20]. A neurophysiological signal is a good alternative due to gender-related anatomical and/or functional differences. Unfortunately, due to large individual differences and the small effect size of gender differences, there is still a lack of powerful and sensitive methods. Some of the previous studies used EEG signal for gender classification [19,20,21,22,23,24,25,26,27]. However, these studies were based on resting-state EEG recordings, which do not capture gender-related differences in cognitive processes, such as attention. Two other studies [37,38] examined the EEG-based gender recognition related to emotions. However, to the best of our knowledge, there are no findings on gender-recognition method based on ERPs in attention.
Therefore, in this paper, we present a hierarchical approach for gender classification based on ERP signal collected during an attention task. This approach consisted of four-step-process, which covered data acquisition, data pre-processing, time series segmentation, and bottom level classification. One of the most important steps was data segmentation, which allowed us to divide the whole ERP epoch into smaller time segments. The highest accuracies obtained by the classifiers were observed for two segments within the timing of P3 waveform, being a neuronal correlated of attention: 9, which corresponded to 300–350 ms time interval, and 16, which corresponded to 650–700 ms time interval. Our previous work, which focused on understanding the underlying neural processes of attention, indicated possible gender-related differences in the speed and character of the P3 waveform. The effect presented here was observed as higher ERP amplitude in men than in women in the earlier time window (300–350 ms) and higher ERP amplitude in women than in men in the later time window (650–700 ms), which suggests that gender-related differences evoked in visual two-stimulus oddball paradigm are complex, and include changes in ERP waveforms generation, and distribution and suppression across the scalp, related to the attention process. The latter effect was in agreement with our previous statistical analysis, which showed significant differences between men and women within this time window. More interestingly, the classification rate was high also for the time window 300–350 ms, i.e., when the P3 waveform was rising and reaching the maximal amplitude. This shows the potential of our ERP-based method for gender classification. Moreover, comparisons of the performance of the most frequently used classifiers for the EEG signals (k-NN, NB, RF, and SVM) suggested the candidates that outperformed the others. In our future work, we are going to further explore the effectiveness of our approach for the classification of ERP signals.

5. Conclusions

In this paper, we proposed a new combined empirical and theoretical approach for an efficient gender recognition based on ERPs, in order to better understand neuronal differences between men’s and women’s brains for the attention task. We developed a new method for the hierarchical classification of ERPs, and broke the analysis into four steps: data acquisition, data pre-processing, time series segmentation, and bottom level classification. We ultimately formed a hierarchy of data processing steps, i.e., we formed a chain of classifiers M M 1 M 2 M 3 , where each classifier is related to each step of the analysis process. Utilizing this, we transformed the addressed problem in a way that on its bottom level (classifier M 3 ) we were able to use a standard, state-of-the-art classifier known from the literature.
The proposed approach now enables us to recognize gender solely using the ERP signal measured in the attention task. As the experiments revealed, the proposed method is highly effective. It indicates gender differences in the attention task, which brings us additional knowledge on the differences between men and women in the functionality of the human brain during such an important cognitive process as attention. What is particularly important is that the classification approach not only correlated with the results obtained in our previous work, where a nonparametric permutation cluster-based analysis was used, but it also outperformed that analysis, and highlights the usefulness of ERP-based classification in electrophysiological signal analysis.
The main limitation of the proposed classification method is its dependency on the high number of parameters that have to be determined on each level of data processing and a small sample size. Further work is necessary towards the automatic adjustment of those parameters, as well as investigating other pre-processing and analysis approaches, e.g., ICA for artifact correction or comparing the results with repeated measures of ANOVA run on the 50 ms time bins.
This work has possible extensions to clinical studies and individual differences. Our focus on gender differences aimed at better understanding neuronal differences between men’s and women’s brains, as a fundamental factor in health and biomedical research at a basic level, but also in preclinical, clinical, and drug studies, which has far-reaching implications in reducing the misdiagnosis and adverse side effects in the under-represented women group. As a more sensitive method to capture gender-related differences in brain electrical activity, it may be a step towards minimizing the under-representation of women in the clinical studies, which in turn may result in smaller health risks, rates of misdiagnosis, and adverse side effects from drug treatment in women.

Author Contributions

Conceptualization, K.M. and W.F.; methodology, K.M. and W.F.; software, K.M. and W.F.; validation, K.M. and W.F.; formal analysis, K.M. and W.F.; investigation, K.M. and W.F.; resources, K.M.; data curation, K.M.; writing—original draft preparation, K.M. and W.F.; writing—review and editing, K.M. and W.F.; visualization, K.M.; supervision, W.F.; project administration, K.M.; funding acquisition, W.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This was a retrospective study conducted on previously acquired data described in [40], according to the guidelines of the Declaration of Helsinki, and approved by the Ethics Committee of University of Silesia on scientific studies conducted on humans (approval number: 1/2018, date of approval: 2018).

Informed Consent Statement

All subjects gave their informed consent for inclusion before they participated in the study.

Data Availability Statement

The data presented in this study are openly available in OSF.IO at https://osf.io/ef7uv/ (accessed on 19 November 2021), reference number [DOI 10.17605/OSF.IO/EF7UV].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nielsen, M.W.; Stefanick, M.L.; Peragine, D.; Neilands, T.B.; Ioannidis, J.P.A.; Pilote, L.; Prochaska, J.J.; Cullen, M.R.; Einstein, G.; Klinge, I.; et al. Gender-related variables for health research. Biol. Sex Differ. 2021, 12, 23. [Google Scholar] [CrossRef]
  2. Rich-Edwards, J.W.; Kaiser, U.B.; Chen, G.L.; Manson, J.E.; Goldstein, J.M. Sex and Gender Differences Research Design for Basic, Clinical, and Population Studies: Essentials for Investigators. Endocr. Rev. 2018, 39, 424–439. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Tannenbaum, C.; Ellis, R.P.; Eyssel, F.; Zou, J.; Schiebinger, L. Sex and gender analysis improves science and engineering. Nature 2019, 575, 137–146. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Thompson, K.; Peters, S.; Woodward, M.; Carcel, C.; Norton, R. Reporting sex and gender in medical research. Rep. Sex Gend. Med. Res. 2019, 393, 2038. [Google Scholar] [CrossRef] [Green Version]
  5. Luck, S.J. An Introduction to the Event-Related Potential Technique; MIT Press: Cambridge, MA, USA, 2014. [Google Scholar]
  6. Conroy, M.A.; Polich, J. Normative variation of p3a and p3b from a large sample. J. Psychophysiol. 2007, 21, 22–32. [Google Scholar] [CrossRef]
  7. Hoffman, L.D.; Polich, J. P300, handedness, and corpus callosal size: Gender, modality, and task. Int. J. Psychophysiol. 1999, 31, 163–174. [Google Scholar] [CrossRef]
  8. Jaušovec, N.; Jaušovec, K. Gender related differences in visual and auditory processing of verbal and figural tasks. Brain Res. 2009, 1300, 135–145. [Google Scholar] [CrossRef]
  9. Jaušovec, N.; Jaušovec, K. Do women see things differently than men do? NeuroImage 2009, 45, 198–207. [Google Scholar] [CrossRef]
  10. Morita, Y.; Morita, K.; Yamamoto, M.; Waseda, Y.; Maeda, H. Effects of facial affect recognition on the auditory P300 in healthy subjects. Neurosci. Res. 2001, 41, 89–95. [Google Scholar] [CrossRef]
  11. Sangal, R.B.; Sangal, J.M. Topography of auditory and visual P300 in normal adults. Clin. Electroencephalogr. 1996, 27, 145–150. [Google Scholar] [CrossRef]
  12. Shelton, P.P.; Hartmann, A.M.; Allen, J. Seasonal photoperiod, gender, and P300. Biol. Psychol. 2002, 60, 151–171. [Google Scholar] [CrossRef]
  13. Steffensen, S.C.; Ohran, A.J.; Shipp, D.N.; Hales, K.; Stobbs, S.H.; Fleming, D.E. Gender-selective effects of the P300 and N400 components of the visual evoked potential. Vis. Res. 2008, 48, 917–925. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Tsolaki, A.; Kosmidou, V.; Hadjileontiadis, L.; Kompatsiaris, I.Y.; Tsolaki, M. Brain source localization of MMN, P300 and N400: Aging and gender differences. Brain Res. 2015, 1603, 32–49. [Google Scholar] [CrossRef] [PubMed]
  15. Vaquero, E.; Cardoso, M.J.; VÁZque, M.; GÓMez, C.M. Gender Differences in Event-Related Potentials during Visual-Spatial Attention. Int. J. Neurosci. 2004, 114, 541–557. [Google Scholar] [CrossRef]
  16. Polich, J.; Herbst, K.L. P300 as a clinical assay: Rationale, evaluation, and findings. Int. J. Psychophysiol. 2000, 38, 3–19. [Google Scholar] [CrossRef]
  17. Lukács, G.; Weiss, B.; Dalos, V.D.; Kilencz, T.; Tudja, S.; Csifcsák, G. The first independent study on the complex trial protocol version of the P300-based concealed information test: Corroboration of previous findings and highlights on vulnerabilities. Int. J. Psychophysiol. 2016, 110, 56–65. [Google Scholar] [CrossRef] [Green Version]
  18. Lin, F.; Wu, Y.; Zhuang, Y. Human Gender Classification: A Review. Int. J. Biom. 2017, 8, 275–300. [Google Scholar] [CrossRef]
  19. Ghani, S.A.; Zaini, N.; Norhazman, H.; Yassin, I.M.; Sani, M.M. Classification of Frontal EEG Signals of Normal Subjects to Differentiate Gender by Using Artificial Neural Network. J. Telecommun. Electron. Comput. Eng. JTEC 2017, 9, 139–143. [Google Scholar]
  20. Hu, J. An approach to EEG-based gender recognition using entropy measurement methods. Knowl.-Based Syst. 2018, 140, 134–141. [Google Scholar] [CrossRef]
  21. Kaur, B.; Singh, D.; Roy, P.P. Age and gender classification using brain–computer interface. Neural Comput. Appl. 2019, 31, 5887–5900. [Google Scholar] [CrossRef]
  22. Kaushik, P.; Gupta, A.; Roy, P.P.; Dogra, D.P. EEG-Based Age and Gender Prediction Using Deep BLSTM-LSTM Network Model. IEEE Sens. J. 2019, 19, 2634–2641. [Google Scholar] [CrossRef]
  23. Li, J.; Deng, J.D.; De Ridder, D.; Adhia, D. Gender classification of EEG signals using a motif attribute classification ensemble. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  24. Nguyen, P.; Tran, D.; Huang, X.; Ma, W. Age and gender classification using EEG paralinguistic features. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 1295–1298. [Google Scholar]
  25. Nguyen, P.; Tran, D.; Vo, T.; Huang, X.; Ma, W.; Phung, D. EEG-Based age and gender recognition using tensor decomposition and speech features. In Neural Information Processing; Springer: Berlin/Heidelberg, Germany, 2013; pp. 632–639. [Google Scholar]
  26. Van Putten, M.J.A.M.; Olbrich, S.; Arns, M. Predicting sex from brain rhythms with deep learning. Sci. Rep. 2018, 8, 3069. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Wang, P.; Hu, J. A hybrid model for EEG-based gender recognition. Cogn. Neurodyn. 2019, 13, 541–554. [Google Scholar] [CrossRef] [Green Version]
  28. Zheng, X.; Cao, Z.; Bai, Q. An evoked potential-guided deep learning brain representation for visual classification. In Neural Information Processing—27th International Conference, ICONIP 2020; Communications in Computer and Information Science; Springer: Berlin/Heidelberg, Germany, 2020; Volume 1333, pp. 54–61. [Google Scholar]
  29. Alam, M.E. Electroencephalographic Signal Processing and Classification Techniques for Noninvasive Motor Imagery Based Brain Computer Interface. Electron. Theses Diss. 2017, 1591, 1–98. [Google Scholar]
  30. Bae, X.G.Y.; Luck, S.J. Dissociable Decoding of Spatial Attention and Working Memory from EEG Oscillations and Sustained Potentials. J. Neurosci. 2018, 38, 409–422. [Google Scholar] [CrossRef] [Green Version]
  31. Bae, G.Y.; Luck, S.J. Reactivation of previous experiences in a working memory task. Psychol. Sci. 2019, 30, 587–595. [Google Scholar] [CrossRef]
  32. Al-Nafjan, A.; Alharthi, K.; Kurdi, H. Lightweight Building of an Electroencephalogram- Based Emotion Detection System. Brain Sci. 2020, 10, 781. [Google Scholar] [CrossRef]
  33. Bajaj, V.; Pachori, R.B. Human Emotion Classification from EEG Signals Using Multiwavelet Transform. In Proceedings of the 2014 International Conference on Medical Biometrics, Shenzhen, China, 30 May–1 June 2014; pp. 125–130. [Google Scholar]
  34. Jalilifard, A.; Pizzolato, E.B.; Islam, M.K. Emotion classification using single-channel scalp-EEG recording. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 845–849. [Google Scholar]
  35. Vaid, S.; Singh, P.; Kaur, C. Classification of Human Emotions using Multiwavelet Transform based Features and Random Forest Technique. Indian J. Sci. Technol. 2015, 8, 1–7. [Google Scholar] [CrossRef]
  36. Aldayel, M.; Ykhlef, M.; Al-Nafjan, A. Deep learning for EEG-Based preference classification in neuromarketing. Appl. Sci. 2020, 10, 1525. [Google Scholar] [CrossRef] [Green Version]
  37. De La Pava, I.; Álvarez, A.; Herrera, P.; Castellanos-Dominguez, G.; Orozco, A. Gender Effects on an EEG-Based Emotion Level Classification System. In Proceedings of the Iberoamerican Congress on Pattern Recognition, Madrid, Spain, 19–22 November 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 810–819. [Google Scholar]
  38. Bilalpur, M.; Kia, S.M.; Chawla, M.; Chua, T.S.; Subramanian, R. Gender and Emotion Recognition with Implicit User Signals. In Proceedings of the ICMI ’17: Proceedings of the 19th ACM International Conference on Multimodal Interaction, Glasgow, UK, 13–17 November 2017; pp. 379–387. [Google Scholar]
  39. Craik, A.; He, Y.; Contreras-Vidal, J. Deep learning for Electroencephalogram (EEG) classification tasks: A review. J. Neural Eng. 2019, 16, 031001. [Google Scholar] [CrossRef]
  40. Maciejewska, K.; Drzazga, Z. Differences in spatio-temporal distribution of the visual P3b event-related potential between young men and women. Acta Neurobiol. Exp. 2019, 79, 25–38. [Google Scholar] [CrossRef] [Green Version]
  41. Duncan, C.C.; Barry, R.J.; Connolly, J.F.; Fischer, C.; Michie, P.T.; Näätänen, R.; Polich, J.; Reinvang, I.; Van Petten, C. Event-related potentials in clinical research: Guidelines for eliciting, recording, and quantifying mismatch negativity, P300, and N400. Clin. Neurophysiol. 2009, 120, 1883–1908. [Google Scholar] [CrossRef] [PubMed]
  42. Pontifex, M.; Gwizdala, K.; Parks, A.; Billinger, M.; Brunner, C. Variability of ICA decomposition may impact EEG signals when used to remove eyeblink artifacts. Psychophysiology 2017, 54, 386–398. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Morabito, F. Independent component analysis and feature extraction techniques for NDT data. Mater. Eval. 2000, 58, 85–92. [Google Scholar]
  44. Calhoun, V.D.; Liu, J.; Adalı, T. A review of group ICA for fMRI data and ICA for joint inference of imaging, genetic, and ERP data. NeuroImage 2009, 45, S163–S172. [Google Scholar] [CrossRef] [Green Version]
  45. Fernandez-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we Need Hundreds of Classifiers to Solve Real World Classification Problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  46. Zhang, Z. Introduction to machine learning: K-nearest neighbors. Ann. Transl. Med. 2016, 4, 218. [Google Scholar] [CrossRef] [Green Version]
  47. Russell, S.; Norvig, P. Artificial Intelligence: A Modern Approach, 3rd ed.; Prentice Hall: Hoboken, NJ, USA, 2010. [Google Scholar]
  48. Ho, T.K. Random Decision Forests. In Proceedings of the Third International Conference on Document Analysis and Recognition (ICDAR’95), Montreal, QC, Canada, 14–16 August 1995; IEEE Computer Society: Washington, DC, USA, 1995; Volume 1, p. 278. [Google Scholar]
  49. Cortes, C.; Vapnik, V. Support Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
Figure 1. Event-related potentials (ERPs) signal pre-processing workflow: (a) raw electroencephalography (EEG) signal, (b) filtered EEG signal, (c) EEG signal after blink correction, (d) EEG signal after baseline correction, epoching and artifact detection, (e) averaged ERPs to target, and standard stimuli, (f) averaged ERP target-standard difference waveform. Please note the reversed polarity of the signal in this figure.
Figure 1. Event-related potentials (ERPs) signal pre-processing workflow: (a) raw electroencephalography (EEG) signal, (b) filtered EEG signal, (c) EEG signal after blink correction, (d) EEG signal after baseline correction, epoching and artifact detection, (e) averaged ERPs to target, and standard stimuli, (f) averaged ERP target-standard difference waveform. Please note the reversed polarity of the signal in this figure.
Entropy 23 01547 g001
Figure 2. Epoched data specification (a) and segment specification (b). Panel (a) represents datasets for each person: from t m i n to t m a x timepoints and for d electrodes. Panel (b) represents the exemplary k-th segment.
Figure 2. Epoched data specification (a) and segment specification (b). Panel (a) represents datasets for each person: from t m i n to t m a x timepoints and for d electrodes. Panel (b) represents the exemplary k-th segment.
Entropy 23 01547 g002
Figure 3. Grand averaged target-standard difference ERPs for men (blue) and women (orange) at an exemplification electrode site Pz (midline parietal). The bottom time scale represents the length of the epoch used for data averaging (−100 to 1000 ms with −100–0 ms baseline). The top time scale represents the number of time segments (1–22) corresponding to 50-ms time bins.
Figure 3. Grand averaged target-standard difference ERPs for men (blue) and women (orange) at an exemplification electrode site Pz (midline parietal). The bottom time scale represents the length of the epoch used for data averaging (−100 to 1000 ms with −100–0 ms baseline). The top time scale represents the number of time segments (1–22) corresponding to 50-ms time bins.
Entropy 23 01547 g003
Figure 4. Classification accuracies for each analyzed classifier: (a) Near-Neighbor (kNN), (b) Naive Bayes (NB), (c) Random Forrest (RF), and (d) Support Vector Machine (SVM), plotted for all segments of the epoched ERP signal.
Figure 4. Classification accuracies for each analyzed classifier: (a) Near-Neighbor (kNN), (b) Naive Bayes (NB), (c) Random Forrest (RF), and (d) Support Vector Machine (SVM), plotted for all segments of the epoched ERP signal.
Entropy 23 01547 g004
Figure 5. A distribution of classification rates ( γ ) for segment 9 (300–350 ms) and segment 16 (600–650 ms). The red horizontal line shows the chance level (0.5).
Figure 5. A distribution of classification rates ( γ ) for segment 9 (300–350 ms) and segment 16 (600–650 ms). The red horizontal line shows the chance level (0.5).
Entropy 23 01547 g005
Figure 6. Individual ERP difference waveform from subject s4 (red) compared with averaged ERP difference waveforms for men (blue) women (orange) at electrode site P4 (right parietal).
Figure 6. Individual ERP difference waveform from subject s4 (red) compared with averaged ERP difference waveforms for men (blue) women (orange) at electrode site P4 (right parietal).
Entropy 23 01547 g006
Table 1. Parameters of the experiments.
Table 1. Parameters of the experiments.
ParameterDescriptionValue
p m a x number of participants20
dnumber of electrodes32
n m a x number of segments22
Table 2. The values of γ ( { W p ( t 1 , t } ) parameters obtained for each analyzed classifier (Near-Neighbor: k-NN, Naive Bayes: NB, Random Forrest: RF, and Support Vector Machine: SVM), for each segment of the epoched ERP signal. The highest accuracy among all the segments of each classifier is marked in bold.
Table 2. The values of γ ( { W p ( t 1 , t } ) parameters obtained for each analyzed classifier (Near-Neighbor: k-NN, Naive Bayes: NB, Random Forrest: RF, and Support Vector Machine: SVM), for each segment of the epoched ERP signal. The highest accuracy among all the segments of each classifier is marked in bold.
k τ 3-NNNBRFSVM
1−100–500.40.350.150.4
2−50–00.550.450.450.2
30–500.350.40.30.45
450–1000.350.650.350.4
5100–1500.40.650.550.4
6150–2000.50.50.20.35
7200–2500.550.50.50.35
8250–3000.550.350.550.55
9300–3500.650.80.750.5
10350–4000.50.60.70.45
11400–4500.70.650.550.5
12450–5000.550.750.50.55
13500–5500.60.60.50.7
14550–6000.50.650.550.75
15600–6500.60.70.750.6
16650–7000.650.80.850.65
17700–7500.650.70.450.45
18750–8000.50.50.550.15
19800–8500.250.40.30.25
20850–9000.40.550.50.5
21900–9500.50.60.20.8
22950–10000.650.70.50.65
Table 3. Classification rate ( γ ) for each experimentation trial of the leave-one-out cross validation, for each used classifier for segment 9 (300–350 [ms]). The number of the trial corresponds to the person p for who the testing was performed.
Table 3. Classification rate ( γ ) for each experimentation trial of the leave-one-out cross validation, for each used classifier for segment 9 (300–350 [ms]). The number of the trial corresponds to the person p for who the testing was performed.
Person3-NNNBRFSVM
11.00.921.00.69
21.01.01.00.0
31.00.921.00.85
40.231.01.00.0
51.01.01.00.0
60.920.920.690.0
71.01.01.00.08
80.380.540.770.54
90.00.230.150.0
100.920.920.850.08
110.461.01.01.0
120.850.850.920.15
130.00.850.540.0
140.00.770.920.92
150.770.690.380.31
161.00.770.851.0
171.00.920.381.0
181.01.01.01.0
190.150.461.01.0
200.310.00.01.0
Table 4. Classification rate ( γ ) for each experimentation trial of the leave-one-out cross validation, for each used classifier for segment 16 (650–700 [ms]). The number of the trial corresponds to the person p for who the testing was performed.
Table 4. Classification rate ( γ ) for each experimentation trial of the leave-one-out cross validation, for each used classifier for segment 16 (650–700 [ms]). The number of the trial corresponds to the person p for who the testing was performed.
Person3-NNNBRFSVM
10.921.01.00.92
20.851.01.00.69
30.620.230.310.0
40.00.380.620.0
51.01.01.00.23
60.620.00.310.38
70.540.690.460.31
81.01.01.01.0
91.01.01.01.0
100.621.01.00.92
110.621.01.00.08
120.01.01.01.0
130.621.01.00.77
141.01.01.01.0
151.01.01.01.0
161.01.01.01.0
171.00.620.920.31
181.01.01.01.0
190.01.01.01.0
200.00.080.380.15
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Maciejewska, K.; Froelich, W. Hierarchical Classification of Event-Related Potentials for the Recognition of Gender Differences in the Attention Task. Entropy 2021, 23, 1547. https://doi.org/10.3390/e23111547

AMA Style

Maciejewska K, Froelich W. Hierarchical Classification of Event-Related Potentials for the Recognition of Gender Differences in the Attention Task. Entropy. 2021; 23(11):1547. https://doi.org/10.3390/e23111547

Chicago/Turabian Style

Maciejewska, Karina, and Wojciech Froelich. 2021. "Hierarchical Classification of Event-Related Potentials for the Recognition of Gender Differences in the Attention Task" Entropy 23, no. 11: 1547. https://doi.org/10.3390/e23111547

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop