Next Article in Journal
Exploring Ohm’s Law: The Randomness of Determinism
Previous Article in Journal
An Information-Theoretic Framework for Understanding Learning and Choice Under Uncertainty
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Seizure Type Classification Based on Hybrid Feature Engineering and Mutual Information Analysis Using Electroencephalogram

College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
Entropy 2025, 27(10), 1057; https://doi.org/10.3390/e27101057
Submission received: 23 August 2025 / Revised: 6 October 2025 / Accepted: 10 October 2025 / Published: 11 October 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

Epilepsy has diverse seizure types that challenge diagnosis and treatment, requiring automated and accurate classification to improve patient outcomes. Traditional electroencephalogram (EEG)-based diagnosis relies on manual interpretation, which is subjective and inefficient, particularly for multi-class differentiation in imbalanced datasets. This study aims to develop a hybrid framework for automated multi-class seizure type classification using segment-wise EEG processing and multi-band feature engineering to enhance precision and address data challenges. EEG signals from the TUSZ dataset were segmented into 1-s windows with 0.5-s overlaps, followed by the extraction of multi-band features, including statistical measures, sample entropy, wavelet energies, Hurst exponent, and Hjorth parameters. The mutual information (MI) approach was employed to select the optimal features, and seven machine learning models (SVM, KNN, DT, RF, XGBoost, CatBoost, LightGBM) were evaluated via 10-fold stratified cross-validation with a class balancing strategy. The results showed the following: (1) XGBoost achieved the highest performance (accuracy: 0.8710, F1 score: 0.8721, AUC: 0.9797), with γ -band features dominating importance. (2) Confusion matrices indicated robust discrimination but noted overlaps in focal subtypes. This framework advances seizure type classification by integrating multi-band features and the MI method, which offers a scalable and interpretable tool for supporting clinical epilepsy diagnostics.

1. Introduction

Epilepsy is one of the most common chronic neurological disorders characterized by unpredictable recurrent seizures, affecting around 70 million people worldwide [1,2,3]. Seizure types vary widely, ranging from focal to generalized, such as focal non-specific seizure (FNSZ), absence seizure (ABSZ), complex partial seizure (CPSZ), tonic clonic seizure (TCSZ), generalized non-specific seizure (GNSZ), and tonic seizure (TNSZ) [4]. These various seizure types are with distinct pathophysiological mechanisms that necessitate precise classification for effective treatment, including antiepileptic drugs and surgical interventions [5,6]. Accurate classification of seizure types is essential for guiding appropriate treatment strategies [7]. Nonetheless, traditional diagnosis relies heavily on clinical expertise and visual EEG inspection, which is time-consuming, subjective, and prone to variability [8,9,10].
Electroencephalogram (EEG) serves as a primary non-invasive tool for capturing epileptic brain activity, offering high temporal resolution to detect abnormal discharges [11,12,13]. However, manual analysis of EEG signals is inefficient and error-prone. Existing methods often focus on binary classification, especially in seizure detection between seizure and non-seizure or single-band analysis [14,15,16,17,18]. A few studies focus on seizure type classification. One study focused on classifying three seizure types of GNSZ, FNSZ, and TCSZ, demonstrating the feasibility of automated classification by a support vector machine (SVM) [19]. Another work computed statistical features such as mean and kurtosis, using an SVM with quadratic kernels for four-class seizure classification at 95% accuracy [20]. Building upon this foundation, one survey highlighted the emerging role of artificial intelligence in automatic seizure type classification, emphasizing the potential of machine learning algorithms to set benchmarks in multi-class scenarios [21]. Feature extraction techniques have also been a focal point, such as the study that compared wavelet-based methods to improve multi-class seizure classification performance, measured through F1-scores [22]. Despite these advances, challenges persist in multi-class seizure classification, including data imbalance such as FNSZ dominating over TCSZ, artificially constructed features, feature redundancy in high-dimensional EEG data, and limited interpretability of models.
This study addresses these pressing challenges by introducing a novel feature-wise and segment-wise framework for multi-class seizure type classification using hybrid EEG feature engineering and mutual information (MI) analysis. Our approach extracts a 65-dimensional feature set from δ , θ , α , β , and γ bands via segment-wise sliding windowing, incorporating time and frequency domain metrics. MI-based selection refines features to an optimal 30-feature subset, reducing redundancy while highlighting γ -band discriminators critical for epileptiform activity. To mitigate class imbalance without oversampling artifacts, we employ weighted ensemble models, including XGBoost. Rigorous experiments on a comprehensive EEG dataset evaluated seven classifiers including accuracy, F1 score, precision, recall, and AUC, with cross-validation ensuring robustness. The results reveal that our XGBoost-integrated pipeline achieves an exceptional AUC of 0.9797, surpassing existing methods by offering interpretable insights through MI-driven feature visualizations. This advancement provides a transparent, efficient diagnostic tool that enhances accuracy in high-dimensional imbalanced datasets, marking a significant step forward in supporting epilepsy diagnosis and treatment.

2. Materials and Methods

As shown in Figure 1, the proposed framework for multi-class seizure classification processes EEG signals through resampling and bandpass filtering (0.5–100 Hz) to remove noise. Signals are then segmented into 1 s epochs with 50% overlap for enhanced temporal resolution. Hybrid feature engineering extracts multi-band features ( δ , θ , α , β , and γ bands), including statistical measures, sample entropy, wavelet energies, Hurst exponents, and Hjorth parameters. Extracted features are followed by MI analysis to select the top 30 discriminative features to reduce dimensionality. Seven machine learning classifiers were evaluated using 10-fold stratified cross-validation, with classifiers incorporating a class weight strategy to address data imbalance, ensuring robust performance across all seizure types.

2.1. EEG Recordings and Preprocessing

This study employed the Temple University Hospital Seizure Corpus (TUSZ), a publicly available dataset containing labeled EEG recordings of epileptic seizures [23,24]. In this study, the EEG recordings of a total of 472 seizures for six seizure types were selected to be analyzed, which consisted of FNSZ, ABSZ, CPSZ, TCSZ, GNSZ, and TNSZ. The selection was guided by criteria aimed at achieving balanced representation and computational feasibility: (1) Prioritizing as many files as possible while capping at levels that prevented over-representation of abundant types. In our study, the selection included 100 files each for the more prevalent seizure types FNSZ, CPSZ, and GNSZ, while fully incorporating the less common types ABSZ (78 files), TCSZ (47 files), and TNSZ (47 files) to ensure comprehensive representation. (2) Ensuring multi-patient diversity to minimize biases. (3) Focusing on various durations sufficient for feature extraction, typically 14.66 s–1659.00 s, as summarized in Table 1. Each seizure EEG signal was segmented using a 1 s sliding window with a 0.5 s step size, increasing the sample size for robust analysis. The sampling frequency was resampled to 250 Hz, consistent with the dataset specifications. Detailed dataset statistics, including patient counts, seizure counts, total durations, and seizure-specific durations, are summarized in Table 1.

2.2. Multi-Domain Feature Extraction

Ictal EEG signals were decomposed into five sub-bands using the Finite Impulse Response (FIR) filter implemented in the MNE Python library (version 1.10.1, Alexandre, G., et al., Espoo, Finland), which ensures linear phase response and minimal distortion [25]. The five sub-bands consisted of δ (1–4 Hz), θ (4–7 Hz), α (8–12 Hz), β (13–30 Hz), and γ (30–80 Hz). For each band, a total of 13 features in the time domain, frequency domain, and non-linear domain were extracted, which included the features of mean, variance, skewness, kurtosis, sample entropy, five wavelet energies, Hurst Index, and Hjorth parameters (mobility and complexity). Specifically, the feature of mean for multi-band means was used to capture band-specific shifts, enhancing seizure discrimination. Then the filtered data were segmented using a 1 s sliding time window with 50% overlap. This resulted in a 65-dimensional-feature vector per segment. Features of all segments were standardized for subsequent analysis.

2.2.1. Statistical Features

Four statistical features (mean, variance, skewness, and kurtosis) were utilized to evaluate EEG signal distribution in this study [26,27]. The mean was selected to calculate the central tendency of the EEG signal due to variations in mean values during seizures reflecting overall discharge intensity, aiding the differentiation of seizure types. The variance was used to measure EEG signal dispersion. Skewness and kurtosis assessed the asymmetry and peakedness of the signal distribution, respectively, aiding in identifying abnormal patterns and extreme values linked to seizure activity. The four statistical features were defined as
μ = 1 N i = 1 N x i ,
σ 2 = 1 N i = 1 N ( x i μ ) 2 ,
γ 1 = 1 N i = 1 N ( x i μ ) 3 1 N i = 1 N ( x i μ ) 2 3 / 2 ,
γ 2 = 1 N i = 1 N ( x i μ ) 4 1 N i = 1 N ( x i μ ) 2 2 3 ,
where N is the window signal length (250 samples), x i is the EEG data value, μ is the mean, σ 2 is the variance, and γ 1 and γ 2 represent skewness and kurtosis, respectively.

2.2.2. Sample Entropy

Sample entropy is a non-linear feature to quantify the complexity of EEG data, which was selected as epileptic EEG signals are chaotic [28]. It can be defined as
S a m p E n ( m , r , N ) = ln A m ( r ) B m ( r ) ,
where m represents the embedding dimension, r is the tolerance threshold, A m ( r ) denotes the number of template matches for sequences of length m, and B m ( r ) represents the number of distance matches. In this study, we set m = 2 and r = 0.2 σ , where σ is the standard deviation of the EEG signal. These values were selected based on established practices in epilepsy EEG analysis as they balance sensitivity to signal complexity and robustness to noise [29,30].

2.2.3. Wavelet Energy

Wavelet energies are frequency-domain features from wavelet decomposition, which was selected to evaluate localized energy distribution [31]. Multi-band wavelet energies were computed to assess spectral differences. This could be computed via Daubechies-4 (db4) wavelet decomposition of the PyWavelets (version 1.8.0, Lee, G., et al.) library at four levels and was defined as [32]
E j = k = 0 L j 1 c j , k 2 ,
where j is the decomposition level and L j is the length of the coefficient vector at level j.

2.2.4. Hurst Index

Hurst index is a non-linear feature used to quantify the long-range dependence or self-similarity of EEG time series data [33]. It was selected in this study because epileptic EEG signals often exhibit persistent discharge patterns, where H > 0.5 indicates positive long-term correlations. It can be defined as
H = log ( R / S ) log ( τ ) ,
where R / S represents the rescaled range, representing the ratio of the range of the cumulative deviations of the EEG time series to its standard deviation over a given time interval. τ is the time scale, ranging from 2 to a maximum value m a x l a g . m a x l a g was set to 20 to limit the maximum subinterval length to 20 samples in this study.

2.2.5. Hjorth Parameters

Hjorth parameters are dynamic features to evaluate the mobility and complexity of an epileptic EEG waveform [34]. The formula is
M o b i l i t y = V a r ( d i f f 1 ) V a r ( x i ) ,
C o m p l e x i t y = V a r ( d i f f 2 ) V a r ( d i f f 1 ) / M o b i l i t y ,
where d i f f 1 and d i f f 2 represent the first-order difference of two EEG signals. The V a r ( d i f f 1 ) and V a r ( d i f f 2 ) denote the variance of d i f f 1 and d i f f 2 .

2.3. Feature Selection

To address the high dimensionality of the extracted EEG features, which included statistical measures (mean, variance, skewness, kurtosis), sample entropy, wavelet energies, Hurst exponent, and Hjorth parameters (mobility and complexity) across five frequency bands, a filter-based feature selection method utilizing MI was applied. MI quantifies the dependency between each feature X and the target class label Y, capturing both linear and non-linear relationships by measuring the reduction in uncertainty about Y given X [35]. For discrete variables, MI is formally defined as
I ( X ; Y ) = x X y Y p ( x , y ) log p ( x , y ) p ( x ) p ( y ) ,
where p ( x , y ) represents the joint probability mass function and p ( x ) and p ( y ) are the marginal probabilities. In the context of continuous EEG features, the mutual_info_classif estimator from scikit-learn was employed to compute non-parametric MI scores, approximating the joint and marginal distributions via k-nearest neighbors (KNNs) [36]. The SelectKBest algorithm then retained the top k = 30 features with the highest MI scores, empirically chosen to balance model complexity, reduce overfitting, and preserve discriminative power while minimizing computational overhead. This threshold was determined through preliminary experiments, where k = 30 yielded optimal cross-validation performance without significant information loss.

2.4. Classification Models

2.4.1. Support Vector Machine (SVM)

A SVM is a supervised learning classifier, which constructs a hyperplane in a high-dimensional space to maximize the margin between classes while minimizing misclassification errors in a soft margin approach [37,38]. The objective function of a soft-margin SVM is defined as
min 1 2 w 2 + C ξ i ,
where w is the normal vector, C is the regularization parameter, and ξ i are slack variables. In this study, the SVM was configured with a radial basis function (RBF) kernel to handle non-linear separability in multi-class seizure classification. The scikit-learn library was used and the class weight parameter was balanced to mitigate the imbalance inherent in the dataset for multi-class seizure types [36].

2.4.2. K-Nearest Neighbor (KNN)

KNN is a lazy learning algorithm that classifies based on the majority vote of its k closest neighbors in the feature space, typically using Euclidean distance as the metric [39,40], defined as
d ( x , y ) = ( x i y i ) 2 ,
where d is Euclidean distance. In this study, KNN was implemented with a neighbor number of 5, integrated with SMOTE for class balancing during cross-validation folds, following MI feature selection with k = 30 to reduce dimensionality.

2.4.3. Decision Tree (DT)

DT is a tree-based model that recursively splits data based on feature thresholds [41]. It was selected for its interpretability in exploring EEG feature hierarchies. It handles non-linear relationships well but is prone to overfitting, which can be mitigated through pruning or ensemble extensions. In our framework, a DT utilized a ‘balanced’ class weight parameter to address seizure type disparities.

2.4.4. Random Forest (RF)

RF is an ensemble method that aggregates multiple decision trees trained on bootstrapped subsets with random feature sampling. It was chosen to reduce overfitting and improve robustness via majority voting [42]. It provides inherent feature importance and handles imbalance through class weighting. In this investigation, RF was parameterized with a balanced class weight parameter and evaluated after the selection of MI features. A balanced variant was also tested to further address class disparities. RF was chosen for its resilience to noise in high-dimensional EEG data, effectively modeling interactions among statistical, entropy, and wavelet features across bands.

2.4.5. Light Gradient Boosting Machine (LightGBM)

LightGBM is an efficient gradient boosting implementation using histogram-based splitting and leaf-wise growth for reduced memory and faster training [43,44]. It incorporates GOSS and EFB for handling large datasets and imbalances. Configured with a ‘balanced’ class weight parameter, LightGBM utilized selected features and 10-fold validation. The classifier was selected for its scalability in high-dimensional EEG processing, prioritizing informative instances in band-specific features.

2.4.6. Categorical Boosting (CatBoost)

CatBoost is a gradient boosting algorithm optimized for categorical features, employing ordered boosting and oblivious trees to reduce overfitting and bias [45]. It automatically handles class imbalance with ‘balanced’ weights. In this study, CatBoost was set with balanced auto class weights and integrated with MI selection (k = 30) and cross-validation. It was chosen for its efficiency in processing categorical-like EEG bands without encoding, effectively modeling complexities in entropy and mobility features.

2.4.7. Extreme Gradient Boosting (XGBoost)

XGBoost is a scalable gradient boosting framework that sequentially builds trees to minimize a regularized loss function, incorporating techniques including shrinkage and column subsampling for enhanced generalization [46]. It supports multi-class objectives and handles imbalance via sample weights. Configured here with objective = ‘multi:softprob’ and inverse class frequency weights, XGBoost followed feature selection and 10-fold validation. The model was selected in this study for its superior performance in EEG analysis, capturing non-linear dependencies in features like Hjorth parameters and wavelet energies, and attributing to regularization that prevents overfitting in noisy, imbalanced datasets.

2.5. Evaluation Methodology

The performance of the proposed classifiers was rigorously evaluated using 10-fold stratified cross-validation, ensuring proportional representation of the six seizure classes (ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, TNSZ) across folds to mitigate bias from dataset imbalance. For each fold, models were trained on nine subsets and tested on the held-out subset, with predictions aggregated for overall metrics. Evaluation metrics consisted of accuracy, weighted precision, weighted recall, weighted F1-score, and area under the receiver operating characteristic curve (AUC) [47,48].
Accuracy is an overall performance metric that evaluates the proportion of correct classifications. In this study, it was used to represent the model’s overall reliability in identifying seizure types. The accuracy was defined as
A c c = T P + T N T P + T N + F P + F N
The F1 score is the harmonic mean of precision and recall, used to balance imbalanced classes. In this study, it represented the model’s comprehensive performance on minority seizure types. The F1 score was defined as
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Precision is the accuracy of positive predictions, evaluating the impact of false positives. In this study, it represented reducing misjudgments of seizures and avoiding unnecessary treatment. The formula was
P = T P T P + F P
Recall is the detection rate of true positives and evaluates omission risks. In this study, it represented reducing missed seizures and ensuring timely diagnosis. It was defined as
R = T P T P + F N
AUC is used to evaluate threshold robustness, where represents the overall ability to distinguish multi-class seizures. The AUC is defined as
AUC = 0 1 T P R ( F P R ) d F P R

3. Results

3.1. Classification Results

The performance of seven machine learning models was evaluated using 10-fold stratified cross-validation on processed EEG segments, representing six seizure types: ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, and TNSZ. Initially, features were extracted from five frequency bands ( δ , θ , α , β , and γ ), encompassing statistical measures (mean, variance, skewness, and kurtosis), sample entropy, wavelet energies, Hurst exponent, and Hjorth parameters (mobility and complexity). Subsequently, an MI-based feature selection method was applied to identify and retain the top 30 features, reducing dimensionality and enhancing generalization. Finally, these selected features were utilized for classification by the models. To establish a baseline for comparison, a null model was implemented, which randomly assigned seizure types with equal probability across the six classes, reflecting a naive approach with no feature-based learning. This model was evaluated using the same 10-fold cross-validation setup to provide a reference point for the performance of trained classifiers.
Table 2 presents the five performance metrics results (accuracy, weighted F1 score, precision, recall, and AUC), along with computational time for seven classifiers. Among the evaluated machine learning classifiers, XGBoost emerged as the top performer, achieving an accuracy of 0.8710 ± 0.0027, weighted F1 score of 0.8721 ± 0.0026, precision of 0.8744 ± 0.0025, recall of 0.8710 ± 0.0027, and AUC of 0.9797 ± 0.0007, with a computational time of 69.39s. The ensemble methods CatBoost followed as a close contender, recording an accuracy of 0.8641 ± 0.0034 and an AUC of 0.9789 ± 0.0009. RF also exhibited better classification performance, with an AUC of 0.9782 ± 0.0007. Methods with their efficiency enhanced by histogram-based optimization techniques, such as LightGBM, with an accuracy of 0.8595 ± 0.0030, also demonstrated strong performance. In contrast, simpler models, such as DT with an accuracy of 0.7899 ± 0.0054 and KNN with an accuracy of 0.5256 ± 0.0022, exhibited lower metrics. The SVM yielded the poorest results, with an accuracy of 0.4709 ± 0.0029. The null model, as expected, yielded an accuracy of 0.0149 and an AUC of 0.5000, consistent with random guessing for a six-class problem. This stark contrast highlights the effectiveness of the feature selection and classification approach.
To further validate the choice of sample entropy parameters used in feature extraction, we conducted a sensitivity analysis specifically on the top-performing XGBoost model. This analysis tested variations in m = 1 , 2 and tolerance thresholds r = 0.1 , 0.15 , 0.2 σ to assess their impact on classification performance. The features of sample entropy were re-extracted using these parameter combinations, and XGBoost was retrained and evaluated under the same 10-fold cross-validation setup. The results, presented in Table 3, demonstrate that m = 2 and r = 0.2 σ yielded the highest metrics across accuracy, F1-score, precision, recall, and AUC, confirming the optimality of these parameters for capturing epileptic EEG signal complexity.
Additionally, to determine the optimal number of features k, we evaluated the XGBoost model with k = 10 , 20 , 40 , 50 , 60 . As presented in Table 4, the results indicate that k = 30 achieved a robust performance. While higher k values, such as k = 60 , yielded a slightly better AUC (0.9844 ± 0.0007), the improvement was marginal, with an increase of 0.0047, and the computational time rose from 69.39 s to 97.95 s. Performance improved steadily from k = 10 to k = 30 , but, beyond k = 30 , the benefits diminished, suggesting potential noise or redundancy. The 10-fold cross-validation confirmed that k = 30 provided an optimal balance between model effectiveness and computational efficiency.
Moreover, to further validate the feature extraction process, we investigated the impact of different step sizes on classification performance using the top-performing XGBoost model. Features were extracted with a fixed time window of 1 s and step sizes of 0.2, 0.4, 0.6, and 0.8 s, compared against the original step size of 0.5 s, all evaluated under 10-fold cross-validation. The results, presented in Table 5, confirm that the 0.5 s step size yielded the highest performance metrics (accuracy: 0.8701 ± 0.0027, AUC: 0.9797 ± 0.0007), outperforming the alternative configurations and justifying its selection.

3.2. Confusion Matrices

To enhance the understanding of model discriminability, confusion matrices were constructed for four classifiers with better performance (XGBoost, CatBoost, LightGBM, and RF), providing a detailed representation of true positives, false positives, and misclassifications of segments across the six seizure types. The matrices are presented in Figure 2. For the XGBoost model (Figure 2a), the high diagonal values indicated robust classification performance, with 9282 (82.1%) true positives for FNSZ and 32,927 (94.19%) true positives for GNSZ. However, a notable overlap was observed, with 720 (3.96%) CPSZ instances misclassified as FNSZ, suggesting shared electrophysiological characteristics within focal seizure subtypes. The CatBoost model (Figure 2b) achieved 9573 (84.67%) true positives for FNSZ and 30,805 (88.12%) for GNSZ, with 1012 (5.57%) CPSZ misclassified as FNSZ, further reinforcing the challenge of distinguishing focal seizures. The LightGBM model (Figure 2c) exhibited similar classification patterns, achieving 9561 (84.57%) true positives for FNSZ and 30,664 (87.72%) for GNSZ, but showed increased errors in rare classes, such as ABSZ, with 975 (83.98%) true positives and 42 (3.62%) misclassifications as CPSZ. This highlights the impact of class imbalance on rarer seizure types. In addition, the RF model (Figure 2d) recorded 8951 (79.17%) true positives for FNSZ and 33,020 (94.46%) for GNSZ, with 581 (3.2%) CPSZ instances misclassified as FNSZ, indicating a similar pattern of focal subtype confusion. Across all models, misclassifications predominantly occurred within focal subtypes, suggesting potential benefits from integrating hybrid features to improve separation. Zero values in the matrices, such as those between ABSZ and TNSZ in XGBoost, indicated no observed misclassifications, reflecting strong class-specific discrimination in certain cases.

3.3. Feature Importance Results

The feature importance analysis was conducted by the MI method, providing insight into the discriminative power of EEG features across five frequency bands, as depicted in Figure 3. This heatmap visualizes feature importance scores within a 5 × 13 matrix for five frequency bands and 13 feature types, which include mean, variance, skewness, kurtosis, sample entropy, wavelet energy 0 through wavelet energy 4, Hurst exponent, Hjorth mobility, and Hjorth complexity. The results reveal that features in the γ band dominated the top rankings, with wavelet energy 4 at the γ exhibiting the highest feature importance score of 0.253, followed by variance at the γ band (0.239) and wavelet energy 3 at the γ band (0.213). This suggests that high-frequency oscillatory components, particularly wavelet-based energy measures, are critical for distinguishing seizure types in the EEG dataset. Other notable contributors include variance at the θ band (0.200) and variance (0.189) at the β band, indicating that variance across mid-frequency bands also plays a significant role. Within the γ band, Hjorth complexity (0.155) and mobility (0.155) further underscored the importance of non-linear dynamics in seizure characterization. Conversely, the α and δ bands showed lower feature importance scores, with Hjorth complexity at the α band (0.003) and Hurst at the δ band (0.032) ranking among the least influential, suggesting limited discriminative power in these ranges.
Common top features (e.g., gamma_hjorth_complexity, beta_hjorth_mobility) across ensembles suggested band-specific complexity metrics as robust discriminators, aligning with prior TUSZ studies where gamma features improved multi-class accuracy by emphasizing oscillatory irregularities. This analysis underscores the value of hybrid statistical–nonlinear features for epilepsy classification, potentially guiding feature engineering in future scalable systems.

4. Discussion

This study proposes a comprehensive framework for multi-class seizure classification, focusing on EEG segments from six seizure types (ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, and TNSZ). Multi-band features ( δ , θ , α , β , and γ ) were extracted through segment-wise processing (1 s windows with 0.5 s overlap), encompassing statistical measures (mean, variance, skewness, kurtosis), sample entropy, wavelet energies, Hurst exponents, and Hjorth parameters. Feature-wise analysis was enhanced via MI-based selection to retain the top 30 features, while imbalance handling was addressed through class-weighted sampling. Evaluated via 10-fold stratified cross-validation, XGBoost emerged as the top performer, with an accuracy of 0.8710 ± 0.0027, F1-score of 0.8721 ± 0.0026, and AUC of 0.9797 ± 0.0007.
The classification results demonstrate robust performance in distinguishing six seizure types (ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, TNSZ), particularly with gradient boosting models such as XGBoost and LightGBM, which outperformed baselines in handling imbalanced distributions. Confusion matrices revealed high true positives along diagonals, such as 32,927 for GNSZ in XGBoost, indicating strong generalization for prevalent classes. However, the 1203 CPSZ segments misclassified as FNSZ by XGBoost revealed subtle but discernible overlaps between distinct focal subtypes, suggesting hidden electrophysiological commonalities. Such overlaps may have been due to shared spike-wave patterns in temporal and frontal lobes [49].
Clinically, this implies the potential for precise antiepileptic therapy tailoring, reducing recurrence risks in idiopathic focal epilepsy by enabling early subtype-specific interventions. Conducting feature importance analysis by introducing the MI selection approach highlighted γ -band dominance, with wavelet energy in the γ band scoring 0.253, which underscores high-frequency energy fluctuations as key discriminators of seizure type [50,51]. Multi-band variance and wavelet features in the θ and β bands further contributed to capturing non-stationary signal variations that aligned with epileptiform oscillations for different seizure types.
There are four key limitations included in this study. Firstly, the dataset-specific biases, such as truncation artifacts, potentially inflated performance for common classes. Secondly, the offline nature restricted real-time deployment. Thirdly, feature selection, while effective, may have overlooked rare inter-band interactions. Fourthly, the current analysis relied on preprocessed EEG data with standard artifact rejection techniques such as bandpass filtering, but lacked specific robustness tests against simulated or real-world noise sources such as eye blinks, muscle activity, and electrode drift. This limits the assurance of model reliability in practical, noisy environments. Future work should incorporate larger datasets or raw EEG for improved generalization, focus on real-time analysis, and include comprehensive noise robustness testing with simulated artifacts and real-world scenarios to enhance the model’s applicability across diverse conditions. Furthermore, to maximize XGBoost’s strength in managing imbalanced classes, we plan to develop and test new statistical features designed to better represent the complex patterns in EEG data, particularly for rare seizure types in future work. Preliminary explorations indicate the potential for improved classification performance, and this integrated approach will strengthen the overall robustness and applicability of our method.
The proposed framework represents a significant advancement in epilepsy classification by combining multi-band feature extraction with MI-based selection, tailored to the imbalanced nature of seizure types. This integration, particularly the emphasis on γ -band features, distinguishes our approach from traditional methods that often focus on narrower frequency ranges or simpler feature sets [8]. The superior performance of XGBoost (accuracy: 0.8701 ± 0.0027, AUC: 0.9797 ± 0.0007) underscores the potential of this method to address clinical challenges. However, the clinical need for automated diagnostic tools remains pressing due to the labor-intensive nature of manual EEG interpretation and the risk of missed diagnoses. This highlights the importance of developing robust, automated systems to support neurologists.
In summary, this study developed a feature-wise and segment-wise framework for EEG analysis, leveraging multi-band feature extraction and an MI selection method, optimizing performance in high-dimensional, imbalanced data. This pipeline integrates segment-wise windowing for enhanced temporal resolution with band-specific features to detect frequency-dependent epileptiform traits, addressing class imbalance through weighted ensemble models like XGBoost, which achieved the highest AUC of 0.9797, without relying on oversampling artifacts, which often compromise generalization. Unlike traditional single-band or unsegmented methods, our methodology advances seizure classification by prioritizing interpretability through MI-based feature importance visualizations, highlighting γ -band features as key discriminators. This interpretability facilitates clinical adoption and diagnostic transparency, while the novel 65-dimensional multi-band EEG extraction pipeline, optimized to 30 features, effectively mitigates feature redundancy and imbalance. The comprehensive evaluation of seven models with tailored imbalance optimizations sets a new performance benchmark and provides a robust, automated diagnostic framework that reduces clinical workload and enhances accuracy, particularly in resource-limited settings.

5. Conclusions

This study proposed a novel framework for seizure type classification by integrating multi-band and segment-wise feature extraction with MI-based selection and strategies for handling imbalanced EEG data. Features were extracted across five frequency bands and encompassed statistical measures, sample entropy, wavelet energies, Hurst exponent, and Hjorth parameters. Moreover, the MI approach was employed to identify the most discriminative features for enhanced performance. Seven classifiers underwent rigorous evaluation through 10-fold stratified cross-validation. The results showed that XGBoost exhibited superior classification performance and the critical role of γ -band features in detecting epileptiform activity. Comprehensive assessments further affirmed the framework’s robustness in managing imbalanced datasets, strengthening classification outcomes. This approach provides a scalable and interpretable methodology, offering substantial advancements for clinical epilepsy diagnosis and supporting the development of tailored treatment strategies.

Funding

This work was funded by the Natural Science Foundation of Top Talent of SZTU (grant no. GDRC202427) and Shenzhen Major Science and Technology Projects (grant no. KJZD20240903102717023).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the use of publicly available datasets.

Informed Consent Statement

This investigation utilized publicly available datasets, where informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets analyzed for this study can be found in the TUSZ dataset by visiting the following link: https://isip.piconepress.com/projects/tuh_eeg/ (accessed on 28 September 2024).

Acknowledgments

The author is grateful to the Temple University Hospital Seizure Corpus dataset and anonymous reviewers for their valuable comments.

Conflicts of Interest

The author declares no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
EEGElectroencephalogram
FNSZFocal non-specific seizure
ABSZAbsence seizure
CPSZComplex partial seizure
TCSZTonic clonic seizure
GNSZGeneralized non-specific seizure
TNSZTonic seizure
MIMutual information
TUSZTemple University Hospital Seizure Corpus
SVMSupport vector machine
KNNK-nearest neighbor
DTDecision tree
RFRandom forest
LightGBMLight gradient boosting machine
CatBoostCategorical boosting
XGBoostExtreme gradient boosting
AUCArea under the receiver operating characteristic curve

References

  1. Ono, T.; Galanopoulou, A.S. Epilepsy and epileptic syndrome. Adv. Exp. Med. Biol. 2012, 724, 99–113. [Google Scholar] [PubMed]
  2. Hirtz, D.; Thurman, D.J.; Gwinn-Hardy, K.; Mohamed, M.; Chaudhuri, A.R.; Zalutsky, R. How common are the “common” neurologic disorders? Neurology 2007, 68, 326–337. [Google Scholar] [CrossRef]
  3. Su, Z.; Lu, J.; Shi, Y.; Qi, B.; Guo, Z. Global, regional, and national children and adolescent epilepsy burden, 1990–2021: An analysis based on the global burden of disease study 2021. Ann. Med. 2025, 57, 2534087. [Google Scholar] [CrossRef] [PubMed]
  4. Scheffer, I.E.; Berkovic, S.; Capovilla, G.; Connolly, M.B.; French, J.; Guilhoto, L.; Hirsch, E.; Jain, S.; Mathern, G.W.; Moshé, S.L.; et al. ILAE classification of the epilepsies: Position paper of the ILAE Commission for Classification and Terminology. Epilepsia 2017, 58, 512–521. [Google Scholar] [CrossRef] [PubMed]
  5. Ahmed, A.; Rahman, A.; Shimul, W.; Toufiq, M.; Eyad, J.; Sheema, C.; Khadijah, A.; Ahmed, M. Seizures and Epilepsy: An Overview for UK Medical Students. Cureus 2024, 16, e70654. [Google Scholar] [CrossRef]
  6. Thijs, R.D.; Surges, R.; O’Brien, T.J.; Sander, J.W. Epilepsy in adults. Lancet 2019, 393, 689–701. [Google Scholar] [CrossRef]
  7. Fisher, R.S.; Acevedo, C.; Arzimanoglou, A.; Bogacz, A.; Cross, J.H.; Elger, C.E.; Engel, J., Jr.; Forsgren, L.; French, J.A.; Glynn, M.; et al. ILAE official report: A practical clinical definition of epilepsy. Epilepsia 2014, 55, 475–482. [Google Scholar] [CrossRef]
  8. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adeli, H.; Subha, D.P. Automated EEG-based screening of depression using deep convolutional neural network. Comput. Methods Programs Biomed. 2018, 161, 103–113. [Google Scholar] [CrossRef]
  9. Joshi, R.K.; Kumar, V.; Agrawal, M.; Rao, A.; Mohan, L.; Jayachandra, M.; Pandya, H.J. Spatiotemporal analysis of interictal eeg for automated seizure detection and classification. Biomed. Signal Process. Control. 2023, 79 Pt 1, 104086. [Google Scholar] [CrossRef]
  10. Chawla, D.; Sharma, E.; Rajab, N.; Łajczak, P.; Silva, Y.P.; Baptista, J.M.; Pomianoski, B.W.; Ahmed, A.R.; Majeed, M.w.; de Sousa, Y.G.; et al. Utilizing machine learning techniques for eeg assessment in the diagnosis of epileptic seizures in the brain: A systematic review and meta-analysis. Seizure Eur. J. Epilepsy 2025, 126, 16–23. [Google Scholar] [CrossRef]
  11. Zhang, H.; Zhou, Q.Q.; Chen, H.; Hu, X.Q.; Li, W.G.; Bai, Y.; Han, J.X.; Wang, Y.; Liang, Z.H.; Chen, D.; et al. The applied principles of EEG analysis methods in neuroscience and clinical neurology. Mil. Med. Res. 2023, 10, 67. [Google Scholar] [CrossRef]
  12. Feng, J.; Jia, W.; Li, Z. Electroencephalography: A valuable tool for assessing motor impairment and recovery post-stroke. J. Neurosci. Methods 2025, 422, 110518. [Google Scholar] [CrossRef]
  13. Dabbabi, T.; Bouafif, L.; Cherif, A. A review of non invasive methods of brain activity measurements via EEG signals analysis. In Proceedings of the 2023 IEEE International Conference on Advanced Systems and Emergent Technologies, Hammamet, Tunisia, 29 April–1 May 2023. [Google Scholar]
  14. Siddiqui, M.K.; Morales-Menendez, R.; Huang, X.; Hussain, N. A review of epileptic seizure detection using machine learning classifiers. Brain Inform. 2020, 7, 5. [Google Scholar] [CrossRef] [PubMed]
  15. Mourad, R.; Diab, A.; Merhi, Z.; Khalil, M.; Jeannès, R.L.B. Machine and deep learning methods for epileptic seizure recognition using eeg data: A systematic review. Brain Res. 2025, 1864, 149797. [Google Scholar] [CrossRef] [PubMed]
  16. Wong, S.; Simmons, A.; Rivera-Villicana, J.; Barnett, S.; Sivathamboo, S.; Perucca, P.; Ge, Z.; Kwan, P.; Kuhlmann, L.; Vasa, R.; et al. EEG datasets for seizure detection and prediction-A review. Epilepsia Open 2023, 8, 252–267. [Google Scholar] [CrossRef]
  17. Shoeibi, A.; Khodatars, M.; Ghassemi, N.; Jafari, M.; Acharya, U.R. Epileptic seizures detection using deep learning techniques: A review. Int. J. Environ. Res. Public Health 2021, 18, 15780. [Google Scholar] [CrossRef]
  18. Lo Giudice, M.; Varone, G.; Ieracitano, C.; Mammone, N.; Tripodi, G.G.; Ferlazzo, E.; Gasparini, S.; Aguglia, U.; Morabito, F.C. Permutation Entropy-Based Interpretability of Convolutional Neural Network Models for Interictal EEG Discrimination of Subjects with Epileptic Seizures vs. Psychogenic Non-Epileptic Seizures. Brain Sci. 2022, 24, 102. [Google Scholar] [CrossRef]
  19. Saputro, I.R.D.; Maryati, N.D.; Solihati, S.R.; Wijayanto, I.; Patmasari, R. Seizure type classification on eeg signal using support vector machine. J. Phys. Conf. Ser. 2019, 1201, 012065. [Google Scholar] [CrossRef]
  20. Wijayanto, I.; Hartanto, R.; Nugroho, H.A.; Winduratna, B. Seizure Type Detection in Epileptic EEG Signal using Empirical Mode Decomposition and Support Vector Machine. In Proceedings of the 2019 International Seminar on Intelligent Technology and Its Applications, Surabaya, Indonesia, 28–29 August 2019. [Google Scholar]
  21. Roy, S.; Asif, U.; Tang, J.; Harrer, S. Seizure type classification using eeg signals and machine learning: Setting a benchmark. In Proceedings of the 2020 IEEE Signal Processing in Medicine and Biology Symposium, Iowa City, IA, USA, 20–24 July 2020. [Google Scholar]
  22. Albaqami, H.; Hassan, G.; Datta, A. Comparison of WPD, DWT and DTCWT for Multi-Class Seizure Type Classification. In Proceedings of the 2021 IEEE Signal Processing in Medicine and Biology Symposium, Philadelphia, PA, USA, 4 December 2021. [Google Scholar]
  23. Obeid, I.; Picone, J. The Temple University Hospital EEG Data Corpus. Front. Neurosci. 2016, 10, 196. [Google Scholar] [CrossRef] [PubMed]
  24. Ochal, D.; Rahman, S.; Ferrell, S.; Elseify, T.; Obeid, I.; Picone, J. The Temple University Hospital EEG Corpus: Annotation Guidelines. Inst. Signal Inf. Process. Rep. 2020, 1, 1–28. [Google Scholar]
  25. Alexandre, G.; Martin, L.; Eric, L.; Denis, A.E.; Daniel, S.; Christian, B.; Roman, G.; Mainak, J.; Teon, B.; Lauri, P.; et al. MEG and EEG data analysis with MNE-Python. Front. Neurosci. 2013, 7, 267. [Google Scholar] [CrossRef]
  26. Subasi, A. EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 2007, 32, 1084–1093. [Google Scholar] [CrossRef]
  27. Joanes, D.N.; Gill, C.A. Comparing measures of sample skewness and kurtosis. J. R. Stat. Soc. 2010, 47. [Google Scholar] [CrossRef]
  28. Lake, D.E.; Richman, J.S.; Griffin, M.P.; Moorman, J.R. Sample entropy analysis of neonatal heart rate variability. Am. J. Physiol.-Regul. Integr. Comp. Physiol. 2002, 283, R789. [Google Scholar] [CrossRef] [PubMed]
  29. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol.-Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  30. Acharya, U.R.; Molinari, F.; Sree, S.V.; Chattopadhyay, S.; Ng, K.H.; Suris, J.S. Automated diagnosis of epileptic EEG using entropies. Biomed. Signal Process. Control 2012, 7, 401–408. [Google Scholar] [CrossRef]
  31. Adeli, H.; Zhou, Z.; Dadmehr, N. Analysis of EEG records in an epileptic patient using wavelet transform. J. Neurosci. Methods 2003, 123, 69–87. [Google Scholar] [CrossRef] [PubMed]
  32. Lee, G.; Gommers, R.; Waselewski, F.; Wohlfahrt, K.; O’Leary, A. Pywavelets: A python package for wavelet analysis. J. Open Source Softw. 2019, 4, 01237. [Google Scholar] [CrossRef]
  33. Hurst, H. Long-term storage capacity of reservoirs. Trans. Am. Soc. Civ. Eng. 1951, 116, 770–799. [Google Scholar] [CrossRef]
  34. Hjorth, B. EEG analysis based on time series properties. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 306–310. [Google Scholar] [CrossRef]
  35. Battiti, R. Using mutual information for selecting features in supervised neural net learning. IEEE Trans. Neural Netw. 1994, 5, 537–550. [Google Scholar] [CrossRef]
  36. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  37. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  38. Guler, I.; Ubeyli, E.D. Multiclass support vector machines for eeg-signals classification. IEEE Trans. Inf. Technol. Biomed. 2007, 11, 117–126. [Google Scholar] [CrossRef]
  39. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  40. Premavathi, T.; Shukla, M.; Jadeja, R. Predicting patient-specific epileptic seizures from scalp EEG signals using kNN model with transfer learning. IEEE Access 2024, 12, 196728–196739. [Google Scholar] [CrossRef]
  41. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  42. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  43. Ke, G.L.; Meng, Q.; Finley, T.; Wang, T.F.; Chen, W.; Ma, W.D.; Ye, Q.W.; Liu, T.Y. LightGBM: A highly efficient gradient boosting decision tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  44. Saranya, D.; Bharathi, A. Automatic detection of epileptic seizure using machine learning-based IANFIS-LightGBM system. J. Intell. Fuzzy Syst. Appl. Eng. Technol. 2024, 1, 46. [Google Scholar] [CrossRef]
  45. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32st International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 2–8 December 2018. [Google Scholar]
  46. Chen, T.Q.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  47. Fawcett, T. An introduction to roc analysis. Pattern Recognit. Lett. 2005, 27, 861–874. [Google Scholar] [CrossRef]
  48. Guo, L.; Rivero, D.; Pazos, A. Epileptic seizure detection using multiwavelet transform based approximate entropy and artificial neural networks. J. Neurosci. Methods 2010, 193, 156–163. [Google Scholar] [CrossRef] [PubMed]
  49. Ye, X.F.; Hu, P.P.; Yang, B.; Yang, Y.; Gao, D.; Zeng, G.Q.H.; Wang, K. Using scalp EEG to predict seizure recurrence and electrical status epilepticus in children with idiopathic focal epilepsy. Seizure Eur. J. Epilepsy 2024, 118, 8–16. [Google Scholar] [CrossRef]
  50. Zweiphenning, W.J.E.M.; Keijzer, H.M.; van Diessen, E.; van ’t Klooster, M.A.; van Klink, N.E.C.; Leijten, F.S.S.; van Rijen, P.C.; van Putten, M.J.A.M.; Braun, K.P.J.; Zijlmans, M. Increased gamma and decreased fast ripple connections of epileptic tissue: A high-frequency directed network approach. Epilepsia 2019, 60, 1908–1920. [Google Scholar] [CrossRef] [PubMed]
  51. Janca, R.; Jahodova, A.; Hlinka, J.; Jezdik, P.; Jiruska, P. Ictal gamma-band interactions localize ictogenic nodes of the epileptic network in focal cortical dysplasia. Clin. Neurophysiol. 2021, 132, 1927–1936. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The proposed scheme. The upper section shows raw EEG signals of FNSZ with the x-axis as time (s) and y-axis as amplitude ( μ V ), marking seizure start/end. The following is the resampling and filtering process. Then there is segmentation by dividing signals into 1 s segments with 50% overlap. After that is hybrid feature engineering by extracting statistical features (mean, variance, skewness, and kurtosis), sample entropy, wavelet energies, Hurst index, and Hjorth parameters from five frequency bands. Then there is MI analysis to select the top 30 features, leading to multi-class classification for ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, and TNSZ.
Figure 1. The proposed scheme. The upper section shows raw EEG signals of FNSZ with the x-axis as time (s) and y-axis as amplitude ( μ V ), marking seizure start/end. The following is the resampling and filtering process. Then there is segmentation by dividing signals into 1 s segments with 50% overlap. After that is hybrid feature engineering by extracting statistical features (mean, variance, skewness, and kurtosis), sample entropy, wavelet energies, Hurst index, and Hjorth parameters from five frequency bands. Then there is MI analysis to select the top 30 features, leading to multi-class classification for ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, and TNSZ.
Entropy 27 01057 g001
Figure 2. Confusion matrix for four classifiers, illustrating the classification performance across six seizure types (ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, TNSZ). The x-axis represents the predicted labels and the y-axis represents the true labels, with colors indicating the log-scaled count of instances and numbers denoting the exact count of each segment classification outcome, accompanied by percentages representing the proportion of each count relative to the total per true label category. (a) Confusion matrix of XGBoost. (b) Confusion matrix of CatBoost. (c) Confusion matrix of LightGBM. (d) Confusion matrix of RF.
Figure 2. Confusion matrix for four classifiers, illustrating the classification performance across six seizure types (ABSZ, CPSZ, FNSZ, GNSZ, TCSZ, TNSZ). The x-axis represents the predicted labels and the y-axis represents the true labels, with colors indicating the log-scaled count of instances and numbers denoting the exact count of each segment classification outcome, accompanied by percentages representing the proportion of each count relative to the total per true label category. (a) Confusion matrix of XGBoost. (b) Confusion matrix of CatBoost. (c) Confusion matrix of LightGBM. (d) Confusion matrix of RF.
Entropy 27 01057 g002
Figure 3. Feature importance scores heatmap of various EEG features across five frequency bands assessed by MI selection. The x-axis represents the feature types, including mean, variance, skewness, kurtosis, sample entropy, wavelet energy 0 through wavelet energy 4, Hurst index, Hjorth mobility, and Hjorth complexity. The y-axis denotes the five frequency bands ( δ , θ , α , β , and γ ). The color intensity, ranging from light to dark within the viridis color map, indicates the magnitude of MI scores, with darker shades representing higher scores (ranging from 0.003 to 0.253). Numbers within each cell, displayed in bold font, represent the specific MI scores, with higher values signifying greater feature importance in classifying seizure types, while zeros indicate no significant contribution.
Figure 3. Feature importance scores heatmap of various EEG features across five frequency bands assessed by MI selection. The x-axis represents the feature types, including mean, variance, skewness, kurtosis, sample entropy, wavelet energy 0 through wavelet energy 4, Hurst index, Hjorth mobility, and Hjorth complexity. The y-axis denotes the five frequency bands ( δ , θ , α , β , and γ ). The color intensity, ranging from light to dark within the viridis color map, indicates the magnitude of MI scores, with darker shades representing higher scores (ranging from 0.003 to 0.253). Numbers within each cell, displayed in bold font, represent the specific MI scores, with higher values signifying greater feature importance in classifying seizure types, while zeros indicate no significant contribution.
Entropy 27 01057 g003
Table 1. Summary of selected dataset information.
Table 1. Summary of selected dataset information.
Seizure TypePatientsSeizuresTotal Duration(s)Seizure Duration Range(s)
FNSZ151005726.8815.95–1552.74
ABSZ1078637.7014.66–202.68
CPSZ1910034,771.3045.74–1478.29
TCSZ647839.4540.62–748.94
GNSZ2010059,616.0021.35–1659.00
TNSZ8475502.9749.90–748.94
Table 2. Seizure type classification performance metrics for the seven classifiers and the null model (values in bold indicate the highest values).
Table 2. Seizure type classification performance metrics for the seven classifiers and the null model (values in bold indicate the highest values).
ClassifierAccuracyF1-ScorePrecisionRecallAUCTime (s)
SVM0.4709 ± 0.00290.4618 ± 0.00340.6249 ± 0.00370.4709 ± 0.00290.8148 ± 0.002632,984.36
KNN0.5256 ± 0.00220.5397 ± 0.00240.5870 ± 0.00380.5256 ± 0.00220.7665 ± 0.003165.80
Decision Tree0.7899 ± 0.00540.7897 ± 0.00530.7896 ± 0.00530.7899 ± 0.00540.8544 ± 0.0039102.09
LightGBM0.8595 ± 0.00300.8612 ± 0.00290.8651 ± 0.00270.8595 ± 0.00300.9772 ± 0.00086208.87
RF0.8684 ± 0.00220.8664 ± 0.00220.8686 ± 0.00230.8684 ± 0.00220.9782 ± 0.0007228.99
CatBoost0.8641 ± 0.00340.8658 ± 0.00330.8697 ± 0.00300.8641 ± 0.00340.9789 ± 0.0009532.23
XGBoost0.8710 ± 0.00270.8721 ± 0.00260.8744 ± 0.00250.8710 ± 0.00270.9797 ± 0.000769.39
Null model0.0149 ± 0.00120.0004 ± 0.00010.0002 ± 0.00000.0149 ± 0.00120.5000 ± 0.00001.79
Table 3. Sensitivity analysis of sample entropy parameters (m and r) on XGBoost performance metrics (values in bold indicate the highest values).
Table 3. Sensitivity analysis of sample entropy parameters (m and r) on XGBoost performance metrics (values in bold indicate the highest values).
mrAccuracyF1-ScorePrecisionRecallAUC
1 0.1 σ 0.8659 ± 0.00240.8672 ± 0.00240.8699 ± 0.00240.8659 ± 0.00240.9782 ± 0.0009
0.15 σ 0.8659 ± 0.00240.8672 ± 0.00240.8699 ± 0.00240.8659 ± 0.00240.9782 ± 0.0009
0.2 σ 0.8659 ± 0.00240.8672 ± 0.00240.8699 ± 0.00240.8659 ± 0.00240.9782 ± 0.0009
2 0.1 σ 0.8693 ± 0.00350.8705 ± 0.00340.8729 ± 0.00330.8693 ± 0.00350.9793 ± 0.0008
0.15 σ 0.8699 ± 0.00360.8710 ± 0.00350.8733 ± 0.00350.8699 ± 0.00360.9796 ± 0.0007
0.2 σ 0.8710 ± 0.00270.8721 ± 0.00260.8744 ± 0.00250.8710 ± 0.00270.9797 ± 0.0007
Table 4. Performance metrics of XGBoost with varying numbers of top features (k).
Table 4. Performance metrics of XGBoost with varying numbers of top features (k).
kAccuracyF1-ScorePrecisionRecallAUCTime (s)
100.7561 ± 0.00400.7602 ± 0.00370.7719 ± 0.00310.7561 ± 0.00400.9397 ± 0.001746.43
200.8463 ± 0.00300.8480 ± 0.00290.8521 ± 0.00290.8463 ± 0.00300.9729 ± 0.000953.16
300.8710 ± 0.00270.8721 ± 0.00260.8744 ± 0.00250.8710 ± 0.00270.9797 ± 0.000769.39
400.8868 ± 0.00270.8875 ± 0.00270.8889 ± 0.00270.8868 ± 0.00270.9836 ± 0.000777.17
500.8862 ± 0.00250.8869 ± 0.00250.8883 ± 0.00250.8862 ± 0.00250.9837 ± 0.000788.24
600.8899 ± 0.00270.8905 ± 0.00270.8917 ± 0.00260.8899 ± 0.00270.9844 ± 0.000797.95
Table 5. Performance metrics of XGBoost with varying step sizes (time window = 1 s, values in bold indicate the highest values).
Table 5. Performance metrics of XGBoost with varying step sizes (time window = 1 s, values in bold indicate the highest values).
Step Size (s)AccuracyF1-ScorePrecisionRecallAUCTime (s)
0.20.8645 ± 0.00170.8662 ± 0.00160.8701 ± 0.00150.8645 ± 0.00170.9795 ± 0.0004159.66
0.40.8583 ± 0.00200.8599 ± 0.00200.8631 ± 0.00210.8583 ± 0.00200.9766 ± 0.001080.60
0.50.8710 ± 0.00270.8721 ± 0.00260.8744 ± 0.00250.8710 ± 0.00270.9797 ± 0.000769.39
0.60.8532 ± 0.00570.8546 ± 0.00560.8572 ± 0.00540.8532 ± 0.00570.9741 ± 0.001557.22
0.80.8495 ± 0.00610.8507 ± 0.00600.8532 ± 0.00570.8495 ± 0.00610.9724 ± 0.001446.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Miao, Y. Seizure Type Classification Based on Hybrid Feature Engineering and Mutual Information Analysis Using Electroencephalogram. Entropy 2025, 27, 1057. https://doi.org/10.3390/e27101057

AMA Style

Miao Y. Seizure Type Classification Based on Hybrid Feature Engineering and Mutual Information Analysis Using Electroencephalogram. Entropy. 2025; 27(10):1057. https://doi.org/10.3390/e27101057

Chicago/Turabian Style

Miao, Yao. 2025. "Seizure Type Classification Based on Hybrid Feature Engineering and Mutual Information Analysis Using Electroencephalogram" Entropy 27, no. 10: 1057. https://doi.org/10.3390/e27101057

APA Style

Miao, Y. (2025). Seizure Type Classification Based on Hybrid Feature Engineering and Mutual Information Analysis Using Electroencephalogram. Entropy, 27(10), 1057. https://doi.org/10.3390/e27101057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop