Next Article in Journal
Optimizing Aircraft Routes in Dynamic Conditions Utilizing Multi-Criteria Parameters
Previous Article in Journal
Special Issue on Recent Processing Technologies for Improving Meat Quality
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Ensemble Selection for EEG Signal Classification in Distributed Data Environments

by
Małgorzata Przybyła-Kasperek
*,† and
Jakub Sacewicz
Institute of Computer Science, University of Silesia in Katowice, Bȩdzińska 39, 41-200 Sosnowiec, Poland
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(11), 6043; https://doi.org/10.3390/app15116043
Submission received: 28 March 2025 / Revised: 23 May 2025 / Accepted: 24 May 2025 / Published: 27 May 2025
(This article belongs to the Special Issue EEG Signal Processing in Medical Diagnosis Applications)

Abstract

This study presents a novel approach to EEG signal classification in distributed environments using dynamic ensemble selection. In scenarios where data dispersion arises due to privacy constraints or decentralized data collection, traditional global modelling is impractical. We propose a framework where classifiers are trained locally on independent subsets of EEG data without requiring centralized access. A dynamic coalition-based ensemble strategy is employed to integrate the outputs of these local models, enabling adaptive and instance-specific decision-making. Coalitions are formed based on conflict analysis between model predictions, allowing either consensus (unified) or diversity (diverse) to guide the ensemble structure. Experiments were conducted on two benchmark datasets: an epilepsy EEG dataset comprising 150 segmented EEG time series from ten patients, and the BCI Competition IV Dataset 1, with continuous recordings from seven subjects performing motor imagery tasks, for which a total of 1400 segments were extracted. In the study, we also evaluated the non-distributed (centralized) approach to provide a comprehensive performance baseline. Additionally, we tested a convolutional neural network specifically designed for EEG data, ensuring our results are compared against advanced deep learning methods. Gradient Boosting combined with measurement-level fusion and unified coalitions consistently achieved the highest performance, with an F1-score, accuracy, and balanced accuracy of 0.987 (for nine local tables). The results demonstrate the effectiveness and scalability of dynamic coalition-based ensembles for EEG diagnosis in distributed settings, highlighting their potential in privacy-sensitive clinical and telemedicine applications.

1. Introduction

Electroencephalography (EEG) is a non-invasive technique that records the electrical activity of the brain and is widely used in clinical diagnostics, especially for neurological disorders such as epilepsy. Accurate and timely classification of EEG signals is essential for effective diagnosis and intervention. However, in modern healthcare and research environments, EEG data are often collected across multiple centers, devices, or sessions, resulting in dispersed datasets. By this statement, we mean that EEG recordings may originate from different patients across multiple hospitals, but some patients may also be tested at various institutions. In many real-world scenarios, especially when data are collected across medical centers, patient identifiers are anonymized or protected due to privacy regulations. As a result, it is not possible to determine whether a given EEG signal comes from the same individual tested in different locations or from different individuals altogether. Our study operates under this constraint, treating the data as independent observations without attempting to resolve patient identity. All of these scenarios—multiple patients across centers, with some cases involving a single patient tested across sessions—are considered possible within our analysis. This fragmentation, coupled with privacy and regulatory concerns, limits the feasibility of centralized processing [1]. Consequently, there is a growing need for intelligent decision-making frameworks that can operate effectively under data distribution constraints while maintaining high diagnostic accuracy. Given the complexity of EEG signals, manual interpretation is a time-consuming task that demands substantial expertise [2]. Automated classification systems can offer consistent and fast support to neurologists, aiding in long-term monitoring, early diagnosis, and personalized therapy strategies.
According to the World Health Organization, epilepsy affects approximately 50 million people worldwide [3], and EEG remains the primary tool for its diagnosis. However, the increasing volume of EEG data collected from patients poses challenges for timely expert analysis. AI-based models, including deep learning and ensemble methods, have been explored to support clinical workflows by identifying abnormal patterns automatically [2,4]. In particular, our ensemble-based approach aims to improve robustness and adaptability across diverse EEG data distributions.
Previous studies have explored decision-making in scenarios involving dispersed or decentralized data, such as federated learning [5,6], local modeling [7,8], and ensemble-based techniques [9,10]. Federated learning approaches enable collaborative model training without direct data sharing, but they often require synchronization and substantial communication overhead. Alternatively, local modeling strategies independently train models on separate data partitions and aggregate their predictions using ensemble methods. While effective, traditional ensemble strategies like majority voting or static averaging do not account for the varying relevance or reliability of models across different input instances, especially in heterogeneous and non-iid data environments.
EEG signal classification has been extensively studied using a range of machine learning and deep learning techniques [11,12,13]. Common approaches include feature-based models utilizing statistical, spectral, and wavelet descriptors, as well as end-to-end deep learning architectures [14,15]. Boosting algorithms, such as AdaBoost and Gradient Boosting, have demonstrated strong performance due to their ability to focus on hard-to-classify instances [16,17]. However, many existing works assume access to a unified dataset and do not address the challenges associated with EEG data dispersion, such as feature inconsistency, sample variability, and localized context.
Despite growing interest in distributed machine learning, there remains a significant gap in effectively integrating classifier outputs from dispersed EEG datasets in a way that adapts to each decision context. Existing ensemble methods often treat all models equally, regardless of their relevance or agreement. Moreover, most EEG classification studies either disregard the implications of data decentralization or lack mechanisms to dynamically select models best suited to individual test cases. While various dynamic ensemble selection techniques have been proposed in the literature [18,19,20,21,22], their application to EEG signal classification remains largely unexplored.
To address these challenges, this study introduces a dynamic ensemble selection framework for EEG signal classification in dispersed data environments. Local models are trained independently on non-overlapping subsets of EEG data, simulating a decentralized scenario. At inference time, we dynamically form coalitions of models based on their predictive behavior for each test instance. These coalitions, formed through conflict analysis, represent either high consensus or high diversity. By using both measurement-level and abstract-level fusion strategies, we adaptively aggregate the most informative predictions. The novelty lies in the context-aware coalition formation and its application to EEG diagnosis under data fragmentation constraints.
The main contributions of this work are as follows:
  • A comprehensive ensemble framework that supports EEG classification with dispersed data without centralized access.
  • A novel coalition-based dynamic model selection mechanism using conflict metrics.
  • An in-depth evaluation across multiple dispersion scenarios with various base classifiers.
  • Empirical evidence showing that unified coalitions with measurement-level fusion achieve good performance, particularly with Gradient Boosting models.
EEG plays a pivotal role in clinical neurology by providing real-time insight into the brain’s electrical activity. In epilepsy diagnosis, EEG allows for the identification of seizure patterns across different stages of the condition, which is crucial for timely intervention and treatment. However, interpreting EEG signals remains a challenging task due to their complexity and variability across patients. In this context, improving the accuracy and robustness of EEG classification models directly benefits medical professionals by offering decision support tools that reduce diagnostic uncertainty. Moreover, such improvements are also vital for the development of brain–computer interfaces (BCIs), which depend on reliable EEG interpretation to enable communication and control in individuals with severe motor impairments. By proposing a framework suitable for distributed data environments, this study addresses real-world constraints, such as data privacy and institutional fragmentation, and aims to support scalable, accurate, and privacy-preserving EEG analysis for both diagnostic and assistive technologies.
The approach proposed in this paper introduces a dynamic ensemble classification framework that adapts to the varying quality and distribution of EEG signals without requiring centralized access to the full dataset. In contrast to existing methods that either assume access to homogeneous data or apply static ensemble techniques, our strategy dynamically selects and integrates classifier outputs based on the consistency or disagreement between models. The originality of the method lies in its coalition-based selection mechanism, which considers the local predictive behavior of models and builds consensus or diversity groups tailored to each test instance. This mechanism is applied to distributed EEG data representing seizure stages across different patients—addressing an important gap in the literature, where classifier integration is often simplistic and lacks adaptability to decentralized clinical data. The results suggest that our unified, context-aware ensemble strategy not only improves performance metrics but also supports realistic deployment scenarios, such as telemedicine and hospital networks, where data cannot be freely shared.
The remainder of this paper is structured as follows. Section 2 describes the materials and methods used, including the classifiers, and coalition formation strategies. Section 3 presents the experimental setup and results. Also, the dataset, feature extraction, and data preparation are discussed in this section. Section 4 discusses the implications of the findings and limitations. Section 5 concludes the paper with key findings and possible future directions.

2. Materials and Methods

In the context of EEG analysis, especially when handling data from multiple sources or collection sites, it is common to encounter dispersed data—datasets that differ in structure, either due to variations in feature representation or the distribution of examples. To address this, and to support privacy-preserving processing, we adopt a local modeling strategy, where models are trained independently on isolated subsets of the data, referred to as local tables. These subsets are processed without requiring global data access, allowing for modular and decentralized analysis.
Different machine learning classifiers could be used for the local table. In this study, to analyze and model localized segments of EEG signals, we systematically compare five widely recognized algorithms: Random Forest, AdaBoost, Gradient Boosting, k-Nearest Neighbours, and Logistic Regression. Each classifier is trained separately on a distinct local dataset, with each dataset comprising a unique subset of EEG recordings. This approach enables us to mimic an object-distributed environment, where data are stored initially across multiple sources or locations, and each model learns from its respective local data without access to the full dataset.
Each trained model outputs a class probability vector for a test instance x ^ , structured as [ μ i , 1 ( x ^ ) , , μ i , c ( x ^ ) ] , where μ i , j ( x ^ ) is the support for class j provided by model i, and c is the number of decision classes. These prediction vectors serve as the basis for our ensemble fusion framework.
To go beyond conventional majority voting, we implement a dynamic ensemble selection framework. This approach dynamically identifies, for each test instance, a subset of models—called a coalition—that will participate in the final decision. The coalition is constructed individually for every tested object, making the ensemble configuration adaptive and context-aware.
We formally represent each local dataset as a decision table: D i = ( U i , A i , d ) denotes the set of EEG signal instances, A i the features (e.g., statistical, spectral, or wavelet-based descriptors), and d the decision label (e.g., ictal vs. inter-ictal state). Due to the distributed nature of the data, features and samples may vary between tables. To form coalitions, we use a conflict matrix approach. Each model’s prediction vector is evaluated to determine its stance on each class. For a class v, a model’s support is encoded by a function v ( i ) { 1 , 0 , 1 } defined as follows:
v ( i ) = 1 if the coordinate μ i , v ( x ^ ) for decision v in the prediction vector of model i has the maximum value of all coordinates in this vector . 0 if the coordinate μ i , v ( x ^ ) for decision v in the prediction vector of model i is this vector s second highest of all coordinates . 1 in other cases
Using these encodings, a conflict function ρ ( i , j ) measures the disagreement between models i and j as the proportion of decision classes where their opinions differ. Formally,
ρ ( i , j ) = c a r d { v V : v ( i ) v ( j ) } c a r d { V } ,
where c a r d { V } is the cardinality of the set of decision classes and i and j are local models.
Two types of coalitions are constructed:
  • Unified coalitions, where ρ ( i , j ) < 0.5 for all pairs in the group, capturing groups of models that largely agree.
  • Diverse coalitions, where ρ ( i , j ) > 0.5 , representing models with significantly differing perspectives.
These coalitions are generated dynamically per test instance, allowing the ensemble structure to adapt to each prediction task. Importantly, models may belong to multiple coalitions simultaneously, reflecting overlapping perspectives, as commonly seen in real-world ensemble behavior.
For decision-making, we focus on the strongest coalition—the one with the highest number of members. Two types of decision fusion are used:
  • Measurement-level fusion: All prediction vectors from coalition members are summed. The final class is the one with the highest aggregated score.
  • Abstract-level fusion: A simple majority vote is conducted among coalition members, based on their most confident class prediction.
In addition to coalition-based methods, we also evaluate traditional ensemble fusion strategies—simple sum of prediction vectors for measurement-level or simple voting for abstract-level from all local models.
By integrating these techniques, we explore both consensus-driven and diversity ensemble strategies for EEG signal classification under distributed constraints. This approach aligns well with clinical scenarios where data are fragmented across institutions or collection systems, and direct data sharing is restricted due to privacy, regulation, or logistical concerns.

3. Datasets and Experimental Results

The aim of this study is to investigate whether EEG data, when distributed across multiple sources, can be effectively utilized for classification using a dynamic ensemble selection approach. The experiments were conducted on two datasets.
The EEG Epilepsy dataset [23] contains exemplary segmented EEG time series recordings from ten epilepsy patients, collected at the Neurology & Sleep Centre in Hauz Khas, New Delhi. Data acquisition was performed using the Grass Telefactor Comet AS40 Amplification System, (manufactured by Grass Technologies, West Warwick, RI, USA). with a sampling frequency of 200 Hz. During the recordings, gold-plated scalp EEG electrodes were placed according to the standard 10–20 system. Although the publicly available dataset contains data for only one channel, the EEG signals were band-pass filtered between 0.5 and 70 Hz and then segmented into three seizure-related stages: pre-ictal, inter-ictal, and ictal. Each single-channel EEG segment was extracted from continuous recordings of multiple patients, and each segment was selected from different channels.
Each downloadable folder corresponds to one of these seizure stages and contains fifty MATLAB version 2022a .mat files. Each file includes a 1024-sample EEG time series segment, representing a 5.12 s duration. From the raw signals, 33 features were extracted per segment, in accordance with the feature extraction protocol described in [14]. These features include the following:
  • Statistical features: mean, variance, skewness, kurtosis, root mean square (RMS), and zero crossings.
  • Spectral power features: delta, theta, alpha, beta, gamma, and total power.
  • Wavelet-based features: for decomposition levels 0 to 4, the mean, variance, skewness, and kurtosis of the wavelet coefficients were computed.
As was mentioned, each observation is a time series of 5.12 s at a frequency of 200 Hz, which gives 1024 points in time, and based on them, the time parameters for the object are calculated (one mat file). As a result, we obtain 150 objects in total. Although the data were originally provided as a single dataset, our objective was to evaluate the performance of a classification system based on distributed data sources, simulating scenarios where EEG data are decentralized. To this end, the original dataset was first split into two parts: a training set containing 120 observations and a test set of 30 observations. For the Epilepsy dataset, the classification was made for three classes—pre-ictal, ictal and post-ictal.
The BCI Competition IV Dataset 1 [24] contains EEG recordings from seven subjects performing motor imagery tasks (e.g., left hand, right hand, or foot) with both calibration and evaluation data. Calibration data include visual cues and labels, while evaluation data simulate real-world conditions with auditory cues and no labels. EEG signals were recorded from 59 channels at 1000 Hz. The dataset incorporates both real and artificially generated EEG data to assess the robustness of classification algorithms. The specific nature (real or artificial) of each dataset is undisclosed to participants. This dataset serves as a benchmark for developing and testing algorithms aimed at real-time, asynchronous detection of motor imagery in EEG signals, a critical component for practical BCI applications. The dataset with extracted features for BCI dataset contains 1400 objects with extracted features for each channel. The data were split into training (80%), evaluating (10%), and testing (10%) datasets. The data were split in a stratified way, preserving the relative proportions of each class.
Next, the training data were dispersed into multiple independent sources, referred to as local tables, to simulate distributed learning environments. To examine the scalability and robustness of the proposed dynamic ensemble selection model, we experimented with varying numbers of these local data sources. Specifically, we tested configurations with 3, 5, 7, 9, and 11 local tables. In each case, the training objects were randomly partitioned among the local tables using a stratified sampling strategy to preserve class distributions across all subsets.
To ensure the stability and reliability of the results, each experiment was repeated 10 times. Given the inherent randomness in some of the evaluated methods, a fixed set of 10 randomly selected seeds was used to enhance reproducibility. Each repetition was executed with a specific seed, consistently applied throughout the entire pipeline.
The evaluation of classification performance was conducted on the test set using a variety of metrics: Classification Accuracy (Acc), Recall, Precision (Prec.), Balanced Accuracy (BAcc), and the F1-score. To provide a more comprehensive assessment, two variants of the F1-score were calculated: macro-averaged F1, which treats all classes equally, and weighted F1, which accounts for class imbalance by weighting classes according to their support.
The results are summarized in Table 1 and Table 2 (for the EEG Epilepsy dataset) and Table 3 and Table 4 (for the BCI Competition IV Dataset 1), presenting averaged metric values from all 10 achieved metrics. The dataset, characterized by varying degrees of dispersion (indicated by annotations such as ‘3LT’, meaning the dataset has been divided into three separate local tables), is evaluated using three approaches: without coalitions, with coalitions representing agents’ agreement, and with coalitions representing agents’ disagreement. The following notations are used:
  • Abstract level, sum: Each local classifier (Random Forest, AdaBoost, Gradient Boosting, k-Nearest Neighbours, and Logistic Regression) makes an independent prediction, one decision class. The final decision is the result of majority voting.
  • Probability level, sum: Each local classifier makes an independent prediction. The resulting probability vectors are summed element-wise, and the class with the highest aggregated score is selected as the final decision.
  • Unified groups, abstract level: Models are grouped using compliance coalition formation. One strongest coalition makes the final decision by voting.
  • Diverse groups, abstract level: Models are grouped using diversified coalition formation. One strongest coalition makes the final decision by voting.
  • Unified groups, measurement level: Models are grouped using compliance coalition formation. One strongest coalition prediction vectors are summed. The final class is the one with the highest aggregated score.
  • Diverse groups, measurement level: Models are grouped using diversified coalition formation. One strongest coalition prediction vectors are summed. The final class is the one with the highest aggregated score.
All methods and experiments were implemented in Python 3.12, utilizing the scikit-learn library for the implementation of classifiers such as Random Forest, AdaBoost, Gradient Boosting, and Logistic Regression, as well as NumPy and Pandas for efficient data processing and analysis. To assess the influence of hyperparameter settings on model performance, a comprehensive evaluation of selected configurations was conducted for each classification algorithm. The following ranges of hyperparameters were examined:
  • Random Forest: The number of estimators (n_estimators) was varied among { 10 , 20 , 50 , 100 , 200 } . Two splitting criteria were evaluated: gini and entropy. The maximum depth of trees was fixed at max_depth = 6, while the minimum number of samples required to split an internal node (min_samples_split) was tested with values { 2 , 3 , 4 } .
  • AdaBoost: The ensemble size was controlled by varying n_estimators in { 10 , 20 , 50 , 100 , 200 } . Two boosting algorithms were considered: SAMME and SAMME.R. Additionally, the learning_rate parameter was evaluated with values { 0.1 , 0.5 , 1.0 , 1.5 } to control the contribution of each weak learner.
  • Gradient Boosting: The number of boosting stages (n_estimators) was set to values among { 10 , 20 , 50 , 100 , 200 } , with the learning_rate adjusted across { 0.1 , 0.5 , 1.0 } . The maximum depth of each individual regression estimator (max_depth) was varied within { 3 , 4 , 5 , 6 } .
  • Logistic Regression: The following solvers were evaluated: lbfgs, liblinear, sag, saga, and newton-cg. The class_weight parameter was set either to balanced or none to assess the impact of handling class imbalance. Regularization was controlled via the penalty parameter, which was tested with values l1, l2, and None.
The tables report the best performance achieved for each configuration, along with the corresponding number of estimators that obtained this result. For each dataset, the highest-performing outcome is indicated, with the best result highlighted to facilitate comparison across different configurations.
The experimental results for the EEG Epilepsy dataset presented in Table 1 and Table 2 provide clear evidence supporting the effectiveness of the proposed approach based on dynamic ensemble selection over distributed EEG data sources. Among the evaluated classifiers, Gradient Boosting demonstrated the most robust and consistently high performance across varying levels of data dispersion. Notably, the unified groups strategy, particularly when applied at the measurement level, achieved the highest classification scores. In the configuration with nine local tables (9LT), Gradient Boosting combined with measurement-level unified coalition selection achieved an F1-score of 0.987, balanced accuracy (BAacc) of 0.987, and overall classification accuracy of 0.987—the best results across all tested configurations.
AdaBoost also yielded high performance, particularly with three to seven local tables, achieving F1-scores above 0.93 in multiple scenarios. In contrast, Random Forest achieved moderately strong results but consistently lagged behind the boosting methods. Logistic Regression and k-Nearest Neighbors (kNN) exhibited significantly lower performance, especially as the number of local tables increased, indicating sensitivity to data sparsity and limitations in capturing complex patterns in a decentralized setting.
Figure 1 illustrates the comparative performance for the EEG Epilepsy dataset of various classifiers across different ensemble strategies and fusion levels. The x-axis represents the ensemble strategy combined with the fusion level (e.g., Sum–Abstract, Unified–Measurement, Diverse–Measurement), while the y-axis lists the classifiers evaluated (Gradient Boosting, AdaBoost, Random Forest, Logistic Regression, kNN). The color-coded values indicate the average F1-weighted score, computed across all levels of data dispersion (3 to 11 local tables). As shown, Gradient Boosting combined with the unified strategy consistently achieves the highest performance, clearly outperforming the other approaches.
A comparison of dynamic ensemble selection strategies with conflict analysis revealed that unified coalition approaches consistently outperformed both simple probability-based voting and diverse coalition models. While the probability sum method remained competitive in several configurations, it was outclassed by coalition-based strategies. In contrast, the diverse groups strategy, which aggregates classifiers with conflicting predictions, generally resulted in reduced classification performance across all metrics, especially when paired with weak learners or simple models.
Furthermore, measurement-level fusion of classifier outputs proved superior to abstract-level fusion. Access to the full probability distribution of predictions enabled more precise and informative coalition formation and decision-making, particularly benefiting complex ensemble techniques.
Importantly, the study demonstrates that the proposed framework is scalable and maintains high classification quality even with an increased number of data sources (up to 11 local tables), particularly when strong base learners are employed and appropriate ensemble strategies are selected.
Figure 2 presents a line chart illustrating the variation in F1-weighted scores across different levels of data dispersion and the EEG Epilepsy dataset (i.e., the number of local tables: 3, 5, 7, 9, 11) for each classifier, based on their best-performing ensemble strategy. The results reveal that Gradient Boosting demonstrates strong robustness to data distribution, achieving its highest performance at nine local tables (LT). AdaBoost maintains consistently high performance up to 7LT, followed by a slight decline. In contrast, Random Forest and Logistic Regression show a gradual decrease in classification quality as dispersion increases. k-Nearest Neighbors (kNN) exhibits the weakest but stable performance across all dispersion levels.
Comparing the results presented in Table 3 and Table 4 for the BCI Competition IV Dataset 1, one can notice small but systematic differences in the effectiveness of various approaches. Generally, the best results in both tables were obtained by AdaBoost and Gradient Boosting models, especially in combination with the measurement level and homogeneous groups. The F1 (weighted) values for these approaches ranged from approximately 0.57 to 0.59, without exceeding 0.6, which indicates a moderate difficulty of the dataset.
Although the differences between the tables are subtle, Table 4—corresponding to experiments with greater data dispersion (9LT and 11LT)—shows slightly higher F1 values for some configurations; e.g., for diverse groups at the measurement level with AdaBoost, an F1 of 0.589 was achieved with 9LT, while analogous approaches in Table 3 reached a maximum of 0.584 with 5LT. On the other hand, in Table 3, methods with fewer local tables (3LT–5LT) often performed better, suggesting that the increase in the number of data sources affects the results—both positively and negatively, depending on the configuration. Overall, however, none of the methods clearly dominated the others, and the classification performance remained at a similar level in all tested variants.
Figure 3 illustrates the comparative F1-measure for the BCI Competition IV Dataset 1 of various classifiers across different ensemble strategies and fusion levels. AdaBoost consistently achieves the highest F1-scores across all strategies, with values ranging from 0.528 to 0.579. The diverse-measurement fusion level brings the best overall score for AdaBoost (0.579), suggesting that incorporating diverse feature sources at the measurement level benefits this classifier the most. Gradient Boosting also performs strongly, with scores clustering around 0.543–0.557. The best result (0.557) is also seen under the diverse-measurement condition. Logistic Regression and Random Forest show more stable but slightly lower performance. Their F1-scores fluctuate minimally across strategies, indicating that their performance is less sensitive to the fusion strategy, with Logistic Regression the highest at 0.536 and Random Forest at 0.530. k-Nearest Neighbors yields the lowest performance, particularly under sum-abstract and unified-abstract, where the F1-score drops to 0.376. However, performance improves substantially under diverse-abstract and diverse-measurement strategies (0.508 and 0.501, respectively), highlighting kNN’s benefit from diversity in feature fusion.
Figure 4 presents a line chart illustrating the variation in F1-weighted scores across different levels of data dispersion and the BCI Competition IV Dataset 1 (i.e., the number of local tables: 3, 5, 7, 9, 11) for each classifier, based on their best-performing ensemble strategy. Among the classifiers, AdaBoost demonstrates the most consistently high performance across all dispersion levels. This stability indicates that AdaBoost is robust to distributional changes and can maintain reliable predictive quality even in more fragmented learning environments. Gradient Boosting starts with strong performance at three local tables, close to that of AdaBoost, but exhibits a gradual decline as the data become more dispersed. Random Forest shows relatively flat performance across all levels of dispersion. While not severely impacted by the increase in local tables, its performance does not reach the levels of the boosting methods, highlighting its limitations in capturing complex patterns in this BCI dataset under decentralized conditions. Logistic Regression exhibits more variability. Finally, k-Nearest Neighbors consistently achieves the lowest F1-weighted scores, although a slight upward trend is visible from three to seven local tables, after which performance plateaus. Overall, the results indicate that AdaBoost offers the most robust classification capability under distributed conditions, followed by Gradient Boosting, while Random Forest and kNN are less effective for this task.
Statistical tests were performed to confirm the observations and F1-weighted values were used for comparison. The F1-weighted values of all six approaches were compared: 1—Abstract level, sum; 2—Measurement level, sum; 3—Unified groups, abstract level; 4—Diverse groups, abstract level; 5—Unified groups, measurement level; 6—Diverse groups, measurement level. So, six dependent samples, each containing 50 observations, were created, representing the results presented in Table 1, Table 2, Table 3 and Table 4. The Friedman test was used to determine whether the differences in F1-weighted values among the approaches were statistically significant. The Friedman test indicated that there is a statistically significant difference in mean F1-weighted among the twenty approaches, χ 2 ( 50 , 5 ) = 14.24 , p = 0.014 . To pinpoint the specific differences, a post hoc Dunn–Bonferroni test was performed, with the significant results highlighted in blue in Table 5. The test revealed significant differences between the diverse groups, abstract level approach, and the diverse groups, measurement level approach. A comparative box plot illustrating the F1-weighted results for the six approaches is provided in Figure 5. Based on the graph, it can be seen that worse results were achieved for the diverse groups with abstract level approach.
As can be seen, the observed performance differences across the two datasets highlight the importance of dataset-specific characteristics in model evaluation. The Epilepsy dataset was carefully curated and preprocessed by its authors to facilitate model training—signals were clearly segmented, of equal length, and correctly labeled for each mental state. In contrast, the BCI dataset was explicitly designed to be challenging: it comprises continuous, unsegmented signals with unknown transitions between mental states, variable trial durations (1.5–8 s), class imbalance, and artificially inserted “no control” periods. These differences substantially impact model performance and suggest that care must be taken when generalizing results across datasets with differing structures and assumptions.
The results presented above are compared to those obtained without applying distributed environments. Admittedly, this comparison is not entirely fair, as the non-dispersed approach allows the model access to all available knowledge simultaneously. Nevertheless, these experiments were conducted to determine whether dispersing knowledge significantly reduces classification quality. The results obtained for both datasets and the approaches used in the paper (Random Forest, AdaBoost, Gradient Boosting, Logistic Regression) are shown in Table 6. The procedure for optimizing the parameters mirrored the strategies employed in the approaches discussed earlier.
Comparing the results from the non-dispersed version (Table 6) with the dispersed approach shows that AdaBoost is the best-performing method overall. For the EEG Epilepsy dataset, the dispersed approach actually performs slightly better, with results like 0.983 accuracy and 0.98 for other metrics. This means that spreading the data out did not hurt the results—in fact, it may have helped a bit. For the BCI Competition IV Dataset 1, the distributed version scores just a little lower, with accuracy values between 0.596 and 0.601. However, this difference is very small and not really significant. Overall, using dispersed data does not reduce the quality of the results. In the case of the EEG Epilepsy data, it even improves it. For the BCI dataset, the drop in performance is so minor that it shows dispersion does not cause any real deterioration. On the contrary, distributed approaches can work just as well, or even better, depending on the dataset.
We also compared our results with state-of-the-art deep learning models commonly used in EEG classification. For this purpose, we selected the approach described in [25]. EEGNet is a lightweight convolutional neural network specifically designed for EEG signal analysis. Its architecture uses depthwise and separable convolutions, which allows it to efficiently capture both spatial and temporal features from EEG data while keeping the model compact and computationally efficient. EEGNet is known for its ability to generalize well across various EEG paradigms, such as P300, ERN, MRCP, and SMR, without the need for extensive task-specific adjustments. It also performs strongly when training data are limited and offers interpretable results, as the features it learns often correspond to known EEG patterns. Specifically, we used EEGNet-SSVEP, as it achieved the best performance on the analyzed datasets. The obtained results are presented in Table 7.
As shown, the results obtained using a dedicated neural network are inferior to those achieved in the distributed environment.

4. Discussion

This study investigated the use of dynamic ensemble selection for EEG signal classification in scenarios where data are distributed due to privacy or logistical constraints. The experimental results across both the EEG Epilepsy dataset and the BCI Competition IV Dataset 1 indicate that the proposed approach—particularly when using Gradient Boosting and unified measurement-level coalitions—achieved consistently strong performance metrics. These outcomes suggest that the coalition-based strategy may offer advantages in adapting to the heterogeneity of EEG signals across distributed sources.
While the unified coalition strategy generally outperformed both diverse coalition and traditional ensemble methods, this should be interpreted within the context of the datasets and settings examined. The dynamic nature of the model selection process appears to support more adjusted decisions, especially when high agreement among local models can be obtained. Moreover, the benefit of measurement-level fusion over abstract-level fusion highlights the value of using detailed probability distributions rather than relying solely on class labels.
This study has several limitations that should be considered. First, the distributed environment was simulated by partitioning centralized EEG datasets, rather than using data collected from genuinely independent sources. This means that important real-world challenges were not fully addressed. Second, the experimental design assumed synchronous model training and evaluation, as well as balanced class distributions and uniform feature representation across all local sites. In real-world scenarios, data are often highly imbalanced or non-iid, and devices may operate asynchronously, which can introduce additional complexity and impact model robustness. Third, the study relied on a relatively small number of subjects and focused on two publicly available EEG datasets related to epilepsy and motor imagery. While stratified sampling was used to maintain class balance, the limited sample size and scope may restrict the generalizability of the findings to other patient populations or clinical applications. Finally, the feature extraction process was based on conventional statistical, spectral, and wavelet-based features. Incorporating more advanced feature representations, such as deep-learned embeddings, could potentially improve classification performance and should be explored in future research.
Despite these limitations, the proposed framework appears promising for applications where data decentralization is essential due to privacy, security, or logistical constraints—such as mobile health (mHealth) platforms, tele-neurology, and collaborative clinical research. Its modular design allows for integration with various base classifiers and coalition strategies and has shown the ability to scale with increasing numbers of distributed data sources in our experiments.
To contextualize our findings, we compared our best results with those reported in recent EEG-based epilepsy classification studies. For instance, Huang et al. [14] reported an F1-score of 0.9516 using a dual-attention deep learning framework on a similarly structured epilepsy dataset, while Kołodziej et al. [12] achieved an accuracy of approximately 0.95 using a CNN-based pipeline on intracranial EEG data. In our study, the unified coalition-based ensemble approach with Gradient Boosting achieved an F1-score and balanced accuracy of 0.987. While direct comparisons are limited by differences in datasets and experimental conditions, these results suggest that our method may offer competitive performance, even in a distributed data setting.
The reliability of EEG features across sessions is a well-documented challenge, particularly in clinical and cognitive applications that rely on single-session recordings for machine learning-based decision-making. Numerous studies have shown that EEG signals can exhibit substantial intra-individual variability across days, devices, or mental states, which may affect the replicability of model outputs [26,27,28].
In our study, this challenge also appears due to data privacy constraints. Because the data used in our decentralised learning framework originate from independent local sources and are subject to anonymisation, we cannot identify whether two recordings originate from the same patient across different sessions or institutions. As such, we cannot explicitly evaluate how likely it is that an AI model would assign the same diagnosis to two EEGs recorded on separate days from the same individual. This limitation reflects a broader issue in privacy-preserving medical AI, where the need to protect patient identity may restrict the evaluation of temporal reliability.
Nevertheless, the proposed dynamic ensemble strategy may help address some of these challenges. By training multiple local models on independently sourced EEG subsets and forming coalitions based on inter-model agreement or conflict, the approach aims to reduce dependence on any single, potentially noisy signal. This instance-specific selection mechanism may contribute to improved resilience against feature noise or anomalies in individual recordings.
In summary, while further refinement and validation in real-world settings are needed, the results suggest that this method could provide a useful foundation for developing distributed and adaptable EEG classification systems.

5. Conclusions

This paper introduced a dynamic ensemble selection framework for EEG signal classification in distributed data environments. The proposed method uses coalition formation based on conflict metrics to dynamically integrate predictions from locally trained models. Our findings suggest that the framework has potential for effective classification in scenarios where data centralization is infeasible.
In particular, unified coalitions at the measurement level consistently achieved strong results, especially when paired with Gradient Boosting classifiers. These results underscore the promise of context-aware ensemble strategies in handling data dispersion and heterogeneity—common challenges in clinical and telemedicine applications. The findings emphasize the following:
  • Coalition-based ensemble models outperform traditional voting approaches.
  • Measurement-level prediction aggregation preserves more discriminative information than abstract label-level fusion.
  • Boosting methods (Gradient Boosting, AdaBoost) are suitable for decentralized EEG classification tasks.
While the results are encouraging, we acknowledge that further validation on larger and more diverse datasets is necessary to fully establish the generalizability of the approach. Future work will focus on expanding the framework to more complex clinical scenarios and integrating it with real-time EEG acquisition systems. Future work should also investigate the robustness of the proposed framework in the presence of noise, missing data, and high variability across different patients. Transfer learning or domain adaptation techniques could be incorporated to improve generalization in dispersed system. Overall, this study contributes to ongoing efforts in developing adaptive, privacy-aware EEG analysis methods, with potential applications in both telemedicine and decentralized healthcare infrastructures.

Author Contributions

Conceptualization, M.P.-K. and J.S.; methodology, M.P.-K. and J.S.; software, J.S.; validation, M.P.-K. and J.S.; formal analysis, M.P.-K. and J.S.; investigation, M.P.-K. and J.S.; writing—original draft preparation, M.P.-K.; writing—review and editing, M.P.-K. and J.S.; visualization, M.P.-K. and J.S.; supervision, M.P.-K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in the paper are freely available [23].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rieke, N.; Hancox, J.; Li, W.; Milletari, F.; Roth, H.R.; Albarqouni, S.; Bakas, S.; Galtier, M.N.; Landman, B.A.; Maier-Hein, K.; et al. The future of digital health with federated learning. NPJ Digit. Med. 2020, 3, 119. [Google Scholar] [CrossRef] [PubMed]
  2. Roy, Y.; Banville, H.; Albuquerque, I.; Gramfort, A.; Falk, T.H.; Faubert, J. Deep learning-based electroencephalography analysis: A systematic review. J. Neural Eng. 2019, 16, 051001. [Google Scholar] [CrossRef] [PubMed]
  3. World Health Organization. Available online: https://www.who.int/news-room/fact-sheets/detail/epilepsy (accessed on 17 May 2025).
  4. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adeli, H.; Subha, D.P. Automated EEG-based screening of depression using deep convolutional neural network. Comput. Methods Programs Biomed. 2018, 161, 103–113. [Google Scholar] [CrossRef]
  5. Pękala, B.; Szkoła, J.; Grochowalski, P.; Gil, D.; Kosior, D.; Dyczkowski, K. A Novel Method for Human Fall Detection Using Federated Learning and Interval-Valued Fuzzy Inference Systems. J. Artif. Intell. Soft Comput. Res. 2025, 15, 77–90. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Li, P.; Al Hammadi, A.Y.; Guo, F.; Damiani, E.; Yeun, C.Y. Reputation-based federated learning defense to mitigate threats in EEG signal classification. In Proceedings of the 2024 16th International Conference on Computer and Automation Engineering (ICCAE), Melbourne, Australia, 14–16 March 2024; IEEE: Piscataway Township, NJ, USA, 2024; pp. 173–180. [Google Scholar]
  7. Burduk, R.; Biedrzycki, J. Subspace-based decision trees integration. Inf. Sci. 2022, 592, 215–226. [Google Scholar] [CrossRef]
  8. Cano, A.; Krawczyk, B. ROSE: Robust online self-adjusting ensemble for continual learning on imbalanced drifting data streams. Mach. Learn. 2022, 111, 2561–2599. [Google Scholar] [CrossRef]
  9. Ruszczak, B.; Rudnik, K. Ensemble-based versus expert-assisted approach to carbon price features selection. In Proceedings of the FedCSIS (Communication Papers), Warsaw, Poland, 17–20 September 2023; pp. 245–250. [Google Scholar]
  10. Singer, G.; Ratnovsky, A.; Naftali, S. Classification of severity of trachea stenosis from EEG signals using ordinal decision-tree based algorithms and ensemble-based ordinal and non-ordinal algorithms. Expert Syst. Appl. 2021, 173, 114707. [Google Scholar] [CrossRef]
  11. Dura, A.; Wosiak, A.; Stasiak, B.; Wojciechowski, A.; Rogowski, J. Reversed correlation-based pairwised EEG channel selection in emotional state recognition. In International Conference on Computational Science; Springer International Publishing: Cham, Switzerland, 2021; pp. 528–541. [Google Scholar]
  12. Kołodziej, M.; Majkowski, A.; Rysz, A. Implementation of machine learning and deep learning techniques for the detection of epileptic seizures using intracranial electroencephalography. Appl. Sci. 2023, 13, 8747. [Google Scholar] [CrossRef]
  13. Krzywicka, M.; Wosiak, A. Efficacy of feature selection and Classification algorithms in cancer remission using medical imaging. Procedia Comput. Sci. 2024, 246, 4572–4581. [Google Scholar] [CrossRef]
  14. Huang, Z.; Yang, Y.; Ma, Y.; Dong, Q.; Su, J.; Shi, H.; Zhang, S.; Hu, L. EEG detection and recognition model for epilepsy based on dual attention mechanism. Sci. Rep. 2025, 15, 9404. [Google Scholar] [CrossRef]
  15. Medhi, K.; Hoque, N.; Dutta, S.K.; Hussain, M.I. An efficient EEG signal classification technique for Brain–Computer Interface using hybrid Deep Learning. Biomed. Signal Process. Control 2022, 78, 104005. [Google Scholar] [CrossRef]
  16. Gosala, B.; Kapgate, P.D.; Jain, P.; Chaurasia, R.N.; Gupta, M. Wavelet transforms for feature engineering in EEG data processing: An application on Schizophrenia. Biomed. Signal Process. Control 2023, 85, 104811. [Google Scholar] [CrossRef]
  17. Lalawat, R.S.; Bajaj, V. Optimal variational mode decomposition based automatic stress classification system using EEG signals. Appl. Acoust. 2025, 231, 110478. [Google Scholar] [CrossRef]
  18. Marfo, K.F.; Przybyła-Kasperek, M. Exploring the Impact of Object Diversity on Classification Quality in Dispersed Data Environments. In Intelligent Information and Database Systems; Nguyen, N.T., Chbeir, R., Manolopoulos, Y., Fujita, H., Hong, T.-P., Nguyen, L.M., Wojtkiewicz, K., Eds.; ACIIDS 2024 Lecture Notes in Computer Science; Springer: Singapore, 2024; Volume 14796. [Google Scholar]
  19. Przybyła-Kasperek, M.; Sacewicz, J. Ensembles of random trees with coalitions-a classification model for dispersed data. Procedia Comput. Sci. 2024, 246, 1599–1608. [Google Scholar] [CrossRef]
  20. Przybyła-Kasperek, M.; Kusztal, K.; Addo, B.A. Dispersed Data Classification Model with Conflict Analysis and Parameterized Allied Relations. Procedia Comput. Sci. 2024, 246, 2215–2224. [Google Scholar] [CrossRef]
  21. Sarnovsky, M.; Kolarik, M. Classification of the drifting data streams using heterogeneous diversified dynamic class-weighted ensemble. PeerJ Comput. Sci. 2021, 7, e459. [Google Scholar] [CrossRef]
  22. Zyblewski, P.; Sabourin, R.; Woźniak, M. Preprocessed dynamic classifier ensemble selection for highly imbalanced drifted data streams. Inf. Fusion 2021, 66, 138–154. [Google Scholar] [CrossRef]
  23. Swami, P.; Gandhi, T.; Panigrahi, B.K.; Tripathi, M.; Anand, S. A novel robust diagnostic model to detect seizures in electroencephalography. Expert Syst. Appl. 2016, 56, 116–130. [Google Scholar] [CrossRef]
  24. Blankertz, B.; Dornhege, G.; Krauledat, M.; Müller, K.R.; Curio, G. The non-invasive Berlin Brain-Computer Interface: Fast acquisition of effective performance in untrained subjects. NeuroImage 2007, 37, 539–550. [Google Scholar] [CrossRef]
  25. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  26. Ancora, L.A.; Blanco-Mora, D.A.; Alves, I.; Bonifácio, A.; Morgado, P.; Miranda, B. Cities and neuroscience research: A systematic literature review. Front. Psychiatry 2022, 13, 983352. [Google Scholar] [CrossRef] [PubMed]
  27. Gerner, N.; Pickerle, D.; Höller, Y.; Hartl, A. Neurophysiological Markers of Design-Induced Cognitive Changes: A Feasibility Study with Consumer-Grade Mobile EEG. Brain Sci. 2025, 15, 432. [Google Scholar] [CrossRef]
  28. He, C.; Chen, Y.Y.; Phang, C.R.; Stevenson, C.; Chen, I.P.; Jung, T.P.; Ko, L.W. Diversity and suitability of the state-of-the-art wearable and wireless EEG systems review. IEEE J. Biomed. Health Inform. 2023, 27, 3830–3843. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Heat map of classifier performance by ensemble strategy and fusion level (F1-weighted) for the EEG Epilepsy dataset.
Figure 1. Heat map of classifier performance by ensemble strategy and fusion level (F1-weighted) for the EEG Epilepsy dataset.
Applsci 15 06043 g001
Figure 2. Classifier performance by ensemble strategy and fusion level (F1-weighted) for the EEG Epilepsy dataset.
Figure 2. Classifier performance by ensemble strategy and fusion level (F1-weighted) for the EEG Epilepsy dataset.
Applsci 15 06043 g002
Figure 3. Heat map of classifier performance by ensemble strategy and fusion level (F1-weighted) for the BCI Competition IV Dataset 1.
Figure 3. Heat map of classifier performance by ensemble strategy and fusion level (F1-weighted) for the BCI Competition IV Dataset 1.
Applsci 15 06043 g003
Figure 4. Classifier performance by ensemble strategy and fusion level (F1-weighted) for the BCI Competition IV Dataset 1.
Figure 4. Classifier performance by ensemble strategy and fusion level (F1-weighted) for the BCI Competition IV Dataset 1.
Applsci 15 06043 g004
Figure 5. Comparison of F1-weighted obtained for all analyzed approaches.
Figure 5. Comparison of F1-weighted obtained for all analyzed approaches.
Applsci 15 06043 g005
Table 1. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the EEG Epilepsy dataset Part 1. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
Table 1. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the EEG Epilepsy dataset Part 1. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
No. of TablesMethod/Model F1F1
Prec.Recall(Weig.)(Macro)BAaccAcc
3LTAbstract level; sum; RF (100,gini)0.9030.8930.8920.8920.8930.893
Measurement level; sum; RF (200,entropy)0.940.9270.9260.9260.9270.927
Unified groups; abstract level; RF (100,gini)0.9030.8930.8920.8920.8930.893
Diverse groups; abstract level; RF (100,gini)0.9030.8930.8920.8920.8930.893
Unified groups; measurement level; RF (200,entropy)0.940.9270.9260.9260.9270.927
Diverse groups; measurement level; RF (200,entropy)0.940.9270.9260.9260.9270.927
Abstract level; sum; kNN (2)0.7790.760.7620.7620.760.76
Measurement level; sum; kNN (2)0.7460.7270.7230.7230.7270.727
Unified groups; abstract level; kNN (2)0.7790.760.7620.7620.760.76
Diverse groups; abstract level; kNN (2)0.750.720.7170.7170.720.72
Unified groups; measurement level; kNN (2)0.7360.720.7210.7210.720.72
Diverse groups; measurement level; kNN (2)0.7120.680.6830.6830.680.68
Abstract level; sum; AdaBoost (50)0.9460.9330.9320.9320.9330.933
Measurement level; sum; AdaBoost (50)0.9460.9330.9320.9320.9330.933
Unified groups; abstract level; AdaBoost (50)0.9460.9330.9320.9320.9330.933
Diverse groups; abstract level; AdaBoost (50)0.9380.9270.9260.9260.9270.927
Unified groups; measurement level; AdaBoost (50)0.9460.9330.9320.9320.9330.933
Diverse groups; measurement level; AdaBoost (50)0.9460.9330.9320.9320.9330.933
Abstract level; sum; Gradient Boosting (100)0.9720.9670.9660.9660.9670.967
Measurement level; sum; Gradient Boosting (50)0.9610.9530.9530.9530.9530.953
Unified groups; abstract level; Gradient Boosting (50)0.9830.980.980.980.980.98
Diverse groups; abstract level; Gradient Boosting (50)0.8390.8270.8230.8230.8270.827
Unified groups; measurement level; Gradient Boosting (50)0.9610.9530.9530.9530.9530.953
Diverse groups; measurement level; Gradient Boosting (200)0.9020.8930.8920.8920.8930.893
Abstract level; sum; Logistic Regression0.8780.8730.8740.8740.8730.873
Measurement level; sum; Logistic Regression0.8990.8930.8930.8930.8930.893
Unified groups; abstract level; Logistic Regression0.8780.8730.8740.8740.8730.873
Diverse groups; abstract level; Logistic Regression0.8320.8070.8090.8090.8070.807
Unified groups; measurement level; Logistic Regression0.8990.8930.8930.8930.8930.893
Diverse groups; measurement level; Logistic Regression0.870.8530.8550.8550.8530.853
5LTAbstract level; sum; RF (10,entropy)0.9090.8670.8560.8560.8670.867
Measurement level; sum; RF (10,gini)0.9180.8870.8820.8820.8870.887
Unified groups; abstract level; RF (10,entropy)0.9090.8670.8560.8560.8670.867
Diverse groups; abstract level; RF (20,gini)0.8660.820.8070.8070.820.82
Unified groups; measurement level; RF (10,entropy)0.9190.8870.8810.8810.8870.887
Diverse groups; measurement level; RF (50,gini)0.90.8670.860.860.8670.867
Abstract level; sum; kNN (2)0.7190.680.6880.6880.680.68
Measurement level; sum; kNN (4)0.7150.6730.6810.6810.6730.673
Unified groups; abstract level; kNN (2)0.7190.680.6880.6880.680.68
Diverse groups; abstract level; kNN (4)0.7430.720.7220.7220.720.72
Unified groups; measurement level; kNN (4)0.7150.6730.6810.6810.6730.673
Diverse groups; measurement level; kNN (2)0.7050.6730.6630.6630.6730.673
Abstract level; sum; AdaBoost (20)0.9570.9470.9460.9460.9470.947
Measurement level; sum; AdaBoost (100)0.9630.9530.9520.9520.9530.953
Unified groups; abstract level; AdaBoost (20)0.9780.9730.9730.9730.9730.973
Diverse groups; abstract level; AdaBoost (10)0.8970.880.8790.8790.880.88
Unified groups; measurement level; AdaBoost (20)0.9720.9670.9660.9660.9670.967
Diverse groups; measurement level; AdaBoost (20)0.910.8930.8910.8910.8930.893
Abstract level; sum; Gradient Boosting (10)0.8990.8670.8570.8570.8670.867
Measurement level; sum; Gradient Boosting (10)0.9070.8930.8920.8920.8930.893
Unified groups; abstract level; Gradient Boosting (200)0.90.8870.8840.8840.8870.887
Diverse groups; abstract level; Gradient Boosting (200)0.770.740.7430.7430.740.74
Unified groups; measurement level; Gradient Boosting (10)0.9070.8930.8920.8920.8930.893
Diverse groups; measurement level; Gradient Boosting (10)0.8780.860.8590.8590.860.86
Abstract level; sum; Logistic Regression0.90.860.860.860.860.86
Measurement level; sum; Logistic Regression0.9270.90.90.90.90.9
Unified groups; abstract level; Logistic Regression0.8740.840.8360.8360.840.84
Diverse groups; abstract level; Logistic Regression0.7620.70.6970.6970.70.7
Unified groups; measurement level; Logistic Regression0.8420.8270.8160.8160.8270.827
Diverse groups; measurement level; Logistic Regression0.8740.8470.8470.8470.8470.847
7LTAbstract level; sum; RF (20,gini)0.8530.8270.8180.8180.8270.827
Measurement level; sum; RF (20,entropy)0.9130.880.8750.8750.880.88
Unified groups; abstract level; RF (10,gini)0.8650.8070.7890.7890.8070.807
Diverse groups; abstract level; RF (200,entropy)0.7780.760.7530.7530.760.76
Unified groups; measurement level; RF (100,entropy)0.9070.8670.8590.8590.8670.867
Diverse groups; measurement level; RF (20,entropy)0.9150.8870.8810.8810.8870.887
Abstract level; sum; kNN (3)0.6190.580.5820.5820.580.58
Measurement level; sum; kNN (3)0.660.6470.6130.6130.6470.647
Unified groups; abstract level; kNN (3)0.6190.580.5820.5820.580.58
Diverse groups; abstract level; kNN (5)0.6630.6330.5950.5950.6330.633
Unified groups; measurement level; kNN (3)0.660.6470.6130.6130.6470.647
Diverse groups; measurement level; kNN (3)0.7070.6870.6640.6640.6870.687
Abstract level; sum; AdaBoost (100)0.9520.940.9390.9390.940.94
Measurement level; sum; AdaBoost (50)0.9520.940.9390.9390.940.94
Unified groups; abstract level; AdaBoost (20)0.9560.940.9390.9390.940.94
Diverse groups; abstract level; AdaBoost (100)0.8990.880.8780.8780.880.88
Unified groups; measurement level; AdaBoost (20)0.9570.9470.9450.9450.9470.947
Diverse groups; measurement level; AdaBoost (100)0.9670.960.9590.9590.960.96
Abstract level; sum; Gradient Boosting (10)0.9520.940.9380.9380.940.94
Measurement level; sum; Gradient Boosting (100)0.9670.960.9590.9590.960.96
Unified groups; abstract level; Gradient Boosting (10)0.9440.9330.9310.9310.9330.933
Diverse groups; abstract level; Gradient Boosting (200)0.760.7330.7170.7170.7330.733
Unified groups; measurement level; Gradient Boosting (100)0.9630.9530.950.950.9530.953
Diverse groups; measurement level; Gradient Boosting (200)0.8990.8870.8820.8820.8870.887
Abstract level; sum; Logistic Regression0.9230.8930.8880.8880.8930.893
Measurement level; sum; Logistic Regression0.9230.8930.8880.8880.8930.893
Unified groups; abstract level; Logistic Regression0.8410.8330.830.830.8330.833
Diverse groups; abstract level; Logistic Regression0.6630.5930.5740.5740.5930.593
Unified groups; measurement level; Logistic Regression0.8650.860.8580.8580.860.86
Diverse groups; measurement level; Logistic Regression0.7340.660.6520.6520.660.66
Table 2. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the EEG Epilepsy dataset Part 2. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color and bold represent best result).
Table 2. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the EEG Epilepsy dataset Part 2. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color and bold represent best result).
No. of TablesMethod/Model F1F1
Prec.Recall(Weig.)(Macro)BAaccAcc
9LTAbstract level; sum; RF (200,gini)0.9180.8870.8820.8820.8870.887
Measurement level; sum; RF (20,gini)0.9060.860.850.850.860.86
Unified groups; abstract level; RF (200,gini)0.9180.8870.8820.8820.8870.887
Diverse groups; abstract level; RF (200,gini)0.8490.8270.8240.8240.8270.827
Unified groups; measurement level; RF (10,gini)0.9070.8670.8590.8590.8670.867
Diverse groups; measurement level; RF (20,entropy)0.8840.840.830.830.840.84
Abstract level; sum; kNN (6)0.7070.6870.6640.6640.6870.687
Measurement level; sum; kNN (6)0.7760.6870.6640.6640.6870.687
Unified groups; abstract level; kNN (6)0.7070.6870.6640.6640.6870.687
Diverse groups; abstract level; kNN (6)0.7880.7270.6810.6810.7270.727
Unified groups; measurement level; kNN (6)0.7760.6870.6640.6640.6870.687
Diverse groups; measurement level; kNN (2)0.7270.6870.6940.6940.6870.687
Abstract level; sum; AdaBoost (100)0.9530.940.9390.9390.940.94
Measurement level; sum; AdaBoost (50)0.9340.9270.9260.9260.9270.927
Unified groups; abstract level; AdaBoost (50)0.9380.9270.9260.9260.9270.927
Diverse groups; abstract level; AdaBoost (100)0.7470.7130.7140.7140.7130.713
Unified groups; measurement level; AdaBoost (20)0.9280.920.920.920.920.92
Diverse groups; measurement level; AdaBoost (10)0.9120.8930.8920.8920.8930.893
Abstract level; sum; Gradient Boosting (10)0.9670.960.9590.9590.960.96
Measurement level; sum; Gradient Boosting (10)0.9640.9530.9520.9520.9530.953
Unified groups; abstract level; Gradient Boosting (10)0.9850.980.980.980.980.98
Diverse groups; abstract level; Gradient Boosting (20)0.6720.6530.6550.6550.6530.653
Unified groups; measurement level; Gradient Boosting (10)0.9890.9870.9870.9870.9870.987
Diverse groups; measurement level; Gradient Boosting (100)0.7630.7330.7380.7380.7330.733
Abstract level; sum; Logistic Regression0.7820.7670.7650.7650.7670.767
Measurement level; sum; Logistic Regression0.7910.780.7760.7760.780.78
Unified groups; abstract level; Logistic Regression0.7710.740.7420.7420.740.74
Diverse groups; abstract level; Logistic Regression0.5410.5530.5350.5350.5530.553
Unified groups; measurement level; Logistic Regression0.7790.7330.7390.7390.7330.733
Diverse groups; measurement level; Logistic Regression0.7120.580.5770.5770.580.58
11LTAbstract level; sum; RF (100,entropy)0.8790.8070.7870.7870.8070.807
Measurement level; sum; RF (10,gini)0.8820.820.7930.7930.820.82
Unified groups; abstract level; RF (200,gini)0.8610.80.780.780.80.8
Diverse groups; abstract level; RF (20,gini)0.7490.7130.6980.6980.7130.713
Unified groups; measurement level; RF (100,gini)0.8820.8130.7950.7950.8130.813
Diverse groups; measurement level; RF (100,entropy)0.8890.8270.810.810.8270.827
Abstract level; sum; kNN (2)0.7420.620.5270.5270.620.62
Measurement level; sum; kNN (5)0.5940.6130.5630.5630.6130.613
Unified groups; abstract level; kNN (2)0.7420.620.5270.5270.620.62
Diverse groups; abstract level; kNN (2)0.6730.640.6410.6410.640.64
Unified groups; measurement level; kNN (5)0.5940.6130.5630.5630.6130.613
Diverse groups; measurement level; kNN (3)0.8180.7070.6790.6790.7070.707
Abstract level; sum; AdaBoost (50)0.9240.920.920.920.920.92
Measurement level; sum; AdaBoost (100)0.9380.920.9180.9180.920.92
Unified groups; abstract level; AdaBoost (50)0.890.880.8790.8790.880.88
Diverse groups; abstract level; AdaBoost (10)0.7090.640.6370.6370.640.64
Unified groups; measurement level; AdaBoost (50)0.9320.920.9180.9180.920.92
Diverse groups; measurement level; AdaBoost (20)0.7860.740.7210.7210.740.74
Abstract level; sum; Gradient Boosting (10)0.870.8330.8250.8250.8330.833
Measurement level; sum; Gradient Boosting (20)0.8530.8270.820.820.8270.827
Unified groups; abstract level; Gradient Boosting (10)0.8440.820.8110.8110.820.82
Diverse groups; abstract level; Gradient Boosting (10)0.6850.6730.6650.6650.6730.673
Unified groups; measurement level; Gradient Boosting (20)0.8130.80.7930.7930.80.8
Diverse groups; measurement level; Gradient Boosting (20)0.860.8530.8510.8510.8530.853
Abstract level; sum; Logistic Regression0.7990.7270.710.710.7270.727
Measurement level; sum; Logistic Regression0.8090.740.7290.7290.740.74
Unified groups; abstract level; Logistic Regression0.7650.6870.680.680.6870.687
Diverse groups; abstract level; Logistic Regression0.5850.540.5290.5290.540.54
Unified groups; measurement level; Logistic Regression0.7770.7070.6870.6870.7070.707
Diverse groups; measurement level; Logistic Regression0.6180.580.5810.5810.580.58
Table 3. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the BCI Competition IV Dataset 1 Part 1. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
Table 3. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the BCI Competition IV Dataset 1 Part 1. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
No. of TablesMethod/Model F1F1
Prec.Recall(Weig.)(Macro)BAaccAcc
3LTAbstract level; sum; RF (10,gini)0.5330.5330.5320.5320.5330.533
Measurement level; sum; RF (50,entropy)0.520.5190.5180.5180.5190.519
Unified groups; abstract level; RF (10,gini)0.5330.5330.5320.5320.5330.533
Diverse groups; abstract level; RF (10,entropy)0.5230.5230.5220.5220.5230.523
Unified groups; measurement level; RF (50,entropy)0.520.5190.5180.5180.5190.519
Diverse groups; measurement level; RF (50,gini)0.5210.520.5180.5180.520.52
Abstract level; sum; kNN (2)0.4670.4840.4080.4080.4840.484
Measurement level; sum; kNN (4)0.4730.4740.4660.4660.4740.474
Unified groups; abstract level; kNN (2)0.4670.4840.4080.4080.4840.484
Diverse groups; abstract level; kNN (2)0.5280.5190.4780.4780.5190.519
Unified groups; measurement level; kNN (4)0.4680.4680.4670.4670.4680.468
Diverse groups; measurement level; kNN (4)0.4830.4840.4830.4830.4840.484
Abstract level; sum; AdaBoost (10)0.5740.5710.5660.5660.5710.571
Measurement level; sum; AdaBoost (200)0.5820.580.5780.5780.580.58
Unified groups; abstract level; AdaBoost (10)0.5740.5710.5660.5660.5710.571
Diverse groups; abstract level; AdaBoost (10)0.5550.5540.5530.5530.5540.554
Unified groups; measurement level; AdaBoost (200)0.5820.580.5780.5780.580.58
Diverse groups; measurement level; AdaBoost (10)0.5810.580.5780.5780.580.58
Abstract level; sum; Gradient Boosting (20)0.5860.5860.5850.5850.5860.586
Measurement level; sum; Gradient Boosting (20)0.5610.560.5590.5590.560.56
Unified groups; abstract level; Gradient Boosting (20)0.5860.5860.5850.5850.5860.586
Diverse groups; abstract level; Gradient Boosting (20)0.5560.5550.5540.5540.5550.555
Unified groups; measurement level; Gradient Boosting (20)0.5610.560.5590.5590.560.56
Diverse groups; measurement level; Gradient Boosting (200)0.5580.5560.5540.5540.5560.556
Abstract level; sum; Logistic Regression0.5370.5360.5360.5360.5360.536
Measurement level; sum; Logistic Regression0.5190.5190.5190.5190.5190.519
Unified groups; abstract level; Logistic Regression0.5370.5360.5360.5360.5360.536
Diverse groups; abstract level; Logistic Regression0.5420.5420.5410.5410.5420.542
Unified groups; measurement level; Logistic Regression0.5190.5190.5190.5190.5190.519
Diverse groups; measurement level; Logistic Regression0.5440.5440.5430.5430.5440.544
5LTAbstract level; sum; RF (10,entropy)0.5060.5060.5040.5040.5060.506
Measurement level; sum; RF (20,entropy)0.5070.5060.5040.5040.5060.506
Unified groups; abstract level; RF (10,entropy)0.5060.5060.5040.5040.5060.506
Diverse groups; abstract level; RF (100,entropy)0.530.530.5290.5290.530.53
Unified groups; measurement level; RF (20,entropy)0.5070.5060.5040.5040.5060.506
Diverse groups; measurement level; RF (200,entropy)0.5210.5210.5190.5190.5210.521
Abstract level; sum; kNN (2)0.4210.4760.3680.3680.4760.476
Measurement level; sum; kNN (3)0.4730.4730.4720.4720.4730.473
Unified groups; abstract level; kNN (2)0.4210.4760.3680.3680.4760.476
Diverse groups; abstract level; kNN (2)0.5170.5160.5020.5020.5160.516
Unified groups; measurement level; kNN (3)0.4730.4730.4720.4720.4730.473
Diverse groups; measurement level; kNN (6)0.5010.5010.50.50.5010.501
Abstract level; sum; AdaBoost (200)0.5680.5670.5660.5660.5670.567
Measurement level; sum; AdaBoost (20)0.5820.5810.5810.5810.5810.581
Unified groups; abstract level; AdaBoost (200)0.5680.5670.5660.5660.5670.567
Diverse groups; abstract level; AdaBoost (20)0.5210.5210.520.520.5210.521
Unified groups; measurement level; AdaBoost (20)0.5820.5810.5810.5810.5810.581
Diverse groups; measurement level; AdaBoost (20)0.5840.5840.5830.5830.5840.584
Abstract level; sum; Gradient Boosting (20)0.5570.5560.5550.5550.5560.556
Measurement level; sum; Gradient Boosting (10)0.5590.5570.5540.5540.5570.557
Unified groups; abstract level; Gradient Boosting (20)0.5570.5560.5550.5550.5560.556
Diverse groups; abstract level; Gradient Boosting (100)0.5340.5340.5330.5330.5340.534
Unified groups; measurement level; Gradient Boosting (10)0.5590.5570.5540.5540.5570.557
Diverse groups; measurement level; Gradient Boosting (50)0.5550.5550.5550.5550.5550.555
Abstract level; sum; Logistic Regression0.5080.5080.5060.5060.5080.508
Measurement level; sum; Logistic Regression0.5130.5130.5130.5130.5130.513
Unified groups; abstract level; Logistic Regression0.5080.5080.5060.5060.5080.508
Diverse groups; abstract level; Logistic Regression0.5310.5310.5290.5290.5310.531
Unified groups; measurement level; Logistic Regression0.5130.5130.5130.5130.5130.513
Diverse groups; measurement level; Logistic Regression0.5550.5550.5550.5550.5550.555
7LTAbstract level; sum; RF (50,entropy)0.5210.5210.520.520.5210.521
Measurement level; sum; RF (10,gini)0.5110.5110.510.510.5110.511
Unified groups; abstract level; RF (50,entropy)0.5210.5210.520.520.5210.521
Diverse groups; abstract level; RF (100,entropy)0.5370.5360.5350.5350.5360.536
Unified groups; measurement level; RF (10,gini)0.5110.5110.510.510.5110.511
Diverse groups; measurement level; RF (10,gini)0.5210.5210.5180.5180.5210.521
Abstract level; sum; kNN (2)0.4190.4830.3570.3570.4830.483
Measurement level; sum; kNN (2)0.4410.4490.4280.4280.4490.449
Unified groups; abstract level; kNN (2)0.4190.4830.3570.3570.4830.483
Diverse groups; abstract level; kNN (2)0.5310.5290.5240.5240.5290.529
Unified groups; measurement level; kNN (2)0.4420.4440.4380.4380.4440.444
Diverse groups; measurement level; kNN (5)0.5190.5180.5120.5120.5180.518
Abstract level; sum; AdaBoost (10)0.6010.5990.5960.5960.5990.599
Measurement level; sum; AdaBoost (50)0.5860.5840.5820.5820.5840.584
Unified groups; abstract level; AdaBoost (10)0.6010.5990.5960.5960.5990.599
Diverse groups; abstract level; AdaBoost (10)0.5240.5240.5240.5240.5240.524
Unified groups; measurement level; AdaBoost (50)0.5860.5840.5820.5820.5840.584
Diverse groups; measurement level; AdaBoost (10)0.5750.5750.5740.5740.5750.575
Abstract level; sum; Gradient Boosting (50)0.5570.5560.5550.5550.5560.556
Measurement level; sum; Gradient Boosting (50)0.5580.5580.5570.5570.5580.558
Unified groups; abstract level; Gradient Boosting (50)0.5570.5560.5550.5550.5560.556
Diverse groups; abstract level; Gradient Boosting (200)0.5290.5290.5280.5280.5290.529
Unified groups; measurement level; Gradient Boosting (50)0.5580.5580.5570.5570.5580.558
Diverse groups; measurement level; Gradient Boosting (20)0.5680.5670.5660.5660.5670.567
Abstract level; sum; Logistic Regression0.4780.4790.4780.4780.4790.479
Measurement level; sum; Logistic Regression0.4810.4810.480.480.4810.481
Unified groups; abstract level; Logistic Regression0.4780.4790.4780.4780.4790.479
Diverse groups; abstract level; Logistic Regression0.5360.5360.5340.5340.5360.536
Unified groups; measurement level; Logistic Regression0.4810.4810.480.480.4810.481
Diverse groups; measurement level; Logistic Regression0.5350.5350.5350.5350.5350.535
Table 4. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the BCI Competition IV Dataset 1 Part 2. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
Table 4. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for the considered approaches for the BCI Competition IV Dataset 1 Part 2. RF is the abbreviation Random Forest and kNN for k-Nearest Neighbors. (Blue text color represent best result).
No. of TablesMethod/Model F1F1
Prec.Recall(Weig.)(Macro)BAaccAcc
9LTAbstract level; sum; RF (10,gini)0.5090.5090.5070.5070.5090.509
Measurement level; sum; RF (10,entropy)0.5180.5180.5160.5160.5180.518
Unified groups; abstract level; RF (10,gini)0.5090.5090.5070.5070.5090.509
Diverse groups; abstract level; RF (100,entropy)0.5330.5330.5320.5320.5330.533
Unified groups; measurement level; RF (10,entropy)0.5180.5180.5160.5160.5180.518
Diverse groups; measurement level; RF (10,entropy)0.5250.5240.5220.5220.5240.524
Abstract level; sum; kNN (4)0.4730.4960.3710.3710.4960.496
Measurement level; sum; kNN (5)0.4540.4540.4540.4540.4540.454
Unified groups; abstract level; kNN (4)0.4730.4960.3710.3710.4960.496
Diverse groups; abstract level; kNN (6)0.5140.5140.5130.5130.5140.514
Unified groups; measurement level; kNN (5)0.4540.4540.4540.4540.4540.454
Diverse groups; measurement level; kNN (5)0.510.510.5070.5070.510.51
Abstract level; sum; AdaBoost (100)0.5650.5650.5650.5650.5650.565
Measurement level; sum; AdaBoost (100)0.5350.5340.5330.5330.5340.534
Unified groups; abstract level; AdaBoost (100)0.5650.5650.5650.5650.5650.565
Diverse groups; abstract level; AdaBoost (20)0.520.520.520.520.520.52
Unified groups; measurement level; AdaBoost (100)0.5350.5340.5330.5330.5340.534
Diverse groups; measurement level; AdaBoost (50)0.5890.5860.5830.5830.5860.586
Abstract level; sum; Gradient Boosting (10)0.5350.5340.5330.5330.5340.534
Measurement level; sum; Gradient Boosting (20)0.5190.5190.5170.5170.5190.519
Unified groups; abstract level; Gradient Boosting (20)0.5350.5340.5330.5330.5340.534
Diverse groups; abstract level; Gradient Boosting (20)0.5290.5290.5280.5280.5290.529
Unified groups; measurement level; Gradient Boosting (50)0.5190.5190.5170.5170.5190.519
Diverse groups; measurement level; Gradient Boosting (20)0.5660.5650.5640.5640.5650.565
Abstract level; sum; Logistic Regression0.5780.5770.5760.5760.5770.577
Measurement level; sum; Logistic Regression0.5740.5740.5720.5720.5740.574
Unified groups; abstract level; Logistic Regression0.5780.5770.5760.5760.5770.577
Diverse groups; abstract level; Logistic Regression0.5390.5390.5370.5370.5390.539
Unified groups; measurement level; Logistic Regression0.5740.5740.5720.5720.5740.574
Diverse groups; measurement level; Logistic Regression0.510.510.510.510.510.51
11LTAbstract level; sum; RF (10,entropy)0.5170.5160.510.510.5160.516
Measurement level; sum; RF (10,gini)0.4970.4970.4950.4950.4970.497
Unified groups; abstract level; RF (10,entropy)0.5170.5160.510.510.5160.516
Diverse groups; abstract level; RF (100,entropy)0.5330.5330.5320.5320.5330.533
Unified groups; measurement level; RF (10,gini)0.4970.4970.4950.4950.4970.497
Diverse groups; measurement level; RF (50,entropy)0.5210.5190.510.510.5190.519
Abstract level; sum; kNN (4)0.5390.5060.3770.3770.5060.506
Measurement level; sum; kNN (7)0.4510.4540.4450.4450.4540.454
Unified groups; abstract level; kNN (4)0.5390.5060.3770.3770.5060.506
Diverse groups; abstract level; kNN (2)0.5240.5240.5210.5210.5240.524
Unified groups; measurement level; kNN (7)0.4510.4540.4450.4450.4540.454
Diverse groups; measurement level; kNN (4)0.5030.5030.5020.5020.5030.503
Abstract level; sum; AdaBoost (10)0.5790.5770.5740.5740.5770.577
Measurement level; sum; AdaBoost (20)0.60.5970.5940.5940.5970.597
Unified groups; abstract level; AdaBoost (10)0.5790.5770.5740.5740.5770.577
Diverse groups; abstract level; AdaBoost (10)0.5220.5220.5220.5220.5220.522
Unified groups; measurement level; AdaBoost (20)0.60.5970.5940.5940.5970.597
Diverse groups; measurement level; AdaBoost (20)0.5790.5760.5730.5730.5760.576
Abstract level; sum; Gradient Boosting (10)0.5280.5280.5270.5270.5280.528
Measurement level; sum; Gradient Boosting (50)0.5270.5260.5260.5260.5260.526
Unified groups; abstract level; Gradient Boosting (10)0.5280.5280.5270.5270.5280.528
Diverse groups; abstract level; Gradient Boosting (200)0.5290.5290.5290.5290.5290.529
Unified groups; measurement level; Gradient Boosting (50)0.5270.5260.5260.5260.5260.526
Diverse groups; measurement level; Gradient Boosting (100)0.5480.5480.5460.5460.5480.548
Abstract level; sum; Logistic Regression0.5280.5280.5270.5270.5280.528
Measurement level; sum; Logistic Regression0.5130.5130.5110.5110.5130.513
Unified groups; abstract level; Logistic Regression0.5280.5280.5270.5270.5280.528
Diverse groups; abstract level; Logistic Regression0.5360.5360.5340.5340.5360.536
Unified groups; measurement level; Logistic Regression0.5130.5130.5110.5110.5130.513
Diverse groups; measurement level; Logistic Regression0.540.5390.5360.5360.5390.539
Table 5. p-values for the post hoc Dunn–Bonferroni test for appeoaches: unified groups unweighted and weighted; diverse groups unweighted and weighted. (Blue text color represent best result).
Table 5. p-values for the post hoc Dunn–Bonferroni test for appeoaches: unified groups unweighted and weighted; diverse groups unweighted and weighted. (Blue text color represent best result).
p-ValueAbstract
Level
Sum
Measurement
Level
Sum
Unified
Groups
Abstract Level
Diverse
Groups
Abstract Level
Unified Groups
Measurement
Level
Diverse Groups
Measurement
Level
Abstract level, sum 110.45611
Measurement level, sum1 10.13211
Unified groups, abstract level11 111
Diverse groups, abstract level0.4560.1321 0.4260.009
Unified groups,
measurement level
1110.426 1
Diverse groups,
measurement level
1110.0091
Table 6. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) in the case where the distributed environments are not applied. RF is the abbreviation Random Forest and kNN stands for k-Nearest Neighbors. (Blue text color represent best result).
Table 6. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) in the case where the distributed environments are not applied. RF is the abbreviation Random Forest and kNN stands for k-Nearest Neighbors. (Blue text color represent best result).
Data SetMethod/ModelPrec.RecallF1 (Weig.)F1 (Macro)BAaccAcc
EEG EpilepsyRF (200,gini)0.9460.9330.9320.9320.9330.933
kNN (4)0.7820.740.7160.7160.740.74
AdaBoost (20)0.9530.9470.9460.9460.9470.947
Gradient Boosting (50)0.9460.9330.9320.9320.9330.933
Logistic Regression0.8880.8670.8690.8690.8670.867
BCI Competition IV Dataset 1RF (20,gini)0.530.5290.5270.5270.5290.529
kNN (2)0.4850.4920.4420.4420.4920.492
AdaBoost (10)0.6060.6040.6030.6030.6040.604
Gradient Boosting (20)0.5970.5970.5970.5970.5970.597
Logistic Regression0.5420.5410.5410.5410.5410.541
Table 7. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for deep learning EEGNet-SSVEP model commonly used in EEG classification.
Table 7. Results of precision (Prec.), recall, F-measure (F-m.), balanced accuracy ( b a c c ), and classification accuracy ( a c c ) for deep learning EEGNet-SSVEP model commonly used in EEG classification.
Data SetPrec.RecallF1 (Weig.)F1 (Macro)BAaccAcc
EEG Epilepsy0.8360.7670.7370.7370.7670.767
BCI Competition IV Dataset 10.5160.5160.5130.5130.5160.516
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Przybyła-Kasperek, M.; Sacewicz, J. Dynamic Ensemble Selection for EEG Signal Classification in Distributed Data Environments. Appl. Sci. 2025, 15, 6043. https://doi.org/10.3390/app15116043

AMA Style

Przybyła-Kasperek M, Sacewicz J. Dynamic Ensemble Selection for EEG Signal Classification in Distributed Data Environments. Applied Sciences. 2025; 15(11):6043. https://doi.org/10.3390/app15116043

Chicago/Turabian Style

Przybyła-Kasperek, Małgorzata, and Jakub Sacewicz. 2025. "Dynamic Ensemble Selection for EEG Signal Classification in Distributed Data Environments" Applied Sciences 15, no. 11: 6043. https://doi.org/10.3390/app15116043

APA Style

Przybyła-Kasperek, M., & Sacewicz, J. (2025). Dynamic Ensemble Selection for EEG Signal Classification in Distributed Data Environments. Applied Sciences, 15(11), 6043. https://doi.org/10.3390/app15116043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop