Next Article in Journal
Adaptive Channel-Aware Garbage Collection Control for Multi-Channel SSDs
Previous Article in Journal
Assessing ChatGPT Adoption in Higher Education: An Empirical Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Ensemble Imbalanced Classification Framework via Dual-Perspective Overlapping Analysis with Multi-Resolution Metrics

1
China Electric Power Research Institute Company Limited, Beijing 100192, China
2
School of Intelligent Engineering and Automation, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(23), 4740; https://doi.org/10.3390/electronics14234740 (registering DOI)
Submission received: 20 October 2025 / Revised: 28 November 2025 / Accepted: 30 November 2025 / Published: 2 December 2025
(This article belongs to the Section Artificial Intelligence)

Abstract

The coexistence of class imbalance and overlap poses a major challenge in classification and significantly limits model accuracy. Data-level methods alleviate class imbalance by generating samples, but without ensuring their rationality, which may introduce noise. Algorithm-level methods are designed based on the model training process, avoiding noise introduction. However, existing methods often fail to consider the potential multiclass scenarios within overlap regions or design targeted solutions for different overlap patterns. This paper proposes an ensemble imbalanced classification framework via dual-perspective overlapping analysis with multi-resolution metrics. The dataset is divided into multiple resolutions for independent analysis, capturing distributional information from local to global levels. For each independent resolution, overlap is analyzed from the perspectives of “feature overlap” and “instance overlap” to derive more refined overlap scores. Flow model mapping and importance weighting are, respectively, applied to refine overlapping samples according to the two criteria. During testing, classifiers are adaptively selected based on the overlap degree of test samples under different criteria, and predictions across resolutions are integrated for the final decision. Experiments on 39 datasets demonstrate that the proposed method outperforms typical imbalanced classification methods in F-measure and G-mean, with particularly notable gains on 15 severely overlapping datasets.

1. Introduction

Classification problems hold a significant position in machine learning, with binary classification as the foundation. Due to data collection biases and natural class rarity, real-world datasets often exhibit long-tailed distributions [1], where minority class samples are critical for decision-making (e.g., disease detection [2], anomaly detection [3] and fraud identification [4]). Recent studies reveal that class imbalance is not the sole factor affecting classification performance. Sample overlap between different classes also significantly impacts model learning ability. Consequently, reasonable partitioning and targeted handling of overlapping regions have become a critical research topic.
In previous research, methods addressing class imbalance have primarily been categorized into two types: algorithmic-level approaches [5,6,7,8] and data-level approaches [9,10,11,12,13,14]. Algorithmic-level methods focus on mitigating feature bias that occurs during the model’s learning process. By artificially adjusting the model structure or designing loss weights, these methods ensure that the model pays more attention to the classification accuracy of minority class samples during training. Data-level methods, on the other hand, address the imbalance in the number of samples directly. Techniques such as oversampling or generative models are employed to augment the number of minority class samples, thereby balancing the numerical disparity between classes and preventing the model from developing biases due to learning from imbalanced datasets. Additionally, some researchers have considered the adverse effects of data overlap on model learning. They often define the degree of overlap autonomously and design rules for region partitioning, which, to some extent, assists the model in distinguishing samples within overlapping regions.
Existing methods address imbalance through data-level approaches (sampling, generation) [9,10,11,12,13,14] or algorithm-level approaches (cost-sensitive learning, ensemble methods) [5,6,7,8]. However, they suffer from two critical limitations: (1) Single-perspective overlap analysis: Current methods primarily rely on instance-level neighborhood analysis (e.g., k-Nearest Neighbors), failing to capture feature-space overlap patterns; (2) Fixed-scale measurement: Using a single neighborhood range cannot comprehensively characterize multi-modal overlap distributions across local and global scopes.
Data-level methods address the issue of class imbalance directly through sampling and generation techniques, thereby mitigating the decision bias problem during model training. Algorithm-level methods, on the other hand, achieve additional focus on minority class samples by means such as adaptive weight adjustment, avoiding overfitting issues caused by numerical imbalance. However, these methods not only suffer from inherent limitations but also lack an effective mechanism for identifying and handling overlapping regions within datasets. The insufficient consideration of overlap issues restricts the classification performance of these methods, preventing them from fully leveraging their algorithmic potential. Therefore, how to conduct a fine-grained analysis of the overlap degree across the entire dataset and design targeted solutions based on overlap metrics is a pressing problem that needs to be addressed. This paper proposes an ensemble imbalanced classification framework via dual-perspective overlapping analysis with multi-resolution metrics (DPOA-MRM). It utilizes “feature overlap” and “instance overlap” as two complementary pieces of information synergistically within the same framework for the first time, and achieves hierarchical overlap measurement from local to global levels through a multi-resolution strategy. This framework not only breaks through the limitations of traditional methods that rely on a single neighborhood range but also avoids information loss caused by resolution selection bias, thereby providing a more reliable foundation for the subsequent refined processing of overlapping samples. Specifically, the “multi-resolution” strategy effectively balances local details and global structure by independently calculating sample overlap degrees under different spatial scales; the “dual-perspective” criteria delve into the complex relationships between classes from the dimensions of feature separability and neighborhood heterogeneity. The integration of the two not only improves the accuracy of identifying overlapping regions but also provides more targeted guidance for the subsequent selection and fusion of classifiers. Building on this, the method further refines the processing of overlapping samples by designing flow model mapping and importance weight calculation techniques tailored to the differences in overlap information obtained from the two perspectives. For each test sample, we calculate its overlap degrees under both perspectives across all resolutions. Classifiers matching the sample’s overlap profile are selected, and their confidence scores are integrated for final prediction. The main contributions of this paper are as follows:
A dual-perspective multi-resolution overlap measurement framework is proposed. We propose a unified framework that analyzes overlap from both “feature” and “instance” perspectives across multiple resolutions. This approach captures distributional information from local to global levels, overcoming the limitations of single-metric methods in identifying complex overlap patterns.
A method for overlapping feature separation based on flow model mapping is proposed. To address the “feature overlap” challenge, we design a flow model-based mapping strategy. This mechanism projects highly entangled samples into a lower-dimensional, separable latent space, significantly reducing the difficulty of learning discriminative features in boundary regions.
A method for calculating attention weights based on multi-nearest-neighbor distribution discrimination in the data neighborhood is proposed. Focusing on “instance overlap,” we develop an adaptive weighting scheme that categorizes samples based on their neighborhood heterogeneity. This enables the model to selectively prioritize hard-to-classify boundary samples while suppressing noise and redundant information from the majority class.
The structure of the subsequent sections of this paper is outlined below. Section 2 provides an overview of the prior research pertinent to this study. Section 3 details the proposed DPOA-MRM methodology. Subsequently, Section 4 outlines the experimental approaches and metrics for performance evaluation that were employed to substantiate our methodology. Finally, Section 5 concludes the paper.

2. Related Work

This paper addresses the shortcomings of existing methods for imbalanced classification and proposes an integrated imbalanced classification method based on multi-resolution overlap measurement and dual-perspective overlap criteria. The related prior work associated with this paper can be mainly divided into three parts: (1) Data-level imbalanced classification methods; (2) Algorithm-level imbalanced classification methods; (3) Relevant work on overlap region division.

2.1. Data-Level Imbalanced Classification Methods

Data-level balancing approaches address class imbalance through dual strategies: either expanding the representation of the minority class or reducing the number of instances in the majority class. Current methodologies predominantly bifurcate into resampling techniques and generation-driven solutions. Resampling mechanisms operate through dataset distribution analysis, either creating synthetic instances (oversampling) or eliminating redundant samples (undersampling). The SMOTE framework (Synthetic Minority Over-sampling Technique) [9] exemplifies this paradigm, employing KNN (k-Nearest Neighbors) [15]-guided random linear interpolations between minority instances to counteract overfitting risks inherent in naive oversampling. Subsequent enhancements to this foundation have emerged, including methodologies detailed in [16] that introduce localized density metrics for minority samples alongside hybrid cleansing protocols. These advanced techniques eliminate majority class intrusions in overlapping feature spaces while establishing weighted sampling criteria informed by decision boundary adjacency and regional density, facilitating synthetic generation across both core and marginal data regions. Nevertheless, such resampling strategies primarily address data distribution disparities without effectively harnessing intrinsic sample characteristics, resulting in synthetic outputs that fail to preserve original feature integrity and may inadvertently incorporate extraneous noise.
Contemporary investigations have concurrently addressed challenges related to boundary ambiguity through enhanced sampling methodologies. Conventional SMOTE implementations [9] inherently exhibit limitations in handling class overlap areas, occasionally exacerbating decision boundary complexity through synthetic instance generation. In response, the SMOTE-NaN-DE framework [17] introduces dual innovations: natural neighbor error analysis for noise/boundary detection and differential evolutionary optimization for refined sample placement. This synergistic mechanism enhances interclass separability while maintaining original data integrity, though its capacity to exploit nuanced overlap patterns remains underdeveloped. Parallel developments include MPP-SMOTE [18], which conducts a comparative analysis of interpolation paradigms, strategically blending linear and feature-space augmentation based on classification complexity metrics. This hybrid approach minimizes emergent overlap artifacts during synthesis by utilizing localized feature interpolation, which is particularly effective for boundary-proximate samples that exhibit high discriminative uncertainty. HSCF [19] accurately locates information-rich areas by dynamically dividing feature intervals and calculating density ratios and proposes a hybrid sampling framework that balances oversampling and undersampling adaptively, thereby avoiding overfitting and information loss and achieving a more balanced data distribution. Nevertheless, both methodologies demonstrate constrained effectiveness in resolving intrinsically ambiguous overlap patterns inherent in high-dimensional imbalanced contexts. Sampling methods fail to learn overlapping region features and rely solely on neighbor class distribution, limiting performance improvements.
Generation-based methods learn the features of minority class samples through deep models and generate new samples to achieve sample balance. The commonly used deep generative models are mainly Variational Autoencoders (VAEs) [13] and Generative Adversarial Networks (GANs) [14]. VAE is an improvement of the Autoencoder (AE), which introduces the Evidence Lower Bound (ELBO) from variational inference to address the issue of difficult-to-predict posterior distributions of generated samples in AE, allowing it to describe observations of the latent space probabilistically. GAN leverages the idea of adversarial training by constructing an additional discriminator that competes with the generator during training, thereby enhancing the realism of the generated samples.
Recent advancements have witnessed domain-specific enhancements of these foundational architectures. Notably, literature [20] synergized variational autoencoders with attention mechanisms for enhanced partial discharge diagnostics in geospatial monitoring systems. Concurrently, literature [21] has devised VLAD, a lifelong learning framework that merges VAE architectures with experience replay and structured memory consolidation for adaptive anomaly detection, while such task-oriented adaptations demonstrate improved performance, they lack generalized quality assurance for synthesized instances. Complementary innovations include WGAN-GP [22] extension incorporating conditional constraints to better approximate sample distributions, thereby enhancing the realism of synthetic outputs for imbalance mitigation. Pioneering work by literature [23] established GAN-based minority sample generation with dual quantitative metrics: MMD for distribution divergence analysis and silhouette coefficients for cluster separability evaluation, setting new benchmarks for synthetic sample assessment. These developments collectively highlight the field’s progression towards more sophisticated generation-evaluation pipelines, though fundamental challenges in sample validity preservation persist.
Deep architectures inherently demand substantial parametric complexity, posing challenges in imbalanced classification scenarios where dataset scales frequently prove inadequate for supporting such intricate models. Particularly in cases of extreme class disproportion or absolute scarcity of minority instances, the paucity of discriminative features severely restricts models’ capacity to synthesize statistically reliable samples for imbalance correction. Targeting this limitation, contemporary research emphasizes mining intricate feature patterns within overlapping instances through enhanced generative frameworks. The ADA-INCVAE [24] methodology, evolved from VAE foundations, implements neighborhood distribution analysis to quantify permissible generation scopes and volumes, thereby strategically constraining synthetic instance production to minimize overlap expansion. Parallel innovations include RGAN-EL [25], a GAN-derived architecture that amalgamates hybrid generator loss functions with selective filtration mechanisms to elevate sample fidelity. This approach further integrates ensemble learning principles to boost model generalizability. Complementarily, RVGAN-TL [26] introduces probabilistic overlap quantification through roulette-based minority sample selection, enabling focused extraction of information-dense patterns from boundary regions via transfer learning paradigms.
These methods have addressed existing issues in each model and proposed solutions for the overlapping region, achieving some degree of additional attention to the feature information of overlapping samples and obtaining certain results. Generation-based methods often choose to perform additional feature learning on overlapping samples to enable models to better synthesize information-rich overlapping samples during the generation of new samples, thereby enhancing the learning ability of subsequent classification models. The problems with this approach are as follows: First, the rationality of the generated new samples lacks a mechanism for assurance. Second, whether the generated overlapping samples are beneficial to the learning of subsequent classifiers cannot be measured. Finally, the deep models used in generation-based methods make it difficult to leverage their advantages under limited sample conditions, and the quality of generated samples is often poor.

2.2. Algorithm-Level Imbalanced Classification Methods

Algorithm-level methods modify the loss functions or weight matrices of existing models. This enables models to focus more on minority class features during training. Cost-sensitive learning [27,28] is a representative approach in this category. Researchers have proposed various characteristic methods for addressing imbalanced classification through mathematical analysis and other means. The authors of [27] considered the differences in modeling errors among samples of different classes and assigned weights to samples based on their modeling errors. By minimizing the mean and variance of modeling errors, the study achieved the classification of imbalanced data. The authors of [28] proposed an adaptive-weight-based cost-sensitive support vector machine (SVM) ensemble method, using cost-sensitive SVM as the base classifier and integrating the results of base classifiers with an improved cost-sensitive Boosting scheme. This approach helps to bias the final decision boundary towards the minority class. However, cost-sensitive learning lacks broad applicability and often requires specific designs for particular problems to achieve satisfactory results. Moreover, the misclassification costs for specific problems are often difficult to obtain, and designing algorithms based solely on past experience is unlikely to yield optimal results, requiring extensive experimentation for validation. Additionally, focusing on minority class samples can easily lead to overfitting.
In addition to cost-sensitive learning, one-class learning and ensemble learning also hold important positions in algorithm-level methods. Typical one-class algorithms include SVDD [6] and iForest [7], which only require learning the features of a single class of samples. Therefore, the decision bias caused by the imbalance in the number of samples does not affect their performance, making them suitable for solving imbalanced classification problems. However, one-class learning ignores the feature information of minority class samples, resulting in suboptimal performance under mild imbalance conditions and limited generalizability. Ensemble learning methods are mainly divided into two categories: Bagging-based methods and Boosting-based methods, which are often combined with data-level sampling methods to obtain more diverse training datasets and more robust integrated results. Common base models such as RF [29], GDBT [30], and XGBoost [31] fall within the scope of ensemble learning. The authors of [32] proposed a hybrid sampling strategy that adaptively samples from the majority class and adds these samples to the minority class dataset. By dynamically adjusting the sampling weights, multiple training sets are generated, and the results of the base classifiers are integrated to obtain the final result. The authors of [33] first divided the dataset into different regions and then used different kernel functions to map boundary samples to higher dimensions. By integrating the results of multiple support vector machines, the study achieved a multiclass classification of imbalanced data. MDSampler [34] proposes an ensemble sampler based on meta-distribution, which integrates semi-supervised learning, imbalanced learning, and ensemble learning through iterative instance undersampling and cascade classifier aggregation. It not only improves the diversity of the model but also further improves the overall performance by integrating the advantages of multiple classifiers. However, although these methods provide more accurate and robust classification results, they are essentially simple combinations of weak classifiers, and their performance is limited by the inherent performance of the classifiers themselves. They fail to explore the feature differences between different classes of samples thoroughly.
Contemporary algorithmic innovations have emerged to tackle interclass data overlap challenges through multiple strategic dimensions. RUE [35] established spatial localization metrics coupled with cost distribution mechanisms, evaluating sample significance through distributional patterns and positional correlations to determine misclassification penalties. PCGDST-IE [36] introduced a combined clustering and dimensional compression approach, restructuring training data into simplified distribution subgroups. Subsequent overlap resolution and rebalancing processes within each subgroup amplified minority class prominence in overlapping zones while optimizing subgroup differentiation. HDAWCR [37] and AWLICSR [38] employed spatial density analyses across intraclass and interclass relationships, utilizing spatial theory principles to augment the classification efficacy of the Collaborative Representation-based Classification (CRC) [39] paradigm. DPHS-MDS [40] implemented zonal partitioning through majority class neighborhood density analysis, segmenting feature space into four distinct regions. This framework adaptively applies region-specific hybrid sampling techniques and deploys tripartite ensemble architectures, dynamically activating optimal classifiers based on the spatial configurations of test instances.
Algorithm-level methods, which are based on expert experience and mathematical derivation, have been designed to address specific tasks by modifying the loss functions or weights. These modifications can, to some extent, alleviate the decision bias caused by data imbalance. However, these methods do not fundamentally solve the problem of data imbalance. Moreover, the approach of modifying weights does not facilitate the model’s ability to extract feature information from the samples more effectively. As a result, algorithm-level methods are often limited by the specific tasks and datasets to which they are applied, lacking generalizability. Additionally, when dealing with datasets that have complex and diverse overlapping situations, a single algorithm may struggle to cover all overlapping scenarios, thus limiting the upper bound of its performance.

2.3. Relevant Work on Overlap Region Division

In imbalanced classification, sample overlap is considered the primary issue affecting model accuracy. Its impact often exceeds that of data imbalance itself. Data overlap and data imbalance are not strongly correlated. They exist independently in various datasets as distinct manifestations of data distribution characteristics. However, for classification problems, the coexistence of data overlap and data imbalance significantly increases the difficulty for models to correctly learn the differences between classes and make accurate class judgments.Limited minority samples make overlapping features harder to learn. On the other hand, the imbalance within the overlapping region further complicates the already complex and difficult-to-handle overlap area, making it harder to locate and address specifically. Overlap handling has become critical in imbalanced classification, with studies in information fusion, finance [41], healthcare [42], software [43], and network systems [44].
Considering that the manifestation of data overlap exists not only at the data distribution level but also at the feature level, researchers have conducted in-depth studies and explorations on the localization and division of overlapping regions, proposing a variety of methods for identifying overlapping areas. Currently, the main methods for dividing overlapping regions can be categorized as follows:
1. Feature Overlap: Feature overlap focuses on regions with low class separability. Researchers use projection and distribution-based identification methods. Literature [45] estimated the descriptors of data space distribution and the shape and size of decision boundaries to measure classification complexity, thereby identifying feature overlap regions in the dataset. The classification complexity can be used to estimate the difficulty of separating data points into their respective classes, thus locating regions with low class separability. However, the measurements derived from feature overlap are based on the assumption of continuous features. Moreover, measurements based on feature space division are susceptible to the concept of separation and noisy data, which limits their applicability.
2. Instance Overlap: Instance overlap measures local characteristics through neighborhood analysis rather than global features. The most common measurement method combines this “complexity” with the k-Nearest Neighbors (kNN) classifier, which measures the overlap degree by analyzing the number of heterogeneous samples within the neighborhood of the current sample. The authors of [46] improved existing model-based measurement standards by incorporating data imbalance rate indicators during the estimation process, proposing an Augmented R-value metric. The study also provided theoretical proof and empirical validation of the effectiveness of the new measurement standard in estimating overlap. However, instance overlap often focuses only on the nearest neighbors surrounding a sample, extracting overlap information from local information and lacking a global perspective. When the dataset contains multiple overlapping regions or multiple distribution patterns, instance overlap often struggles to identify complex overlapping ranges accurately.
3. Structural Overlap: Structural overlap measurement, which is related to the complexity of data morphology, is considered a good indicator for identifying class overlap. As previously analyzed, class overlap accumulates a large number of complexity sources, including both feature space overlap and neighborhood overlap. In addition, the topology, shape, or structure of the data itself may have hidden dependencies on the degree of overlap. On the one hand, the global characteristics of the domain affect the identification of problematic regions, thereby influencing the quantification and representation of class overlap. On the other hand, class overlap directly affects the shape of decision boundaries between classes and may introduce additional complexities, such as class skewness, which alters the structural properties of the domain. The authors of [47], from the perspective of meta-learning, analyzed the relationship between the performance of existing methods and various data complexity metrics, using only metrics calculated from the data to determine the best-performing method for a given dataset. This study provided targeted improvements to existing metrics.
Existing methods identify overlaps independently without integration, limiting their ability to precisely localize overlapping regions. From the analysis of existing work in the field of imbalanced classification, it is evident that the current methods for dividing overlapping regions often rely solely on instance-based division, which means that only local information within the sample neighborhood can be captured. Existing methods neither consider the differences in overlap degrees of the same samples from different perspectives nor conduct refined processing of overlapping samples obtained under different overlap measurement conditions. How to comprehensively assess the degree of overlap from a multi-resolution perspective, ranging from global to local, and design corresponding overlap sample processing methods that match the characteristics of different overlap measurements is a key challenge in solving imbalanced classification problems.

3. The Proposed Method: DPOA-MRM

Algorithm-level methods address classification boundary shift by designing loss functions or modifying weight matrices, while achieving relatively accurate results, these methods lack generalizability. Data-level methods synthesize new samples by learning features and spatial distributions. Sampling methods, however, do not learn inherent data features. Generation-based methods cannot verify sample rationality, potentially introducing noise. Generation-based methods, although learning sample features through deep networks, cannot verify the rationality and effectiveness of the generated samples, potentially introducing additional noise. Existing methods use only local neighborhood information and single-scale measurement, causing inaccurate overlap localization. The proposed framework integrates multi-resolution analysis (global to local) with dual-perspective criteria (feature and instance) to accurately identify overlapping regions. This addresses neighborhood range selection challenges while capturing comprehensive overlap information. On this basis, for each independent resolution scenario, the degree of data overlap is calculated from both the “feature overlap” and “instance overlap” perspectives. The dual-perspective overlap calculation method facilitates an in-depth exploration of the complex feature relationships between samples of different classes and the distribution information within the neighborhood space, thereby yielding more refined and accurate overlap scores and avoiding the one-sidedness of single-perspective information. To further analyze and process the identified overlapping regions, the method employs flow model mapping and importance weight calculation techniques to refine the handling of two types of overlapping samples: those identified by feature overlap and those identified by instance overlap. During the testing phase, the feature overlap and instance overlap degrees of each test sample are first calculated. Based on these, a set of classifiers matching its overlap state is selected, and the final precise and robust classification result is obtained by integrating the class confidence scores from these classifiers.

3.1. The Framework for Overlap Degree Analysis Based on Dual-Perspective and Multi-Resolution Measurements

Current methods face two critical limitations: (1) fixed neighborhood ranges fail to capture multi-scale overlap patterns; (2) single-perspective analysis (instance-only or feature-only) cannot comprehensively characterize complex overlaps. Our multi-resolution dual-perspective framework directly addresses these limitations. This is essential for subsequent targeted processing of overlapping samples. Additionally, instance-based overlap analysis focuses on the distribution of samples within the neighborhood, primarily considering the internal spatial distribution patterns of the dataset. However, overlap situations in various datasets are complex and diverse, and analyzing overlap from a single perspective is insufficient for obtaining detailed and accurate results. To address these issues, a framework for overlap degree analysis based on dual-perspective and multi-resolution measurements is proposed. This framework aims to provide a more comprehensive understanding of overlap by integrating different perspectives and resolutions. The entire process of DPOA-MRM is shown in Figure 1.
Firstly, for the dataset to be processed, the degree of overlap is analyzed from the dual perspectives: “feature overlap” and “instance overlap.” The feature overlap analysis consists of two parts: feature selection and overlap degree calculation. Initially, the complexity and importance of the original features are analyzed using the F1 (Maximum Fisher’s discriminant ratio) metric and the Random Forest (RF) algorithm. Based on this, important and less complex features are selected to form a new feature space, and the distribution of the current feature space is analyzed to obtain the overlap degree for each sample. The instance overlap analysis is based on the k-Nearest Neighbors (kNN) algorithm to explore the neighborhood range of samples. Different overlap categories are divided according to the data distribution characteristics and the importance of different class samples, thereby achieving a refined instance overlap analysis. Multi-resolution measurement addresses single-range limitations by analyzing overlap at five scales. At each resolution:
Feature overlap: Flow model mapping reduces feature complexity.
Instance overlap: Overlap scores augment training data as attention weights.
This dual processing strategy matches the distinct characteristics of each overlap type.
Based on the predefined multi-resolution parameters, multiple sets of feature and instance overlap analysis results and multiple classifiers are obtained under different analysis ranges. By placing the test samples into the original feature space for overlap degree analysis and calculating the corresponding classification results, the final prediction class is obtained by integrating multiple sets of classification results.
Adopting a multi-resolution strategy is crucial because overlap phenomena are scale-dependent. A fixed neighborhood size may miss global structural overlaps or overfit to local noise. By analyzing data across varying resolutions, our framework captures both fine-grained boundary details and broader distributional trends. Resolution-adaptive weighting (Equation (1)) automatically balances the two perspectives. With a small neighborhood (S), the instance-based metric is susceptible to local noise. In contrast, feature-based metrics (calculated globally or on subsets) provide more stable guidance at this fine granularity. Therefore, as S increases, reliance is shifted towards the instance distribution.
δ = 1 S m S m ,
where m is a hyperparameter used to control the magnitude of the weight δ ; S represents the number of neighboring samples examined in the feature space at the current resolution, which reflects the spatial scope considered when analyzing the degree of overlap to some extent.
This integrated approach compensates for single-scale, single-perspective limitations, effectively addressing multi-modal overlap phenomena for targeted processing.

3.2. Hybrid Feature Separation Method Based on Flow Model Mapping

Most methods classify samples in the original feature space. Classifier performance is often limited by complex decision boundaries and severe interclass overlap. Mapping samples to a latent space with simpler distributions can effectively alleviate these issues. This reduces classification difficulty. Following this rationale, for the mixed features identified through feature overlap calculation, a flow model is introduced to map them into a low-dimensional, highly separable space. These mapped features are then fed into the classifier for feature learning, effectively mitigating the challenge of learning overlapping features in the original space, as illustrated in Figure 2 below.
Flow models employ reversible transformations to map overlapping samples into low-dimensional spaces. This preserves discriminative information while reducing feature complexity. The three-component architecture (Figure 3), comprising a normalization layer, a reversible transformation layer, and an affine coupling layer, enables effective nonlinear feature disentanglement for improved separability.
In the model, the input sample is denoted as X = x 1 , x 2 , , x m , and the inputs and outputs of each module are represented as u and v, respectively. The flow model consists of three components:
(1) Sample Normalization Layer:
Stabilizes training via learnable transformations (Equation (2)):
v = p u + d ,
where ⊙ denotes the element-wise multiplication of vectors. The loss for the sample normalization layer can be expressed as:
L o s s = k = 1 n log s k ,
where s k represents the k t h element of the vector.
(2) Invertible Transformation Layer:
In the invertible transformation layer, the sample dimensions consist of two parts: the first part remains consistent with the original dimensions, while the other part undergoes an affine transformation. To preserve the association information between original samples to the greatest extent during the transformation, a certain degree of permutation transformation is applied to the sample feature dimensions before the sample enters the affine coupling layer. To ensure the invertibility of the mapping, a 1 × 1 invertible transformation is introduced to replace the random shuffling operation in traditional flow models. The sample after the invertible transformation can be represented as:
v = H u ,
where H represents a randomly initialized
n × n
matrix. The loss function for the invertible transformation layer can be expressed as:
L o s s P = log det ( H ) .
(3) Affine Coupling Layer:
The affine coupling layer was first proposed in [48], which addressed the issue of the difficult computation of the Jacobian determinant in the original flow model and further maintained the invertibility of the model. The affine coupling layer divides the original sample into two parts, u 1 : d and u d + 1 : n , where 1 : d denotes the first to the d t h dimensions of the sample, and  d + 1 : n denotes the ( d + 1 ) t h to the n t h dimensions of the sample. The affine coupling layer applies an affine transformation to u d + 1 : n , while u 1 : d remains unchanged. The scale transformation p and the shift transformation d of the affine transformation are obtained by the nonlinear transformation N N , and can be specifically expressed as:
p , d = N N u 1 : d ,
The features u d + 1 : n after the affine transformation can be expressed as:
v d + 1 : n = p u d + 1 : n + d ,
Therefore, the Jacobian matrix of the invertible transformation can be expressed as:
z x = I d 0 v d + 1 : D u 1 : d v d + 1 : D u d + 1 : D = I d 0 v d + 1 : D u 1 : d d i a g ( s ) ,
where I d denotes the d × d identity matrix, and  d i a g ( s ) represents the diagonal matrix with s as its diagonal elements. The loss for the affine coupling layer can be expressed as:
L o s s A = k = 1 n d log s k .
Directly classifying samples in the original feature space is often suboptimal due to the intricate entanglement of minority and majority class features. The flow model is essential here as it creates a nonlinear mapping to a latent space where class boundaries become more distinct, thereby simplifying the subsequent classification task.
(1) Specific Structure of the Flow Model
The adopted flow model is mainly composed of three reversible transformation layers stacked sequentially. Its core idea is to transform the complex original feature distribution into a tractable base distribution (such as a standard normal distribution) through a series of bijective mappings. The overall architecture is shown in Figure 3. The model input is the selected feature-overlapping sample set X. This model is a sequence composed of L flow steps. Each flow step contains the following three sub-layers:
Activation Normalization Layer: As the first sub-layer, it applies an affine transformation to the data of each dimension for initial normalization. Its parameters (scale s and bias b) are calculated based on the mean and variance of the first batch of training data during initialization and are updated during training.
Invertible 1 × 1 Convolution Layer: As the second sub-layer, it replaces the traditional random permutation operation. It performs an intelligent shuffle of feature dimensions through a learnable rotation matrix W (initialized as an identity matrix) to enhance the model’s expressive power.
Affine Coupling Layer: As the core sub-layer, it splits the input u into u 1 and u 2 . u 1 remains unchanged and is input into a sub-neural network, which outputs the scale factor (s) and translation amount (t) used to transform u 2 . The transformation formula is: v 2 = u 2 exp s u 1 + t u 1 , while v 1 = u 1 . Finally, v 1 and v 2 are concatenated as the output.
The sub-neural network in the coupling layer is a simple Multi-Layer Perceptron (MLP), with the structure including: (i) Input layer: dimension d (when the dimension of u is n, d = n / 2 ); (ii) Hidden layers: 2 layers, with  4 d neurons per layer; (iii) Activation function: LeakyReLU (negative slope = 0.01) for hidden layers; (iv) Output layer: dimension n d , the output vector is split equally into scale factor s and translation amount t.
(2) Detailed Description of the Training Process
The training objective of the flow model is to maximize the probability of the mapped data in the log-likelihood space. According to the change of variable formula, the loss function is defined as the Negative Log-Likelihood (NLL):
L = E log p Z f x + log d e t f x
where f is the entire flow transformation, p Z is the prior distribution (standard normal distribution), and the second term is the logarithm of the determinant of the Jacobian matrix of the transformation, used to measure volume change.
Regarding training settings, the feature-overlapping sample set was randomly split into a training set and validation set at a 4:1 ratio; use the Adam optimizer; set the initial learning rate to 1 × 10 3 ; dynamically adjust the batch size based on the dataset size, set to min(256, N/10), where N is the number of training samples; employ an early stopping strategy for training epochs, stopping when the loss on the validation set does not decrease for 20 consecutive epochs, with the maximum epochs set to 1000.
(3) Explanation of Key Parameter Settings
Flow Steps (L): Selection was made from 4, 6, 8, 10 via grid search and it was determined that L = 8 achieves the best results on most datasets, striking a balance between complexity and expressive power.
Latent Space Dimension: The input and output dimensions of the flow model remain consistent with the original feature dimensions. The goal is feature disentanglement rather than dimensionality reduction. The mapped features retain the integrity of the original information, but the distribution is easier for the classifier to learn.
Sub-network Hidden Layer Size: Set to 4 d . This is an empirical value intended to provide sufficient nonlinear transformation capability while avoiding overfitting.
The training process of the flow model is shown in Algorithm 1.
Theoretical advantages of the flow model and the argumentation for improving interclass separability are as follows:
(1) Theoretical Value of Reversibility:
The flow model is based on the essence of reversible transformation, ensuring information losslessness and precise probability density transformation.
Information Losslessness: The invertible mapping f : R d R d ensures that all discriminative information of the original feature space can be retained in the low-dimensional space. Mathematically, reversibility ensures f 1 ( f ( x ) ) = x , meaning the low-dimensional representation z = f ( x ) preserves all feature information of the original sample x. In class overlap regions, discriminative features between samples of different classes are often very subtle, possibly reflected only in minute differences in certain feature dimensions. Traditional dimensionality reduction methods (like PCA retaining principal components) might lose this crucial but low-variance discriminative information, causing originally separable samples to become inseparable after reduction. The information losslessness of the flow model ensures that even the subtlest discriminative features can be preserved in the low-dimensional space, allowing the classifier to accurately distinguish overlapping samples in the simplified feature space.
Algorithm 1 The calculation process of feature overlap.
Require: Training dataset D t r a i n = { ( x i , y i ) | i N } , the training epochs of flow model e p o c h , The number of samples sampled during each training of the flow model b a t c h s i z e .
Ensure: Feature overlapping sample set D o v e r l a p ; Non overlapping feature sample set D n o n o v e r l a p ; Trained Flow Model θ f l o w ; Trained feature overlapping sample classifier θ o v e r l a p c l s ; Trained feature non overlapping sample classifier θ n o n o v e r l a p c l s ;
  1:
for  f i in [ f 1 ,   f 2 ,   . . . ,   f n ]  do
  2:
      F i = ( μ f i 0 μ f i 1 ) 2 σ f i 0 2 σ f i 1 2
  3:
end for
  4:
Obtain the degree of overlap of all features F = [ F 0 ,   F 1 ,   . . . ,   F n ]
  5:
Use random forest to calculate the importance of all features I = [ I 0 ,   I 1 ,   . . . ,   I n ]
  6:
Obtain a comprehensive discrimination index for feature importance based on the degree of feature overlap and importance P = F I = [ F 0 · I 0 ,   F 1 · I 1 ,   . . . ,   F n · I n ] = [ P 0 ,   P 1 ,   . . . ,   P n ]
  7:
Obtain the maximum feature importance level P s u b = [ P i ,   P j ,   . . . ,   P t ] that occupies the top of the total and its corresponding feature F s u b = [ f i ,   f j ,   . . . ,   f t ]
  8:
for  f i in F s u b  do
  9:
      subD = [ D t r a i n [ : ,   f i ] ,   y ]
10:
end for
11:
Obtain the feature overlap degree of each sample by integrating the importance of features and the instance overlap degree of subsets F e a = [ I n s subD i T , I n s subD j T , . . . , I n s subD t T ] P s u b
12:
φ = F o c S N d i f f + 1 I g
13:
Divide the training set into feature overlapping sample set D o v e r l a p and feature non overlapping sample set D n o n o v e r l a p based on the feature overlap degree
14:
Initialize flow model θ f l o w and feature overlapping sample classifier θ o v e r l a p c l s
15:
θ f l o w θ f l o w ( θ f l o w ( D o v e r l a p ) ) , θ o v e r l a p c l s = R F ( D o v e r l a p )
16:
θ n o n o v e r l a p c l s = R F ( D n o n o v e r l a p )
17:
return  D o v e r l a p , D n o n o v e r l a p , θ f l o w , θ o v e r l a p c l s , θ n o n o v e r l a p c l s
Precise Density Transformation: Through the precise calculation of the Jacobian determinant (Equation (8)), the probability density of the transformed space is p Z ( z ) = p X ( x ) d e t ( f / x ) 1 . This allows the model to accurately model sample distribution in the low-dimensional space, rather than relying on approximate inference like VAE. Precise probability density modeling is crucial for the classification of overlapping samples. In overlapping regions, probability distributions of different classes intertwine, and classification decisions inherently rely on the accurate estimation of the posterior probability p y x . VAE uses the Evidence Lower Bound (ELBO) for approximate inference; this approximation introduces large errors in highly overlapping regions, leading to blurred class boundaries. In contrast, the flow model, through precise density transformation, can accurately calculate the class-conditional probability density of each sample point in the transformed space, enabling the Bayesian classifier to make more precise decisions.
(2) Mechanism for Overlapping Sample Separation:
How the flow model achieves separation of overlapping samples was explained in detail:
Nonlinear Feature Decoupling: The affine coupling layer (Equations (6) and (7)) decouples the entangled feature dimensions in the original space by learning the nonlinear transformation NN ( · ) . Specifically, for samples in overlapping regions where interclass features are highly correlated, the flow model can learn a transformation such that the interclass differences of the transformed features in the new space are amplified, while intraclass similarities are maintained.
Adaptive Dimension Compression: Unlike linear dimensionality reduction with fixed projection directions, the flow model’s invertible transformation layer (Equation (4)) adaptively finds the optimal feature reorganization direction by learning the matrix H. This adaptivity is particularly important for overlapping samples because overlap often occurs in specific feature subspaces, and the flow model can map these problematic dimensions to a more separable space.
Hierarchical Feature Learning: The multi-layer flow model (Figure 3) achieves feature separation from coarse to fine granularity through layer-by-layer transformation. The first layer may separate major class features, while subsequent layers focus on subtle differences in boundary regions. This hierarchical mechanism is particularly suitable for handling complex situations with multi-modal overlap.
(3) Theoretical Demonstration of Improved Interclass Separability:
According to Bayes’ error rate theory, the lower bound of classification error depends on the degree of overlap of class-conditional probability densities:
P error min p x C 0 , p x C 1 d x .
Through the reversible transformation f, the flow model, while keeping the total probability mass unchanged, is able to learn a transformation such that:
min p z C 0 , p z C 1 d z < min p x C 0 , p x C 1 d x .
This reduces the overlap of class-conditional distributions in the transformed space, thereby lowering the lower bound of the theoretical error rate.
Through the above multilevel and multi-angle supplements, we believe we have fully clarified the theoretical advantages and practical value of flow model mapping.

3.3. Attention Weight Calculation Method Based on Multi-Neighbor Distribution Discrimination

During the analysis of instance overlap, the degree of overlap for a given sample is calculated by examining the number of heterogeneous samples within its neighborhood. However, not all samples warrant additional weighting or attention, as the importance of samples varies depending on their distribution patterns. Therefore, to comprehensively and accurately determine the overlap scores for different samples, a novel attention weight calculation method based on multi-neighborhood distribution discrimination is proposed. First, samples are categorized into the following three types based on their distribution characteristics, which is illustrated in Figure 4.
1. Low-overlap samples ( N d i f f 1 ): Predominantly homogeneous neighborhoods indicate easy-to-learn patterns. Lower weights (Equation (13)) prevent redundant learning.
2. High-overlap samples ( S N d i f f 1 ): Predominantly heterogeneous neighborhoods represent rare patterns. Moderate weights (Equation (14)) ensure adequate learning without overemphasis.
3. Boundary samples (otherwise): Mixed neighborhoods indicate critical boundary regions. Highest weights (Equation (15)) ensure comprehensive learning of discriminative features.
For each sample x with S nearest neighbors, the overlap weight φ is calculated based on heterogeneous neighbor count N d i f f :
When N d i f f 1 : In this case, only a small portion of the samples in the examined neighborhood range belong to different classes. Based on the earlier analysis, this distribution pattern is relatively common in the original data space, and the model can easily learn from such samples. For these samples, the model should pay relatively less attention during training to avoid overfitting, which could diminish its ability to distinguish challenging samples in overlapping regions. The corresponding overlap weight φ is calculated as follows:
φ = F o c S N d i f f + 1 I g ,
where F o c represents the base attention level, a predefined hyperparameter that denotes the model’s baseline attention value for each individual sample. I g represents the ignore coefficient, which is set for samples with easily learnable distribution patterns and indicates the degree to which the model ignores such samples. For these samples, because their features are easy to learn, a certain reduction in attention is applied on top of the base attention level to prevent the model from overemphasizing easily learnable features at the expense of critical features in overlapping regions. This results in the final overlap weight.
When S N d i f f < = 1 : In this case, only a small portion of the samples in the examined neighborhood range belong to the same class. This distribution pattern occurs infrequently and may contain potential sample distribution patterns. Therefore, an appropriate increase in attention is considered to ensure that the model can recognize such distribution patterns. The overlap weight is calculated as follows:
φ = F o c 1 N d i f f + 1 .
For samples that do not fall into the above two categories, it can be assumed that they exhibit varying degrees of overlap. These samples contain relatively important overlapping feature information for classification tasks. Due to the mixed nature of their features and the spatially close distribution of samples from different classes, it is challenging for the model to learn the characteristics of such samples fully. Therefore, a higher level of attention is required for these samples to ensure that the model adequately learns the information they contain. The overlap weight is calculated as follows:
φ = F o c N d i f f + 1 .
The calculated overlap weight is incorporated as an additional attribute of the corresponding target sample and is jointly input into the model for learning.
Not all samples contribute equally to the decision boundary. This adaptive weighting is important to prevent the model from being overwhelmed by the sheer volume of easy-to-classify majority samples (redundancy), ensuring it focuses its capacity on the information-rich but scarce samples located in the critical overlap regions.
The overlap weights calculation process is shown in Algorithm 2.
Algorithm 2 The overlap weights calculation process.
Require: Original dataset X, the size of the nearest neighbor pool S, The overall expansion factor of the dataset F o c .
Ensure: Target-neighbor sample combination dataset A;
      for each x in X do
  2:
     N e i = K N N ( x , X , S )
  /* Among them, N e i represents all samples except x, and  K N N ( x , X , S ) represents the S neighbor samples of x except itself in X. */
     N d i f f = J u d g e ( N e i )
        /* Among them, N d i f f the number of different class neighbors of sample x, J u d g e ( N e i ) represents check every neighbor in the neighbor space N e i and count the different class samples. */
  4:
     i f : S N d i f f < = 1
     φ = F o c 1 N d i f f + 1
  6:
     e l s e i f : N d i f f < = 1
     φ = F o c S N d i f f + 1 I g
  8:
     e l s e :
     φ = F o c N d i f f + 1
10:
     x = ( x , φ )
     A d d ( x , A )
12:
end for
return A
In summary, the method proposed in this paper has significant innovations in the following three aspects: (1) it proposes a dual overlap measurement mechanism fusing feature perspective and instance perspective for the first time, breaking through the information bottleneck of traditional single perspectives; (2) it introduces a multi-resolution analysis framework to achieve multilevel overlap information extraction from local neighborhoods to global distributions; (3) it designs a differentiated processing strategy for overlapping samples based on flow model mapping and importance weighting, enhancing the model’s discriminative ability for hard-to-classify regions.The overall flowchart of DPOA-MRM is shown in Figure 5.

3.4. Algorithm Complexity Analysis

This section derives the DPOA-MRM from two dimensions: time complexity and space complexity, and compares it with existing mainstream ensemble imbalanced classification methods (based on MDSampler, DPHS-MDS, AWLICSR, and BRAF). All symbol definitions are consistent with the main text: N is the number of training samples, d is the original feature dimension, K is the number of resolution layers (default K = 5 ), S m a x is the maximum neighborhood radius (default S m a x = 9 ), T f l o w is the number of iterations of the flow model, T b a s e is the number of iterations of the base learner, E is the number of base learners (ensemble size).

3.4.1. Time Complexity

(1) Multi-resolution Overlap Measurement Stage
Feature Perspective: Calculating the importance of F1 and RF requires traversing all samples and features, with a complexity of O ( N d ) ; for K resolutions, scanning feature subsets S k 1 , 3 , 5 , 7 , 9 time, the total complexity is T 1   =   O K · N · d · S m a x   =   O 5 N d · 9   =   O N d .
Instance Perspective: KNN search is performed at K resolutions. After acceleration with KD-Tree, the average single search is O l o g N , and the total for N samples is T 2   =   O K · N · l o g N   =   O N l o g N . Dual Perspective Fusion: Linear weighting, which O N can be ignored. In summary, the total time complexity of the overlap measurement stage is T o v e r l a p = T 1 + T 2 = O N d + N l o g N .
(2) Overlap Sample Processing Stage
Flow Model Mapping: The network forward/backward propagation only applies to the subset of samples judged as “feature overlap”. Let its ratio be α , and each round traverses α N samples, reducing the dimension to d = d / 2 , with a single round complexity of O ( α N d ) , and after T f l o w rounds, T f l o w m a p   =   O ( α N d T f l o w ) .
Importance Weighting: KNN searches again, with the same complexity as T 2 , and the index can be reused, so the additional overhead is O N .
The total time complexity of this stage is T p r o c e s s   =   O α N d T f l o w   +   N .
(3) Ensemble Training Stage
Train two RFs (overlapping/non-overlapping subsets) for each resolution. The RF training complexity is O N s u b · d · l o g N s u b · T b a s e , and the subset sizes are approximately α N and 1 α N , respectively. The ensemble of trees is T t r a i n = K · 2 · E · O ( α N d l o g ( α N ) + ( 1 α ) N d l o g ( ( 1 α ) N ) ) =   O ( K E N d l o g N ) .
In summary, the total time complexity of the DPOA-MRM training stage is T t r a in t o t a l = O ( N d + N l o g N + 0.15 N d T f l o w + K E N d l o g N ) = O ( K E N d l o g N ) (because K , E , T f l o w are of constant order of magnitude).
(4) Testing Stage
For a single test sample, perform a dual-view overlap calculation and tree prediction once at K resolution, with a complexity of T t e s t = O K d + l o g N + E l o g N = O d + E l o g N .

3.4.2. Space Complexity

Original Data Storage: O N d .
Multi-resolution KNN Index: KD-Tree has K trees in total, each tree has O N , totaling O K N .
Flow Model Parameters: The network weight size is proportional to d, approximately O d .
Ensemble Model: There are E trees, the upper bound of the number of nodes per tree is O N s u b l o g N s u b , totaling O K · 2 · E · N l o g N = O K E N l o g N .
In summary, the peak space complexity is S t o t a l = O ( K E N l o g N + N d ) .

3.4.3. Comparison of Complexity with Existing Ensemble Methods

In Table 1, k , R , w , α represents the hyperparameters mentioned in the corresponding methods. The BRAF method reduces the d dimensional features to k through projection; the SWSEL method uses a sliding window to resample the nearest w samples in each round, and after training the subforest, only the best-performing α E trees are retained. The training round R is triggered by concept drift detection. The average depth of a single tree is uniformly estimated using O l o g N .
It can be seen that DPOA-MRM has the same order of magnitude in complexity as most mainstream ensemble methods, with training time complexity at the O ( N l o g N ) level, testing time complexity at the O ( l o g N ) level, and peak space complexity at the O ( N l o g N ) level, but the difference in complexity between it and other methods is small. The multi-resolution strategy introduces a multiplier K that is acceptable at the constant level ( K = 5 ), and each resolution can be parallelized, further reducing the actual wall-clock time.

4. Experiments

This section conducts comprehensive performance evaluations of our ensemble classification framework for imbalanced data, which employs multi-resolution overlap quantification and dual-aspect boundary analysis. The experimental architecture comprises five components: Section 4.1 delineates the benchmark datasets and performance indicators. Section 4.2 specifies critical hyperparameter configurations, including multi-resolution granularities and neighborhood radii for overlap computation. Section 4.3 performs comparative analyses between our method and existing algorithm-level/data-level approaches in addressing class imbalance challenges. Section 4.4 conducts focused case studies on datasets exhibiting significant interclass overlap to demonstrate methodological superiority. Section 4.5 applies Nemenyi post hoc testing to validate statistical significance across comparative methods. Section 4.6 explores the sensitivity of DPOA-MRM to adjustable hyperparameters and their optimal value ranges through parameter sensitivity experiments. Ablation experiments are performed in Section 4.7 to verify the effectiveness and necessity of each module of DPOA-MRM.The complete specifications of all benchmarked techniques are systematically organized in Table 2. Furthermore, the source code for DPOA-MRM is publicly available at GitHub (https://github.com/bupt-lqw/DPOA-MRM) (accessed on 29 November 2025).

4.1. Datasets and Evaluation Metrics

To comprehensively evaluate the proposed method’s performance, we selected 39 benchmark datasets from the KEEL machine learning repository [58]. This selection aims to minimize result bias by ensuring diversity in dataset attributes and imbalance characteristics. During dataset selection, we ensured maximum variation in imbalance ratios. We also considered additional factors, including sample size, overlap degree, and feature dimensions. This approach guarantees a reasonable distribution of attributes across different ranges. To demonstrate the superiority of our algorithm in handling significantly overlapping datasets, we adopted the F1 metric (Maximum Fisher’s Discriminant Ratio) [59] as a data complexity indicator. This measure computes feature-wise discriminant ratios and selects the maximum value across all features to quantify dataset intricacy. Elevated F1 scores correspond to intensified feature-space overlaps, reflecting heightened interclass ambiguity. Datasets exceeding threshold 0.4 in F1 values were designated as having critical overlap severity. Key characteristics of these selected datasets are systematically presented in Table 3.
In the field of imbalanced classification, minority class samples are often of greater concern. However, while maintaining model comprehensiveness and robustness, the classifier’s capacity to process majority instances remains crucial. To holistically evaluate performance across both classes, we adopt F-measure and G-mean as dual evaluation metrics—the former emphasizing minority class precision–recall balance and the latter quantifying balanced recognition capability across classes. Following standard experimental protocols, minority and majority classes are designated as positive and negative categories, respectively, for evaluation consistency. The quantitative expressions for minority/majority class recall rates and minority precision are formally defined in Equation (16).
r e c a l l maj = T N T N + F P r e c a l l min = T P T P + F N p r e c i s i o n min = T P T P + F P .
The F-measure is a metric designed to evaluate the classification performance on the minority class. It takes into account both the precision and recall of the minority class samples in its calculation, and to some extent, it also reflects the model’s performance on the majority class samples. Its formula is shown in Equation (17). In the F-measure, the parameter β plays a crucial role. The parameter β intuitively determines the degree of involvement of the minority class recall ( r e c a l l min ) and precision ( a c c u r a c y min ) in the calculation. When β is increased, the weight of the minority class precision in the F-measure becomes larger; conversely, the weight of recall becomes larger. In the experiments of this paper, considering the equal importance of both in imbalanced classification problems, β is chosen to be 1, i.e., both are equally important. At this point, the F-measure is regarded as the F1-score. Meanwhile, to further reflect the model’s performance on the majority class samples, this paper additionally selects the G-mean metric to examine the recall of the majority class samples based on the F1-measure. The G-mean is calculated by taking the geometric mean of the minority class recall ( r e c a l l min ) and the majority class recall ( r e c a l l maj ). This method of calculation can comprehensively reflect the classifier’s performance across different classes. Its formula is shown in Equation (18).
F-measure = ( 1 + β 2 ) · r e c a l l min · p r e c i s i o n min β 2 · r e c a l l min + p r e c i s i o n min ,
G-mean = r e c a l l min · r e c a l l maj .
To statistically validate the proposed algorithm’s efficacy and generalization capacity, this study implements rigorous hypothesis testing through two non-parametric approaches: The Friedman test [60] for multi-method ranking analysis and the Wilcoxon signed-rank test [61] for pairwise comparative assessment. The Friedman procedure establishes performance hierarchies across all examined techniques, whereas the Wilcoxon method quantifies significance levels in direct method comparisons. Our experimental protocol establishes statistical significance at α = 0.05 confidence level, with p-values below this threshold indicating superior performance of the developed methodology over existing balancing techniques.

4.2. Parameter Settings

In this section, the parameter settings mentioned involve the spatial scope during overlap degree calculation under different resolutions, the base attention coefficient of the model for samples, and the neglect coefficient. The following will introduce the actual effects and setting methods of the above parameters, respectively.
This paper proposes a dual-perspective multi-resolution overlap degree analysis framework. The multi-resolution approach enables the model to comprehensively learn the distribution patterns and feature information of samples in the original dataset, allowing for more accurate determination of the overlap situation within the dataset and laying the foundation for subsequent targeted processing. Among them, the spatial scope S during overlap degree determination under different resolutions directly determines the actual effect of the multi-resolution measurement framework. The spatial scope S represents a parameter array containing the number of neighbors considered in overlap degree calculations under different resolution conditions. When the value of parameter S is relatively small, it means that in the process of overlap degree calculation, the model pays more attention to the feature correlation between the current sample and its nearby samples. At this time, the resolution is high, and the feature overlap perspective in the dual-perspective framework has a larger decision weight. Conversely, when the value of parameter S is relatively large, it means that in the process of overlap degree calculation, the model pays more attention to the distribution pattern of samples within the range where the current sample is located. At this time, the resolution is low, and the instance overlap perspective in the dual-perspective framework has a larger decision weight. Therefore, determining an appropriate value for S can effectively help the model focus on the key information at different resolutions, thereby accurately and comprehensively locating the overlapping samples in the dataset for subsequent targeted processing. The different decision weights corresponding to the dual perspectives under different resolutions can also help the model achieve more accurate and robust prediction results. Through experiments, the parameter S is set as an array, with the spatial scope parameters included being: 1, 3, 5, 7, 9.
The attention weight calculation based on multi-nearest-neighbor distribution discrimination in the data neighborhood is conducive to the model’s ability to compute the corresponding sample attention levels according to different distribution patterns, thereby selectively learning more important overlapping samples. The main parameters involved in this process include the base attention F o c and the neglect coefficient I g . The base attention F o c represents the model’s attention level to individual samples in the original dataset during the training process. The greater the attention level, the larger the gradient values computed by the model during training. However, excessively large gradient values may cause the model to fall into local optima, while excessively small gradient values may slow down the model’s convergence speed. The neglect coefficient I g is aimed at the neglect degree of easily learnable samples in non-overlapping regions. For samples in non-overlapping regions, most of their neighbors are of the same class, making the judgment of sample similarity relatively easy and resulting in a higher redundancy of information among samples. If the model pays too much attention to such samples, it may neglect the more difficult-to-classify samples in the overlapping regions. Therefore, setting the neglect coefficient Ig helps to constrain the expansion of such samples. Through experiments, the base attention F o c is set to 10, and the neglect coefficient I g is set to 0.5.

4.3. Comparison with Imbalanced Classification Methods

The proposed DPOA-MRM method employs a multi-resolution approach to capture key features across different spatial scopes. The dual-perspective overlap criteria comprehensively analyze complex overlap situations. This design specifically targets intermingled features in overlapping samples. To verify the effectiveness of the proposed method in solving imbalanced classification problems, typical data-level and algorithm-level methods from the current state-of-the-art in this field were selected for comparison.
The algorithm-level methods include iForest [7], SVDD [6], RF [29], GDBT [30], BRAF [49], DPHS-MDS [40], PCGDST-IE [36], SWSEL [50], HDAWCR [37], AWLICSR [38], and MDSampler [34] a total of 11 algorithmic methods. All methods adopted were strictly configured according to the parameters described in the original papers. Meanwhile, to minimize the additional impact of parameter tuning on the experimental results, the implementations of the four conventional classifiers were based on the default parameters provided by the Scikit-Learn library (version 0.24.2) in Python (version 3.7). The data-level methods include seven oversampling methods: SMOTE [9], Borderline-SMOTE [51], G-SMOTE [52], SMOTE-NaN-DE [17], TSSE-BIM [53], MPP-SMOTE [18], and HSCF [19] as well as seven sample generation methods: CWGAN-GP [22], ADA-INCVAE [24], RVGAN-TL [26], CTGAN [54], CDC-Glow [55], ConvGeN [56], and SSG [57]. Since data-level methods themselves do not possess classification capabilities, RF [29], SVM [28], and GBDT [30] classifiers were selected to learn from the balanced datasets and compare the classification performance. Among the six oversampling methods, except for MPP-SMOTE and TSSE-BIM, the remaining four methods were implemented directly through the Smote-variants imbalanced learning library. The remaining methods were strictly configured according to the parameters described in the original papers. Similarly, to minimize the additional impact of parameter tuning on the experimental results, the implementations of the RF, SVM, and GBDT classifiers were also based on the default parameters provided by the Scikit-Learn library in Python. Due to space limitations, only the results related to the RF classifier, which showed the best comprehensive performance, are presented here. The experimental results for the SVM and GBDT classifiers are provided in the Appendix A.
As shown in Table 4 and Table 5, experimental evaluations demonstrate the superior performance of our multi-resolution ensemble framework for imbalanced classification, which achieves top-2 rankings across 28 datasets in F1-measure and 21 datasets in G-mean among contemporary algorithm-level approaches. The methodology’s effectiveness stems from its innovative dual-aspect boundary analysis mechanism—through multi-scale quantification of class overlap patterns, it constructs complementary feature-space and instance-space classifiers at varying resolution levels. Each resolution-specific classifier specializes in identifying challenging samples within particular boundary ambiguity scopes through targeted training on corresponding overlap subsets. The ensemble architecture strategically combines these diverse perspectives by aggregating predictions across multiple spatial resolutions, enabling comprehensive consideration of test instances’ positional characteristics relative to dynamic decision boundaries. This multi-faceted analysis capacity ensures enhanced classification robustness and precision when handling complex interclass ambiguity scenarios.
The experimental findings reveal a notable performance divergence between F1-measure and G-mean metrics, with our method demonstrating superior F1 outcomes. This discrepancy originates from the algorithm’s overlap quantification mechanism that evaluates sample distribution patterns within defined spatial boundaries. Given the inherent class imbalance in benchmark datasets, decision boundary regions predominantly exhibit majority class dominance. Consequently, minority class instances in these ambiguous zones encounter heightened concentrations of dissimilar neighbors during overlap assessment, increasing their probability of being categorized as overlapping samples.
This operational characteristic elevates the minority class representation within identified overlap subsets compared to original data distributions, while such redistribution mitigates initial imbalance severity, it inadvertently biases subsequent feature learning processes toward minority class patterns in overlap regions. The classifiers thus develop enhanced sensitivity to minority class discriminative features, optimizing F1-measure performance through improved precision–recall balance for minority predictions. However, this focused adaptation slightly compromises majority class recognition efficacy, as reflected in relatively moderated G-mean scores. The empirical evidence corroborates that our multi-resolution overlap analysis effectively amplifies minority class discriminability at boundary regions, albeit with measurable trade-offs in holistic class-wise equilibrium.
As can be seen from Table 6, the statistical analysis further verified the significant advantages: compared with traditional classifiers (iForest, SVDD, RF, GBDT), the p-values for F1-score ranged from 6.64 × 10 8 to 2.25 × 10 6 , and for G-mean from 1.94 × 10 7 to 2.35 × 10 5 . Compared with ensemble methods (BRAF, DPHS-MDS, SWSEL), the p-values for F1-score ranged from 1.27 × 10 3 to 2.04 × 10 3 , and for G-mean from 9.33 × 10 5 to 4.46 × 10 3 . It is noteworthy that for cost-sensitive methods (HDAWCR, AWLICSR), the p-values for F1-score range from 3.62 × 10 5 to 7.33 × 10 4 , and the p-values for G-mean range from 2.85 × 10 4 to 5.81 × 10 5 , indicating that our proposed method has robust advantages across different algorithmic paradigms.
The proposed method does not show significant differences in G-mean compared to the two algorithmic methods, PCGDST-IE and MDSampler. PCGDST-IE’s clustering-based partitioning strategy effectively handles majority class samples, making its G-mean highly competitive, but its significantly lower F1-score indicates weaker minority class precision. MDSampler’s meta-distribution ensemble achieves a relatively balanced distribution between the two classes, but lacks a dedicated mechanism for overlapping regions, resulting in a significantly lower F1-score. These factors reveal the strategic trade-offs of our proposed method: prioritizing minority class discrimination while maintaining acceptable majority class recognition.
The proposed method shows a more significant improvement in F1-score (all p < 0.01), while the two methods show no significant difference in G-mean: PCGDST-IE and MDSampler. This difference stems from the following reason: In imbalanced data, overlapping regions usually contain a larger proportion of majority class samples. The overlap identification mechanism in Section 3.1 classifies these boundary samples as overlapping subsets, thus increasing the proportion of minority class samples in overlapping regions. The streaming model mapping in Section 3.2 and the attention weighting mechanism in Section 3.3 focus on strengthening the learning of minority class features in these regions, thereby improving precision and recall (components of F1), while the improvement on balanced precision (G-mean) is relatively limited.
As shown in Table 7 and Table 8, compared with data-level methods, the proposed algorithm ranks in the top two on 14 datasets in terms of the F1-measure and on 18 datasets in terms of the G-mean. Compared with algorithm-level methods, its advantage is relatively less pronounced. This is partly because the RF classifier, which showed the best performance for each comparison algorithm, was selected for the result presentation. Additionally, the RF classifier can handle difficult-to-classify samples in the overlapping region to some extent, which weakens the main advantage of the proposed algorithm. However, it is still evident that, even under these circumstances, the proposed algorithm achieves the highest mean ranking in the Friedman test, indicating a statistically significant advantage. Moreover, the results show a more pronounced advantage in G-mean compared to the F1-measure. This is because data-level methods often employ the approach of supplementing minority class samples to achieve data balance, which, to some extent, compensates for the missing sample information of the minority class. As a result, classifiers tend to focus more on learning minority class samples during subsequent training, leading to poor classification performance on the majority class.
As shown in Table 9, all sampling methods showed significant differences in both metrics (p < 0.05). The traditional SMOTE variant had p-values of 9.69 × 10 4 to 1.67 × 10 5 on the F1-score, while the latest methods, such as TSSE-BIM, had p = 8.52 × 10 6 . Generative methods performed more complexly, with 4 out of 7 methods significantly outperforming the proposed methods on the F1-score (p < 0.05). ADA-INCVAE (p = 0.199) and CDC-Glow (p = 0.194), due to their specialized overlap handling mechanisms, effectively improved classification performance and did not show significant differences compared to the proposed methods.

4.4. Comparison of Experimental Results When the Datasets Overlap Seriously

Accurately identifying overlapping samples is a key challenge in imbalanced classification. These samples contain significant common information, with intermingled class features that obscure classification boundaries. Traditional data-level methods lack the ability to uncover differences between overlapping samples. They often sacrifice some samples to achieve better results, limiting model performance potential. Moreover, existing methods typically rely solely on the class distribution within the neighborhood of samples to identify overlapping regions, ignoring the intermingling of features themselves.
In the proposed algorithm of this paper, dual-perspective overlap criteria are introduced to assess the overlap status of individual samples from both feature and distribution perspectives, maximizing the utilization of the feature information contained in the samples. On this basis, the multi-resolution overlap measurement approach helps the model achieve complementary overlap information under different discrimination ranges, comprehensively reflecting the complex and diverse overlap situations in the dataset. Furthermore, the algorithm targets different types of overlapping samples, enhancing the accuracy and robustness of the model. Therefore, to better highlight the advantages of the proposed method in this paper, the F1-measure was selected as the evaluation metric for the complexity of the datasets, and 15 datasets with severe overlap were chosen for comparison. The specific experimental results are shown in Table 10 and Table 11.
On 15 severely overlapping datasets, our method ranks top-2 on 13 datasets (F1-measure) and 9 datasets (G-mean). The method shows relatively excellent performance under both metrics. It is worth noting that for overlapping regions, the primary issue to address is how to avoid misclassifying the already limited number of samples as majority class samples due to feature overlap. Therefore, in the experimental analysis of this section, the results of the F1-measure are more valuable for reference, which further illustrates the advantages of the proposed method in dealing with overlapping regions.
Severe overlap often involves multiple overlapping regions at different scales, making it difficult for single-resolution methods to simultaneously capture fine-grained boundaries and global patterns. Our five-scale analysis (S = 1, 3, 5, 7, 9) system covers the local to global levels, achieving a greater improvement over single-scale methods on highly overlapping datasets. The proposed method leverages feature similarity and spatial adjacency to provide complementary information in severely overlapping regions, and uses a flow model mapping to reduce feature complexity without information loss. Targeted processing avoids information loss, which is particularly important for preserving subtle but crucial discriminative information.

4.5. Nemenyi Post Hoc Test

To further verify the significant differences between the proposed algorithm in this paper and other comparison methods, the Nemenyi post hoc test [62] was employed. The test results are shown in Figure 6 and Figure 7.
The experimental results demonstrate that, compared with other algorithmic methods, the proposed ensemble imbalanced classification method based on multi-resolution overlap measurement and dual-perspective overlap criteria significantly outperforms all other methods in terms of the F1-measure and also shows a distinct advantage under the G-mean metric. Compared with the data-level methods, the proposed method significantly outperforms all comparison methods in terms of both the F1-measure and the G-mean under the RF classifier, which showed the best comprehensive performance among all algorithms. This further demonstrates the effectiveness of the proposed algorithm in dealing with imbalanced classification problems.

4.6. Parameter Sensitivity Experiments

The proposed DPOA-MRM mainly involves a key hyperparameter: the parameter m used to control the magnitude of overlapping sample weights. The selection of hyperparameters is crucial to model performance. To determine whether the proposed method is sensitive to hyperparameters and the optimal range of hyperparameter values, we selected the control variable method for parameter sensitivity experiments. For a selected hyperparameter, we varied it uniformly over as large a range as possible while keeping other hyperparameters at their optimal values. We determined the sensitivity of the proposed method to this hyperparameter and its optimal value range through the mean of the average performance on intraclass imbalance datasets under different hyperparameter conditions. The average performance under different hyperparameter m conditions is shown in Table 12.
From the results in Table 12, it can be seen that changes in the hyperparameter m lead to significant fluctuations in the performance of the proposed method; thus, the proposed method is observed to be very sensitive to changes in hyperparameter m. As the value of m continues to increase, a downward trend is observed in the average performance of DPOA-MRM on both F1-measure and G-mean metrics. According to Equation (1), m is used to control the magnitude of the overlapping sample weight. When m is large, feature overlap receives a larger weight; when m is small, instance overlap receives a larger weight. In the dual-perspective overlap degree analysis method used in the proposed approach, the weights occupied by feature overlap and instance overlap can be effectively balanced by an appropriate value of m. This allows the more effective instance overlap perspective to dominate while the feature overlap perspective serves as a supplement, effectively coping with the multi-modal overlap phenomenon in the dataset and laying a solid foundation for subsequent targeted processing. Based on the above synthesis, the optimal value for the hyperparameter m is determined to be 10.

4.7. Ablation Study

To further quantify the impact of each module on model performance, the model’s performance was tested by removing each module individually. The comparison of the average results of the ablation experiments on F-measure and G-mean metrics is shown in Table 13.
From the above ablation experiment results, it can be seen that the instance branch module has the greatest impact on model performance, followed by the multi-resolution module, while the feature branch and flow model have a relatively smaller impact. Specifically, the instance branch directly identifies hard-to-classify samples located at the decision boundary by analyzing the class distribution of the sample neighborhood via k-Nearest Neighbors. Removing the instance branch causes the model to lose its perception of the spatial position of samples, leading to a significant decline in classification performance. The multi-resolution strategy achieves hierarchical information extraction from local to global levels by independently calculating overlap degrees under different neighborhood ranges. After removing the multi-resolution module, the model fails to synergistically identify samples with small-scale feature entanglement and grasp large-scale class distribution trends, significantly damaging generalization ability. The feature branch identifies overlapping samples at the feature space level by analyzing the class separability of feature dimensions using the Maximum Fisher’s discriminant ratio. The synergy between the feature branch and the instance branch ensures the completeness of the framework. Although the independent contribution of the flow model module is small, it enhances the separability of samples with severe feature entanglement identified by the feature branch, possessing irreplaceability in certain high-complexity scenarios.

5. Conclusions

This paper proposes an ensemble imbalanced classification framework using dual-perspective overlap analysis with multi-resolution metrics. The method divides datasets into multiple resolutions and calculates sample overlap degrees under dual-perspective criteria. This maximizes consideration of multilevel distribution information from local to global perspectives. The differentiated overlap information obtained from the dual-perspective criteria effectively complements the features, ensuring that the model acquires more refined and accurate sample overlap information. In view of the characteristics and differences between the two criteria, the method employs flow model mapping and importance weight calculation for targeted processing, enhancing the model’s ability to distinguish difficult samples in the overlapping regions. Experiments on 39 datasets show significant advantages over data-level and algorithm-level methods on F1-measure and G-mean. In credit card fraud detection, loan default prediction, medical diagnosis, and equipment failure forecasting, minority-class samples are highly scarce and often overlap with majority-class features. The proposed method effectively captures shared and discriminative patterns within overlapping regions, enhancing minority-class recognition. It can be integrated as a preprocessing or joint-learning module in real-time risk-control systems, medical image or record-analysis workflows, and industrial IoT platforms. Since these tasks all suffer from class imbalance and class overlap, the method offers more robust and interpretable decision support across such application domains.
In summary, the multi-resolution overlap measurement framework can effectively extract sample overlap information from local to global perspectives, and the dual-perspective overlap criteria can effectively analyze the complex and diverse overlap patterns contained in the dataset. Future work may focus on developing adaptive resolution acquisition schemes in the resolution division stage and more targeted overlap sample processing methods to further enhance the generalizability and classification accuracy of the proposed framework.

Author Contributions

Conceptualization, X.D., Q.L. and X.G.; methodology, Y.L. (Yuan Li) and X.G.; software, Y.L. (Yuan Li), Z.M., T.C. and Y.L. (Yukun Lin); validation, Z.M., T.C. and Y.L. (Yukun Lin); formal analysis, Y.L. (Yuan Li) and X.G.; resources, X.D., Q.L. and X.G.; writing—original draft preparation, Y.L. (Yuan Li) and X.G.; writing—review and editing, Y.L. (Yuan Li), Y.H. and X.G.; supervision, X.G.; funding acquisition, X.D. and X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science & Technology Project of State Grid Corporation of China (5216A025000M-057-ZN).

Data Availability Statement

The benchmark datasets from the KEEL machine learning repository used in this study are publicly available, and is accessible through the open repository at https://sci2s.ugr.es/keel/imbalanced.php (accessed on 29 November 2025).

Acknowledgments

The authors would like to thank their colleagues from the Machine learning group for discussions on this paper.

Conflicts of Interest

Authors Yuan Li, Xinping Diao, Tianyang Chen and Yukun Lin were employed by the company China Electric Power Research Institute Company Limited. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. The Supplementary Experimental Results

Appendix A contains additional experimental outcomes comparing data-level methods with DPOA-MRM. Table A1 and Table A2 present the F1-measure and G-mean scores, respectively, of DPOA-MRM in comparison with standard data-level methods on SVM. Table A3 and Table A4 display the F1-measure and G-mean scores, respectively, of DPOA-MRM when contrasted with standard data-level methods on GBDT.
Table A1. F1-measure results for data-level methods on SVM and DPOA-MRM.
Table A1. F1-measure results for data-level methods on SVM and DPOA-MRM.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.98710.98620.98620.98620.98060.98060.97090.98710.93670.68330.89660.48320.96770.97980.9802
wisconsin0.95750.95170.95340.95710.95320.94220.95840.95920.93860.98620.91360.00000.96970.96090.9617
pima0.66630.68790.66220.66390.67240.60680.67210.64780.63570.33830.00000.49240.68250.64600.6786
vehicle_20.95490.94460.94820.94790.93090.94420.84600.93140.90740.65850.92110.65550.97780.94300.9752
vehicle_10.67350.67730.68850.63650.66480.59340.57480.47780.55580.66150.41510.74050.63330.42090.6532
vehicle_30.63900.64130.64750.61190.64850.56980.55230.38830.56850.36670.00000.49310.58620.33070.6099
vehicle_00.93390.90290.91750.92000.89840.89190.76620.90620.90550.49430.93670.79370.95240.92580.9520
ecoli_10.76150.76440.77090.78690.76970.76510.76110.79460.66480.77780.11110.69170.72220.77600.8321
new-thyroid10.97330.95810.95810.95810.97330.95810.97140.96000.87000.71770.94120.93031.00000.91790.9647
new-thyroid20.97330.97500.97500.97500.97330.89100.94070.96000.83380.73570.85710.80691.00000.91471.0000
ecoli20.86780.86950.87780.87860.84550.78670.74280.83900.80900.97240.12500.51840.90910.82120.8816
segment00.99080.96740.98930.98920.94420.98040.97470.97180.98620.85460.70710.94260.98460.97660.9909
yeast30.73140.75220.74300.76510.71530.76330.72860.70630.73720.03330.00000.16650.72220.76330.7948
ecoli30.67380.66730.65150.67340.65220.68490.52510.64900.62360.70120.00000.73510.57140.63330.7029
page-blocks00.73870.72840.73570.74410.72510.83890.67950.74710.69480.76110.48480.75890.62300.65140.8533
yeast-2_vs_40.76220.77810.74700.75890.70530.72960.73220.69910.73460.44460.00000.68710.76920.75750.8225
ecoli-0-6-7_vs_3-50.77950.69760.82080.79050.67880.69290.66560.74220.77220.80000.33330.35080.72730.80060.7802
yeast-0-2-5-6_vs_3-7-8-90.58300.43700.61990.60070.57970.55930.62860.55650.36270.92910.00000.95530.61540.55250.6735
yeast-0-2-5-7-9_vs_3-6-80.80100.64130.82290.80000.76930.73960.77930.72610.66250.46260.00000.72060.85710.83070.8231
glass-0-4_vs_50.63330.70000.80000.80000.66670.96000.68670.82860.58500.75240.00000.73330.66670.13331.0000
ecoli-0-1-4-7_vs_5-60.78050.79330.83780.83520.62410.75430.75820.71100.52390.05000.00000.48600.75000.82960.8039
glass40.80760.79330.75330.75330.72570.84000.49090.76710.65000.03640.50000.16620.80000.21330.8305
ecoli40.77110.81110.76530.78750.82320.68590.73880.88330.75610.84830.66670.77760.75000.84760.8292
page-blocks-1-3_vs_40.90710.91190.87990.88250.90510.94830.51360.62050.55780.72650.75000.84600.90910.60320.9846
abalone9-180.31520.31640.31830.36800.38140.28810.21650.28240.24350.08030.00000.34750.35290.00000.4340
MEU-Mobile KSD0.81680.85200.83730.84890.82790.86460.92400.89240.69230.65040.90000.85140.84210.90970.9204
yeast-2_vs_80.34130.55240.34840.38170.63810.40190.66810.62760.60000.00000.00000.19670.40000.68570.7405
flare-F0.26760.32120.25450.24860.25220.20730.18350.27700.13490.67600.00000.62510.27270.18120.2503
kr-vs-k-zero-one_vs_draw0.24410.00000.21140.15200.30270.37210.46600.79380.67370.29030.30000.72330.27850.86510.0000
kr-vs-k-one_vs_fifteen0.57250.81470.71830.25510.58650.81480.86700.99350.66890.03641.00000.22950.59571.00000.8167
winequality-red-40.16440.14600.18950.18280.15610.18190.17240.20300.09590.31330.00000.43810.12770.00000.1783
yeast-1-2-8-9_vs_70.08690.30340.12980.13990.12630.22480.18360.19440.06060.94520.00000.96650.23530.05000.3248
abalone-3_vs_110.94290.97140.97140.97140.85480.97140.42610.97140.92140.08231.00000.32960.80001.00000.9714
kr-vs-k-three_vs_eleven0.15400.00000.15430.15430.15400.15960.81590.91110.78660.29170.97440.65000.14980.99430.0000
ecoli-0-1-3-7_vs_2-60.46670.73330.41430.42760.34670.71330.63330.71330.60000.97140.00001.00000.66670.73330.5800
abalone-17_vs_7-8-9-100.20420.28820.23910.23650.21850.28750.12120.16250.15650.89420.00000.84140.21740.16420.3235
yeast60.40800.45520.41230.39900.32250.41960.28760.36970.42180.96170.00000.95060.60870.50630.6467
poker-8-9vs60.84210.73650.75870.84210.65110.27920.92780.19310.34590.24641.00000.34960.80000.58100.7336
poker-8vs60.89330.70290.67330.84290.74810.38410.80290.09950.21110.80771.00000.93140.66670.31330.7867
Average0.66840.67260.67140.66550.65110.65840.65010.67040.61240.56510.40340.62470.67080.64650.7201
Average Friedman-rank7.08977.01286.73086.56418.53857.80779.05137.628210.76928.807711.85908.76927.10268.03854.2308
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table A2. G-mean results for data-level methods on SVM and DPOA-MRM.
Table A2. G-mean results for data-level methods on SVM and DPOA-MRM.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.98730.98640.98640.98640.98370.98350.97210.98730.94180.69160.90140.53230.96820.98010.9833
wisconsin0.97240.96770.96790.97020.96830.95870.97060.97170.95900.99010.93500.00000.98300.97270.9774
pima0.74060.75720.73700.73810.74520.69110.74610.72550.71130.68940.00000.90390.75190.71560.7457
vehicle_20.97470.97250.96950.96800.96830.96480.92770.94590.95360.72880.92390.73070.99200.95000.9867
vehicle_10.80940.81470.82160.76930.80060.72380.72070.59630.68130.79900.52040.92530.77660.53420.7802
vehicle_30.78640.78890.79400.75730.79480.70120.70760.51600.70350.39930.00000.75650.74020.46040.7540
vehicle_00.97300.96040.96250.96660.96130.93220.90060.95070.95170.59060.97170.95120.98450.95560.9792
ecoli_10.87350.87380.88390.87570.88390.85640.88600.89070.76340.86930.24250.82360.83850.83730.9168
new-thyroid10.99440.97980.97980.97980.99440.97980.98260.99160.94630.93090.98560.95721.00000.92320.9915
new-thyroid20.99440.99440.99440.99440.99440.94000.95290.99160.89980.93150.86600.94921.00000.92151.0000
ecoli20.94100.91690.94250.94250.93580.86170.90770.93390.91100.97310.25820.00000.94510.88160.9370
segment00.99210.98800.99310.99180.99010.99040.98930.98880.98880.92040.73950.96470.98470.98960.9934
yeast30.89500.89900.90000.90380.90680.88160.88640.89860.83640.06030.00000.48300.86560.84100.9221
ecoli30.89580.88350.87220.88130.90090.80480.76550.87890.81870.79990.00000.82380.86300.77310.8667
page-blocks00.92500.92150.91990.92070.91940.94820.87510.87390.80970.91540.57590.95120.84790.79070.9401
yeast-2_vs_40.89740.87330.89360.89550.87030.87330.84930.86940.86590.56250.00000.79210.92750.82430.9169
ecoli-0-6-7_vs_3-50.86340.80000.88100.87590.86850.76520.85460.89620.86170.80000.44720.73880.87180.82530.8602
yeast-0-2-5-6_vs_3-7-8-90.76470.74240.78750.78910.79900.75050.78420.78930.53250.95830.00000.97370.85400.65550.7949
yeast-0-2-5-7-9_vs_3-6-80.88290.84850.90010.89700.90190.85080.89920.91240.79980.56920.00000.82280.93810.88090.8939
glass-0-4_vs_50.67840.76150.82430.82430.85560.99360.95010.96180.68860.86600.00000.87060.70710.14141.0000
ecoli-0-1-4-7_vs_5-60.87970.87740.90660.90680.85980.83970.85240.94580.67460.11030.00000.76630.77460.88700.9034
glass40.94800.90530.90280.90280.94040.92160.89050.89240.71040.06650.57740.56880.81650.25550.9796
ecoli40.88030.88350.90410.90570.93730.81170.92860.96710.88890.93590.70710.90220.85920.86100.9374
page-blocks-1-3_vs_40.97710.96180.95600.95590.97700.98030.81390.77810.65390.92910.77460.94730.91290.66110.9989
abalone9-180.76370.53960.72460.77340.73520.57920.41800.44980.51340.19920.00000.69240.75820.00000.6506
MEU-Mobile KSD0.84650.87020.85780.86860.90050.93350.92700.91460.88990.71360.90450.90390.89220.91480.9467
yeast-2_vs_80.60500.65090.63600.70280.75670.69750.72830.75650.72030.00000.00000.52880.50000.73810.8008
flare-F0.75590.70250.69580.75220.72520.62020.77360.81550.27270.80800.00000.87380.75650.30740.3976
kr-vs-k-zero-one_vs_draw0.77270.00000.79170.73440.79310.78880.95590.98460.79900.41730.42010.89100.69270.95160.0000
kr-vs-k-one_vs_fifteen0.86010.83010.82560.83580.85100.83030.99420.99980.91260.06651.00000.64570.91691.00000.8311
winequality-red-40.60560.37570.59720.63600.64680.64020.67430.47390.30210.35690.00000.48830.49370.00000.4730
yeast-1-2-8-9_vs_70.43380.62400.61160.55040.56520.70980.67230.70000.21860.95240.00000.98200.76140.08140.4755
abalone-3_vs_110.99790.99900.99900.99900.99380.99900.85220.99900.99690.21201.00000.69710.81651.00000.9990
kr-vs-k-three_vs_eleven0.82930.00000.82930.82930.82930.78010.97470.99680.90490.35360.99910.75810.81360.99980.0000
ecoli-0-1-3-7_vs_2-60.73040.74140.72480.72670.71980.87710.73780.93140.79260.99900.00001.00000.70710.74140.7851
abalone-17_vs_7-8-9-100.80540.71240.85510.85410.78480.72020.25730.33470.41050.91570.00000.88280.83990.30880.5917
yeast60.82630.79690.84090.83990.83830.79440.88600.87620.68320.97750.00000.96500.98440.67260.8178
poker-8-9vs60.86030.77370.79480.86030.83250.74520.93380.71910.51860.59971.00000.88040.81650.65200.8098
poker-8vs60.90470.70970.72020.86190.87630.86930.82520.74840.37620.91011.00000.97420.70710.35690.8094
Average0.84930.78680.85090.85700.86170.83560.83650.84240.74010.65560.42950.77690.83740.69860.8063
Average Friedman-rank6.14108.25646.58976.32055.73088.42317.66676.435910.948710.012812.11548.33337.038510.34625.6410
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table A3. F1-measure results for data-level methods on GBDT and DPOA-MRM.
Table A3. F1-measure results for data-level methods on GBDT and DPOA-MRM.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.98100.97410.96760.98620.97460.97380.97980.97500.95830.65400.93330.72740.93750.95950.9802
wisconsin0.95470.95050.94790.93390.95460.94690.93250.95550.94380.98770.92680.00000.95920.95170.9617
pima0.68090.66440.64910.62290.67540.61860.66310.64040.64190.58000.60000.47930.69030.61700.6786
vehicle_20.96560.95890.95900.96530.96420.95090.88090.94640.95960.64990.95120.65570.95560.96220.9752
vehicle_10.62310.63410.64780.58790.62670.58460.62010.60530.55830.79790.51020.85720.60240.52790.6532
vehicle_30.64090.60810.62970.54780.62170.53540.61700.54550.58510.69210.52380.48690.60000.49400.6099
vehicle_00.92050.92940.92710.90570.92640.90430.74860.91750.91080.80610.87800.75160.85000.91310.9520
ecoli_10.78790.76420.79540.80990.78750.74680.77090.78790.74500.78010.73680.62320.66670.78400.8321
new-thyroid10.95590.88240.88240.92000.90810.91790.91190.94830.86670.81730.94120.92561.00000.91960.9647
new-thyroid20.90140.92630.95960.91380.93140.90080.88980.90900.91290.81730.93330.85420.92310.95131.0000
ecoli20.82320.78690.76520.75710.79490.81990.69150.86110.78670.96130.76920.00000.76190.82460.8816
segment00.98790.98020.98790.98320.98360.98040.95870.98800.98300.93100.97670.84700.98480.98630.9909
yeast30.77040.76910.77110.77620.78920.73100.77550.78630.75420.05580.66670.19010.66670.76140.7948
ecoli30.62400.60730.56030.63210.62370.60320.55600.52950.59470.74300.54550.72480.75000.58620.7029
page-blocks00.85200.84270.85110.85060.82420.84150.66810.86210.79110.79190.85020.69430.77780.87440.8533
yeast-2_vs_40.77290.79780.76190.77290.74530.72530.64840.78250.76360.54270.61540.68760.75000.72480.8225
ecoli-0-6-7_vs_3-50.67950.72880.74570.64760.71050.70670.63740.78810.75000.51110.66670.54670.83330.73920.7802
yeast-0-2-5-6_vs_3-7-8-90.57080.50300.57000.57290.57450.55910.48920.62560.48830.92800.64860.95000.72220.58580.6735
yeast-0-2-5-7-9_vs_3-6-80.79030.79990.81310.82420.75160.77320.72100.80360.71280.56730.72730.70320.80950.81570.8231
glass-0-4_vs_50.93330.96000.96000.96000.85330.96000.85330.89330.75080.72020.66670.72541.00000.92001.0000
ecoli-0-1-4-7_vs_5-60.82440.76280.80990.83780.73380.72300.62580.76110.73350.36140.66670.50490.66670.75440.8039
glass40.83140.86480.86290.89140.62140.76000.54830.75420.70620.15310.66670.19820.80000.71330.8305
ecoli40.81590.72860.78000.72440.83490.73570.61950.88830.75800.75330.75000.75200.54550.65140.8292
page-blocks-1-3_vs_40.96360.95130.96920.96920.92050.91360.56390.87910.84260.74120.90910.73441.00000.93310.9846
abalone9-180.28280.34800.36520.35200.31290.31720.19350.38680.28590.19180.50000.37360.16670.34590.4340
MEU-Mobile KSD0.90970.89620.89920.89920.86100.86370.80440.89600.63900.57190.84210.72050.88890.88360.9204
yeast-2_vs_80.45230.46100.53110.45270.52290.40950.53020.63950.64440.13870.33330.20650.40000.57900.7405
flare-F0.24120.27790.18680.21020.21750.24250.15840.25260.15310.62520.00000.61570.00000.19050.2503
kr-vs-k-zero-one_vs_draw0.24410.00000.22270.15200.30270.37280.16920.93950.78700.66190.97140.67000.00000.96560.0000
kr-vs-k-one_vs_fifteen0.58080.81470.71830.25510.72250.81750.29111.00000.84170.18951.00000.25610.89661.00000.8167
winequality-red-40.14210.15250.13190.16730.15940.16400.16630.08820.09630.38000.22220.45000.20000.06410.1783
yeast-1-2-8-9_vs_70.15780.31610.23980.17010.22300.17990.27220.25330.32070.96310.25000.96230.60000.33330.3248
abalone-3_vs_111.00001.00001.00001.00000.94291.00001.00001.00000.97140.18661.00000.31221.00001.00000.9714
kr-vs-k-three_vs_eleven0.15400.00000.15430.15430.15400.15680.31021.00000.76120.82860.97440.71710.00001.00000.0000
ecoli-0-1-3-7_vs_2-60.37330.54670.46670.63330.37330.39050.47330.54670.57780.97140.66670.96000.40000.34670.5800
abalone-17_vs_7-8-9-100.34710.35210.36980.33250.30770.25420.20850.31230.17300.64210.37500.67730.35290.31350.3235
yeast60.42290.48890.52470.46290.38990.36220.36840.56510.49640.94160.30770.94010.46150.47990.6467
poker-8-9vs60.88740.88330.50870.92780.95000.16110.85560.42540.70870.51050.33330.68180.50000.44210.7336
poker-8vs60.97140.66480.48330.86480.94290.25260.76290.34000.58000.93790.40000.87781.00000.35140.7867
Average0.68770.68150.67630.67760.67990.63740.61370.72000.68040.64320.67270.61640.66970.69860.7201
Average Friedman-rank6.52567.33336.82057.03857.78219.820510.66676.44879.38469.17959.62829.74367.79497.84623.9872
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table A4. G-mean results for data-level methods on GBDT and DPOA-MRM.
Table A4. G-mean results for data-level methods on GBDT and DPOA-MRM.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.98390.97950.97610.98640.98030.97980.98310.98060.96940.79380.93540.90090.95140.96630.9833
wisconsin0.96740.96400.96000.95280.96740.96260.94780.96940.95770.99160.94760.00000.97270.96400.9774
pima0.75360.73850.72630.70370.74900.70090.73440.71960.71930.77390.67330.85310.76010.69770.7457
vehicle_20.97580.97350.97500.97590.98120.97490.94700.97040.97960.72510.96770.73150.97670.97180.9867
vehicle_10.74950.75470.76750.71490.75480.71730.76020.71160.68420.86880.68290.94330.71070.65000.7802
vehicle_30.76430.74230.75730.68860.75340.67800.76050.66380.71040.70540.61590.68180.70780.62110.7540
vehicle_00.95670.96350.95600.94510.96710.94320.88240.94800.95190.92620.94340.90950.90040.94090.9792
ecoli_10.87470.86280.87910.89220.88020.84480.89520.85690.81660.88870.84290.78980.79660.86180.9168
new-thyroid10.96750.92350.92350.94630.95710.95950.94640.98880.94620.95020.98560.93681.00000.95950.9915
new-thyroid20.91870.94860.97950.94660.97420.94380.94050.97720.94580.95020.93540.95230.92580.95421.0000
ecoli20.89940.85870.85450.85190.87650.90000.86200.92450.87630.97000.80880.00000.83770.86450.9370
segment00.99420.98780.99290.98960.99590.98780.98650.99540.98440.96210.98970.88670.99110.99260.9934
yeast30.89400.89420.89980.89220.89970.86470.88950.87680.84520.12010.77920.56660.75160.85260.9221
ecoli30.79730.76570.75090.83170.82350.75410.78180.68000.71880.81450.61240.80870.90280.71200.8667
page-blocks00.95740.94670.95030.95170.95010.92510.92830.91780.87480.95570.92480.92570.91370.92100.9401
yeast-2_vs_40.89080.88630.86150.91020.88660.86450.91040.88210.85850.66390.70340.78540.88490.81160.9169
ecoli-0-6-7_vs_3-50.84940.82460.82730.81330.83360.80330.83940.86410.84290.72640.76490.78700.97470.79980.8602
yeast-0-2-5-6_vs_3-7-8-90.74180.70580.73780.74890.77790.75840.77020.72970.64940.95450.82170.96680.79950.69360.7949
yeast-0-2-5-7-9_vs_3-6-80.89100.90660.89380.90500.88520.86990.89030.87820.79650.67570.83100.79930.90910.88450.8939
glass-0-4_vs_50.99360.99400.99400.99400.97510.99400.98130.98770.89950.87230.97180.89351.00000.98811.0000
ecoli-0-1-4-7_vs_5-60.90780.83220.90640.92740.85280.84780.83730.85860.88030.50000.76830.75780.87250.81850.9034
glass40.95050.95570.95580.95830.83730.87230.91490.93740.81350.25890.80620.48290.81650.81500.9796
ecoli40.88350.81080.83920.83490.88510.85150.88720.94170.87700.84280.85920.85800.83850.75110.9374
page-blocks-1-3_vs_40.99770.98040.99770.99770.99430.97690.93510.99090.94770.95520.99440.93111.00000.97930.9989
abalone9-180.61150.55860.66590.67170.57180.66720.64460.58470.40570.35940.63020.70440.33090.50190.6506
MEU-Mobile KSD0.91480.90300.90460.90460.93280.91260.87730.93460.90150.79650.85280.84980.89440.89170.9467
yeast-2_vs_80.68630.56180.72990.74810.69100.73760.68200.71810.72260.26790.44720.54670.50000.65530.8008
flare-F0.50840.48160.36630.64060.54370.71430.73390.37860.25890.74000.00000.83360.00000.27920.3976
kr-vs-k-zero-one_vs_draw0.77270.00000.79440.73440.79310.79130.79430.94570.90730.78850.99910.81080.00000.97070.0000
kr-vs-k-one_vs_fifteen0.84600.83010.82560.83580.84560.83200.88591.00000.99040.32921.00000.57920.90141.00000.8311
winequality-red-40.40570.31270.36790.48030.46680.50010.50870.22910.30780.41550.40690.55600.42160.12330.4730
yeast-1-2-8-9_vs_70.33200.55450.41490.43960.52270.61060.48090.33740.45180.97370.44480.98050.70520.41970.4755
abalone-3_vs_111.00001.00001.00001.00000.99791.00001.00001.00000.99900.35101.00000.63031.00001.00000.9990
kr-vs-k-three_vs_eleven0.82930.00000.82930.82930.82930.79920.92991.00000.95320.90000.99910.89710.00001.00000.0000
ecoli-0-1-3-7_vs_2-60.53520.67840.48280.68160.53520.72340.73040.73330.78500.99900.70710.96330.69410.53400.7851
abalone-17_vs_7-8-9-100.81140.69140.76010.80600.73310.71120.68240.48150.31850.83730.57480.80260.49890.47630.5917
yeast60.73930.76070.81750.79520.74950.77290.81920.72890.66810.95560.49740.95650.65130.64790.8178
poker-8-9vs60.91230.89160.54970.93380.95490.70070.86760.48430.74970.64840.44720.91880.57740.46030.8098
poker-8vs60.97320.70890.57170.87790.94640.65310.78850.37880.64210.95140.50000.93751.00000.38870.8094
Average0.83180.77270.80620.83940.83470.82310.83690.79960.78480.74770.76080.77220.75310.76460.8063
Average Friedman-rank6.19238.50007.56416.60266.15388.17957.39747.79499.83338.871810.66679.02568.294910.46154.4615
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.

References

  1. He, H.; Garcia, E.A. Learning from Imbalanced Data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  2. Devarriya, D.; Gulati, C.; Mansharamani, V.; Sakalle, A.; Bhardwaj, A. Unbalanced breast cancer data classification using novel fitness functions in genetic programming. Expert Syst. Appl. 2020, 140, 112866. [Google Scholar] [CrossRef]
  3. Zavrtanik, V.; Kristan, M.; Skočaj, D. Reconstruction by inpainting for visual anomaly detection. Pattern Recognit. 2021, 112, 107706. [Google Scholar] [CrossRef]
  4. Zhu, H.; Liu, G.; Zhou, M.; Xie, Y.; Abusorrah, A.; Kang, Q. Optimizing Weighted Extreme Learning Machines for imbalanced classification and application to credit card fraud detection. Neurocomputing 2020, 407, 50–62. [Google Scholar] [CrossRef]
  5. Yang, Y.; Huang, S.; Huang, W.; Chang, X. Privacy-Preserving Cost-Sensitive Learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2105–2116. [Google Scholar] [CrossRef]
  6. Zhao, Y.; Wang, S.; Xiao, F. Pattern recognition-based chillers fault detection method using Support Vector Data Description (SVDD). Appl. Energy 2013, 112, 1041–1048. [Google Scholar] [CrossRef]
  7. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 413–422. [Google Scholar] [CrossRef]
  8. Wang, B.; Mao, Z. Outlier detection based on a dynamic ensemble model: Applied to process monitoring. Inf. Fusion 2019, 51, 244–258. [Google Scholar] [CrossRef]
  9. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  10. Krawczyk, B.; Galar, M.; Jeleń, Ł.; Herrera, F. Evolutionary undersampling boosting for imbalanced classification of breast cancer malignancy. Appl. Soft Comput. 2016, 38, 714–726. [Google Scholar] [CrossRef]
  11. Yen, S.J.; Lee, Y.S. Cluster-based under-sampling approaches for imbalanced data distributions. Expert Syst. Appl. 2009, 36, 5718–5727. [Google Scholar] [CrossRef]
  12. Vuttipittayamongkol, P.; Elyan, E. Neighbourhood-based undersampling approach for handling imbalanced and overlapped data. Inf. Sci. 2020, 509, 47–70. [Google Scholar] [CrossRef]
  13. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014—Conference Track Proceedings, Banff, AB, Canada, 14–16 April 2014. [Google Scholar] [CrossRef]
  14. Creswell, A.; White, T.; Dumoulin, V.; Arulkumaran, K.; Sengupta, B.; Bharath, A.A. Generative Adversarial Networks: An Overview. IEEE Signal Process. Mag. 2018, 35, 53–65. [Google Scholar] [CrossRef]
  15. Cover, T.M.; Hart, P.E. Nearest Neighbor Pattern Classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  16. Tao, X.; Zheng, Y.; Chen, W.; Zhang, X.; Qi, L.; Fan, Z.; Huang, S. SVDD-based weighted oversampling technique for imbalanced and overlapped dataset learning. Inf. Sci. 2022, 588, 13–51. [Google Scholar] [CrossRef]
  17. Li, J.; Zhu, Q.; Wu, Q.; Zhang, Z.; Gong, Y.; He, Z.; Zhu, F. SMOTE-NaN-DE: Addressing the noisy and borderline examples problem in imbalanced classification by natural neighbors and differential evolution. Knowl.-Based Syst. 2021, 223, 107056. [Google Scholar] [CrossRef]
  18. Wei, Z.; Zhang, L.; Zhao, L. Minority-prediction-probability-based oversampling technique for imbalanced learning. Inf. Sci. 2023, 622, 1273–1295. [Google Scholar] [CrossRef]
  19. Wang, F.; Zheng, M.; Ma, K.; Hu, X. Resampling approach for imbalanced data classification based on class instance density per feature value intervals. Inf. Sci. 2025, 692, 121570. [Google Scholar] [CrossRef]
  20. Qian, Q.; Sun, W.; Wang, Z.; Lu, Y.; Li, Y.; Jiang, X. GIS partial discharge data enhancement method based on self attention mechanism VAE-GAN. Glob. Energy Interconnect. 2023, 6, 601–613. [Google Scholar] [CrossRef]
  21. Faber, K.; Corizzo, R.; Sniezynski, B.; Japkowicz, N. VLAD: Task-agnostic VAE-based lifelong anomaly detection. Neural Netw. 2023, 165, 248–273. [Google Scholar] [CrossRef]
  22. Zheng, M.; Li, T.; Zhu, R.; Tang, Y.; Tang, M.; Lin, L.; Ma, Z. Conditional Wasserstein generative adversarial network-gradient penalty-based approach to alleviating imbalanced data classification. Inf. Sci. 2020, 512, 1009–1023. [Google Scholar] [CrossRef]
  23. Zhai, J.; Qi, J.; Shen, C. Binary imbalanced data classification based on diversity oversampling by generative models. Inf. Sci. 2022, 585, 313–343. [Google Scholar] [CrossRef]
  24. Huang, K.; Wang, X. ADA-INCVAE: Improved data generation using variational autoencoder for imbalanced classification. Appl. Intell. 2021, 52, 2838–2853. [Google Scholar] [CrossRef]
  25. Ding, H.; Sun, Y.; Wang, Z.; Huang, N.; Shen, Z.; Cui, X. RGAN-EL: A GAN and ensemble learning-based hybrid approach for imbalanced data classification. Inf. Process. Manag. 2023, 60, 103235. [Google Scholar] [CrossRef]
  26. Ding, H.; Sun, Y.; Huang, N.; Shen, Z.; Wang, Z.; Iftekhar, A.; Cui, X. RVGAN-TL: A generative adversarial networks and transfer learning-based hybrid approach for imbalanced data classification. Inf. Sci. 2023, 629, 184–203. [Google Scholar] [CrossRef]
  27. Ma, J.; Zhou, S.; Li, D. Robust multiclass least squares support vector classifier with optimal error distribution. Knowl.-Based Syst. 2021, 215, 106652. [Google Scholar] [CrossRef]
  28. Tao, X.; Li, Q.; Guo, W.; Ren, C.; Li, C.; Liu, R.; Zou, J. Self-adaptive cost weights-based support vector machine cost-sensitive ensemble for imbalanced data classification. Inf. Sci. 2019, 487, 31–56. [Google Scholar] [CrossRef]
  29. Svetnik, V.; Liaw, A.; Tong, C.; Culberson, J.C.; Sheridan, R.P.; Feuston, B.P. Random Forest: A Classification and Regression Tool for Compound Classification and QSAR Modeling. J. Chem. Inf. Comput. Sci. 2003, 43, 1947–1958. [Google Scholar] [CrossRef]
  30. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  31. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar] [CrossRef]
  32. Wang, G.; Wang, J.; He, K. Majority-to-minority resampling for boosting-based classification under imbalanced data. Appl. Intell. 2023, 53, 4541–4562. [Google Scholar] [CrossRef]
  33. Mehmood, Z.; Asghar, S. Customizing SVM as a base learner with AdaBoost ensemble to learn from multi-class problems: A hybrid approach AdaBoost-MSVM. Knowl.-Based Syst. 2021, 217, 106845. [Google Scholar] [CrossRef]
  34. Ning, Z.; Guo, C.; Zhang, D. Meta-distribution-based ensemble sampler for imbalanced semi-supervised learning. Pattern Recognit. 2025, 164, 111552. [Google Scholar] [CrossRef]
  35. Zhou, S.; Gu, Y.; Yu, H.; Yang, X.; Gao, S. RUE: A robust personalized cost assignment strategy for class imbalance cost-sensitive learning. J. King Saud Univ.-Comput. Inf. Sci. 2023, 35, 36–49. [Google Scholar] [CrossRef]
  36. Li, F.; Wang, B.; Shen, Y.; Wang, P.; Li, Y. An overlapping oriented imbalanced ensemble learning algorithm with weighted projection clustering grouping and consistent fuzzy sample transformation. Inf. Sci. 2023, 637, 118955. [Google Scholar] [CrossRef]
  37. Li, Y.; Wang, S.; Jin, J.; Tao, H.; Han, C.; Chen, C.P. Hybrid density-based adaptive weighted collaborative representation for imbalanced learning. Appl. Intell. 2024, 54, 4334–4351. [Google Scholar] [CrossRef]
  38. Li, Y.; Wang, S.; Jin, J.; Zhu, F.; Zhao, L.; Liang, J.; Philip Chen, C. Imbalanced complemented subspace representation with adaptive weight learning. Expert Syst. Appl. 2024, 249, 123555. [Google Scholar] [CrossRef]
  39. Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 471–478. [Google Scholar] [CrossRef]
  40. Gao, X.; Ren, B.; Zhang, H.; Sun, B.; Li, J.; Xu, J.; He, Y.; Li, K. An ensemble imbalanced classification method based on model dynamic selection driven by data partition hybrid sampling. Expert Syst. Appl. 2020, 160, 113660. [Google Scholar] [CrossRef]
  41. Sun, J.; Li, H.; Fujita, H.; Fu, B.; Ai, W. Class-imbalanced dynamic financial distress prediction based on Adaboost-SVM ensemble combined with SMOTE and time weighting. Inf. Fusion 2020, 54, 128–144. [Google Scholar] [CrossRef]
  42. Ali, F.; El-Sappagh, S.; Islam, S.R.; Kwak, D.; Ali, A.; Imran, M.; Kwak, K.S. A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion. Inf. Fusion 2020, 63, 208–222. [Google Scholar] [CrossRef]
  43. Chen, L.; Fang, B.; Shang, Z.; Tang, Y. Tackling class overlap and imbalance problems in software defect prediction. Softw. Qual. J. 2018, 26, 97–125. [Google Scholar] [CrossRef]
  44. Lopez-Martin, M.; Sanchez-Esguevillas, A.; Arribas, J.I.; Carro, B. Supervised contrastive learning over prototype-label embeddings for network intrusion detection. Inf. Fusion 2022, 79, 200–228. [Google Scholar] [CrossRef]
  45. Lorena, A.C.; Garcia, L.P.F.; Lehmann, J.; Souto, M.C.P.; Ho, T.K. How Complex Is Your Classification Problem? A Survey on Measuring Classification Complexity. ACM Comput. Surv. 2019, 52, 1–34. [Google Scholar] [CrossRef]
  46. Oh, S. A new dataset evaluation method based on category overlap. Comput. Biol. Med. 2011, 41, 115–122. [Google Scholar] [CrossRef]
  47. van der Walt, C.M. Data Measures that Characterise Classification Problems. Master’s Thesis, University of Pretoria, Pretoria, South Africa, 2008. [Google Scholar]
  48. Kingma, D.P.; Dhariwal, P. Glow: Generative Flow with Invertible 1 × 1 Convolutions. arXiv 2018. [Google Scholar] [CrossRef]
  49. Bader-El-Den, M.; Teitei, E.; Perry, T. Biased random forest for dealing with the class imbalance problem. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 2163–2172. [Google Scholar] [CrossRef]
  50. Dai, Q.; Liu, J.W.; Yang, J.P. SWSEL: Sliding Window-based Selective Ensemble Learning for class-imbalance problems. Eng. Appl. Artif. Intell. 2023, 121, 105959. [Google Scholar] [CrossRef]
  51. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning. Lect. Notes Comput. Sci. 2005, 3644, 878–887. [Google Scholar] [CrossRef]
  52. Sandhan, T.; Choi, J.Y. Handling imbalanced datasets by partially guided hybrid sampling for pattern recognition. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1449–1453. [Google Scholar] [CrossRef]
  53. Bai, L.; Ju, T.; Wang, H.; Lei, M.; Pan, X. Two-step ensemble under-sampling algorithm for massive imbalanced data classification. Inf. Sci. 2024, 665, 120351. [Google Scholar] [CrossRef]
  54. Xu, L.; Skoularidou, M.; Cuesta-Infante, A.; Veeramachaneni, K. Modeling Tabular data using Conditional GAN. Adv. Neural Inf. Process. Syst. 2019, 32, 1–11. [Google Scholar]
  55. Huang, Z.; Gao, X.; Chen, W.; Cheng, Y.; Xue, B.; Meng, Z.; Zhang, G.; Fu, S. An imbalanced binary classification method via space mapping using normalizing flows with class discrepancy constraints. Inf. Sci. 2023, 623, 493–523. [Google Scholar] [CrossRef]
  56. Schultz, K.; Bej, S.; Hahn, W.; Wolfien, M.; Srivastava, P.; Wolkenhauer, O. ConvGeN: A convex space learning approach for deep-generative oversampling and imbalanced classification of small tabular datasets. Pattern Recognit. 2024, 147, 110138. [Google Scholar] [CrossRef]
  57. Ahsan, M.M.; Ali, M.S.; Siddique, Z. Enhancing and improving the performance of imbalanced class data using novel GBO and SSG: A comparative analysis. Neural Netw. 2024, 173, 106157. [Google Scholar] [CrossRef] [PubMed]
  58. Derrac, J.; Garcia, S.; Sanchez, L.; Herrera, F. Keel data-mining software tool: Data set repository, integration of algorithms and experimental analysis framework. J. Mult. Valued Log. Soft Comput 2015, 17, 255–287. [Google Scholar]
  59. Barella, V.H.; Garcia, L.P.; de Souto, M.C.; Lorena, A.C.; de Carvalho, A.C. Assessing the data complexity of imbalanced datasets. Inf. Sci. 2021, 553, 83–109. [Google Scholar] [CrossRef]
  60. García, S.; Fernández, A.; Luengo, J.; Herrera, F. Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf. Sci. 2010, 180, 2044–2064. [Google Scholar] [CrossRef]
  61. Taheri, S.M.; Hesamian, G. A generalization of the Wilcoxon signed-rank test and its applications. Stat. Pap. 2012, 54, 457–470. [Google Scholar] [CrossRef]
  62. Pereira, D.G.; Afonso, A.; Medeiros, F.M. Overview of Friedman’s Test and Post-hoc Analysis. Commun. Stat.-Simul. Comput. 2015, 44, 2636–2653. [Google Scholar] [CrossRef]
Figure 1. The entire process of DPOA-MRM.
Figure 1. The entire process of DPOA-MRM.
Electronics 14 04740 g001
Figure 2. The advantage of mapping original samples to low dimensional space. The red circles highlight the class-overlapping regions.
Figure 2. The advantage of mapping original samples to low dimensional space. The red circles highlight the class-overlapping regions.
Electronics 14 04740 g002
Figure 3. The structure of the flow model.
Figure 3. The structure of the flow model.
Electronics 14 04740 g003
Figure 4. Classify the original samples into three types based on spatial distribution.
Figure 4. Classify the original samples into three types based on spatial distribution.
Electronics 14 04740 g004
Figure 5. The overall flowchart of DPOA-MRM.
Figure 5. The overall flowchart of DPOA-MRM.
Electronics 14 04740 g005
Figure 6. The Nemenyi post hoc test results for DPOA-MRM and methods at the algorithm level.
Figure 6. The Nemenyi post hoc test results for DPOA-MRM and methods at the algorithm level.
Electronics 14 04740 g006
Figure 7. The Nemenyi post hoc test results for DPOA-MRM and methods at the data level.
Figure 7. The Nemenyi post hoc test results for DPOA-MRM and methods at the data level.
Electronics 14 04740 g007
Table 1. Asymptotic Complexity of Various Ensemble Learning Methods at the Same Ensemble Scale (Constant Terms Omitted).
Table 1. Asymptotic Complexity of Various Ensemble Learning Methods at the Same Ensemble Scale (Constant Terms Omitted).
MethodTraining Time
Complexity
Testing Time
Complexity
Peak Space
Complexity
DPOA-MRM (ours) O ( K E N d l o g N ) O d + E l o g N O K E N l o g N + N d
GBDT O E N d l o g N O E l o g N O E N
RF O E N d l o g N O E l o g N O E N
BRAF O d k + E N k l o g N O l o g d + E l o g N O E N
DPHS-MDS O N d l o g N O E l o g N O E N
SWSEL O R E w d l o g w O α E l o g w O α E w
MDSampler O E N l o g N · l o g d O E l o g N O E N
Table 2. Compared methods information.
Table 2. Compared methods information.
LevelMethodBasisPublication and Year
AlgorithmiForest [7]One-class learningICDM 2008
SVDD [6]Applied Energy 2013
GDBT [30]Ensemble learningAnnals of statistics 2001
RF [29]Journal of Chemical Information and Computer Sciences 2003
BRAF [49]IEEE Transactions on Neural Networks and Learning Systems 2018
DPHS-MDS [40]Expert Systems with Applications 2020
SWSEL [50]Engineering Applications of Artificial Intelligence 2023
MDSampler [34]Pattern Recognition 2025
PCGDST-IE [36]Cost-sensitive learningInformation Sciences 2023
HDAWCR [37]Applied Intelligence 2024
AWLICSR [38]Expert Systems with Applications 2024
DataSMOTE [9]SamplingJournal of Artificial Intelligence Research 2002
Borderline-SMOTE [51]ICIC 2005
G-SMOTE [52]ICPR 2014
SMOTE-NaN-DE [17]Knowledge-Based Systems 2021
MPP-SMOTE [18]Information Sciences 2023
TSSE-BIM [53]Information Sciences 2024
HSCF [19]Information Sciences 2025
CTGAN [54]GeneratingNIPS 2019
CWGAN-GP [22]Information Sciences 2020
ADA-INCVAE [24]Applied Intelligence 2022
RVGAN-TL [26]Information Sciences 2023
CDC-Glow [55]Information Sciences 2023
ConvGeN [56]Pattern Recognition 2024
SSG [57]Neural Networks 2024
Table 3. Characteristics of the selected datasets in the experimental verification.
Table 3. Characteristics of the selected datasets in the experimental verification.
DatasetInstancesFeaturesMinority
Instances
Majority
Instances
IROverlap
Degree
ecoli-0vs12208771431.860.0367
wisconsin683102394441.860.0405
pima76892685001.870.5625
vehicle_2846192186282.880.6811
vehicle_1846192176292.90.8221
vehicle_3846192126342.990.8284
vehicle_0846191996473.250.5682
ecoli_13368772593.360.3566
newthyroid12156351805.140.0819
newthyroid22156351805.140.0727
ecoli23368522845.460.3416
segment023082032919796.020.4925
yeast31484916313218.10.3092
ecoli33368353018.60.4916
page-blocks054721155949138.790.0640
yeast-2_vs_45149514639.080.2896
ecoli-0-6-7_vs_3-52228222009.090.0527
yeast-0-2-5-6_vs_3-7-8-910049999059.140.3108
yeast-0-2-5-7-9_vs_3-6-810049999059.140.3058
glass-0-4_vs_592109839.220.2508
ecoli-0-1-4-7_vs_5-633272530712.280.2341
glass4214101320115.460.2434
ecoli433682031615.80.2631
page-blocks-1-3_vs_4472112844415.860.0731
abalone9-1873194268916.40.4568
MEU-Mobile KSD107172511020200.2848
yeast-2_vs_848292046223.10.0000
flare-F10661243102323.790.4131
kr-vs-k-zero-one_vs_draw29017105279626.630.2681
kr-vs-k-one_vs_fifteen2244778216627.770.0000
winequality-red-415991253154629.170.6376
yeast-1-2-8-9_vs_794793091730.570.6547
abalone-3_vs_1150291548732.470.4092
kr-vs-k-three_vs_eleven2935781285435.230.5400
ecoli-0-1-3-7_vs_2-62818727439.140.0499
abalone-17_vs_7-8-9-102338958228039.310.0747
yeast61484935144941.40.3550
poker-8-9vs614851125146058.40.9367
poker-8vs614771117146085.880.9297
Table 4. F1-measure results for DPOA-MRM and compared methods at the algorithm level.
Table 4. F1-measure results for DPOA-MRM and compared methods at the algorithm level.
DatasetiForestSVDDGBDTRFBRAFDPHS-MDSSWSELPCGDST-IEHDAWCRAWLICSRMDSamplerDPOA-MRM
ecoli-0vs10.35930.65750.98100.98710.98710.98190.98620.95430.00000.28570.96040.9802
wisconsin0.94200.74300.95440.96090.95790.95970.93730.95390.16660.06630.96080.9617
pima0.35440.57460.64960.63750.63820.66570.67150.61820.63760.66660.65730.6786
vehicle_20.20420.16240.95800.97000.97240.97100.85430.76540.44770.52940.97040.9752
vehicle_10.12010.30490.54640.52030.54980.60950.56480.57070.96200.96200.64790.6532
vehicle_30.10320.44510.51940.49640.51120.56880.54340.54340.95550.96620.66610.6099
vehicle_00.22080.32180.92440.93550.94300.94080.81680.80440.25000.25000.92950.9520
ecoli_10.19330.49440.76810.78880.78710.76660.78580.74720.84840.96960.78680.8321
new-thyroid10.45840.50240.88720.95600.95160.92760.91330.90260.06980.06980.97130.9647
new-thyroid20.43660.47310.92260.95130.96670.94190.95140.90860.93330.93330.95791.0000
ecoli20.04220.26760.81430.79940.82580.82000.67410.64040.73330.78570.80320.8816
segment00.00640.04110.98030.99230.98180.98840.97760.90230.02670.02670.98040.9909
yeast30.03180.17460.75190.74600.77110.78600.68770.75090.00000.00000.75660.7948
ecoli30.14020.30090.63370.57720.57390.57970.61540.54830.47050.79990.59720.7029
page-blocks00.42220.39740.87210.88010.88690.87750.69560.58590.11880.11880.88260.8533
yeast-2_vs_40.34650.42060.70150.75990.75550.73590.73870.65070.00000.00000.79550.8225
ecoli-0-6-7_vs_3-50.33750.28230.74570.80710.74480.75790.69150.70740.37030.52630.66010.7802
yeast-0-2-5-6_vs_3-7-8-90.29880.39030.59820.62620.63690.64590.35060.50890.51890.51890.55580.6735
yeast-0-2-5-7-9_vs_3-6-80.28530.46430.77460.79520.78420.79460.63340.74250.36730.37490.80600.8231
glass-0-4_vs_50.29050.17460.93330.93330.96000.96000.86670.61310.08070.08070.93331.0000
ecoli-0-1-4-7_vs_5-60.31990.19230.80330.78330.81540.78980.51530.58560.18180.99990.70220.8039
glass40.23240.20300.51330.62000.69780.68950.63430.56950.99990.00000.62670.8305
ecoli40.19920.20920.78620.75620.75630.75430.78890.59090.00000.54540.71870.8292
page-blocks-1-3_vs_40.41100.13460.98180.94180.97870.94470.68900.58700.76920.93330.98180.9846
abalone9-180.18000.10770.31930.15720.33260.28890.13540.29030.40000.79990.24920.4340
MEU-Mobile KSD0.02500.26110.79420.86270.93320.89240.50860.86090.09170.08930.90650.9204
yeast-2_vs_80.55950.12080.52760.54670.57710.58380.26440.22890.28570.62490.23170.7405
flare-F0.15850.12160.12450.11680.03230.03230.19420.58480.50000.72720.19320.2503
kr-vs-k-zero-one_vs_draw0.11180.08280.00000.00000.91100.90140.00000.92250.05620.05620.37410.0000
kr-vs-k-one_vs_fifteen0.23290.19300.81480.81481.00001.00000.81470.99310.49990.49990.81480.8167
winequality-red-40.06640.04720.00000.00000.02180.02380.11960.30600.40620.46370.14110.1783
yeast-1-2-8-9_vs_70.04610.07220.31440.23000.22510.18080.11120.22210.59250.67990.12600.3248
abalone-3_vs_110.17950.23671.00001.00001.00001.00000.77140.88150.05000.05000.93140.9714
kr-vs-k-three_vs_eleven0.09410.08000.00000.00000.95700.96520.00000.94600.06870.06870.15650.0000
ecoli-0-1-3-7_vs_2-60.15930.08010.34670.46670.57930.74670.37330.29820.20000.20000.19210.5800
abalone-17_vs_7-8-9-100.09210.08070.21050.03330.24630.24330.06640.46070.08930.08930.27770.3235
yeast60.13030.13440.34350.49550.50970.49760.32210.44680.19940.19940.31420.6467
poker-8-9vs60.03020.03160.06670.06670.33240.12480.06920.27510.10000.23530.06850.7336
poker-8vs60.01890.02210.10000.00000.00800.00000.02370.34440.04230.04490.03550.7867
Average0.22670.25650.61440.61570.69490.69070.54760.63620.34590.41640.62360.7201
Average Friedman-rank10.153810.10266.20516.03854.14104.60267.17956.66678.03857.06415.21792.5897
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 5. G-mean results for DPOA-MRM and compared methods at the algorithm level.
Table 5. G-mean results for DPOA-MRM and compared methods at the algorithm level.
DatasetiForestSVDDGBDTRFBRAFDPHS-MDSSWSELPCGDST-IEHDAWCRAWLICSRMDSamplerDPOA-MRM
ecoli-0vs10.43630.61160.98390.98730.98730.98440.98640.96560.00000.44160.96690.9833
wisconsin0.96240.76890.96650.97190.96950.97190.96230.96660.30150.00000.97180.9774
pima0.47610.66300.72360.71200.71400.74060.74480.69000.85140.78400.73390.7457
vehicle_20.36450.32980.96580.97740.98180.97910.93350.86560.56500.62420.98060.9867
vehicle_10.27170.40260.65600.64180.66550.73280.71770.72120.97090.97090.78030.7802
vehicle_30.25440.55650.63490.60670.62740.69880.70300.69890.97670.98070.80120.7540
vehicle_00.38250.40480.94980.95520.96700.96740.92250.90820.00000.00000.96180.9792
ecoli_10.28270.64910.83920.84910.86460.85130.86380.87760.88570.98260.87580.9168
new-thyroid10.64300.77100.92730.96770.96350.93930.95780.96310.00000.00000.98240.9915
new-thyroid20.63070.74770.93660.95420.97200.95220.98870.96350.98600.98600.97961.0000
ecoli20.11730.36220.86480.83520.88210.85570.85220.86730.80490.82110.90190.9370
segment00.03240.14420.98770.99240.99310.98970.99110.96680.00000.00000.99030.9934
yeast30.09360.46470.84430.82240.85200.87000.91310.93870.00000.00000.87690.9221
ecoli30.31560.62950.74970.68120.68380.69010.88980.86780.59230.91940.82500.8667
page-blocks00.63220.58740.91650.92500.93570.94200.93960.84730.00000.00000.93850.9401
yeast-2_vs_40.52420.79190.82350.83150.82910.80540.92790.88160.00000.00000.92100.9169
ecoli-0-6-7_vs_3-50.64180.66140.81840.84390.80300.80210.83600.85700.85190.92460.84150.8602
yeast-0-2-5-6_vs_3-7-8-90.48770.74970.70360.70790.72030.73280.74690.75940.00000.00000.79890.7949
yeast-0-2-5-7-9_vs_3-6-80.47380.82650.85530.85750.86180.85800.89800.91800.63270.63460.90310.8939
glass-0-4_vs_50.40200.43070.99360.99360.99360.99360.98150.84620.00000.00000.99361.0000
ecoli-0-1-4-7_vs_5-60.69590.55060.86100.83400.86040.83330.86000.90380.82021.00000.89160.9034
glass40.50220.69540.59470.71950.75270.76390.96180.87901.00000.00000.92470.9796
ecoli40.39740.64530.85650.78780.83500.78020.92960.91670.00000.64930.89750.9374
page-blocks-1-3_vs_40.75010.51200.99890.96050.98230.94820.97020.93370.83330.98600.99890.9989
abalone9-180.57610.32660.38840.23540.42140.39450.51080.72590.79540.81650.72540.6506
MEU-Mobile KSD0.06250.79500.82100.87130.93610.89860.89750.91950.17150.00000.94490.9467
yeast-2_vs_80.78590.51790.62300.62370.64910.65540.76390.68050.42410.67420.76240.8008
flare-F0.71240.54300.24950.23110.06630.06630.74740.80700.93540.97630.69110.3976
kr-vs-k-zero-one_vs_draw0.62480.34390.00000.00000.94100.92510.00000.96520.00000.00000.78880.0000
kr-vs-k-one_vs_fifteen0.80960.80200.83030.83031.00001.00000.83010.99740.57730.57730.83030.8311
winequality-red-40.29920.31630.00000.00000.03790.04360.62740.59130.53220.58550.68860.4730
yeast-1-2-8-9_vs_70.15860.42670.41910.31170.30760.27090.61790.65710.84880.88820.67720.4755
abalone-3_vs_110.84640.72891.00001.00001.00001.00000.99070.99520.00000.00000.96230.9990
kr-vs-k-three_vs_eleven0.67250.61140.00000.00000.96850.97060.00000.98310.00000.00000.79780.0000
ecoli-0-1-3-7_vs_2-60.66520.63710.53330.48280.66620.87840.71910.83640.33330.33330.72110.7851
abalone-17_vs_7-8-9-100.61390.49700.35920.06030.38560.36970.57190.66560.00000.00000.83200.5917
yeast60.39760.67030.48040.59490.63690.61090.85630.83050.00000.00000.81980.8178
poker-8-9vs60.25650.47770.08940.08940.39790.16470.66870.71620.23210.37320.71080.8098
poker-8vs60.24850.48860.11550.00000.01000.00000.43270.70300.55390.58050.61390.8094
Average0.47440.56770.67590.66020.74670.74180.78750.84810.42250.44900.85400.8063
Average Friedman-rank9.46159.02567.65387.51285.80776.20515.12824.43598.42317.47443.75643.1154
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 6. Wilcoxon signed-rank test results for DPOA-MRM and compared methods at the algorithm level.
Table 6. Wilcoxon signed-rank test results for DPOA-MRM and compared methods at the algorithm level.
F1-ScoreG-Mean
VSR+R−p-ValueAssumingR+R−p-ValueAssuming
iForest7737 9.05 × 10 8 rejected69387 2.35 × 10 5 rejected
SVDD7773 6.64 × 10 8 rejected70872 9.09 × 10 6 rejected
GBDT75822 6.66 × 10 7 rejected7764 2.36 × 10 7 rejected
RF74238 2.25 × 10 6 rejected773.56.5 1.94 × 10 7 rejected
BRAF621159 1.27 × 10 3 rejected670110 9.33 × 10 5 rejected
DPHS-MDS611169 2.04 × 10 3 rejected647133 3.35 × 10 4 rejected
SWSEL7782 1.34 × 10 7 rejected617163 4.46 × 10 3 rejected
PSGDST-IE615165 1.69 × 10 3 rejected469311 2.70 × 10 1 not rejected
HDAWCR68694 3.62 × 10 5 rejected69585 5.81 × 10 5 rejected
AWLICSR632148 7.33 × 10 4 rejected669111 2.85 × 10 4 rejected
MDSampler674106 7.39 × 10 5 rejected476304 3.35 × 10 1 not rejected
Table 7. F1-measure results for DPOA-MRM and compared methods at the data level.
Table 7. F1-measure results for DPOA-MRM and compared methods at the data level.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.97460.98620.98620.98620.97460.97930.97980.98100.98020.65470.96770.78470.96770.98670.9802
wisconsin0.95920.95500.95690.94430.96740.95050.93740.96340.94890.99230.93980.04950.95920.96050.9617
pima0.65100.65580.66290.62310.66750.62550.67780.65460.63130.69590.52430.47430.69030.63290.6786
vehicle_20.97500.97000.97040.96750.96400.96360.87430.95320.96450.64330.97560.64440.94510.97160.9752
vehicle_10.59710.61050.62140.58430.61660.58790.63370.57710.56400.80200.58140.87870.60000.50540.6532
vehicle_30.63290.57140.58380.54330.60500.56400.59350.54350.56610.55950.37140.51520.57500.51910.6099
vehicle_00.91990.91750.92560.91140.92270.91580.75060.93460.92190.83760.93510.71970.89740.93750.9520
ecoli_10.78880.78690.81090.80880.80050.77040.76040.80710.73210.82780.77780.70630.74290.75400.8321
new-thyroid10.97130.91880.90350.92000.95590.88510.94270.94830.85330.88721.00000.94090.92310.93790.9647
new-thyroid20.97130.92630.95330.94170.93090.96000.91500.95140.93790.90770.93330.92231.00000.96671.0000
ecoli20.85900.81990.80490.81920.81040.76950.74040.82630.78020.98020.75000.20840.85710.81450.8816
segment00.99240.98920.99240.98470.98500.97860.97090.98800.98920.95590.98440.95590.99240.99230.9909
yeast30.78960.75560.77280.79550.76310.75050.75910.74970.73710.00000.72460.15960.60710.74630.7948
ecoli30.62130.67010.62250.64160.62560.63360.63890.62650.56330.71640.54550.72950.80000.53530.7029
page-blocks00.86320.87610.88880.85090.82920.82410.66700.87890.79770.88240.87240.78880.78430.89500.8533
yeast-2_vs_40.78520.77360.79450.77810.77310.71090.69590.76950.75530.47360.80000.70300.75000.76190.8225
ecoli-0-6-7_vs_3-50.72950.75760.76520.74700.65890.74700.69980.82530.78330.66670.75000.66670.90910.72250.7802
yeast-0-2-5-6_vs_3-7-8-90.61720.52840.58650.58220.60010.51870.55810.62000.46890.93590.64000.95760.72220.62970.6735
yeast-0-2-5-7-9_vs_3-6-80.80040.76760.79170.82970.79060.78740.74980.80010.68810.48860.80000.70860.85710.83250.8231
glass-0-4_vs_51.00001.00001.00001.00001.00000.96000.96000.93330.79080.70790.66670.73461.00000.80001.0000
ecoli-0-1-4-7_vs_5-60.76870.76110.74830.83430.72600.73110.75660.80990.70560.25580.66670.54490.80000.79440.8039
glass40.83430.86480.76670.89330.65480.72670.61180.76140.70680.09830.50000.17460.80000.62000.8305
ecoli40.78890.75140.78000.75780.83250.75250.73550.92060.75240.75950.40000.80400.80000.76510.8292
page-blocks-1-3_vs_40.98180.98180.96360.96360.95320.93590.49430.96640.72260.91851.00000.78711.00000.90180.9846
abalone9-180.26490.31300.30470.33420.35000.28000.22720.32690.32190.09350.00000.26950.00000.12000.4340
MEU-Mobile KSD0.92260.84800.88610.93570.90500.87640.91230.92360.76670.75780.77780.83600.88890.87310.9204
yeast-2_vs_80.58860.63240.67140.52020.66730.43600.60140.66100.62440.00000.33330.17690.40000.63240.7405
flare-F0.26220.12910.16910.20890.17410.23860.16300.13120.10670.59480.11110.64310.00000.09330.2503
kr-vs-k-zero-one_vs_draw0.24410.00000.21140.15200.30270.33930.25400.96120.68550.63140.97140.65000.00000.96100.0000
kr-vs-k-one_vs_fifteen0.58080.81470.71830.25510.72250.81350.52311.00000.92420.03331.00000.23080.89661.00000.8167
winequality-red-40.16060.05710.07110.10540.19370.17370.17440.11880.06970.20000.00000.50000.14290.00000.1783
yeast-1-2-8-9_vs_70.18050.19160.26220.12380.14770.18910.35870.25780.23890.97680.00000.97330.28570.17140.3248
abalone-3_vs_111.00001.00001.00001.00000.94290.96001.00001.00000.89290.07941.00000.32121.00001.00000.9714
kr-vs-k-three_vs_eleven0.15400.00000.15430.15430.15400.15680.37061.00000.82290.76001.00000.67110.00001.00000.0000
ecoli-0-1-3-7_vs_2-60.46000.46670.46670.66670.49330.43560.61330.69330.66671.00000.40000.96000.66670.53330.5800
abalone-17_vs_7-8-9-100.27660.38560.26350.33340.33540.30940.21120.24760.11970.81390.00000.83250.25000.21330.3235
yeast60.47500.44370.56370.51070.44220.39940.37380.58660.45500.95830.00000.95070.40000.43960.6467
poker-8-9vs60.60320.62300.41430.48890.82220.19730.79210.44440.55630.59250.33330.67250.50000.28330.7336
poker-8vs60.49330.38000.28000.08000.59330.21890.38000.24000.52000.95590.40000.95790.40000.16000.7867
Average0.68050.66360.66900.65580.68340.64240.64250.72780.66960.64350.62650.64630.66180.67860.7201
Average Friedman-rank6.57697.93596.96157.69237.30779.91039.58976.25649.76928.60269.80779.42317.60268.37184.1923
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 8. G-mean results for DPOA-MRM and compared methods at the data level.
Table 8. G-mean results for DPOA-MRM and compared methods at the data level.
DatasetSMOTEBorderline-SMOTEG-SMOTESMOTE-NaN-DEMPP-SMOTETSSE-BIMHSCFCTGANCWGAN-GPADA-INCVAERVGAN-TLCDC-GlowConvGeNSSGDPOA-MRM
ecoli-0vs10.98000.98640.98640.98640.98030.98260.98310.98390.98350.78600.96820.91110.96820.98690.9833
wisconsin0.97170.96820.96930.96050.97820.96410.95250.97580.96410.99240.95990.00000.97270.97080.9774
pima0.72800.73190.73680.70420.74150.70600.74810.73110.70810.84330.61300.85520.76010.70730.7457
vehicle_20.98370.97910.98220.97660.98120.97810.94510.97570.98570.71590.98390.72070.97280.97660.9867
vehicle_10.72540.73630.74300.70870.75080.71930.77110.69570.69430.87050.72390.94130.72190.63030.7802
vehicle_30.75480.70710.71300.68380.73770.70300.74130.65770.69140.62380.48430.62310.68990.62490.7540
vehicle_00.95830.95790.95730.94390.96900.94880.88830.96690.96210.91850.96220.88240.92460.95940.9792
ecoli_10.87480.87990.88830.89210.89120.84820.88690.86880.80350.89580.86190.83140.84780.83740.9168
new-thyroid10.98240.94560.93100.93090.96750.92900.96490.98880.92080.93511.00000.95110.92580.95140.9915
new-thyroid20.98240.94860.96620.96340.96190.97990.95780.98870.95140.95600.93540.95671.00000.96901.0000
ecoli20.90610.86520.86980.88210.89000.86350.89100.89300.85970.98350.77460.00000.89660.84470.9370
segment00.99490.99060.99490.98850.99620.98620.99240.99540.99060.96750.99090.96750.99240.99230.9934
yeast30.89160.86920.88290.90350.88450.86560.87570.84110.81240.00000.80170.41240.70960.82650.9221
ecoli30.79810.83200.79650.81820.82580.78620.81150.76850.68490.78950.61240.80290.91050.65030.8667
page-blocks00.94730.94740.95220.94860.94550.93410.93290.93240.88590.95650.93160.94380.89640.93430.9401
yeast-2_vs_40.87130.86400.88400.90970.91600.84240.90240.87040.84860.59260.86150.79430.88490.82690.9169
ecoli-0-6-7_vs_3-50.85330.82870.84780.84530.84200.84110.85680.89150.86990.73960.77460.79260.98740.79750.8602
yeast-0-2-5-6_vs_3-7-8-90.75770.67840.72940.74450.78280.74040.79040.71640.60390.96200.68600.97600.79950.71140.7949
yeast-0-2-5-7-9_vs_3-6-80.88790.84280.87700.90570.90070.87580.89500.88230.77050.61560.83790.80160.93810.88150.8939
glass-0-4_vs_51.00001.00001.00001.00001.00000.99400.99400.99360.90580.83840.97180.88761.00000.80001.0000
ecoli-0-1-4-7_vs_5-60.83440.81290.83610.88730.85570.83530.89800.90640.82700.36000.76830.76270.88720.84110.9034
glass40.91900.90220.81810.92450.83900.88780.92300.94010.85050.20270.57740.43280.81650.68380.9796
ecoli40.88190.81240.83920.83760.88490.88300.92740.94480.86600.84460.50000.89360.98430.82580.9374
page-blocks-1-3_vs_40.98260.98260.96510.96510.99660.97930.93230.99770.87930.96441.00000.92811.00000.93570.9989
abalone9-180.57140.49210.55090.63680.59720.58830.66030.50800.47050.19980.00000.50200.00000.20760.6506
MEU-Mobile KSD0.92630.85890.89300.93780.93520.92340.91610.93610.89840.84360.79770.88200.89440.88150.9467
yeast-2_vs_80.74060.69670.73730.72900.74480.71150.68640.71910.72170.00000.44720.37450.50000.69670.8008
flare-F0.52760.27250.30200.61420.47510.69180.74450.25780.20320.68850.36870.83840.00000.16000.3976
kr-vs-k-zero-one_vs_draw0.77270.00000.79170.73440.79310.78440.87750.97060.78710.75370.99910.80900.00000.97050.0000
kr-vs-k-one_vs_fifteen0.84600.83010.82560.83580.84560.82920.96391.00000.99580.06631.00000.50120.90141.00000.8311
winequality-red-40.33430.08510.14460.31000.42730.60760.57220.34720.24180.23090.00000.55670.30050.00000.4730
yeast-1-2-8-9_vs_70.33360.32530.39210.27440.33060.60360.56420.33750.31210.98290.00000.98740.40820.24490.4755
abalone-3_vs_111.00001.00001.00001.00000.99790.96331.00001.00000.99590.14161.00000.58681.00001.00000.9990
kr-vs-k-three_vs_eleven0.82930.00000.82930.82930.82930.79950.94951.00000.95820.82801.00000.82690.00001.00000.0000
ecoli-0-1-3-7_vs_2-60.53830.48280.48280.68280.53960.72580.73590.73960.79631.00000.50000.96330.70710.54140.7851
abalone-17_vs_7-8-9-100.59310.61240.50720.67660.65220.72260.66180.48310.29230.84730.00000.85980.40740.32680.5917
yeast60.71010.64670.74940.76440.70830.79010.80630.69040.59030.96860.00000.96500.53360.55780.8178
poker-8-9vs60.66310.67620.43630.53670.84720.71670.81520.53670.58680.78660.44720.92810.57740.33380.8098
poker-8vs60.50470.41550.33090.10000.62020.84110.44640.26330.59420.96750.50000.97960.50000.20000.8094
Average0.80410.72980.77280.79420.81700.83010.84260.79990.76320.70920.68310.75970.72350.72530.8063
Average Friedman-rank6.48729.25647.93597.52565.67958.37186.21796.756410.05139.217910.73089.11548.166710.26924.2179
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 9. Wilcoxon signed-rank test results for DPOA-MRM and compared methods at the data level.
Table 9. Wilcoxon signed-rank test results for DPOA-MRM and compared methods at the data level.
F1-ScoreG-Mean
VSR+R−p-ValueAssumingR+R−p-ValueAssuming
SMOTE637143 9.69 × 10 4 rejected642138 7.47 × 10 4 rejected
Borderline-SMOTE72159 1.67 × 10 5 rejected759.520.5 9.12 × 10 7 rejected
G-SMOTE674106 1.25 × 10 4 rejected69189 4.46 × 10 5 rejected
SMOTE-NaN-DE615165 2.88 × 10 3 rejected621159 2.16 × 10 3 rejected
MPP-SMOTE654126 3.91 × 10 4 rejected586194 1.05 × 10 2 rejected
TSSE-BIM70971 8.52 × 10 6 rejected546234 2.95 × 10 2 rejected
HSCF660120 1.65 × 10 4 rejected504.5275.5 1.10 × 10 1 not rejected
CTGAN549231 2.65 × 10 2 rejected627153 9.42 × 10 4 not rejected
CWGAN-GP668112 1.78 × 10 4 rejected663117 1.39 × 10 4 rejected
ADA-INCVAE482298 1.99 × 10 1 not rejected508272 9.96 × 10 2 rejected
RVGAN-TL643137 4.15 × 10 4 rejected671109 8.81 × 10 5 rejected
CDC-Glow483297 1.94 × 10 1 not rejected450330 4.02 × 10 1 not rejected
ConvGeN626154 8.36 × 10 3 rejected674.5105.5 6.00 × 10 4 rejected
SSG641139 4.61 × 10 4 rejected674106 7.39 × 10 5 rejected
Table 10. F1-measure results for algorithm-level methods and DPOA-MRM when the datasets overlap seriously.
Table 10. F1-measure results for algorithm-level methods and DPOA-MRM when the datasets overlap seriously.
DatasetiForestSVDDGBDTRFBRAFDPHS-MDSSWSELPCGDST-IEHDAWCRAWLICSRMDSamplerDPOA-MRM
pima0.35440.57460.64960.63750.63820.66570.67150.61820.10000.23530.65730.6786
vehicle_20.20420.16240.95800.97000.97240.97100.85430.76540.95550.96620.97040.9752
vehicle_10.12010.30490.54640.52030.54980.60950.56480.57070.44770.52940.64790.6532
vehicle_30.10320.44510.51940.49640.51120.56880.54340.54340.40620.46370.66610.6099
vehicle_00.22080.32180.92440.93550.94300.94080.81680.80440.96200.96200.92950.9520
segment00.00640.04110.98030.99230.98180.98840.97760.90230.25000.25000.98040.9909
ecoli30.14020.30090.63370.57720.57390.57970.61540.54830.00000.54540.59720.7029
abalone9-180.18000.10770.31930.15720.33260.28890.13540.29030.20000.20000.24920.4340
flare-F0.15850.12160.12450.11680.03230.03230.19420.58480.08070.08070.19320.2503
winequality-red-40.06640.04720.00000.00000.02180.02380.11960.30600.16660.06630.14110.1783
yeast-1-2-8-9_vs_70.04610.07220.31440.23000.22510.18080.11120.22210.00000.00000.12600.3248
kr-vs-k-three_vs_eleven0.09410.08000.00000.00000.95700.96520.00000.94600.05620.05620.15650.0000
abalone-17_vs_7-8-9-100.09210.08070.21050.03330.24630.24330.06640.46070.05000.05000.27770.3235
poker-8-9vs60.03020.03160.06670.06670.33240.12480.06920.27510.04230.04490.06850.7336
poker-8vs60.01890.02210.10000.00000.00800.00000.02370.34440.02670.02670.03550.7867
Average0.12240.18090.42310.38220.48840.47890.38420.54550.24960.29840.44640.5729
Average Friedman-rank9.53339.53336.03337.73335.43335.20006.56674.80008.60008.13334.33332.1000
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 11. G-mean results for algorithm-level methods and DPOA-MRM when the datasets overlap seriously.
Table 11. G-mean results for algorithm-level methods and DPOA-MRM when the datasets overlap seriously.
DatasetiForestSVDDGBDTRFBRAFDPHS-MDSSWSELPCGDST-IEHDAWCRAWLICSRMDSamplerDPOA-MRM
pima0.47610.66300.72360.71200.71400.74060.74480.69000.23210.37320.73390.7457
vehicle_20.36450.32980.96580.97740.98180.97910.93350.86560.97670.98070.98060.9867
vehicle_10.27170.40260.65600.64180.66550.73280.71770.72120.56500.62420.78030.7802
vehicle_30.25440.55650.63490.60670.62740.69880.70300.69890.53220.58550.80120.7540
vehicle_00.38250.40480.94980.95520.96700.96740.92250.90820.97090.97090.96180.9792
segment00.03240.14420.98770.99240.99310.98970.99110.96680.00000.00000.99030.9934
ecoli30.31560.62950.74970.68120.68380.69010.88980.86780.00000.64930.82500.8667
abalone9-180.57610.32660.38840.23540.42140.39450.51080.72590.33330.33330.72540.6506
flare-F0.71240.54300.24950.23110.06630.06630.74740.80700.00000.00000.69110.3976
winequality-red-40.29920.31630.00000.00000.03790.04360.62740.59130.30150.00000.68860.4730
yeast-1-2-8-9_vs_70.15860.42670.41910.31170.30760.27090.61790.65710.00000.00000.67720.4755
kr-vs-k-three_vs_eleven0.67250.61140.00000.00000.96850.97060.00000.98310.00000.00000.79780.0000
abalone-17_vs_7-8-9-100.61390.49700.35920.06030.38560.36970.57190.66560.00000.00000.83200.5917
poker-8-9vs60.25650.47770.08940.08940.39790.16470.66870.71620.55390.58050.71080.8098
poker-8vs60.24850.48860.11550.00000.01000.00000.43270.70300.00000.00000.61390.8094
Average0.37570.45450.48590.43300.54850.53860.67200.77120.29770.33980.78730.6876
Average Friedman-rank8.33338.06677.66678.43336.23336.40004.56674.00009.46679.00002.93332.9000
Note: The bold data represents optimal results on each dataset, and the underlined data represents second-best results on each dataset.
Table 12. Average Performance Under Different Hyperparameter m Conditions.
Table 12. Average Performance Under Different Hyperparameter m Conditions.
mAverage F1-ScoreAverage G-Mean
100.72010.8063
200.71560.7996
300.7080.797
400.70430.7883
500.69820.7743
Table 13. Ablation Experiment Settings and Average Performance.
Table 13. Ablation Experiment Settings and Average Performance.
Average F1-ScoreAverage G-Mean
Full Model0.72010.8063
w/o Multi-resolution0.63440.7318
w/o Instance Branch0.58330.6863
w/o Feature Branch0.70130.7736
w/o Flow Model0.70960.7821
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Diao, X.; Li, Q.; Meng, Z.; Chen, T.; Lin, Y.; Hao, Y.; Gao, X. An Ensemble Imbalanced Classification Framework via Dual-Perspective Overlapping Analysis with Multi-Resolution Metrics. Electronics 2025, 14, 4740. https://doi.org/10.3390/electronics14234740

AMA Style

Li Y, Diao X, Li Q, Meng Z, Chen T, Lin Y, Hao Y, Gao X. An Ensemble Imbalanced Classification Framework via Dual-Perspective Overlapping Analysis with Multi-Resolution Metrics. Electronics. 2025; 14(23):4740. https://doi.org/10.3390/electronics14234740

Chicago/Turabian Style

Li, Yuan, Xinping Diao, Qiangwei Li, Zhihang Meng, Tianyang Chen, Yukun Lin, Yu Hao, and Xin Gao. 2025. "An Ensemble Imbalanced Classification Framework via Dual-Perspective Overlapping Analysis with Multi-Resolution Metrics" Electronics 14, no. 23: 4740. https://doi.org/10.3390/electronics14234740

APA Style

Li, Y., Diao, X., Li, Q., Meng, Z., Chen, T., Lin, Y., Hao, Y., & Gao, X. (2025). An Ensemble Imbalanced Classification Framework via Dual-Perspective Overlapping Analysis with Multi-Resolution Metrics. Electronics, 14(23), 4740. https://doi.org/10.3390/electronics14234740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop