Next Article in Journal
The Inverted Hjorth Distribution and Its Applications in Environmental and Pharmaceutical Sciences
Next Article in Special Issue
Divide-and-Merge Parallel Hierarchical Ensemble DNNs with Local Knowledge Augmentation
Previous Article in Journal
Symmetry-Aware Code Generation: Distilling Pseudocode Reasoning for Lightweight Deployment of Large Language Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TS-SMOTE: An Improved SMOTE Method Based on Symmetric Triangle Scoring Mechanism for Solving Class-Imbalanced Problems

School of Science, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(8), 1326; https://doi.org/10.3390/sym17081326
Submission received: 14 July 2025 / Revised: 1 August 2025 / Accepted: 7 August 2025 / Published: 14 August 2025
(This article belongs to the Special Issue Advances in Neural Network/Deep Learning and Symmetry/Asymmetry)

Abstract

The imbalanced classification problem is a key research in machine learning as the relevant algorithms tend to focus on the features and patterns of the majority class instead of insufficient learning of the minority class, resulting in unsatisfactory performance of machine learning. Scholars have attempted to solve this problem and proposed many ideas at the data and algorithm levels. The SMOTE (Synthetic Minority Over-sampling Technique) method is an effective approach at the data level. In this paper, we propose an oversampling method based on SMOTE and symmetric regular triangles scoring mechanism. This method uses symmetrical triangles to flatten the plane, and then establishes a suitable scoring mechanism to select the minority samples that participate in the synthesis. After selecting the minority samples, it conducts multiple linear interpolations according to the established rules to generate new minority samples. In the experimental section, we select 30 imbalanced datasets to test their performance of the proposed method and some classical oversampling methods under different indicators. In order to demonstrate the performance of these oversampling methods with classifiers, we select three different classifiers and test their performance. The experimental results show that the TS-SMOTE method has the best performance.

1. Introduction

In recent years, machine learning has become increasingly popular [1,2] and has gradually become key research. In real life, as one of the important areas of machine learning, the classification problem [3,4] can be felt in various aspects. In classification problems, we often encounter binary classification problems [5,6], and many binary classification problems in life are class imbalance problems, for example, medical diagnosis, financial fraud detection, natural disaster prediction, facial recognition, speech recognition and so on. However, when dealing with classification problems in real-world applications, it is inevitable that we encounter significant imbalances in the proportion between classes. The case is also known as imbalanced classification and has drawn considerable attention and research efforts in the industry. For example, in the employment field, there is the problem of gender minority class imbalance [7]. In certain specific industries or occupations, there is a serious imbalance in the gender ratio. The number of cases of a certain gender is relatively small, while the number of cases of the other gender is extremely large. This will lead to the traditional neural network not paying enough attention to the minority samples, thus overly learning the data of the majority samples, resulting in poor learning performance of the network. Typically, the classes that possess a smaller number of samples are referred to as minority classes. On the other hand, those classes that have a larger quantity of samples are known as majority classes. The imbalanced ratio (IR) [8] is calculated by dividing the number of majority samples by that of minority samples. We measure the degree of class imbalance through the imbalance ratio. The larger the ratio, the more imbalanced the data is. For an imbalanced dataset, the samples belonging to the class with a relatively small sample number are called minority samples. The case that is the opposite of the minority samples is called the majority samples. When dealing with the problem of class imbalance, it is undeniable that due to the significant difference in the number of majority and minority samples in the collected data, traditional classifiers have no effective means to identify minority class samples, resulting in insignificant classification results. Therefore, in order to solve this problem, many scholars have overcome difficulties and have proposed a large number of methods from the aspects of data and algorithms [9,10] to improve the recognition accuracy of classifiers. At the data level, there are two main methods: undersampling [11,12] and oversampling [13,14,15]. Undersampling refers to reducing the number of majority samples when dealing with imbalanced data to make it equal to the number of minority samples. In contrast, oversampling performs the opposite, increasing the number of minority samples to make it the same as the number of majority samples. Obviously, the method of undersampling, which reduces the number of majority samples, may lead to the deletion of some important information and features, thus affecting the model’s learning and understanding of majority class data and making it prone to the problem of overfitting. Oversampling, by contrast, preserves more information and represents a more scientific approach. Currently, the most widely used oversampling method is the Synthetic Minority Over-sampling Technique (SMOTE) [16,17,18,19].
The operation of SMOTE is as follows: First, select a minority sample. Then, randomly pick a minority sample from the k-nearest neighbors of this selected sample. After linear interpolation is employed between these two samples, then generate a new minority sample (see Figure 1A). Repeat these processes continuously until the number of samples of the majority class and the minority class is balanced or meets the set proportion. The samples synthesized by this method are generally located in the region of the minority class. It can effectively expand the minority samples in the feature space, assist the network model in learning more abundant feature information, and improve the classification performance of the model for minority samples.
Subsequently, numerous scholars carried out optimization work on the SMOTE algorithm. With the booming development of the fields of deep learning and neural networks, an increasing number of SMOTE-based models have started to integrate with these cutting-edge technologies. Cost-sensitive learning focuses on algorithm-level optimization, adjusting the model’s learning focus by setting different misclassification costs for different classes. For instance, the cost of misclassifying minority class samples is set much higher than that of the majority class, prompting the model to pay more attention to minority samples during training. This method does not require altering the original data distribution and is highly suitable for scenarios where samples are scarce and generating new samples is difficult, such as precious case data in medical diagnosis. Anomaly-detection-based classifiers adopt a different approach, treating minority class samples as “anomalies” and achieving classification by modeling the distribution of majority class samples. This method has certain applicability in extremely imbalanced scenarios, that is, when the proportion of minority class samples is extremely low (usually less than 1%). AutoSMOTE [20] proposes a method for achieving automated imbalanced learning based on deep hierarchical reinforcement learning. This method is capable of automatically selecting appropriate sampling and ensemble strategies to enhance the classification performance on imbalanced data, thereby realizing an automated imbalanced learning process. In addition, some scholars have combined generative adaptive networks (GANs) with oversampling techniques to develop new SMOTE methods. The latest research achievements include methods such as SMOTified-GAN [21] and GAN-SMOTE [22]. In addition, there are many other methods based on boosting or ensemble learning techniques [23], such as SMOTEBoost [24] and Ensemble- SMOTE [25]. The concept of ensemble learning holds significant importance in cost-sensitive learning. A prime illustration of this is AdaCost [26], which is founded on AdaBoost. It is crucial to note that ensemble learning does not inherently define an algorithm’s unique concept; Rather, it serves as a versatile approach that can be integrated with nearly all algorithms. Moreover, undersampling presents a notable drawback as it discards a substantial amount of valuable data, potentially leading to the loss of important information. On the other hand, generative adaptive networks (GANs) come with their own set of challenges. They are relatively more intricate in their implementation, which can limit their reliability. In contrast, oversampling methods based on the SMOTE tend to offer the most promising results. SMOTE-based approaches strike a balance between preserving data integrity and enhancing the performance of models dealing with imbalanced datasets, making them a preferred choice in many practical applications.
Although SMOTE is an effective method, it still has some problems. Firstly, the selection of k-nearest neighbors in SMOTE is blind [27]. The lower limit of the parameter k is 1, but there is no fixed upper limit value for different datasets.
In the SMOTE algorithm, k is usually set to 5, which is a common practice. This has a certain degree of rationality, but there are also corresponding limitations. For some datasets with complex data distributions, multiple sub-clusters, or significant changes in data density, k = 5 may not be applicable. Especially when the dataset is at small scale or the number of minority samples is negligible, SMOTE can give rise to sample overlap because of repeated sampling. This situation will further magnify the noise problem. For example, in a sparse sub-cluster of data, five nearest neighbors may not be sufficient to fully represent the characteristics of that region. On the other hand, in a data-dense area, five nearest neighbors may be excessive, resulting in synthetic samples that cannot accurately reflect the local structure [28].
In specific experiments, due to the different characteristics of various datasets, a fixed k value of 5 obviously cannot meet the requirements of different datasets. Therefore, we need to continuously adjust the value of k and ensure that the k value is optimized to yield the best experimental results. For example, if unreasonable synthetic samples are generated near the class boundaries, it may make it difficult for the classifier to accurately distinguish between different classes, thus leading to a decline in classification performance. As a result, the boundary between the majority samples and the minority samples becomes increasingly blurred [29]. Another crucial problem is that it does not consider the distribution patterns of both the samples and their neighboring points. Instead, it merely performs an average sampling operation on the points of the minority class. This approach renders it highly susceptible to causing overfitting [30].
Considering all the above problems, we present three common issues that arise during the synthesis of new samples by SMOTE, and will explain them with the aid of images. In the first situation, for a certain noisy point x, if the selected k-nearest neighbor point x is within the region of minority samples (see Figure 1B), then the newly synthesized sample x n e w is highly likely to be situated near the boundary, which is unfavorable for the subsequent learning. In the second case, both x and x are noisy points. Then, the line segment connecting them lies within the boundary of the majority samples. This will cause the newly generated sample x n e w to also be a noisy point (see Figure 1C), resulting in a poor generation effect. In the third case, although both x and x are located in the minority region, due to the excessively close distance between these two points, the newly generated sample x n e w almost coincides with the original two samples. Even though this sample is indeed synthesized in the minority region (see Figure 1D), it does not provide much help for our subsequent analysis. Moreover, it may lead to overfitting during the learning process.
Based on the above-mentioned deficiencies of the SMOTE method, we propose an improved SMOTE method (TS-CMOTE) based on regular triangle scoring mechanism. Firstly, the PCA dimensionality reduction technique is used to reduce multidimensional samples to a two-dimensional space. Then, according to the obtained new data, an appropriate side length of the regular symmetric triangle is calculated, and they are closely connected in the two-dimensional space in the form of tessellation. After that, we establish a regular triangle scoring mechanism to dynamically select samples. Finally, the multidimensional samples corresponding to the selected two-dimensional samples are found and used to participate in the synthesis of minority class samples. By calculating and limiting the range of the side length of the regular triangle, the TS-SMOTE method can effectively reduce the possibility of synthesizing new noise samples. In addition, instead of performing linear interpolation between two samples as in SMOTE, TS-SMOTE selects multiple samples simultaneously and synthesizes a new sample within the closed region of these samples. This expands the sample selection and can enrich the information content of the new samples.
The symmetric triangle scoring mechanism has an outstanding value. Our method fully takes into account the impact brought by the distribution of the majority samples. By considering different categories and combining it with our method, we can generate minority samples reasonably. In particular, because our method can effectively reduce the generation of noise points, the generated points at the boundaries are also more reasonable.
We conducted experiments using 30 imbalanced datasets from the website https://sci2s.ugr.es/keel/imbalanced.php, accessed on 4 March 2025. The experimental results show that compared with the original SMOTE method and the improved SMOTE methods, the network model trained by the TS-SMOTE method has better classification performance. In addition, to prevent the uncertainty of the classifier, we tested the performance of each oversampling method on three commonly used classifiers: multilayer perceptron (MLP) [31], support vector machine (SVM) [32], and Adaptive Boosting (AdaBoost) [33]. By comparing the experimental results, it can be seen that TS-SMOTE can be effectively applied to solve the class imbalance problem.
The remaining content of this paper is arranged as follows. In Section 2, some related work will be briefly introduced first. Section 3 will provide a detailed introduction to the TS-SMOTE method. In Section 4, data experiments and corresponding analyses will be conducted. Some conclusions will be drawn in Section 5.

2. Related Works

Regarding the decision-making methods for handling imbalanced classification problems, we have introduced many methods previously. However, since our research mainly focuses on SMOTE and the improved SMOTE, we will no longer cover other non-SMOTE methods. Instead, we will concentrate our efforts on the discussion of SMOTE and the improved SMOTE.
The Synthetic Minority Over-sampling Technique (SMOTE) is a method for oversampling minority classes. It was proposed by Nitesh V. Chawla and other researchers in 2002 [16]. This method is an oversampling algorithm specifically designed for imbalanced datasets.
The specific method is as follows:
  • Determine the minority class sample set S m i n o r i t y . For each sample x i S m i n o r i t y in it, calculate and find its k nearest neighbors through methods such as the Euclidean distance.
  • For each minority class sample x i , randomly select a nearest neighbor sample x j from its k nearest neighbors.
  • Synthesize a new sample through the formula x n e w = x i + λ × ( x j x i ) , where λ [ 0 , 1 ] is a random number.
  • Repeat the above steps to generate a sufficient number of minority class samples, so as to balance the class distribution of the dataset and solve the problem of data imbalance.
In the previous chapter, we pointed out some problems existing in SMOTE, such as blurred boundaries, sample overlap, noise expansion, overfitting, and the blindness of k-nearest neighbors, among others (see Figure 1). Subsequent scholars have continuously proposed new improved methods of SMOTE. Borderline-SMOTE [34] is an improved algorithm of SMOTE, proposed by Han et al., and it is used to deal with imbalanced datasets. The traditional SMOTE oversamples all minority class samples, which is likely to lead to data overlap and overfitting. Borderline-SMOTE focuses on the boundary samples of the minority class. First, the minority class samples are classified: for a sample x, its k nearest neighbors are found. If the number of majority class samples among them exceeds half, x is a dangerous sample; if the number of majority class samples is 0, x is a safe sample; and the rest are uncertain samples. This algorithm only oversamples the dangerous samples. For a dangerous sample x, a sample y is selected from the minority class samples among its k nearest neighbors, and a new sample is generated according to the formula x n e w = x + λ × ( y x ) ( λ [ 0 , 1 ] is a random number). In this way, the new samples are closer to the classification boundary and can better reflect the class distribution relationship. When dealing with imbalanced data, compared with SMOTE, it can reduce overfitting and improve the classification performance.
The blindness of the SMOTE algorithm has been a long-standing issue. Some algorithms place more emphasis on how to select and generate points. MWMOTE (Majority Weighted Minority Oversampling Technique) [35] is used to solve the problem of data imbalance. MWMOTE can effectively adjust the data distribution, making the generated minority class samples more representative. The models trained on imbalanced datasets can obtain better classification performance and generalization ability. DTO-SMOTE [36] is an improved data oversampling method. It combines density and topological structure information to process minority class samples. Therefore, it can more reasonably increase the number of minority class samples, improve the data imbalance situation, and enhance the performance of models trained on imbalanced data. Safe-Level SMOTE [37] is a technique used to deal with data class imbalance. It generates samples based on the information within the safe region, which can better improve the imbalance and enhance the model’s ability to recognize the minority classes. SMOTE-Tomek Links [38] is an algorithm that combines SMOTE and Tomek Links, and it is used to deal with the problem of class imbalance in data. This algorithm first uses SMOTE to oversample the minority classes to increase the number of their samples, and then removes those sample pairs that may be confusing at the class boundaries (usually removing the majority class samples) through Tomek Links, so as to clean the data, reduce noise interference, optimize the distribution of the dataset, and improve the classification performance of the model for the minority classes. SASMOTE (Self-Inspected Adaptive SMOTE) [39] is an algorithm used to address the problem of highly imbalanced data classification. This algorithm uses an adaptive nearest neighbor selection algorithm to identify “visible” nearest neighbors for generating minority class samples, improving sample quality. At the same time, it introduces a self-inspection uncertainty elimination method to filter out low-quality synthetic samples that are difficult to distinguish from the majority class. SMOTE-ENN [40] is an algorithm used to address the problem of data imbalance. It combines SMOTE and ENN (Edited Nearest Neighbors). SMOTE generates new samples by interpolating between the neighbors of minority class samples to expand the number of minority class samples. ENN, based on the nearest neighbor rule, removes noisy samples from the dataset, that is, those samples whose class is different from the majority of their neighbors. SMOTE-ENN first uses SMOTE to oversample the minority class to increase its sample size, and then applies ENN to clean the oversampled data, removing noisy and misclassified samples, optimizing the dataset, and enhancing the classification performance of the model on imbalanced data. KNNOR (K-Nearest Neighbor Oversampling Approach) [13], that is, the K-nearest neighbor oversampling method, is an oversampling technique for dealing with imbalanced datasets. Based on the SMOTE algorithm, it determines the key and safe enhancement regions of minority class samples through a three-step process and generates synthetic data points. When generating artificial points, it takes into account the relative density of the entire dataset, enabling more reliable oversampling of the minority class and having stronger robustness to noise.
Simultaneously, due to the overfitting problem resulting from randomness, a greater number of algorithms focus on researching the distribution of samples. ADASYN (Adaptive Synthetic Sampling Approach) [41] is an adaptive synthetic sampling method and an oversampling technique for dealing with imbalanced datasets. It adaptively determines the number of synthetic samples to be generated for each minority class sample based on the local density of the minority class samples. In this way, the distribution of minority class samples can be made more reasonable, effectively alleviating the problem of data imbalance, enabling the learning algorithm to better learn the features of minority class samples during training, and improving the classification performance of the model on imbalanced data. Gaussian-SMOTE [42] is an algorithm for handling the problem of data imbalance. It combines the Gaussian distribution with SMOTE. Based on SMOTE, it introduces the Gaussian distribution to generate new samples. Gaussian-SMOTE utilizes the characteristics of the Gaussian distribution to make the newly generated samples more diverse and reasonable. This algorithm first determines the k-nearest neighbors of minority class samples, and then generates new samples between the samples and their neighbors according to the Gaussian distribution, expanding the number of minority class samples, alleviating data imbalance, and effectively improving the classification performance of the model on imbalanced data. Geometric SMOTE (Geometric Synthetic Minority Over-sampling Technique) [43] is an improved method of the traditional SMOTE. Based on geometric principles, it takes into account the spatial distribution patterns of minority class samples. When generating new samples, instead of simply performing linear interpolation between minority class samples and their nearest neighbors, it determines a more reasonable position for sample generation according to the geometric structure. For example, geometric figures formed by sample points (such as convex hulls, etc.) are used to restrict the generation area, making the new samples more consistent with the inherent geometric characteristics of the minority class data. This can more effectively increase the number of minority class samples, improve the distribution of imbalanced datasets, and enhance the model’s ability to learn and classify minority class samples.
The original purpose of most of the algorithms described above is to solve the balance between classes. However, the uneven distribution of similar samples in the feature space will also have an impact on the classification task, which is referred to as intra-class imbalance. ADASYN [41], namely, the Adaptive Synthetic Sampling Approach, is used to deal with imbalanced datasets. It improves the data distribution by generating samples of the minority class so as to enhance the performance of the model. Some scholars have also combined the clustering algorithm with SMOTE to reduce the intra-class imbalance. K-means SMOTE [44] is a method that combines the K-means clustering algorithm and SMOTE. SMOTE is an oversampling algorithm. By analyzing the feature space of minority class samples, it synthesizes new samples around them to alleviate the problem of data imbalance. However, it does not take into account the sample distribution density. K-means is a commonly used clustering algorithm that can divide a dataset into K clusters, making the samples within a cluster highly similar and the differences between clusters large. K-means SMOTE first uses K-means to cluster the minority class samples to obtain different clusters. Then, SMOTE is applied to each cluster separately to generate new minority class samples within the clusters. In this way, the generated samples are more in line with the distribution characteristics of the samples within each cluster, and it can more reasonably expand the number of minority class samples, avoiding the problem of unreasonable sample distribution that may occur when SMOTE is simply applied. When dealing with imbalanced datasets, it can improve the model’s ability to recognize the minority class and enhance the overall performance of the model.
Other non-SMOTE algorithms have also been further developed. GB-SMOTE [45] is used to solve the problem of imbalanced data classification. Firstly, it uses the slack variables of Support Vector Machine (SVM) to divide the minority class samples into a misclassification set, a margin set, and a correctly classified set. When selecting samples, samples are, respectively, selected from the margin set and the correctly classified set, and the selection is carried out according to the weights determined by the distances between the samples and the decision hyperplane. In the sample generation stage, the occurrence frequency of sample pairs and the distances in the feature space are calculated. The distances are evenly divided into corresponding sub-segments, and the midpoints are taken as new samples. NBG [46] is used to solve the problem of imbalanced data classification. It combines the NPSMOTE oversampling algorithm, the BALO feature selection algorithm, and the GVM classification algorithm. NPSMOTE generates effective positive class samples by removing noisy samples, assigning weights, and limiting distances. BALO can adaptively search the feature space and select important features. GVM has strong generalization ability. Experiments show that when dealing with the classification of imbalanced small-sample datasets, the NBG algorithm has significant advantages compared with a variety of existing algorithms and can effectively improve the recognition rate of the minority class. In practical applications, the performance of non-SMOTE algorithms is inferior to that of the variants of SMOTE.

3. Specific Methods of the Improved Symmetric Triangle Scoring Mechanism

3.1. Detailed Introduction of Methods

In the previous sections, for different oversampling methods, we explained their basic ideas. SMOTE is the most fundamental algorithm among them. However, SMOTE still has many drawbacks. For example, the selection of new samples for SMOTE is randomly selected between the line of two minority samples with equal possibility. This will lead to the generation of many noise samples. In addition, SMOTE has problems with samples overlapping and overgeneralization, these issues will continue to expand the scope of the original noise. Regarding the above-mentioned issues with SMOTE, many scholars have proposed improvement methods based on SMOTE. The MWMOTE provided us with some ideas for our work. We mainly borrowed the scoring weights and selection probabilities from MWMOTE to distinguish different samples. For linear interpolation in Borderline-SMOTE, we discard this simple method, choosing synthesize samples in the geometric area. In summary, T-SMOTE is an oversampling method based on the regular triangle scoring mechanism, used for multiple dynamic interpolation. The detailed introduction is as follows.
  • Step 1. Multidimensional data is projected onto a 2D plane via PCA.
PCA (principal component analysis [47], as a commonly used dimensionality reduction technique, plays an important role in reducing dimensions from imbalanced multidimensional datasets to two-dimensional data.
In this tessellation problem, we choose triangles to cover a plane. For a regular triangle, there are three surrounding regular triangles with sides connected and nine regular triangles angles connected. When scoring, we will select different coefficients to calculate the score of this regular triangle area.
Nevertheless, discovering a regular triangular region that yields consistent effects across various high-dimensional spaces is arduous. Thus, we plan to make use of PCA to obtain the first two-dimensional features of the data. In an effort to minimize the losses resulting from PCA, the processed data is only utilized for the purpose of sample selection. From this perspective, PCA can mitigate noise to a certain extent as well. Finally, the creation of new samples remains in the original dimension.
  • Step 2. Pave the entire two-dimensional plane with regular triangles.
In the two-dimensional plane, we choose regular triangles and pave the entire two-dimensional plane according to the method of tessellation. Now we will place the dimensionality reduction data obtained in step 1 into their respective triangles. However, at this time, we have not determined the side lengths of the triangle, so we need to determine some parameters to determine the side length of the triangle. Since the distance in the data points is easy to measure, we use the Euclidean distance [48] as the measure of distance. For each regular triangle, we assign it a field G. On this two-dimensional plane, we record the entire sample set as X = { x 1 , x 2 , , x n } , where n is the number of samples in the entire dataset, and we record the minority samples as X m i n = { x 1 m i n , x 2 m i n , , x m m i n } , where m is the number of minority samples in the entire dataset. We define the distance between two points a and b as d i s t a n c e ( a , b ) ; the coordinate of a is ( a 1 , a 2 ) , The coordinate of b is ( b 1 , b 2 ) .
d i s t a n c e ( a , b ) = ( a 1 b 1 ) 2 + ( a 2 b 2 ) 2 .
Thus far, we define the following four distances for measuring the length of the sides of a triangle.
d ¯ m i n = i = 1 m j = 1 m distance ( x i m i n , x j m i n ) m · ( m 1 ) , j i and j , i = 1 , 2 , , m , d ¯ m i n m i n = i = 1 m min ( distance ( x i m i n , x j m i n ) ) m , j i and j , i = 1 , 2 , , m , d ¯ = i = 1 n j = 1 n distance ( x i , x j ) n · ( n 1 ) , j i and j , i = 1 , 2 , , n , d ¯ m i n = i = 1 n min ( distance ( x i , x j ) ) n , j i and j , i = 1 , 2 , , n .
d ¯ m i n is used to calculate the average distance between pairs of minority samples, d ¯ m i n m i n represents the mean of the minimum distances from every minority sample to the other minority samples, d ¯ indicates the average of the distances that each sample has to other samples, and d ¯ m i n represents the average value of the minimum distances from each individual sample to the rest of the samples. The side length d G of field G, at this point in time, ought to conform to the subsequent requirements:
d G = 3 a d ¯ m i n .
In this case, a represents the adaptation factor for the side length and complies with the subsequent requirements.
d ¯ min d ¯ a d ¯ min min d ¯ min .
Through the above equation, we can obtain a relatively suitable size for the side length of a triangle. For the side length of a triangle, if the side length is too large, it will result in all points being within certain regions. If the side length is too small, there will be too few points in the field, losing the meaning of dividing the field.
Note: For rare cases where the sample may fall on the edge of a triangle, the sample is considered as noise and does not participate in the synthesis.
  • Step 3. Rules for naming and assimilation.
  • Rules for naming.
Based on the above rules, we can determine the side length of a triangle. The following steps require us to establish some rules to determine the type of field for triangles, which can serve as a prerequisite for our scoring mechanism. Our rule is based on the number of minority samples and majority samples for classification.
For a regular triangle, it has a total of 12 adjacent triangles. Three adjacent triangles are connected by their edges, while the remaining nine triangles are connected by their vertices.
As shown in Figure 2, Figure 2a shows the situation of a regular triangle and its 12 adjacent regular triangles, which make up a symmetrical shape.
Figure 2b shows the first situation, where three triangles are connected to the sides of a regular triangle, with the red triangle representing the initial triangle and the blue triangle representing the three triangles connected to its sides. Obviously, these three blue triangles are also symmetrical to each other.
Figure 2c shows nine triangles connected by vertices, where the green triangles are what we call triangles connected by vertices, which is the second situation. It is easy to see that these nine green triangles are also symmetrical about their centers (red triangle).
After introducing the above rules, we explain the meanings of the following symbols:
  • T: Each regular triangles is recorded as a field T.
  • NGB T : The remaining 12 regular triangles adjacent to field T are called neighbors of field T and recorded as NGB T . They include three triangles with connected sides and nine triangles with adjacent vertices.
  • NGB T 1 : Triangles connected to a regular triangle by their sides.
  • NGB T 2 : Triangles connected to a regular triangle by their vertices.
  • T m i n : A triangle containing only minority samples.
  • T m a j : A triangle containing only majority samples.
  • T e m p t y : A triangle that does not contain any samples.
  • T d e b : A triangle that contains both majority and minority samples, which requires further debate.
  • Rules for Assimilation.
For an empty field T e m p t y , when all of its neighboring fields NGB T e m p t y belong to T G , where T G denotes any type of field except for T d e b , if T e m p t y is not an element of T G , an assimilation process is initiated to make T e m p t y an element of T G . The assimilation rules for T m i n and T m a j are identical to those stated above (see Figure 3). Nevertheless, for the controversial field T d e b , if all of its neighboring fields NGB T d e b are elements of T D (where D = T m i n or D = T m a j ), there exists noise in T d e b , and the noisy samples are either majority samples or minority samples within T d e b . Ultimately, T d e b in T D is achieved via the assimilation operation.
  • Step 4. Obtain marking mechanism.
After following the above steps, we will conduct the corresponding marking mechanism. Based on our previous steps, we have a premise: only T m i n and T d e b that have not been assimilated can be selected for sampling, and the fields have been assimilated merely to take part in the scoring and weighting processes. Subsequently, we assign different scores based on the types of fields selected. For a regular triangle, our dense tiling method will have 12 triangles around it. We set a selected field as T o n e , as his neighbor N G B T o n e . We have 12 choices. Let us consider these neighbors and T o n e as T m , then m = 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 , 12 , 13 . Below are the specific rules.
(1)
The overall score is 18 points, and the score assigned to T m is represented as mark ( T m ) .
(2)
If T m T m i n and they are connected by side, called T m i n s , then mark ( T m ) = 1.9 .
(3)
If T m T m i n and they are connected by vertex, called T m i n v , then mark ( T m ) = 0.45 .
(4)
If T m T d e b and they are connected by side, called T d e b s , then mark ( T m ) = 0.6 .
(5)
If T m T d e b and they are connected by vertex, called T d e b v , then mark ( T m ) = 0.15 .
(6)
For itself, it is calculated according to the highest score.
(7)
If T m { T e m p t y , T m a j } then mark ( T m ) = 0 .
Therefore, according to the scoring rules, the total score of the T m is
Point ( T one ) = m = 1 13 Point ( T m ) .
If the total score is less than 3 points, then this field does not participate in sampling, expressed as follows:
Point ( T one ) = 0 .
This means that the field will not participate in the rating and will be simplified as an empty field.
Then, the probability formula of T m is expressed as
point ( T one ) × b point ( T one ) 13 × b 13 ,
where b is the boundary-weight factor; taking b < 1 ( b > 1 ) can strengthen (weaken) the selection frequency of boundary samples. In order to avoid the situation that our sample population is not just boundary samples, the value of b in this paper is 1.
After normalization, the selection probability of T o n e is
P select = p p = point ( T one ) × b Point ( T one ) 13 × b 13 i = 1 K point ( T one i ) × b Point ( T one i ) 13 × b 13 ,
where K represents the total quantity of all T o n e fields. Figure 4 provides an example of a specific scoring mechanism.
  • Step 5. Sampling and synthesizing new samples.
We need to randomly select a field as T o n e from the unadulterated T m i n and T d e b based on the selection probability formula. Afterwards, we will observe neighbors of T o n e : N G B T o n e , check valid T m i n and T d e b in N G B T o n e . If they exist, we record the sum of the quantities of T m i n and T d e b as m. If the quantity is zero, we discard it.
As shown in Figure 5, this is our rough diagram, first selecting samples and then generating new samples. After selecting T o n e , we randomly select a minority sample from it. Similarly, we select a minority sample from each N G B T o n e that meets the criteria. Let us peel off these points. We hope that the sum of the generated minority samples and the original minority samples will balance with the majority class samples. Next, let us peel off these points. We select m + 1 sample, and let them be a 0 , a 1 , a 2 , …, a m . In the line connecting a 0 and  a 1 , we generate A 1 using linear interpolation method. Then, we continue to generate A 2  using linear interpolation in the line connecting A 1 and a 2 . Similarly, we will generate the final sample A m on the line connecting A m 1 and a m . Letting r be a random number between 0 and 1, K = 1 , 2 , , m , we have the following synthesis formula:
A K = A K 1 + r ( a K A K 1 ) .
As shown in Figure 5, the points generated are ultimately equivalent to conducting random sampling. This sampling occurs within the closed area formed by the selected points. Overall, different partitioning methods have different selection probabilities, and the final synthesis situation is also different.
Note: To ensure that the data information is not distorted, we will not directly interpolate linearly in two-dimensional space. The specific approach is as follows. First, we select the reduced dimensional samples based on scores and probabilities; secondly, we find the high-dimensional samples corresponding to these samples before dimensionality reduction; finally, linear interpolation is applied to high-dimensional samples. In this way, we can ensure that the information of multidimensional samples is not lost.
  • TS-SMOTE algorithm
To sum up, the TS-SMOTE algorithm (Algorithm 1) process is as follows:
Algorithm 1 TS-SMOTE
 1:
Input: Total sample set S, minority sample set S m i n , total number of samples N, total number of minority samples m.
 2:
Output: Synthetic balanced dataset.
 3:
if  m = = ( N m )  then
 4:
    Print “The original sample dataset is balanced.”
 5:
    Break
 6:
else
 7:
    Step.1: Multidimensional data is projected onto a 2D plane via PCA
 8:
    for  S i , S j m i n  do
 9:
        By PCA, X i S i , X j m i n S j m i n , where i = 1 , 2 , . . . , N , j = 1 , 2 , . . . , m ;
10:
    end for
11:
    Step.2: Pave the entire two-dimensional plane with regular triangles
12:
     a X i , X j m i n following the relevant Formulas (1) and (3);
13:
     d G a , X i , X j m i n following the relevant Formula (2);
14:
    Obtain all the fields, G t d G , where t = 1 , 2 , . . . , n , and n is the number of fields;
15:
    Step.3: Rules for Naming and Assimilation
16:
    for  i = 1 : n  do
17:
        if  X i T t  then
18:
           remove X i ;
19:
        end if
20:
    end for
21:
    for  t = 1 : n  do
22:
        Name T t ;
23:
        Assimilate T t ;
24:
        Collect all T o n e ;
25:
    end for
26:
    Step.4: Obtain marking mechanism
27:
    for  T o n e  do
28:
         P o i n t ( T o n e ) T o n e following the scoring rules and the associated Formulas (4) and (5);
29:
         P s e l e c t P o i n t ( T o n e ) according to Formulas (6) and (7);
30:
    end for
31:
    Step.5: Sampling and synthesis
32:
    while  m N m  do
33:
        Choose one randomly T o n e P s e l e c t ;
34:
        Obtain the sampling neighborhood N G B T o n e T o n e ;
35:
         M the total number of T o n e and N G B T o n e ;
36:
        Randomly choose X M m i n T o n e , N G B T o n e ;
37:
         S M m i n X M m i n ;
38:
         N e w S a m p l e S M m i n following the Formula (8);
39:
         m m + 1 ;
40:
    end while
41:
    Print dataset { N e w S a m p l e } , { S , N e w S a m p l e } ;
42:
end if

3.2. Discussion of Parameters

3.2.1. Discussion on Factor a

In the previous discussion, we examined the effects of factor a being either too large or too small and confined the edge adaptation factor a to a relatively appropriate range. Then, within this known range, we will discuss the optimal value of a from the perspectives of performance and time complexity.
Let us consider a dataset containing n samples with an imbalance ratio of r. The number of minority samples can then be expressed as n / ( r + 1 ) .
Given that each regular triangle has a side length of a, the space can accommodate i triangles horizontally and j triangles vertically. Thus, we can approximate the required number of fields to construct as m = i j .
The original SMOTE algorithm requires iterating through all minority-class samples and searching for their k-nearest neighbors among all samples, resulting in a time complexity of O ( n n / ( r + 1 ) ) = O ( n 2 ) .
Our new method differs from the original SMOTE approach primarily in three aspects: application of PCA dimensionality reduction, selection criteria for data points, and strategy for generating new samples. Next, we will analyze the time complexity of each component. First, we observe that traditional PCA requires full eigendecomposition of an n × n matrix, resulting in O ( n 3 ) time complexity. Based on principal component analysis theory, the dimensionality reduction process requires us to project the dataset onto the first q principal components. This involves computing the top q eigenvalues and eigenvectors, yielding an overall time complexity of O ( q · n 2 ) [49]. In our preliminary work, we specifically reduce the data to two dimensions ( q = 2 ). Since two is negligible compared to n, the effective time complexity of PCA in our case becomes O ( n 2 ) . Next, we analyze the time complexity for the triangular tessellation space. Based on our previous work, we have determined the length i and width j of the triangular tessellation space. We generate approximately m = i × j regular triangles, resulting in a time complexity of O ( i × j × m ) = O ( m 2 ) . Subsequently, according to our defined rules, each triangular region requires four operations: assimilation, scoring, probability calculation, and synthesis. Since these operations involve traversing both the current region and its 12 adjacent triangular regions, the time complexity becomes O ( 13 × m ) = O ( m ) .
Finally, when we have a grasp of the sample and the divided regions, we need to match them one by one. Since the number of samples is n and the number of regions is approximately m, considering the matching between samples and regions, it is necessary to traverse the samples and regions. We can obtain a time complexity of O ( n m ) .
In summary, the overall time complexity of our algorithm is dominated by the biggest component, which can be expressed as O ( max ( n 2 , m 2 , n × m ) ) . We should select the smallest possible value of m with m < n to ensure the optimal performance of the algorithm. Notably, the smaller the value of m is, the larger the side length of the triangle needs to be, which in turn indicates that the value of factor a should also increase accordingly. At this point, the difference in time complexity between our algorithm and SMOTE is not pronounced.
Based on the above analysis, when the performance remains consistent, a larger value of factor a corresponds to a lower time complexity. Therefore, in the experimental section of Section 4, we choose the maximum value of a for our experiments.

3.2.2. Discussion on Factor b

We define b as the boundary weight factor because it significantly influences the proportion of new sample generation at the boundary. The scores of boundary regions are relatively lower compared to other regions. Moreover, all regions participating in the synthesis of new samples have scores no less than 3. In the probability calculation Formula (8), the scores of these regions serve as the exponent of b. Therefore, when the value of b is less than 1, the smaller b is, and the higher the probability of selecting from the lower-scoring boundary regions. In this case, we can ensure that more samples are generated near the boundary. This strategy is similar to the concept of borderline-SMOTE and helps to obtain clearer boundaries. When the value of b is greater than 1, the larger b is, the lower the probability of selecting boundary regions, thus ensuring that more generated samples are distributed away from the boundary. In this paper, b is set to 1. Since we aim for a more universal approach and do not overly focus on the generation of boundary samples, we expect to achieve a compromise.

4. Specific Settings of the Experiment and Comparative Analysis

After establishing the above rules, we need to conduct experiments to analyze the effectiveness of this method. We compare with other methods to demonstrate the advantages of TS-SMOTE. We first introduce the experimental setup, which mainly includes the dataset, classifiers, and comparison parameters. In the following chapters, we will present the experimental results and the analysis.

4.1. Related Methods and Experimental Settings

The dataset for this experiment mainly comes from the imbalanced dataset in the KEEL dataset. We selected 30 imbalanced datasets among them. Table 1 shows the specific information of these datasets.
For SMOTE, there are many improved methods. Among these methods, we choose ADASYN, G-SMOTE, MWMOTE, GAUSSIAN-SMOTE, DTO-SMOTE, KNNOR-SMOTE, and SMOTE to conduct experimental comparisons. We simplify the class-imbalanced problem by using synthesized minority samples transformed into a conventional binary classification problem. This can make the experiment more convenient and facilitate our further analysis. In the experiments, multi-layer perceptron (MLP) [50] networks, support vector machines (SVM) [51], and Adaptive Boosting (AdaBoost) [33] are used to assess the rationality of the synthesized samples under diverse methods.
For TS-SMOTE and other methods, we need some standards in the experiment to measure their performance and help us make better comparisons. Here is the confusion matrix (shown in Figure 6) for binary classification, along with the formula representations of A c c u r a c y , P r e c i s i o n , a n d R e c a l l [52]:
Accuracy = T P + T N T P + T N + F P + F N , Precision = T P T P + F P , Recall = T P T P + F N .
However, among the three evaluation indicators stated earlier, accuracy predominantly emphasizes the majority samples. In contrast, precision and recall place greater emphasis on the performance aspects of the minority samples. Therefore, in order to evaluate the ability of comprehensive classification, we need other indicators to measure [53].
Here are the three additional metrics in detail.
G-mean (geometric mean) is a comprehensive metric that takes into account both the precision and recall of a classifier. It is often used to evaluate the performance of binary classification models with imbalanced positive and negative samples. A larger G-mean indicates better model performance.
G m e a n = T P T P + F N T N T N + F P .
The F 1 m e a s u r e is an indicator used to evaluate classification problems. In many multi-classification machine learning problems, the F 1 m e a s u r e is often chosen as the final evaluation metric. It is the harmonic mean of precision and recall, with a value range from 0 to 1, where 1 is the maximum and 0 is the minimum.
F 1 m e a s u r e = 2 Precision Precision + Recall .
AUC stands for area under the curve. Here, the curve denotes the ROC curve, and AUC represents the area beneath the ROC curve. The area ranges from 0.1 to 1. As a number, AUC (area under the curve) can intuitively evaluate the advantages and disadvantages of the classifier. The larger the value, the better the performance of the classifier.
To solve various class imbalance problems, we use the five-fold cross-validation method. That is, the dataset is randomly divided into five parts with equal or similar numbers; one part is selected as the test sample each time, and the other four parts serve as the training sample. By transforming the part of the test set, five results are obtained. The process was then repeated 20 times to obtain 100 sets of results. Finally, the average of these 100 groups of results was calculated for subsequent comparison. Using average value for comparison can avoid the extreme results in a single experiment and make the comparison conclusion more convincing.

4.2. Experimental Results and Comparative Analysis

We set up three sections to describe the our experiment. In part one, we analyze the results of our experiment. The effect of this part is to show the advantages of the TS-SMOTE. In part two, we want to make the data more intuitive, so we list the images of data visualization and related experimental data. In part three, we use some statistical methods to test the significant effect of our method.

4.2.1. Comparison and Analysis of Various Metric Evaluations

After we finished our TS-SMOTE method, we chose seven different oversamling methods for comparison. They are, respectively, ADASYN, BORDERLINE, MWMOTE, GAUSSIAN-SMOTE, DTO-SMOTE, KNNOR-SMOTE, and SMOTE. In the table, we use A, B, M, G, D, K, and S as substitutes. The TS-SMOTE method uses T instead. Through these eight methods, we obtain the oversampled data. These datasets have been processed and are now balanced datasets. Next, our focus will be on detecting which method holds the advantage. To avoid potential anomalies caused by a single classifier, we selected three classifiers: MLP, SVM, and AdaBoost. We choose 30 different datasets, and the related results are shown in Table 2, Table 3 and Table 4.
In the previous section, we selected three evaluation metrics: F1-score, G-mean, and AUC. In the 30 datasets under 3 classifiers, we conducted 90 sets of experiments.
Under the three classifiers (MLP, SVM, and AdaBoost), with 90 rankings per classifier (3 metrics * 30 datasets), our TS-SMOTE method achieved first place 53 times for MLP, 60 times for SVM, and 44 times for AdaBoost. The results confirm TS-SMOTE’s unequivocal superiority. The data demonstrates that under the SVM classifier, our TS-SMOTE method achieved the strongest performance compared to the other two classifiers.
To provide a clearer visualization of the rankings, we calculated the average rank of each method across different classifiers under various evaluation metrics, as presented in Table 5. Across all three classifiers (MLP, SVM, AdaBoost) and three evaluation metrics (F1-score, G-mean, AUC), TS-SMOTE consistently ranked first in all nine comparative assessments. TS-SMOTE secured first place in all three classifiers’ aggregated rankings (shown in Table 6), demonstrating consistent dominance across MLP, SVM, and AdaBoost. Furthermore, we provide violin-box plots to comprehensively visualize the comparative performance of all methods. First, let us briefly introduce the violin-box plot. Violin plots reveal the probability density of evaluation scores, and box plots superimposed on violins quantify quartile statistics. Among them, the solid lines from top to bottom represent the maximum, upper quartile, median, lower quartile, and minimum values, respectively. We also computed the mean values of TS-SMOTE (shown in Table 7), demonstrating that TS-SMOTE achieves the highest average scores across all groups. The violin plot in Figure 7 reveals that TS-SMOTE exhibits superior stability, with its performance metrics consistently clustering in the higher value range, thereby confirming its exceptional robustness. This represents the scenario under the MLP classifier. Figure 8 and Figure 9 illustrate the performance under SVM and AdaBoost, showing similar trends to MLP. In summary, TS-SMOTE demonstrates high performance and robust stability. Regarding median rankings, although TS-SMOTE’s G-mean performance under the MLP classifier is slightly inferior to BORDERLINE (by only 0.008), its average ranking is higher than BORDERLINE. Therefore, we can conclude that TS-SMOTE outperforms BORDERLINE in this scenario. A similar situation is observed for AUC performance under AdaBoost, where the difference between TS-SMOTE and GAUSSIAN-SMOTE is even smaller (only 0.003). Moreover, TS-SMOTE achieves a better (lower) average ranking than GAUSSIAN-SMOTE, further confirming its superior performance in this case.
Additionally, we calculated the percentage difference in mean values between TS-SMOTE and other methods (shown in Table 8). The percentage difference serves to compare values of the same type, helping us understand their relative differences. The larger the difference, the more significant the disparity. Since TS-SMOTE’s mean values exceed all other methods, all computed differences are greater than 0, demonstrating that our method is optimal.

4.2.2. Data Visualization of Some Characteristic Datasets

After completing the data-level analysis, it is necessary to demonstrate the generation effects of TS-SMOTE. We selected three sample datasets (haberman, winequality-red-4, and yeast3) to show the generation results.
ADASYN adjusts the number of generated samples dynamically, according to the “learning difficulty” of the minority class samples. BORDERLINE-SMOTE generates samples concentrated in the boundary region rather than being uniformly distributed. MWMOTE identifies and filters noisy samples through the k-nearest neighbors (k-NN) method to avoid generating invalid data. The core characteristic of GAUSSIAN-SMOTE is that it enhances the diversity of the generated samples and the rationality of their distribution through the Gaussian distribution. DTO-SMOTE is suitable for scenarios of imbalanced data with high noise levels and blurred boundaries. KNNOR-SMOTE analyzes the local neighborhood (k-nearest neighbors) of minority class samples and decides whether it is necessary to generate new samples dynamically. SMOTE generates new data among minority class samples through linear interpolation. Finally, TS-SMOTE filters out the noise points and outliers, and then conducts multiple linear interpolations in the selected area to generate new samples.
TS-SMOTE can filter out the noise samples effectively, and generate new samples that are more concentrated. For example, as presented in Figure 10, we can observe that under the dataset yeast3, the samples generated by TS-SMOTE are more concentrated compared to BORDERLINE-SMOTE. The essential reason is that BORDERLINE-SMOTE uses a clustering method to generate new samples. In contrast, TS-SMOTE employs multiple linear interpolations and combines them with a scoring mechanism. This approach effectively ensures that new samples can be generated in an orderly manner within the regions between the minority samples, thereby further guaranteeing the effectiveness of the newly generated samples.
Particularly evident in Figure 10 and Figure 11, when the samples are dispersed in a primary region along with some outliers, traditional oversampling techniques such as SMOTE, Borderline-SMOTE, ADASYN, Guassian-SMOTE, and DTO-SMOTE tend to produce a significant amount of noise in the area between the main region and the outliers. Subsequently, as illustrated in Figure 10, Figure 11 and Figure 12, Borderline-SMOTE, MWMOTE, and KNNOR-SMOTE are limited to generating new samples within a relatively small geometric area. This limitation makes them highly susceptible to overfitting issues. Moreover, Borderline-SMOTE specifically generates samples only around the boundary, which results in the loss of a substantial amount of information regarding the original data distribution.
In stark contrast to these conventional methods, our proposed approach, TS-SMOTE, effectively prevents the generation of samples either between the outliers and the main regions or between separate distinct regions. Through practical applications and experiments, it has been convincingly demonstrated that TS-SMOTE is capable of efficiently addressing the problem of imbalanced classification in scenarios involving multiple regions and outliers, offering a more reliable and effective solution compared to the existing techniques.

4.2.3. Friedman Test and Wilcoxon Signed Rank Test

The Friedman test is a non-parametric statistical test that is widely used to compare multiple classifiers over multiple datasets. The role of the Friedman test is to determine whether there are significant differences in the performance of multiple classifiers across different datasets. For L classifiers and Q datasets, the Friedman test statistic χ F 2 is calculated as follows:
χ F 2 = 12 Q L ( L + 1 ) i = 1 L ( R i L + 1 2 ) 2 ,
where R i is the average rank of the i-th classifier, 1 i L . The null hypothesis H 0 is that all classifiers have identical performance distributions. Under H 0 , χ F 2 follows a chi-squared distribution with ( L 1 ) degrees of freedom. A small p-value (e.g., p < 0.05 ) allows us to reject H 0 , indicating significant differences among classifiers.
The Wilcoxon signed-rank test is a non-parametric test for paired data. Its role is to compare the performance of two related samples, in this case, the performance of TS-SMOTE and another method. For each dataset, let d j = x j , 1 x j , 2 be the difference between the performance of TS-SMOTE ( x j , 1 ) and another method ( x j , 2 ). The test statistic W is computed as follows:
W = j : d j > 0 r j ,
where r j is the rank of | d j | . Under the null hypothesis that d j has a symmetric distribution around zero, W follows a Wilcoxon distribution. A p-value < 0.05 indicates that TS-SMOTE’s performance differs significantly from the compared method.
The two tables (Table 9 and Table 10) below show the results of the Friedman test [54] and the Wilcoxon signed-rank test [55]. From the results of these two tests, we can observe that both tests yield statistically significant results ( p < 0.05 ), leading us to reject the null hypothesis. This indicates a significant difference between TS-SMOTE and the other methods, demonstrating the uniqueness of TS-SMOTE.
Therefore, by combining statistical testing with experimental analysis, we compared TS-SMOTE against the seven other methods across multiple dimensions. The comprehensive results confirm that TS-SMOTE achieves the best overall performance.

5. Conclusions

This paper proposes a novel SMOTE method based on the symmetric triangle scoring mechanism. This improved method first reduces the dimensionality of the basic data, divides regions according to the properties of symmetric triangles (regular triangles), and establishes a corresponding scoring mechanism. Then, samples are dynamically selected based on different regions, and new samples are generated through multiple linear interpolation. We tested with 30 imbalanced datasets and compared with 7 other methods. Three classifiers were adopted, and three metrics were selected for the comparative experiments. We integrated the obtained data, and verified through the Friedman test and the Wilcoxon signed rank test that TS-SMOTE has the best comprehensive performance and can effectively handle the classification problems of imbalanced datasets.
The symmetric triangle scoring mechanism emerges as a valuable tool for imbalanced classification. TS-SMOTE fully considers the impact of the distribution of majority samples on the synthetic points of minority samples and obtains more trustworthy results in imbalanced classification tasks.
There are many directions for future expansion in this paper. For example, only the case of symmetric triangles has been considered in this paper, and it can also be extended to three-dimensional space based on symmetry. Meanwhile, only binary classification problems were considered in this paper, and multi-classification problems are also an area to be explored. Therefore, in the future, we will study more methods based on symmetry.

Author Contributions

Conceptualization, S.S.; methodology, S.S.; software, S.S.; validation, S.S.; formal analysis, S.S.; investigation, S.S.; writing—original draft, S.S.; writing—review and editing, S.Y.; visualization, S.S.; supervision, S.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef]
  2. Lee, A.; Taylor, P.; Kalpathy-Cramer, J.; Tufail, A. Machine learning has arrived! Ophthalmology 2017, 124, 1726–1728. [Google Scholar] [CrossRef]
  3. Mukhamediev, R.I.; Popova, Y.; Kuchin, Y.; Zaitseva, E.; Kalimoldayev, A.; Symagulov, A.; Levashenko, V.; Abdoldina, F.; Gopejenko, V.; Yakunin, K.; et al. Review of artificial intelligence and machine learning technologies: Classification, restrictions, opportunities and challenges. Mathematics 2022, 10, 2552. [Google Scholar] [CrossRef]
  4. Tarekegn, A.N.; Giacobini, M.; Michalak, K. A review of methods for imbalanced multi-label classification. Pattern Recognit. 2021, 118, 107965. [Google Scholar] [CrossRef]
  5. Yang, P.; Yu, J. Challenges in Binary Classification. arXiv 2024, arXiv:2406.13665. [Google Scholar] [CrossRef]
  6. Singh, S.; Khim, J.T. Optimal binary classification beyond accuracy. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 18226–18240. [Google Scholar]
  7. Casebolt, M.K. Gender Diversity In Aviation: What Is It Like To Be In The Female Minority? J. Aviat. Educ. Res. 2023, 32, 4. [Google Scholar] [CrossRef]
  8. Guo, L.Z.; Li, Y.F. Class-imbalanced semi-supervised learning with adaptive thresholding. In Proceedings of the International Conference on Machine Learning (PMLR), Baltimore, MD, USA, 17–23 July 2022; pp. 8082–8094. [Google Scholar]
  9. Fan, J.; Yuan, B.; Chen, Y. Improved dimension dependence of a proximal algorithm for sampling. In Proceedings of the 36th Annual Conference on Learning Theory (PMLR), Bangalore, India, 12–15 July 2023; pp. 1473–1521. [Google Scholar]
  10. Ding, S.; Li, C.; Xu, X.; Ding, L.; Zhang, J.; Guo, L.; Shi, T. A sampling-based density peaks clustering algorithm for large-scale data. Pattern Recognit. 2023, 136, 109238. [Google Scholar] [CrossRef]
  11. Zhang, R.; Zhang, Z.; Wang, D. RFCL: A new under-sampling method of reducing the degree of imbalance and overlap. Pattern Anal. Appl. 2021, 24, 641–654. [Google Scholar] [CrossRef]
  12. Goyal, S. Handling class-imbalance with KNN (neighbourhood) under-sampling for software defect prediction. Artif. Intell. Rev. 2022, 55, 2023–2064. [Google Scholar] [CrossRef]
  13. Islam, A.; Belhaouari, S.B.; Rehman, A.U.; Bensmail, H. KNNOR: An oversampling technique for imbalanced datasets. Appl. Soft Comput. 2022, 115, 108288. [Google Scholar] [CrossRef]
  14. Feng, S.; Keung, J.; Yu, X.; Xiao, Y.; Zhang, M. Investigation on the stability of SMOTE-based oversampling techniques in software defect prediction. Inf. Softw. Technol. 2021, 139, 106662. [Google Scholar] [CrossRef]
  15. Jiang, Z.; Pan, T.; Zhang, C.; Yang, J. A new oversampling method based on the classification contribution degree. Symmetry 2021, 13, 194. [Google Scholar] [CrossRef]
  16. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  17. Fernández, A.; Garcia, S.; Herrera, F.; Chawla, N.V. SMOTE for learning from imbalanced data: Progress and challenges, marking the 15-year anniversary. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
  18. Blagus, R.; Lusa, L. SMOTE for high-dimensional class-imbalanced data. BMC Bioinform. 2013, 14, 106. [Google Scholar] [CrossRef]
  19. Camacho, L.; Douzas, G.; Bacao, F. Geometric SMOTE for regression. Expert Syst. Appl. 2022, 193, 116387. [Google Scholar] [CrossRef]
  20. Zha, D.; Lai, K.H.; Tan, Q.; Ding, S.; Zou, N.; Hu, X.B. Towards automated imbalanced learning with deep hierarchical reinforcement learning. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 2476–2485. [Google Scholar]
  21. Sharma, A.; Singh, P.K.; Chandra, R. SMOTified-GAN for class imbalanced pattern classification problems. IEEE Access 2022, 10, 30655–30665. [Google Scholar] [CrossRef]
  22. Liu, Y.; Liu, Q. SMOTE oversampling algorithm based on generative adversarial network. Clust. Comput. 2025, 28, 271. [Google Scholar] [CrossRef]
  23. Puri, A.; Kumar Gupta, M. Improved hybrid bag-boost ensemble with K-means-SMOTE–ENN technique for handling noisy class imbalanced data. Comput. J. 2022, 65, 124–138. [Google Scholar] [CrossRef]
  24. Chawla, N.V.; Lazarevic, A.; Hall, L.O.; Bowyer, K.W. SMOTEBoost: Improving prediction of the minority class in boosting. In Proceedings of the European Conference on Principles of Data Mining and Knowledge Discovery, Dubrovnik, Croatia, 22–26 September 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 107–119. [Google Scholar]
  25. Law, T.J.; Ting, C.Y.; Ng, H.; Goh, H.N.; Quek, A. Ensemble-SMOTE: Mitigating class imbalance in graduate on time detection. J. Inform. Web Eng. 2024, 3, 229–250. [Google Scholar] [CrossRef]
  26. Fan, W.; Stolfo, S.J.; Zhang, J.; Chan, P.K. AdaCost: Misclassification cost-sensitive boosting. In Proceedings of the International Conference on Machine Learning, Bled, Slovenia, 27–30 June 1999; Volume 99, pp. 97–105. [Google Scholar]
  27. Wang, J.B.; Zou, C.A.; Fu, G.H. AWSMOTE: An SVM-Based Adaptive Weighted SMOTE for Class-Imbalance Learning. Sci. Program. 2021, 2021, 9947621. [Google Scholar] [CrossRef]
  28. Turlapati, V.P.K.; Prusty, M.R. Outlier-SMOTE: A refined oversampling technique for improved detection of COVID-19. Intell.-Based Med. 2020, 3, 100023. [Google Scholar] [CrossRef]
  29. Li, J.; Zhu, Q.; Wu, Q.; Zhang, Z.; Gong, Y.; He, Z.; Zhu, F. SMOTE-NaN-DE: Addressing the noisy and borderline examples problem in imbalanced classification by natural neighbors and differential evolution. Knowl.-Based Syst. 2021, 223, 107056. [Google Scholar] [CrossRef]
  30. Meng, D.; Li, Y. An imbalanced learning method by combining SMOTE with Center Offset Factor. Appl. Soft Comput. 2022, 120, 108618. [Google Scholar] [CrossRef]
  31. Pinkus, A. Approximation theory of the MLP model in neural networks. Acta Numer. 1999, 8, 143–195. [Google Scholar] [CrossRef]
  32. Jakkula, V. Tutorial on Support Vector Machine (SVM); School of EECS, Washington State University: Pullman, WA, USA, 2006; Volume 37, p. 3. [Google Scholar]
  33. Schapire, R.E. Explaining adaboost. In Empirical Inference: Festschrift in Honor of Vladimir N. Vapnik; Springer: Berlin/Heidelberg, Germany, 2013; pp. 37–52. [Google Scholar]
  34. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  35. Barua, S.; Islam, M.M.; Yao, X.; Murase, K. MWMOTE—Majority weighted minority oversampling technique for imbalanced data set learning. IEEE Trans. Knowl. Data Eng. 2012, 26, 405–425. [Google Scholar] [CrossRef]
  36. de Carvalho, A.M.; Prati, R.C. DTO-SMOTE: Delaunay tessellation oversampling for imbalanced data sets. Information 2020, 11, 557. [Google Scholar] [CrossRef]
  37. Bunkhumpornpat, C.; Sinapiromsaran, K.; Lursinsap, C. Safe-level-smote: Safe-level-synthetic minority over-sampling technique for handling the class imbalanced problem. In Proceedings of the 13th Pacific-Asia conference of the Advances in Knowledge Discovery and Data Mining (PAKDD), Bangkok, Thailand, 27–30 April 2009; Proceedings 13. Springer: Berlin/Heidelberg, Germany, 2009; pp. 475–482. [Google Scholar]
  38. Viadinugroho, R.A.A. Imbalanced Classification in Python: SMOTE-Tomek Links Method; Medium: San Francisco, CA, USA, 2021. [Google Scholar]
  39. Kosolwattana, T.; Liu, C.; Hu, R.; Han, S.; Chen, H.; Lin, Y. A self-inspected adaptive SMOTE algorithm (SASMOTE) for highly imbalanced data classification in healthcare. BioData Min. 2023, 16, 15. [Google Scholar] [CrossRef]
  40. Muntasir Nishat, M.; Faisal, F.; Jahan Ratul, I.; Al-Monsur, A.; Ar-Rafi, A.M.; Nasrullah, S.M.; Reza, M.T.; Khan, M.R.H. A Comprehensive Investigation of the Performances of Different Machine Learning Classifiers with SMOTE-ENN Oversampling Technique and Hyperparameter Optimization for Imbalanced Heart Failure Dataset. Sci. Program. 2022, 2022, 3649406. [Google Scholar] [CrossRef]
  41. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; IEEE: New York, NY, USA, 2008; pp. 1322–1328. [Google Scholar]
  42. Lee, H.; Kim, J.; Kim, S. Gaussian-based SMOTE algorithm for solving skewed class distributions. Int. J. Fuzzy Log. Intell. Syst. 2017, 17, 229–234. [Google Scholar] [CrossRef]
  43. Douzas, G.; Bacao, F. Geometric SMOTE a geometrically enhanced drop-in replacement for SMOTE. Inf. Sci. 2019, 501, 118–135. [Google Scholar] [CrossRef]
  44. Santos, M.S.; Abreu, P.H.; García-Laencina, P.J.; Simão, A.; Carvalho, A. A new cluster-based oversampling method for improving survival prediction of hepatocellular carcinoma patients. J. Biomed. Inform. 2015, 58, 49–59. [Google Scholar] [CrossRef]
  45. Ren, J.; Wang, Y.; Cheung, Y.m.; Gao, X.Z.; Guo, X. Grouping-based oversampling in kernel space for imbalanced data classification. Pattern Recognit. 2023, 133, 108992. [Google Scholar] [CrossRef]
  46. Feng, F.; Li, K.C.; Yang, E.; Zhou, Q.; Han, L.; Hussain, A.; Cai, M. A novel oversampling and feature selection hybrid algorithm for imbalanced data classification. Multimed. Tools Appl. 2023, 82, 3231–3267. [Google Scholar] [CrossRef]
  47. Maćkiewicz, A.; Ratajczak, W. Principal components analysis (PCA). Comput. Geosci. 1993, 19, 303–342. [Google Scholar] [CrossRef]
  48. Liberti, L.; Lavor, C.; Maculan, N.; Mucherino, A. Euclidean distance geometry and applications. SIAM Rev. 2014, 56, 3–69. [Google Scholar] [CrossRef]
  49. Demmel, J.W. Matrix computations (gene h. golub and charles f. van loan). SIAM Rev. 1986, 28, 252–255. [Google Scholar] [CrossRef]
  50. Popescu, M.C.; Balas, V.E.; Perescu-Popescu, L.; Mastorakis, N. Multilayer perceptron and neural networks. WSEAS Trans. Circuits Syst. 2009, 8, 579–588. [Google Scholar]
  51. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
  52. He, H.; Garcia, E. Learning from Imbalanced Data IEEE Transactions on Knowledge and Data Engineering; IEEE: New York, NY, USA, 2009. [Google Scholar]
  53. Douzas, G.; Rauch, R.; Bacao, F. G-SOMO: An oversampling approach based on self-organized maps and geometric SMOTE. Expert Syst. Appl. 2021, 183, 115230. [Google Scholar] [CrossRef]
  54. Sheldon, M.R.; Fillyaw, M.J.; Thompson, W.D. The use and interpretation of the Friedman test in the analysis of ordinal-scale data in repeated measures designs. Physiother. Res. Int. 1996, 1, 221–228. [Google Scholar] [CrossRef] [PubMed]
  55. Woolson, R.F. Wilcoxon signed-rank test. In Encyclopedia of Biostatistics; Wiley: Hoboken, NJ, USA, 2005; Volume 8. [Google Scholar]
Figure 1. The main problems of SMOTE: Noise samples are usually referred to as minority samples in the majority regions. (A) presents the situation of newly synthesized samples under ideal conditions, where the new samples are within the minority region. (B,C) show that during the synthesis process, noise samples are selected and synthesized with other minority samples. Although the samples participate in the synthesis, (B) leads to blurred boundaries, while (C) generates new noise points. (D) Two minority samples that are too close to each other are selected and involved in the synthesis, resulting in overlapping samples, which may lead to overfitting.
Figure 1. The main problems of SMOTE: Noise samples are usually referred to as minority samples in the majority regions. (A) presents the situation of newly synthesized samples under ideal conditions, where the new samples are within the minority region. (B,C) show that during the synthesis process, noise samples are selected and synthesized with other minority samples. Although the samples participate in the synthesis, (B) leads to blurred boundaries, while (C) generates new noise points. (D) Two minority samples that are too close to each other are selected and involved in the synthesis, resulting in overlapping samples, which may lead to overfitting.
Symmetry 17 01326 g001
Figure 2. Two types of neighbors of a regular triangle. Panel (a) shows 12 neighbors near a certain red regular triangle, which is called NGB T . The blue triangle in (b) represents the first type of neighbor of the red triangle, and it is called NGB T 1 . The green triangle in (c) represents the second type of neighbor of the red triangle, and it is called NGB T 2 .
Figure 2. Two types of neighbors of a regular triangle. Panel (a) shows 12 neighbors near a certain red regular triangle, which is called NGB T . The blue triangle in (b) represents the first type of neighbor of the red triangle, and it is called NGB T 1 . The green triangle in (c) represents the second type of neighbor of the red triangle, and it is called NGB T 2 .
Symmetry 17 01326 g002
Figure 3. Diagrams of naming and assimilation. As shown in the figure, we can see that S is a minority sample, so the field T S belongs to T m i n , and NGB T S belong to T e m p t y . Thus, according to the assimilation rules, T S is assimilated into T m i n .
Figure 3. Diagrams of naming and assimilation. As shown in the figure, we can see that S is a minority sample, so the field T S belongs to T m i n , and NGB T S belong to T e m p t y . Thus, according to the assimilation rules, T S is assimilated into T m i n .
Symmetry 17 01326 g003
Figure 4. This image shows the score of T n at different side lengths. The red score indicates elimination, while the blue score indicates retention.
Figure 4. This image shows the score of T n at different side lengths. The red score indicates elimination, while the blue score indicates retention.
Symmetry 17 01326 g004
Figure 5. This image shows the process of selecting old samples and synthesizing new samples. Among them, the blue dots represent minority samples, the red dots represent majority samples, and the purple dots represent the newly generated sample points. a 1 , a 2 , a 3 are selected minority samples, while A 1 , A 2 are minority samples synthesized according to the synthesis rules. (ac) represent the schematic diagrams of the synthesis process.
Figure 5. This image shows the process of selecting old samples and synthesizing new samples. Among them, the blue dots represent minority samples, the red dots represent majority samples, and the purple dots represent the newly generated sample points. a 1 , a 2 , a 3 are selected minority samples, while A 1 , A 2 are minority samples synthesized according to the synthesis rules. (ac) represent the schematic diagrams of the synthesis process.
Symmetry 17 01326 g005
Figure 6. Confusion matrix.
Figure 6. Confusion matrix.
Symmetry 17 01326 g006
Figure 7. Box-violin plot description: The figure presents the results of three different evaluation metrics under the MLP classifier. The blue components and connecting lines (arranged top to bottom) represent maximum value, upper quartile, median, lower quartile, and minimum value. Blue annotations indicate mean values, while red annotations mark median values. The green shaded area illustrates the probability density distribution of the dataset. Certain data points are identified as outliers and are excluded from the aforementioned process. The blue part is a box plot, showing quartiles, the median, and the mean; the green part is a violin plot, using width to reflect data density and distribution shape.
Figure 7. Box-violin plot description: The figure presents the results of three different evaluation metrics under the MLP classifier. The blue components and connecting lines (arranged top to bottom) represent maximum value, upper quartile, median, lower quartile, and minimum value. Blue annotations indicate mean values, while red annotations mark median values. The green shaded area illustrates the probability density distribution of the dataset. Certain data points are identified as outliers and are excluded from the aforementioned process. The blue part is a box plot, showing quartiles, the median, and the mean; the green part is a violin plot, using width to reflect data density and distribution shape.
Symmetry 17 01326 g007
Figure 8. Box-violin plot under SVM classifier.
Figure 8. Box-violin plot under SVM classifier.
Symmetry 17 01326 g008
Figure 9. Box-violin plot under AdaBoost classifier.
Figure 9. Box-violin plot under AdaBoost classifier.
Symmetry 17 01326 g009
Figure 10. Under the Haberman dataset, the final results generated by eight methods. Red dots are the minority samples, blue are the majority ones, and grey are the synthetic samples.
Figure 10. Under the Haberman dataset, the final results generated by eight methods. Red dots are the minority samples, blue are the majority ones, and grey are the synthetic samples.
Symmetry 17 01326 g010
Figure 11. Under the vowel0 dataset, the final results generated by eight methods.
Figure 11. Under the vowel0 dataset, the final results generated by eight methods.
Symmetry 17 01326 g011aSymmetry 17 01326 g011b
Figure 12. Under the vehicle0 dataset, the final results generated by eight methods.
Figure 12. Under the vehicle0 dataset, the final results generated by eight methods.
Symmetry 17 01326 g012
Table 1. The imbalanced dataset required for the experiment.
Table 1. The imbalanced dataset required for the experiment.
DatesetsSamplesMinority SamplesMajority SamplesIR
glass1214761381.82
wisconsin6832384441.86
pima7682685001.87
glass0214701442.06
yeast1148442910552.46
haberman306812252.78
vehicle38462126342.99
vehicle08461996473.25
ecoli1336772593.36
new-thyroid2215351805.14
ecoli2336522845.46
segment0230832919796.02
yeast3148416313218.10
yeast-2_vs_4514514639.08
yeast-0-2-5-7-9_vs_3-6-81004999059.14
yeast-0-5-6-7-9_vs_4528514779.35
vowel0988908989.98
yeast-1_vs_74593042914.30
ecoli43362031615.80
page-blocks-1-3_vs_44722844415.86
dermatology-63582033816.9
yeast-1-4-5-8_vs_76933066322.10
yeast4148451143328.1
winequality-red-4159953154629.17
yeast-1-2-8-9_vs_79473091730.57
yeast5148444144032.73
yeast6148435144941.4
poker-8-9_vs_5148525146058.4
poker-8-9_vs_6207525205082
poker-8_vs_6147717146085.88
Table 2. The scores of each dataset across different metrics when using MLP as a classifier (The bold numbers representing the corresponding methods is the best).
Table 2. The scores of each dataset across different metrics when using MLP as a classifier (The bold numbers representing the corresponding methods is the best).
Datasets EstimatorsIR
ABMGDKST
glass1F1-score0.79620.79690.55840.57860.54250.54850.63090.79941.82
G-mean0.78530.78160.29050.31820.45030.44440.70500.79241.82
AUC0.84390.85380.40160.40580.52650.51830.80080.87081.82
wisconsinF1-score0.97910.97270.97930.96710.97400.97520.95540.97411.86
G-mean0.97860.97200.97880.96690.97380.97490.96900.97391.86
AUC0.98940.98850.99180.99450.99470.99440.99530.99531.86
pimaF1-score0.78000.79490.69340.69500.70950.71770.66650.80021.87
G-mean0.77700.78570.67480.67630.68480.69650.74090.80191.87
AUC0.84100.84780.75390.75710.77500.78390.82080.88431.87
glass0F1-score0.85910.80000.74660.73880.77440.77530.72010.85132.06
G-mean0.84160.79280.68090.68110.70420.70280.79490.84302.06
AUC0.90780.86930.80830.81060.82170.82030.85970.91832.06
yeast1F1-score0.77650.74610.73790.71160.74760.74710.59810.82622.46
G-mean0.75250.74080.70770.70920.73760.73590.71950.83212.46
AUC0.82920.81710.78990.79300.81540.81720.79290.91432.46
habermanF1-score0.67500.69720.64340.60700.62660.61140.47230.77972.78
G-mean0.66760.69880.62720.64020.65690.64640.62550.78262.78
AUC0.74930.75970.68580.70970.73890.74900.69490.84472.78
vehicle3F1-score0.89110.89840.58270.60560.62850.61280.67320.88732.99
G-mean0.88480.89330.56890.57190.55970.57490.79160.88722.99
AUC0.94000.95510.73650.75500.76600.76560.89520.96212.99
vehicle0F1-score0.98850.98880.93640.91780.93470.93520.95500.98633.25
G-mean0.98850.98860.93340.91230.93260.93190.97870.98623.25
AUC0.99810.99840.97990.98110.98510.98550.99780.99883.25
ecoli1F1-score0.89990.91140.90630.87320.89530.89190.78640.92383.36
G-mean0.89520.91140.89770.87040.89110.88790.88620.92233.36
AUC0.95580.95950.94130.95160.96330.96470.95460.97583.36
new-thyroid2F1-score0.99350.99370.99730.99110.99730.99730.97460.99615.14
G-mean0.99330.99360.99720.99110.99720.99720.99230.99615.14
AUC0.99981.00000.99990.99920.99970.99960.99960.99975.14
ecoli2F1-score0.96670.98750.86710.91300.94030.93650.84800.96055.46
G-mean0.96540.98830.86690.91130.93840.93450.91940.96055.46
AUC0.98410.99630.94970.95250.97530.97170.96710.98835.46
segment0F1-score0.99880.99890.99860.99740.99860.99860.99220.99826.02
G-mean0.99880.99890.99860.99740.99860.99860.99510.99826.02
AUC0.99980.99990.99980.99970.99980.99980.99970.99946.02
yeast3F1-score0.96650.96780.94310.90650.95280.95130.75440.96828.10
G-mean0.96560.96700.94150.90700.95210.95060.89380.96848.10
AUC0.98280.98940.97650.97000.98460.98480.96520.99488.10
yeast-2_vs_4F1-score0.97870.97730.87110.92050.91680.91590.72940.97429.08
G-mean0.97810.97970.87290.91670.91720.91600.86620.97429.08
AUC0.98900.99140.95790.96210.96790.96710.95060.99499.08
yeast-0-2-5-7-9_vs_3-6-8F1-score0.97410.97830.82180.90890.90630.90860.76480.97929.14
G-mean0.97300.97770.81870.91170.90710.90970.88220.97939.14
AUC0.98870.99350.90460.95650.95190.95600.93370.99259.14
yeast-0-5-6-7-9_vs_4F1-score0.95260.95620.81140.94900.82240.82270.50530.94949.35
G-mean0.95050.95460.80610.94980.82190.82490.74320.94999.35
AUC0.98080.98100.90020.97850.91230.90910.86350.98219.35
vowel0F1-score0.99900.99860.99840.99770.99830.99840.98930.99929.98
G-mean0.99900.99860.99840.99770.99830.99840.99890.99929.98
AUC1.00001.00001.00001.00001.00001.00001.00001.00009.98
yeast-1_vs_7F1-score0.93420.96210.82620.79030.80400.82200.30210.961214.3
G-mean0.92880.96140.81360.78860.79950.81630.65790.961514.3
AUC0.97540.98570.88030.87220.89260.89980.76900.979614.3
ecoli4F1-score0.98740.97650.98560.97410.98560.98710.79840.984515.8
G-mean0.98700.98110.98510.97380.98540.98700.90030.984315.8
AUC0.99550.99850.99800.99640.99900.99930.98630.996315.8
page-blocks-1-3_vs_4F1-score0.99700.99650.99590.97640.99430.99500.91070.993515.86
G-mean0.99690.99650.99580.97610.99420.99490.99180.993515.86
AUC0.99730.99710.99700.98060.99590.99610.99820.997815.86
dermatology-6F1-score1.00001.00001.00000.99971.00001.00001.00001.000016.9
G-mean1.00001.00001.00000.99971.00001.00001.00001.000016.9
AUC1.00001.00001.00001.00001.00001.00001.00001.000016.9
yeast-1-4-5-8_vs_7F1-score0.94090.94900.79080.72480.78530.79050.13990.973922.1
G-mean0.93620.94670.77450.71030.76750.77070.50110.974222.1
AUC0.97690.98240.86770.80680.85980.85230.64120.985422.1
yeast4F1-score0.97470.97970.87990.85530.88570.89050.35940.981228.1
G-mean0.97380.97930.87660.85580.88330.88850.71870.981328.1
AUC0.98740.99070.95020.92800.95140.95540.86270.994828.1
winequality-red-4F1-score0.97570.97710.87020.86280.85860.87260.13360.976029.17
G-mean0.97470.97680.86540.86200.85220.86770.42500.976229.17
AUC0.99060.98870.93360.93500.92860.93650.67190.989029.17
yeast-1-2-8-9_vs_7F1-score0.95880.96370.79350.77160.78600.79070.15290.983730.57
G-mean0.95710.96330.77710.77270.77960.78520.50710.983830.57
AUC0.98620.99200.86230.85520.87900.88210.70090.990730.57
yeast5F1-score0.99220.99200.98140.96420.98190.98240.69450.988632.73
G-mean0.99220.99190.98080.96300.98130.98180.91180.988732.73
AUC0.99640.99720.99030.98680.99250.99230.98740.999532.73
yeast6F1-score0.98610.98830.93960.91020.93660.93730.45510.991041.4
G-mean0.98570.98820.93700.90980.93560.93630.79650.991041.4
AUC0.99390.99580.97980.96700.98450.98440.92260.998241.4
poker-8-9_vs_5F1-score0.99310.99130.99200.91900.99270.99130.22770.992858.4
G-mean0.99300.99120.99190.91820.99260.99110.58580.992958.4
AUC0.99920.99790.99900.97710.99930.99900.79830.996758.4
poker-8-9_vs_6F1-score0.99991.00001.00001.00001.00000.99991.00001.000082
G-mean0.99991.00001.00001.00001.00000.99991.00001.000082
AUC1.00001.00001.00001.00001.00001.00001.00001.000082
poker-8_vs_6F1-score1.00001.00001.00001.00000.99991.00000.99381.000085.88
G-mean1.00001.00001.00001.00000.99991.00000.99631.000085.88
AUC1.00001.00001.00001.00001.00001.00001.00001.000085.88
Table 3. When using SVM as a classifier, the scores of each dataset across different metrics.
Table 3. When using SVM as a classifier, the scores of each dataset across different metrics.
Datasets EstimatorsIR
ABMGDKST
glass1F1-score0.77050.76390.40110.40270.40310.40410.6260.79181.82
G-mean0.73030.70460.0940.09120.08450.08660.69850.77091.82
AUC0.81760.82930.36230.39610.43340.42710.78190.85191.82
wisconsinF1-score0.97720.97030.97780.97340.9770.97830.95750.97751.86
G-mean0.97660.96910.97690.9730.97640.97780.97230.97711.86
AUC0.98520.97650.98430.98910.9860.98690.98640.9911.86
pimaF1-score0.77120.79630.69350.70380.71140.7170.66570.80421.87
G-mean0.76470.77590.68850.70850.72080.72510.74020.80631.87
AUC0.83380.83390.77170.80510.81520.81920.82480.8941.87
glass0F1-score0.83340.80850.5840.58870.58830.58770.71570.84782.06
G-mean0.78750.78950.34910.36340.36210.36020.79130.82842.06
AUC0.86770.86350.4520.44080.44930.44760.85230.88312.06
yeast1F1-score0.75860.74610.74770.71790.73660.73950.58590.7442.46
G-mean0.71560.71340.69680.71340.72360.72530.70910.75822.46
AUC0.7950.81170.77980.80040.8110.81450.78570.83212.46
habermanF1-score0.64730.64830.59390.53620.48950.48030.43230.7442.78
G-mean0.65020.66720.4370.60010.56370.55710.5890.75822.78
AUC0.73310.73720.44580.71640.71660.71840.680.83212.78
vehicle3F1-score0.86060.86790.62650.64450.6450.64580.65370.83472.99
G-mean0.8350.84740.64860.6610.66420.6640.79510.83562.99
AUC0.89760.91680.72440.73510.73880.73960.87130.91852.99
vehicle0F1-score0.97420.97360.85150.84140.84190.84260.92480.96793.25
G-mean0.97370.97250.80660.78840.78930.79050.96990.96773.25
AUC0.9980.99480.94790.95520.96190.96240.99420.99723.25
ecoli1F1-score0.91060.90570.91710.90990.90960.91330.77070.91883.36
G-mean0.90460.90480.90870.90240.90310.90720.88240.91853.36
AUC0.9520.95110.95690.95040.96760.96790.94690.9763.36
new-thyroid2F1-score0.99450.99450.91380.89220.86520.86490.96710.98825.14
G-mean0.99440.99440.90820.89790.87370.87350.98720.98825.14
AUC110.97860.99130.98170.99160.99940.99935.14
ecoli2F1-score0.91530.98140.91270.93280.95880.95950.87940.96185.46
G-mean0.91660.98270.91080.93350.95880.95970.94140.96225.46
AUC0.97810.98910.97760.97160.98640.9860.95970.98935.46
segment0F1-score0.99430.99640.9480.98510.98820.98710.98920.99686.02
G-mean0.99430.99640.94350.98510.98810.98710.99110.99686.02
AUC0.99980.99990.99860.99770.99890.99880.99980.99996.02
yeast3F1-score0.94710.9370.94160.90860.95590.9560.73520.95778.1
G-mean0.94470.93350.93890.90950.95480.9550.89880.95768.1
AUC0.98010.97920.97640.97220.98510.98520.96880.98618.1
yeast-2_vs_4F1-score0.9660.96630.94760.93280.92110.9310.74520.97689.08
G-mean0.96430.96930.94290.92790.92290.9320.8820.97719.08
AUC0.99180.98710.98420.96980.98850.98950.96960.99619.08
yeast-0-2-5-7-9_vs_3-6-8F1-score0.89780.90780.81180.91270.92110.9240.79040.96569.14
G-mean0.89260.90050.79690.91550.92290.92610.88250.96619.14
AUC0.96020.97580.89750.94540.98850.95590.9410.99079.14
yeast-0-5-6-7-9_vs_4F1-score0.90620.90510.83880.80780.83360.85060.47270.9349.35
G-mean0.90080.90460.83210.81230.83890.85430.74330.93559.35
AUC0.95840.96930.91380.90150.92640.92980.87780.98289.35
vowel0F1-score0.99910.99970.98410.97870.98190.98190.99680.99949.98
G-mean0.99910.99970.98410.9780.98130.98140.99970.99949.98
AUC110.99750.99910.99920.9994119.98
yeast-1_vs_7F1-score0.86990.93540.84450.82010.83290.85110.2750.948614.3
G-mean0.85940.93440.82590.8170.82640.84180.66560.948914.3
AUC0.93980.97470.91160.89750.91790.92540.76820.977714.3
ecoli4F1-score0.98750.97650.98710.98540.98780.99030.77480.9915.8
G-mean0.98720.98110.98670.9850.98780.99010.88310.9915.8
AUC0.99820.99610.99910.99740.99950.99970.99090.999415.8
page-blocks-1-3_vs_4F1-score0.98240.99760.92010.71860.75950.78690.87410.988215.86
G-mean0.98240.99760.91990.7470.78130.80360.9680.988215.86
AUC0.999410.96750.90640.92810.91690.99810.999515.86
dermatology-6F1-score110.99560.98450.98760.983910.998116.9
G-mean110.99550.98420.98730.983710.998116.9
AUC1110.99860.99950.99951116.9
yeast-1-4-5-8_vs_7F1-score0.86970.89390.78470.74950.77660.76560.13770.974222.1
G-mean0.85730.88630.75560.74710.74460.73210.56720.974522.1
AUC0.93390.95560.86110.82430.85010.84480.64610.980822.1
yeast4F1-score0.92860.95320.8840.87390.87690.89550.29420.978628.1
G-mean0.92350.95280.87890.87410.8790.8950.74750.978928.1
AUC0.97530.98570.94850.93650.95240.95770.88310.993228.1
winequality-red-4F1-score0.91730.94720.65720.60320.62610.62410.17190.977229.17
G-mean0.91150.94620.67770.63870.64620.64810.6140.977429.17
AUC0.96360.97930.75170.71230.71470.71450.71630.98729.17
yeast-1-2-8-9_vs_7F1-score0.85480.92840.80370.76550.80160.79740.11560.982330.57
G-mean0.84630.92730.77150.77730.78780.78540.57310.982530.57
AUC0.94260.97820.8570.86380.88040.88180.68940.987930.57
yeast5F1-score0.9850.98520.97410.96470.97490.97480.63320.988532.73
G-mean0.98470.98490.9730.9630.97390.97370.91910.988532.73
AUC0.99370.99350.98860.98790.99080.99060.98540.999132.73
yeast6F1-score0.94940.98060.90460.91540.92470.92660.3940.98941.4
G-mean0.94760.98020.90220.91560.92530.92710.8150.989141.4
AUC0.98630.99370.96890.96660.97990.98020.92420.997641.4
poker-8-9_vs_5F1-score0.97140.970.96330.9090.96560.96320.120.993958.4
G-mean0.970.96880.96140.90690.96410.96120.59990.993958.4
AUC0.99570.99520.99140.96830.99410.99090.77490.994658.4
poker-8-9_vs_6F1-score10.99930.91860.86970.94080.93310.8210.994782
G-mean10.99930.91020.8770.93830.92730.84490.994782
AUC10.99990.98810.96770.98980.99040.98050.999882
poker-8_vs_6F1-score10.99970.92970.89970.93820.92920.78030.99785.88
G-mean10.99970.92120.90370.93160.92050.80930.99785.88
AUC110.99530.97910.99670.9950.98540.999785.88
Table 4. When using AdaBoost as a classifier, the scores of each dataset across different metrics.
Table 4. When using AdaBoost as a classifier, the scores of each dataset across different metrics.
Datasets EstimatorsIR
ABMGDKST
glass1F1-score0.79140.81750.79060.8050.82160.79870.67420.81281.82
G-mean0.78970.81440.78580.80530.81530.79430.74070.81011.82
AUC0.83760.84650.83960.86620.85740.84530.79750.87221.82
wisconsinF1-score0.96950.96370.97270.9690.96970.97020.9430.96711.86
G-mean0.96950.96320.97260.9690.96960.970.95750.96711.86
AUC0.98970.9840.99070.99480.99370.99390.9910.99281.86
pimaF1-score0.74260.7460.75810.77070.77420.77730.65390.80031.87
G-mean0.74590.73580.75470.77470.77210.77570.73010.80071.87
AUC0.81970.79940.82790.86330.84880.84920.81060.8841.87
glass0F1-score0.82590.87640.81150.82990.85250.86690.71010.84712.06
G-mean0.81280.87380.80120.82820.84450.86220.78180.84522.06
AUC0.87250.89520.86560.90380.90460.90870.8470.9092.06
yeast1F1-score0.76940.73950.75710.83010.76930.77210.59130.82972.46
G-mean0.75090.73630.74270.83560.76410.76680.71260.83542.46
AUC0.82670.82210.82260.91760.84440.84390.78910.91572.46
habermanF1-score0.70910.62880.71260.75320.70630.69930.38790.7882.78
G-mean0.70280.63850.70830.76210.70770.70340.55050.7942.78
AUC0.77130.68290.77240.82650.75810.76480.60570.86122.78
vehicle3F1-score0.83120.84290.82870.81170.8240.82330.59010.83752.99
G-mean0.830.83710.82410.81290.82110.82090.72860.84172.99
AUC0.90340.90780.89980.90470.90080.90010.83480.92832.99
vehicle0F1-score0.97910.97910.97710.97430.97690.97990.91710.97093.25
G-mean0.97910.97870.97660.9740.97650.97960.95770.97073.25
AUC0.99590.99540.99550.99340.99590.99660.98710.99383.25
ecoli1F1-score0.88550.9150.89910.92370.91530.91030.75750.92823.36
G-mean0.88450.91770.89490.92270.91360.90870.85910.92723.36
AUC0.93820.95090.94440.97310.95650.95480.92180.97413.36
new-thyroid2F1-score0.9940.99040.99280.98420.99090.99060.92860.9815.14
G-mean0.99380.99020.99270.9840.99070.99050.95210.98095.14
AUC0.99970.99960.99970.9990.99980.99970.99710.9995.14
ecoli2F1-score0.93630.96860.9290.96360.95760.94920.78020.96185.46
G-mean0.93570.97020.92710.96390.95750.94920.8740.9625.46
AUC0.97230.97380.9670.99060.98190.98020.93370.98675.46
segment0F1-score0.99820.99830.99820.99770.99820.99850.98840.99756.02
G-mean0.99820.99830.99820.99770.99820.99850.99270.99756.02
AUC110.99990.99990.99980.99990.99950.99986.02
yeast3F1-score0.9530.94680.95040.96970.95860.95850.77090.9718.1
G-mean0.95230.94560.94980.96960.95830.95820.90280.97118.1
AUC0.98070.98280.98170.9960.98730.98730.96840.99258.1
yeast-2_vs_4F1-score0.96830.96550.96990.94520.96990.9670.73650.97449.08
G-mean0.96790.96820.96920.94540.96940.96670.85110.97459.08
AUC0.98670.97430.98550.98720.98740.98880.90970.98939.08
yeast-0-2-5-7-9_vs_3-6-8F1-score0.92240.93620.93390.97720.94210.93170.71880.97819.14
G-mean0.92180.93550.93190.97740.94270.93260.87460.97829.14
AUC0.97120.97530.97010.99340.98270.98210.93240.9929.14
yeast-0-5-6-7-9_vs_4F1-score0.88990.92630.90370.9490.90930.90250.48290.94759.35
G-mean0.88920.92440.9010.94980.90830.9020.73920.94829.35
AUC0.9490.96340.95090.97850.95890.95860.80720.97549.35
vowel0F1-score0.99640.99660.99590.99640.99630.99710.96280.99089.98
G-mean0.99640.99660.99590.99640.99630.9970.9840.99099.98
AUC0.99960.99960.99980.99970.99970.99980.99820.99769.98
yeast-1_vs_7F1-score0.8860.93330.89760.95570.89660.89930.2940.954614.3
G-mean0.88230.92950.89460.95610.89440.8970.61980.954814.3
AUC0.94590.96090.94880.98230.95230.95380.75810.979514.3
ecoli4F1-score0.98730.99260.98880.98640.98820.98850.78370.983815.8
G-mean0.98720.99250.98870.98630.98810.98840.88260.983815.8
AUC0.99840.99750.9990.99920.99920.99920.98780.999215.8
page-blocks-1-3_vs_4F1-score0.99810.99690.99790.99110.99890.99840.95510.996415.86
G-mean0.99810.9970.99790.99110.99890.99840.9720.996415.86
AUC10.99880.99890.99820.99890.99890.99890.999815.86
dermatology-6F1-score0.99720.99840.99790.99640.99820.99850.93780.998516.9
G-mean0.99720.99840.99790.99640.99820.99850.96340.998516.9
AUC0.99850.99850.99850.99940.99850.99850.99850.998516.9
yeast-1-4-5-8_vs_7F1-score0.87470.92430.8830.97080.88270.88110.14190.970222.1
G-mean0.86940.92220.87920.97120.87870.87650.48930.970522.1
AUC0.94190.9690.94940.98650.94850.94330.65430.985722.1
yeast4F1-score0.9320.96250.93220.97870.94070.9320.34620.978928.1
G-mean0.93130.9620.93120.97880.94010.93180.74090.979128.1
AUC0.97240.98440.97170.99370.9770.9760.83920.992328.1
winequality-red-4F1-score0.84460.95310.86330.8810.8480.84490.12270.967529.17
G-mean0.84190.95250.86170.88390.84540.84340.5380.967629.17
AUC0.91910.98270.93310.95490.91930.91970.63050.986329.17
yeast-1-2-8-9_vs_7F1-score0.88950.94840.89450.98130.90250.89690.14960.980330.57
G-mean0.88630.94810.89080.98140.90090.89490.53980.980430.57
AUC0.95340.98270.95610.99070.96350.95880.6950.989530.57
yeast5F1-score0.98630.98810.98630.98920.9890.98830.67450.990232.73
G-mean0.98620.98790.98610.98920.98890.98820.89380.990232.73
AUC0.99370.99470.99390.99960.99550.99540.98350.999432.73
yeast6F1-score0.95780.98420.95860.98810.95950.96230.37680.989441.4
G-mean0.95730.9840.95790.98810.95920.96210.76750.989441.4
AUC0.98520.99340.98550.99820.98890.98870.87560.997541.4
poker-8-9_vs_5F1-score0.92290.8510.94340.99320.94670.93630.02510.993958.4
G-mean0.9190.84310.94250.99320.94660.93540.37270.993958.4
AUC0.97140.9020.9850.99310.98850.98480.4710.99458.4
poker-8-9_vs_6F1-score0.90040.79330.94610.98960.95840.94210.02790.991382
G-mean0.89850.79340.94650.98970.95870.94270.35350.991382
AUC0.96240.87550.98380.98890.98750.98580.41550.991282
poker-8_vs_6F1-score0.92470.91450.96590.99310.96690.96350.03610.993385.88
G-mean0.92340.91080.96570.99320.96680.96310.41120.993485.88
AUC0.98160.96490.99220.99440.99280.99170.55090.995185.88
Table 5. Average rankings of different metrics across classifiers.
Table 5. Average rankings of different metrics across classifiers.
ClassifierMLPSVMAdaBoost
EstimatorsF1-ScoreG-MeanAUCF1-ScoreG-MeanAUCF1-ScoreG-MeanAUC
ADASYN2.472.532.932.773.102.905.275.375.40
BORDERLINE2.302.532.932.703.102.903.973.374.87
MWMOTE4.475.005.535.235.836.134.604.805.00
GAUSSIAN-SMOTE6.206.475.736.506.436.833.433.403.00
DTO-SMOTE4.804.474.035.275.204.773.673.703.00
KNNOR-SMOTE4.404.434.004.904.874.574.003.903.00
SMOTE7.136.375.536.736.076.008.008.008.00
TS-SMOTE2.202.101.871.731.631.432.772.702.00
Table 6. Average rankings across different classifiers.
Table 6. Average rankings across different classifiers.
ClassifierADASYNBORDERLINEMWMOTEGAUSSIANDTO-SMOTEKNNOR-SMOTESMOTETS-SMOTE
MLP2.642.304.906.134.434.286.342.06
SVM2.922.685.736.595.084.786.271.60
AdaBoost5.344.274.803.043.613.867.822.58
Table 7. Average values of different metrics across classifiers.
Table 7. Average values of different metrics across classifiers.
ClassifierMLPSVMAdaBoost
EstimatorsF1-ScoreG-MeanAUCF1-ScoreG-MeanAUCF1-ScoreG-MeanAUC
ADASYN0.94050.93730.96660.91470.90720.94920.90880.90660.9479
BORDERLINE0.94140.94000.96420.92450.91950.95560.91590.91490.9453
MWMOTE0.87160.85530.90790.84200.81140.87930.91460.91220.9503
GAUSSIAN-SMOTE0.86760.85530.90940.82430.80990.88480.93820.93890.9689
DTO-SMOTE0.87920.86980.92200.83740.82010.89760.92030.91900.9556
KNNOR-SMOTE0.88010.87150.92280.83950.82180.89690.91750.91660.9549
SMOTE0.67280.81650.89430.64330.81600.89270.58220.75780.8330
TS-SMOTE0.94930.94920.97480.94040.94020.96790.94570.94620.9724
Table 8. Percentage differences between TS-SMOTE and other methods.
Table 8. Percentage differences between TS-SMOTE and other methods.
ClassifierMLPSVMAdaBoost
EstimatorsF1-ScoreG-MeanAUCF1-ScoreG-MeanAUCF1-ScoreG-MeanAUC
ADASYN0.93%1.25%0.85%2.77%3.58%1.95%3.98%4.27%2.55%
BORDERLINE0.84%0.97%1.09%1.70%2.23%1.28%3.20%3.36%2.83%
MWMOTE8.54%10.41%7.11%11.04%14.70%9.59%3.35%3.65%2.30%
GAUSSIAN9.00%10.40%6.94%13.16%14.89%8.97%0.80%0.77%0.36%
DTO-SMOTE7.67%8.73%5.57%11.59%13.65%7.53%2.73%2.91%1.74%
KNNOR-SMOTE7.57%8.53%5.48%11.33%13.44%7.61%3.03%3.18%1.81%
SMOTE34.09%15.03%8.61%37.51%14.14%8.08%47.59%22.11%15.44%
Table 9. Results for Friedman test of different estimators.
Table 9. Results for Friedman test of different estimators.
ClassifierMLPSVMAdaBoost
EstimatorsF1-ScoreG-MeanAUCF1-ScoreG-MeanAUCF1-ScoreG-MeanAUC
p-value2.968 × 10 23 5.305 × 10 21 6.116 × 10 18 6.467 × 10 23 9.582 × 10 21 9.465 × 10 24 6.864 × 10 17 1.808 × 10 17 1.197 × 10 20
Table 10. Results for the Wilcoxon signed rank test.
Table 10. Results for the Wilcoxon signed rank test.
ClassifierADASYNBORDERLINEMWMOTEGAUSSIANDTO-SMOTEKNNOR-SMOTESMOTE
MLP8.691 × 10 6 9.346 × 10 4 2.041 × 10 12 4.654 × 10 15 1.310 × 10 12 1.540 × 10 12 6.963 × 10 15
SVM7.957 × 10 11 1.683 × 10 8 4.169 × 10 16 1.743 × 10 16 1.802 × 10 16 3.293 × 10 16 8.920 × 10 16
AdaBoost1.934 × 10 12 6.107 × 10 9 5.350 × 10 12 2.586 × 10 2 1.068 × 10 9 3.023 × 10 10 2.730 × 10 16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, S.; Yang, S. TS-SMOTE: An Improved SMOTE Method Based on Symmetric Triangle Scoring Mechanism for Solving Class-Imbalanced Problems. Symmetry 2025, 17, 1326. https://doi.org/10.3390/sym17081326

AMA Style

Song S, Yang S. TS-SMOTE: An Improved SMOTE Method Based on Symmetric Triangle Scoring Mechanism for Solving Class-Imbalanced Problems. Symmetry. 2025; 17(8):1326. https://doi.org/10.3390/sym17081326

Chicago/Turabian Style

Song, Shihao, and Sibo Yang. 2025. "TS-SMOTE: An Improved SMOTE Method Based on Symmetric Triangle Scoring Mechanism for Solving Class-Imbalanced Problems" Symmetry 17, no. 8: 1326. https://doi.org/10.3390/sym17081326

APA Style

Song, S., & Yang, S. (2025). TS-SMOTE: An Improved SMOTE Method Based on Symmetric Triangle Scoring Mechanism for Solving Class-Imbalanced Problems. Symmetry, 17(8), 1326. https://doi.org/10.3390/sym17081326

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop