Next Article in Journal
Constrained Optimal Control for Nonlinear Multi-Input Safety-Critical Systems with Time-Varying Safety Constraints
Next Article in Special Issue
CNN-Hyperparameter Optimization for Diabetic Maculopathy Diagnosis in Optical Coherence Tomography and Fundus Retinography
Previous Article in Journal
Feature Selection Using Artificial Gorilla Troop Optimization for Biomedical Data: A Case Analysis with COVID-19 Data
Previous Article in Special Issue
Mispronunciation Detection and Diagnosis with Articulatory-Level Feedback Generation for Non-Native Arabic Speech
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Ensemble and Iterative Recovery Strategy Based kGNN Method to Edit Data with Label Noise

1
Chongqing Key Laboratory of Computational Intelligence, College of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
Department of Computer Science and Engineering, University of California, Riverside, CA 92521, USA
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(15), 2743; https://doi.org/10.3390/math10152743
Submission received: 6 June 2022 / Revised: 25 July 2022 / Accepted: 1 August 2022 / Published: 3 August 2022
(This article belongs to the Special Issue Recent Advances in Artificial Intelligence and Machine Learning)

Abstract

:
Learning label noise is gaining increasing attention from a variety of disciplines, particularly in supervised machine learning for classification tasks. The k nearest neighbors (kNN) classifier is often used as a natural way to edit the training sets due to its sensitivity to label noise. However, the kNN-based editor may remove too many instances if not designed to take care of the label noise. In addition, the one-sided nearest neighbor (NN) rule is unconvincing, as it just considers the nearest neighbors from the perspective of the query sample. In this paper, we propose an ensemble and iterative recovery strategy-based kGNN method (EIRS-kGNN) to edit data with label noise. EIRS-kGNN first uses the general nearest neighbors (GNN) to expand the one-sided NN rule to a binary-sided NN rule, taking the neighborhood of the queried samples into account. Then, it ensembles the prediction results of a finite set of ks in the kGNN to prudently judge the noise levels for each sample. Finally, two loops, i.e., the inner loop and the outer loop, are leveraged to iteratively detect label noise. A frequency indicator is derived from the iterative processes to guide the mixture approaches, including relabeling and removing, to deal with the detected label noise. The goal of EIRS-kGNN is to recover the distribution of the data set as if it were not corrupted. Experimental results on both synthetic data sets and UCI benchmarks, including binary data sets and multi-class data sets, demonstrate the effectiveness of the proposed EIRS-kGNN method.

1. Introduction

Nowadays, the high expense of correctly and reliably annotating large-scale data sets, along with the rapid increase in data scale, is leading to cheap data sets with label noise, posing a severe barrier to supervised machine learning, particularly in classification. Label noise, also known as class noise [1], usually refers to the class or label of an instance that is corrupted. “Corrupted” specifically indicates that the observed label of the instance is not its real label. Label noise can be triggered by insufficient information being provided to tag, flaws in expert labeling, subjectivity in the labeling process, data coding, or communication issues [2]. Many detrimental influences will be introduced by the label noise, including decreasing the generalizability of classical classifiers [3], increasing the number of samples that training models require [4], boosting the complexity of the trained models [5], and yielding the inaccurate prediction results [6]. Hence, it is significant and necessary to detect label noise to mitigate the negative influences aroused by the label noise.
Many label noise reduction methods have been developed in response to the sensitivities of k nearest neighbors classifiers—one of the most straightforward instance-based learning algorithms—to label noise [7]. Edited nearest neighbor (ENN) is the most fundamental algorithm developed by Wilson in 1972 to remove instances misclassified by k nearest neighbors [8]. Following that, numerous extensions of this method have been developed to compensate for the drawbacks of this rough editing [9,10]. Despite its flourishing development, the kNN-based label noise editing methods still have some limitations. An important danger of instance selection methods of this kind is removing too many instances if it is not designed to take care of the label noise [2,7,11]. Furthermore, given a testing sample, the traditional NN rule just selects the k nearest neighbors from the perspective of the testing sample, without considering the k nearest neighbors of the training samples. For example, given a testing sample A, its 3-nearest neighbors in the training sets are B, C, and D. Sample E is A’s 4th nearest neighbor, and E’s nearest neighbor is A. Based on the traditional NN rule, E will be excluded from the 3-nearest neighbors of A. However, from the perspective of E, E and A are closest to each other and may also contribute to classifying A. Consequently, only considering the neighborhood of the testing sample and neglecting the neighborhood of the training sample might result in the one-sided NN rule being unconvincing [12,13,14,15].
In this paper, we propose a novel ensemble and iterative recovery strategy-based kGNN method (EIRS-kGNN) to edit the data with label noise. The main contributions can be ascribed as follows:
  • We introduce the general nearest neighbor (GNN) [12] method to take the neighborhood information of all samples into account. For any query sample Q, its traditional k-general nearest neighbors and samples whose k nearest neighbors contain Q constitute Q’s k general nearest neighbors. In this way, the one-sided NN rule is substituted with the bilateral NN rule, improving the convincingness of the NN rule.
  • We ensemble the prediction results of a finite set of ks to produce the inconsistent number as a vote to measure the noise levels for each sample using the kGNN classifier. A sample with a higher vote value is obviously much easier to be detected as label noise, as it has more predicted results contradictory to its given label.
  • We flip the labels of easy-to-learn label noise to the expected ones and repeat the ensemble classifying process to detect more difficult-to-learn label noise, enhancing the precision of the detected label noise. Here, the samples selected to be flipped have different vote values. This happens one level each time, which is called cascade recovery.
  • Two loops, i.e., the inner loop and the outer loop, are utilized to iteratively detect the suspected label noise. The inner loop will output the suspected label-noise samples for each fold, and the outer loop will produce the frequency of each sample being detected as label noise among the R folds. Frequency is later used as an indicator to categorize the detected label noise into two types, i.e., boundary noise and definite noise. Boundary noise is removed, and definite noise is relabeled to better recover the original distribution of the training set.
The rest of the paper is organized as follows. In Section 2, related methods to edit the training sets are introduced. We discuss the proposed EIRS-kGNN method in detail in Section 3. Experimental results are shown in Section 4, and the conclusions are drawn in Section 5.

2. Related Work

Cover and Hart first proposed the k nearest neighbors (kNN) classifier in 1967 [16], and it has been widely used and prospered ever since then. It predicts a sample’s label using the majority label of its k nearest neighbors, which has the advantage of converging quickly in the training phase and learning new samples incrementally. However, the kNN classifier is vulnerable to label noise, as demonstrated by Sánchez and Wilson in [7,17]. Due to its sensitivity, the kNN classifier is frequently used to detect label noise.
The edited nearest neighbor (ENN) method was invented by Wilson to eliminate instances that were misclassified by their k nearest neighbors [6]. ENN eliminates instances whose labels disagree with the majority label of their k neighbors, and samples located around the border or in the overlapping areas, resulting in smoother decision boundaries. Tomek extended ENN in two ways [9]. Repeated edited nearest neighbors (RENN) is the first extension. It repeats the ENN algorithms on the training set multiple times until no more instances will be removed. The all kNN method (AllkNN) is the second extension. It repeatedly conducts ENN on the training set with k ranging from 1 to k. Instances misclassified by any k value will be removed. To limit the amount of reduction that it can accomplish, there are some criteria by which to terminate the algorithm. For example, once the majority class is reduced to a minority class, the iteration will be stopped. These two extensions can keep the main data intact and generally serve more as noise filters rather than serious reduction techniques [7].
Sánchez et al. used the k nearest centroid neighbors that were closest to the target but also as symmetrically distributed around the target as possible [18], to substitute the k nearest neighbors. Koplowitz and Brown developed a generalized editing method [11] to improve the performance of ENN by reducing the proportion of deleted samples. Devijver presented a multi-edit algorithm [19] that splits the training data into subsets and estimates the mislabeled instance separately. Hattori and Takahashi proposed a modified edited kNN rule (MEkNN) [20]. In MEkNN, all of k or ( k + 1 ) -nearest neighbors of y must be in the class to which y belongs. Hart proposed a condensed nearest neighbor (CNN) rule to edit the training data of the kNN [21]. It is achieved by enumerating instances in the data set and adding them to "storage" only when the instance cannot be properly classified by the current instances of the "storage." However, CNN is easily influenced by the label noise. If there is label noise, it will be retained in the condensed data set, so more samples will be incorrectly classified. Kubat presented the one-sided selection (OSS) algorithm [22], which mainly focuses on reducing the number of the majority class to mitigate the imbalance distribution between classes. On the one hand, it uses the CNN algorithm to remove correctly classified instances from the majority class; and on the other hand, it removed the borderline instances appearing in Tomek links [23]. The major drawback of OSS is that the CNN rule is extremely susceptible to noise. Furthermore, problems will deteriorate once the label noise falls into the minority class. Laurikkala proposed a neighborhood cleaning rule (NCR). The basic idea is similar to OSS in that it only edits the samples from the majority class. However, it emphasizes data cleaning more than data reduction. NCR employs the ENN algorithm to remove the misclassified samples in the majority class and cleans the neighborhood of the minority samples by removing their neighbors, which belong to the majority class and misclassify the minority samples [24].
The k nearest neighbors algorithms can also be effective at describing the local characteristics of label noise, rather than detecting label noise with direct application of the kNN-based classifiers. A series of methods have been developed based on this observation. First introduced by Xia et al., the relative density (RD) method is a highly successful method for detecting label noise [25]. It utilizes the attributes of training samples to estimate a distance ratio between its nearest k-homogeneous neighbors and k-heterogeneous neighbors to detect label noise. RD predefined a hard threshold to detect label noise. Prior to the training process, samples with relative density values greater than one are detected as label noise by RD and are to be removed from the training set. RD has proven to be successful in describing the local characteristics of each sample [26,27]. However, in some cases, the hard threshold value of one may lose its effects because it ignores the varied data distributions between and within various data sets. Still, there is potential improvement in RD. Liu et al. [28] and Huang et al. [29] have discovered the threshold problem and tried to overcome it by either making the threshold grow by 0.1 intervals to adapt to each data set or by voting on a set of noise corresponding to different thresholds. Only limited benefits can be achieved with these methods though.
Relying on only one base learner is risky; thus, the ensemble filter is appealing in case the given learner is inappropriate for learning the concepts of the given domain problem [30]. The assumption behind the ensemble idea is that the inconsistencies in the prediction results can indicate potential noise [31]. Brodley and Friedl investigated the performances of ensemble classifiers by using a set of learning algorithms that act as a filter for the training data [30]. Subsequently, the majority vote and consensus vote are compared to evaluate the probability of each type of error that can be made in identifying noisy instances. Garcia et al. extended this idea by choosing m models with the best predictions from a given number of well-known classifiers to compose the dynamic ensemble filter [31]. Sluban et al. proposed a novel high-agreement random forest filter to enhance the precision of label noise detection [32]. In the high-agreement random forest filter, an instance is identified as noisy when it is classified into the opposite class by over 60, 70, 80, or 90% of the randomly generated decision trees in the forest. The concept of a partitioning filter was introduced by Zhu et al. to address the data size limitation of the ensemble filters [33]. Khoshgoftaar and Rebours developed two variants of the partitioning filter, i.e., the multiple-partitioning filter (MPF) and the iterative-partitioning filter (IPF) to detect label noise in [34]. For either version, the training data are first split into n subsets. The difference lies in that MPF uses several classifiers on each split, whereas the IPF uses only one base classifier but performs multiple iterations.
There are also some classification algorithms that can be improved by modifying them heuristically to better tolerate the existence of label noise [35,36,37]. An additional issue arises in parameter selection when these methods are used to reduce the impact of label noise. In addition, after focusing on a specific classifier, these modifications are challenging to implement in another classifier, restricting their applicability.

3. An Ensemble and Iterative Recovery Strategy-Based kGNN Method

3.1. k Nearest Neighbors Algorithm

After being developed, the k nearest neighbors algorithm went on to become one of the top 10 algorithms in data mining in 2008 [38] due to its simplicity, effectiveness, and intuitiveness. It predicts the label of new sample based on the majority label of its k nearest neighbors in the training set. However, its predicting performance is biased in data sets with label noise or in the data sets with asymmetric proximity relations.
Figure 1 shows the limited cases of the kNN algorithm. Let the positive class be represented by blue squares, the negative class by yellow triangles, and the query point whose label is unknown by the green circle. Figure 1a shows that the predicted results are highly sensitive to label noise. For sample B, whose true label should be positive, though being corrupted, it is observed as negative. When using the kNN algorithm to predict the label of B, it is easy to output the positive prediction result despite the observed negative label. Due to its sensitivity to label noise, the kNN algorithm is frequently used as a natural way to detect label noise. However, its sensitivity to label noise also leads to unreliable prediction results in training sets. For the query point Q in Figure 1a, when k = 3, Q is predicted as a negative sample, whereas it should be predicted as positive. In addition, the closer the label noise is to the query point, the more difficult it is to provide an accurate prediction result. If point A is label noise, no matter whether k is 1, 3, or 5, the query point Q will be predicted as negative. Figure 1a also demonstrates that the prediction result of the kNN algorithm is dependent on the k values. When k = 1, Q will be predicted as positive, but when k = 3 or 5, Q will be predicted as negative. How to find an appropriate k value for all training samples is crucial to obtaining high accuracy.
One case easily overlooked is that the nearest neighbor (NN) rule just considers the k nearest neighbors of the query point, ignoring the k nearest neighbors of the queried sample. As shown in Figure 1b, given a negative sample Q, Q is easily classified as a positive sample under the conventional NN rule, no matter whether k = 1, 3, or 5. Let k = 1. Q is classified as positive, as its closest neighbor is A. However, Q’s nearest neighbor is A, but A’s nearest neighbor is not Q. In fact, A’s nearest neighbor is B. D and E’s nearest neighbor is Q. We mark the distance ranks of selected samples among their nearest neighbors by the arrows. It can be easily found that the closeness of any two samples can be asymmetric. In this case, samples D and E can hardly participate the voting process of Q when k = 1. The one-sided NN rule disregards the distance-wise closeness of the queried samples to the query sample. Fortunately, we can introduce the general nearest neighbor (GNN) rule to solve this issue [12]. In GNN, except for the k nearest neighbors of the query sample Q, the queried samples whose k nearest neighbors include Q will also be considered as Q’s k general nearest neighbors. Therefore, the traditional k nearest neighbors will be expanded under the GNN rule. When using GNN to classify Q at k = 1, Q’s 1-general nearest neighbors consist of A, D, and E, and then the label of sample Q is determined via a majority voting using Q’s 1-general nearest neighbors, which is negative. Similarly, when k = 3 , Q’s 3-general nearest neighbors consist of A, B, D, E, and F, and Q will also be classified as negative.
Figure 1 indicates that although the kNN algorithm can be regarded as a natural way to detect label noise, its performance is limited by either the selection of the neighborhood size k, or the existence of label noise, or the asymmetric proximity relations, all of which easily arouse the over-cleaning problems in the filtering process.

3.2. An Ensemble and Iterative Recovery Strategy-Based kGNN Method

To mitigate the over-cleaning problem, we propose a novel ensemble and iterative recovery strategy based on the kGNN (EIRS-kGNN) method in this paper. EIRS-kGNN mainly consists of two iterative procedures, i.e., the inner loop and the outer loop. The inner loop is to output the suspected label-noise samples and the expected labels in each fold. The outer loop is to split the training set into R folds, each fold corresponding to an input of the inner loop, and generate the frequency of each sample being detected as label noise in the inner loop. Compared with the standard kNN algorithm, four benefits can be ascribed to EIRS-kGNN in alleviating the over-cleaning problem.
First of all, it considers the mutual neighborhood information of all samples, not only from the perspective of the query sample, but also from the perspective of the queried samples by introducing the general nearest neighbor (GNN) [12], improving the convincingness of the NN rule. Secondly, it ensembles a finite set of neighborhood sizes to detect the suspected label noise instead of relying on a single k, and records the number of prediction results that are inconsistent with its observed label for each sample, i.e., vote, to measure the different noisy levels, contributing to distinguishing easy-to-learn label noise and difficult-to-learn label noise. Thirdly, it flips the labels of easy-to-learn label noise to facilitate detecting difficult-to-learn label noise, enhancing the detection precision of label noise. Here, the samples selected to be flipped are samples with higher vote scores, which are flipped to samples with lower votes by one level each time, which is called “cascade recovery”. Finally, the outer loop process is leveraged to output the frequency, and frequency is later used as an indicator to categorize the detected label noise into two types, i.e., boundary noise and definite noise. Boundary noise will be removed and definite noise will be relabeled to better recover the original distribution of the training set.

3.2.1. kGNN Algorithm

Pan et al. have introduced the concept of general nearest neighbors (GNN) to consider the mutual neighborhood information of both query samples and queried samples [12]. Given a sample X i , its general nearest neighbors mainly consist of two groups. The first group has the traditional k-nearest neighbors of X i when using kNN rule. The second group has the queried samples whose traditional k-nearest neighbors include X i .
Specifically, the procedure designed to calculate the GNN of all samples is as follows [12]:
  • Given a data set D n × ( d + 1 ) = { ( X i , Y i ) } i = 1 n , which contains n samples with d dimensions, for each sample X i , find its k nearest neighbors using the NN rule and generate an index matrix M ( i d x ) n × ( k + 1 ) , where the first column refers to the index of each sample, the second column refers to the index of its first nearest neighbor, the third column refers to the index of its second nearest neighbor, and so on until all k nearest neighbors are found. Euclidean distance is used in this step. Then, the list L i in each row represents k + 1 nearest neighbors of X i , where the nearest neighbor is X i  itself.
  • For each sample X i , let us check which rows in M ( i d x ) n × ( k + 1 ) contain its index i, and append the index of this row j to L i to form X i ’s general nearest neighbors. Additionally, a difference operation between sets L i and i is conducted, as the index of each sample i in L i is redundant. Finally, as the length of each sample’s GNN may be different, we use a dictionary to store the indices of each sample’s general nearest neighbors.
Thus, it uses the broad concept of nearest neighbors by leveraging the GNN rule to consider the mutual nearest neighborhood. GNN sets a limitation to the queried samples within k nearest neighbors, thereby improving the reliability and minimizing the searching space. Following that, the label of X i is decided via majority voting using X i ’s k-general nearest neighbors. Note that, with the extension of the k nearest neighbors, the number of labels used to vote is possibly larger than k, and there may occur equal votes for different labels. For any sample in the training set, if they got equal votes for different labels, we keep their labels unchanged to maintain the original distribution.

3.2.2. An Ensemble and Iterative Recovery Strategy

Directly using the kNN algorithm to edit the data sets will remove too many samples, especially in the data sets with label noise. Despite replacing the NN rule with GNN rule, we propose an ensemble and iterative recovery strategy (EIRS) to address this issue.
EIRS mainly consists of two iterative procedures, i.e., the outer loop and the inner loop. The outer loop is to split the training set into R folds, and each fold corresponds to an input of the inner loop. The inner loop prudently detects the suspected label noise of each fold and outputs their expected labels. The outer loop generates the frequency to indicate the number of each sample being detected as label noise among R folds, facilitating categorizing the detected label noise into two types. The two iterative procedures provide meaningful guidance to deal with different types of label noise, i.e., the boundary noise and definite noise. Figure 2 and Figure 3 demonstrate the iterative procedures of the inner loop and outer loop, respectively, and pseudo code is provided in Algorithm 1.
Give a training set D = { ( X i , Y i ) } i = 1 n , Y i = 1 , 2 , , m , D is firstly split into R folds at the beginning of the outer loop. We take ( R 1 ) folds as the input of the inner loop each time, as shown in Figure 3. We call the ( R 1 ) folds the training validation set. In the inner loop, a finite set of different neighborhood sizes ks, i.e., ks={ k 1 , k 2 , k 3 , k 4 , k 5 }, will be used first to predict the label of each sample X i in the training validation set with the kGNN algorithm, and the label predicted by each k is denoted as Y ^ i k . Then, we compare Y ^ i k with its observed label Y i to record the number of how many times Y ^ i k is inconsistent with Y i as the vote, where k k s . That is V o t e ( X i ) = k = k 1 k s 1 Y ˜ i k Y i , where 1 ( · ) represents the indicator function, and only when the condition in the subscript is satisfied is its value 1; otherwise, it is 0. This step corresponds to step 1 in Algorithm 1. After that, the expected label will be derived from the prediction results of the ensemble ks and denoted as Y ˜ i r , Y ˜ i r = m o d e ( Y ^ i k ) , k k s.
Algorithm 1 EIRS-kGNN.
Input:
A corrupted training data set D = { ( X i , Y i ) } i = 1 n , Y i = 1 , 2 , , m , splits R, a set of neighborhood size ks={ k 1 , k 2 , k 3 , k 4 , k 5 }.
Output:
D e d i t , D n o i s e .
//Outerloop:
1:
Split D into R disjoint subsets D 1 , D 2 ,…, D R ;
2:
for each train split D r , r = 1 , 2 , , R  do
 //Step 1. Utilize ensemble ks to predict the expected label Y i ˜ of X i in D r .
3:
    for each k, k k s do
4:
        for each sample X i , X i D r  do
5:
           Predict the label Y ^ i k of X i using kGNN;
6:
           if  Y ^ i k Y i  then
7:
                V o t e ( X i ) + = 1 ; //Statistic the votes of X i being detected as label noise by k s .
8:
           end if
9:
        end for
10:
    end for
11:
     Y ˜ i r = m o d e ( Y ^ i k ) , k k s; //Expected label Y ˜ i r refers to the most frequent label in Y ^ i k , k k s.
//Innerloop:
12:
    Initialize V o t e = 0 , S u s p = , v = 5 , S u s p e c t _ n o i s e = 0 , s u s p _ l a s t = ;
//Step2: Cascadedly recover and detect the noisy labels in D r ;
13:
    while  v 3  do
14:
         s u s p _ c u r = { X i | V o t e = v };
15:
        if  s u s p _ c u r  then
16:
            S u s p = S u s p s u s p _ c u r ;
17:
            S u s p e c t _ n o i s e ( X i ) + = 1 , X i s u s p _ c u r ;
18:
        else
19:
            v = 1 ;
20:
           Repeat 14–17 to find out the suspected label noise in current iteration cascadedly;
21:
        end if
22:
    end while
23:
    if  s u s p _ c u r s u s p _ l a s t  then
24:
        Edit D r by flipping the label Y i of X i in s u s p _ c u r into Y ˜ i r ;
25:
        Repeat 3–22 to iteratively edit the data sets and detect label noise;
26:
         s u s p _ l a s t = s u s p _ c u r ;
27:
    else
28:
         N o i s e r = { S u s p | S u s p e c t _ n o i s e ( X i ) % 2 = 1 } ;
29:
    end if
30:
end for
31:
F r e q ( X i ) = r = 1 R 1 X i N o i s e r ;
32:
D n o i s e = { X i | F r e q ( X i ) R / 2 } ;
//Step 3: Use mixture strategy to deal with the detected label noise.
33:
for each sample X i , X i D  do
34:
    if  F r e q ( X i ) > R / 2  then
35:
         Y i ˜ = m o d e ( Y ˜ i r ) , r = 1 , 2 , , R ;
36:
         D e d i t = Flip the label Y i of X i to Y i ˜ ;
37:
    else if  R / 2 F r e q ( X i ) R / 2  then
38:
         D e d i t = Remove X i from D e d i t ;
39:
    end if
40:
end for
41:
return D n o i s e , D e d i t .
Different vote values reflect the different noise levels. Samples with higher vote values are more likely to be labeled as noise, which is also the underlying rationale behind the ensemble filters. We call these samples with higher vote values easy-to-learn label noise. In contrast, samples with lower vote values are called difficult-to-learn label noise. For instance, samples with v o t e = 4 are more difficult to learn than samples with v o t e = 5 , but much easier to be detected than samples with v o t e = 3 .
EIRS flips the labels of easy-to-learn label noise to its expected ones for further detecting the difficult-to-learn label noise one level at each time. This process is named the cascade recovery strategy. Specifically, this strategy initializes a S u s p set to store the suspected noise label and start the recovery strategy from the noisiest samples with vote value v = 5. The cascade recovery strategy starts by checking whether there are samples { X i | V o t e ( X i ) = 5 } . If X i exists, it will be added to S u s p , and the S u s p e c t _ n o i s e ( X i ) will be increased by one; if it does not exist, the vote value v will be reduced by one, and it will go to step 14 to check the noisy samples in the next level. The process will run iteratively until there does not exist any sample X i with V o t e ( X i 3 ) , as shown in lines 13–22 in Algorithm 1. Then it flips the labels of suspected noise samples in the current iteration to their expected labels as a recovery step. To avoid the algorithm falling into an infinite loop, we make two temp sets, i.e., s u s p _ c u r and s u s p _ l a s t , to store the flipped samples in the current and last iterations, separately. Once s u s p _ c u r s u s p _ l a s t , it will flip the label Y i of X i in s u s p _ c u r to the expected one Y ˜ i r , and repeat lines 3–22 to iteratively detect label noise; otherwise, it will output the suspected label noise in the current fold.
One case to keep in mind is that some samples, particularly those in the boundary areas, may flip more than once. We should treat carefully when deciding whether to remove or retain these border samples, as boundary samples possibly contain important information for classification [39]. We added a condition in line 28 in Algorithm 1, i.e., N o i s e r = { S u s p | S u s p e c t _ n o i s e ( X i ) % 2 = 1 } , to help retain the important boundary samples in the data sets. This condition is used to judge whether a sample is detected as label noise an even number of times in the inner loop or not. If a sample is detected as label noise an even number of times, we consider it as an important boundary sample and keep it in the data sets; otherwise, it will be put into the noisy set N o i s e r to be further considered.
The inner loop will not terminate until either of the following two conditions is satisfied:
  • Vote values of all samples V o t e ( X ) 2, where 2 = 5 / 2 , · represents the rounding down function, and 5 is the size of ks;
  • s u s p _ c u r = s u s p _ l a s t .
Figure 3 illustrates the outer loop of the EIRS-kGNN method for detecting label noise. It splits the training set into R folds and takes ( R 1 ) out of R folds as the training validation set in each iteration. The training validation set will be the input data of the inner loop, and the inner loop will output a noise set N o i s e i for each fold. F r e q u e n c y is an indicator to record the total number of each sample being identified as label noise across the R folds, as computed in Equation (1).
F r e q ( X i ) = s = 1 R 1 X i N o i s e s
In each fold, if a sample is detected as label noise, its frequency will add 1. Finally, samples with F r e q ( X i ) > R / 2 will be regarded as definite label noise, where · represents the rounding up function, and their labels should be relabeled to the expected ones Y ˜ i r ; in contrast, samples with R / 2 F r e q ( X i ) R / 2 will be regarded as boundary noise, and these samples should be deleted to simplify the classification model. This corresponds to Step 3 in lines 33 to 40 in the Algorithm 1.
There are two advantages of the outer loop. One is that 5-fold splits will bring a slight variation to the data set, which helps to find hard-to-learn label noise. For example, when the noise samples are densely distributed, taking part of the data away in the dense area can make the region sparse, thereby helping to learn the label noise in the dense region. Secondly, the frequency output in the outer loop can be used to guide the next step of processing label noise.
EIRS strikes a balance between the removal and the reserving of the boundary samples by leveraging the remainder function of S u s p e c t _ n o i s e ( X i ) from two, and using frequency indicator to take the important boundary samples into account; and we take a prudent attitude to the relabeling or removal of any detected label noise, enhancing the performance of recovering the original distribution of the data sets.

3.3. Time Complexity Analysis

The time complexity of the proposed EIRS-kGNN is mainly dominated by three steps: finding the k general nearest neighbors, using the inner loop to output the suspected label noise, and using the outer loop to generate the frequency value to deal with the detected label noise. Given a training set D, which contains n samples with d dimensions, and given the finite set of ks, whose number is s, D is split into R splits first of all. For each fold D r , due to the invariance of feature space, the k-general nearest neighbors of samples in D r can be calculated and stored to save time in the first step. The time complexity of using O ( G N N ) to find the ks general nearest neighbors would be O ( s ( n log n + n k ) ) , where O ( n log n ) is the average time complexity of searching the k-nearest neighbors using the k D t r e e and O ( n k ) is the time complexity of searching samples whose k-nearest neighbors contain the query one. Once the k-general nearest neighbors are obtained, the predicted labels and vote values corresponding to various ks can be counted instantly; let it be O ( n ) . In the inner loop, as the GNN of all samples has been stored in a dictionary in advance, it only needs to find the GNN in the dictionary of each suspected label noise, and predict and flip their labels accordingly. In addition, as only some of the samples will be detected as label noise, let it be 1 2 n ; hence, the time of predicting and flipping labels can be O ( 1 2 n ) . The main time overhead is dominated by the iterations that are needed to converge the inner loop. Taking that to be t, the time complexity of the inner loop O ( L i n ) is O ( t 1 2 n ) . In the outer loop, the time complexity mainly depends on the number of splits of the data set. Let us split the data set into R folds, and each iteration takes ( R 1 ) folds as the input data for the inner loop; the time complexity for the outer loop O ( L o u t ) is R R 1 R ( O ( G N N ) + O ( L i n ) ) . Hence, the total time complexity of EIRS-kGNN is O ( ( R 1 ) ( s ( n log n + n k ) + 1 2 t n ) ) .

4. Experimental Results, Analyses, and Discussions

To verify the effectiveness of the proposed EIRS-kGNN method, adequate experiments were conducted on synthetic data sets, binary UCI data sets, and multi-class UCI data sets (https://archive.ics.uci.edu/ml/index.php, accessed on 28 March 2022). Multiple representative kNN-based editing methods, including ENN [6], RENN [9], AllkNN [9], NCR [24], RD [25], and ensemble editing methods, including IPF [34] and DynamicCF [31], were used in the experiments as comparisons. All the kNN-based editing methods were implemented in Python 3.8.5: ENN, RENN, AllkNN, and NCR were implemented with the "imblearn.under_sampling" package; the ensemble editing methods, IPF and DynamicCF, were implemented with the "NoiseFiltersR" package in R-4.2.1 [40]. In this experiment, the noise rate γ was set to {0.1,0.2,0.3}. Given the noise rate γ , γ | P | positive labels were randomly flipped to the negative class, and similarly, γ | N | negative labels were randomly flipped to the positive class, where | P | and | N | denote the numbers of positive samples and negative samples, separately.
With regard to different editing methods, the parameters for each method can be enumerated as follows to search for the best editing results:
  • ENN: k { 3 , 5 , 7 , 9 , 11 } , sampling_strategy = all
  • RENN: k { 3 , 5 , 7 , 9 , 11 } , sampling_strategy = all
  • AllkNN: k { 3 , 5 , 7 , 9 , 11 } , sampling_strategy = all
  • NCR: k { 3 , 5 , 7 , 9 , 11 } , sampling_strategy = all
  • RD: k { 3 , 5 , 7 , 9 , 11 }
  • IPF: R = 5 , p = 0.01 , y = 0.5 , s = 3
  • DynamicCF: R = 5 , m = 3
  • EIRS-kGNN: k s = { 3 , 5 , 7 , 9 , 11 } , R = 5
Here, k denotes the number of nearest neighbors used; sampling_strategy was set as “all” to edit samples from all classes instead of only the “majority” class in the data sets; R denotes the number of partitions in each iteration; p is the minimum proportion of original instances which must be tagged as noisy in order to go for another iteration; y sets the proportion of good instances which must be stored in each iteration; s denotes the iteration stopping criteria; m sets the number of classifiers to make up the ensemble. In this experiment, as R = 5 , EIRS-kGNN regarded samples with frequencies larger than three as definite noise and samples with frequencies between 2 and 3 as boundary noise. Based on the mixture strategy, the labels of the definite noise samples were relabeled to the expected ones, and the boundary noise was removed directly from the training set.

4.1. Experimental Results on Two-Dimensional Data Sets

We first experimented on three two-dimensional data sets, i.e., fourclass, moons, and circles. Figure 4 shows the clean distribution of each data set, and Figure 5, Figure 6 and Figure 7 intuitively demonstrate the varying performances of various editing methods on each of the data sets. Among the data sets, “fourclass” is a frequently used data set from the UCI repository; “moons” and “circles” are synthetic data sets generated by the Python library scikit-learn [27].
Figure 4 contains three subfigures, and each subfigure corresponds to one data set. In each subfigure, the x-axis and y-axis represent each dimension of the two-dimensional data sets. Cyan points represent the positive samples, and green points represent the negative samples. Due to the limitation on the paper length, here, we just show the editing performances of varying methods at γ = 30%. Though omitted, intermediate values do not change the conclusions drawn.
In Figure 5, Figure 6 and Figure 7, each figure consists of nine subfigures. Figure 5a, Figure 6a and Figure 7a show the noisy distribution of every data set, separately, and the noisy samples are colored in red and marked as stars. Figure 5b–i, Figure 6b–i and Figure 7b–i depict the diverse results of various editing methods on each data set, with the eliminated samples highlighted in red and marked with stars. For the EIRS-kGNN method, as it recovers the label of definite noise samples, we colored the recovered noise samples in the positive class light cyan, the recovered noise samples in the negative class light green, and the deleted noisy samples as red stars to facilitate demonstration.
In Figure 5a, Figure 6a and Figure 7a, it can be seen that in the noisy data sets, the cyan samples and the green samples are mixed together, causing difficulty in searching for the optimal classification boundary to separate the two classes appropriately. Figure 5b, Figure 6b and Figure 7b show that though ENN keeps the internal green points intact, it is prone to over-cleaning in the presence of label noise, leaving only a few samples in the training set.
RENN and AllkNN methods are iterative variants based on ENN, and the iteration process is terminated when the majority class becomes the minority one. In these three cases, both RENN and AllkNN methods were stopped at first iteration. RENN directly repeats the ENN method multiple times, so Figure 5c, Figure 6c and Figure 7c are the same as Figure 5b, Figure 6b and Figure 7b. AllkNN repeats the ENN method from k = 1 to the default neighborhood size, which was set to 3, so the removed samples were relatively fewer than ENN, as shown in Figure 5d, Figure 6d and Figure 7d. On the three data sets, AllkNN performed slightly better than ENN and RENN, but still had over-cleaning issues and left some noisy samples undetected.
NCR focuses more on data cleaning rather than data reduction. It not only removes the samples whose labels differ from the majority labels of their neighbors, but also cleans the neighborhoods of the retained samples. In this way, NCR retains the majority of positive and negative class samples and cleans most of the noisy data, as shown in Figure 5e, Figure 6e and Figure 7e. However, when relying on the single ENN to edit the misclassified samples, over-cleaning problems still remain in the presence of label noise, especially in the highly overlapping area.
By leveraging the feature that label-noise samples are generally located far away from their homogeneous samples and closer to their heterogeneous samples, RD can detect most of the label noise samples, as shown in Figure 5f, Figure 6f and Figure 7f. However, relying on a hard threshold, i.e., one, to process all samples in all data sets, it is difficult to produce precise detection results for uncertain samples hovering near 1, resulting in some boundary samples being mistakenly filtered out or noisy ones in overlapping regions being overlooked.
IPF iteratively detects and filters label noise in the data sets using the classification results of a certain classifier. Its detection performance highly relies on the classifying performance. However, as we analyzed previously, label noise negatively affects the performance of classifiers, and this can be observed in Figure 5g, where some of the noisy samples of the negative class whose labels have been flipped from green to cyan are shown to have been undetected, as the classifier regards them as positive cyan samples. Similar results can be found in Figure 6g and Figure 7g. It also can be seen in Figure 5g that, for the imbalanced fourclass data set, the classifier was biased to the majority cyan class, thereby removing too many negative green samples.
DynamicCF is a powerful competitor by dynamically ensembling the best predictions of different classifiers. It can detect and filter label noise at a low error rate. However, some difficult-to-learn label noise, especially the center of the label noise cluster, can hardly be detected and will be left, as shown in Figure 5h, Figure 6h and Figure 7h.
In Figure 5i, Figure 6i and Figure 7i, it can be seen that EIRS-kGNN, with the introduction of general nearest neighbors, the integrated k values, multi-splits of the training set, and hybrid operation of relabeling and filtering, was able to alleviate the over-filtering problem and recover the original distribution of the data set as much as possible, assisting in the discovery of classification boundary in the noisy data set as if it were not corrupted. Moreover, by cascade retrieving the label of the training samples for subsequent detection of the label noise, it is helpful to detect the difficult-to-learn noisy samples in the noisy clusters, which can hardly be detected by the conventional comparative methods.

4.2. Experimental Results on Binary UCI Data Sets

This section exhibits the comparison results of the proposed EIRS-kGNN and the comparable methods on 10 UCI binary data sets. Information about the data sets is listed in Table 1. To begin, we investigate the detection performances of various methods in regard to identifying label noise in Section 4.2.1. Next, we contrast the editing performances of various methods in classification with label noise from the average perspective across classifiers and data sets, separately, in Section 4.2.2. Finally, Section 4.2.3 presents time efficiency comparison of the different methods.
In this experiment, each data set was randomly divided into two parts, 80% for training and 20% for testing. Only the labels in training sets were flipped according to the given noise rate γ . All methods were conducted on the training set to edit the data with label noise, and the edited data were used to train the classification model. Five widely-used classifiers, including adaboost, decision tree (DT), k nearest neighbor (kNN), logistic regression (LR), and gradient boosting decision tree (GBDT), were employed to avoid the classification performance being limited to a specific classifier. For fair comparison, all the parameters were set to their defaults, as listed in Appendix A.1. Results were obtained by repeating the experiment five times to minimize the biases that could be introduced by arbitrary data splitting and label noise flipping. The average scores across runs are recorded as final scores per data set, classifier, and noise rate.

4.2.1. Detection Performance of Label Noise

Successful identification of label noise is one of the important indicators to verify the effectiveness of each editing method. Here, we employ two metrics to assess how successful each method is. One is recall (the proportion of how many noisy samples are successfully detected) and the other is false (the proportion of how many non-noisy samples are mistaken as label noise) [41]. Obviously, a method with high recall and lower false scores performs effectively in detecting label noise. Detailed results are reported in Table 2. The highest scores of recall and lowest scores of false are marked in bold.
As reported in Table 2, in terms of the recall metric, ENN and RENN obtained the highest scores, whereas RD obtained the lowest scores in most cases. AllkNN was able to obtain high recall scores at a low noise rate, but as the noise rate increases to 0.3, its average recall score decreased significantly from 88.20% to 69.00%, with a decrease of 19.20%. This suggests that the performance of AllkNN is greatly influenced by label noise. The recall values obtained by NCR are comparable to those of the ensemble methods. In contrast, the high false scores of ENN, RENN, and AllkNN show that these comparison algorithms sacrifice correct classification of many normal samples to get high recall scores, resulting in the removal of too many samples. RD performed less effectively than other methods in detecting label noise, with the lowest recall and highest false scores.
The ensemble methods, including IPF, DynamicCF, and EIRS-kGNN, achieved higher recall scores and lower false scores in all data sets and noise rates. Their abilities to detect label noise are on par with one another. EIRS-kGNN was slightly superior to the other two approaches at γ = 0.3 , with the highest average recall value 79.54% and the lowest average false value 31.67%, indicating its reliability and stability in identifying label noise, as it can detect more label noise at a lower cost, even if there is a high noise rate.
Recall and false are two metrics that are incompatible with one another, and it is challenging to develop an algorithm that can outperform others in recall and lowest false scores on all data sets at various noise rates. Overall, EIRS-kGNN strikes a great balance between these two measures and exhibits comparable performance with other ensemble methods in detecting label noise.

4.2.2. Classification Performance after Editing

The ultimate goal of any editing method is to improve classification results. To put it another way, an editing method is successful if it produces data that enhance the predicted performance of a classifier. Classifiers using a threshold to discriminate between classes can be used to produce several pairs of TPR (TPR = T P P )–FPR (FPR = F P N ) values by using different thresholds. This results in a receiver operating characteristic (ROC) curve, and the area under the ROC curve—AUC, which generally regarded as an important metric to estimate the ability of a classifier to distinguish between classes [42]. Therefore, AUC was employed to compare the classification performances on the data sets edited by different methods. Table 3 compares the average AUC scores for each method across five classifiers at varying noise rates γ , and the highest scores are highlighted in bold. In Table 3, classification results on the noisy data sets are reported and denoted as "NSY" to compare the improvements of different methods.
As shown in Table 3, label noise does have a negative impact on classification performance. When the noise rate γ was increased from 0.1 to 0.3, the average AUC score on NSY across all data sets decreased from 80.68% to 73.75%, a decrease of 6.93%. After applying different methods to edit the noisy data, the AUC scores improved to varying degrees. When γ = 0.1, most editing methods, including ENN, AllkNN, NCR, RD, DynamicCF, and EIRS-kGNN, were able to bring improvements to AUC scores. The highest improvement of 3.67% was obtained by EIRS-kGNN. The improvement of 3.08% by NCR was second best. RD showed certain resistance to label noise at γ = 0.1 , with an average AUC score of 83.42%, ranking third among all editing methods. ENN and RENN have a tendency to over-clean data sets, making them unfit to attain satisfactory AUC scores. By repeating ENN, RENN tends to cause more serious over-cleaning issues, which is particularly evident for the data set ecoli. RENN removed the entire positive class, resulting in a AUC score of 0 on data set ecoli and decreasing its average AUC score across all data sets.
The average improvement obtained by AllkNN was limited, as the improvement on every data set was limited. On the contrary, the negative improvement obtained by IPF was mainly due to its poor performance on certain data sets, such as bupa and fourclass. This suggests that the suitability of the base learner for a given data set has a significant impact on IPF’s performance. It also can be seen in Table 2 that IPF’s detection performance on these two data sets is unsatisfying, with low recall scores of 44.44% and 77.97% and high false scores of 88.79% and 68.07%. The average improvement of DynamicCF was 2.38% compared to the noisy results at γ = 0.1 , showing the effectiveness of the dynamic ensemble of different kinds of classifiers, but this was still less than the improvement of EIRS-kGNN.
In Table 3, it can be seen that when γ = 0.2 , EIRS-kGNN ranks first with the average AUC score of 83.76%, an average improvement of 3.39%; when γ = 0.3 , its average AUC score remains the highest of 83.23%, resulting in an improvement of 9.48% over the noisy data. This demonstrates that our proposed EIRS-kGNN method contributes to recovering the distribution of the corrupted data set as if it were not corrupted, even with highly corrupted data sets. In contrast, the performances of ENN, AllkNN, NCR, and RD deteriorate gradually as γ grows. Though improvements were obtained compared to the noisy data, their average AUC scores, 78.29%, 78.30%, 79.63%, and 79.84%, are substantially lower than that of EIRS-kGNN.
IPF’s performance varies depending on the data set. On one hand, it obtained high AUC scores on some data sets, such as newthyroid and votes, with scores of 98.49% and 95.90%, respectively, at γ = 0.3 . On the other hand, with the iterative removal of samples, it is possible to delete all the samples in a certain class, leading to 0 AUC scores on these data sets, which occurred with ecoli and haberman. The over-cleaning problem is more likely to occur in tiny, imbalanced data sets. In terms of identifying label noise and enhancing classification performance, DynamicCF is a strong, competitive method. It achieved performance comparable with that of EIRS-kGNN, and achieved a high average AUC score at γ = 0.3 , i.e., 82.60%, just slightly less than the 83.23% of EIRS-kGNN. The excellent performance of DynamicCF also encouraged us to improve EIRS-kGNN with the ensemble classifiers of different types in future work.
We also investigated the combined performance of five classification algorithms with various editing methods, in addition to comparing the average AUC scores obtained on each data set. Following the guidance of [43], generated scores were not compared directly, but rather sorted to get a rank for evaluating the performances of classifiers across multiple data sets. Here, editing methods were used to rank instead of classification algorithms to assess the performances of editing methods. We allocated rank one to the editing method with the best performance and rank nine to the method with the worst performance. The ranking results are the averages across the 10 data sets. Hence, the mean rank of each method is a real number within [1.00, 9.00]. Figure 8 depicts the mean ranking results for each combination of different editing methods and classifiers at noise rate γ . Each red column corresponds to the best-performing editing method with the lowest mean rank among the nine methods.
As shown in Figure 8, different methods achieved the best performance in different combinations with classifiers and at different noise rates. Classifying on the noisy data sets generally resulted in the worse classifying performance, especially at γ = 0.3 . At γ = 0.1 , NCR, RD, DynamicCF, and EIRS-kGNN showed improvements over others in the combinations with well selected classifiers: NCR combined with LR, RD with GBDT, DynamicCF with Adaboost and KNN, and EIRS-kGNN with Adaboost and DT achieved the best performance of all in each combination. EIRS-kGNN is prone to retaining more samples in the training set at the cost of some label noise remaining undetected. At the low level of noise rate, its superiority is not that significant. However, the advantages of EIRS-kGNN progressively become apparent as the noise rate γ rises. EIRS-kGNN obtained the highest improvements in 4/5 subfigures at γ = 0.3 .
Figure 8 further demonstrates that EIRS-kGNN could adapt to most classifiers except LR. This is because, in this experiment, the linear LR model was used as the LR classifier. However, EIRS-kGNN retains as many samples as is feasible to recover the true distribution of the data, which can hardly be separated linearly due to the intrinsic intricacy of the data distribution.
In general, EIRS-kGNN shows certain robustness to high levels of label noise; it can not only accurately identify label noise but also recover the true data distribution as if it were not corrupted, thereby improving the data quality and effectively improving the performance of classification with label noise.

4.2.3. Time Efficiency Comparison

We demonstrate the efficiency of the EIRS-kGNN method by evaluating its performance on the 10 UCI data sets and comparing it with other editing methods. Only the time required to detect the label noise is compared among them. As the time efficiency is independent of noise rates and classifiers, here, for simplicity, we directly show the average running time of each method for all noise rates and classifiers on each data set in Table 4.
Table 4 shows that the runtimes of the non-ensemble methods were less than those of the ensemble methods, such as ENN, RENN, AllkNN, and NCR. Due to the requirement of computing the distance matrix of all samples to calculate the relative density value of each sample, the time complexity of RD is O ( n 2 ) , leading to its less efficient results. The ensemble methods, such as IPF, DynamicCF, and EIRS-kGNN, take more time to output the detected results, especially EIRS-kGNN. IPF repeated the classifying process to detect label noise on a gradually decreasing data set, and its maximum number of iterations was set to three by default, so it took the shortest time among the three ensemble methods. DynamicCF trained nine different classifiers to classify a given data set and selected the top three classifiers with the best prediction results to form the ensembler. Its time complexity mainly depended on the time needed by the nine classifiers to classify the data. EIRS-kGNN repeats the label flipping and label-noise detection processes until all suspected label-noise samples are discovered. Its running time is dominated by the convergence speed of the algorithm, i.e., how many iterations are needed. The experimental findings show that EIRS-kGNN typically needs to run 6–8 iterations to converge. Consequently, EIRS-kGNN required the longest average running time, which was approximately seven times that of DynamicCF.
Although it is somewhat time-consuming compared to the aforementioned methods, EIRS-kGNN tries its best to recover the potentially true labels of each data set and can be regarded as a useful auxiliary tool to expensive manual annotation, contributing to obtaining the high-quality data sets while lowering labeling overheads.

4.3. Experimental Results on Multi-Class UCI Data Sets

EIRS-kGNN can be easily extended to multi-class problems. One just needs to predict and record the expected label of each suspected label noise of the inner loop, providing the target label for relabeling operation in the outer loop. Other processes are the same as those in dealing with binary data sets. We employed eight multi-class data sets from the UCI repository, as listed in Table 5, to compare the performances of various editing methods. Similarly to binary classification, each data set was randomly divided into two parts, 80% for training and 20% for testing. To add noise into multi-class data sets, given a noise rate γ , the labels Y i of the ith class were flipped to any other label Y i except Y i , and the number of labels to be flipped was determined by γ and the size of each class.

4.3.1. Performance of Detection of Label Noise in Multi-Class Data Sets

Recall and false metrics were also used with multi-class data sets to estimate the label noise detection performance of the proposed EIRS-kGNN and the performances of other methods. Detailed results are reported in Table 6. The highest scores in recall and lowest scores in false are marked in bold.
In Table 6, it can be seen that the performances of the various editing methods on multi-class data sets were consistent with their performances on binary data sets. ENN and RENN most often achieved the highest recall scores, coupled with high false scores. RD was still the worst one at detecting label noise, with the lowest recall scores and highest false scores for all the data sets. As the noise rate grew from 0.1 to 0.3, AllkNN’s performance remained limited by the high level of label noise; its average recall score decreasing from 88.20% to 69.00%. NCR could achieve comparable performance with the ensemble methods in terms of recall scores, but its higher false scores indicate that it is inclined to remove too many samples from the training set. The three ensemble methods, IPF, DynamicCF, and EIRS-kGNN, had a better balance between detecting more label noise and avoiding removing too many samples. Among the three, EIRS-kGNN outperformed the other two methods, with slightly higher average recall scores and significantly lower false scores at γ = 0.3 , demonstrating its stability and reliability in detecting label noise, even with a high level of noise.

4.3.2. Classification Performance on Multi-Class Data Sets after Editing

Similarly, the classification results were employed to estimate the performances of different editing methods. Two popular strategies, i.e., one against one (OAO) and one against all (OAA), are generally used to decompose the multi-class classification problems into binary ones [44]. We took the one-against-all decomposing strategy to tackle the multi-class problem, avoiding generating too many median classifiers. Five classifiers, LR, DT, Adaboost, kNN, and GBDT, were also employed to ensure that the evaluation of the performances was not limited to a specific classifier. In this work, all the multi-class classification tasks were implemented with the “sklearn.multiclass” package in Python 3.8, with all parameters set to their defaults. The overall performance of multi-class classification is usually estimated in two ways: macro-averaging (averaging the same metrics calculated for different independent binary classifiers C 1 , C 2 , , C m ) or micro-averaging (counting the cumulative t p i and f p i ; t n i and f n i ; and then calculating a performance metric) [45]. A macro-average will compute the metric independently for each class and then take the average, hence treating all classes equally, whereas a micro-average will aggregate the contributions of all classes to compute the average metric and favor bigger classes. Here, we utilize the macro-AUC metric to ensure the reliability of comparing the results of multi-class classification performance. Table 7 illustrates the comparison results of all editing methods for each multi-class data set at various noise rates, with the highest two values marked in bold.
Table 7 demonstrates that various editing methods can help to improve the classification performance to varying degrees in multi-class data sets. Among all the methods, EIRS-kGNN achieved the second highest average AUC scores under different noise rates, and still maintained an average AUC score of approximately 82% at γ = 0.3 , suggesting its reliability and stability in detecting and dealing with label noise. DynamicCF is a highly competent method whose average AUC scores were next to those of EIRS-kGNN at different noise rates. It would be worth learning from its ensemble strategy of diversified classifiers to improve our proposed EIRS-kGNN method. The performance of IPF on multi-class data sets is more stable than its performance on binary data sets, as it is less likely to delete an entire class from a given data set in multi-class data sets. However, IPF is inferior to DynamicCF or EIRS-kGNN, since it relies on only one classifier to repeatedly filter data. Notably, AllkNN had certain advantages at γ = 0.1 and γ = 0.2. When γ = 0.1, AllkNN achieved the highest AUC scores on the data sets glass and seeds; at γ = 0.2, the number of data sets highlighted in bold increased to 5, which is comparable to how EIRS-kGNN performed later. However, when γ grew to 0.3, the enhancement of AllkNN on each data set was less than that of EIRS-kGNN, showing that the resistance of AllkNN to label noise is limited at a high noise level. Like AllkNN, other approaches such as ENN, RENN, NCR, and RD exhibit limited resistance to label noise at high noise levels.

5. Conclusions

In this paper, we proposed an ensemble and iterative recovery strategy-based kGNN (EIRS-kGNN) method to detect and deal with label noise. It displays great label noise detection performance and achieves successful improvements in classification performance by exploiting the ensemble k values and iterative cascade recovery strategy to seek the difficult-to-learn label noise samples. Furthermore, by substituting the unilateral neighborhood rule with the bilateral neighborhood rule, by introducing the general nearest neighbors, the kGNN algorithm can better identify the label-noise samples, lowering the false rate in label noise detection. Moreover, the distribution of the corrupted data sets can be better recovered by using the mixed strategy to process the label noise, meaning definite noise samples relabeled and overlapping samples removed. Hence, a smoother classification boundary can be obtained through the learning process, effectively improving the classification performance. Experimental results showed that the benefits of the EIRS-kGNN method become increasingly significant at a higher level of noise rate. However, it sacrifices time efficiency to achieve high accuracy in detecting label noise. Considering its convergence speed, it is more suitable for processing small- to moderate-sized data sets. The results encourage us to continue further research on enhancing the time efficiency of the proposed method. Furthermore, impressed by the excellent performance of DynamicCF, we are also inspired to incorporate diversified classifiers in our method to better detect label noise in various data sets in our future work.

Author Contributions

Conceptualization, B.C. and L.H.; methodology, B.C. and L.H.; software, L.H.; validation, B.C.; formal analysis, B.C.; investigation, B.C.; resources, G.W.; data curation, B.C. and L.H.; writing—original draft preparation, B.C.; writing—review and editing, Z.C.; visualization, B.C.; supervision, Z.C. and G.W.; project administration, G.W.; funding acquisition, G.W. and B.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the National Key Research and Development Program of China (grant number 2019QY(Y)0301); State Scholarship Fund of China Scholarship Council (grant number 202008500187); and Doctor Innovative High-end Talents Project of Chongqing University of Posts and Telecommunications (grant number BYJS201901).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available data sets were analyzed in this study. These data sets can be found here: https://archive.ics.uci.edu/ml/index.php, accessed on 28 March 2022.

Acknowledgments

We are grateful for the support of Ling Bai in recommendation of this journal.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Parameters of the Classification Algorithms

Five classifiers were used to ensure that the evaluation of the performances was not limited to a specific classifier: logistic regression (LR), decision tree (DT), adaboost, k-nearest neighbors (kNN), gradient boosting decision tree (GBDT). The parameters of all classifiers in “scikit-learn” packages in Python 3.8 were set defaults as follows:

References

  1. Zhu, X.; Wu, X. Class noise vs. attribute noise: A quantitative study. Artif. Intell. Rev. 2004, 22, 177–210. [Google Scholar] [CrossRef]
  2. Frénay, B.; Verleysen, M. Classification in the presence of label noise: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2013, 25, 845–869. [Google Scholar] [CrossRef] [PubMed]
  3. Bi, Y.; Jeske, D.R. The efficiency of logistic regression compared to normal discriminant analysis under class-conditional classification noise. J. Multivar. Anal. 2010, 101, 1622–1637. [Google Scholar] [CrossRef] [Green Version]
  4. Laird, P.D. Learning from Good and Bad Data; Springer Science & Business Media: New York, NY, USA, 2012; Volume 47. [Google Scholar]
  5. Abellán, J.; Masegosa, A.R. Bagging decision trees on data sets with classification noise. In Proceedings of the International Symposium on Foundations of Information and Knowledge Systems, Sofia, Bulgaria, 15–19 February 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 248–265. [Google Scholar]
  6. Bross, I. Misclassification in 2 × 2 tables. Biometrics 1954, 10, 478–486. [Google Scholar] [CrossRef]
  7. Wilson, D.R.; Martinez, T.R. Reduction techniques for instance-based learning algorithms. Mach. Learn. 2000, 38, 257–286. [Google Scholar] [CrossRef]
  8. Wilson, D.L. Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans. Syst. Man Cybern. 1972, SMC-2, 408–421. [Google Scholar] [CrossRef] [Green Version]
  9. Tomek, I. An Experiment with the Edited Nearest-Neighbor Rule. IEEE Trans. Syst. Man Cybern. 1976, SMC-6, 448–452. [Google Scholar] [CrossRef]
  10. Sánchez, J.S.; Barandela, R.; Marqués, A.I.; Alejo, R.; Badenas, J. Analysis of new techniques to obtain quality training sets. Pattern Recognit. Lett. 2003, 24, 1015–1022. [Google Scholar] [CrossRef]
  11. Koplowitz, J.; Brown, T.A. On the relation of performance to editing in nearest neighbor rules. Pattern Recognit. 1981, 13, 251–255. [Google Scholar] [CrossRef]
  12. Pan, Z.; Wang, Y.; Ku, W. A new general nearest neighbor classification based on the mutual neighborhood information. Knowl.-Based Syst. 2017, 121, 142–152. [Google Scholar] [CrossRef]
  13. Liu, H.; Zhang, S. Noisy data elimination using mutual k-nearest neighbor for classification mining. J. Syst. Softw. 2012, 85, 1067–1074. [Google Scholar] [CrossRef]
  14. Gowda, K.C.; Krishna, G. The condensed nearest neighbor rule using the concept of mutual nearest neighborhood. IEEE Trans. Inf. Theory 1979, 25, 488–490. [Google Scholar] [CrossRef] [Green Version]
  15. Gowda, K.C.; Krishna, G. Agglomerative clustering using the concept of mutual nearest neighbourhood. Pattern Recognit. 1978, 10, 105–112. [Google Scholar] [CrossRef]
  16. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef] [Green Version]
  17. Sánchez, J.S.; Pla, F.; Ferri, F.J. Prototype selection for the nearest neighbour rule through proximity graphs. Pattern Recognit. Lett. 1997, 18, 507–513. [Google Scholar] [CrossRef]
  18. Chaudhuri, B. A new definition of neighborhood of a point in multi-dimensional space. Pattern Recognit. Lett. 1996, 17, 11–17. [Google Scholar] [CrossRef]
  19. Devijver, P.A. On the editing rate of the multiedit algorithm. Pattern Recognit. Lett. 1986, 4, 9–12. [Google Scholar] [CrossRef]
  20. Hattori, K.; Takahashi, M. A new edited k-nearest neighbor rule in the pattern classification problem. Pattern Recognit. 2000, 33, 521–528. [Google Scholar] [CrossRef]
  21. Hart, P. The condensed nearest neighbor rule. IEEE Trans. Inf. Theory 1968, 14, 515–516. [Google Scholar] [CrossRef]
  22. Kubat, M.; Matwin, S. Addressing the curse of imbalanced training sets: One-sided selection. In Proceedings of the ICML, Citeseer, Nashville, TN, USA, 8–12 July 1997; Volume 97, p. 179. [Google Scholar]
  23. Tomek, I. Two Modifications of CNN. IEEE Trans. Syst. Man Cybern. 1976, SMC-6, 769–772. [Google Scholar]
  24. Laurikkala, J. Improving identification of difficult small classes by balancing class distribution. In Proceedings of the Conference on Artificial Intelligence in Medicine in Europe, Cascais, Portugal, 1–4 July 2001; Springer: Berlin/Heidelberg, Germany; pp. 63–66. [Google Scholar]
  25. Xia, S.; Xiong, Z.; Luo, Y.; Dong, L.; Xing, C. Relative density based support vector machine. Neurocomputing 2015, 149, 1424–1432. [Google Scholar] [CrossRef]
  26. Xia, S.; Chen, B.; Wang, G.; Zheng, Y.; Gao, X.; Giem, E.; Chen, Z. mCRF and mRD: Two classification methods based on a novel multiclass label noise filtering learning framework. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 2916–2930. [Google Scholar] [CrossRef] [PubMed]
  27. Chen, B.; Xia, S.; Chen, Z.; Wang, B.; Wang, G. RSMOTE: A self-adaptive robust SMOTE for imbalanced problems with label noise. Inf. Sci. 2021, 553, 397–428. [Google Scholar] [CrossRef]
  28. Liu, Y.; Xia, S.; Yu, H.; Luo, Y.; Chen, B.; Liu, K.; Wang, G. Prediction of Aluminum Electrolysis Superheat Based on Improved Relative Density Noise Filter SMO. In Proceedings of the 2018 IEEE International Conference on Big Knowledge (ICBK), Singapore, 17–18 November 2018; pp. 376–381. [Google Scholar]
  29. Huang, L.; Shao, Y.; Peng, J. An Adaptive Voting Mechanism Based on Relative Density for Filtering Label Noises. In Proceedings of the 2022 IEEE 5th International Conference on Electronics Technology (ICET), Chengdu, China, 13–16 May 2022. [Google Scholar]
  30. Brodley, C.E.; Friedl, M.A. Identifying mislabeled training data. J. Artif. Intell. Res. 1999, 11, 131–167. [Google Scholar] [CrossRef]
  31. Garcia, L.P.F.; Lorena, A.C.; Carvalho, A.C. A study on class noise detection and elimination. In Proceedings of the 2012 Brazilian Symposium on Neural Networks, Curitiba, Brazil, 20–25 October 2012; pp. 13–18. [Google Scholar]
  32. Sluban, B.; Gamberger, D.; Lavraě, N. Advances in class noise detection. In Proceedings of the ECAI 2010, Lisbon, Portugal, 16–20 August 2010; IOS Press: Amsterdam, The Netherlands, 2010; pp. 1105–1106. [Google Scholar]
  33. Zhu, X.; Wu, X.; Chen, Q. Eliminating class noise in large datasets. In Proceedings of the Proceedings of the 20th International Conference on Machine Learning (ICML-03), Washington, DC, USA, 21–24 August 2003; pp. 920–927. [Google Scholar]
  34. Khoshgoftaar, T.M.; Rebours, P. Improving software quality prediction by noise filtering techniques. J. Comput. Sci. Technol. 2007, 22, 387–396. [Google Scholar] [CrossRef]
  35. Biggio, B.; Nelson, B.; Laskov, P. Support vector machines under adversarial label noise. In Proceedings of the Asian Conference on Machine Learning, PMLR, Taoyuan, Taiwan, 13–15 November 2011; pp. 97–112. [Google Scholar]
  36. Jin, R.; Liu, Y.; Si, L.; Carbonell, J.G.; Hauptmann, A. A new boosting algorithm using input-dependent regularizer. In Proceedings of the Twentieth International Conference on Machine Learning, Washington, DC, USA, 21–24 August 2003; Carnegie Mellon University: Pittsburgh, PA, USA, 2003. [Google Scholar]
  37. Khardon, R.; Wachman, G. Noise Tolerant Variants of the Perceptron Algorithm. J. Mach. Learn. Res. 2007, 8, 227–248. [Google Scholar]
  38. Wu, X.; Kumar, V.; Ross Quinlan, J.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.J.; Ng, A.; Liu, B.; Yu, P.S.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef] [Green Version]
  39. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  40. Morales, P.; Luengo, J.; Garcia, L.P.F.; Lorena, A.C.; de Carvalho, A.C.; Herrera, F. The NoiseFiltersR Package: Label Noise Preprocessing in R. R J. 2017, 9, 219. [Google Scholar] [CrossRef] [Green Version]
  41. Xia, S.; Huang, L.; Wang, G.; Gao, X.; Shao, Y.; Chen, Z. An adaptive and general model for label noise detection using relative probabilistic density. Knowl.-Based Syst. 2022, 239, 107907. [Google Scholar] [CrossRef]
  42. Fawcett, T. ROC graphs: Notes and practical considerations for researchers. Mach. Learn. 2004, 31, 1–38. [Google Scholar]
  43. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  44. Galar, M.; Fernández, A.; Barrenechea, E.; Bustince, H.; Herrera, F. An overview of ensemble methods for binary classifiers in multi-class problems: Experimental study on one-vs-one and one-vs-all schemes. Pattern Recognit. 2011, 44, 1761–1776. [Google Scholar] [CrossRef]
  45. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
Figure 1. Illustration of the limitations of kNN. (a) kNN in data sets with label noise. (b) kNN in data sets with asymmetric proximity relations.
Figure 1. Illustration of the limitations of kNN. (a) kNN in data sets with label noise. (b) kNN in data sets with asymmetric proximity relations.
Mathematics 10 02743 g001
Figure 2. The inner loop of EIRS-kGNN to detect label noise.
Figure 2. The inner loop of EIRS-kGNN to detect label noise.
Mathematics 10 02743 g002
Figure 3. The outer loop of EIRS-kGNN to detect label noise.
Figure 3. The outer loop of EIRS-kGNN to detect label noise.
Mathematics 10 02743 g003
Figure 4. Clean distributions of two-dimensional data sets. (a) fourclass. (b) moons. (c) circles.
Figure 4. Clean distributions of two-dimensional data sets. (a) fourclass. (b) moons. (c) circles.
Mathematics 10 02743 g004
Figure 5. Different performances of various editing methods on the fourclass data set.
Figure 5. Different performances of various editing methods on the fourclass data set.
Mathematics 10 02743 g005
Figure 6. Different performances of various editing methods on the moons data set.
Figure 6. Different performances of various editing methods on the moons data set.
Mathematics 10 02743 g006
Figure 7. Different performances of various editing methods on the circles data set.
Figure 7. Different performances of various editing methods on the circles data set.
Mathematics 10 02743 g007
Figure 8. Mean ranks of various editing methods combined with different classifiers at varying γ .
Figure 8. Mean ranks of various editing methods combined with different classifiers at varying γ .
Mathematics 10 02743 g008
Table 1. Information of the binary data sets.
Table 1. Information of the binary data sets.
DataClassesFeaturesSamplesTrainPosNegTest
bupa2634527616011669
diabetes28768614214400154
ecoli273362682824068
fourclass22862689245444173
haberman233062446517962
heart2132702169612054
image21820861668950718418
newthyroid252151722814443
pima28768614214400154
votes21643534813421487
Table 2. Average results (%) comparison of label-noise detection performances at varying γ .
Table 2. Average results (%) comparison of label-noise detection performances at varying γ .
RecallFalse
ENNRENNAllkNNNCRRDIPFDynamic
CF
EIRS-
k GNN
ENNRENNAllkNNNCRRDIPFDynamic
CF
EIRS-
k GNN
γ = 0.1 96.8696.8688.2086.3625.1280.4583.9586.6181.3981.5074.2777.3990.5557.5652.3455.50
bupa85.1985.1937.0462.9633.3344.4451.8562.9688.5688.5688.8987.5091.0988.7983.9182.60
ecoli100.00100.0096.1588.4611.54100.00100.0094.6181.5682.0772.8379.8294.1246.9450.9446.03
fourclass100.00100.00100.0097.0617.6577.9497.06100.0072.2472.4752.7861.8584.6268.0710.813.41
haberman95.6595.6595.6595.6556.5273.9173.9177.3987.8587.8583.3386.9886.1776.3979.5274.55
heart100.00100.0080.9580.9538.1080.9571.4390.4884.3384.3376.0683.9686.4475.7172.2265.55
image97.5997.5995.1894.5815.0692.1796.9994.4673.8373.8359.8054.2390.0826.4424.4138.59
newthyroid100.00100.00100.00100.006.2593.75100.00100.0075.0075.3865.9672.8896.156.2515.7936.00
pima95.0895.0886.8981.9740.9870.4975.4172.7986.5186.5183.8486.2389.0876.8874.0176.50
votes100.00100.00100.0088.248.8297.0694.1297.0677.3377.3375.3674.1494.8335.2938.4654.89
diabetes95.0895.0890.1673.7722.9573.7778.6976.3986.7086.7083.8286.2892.9374.8673.3376.85
γ = 0.2 95.2695.4577.3582.9534.8982.4682.1782.4371.8972.2264.1569.0779.8443.2042.7840.89
bupa89.0989.0958.1863.6447.2763.6469.0968.3677.7377.7373.9877.2778.6972.4466.3769.19
ecoli98.11100.0090.5792.4528.3086.7986.7988.6865.5665.8157.8970.4882.3520.6931.3429.21
fourclass99.2799.2798.5490.5128.4778.1096.3599.2767.7067.7051.6159.7476.7954.0820.000.73
haberman95.8395.8385.4285.4247.9272.9266.6763.7577.0077.0073.7277.7279.0959.7764.0463.58
heart88.3788.3779.0793.0239.5383.7279.0783.7275.1675.1663.0469.7078.7552.6351.4341.16
image98.2098.2076.2887.3926.7391.8992.7993.9968.3168.3153.9050.0079.7319.6919.5321.86
newthyroid100.00100.0090.9181.8224.2493.9490.9190.3062.9265.9862.9668.9780.4916.2228.5725.13
pima93.4493.4460.6672.1336.8978.6972.9569.8377.0277.0271.2176.7281.4057.7159.1762.27
votes98.5398.5369.1285.2929.4197.0694.1289.7170.6170.6162.4064.8580.5819.5128.8935.51
diabetes91.8091.8064.7577.8740.1677.8772.9576.7276.8676.8670.7475.2680.4859.2358.4160.30
γ = 0.3 91.8391.8369.0075.1640.2077.4778.9379.5465.1865.1857.5663.0170.9737.8038.2631.67
bupa92.6892.6863.4168.2943.9064.6371.9566.8368.7268.7264.6365.2272.7358.2752.0353.37
ecoli91.2591.2596.2573.7531.2590.0083.7585.2561.7861.7850.3265.0975.7321.7434.3119.35
fourclass97.0997.0971.8484.4736.8978.6489.3299.5161.6161.6150.9957.1469.3536.9623.650.68
haberman86.1186.1176.3969.4440.2873.6165.2863.8968.6968.6963.5870.9372.9046.4657.2752.08
heart89.0689.0653.1273.4439.0670.3167.1976.8765.6665.6660.9264.9375.2543.7553.2634.04
image96.0096.0072.6082.2042.8078.0092.4089.6063.9663.9652.8050.1269.5241.7923.8919.86
newthyroid94.1294.1282.3578.4341.1894.1296.0893.7362.7962.7952.8163.3068.1811.1112.5019.17
pima88.0488.0454.8971.7443.4872.8371.7463.5967.9267.9263.9367.4971.4347.8648.4451.49
votes94.2394.2357.6975.9637.5088.4683.6586.7363.9763.9757.4559.4966.3822.0328.6919.04
diabetes89.6789.6761.4173.9145.6564.1367.9369.4666.7366.7358.1566.4268.1848.0248.5647.60
Average94.6594.7178.1881.4933.4080.1381.6882.8672.8272.9765.3269.8280.4546.1944.4642.69
Table 3. Average AUC (%) comparison of classification results at varying γ .
Table 3. Average AUC (%) comparison of classification results at varying γ .
NSYENNRENNAllkNNNCRRDIPFDynamic
CF
EIRS-
kGNN
γ = 0.1
bupa67.7868.4868.5569.1568.2365.7852.1562.1764.78
ecoli81.2981.920.0083.4083.3080.4475.3681.7187.52
fourclass92.9294.1493.9894.0594.4494.9781.9294.6094.83
haberman57.6565.9265.9258.7465.6466.3368.5765.4670.10
heart78.4483.6983.6982.5187.3886.8691.4987.1991.38
image92.3194.0794.1094.5894.7394.8794.1795.0192.59
newthyroid91.4391.3589.7689.2991.2790.4091.3990.0889.48
pima74.3674.6575.0375.0877.8678.6572.0077.9176.52
votes93.5095.6695.9593.6395.6696.6595.7796.2395.50
diabetes77.1579.1679.3578.3779.1379.2977.3680.2580.79
Average80.6882.9074.6381.8883.7683.4280.0283.0684.35
γ = 0.2
bupa71.5265.6665.6670.8668.3169.8557.1872.3470.47
ecoli82.7677.0782.0481.0882.9980.9482.2073.7979.86
fourclass90.8493.7193.7195.5793.9095.8174.6695.0196.35
haberman68.7473.1973.1970.6073.1571.660.0072.2373.30
heart73.8278.1177.2879.0078.6780.0483.6977.7684.74
image90.2893.5293.5293.2594.6494.8495.3395.6394.35
newthyroid89.2896.9875.7175.7197.6294.2986.3187.0693.06
pima72.2271.3571.4876.2075.1476.0674.0276.7474.87
votes92.8797.1097.1092.7497.3696.3697.8698.5097.51
diabetes71.3770.3171.0569.6974.1569.9572.7775.3473.08
Average80.3781.7080.0780.4783.5982.9872.4082.4483.76
γ = 0.3
bupa54.6861.9561.0756.9057.2759.0751.6563.6665.25
ecoli80.8796.0295.9388.2789.3288.830.0088.1595.63
fourclass85.9686.1886.2086.8990.3290.7079.5292.3495.65
haberman52.7650.6950.3958.0656.3556.450.0062.0461.10
heart81.8681.1780.7584.2686.5790.1082.2487.4495.26
image84.0488.5288.5288.7391.4689.8778.2493.5693.12
newthyroid74.2587.5086.6394.4091.8392.6698.4996.2790.87
pima70.6171.9471.9072.3173.4468.8272.1678.4674.52
votes86.7191.2291.1987.2592.4591.2795.9093.3196.70
diabetes65.7467.6667.4465.9467.3170.6869.2570.8068.65
Average73.7578.2978.0078.3079.6379.8462.7482.6083.23
Table 4. Average running time (s) comparison.
Table 4. Average running time (s) comparison.
ENNRENNAllkNNNCRRDIPFDynamicCFEIRS-kGNN
bupa0.00400.00400.00400.01710.57312.82404.663115.5996
ecoli0.00460.00820.01450.01380.48461.53133.687511.9451
fourclass0.00500.00840.01160.03643.55481.93048.470346.4174
haberman0.00330.00390.00890.01380.47981.86603.617011.7543
heart0.00490.00760.00530.01760.34605.572110.22619.1362
image0.07540.08190.07740.201420.05882.16621.2292252.4761
newthyroid0.00400.00960.01100.01030.25541.66952.31696.1508
pima0.01080.00830.00830.04492.53372.623412.090349.4904
votes0.00490.00500.00880.01960.85742.50509.069919.4140
diabetes0.01280.01190.01440.04012.78522.405212.579752.3045
Average0.01300.01490.01640.04153.19292.50936.795047.4688
Table 5. Information about the multi-class data sets.
Table 5. Information about the multi-class data sets.
DataSamplesClassesFeaturesEach Class
glass21469[7, 23, 56, 60, 13, 10]
iris15034[40, 40, 40]
newthyroid21535[120, 28, 24]
seeds21037[56, 56, 56]
segmentation2310716[264, 264, 264, 264, 264, 264, 264]
vertebralColumn31036[48, 80, 120]
wine178313[47, 56, 38]
yeast1484108[370, 343, 195, 35, 130, 24, 16, 40, 28, 4]
Table 6. Average results (%) comparison of label-noise detection performances on multi-class data sets at varying γ .
Table 6. Average results (%) comparison of label-noise detection performances on multi-class data sets at varying γ .
RecallFalse
DataENNRENNAll k NNNCRRDIPFDynamic
CF
EIRS-
k GNN
ENNRENNAll k NNNCRRDIPFDynamic
CF
EIRS-
k GNN
0.10100.00100.0093.6797.3427.8297.3096.1893.9080.6880.6867.6454.1890.1247.7851.0052.24
glass100.00100.0093.3393.3340.00100.00100.0088.0086.7386.7379.1077.0591.0475.4177.9480.45
iris100.00100.0091.67100.0025.0083.33100.00100.0073.9173.9157.6940.0083.3344.4425.0028.53
newthyroid100.00100.0087.50100.0018.75100.0093.7591.2576.1276.1261.1140.7488.4640.7440.0050.89
seeds100.00100.00100.00100.006.67100.0093.3397.3379.4579.4557.1446.4396.3037.5046.1547.86
segmentation100.00100.0089.5698.9015.3899.4598.3597.8074.7674.7655.7135.7190.1123.6331.6836.51
vertebralColumn100.00100.0091.6791.6745.83100.0095.8389.1786.2986.2977.5577.5587.9160.6663.4973.86
wine100.00100.00100.00100.008.33100.0091.67100.0079.6679.6669.2333.3394.1220.0042.1118.46
yeast100.00100.0095.6594.7862.6195.6596.5287.6688.5388.5383.5882.6489.6879.8581.6281.39
0.2099.0899.0889.7391.0538.8795.3094.1491.1972.2372.2357.1646.5380.3937.6839.5437.51
glass100.00100.00100.0096.8843.75100.0096.8889.3775.3875.3862.7962.6584.2759.4964.7763.62
iris100.00100.0091.6783.3320.8395.8395.83100.0070.3770.3751.1142.8686.1128.1214.8115.42
newthyroid100.00100.0096.9793.9439.3996.9790.9188.4968.2768.2753.6232.6175.0021.9534.7835.11
seeds96.9796.9778.7981.8227.2784.8590.9190.3071.6871.6851.8538.6479.5530.0028.5732.27
segmentation99.7399.7389.2998.0830.4999.7398.0897.5867.5667.5646.8129.5979.9313.1618.8618.15
vertebralColumn95.9295.9275.5181.6353.0691.8495.9284.0875.5275.5267.2661.5476.3643.7550.0057.44
wine100.00100.0096.30100.0033.33100.0088.89100.0070.9770.9751.8535.7180.4341.3035.1410.00
yeast100.00100.0089.3292.7462.8293.1695.7379.6678.0778.0771.9568.6481.4963.6769.4068.07
0.3098.5798.5784.2389.0654.9392.5893.8990.8264.8364.8351.6142.6169.0934.6630.5627.12
glass97.9297.9297.9293.7575.0097.9291.6789.1768.8768.8753.4751.6166.6751.5557.6949.99
iris97.2297.2280.5688.8950.0086.1197.2296.1164.2964.2945.2836.0066.6724.3920.455.47
newthyroid98.0498.0484.3196.0835.2996.0890.2085.8863.7763.7750.5731.9475.6822.2231.3421.79
seeds100.00100.0079.1785.4247.9279.1789.5892.0963.3663.3648.6537.8867.1435.5915.6923.25
segmentation100.00100.0088.6195.4845.9399.6498.5597.6561.5261.5244.2530.8071.2314.7116.1512.31
vertebralColumn95.9595.9574.3277.0356.7690.5494.5982.1666.6766.6758.6557.4670.8341.2335.1945.90
wine100.00100.0078.0582.9356.1095.1292.68100.0061.6861.6850.7738.1863.4936.0711.634.20
yeast99.4399.4390.9192.9072.4496.0296.5983.5268.5068.5061.2157.0370.9951.5156.3554.02
Table 7. Average macro-AUC comparison (%) of classification results at varying γ .
Table 7. Average macro-AUC comparison (%) of classification results at varying γ .
DataNSYENNRENNAllkNNNCRRDIPFDynamic
CF
EIRS-
kGNN
γ = 0.1
glass69.2962.0762.0770.4869.2766.5463.2065.8765.34
iris91.5089.8389.8390.6791.0091.0091.0088.5094.83
newthyroid77.4382.4082.4979.2273.8273.8575.8378.3476.14
seeds87.3887.5086.9091.6790.8390.8388.4591.0790.24
segmentation92.7794.3294.2594.6095.1995.3095.3995.3994.73
vertebralColumn78.1576.5076.6980.7082.6181.3379.4181.5683.35
wine91.9389.9688.8887.8890.3590.1390.9790.5992.62
yeast66.1568.1068.0370.0369.2969.8963.8770.0367.01
Average81.8381.3381.1483.1682.7982.3681.0282.6783.03
γ = 0.2
glass60.5857.0357.1264.3761.6562.7361.1463.6464.19
iris90.1791.5091.5090.3390.3391.0094.1791.1794.83
newthyroid74.9272.7771.7576.8075.8376.4374.4675.8374.20
seeds85.8387.5088.1089.8889.4089.4090.0087.8688.57
segmentation89.0792.4692.3593.7094.5394.4394.3194.9794.61
vertebralColumn74.5675.6775.6081.0779.8077.6479.4777.9780.43
wine84.1586.8785.5689.8990.7690.0590.4592.8492.26
yeast63.2864.9865.0369.1469.0969.5164.0767.9967.94
Average77.8278.6078.3881.9081.4281.4081.0181.5382.13
γ = 0.3
glass60.4557.5058.8767.5664.9766.6959.9468.1764.15
iris82.8385.5086.3387.3390.6790.3393.3392.0094.83
newthyroid68.4171.3170.6773.3971.9373.2172.1877.5574.68
seeds80.7185.6086.0784.2988.1086.1989.2984.6488.10
segmentation83.8191.0591.2192.2593.8394.2693.9994.7994.37
vertebralColumn70.2072.0572.6177.7471.9878.7978.2576.9579.31
wine83.9886.8986.2383.6685.7488.9691.1290.5593.48
yeast57.8558.6559.0264.3268.5267.2861.3864.4166.71
Average73.5376.0776.3878.8279.4780.7179.9481.1381.95
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, B.; Huang, L.; Chen, Z.; Wang, G. An Ensemble and Iterative Recovery Strategy Based kGNN Method to Edit Data with Label Noise. Mathematics 2022, 10, 2743. https://doi.org/10.3390/math10152743

AMA Style

Chen B, Huang L, Chen Z, Wang G. An Ensemble and Iterative Recovery Strategy Based kGNN Method to Edit Data with Label Noise. Mathematics. 2022; 10(15):2743. https://doi.org/10.3390/math10152743

Chicago/Turabian Style

Chen, Baiyun, Longhai Huang, Zizhong Chen, and Guoyin Wang. 2022. "An Ensemble and Iterative Recovery Strategy Based kGNN Method to Edit Data with Label Noise" Mathematics 10, no. 15: 2743. https://doi.org/10.3390/math10152743

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop