Next Article in Journal
In Situ Evaluation of the Self-Heating Effect in Resistance Temperature Sensors
Previous Article in Journal
Received Signal Strength Indicator Measurements and Simulations for Radio Frequency Identification Tag Identification and Location in Beehives
Previous Article in Special Issue
Embedded Vision System for Thermal Face Detection Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tripartite: Tackling Realistic Noisy Labels with More Precise Partitions

1
School of Arts and Sciences, Beijing Normal University, Beijing 100875, China
2
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
3
Guangzhou Institute of Technology, Xidian University, Guangzhou 510555, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3369; https://doi.org/10.3390/s25113369
Submission received: 4 April 2025 / Revised: 22 May 2025 / Accepted: 23 May 2025 / Published: 27 May 2025
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems)

Abstract

:
Samples in large-scale datasets may be mislabeled for various reasons, and deep models are inclined to over-fit some noisy samples using conventional training procedures. The key solution is to alleviate the harm of these noisy labels. Many existing methods try to divide training data into clean and noisy subsets in terms of loss values. We observe that a reason hindering the better performance of deep models is the uncertain samples, which have relatively small losses and often appear in real-world datasets. Due to small losses, many uncertain noisy samples are divided into the clean subset and then degrade models’ performance. Instead, we propose a Tripartite solution to partition training data into three subsets, uncertain, clean and noisy according to the following criteria: the inconsistency of the predictions of two networks and the given labels. Tripartite considerably improves the quality of the clean subset. Moreover, to maximize the value of clean samples in the uncertain subset and minimize the harm of noisy labels, we apply low-weight learning and a semi-supervised learning, respectively. Extensive experiments demonstrate that Tripartite can filter out noisy samples more precisely and outperforms most state-of-the-art methods on four benchmark datasets and especially real-world datasets.

1. Introduction

The success of DNNs in AI-based computer vision systems—such as autonomous vehicles and surveillance networks—relies heavily on large-scale datasets with precise human annotations. However, acquiring high-quality labeled data for sensor-driven applications is prohibitively costly and time-consuming. To address this, practitioners often resort to cost-efficient alternatives such as web-crawling [1], online queries [2,3,4], crowdsourcing [5,6], and so on. Unfortunately, these methods introduce label noise due to sensor-specific challenges: motion-blurred LiDAR scans may be mislabeled as static objects, or occluded pedestrians in camera footage might be ambiguously annotated. Studies [7,8,9,10] demonstrate that DNNs can easily over-fit to noise, which significantly degrades their generalization ability, severely degrading their ability to generalize to real-world scenarios—for instance, failing to recognize partially obscured traffic signs or misclassifying low-resolution objects in edge-device deployments [11,12].
There have been many efforts [13,14,15] to tackle noisy labels. Most of them reach a consensus that the key is to alleviate the impact of noisy labels for network training. Loss-correction-based methods [16,17,18,19] aim to rectify the losses of noisy samples during training but may mistakenly rectify some clean samples. Sample-selection based methods [20,21,22,23] tackle noisy labels by partitioning the training data into the clean and noisy subsets separately, then using them for network training in different ways. These bipartition methods commonly utilize the following partition criteria: (1) The Small-loss criterion  [20] assumes that the losses of noisy samples are significantly greater than those of the clean samples during training. Therefore, they try to find a threshold, T, and select the samples, whose losses < T , as clean samples. The others are treated as noisy samples. (2) The Gaussian Mixture Model (GMM) criterion  [21] believes that the statistical distribution of noisy samples’ losses is different from that of clean samples’ losses. It aims to find the probability of a sample being noisy or clean by fitting a mixture model. (3) The consistency criterion  [23,24,25] assumes that noisy samples exhibit inconsistencies in predictions or inconsistencies in label distributions when compared to their neighboring samples. Based on this, the samples are divided into clean and noisy subsets.
However, in the early stage of training, the network’s capacity is limited, posing a considerable challenge of identifying some samples as clean or noisy. We refer to samples that are difficult to correctly classify as either clean or noisy as uncertain samples. Our observations indicate that a substantial cohort of uncertain samples are distributed close to the classification boundaries, e.g., samples grouped by the red curve shown in Figure 1a, especially in real-world datasets. Given that uncertain samples encompass both noisy and clean instances, bipartition methods may misclassify some noisy uncertain samples as clean. Subsequently, as vision systems fit these samples during training, they may result in smaller losses, as illustrated in the green area between the two red dashed lines in Figure 1b. These samples are very likely to be included in the clean subset again and then be over-fitted by the vision system during later training phases. This over-fitting can lead to performance degradation in the vision system, particularly in applications where robustness to label noise is critical. In the context of AI-based computer vision sensors and systems, the challenge of handling uncertain samples becomes even more pronounced.
In this case, we face a challenging issue: loss-based methods cannot accurately divide samples when some noisy samples also exhibit smaller losses. This is particularly problematic in AI-based computer vision systems, where uncertain samples are often distributed close to the classification boundaries and are prone to being misclassified into either the clean or the noisy subset. As a result, traditional bipartition methods struggle to handle uncertain samples effectively, which can lead to performance degradation in vision systems, especially in applications requiring high robustness to label noise. To address these challenges, we propose a novel method, Tripartite, which divides training data into three subsets: clean, noisy, and uncertain. This approach aims to improve the quality of the clean subset by leveraging the consistencies of predictions from two different networks and the given label. In the context of computer vision sensors and systems, this method can enhance the reliability of the system by reducing the impact of label noise on the training process. In the Tripartite method, samples in the uncertain subset typically exhibit lower classification confidence from the networks, meaning that the two networks are likely to give inconsistent predictions. Conversely, samples with higher classification confidence, where the two networks provide consistent predictions, are assigned to either the clean or the noisy subset. If the prediction matches the given label, the sample is placed in the clean subset; if it differs, it is placed in the noisy subset. To effectively utilize the data, we apply a low-weight training strategy for samples in the uncertain subset, a semi-supervised training strategy for samples in the noisy subset using pseudo labels instead of given labels, and a traditional supervised training strategy for samples in the clean subset.
Unlike bipartition methods, which may misclassify noisy, uncertain samples into the clean subset, Tripartite enhances the quality of the clean subset by relocating such samples into the uncertain subset. This approach is specifically designed to minimize the impact of noisy labels within the clean subset and maximize the utility of clean uncertain samples within the uncertain subset, which is crucial for maintaining the robustness and accuracy of AI-based computer vision systems in real-world applications. Extensive experiments demonstrate that Tripartite outperforms state-of-the-art (SOTA) methods in handling label noise, making it particularly suitable for vision sensors and systems that require high reliability.
In particular, sensor-based data systems—such as wearables, autonomous platforms, or IoT devices—frequently suffer from ambiguous annotations due to noise from environmental conditions, hardware limitations, or manual labeling errors. These challenges often lead to boundary samples or class confusion in readings (e.g., transition between similar motion states), which are difficult to handle using conventional clean/noisy bipartition schemes. The Tripartite framework is well suited for such sensor environments. By explicitly introducing an “uncertain” subset and applying low-weight and semi-supervised strategies, Tripartite effectively isolates ambiguous samples and enhances training robustness. Moreover, the proposed realistic noise generation approach mimics the semantic confusion patterns often found in sensor data, making the framework adaptable to multimodal, dynamic, or edge-deployed AI systems.
The key contributions of our work are threefold:
  • We propose a novel partition criterion, which divides training data into three subsets: uncertain, noisy, and clean. It alleviates the uncertain sample division problem of bipartition methods and minimizes the harm of noisy labels by improving the quality of the clean subset.
  • We design a low-weight training strategy that aims to maximize the value of clean samples in the uncertain subset while minimizing the influence of potential noise. This strategy is particularly beneficial for vision systems operating in environments with ambiguous or noisy sensor data.
  • To simulate the characteristics of noisy labels in real-world datasets, we design a synthetic class-dependent label noise model on CIFAR datasets, referred to as realistic noise. This model flips labels of samples from two different classes at controlled ratios based on their class similarity, providing a more accurate representation of label noise challenges faced by computer vision systems in practical scenarios.

2. Related Work

There have been many studies to address noisy labels. They all try to lower the impact of noisy labels to improve the classification performance of methods during training. To this end, two different ideas were proposed: loss correction and sample selection.

2.1. Loss Correction

The specific methods include noise transition matrix  [16,27,28,29], robust loss functions  [30,31,32,33,34], label correction  [19,35,36], etc. Noise transition matrix methods construct the label transition matrix to estimate the possibilities of noisy labels transitioning among multiple classes. F-correction [16] performs forward correction by multiplying the transition matrix with the softmax outputs in the forward propagation. Later, Hendrycks et al. [37] improved the corruption matrix using a small clean dataset. Robust-loss methods aim to design loss functions that are robust to noisy labels, such as Mean Absolute Error (MAE) [30]; Improved MAE [32], which is a reweighted MAE; Generalized Cross-Entropy loss (GCE) [33], which is a generalization of MAE; and Symmetric Cross-Entropy [34], which adds a reverse cross-entropy term to the usual cross-entropy loss. Label correction methods try to rectify the noisy labels according to the network predictions during the training. Joint optimization [19] learns network parameters and infers the true labels simultaneously. PENCIL [35] adopts label probability distributions to supervise network learning and updates these distributions through back-propagation in each epoch. Inspired by “early learning”, ELR+ [36] uses the network predictions in the early training stage to correct the noisy labels.
The above methods do not divide training data into clean and noisy samples. They may mistakenly rectify the losses of clean samples and introduce new, noisy labels into the training data. Therefore, sample selection methods were proposed.

2.2. Sample Selection

These methods try to divide training data into “clean” and “noisy” subsets according to a specific partition criterion. Afterward, they apply different strategies to train the model on two subsets separately. The existing partition criteria are mainly based on the training loss, e.g., the small-loss criterion  [20], the Gaussian Mixture Model (GMM) criterion  [21] and the consistency-based criterion. The small-loss criterion selects training samples with small loss as clean ones by setting a threshold T. In particular, MentorNet [38] pretrains a teacher network to select clean samples to guide the training of a student network. Co-teaching [20] trains two networks where each network selects small-loss samples to feed its peer network for updating parameters. Furthermore, Co-teaching+ [22] emphasizes the help of inconsistency to the network. However, all these methods only use the data with smaller losses for training without considering the data with larger losses. Meanwhile, finding a feasible T is very challenging. GMM assumes that the distributions of losses of noisy samples and clean samples follow two normal distributions, respectively. DivideMix [21] employs two networks, each of which selects data for the other in the training and then applies a semi-supervised method to learn the noisy samples. This greatly improves the performance of networks. However, in real-world datasets, the noisy labels often appear between similar classes, such as whale and dolphin, and oak_tree and maple_tree in CIFAR-100. GMM may mistake the noisy, uncertain samples and the clean, uncertain samples because of the small difference between their losses. This leads to a low quality of training data partition. TCL [39] presents a novel twin contrastive learning model, which constructs a GMM model over the representations and detects the examples with wrong labels by another two-component GMM and then infers the posterior probability of one example having clean labels through the two-component GMM. It can learn discriminative representations aligned with estimated labels through mixup and contrastive learning, but the posterior probability may not be precise for some uncertain samples. PES [40] finds that the latter layers in a DNN are much more sensitive to label noises, so it separates a DNN into different parts and progressively trains them. It selects a training sample into the clean subset or the noisy subset according to whether the network prediction is consistent with the given label and then applies a semi-supervised learning on the noisy subset. DLT [41] proposes a novel noisy label detection method that records the loss value of each sample and calculates dynamic loss thresholds, then uses these thresholds to detect noisy labels. These two methods may mistake the noisy, uncertain samples and the clear, uncertain samples as well. The consistency-based criterion relies on the inconsistency in predictions or label distributions of noisy samples to select data. JoCoR [23] emphasizes the consistency of the two networks. TC-Net [24] constrains the transform consistency between the original images and the transformed images for network training. NCE [25] determines whether a candidate is a noisy sample by estimating the inconsistency in the label distribution between the candidate sample and its contrastive neighbors. These methods still partition samples into clean and noisy subsets, thereby risking confusion between noisy, uncertain samples and clean, uncertain samples. To address the web label noise, DSOS [42] is proposed to divide data into clean, in-vocabulary noise, and out-of-vocabulary noise subsets. It reported the best performance among SOTA methods on web-crawled datasets without considering uncertain samples. By contrast, we propose the Tripartition criterion, which divides training data into clean, uncertain, and noisy subsets. A more precise partition can effectively lower the harm of noisy labels for network training. More recent efforts have also explored end-to-end frameworks to address the challenges of noisy labels. For instance, SplitNet [43] introduces a learnable clean–noisy partitioning strategy that dynamically assigns training samples into subsets based on representation-level features. It enables more adaptive filtering during training without relying on static thresholds. Additionally, AEON [44] tackles instance-dependent and out-of-distribution noise through an adaptive estimation mechanism that separates the noise types and applies targeted training objectives. These approaches further emphasize the importance of noise-aware partitioning and provide complementary perspectives to our proposed Tripartition strategy.

3. The Design of Realistic Label Noise

In real-world noisy datasets, noisy labels often come from similar classes. We observe that many noisy labels are introduced into the aforementioned uncertain data. However, symmetric noise, which randomly replaces labels with other labels, ignores inter-class similarities. Thus, while many existing methods achieve remarkable performance on such types of label noise, they may not perform well on real-world noisy datasets. Recent studies have increasingly addressed label noise across different domains. For example, Liu et al. [45] proposed a discriminative approximate regression projection method to enhance robust feature extraction. Wang et al. [46] adopted a meta-learning framework to optimize lifetime prediction under label noise in battery data. Zhang et al. [47] tackled label noise in cross-modal person re-identification using deep learning. These methods emphasize robustness in feature learning or task adaptation, which complements our focus on partitioning strategies under noisy supervision. Therefore, we propose a new noise label on CIFAR datasets called realistic noise to simulate real-world noise better. It is used to evaluate the performance of all methods and explain the mechanism of Tripartite. It flips labels in class pairs at a certain ratio, which is proportional to the similarity between two classes. The more similar the two classes are, the more noisy labels are introduced. Although the asymmetric noise in previous works also attempts to mimic the structure of real-world label noise, the assigned classes are often defined based on intuitive judgments rather than the nature of the data. In contrast, our proposed realistic noise calculates the similarity between classes and transfers noisy labels, better simulating the noise in the real world. The detailed procedure for generating the noise transition matrix of realistic noise is shown below.
Step 1. We first train a ResNet50 on CIFAR-10 and CIFAR-100 until it converges. The trained model achieves 94.67% test accuracy on CIFAR-10 and 78.85% on CIFAR-100. We then consider the weight vector between the last fully connected layer and a node (a class) in the output layer as the prototype of the class [48].
Step 2. We pair all prototypes into N class pairs ( c i , c j )   i j , then compute their cosine similarities and sort them in descending order. We select the top K pairs and flip some labels in each pair, where K is set to 10 for CIFAR-10 (Table 1) and 60 for CIFAR-100 (Table 2).
Step 3. The K pairs are partitioned into three similarity levels. Each level has a weight w, which is proportional to the similarity between two classes in the pair. The higher w is, the more labels will be flipped. w is selected from { 0.9 , 0.6 , 0.3 } for CIFAR-100, and { 0.9 , 0.8 , 0.7 } for CIFAR-10. As more similar pairs are selected from CIFAR-100, the weight step between different levels becomes larger. If the similarity of a class pair is not in the top k pairs, the w is set as 0.
Step 4. To construct the noise transition matrix, we temporarily fill in the positions of the top K class pairs with the corresponding weights w i , j , where i and j denote the i t h row and j t h column in the matrix. If a row is empty, we will fill the non-diagonal position uniformly with 1. Other positions are filled in with 0. In order to make the sum of all weights in each row to be 1, the w i , j is replaced by
w ^ i , j = w i , j j = 1 n w i , j ,
where n is the length of columns.
Step 5. Given a noise ratio r, we fill in the diagonal of the matrix with ( 1 r ) and multiply the other elements by r.
Figure 2 shows the generated noise transition matrices of CIFAR-10 with noise ratios of 20% and 50%, respectively.
Figure 3 visualizes the feature distributions of three similar class-pairs in CIFAR-100 using t-SNE. We train a ResNet50 for 200 epochs on CIFAR-100 without noisy labels and use the output of the last convolutional layer as the feature of each sample. We can see that many uncertain samples are distributed around the classification boundary between two similar classes.

4. The Proposed Method

To explain the mechanism of Tripartite, we first analyze the distributions of uncertain, clean, and noisy samples and then derive the logic of the data partition in Section 4.2. Section 4.3 details the proposed Tripartition criterion. Section 4.4 presents the training strategies for three subsets.

4.1. Preliminaries

In our method, we train two networks, denoted by f ( θ k ) , k = { 1 , 2 } , which are initialized with different weights. We assume that they can learn not only the reliable information from training data but also some different views, which may result in the inconsistent predictions of networks on uncertain samples. P k is the predicted label of f ( x ; θ k ) on an input x. The proposed methods mentioned hereinafter are all based on the predictions of these two networks.
Compared to DISC [50], which uses prediction consistency for binary partitioning, Tripartite introduces a third “uncertain” subset and applies subset-specific learning strategies. This improves robustness under ambiguous or instance-dependent noise, where binary categorization may be insufficient.

4.2. Distribution of Training Data and the Logic of Tripartite

For an understanding of the noisy label problem on realistic datasets, we visualize the feature distribution of samples from the realistic CIFAR-10 in Figure 1a, where each color represents a class, and a black dot denotes a noisy sample. One can see that each class has a clear cluster. There are also many samples distributed between clusters; e.g., the dots grouped by the red curve. These are the aforementioned uncertain samples, which have some shared features of two or even more classes. It is very challenging for networks to distinguish them as clean or noisy samples. We also observe a small portion of uncertain samples that are very special. As two networks in our method could learn different views from a sample, these special samples may be distributed close to the classification boundary in the feature space of one network, but distribute in the clusters in the feature space of another network. The detailed discussion is given in Section 5.4.
To simplify the discussion, we analyze the distribution in the feature space of one network. Let us take the categories cat and dog as an example, where 50 samples are uniformly selected from each category for a clearer visualization, as shown in Figure 4. The samples on the outside of the bounding box in Figure 4b are the certain samples. They belong to two cases. The first is certain samples with clean labels, which are red dots and blue triangles. As the label information is consistent with the sample features, networks will fit these samples in the early training stage. Moreover, they are distant from the classification boundary, so networks can correctly recognize them. The second is certain samples with noisy labels, which are blue dots, red triangles, and black stars. As the information provided by the given labels is inconsistent with the features of these samples, networks would not fit them in the early training stage. These samples could also be correctly predicted by both networks, but the prediction is obviously inconsistent with the given label.
The samples in the bounding box in Figure 4b are considered as the uncertain samples. They belong to three cases. The first is uncertain samples with clean labels, which are red dots and blue triangles. Although the label information is correlated with the sample features, they have some shared features between the similar categories due to being distributed close to the classification boundary. The second is uncertain samples that are mislabeled into similar categories, which are blue dots and red triangles. They also have some shared features between the similar categories for the same reason. The third is uncertain samples that are mislabeled into dissimilar categories, which are black stars. Their features are irrelevant to the information provided by the given labels, so their labels are similar to the symmetric noise, which are transferred to dissimilar categories. As many samples in the aforementioned three cases are distributed close to the classification boundary and have shared features of two categories, two networks in our method are likely to give inconsistent predictions.

4.3. Tripartition Method and Criteria

For the realistic label noise, noisy labels tend to happen between similar categories. Our investigation shows that networks with varied initialization may learn different views from a sample. When two networks give inconsistent predictions, it indicates their classification confidences are low. Then, we consider such samples as uncertain samples. When two networks give a consistent prediction, their classification confidences are usually high. These samples are considered as certain samples and then divided into the clean or noisy subset according to the consistency between the prediction and the given label. The logic of Tripartite is shown in Figure 5. The detailed partition criteria are given below.

4.3.1. The Selection Criterion for the Uncertain Subset

Let p 1 and p 2 denote the predictions of the two networks f ( θ 1 ) and f ( θ 2 ) , respectively.
P 1 P 2
Samples in the uncertain subset mainly have clean or noisy labels of similar classes. Moreover, some uncertain samples have noisy labels of dissimilar classes. As many uncertain samples are distributed close to the classification boundary, different networks likely give inconsistent predictions on such samples, especially in the early training stage. We then apply the inconsistent predictions of two networks, Equation (2), to select them. Our experiments demonstrate that the population of these data will gradually decrease with the increase in the classification confidence of networks.
Although this inconsistency arises from the predictions of two independently initialized networks, such an initialization difference is an intentional design choice (see Section 4.1). It allows the two networks to learn different feature perspectives, which helps identify ambiguous or boundary samples more effectively. As shown in the visualization in Section 5.4, these uncertain samples often appear inconsistent in one network while clustered in another, validating the effectiveness and stability of this criterion despite initialization randomness.

4.3.2. The Selection Criterion for the Noisy Subset

P 1 = P 2 G L
The noisy subset mainly includes certain samples with noisy labels. As the predictions of two networks are consistent and different from the given label, the sample is very likely to be a noisy sample. Meanwhile, the networks will not fit them in the early training stage and can provide a high-confidence prediction; we then apply this inconsistency, Equation (3), to select them.

4.3.3. The Selection Criterion for the Clean Subset

P 1 = G L P 2 = G L
The clean subset mainly includes certain samples with clean labels. As these samples are distributed far from the classification boundary and their features and the label information are rather correlated, networks can learn the mapping between data features and labels in the early training stage. So the predictions of the two networks and the given label should be consistent. We apply this consistency, Equation (4), to select clean samples.
Our realistic noise generation strategy is conceptually related to instance-dependent label noise simulation methods, such as RoG [51]. However, there are fundamental differences. RoG generates noise by first training a deep discriminative model and then fitting a generative classifier (e.g., Gaussian) over the learned feature representations, and finally identifying low-likelihood instances—via the Mahalanobis distance—to relabel them. This approach is highly model-dependent and relies on instance-level statistical estimation over the learned feature space.
In contrast, our method constructs a class-level confusion matrix based on cosine similarity between class prototypes or semantic embeddings. It introduces label noise by probabilistically flipping labels between semantically similar classes without requiring any model training or feature inference. This makes our method simpler, model-agnostic, and more scalable.

4.4. Learning Strategies for the Three Subsets

We design different learning strategies for three subsets, respectively. Let us consider a classification problem with C classes. The training set consists of n examples, where ( x i , y i ) is the i t h input sample and y i { 0 , 1 } C is a one-hot label of C dimensions vector corresponding to the class of x i . The network maps the input x i into a C-dimensional encoding and feeds it into a softmax function to estimate the probability p i of x i for each class.

4.4.1. The Learning Strategy for the Uncertain Subset

The uncertain subset includes samples with either clean labels or noisy labels. We wish to utilize the valuable information of clean samples to improve the discriminative ability of networks, while lowering the harm of noisy labels in this subset. Therefore, we apply a low-weight cross-entropy loss for the network training.
L o s s u = λ u ( 1 n i = 1 n c = 1 C y c i l o g p c i ) ,
where λ u is a weight in the range (0, 1), p c i is the estimated probability of x i being in class c.

4.4.2. The Learning Strategy for the Noisy Subset

To lower the impact of noisy labels, it would be better to reassign a reasonable label for noisy samples. To this end, we utilize the ensemble of two networks’ predictions to “co-guess” labels. This strategy is similar to the method in DivideMix [21], which had reported a reliable “guess”.
q ^ i = 1 2 M m = 1 M f x ^ m i , θ 1 + f x ^ m i , θ 2 ,
q i = q ^ c i 1 T / c = 1 C q ^ c i 1 T , where c = 1 , 2 , , C .
Given a noisy labeled sample x i , we obtain its M augmentations { x ^ 1 i , , x ^ M i } randomly by rotation, flip, cropping, desaturation, contrast, blurring, mixup [52], etc. A new label q ^ i is guessed by Equation (6) before each epoch, where θ 1 and θ 2 are the parameters of two networks, respectively. Then, the guessed label q i is sharpened by Equation (7), where q ^ c i is the guessed probability for class c of sample x i and T is the sharpening temperature. We use the CE loss to train each network.
L o s s n = λ n 1 n i = 1 n c = 1 C q c i log p c i ,
where p i is the label probability of x i predicted by the k t h network during training, p i = f x i , θ k and k = { 1 , 2 } , q i is the guessed label probability done by two networks, and λ n is a parameter to control the contribution of the noisy subset.

4.4.3. The Learning Strategy for the Clean Subset

As Tripartite ensures the high quality of the clean subset, we apply the cross-entropy loss for the network training,
L o s s c = 1 n i = 1 n c = 1 C y c i l o g p c i .
Finally, the total loss is
L o s s t r i = L o s s c + L o s s n + L o s s u .

4.5. Pseudo-Code

The pseudo-code of Tripartite is shown in Algorithm 1.
Algorithm 1: Tripartite
Sensors 25 03369 i001

5. Experiments

5.1. Datasets and Noise Types

We verify the effectiveness of our Tripartite extensively on two datasets with artificial label noises (CIFAR-10 [53], CIFAR-100 [53]) and two real-world datasets (Clothing1M [54], WebVision1.0 [4]).
Both CIFAR-10 and CIFAR-100 contain 60,000 images of size 32 × 32. CIFAR-10 includes 10 super-classes. CIFAR-100 has 100 subclasses. Three different label noises are set on them. (1) Symmetric noise: Labels of a certain percentage of training data are flipped to other labels randomly and uniformly. (2) Instance-dependent noise: The noisy labels are generated by the method proposed in [55]. (3) Realistic noise: The noisy labels are generated based on the similarities between classes introduced in Section 3.
Clothing1M consists of 1 million training images collected from online shopping websites with labels generated according to the surrounding text. Its noise ratio is about 38.5%.
WebVision1.0 is a large-scale dataset with real-world noisy labels. It contains more than 2.4 million images crawled from the web using the 1000 concepts in ImageNet ILSVRC2012 [56]. Its noise ratio is about 20%. For ease of comparison with competing methods, we follow the previous work [57] and then perform the test on a subset comprising the first 50 classes crawled from Google image. The details of these six datasets are listed in Table 3.

5.2. Models and Parameters

Following previous works [7,21,36,40,42], we train the 18-layer PreAct Resnet [26] on CIFAR-10 and CIFAR-100 using the SGD optimizer with a learning rate of 0.02, a momentum of 0.9, and a weight-decay of 0.0005 for 300 epochs. The batch size is set to 128. The architectures of the two networks used in our method are the same. However, the networks are randomly initialized with different weights and are fed different randomly augmented data each time, ensuring their differences in the early stages of training. The learning rate is reduced by a factor of 10 after 150 and 200 epochs. We set the warm-up as 10 epochs for CIFAR-10 with all label noises, 30 epochs for CIFAR-100 with symmetric noise and instance-dependent noise, and 20 epochs for CIFAR-100 with realistic noise. The most critical parameters in Tripartite are λ n and λ u . The λ n is selected from {0.2, 0.3, 0.4, 0.5, 0.6}, λ u is 0.3 on both CIFAR datasets.
For real-world datasets, the experiment settings and parameters are the same as CIFAR datasets except training epochs, learning rate, and network architectures. The batch size is set to 64. For Webvision1.0, we set the warm-up to 20 epochs. We train the network for 100 epochs, and the learning rate is reduced by a factor of 10 after 60 and 100 epochs. For Clothing1M, we set the warm-up to five epochs. We train the network for 60 epochs, and the learning rate is reduced by a factor of 10 after 20 and 40 epochs. Resnet-50 [58], pre-trained on ImageNet (following previous work [7]), is for Clothing1M, and inception-resnet v2 [59] is for WebVsion1.0. The λ u is 0.6 for Webvision1.0 and 0.2 for Clothing1M.
The values of λ n and λ u were determined empirically. For CIFAR datasets, λ n was chosen from the candidate set 0.2, 0.3, 0.4, 0.5, 0.6, and λ u was fixed at 0.3, which provided stable performance across different noise types. For real-world datasets, these parameters were tuned separately based on experimental validation.

5.3. Comparison with SOTA Methods

We compared our method with the baseline Standard CE (the backbone + Cross Entropy), and the recent SOTA methods including Bootstrap [18], F-correction [16], P-correction (PENCIL) [35], Meta-Learning [7], M-correction [60], Co-teaching [20], JoCoR [23], DivideMix [21], ELR+ [36], Co-learning [61], DSOS [42], JNPL [62], PES [40], TCL [39], and DLT [41]. For a fair comparison, we followed previous works and report the combined predictions of two networks on the test set. To ensure statistical reliability, we conducted the experiments five times and report the mean results. Since Tripartite aims at real-world noisy data, in which the noise ratio over 50% is unlikely to happen [61], we focus on the data with the noise ratio less than 50%.

5.3.1. Results on CIFAR-10 and CIFAR-100

Table 4 shows the test accuracy on CIFAR-10 and CIFAR-100 with symmetric noise. We can see that Tripartite outperforms competing methods on CIFAR-100 with all noise ratios and on CIFAR-10 with a 20% noise ratio. Table 5 shows the test accuracy on CIFAR-10 and CIFAR-100 with instance-dependent noise. Tripartite outperforms competing methods at all settings. Table 6 shows the test accuracy on CIFAR-10 and CIFAR-100 with realistic noise. Tripartite achieves a significant performance gain compared to other methods across all noise ratios.

5.3.2. Results on Real-World Datasets

Table 7 and Table 8 list the results on real-world noisy datasets, Clothing1M and WebVision1.0, respectively. We also evaluate the generalization ability of these methods on ImageNet ILSVRC12. Tripartite consistently outperforms competing methods across all large-scale datasets. One can see the performance gains are not very significant on Clothing1M. The possible reason is that some images contain more than one subject. For example, images in Clothing1M are usually full-body shots that include top and bottom clothing but are labeled as either of them. In this case, the networks have difficulty learning the discriminative feature. Hence, Tripartite may partition data with some uncertainty.
As WebVision has more categories and more diverse categories, it allows Tripartite to achieve a considerable improvement than other methods. TCL [39] also performs well on WebVision and ILSVRC12 because of the twin contrastive learning structure. Tripartite outperforms it by about 1% across all criteria. Our investigation reveals that the posterior probabilities inferred by TCL are inaccurate for many uncertain samples that are subsequently selected into the clean subset. Conversely, Tripartite exhibits its robustness in selecting uncertain samples, resulting in a high-quality clean subset.

5.4. Visualization of Uncertain Samples

Selecting uncertain samples plays a crucial role in improving the quality of the clean subset. We think that this is the key reason that Tripartite can achieve better performance in the above experiments, especially on realistic label noise. It is worth having an insight into these samples to better understand the mechanism of our Tripartite.
Figure 6a shows the distribution of clean samples (color dots) and uncertain samples (black dots) in CIFAR-10 in the feature space of network 1 in the later training stage. One can see that most uncertain samples are distributed close to the classification boundary. Therefore, network 1 has lower confidence to classify them into either of the classes.
Meanwhile, a few black dots are distributed in the clusters. We plot them only in Figure 6b. This indicates that network 1 has higher confidence to classify them, which seems to contradict our definition of uncertain samples. We then visualize their distribution in the feature space of trained network 2 in Figure 6c. One can see that they are almost distributed at the classification boundary again. This means that network 2 has lower confidence to classify them into either of the classes. The possible reason is that two networks have varied initialization; then, network 1 learns a better discriminative feature in the current training stage, but network 2 does not, and vice versa. As two networks learn different features from such samples, it is difficult to classify them precisely. We consider them as uncertain samples as well.
In summary, uncertain samples consist of two cases. (1) The majority are distributed close to the classification boundary in both feature spaces of the two networks. (2) The minority are distributed at the classification boundary in the space of one network but in the clusters in the space of another network. Neither of the two cases allows two networks to give the same prediction. Therefore, we apply the partition criterion (2) in the Section 4.3.1 to select uncertain samples.

5.5. The Quality of Training Data Partition

As the quality of the clean subset is the key reason that Tripartite performs better than the competing methods, we test the partition accuracies of clean and noisy subsets on the realistic label noise and the instance-dependent noise. The partition accuracy of the clean subset is defined as
P a r t i t i o n   A c c u r a c y = 1 N u m c i = 1 N u m c I ( G L i = T L i ) ,
where N u m c denotes the number of samples in the clean subset and G L i and T L i are the given label and the ground-truth label of the i t h sample in the clean subset. I ( · ) is an indicator function, where I ( · ) = 1 if · is true; otherwise, I ( · ) = 0 . The partition accuracy of the noisy subset is analogous to Equation (11).
The comparison is conducted among Tripartite, the small-loss criterion [20], and the GMM criterion [21]. The results are shown in Figure 7. We can see that Tripartite largely outperforms the other two criteria on the clean subset. The two loss-based criteria perform worse due to uncertain, noisy samples, which blur the distinction between losses of clean and noisy samples in the training set and then further weakens the discrimination of the two loss-based criteria. Table 9 and Table 10 are the numerical versions of Figure 7.

5.6. The Effectiveness of Tripartition for the Other Methods

It is well worth testing whether the data divided by Tripartition can also help other methods. Then, we conduct a control experiment for the small-loss-based and GMM-based methods. Specifically, we replace their training data with the one partitioned by our Tripartition and apply their own learning strategies to verify if their performance would be improved. Firstly, we train two networks with different initializations (PreActResNet-18) on CIFAR-100 without label noise for 20 epochs. They achieve accuracies of 64.18% and 64.56%, respectively. Later on, networks are trained on CIFAR-100 with 30% instance-dependent noise and realistic noise. Take co-teaching on realistic noise as an example. In the experimental group, co-teaching applies its own learning strategy to the data divided by the small-loss criterion. In the control group, co-teaching uses the same learning strategy on the data divided by our Tripartition criterion. Two networks usually converge after 200 epochs. Their classification accuracies are listed in Table 11. For DivideMix, we follow the same protocol. Networks are trained by the data divided by the GMM criterion and the Tripartition criterion, respectively.
We can observe that our Tripartition criterion improves co-teaching and DivideMix by about 1.04% and 0.35% on instance-dependent noise, respectively, and improves them by about 0.99% and 0.05% on realistic noise. These results indicate that Tripartition selects a better clean subset than Small-loss and thus improves Co-teaching considerably. But we can see Tripartition is less helpful for DivideMix. The reason is that the uncertain subset divided by Tripartition is also learned by a semi-supervised strategy in DivideMix, in which clean samples are likely to be misused. The above experiments demonstrate that our Tripartition does reach a better partition of training data and can improve on other methods.

5.7. Ablation Study

To verify the effectiveness of our training strategies, the ablation study is conducted on CIFAR-100 using symmetric and realistic settings with varied noise ratios.

5.7.1. The Training Strategy for the Uncertain Subset

The weight λ u is a key parameter for training the uncertain subset. Our motivation is to lower the harm of noisy labels; therefore, we test the λ u in the range of (0, 1]; and list results in Table 12. Obviously, λ u = 1 reaches the worst result because it preserves the harm. When λ u = 0.3 , we have the best performance. However, λ u = 0 results in a worse result. The possible reason is that there are some uncertain clean samples, which are critical to the performance of the network. If they are excluded from training, the network can not learn valuable information from them.

5.7.2. The Training Strategy for the Noisy Subset

To evaluate the effectiveness of the training strategy for the noisy subset, we test two strategies and report the results in Table 13. “Drop” means the noisy subset will be excluded from training. One can see that our semi-supervised strategy performs better because the given labels of noisy samples are replaced by the more reliable “co-guessed” labels from the predictions of two networks. This can minimize the harm of noisy labels and fully use these data. By contrast, dropping the mislabeled data performs worse because it does not use these data.
The λ n is a parameter to control the contribution of the noisy subset. Table 14 shows the detailed results with varied λ n . One can see that Tripartite is not sensitive to λ n , particularly with a low noise ratio and realistic noise. With symmetric noise with a ratio of 50%, the difference between the best and worst performances is less than 2%. In short, the dataset with a higher noise ratio and more complex data requires a larger λ n to learn sufficient information.

6. Conclusions

We present Tripartite—a novel partition method to prevent the model from misusing uncertain samples in training data by partitioning training data into three subsets: uncertain, noisy, and clean. Despite appearing to be an upgrade of the bipartition strategy, Tripartite addresses a core issue that is overlooked by binary approaches by carefully examining the nature of the distribution of noisy labeled data, especially on real-world label noise. Thus, it demonstrates a significant performance gain over SOTA methods on multiple types of label noise. Explorative analysis, such as the visualization of uncertain samples, the quality of data partition, and Tripartition for other methods, reveals the desirable trait, which is that Tripartition improves the quality of the clean subset and then the performance of models. We think Tripartite provides a new perspective on data partition in label noise learning.
While our proposed Tripartite demonstrates promising performance, it is pertinent to recognize some limitations. In an uncertain subset, some noisy samples may still exist. In future work, we will investigate other solutions to enhance the Tripartition strategy, in particular, more effectively utilizing the clean samples in the uncertain subset.

Author Contributions

Conceptualization, X.L. (Xuefeng Liang) and L.Y. (Longshan Yao); methodology, X.L. (Xuefeng Liang); software, L.Y. (Lida Yu), C.C., and X.L. (Xingyu Liu); validation, L.Y. (Lida Yu), C.C., L.Y. (Longshan Yao) and X.L. (Xingyu Liu); formal analysis, X.L. (Xuefeng Liang) and L.Y. (Lida Yu); investigation, L.Y. (Lida Yu), C.C., L.Y. (Longshan Yao) and X.L. (Xingyu Liu); data curation, L.Y. (Lida Yu), C.C. and X.L. (Xingyu Liu); writing—original draft preparation, X.L. (Xuefeng Liang) and L.Y. (Longshan Yao); writing—review and editing, L.Y. (Lida Yu), X.L. (Xuefeng Liang) and L.Y. (Longshan Yao); visualization, L.Y. (Lida Yu) and C.C.; supervision, X.L. (Xuefeng Liang); project administration, X.L. (Xuefeng Liang); funding acquisition, X.L. (Xuefeng Liang). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the research funding of GuangZhou DaYou Network Technology Co., Ltd, China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are publicly available benchmark datasets commonly used in machine learning research. The specific datasets and their sources are detailed in Section 4 of the manuscript. The code and experimental results supporting the conclusions of this article will be made available by the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mahajan, D.; Girshick, R.; Ramanathan, V.; He, K.; Paluri, M.; Li, Y.; Bharambe, A.; Van Der Maaten, L. Exploring the limits of weakly supervised pretraining. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 181–196. [Google Scholar]
  2. Blum, A.; Kalai, A.; Wasserman, H. Noise-tolerant learning, the parity problem, and the statistical query model. J. ACM 2003, 50, 506–519. [Google Scholar] [CrossRef]
  3. Thomee, B.; Shamma, D.A.; Friedland, G.; Elizalde, B.; Ni, K.; Poland, D.; Borth, D.; Li, L.J. YFCC100M: The new data in multimedia research. Commun. ACM 2016, 59, 64–73. [Google Scholar] [CrossRef]
  4. Li, W.; Wang, L.; Li, W.; Agustsson, E.; Van Gool, L. Webvision database: Visual learning and understanding from web data. arXiv 2017, arXiv:1708.02862. [Google Scholar]
  5. Yan, Y.; Rosales, R.; Fung, G.; Subramanian, R.; Dy, J. Learning from multiple annotators with varying expertise. Mach. Learn. 2014, 95, 291–327. [Google Scholar] [CrossRef]
  6. Yu, X.; Liu, T.; Gong, M.; Tao, D. Learning with biased complementary labels. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 68–83. [Google Scholar]
  7. Li, J.; Wong, Y.; Zhao, Q.; Kankanhalli, M.S. Learning to learn from noisy labeled data. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 5051–5059. [Google Scholar]
  8. Tanno, R.; Saeedi, A.; Sankaranarayanan, S.; Alexander, D.C.; Silberman, N. Learning from noisy labels by regularized estimation of annotator confusion. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 11244–11253. [Google Scholar]
  9. Zhang, C.; Bengio, S.; Hardt, M.; Recht, B.; Vinyals, O. Understanding deep learning (still) requires rethinking generalization. Commun. ACM 2021, 64, 107–115. [Google Scholar] [CrossRef]
  10. Arpit, D.; Jastrzebski, S.; Ballas, N.; Krueger, D.; Bengio, E.; Kanwal, M.S.; Maharaj, T.; Fischer, A.; Courville, A.; Bengio, Y.; et al. A closer look at memorization in deep networks. In Proceedings of the ICML, Sydney, Australia, 6–11 August 2017; PMLR: Cambridge, MA, USA, 2017; pp. 233–242. [Google Scholar]
  11. Ahmed, M.T.; Monjur, O.; Khaliduzzaman, A.; Kamruzzaman, M. A comprehensive review of deep learning-based hyperspectral image reconstruction for agri-food quality appraisal. Artif. Intell. Rev. 2025, 58, 96. [Google Scholar] [CrossRef]
  12. Gimaletdinova, G.; Shaiakhmetov, D.; Akpaeva, M.; Abduzhabbarov, M.; Momunov, K. Training a Neural Network for Partially Occluded Road Sign Identification in the Context of Autonomous Vehicles. arXiv 2025, arXiv:2503.18177. [Google Scholar]
  13. González-Santoyo, C.; Renza, D.; Moya-Albor, E. Identifying and Mitigating Label Noise in Deep Learning for Image Classification. Technologies 2025, 13, 132. [Google Scholar] [CrossRef]
  14. Zhang, X.Y.; Zhang, X.P.; Yu, H.G.; Liu, Q.S. A confident learning-based support vector machine for robust ground classification in noisy label environments. Tunn. Undergr. Space Technol. 2025, 155, 106128. [Google Scholar] [CrossRef]
  15. Huang, B.; Xie, Y.; Xu, C. Learning with noisy labels via clean aware sharpness aware minimization. Sci. Rep. 2025, 15, 1350. [Google Scholar] [CrossRef] [PubMed]
  16. Patrini, G.; Rozza, A.; Krishna Menon, A.; Nock, R.; Qu, L. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1944–1952. [Google Scholar]
  17. Goldberger, J.; Ben-Reuven, E. Training deep neural-networks using a noise adaptation layer. In Proceedings of the ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  18. Reed, S.; Lee, H.; Anguelov, D.; Szegedy, C.; Erhan, D.; Rabinovich, A. Training deep neural networks on noisy labels with bootstrapping. In Proceedings of the ICLR, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  19. Tanaka, D.; Ikami, D.; Yamasaki, T.; Aizawa, K. Joint optimization framework for learning with noisy labels. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–22 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 5552–5560. [Google Scholar]
  20. Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.W.; Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of the NeurIPS, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  21. Li, J.; Socher, R.; Hoi, S.C. DivideMix: Learning with Noisy Labels as Semi-supervised Learning. In Proceedings of the ICLR, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  22. Yu, X.; Han, B.; Yao, J.; Niu, G.; Tsang, I.; Sugiyama, M. How does disagreement help generalization against label corruption? In Proceedings of the ICML, Long Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 7164–7173. [Google Scholar]
  23. Wei, H.; Feng, L.; Chen, X.; An, B. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the CVPR, Seattle, WA, USA, 14–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 13726–13735. [Google Scholar]
  24. Yi, R.; Huang, Y. TC-Net: Detecting Noisy Labels Via Transform Consistency. IEEE Trans. Multimed. 2022, 24, 4328–4341. [Google Scholar] [CrossRef]
  25. Li, J.; Li, G.; Liu, F.; Yu, Y. Neighborhood collective estimation for noisy label identification and correction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 128–145. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 630–645. [Google Scholar]
  27. Menon, A.; Van Rooyen, B.; Ong, C.S.; Williamson, B. Learning from corrupted binary labels via class-probability estimation. In Proceedings of the ICML, Lille, France, 6–11 July 2015; PMLR: Cambridge, MA, USA, 2015; pp. 125–134. [Google Scholar]
  28. Natarajan, N.; Dhillon, I.S.; Ravikumar, P.; Tewari, A. Learning with Noisy Labels. In Proceedings of the NeurIPS, Lake Tahoe, NV, USA, 5–10 December 2013. [Google Scholar]
  29. Xia, X.; Liu, T.; Wang, N.; Han, B.; Gong, C.; Niu, G.; Sugiyama, M. Are Anchor Points Really Indispensable in Label-Noise Learning? In Proceedings of the NeurIPS, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  30. Ghosh, A.; Kumar, H.; Sastry, P. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI, San Francisco, CA, USA, 4–9 February 2017; Volume 31, pp. 1919–1925. [Google Scholar]
  31. Xu, Y.; Cao, P.; Kong, Y.; Wang, Y. LDMI: A Novel Information-theoretic Loss Function for Training Deep Nets Robust to Label Noise. In Proceedings of the NeurIPS, Vancouver, BC, Canada, 8–14 December 2019; pp. 6222–6233. [Google Scholar]
  32. Wang, X.; Hua, Y.; Kodirov, E.; Robertson, N.M. IMAE for Noise-Robust Learning: Mean Absolute Error Does Not Treat Examples Equally and Gradient Magnitude’s Variance Matters. arXiv 2019, arXiv:1903.12141. [Google Scholar]
  33. Zhang, Z.; Sabuncu, M.R. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the NeurIPS, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  34. Wang, Y.; Ma, X.; Chen, Z.; Luo, Y.; Yi, J.; Bailey, J. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 322–330. [Google Scholar]
  35. Yi, K.; Wu, J. Probabilistic end-to-end noise correction for learning with noisy labels. In Proceedings of the CVPR, Long Beach, CA, USA, 15–21 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7017–7025. [Google Scholar]
  36. Liu, S.; Niles-Weed, J.; Razavian, N.; Fernandez-Granda, C. Early-Learning Regularization Prevents Memorization of Noisy Labels. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020. [Google Scholar]
  37. Hendrycks, D.; Mazeika, M.; Wilson, D.; Gimpel, K. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise. In Proceedings of the NeurIPS, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  38. Jiang, L.; Zhou, Z.; Leung, T.; Li, L.J.; Fei-Fei, L. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In Proceedings of the ICML, Stockholm, Sweden, 10–15 July 2018; PMLR: Cambridge, MA, USA, 2018; pp. 2304–2313. [Google Scholar]
  39. Huang, Z.; Zhang, J.; Shan, H. Twin contrastive learning with noisy labels. In Proceedings of the CVPR, Vancouver, BC, Canada, 18–22 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 11661–11670. [Google Scholar]
  40. Bai, Y.; Yang, E.; Han, B.; Yang, Y.; Li, J.; Mao, Y.; Niu, G.; Liu, T. Understanding and improving early stopping for learning with noisy labels. In Proceedings of the NeurIPS, Virtual, 6–14 December 2021. [Google Scholar]
  41. Yang, H.; Jin, Y.Z.; Li, Z.Y.; Wang, D.B.; Geng, X.; Zhang, M.L. Learning from noisy labels via dynamic loss thresholding. IEEE Trans. Knowl. Data. Eng. 2023, 36, 6503–6516. [Google Scholar] [CrossRef]
  42. Albert, P.; Ortego, D.; Arazo, E.; O’Connor, N.E.; McGuinness, K. Addressing out-of-distribution label noise in webly-labelled data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 392–401. [Google Scholar]
  43. Kim, D.; Ryoo, K.; Cho, H.; Kim, S. SplitNet: Learnable clean-noisy label splitting for learning with noisy labels. Int. J. Comput. Vis. 2025, 133, 549–566. [Google Scholar] [CrossRef]
  44. Garg, A.; Nguyen, C.; Felix, R.; Liu, Y.; Do, T.T.; Carneiro, G. AEON: Adaptive Estimation of Instance-Dependent In-Distribution and Out-of-Distribution Label Noise for Robust Learning. arXiv 2025, arXiv:2501.13389. [Google Scholar]
  45. Liu, Z.; Zhu, F.; Vasilakos, A.V.; Chen, X.; Zhao, Q.; Camacho, D. Discriminative approximate regression projection for feature extraction. Inf. Fusion 2025, 120, 103088. [Google Scholar] [CrossRef]
  46. Wang, G.; Wang, C.; Chen, Y.; Liu, J. Meta-learning collaborative optimization for lifetime prediction of lithium-ion batteries considering label noise. J. Energy Storage 2025, 107, 114928. [Google Scholar] [CrossRef]
  47. Zhang, R.; Cao, Z.; Huang, Y.; Yang, S.; Xu, L.; Xu, M. Visible-Infrared Person Re-identification with Real-world Label Noise. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 4857–4869. [Google Scholar] [CrossRef]
  48. Bukchin, G.; Schwartz, E.; Saenko, K.; Shahar, O.; Feris, R.S.; Giryes, R.; Karlinsky, L. Fine-grained Angular Contrastive Learning with Coarse Labels. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 8726–8736. [Google Scholar]
  49. van der Maaten, L.; Hinton, G.E. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  50. Li, Y.; Han, H.; Shan, S.; Chen, X. Disc: Learning from noisy labels via dynamic instance-specific selection and correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 24070–24079. [Google Scholar]
  51. Lee, K.; Yun, S.; Lee, K.; Lee, H.; Li, B.; Shin, J. Robust inference via generative classifiers for handling noisy labels. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 3763–3772. [Google Scholar]
  52. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. In Proceedings of the ICLR, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  53. Krizhevsky, A.; Hinton, G.E. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
  54. Xiao, T.; Xia, T.; Yang, Y.; Huang, C.; Wang, X. Learning from massive noisy labeled data for image classification. In Proceedings of the CVPR, Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 2691–2699. [Google Scholar]
  55. Xia, X.; Liu, T.; Han, B.; Wang, N.; Gong, M.; Liu, H.; Niu, G.; Tao, D.; Sugiyama, M. Part-dependent label noise: Towards instance-dependent label noise. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020. [Google Scholar]
  56. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the CVPR, Miami, FL, USA, 20–25 June 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 248–255. [Google Scholar]
  57. Chen, P.; Liao, B.B.; Chen, G.; Zhang, S. Understanding and utilizing deep neural networks trained with noisy labels. In Proceedings of the ICML, Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 1062–1070. [Google Scholar]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the CVPR, Las Vegas, NV, USA, 26 June–1 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  59. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  60. Arazo, E.; Ortego, D.; Albert, P.; O’Connor, N.; McGuinness, K. Unsupervised label noise modeling and loss correction. In Proceedings of the ICML, Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 312–321. [Google Scholar]
  61. Tan, C.; Xia, J.; Wu, L.; Li, S.Z. Co-learning: Learning from noisy labels with self-supervision. In Proceedings of the ACM MM, Virtual Event, China, 20–24 October 2021; ACM: New York, NY, USA, 2021; pp. 1405–1413. [Google Scholar]
  62. Kim, Y.; Yun, J.; Shon, H.; Kim, J. Joint Negative and Positive Learning for Noisy Labels. In Proceedings of the CVPR, Virtual, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 9442–9451. [Google Scholar]
  63. Garg, A.; Nguyen, C.; Felix, R.; Do, T.T.; Carneiro, G. Instance-dependent noisy label learning via graphical modelling. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2023; pp. 2288–2298. [Google Scholar]
  64. Malach, E.; Shalev-Shwartz, S. Decoupling “when to update” from “how to update”. In Proceedings of the NeurIPS, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  65. Ma, X.; Wang, Y.; Houle, M.E.; Zhou, S.; Erfani, S.; Xia, S.; Wijewickrema, S.; Bailey, J. Dimensionality-driven learning with noisy labels. In Proceedings of the ICML, Stockholm, Sweden, 10–15 July 2018; PMLR: Cambridge, MA, USA, 2018; pp. 3355–3364. [Google Scholar]
Figure 1. The feature distribution and loss distribution of the entire training set in CIFAR-10 with a realistic noise ratio of 50%. The testing model is a trained 18-layer PreAct Resnet [26] stopped at the early stage (40 epochs) using the DivideMix partition criterion [21]. (a) The feature distribution. The black dots represent the noisy samples. The colored dots represent the clean samples. Each color denotes a category. There are many uncertain samples distributed close to the classification boundary, in which some of them are grouped by the red curve. (b) The loss distribution. Some noisy samples also have small losses, as shown between two red dashed lines.
Figure 1. The feature distribution and loss distribution of the entire training set in CIFAR-10 with a realistic noise ratio of 50%. The testing model is a trained 18-layer PreAct Resnet [26] stopped at the early stage (40 epochs) using the DivideMix partition criterion [21]. (a) The feature distribution. The black dots represent the noisy samples. The colored dots represent the clean samples. Each color denotes a category. There are many uncertain samples distributed close to the classification boundary, in which some of them are grouped by the red curve. (b) The loss distribution. Some noisy samples also have small losses, as shown between two red dashed lines.
Sensors 25 03369 g001
Figure 2. The generated noise transition matrices of CIFAR-10 with noise ratios of 20% and 50%.
Figure 2. The generated noise transition matrices of CIFAR-10 with noise ratios of 20% and 50%.
Sensors 25 03369 g002
Figure 3. The visualization of the feature distributions of three similar class-pairs in CIFAR-100 using t-SNE [49].
Figure 3. The visualization of the feature distributions of three similar class-pairs in CIFAR-100 using t-SNE [49].
Sensors 25 03369 g003
Figure 4. The distribution of 100 samples from categories “cat” and “dog” in the realistic CIFAR-10. Red and blue represent that given labels are cat and dog, respectively. Dot and triangle represent that the ground-truth labels are cat and dog, respectively. TL and GL denote the ground-truth label and the given label, respectively. Samples in the bounding box are considered as uncertain samples because they are close to the classification boundary. (a) The given labels are clean, equal to the ground-truth labels. (b) The given labels are noisy and differ from the ground-truth labels. Star represents a sample that is either a cat or dog but is given a noisy label from other classes.
Figure 4. The distribution of 100 samples from categories “cat” and “dog” in the realistic CIFAR-10. Red and blue represent that given labels are cat and dog, respectively. Dot and triangle represent that the ground-truth labels are cat and dog, respectively. TL and GL denote the ground-truth label and the given label, respectively. Samples in the bounding box are considered as uncertain samples because they are close to the classification boundary. (a) The given labels are clean, equal to the ground-truth labels. (b) The given labels are noisy and differ from the ground-truth labels. Star represents a sample that is either a cat or dog but is given a noisy label from other classes.
Sensors 25 03369 g004
Figure 5. The logic of our Tripartite method. GL denotes the given label.
Figure 5. The logic of our Tripartite method. GL denotes the given label.
Sensors 25 03369 g005
Figure 6. The networks are trained by our Tripartite solution in the later stage. To have a clearer visual effect, only clean and uncertain samples are visualized. (a) The distribution of uncertain samples (black dots) in the feature space of network 1. (b) A few uncertain samples distribute in the cluster in the feature space of network 1. (c) The uncertain samples shown in (b) distribute close to the classification boundary in the feature space of network 2.
Figure 6. The networks are trained by our Tripartite solution in the later stage. To have a clearer visual effect, only clean and uncertain samples are visualized. (a) The distribution of uncertain samples (black dots) in the feature space of network 1. (b) A few uncertain samples distribute in the cluster in the feature space of network 1. (c) The uncertain samples shown in (b) distribute close to the classification boundary in the feature space of network 2.
Sensors 25 03369 g006
Figure 7. Comparison of three partition criteria on the CIFAR-100 dataset with 50% realistic label noise and 40% instance-dependent label noise. (a,c) The partition accuracy of clean samples in the clean subset against training epochs on two label noises. (b,d) The partition accuracy of noisy samples in the noisy subset against training epochs on two label noises.
Figure 7. Comparison of three partition criteria on the CIFAR-100 dataset with 50% realistic label noise and 40% instance-dependent label noise. (a,c) The partition accuracy of clean samples in the clean subset against training epochs on two label noises. (b,d) The partition accuracy of noisy samples in the noisy subset against training epochs on two label noises.
Sensors 25 03369 g007
Table 1. The top 10 similar class pairs in descending order of the cosine similarity from CIFAR-10. Table headers are bolded to enhances the readability of the table and makes the structural hierarchy clearer, Other forms are handled in the same way.
Table 1. The top 10 similar class pairs in descending order of the cosine similarity from CIFAR-10. Table headers are bolded to enhances the readability of the table and makes the structural hierarchy clearer, Other forms are handled in the same way.
Similar ClassesSimilarity
automobiletruck0.075
automobileship0.062
catdog0.062
doghorse0.054
deerhorse0.052
birdfrog0.046
airplaneship0.044
horsetruck0.025
catfrog0.021
airplanetruck0.016
Table 2. The top 60 similar class pairs in descending order of the cosine similarity from CIFAR-100.
Table 2. The top 60 similar class pairs in descending order of the cosine similarity from CIFAR-100.
Similar ClassesSimilaritySimilar ClassesSimilaritySimilar ClassesSimilarity
maple_treeoak_tree0.337poppytulip0.249beaverotter0.222
dolphinwhale0.312bedcouch0.246liontiger0.220
plainsea0.306poppyrose0.245clockplate0.220
buspickup_truck0.301snakeworm0.245pickup_trucktank0.219
girlwoman0.298bottlecan0.242flatfishray0.217
busstreetcar0.296applepear0.242plainroad0.216
bicyclemotorcycle0.275dolphinshark0.237chimpanzeewoman0.216
babygirl0.273babyboy0.232sharktrout0.215
maple_treewillow_tree0.270poppysunflower0.232lawn_mowertractor0.214
castlehouse0.260bowlplate0.231housestreetcar0.211
oak_treepine_tree0.258cloudplain0.231leopardlion0.210
cloudsea0.257hamstermouse0.230otterseal0.210
oak_treewillow_tree0.255chaircouch0.230beetlespider0.209
orchidtulip0.255bowlcup0.229bedchair0.209
beetlecockroach0.254rosetulip0.229orangepear0.209
streetcartrain0.253appleorange0.229maple_treerose0.207
leopardtiger0.253televisionwardrobe0.227pickup_trucktractor0.205
manwoman0.251keyboardtelephone0.227forestwillow_tree0.205
orangesweet_pepper0.250boygirl0.224crablobster0.205
applesweet_pepper0.250mouseshrew0.223mountainsea0.204
Table 3. The detailed configurations of four benchmark datasets.
Table 3. The detailed configurations of four benchmark datasets.
Datasets# of Training# of Test# of ClassSize
CIFAR-1050,00010,0001032 × 32
CIFAR-10050,00010,00010032 × 32
Clothing1M1,000,00010,52614*
WebVision1.02,400,00050,0001000*
Table 4. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20% and 50% symmetric noises. We run the open-source code of DSOS and report the result. The results of ELR+, Co-learning, PES, TCL+, and DLT are from their papers, and the rest of the results are from [21]. The best results are in bold, and the second-best results are underlined. * means that the methods did not have results provided.
Table 4. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20% and 50% symmetric noises. We run the open-source code of DSOS and report the result. The results of ELR+, Co-learning, PES, TCL+, and DLT are from their papers, and the rest of the results are from [21]. The best results are in bold, and the second-best results are underlined. * means that the methods did not have results provided.
Dataset
Symmetric
CIFAR-10CIFAR-100
Method 20%50%20%50%
Standard CE86.879.462.046.7
Bootstrap (2015) [18]86.879.862.146.6
F-correction (2017) [16]86.879.861.546.6
Co-teaching + (2019) [22]89.585.765.651.8
P-correction (2019) [35]92.489.169.457.5
Meta-Learning (2019) [7]92.989.368.559.2
M-correction (2019) [60]94.092.073.966.1
DivideMix (2020) [21]96.194.677.374.6
ELR + (2020) [36]95.894.877.673.6
Co-learning (2021) [61]92.584.866.755.0
PES(semi) (2021) [40]95.995.177.474.3
DSOS (2022) [42]92.787.475.166.2
TCL + (2023) [39]96.094.579.374.6
DLT (2023) [41]96.3*77.1*
Tripartite96.495.181.176.6
Table 5. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20% and 40% instance-dependent noises. The result of InstanceGM is from [63], and the rest of the results are from PES [40]. The best results are in bold and the second-best results are underlined.
Table 5. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20% and 40% instance-dependent noises. The result of InstanceGM is from [63], and the rest of the results are from PES [40]. The best results are in bold and the second-best results are underlined.
Dataset
Inst
CIFAR-10CIFAR-100
Method 20%40%20%40%
Standard CE87.578.956.848.2
DivideMix (2020) [21]95.594.575.270.9
ELR + (2020) [36]94.994.375.874.3
PES(semi) (2021) [40]95.995.377.676.1
InstanceGM (2023) [63]96.796.479.778.5
Tripartite96.896.680.178.8
Table 6. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20%, 40%, and 50% realistic noises. We strictly implemented the competing methods according to their open codes and papers. The backbone is the Pre-ResNet18. The best results are in bold and the second-best results are underlined.
Table 6. Accuracy comparisons on CIFAR-10 and CIFAR-100 with 20%, 40%, and 50% realistic noises. We strictly implemented the competing methods according to their open codes and papers. The backbone is the Pre-ResNet18. The best results are in bold and the second-best results are underlined.
Dataset
Realistic
CIFAR-10CIFAR-100
Method 20%40%50%20%40%50%
Co-teaching (2018) [20]82.774.455.850.340.932.5
JoCoR (2020) [23]82.268.755.849.735.129.1
DivideMix (2020) [21]91.192.391.876.266.159.5
ELR + (2020) [36]95.192.991.275.872.560.6
Co-learning (2021) [61]91.675.561.768.558.550.2
PES(semi) (2021) [40]95.993.991.176.373.266.7
DSOS (2022) [42]92.589.081.175.364.852.6
Tripartite96.295.994.480.975.571.4
Table 7. Accuracy comparison on Clothing1M. The results of JNPL, DSOS, and TCL are from [39,42,62]. The rest of the results are from [36]. The best results are in bold and the second-best results are underlined.
Table 7. Accuracy comparison on Clothing1M. The results of JNPL, DSOS, and TCL are from [39,42,62]. The rest of the results are from [36]. The best results are in bold and the second-best results are underlined.
Clothing1M
MethodsAcc.
Standard CE69.21
F-correction (2017) [16]69.84
Joint (2018) [19]72.16
Meta-Learning (2019) [7]73.47
P-correction (2019) [35]73.49
DivideMix (2020) [21]74.76
ELR + (2020) [36]74.81
JNPL (2021) [62]74.15
PES(semi) (2021) [40]74.99
DSOS (2022) [42]73.63
TCL (2023) [39]74.80
Tripartite75.23
Table 8. Top-1 (Top-5) accuracy comparison on the WebVision1.0 and ILSVRC12. The result of DSOS is from [42], the result of TCL is from [39], and the rest of the results are from [36]. The best results are in bold and the second-best results are underlined.
Table 8. Top-1 (Top-5) accuracy comparison on the WebVision1.0 and ILSVRC12. The result of DSOS is from [42], the result of TCL is from [39], and the rest of the results are from [36]. The best results are in bold and the second-best results are underlined.
DatasetWebVisionILSVRC12
Method Top-1Top-5Top-1Top-5
F-correction (2017) [16]61.1282.6857.3682.36
Decoupling (2017) [64]62.5484.7458.2682.26
D2L (2018) [65]62.6884.0057.8081.36
MentorNet (2018) [38]63.0081.4057.8079.92
Co-teaching (2018) [20]63.5885.2061.4884.70
Iterative-CV (2019) [57]65.2485.3461.6084.98
DivideMix (2020) [21]77.3291.6475.2090.84
ELR+ (2020) [36]77.7891.6870.2989.76
DSOS (2022) [42]77.7692.0474.3690.80
TCL (2023) [39]79.1092.3075.4092.40
Tripartite80.1293.8876.4093.20
Table 9. The partition accuracy of Tripartition on CIFAR-100 with 50% realistic noise.
Table 9. The partition accuracy of Tripartition on CIFAR-100 with 50% realistic noise.
EpochAcc. (%)Clean SubsetNoisy SubsetUncertain Subset
Partition Acc.Total NumsPartition Acc.Total Nums
2046.4773.9412,89671.75959727,507
5059.4583.3418,92777.6619,90211,171
10062.5782.6920,72279.8119,5519727
15064.7782.3121,63481.1619,0859281
20066.1578.0125,93681.5520,3693695
25065.9977.2226,50181.5120,5372962
30065.8076.7526,83181.7320,3702799
Table 10. The partition accuracy of Tripartition on CIFAR-100 with 40% instance-dependent noise.
Table 10. The partition accuracy of Tripartition on CIFAR-100 with 40% instance-dependent noise.
EpochAcc. (%)Clean SubsetNoisy SubsetUncertain Subset
Partition Acc.Total NumsPartition Acc.Total Nums
3049.4278.1521,95477.71364724,399
5067.1399.4821,58582.5817,87110,544
10070.2399.4423,07784.2618,4878436
15071.5799.2323,83685.0918,4927672
20075.0397.4227,21886.0720,1702612
25075.7096.9327,55486.1420,3352111
30075.4096.6127,73286.2020,3441924
Table 11. The effectiveness of Tripartition for other methods on CIFAR-100 with two label noises. Bold values indicate the best performance results in each comparison group.
Table 11. The effectiveness of Tripartition for other methods on CIFAR-100 with two label noises. Bold values indicate the best performance results in each comparison group.
Noisy TypeMethodPartition CriterionAcc. %
Instance-dependent
30%
Co-teachingSmall-loss71.30
Tripartition72.34
DivideMixGMM75.94
Tripartition76.29
Realistic
30%
Co-teachingSmall-loss71.04
Tripartition72.03
DivideMixGMM76.44
Tripartition76.49
Table 12. Evaluation of λ u in the training strategy for the uncertain subset on CIFAR-100. Bold values highlight the optimal performance for each noise condition with the corresponding best $\lambda_u$ parameter setting.
Table 12. Evaluation of λ u in the training strategy for the uncertain subset on CIFAR-100. Bold values highlight the optimal performance for each noise condition with the corresponding best $\lambda_u$ parameter setting.
λ u = 10.50.30.10
Noise Type
Symmetric20%78.1179.9581.1480.6078.84
50%67.8075.2676.6375.0472.53
Realistic20%77.5380.0180.8580.0077.63
50%67.7870.0771.4070.9964.6
Table 13. Evaluation of the training strategy for the noisy subset on CIFAR-100. Bold values indicate the best performance achieved by each training strategy under different noise conditions.
Table 13. Evaluation of the training strategy for the noisy subset on CIFAR-100. Bold values indicate the best performance achieved by each training strategy under different noise conditions.
Noise TypeSymmetricRealistic
Strategy 20%50%20%50%
Drop79.872.9779.768.6
Semi-supervised81.1476.6380.8571.40
Table 14. Evaluation of λ n in the training strategy for the noisy subset on CIFAR-100. Bold values highlight the optimal performance for each noise condition with the corresponding best $\lambda_n$ parameter setting.
Table 14. Evaluation of λ n in the training strategy for the noisy subset on CIFAR-100. Bold values highlight the optimal performance for each noise condition with the corresponding best $\lambda_n$ parameter setting.
λ n = 0.60.50.40.30.20.1
Noise Type
Symmetric20%80.1080.2680.5381.0081.1480.29
50%75.9176.6376.4676.0374.9774.85
Realistic20%79.9680.0280.6080.8580.6579.93
50%70.6071.4071.1271.3770.1969.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, L.; Liang, X.; Cao, C.; Yao, L.; Liu, X. Tripartite: Tackling Realistic Noisy Labels with More Precise Partitions. Sensors 2025, 25, 3369. https://doi.org/10.3390/s25113369

AMA Style

Yu L, Liang X, Cao C, Yao L, Liu X. Tripartite: Tackling Realistic Noisy Labels with More Precise Partitions. Sensors. 2025; 25(11):3369. https://doi.org/10.3390/s25113369

Chicago/Turabian Style

Yu, Lida, Xuefeng Liang, Chang Cao, Longshan Yao, and Xingyu Liu. 2025. "Tripartite: Tackling Realistic Noisy Labels with More Precise Partitions" Sensors 25, no. 11: 3369. https://doi.org/10.3390/s25113369

APA Style

Yu, L., Liang, X., Cao, C., Yao, L., & Liu, X. (2025). Tripartite: Tackling Realistic Noisy Labels with More Precise Partitions. Sensors, 25(11), 3369. https://doi.org/10.3390/s25113369

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop