FedENLC: An End-to-End Noisy Label Correction Framework in Federated Learning
Abstract
1. Introduction
- SCE is used as the loss function in both stage-1 and stage-2, and label smoothing is applied in stage-1. SCE combines CE and RCE and is a loss function robust to noisy labels, while label smoothing distributes part of the probability mass of the ground-truth label to neighboring classes, mitigating overconfidence and preventing extreme probability shifts caused by noisy labels.
- A three-component BGMM is used to distinguish noisy clients. By leveraging Bayesian priors, BGMM assigns prior distributions to parameters to alleviate extreme bias, making it more suitable than EM-based GMM for federated learning environments that are already biased due to label noise and data heterogeneity. In particular, a three-component BGMM allows more fine-grained client separation and can detect noisy clients more accurately and stably than a two-component GMM.
- Label correction is applied only to clients with high noise ratios among the noisy clients. This approach selectively applies the computationally expensive label correction procedure only to high-noise clients, thereby reducing performance degradation caused by misclassification, improving the stability of federated learning, and increasing the overall training efficiency.
2. Related Works
2.1. Federated Learning
2.2. Federated Noisy Label Learning
3. Preliminaries
3.1. Problem Definition
3.2. Label Noise
4. Methodology
4.1. Proposed Model Framework
4.1.1. Stage-1: Noisy Client Detection
4.1.2. Stage-2: End-to-End Label Correction
- First, unlike the existing model that uses CE as the classification loss, the proposed model employs SCE. Instead of using the one-hot hard label , which may contain noise, the classification loss is computed between the model prediction p and the learnable distribution , and is defined as follows:where and are hyperparameters that balance CE and RCE. Through empirical tuning, we set them to 1 and 0.1 for CIFAR-10, and 1 and 2 for CIFAR-100.
- Second, the compatibility regularization loss is used to encourage the soft label does not deviate significantly from the hard label , and is defined as follows:where M denotes the number of classes in the dataset.
- Third, the entropy regularization loss encourages the model to produce more confident and reliable predictions, and is defined as:where is the softmax probability of the model output for class m.
4.2. Bayesian Gaussian Mixture Model
4.3. Label Correction for Top Noisy Clients
4.4. Experimental Datasets
4.5. Evaluation Metrics
4.6. Implementation Details
5. Results
6. Discussion
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated Learning: Challenges, Methods, and Future Directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Agüera y Arcas, B. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
- Wang, Y.; Li, R.; Tan, H.; Jiang, X.; Sun, S.; Liu, M.; Gao, B.; Wu, Z. Federated skewed label learning with logits fusion. arXiv 2023, arXiv:2311.08202. [Google Scholar] [CrossRef]
- Wu, N.; Yu, L.; Jiang, X.; Cheng, K.-T.; Yan, Z. FedNoRo: Towards Noise-Robust Federated Learning by Addressing Class Imbalance and Label Noise Heterogeneity. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, Macao, China, 19–25 August 2023; pp. 4424–4432. [Google Scholar]
- Yan, B.; Cao, D.; Jiang, X.; Chen, Y.; Dai, W.; Dong, F.; Huang, W.; Zhang, T.; Gao, C.; Chen, Q.; et al. FedEYE: A scalable and flexible end-to-end federated learning platform for ophthalmology. Patterns 2024, 5, 100928. [Google Scholar] [CrossRef] [PubMed]
- Tan, B.; Liu, B.; Zheng, V.; Yang, Q. A federated recommender system for online services. In Proceedings of the ACM Conference on Recommender Systems, Online, 22–26 September 2020; pp. 579–581. [Google Scholar]
- Yang, L.; Tan, B.; Zheng, V.W.; Chen, K.; Yang, Q. Federated Recommendation Systems. In Federated Learning; Springer: Cham, Switzerland, 2020; Volume 12500, pp. 225–239. [Google Scholar]
- Kang, B.; Xie, S.; Rohrbach, M.; Yan, Z.; Gordo, A.; Feng, J.; Kalantidis, Y. Decoupling representation and classifier for long-tailed recognition. In Proceedings of the International Conference on Learning Representations, Online, 26 April–1 May 2020; pp. 1–16. [Google Scholar]
- Zhou, B.; Cui, Q.; Wei, X.-S.; Chen, Z.-M. BBN: Bilateral-branch network with cumulative learning for long-tailed visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9719–9728. [Google Scholar]
- Cao, K.; Wei, C.; Gaidon, A.; Arechiga, N.; Ma, T. Learning imbalanced datasets with label-distribution-aware margin loss. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 1567–1578. [Google Scholar]
- Menon, A.K.; Jayasumana, S.; Rawat, A.S.; Jain, H.; Veit, A.; Kumar, S. Long-tail learning via logit adjustment. In Proceedings of the International Conference on Learning Representations, Online, 3–7 May 2021; pp. 1–24. [Google Scholar]
- Yang, S.; Park, H.; Byun, J.; Kim, C. Robust Federated Learning with Noisy Labels. IEEE Intell. Syst. 2022, 37, 35–43. [Google Scholar] [CrossRef]
- Jiang, X.; Sun, S.; Li, J.; Xue, J.; Li, R.; Wu, Z.; Xu, G.; Wang, Y.; Liu, M. Tackling noisy clients in federated learning with end-to-end label correction. In Proceedings of the ACM International Conference on Information and Knowledge Management, Boise, ID, USA, 21–25 October 2024; pp. 1015–1026. [Google Scholar]
- Wang, Y.; Ma, X.; Chen, Z.; Luo, Y.; Yi, J.; Bailey, J. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 322–330. [Google Scholar]
- Pereyra, G.; Tucker, G.; Chorowski, J.; Kaiser, Ł; Hinton, G. Regularizing neural networks by penalizing confident output distributions. arXiv 2017, arXiv:1701.06548. [Google Scholar] [CrossRef]
- Lu, J. A survey on Bayesian inference for Gaussian mixture model. arXiv 2021, arXiv:2108.11753. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. In Proceedings of the Advances in Machine Learning and Systems, Austin, TX, USA, 2–4 March 2020; pp. 429–450. [Google Scholar]
- Jhunjhunwala, D.; Wang, S.; Joshi, G. FedExP: Speeding Up Federated Averaging via Extrapolation. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; pp. 5693–5700. [Google Scholar]
- Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.; Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 2–8 December 2018; pp. 8536–8546. [Google Scholar]
- Yu, X.; Han, B.; Yao, J.; Niu, G.; Tsang, I.; Sugiyama, M. How Does Disagreement Help Generalization against Label Corruption? In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7164–7173. [Google Scholar]
- Tanaka, D.; Ikami, D.; Yamasaki, T.; Aizawa, K. Joint Optimization Framework for Learning with Noisy Labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 5552–5560. [Google Scholar]
- Song, H.; Kim, M.; Lee, J.-G. SELFIE: Refurbishing Unclean Samples for Robust Deep Learning. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 5907–5915. [Google Scholar]
- Li, J.; Socher, R.; Hoi, S. DivideMix: Learning with Noisy Labels as Semi-supervised Learning. In Proceedings of the International Conference on Learning Representations, Online, 26 April–1 May 2020; pp. 1–14. [Google Scholar]
- Zhang, H.; Cissé, M.; Dauphin, Y.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–13. [Google Scholar]
- Berthelot, D.; Carlini, N.; Goodfellow, I.; Papernot, N.; Oliver, A.; Raffel, C. MixMatch: A Holistic Approach to Semi-Supervised Learning. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 5050–5060. [Google Scholar]
- Zhang, C.; Wu, F.; Yi, J.; Xu, D.; Yu, Y.; Wang, J.; Wang, Y.; Xu, T.; Xie, X.; Chen, E. Non-IID Always Bad? Semi-Supervised Heterogeneous Federated Learning with Local Knowledge Enhancement. In Proceedings of the ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; pp. 3257–3267. [Google Scholar]
- Li, T.; Hu, S.; Beirami, A.; Smith, V. Ditto: Fair and Robust Federated Learning Through Personalization. In Proceedings of the International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 6357–6368. [Google Scholar]
- Yin, D.; Chen, Y.; Ramchandran, K.; Bartlett, P. Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5636–5645. [Google Scholar]
- Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 119–129. [Google Scholar]
- Jiang, X.; Sun, S.; Wang, Y.; Liu, M. Towards Federated Learning against Noisy Labels via Local Self-Regularization. In Proceedings of the ACM International Conference on Information and Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 862–873. [Google Scholar]
- Xu, J.; Chen, Z.; Quek, T.; Chong, K.F.E. FedCorr: Multi-Stage Federated Learning for Label Noise Correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 10174–10183. [Google Scholar]
- Allen-Zhu, Z.; Li, Y. Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning. In Proceedings of the International Conference on Learning Representations, Online, 25–29 April 2022; pp. 1–12. [Google Scholar]
- Kim, S.; Shin, W.; Jang, S.; Song, H.; Yun, S.-Y. FedRN: Exploiting k-Reliable Neighbors Towards Robust Federated Learning. In Proceedings of the ACM International Conference on Information and Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 972–981. [Google Scholar]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006; Volume 4, pp. 78–110. [Google Scholar]
- Blei, D.M.; Kucukelbir, A.; McAuliffe, J.D. Variational Inference: A Review for Statisticians. J. Am. Stat. Assoc. 2017, 112, 859–877. [Google Scholar] [CrossRef]




| Method | Dirichlet ( = 1.0) | Dirichlet ( = 0.5) | ||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Sym. (0.0–0.4) | Sym. (0.0–0.8) | Asym. (0.0–0.4) | Mixed (0.0–0.4) | Sym. (0.0–0.4) | Sym. (0.0–0.8) | Asym. (0.0–0.4) | Mixed (0.0–0.4) | |||||||||
| FedAvg [2] | 75.85 | 73.58 | 57.41 | 54.74 | 77.60 | 76.08 | 77.54 | 75.72 | 72.11 | 62.98 | 49.44 | 45.34 | 74.29 | 64.66 | 73.76 | 67.55 |
| FedProx [17] | 75.83 | 73.60 | 51.59 | 53.81 | 77.41 | 77.40 | 77.39 | 75.41 | 69.50 | 63.91 | 49.01 | 45.56 | 74.49 | 64.63 | 73.52 | 66.94 |
| FedExP [18] | 71.34 | 70.15 | 50.43 | 51.74 | 75.90 | 75.98 | 77.39 | 74.21 | 71.04 | 62.91 | 50.21 | 45.56 | 74.94 | 64.44 | 73.88 | 67.01 |
| TrimmedMean [28] | 71.21 | 64.29 | 47.93 | 44.41 | 74.06 | 66.06 | 72.83 | 64.71 | 69.47 | 59.82 | 49.47 | 42.40 | 69.44 | 57.17 | 71.12 | 58.38 |
| Krum [29] | 70.99 | 65.35 | 50.09 | 47.53 | 75.84 | 70.63 | 76.07 | 68.46 | 68.54 | 58.49 | 49.97 | 42.43 | 70.22 | 58.43 | 72.28 | 59.60 |
| Median [27] | 72.52 | 70.51 | 58.45 | 56.56 | 75.36 | 73.06 | 73.26 | 71.53 | 65.83 | 56.34 | 48.23 | 43.28 | 72.06 | 64.85 | 72.05 | 63.73 |
| Co-teaching [19] | 73.59 | 71.70 | 64.44 | 61.45 | 75.60 | 73.64 | 76.59 | 74.82 | 70.58 | 63.82 | 49.84 | 43.82 | 72.60 | 64.64 | 74.11 | 65.91 |
| Co-teaching+ [20] | 69.70 | 65.47 | 47.11 | 49.07 | 59.07 | 60.67 | 74.30 | 65.77 | 68.90 | 64.91 | 49.03 | 44.62 | 74.79 | 61.13 | 72.52 | 55.72 |
| Joint Optim [21] | 64.74 | 64.69 | 59.78 | 59.58 | 75.04 | 64.77 | 64.43 | 64.77 | 57.76 | 57.37 | 52.18 | 52.18 | 75.02 | 64.54 | 58.74 | 52.87 |
| SELFIE [22] | 73.74 | 73.58 | 62.86 | 60.51 | 76.79 | 76.14 | 76.70 | 76.14 | 72.06 | 61.48 | 62.39 | 60.14 | 70.93 | 64.03 | 71.74 | 58.39 |
| Symmetric CE [14] | 76.32 | 73.38 | 70.96 | 66.25 | 76.94 | 76.34 | 76.73 | 72.83 | 72.06 | 61.46 | 62.39 | 60.14 | 70.93 | 64.03 | 71.74 | 58.39 |
| DivideMix [23] | 73.68 | 61.94 | 61.51 | 59.51 | 76.70 | 76.26 | 76.73 | 76.23 | 69.57 | 58.65 | 62.39 | 61.40 | 70.93 | 63.04 | 71.74 | 58.39 |
| Robust FL [12] | 63.17 | 61.61 | 51.34 | 51.34 | 65.22 | 65.24 | 66.25 | 66.23 | 62.65 | 54.45 | 55.47 | 45.06 | 54.08 | 42.86 | 59.43 | 43.41 |
| FedLSR [30] | 71.65 | 66.92 | 70.64 | 66.23 | 75.13 | 74.61 | 73.83 | 68.52 | 68.42 | 57.54 | 51.47 | 55.97 | 70.64 | 61.54 | 70.64 | 57.04 |
| FedRDN [33] | 72.66 | 67.74 | 62.60 | 60.71 | 72.93 | 70.64 | 70.93 | 68.52 | 65.84 | 54.42 | 50.44 | 51.30 | 70.44 | 60.29 | 70.64 | 57.32 |
| FedNoRo [4] | 73.76 | 73.52 | 66.50 | 66.18 | 77.55 | 77.36 | 77.55 | 75.52 | 71.09 | 57.11 | 59.11 | 57.41 | 73.77 | 72.77 | 73.55 | 72.32 |
| FedELC [13] | 76.81 | 76.72 | 68.22 | 66.61 | 77.97 | 76.98 | 77.98 | 77.24 | 73.13 | 71.31 | 60.31 | 59.67 | 75.78 | 74.65 | 74.55 | 73.80 |
| FedENLC | 82.80 | 81.62 | 75.79 | 74.48 | 82.49 | 81.06 | 82.79 | 81.24 | 79.84 | 75.28 | 69.93 | 64.56 | 79.32 | 73.34 | 80.03 | 75.91 |
| Method | Dirichlet ( = 1.0) | Dirichlet ( = 0.5) | ||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Sym. (0.0–0.4) | Sym. (0.0–0.8) | Asym. (0.0–0.4) | Mixed (0.0–0.4) | Sym. (0.0–0.4) | Sym. (0.0–0.8) | Asym. (0.0–0.4) | Mixed (0.0–0.4) | |||||||||
| FedAvg [2] | 43.51 | 41.01 | 31.38 | 28.38 | 43.57 | 41.52 | 42.06 | 40.37 | 42.11 | 40.49 | 30.27 | 27.27 | 44.42 | 43.36 | 41.84 | 40.74 |
| FedProx [17] | 40.82 | 38.27 | 29.16 | 27.33 | 43.33 | 41.76 | 42.98 | 39.97 | 40.48 | 37.44 | 27.49 | 25.17 | 44.30 | 42.56 | 42.66 | 40.56 |
| FedExP [18] | 43.88 | 40.36 | 28.23 | 28.23 | 44.46 | 42.62 | 41.48 | 39.57 | 44.30 | 40.64 | 29.26 | 26.93 | 44.91 | 43.57 | 42.65 | 40.79 |
| TrimmedMean [28] | 34.48 | 34.48 | 28.33 | 25.44 | 41.33 | 39.38 | 41.13 | 38.47 | 35.14 | 34.40 | 27.60 | 25.14 | 40.58 | 39.11 | 39.14 | 37.91 |
| Krum [29] | 22.99 | 17.34 | 17.36 | 14.38 | 29.43 | 26.38 | 32.44 | 18.61 | 14.40 | 11.97 | 7.95 | 7.55 | 13.12 | 11.35 | 13.91 | 11.94 |
| Median [27] | 43.90 | 34.47 | 31.78 | 27.36 | 43.44 | 42.12 | 43.44 | 41.00 | 43.17 | 34.10 | 31.21 | 28.84 | 42.02 | 41.30 | 42.11 | 40.27 |
| Co-teaching [19] | 44.41 | 42.53 | 32.74 | 29.36 | 44.78 | 42.83 | 43.21 | 41.62 | 47.83 | 43.01 | 30.44 | 29.30 | 45.52 | 43.10 | 45.37 | 42.73 |
| Co-teaching+ [20] | 36.67 | 34.42 | 27.17 | 27.17 | 43.71 | 41.33 | 38.27 | 38.67 | 36.99 | 34.93 | 28.20 | 26.80 | 42.06 | 40.94 | 36.03 | 35.70 |
| Joint Optim [21] | 26.87 | 22.67 | 23.59 | 23.59 | 27.88 | 27.44 | 31.83 | 27.44 | 29.37 | 27.01 | 21.95 | 22.80 | 26.00 | 27.37 | 31.03 | 27.50 |
| SELFIE [22] | 44.62 | 42.92 | 32.66 | 30.54 | 44.90 | 42.74 | 43.42 | 41.95 | 45.00 | 42.53 | 33.20 | 32.13 | 40.60 | 42.54 | 42.04 | 41.97 |
| Symmetric CE [14] | 43.74 | 41.36 | 33.67 | 31.97 | 44.04 | 42.47 | 43.72 | 42.18 | 42.16 | 40.93 | 31.90 | 31.17 | 42.44 | 42.20 | 42.20 | 42.18 |
| DivideMix [23] | 37.42 | 37.68 | 30.74 | 36.78 | 37.18 | 38.39 | 38.59 | 39.01 | 37.23 | 37.39 | 31.79 | 31.87 | 36.98 | 37.46 | 36.86 | 37.01 |
| Robust FL [12] | 17.41 | 16.09 | 15.08 | 5.02 | 19.67 | 17.59 | 17.59 | 9.45 | 17.57 | 9.45 | 7.50 | 5.41 | 10.50 | 7.44 | 13.54 | 8.19 |
| FedLSR [30] | 36.10 | 28.13 | 25.30 | 18.20 | 43.74 | 41.42 | 43.92 | 40.23 | 35.07 | 27.97 | 24.23 | 20.73 | 42.11 | 40.29 | 35.45 | 34.14 |
| FedRN [33] | 20.28 | 19.73 | 19.53 | 17.32 | 43.32 | 42.51 | 35.74 | 21.55 | 19.80 | 18.94 | 19.07 | 17.88 | 42.71 | 40.59 | 34.59 | 34.27 |
| FedNoRo [4] | 44.70 | 43.41 | 32.98 | 31.41 | 43.51 | 43.56 | 43.52 | 41.51 | 45.09 | 43.79 | 32.39 | 30.33 | 43.97 | 43.24 | 43.54 | 43.27 |
| FedELC [13] | 44.83 | 43.65 | 33.07 | 32.95 | 44.01 | 43.83 | 43.87 | 42.73 | 45.22 | 43.12 | 32.98 | 32.67 | 44.97 | 43.64 | 45.39 | 42.81 |
| FedENLC | 51.56 | 49.89 | 39.29 | 37.76 | 51.16 | 49.13 | 50.29 | 48.58 | 49.84 | 47.68 | 40.83 | 39.14 | 50.01 | 47.40 | 51.95 | 49.02 |
| Method | CIFAR-10 | CIFAR-100 | ||
|---|---|---|---|---|
| Training Time | Test Time | Training Time | Test Time | |
| FedELC [13] | 37.19 s | 0.974 s | 58.789 s | 1.498 s |
| FedENLC | 33.49 s | 0.947 s | 53.581 s | 1.482 s |
| Method | Dirichlet ( = 1.0) | Dirichlet ( = 0.5) | ||
|---|---|---|---|---|
| Mixed (0.0–0.4) | Mixed (0.0–0.4) | |||
| FedENLC | 82.79 | 81.24 | 80.03 | 75.91 |
| Ablation1 | 82.38 | 81.23 | 79.45 | 74.87 |
| Ablation2 | 75.99 | 75.37 | 73.70 | 72.23 |
| Ablation3 | 81.80 | 80.06 | 77.63 | 72.62 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Cho, Y.; Kim, J. FedENLC: An End-to-End Noisy Label Correction Framework in Federated Learning. Mathematics 2026, 14, 290. https://doi.org/10.3390/math14020290
Cho Y, Kim J. FedENLC: An End-to-End Noisy Label Correction Framework in Federated Learning. Mathematics. 2026; 14(2):290. https://doi.org/10.3390/math14020290
Chicago/Turabian StyleCho, Yeji, and Junghyun Kim. 2026. "FedENLC: An End-to-End Noisy Label Correction Framework in Federated Learning" Mathematics 14, no. 2: 290. https://doi.org/10.3390/math14020290
APA StyleCho, Y., & Kim, J. (2026). FedENLC: An End-to-End Noisy Label Correction Framework in Federated Learning. Mathematics, 14(2), 290. https://doi.org/10.3390/math14020290

