Noise-Resilient Masked Face Detection Using Quantized DnCNN and YOLO
Abstract
1. Introduction
- End-to-End Noise-Resilient Detection Pipeline for Edge Deployment: We propose an end-to-end denoising–detection pipeline that integrates a lightweight DnCNN-based denoiser with the YOLOv11 detector, enabling robust masked-face detection under noisy imaging conditions on resource-constrained edge devices.
- Systematic Evaluation across Desktop and Embedded Platforms: We perform a comprehensive and controlled evaluation of the proposed pipeline on both a desktop workstation and the NVIDIA Jetson AGX Orin, focusing on detection robustness, quantization stability, and real-time feasibility under multiple noise levels.
- Practical INT8 Deployment via QDQ-Based Post-Training Quantization: We demonstrate that ONNX-compliant QDQ-based post-training quantization enables efficient INT8 acceleration of the denoising stage with minimal accuracy degradation, supporting practical deployment in latency-tolerant edge scenarios.
- Parallelized Edge-AI Execution Pipeline: We implement a parallelized CPU–GPU execution pipeline that overlaps preprocessing, denoising, and detection, significantly improving hardware utilization and increasing end-to-end throughput on the Jetson AGX Orin.
- Comprehensive Validation across Noise Levels and Hardware Settings: Extensive experiments on the FMLD dataset across noise variances from 0.01 to 0.10 confirm consistent detection improvements and demonstrate the robustness and deployability of the proposed system in real-world edge environments.
2. Related Work
2.1. Image Denoising and Restoration
2.2. Masked Face Detection Under Adverse Conditions
2.3. Quantization for Efficient and Robust Edge Deployment
3. Proposed Method
3.1. System Architecture: Initialization, Validation, and Edge Deployment
3.2. Denoising Strategy Using DnCNN
3.2.1. Comparison with Other Denoising Models
- Structural Difference (DnCNN vs. FFDNet): FFDNet is a non-blind denoiser that requires a noise level as an additional input channel. This dependency is impractical in dynamic scenes where the noise level varies unpredictably and cannot be measured in advance.
- Blind Denoising Capability: DnCNN operates as a blind denoiser, removing noise without any external knowledge of . Its residual-learning structure allows it to handle diverse and unknown degradation patterns, ensuring stable preprocessing across a wide range of conditions.
3.2.2. Noise Construction and Parameter Definition
3.3. Quantized DnCNN
3.3.1. Post-Training Quantization
3.3.2. QDQ-Based Post-Training Quantization (TensorRT INT8)
3.4. YOLOv11-Based Masked Face Detection with Frame Reconstruction
3.5. Edge Device Implementation on Jetson AGX Orin
3.6. Overall System Operation
4. Experiment Results
4.1. Evaluation Metrics
4.2. Dataset Preparation and Settings
4.3. Noise Robustness on Desktop
4.4. Quantization Stability on Jetson AGX Orin
4.5. Real-Time Edge Efficiency on Jetson AGX Orin
5. Discussion
5.1. Quantization Effects and Interpretation
5.2. Noise Modeling and Robustness Scope
5.3. Performance Boundaries and Failure Modes
5.4. System-Level Implications for Edge Deployment
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Chen, L.; Wang, H.; Wang, X.; Gao, J.; Deng, W. The Devil of Face Recognition Is in the Noise. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 765–782. [Google Scholar]
- Esmaeilzehi, A.; Ahmad, M.O.; Swamy, M.N.S. SRNHARB: A deep light-weight image super resolution network using hybrid activation residual blocks. Signal Processing Image Commun. 2021, 99, 116509. [Google Scholar] [CrossRef]
- Guo, Y.; Zhang, L.; Hu, Y.; He, X.; Gao, J. A Dataset and Benchmark for Large-Scale Face Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4873–4882. [Google Scholar]
- Batagelj, B.; Peer, P.; Struc, V.; Dobrisek, S. How to Correctly Detect Face-Masks for COVID-19 from Visual Information? Appl. Sci. 2021, 11, 2070. [Google Scholar] [CrossRef]
- Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
- He, L.; Zhou, Y.; Liu, L.; Ma, J. Research and Application of YOLOv11-Based Object Segmentation in Intelligent Construction-Site Recognition. Buildings 2024, 14, 3777. [Google Scholar] [CrossRef]
- Jacob, B.; Kligys, S.; Chen, B.; Zhu, M.; Tang, M.; Howard, A.; Adam, H. Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2704–2713. [Google Scholar]
- Nagel, S.; Shao, R.; Gouk, H.; Hospedales, T.M. QReg: On Regularization Effects of Quantization. arXiv 2022, arXiv:2206.12372. [Google Scholar] [CrossRef]
- Xu, Y.; Wang, Z.; Li, Z.; Xu, Y.; Tao, D. Fighting Overfitting with Quantization for Deep Neural Networks on Noisy Labels. arXiv 2023, arXiv:2303.11803. [Google Scholar]
- Wang, R.; Tang, Y.; Gong, C.; Liu, Y. In-Distribution Consistency Regularization for QAT Generalization. arXiv 2024, arXiv:2402.13497. [Google Scholar]
- Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
- Pande-Chhetri, R.; Abd-Elrahman, A. De-Striping Hyperspectral Imagery Using Wavelet Transform and Adaptive Frequency Domain Filtering. ISPRS J. Photogramm. Remote Sens. 2011, 66, 620–636. [Google Scholar] [CrossRef]
- Cui, H.; Jia, P.; Zhang, G.; Jiang, Y.-H.; Li, L.-T.; Wang, J.-Y.; Hao, X.-Y. Multiscale Intensity Propagation to Remove Multiplicative Stripe Noise From Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2308–2323. [Google Scholar] [CrossRef]
- Chang, Y.; Yan, L.; Liu, L.; Fang, H.; Zhong, S. Infrared Aerothermal Nonuniform Correction via Deep Multiscale Residual Network. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1120–1124. [Google Scholar] [CrossRef]
- Sun, W.; Ren, K.; Meng, X.; Yang, G.; Xiao, C.; Peng, J.; Huang, J. MLR-DBPFN: A Multi-scale Low Rank Deep Back Projection Fusion Network for Anti-noise Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522914. [Google Scholar] [CrossRef]
- Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 1833–1844. [Google Scholar]
- Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible CNN-Based Image Denoiser. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed]
- Guo, S.; Li, Q.; Zuo, W. Toward Convolutional Blind Denoising of Real Photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1712–1722. [Google Scholar]
- Li, L.; Jiang, L.; Zhang, J.; Wang, S.; Chen, F. A Complete YOLO-Based Ship Detection Method for Thermal Infrared Remote Sensing Images under Complex Backgrounds. Remote Sensing 2022, 14, 1534. [Google Scholar] [CrossRef]
- Rodríguez-Rodríguez, J.A.; López-Rubio, E.; Ángel-Ruiz, J.A.; Molina-Cabello, M.A. The Impact of Noise and Brightness on Object Detection Methods. Sensors 2024, 24, 821. [Google Scholar] [CrossRef] [PubMed]
- Li, J.; Zhang, S.; Zhang, X.; Wang, A. DiffuYOLO: A Novel Method for Small Vehicle Detection in Remote Sensing Based on Diffusion Models. Alex. Eng. J. 2025, 114, 485–496. [Google Scholar] [CrossRef]
- Liu, Y.; Li, S.; Zhou, L.; Liu, H.; Li, Z. Dark-YOLO: A Low-Light Object Detection Algorithm Integrating Multiple Attention Mechanisms. Appl. Sci. 2025, 15, 5170. [Google Scholar] [CrossRef]
- Ge, S.; Li, J.; Ye, Q.; Luo, Z. Detecting Masked Faces with LLE-CNNs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2682–2690. [Google Scholar]
- Batagelj, B. Face Mask Label Dataset (FMLD). GitHub Repository. 2024. Available online: https://github.com/borutb-fri/FMLD (accessed on 3 September 2025).
- Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. In Advances in Neural Information Processing Systems (NeurIPS); Vancouver, BC, Canada, Curran Associates, Inc.: Red Hook, NY, USA, 10-16 December 2024; Volume 37. [Google Scholar]
- Zhang, K.; Zuo, W.; Zhang, L. KAIR: Image Restoration Toolbox for Reproducible Research. GitHub Repository. 2021. Available online: https://github.com/cszn/KAIR (accessed on 3 September 2025).
- MathWorks. Denoising Network (MATLAB R2024b, Image Processing Toolbox). Available online: https://www.mathworks.com/help/images/ref/denoisingnetwork.html (accessed on 12 February 2025).
- Ke, R. Deep Variation Prior: Joint Image Denoising and Noise Variance Estimation Without Clean Data. IEEE Trans. Image Process. 2024, 33, 2908–2923. [Google Scholar] [CrossRef] [PubMed]
- NVIDIA Corporation. TensorRT Developer Guide: Quantization and Calibration; NVIDIA: Santa Clara, CA, USA, 2023. [Google Scholar]
- ONNX Working Group. ONNX Quantization Specification (Q/DQ Format). 2022. Available online: https://github.com/onnx/onnx/blob/main/docs/Operators.md (accessed on 17 December 2025).
- Apostolidis, K.D.; Papakostas, G.A. Delving into YOLO Object Detection Models: Insights into Adversarial Robustness. Electronics 2025, 14, 1624. [Google Scholar] [CrossRef]
- Ultralytics. Detection Datasets. Available online: https://docs.ultralytics.com/datasets/detect/ (accessed on 17 December 2025).
- Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and Flexible Image Augmentations. GitHub Repository. 2020. Available online: https://github.com/albumentations-team/albumentations (accessed on 3 September 2025).
- Hendrycks, D.; Dietterich, T. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019; Available online: https://openreview.net/forum?id=HJz6tiCqYm (accessed on 3 September 2025).
- Lee, J.; Hwang, K.-I. YOLO with Adaptive Frame Control for Real-Time Object Detection Applications. Multimed. Tools Appl. 2022, 81, 36375–36396. [Google Scholar] [CrossRef]










| Algorithm | Noise Level | PSNR (dB) | SSIM |
|---|---|---|---|
| DnCNN [5] | 36.72 | 0.945 | |
| FFDNet [17] | 35.62 | 0.936 | |
| SwinIR [16] | 37.69 | 0.955 | |
| DnCNN [5] | 34.70 | 0.920 | |
| FFDNet [17] | 34.69 | 0.925 | |
| SwinIR [16] | 35.45 | 0.935 | |
| DnCNN [5] | 31.14 | 0.868 | |
| FFDNet [17] | 31.55 | 0.880 | |
| SwinIR [16] | 32.41 | 0.897 |
| Algorithm | FPS |
|---|---|
| DnCNN [5] | 17.02 |
| FFDNet [17] | 64.91 |
| SwinIR [16] | 0.17 |
| Dataset | Images | Instances | Masked Face | Incorrectly Masked Face | Unmasked Face |
|---|---|---|---|---|---|
| FMLD (Updated) | 34,781 | 50,384 | 24,603 | 1204 | 24,576 |
| Validation (Updated) | 7148 | 12,675 | 7423 | 324 | 4928 |
| Totals | 41,934 | 63,059 | 32,026 | 1528 | 29,505 |
| Class | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
|---|---|---|---|---|
| Face | 0.817 | 0.824 | 0.880 | 0.623 |
| Mask | 0.940 | 0.955 | 0.978 | 0.654 |
| Incorrect Mask | 0.796 | 0.670 | 0.720 | 0.443 |
| Average | 0.851 | 0.816 | 0.859 | 0.573 |
| Distortion | Intensity | Method | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
|---|---|---|---|---|---|---|
| Baseline | – | Original YOLOv11 | 0.851 | 0.816 | 0.859 | 0.573 |
| Primary Target: Gaussian Noise | ||||||
| Gaussian Noise | Noise only | 0.830 | 0.712 | 0.793 | 0.510 | |
| DnCNN (FP32) | 0.842 | 0.800 | 0.844 | 0.619 | ||
| Q-DnCNN (FP16) | 0.849 | 0.793 | 0.846 | 0.561 | ||
| Q-DnCNN (INT8) | 0.841 | 0.780 | 0.760 | 0.554 | ||
| Noise only | 0.742 | 0.416 | 0.489 | 0.295 | ||
| DnCNN (FP32) | 0.831 | 0.733 | 0.812 | 0.527 | ||
| Q-DnCNN (FP16) | 0.819 | 0.722 | 0.801 | 0.517 | ||
| Q-DnCNN (INT8) | 0.761 | 0.700 | 0.768 | 0.492 | ||
| Noise only | 0.577 | 0.214 | 0.262 | 0.151 | ||
| DnCNN (FP32) | 0.751 | 0.655 | 0.723 | 0.456 | ||
| Q-DnCNN (FP16) | 0.732 | 0.637 | 0.704 | 0.441 | ||
| Q-DnCNN (INT8) | 0.708 | 0.598 | 0.659 | 0.414 | ||
| Verification: Real-world Distortions | ||||||
| Motion Blur | L1 | Noise only | 0.806 | 0.594 | 0.676 | 0.426 |
| L3 | Noise only | 0.677 | 0.348 | 0.406 | 0.235 | |
| L5 | Noise only | 0.578 | 0.208 | 0.241 | 0.130 | |
| Low Illumination | L1 | Noise only | 0.850 | 0.815 | 0.857 | 0.572 |
| L3 | Noise only | 0.831 | 0.797 | 0.838 | 0.545 | |
| L5 | Noise only | 0.802 | 0.710 | 0.761 | 0.470 | |
| JPEG Compression | L1 | Noise only | 0.846 | 0.815 | 0.855 | 0.572 |
| L3 | Noise only | 0.843 | 0.811 | 0.854 | 0.570 | |
| L5 | Noise only | 0.843 | 0.807 | 0.850 | 0.567 | |
| Model/Setting | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 | FPS |
|---|---|---|---|---|---|
| YOLO (Clean) | 0.8649 | 0.8031 | 0.8426 | 0.5893 | 14.58 |
| YOLO (Noise) | 0.8921 | 0.1427 | 0.5223 | 0.3238 | 26.09 |
| DnCNN FP16 (Serial) | 0.8387 | 0.4269 | 0.6427 | 0.4369 | 4.42 |
| DnCNN FP16 (Parallel) | 0.8465 | 0.4274 | 0.6464 | 0.4397 | 5.82 |
| DnCNN INT8 (Serial) | 0.8384 | 0.4172 | 0.6367 | 0.4364 | 6.24 |
| DnCNN INT8 (Parallel) (Ours) | 0.8471 | 0.4154 | 0.6402 | 0.4407 | 7.66 |
| Pipeline | FPS | Avg Power (W) | Max Power (W) | Energy (Wh) | FPS/W |
|---|---|---|---|---|---|
| YOLO-only (Clean) | 14.58 | 5.59 | 6.99 | 0.010 | 2.671 |
| YOLO-only (Noise) | 26.09 | 6.77 | 7.19 | 0.007 | 4.000 |
| DnCNN FP16 (Parallel) | 5.82 | 5.79 | 6.08 | 0.028 | 0.982 |
| DnCNN INT8 (Parallel) | 7.66 | 5.93 | 6.18 | 0.023 | 1.222 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Choi, R.; Lee, H.; Kim, B.-s.; Kim, S.; Kim, M.Y. Noise-Resilient Masked Face Detection Using Quantized DnCNN and YOLO. Electronics 2026, 15, 143. https://doi.org/10.3390/electronics15010143
Choi R, Lee H, Kim B-s, Kim S, Kim MY. Noise-Resilient Masked Face Detection Using Quantized DnCNN and YOLO. Electronics. 2026; 15(1):143. https://doi.org/10.3390/electronics15010143
Chicago/Turabian StyleChoi, Rockhyun, Hyunki Lee, Bong-seok Kim, Sangdong Kim, and Min Young Kim. 2026. "Noise-Resilient Masked Face Detection Using Quantized DnCNN and YOLO" Electronics 15, no. 1: 143. https://doi.org/10.3390/electronics15010143
APA StyleChoi, R., Lee, H., Kim, B.-s., Kim, S., & Kim, M. Y. (2026). Noise-Resilient Masked Face Detection Using Quantized DnCNN and YOLO. Electronics, 15(1), 143. https://doi.org/10.3390/electronics15010143

