Efficient Synthetic Defect on 3D Object Reconstruction and Generation Pipeline for Digital Twins Smart Factory
Abstract
1. Introduction
- We created a novel dataset focused on synthetic scratches on 3D industrial object surfaces by combining NeRF-based 3D reconstruction with synthetic defect generation.
- We introduce a framework utilizing NeRF to reconstruct high-fidelity 3D models of industrial objects from 2D images. This approach circumvents the limitations of CAD, particularly licensing restrictions.
- Our framework incorporates the NVIDIA Omniverse Replicator to simulate synthetic scratches on reconstructed 3D objects. This work enables the generation of diverse, annotated datasets critical for training robust deep learning-based defect detectors.
- We benchmark the performance of several YOLO object detectors on our datasets. The experiment’s results demonstrate that incorporating synthetic defects along with real ones in the training process significantly enhances the generalization performance of these models.
2. Related Work
3. Proposed Method
3.1. NeRF-Based Model for 3D Reconstruction
3.2. Synthetic Surface Defect Generation
4. Experiment Setup and Results
4.1. NeRF-Based 3D Reconstruction Evaluation Metrics
4.2. YOLO-Based Object Detection Evaluation Metrics
4.3. Experiment Results
4.3.1. 3D Reconstruction Results
4.3.2. Synthetic Defect Rendering Results
4.3.3. Deep Learning Object Detection Models Results
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- NVIDIA. Transforming Industrial Defect Detection with NVIDIA TAO and Vision AI Models. Available online: https://developer.nvidia.com/blog/transforming-industrial-defect-detection-with-nvidia-tao-and-vision-ai-models/ (accessed on 10 September 2025).
- Schieber, H.; Demir, K.; Kleinbeck, C.; Yang, S.; Roth, D. Indoor synthetic data generation: A systematic review. Comput. Vis. Image Underst. 2024, 240, 103907. [Google Scholar] [CrossRef]
- Unity Technologies. Unity Documentation. Available online: https://docs.unity.com/ (accessed on 10 September 2025).
- Epic Games. Unreal Engine. Available online: https://www.unrealengine.com/en-US (accessed on 10 September 2025).
- NVIDIA. Omniverse Replicator Extension. Available online: https://docs.omniverse.nvidia.com/extensions/latest/ext_replicator.html (accessed on 10 September 2025).
- Blender Foundation. Blender. Available online: https://www.blender.org/ (accessed on 10 September 2025).
- Aerial Precision. Exploring the Precision of LiDAR Mapping: Unveiling Its Accuracy. Available online: https://www.aerial-precision.com/blogs/blog/unveiling-the-precision-of-lidar-mapping-how-accurate-is-it-really (accessed on 10 September 2025).
- Fraser, C.; Brow, D. Industrial photogrammetry: New developments and recent applications. Photogramm. Rec. 2006, 12, 197–217. [Google Scholar] [CrossRef]
- Sansoni, G.; Trebeschi, M.; Docchio, F. State-of-the-art and applications of 3D imaging sensors in industry, cultural heritage, medicine, and criminal investigation. Sensors 2009, 9, 568–601. [Google Scholar] [CrossRef] [PubMed]
- Rodríguez-Martín, M.; Lagüela, S.; González-Aguilera, D.; Rodríguez-Gonzálvez, P. Procedure for quality inspection of welds based on macro-photogrammetric three-dimensional reconstruction. Opt. Laser Technol. 2015, 73, 54–62. [Google Scholar] [CrossRef]
- Remondino, F. Heritage recording and 3D modeling with photogrammetry and 3D scanning. Remote Sens. 2011, 3, 1104–1138. [Google Scholar] [CrossRef]
- Menna, F.; Nocerino, E.; Remondino, F.; Dellepiane, M.; Callieri, M.; Scopigno, R. 3D digitization of an heritage masterpiece—A critical analysis on quality assessment. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 675–683. [Google Scholar] [CrossRef]
- Stylianidis, E.; Remondino, F. 3D Recording, Documentation and Management of Cultural Heritage; Whittles Publishing: Dunbeath, UK, 2016; p. 388. ISBN 978-184995-168-5. [Google Scholar]
- Verhoeven, G.; Wild, B.; Schlegel, J.; Wieser, M.; Pfeifer, N.; Wogrin, S.; Eysn, L.; Carloni, M.; Koschiček-Krombholz, B.; Molada-Tebar, A.; et al. Project Indigo—Document, disseminate & analyse a graffiti-scape. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 46, 513–520. [Google Scholar]
- Remondino, F.; El-Hakim, S. Image-based 3D modelling: A review. Photogramm. Rec. 2006, 21, 269–291. [Google Scholar] [CrossRef]
- Wu, B.; Zhou, Y.; Qian, Y.; Gong, M.; Huang, H. Full 3D reconstruction of transparent objects. arXiv 2018, arXiv:1805.03482. [Google Scholar] [CrossRef]
- Santoši, Ž.; Budak, I.; Stojaković, V.; Šokac, M.; Vukelić, Đ. Evaluation of synthetically generated patterns for image-based 3D reconstruction of texture-less objects. Measurement 2019, 147, 106883. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. All you need is attention. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. NeRF: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
- Nerfstudio Project. Nerfstudio: A Modular Framework for Neural Radiance Field Development. Available online: https://github.com/nerfstudio-project/nerfstudio/ (accessed on 10 September 2025).
- NVLabs. Instant Neural Graphics Primitives. Available online: https://github.com/NVlabs/instant-ngp (accessed on 10 September 2025).
- Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 2022, 41, 102–117. [Google Scholar] [CrossRef]
- Tancik, M.; Weber, E.; Ng, E.; Li, R.; Yi, B.; Wang, T.; Kristoffersen, A.; Austin, J.; Salahi, K.; Ahuja, A.; et al. Nerfstudio: A modular framework for neural radiance field development. arXiv 2023, arXiv:2302.04264. [Google Scholar] [CrossRef]
- Volinga. Volinga Model. Available online: https://github.com/Volinga/volinga-model (accessed on 10 September 2025).
- Chen, A.; Xu, Z.; Geiger, A.; Yu, J.; Su, H. TensoRF: Tensorial radiance fields. In Proceedings of the 17th European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
- Wang, H.; Sridhar, S.; Huang, J.; Valentin, J.; Song, S.; Guibas, L.J. Normalized object coordinate space for category-level 6D object pose and size estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2642–2651. [Google Scholar]
- Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. ShapeNet: An information-rich 3D model repository. arXiv 2015, arXiv:1512.03012. [Google Scholar]
- Shi, G.; Zhu, Y.; Tremblay, J.; Birchfield, S.; Ramos, F.; Anandkumar, A.; Zhu, Y. Fast uncertainty quantification for deep object pose estimation. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021. [Google Scholar]
- To, T.; Tremblay, J.; McKay, D.; Yamaguchi, Y.; Leung, K.; Balanon, A.; Cheng, J.; Hodge, W.; Birchfield, S. NDDS: NVIDIA deep learning dataset synthesizer. In Proceedings of the CVPR 2018 Workshop on Real World Challenges and New Benchmarks for Deep Learning in Robotic Vision, Salt Lake City, UT, USA, 22 June 2018. [Google Scholar]
- Back, S.; Lee, J.; Kim, T.; Noh, S.; Kang, R.; Bak, S.; Lee, K. Unseen object amodal instance segmentation via hierarchical occlusion modeling. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 5085–5092. [Google Scholar]
- Denninger, M.; Sundermeyer, M.; Winkelbauer, D.; Olefir, D.; Hodan, T.; Zidan, Y.; Elbadrawy, M.; Knauer, M.; Katam, H.; Lodhi, A. BlenderProc: Reducing the reality gap with photorealistic rendering. In Proceedings of the Robotics: Science and Systems (RSS) Workshops, Corvalis, OR, USA, 12–16 June 2020. [Google Scholar]
- Kasper, A.; Xue, Z.; Dillmann, R. The KIT object models database: An object model database for object recognition, localization and manipulation in service robotics. Int. J. Robot. Res. 2012, 31, 927–934. [Google Scholar] [CrossRef]
- Singh, A.; Sha, J.; Narayan, K.S.; Achim, T.; Abbeel, P. BigBIRD: A large-scale 3D database of object instances. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; IEEE: New York, NY, USA, 2014; pp. 509–516. [Google Scholar]
- Hodaň, T.; Sundermeyer, M.; Drost, B.; Labbé, Y.; Brachmann, E.; Michel, F.; Rother, C.; Matas, J. BOP challenge 2020 on 6D object localization. In Proceedings of the European Conference on Computer Vision Workshops, Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Ng, Z.; Wang, H.; Zhang, Z.; Hock, F.T.E.; Ang, M.H., Jr. SynTable: A synthetic data generation pipeline for unseen object amodal instance segmentation of cluttered tabletop scenes. arXiv 2023, arXiv:2307.07333. [Google Scholar]
- NVIDIA. Omniverse Isaac Sim Replicator Tutorial. Available online: https://docs.omniverse.nvidia.com/extensions/latest/ext_replicator/getting_started.html (accessed on 10 September 2025).
- Downs, L.; Francis, A.; Koenig, N.; Kinman, B.; Hickman, R.; Reymann, K.; McHugh, T.B.; Vanhoucke, V. Google scanned objects: A high-quality dataset of 3D scanned household items. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 2553–2560. [Google Scholar]
- NVIDIA. NVIDIA Kit Extension Sample—Defects Generation. Available online: https://github.com/NVIDIA-Omniverse/kit-extension-sample-defectsgen (accessed on 10 September 2025).
- Zeng, Y.; Lei, J.; Feng, T.; Qin, X.; Li, B.; Wang, Y.; Wang, D.; Song, J. Neural radiance fields-based 3D reconstruction of power transmission lines using progressive motion sequence images. Sensors 2023, 23, 9537. [Google Scholar] [CrossRef]
- Mazzacca, G.; Karami, A.; Rigon, S.; Farella, E.M.; Trybala, P.; Remondino, F. NeRF for heritage 3D reconstruction. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 48, 1051–1058. [Google Scholar] [CrossRef]
- Remondino, F.; Karami, A.; Yan, Z.; Mazzacca, G.; Rigon, S.; Qin, R. A critical analysis of NeRF-based 3D reconstruction. Remote Sens. 2023, 15, 3585. [Google Scholar] [CrossRef]
- Caruso, B.; Mahendrakar, T.; Nguyen, V.M.; White, R.T.; Steffen, T. 3D reconstruction of non-cooperative resident space objects using instant NGP-accelerated NeRF and D-NeRF. arXiv 2023, arXiv:2301.09060. [Google Scholar]
- Pei, S.; Yang, R.; Liu, Y.; Xu, W.; Zhang, G. Research on 3D reconstruction technology of large-scale substation equipment based on NeRF. IET Sci. Meas. Technol. 2023, 17, 71–83. [Google Scholar] [CrossRef]
- COLMAP. Available online: https://github.com/colmap/colmap (accessed on 10 September 2025).
- Gao, K.; Gao, Y.; He, H.; Lu, D.; Xu, L.; Li, J. NeRF: Neural radiance field in 3D vision, a comprehensive review. arXiv 2022, arXiv:2210.00379. [Google Scholar]
- Meshroom 3D Reconstruction Software. Available online: https://github.com/alicevision/Meshroom (accessed on 10 September 2025).
- NeRF: Neural Radiance Fields. Available online: https://github.com/bmild/nerf (accessed on 10 September 2025).
- SP Auto. SP Auto Product Page. Available online: http://m.spauto.co.kr/en/m/sub/product/auto.asp (accessed on 10 September 2025).
- Ultralytics. YOLOv5. Available online: https://github.com/ultralytics/yolov5/releases/tag/v6.0 (accessed on 10 September 2025).
- Meituan. YOLOv6. Available online: https://github.com/meituan/YOLOv6 (accessed on 10 September 2025).
- Megvii. YOLOX. Available online: https://github.com/Megvii-BaseDetection/YOLOX (accessed on 10 September 2025).
- Wong, K.Y. YOLOv7. Available online: https://github.com/WongKinYiu/yolov7 (accessed on 10 September 2025).
- Ultralytics. YOLOv8. Available online: https://github.com/ultralytics/ultralytics (accessed on 10 September 2025).
- THU-MIG. YOLOv10. Available online: https://github.com/THU-MIG/yolov10 (accessed on 10 September 2025).
- Wong, K.Y. YOLOv9. Available online: https://github.com/WongKinYiu/yolov9 (accessed on 10 September 2025).
- NEU-DET. IEEE Dataport. 7 November 2024. Available online: https://ieee-dataport.org/documents/neu-det (accessed on 21 October 2025).
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
- Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In Proceedings of the 13th European Conference on Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]














| Reference | Object | Methods |
|---|---|---|
| SynTable 2023 [35] | CAD/YCB | NVIDIA Omniverse Issac Sim |
| UOAIS-SIM 2023 [31] | 3D Assets | Blender |
| Synthetic HOPE 2021 [28] | Daily Object | NDDS Unreal |
| NOCS 2019 [26] | ShapeNetCore | Unity |
| Reference | Object | Methods | Evaluation Metrics |
|---|---|---|---|
| 2023 [39] | Power Transmission Line | Instant-NGP Volinga | PSNR SSIM LIPPS fps |
| 2023 [40] | Heritage | Instant-NGP Nerfacto TensoRF | RMSE MAE STD |
| 2023 [41] | Ignatius Statue Truck Stair Synthetic Industrial Bottle | Nerfacto Instant-NGP TensoRF | RMSE MAE STD |
| 2023 [42] | Space Object | Instant-NGP D-Nerf | PSNR SSIM LPIPS |
| 2023 [43] | Substation Equipment | COLMAP NeRF Instant-NGP | PSNR SSIM LPIPS speed |
| Dataset | Images | Scratch Labels | |
|---|---|---|---|
| Our synthetic dataset | 1890 | Training set: 1512 | ./. |
| Validation set: 378 | 411 | ||
| NEU-DET real scratch | 300 | Training set: 240 | ./. |
| Validation set: 60 | 121 |
| Models | Training Iteration (ms) | Industrial Object 1 | Industrial Object 2 | |||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Dataset 1: 102 Images Train/Val: 92/10 | Dataset 2: 192 Images Train/Val: 173/19 | Dataset 3: 192 Images Train/Val: 173/19 | ||||||||
| PSNR | SSIM | LPIPS | PSNR | SSIM | LPIPS | PSNR | SSIM | LPIPS | ||
| Instant-NPG | 16.192 | 18.90 | 0.78 | 0.57 | 22.83 | 0.85 | 0.46 | 16.28 | 0.87 | 0.59 |
| Nerfacto | 24.356 | 14.90 | 0.74 | 0.60 | 21.48 | 0.84 | 0.44 | 21.58 | 0.89 | 0.44 |
| Volinga | 23.443 | 14.04 | 0.73 | 0.63 | 20.65 | 0.83 | 0.44 | 20.38 | 0.88 | 0.45 |
| TensoRF | 61.196 | 14.52 | 0.76 | 0.64 | 17.51 | 0.83 | 0.5 | 16.96 | 0.86 | 0.58 |
| Method | Paras (M) | FLOPs (G) | mAP@0.5 | mAP@0.5:0.95 | Precision | Recall | F1-Score |
|---|---|---|---|---|---|---|---|
| YOLOv5-n | 1.76 | 4.1 | 70.4 | 30.7 | 73.2 | 63.7 | 69 |
| YOLOv5-s | 7.01 | 15.8 | 85.1 | 41 | 84.2 | 80.3 | 82 |
| YOLOv6-n | 4.63 | 11.34 | 87.2 | 41.1 | 88.8 | 83.2 | 85.9 |
| YOLOv6-s | 18.05 | 45.17 | 91.3 | 41.2 | 92.4 | 89.1 | 90.7 |
| YOLOX-n | 0.91 | 1.08 | 86.97 | 40.59 | -/- | -/- | -/- |
| YOLOX-s | 8.94 | 26.76 | 90.6 | 48.42 | -/- | -/- | -/- |
| YOLOv7-Tiny | 6.01 | 13.2 | 53.8 | 20.8 | 61.4 | 51.6 | 56 |
| YOLOv7 | 37.19 | 105.1 | 44.8 | 13.9 | 51.1 | 49.4 | 50 |
| YOLOv8-n | 3.0 | 8.1 | 92.7 | 49.5 | 92.3 | 87 | 89 |
| YOLOv8-s | 11.12 | 28.4 | 96.4 | 53 | 96.3 | 92 | 94 |
| YOLOv10-n | 2.69 | 8.2 | 81.6 | 40.2 | 80.2 | 73.1 | 76 |
| YOLOv10-s | 8.03 | 24.4 | 89.4 | 48.5 | 86.6 | 82 | 84 |
| Train: Real, Val: Real | Train: Real + Synthetic, Val: Real | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Method | Paras (M) | FLOPs (G) | mAP@ 0.5 | mA@ 0.5:0.95 | p | R | F1-Score | mAP@ 0.5 | mAP@ 0.5:0.95 | p | R | F1-Score |
| YOLOv5n | 1.76 | 4.1 | 22 | 7.49 | 20.2 | 0.43 | 25 | 33.5 | 11.1 | 29.3 | 62.9 | 34 |
| YOLOv5s | 7.01 | 15.8 | 54.2 | 19 | 52.2 | 56.9 | 54 | 63.4 | 25 | 50.8 | 76 | 61 |
| YOLOv6n | 4.63 | 11.34 | 61.7 | 0.283 | 61 | 59.5 | 60.2 | 80.5 | 41.6 | 70.4 | 82.6 | 76 |
| YOLOv6s | 18.05 | 45.17 | 77.4 | 41.6 | 73.6 | 76 | 74.8 | 81.9 | 45.1 | 78.2 | 74.4 | 76.2 |
| YOLOv8n | 3.0 | 8.1 | 59.2 | 28.6 | 48.8 | 68.6 | 56 | 62.4 | 28.8 | 62 | 62 | 61 |
| YOLOv8s | 11.12 | 28.4 | 73.5 | 33.3 | 72.3 | 70.2 | 70 | 75 | 36.2 | 65 | 78.5 | 72 |
| YOLO10n | 2.69 | 8.2 | 36.7 | 18 | 44.7 | 44.6 | 41 | 49.3 | 21.7 | 46.4 | 49.6 | 49 |
| YOLO10s | 8.03 | 24.4 | 61.8 | 29.6 | 65.9 | 61.2 | 61.8 | 65.3 | 22.2 | 63.4 | 70.1 | 66 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Nguyen, V.-H.; Pham, T.-N.; Huh, J.-H.; Choi, P.-J.; Kim, Y.-B.; Kwon, O.-H.; Kwon, K.-R. Efficient Synthetic Defect on 3D Object Reconstruction and Generation Pipeline for Digital Twins Smart Factory. Sensors 2025, 25, 6908. https://doi.org/10.3390/s25226908
Nguyen V-H, Pham T-N, Huh J-H, Choi P-J, Kim Y-B, Kwon O-H, Kwon K-R. Efficient Synthetic Defect on 3D Object Reconstruction and Generation Pipeline for Digital Twins Smart Factory. Sensors. 2025; 25(22):6908. https://doi.org/10.3390/s25226908
Chicago/Turabian StyleNguyen, Viet-Hoan, Thi-Ngot Pham, Jun-Ho Huh, Pil-Joo Choi, Young-Bong Kim, Oh-Heum Kwon, and Ki-Ryong Kwon. 2025. "Efficient Synthetic Defect on 3D Object Reconstruction and Generation Pipeline for Digital Twins Smart Factory" Sensors 25, no. 22: 6908. https://doi.org/10.3390/s25226908
APA StyleNguyen, V.-H., Pham, T.-N., Huh, J.-H., Choi, P.-J., Kim, Y.-B., Kwon, O.-H., & Kwon, K.-R. (2025). Efficient Synthetic Defect on 3D Object Reconstruction and Generation Pipeline for Digital Twins Smart Factory. Sensors, 25(22), 6908. https://doi.org/10.3390/s25226908

