An Automated Training of Deep Learning Networks by 3D Virtual Models for Object Recognition
Abstract
:1. Introduction
1.1. Problem Specification and Motivation
1.2. Research Idea and Benefits
2. Relevant Previous Works
3. Proposed Work
3.1. The Principle of 2D Sample Generation
- Import the 3D model from any 3D CAD software (CATIA, CREO, Autodesk Inventor, etc.) to 3D rendering software with integrated scripting language.
- Generate 2D sample images with different object rotations/position, textures, and backgrounds, including standard views.
- Generate the object bounding box in the 2D image with basic parameters (position/dimension) for part localization by standard image processing techniques (Python OpenCV library) used for single shot detection (SSD) during training.
- Generate the text description file (XML—eXtensible Markup Language) with the following basic image parameters: Name of file, size and position, and image resolution.
- Separate the random samples into two groups: Training set (80%) and validation set (20%).
- Ry, Rz—matrix of virtual 3D rotation of part.
- β, γ—rotation angle for each additional generated view.
- Txyz—matrix of the virtual 3D translation of a part.
- tx, ty, tz—random translation in pixels for each additional generated view.
3.2. Input Images Used for the Training Process
- Chrome part material—wood table base.
- Brass part material—base steel plate.
- Steel material—polished rock workplace.
3.3. Training Process
4. Experiments
4.1. Experiment Environment
- TensorFlow (Python)—teaching CNN (model export and optimization).
- Android Studio (Java/C++)—CNN execution (recognition processing delay minimization).
- Unity (C#)—3D engine—data visualization (3D model visualization in augmented reality).
4.2. Evaluation Measure
4.3. Experiment Process
4.4. Experiment Results
4.5. Discussion
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Pavlenko, I.; Simonovskiy, V.; Ivanov, V.; Zajac, J.; Pitel, J. Application of artificial neural network for identification of bearing stiffness characteristics in rotor dynamics analysis. In Advances in Design, Simulation and Manufacturing; Lecture Notes in Mechanical Engineering; Springer: Cham, Switzerland, 2019; pp. 325–335. [Google Scholar]
- Zidek, K.; Piteľ, J.; Hošovský, A. Machine learning algorithms implementation into embedded systems with web application user interface. In Proceedings of the IEEE 21st International Conference on Intelligent Engineering Systems (INES 2017), Larnaca, Cyprus, 20–23 October 2017; pp. 77–81. [Google Scholar]
- Zidek, K.; Maxim, V.; Pitel, J.; Hosovsky, A. Embedded vision equipment of industrial robot for inline detection of product errors by clustering-classification algorithms. Int. J. Adv. Robot. Syst. 2016, 13. [Google Scholar] [CrossRef]
- Shang, Y. Deffuant model of opinion formation in one-dimensional multiplex networks. J. Phys. A Math. Theor. 2015, 48, 395101. [Google Scholar] [CrossRef]
- Pavlenko, I.; Trojanowska, J.; Ivanov, V.; Liaposhchenko, O. Scientific and methodological approach for the identification of mathematical models of mechanical systems by using artificial neural networks. In Innovation, Engineering and Entrepreneurship; Lecture Notes in Electrical Engineering; Springer: Cham, Switzerland, 2019; pp. 299–306. [Google Scholar]
- Židek, K.; Maxim, V.; Sadecký, R. Diagnostics of errors at component surface by vision recognition in production systems. Appl. Mech. Mater. 2014, 616, 227–235. [Google Scholar] [CrossRef]
- Židek, K.; Hošovský, A.; Dubjak, J. Diagnostics of surface errors by embedded vision system and its classification by machine learning algorithms. Key Eng. Mater. 2016, 669, 459–466. [Google Scholar] [CrossRef]
- Židek, K.; Hosovsky, A.; Piteľ, J.; Bednár, S. Recognition of assembly parts by convolutional neural networks. In Advances in Manufacturing Engineering and Materials; Lecture Notes in Mechanical Engineering; Springer: Cham, Switzerland, 2019; pp. 281–289. [Google Scholar]
- Sarkar, K.; Varanasi, K.; Stricker, D. Trained 3D models for CNN based object recognition. In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (Visigrapp 2017), Porto, Portugal, 27 February–1 March 2017; pp. 130–137. [Google Scholar]
- Socher, R.; Huval, B.; Bhat, B.; Manning, C.D.; Ng, A.Y. Convolutional-recursive deep learning for 3D object classification. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS’12), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 656–664. [Google Scholar]
- Su, H.; Qi, C.R.; Li, Y.; Guibas, L.J. Render for CNN: Viewpoint estimation in images using CNNs trained with rendered 3D model views. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV’15), Santiago, Chile, 7–13 December 2015; pp. 2686–2694. [Google Scholar]
- Tian, Y.; Li, X.; Wang, K.; Wang, F.Y. Training and testing object detectors with virtual images. IEEE/CAA J. Autom. Sin. 2018, 5, 539–546. [Google Scholar] [CrossRef]
- Nie, W.Z.; Cao, Q.; Liu, A.A.; Su, Y.T. Convolutional deep learning for 3D object retrieval. Multimed. Syst. 2017, 23, 325–332. [Google Scholar] [CrossRef]
- Peng, X.C.; Sun, B.C.; Ali, K.; Saenko, K. Learning deep object detectors from 3D models. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV’15), Santiago, Chile, 7–13 December 2015; pp. 1278–1286. [Google Scholar]
- Cucos, M.M.; Pista, I.M.; Ripanu, M.I. Product engineering design enhancing by parameterizing the 3D solid model. MATEC Web Conf. 2018, 178, 05011. [Google Scholar] [CrossRef]
- Nguyen, C.H.P.; Choi, Y. Triangular mesh and boundary representation Combined Approach for 3D CAD Lightweight Representation for Collaborative Product Development. J. Comput. Inf. Sci. Eng. 2018, 19, 011009. [Google Scholar] [CrossRef]
- Pollák, M.; Baron, P.; Zajac, J. Automatization process of designing the technological documentation by tools of comprehensive CAD-CAM-CAE system. MM Sci. J. 2016, 942–946. [Google Scholar] [CrossRef]
- Pollák, M.; Baron, P.; Telišková, M.; Kočiško, M.; Török, J. Design of the web interface to manage automatically generated production documentation. Tech. Technol. Educ. Manag. TTEM 2018, 7, 703–707. [Google Scholar]
- Gopalakrishnan, K. Deep Learning in data-driven pavement image analysis and automated distress detection: A review. Data 2018, 3, 28. [Google Scholar] [CrossRef]
- Mao, K.; Lu, D.; E, D.; Tan, Z. A case study on attribute recognition of heated metal mark image using deep convolutional neural networks. Sensors 2018, 18, 1871. [Google Scholar] [CrossRef] [PubMed]
- Pham, G.; Lee, S.-H.; Kwon, O.-H.; Kwon, K.-R. Anti-3D weapon model detection for safe 3D printing based on convolutional neural networks and D2 shape distribution. Symmetry 2018, 10, 90. [Google Scholar] [CrossRef]
- Liu, H.; Wang, L. An AR-based worker support system for human-robot collaboration. Procedia Manuf. 2017, 11, 22–30. [Google Scholar] [CrossRef]
- Gerstweiler, G.; Furlan, L.; Timofeev, M.; Kaufmann, H. Extraction of structural and semantic data from 2D floor plans for interactive and immersive VR real estate exploration. Technologies 2018, 6, 101. [Google Scholar] [CrossRef]
- Gao, Q.H.; Wan, T.R.; Tang, W.; Chen, L.; Zhang, K.B. An improved augmented reality registration method based on Visual SLAM. In E-Learning and Games; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2017; pp. 11–19. [Google Scholar]
- Židek, K.; Hošovský, A. Wireless device based on MEMS sensors and bluetooth low energy (LE/smart) technology for diagnostics of mechatronic systems. Appl. Mech. Mater. 2014, 460, 13–21. [Google Scholar] [CrossRef]
- Hruboš, M.; Svetlík, J.; Nikitin, Y.; Pirník, R.; Nemec, D.; Šimák, V.; Janota, A.; Hrbček, J.; Gregor, M. Searching for collisions between mobile robot and environment. Int. J. Adv. Robot. Syst. 2016, 13. [Google Scholar] [CrossRef] [Green Version]
- Shang, Y. On the delayed scaled consensus problems. Appl. Sci. 2017, 7, 713. [Google Scholar] [CrossRef]
Part Name | Training: Example of Virtual Part Samples (Changed Material/Background) | Testing with TensorFlow Framework (Python Execution Experiments) | mAP min/max | |
---|---|---|---|---|
Inception V2 | Mobilenet V2 | |||
Screw M12 | 0.96/0.99 | |||
Nut M12 | 0.93/0.86 | |||
Washer 12 | 0.99/1 |
Processor Type\Recognition Processing Delays [ms] | Screw (700 × 525) Inception/Mobilenet | Nut (390 × 300) Inception/Mobilenet | Washer (700 × 525) Inception/Mobilenet |
---|---|---|---|
Intel Core i5-8400 2.8 GHz (HP Omen) | 771.09/109.83 | 672.04/43.39 | 821.40/106.18 |
Amlogic A57 1.5 GHz (Odroid C2) | 14,192.91/1339.74 | 10,924.65/675.78 | 14,311.28/1308.42 |
Allwinner H3 A7 1.6 GHz (Orange Pi Lite Plus) | 48,482.33/3171.07 | -/1484.84 | 52,294.71/3661.11 |
Broadcom A53 1.4 GHz (Raspberry Pi 3) | 44,886.64/6814.01 | 33,707.40/2488.06 | 44,661.78/6933.43 |
Device Type\Delay [ms] | Screw (320 × 240) | Nut (320 × 240) | Washer (320 × 240) |
---|---|---|---|
Cardboard VR Exynos A53 4 × 2.3 GHz Android 8.1 (Samsung S7) | 1347.36 | 1375.42 | 1389.36 |
Smart Glasses AR Atom x5 4 × 1.44 GHz Android 5.1 (Epson Moverio bt-350) | 653.30 | 650.06 | 634.48 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Židek, K.; Lazorík, P.; Piteľ, J.; Hošovský, A. An Automated Training of Deep Learning Networks by 3D Virtual Models for Object Recognition. Symmetry 2019, 11, 496. https://doi.org/10.3390/sym11040496
Židek K, Lazorík P, Piteľ J, Hošovský A. An Automated Training of Deep Learning Networks by 3D Virtual Models for Object Recognition. Symmetry. 2019; 11(4):496. https://doi.org/10.3390/sym11040496
Chicago/Turabian StyleŽidek, Kamil, Peter Lazorík, Ján Piteľ, and Alexander Hošovský. 2019. "An Automated Training of Deep Learning Networks by 3D Virtual Models for Object Recognition" Symmetry 11, no. 4: 496. https://doi.org/10.3390/sym11040496
APA StyleŽidek, K., Lazorík, P., Piteľ, J., & Hošovský, A. (2019). An Automated Training of Deep Learning Networks by 3D Virtual Models for Object Recognition. Symmetry, 11(4), 496. https://doi.org/10.3390/sym11040496