Next Article in Journal
Application of Biographical Data in Student’s Major Selection
Next Article in Special Issue
Deep Reinforcement Learning-Based Scheduler on Parallel Dedicated Machine Scheduling Problem towards Minimizing Total Tardiness
Previous Article in Journal
How Engagement in the Industry-Education Integration Promotes One’s Attitudes toward Energy Efficiency. Evidence from Chinese University Students
Previous Article in Special Issue
Identification of Machine Learning Relevant Energy and Resource Manufacturing Efficiency Levers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of YOLO and ResNet in Heat Staking Process Inspection

1
Graduate School of Carbon Neutrality, Ulsan National Institute of Science and Technology (UNIST), Ulsan 44919, Republic of Korea
2
Data Science Group, INTERX, Seoul 08503, Republic of Korea
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(23), 15892; https://doi.org/10.3390/su142315892
Submission received: 30 October 2022 / Revised: 24 November 2022 / Accepted: 25 November 2022 / Published: 29 November 2022
(This article belongs to the Special Issue Application of Artificial Intelligence in Sustainable Manufacturing)

Abstract

:
In the automobile manufacturing industry, inspecting the quality of heat staking points in a door trim involves significant labor, leading to human errors and increased costs. Artificial intelligence has provided the industry some aid, and studies have explored using deep learning models for object detection and image classification. However, their application to the heat staking process has been limited. This study applied an object detection algorithm, the You Only Look Once (YOLO) framework, and a classification algorithm, residual network (ResNet), to a real heat staking process image dataset. The study leverages the advantages of YOLO models and ResNet to increase the overall efficiency and accuracy of detecting heat staking points from door trim images and classify whether the detected heat staking points are defected or not. The proposed model achieved high accuracy in both object detection (mAP of 95.1%) and classification (F1-score of 98%). These results show that the developed deep learning models can be applied to the real-time inspection of the heat staking process. The models can increase productivity and quality while decreasing human labor cost, ultimately improving a firm’s competitiveness.

1. Introduction

Technological advancement has enabled the development of various practical deep learning methodologies. Deep learning frameworks and architectures, such as YOLO (You Only Look Once) or ResNet, provide highly accurate and precise real-time identifications of objects [1]. These models have been used in solving on-site issues in diverse fields. This study attempts to further test the validity of recent deep learning models by identifying and classifying the quality of heat staking points. This study specifically focuses on the heat staking process of points on automobile door trims.
Employing deep learning-based quality prediction in the manufacturing process is particularly valuable because there are over sixty staking points in a single door trim and inspecting the quality of all points in a limited takt time is difficult. Furthermore, human errors are inevitable during measurements [2]. Because of these errors, acceptable product points are sometimes rejected (also known as “overkilled”), and defective product points are accepted as acceptable product points (also known as “escaped”). Both overkilling and escaping lead to tragic results, as overkill increases production costs and escape causes critical customer dissatisfaction.
One solution to partially alleviate these hurdles is to employ a machine vision system. The idea is to set up an environment for machine vision, take vision images, and run rule-based tests to check the quality of products. However, applying rule-based methods in a heat staking process is difficult because the locations and sizes of staking points vary from product to product. Therefore, it is recommended to employ a deep learning framework that is relatively free from the problems of rule-based algorithms. In this manner, studies have employed various deep learning frameworks and discuss the results.
The problems of the current inspection process of heat staking points in the automotive industry are that the process fully relies on the human labor and that human labor often incorporates inspection errors. That is, due to various reasons such as immature work level, tiredness, and so on, inspection errors exist. Therefore, this study tried to employ various deep learning models to determine whether the artificial intelligence technology is an effective strategy to enhance the inspection process of heat staking points. In terms of the methodology, this paper tried to apply two different objectives—object detection and classification—and combine them into one deep learning model to apply in the inspection process. For the object detection, this study applied the YOLO methodology as it is a powerful algorithm and one frequently used in object detection problems. Using the YOLO network, this study detected all heat staking points, regardless of their quality. From the detected heat staking points, the study then used the ResNet classification model to further classify whether the detected heat staking point was defected or not. This process, all connected into one algorithm, can be a powerful alternative for heat staking manufacturing firms that have problems in inspection processes.
The advent of AlexNet was a huge turning point in deep learning applications [3]. Since the introduction of the AlexNet framework, models have been applied to various fields including the manufacturing industry [4,5]; however, slow detection rate was a practical problem when applying deep learning to the manufacturing industry. The YOLO series are representative one-stage detectors, and fast detection speed, a critical index in real-time application, is their most prominent feature [6,7,8,9,10,11].
Among the YOLO series, this study used the most recent, YOLOv5, for detecting staking points from automobile door trims and the ResNet classification model for classifying the quality of detected staking points. This study ensembled two different methods to increase both the efficiency and accuracy of the given task. In this study, the function of YOLOv5 was to localize the abnormal regions surrounding staking points. The advantage of using ResNet is in avoiding gradient explosion problems in deep learning for classification. Using both models, this study dynamically filtered the result from the YOLOv5 and ResNet models. Because of several advantages, prior studies have well-used object detection models and classification models simultaneously in solving specific given tasks [1,12,13,14,15,16,17,18,19,20,21].
Following prior findings, this study used YOLOv5 with ResNet and obtained robust results. Using 2400 training door trim images (100,310 staking points labeled) and 600 test images, this study first found that the training result of the YOLOv5x model was significantly accurate; the mAP was 0.951, precision was 0.934, and recall was 0.939. The highly accurate result showed the good ability of the YOLO model to detect the heat staking points in the door trim image. Furthermore, the ResNet classification model also showed a noteworthy result; the accuracy of the model was 0.98, and the F1-score (the harmonic mean of the precision and recall) was 0.98. These results imply that the ResNet model classified the quality of the staking points detected by the YOLOv5 model effectively. The high F1-scores also showed that the results were relatively free from type 1 or type 2 errors.
This paper highlights a technical innovation in the deep learning field. By joining the YOLO and ResNet models, it provides a novel method to more accurately simultaneously detect inspection points and classify their quality both accurately and reliably.
This research makes the following contributions: first, it applies a deep learning framework to a real-time problem, particularly in the heat staking process. Inspecting the quality of a manufactured product and guaranteeing high quality for customers are critical for a business’s sustainable growth. Manual inspection has long encompassed problems. Immature work skills because of frequent labor changes and increased process complexity are typical examples that lead to human errors in the inspection process [4]. With the necessity to employ a deep learning-based vision system into the manufacturing process, this paper shows that the combination of the YOLO and ResNet frameworks can reduce costs and ultimately increase productivity. This paper also contributes to the literature on applying ensemble methods, particularly by combining objection detection and classification methodologies. It shows that the model’s performance is improved to the extent that it could be used on a real-time heat staking process.
The rest of this paper is organized as follows: In Section 2, it reviews relevant studies to this system and discusses key takeaway messages. It outlines the object detection and classification framework in Section 3. It provides the experimental results, consisting of object detection and quality classification on a real dataset, in Section 4. Finally, Section 5 discusses the results and provides potential recommendations for future research.

2. Related Work

2.1. YOLO Framework

The YOLO framework, a popular object detection model, was introduced in 2016. YOLO refers to the ability of the human visual system to immediately detect objects. Therefore, the YOLO framework was designed to detect objects similarly to the human visual system. The first YOLO consisted of a 24-layer convolutional neural network for feature extraction and two fully connected layers for predicting the probability and coordinates of objects. The latest version, YOLOv5, outperforms previous YOLO versions.
Because of its intuitiveness and innovativeness, the framework is renowned for its fast speed and accuracy. The YOLO framework introduces a new structure for object recognition systems and has received much attention; accordingly, it has been widely used and applied in various applications [22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43].

2.2. ResNet Framework

The ResNet framework is a residual neural network, which is a gateless or open-gated variant of the HighwayNet, a deep feedforward neural network with hundreds of layers [44]. The ResNet classification models are implemented with double layer skips that contain nonlinearities (ReLU) and batch normalization in between. Like the long short-term memory (LSTM) model, ResNet skips connections to avoid the gradient vanishing problem (leading to easier optimization of neural networks) or to mitigate the degradation problem [45]. Accuracy saturates when additional layers in a neural network increase training errors [44].
This skipping effectively simplifies and lightens the neural network, and the neural network learns by reducing the impact of vanishing gradients, as there are fewer layers to propagate through. The network then gradually restores the skipped layers. When all layers are expanded, it stays closer to the manifold. A neural network without residual parts explores more of the feature space. This is more vulnerable to perturbations, and it thus necessitates additional training to recover. Because of its advantages, ResNet is widely used and is one of the most cited neural network frameworks. Specifically, it is widely accepted in industries where takt time and accuracy are two important criteria. As ResNet is lighter than other deep learning-based classification models but a powerful classification learner, researchers and practitioners have long developed and employed it in their domain [10,11,14,15,46,47,48,49,50,51,52].

3. Proposed System

3.1. Shortcomings of Existing Models

There are small, multiple points in each door trim that require heat staking. In fact, in this sample dataset, although the number of points varied among automobile types, each image had approximately over sixty points. Multiple object detection and recognition tasks are important topics in manufacturing. Figure 1a shows the outline of the model for applying YOLO to heat staking process inspection. If this study employed only the YOLO model, then for each input door trim image it would search for points and show bounding boxes with a confidence level. Final detection results would be shown accordingly. However, YOLO frameworks have shortcomings in identifying positions of multiple objects accurately. Furthermore, the accuracy of classifying multiple small objects is low, which also reduces the recall rate, ultimately causing problems with adopting YOLO models on real heat staking process sites. Figure 1b illustrates the outline of the classification model ResNet. If this study only employed the ResNet model to the problem, then it would be able to classify whether the door trim image was of a defect or not, but the classification model has difficulties in illustrating where the defect parts are. Thus, for industrial usage where the workers must understand where the defects are located, classification models are hard to use.
To mitigate the issue to an extent, this study used the improved and modified YOLOv5 model by adding a deep ResNet framework with the same number of layers as the Darknet network in the feature extraction part of YOLOv5 [53]. Afterwards, the mean value was reduced to generate feature graphs of three scales after outputting the two feature extraction models. This process enabled more efficient information extraction of heat staking points in the door trim images and was more efficient in conducting object detection. Figure 2a shows an example of a full door trim image, and Figure 2b illustrates the real door trim images with heat staking points boxed around. Note that the heat staking points take a small portion of the entire door trim image.
Figure 3 illustrates the larger size of the heat staking points grouped by quality. For the classification, this study manually grouped the images into three different categories. The images were collected with a camera with a resolution of 5 M and an effective number of pixels (H × V) of 2592 × 1944: Figure 3a shows good quality points, Figure 3b shows defected points, and Figure 3c shows unknown points. Unknown points consist of points that are either partially covered by other cables, or out-of-focus and blurred images. Because the images are small-sized and the differences between defects and non-defects are not large, deepening the neural network will lead to the gradient vanishing problem. Therefore, this study chose ResNet, as it has the advantage of solving the gradient vanishing problem while deepening the network. This is typically important in the inspection of the heat staking process because the heat staking points are extremely small compared with the door trim images.

3.2. Suggested YOLO-ResNet Model

Figure 4 shows the flowchart of the proposed network that combines YOLOv5 and ResNet. Based on the Darknet network structure for feature extraction, ResNet was added for feature extraction, which solved the problem of poor accuracy in object detection.
The model is composed of six main parts: a deep fully convolutional network, region proposal network, ROI pooling and fully connected networks, bounding box regressor, classifier for object detection and ResNet classifier. For consistency, this paper employed a deep fully convolutional network. The input image was put into the initial stage and extracted to 256 × n × n feature map, which is the input of region proposal network and ROI pooling layer. In the region proposal network, there are k anchors with different scales and ratios for each point on feature map. There will be n × n × k candidate windows that are ranked according to the score, and then 2000 candidate windows are obtained through non-maximum suppression. Overall, complexity is O(N^2/2), which is consistent with the classic YOLO model. This is because the ResNet classifier was attached as a final layer to the current YOLO network.
Furthermore, in terms of the industrial usage, this proposed network is useful. That is, using solely the YOLO model results in difficulties in classifying points by their quality, and using only the image classification model results in difficulties in visualizing where the defect points are located. In this manner, taking only the advantages of both the object detection and the classification models enables to first detect the heat staking points, and then use the classifier to classify whether the detected points are defected or not.

4. Experimental Results

4.1. Experimental Dataset

This study used a novel and rich dataset provided by SEOYON E-HWA, from South Korea. The company’s main products are interior parts such as door trims, consoles, head linings, and package trays inside automobiles. They also manufacture exterior parts such as bumpers and seats for commercial vehicles. Thanks to their cooperation, this study was able to retrieve 3000 door trim images that had over sixty heat staking points. This study split the collected images such that there were 2400 training images and 600 test images.

4.2. Experimental Results

Table 1 first provides the hyperparameter tuning results for five different schemes (One to Five). Following prior YOLO literature, this study tuned the following hyperparameters that have been frequently tuned in past studies, to the five different schemes: lr0, lrf, momentum, weight_decay, warmup_epochs, warmup_momentum, warmup_bias_lr, box, cls, cls_pw, obj, obj_pw, iou_t, anchor_t, hsv_h, hsv_s, hsv_v, translate, scale, fliplr, and mosaic. The hyperparameter settings in the fourth scheme provided the best performance. Table 1 also shows the entire hyperparameter values and their descriptions. For easier crosschecking with other related literature, this study followed the default hyperparameter value settings. The trained model showed a precision of 0.995, a recall of 0.996, and a [email protected] of 0.994. The results imply that the parameters are well tuned.
This study trained the YOLO model on the training dataset that included 2400 door trim images, which consisted of 200,620 heat staking points. The accuracy of the trained model was then evaluated on the test set of 600 images. Table 2 presents the results that show that YOLOv5m outperformed YOLOv5n, YOLOv5s, YOLOv5l, and YOLOv5x, with a precision of 0.947, F1-score of 0.930, and [email protected] of 0.956. These results are significant. It should be noted that the performance of the YOLO v5m model outperformed the YOLO v5l and YOLO v5x model. Theoretically, YOLO v5l and YOLO v5x should outperform the YOLO v5m model as v5l and v5x are larger size in terms of the number of extracted features thus are able to train deeper. However, the results show that the optimum number of features to be extracted do not monotonically increase. This implies that there are a smaller number of features to be extracted from images of heat staking points, and that a certain number of features provided by YOLO v5m model is enough. Therefore, this study used the optimal model, YOLO v5m, as the baseline object detection model which was then used to merge with the classification model.
As the object detection results are significant, this study used the obtained heat staking points to further develop the ResNet classification model. Table 3 and Figure 5 report the summary statistics of the ResNet classification model. It should be noted that the F1-scores for non-defected, defected, and unknown classes were generally above 0.97. In the manufacturing industry, it is important to obtain a high F1-score, as recall and precision are both important factors [4]. This classification model’s high accuracy shows that ResNet well captured the different distribution of features between defected and non-defected heat staking points of automobile door trims. Furthermore, Table 4 reports performance metrics of the YOLO-ResNet model with the statistics for recall, false negative rate, precision, false negative rate, and F1-score of the model. It should be noted that the overall performance metrics is generally high and significant. Specifically, important metrics for manufacturing firms such as recall, and precision are approximately 98%. The overall results show that the generated YOLO-ResNet model is a significant model that may somewhat replace human labor inspection process.

5. Conclusions

Increased labor cost adds to the overall costs suffered by manufacturing firms in developed nations. One potential breakthrough to overcome the issue is to employ Artificial Intelligence models in their manufacturing process. Many studies have investigated the usage of object detection and image classification in deep learning models. However, there are few studies on the application of these models in the inspection of the heat staking process. In this respect, this study proposed a new YOLO–ResNet model for detection and classification of heat staking points in the automobile manufacturing industry. In this model, the YOLOv5 framework first detects the heat staking point accurately and then ResNet classifies whether the detected points are defected or not. The ability to detect the points and classify their quality is sufficiently accurate to be applied to real manufacturing sites.
This study used YOLO and ResNet models to detect the heat staking points and classify their quality. In the future, it is recommended to apply other object detection models or classification models and improve the current models. Furthermore, future researchers may also consider adding more defect categories for the heat staking process. The current model only considered whether the detected images are defected, non-defected, or unknown. However, there are multiple defect categories in the heat staking process such as overstaking and understaking. More accurate categorization may be beneficial for industries.
This study makes a contribution to the manufacturing industry that suffers from high inspection costs. The F1-score of the model is well-above 97%, which implies that if the model is used in the manufacturing inspection process, it would perform like or better than the human labor inspection system. Furthermore, if the model can train multiple defect types of heat staking points, then classifying detected heat staking points into different categories by their defect type would also be possible. Such multi-class object detection and classification hybrid model would be an important asset for the sustainable development and growth of manufacturing firms.
Furthermore, this study makes a contribution to the growing deep learning application literature. There are recent studies that employ image classification, multi-strategy particle swarm and ant colony hybrid optimization and optimal search mapping on various fields [54,55,56,57]. The underlying mechanism is to use modern algorithms and understand how to employ them in a certain domain field. In this manner, this study contributes by showing how YOLO and ResNet model could be used in the manufacturing inspection process.

Author Contributions

Conceptualization, H.J.; methodology, H.J.; software, J.R.; validation, J.R.; formal analysis, J.R.; investigation, J.R.; resources, H.J.; data curation, H.J.; writing—original draft preparation, H.J.; writing—review and editing, H.J.; visualization, H.J.; supervision, H.J.; project administration, H.J.; funding acquisition, H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Korea Technology and Information Promotion Agency for SMEs (TIPA) funded by the Korean Government [Ministry of SMEs and Startups (MSS)] under Grant RS-2022-00140573.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The heat staking points dataset is available upon request. Please contact the corresponding author for data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Luo, Y.; Zhang, Y.; Sun, X.; Dai, H.; Chen, X. Intelligent solutions in chest abnormality detection based on YOLOv5 and ResNet50. J. Healthc. Eng. 2021, 2021, 2267635. [Google Scholar] [CrossRef] [PubMed]
  2. Fu, S.; Kauppila, O.; Mottonen, M. Measurement system escape and overkill rate analysis. Int. J. Adv. Manuf. Technol. 2011, 57, 1079–1086. [Google Scholar] [CrossRef]
  3. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  4. Jung, H.; Jeon, J.; Choi, D.; Park, J.Y. Application of machine learning techniques in injection molding quality prediction: Implications on sustainable manufacturing industry. Sustainability 2021, 13, 4120. [Google Scholar] [CrossRef]
  5. Jamwal, A.; Agrawal, R.; Sharma, M.; Giallanza, A. Industry 4.0 technologies for manufacturing sustainability: A systematic review and future research directions. Appl. Sci. 2021, 11, 5725. [Google Scholar] [CrossRef]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  7. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  8. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  9. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  10. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. Scaled-yolov4: Scaling cross stage partial network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  11. Wang, G.; Yu, H.; Sui, Y. Research on maize disease recognition method based on improved ResNet50. Mob. Inf. Syst. 2021, 2021, 9110866. [Google Scholar] [CrossRef]
  12. Chaschatzis, C.; Karaiskou, C.; Mouratidis, E.G.; Karagiannis, E.; Sarigiannidis, P.G. Detection and characterization of stressed sweet cherry tissues using machine learning. Drones 2021, 6, 3. [Google Scholar] [CrossRef]
  13. Kwan, C.; Gribben, D.; Tran, T. Tracking and classification of multiple human objects directly in compressive measurement domain for low quality optical videos. In Proceedings of the IEEE Publications 10th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), New York, NY, USA, 10–12 October 2019; IEEE Publications: Piscataway, NJ, USA, 2019; Volume 2019, pp. 488–494. [Google Scholar]
  14. Yuan, X.; Xia, J.; Wu, J.; Shi, J.; Deng, L. Low Altitude Small UAV Detection Based on YOLO model. In Proceedings of the 2020 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; IEEE Publications: Piscataway, NJ, USA, 2020; Volume 2020, pp. 7362–7366. [Google Scholar]
  15. Yuan, J.; Fan, Y.; Lv, X.; Chen, C.; Li, D.; Hong, Y.; Wang, Y. Research on the practical classification and privacy protection of CT images of parotid tumors based on. In Proceedings of the Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2020; Volume 1576, p. 012040. [Google Scholar]
  16. Hendryli, J.; Herwindiati, D.E. Automatic license plate recognition for parking system using convolutional neural networks. In Proceedings of the International Conference on Information Management and Technology (ICIMTech), Indonesia, Southeast Asia, 13–14 August 2020; IEEE Publications: Piscataway, NJ, USA, 2020; Volume 2020, pp. 71–74. [Google Scholar]
  17. Chakravarthy, A.S.; Raman, S. Early blight identification in tomato leaves using deep learning. In Proceedings of the International Conference on Contemporary Computing and Applications (IC3A), Lucknow, India, 5–7 February 2020; IEEE Publications: Piscataway, NJ, USA, 2020; Volume 2020, pp. 154–158. [Google Scholar]
  18. Shah, S.; Deshmukh, C. Pothole and bump detection using convolution neural networks. In Proceedings of the IEEE Transportation Electrification Conference (ITEC-India), Bengaluru, India, 17–19 December 2019; IEEE Publications: Piscataway, NJ, USA, 2019; Volume 2019, pp. 1–4. [Google Scholar]
  19. Abdullah, S.; Hasan, M.M.; Islam, S.M.S. YOLO-based three-stage network for Bangla license plate recognition in Dhaka metropolitan city. In Proceedings of the International Conference on Bangla Speech and Language Processing (ICBSLP), Sylhet, Bangladesh, 21–22 September 2018; IEEE Publications: Piscataway, NJ, USA, 2018; Volume 2018, pp. 1–6. [Google Scholar]
  20. Hişam, D.; Hişam, E. Deep learning models for classifying cancer and COVID-19 lung diseases. In Proceedings of the Innovations in Intelligent Systems and Applications Conference (ASYU), Elazig, Turkey, 6–8 October 2021; IEEE Publications: Piscataway, NJ, USA, 2021; Volume 2021, pp. 1–4. [Google Scholar]
  21. Park, G.; Park, N.; Kim, S.; Kim, S.; Kim, J.; Ko, B. Real-time mask facial expression recognition using Tiny-YOLOv3 and ResNet50. In Proceedings of the Korean Society of Broadcast Engineers Conference; The Korean Institute of Broadcast and Media Engineers: Seoul, Republic of Korea, 2021; pp. 232–234. [Google Scholar]
  22. Li, Y.; Han, Z.; Xu, H.; Liu, L.; Li, X.; Zhang, K. YOLOv3-lite: A lightweight crack detection network for aircraft structure based on depthwise separable convolutions. Appl. Sci. 2019, 9, 3781. [Google Scholar] [CrossRef] [Green Version]
  23. Tian, Y.; Yang, G.; Wang, Z.; Wang, H.; Li, E.; Liang, Z. Apple detection during different growth stages in orchards using the improved YOLO-V3 model. Comput. Electron. Agric. 2019, 157, 417–426. [Google Scholar] [CrossRef]
  24. Liu, W.; Wang, Z.; Zhou, B.; Yang, S.; Gong, Z. Real-time Signal Light Detection based on Yolov5 for Railway. In Proceedings of the IOP Conference Series. Earth and Environmental Science, Jakarta, Indonesia, 25–26 September 2021; Volume 769, p. 042069. [Google Scholar]
  25. Liu, X.; Jiang, X.; Hu, H.; Ding, R.; Li, H.; Da, C. Traffic sign recognition algorithm based on improved YOLOv5s. In Proceedings of the International Conference on Control, Automation and Information Sciences (ICCAIS), Xi’an, China, 14–17 October 2021; Volume 2021, pp. 980–985. [Google Scholar]
  26. Kim, J.A.; Sung, J.Y.; Park, S.H. Comparison of Faster-RCNN, YOLO, and SSD for real-time vehicle type recognition. In Proceedings of the IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Yeosu, Republic of Korea, 26–28 October 2020; Volume 2020, pp. 1–4. [Google Scholar]
  27. Jia, W.; Xu, S.; Liang, Z.; Zhao, Y.; Min, H.; Li, S.; Yu, Y. Real-time automatic helmet detection of motorcyclists in urban traffic using improved YOLOv5 detector. IET Image Process. 2021, 15, 3623–3637. [Google Scholar] [CrossRef]
  28. Dharneeshkar, J.; Aniruthan, S.A.; Karthika, R.; Parameswaran, L. Deep Learning based Detection of potholes in Indian roads using YOLO. In Proceedings of the International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020; Volume 2020, pp. 381–385. [Google Scholar]
  29. Junos, M.H.; Mohd Khairuddin, A.S.; Thannirmalai, S.; Dahari, M. Automatic detection of oil palm fruits from UAV images using an improved YOLO model. Vis. Comput. 2022, 38, 2341–2355. [Google Scholar] [CrossRef]
  30. Kuznetsova, A.; Maleva, T.; Soloviev, V. Detecting apples in orchards using YOLOv3 and YOLOv5 in general and close-up images. In Lecture Notes in Computer Science International Symposium on Neural Networks; Springer: Cham, Swizerland, 2020; pp. 233–243. [Google Scholar] [CrossRef]
  31. Lyu, S.; Li, R.; Zhao, Y.; Li, Z.; Fan, R.; Liu, S. Green citrus detection and counting in orchards based on YOLOv5-CS and AI edge system. Sensors 2022, 22, 576. [Google Scholar] [CrossRef]
  32. Mukti, I.Z.; Biswas, D. Transfer learning based plant diseases detection using. In Proceedings of the 2019 4th International Conference on Electrical Information and Communication Technology (EICT), Khulna, Bangladesh, 20–22 December 2019; pp. 1–6. [Google Scholar]
  33. Zhang, L.; Yin, L.; Liu, L.; Zhuo, R.; Zhuo, Y. Forestry pests identification and classification based on improved YOLO v5s. In Proceedings of the International Conference on Electronic Information Engineering and Computer Science (EIECS), Changchun, China, 23–26 September 2021; Volume 2021, pp. 670–673. [Google Scholar]
  34. Zhang, D.; Huang, Z.; Wang, H.; Wu, Y.; Wang, Y.; Zou, J. Research and application of wild edible mushroom detection based on multi-scale features. In Proceedings of the 2nd International Conference on Information Science and Education (ICISE-IE), Nanchang, China, 26–28 November 2021; Volume 2021, pp. 455–459. [Google Scholar]
  35. Dima, T.F.; Ahmed, M.E. Using YOLOv5 algorithm to detect and recognize American sign language. In Proceedings of the International Conference on Information Technology (ICIT), Amman, Jordan, 14–15 July 2021; Volume 2021, pp. 603–607. [Google Scholar]
  36. Yang, Y.C.; Chen, W. An improved YOLO leucocyte classification and recognition method. In Proceedings of the International Conference on Intelligent Transportation, Big Data & Smart City (ICITBS), Xi’an, China, 27–28 March 2021; Volume 2021, pp. 618–621. [Google Scholar]
  37. Wu, Z.; Zhang, D.; Shao, Y.; Zhang, X.; Zhang, X.; Feng, Y.; Cui, P. Using YOLOv5 for garbage classification. In Proceedings of the 4th International Conference on Pattern Recognition and Artificial Intelligence (PRAI), Yibin, China, 20–22 August 2021; Volume 2021, pp. 35–38. [Google Scholar]
  38. Cengil, E.; Çinar, A.; Yildirim, M. A case study: Cat-dog face detector based on YOLOv5. In Proceedings of the International Conference on Innovation and Intelligence for Informatics, Computing, and Technologies (3ICT), Virtual Conference, Zallaq, Bahrain, 29–30 September 2021; Volume 2021, pp. 149–153. [Google Scholar]
  39. Zhou, Y.; Wu, M.; Bai, Y.; Guo, C. Flame detection with pruned and knowledge distilled YOLOv5. In Proceedings of the 5th Asian Conference on Artificial Intelligence Technology (ACAIT), Haikou, China, 29–31 October 2021; Volume 2021, pp. 780–785. [Google Scholar]
  40. Bushra, S.N.; Shobana, G.; Maheswari, K.U.; Subramanian, N. Smart video Survillance based weapon identification using Yolov5. In Proceedings of the International Conference on Electronic Systems and Intelligent Computing (ICESIC), Chennai, India, 22–23 April 2022; Volume 2022, pp. 351–357. [Google Scholar]
  41. Wang, X.; Niu, D.; Luo, P.; Zhu, C.; Ding, L.; Huang, K. A Safety Helmet and Protective Clothing Detection Method based on Improved-Yolo. In Proceedings of the 2020 Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020; Volume 3, pp. 5437–5441. [Google Scholar]
  42. Patel, D.; Patel, F.; Patel, S.; Patel, N.; Shah, D.; Patel, V. Garbage detection using advanced object detection techniques. In Proceedings of the International Conference on Artificial Intelligence and Smart Systems (ICAIS), Coimbatore, India, 25–27 March 2021; Volume 2021, pp. 526–531. [Google Scholar]
  43. Ji, S.Y.; Jun, H.J. Deep learning model for form recognition and structural member classification of east Asian traditional buildings. Sustainability 2020, 12, 5292. [Google Scholar] [CrossRef]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  45. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  46. Deshpande, A.; Estrela, V.V.; Patavardhan, P. The DCT-CNN-ResNet50 architecture to classify brain tumors with super-resolution, convolutional neural network, and the ResNet50. Neurosci. Inform. 2021, 1, 100013. [Google Scholar] [CrossRef]
  47. Chen, M.; Chen, W.; Chen, W.; Cai, L.; Chai, G. Skin cancer classification with deep convolutional neural networks. J. Med. Imaging Health Inform. 2020, 10, 1707–1713. [Google Scholar] [CrossRef]
  48. Luo, W.; Liu, J.; Huang, Y.; Zhao, N. An effective vitiligo intelligent classification system. J. Ambient Intell. Human. Comput. 2020, 1–10. [Google Scholar] [CrossRef]
  49. Jung, H.; Choi, M.K.; Jung, J.; Lee, J.H.; Kwon, S.; Young Jung, W. ResNet-based vehicle classification and localization in traffic surveillance systems. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  50. Zheng, Z.; Zhang, H.; Li, X.; Liu, S.; Teng, Y. Resnet-based model for cancer detection. In Proceedings of the IEEE International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 15–17 January 2021; Volume 2021, pp. 325–328. [Google Scholar]
  51. Kutlu, H.; Avci, E.; Özyurt, F. White blood cells detection and classification based on regional convolutional neural networks. Med. Hypotheses 2020, 135, 109472. [Google Scholar] [CrossRef]
  52. Polat, Ö. Detection of pediatric pneumonia from X-ray images using. In Proceedings of the 2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; pp. 1–6. [Google Scholar]
  53. Lu, Z.; Lu, J.; Ge, Q.; Zhan, T. Multi-object detection method based on YOLO and ResNet hybrid networks. In Proceedings of the Conference on Advanced Robotics and Mechatronics (ICARM), Toyonaka, Japan, 3–5 July 2019; Volume 2019, pp. 827–832. [Google Scholar]
  54. Chen, H.; Miao, F.; Chen, Y.; Xiong, Y.; Chen, T. A hyperspectral image classification method using multifeature vectors and optimized KELM. Proc. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2781–2795. [Google Scholar] [CrossRef]
  55. Deng, W.; Zhang, L.; Zhou, X.; Zhou, Y.; Sun, Y.; Zhu, W.; Chen, H.; Deng, W.; Chen, H.; Zhao, H. Multi-strategy particle swarm and ant colony hybrid optimization for airport taxiway planning problem. In Information Sciences; Elsevier: Amsterdam, The Netherlands, 2022; Volume 612, pp. 576–593. [Google Scholar]
  56. Ren, Z.; Han, X.; Yu, X.; Skjetne, R.; Leira, B.J.; Sævik, S.; Zhu, M. Data-driven simultaneous identification of the 6DOF dynamic model and wave load for a ship in waves. In Mechanical Systems and Signal Processing; Elsevier: Amsterdam, The Netherlands, 2022; Volume 184, p. 109422. [Google Scholar]
  57. Yu, Y.; Hao, Z.; Li, G.; Liu, Y.; Yang, R.; Liu, H. Optimal search mapping among sensors in heterogeneous smart homes. In Mathematical Biosciences and Engineering; AIMS Press: Springfield, IL, USA, 2023; Volume 20, pp. 1960–1980. [Google Scholar]
Figure 1. (a) YOLO flowchart for detecting quality of heat staking points in door trim images and (b) ResNet flowchart for classifying the quality of heat staking points.
Figure 1. (a) YOLO flowchart for detecting quality of heat staking points in door trim images and (b) ResNet flowchart for classifying the quality of heat staking points.
Sustainability 14 15892 g001
Figure 2. Example image of a (a) full door trim and (b) detected heat staking points.
Figure 2. Example image of a (a) full door trim and (b) detected heat staking points.
Sustainability 14 15892 g002
Figure 3. Example of a (a) non-defected heat staking points, (b) defected heat staking points, and (c) unknown points.
Figure 3. Example of a (a) non-defected heat staking points, (b) defected heat staking points, and (c) unknown points.
Sustainability 14 15892 g003
Figure 4. Flowchart of the proposed model, YOLOv5 with ResNet.
Figure 4. Flowchart of the proposed model, YOLOv5 with ResNet.
Sustainability 14 15892 g004
Figure 5. Confusion matrix of the ResNet classification model.
Figure 5. Confusion matrix of the ResNet classification model.
Sustainability 14 15892 g005
Table 1. Hyperparameter tuning results of YOLO model.
Table 1. Hyperparameter tuning results of YOLO model.
SchemeOptimizerClassImagesLabelsPrecisionRecall[email protected]mAP@
OneAdamWAll60025,9990.9830.9890.9890.652
Points16,1630.9890.9990.990.699
Screw95230.980.9960.9920.679
Hide3130.9810.9710.9860.578
TwoAdamWAll60025,9990.9460.9630.9610.596
Points16,1630.980.9980.9880.658
Screw95230.9720.9930.9910.646
Hide3130.8860.8980.9030.485
ThreeAdamWAll60025,9990.9860.9690.9770.594
Points16,1630.990.9970.9880.645
Screw95230.9850.9870.990.624
Hide3130.9830.9230.9530.514
FourAdamWAll60025,9990.9950.9960.9940.664
Points16,1630.9980.9990.9950.693
Screw95230.9950.9960.9950.647
Hide3130.9930.9940.9940.651
FiveAdamWAll60025,9990.9920.9850.990.744
Points16,1630.99710.9950.785
Screw95230.9960.9940.9950.782
Hide3130.9840.9620.980.665
HyperparametersDescriptionValue
lr0Initial learning rate0.01
lrfFinal OneCycleLR learning rate (lr0 * lrf)0.01
momentumTuning parameter for the gradient descent algorithm0.937
weight_decayOptimizer weight decay0.0005
warmup_epochsWarmup epochs3.0
warmup_momentumWarmup initial momentum0.8
warmup_bias_lrWarmup initial bias learning rate0.1
boxBox loss gain0.05
clsClass loss gain0.5
cls_pwClass BCELoss positive weight1.0
objObject loss gain (scale with pixels)1.0
obj_pwObject BCELoss positive weight1.0
iou_tIoU training threshold0.20
anchor_tAnchor-multiple threshold4.0
hsv_hImage HSV-Hue augmentation (fraction)0.015
hsv_sImage HSV-Saturation augmentation (fraction)0.7
hsv_vImage HSV-Value augmentation (fraction)0.4
translateImage translation (+/− fraction)0.1
scaleImage scale (+/− gain)0.5
fliplrImage flip left-right (probability)0.5
mosaicImage mosaic (probability)1.0
Table 2. YOLO model heat staking points detection results.
Table 2. YOLO model heat staking points detection results.
ModelClassOpt.ImagesLabelsPrecisionRecall[email protected]mAP@
Yolov5nAllAdamW600100,3100.9370.9140.9450.592
Points54,0600.9780.990.9920.622
Screw43,5000.9720.9890.9860.692
Hide27500.8630.7620.8570.461
Yolov5sAllAdamW600100,3100.9240.9440.9450.592
Points54,0600.9780.990.9920.622
Screw43,5000.9690.9930.9870.692
Hide27500.8260.8480.8550.461
Yolov5mAllAdamW600100,3100.9470.930.9560.591
Points54,0600.9870.9870.9920.627
Screw43,5000.980.9880.9870.689
Hide27500.8740.8160.890.456
Yolov5lAllAdamW600100,3100.9340.9290.9520.594
Points54,0600.9820.990.9920.614
Screw43,5000.9790.9890.9880.689
Hide27500.8410.8080.8770.479
Yolov5xAllAdamW600100,3100.9340.9390.9510.595
Points54,0600.9840.9880.990.619
Screw43,5000.9820.990.9880.696
Hide27500.8360.8390.8760.469
Table 3. ResNet model heat staking points quality classification results.
Table 3. ResNet model heat staking points quality classification results.
PrecisionRecallF1-ScoreSupport
Non-defect0.970.990.98306
Defect0.960.930.95105
Unknown0.990.970.98160
Accuracy 0.98571
Table 4. YOLO-ResNet model performance metrics.
Table 4. YOLO-ResNet model performance metrics.
R e c a l l   ( R ) = T P T P + F N = 0.98
F a l s e   N e g a t i v e   R a t e   ( F N R ) = 1.00 R = 0.02
P r e c i s i o n   ( P ) = T P T P + F P = 0.98
F a l s e   N e g a t i v e   R a t e   ( F P R ) = 1.00 P = 0.02
F 1   s c o r e = 2   ·   P   ·   R P + R = 0.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jung, H.; Rhee, J. Application of YOLO and ResNet in Heat Staking Process Inspection. Sustainability 2022, 14, 15892. https://doi.org/10.3390/su142315892

AMA Style

Jung H, Rhee J. Application of YOLO and ResNet in Heat Staking Process Inspection. Sustainability. 2022; 14(23):15892. https://doi.org/10.3390/su142315892

Chicago/Turabian Style

Jung, Hail, and Jeongjin Rhee. 2022. "Application of YOLO and ResNet in Heat Staking Process Inspection" Sustainability 14, no. 23: 15892. https://doi.org/10.3390/su142315892

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop