You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

20 June 2024

Steel Surface Defect Detection Algorithm Based on Improved YOLOv8n

,
,
and
School of Mechanical Engineering, Shenyang Jianzhu University, Shenyang 110168, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Image Recognition and Processing Technologies

Abstract

The traditional detection methods of steel surface defects have some problems, such as a lack of feature extraction ability, sluggish detection speed, and subpar detection performance. In this paper, a YOLOv8-based DDI-YOLO model is suggested for effective steel surface defect detection. First, on the Backbone network, the extended residual module (DWR) is fused with the C2f module to obtain C2f_DWR, and the two-step approach is used to carry out the effective extraction of multiscale contextual information, and then fusing feature maps formed from the multiscale receptive fields to enhance the capacity for feature extraction. Also based on the above, an extended heavy parameter module (DRB) is added to the structure of C2f_DWR to make up for the lack of C2f’s ability to capture small-scale pattern defects between training to enhance the training fluency of the model. Finally, the Inner-IoU loss function is employed to enhance the regression accuracy and training speed of the model. The experimental results show that the detection of DDI-YOLO on the NEU-DET dataset improves the mAP by 2.4%, the accuracy by 3.3%, and the FPS by 59 frames/s compared with the original YOLOv8n.Therefore, this paper’s proposed model has a superior mAP, accuracy, and FPS in identifying surface defects in steel.

1. Introduction

Nowadays, our country vigorously develops new industries, vigorously promotes industry, and its information business is a very important asset. In steel production processes, the detection of surface defects is a very critical part, as it is related to the quality of the entire industrial products and production safety aspects. Steel is extensively employed in many fields, including, in the past few years in our country, the rapidly developing aerospace and aviation industry, the oil industry, the automobile industry, the shipbuilding industry, the large-scale equipment industry, and other important industrial areas. Steel surface defects in these important industry areas are very important and have an immediate impact on the caliber of the industrial goods we produce and manufacture out of the finished product quality. If the quality of the finished product is poor, the economy will suffer. Therefore, it is urgent to improve the accuracy of steel surface defect detection.
Current surface defect detection methods fall into one of three categories: deep-learning-based, traditional-technology-based, and machine vision methods. These include manual inspections using magnetic particles, eddy currents, vision, and visible light inspection. Artificial visual inspection consists of checking the appearance or performance defects of the product by visual or tactile means, and the shortcomings of artificial visual inspection are also very obvious, e.g., too subjective, low efficiency, and the influence of external factors. The eddy-current-based inspection method is a non-destructive surface defect detection method. If there are some common defects on the steel surface, these defects will change the distribution of eddy currents and the characteristics of the magnetic field, so that the defects can be identified and localized by detecting changes in the magnetic field. However, when it comes to the detection of more complex defects, eddy current detection has relatively low sensitivity and also needs experienced technicians to interpret and analyze the data; therefore, from another economic point of view, the labor cost is relatively high. Magnetic particle inspection is widely used in welding, manufacturing, foundry, aerospace, and other fields relying on steel production. It can be used to detect cracks, inclusions, porosity, fatigue cracks, and other defects on the steel surface. It has the advantage of being fast, economical, easy to operate, and able to identify items of various sizes and shapes. The disadvantage of magnetic particle inspection is that it is only applicable to materials with magnetic properties, and it does not work for the defect detection of non-magnetic materials. It may be difficult to implement with respect to parts characterized by complex surface shapes.
Machine vision surface defect detection is a method that combines algorithms and vision techniques. The input image is processed by a computer to achieve fast and accurate recognition and localization of defects. The visual light inspection method is a commonly used method for detecting surface defects on steel, such as dents, blemishes, and scratches. It usually involves the use of a light source and optical equipment to observe and record an image of the surface of the material being inspected. The advantage is that it is relatively fast, especially for single-object detection. The disadvantage is that it mainly depends on the subjective judgment and experience of the observer when detecting the surface defects of steel which will be greatly affected by human factors, and the detection accuracy is low for tiny objective or hidden defects. Machine vision defect detection is frequently used to identify and detect defects on the surface of objects. It processes and analyzes the image by computer to achieve fast and accurate recognition and localization of defects.
Compared with the traditional manual visualization technology, defect detection by machine vision includes the advantages of quick speed and good versatility. It can lessen the impact of human factors on the detection results and, at the same time, it can realize automated inspections on a large scale and in relation to continuous production lines, so it is often used for inspection of actual production. Advantages of the automated inspection process include the improvement of production efficiency, the ability to handle a large amount of image data, and the ability to distinguish production lines. However, this method also has a lot of defects, namely the classification of defect types is relatively poor, it requires manual extraction of the characteristics of various defects, and the precision of defect detection is relatively poor. Therefore, deep-learning-based industrial defect detection technologies have a broad prospect in our intelligent manufacturing sector, which needs our continuous exploration and innovation.
The advantage of deep learning is that, by simply inputting specific data, the computer can automatically recognize various data and extract features, avoiding the disadvantage of manually extracting features. Currently, deep-learning-driven target recognition techniques are advancing rapidly. These target detection techniques for deep learning [1] methods are primarily categorized into the following two types. The main features of the one-stage methods are a good performance, a faster speed, and ease of practical application. The main one-stage target detection methods are YOLO [2] and SSD [3]. The two-stage detection methods need to generate the required region in advance and then the network classifies the corresponding region and obtains the prediction information. Typical representatives of two-stage algorithms include U-net [4], R-CNN [5], fast R-CNN [6], RetinaNet [7], faster R-CNN [8], and mask R-CNN [9]. These algorithms first generate the corresponding alternative regions in stage one, followed by classification and bounding box adjustment of these regions in stage two to arrive at the final recognition result. In addition, there are transformer [10,11,12,13] and unsupervised learning [14,15] approaches that have been brought to the computer vision field to solve the related problems in target detection.
There have been a lot of scholars who have studied the application of algorithms deriving from deep learning target detection of steel surface defects. Ren [16] proposed that the information based on the model should incorporate the ECA-SimSPPFSIoU-Yolov5 effective channel attention to enhance the significant weights. Douzhi et al. [17] proposed a small sample steel plate defect detection algorithm based on lightweight YOLOv8 to bypass the difficult-to-apply deep learning approach when it comes to the detection of small sample defects. The LMSRNet network was designed to replace the YOLOv8 backbone, and the CBFPN and ECSA modules were developed to address the issue of the light weight of the model. Guo et al. [18] proposed an algorithm based on the improved MSFT-YOLO model, integrating the TRANS module into the network and data enhancement processing to increase the accuracy of steel defect detection. Cui Kebin et al. [19] proposed MCB-FAH-YOLOv8 to address the problems of defect detection, such as false detection and high leakage rate, by adding the improved convolutional attention mechanism module (CBAM) and the replaceable four-head ASFF predictor, so that the network could enhance the ability to detect tiny objects and dense targets and improve the accuracy at the expense of speed. The above methods are all based on YOLO for steel surface defects.
Zhou Yan et al. [20] proposed a method for detecting steel using multiscale lightweight attention, using a channel attention module for multilevel feature maps to reconstruct their channel-related information and improve the detection effect. Zhu et al. [21] used an efficient swin transformer to detect and classify steel surface defects, an approach which strengthened the connection between the feature mapping channels, reduced the resolution, and solved the image information retention problem. He [22] suggested that an MFN network can effectively improve the ability to obtain information about steel surface defects, and its features are able to integrate information from various levels to detect steel surface defects. A baseline convolutional neural network (CNN) is employed to produce feature maps for subsequent stages. The proposed network (MFN) is capable of fusing multiple features into a single feature and, based on these multilevel features, the region proposal network (RPN) is employed to produce the region of interest (ROI); finally, the baseline network ResNet 34 is used to achieve 74.8% mAP. The above methods have been designed and improved based on transformer and convolutional neural networks, respectively. Although all of the above methods improve the performance of the target detection algorithm to some degree, they still fall short in terms of accuracy and some other aspects. To summarize, we propose the steel surface defect detection algorithm DDI-YOLO, and the contributions made in this essay are the following:
(1) The resolution of the issue of the original C2f module of YOLOv8 having insufficient ability to extract defective features on the steel surface and being unable to extract multiscale contextual information. Therefore, the dilation-wise residual module is added to the original C2f to further enhance the capacity of the network to extract multiscale contextual information.
(2) Since C2f in YOLOv8 cannot detect small-scale patterns during training, the dilated reparam block module is used to enable the detection of defects by C2f on small-scale patterns, thus enhancing the training ability of the model.
(3) The C2f_DWR module and the C2f_DRB module are combined to form the C2f_DWR_DRB module, which blends the benefits of every module to improve the model’s comprehensiveness.
(4) A faster and more accurate convergence by making use of the Inner-IoU loss function to take the place of YOLOv8’s original CIoU loss function.

3. Results

3.1. Image Dataset

The NEU-DET [32] dataset was utilized in this study, consisting of a steel surface dataset from Northeastern University on which training and validation were performed. This dataset contains 1800 images in total, consisting of six categories of classified defects, and each type of defect is exemplified by 300 images which contain rolled-in scale (RS), patches (Pa), crazing (Cr), scratches (Sc), pitting surfaces (Ps), and inclusions (In), as shown in Figure 5. The dataset images contain 200 × 200 pixels and the total of 1800 images were randomly divided into an 8:1:1 ratio to create NEU-DET training, test, and validation sets, i.e., 1440 training samples, 180 test samples, and 180 validation samples.
Figure 5. Schematic diagram of six defect samples.

3.2. Experimental Environment

The operating system used for the experiments in this paper was Windows 11, the CPU was a 13th Gen Intel(R) Core(TM) i5-13500HX, the GPU used was the NVIDIA GeForce RTX 4060 GPU, and the RAM was 16 GB. The deep learning training architecture used was Pytorch 1.13.1. The specific parameters of the experiment were as follows: the learning rate was 0.01, the image size was 640 × 640, the number of iteration rounds (Epochs) was 200, and the optimizer chosen was SGD.

3.3. Experimental Metrics

The data evaluation metrics used in this study were precision and recall curves P–R (precision–recall curve), the number of parameters x (parameters), the average precision (AP) of each defect category, the average accuracy (mAP), and the frames per second (FPS), with the AP determined by the precision (Precision, P) and the recall (Recall, R). The P–R curve is a curve formed by the coordinate system of test precision and recall, and the area enclosed by the curve is mAP. The related calculation formula is:
P r e c i s i o n = T P ( T P + F P )
where T P is the number of predicted positive samples that are actually positive, i.e., positive samples correctly identified. F P is the number of predicted positive samples that are actually negative, i.e., negative samples misreported.
R e c a l l = T P ( T P + F N )
where F N is the number of predicted negative samples that are actually positive, i.e., positive samples missed.
A P = 0 1 p r d r
where A P is the accuracy of each detection type.
m A P = 1 n i = 1 n A P i
where m A P denotes the mean of the values in all categories and n is the number of categories.

3.4. Analysis of Experimental Results

3.4.1. Ablation Experiments

To demonstrate the enhancement effect of each of our improvements to YOLOv8, we conducted five sets of ablation experiments separately. First, the DWR module was added to the sixth and eighth layers of the C2f structure in the backbone network, and then the DRB module was added outside the second, fourth, sixth, and eighth layers of the C2f structure. The DRB module was also added to the neck and head layers. In addition to that, a combination of DRB and DWR modules were added to the sixth and eighth layers in the C2f structure in the backbone network, and, lastly, the Inner-IoU loss function took the place of the original loss function. Each part of the modules corresponding to the above was added to the YOLOv8 model accordingly, and we report the experimental results of the various combinations of the improved modules. Table 1 shows the outcomes of our ablation studies.
Table 1. Ablation experiment results.
As it can be seen from Table 1, after replacing C2f in the backbone network with C2f_DWR, except for a slight decrease in recall and FPS, the rest of the metrics, such as mAP and precision, improved, with the mAP item experiencing a notable upgrade of 2.5 percentage points which was attributed to the excellent ability of DWR to extract the multiscale contextual information and then fuse the wildly formed by the multiscale fields into the feature maps, with a decrease in the number of parameters and computational effort. After substituting the C2f structure in the main network with the C2f_DRB structure, the DRB reparameterization module used a non-inflated small kernel and multiple inflated small kernel layers to augment a non-inflated large kernel convolutional layer, an approach which greatly reduced the number of parameters and significantly increased mAP and FPS, as well as FPS to 104. This was because DRB compensated for the inability of Cf2 to capture small-scale patterns between training sessions, thus improving the model’s training and detection capabilities. Then, after substituting the loss function of YOLOv8 with the Inner-IoU loss function alone, while not affecting the size of the model parameters, there was a notable increase in accuracy, mAP, and FPS, with a 5.4% increase in accuracy, a 1.5% increase in mAP, and a 9% increase in FPS. Therefore, our improvement not only optimized the speed of the model, but also enhanced the detection of defects by the model, the average accuracy rate, and the speed of defect detection, indicating that our improvement is effective. The fourth set of experiments entailed the DWR module and the DRB module being combined to form a new C2f_DWR_DRB structure and, from Table 1, we can see that, except for the recall rate having declined, all other indexes experienced a great improvement, with the precision rate, mAP, and FPS improving by 2.1%, 4.6%, and 27%, respectively. The most obvious improvement was that experienced by the mAP and FPS indexes, as well as the quantity of calculations and the number of parameters. There was a decrease in the number of calculations and the number of parameters, showing that the C2f_DWR_DRB structure combined the advantages of the two respective modules. Finally, our DDI-YOLO experienced a decrease in recall, but all other metrics remained optimal. In summary, the improvement method of the model put forth in this research works well.

3.4.2. Comparative Experiments

To confirm that our suggested DDI-YOLO modeling algorithm is effective, in this paper, we used the original YOLOv8n as the baseline and tested YOLOv8n and DDI-YOLO using the NEU-DET dataset. Table 2 provides a summary of the findings. The indicator “↑1.8” indicates that for the defect type In, the mAP of DDI-YOLO was 1.8% greater than the value associated with the baseline model YOLOv8n, as well as the other models. As shown in Table 2, the accuracy was improved for the remaining four defect types, except for Cr defects and Ps defects for which the map values experienced a slight decrease. In addition, the AP values of In defects and Rs defects for the original YOLOv8n were only 84.7 and 74.6, respectively, and our model improved the mAP values of these two defects by 1.8% and 10.6%, respectively. Our introduction of the C2f_DWR_DRB module in place of the C2f module enhanced the extraction of features from the scalable receptive fields at the higher level of the network. In addition to this, our method improved the accuracy of Cr defects from 35.1% to 49.7% and of Rs defects from 52.1% to 67.4%, that is, an increase of 14.6% and 15.3%, respectively. Our approach offers a notable enhancement of the detection accuracy and precision of these defects, thus demonstrating the efficacy of our approach.
Table 2. The detection performance of DDI-YOLO using the NEU-DET datasets.
To further learn about the detection capability of our proposed DDI-YOLO and baseline model YOLOv8n, the P–R curves of the two methods are shown in Figure 6a,b. The region enclosed by our proposed DDI-YOLO was to be greater than the region enclosed by YOLOv8. The NEU-DET dataset’s overall mAP was 78.3%, which was a significant improvement of 2.4 percentage points over YOLOv8n.
Figure 6. Comparison of different loss functions.
To confirm the efficacy of the network improvement proposed in this work, the improved YOLOv8n algorithmic model was contrasted with other algorithmic models for experimentation purposes. In this paper, the classical SSD, YOLOv3-tiny, YOLOv5n, YOLOv6, YOLOv7-tiny [33], YOLOv8n (baseline), and the newer RT-DETR [34] were selected for comparative studies. These experiments were carried out utilizing an identical environment, dataset, and equipment, as shown in the graphs for the outcomes of the studies conducted to compare various algorithms.
Table 3 shows that, compared to several other algorithms, this paper’s algorithm achieved the best performance (78.3% and 158 frames/s) in terms of mAP and FPS values, and was only second to YOLOv8n and YOLOv5n in terms of recall, but performed better than YOLOv8n with respect to all other metrics. This demonstrates that our improved model is superior to other algorithmic models in terms of detection accuracy and detection speed. Regarding the quantity of model parameters, compared to other algorithms, it was only a little higher than that characterizing YOLOv5, which is much smaller than most other algorithmic models. Figure 7 displays the experimental detection outcomes of several different algorithms, and it can be observed from the figure that the other algorithmic models were associated with wrong detections, missed detections, etc., while this paper’s proposed algorithmic model had the best detection effect. To sum up, the improved algorithm presented in this essay not only performed better in terms of precision of measurements, accuracy, and detection speed, but also reduced the amount of arithmetic involved, something which, in turn, improved the detection efficiency. Therefore, the improved algorithm proposed in this work has a rather high generalization value and practicality.
Table 3. Comparative experimental results.
Figure 7. Comparison of the experimental results.

4. Conclusions

In response to the necessity of detecting steel surface defects in actual production, our institute proposes the defect detection algorithm DDI-YOLO. First, the backbone layer of YOLOv8 was improved by proposing the C2f_DWR_DRB module structure, which improved the backbone’s ability to extract features. Second, in the neck structure, the C2f_DRB module was proposed to compensate for the inability of the C2f module in the neck network to capture the defects of small-scale patterns during training, and it improved the model’s training ability. Finally, the Inner-IoU loss function was used to replace the loss function CIoU of the initial YOLOv8n model, leading to a more accurate model and faster training. Based on the experimental results, the following conclusions can be drawn: compared with YOLOv8n, DDI-YOLO enhanced the mAP by 2.4%, the accuracy by 3.3%, and the FPS by 59 frames/s, all of which can satisfy more practical needs. In summary, our proposed algorithmic model can meet the requirements of industrial detection, offers advantages in practical industrial applications, and can be applied to a variety of industrial scenarios. We will assess the suggested method’s effectiveness in actual industrial defect detection in the future and make the necessary improvements based on the actual detection outcomes.

Author Contributions

Writing—review & editing, T.Z., P.P. and J.Z.; supervision, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China [grant numbers 52075348, 52175107, 52275119], National Key Research and Development Program of China [Grant numbers 2023YFB3408502], Natural Science Foundation of Liaoning Province [grant numbers 2022-MS-280], Shenyang Outstanding Young and Middle-aged Science and Technology Talents Project [Grant numbers RC230739], Basic Scientific Research Project of Liaoning Provincial Department of Education [Grant numbers JYTMS20231568].

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://faculty.neu.edu.cn/songkechen/zh_CN/zdylm/263270/list/index.htm (accessed on 25 May 2024).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yu, Q.; Wu, Q.; Liu, H. Research on X-Ray Contraband Detection and Overlapping Target Detection Based on Convolutional Network. In Proceedings of the 2022 4th International Conference on Frontiers Technology of Information and Computer (ICFTIC), Qingdao, China, 2–4 December 2022; pp. 736–741. [Google Scholar]
  2. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  3. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science. Springer International Publishing: Cham, Switzerland, 2016; Volume 9905, pp. 21–37, ISBN 978-3-319-46447-3. [Google Scholar]
  4. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science. Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241, ISBN 978-3-319-24573-7. [Google Scholar]
  5. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  6. Girshick, R. Fast R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  7. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  9. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  10. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  11. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Lecture Notes in Computer Science. Springer International Publishing: Cham, Switzerland, 2020; Volume 12346, pp. 213–229, ISBN 978-3-030-58451-1. [Google Scholar]
  12. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  13. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  14. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2016, arXiv:1511.06434. [Google Scholar]
  15. Applied Sciences|Free Full-Text|A Feature-Oriented Reconstruction Method for Surface-Defect Detection on Aluminum Profiles. Available online: https://www.mdpi.com/2076-3417/14/1/386 (accessed on 8 June 2024).
  16. Ren, F.; Fei, J.; Li, H.; Doma, B.T. Steel Surface Defect Detection Using Improved Deep Learning Algorithm: ECA-SimSPPF-SIoU-Yolov5. IEEE Access 2024, 12, 32545–32553. [Google Scholar] [CrossRef]
  17. Dou, Z.; Gao, H.; Liu, B.; Chang, F. Small sample steel plate defect detection algorithm of lightweight YOLOv8. Comput. Eng. Appl. 2024, 60, 90–100. [Google Scholar] [CrossRef]
  18. Guo, Z.; Wang, C.; Yang, G.; Huang, Z.; Li, G. Msft-Yolo: Improved Yolov5 Based on Transformer for Detecting Defects of Steel Surface. Sensors 2022, 22, 3467. [Google Scholar] [CrossRef]
  19. Cui, K.; Jiao, J. Steel surface defect detection algorithm based on MCB-FAH-YOLOv8. J. Graph. 2024, 45, 112–125. [Google Scholar]
  20. Zhou, Y.; Meng, J.; Wang, D.; Tang, Y. Steel defect detection based on multi-scale lightweight attention. Control Decis. 2024, 39, 901–909. [Google Scholar] [CrossRef]
  21. Zhu, W.; Zhang, H.; Zhang, C.; Zhu, X.; Guan, Z.; Jia, J. Surface Defect Detection and Classification of Steel Using an Efficient Swin Transformer. Adv. Eng. Inform. 2023, 57, 102061. [Google Scholar] [CrossRef]
  22. He, Y.; Song, K.; Meng, Q.; Yan, Y. An End-to-End Steel Surface Defect Detection Approach via Fusing Multiple Hierarchical Features. IEEE Trans. Instrum. Meas. 2019, 69, 1493–1504. [Google Scholar] [CrossRef]
  23. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  24. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  25. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  26. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  27. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Fang, J.; Yifu, Z.; Wong, C.; Montes, D. Ultralytics/Yolov5: V7. 0-Yolov5 Sota Realtime Instance Segmentation. Zenodo 2022. [Google Scholar] [CrossRef]
  28. Wei, H.; Liu, X.; Xu, S.; Dai, Z.; Dai, Y.; Xu, X. DWRSeg: Rethinking Efficient Acquisition of Multi-Scale Contextual Information for Real-Time Semantic Segmentation. arXiv 2023, arXiv:2212.01173. [Google Scholar]
  29. Ding, X.; Zhang, Y.; Ge, Y.; Zhao, S.; Song, L.; Yue, X.; Shan, Y. UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition. arXiv 2024, arXiv:2311.15599. [Google Scholar]
  30. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  31. Zhang, H.; Xu, C.; Zhang, S. Inner-IoU: More Effective Intersection over Union Loss with Auxiliary Bounding Box. arXiv 2023, arXiv:2311.02877. [Google Scholar]
  32. Song, K.; Yan, Y. A Noise Robust Method Based on Completed Local Binary Patterns for Hot-Rolled Steel Strip Surface Defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
  33. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  34. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-Time Object Detection. arXiv 2024, arXiv:2304.08069. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.