Next Article in Journal
Time Series Prediction Method of Clean Coal Ash Content in Dense Medium Separation Based on the Improved EMD-LSTM Model
Previous Article in Journal
Analysis of Shots Trajectory and Effectiveness in Women’s and Men’s Football European Championship Matches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FAD-Net: Automated Framework for Steel Surface Defect Detection in Urban Infrastructure Health Monitoring

1
School of Engineering, Zhejiang Normal University, Jinhua 321004, China
2
School of Artificial Intelligence, Zhejiang Normal University, Jinhua 321004, China
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2025, 9(6), 158; https://doi.org/10.3390/bdcc9060158
Submission received: 7 April 2025 / Revised: 23 May 2025 / Accepted: 10 June 2025 / Published: 13 June 2025

Abstract

Steel plays a fundamental role in modern smart city development, where its surface structural integrity is decisive for operational safety and long-term sustainability. While deep learning approaches show promise, their effectiveness remains limited by inadequate receptive field adaptability, suboptimal feature fusion strategies, and insufficient sensitivity to small defects. To overcome these limitations, we propose FAD-Net, a deep learning framework specifically designed for surface defect detection in steel materials within urban infrastructure. The network incorporates three key innovations: The RFCAConv module, which leverages dynamic receptive field construction and coordinate attention mechanisms to enhance feature representation for defects with long-range spatial dependencies and low-contrast characteristics. The MSDFConv module, employing multi-scale dilated convolutions with optimized dilation rates to preserve fine details while expanding the receptive field. An Auxiliary Head that introduces hierarchical supervision to improve the detection of small-scale defects. Experiments on the GC10-DET dataset showed that FAD-Net achieved 5.0% higher mAP@0.5 than baseline models. Cross-dataset validation with NEU and RDD2022 further confirmed its robustness. These results demonstrate FAD-Net’s effectiveness for automated infrastructure health monitoring.

1. Introduction

Steel, as the fundamental material for modern urban infrastructure, plays a decisive role in ensuring the structural safety of critical facilities such as bridges and tunnels. During steel manufacturing processes, surface defects inevitably emerge due to rolling parameter variations and non-uniform heating conditions. These production-induced defects subsequently become potential structural weak points when the material enters service, ultimately compromising the long-term reliability of urban infrastructure [1,2]. Characteristic surface defect categories in steel materials include Crease, Waist folding, Silkspot, Water spot, Punching, and Inclusion, as shown in Figure 1. These defects are often complex, with irregular shapes and long-range spatial dependencies, and some are even highly similar to the texture or color of the steel. Moreover, these defects vary widely in size and morphology. Small defects, such as Punching and Inclusion, are easily overlooked but can gradually develop into more serious problems over time.
Traditional manual visual inspection for steel surface defect detection is inefficient and subjective. With the rapid development of machine vision [3] and deep learning [4], image-based steel surface defect detection methods have been widely studied and applied. Machine vision, a powerful tool enabled by computer vision technology, enables the automatic detection of steel surface defects [5]. Deep learning methods can automatically extract high-level features from raw images, significantly improving detection accuracy and efficiency. However, current detection technologies suffer from high false detection and miss rates, making the improvement of algorithm accuracy and stability a key challenge [6,7].
Computer vision, a fundamental and challenging area in AI, has gained recent attention. Algorithm performance is typically evaluated based on accuracy and speed. With the fast progress of deep learning, deep-learning-based object detection has become mainstream. Currently, there are two main types of object detectors: two-stage regional detectors like the R-CNN [8] series and one-stage regression-based detectors like the YOLO series. Computer vision-based methods are now used in many real-world applications, such as defect inspection, autonomous driving, and medical imaging. The YOLO series [9,10,11,12,13,14,15,16], able to predict all objects in an image via a single forward pass, stands out among object detection algorithms due to its speed–accuracy balance [17]. Since its introduction, YOLO has evolved through several versions, each building on previous ones to address limitations and enhance performance. YOLO11, the latest version and a SOTA detector in object detection, is celebrated for its fast and stable detection. But YOLO11 still has shortcomings in handling complex steel surface defects. To address these issues, we present an optimized FAD-Net (Fusion-Aware Detection Network), which makes significant contributions in three key aspects:
  • Traditional convolution operations are limited by a fixed receptive field, which restricts the ability to capture both global features and detailed information in complex defect patterns. This paper introduces the RFCAConv (Receptive Field and Coordinate Attention Convolution) module, which combines the advantages of coordinate attention for spatial modeling with a dynamic receptive field construction strategy. By employing a non-parametric shared convolution structure, it reduces feature overlap and information loss, improving the model’s ability to capture long-range spatial dependencies and enhancing its ability to detect low-contrast defects.
  • In Feature Pyramid Networks (FPNs), traditional pyramid pooling can result in the loss of detailed information, especially when handling multi-scale defects. This paper proposes the MSDFConv (Multi-Scale Dilated Fusion Convolution) module, which uses stacked dilated convolutions instead of pooling to expand the receptive field, preserving detailed features while improving the model’s ability to detect defects at different scales.
  • Existing algorithms struggle with small target detection, especially when dealing with small defects on steel surfaces, often leading to missed detections. Therefore, this paper introduces the Auxiliary Head, which introduces additional supervisory signals at the intermediate layers of the network. Through multi-level feature fusion, it increases the network’s sensitivity to small defects.
The rest of this paper is organized as follows: Section 2 reviews the research progress in the field of defect detection; Section 3 introduces the main principles of the proposed algorithm; Section 4 analyzes the experimental results and data; Section 5 concludes the paper and discusses future directions.

2. Related Work

Surface defect inspection of steel materials is critical for ensuring the safety and durability of key structures like bridges and tunnels in urban infrastructure [1]. Existing steel surface defect detection methods are categorized into four types: structure-based, statistical-based, filtering-based, and learning-based [18].
Structure-based methods identify defects by extracting texture primitives and characterizing image regions according to their placement rules. For example, Zhang et al. [19] used an adaptive thresholding and morphological reconstruction approach for detecting defects in aluminum alloy wheels. Shi et al. [20] improved the Sobel edge detection algorithm to reduce noise impact and achieve more precise defect localization. However, the accuracy of this method is limited when detecting low-contrast defects. Liu et al. [21] proposed a mathematical morphology-based enhancement operator (EOBMM) to address uneven illumination in steel strip defect images and enhance image details. However, this method has high computational complexity when dealing with multi-scale defects. Statistical-based methods describe texture features using the grayscale distribution of image regions, and common techniques include histograms, co-occurrence matrices, and local binary patterns. For instance, Aminzadeh M et al. [22] presented a thresholding method for defect identification based on histogram information. While histogram methods have been used in modeling and visualization, the information they can extract is limited. Yan K et al. [23] enhanced the LBP method by introducing the Complete Local Ternary Pattern (CLTP) to detect welds, improving performance across different scenarios such as varying lighting conditions. Nevertheless, LBP-based methods remain sensitive to noise and directional information and are prone to interference from complex backgrounds.
Filtering-based methods analyze images as 2D signals using techniques like curvelet transforms, Gabor filters, and wavelet transforms. For example, Cong et al. [24] used Gabor filters to extract texture features of surface defects in steel strips, determining optimal filter parameters by maximizing an evaluation function that quantified the energy response difference between defective and non-defective images. Wu et al. [25] decomposed surface defect images of hot-rolled steel plates into 40 components across five scales and eight directions using Gabor wavelets. They extracted the mean and variance of the real and imaginary parts for each component and the original image, forming a 162-dimensional feature vector. However, these methods have high computational complexity due to localized analysis in both spatial and frequency domains, making it challenging to meet real-time detection requirements.
With the advancement of computational power and deep learning technologies, deep learning-based methods have gradually become mainstream in steel strip surface defect detection. These methods can automatically learn feature representations without manual feature extraction rule design, and they offer stronger representation and adaptability. Current deep learning-based defect detection methods are mainly divided into single-stage and two-stage algorithms. Two-stage algorithms first extract candidate boxes from an image and then refine the detection results based on the candidate regions. For example, Zhao et al. [26] proposed a transmission line insulator identification and fault detection model based on improved Faster R-CNN. This model integrates feature pyramid networks (FPNs) to enhance insulator detection in complex background images and employs an HSV color space adaptive threshold algorithm for image segmentation to mitigate the effects of lighting, background noise, and shooting angles. Liu et al. [27] proposed a multi-scale context detection network (MSC-DNet) that introduces auxiliary image-level supervision (AIS) to enhance defect-distinguishing features, achieving excellent performance in complex backgrounds. However, two-stage detection models have high computational complexity, making it difficult to meet real-time detection requirements. Single-stage algorithms directly compute results from images and are fast. For example, Wu et al. [28] proposed a lightweight feature fusion method based on YOLOv3, enabling the detection network to operate on single-scale feature maps and reducing computational complexity. Nevertheless, its adaptability to different materials and textures is limited, especially in complex steel surface textures and multi-scale defect scenarios. Zhang et al. [29] proposed CrackNet, which eliminates pooling layers to maintain original image resolution and uses feature extractors to generate multi-directional and multi-scale feature maps. Yin et al. [30] developed an automatic sewer pipeline defect detection framework based on YOLOv3, employing a three-scale prediction method to identify defects of various sizes. Zhao et al. [31] proposed an improved RDD-YOLO model based on YOLOv5, integrating Res2Net as the backbone network, dual feature pyramid networks, and a decoupled head structure, achieving a balance between detection accuracy and real-time performance. However, this model still has limitations in fine-grained feature extraction and the accurate identification of complex defects.
Although existing research has made significant progress in steel defect detection for urban infrastructure, current methods still have obvious limitations in handling the unique texture characteristics and complex shape variations of construction steel surfaces. They also struggle to meet the low computational complexity requirements of real-time monitoring scenarios. This study proposes a surface defect detection model specifically designed for construction steel materials, offering a reliable solution for the intelligent safety maintenance of urban infrastructure [32].

3. Methods

To tackle the challenges of long-range saptial dependencies, low contrast in some defects, significant scale differences, diverse defect shapes, and limited small target detection capabilities on steel surfaces, we propose the FAD-Net model. To enhance the model’s ability to recognize long-range, complex defects and improve the representation of defect region features, we introduce the RFCAConv module, which replaces the conventional convolutional modules in the network architecture as well as those in the C3k2 feature extraction module. This module applies the CA mechanism to the receptive field’ s spatial features, focusing on regions with significant information, and uses them in place of standard convolution operations. To address the issue of local detail loss caused by pooling in traditional spatial pyramid pooling operations, we propose the MSDFConv module to replace the spatial pooling pyramid module. This module captures multi-scale information by stacking convolutions with varying dilation rates and uses a shared convolution kernel design to reduce computational complexity. Finally, we propose the Auxiliary Head. By integrating multi-scale feature learning modules P3, P4, and P5 and incorporating a larger positive sample allocation strategy (top-K = 13) and a dynamic task alignment allocator (Task-Aligned Assigner), this model effectively enhances the network’s sensitivity to small defects, thereby improving detection completeness.
The architecture of the FAD-Net network is illustrated in Figure 2.

3.1. RFCAConv

Steel surfaces exhibit unique defect distribution patterns. Some defects span the entire width or concentrate at edges, featuring complex shapes and long-range spatial dependence. Additionally, low-contrast defects possess textural features highly similar to the background, significantly increasing detection difficulty. Conventional convolutional operations, due to their parameter-sharing mechanism, cannot capture differences in information from various positions. Although existing spatial attention mechanisms enhance model focus on key areas by adjusting feature weight distributions, the issue of shared attention weights persists. To address these challenges, this paper proposes an innovative RFCAConv module that combines dynamic receptive field strategies with a CA [33] mechanism.
In the RFCAConv module, as illustrated in Figure 3, the input feature map X first undergoes a set of 3 × 3 group convolutions, increasing the number of channels ninefold. The feature map is then reshaped into a C × 3H × 3W receptive field spatial feature map. This rearrangement expands each point in the original feature map into a 3 × 3 matrix, thereby significantly enriching the feature information across different spatial locations. Subsequently, average pooling operations are performed on the receptive field spatial feature map along the H and W dimensions to aggregate global information within the receptive field. The features from these two dimensions are stacked, and a 1 × 1 convolution is applied to facilitate information exchange among channels. This step reduces redundancy among channels and enhances feature representation.
The aggregated feature vector is divided into two parts, corresponding to the height and width information of the receptive field spatial feature map. Separate 1 × 1 convolution operations are performed on these two parts to independently generate attention weight matrices for height and width. The attention weights in both the height and width directions are element-wise multiplied with the receptive field spatial feature map. This operation enhances key regional information within the feature map while suppressing irrelevant areas. When detecting distant defects, dynamic weights place greater emphasis on the defect’s extension direction, enabling the model to better capture long-range dependent defects. For low-contrast defects, the model focuses on local minute features, significantly improving detection accuracy.
Finally, a 3 × 3 convolutional layer with a stride of 3 is employed to restore the spatial dimensions of the feature map to match the original input, ensuring seamless integration with subsequent network layers. In the network implementation, all standard convolutional modules in the backbone and neck networks are replaced by RFCAConv modules. Within the C3k2 Bottleneck structure, the RFCAConv module replaces the second standard convolutional module. The detailed structure of the improved C3k2_RFCAConv is depicted in Figure 4.

3.2. MSDFConv

The surface of construction steels exhibits distinct multi-scale characteristics. Traditional multi-scale feature extraction and fusion modules employ fixed-size pooling windows to integrate information across different scales. While this approach mitigates scale differences to some extent, it falls short in capturing detailed features. Moreover, the fixed window size restricts the expansion of the receptive field, hindering the balance between local details and global context.
To address these limitations, we propose the MSDFConv module, which incorporates dilated convolutions with varying dilation rates within the convolutional layers. This approach enables feature extraction across different scales, which is highly beneficial for capturing objects of varying sizes and contextual information within images. Specifically, convolutional layers with lower dilation rates excel at capturing fine local details such as edges and textures. In contrast, those with higher dilation rates perform better in capturing the global context, including the relative positions of objects and background information.
In order to boost the efficiency of the model and reduce redundancy, we implemented Shared Convolutional Layers (ShareConv). This approach reduces the number of parameters by sharing the weights of the convolutional kernel W s h a r e .
As shown in Figure 5, the input feature map X is first transformed by a 1 × 1 convolution, which reduces the number of channels to a hidden dimension C h i d , producing the feature map Y 0 :
Y 0 = C o n v ( X , W c o n v )
Then, for each dilation rate r r 1 , r 2 , , r n , the shared convolutional kernel weights W s h a r e are applied to the previous feature map Y i 1 to generate a new feature map Y i :
Y i = C o n v ( Y i 1 , W s h a r e , d i l a t i o n = r )
To ensure the size consistency of input and output feature maps, dynamic padding adjustments are applied to each dilated convolution layer, ensuring a smooth transition between convolutions with different dilation rates, and the calculation formula is as follows:
p = ( k 1 ) · r 2
Finally, all the feature maps Y 0 , Y 1 , , Y n are concatenated, and a 1 × 1 convolution is applied to adjust the channel dimensions of the concatenated feature maps, resulting in the final output feature map:
O u t p u t = C o n v ( C o n c a t ( Y 0 , Y 1 , , Y n ) , W f i n a l )
Through parameter-sharing and multi-scale feature extraction, the model maintains its performance while achieving greater computational efficiency.

3.3. Auxiliary Head

Surfaces of steel materials in urban infrastructure often exhibit minute surface anomalies. Although dilated convolutions expand the receptive field and effectively prevent detail loss from traditional pooling operations, larger dilation rates may still overlook high-resolution details. To address this issue, we introduced an Auxiliary Head into the baseline detection framework. By adding extra loss terms to intermediate feature maps, the network is guided to enhance sensitivity to small-scale defects.
Specifically, this design incorporates additional supervisory signals into the network’s intermediate layers (P3, P4, P5) to guide feature learning across different scales. The P3 layer, being the lowest-level feature map, captures the contours and textures of small targets. The P4 layer integrates the detailed features from P3, combining local and contextual information to improve the accuracy of small target detection. The P5 layer provides high-level semantic information that lower layers fail to capture effectively. Through multi-layer feature fusion and additional supervisory signals, the Auxiliary Head helps the network focus on these challenging-to-detect small targets.
In the label assignment process, we adopt the dynamic task alignment strategy (Task-Aligned Assigner) from the baseline model, calculating the alignment score between ground truth labels and predicted results to determine anchor-level alignment [34]. The specific calculation formula is as follows:
t = s α + u β
Here, s denotes the classification score, u is the Intersection over Union (IoU) value, and α and β are weight hyperparameters. During positive sample allocation, each ground truth box is ranked by its alignment score with the predicted boxes. During training, the model selects the top n predicted boxes as positive samples for learning. In the baseline model, the number of positive samples n for the original detection head is set to 10. Since the Auxiliary Head has a weaker learning capacity, we set the number of positive samples n for the Auxiliary Head to 13 to provide more learning opportunities. As shown in Figure 6, the Auxiliary Head is positioned in the intermediate layers of the network, providing additional supervisory signals for coarse features. Meanwhile, the original detection head focuses on learning residual information that has not been captured, gradually improving overall model performance.
As shown in Figure 7, the training strategy for the Auxiliary Head consists of several key steps. First, during label assignment, the original detection head guides the label allocator to generate fine samples, improving the accuracy and localization of target detection. The Auxiliary Head, in contrast, guides the label allocator to generate coarse samples, helping the model learn more diverse features and background information. By calculating the loss for each sample separately, this design helps reduce missed detections and prevents the negative impact of additional coarse samples on prediction results. During training, the additional supervisory signals from the Auxiliary Head improve the network’s stability and generalization. However, during inference and validation, the Auxiliary Head is removed to ensure the model’s final accuracy and robustness.

4. Results

4.1. Datasets

For model performance evaluation, this study employed the publicly available GC10-DET dataset, which originates from the production process of hot-rolled strip steel, the most extensively utilized structural material in urban infrastructure applications. As the fundamental material for critical components including building steel structures and bridge load-bearing elements, the dataset comprehensively documents ten characteristic defect types on hot-rolled steel surfaces: Punching, Welding line, Water spot, Crescent gap, Oil spot, Inclusion, Silkspot, Rolled pit, Crease, Waist folding. Correspondence between the types of defects and the labels of the dataset can be seen in Table 1. The dataset contains 2280 high-quality grayscale images, split into training and validation sets at a ratio of 4:1.

4.2. Experimental Environment

The experiments in this study were run on a server leased by AutoDL. The experimental environment was as follows: the CPU was an Intel(R) Xeon(R) Platinum 8362V @ 2.80GHz and the GPU was an Nvidia GeForce RTX 3090 with 24 GB VRAM. The programming language used was Python 3.8.10, with CUDA driver version 11.3. The deep learning framework was Pytorch-GPU 1.10.0 + cu113. The default settings and hyperparameters [35] for the experiment are shown in Table 2.

4.3. Evaluation Metrics

To analyze the experimental results, this study used mean Average Precision (mAP), number of parameters (Parameters), floating-point operations per gigaflop (FLOPs/G), and frames per second (FPSs) to evaluate model performance. In addition, Precision and Recall were used to measure false positive and false negative rates. The formulas for calculating Precision and Recall were as follows:
P = T P T P + F P
R = T P T P + F N
where TP denotes true positives (products with defects and correctly detected), FP denotes false positives (products without defects but detected as defective), and FN denotes false negatives (products with defects but not detected). Average Precision (AP) is the area under the Precision–Recall (PR) curve and mAP is the average of AP values across all categories. A higher mAP indicated better model detection accuracy. The calculation of AP and mAP is shown in Formulas (8) and (9), where N represents the number of categories in the dataset:
A P = 0 1 P ( R ) d R
m A P = i = 1 k A P i N
Intersection over Union (IoU) measured the overlap of predicted bounding boxes, with a threshold to distinguish between positive and negative samples. This study evaluated model performance using mAP@0.5 and mAP@0.5–0.95. The parameter count (Params) represented the sum of all trainable parameters within a model, serving as a metric for the spatial complexity of the model. The floating-point operations per second (FLOPs) quantified the computational load during a single forward propagation through the model and were typically utilized to evaluate the temporal complexity. The frames per second (FPSs) metric assessed the model’s processing velocity on hardware, reflecting the rate at which the model could handle images per second.

4.4. Experimental Results and Analysis

To verify that the improved modules did not conflict with the model and positively impacted its performance, we conducted ablation experiments on the model with the selected modules. The experimental results are shown in Table 3. The results show that the improved model achieved peak performance, with mAP@0.5 and mAP@0.5–0.95 increasing to 68.3% and 34.0%, respectively, with only a slight increase in computational complexity. Overall, the detection performance was significantly improved with minimal additional computational burden.
To verify the advantages of the RFCAConv module in detecting long-range spatial dependencies and low-contrast defects, we conducted comprehensive comparison experiments based on the baseline model. As shown in Table 4 and Table 5, for the two typical long-range spatial dependency defects (Crease and Waist folding), RFCAConv improved mAP@0.5 by 27.8% and 2.3% and mAP@50–95 by 8.4% and 0.7%. For the two typical low-contrast defects (Silkspot and Water spot), it enhanced mAP@0.5 by 8% and 4.4% and mAP@0.5–95 by 2.9% and 2.1%.
To achieve the best balance between receptive field expansion, edge information retention, and global feature capture, we explored various dilation rate combinations. The pixel utilization distribution for each combination is shown in Figure 8a–e. The r = [1,1,1] combination had slow receptive field expansion and focused on local information. The r = [1,2,3] combination’s receptive field grew with increasing dilation rates but payed less attention to edge details. The r = [2,1,2] combination balanced receptive field expansion and edge retention but had limited global feature capture. The r = [2,2,2] combination, with a common factor greater than 1, caused detail loss. Finally, the r = [1,3,5] combination rapidly expanded the receptive field and achieved the best balance between global features and edge information. We then verified the impact of dilated convolutions with different dilation rates on network predictions, with the results shown in Table 6. The r = [1,3,5] combination showed the best performance in mAP@0.5, mAP@0.5–0.95, Precision, and Recall.
To validate the effectiveness of the Aux Head module for small-target defect detection, we performed multiple comparative experiments based on the baseline model, with detailed results presented in Table 7. The experimental results demonstrated that for two representative small-target defect categories—Punching and Waist folding—the Aux Head module achieved optimal performance, showing improvements of 1.1% and 2.4% in mAP@0.5, along with enhancements of 2.0% and 0.8% in mAP@0.5–0.95 compared to the baseline, respectively.
To further validate the outstanding detection performance of FAD-Net, we conducted a comparative experiment on the GC10-DET dataset with mainstream detection algorithms: Faster-RCNN, DINO, Retinanet, RTMDET, RT-DETR, Swin-Transformer, and other YOLO series models. The experimental results are shown in Table 8. FAD-Net achieved the best performance in terms of detection accuracy. Specifically, FAD-Net improved mAP@0.5 by 9% over RetinaNet and outperformed the high-precision two-stage detectors Faster-RCNN and DINO by 3.4% and 1.8%, respectively. When compared with other YOLO series algorithms, FAD-Net also achieved the highest detection accuracy. Notably, FAD-Net demonstrated superior computational efficiency while maintaining exceptional detection accuracy. Comparative analysis of floating-point operations (FLOPs) and parameter counts (Params) confirmed FAD-Net’s effective complexity control—a critical advantage for real-time structural health monitoring in urban infrastructure applications.
Figure 9 presents the model comparison results. YOLO-series algorithms, with their compact architectures and lower FLOPs/G, are more suitable for real-time inspection requirements of urban infrastructure steel structures, achieving superior balance between computational efficiency and detection speed. Comparative experiments revealed FAD-Net’s 5% mAP improvement over baseline models, demonstrating its distinct advantages.
To verify the generalization ability of the proposed framework for different types of defects in both structural components and road scenarios, cross-dataset evaluations were performed using the NEU dataset from Northeastern University and the RDD2022 road defect dataset, respectively. The NEU dataset comprises six common characteristic defect categories in structural components: rolled scale, patches, cracks, scratches, pitted surfaces, and inclusions, with 300 samples per class (1800 samples in total). The RDD2022 road defect dataset contains four common types of road surface defects: longitudinal cracks, transverse cracks, alligator cracks, and potholes, with a total of 173,767 samples. Both were split into training and validation sets at a ratio of 4:1. Generalization experiment results are shown in Table 9. All models were trained and tested using the same hyperparameters as in Table 2.
Figure 10 presents a comparison of key metrics between the baseline model and our proposed FAD-Net model over 300 training epochs on the validation set. Figure 10a shows that FAD-Net significantly outperformed the baseline model in terms of mAP@0.5. Figure 10b further illustrates that under the mAP@0.5–0.95 evaluation standard, FAD-Net continued to perform excellently, with improvements observed at multiple IoU thresholds. Figure 10c,d show comparisons of Precision and Recall, where FAD-Net also significantly outperformed the baseline model, indicating its advantage in detection accuracy and recall rate. In conclusion, the FAD-Net model demonstrates higher accuracy and reliability in object detection tasks, validating its effectiveness.
Figure 11 presents the model’s prediction results, demonstrating the improvements of FAD-Net over the baseline model in terms of missed detection, false detection, and confidence level. Specifically, lower missed and false detection rates showed that the model’s feature extraction was more accurate and its ability to suppress background interference was stronger, highlighting the model’s robustness and generalization ability in complex scenarios. Higher confidence in detection boxes indicated a higher probability of the target being present, reflecting the model’s ability to capture more target details. The first set of images shows that FAD-Net detected defects missed by the baseline model, reducing the missed detection rate. The second set shows false detections from the baseline model, which FAD-Net successfully avoided. The third set further demonstrates that FAD-Net improved the detection accuracy of the baseline model.

5. Conclusions and Discussion

The proposed FAD-Net is an advanced deep learning framework targeting the detection of surface defects in steel materials used in urban infrastructure. It has three main innovations. First, the RFCAConv module, which uses a coordinated attention mechanism, can dynamically focus on key areas in the receptive field, thereby significantly enhancing the detection of long-distance and low-contrast defects. Second, the MSDFConv module, which uses dilated convolutions with varying dilation rates, effectively captures multi-scale information and addresses the problem of local detail loss in traditional spatial pyramid pooling. Third, the auxiliary head, which combines an optimized positive sample allocation strategy and a dynamic task-aligned assigner, performs feature fusion across multiple layers to improve the detection of small defects. The experimental results demonstrate the effectiveness of FAD-Net, with a 5% improvement in mAP@0.5 to 68.3% on the GC10-DET dataset. Cross-dataset validation on the NEU and RDD2022 datasets showed respective improvements of 2.5% and 2.8% in mAP@0.5 compared to the baseline model, indicating good generalization in defect detection.
Current algorithm optimization is mainly based on laboratory-environment validation and does not consider the computational constraints in industrial deployment. In the future, we will actively explore the deployment and application of the algorithm on edge devices and migrate it from the laboratory to real-world factory production processes to contribute to more reliable urban infrastructure inspection.

Author Contributions

Conceptualization, N.W.; methodology, N.W. and Y.C.; software, N.W. and W.L.; validation, N.W., Y.C., W.L., and L.Z.; formal analysis, L.Z.; investigation, N.W. and W.L.; resources, J.T.; data curation, N.W. and L.Z.; writing—original draft preparation, N.W.; writing—review and editing, J.T.; visualization, W.L.; supervision, J.T.; project administration, J.T.; funding acquisition, J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Wenzhou city “unveiled the list of marshals—global attraction of talent” special program grant number ZR2022004 and the APC was funded by Jinghong Tian.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available upon reasonable request to the author, Nian Wang.

Acknowledgments

The authors acknowledge the valuable support of the editors and reviewers.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mostafa, K.; Hegazy, T. Review of image-based analysis and applications in construction. Autom. Constr. 2021, 122, 103516. [Google Scholar] [CrossRef]
  2. Spencer, B.F., Jr.; Hoskere, V.; Narazaki, Y. Advances in computer vision-based civil infrastructure inspection and monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  3. Wiley, V.; Lucas, T. Computer vision and image processing: A paper review. Int. J. Artif. Intell. Res. 2018, 2, 29–36. [Google Scholar] [CrossRef]
  4. Dong, S.; Wang, P.; Abbas, K. A survey on deep learning and its applications. Comput. Sci. Rev. 2021, 40, 100379. [Google Scholar] [CrossRef]
  5. Jia, F.; Lei, Y.; Lu, N.; Xing, S. Deep normalized convolutional neural network for imbalanced fault classification of machinery and its understanding via visualization. Mech. Syst. Signal Process. 2018, 110, 349–367. [Google Scholar] [CrossRef]
  6. Li, W.; Zhang, H.; Wang, G.; Xiong, G.; Zhao, M.; Li, G.; Li, R. Deep learning based online metallic surface defect detection method for wire and arc additive manufacturing. Robot. Comput.-Integr. Manuf. 2023, 80, 102470. [Google Scholar] [CrossRef]
  7. Luo, Q.; Fang, X.; Liu, L.; Yang, C.; Sun, Y. Automated visual defect detection for flat steel surface: A survey. IEEE Trans. Instrum. Meas. 2020, 69, 626–644. [Google Scholar] [CrossRef]
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  10. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  12. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  13. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  14. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  15. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning what you want to learn using programmable gradient information. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 15–20 September 2024; pp. 1–21. [Google Scholar]
  16. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. YOLOv10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  17. Czimmermann, T.; Ciuti, G.; Milazzo, M.; Chiurazzi, M.; Roccella, S.; Oddo, C.M.; Dario, P. Visual-based defect detection and classification approaches for industrial applications—A survey. Sensors 2020, 20, 1459. [Google Scholar] [CrossRef] [PubMed]
  18. Wen, X.; Shan, J.; He, Y.; Song, K. Steel surface defect recognition: A survey. Coatings 2022, 13, 17. [Google Scholar] [CrossRef]
  19. Zhang, J.; Guo, Z.; Jiao, T.; Wang, M. Defect detection of aluminum alloy wheels in radiography images using adaptive threshold and morphological reconstruction. Appl. Sci. 2018, 8, 2365. [Google Scholar] [CrossRef]
  20. Shi, T.; Kong, J.-Y.; Wang, X.-D.; Liu, Z.; Zheng, G. Improved Sobel algorithm for defect detection of rail surfaces with enhanced efficiency and accuracy. J. Cent. South Univ. 2016, 23, 2867–2875. [Google Scholar] [CrossRef]
  21. Liu, M.; Liu, Y.; Hu, H.; Nie, L. Genetic algorithm and mathematical morphology based binarization method for strip steel defect image with non-uniform illumination. J. Vis. Commun. Image Represent. 2016, 37, 70–77. [Google Scholar] [CrossRef]
  22. Aminzadeh, M.; Kurfess, T. Automatic thresholding for defect detection by background histogram mode extents. J. Manuf. Syst. 2015, 37, 83–92. [Google Scholar] [CrossRef]
  23. Yan, K.; Dong, Q.; Sun, T.; Zhang, M.; Zhang, S. Weld defect detection based on completed local ternary patterns. In Proceedings of the International Conference on Video and Image Processing, Singapore, 27–29 December 2017; pp. 6–14. [Google Scholar]
  24. Cong, J.-H.; Yan, Y.-H.; Dong, D.-W. Application of Gabor filter to strip surface defect detection. J. Northeast. Univ. Nat. Sci. 2010, 31, 257. [Google Scholar]
  25. Wu, X.-Y.; Xu, K.; Xu, J.-W. Automatic recognition method of surface defects based on Gabor wavelet and kernel locality preserving projections. Acta Autom. Sin. 2010, 36, 199–222. [Google Scholar] [CrossRef]
  26. Zhao, W.; Xu, M.; Cheng, X.; Zhao, Z. An insulator in transmission lines recognition and fault detection model based on improved faster RCNN. IEEE Trans. Instrum. Meas. 2021, 70, 5016408. [Google Scholar] [CrossRef]
  27. Liu, R.; Huang, M.; Gao, Z.; Cao, Z.; Cao, P. MSC-DNet: An efficient detector with multi-scale context for defect detection on strip steel surface. Measurement 2023, 209, 112467. [Google Scholar] [CrossRef]
  28. Wu, H.; Li, B.; Tian, L.; Feng, J.; Dong, C. An adaptive loss weighting multi-task network with attention-guide proposal generation for small size defect inspection. Vis. Comput. 2024, 40, 681–698. [Google Scholar] [CrossRef]
  29. Zhang, A.; Wang, K.C.P.; Fei, Y.; Liu, Y.; Chen, C.; Yang, G.; Li, J.Q.; Yang, E.; Qiu, S. Automated pixel-level pavement crack detection on 3D asphalt surfaces with a recurrent neural network. Comput.-Aided Civ. Infrastruct. Eng. 2019, 34, 213–229. [Google Scholar] [CrossRef]
  30. Yin, X.; Chen, Y.; Bouferguene, A.; Zaman, H.; Al-Hussein, M.; Kurach, L. A deep learning-based framework for an automated defect detection system for sewer pipes. Autom. Constr. 2020, 109, 102967. [Google Scholar] [CrossRef]
  31. Zhao, C.; Shu, X.; Yan, X.; Zuo, X.; Zhu, F. RDD-YOLO: A modified YOLO for detection of steel surface defects. Measurement 2023, 214, 112776. [Google Scholar] [CrossRef]
  32. Yang, J.; Li, S.; Wang, Z.; Dong, H.; Wang, J.; Tang, S. Using deep learning to detect defects in manufacturing: A comprehensive survey and current challenges. Materials 2020, 13, 5755. [Google Scholar] [CrossRef]
  33. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  34. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-aligned one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3490–3499. [Google Scholar]
  35. Bengio, Y. Practical recommendations for gradient-based training of deep architectures. In Neural Networks: Tricks of the Trade, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 437–478. [Google Scholar]
  36. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  37. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Yu, Z.; Huang, H.; Chen, W.; Su, Y.; Liu, Y.; Wang, X. YOLO-FaceV2: A scale and occlusion aware face detector. Pattern Recognit. 2024, 155, 110714. [Google Scholar] [CrossRef]
  39. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  40. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  41. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.-Y. DINO: DETR with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  42. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  43. Lyu, C.; Zhang, W.; Huang, H.; Zhou, Y.; Wang, Y.; Liu, Y.; Zhang, S.; Chen, K. RTMDet: An empirical study of designing real-time object detectors. arXiv 2022, arXiv:2212.07784. [Google Scholar]
  44. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs beat YOLOs on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 16965–16974. [Google Scholar]
  45. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
Figure 1. Example diagram of steel surface defects. (a) Crease. (b) Waist folding. (c) Silkspot. (d) Water spot. (e) Punching. (f) Inclusion.
Figure 1. Example diagram of steel surface defects. (a) Crease. (b) Waist folding. (c) Silkspot. (d) Water spot. (e) Punching. (f) Inclusion.
Bdcc 09 00158 g001
Figure 2. Structure of the FAD-NET network.
Figure 2. Structure of the FAD-NET network.
Bdcc 09 00158 g002
Figure 3. Structure of RFCAConv.
Figure 3. Structure of RFCAConv.
Bdcc 09 00158 g003
Figure 4. Structure of C3k2_RFCAConv.
Figure 4. Structure of C3k2_RFCAConv.
Bdcc 09 00158 g004
Figure 5. Structure diagram of the MSDFConv module.
Figure 5. Structure diagram of the MSDFConv module.
Bdcc 09 00158 g005
Figure 6. Improved model with Auxiliary Head.
Figure 6. Improved model with Auxiliary Head.
Bdcc 09 00158 g006
Figure 7. Auxiliary Head training strategy.
Figure 7. Auxiliary Head training strategy.
Bdcc 09 00158 g007
Figure 8. Distribution of pixel utilization for different combinations of dilation rates. (a) r = [1,1,1]. (b) r = [1,2,3]. (c) r = [2,1,2]. (d) r = [2,2,2]. (e) r = [1,3,5].
Figure 8. Distribution of pixel utilization for different combinations of dilation rates. (a) r = [1,1,1]. (b) r = [1,2,3]. (c) r = [2,1,2]. (d) r = [2,2,2]. (e) r = [1,3,5].
Bdcc 09 00158 g008
Figure 9. Comparison of the experimental results across models.
Figure 9. Comparison of the experimental results across models.
Bdcc 09 00158 g009
Figure 10. Comparison of key indicators. (a) mAP50 curve graph. (b) mAP50-95 curve graph. (c) Precision curve graph. (d) Recall curve graph.
Figure 10. Comparison of key indicators. (a) mAP50 curve graph. (b) mAP50-95 curve graph. (c) Precision curve graph. (d) Recall curve graph.
Bdcc 09 00158 g010
Figure 11. Model prediction results. (a) Original label. (b) Baseline model prediction result. (c) FAD-Net prediction result.
Figure 11. Model prediction results. (a) Original label. (b) Baseline model prediction result. (c) FAD-Net prediction result.
Bdcc 09 00158 g011
Table 1. Defect categories and validation set distribution.
Table 1. Defect categories and validation set distribution.
Category No.Defect TypesLabelsValidation Instances
0Crescent gap3_yueyawan52
1Crease9_zhehen13
2Silkspot6_siban181
3Water spot4_shuiban92
4Welding line2_hanfeng102
5Inclusion7_yiwu88
6Oil spot5_youban77
7Rolled pit8_yahen21
8Punching1_chongkong66
9Waist folding10_yaozhe26
Table 2. Default settings and hyperparameters.
Table 2. Default settings and hyperparameters.
Default SettingsHyperparameters
epoch300h0/trf0.01
batchsize32box7.5
workers4cls0.5
optimizerSGDdfl1.5
imgsz640momentum0.937
resumeFalsewarmup_epochs3.0
overlap_maskTruewarmup_momentum0.8
mask_ratio4warmup_bias_lr0.1
iou0.7weight_decay0.0005
Table 3. Comparison of ablation experiment results.
Table 3. Comparison of ablation experiment results.
YOLO11RFCAConvMSDFConvAux HeadmAP@0.5%mAP@0.5–0.95%FLOPs/GParams/M
---63.332.76.32.5
--66.633.26.92.7
--65.333.26.32.7
--64.833.36.32.5
-67.434.36.92.8
-67.433.76.92.7
68.334.06.92.8
Table 4. Performance comparison of attention modules in detecting long-range spatial dependency defects.
Table 4. Performance comparison of attention modules in detecting long-range spatial dependency defects.
ModuleCrease (Long-Range Defects1)Waist Folding (Long-Range Defects2)
mAP@0.5/%mAP@0.5–0.95/%mAP@0.5/%mAP@0.5–0.95/%
YOLO11 (Baseline)25.212.490.249.7
+SE [36]31.215.891.350.2
+CBAM [37]35.818.291.950.2
+RFCAConv53.020.892.550.4
Table 5. Performance comparison of attention modules in detecting low-contrast defects.
Table 5. Performance comparison of attention modules in detecting low-contrast defects.
ModuleSilkspot (Low-Contrast Defects1)Water Spot (Low-Contrast Defects2)
mAP@0.5/%mAP@0.5–0.95/%mAP@0.5/%mAP@0.5–0.95/%
YOLO11 (Baseline)52.223.271.438.2
+SE [36]55.824.172.839.1
+CBAM [37]58.125.374.239.6
+RFCAConv60.226.175.840.3
Table 6. Comparison of network performance with different dilation rates in dilated convolutions.
Table 6. Comparison of network performance with different dilation rates in dilated convolutions.
Dilation RatesmAP@0.5/%mAP@0.5–0.95/%PrecisionRecall
r = [ 1 , 1 , 1 ] 66.033.163.260.7
r = [ 1 , 2 , 3 ] 65.033.263.864.0
r = [ 2 , 1 , 2 ] 65.230.662.565.3
r = [ 2 , 2 , 2 ] 62.131.261.860.4
r = [ 1 , 3 , 5 ] 65.333.264.065.7
Table 7. Performance comparison of different detection modules in detecting small defects.
Table 7. Performance comparison of different detection modules in detecting small defects.
ModulePunching (Small Defect1)Inclusion (Small Defect2)
mAP@0.5/%mAP@0.5–0.95/%mAP@0.5/%mAP@0.5–0.95/%
YOLO11 (Baseline)97.655.541.413.0
+SEAMHead [38]96.854.240.112.3
+EfficientHead [39]95.452.738.511.8
+PGI [40]98.156.342.013.2
+Aux Head98.757.543.813.8
Table 8. Comparison of the performance of different algorithms on the GC10-DET dataset.
Table 8. Comparison of the performance of different algorithms on the GC10-DET dataset.
ModelmAP@0.5/%mAP@0.5–0.95/%FLOPs/GParams/MFPS
Faster-RCNN-r50 [8]64.931.6156.041.559.3
DINO-4scale [41]66.533.3205.047.625.2
Retinanet [42]59.328.18.04.9104.0
RTMDET-tiny [43]66.633.3170.137.964.7
RT-DETR-r50 [44]62.030.4125.74.276.9
Swin-Transformer-tiny [45]63.029.977.629.7135.3
YOLOv5n62.631.97.12.5395.5
YOLOv8n63.230.48.13.0354.8
YOLOv9t [15]63.132.57.61.97329.9
YOLOv10n [16]60.530.26.52.3305.3
YOLO11n (baseline)63.332.76.32.5402.0
FAD-Net (ours)68.334.06.92.8362.9
Table 9. Generalizability experiment.
Table 9. Generalizability experiment.
DatasetModelmAP@0.5/%mAP@0.5–0.95/%FLOPs/GParams/M
NEUYOLO11 (baseline)76.542.76.32.5
FAD-Net (ours)79.044.66.92.8
RDD2022YOLO11 (baseline)56.428.96.32.5
FAD-Net (ours)59.230.56.92.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, N.; Chen, Y.; Li, W.; Zhang, L.; Tian, J. FAD-Net: Automated Framework for Steel Surface Defect Detection in Urban Infrastructure Health Monitoring. Big Data Cogn. Comput. 2025, 9, 158. https://doi.org/10.3390/bdcc9060158

AMA Style

Wang N, Chen Y, Li W, Zhang L, Tian J. FAD-Net: Automated Framework for Steel Surface Defect Detection in Urban Infrastructure Health Monitoring. Big Data and Cognitive Computing. 2025; 9(6):158. https://doi.org/10.3390/bdcc9060158

Chicago/Turabian Style

Wang, Nian, Yue Chen, Weiang Li, Liyang Zhang, and Jinghong Tian. 2025. "FAD-Net: Automated Framework for Steel Surface Defect Detection in Urban Infrastructure Health Monitoring" Big Data and Cognitive Computing 9, no. 6: 158. https://doi.org/10.3390/bdcc9060158

APA Style

Wang, N., Chen, Y., Li, W., Zhang, L., & Tian, J. (2025). FAD-Net: Automated Framework for Steel Surface Defect Detection in Urban Infrastructure Health Monitoring. Big Data and Cognitive Computing, 9(6), 158. https://doi.org/10.3390/bdcc9060158

Article Metrics

Back to TopTop