Next Article in Journal
ZnO Thin Films as Promising Corrosion Protection on Mg-Based Alloys
Previous Article in Journal
Numerical and Experimental Investigation of Parameters in Cement Delivery Through Spinal Implants
Previous Article in Special Issue
Evaluation of the Effect of Post-Processing Methods on the Surface Parameters of Parts Produced by FFF/FDM Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Cross-Scale Spatial–Semantic Feature Aggregation Network for Strip Steel Surface Defect Detection

School of Computer and Information Technology, Xinyang Normal University, Xinyang 464000, China
*
Authors to whom correspondence should be addressed.
Materials 2025, 18(24), 5567; https://doi.org/10.3390/ma18245567
Submission received: 10 November 2025 / Revised: 1 December 2025 / Accepted: 5 December 2025 / Published: 11 December 2025

Abstract

Strip steel surface defect detection remains a challenging task due to the diverse scales and uneven spatial distribution of defects, which often lead to incomplete feature representation and missed detections in sparsely distributed regions. To address these challenges, we propose a novel cross-scale spatial–semantic feature aggregation network (CSSFAN) that achieves fine-grained and semantically consistent feature fusion across multiple scales. Specifically, CSSFAN adopts a bottom-up feature aggregation strategy equipped with a series of cross-scale spatial–semantic aggregation modules (CSSAMs). Each CSSAM first establishes a mapping relationship between high-level feature points and low-level feature regions and then introduces a cross-scale attention mechanism that adaptively injects spatial details from low-level features into high-level semantic representations. This aggregation strategy bridges the gap between spatial precision and semantic abstraction, enabling the network to capture subtle and irregular defect patterns. Furthermore, we introduce an adaptive region proposal network (ARPN) to cope with the uneven spatial distribution of defects. ARPN dynamically adjusts the number of region proposals according to the local feature complexity, ensuring that regions with dense or subtle defects receive more proposal attention, while sparse or background regions are adaptively suppressed, thereby enhancing the model’s sensitivity to defect-prone areas. Extensive experiments on two strip steel surface defect datasets demonstrate that our method significantly improves detection performance, validating its effectiveness and robustness.

1. Introduction

Strip steel serves as a fundamental material in the steel industry and is widely utilized in automotive manufacturing, construction, household appliances, and energy sectors, playing a critical role in industrial production and national economic development [1,2,3]. However, during production and storage, the surface of strip steel inevitably suffers from various defects caused by factors such as rolling, cooling, transportation, and environmental corrosion. Typical defects include scratches, pits, folds, spots, and cracks [4,5,6]. These defects not only degrade the surface quality and mechanical properties of strip steel but also can cause fracture, stress concentration, or even product scrapping in subsequent processing, ultimately reducing production efficiency and causing economic losses. Therefore, accurate detection of strip steel surface defects holds significant theoretical importance and practical value, particularly for ensuring product quality [7,8,9].
Early approaches to strip steel surface defect detection primarily relied on handcrafted features and conventional image processing techniques, including gray-level statistics, edge detection, and texture modeling [10,11,12]. While these methods are relatively simple to implement and require limited computational resources, they suffer from several inherent limitations [13]. Specifically, their effectiveness is highly dependent on expert prior knowledge and carefully designed feature descriptors, which restricts adaptability to new scenarios. Moreover, due to their shallow representational capacity, these approaches struggle to capture the complex appearance variations of defects caused by changes in illumination, surface roughness, and background noise [14]. As a result, they exhibit poor robustness and limited generalization capability, often failing to deliver reliable detection performance in real-world industrial environments where defect types are diverse and imaging conditions are unconstrained [15].
In recent years, deep learning techniques have achieved significant progress in the field of computer vision and have been increasingly applied to strip steel surface defect detection tasks [16,17,18]. Convolutional neural network (CNN)-based detection methods can automatically learn hierarchical feature representations, effectively overcoming the limitations of traditional handcrafted features [19]. To further improve the performance of CNN-based methods, researchers have proposed various improvement strategies. For instance, Sohag et al. [20] proposed an improved Faster R-CNN by integrating the Swin Transformer with a path aggregation feature pyramid network (PAFPN) [21], strengthening the model’s capability to capture complex defect. Han et al. [22] developed a scale-aware feature pyramid network that leverages an attention mechanism to reduce semantic discrepancies across pyramid levels, improving multiscale feature consistency. Hou et al. [23] designed a spatial attention encoder that establishes long-range dependencies to capture global contextual cues, enhancing the perception of defects. Similarly, Lu et al. [24] designed an anchor-free detector based on the AutoAssign framework, which enhances the semantic representation of defects while suppressing background interference.
Despite the remarkable progress achieved by the aforementioned CNN-based methods, challenges remain when dealing with defects of varying scales and irregular distributions, which often lead to incomplete feature representation and inaccurate localization in complex industrial scenarios. As shown in Figure 1, STD2 [20] generates a large number of redundant bounding boxes when handling defects with diverse shapes and textures, making it difficult to accurately distinguish adjacent regions. SA-FPN [22], limited by its ability to discriminate complex background features, tends to misclassify true defect regions as background, resulting in missed detections. CANet [23] struggles to differentiate small-scale defects from background textures, often misidentifying background patterns as defects, which leads to false positives. In addition, CA-Auoassign [24] excessively focuses on background noise, which introduces false detections.
To address the aforementioned challenges, we propose a novel cross-scale spatial–semantic feature aggregation network (CSSFAN) for steel surface defect detection. Specifically, CSSFAN adopts a bottom-up feature aggregation strategy equipped with a series of cross-scale spatial–semantic aggregation modules (CSSAMs). Each CSSAM first establishes a spatial correspondence between high-level feature points and low-level feature regions, and then introduces a cross-scale attention mechanism to adaptively inject spatial details from low-level features into high-level semantic representations. This design bridges the gap between spatial precision and semantic abstraction, enabling the network to capture subtle and irregular defect patterns that are often overlooked by conventional multiscale fusion methods. Furthermore, we introduce an adaptive region proposal network (ARPN) to handle the uneven spatial distribution of defects. Unlike traditional RPNs that use a fixed number of anchors across the entire image, ARPN dynamically adjusts the number of region proposals based on local feature complexity.
In summary, the main contributions of this paper are as follows:
  • We design a novel CSSFAN that adopts a bottom-up pyramid-aware feature aggregation strategy combined with CSSAMs, enhancing the spatial representation of high-level features while maintaining semantic consistency.
  • We develop an ARPN that dynamically adjusts the number and spatial density of proposals according to local feature complexity, addressing the uneven spatial distribution of defects.
  • Extensive experiments conducted on two benchmark datasets for strip steel surface defect detection demonstrate that the proposed method consistently outperforms state-of-the-art approaches in terms of detection accuracy and robustness.
The structure of this paper is as follows: Section 2 reviews related works; Section 3 details the components of our method; Section 4 presents the experimental evaluation of the proposed method; and Section 5 concludes the paper.

2. Related Work

2.1. Strip Steel Surface Defect Detection

Many studies have proposed solutions for strip steel surface defect detection, categorized into traditional and deep-learning methods [25]. Traditional detection methods for strip steel surface defects mainly include manual inspection [26], magnetic flux leakage detection [27], and infrared detection [28]. Manual inspection relies heavily on human vision and experience, which is time-consuming, labor-intensive, and prone to subjective errors. Magnetic flux leakage detection analyzes disturbances in the magnetic field to reveal hidden or surface defects, but it often requires strict testing conditions and suffers from limited sensitivity to small flaws. Infrared detection identifies defects by capturing temperature variations caused by surface irregularities, yet its accuracy can be easily influenced by environmental factors such as heat radiation and ambient temperature. Overall, while these traditional approaches provide valuable tools for defect detection, their limited efficiency, robustness, and adaptability to complex industrial environments restrict their widespread application [29].
Deep learning-based methods have been widely applied in strip steel surface defect detection, demonstrating superior performance compared to traditional approaches [30]. For instance, Shi et al. [31] employed ConvNeXt [32] as the backbone of Faster-RCNN and incorporated the CBAM module [33] to enhance feature representation, improving detection accuracy. Chen et al. [34] designed a defect detection network that integrates deformable convolution with coordinate attention, enabling the model to capture geometric transformations and long-range dependencies for more precise and efficient detection. Huang et al. [35] proposed a lightweight method based on YOLOv8n, where the GhostNetv2 [36] module was introduced to enhance feature extraction capability while maintaining computational efficiency. In addition, Shi et al. [31] used ConvNeXt as the feature extractor for Faster-RCNN and applied the CBAM to enhance the model’s feature expression capability.

2.2. Multiscale Feature Fusion

Multiscale feature fusion has been widely applied in steel surface defect detection, aiming to integrate information across different resolutions so that models can simultaneously capture fine-grained local structures and high-level semantic context [37]. feature pyramid networks (FPN) [38] achieve this by constructing a top-down architecture that progressively fuses semantic-rich high-level features with detail-preserving low-level features, significantly enhancing multiscale object detection performance. Building upon FPN, PAFPN [21] introduces a bottom-up path to strengthen cross-scale feature propagation, thereby improving localization precision and robustness. To further optimize feature fusion strategies, neural architecture search FPN (NAS-FPN) [39] employs automated neural architecture search to explore the design space, generating fusion schemes that outperform manually designed pyramids. In addition, bidirectional FPN (BiFPN) [40] introduces efficient bidirectional cross-scale connections with learnable weights, enabling adaptive balancing of feature importance across levels and achieving a better trade-off between accuracy and computational efficiency. Overall, these studies demonstrate that effective multiscale feature fusion plays a critical role in enhancing the robustness and adaptability of modern object detection frameworks.

3. Method

3.1. Overall

We propose a novel strip steel surface defect detector, as illustrated in Figure 2. Our method consists of four main components: (1) the backbone network, (2) the cross-scale spatial–semantic feature aggregation network (CSSFAN), (3) the adaptive region proposal network (ARPN), and (4) the detection head.
In the feature extraction stage, the backbone network (i.e., ResNet) [41] is employed to generate hierarchical multiscale feature representations, providing a solid foundation for subsequent feature fusion and region proposal generation. To further enhance multiscale representation, we adopt a bottom-up fusion strategy and integrate cross-scale spatial–semantic aggregation modules (CSSAMs). Specifically, CSSAM establishes cross-attention mappings between high-level semantic features and their corresponding low-level feature regions, enabling bidirectional information exchange across scales. This mechanism aims to preserve semantic consistency and inject fine-grained spatial details into high-level feature maps, improving the discriminative power of fused representations. In the proposal generation stage, the adaptive region proposal network (ARPN) is introduced to address the limitations of conventional RPNs. Unlike fixed-anchor strategies, ARPN dynamically adjusts the number and distribution of anchors based on local feature complexity. This adaptive mechanism effectively reduces redundant proposals in relatively simple regions while increasing proposal density in complex regions, ensuring more balanced and efficient coverage across varying difficulty levels. Finally, the detection head refines the proposals through classification and regression, achieving fine-grained defect localization and recognition.
The primary contributions of this work lie in the proposed CSSAM and ARPN, which are elaborated in Section 3.2 and Section 3.3, respectively.

3.2. Cross-Scale Spatial–Semantic Aggregation Modules (CSSAM)

As illustrated in Figure 3, we propose CSSAM to facilitate spatial information transfer between high-level semantic feature points and their corresponding low-level feature regions. This mechanism adaptively injects spatial details from low-level features into high-level semantic representations, enhancing spatial precision while maintaining semantic consistency. This design enables for more accurate defect location in complex industrial scenarios. Let the low-level feature map be denoted as f i R C × H × W and the high-level semantic feature map as f i + 1 R C × H × W . The CSSAM operates through three sequential stages: (1) query construction and key–value extraction; (2) cross-scale attention computation; (3) feature reconstruction and aggregation.

3.2.1. Query Construction and Key-Value Extraction

To enable efficient cross-scale feature interaction, we first sample query tokens from the high-level feature map and extract key–value pairs from the corresponding low-level feature map. This design allows the network to establish a bidirectional correspondence between fine-grained spatial information and high-level semantic representations, promoting effective multiscale information exchange.
Specifically, for a high-level feature map f i + 1 containing rich semantic cues but limited spatial resolution, we treat each spatial position as an independent query token, forming:
Q i + 1 = U k , s ( f i + 1 ) R ( k 2 C ) × L .
where C denotes the channel dimension and L = H × W represents the total number of spatial positions. Each query token encapsulates semantic information at a specific location, serving as an anchor for cross-scale interaction.
In parallel, the corresponding low-level feature map f i , which retains fine-grained texture and boundary information, is processed using an unfolding operator U k , s ( · ) with kernel size k × k and stride s to partition it into a series of non-overlapping local patches:
K i = V i = U k , s ( f i ) R L × C ,
Each patch preserves detailed local structural information and serves as a spatial reference during attention computation, while the transposition ( · ) aligns spatial and channel dimensions for efficient matrix operations. By aligning the high-level queries Q i + 1 with the low-level keys and values ( K i , V i ) , the subsequent cross-attention mechanism adaptively transfers spatial details from the low-level domain to enhance high-level semantic features. This process effectively bridges the gap between spatial precision and semantic abstraction, enabling the network to accurately locate subtle and irregular defect patterns across multiple scales.

3.2.2. Cross-Attention Computation

We compute cross-scale attention using the constructed query, key, and value matrices to enable semantic propagation from high-resolution to low-resolution features. The cross-scale attention is formally defined as:
A i + 1 = Softmax Q i + 1 ( K i ) d V i ,
where d denotes the embedding dimension used for normalization. The dot-product operation measures the similarity between each query and key, while the Softmax function converts this similarity map into a probabilistic attention distribution. Through this adaptive weighting mechanism, the network selectively aggregates high-resolution spatial information that is most relevant to each low-resolution position. Consequently, high-level features are enriched with fine-grained spatial cues, whereas low-level features gain enhanced semantic coherence. This process achieves a balanced representation between spatial precision and semantic abstraction, benefiting both object localization and semantic recognition.

3.2.3. Feature Reconstruction and Fusion

We reconstruct and fuse the attention-enhanced representation A i + 1 with the original high-level feature f i + 1 to obtain a unified multiscale representation. The attention output A i + 1 carries spatially detailed information propagated from low-level features, serving as an effective modulation signal to refine the high-level semantics.
To integrate these complementary cues, A i + 1 is first reshaped to match the spatial structure of the original feature map and then fused via element-wise multiplication:
f ^ i + 1 = Reshape ( A i + 1 ) · f i + 1 .
This operation injects fine-grained spatial information into the semantic representation, producing an enhanced feature map f ^ i + 1 that preserves both spatial fidelity and semantic richness.
In contrast to conventional feature addition or concatenation, the proposed multiplicative fusion adaptively modulates each spatial position according to its contextual relevance. This design facilitates more precise structural alignment and promotes smoother optimization during cross-scale fusion, ultimately yielding a more discriminative and robust representation for downstream detection and recognition tasks.

3.3. Adaptive Region Proposal Network (ARPN)

As shown in Figure 4a, conventional RPN generates a fixed number of anchors at each spatial location of the feature map, regardless of the local visual complexity. This design inevitably leads to redundant anchors in simple regions and insufficient coverage in complex regions with rich structures. To address this limitation, we propose an ARPN (Figure 4b) that dynamically adjusts the anchor density according to the local feature complexity, thereby improving proposal quality and reducing computational redundancy.

3.3.1. Density Score Estimation

Given an input feature map f ^ i R C × H × W , a lightweight convolutional predictor is designed to estimate a density score map  D ( 0 , 1 ) H × W , which quantifies the structural complexity of each spatial region. Intuitively, areas containing rich texture details or multiple defect cues are more likely to require denser anchor sampling for precise localization. Each element D i represents the local structural complexity at spatial position i, where a higher value implies a more intricate region that requires a finer anchor coverage. The density score is computed as:
D i = σ W 2 · δ ( W 1 f ^ i ) ,
where ∗ denotes convolution, δ ( · ) is the ReLU activation, and σ ( · ) is the Sigmoid function that normalizes responses into ( 0 , 1 ) . The learnable parameters W 1 and W 2 are jointly optimized with the entire detection framework in an end-to-end manner, enabling the model to automatically learn an adaptive density representation consistent with the spatial distribution of surface defects.

3.3.2. Adaptive Anchor Reweighting

During candidate proposal generation, the estimated density score D i is integrated with the classification confidence score s i predicted by the RPN head to produce an adaptive confidence score
s ˜ i = s i · D i .
This reweighting mechanism adaptively balances the importance of each anchor according to the underlying spatial complexity. Specifically, anchors located in regions with high density scores, which typically contain fine-grained textures or multiple small-scale defects, are assigned larger weights to ensure adequate proposal generation and reduce the risk of missed detections. In contrast, anchors in smooth or homogeneous regions are down-weighted, suppressing redundant or low-quality proposals. Through this dynamic reweighting strategy, the proposed method achieves an improved trade-off between recall and precision, enabling the RPN to generate more informative and spatially balanced proposals that subsequently enhance feature refinement and detection accuracy.

3.4. Loss Function

Our method is optimized using a composite loss function consisting of a classification term [42] and a bounding box regression term [43], expressed as
L = L cls + L box .
Here, L cls evaluates the discrepancy between predicted class probabilities and ground-truth labels, and is formulated as
L cls = g log ( p ) ( 1 g ) log ( 1 p ) ,
where p denotes the predicted confidence score and g { 0 , 1 } represents the ground-truth class indicator. The second term, L box , enforces accurate localization by penalizing the deviation between the predicted bounding box and its ground truth:
L box = j x , y , w , h L 1 t j t ^ j ,
where t j = ( t x , t y , t w , t h ) indicates the ground-truth box and t ^ j = ( t ^ x , t ^ y , t ^ w , t ^ h ) denotes the predicted box.

4. Experiments

4.1. Experimental Setup

4.1.1. Implementation Details

Our method is implemented based on the MMDetection framework [44]. All experiments are conducted on a computing platform running Ubuntu 20.04 with a single NVIDIA A100 GPU (80 GB memory). The software environment includes Python 3.9, PyTorch 1.13.1, and CUDA 11.7. Our code is available at https://github.com/hpguo1982/CSSFAN (accessed on 11 November 2025). Detailed hyperparameter settings are listed in Table 1.

4.1.2. Datasets

To validate the effectiveness of the proposed method, we conduct extensive experiments on two challenging steel surface defect datasets: NEU-DET [45,46,47] and GC10-DET [48].
NEU-DET Dataset: The NEU-DET dataset, released by Northeastern University, is one of the most representative benchmark datasets for industrial surface defect detection. It consists of 1800 grayscale images collected from hot-rolled steel strips, covering six typical types of surface defects commonly encountered in industrial production: scratch (Sc), patch (Pa), inclusion (In), rolled-in scale (Rs), pitted surface (Ps), and crazing (Cr). Each category contains 300 images with a resolution of 200 × 200 pixels. Due to its balanced distribution across defect types and its high relevance to real-world inspection scenarios, NEU-DET has become a widely adopted dataset for evaluating the robustness, recognition accuracy, and generalization capability of defect detection models.
GC10-DET Dataset: The GC10-DET dataset was collected from real-world hot-rolled steel production lines and contains 2294 images with varying resolutions. It covers ten representative categories of surface defects, namely punch (Pu), welding line (Wl), crescent gap (Cg), water spot (Ws), oil spot (Os), silk spot (Ss), inclusion (In), roll pit (Rp), crease (Cr), and waist fold (Wf). Each category contains dozens to hundreds of samples, exhibiting a wide range of variations in scale, shape, texture, and intensity. Compared with NEU-DET, GC10-DET is more challenging due to its higher intra-class diversity, complex backgrounds, and inter-class similarities. As such, it provides a more realistic benchmark for evaluating the robustness, adaptability, and generalization ability of defect detection algorithms in industrial inspection tasks.

4.1.3. Evaluation Metrics

We evaluate the detection performance following the COCO evaluation protocol [49]. Specifically, we report AP 50 (IoU threshold = 0.5), AP 75 (IoU threshold = 0.75), and the overall AP averaged over IoU thresholds from 0.5 to 0.95. Furthermore, scale-aware metrics are also included, i.e., AP S , AP M , and AP L , which correspond to objects of small ( area < 32 2 ), medium ( 32 2 area < 96 2 ), and large sizes ( area 96 2 ), respectively.

4.2. Quantitative Comparison

We compare our method with fifteen state-of-the-art methods on the NEU-DET and GC10-DET datasets, including SSD300 [50], TOOD [51], Faster-RCNN [52], DH-RCNN [53], Cascade-RCNN [54], Dynamic-RCNN [55], Grid-RCNN [56], Libra-RCNN [57], Sparse-RCNN [58], YOLOv9 [59], YOLOv10 [60], DETR [61], RT-DETR [62], CA-Autoassign [24] and STD2 [20], the results are shown in Table 2 and Table 3.

4.2.1. Quantitative Comparison on NEU-DET Dataset

As reported in Table 2, our proposed method demonstrates superior performance on the NEU-DET dataset across multiple evaluation metrics. Specifically, it attains the highest overall AP of 45.1%, along with the best scores in AP50 (81.1%), AP75 (46.5%), APS (43.3%), and APL (56.9%), indicating its strong capability in achieving both coarse and fine-grained localization for small and large defects. Although its APM ranks third, slightly below Cascade-RCNN and DETR, the performance gap is marginal, showing the competitive balance of CA-RCNN across object scales. In terms of category-level evaluation, our method achieves the highest AP values on four representative defect types, namely Cr (53.6%), In (86.8%), Pa (95.1%), and Ps (93.2%). For the remaining Rs and Sc categories, our method also delivers competitive results, ranking closely to the best-performing methods.

4.2.2. Quantitative Comparison on GC10-DET Dataset

As shown in Table 3, our method demonstrates consistently superior performance on the GC10-DET dataset. our method achieves the highest scores in overall AP (35.9%), AP50 (72.3%), APS (14.1%), and APL (40.1%), highlighting its effectiveness in accurately detecting defects across varying object sizes, particularly small-scale and large-scale defects. In addition, it delivers competitive results on AP75 and APM, further illustrating the robustness of the proposed framework across different IoU thresholds and object scales. At the category level, our method outperforms other approaches in several representative defect types, achieving the highest AP values for Pu (97.4%), Wl (98.3%), Ss (67.8%), In (43.7%), Rp (50.6%), and Cr (56.5%). Moreover, our method maintains strong competitiveness on other categories, such as Ws (78.9%) and Wf (72.8%), which involve challenging backgrounds and complex textures.

4.3. Visual Comparison

We present the visualization results of six methods, i.e., Tood [51], Faster-RCNN [52], Cascade-RCNN [54], RT-DETR [62], STD2 [20], and our method, on the NEU-DET and GC10-DET datasets, as illustrated in Figure 5 and Figure 6, respectively.

4.3.1. Visual Comparison on NEU-DET

As shown in Figure 5, the proposed method (column 2) exhibits the most favorable detection performance on the NEU-DET dataset, with results closely aligned with the ground truth. Specifically, for low-contrast defects (row 1), Faster-RCNN and Cascade-RCNN erroneously classify background regions as defect objects, while TOOD and RT-DETR fail to detect certain instances. For point-like defects (row 2), RT-DETR and STD2 both suffer from false detections. In cases of elongated or small defects (row 3), Faster-RCNN generates redundant bounding boxes, whereas Cascade-RCNN fails to detect some defect instances. For defects with diverse shapes (row 4), Faster-RCNN and Cascade-RCNN do not accurately localize the targets. In contrast, the proposed method effectively suppresses background noise, mitigates misidentification, and demonstrates superior precision and robustness across various defect types.

4.3.2. Visual Comparison on GC10-DET

As illustrated in Figure 6, we present the visual comparison between our method and other state-of-the-art detectors on the GC10-DET dataset. From Figure 6, it can be observed that in cases of multiple defect detection (rows 1 and 2), TOOD suffers from false negatives (i.e., misclassifying defects as background), while Faster-RCNN and Cascade-RCNN exhibit false positives (i.e., misclassifying background as defects). For low-contrast defects (rows 3 and 4), TOOD, Faster-RCNN, and Cascade-RCNN all produce false positives, whereas RT-DETR and STD2 show false negatives. In the last row of defects, both Faster-RCNN and Cascade-RCNN generate false positives. In contrast, our method effectively suppresses these errors and achieves precise localization. Furthermore, although all methods can successfully detect defects in the last two rows, our method demonstrates superior stability and a clear advantage when addressing low-contrast and complex defect scenarios.

4.4. Ablation Experiments

To evaluate the effectiveness of each component in our method, we progressively incorporate CSSFAN, and ARPN into the baseline model, the ablation results are shown in Table 4.
From Table 4, the introduction of CSSFAN improves AP from 43.4% to 44.5%, with consistent gains observed in AP50 and AP75, indicating that CSSFAN effectively enhances multiscale feature representation. Incorporating ARPN alone yields an AP of 43.9%, and further contributes to the detection of small and large objects, as reflected by improvements in APS (42.6%) and APL (53.6%). When both CSSFAN and ARPN are simultaneously integrated, the proposed framework achieves the highest overall performance with an AP of 45.1%, along with consistent improvements across all evaluation metrics, including AP50, AP75, and object scales. These results confirm the complementary nature of CSSFAN and ARPN, and validate the effectiveness of the proposed method.

4.5. Effectiveness of CSSFAN

To further validate the effectiveness of CSSFAN, we conduct comparative experiments with five mainstream feature fusion networks (FPN [38], NAS-FPN [39], PAFPN [21], BiFPN [40], and AFPN [63]) on the NEU-DET dataset. The comparison results are shown in Table 5.
From Table 5, our CSSFAN achieves the best performance across almost all evaluation metrics. Specifically, our method attains the highest AP of 45.1%, outperforming the second-best PAFPN (44.0%) by 1.1 percentage points. In terms of AP50 and AP75, CSSFAN reaches 81.1% and 46.5%, respectively, indicating stronger localization accuracy and robustness. Furthermore, CSSFAN demonstrates consistent improvements across different object scales, achieving the best APS, and APL scores of 43.3%, and 56.9%, respectively. These results clearly verify that the proposed cross-scale and semantic feature aggregation strategy in CSSFAN effectively enhances multiscale feature representation and detection precision compared to conventional and state-of-the-art feature fusion architectures.

4.6. Generalization Experiments

To further validate the effectiveness of the proposed method, we conduct generalization experiments on the resistance spot welding defect (RSW-D) dataset [64]. The RSW-D dataset comprises 4134 images, covering seven defect categories: Normal (No), Edge (Ed), Copper-adhesion (Ca), Overlap (Ov), Mutilation (Mu), Splash (Sp), and Twist (Tw). Table 6 shows the corresponding results on the RSW-D datasets.
As shown in Table 6, the proposed method consistently surpasses all compared methods in terms of AP, AP50, AP75, APM, and APL, achieving the highest scores of 70.9%, 97.2%, 87.6%, 67.7%, and 71.8%, respectively, while ranking second on APS. At the category level, our method achieves the best AP50 results in Wl (99.0%), Cg (98.0%), Os (97.6%), Ss (96.6%), and In (97.0%), and obtains comparable performance in Pu and Ws. Overall, these results clearly verify that our method achieves superior generalization and detection stability across both challenging defect scales and heterogeneous categories, further validating its effectiveness and robustness across diverse defect types.

5. Conclusions

In this paper, we proposed a novel Cross-Scale Spatial–Semantic Feature Aggregation Network (CSSFAN) for steel surface defect detection, addressing the challenges posed by diverse defect scales and uneven spatial distributions. The proposed network integrates multi-level features through a bottom-up aggregation strategy, where the Cross-Scale Spatial–Semantic Aggregation Module (CSSAM) adaptively fuses low-level spatial details with high-level semantic information, enhancing feature completeness and strengthening the representation of subtle and irregular defect patterns. To further alleviate the problem of spatial imbalance, an Adaptive Region Proposal Network (ARPN) was introduced to dynamically adjust the number and distribution of proposals according to local feature complexity, allowing the network to focus on defect-prone regions while suppressing redundant proposals in homogeneous areas. Extensive experiments conducted on two benchmark strip steel defect datasets demonstrate that CSSFAN consistently outperforms existing detection methods in both accuracy and robustness. These results confirm the effectiveness of the proposed cross-scale aggregation and adaptive proposal mechanisms, offering a promising direction for high-precision industrial surface defect inspection.

Author Contributions

C.X.: Methodology, Writing—original draft, Software, Formal analysis, Resources, Writing—review & editing; Y.S.: Conceptualization, Writing—review & editing; L.H.: Writing—review & editing, Resources; H.G.: Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded in part by Henan Provincial Science and Technology Program under Grant 241111212200 and 252102220046, in part by Henan Joint Fund for Science and Technology Research under Grant 20240012, in part by Key Scientific Research Projects of Higher Education Institutions in Henan Province under Grants 26A520036 and 26A520037, and in part by Henan Key Laboratory of Education Big Data Analysis and Application under Grant 2025JYDSJ01.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, H.; Hu, R.; Dong, H.; Liu, Z. SFC-YOLOv8: Enhanced Strip Steel Surface Defect Detection Using Spatial-Frequency Domain-Optimized YOLOv8. IEEE Trans. Instrum. Meas. 2025, 74, 9700111. [Google Scholar] [CrossRef]
  2. Li, Z.; Wei, X.; Hassaballah, M.; Li, Y.; Jiang, X. A deep learning model for steel surface defect detection. Complex Intell. Syst. 2024, 10, 885–897. [Google Scholar] [CrossRef]
  3. Wang, H.; Li, W.; Zhang, B.; Gu, Z. YOLOv8n-GSE: Efficient Steel Surface Defect Detection Method. IEEE Access 2025, 13, 166343–166356. [Google Scholar] [CrossRef]
  4. Yu, F.; Zhang, J.; Mu, D. Steel Defect Detection Based on YOLO-SAFD. IEEE Access 2025, 13, 77291–77304. [Google Scholar] [CrossRef]
  5. Yeung, C.C.; Lam, K.M. Efficient fused-attention model for steel surface defect detection. IEEE Trans. Instrum. Meas. 2022, 71, 2510011. [Google Scholar] [CrossRef]
  6. Cheng, Z.; Gao, L.; Wang, Y.; Deng, Z.; Tao, Y. EC-YOLO: Effectual Detection Model for Steel Strip Surface Defects Based on YOLO-V5. IEEE Access 2024, 12, 62765–62778. [Google Scholar] [CrossRef]
  7. Li, C.; Xu, A.; Zhang, Q.; Cai, Y. Steel Surface Defect Detection Method Based on Improved YOLOX. IEEE Access 2024, 12, 37643–37652. [Google Scholar] [CrossRef]
  8. Liang, C.; Wang, Z.Z.; Liu, X.L.; Zhang, P.; Tian, Z.W.; Qian, R.L. SDD-Net: A Steel Surface Defect Detection Method Based on Contextual Enhancement and Multiscale Feature Fusion. IEEE Access 2024, 12, 185740–185756. [Google Scholar] [CrossRef]
  9. Song, X.; Cao, S.; Zhang, J.; Hou, Z. Steel surface defect detection algorithm based on YOLOv8. Electronics 2024, 13, 988. [Google Scholar] [CrossRef]
  10. Tang, B.; Chen, L.; Sun, W.; Lin, Z.k. Review of surface defect detection of steel products based on machine vision. IET Image Process. 2023, 17, 303–322. [Google Scholar] [CrossRef]
  11. Liu, H.; Chen, C.; Hu, R.; Bin, J.; Dong, H.; Liu, Z. CGTD-net: Channel-wise global transformer-based dual-branch network for industrial strip steel surface defect detection. IEEE Sens. J. 2024, 24, 4863–4873. [Google Scholar] [CrossRef]
  12. Wen, X.; Shan, J.; He, Y.; Song, K. Steel surface defect recognition: A survey. Coatings 2022, 13, 17. [Google Scholar] [CrossRef]
  13. Gao, S.; Chu, M.; Zhang, L. A detection network for small defects of steel surface based on YOLOv7. Digit. Signal Process. 2024, 149, 104484. [Google Scholar] [CrossRef]
  14. Lu, J.; Yu, M.; Liu, J. Lightweight strip steel defect detection algorithm based on improved YOLOv7. Sci. Rep. 2024, 14, 13267. [Google Scholar] [CrossRef]
  15. Chen, Y.; Ding, Y.; Zhao, F.; Zhang, E.; Wu, Z.; Shao, L. Surface defect detection methods for industrial products: A review. Appl. Sci. 2021, 11, 7657. [Google Scholar] [CrossRef]
  16. Tian, R.; Jia, M. DCC-CenterNet: A rapid detection method for steel surface defects. Measurement 2022, 187, 110211. [Google Scholar] [CrossRef]
  17. Ashrafi, S.; Teymouri, S.; Etaati, S.; Khoramdel, J.; Borhani, Y.; Najafi, E. Steel surface defect detection and segmentation using deep neural networks. Results Eng. 2025, 25, 103972. [Google Scholar] [CrossRef]
  18. Yuan, Z.; Ning, H.; Tang, X.; Yang, Z. GDCP-YOLO: Enhancing steel surface defect detection using lightweight machine learning approach. Electronics 2024, 13, 1388. [Google Scholar] [CrossRef]
  19. Singh, S.A.; Desai, K.A. Automated surface defect detection framework using machine vision and convolutional neural networks. J. Intell. Manuf. 2023, 34, 1995–2011. [Google Scholar] [CrossRef]
  20. Sohag Mia, M.; Li, C. STD2: Swin Transformer-Based Defect Detector for Surface Anomaly Detection. IEEE Trans. Instrum. Meas. 2025, 74, 3492728. [Google Scholar] [CrossRef]
  21. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  22. Han, L.; Li, N.; Li, J.; Gao, B.; Niu, D. SA-FPN: Scale-aware attention-guided feature pyramid network for small object detection on surface defect detection of steel strips. Measurement 2025, 249, 117019. [Google Scholar] [CrossRef]
  23. Hou, X.; Liu, M.; Zhang, S.; Wei, P.; Chen, B. CANet: Contextual information and spatial attention based network for detecting small defects in manufacturing industry. Pattern Recognit. 2023, 140, 109558. [Google Scholar] [CrossRef]
  24. Lu, H.; Fang, M.; Qiu, Y.; Xu, W. An anchor-free defect detector for complex background based on pixelwise adaptive multiscale feature fusion. IEEE Trans. Instrum. Meas. 2022, 72, 5002312. [Google Scholar] [CrossRef]
  25. Jiang, X.; Cui, Y.; Cui, Y.; Xu, R.; Yang, J.; Zhou, J. Optimization algorithm of steel surface defect detection based on YOLOv8n-SDEC. IEEE Access 2024, 12, 95106–95117. [Google Scholar] [CrossRef]
  26. Soukup, D.; Huber-Mörk, R. Convolutional neural networks for steel surface defect detection from photometric stereo images. In Advances in Visual Computing; Springer: Cham, Switzerland, 2014; pp. 668–677. [Google Scholar]
  27. Sophian, A.; Tian, G.Y.; Zairi, S. Pulsed magnetic flux leakage techniques for crack detection and characterisation. Sens. Actuators A Phys. 2006, 125, 186–191. [Google Scholar] [CrossRef]
  28. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional local contrast networks for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  29. Xie, W.; Ma, W.; Sun, X. An efficient re-parameterization feature pyramid network on YOLOv8 to the detection of steel surface defect. Neurocomputing 2025, 614, 128775. [Google Scholar] [CrossRef]
  30. Zhao, B.; Chen, Y.; Jia, X.; Ma, T. Steel surface defect detection algorithm in complex background scenarios. Measurement 2024, 237, 115189. [Google Scholar] [CrossRef]
  31. Shi, X.; Zhou, S.; Tai, Y.; Wang, J.; Wu, S.; Liu, J.; Xu, K.; Peng, T.; Zhang, Z. An improved faster R-CNN for steel surface defect detection. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), Shanghai, China, 26–28 September 2022; pp. 1–5. [Google Scholar]
  32. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  33. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  34. Chen, H.; Du, Y.; Fu, Y.; Zhu, J.; Zeng, H. DCAM-Net: A rapid detection network for strip steel surface defects based on deformable convolution and attention mechanism. IEEE Trans. Instrum. Meas. 2023, 72, 5005312. [Google Scholar] [CrossRef]
  35. Huang, M.; Cai, Z. Steel surface defect detection based on improved YOLOv8. In Proceedings of the International Conference on Algorithms, High Performance Computing, and Artificial Intelligence (AHPCAI 2023), Yinchuan, China, 18–19 August 2023; SPIE: Bellingham, WA, USA, 2023; Volume 12941, pp. 1356–1360. [Google Scholar]
  36. Tang, Y.; Han, K.; Guo, J.; Xu, C.; Xu, C.; Wang, Y. GhostNetv2: Enhance cheap operation with long-range attention. Adv. Neural Inf. Process. Syst. 2022, 35, 9969–9982. [Google Scholar]
  37. Liu, G.; Chu, M.; Gong, R.; Zheng, Z. Global attention module and cascade fusion network for steel surface defect detection. Pattern Recognit. 2025, 158, 110979. [Google Scholar] [CrossRef]
  38. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  39. Ghiasi, G.; Lin, T.Y.; Le, Q.V. NAS-FPN: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7036–7045. [Google Scholar]
  40. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 23803–23828. [Google Scholar]
  43. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 516–520. [Google Scholar]
  44. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar] [CrossRef]
  45. Bao, Y.; Song, K.; Liu, J.; Wang, Y.; Yan, Y.; Yu, H.; Li, X. Triplet-graph reasoning network for few-shot metal generic surface defect segmentation. IEEE Trans. Instrum. Meas. 2021, 70, 5011111. [Google Scholar] [CrossRef]
  46. Song, K.; Yan, Y. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
  47. He, Y.; Song, K.; Meng, Q.; Yan, Y. An end-to-end steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans. Instrum. Meas. 2019, 69, 1493–1504. [Google Scholar] [CrossRef]
  48. Lv, X.; Duan, F.; Jiang, J.j.; Fu, X.; Gan, L. Deep metallic surface defect detection: The new benchmark and detection network. Sensors 2020, 20, 1562. [Google Scholar] [CrossRef] [PubMed]
  49. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  50. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  51. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. Tood: Task-aligned one-stage object detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3490–3499. [Google Scholar]
  52. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  53. Wu, Y.; Chen, Y.; Yuan, L.; Liu, Z.; Wang, L.; Li, H.; Fu, Y. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10186–10195. [Google Scholar]
  54. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  55. Zhang, H.; Chang, H.; Ma, B.; Wang, N.; Chen, X. Dynamic R-CNN: Towards high quality object detection via dynamic training. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XV 16. Springer: Cham, Switzerland, 2020; pp. 260–275. [Google Scholar]
  56. Lu, X.; Li, B.; Yue, Y.; Li, Q.; Yan, J. Grid R-CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7363–7372. [Google Scholar]
  57. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 821–830. [Google Scholar]
  58. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. SparseR-CNN: End-to-End Object Detection with Learnable Proposals. arXiv 2020, arXiv:2011.12450. [Google Scholar]
  59. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. Yolov9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  60. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  61. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Computer Vision—ECCV 2020; Springer: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  62. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  63. Yang, G.; Lei, J.; Zhu, Z.; Cheng, S.; Feng, Z.; Liang, R. AFPN: Asymptotic feature pyramid network for object detection. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, HI, USA, 1–4 October 2023; pp. 2184–2189. [Google Scholar]
  64. Xiao, M.; Yang, B.; Wang, S.; Zhang, Z.; He, Y. Fine coordinate attention for surface defect detection. Eng. Appl. Artif. Intell. 2023, 123, 106368. [Google Scholar] [CrossRef]
Figure 1. Comparison of detection results between our method and other state-of-the-art strip steel surface defect detection approaches.
Figure 1. Comparison of detection results between our method and other state-of-the-art strip steel surface defect detection approaches.
Materials 18 05567 g001
Figure 2. Overall Architecture of the Proposed Method. (a) Backbone network. (b) Cross-Scale Spatial–Semantic Feature Aggregation Network. (c) Adaptive Region Proposal Network, (d) Detection head.
Figure 2. Overall Architecture of the Proposed Method. (a) Backbone network. (b) Cross-Scale Spatial–Semantic Feature Aggregation Network. (c) Adaptive Region Proposal Network, (d) Detection head.
Materials 18 05567 g002
Figure 3. The structure of the proposed CSSAM.
Figure 3. The structure of the proposed CSSAM.
Materials 18 05567 g003
Figure 4. The structure of the proposed ARPN.
Figure 4. The structure of the proposed ARPN.
Materials 18 05567 g004
Figure 5. Visual Comparison on the NEU-DET Dataset. (a) GT. (b) Ours. (c) TOOD. (d) Faster-RCNN. (e) Cascade-RCNN. (f) RT-DETR. (g) STD2.
Figure 5. Visual Comparison on the NEU-DET Dataset. (a) GT. (b) Ours. (c) TOOD. (d) Faster-RCNN. (e) Cascade-RCNN. (f) RT-DETR. (g) STD2.
Materials 18 05567 g005
Figure 6. Visual Comparison on the GC10-DET Dataset. (a) GT. (b) Ours. (c) TOOD. (d) Faster-RCNN. (e) Cascade-RCNN. (f) RT-DETR. (g) STD2.
Figure 6. Visual Comparison on the GC10-DET Dataset. (a) GT. (b) Ours. (c) TOOD. (d) Faster-RCNN. (e) Cascade-RCNN. (f) RT-DETR. (g) STD2.
Materials 18 05567 g006
Table 1. Hyperparameter setting of CA-RCNN.
Table 1. Hyperparameter setting of CA-RCNN.
HyperparametersLearning RateWeight DecayMomentum CoefficienOptimizerBatch SizeTrain:TestEpoch
Value0.0080.00010.9SGD28:236
Table 2. Quantitative comparison with 15 SOTA methods on the NEU-DET dataset.
Table 2. Quantitative comparison with 15 SOTA methods on the NEU-DET dataset.
MethodsAPAP50AP75APSAPMAPLCrInPaPsRsSc
SSD300 [50]39.773.734.433.833.351.248.882.794.285.963.866.9
TOOD [51]41.276.738.736.731.952.538.686.490.990.565.088.8
Faster-RCNN [52]43.478.943.241.737.751.944.785.493.292.565.795.3
DH-RCNN [53]36.775.033.540.329.140.740.479.487.688.563.091.2
Cascade-RCNN [54]44.079.344.041.639.853.947.084.591.789.566.095.3
Dynamic-RCNN [55]40.576.538.941.134.849.943.783.293.089.257.292.7
Grid-RCNN [56]41.875.941.438.734.053.340.185.792.589.956.290.8
Libra-RCNN [57]40.074.638.242.535.444.240.184.590.783.959.588.9
Sparse-RCNN [58]39.771.339.635.933.147.937.480.889.190.149.381.1
YOLOv9 [59]42.576.041.943.136.250.144.684.792.587.852.792.8
YOLOv10 [60]41.377.341.542.638.249.645.482.491.790.462.990.8
DETR [61]44.173.244.134.541.852.339.781.188.278.055.494.8
RT-DETR [62]44.575.544.437.639.053.743.686.391.883.058.690.0
CA-Autoassign [24]39.577.041.534.624.045.444.484.190.483.465.893.6
STD2 [20]43.180.441.736.738.753.152.985.394.193.164.093.1
Ours45.181.146.543.339.256.953.686.895.193.264.893.3
The bold text indicates the best result among all methods.
Table 3. Quantitative comparison with 15 SOTA methods on the GC10-DET dataset.
Table 3. Quantitative comparison with 15 SOTA methods on the GC10-DET dataset.
MethodsAPAP50AP75APSAPMAPLPuWlCgWsOsSsInRpCrWf
SSD300 [50]27.858.120.18.724.529.493.676.490.873.555.160.037.314.544.835.4
TOOD [51]34.565.231.312.628.939.093.279.890.176.566.059.726.233.555.970.7
Faster-RCNN [52]34.268.029.610.230.534.295.795.290.576.265.063.137.648.837.270.5
DH-RCNN [53]30.665.826.49.929.430.496.979.484.080.870.267.134.933.339.070.2
Cascade-RCNN [54]34.869.930.911.930.038.794.596.988.675.071.363.738.046.550.473.8
Dynamic-RCNN [55]30.362.223.910.629.930.697.296.486.569.367.857.931.425.931.058.5
Grid-RCNN [56]30.163.124.211.328.730.996.392.385.667.369.454.230.532.337.268.0
Libra-RCNN [57]27.557.922.112.928.926.697.493.688.267.465.354.817.518.227.049.9
Sparse-RCNN [58]32.365.830.08.127.734.796.998.287.471.067.262.429.030.142.472.4
YOLOv9 [59]33.667.631.514.229.735.991.582.892.979.670.066.825.741.156.170.8
YOLOv10 [60]32.964.231.013.530.634.793.279.896.177.565.061.719.230.555.973.7
DETR [61]31.768.830.112.127.036.995.890.390.980.455.460.143.043.355.471.5
RT-DETR [62]30.169.722.311.625.535.096.788.792.180.462.365.241.545.355.070.1
CA-Autoassign [24]24.362.627.59.722.732.995.977.792.770.562.061.735.331.832.565.7
STD2 [20]35.071.029.812.132.038.395.696.587.077.770.863.942.348.556.470.0
Ours35.972.330.014.131.540.197.498.388.678.967.967.843.750.656.572.8
The bold text indicates the best result among all methods.
Table 4. Ablation study results on the NEU-DET datasets.
Table 4. Ablation study results on the NEU-DET datasets.
MethodsAPAP50AP75APSAPMAPL
Baseline43.478.943.241.737.751.9
Baseline + CSSFAN44.580.045.442.438.154.2
Baseline + ARPN43.979.844.342.638.953.6
Baseline + CSSFAN + ARPN45.181.146.543.339.256.9
Table 5. Comparison results between CSSFAN and five mainstream feature fusion networks on the NEU-DET dataset.
Table 5. Comparison results between CSSFAN and five mainstream feature fusion networks on the NEU-DET dataset.
MethodsAPAP50AP75APSAPMAPL
FPN [38]43.478.943.241.737.751.9
NAS-FPN [39]41.777.142.440.234.949.5
PAFPN [21]44.079.344.942.639.954.3
BiFPN [40]42.978.547.142.338.753.6
AFPN [63]42.676.340.240.035.150.3
CSSFAN (Ours)45.181.146.543.339.256.9
The bold text indicates the best result among all methods.
Table 6. Generalization experiments on the RSW-D dataset.
Table 6. Generalization experiments on the RSW-D dataset.
MethodsAPAP50AP75APSAPMAPLPuWlCgWsOsSsIn
SSD300 [50]66.095.579.439.555.767.095.798.396.893.897.694.791.7
TOOD [51]70.595.986.248.364.871.597.598.495.096.498.096.190.1
Faster-RCNN [52]68.395.782.441.763.169.295.496.996.795.495.595.593.6
DH-RCNN [53]66.994.283.043.162.967.394.295.095.093.894.792.694.5
Cascade-RCNN [54]69.996.585.443.267.970.695.598.996.896.097.495.695.0
Dynamic-RCNN [55]62.191.880.040.657.564.191.193.392.690.492.591.391.4
Grid-RCNN [56]63.793.881.541.660.365.995.296.493.394.795.294.188.1
Libra-RCNN [57]60.690.776.342.151.562.591.793.491.990.392.589.596.4
Sparse-RCNN [58]62.692.881.140.356.464.992.394.393.591.393.892.792.4
YOLOv9 [59]65.594.570.644.157.761.494.897.295.792.896.793.690.8
YOLOv10 [60]66.194.780.140.661.965.294.595.895.694.595.394.792.6
DETR [61]64.592.983.146.361.568.294.395.292.293.595.193.187.0
RT-DETR [62]67.295.481.845.554.168.396.798.595.795.296.094.791.1
CA-Autoassign [24]67.695.781.447.356.868.596.598.696.895.496.394.991.2
STD2 [20]68.196.884.044.662.569.196.498.097.895.897.596.196.0
Ours70.997.287.647.667.771.896.199.098.096.297.696.697.0
The bold text indicates the best result among all methods.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, C.; Sun, Y.; Huang, L.; Guo, H. A Cross-Scale Spatial–Semantic Feature Aggregation Network for Strip Steel Surface Defect Detection. Materials 2025, 18, 5567. https://doi.org/10.3390/ma18245567

AMA Style

Xu C, Sun Y, Huang L, Guo H. A Cross-Scale Spatial–Semantic Feature Aggregation Network for Strip Steel Surface Defect Detection. Materials. 2025; 18(24):5567. https://doi.org/10.3390/ma18245567

Chicago/Turabian Style

Xu, Chenglong, Yange Sun, Linlin Huang, and Huaping Guo. 2025. "A Cross-Scale Spatial–Semantic Feature Aggregation Network for Strip Steel Surface Defect Detection" Materials 18, no. 24: 5567. https://doi.org/10.3390/ma18245567

APA Style

Xu, C., Sun, Y., Huang, L., & Guo, H. (2025). A Cross-Scale Spatial–Semantic Feature Aggregation Network for Strip Steel Surface Defect Detection. Materials, 18(24), 5567. https://doi.org/10.3390/ma18245567

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop