Next Article in Journal
Development and Field Deployment of a Compact Dual-Range Infrared Carbon Dioxide Sensor System
Previous Article in Journal
A Two-Phase Embedding Approach for Secure Distributed Steganography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Detection Algorithm for Concrete Bridge Defects Based on SATH–YOLO Model

College of Automotive and Transportation Engineering, Wuhan University of Science and Technology, Wuhan 430081, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(5), 1449; https://doi.org/10.3390/s25051449
Submission received: 31 January 2025 / Revised: 20 February 2025 / Accepted: 23 February 2025 / Published: 27 February 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Amid the era of intelligent construction and inspection, traditional object detection models like YOLOv8 struggle in bridge defect detection due to high computational complexity and limited speed. To address this, the lightweight SATH–YOLO model was proposed in this paper. First, the Star Block from StarNet was used to build the STNC2f module, enriching semantic information and improving multi-scale feature fusion while reducing parameters and computation. Second, the SPPF module was replaced with an AIFI module to capture finer-grained local features, improving feature-fusion precision and adaptability in complex scenarios. Lastly, a lightweight TDMDH detection head with shared convolution and dynamic feature selection further reduced computational costs. With the SATH–YOLO model, parameter count, computation, and model size were reduced significantly by 39.9%, 8.6%, and 36.2%, respectively. Meanwhile, the average detection precision was not impacted but improved by 1%, which meets the demands of edge devices and resource-constrained environments.

1. Introduction

With the rapid pace of urbanization and the accelerated development of infrastructure, bridges serve as critical nodes in transportation networks, where their safety and reliability directly affect public safety and the seamless operation of economic and social systems. Among existing bridge structures, concrete bridges are one of the most widely utilized types, owing to their superior performance, long service life, and low maintenance costs [1]. However, as bridges age, internal damage and deterioration diminish their load-bearing capacity. Cracks, as a direct manifestation of structural damage, are a primary focus in visual bridge inspections [2]. The presence of cracks accelerates the corrosion of reinforcement steel, significantly shortens the lifespan of bridge structures, and undermines their overall strength and stability. Severe penetrating cracks can result in critical structural damage, posing substantial threats to bridge safety. Consequently, the development of efficient and precise methods for detecting concrete bridge defects has become a pressing issue in the maintenance and management of bridges.
Traditional bridge inspection methods primarily rely on manual inspections and basic physical equipment, which are characterized by low efficiency, high costs, and a dependency on subjective judgment. These limitations make it challenging to detect defects in a timely and accurate manner, especially given the growing demand for routine maintenance of a large number of bridges. Against this backdrop, intelligent detection technologies have gained significant attention in recent years.
The rise of computer vision, particularly the rapid advancement of deep learning techniques, has offered novel solutions for bridge defect detection [3]. For example, Zheng et al. [4] proposed a high-precision concrete crack detection method based on Seg-Net and Bottleneck Architectures. Wang et al. [5] developed a crack detection model using the Inception-ResNet-v2 algorithm, enabling precise segmentation of target areas. Zhang et al. [6] introduced an improved You Only Look Once v3 (YOLOv3)-based method for bridge crack recognition, combining MobileNet with convolutional attention modules to achieve real-time detection of bridge deck cracks. Liao Yanna et al. [7] enhanced the multi-scale prediction module in the YOLOv3 network to leverage shallow semantic features, improving crack detection accuracy. Huang Keyuan et al. [8] utilized YOLOv5 with attention modules to further enhance the accuracy and efficiency of concrete bridge crack recognition. Li Pei et al. [9] proposed an automatic crack detection and classification approach using YOLOv5s with SENet and Coordinate Attention modules. INAM et al. [10] developed a two-stage algorithm that combines YOLOv5 and U-Net. Yang Guojun et al. [2] designed an integrated model for bridge crack recognition and segmentation based on an improved YOLOv7 and SeaFormer, addressing issues such as large model parameters, low detection efficiency, and high latency. Xu Yang et al. [11] applied the Faster R-CNN algorithm to the detection of surface defects of concrete buildings after an earthquake and achieved the recognition and location of four types of defects, including cracks, spalling, exposed steel bars, and buckling steel bars, with an overall average accuracy of 80%. Li Ruoxing et al. [12] proposed an improved Faster R-CNN algorithm to classify and locate concrete defects such as spalling, cracks, and exposed steel bars, with an accuracy of 80.7% for defect classification and 86% for location. Maeda et al. [13] embedded the single-stage SSD algorithm into mobile devices for the detection of concrete pavement defects, successfully detecting a variety of common concrete surface defects, achieving real-time detection of defects, and improving the efficiency of concrete pavement defect detection.
While these methods have achieved notable success in crack detection, their practical application in real-world bridge engineering is often constrained by factors such as limited computational resources, model complexity, and high parameter and computational requirements [14,15]. To address these challenges, this study proposes SATH–YOLO, an improved lightweight model based on YOLOv8. SATH–YOLO aims to facilitate efficient and precise crack detection in practical bridge engineering scenarios.

2. SATH–YOLO Algorithm Framework

2.1. Overview of the YOLOv8 Algorithm

As depicted in Figure 1, the YOLOv8 algorithm comprises three primary components: the Backbone, the Neck, and the Prediction Head. Compared to YOLOv5, YOLOv8 introduces the C2f module in both the Backbone and Neck, replacing the C3 module. This enhancement improves the model’s lightweight characteristics and facilitates gradient flow information, thereby boosting feature-representation capabilities.
In the Prediction Head, YOLOv8 employs a decoupled head design, separating the classification and detection tasks. This decoupling allows for independent task optimization, resulting in improved average precision and detection stability. Furthermore, YOLOv8 eliminates the traditional anchor-based mechanism, opting instead for an anchor-free approach. This transition simplifies the processes of anchor generation and bounding box refinement, reduces reliance on IoU computations, and significantly decreases memory and computational demands during training.
For positive and negative sample assignment, YOLOv8 introduces the Task-Aligned Assigner mechanism, which dynamically adjusts the ratio of positive to negative samples during training. This feature enhances the model’s adaptability across diverse datasets and tasks.
In terms of loss functions, YOLOv8 employs Binary Cross-Entropy (BCE) loss for classification and integrates Complete IoU (CIoU) loss with Distribution Focal Loss (DFL) for regression. This design aligns seamlessly with the anchor-free paradigm, accelerating the learning of target positional distributions and enabling the model to locate objects more quickly and accurately.

2.2. STNC2f Module

The YOLO series models [16,17,18,19] commonly employ a combination of linear projection and nonlinear activation functions. To enhance the feature extraction capabilities of these models, researchers often introduce self-attention mechanism modules based on this combination. These modules construct attention matrices through dot-product operations; however, the computational complexity increases significantly as the feature extraction capability improves.
To address this issue, Microsoft proposed a novel lightweight network named StarNet in 2024 [20], whose architecture is illustrated in Figure 2. StarNet adopts a model design based on star-shaped operations, significantly improving the ability of the model to map inputs to high-dimensional feature spaces, while maintaining low computational complexity. This enhancement ultimately bolsters the model’s feature extraction capacity [21].
The network employs a traditional hierarchical structure, divided into four stages, with the number of channels doubling progressively in each stage at an expansion factor of four. The network achieves down-sampling through convolutional layers and incorporates Star Blocks, which include depthwise convolution, for feature extraction. After the depthwise convolution layers, batch normalization (BN) is utilized instead of layer normalization (LN) to improve computational efficiency.
This design demonstrates exceptional performance and computational efficiency in fields such as natural language processing and computer vision, gradually being applied to lightweight network architectures. The star-shaped operation integrates features from two linear transformations through element-wise multiplication, with the computational formula shown below:
O s t a r = W 1 T X W 2 T X
W = W B
X a u g = X 1
In the equations: O s t a r represents the result of the star-shaped operation; W denotes the weight matrix of the linear layer; B is the bias of the linear layer; X a u g represents the input; and signifies the star-shaped operation. Specifically, W 1 , W 2 R ( q + 1 ) ( q + 1 )   , X R d + 1 × n .
To simplify the analysis, StarNet focuses on scenarios involving single-output channel transformations and single-element inputs. Based on Equations (1)–(3), it is further defined that W 1 ,   W 2 , X R q + 1 × 1 , here q is the number of input channels. Thus, the star-shaped operation can be further expressed as:
( w 1 T   x ) ( w 2 T   x ) = m = 1 q + 1 w 1 m   x m   n = 1 q + 1 w 2 n   x n   = m = 1 q + 1   n = 1 q + 1 w 1 m   x m   w 2 n   x n = α ( 1 , 1 )   x 1 x 1 + + α ( 4 , 5 )   x 4 x 5 + + α ( q + 1 , q + 1 )   x q + 1 x q + 1 q + 2 q + 1 2
where m , n represents the channel index, and α denotes the coefficient for each term.
α ( m , n ) = w 1 m w 2 n   m = n , w 1 m w 2 n + w 1 n w 2 m   m n .
By further elaborating on the star-shaped operation described in Equation (1), it can be expanded into a combination of different terms, as shown in Equation (4). Apart from the special terms, each remaining term exhibits a nonlinear relationship with x, indicating that these terms represent independent, implicit dimensions. After computation in a q-dimensional space using the star-shaped operation, the resulting implicit dimensional feature space has a dimensionality of ( q + 2 ) ( q + 1 ) 2 ( d 2 ) 2 . This significantly amplifies the feature dimensions without incurring any additional computational overhead within a single layer.
By stacking multiple layers, the star-shaped operation can recursively increase the implicit dimensionality to nearly infinite levels:
O l = W l , 1 T O l 1 W l , 2 T O l 1 R d 2 2   l
where O l represents the output of the l-th star-shaped operation.
As expressed in Equation (6), using l layers of star-shaped operations enables the construction of an implicit dimensional feature space belonging to R ( d 2 ) 2 l .
Inspired by the outstanding lightweight design and information extraction capabilities of StarNet, the YOLOv8 baseline model is improved in this work. Given that the overall architecture of StarNet is relatively complex, a direct introduction would result in model redundancy, which could negatively affect the lightweight characteristics of the model. Therefore, this work only introduces the Star Blocks structure, replacing the Bottleneck in the original C2f module in the Backbone and Neck parts with Star Blocks of StarNet, forming the STNC2f module. This modification aims to enhance the network’s feature extraction capabilities and multi-scale information fusion, while reducing both the model’s parameter count and computational complexity. The improved STNC2f module structure is shown in Figure 3.

2.3. Intrascale Feature Interaction

The Spatial Pyramid Pooling Fast (SPPF) module is a crucial component in YOLOv8 for enhancing feature-extraction efficiency and performance. By pooling features at different scales, SPPF effectively captures contextual information and improves the spatial invariance of the model. The Attention-based Intrascale Feature Interaction (AIFI) module, introduced in the RT-DETR model [22], is a key element that employs an attention mechanism for intrascale feature interaction, thus improving the efficiency and effectiveness of feature extraction. Its core lies in the application of attention mechanisms between features of the same scale, enhancing the model’s focusing ability and facilitating richer feature fusion. The structure of the AIFI module is illustrated in Figure 4.
The feature fusion process of the AIFI module is shown in Figure 5: First, the 2D S5 feature map is converted into high-dimensional vectors, which are then processed through the AIFI module. The mathematical process involves the output of Multi-Head Self-Attention (MHSA) being added to the input via residual connections and normalized by Layer Normalization to efficiently capture complex relationships within the input sequence. Subsequently, these features are passed through a Feed-Forward Network (FFN) for nonlinear transformation and feature extraction. Finally, the output is normalized again, generating attention scores that encapsulate the critical feature information, which is then returned to 2D form, denoted as F5, for subsequent cross-scale feature fusion. AIFI processes only the S5 feature map to leverage its rich semantic information, reducing computational complexity while preserving detection performance.
Based on this, the improved model in this study replaced the SPPF module with AIFI and added an additional convolutional layer (Conv) before AIFI. This combination helps capture the dependencies between features and improve the efficiency of feature handling and fusion, thereby it is promising in reinforcing the overall detection performance and maintaining good precision for large, medium, and small object detection.

2.4. Task Dynamic Mutual Detection Head (TDMDH)

YOLOv8 employs the widely used decoupled head structure in the detection head, which splits the regression and classification tasks into independent branches, each containing two 3 × 3 convolutions and one 1 × 1 convolution. While this design improves detection performance, it significantly increases both the parameter count and computational load. Moreover, YOLOv8 uses a traditional single-scale feature prediction module, which overlooks the impact of features at other scales on detection performance. As a result, YOLOv8 struggles to fully utilize its performance advantages in resource-constrained environments, such as edge devices. To clear this hurdle, we are proposing a customized task-aligned detection head structure, Task Dynamic Mutual Detection Head (TDMDH). Inspired by the concepts of Tood [23], this structure further optimizes model performance through task-aligned label assignment strategies. TDMDH generates joint features by interacting with information from different scales, significantly reducing the model’s parameter count while ensuring that the average precision remains consistent. This makes the model more lightweight and better suited for deployment in resource-constrained environments.
Regarding normalization techniques, Batch Normalization (BN) has proven effective in improving training speed and convergence. However, its strong dependence on batch size can negatively affect performance. Therefore, we replace BN in the convolutional layers with Group Normalization (GN) to overcome the limitations of BN and prevent accuracy degradation due to lightweight operations. The structure is illustrated in Figure 6.
The TDMDH structure is shown in Figure 7. The three feature layers ( P 3 , P 4 , P 5 ) from the Neck are processed sequentially through two shared convolutional modules (Conv_GN), each with a 3 × 3 kernel. The information sharing between the two convolution layers effectively reduces both the parameter count and computational load. These features are then concatenated using a Concat operation, forming interactive features with combined information. These interactive features are then separated into regression and classification branches via a Task Decomposition module.
In the regression-shared branch, the interactive features generate offsets and masks, which are combined with Deformable Convolutional Networks v2 (DCNv2) [24]. DCNv2 adjusts the convolutional kernel positions and weights more flexibly, especially effective for handling complex geometric transformations such as cracks. Finally, a 1 × 1 convolution predicts the boundary box offsets, and a Scale layer adjusts the output features to accommodate different object scales.
In the classification shared branch, the Dynamic Features Select (DYFS) module performs dynamic feature selection on the interactive features and generates corresponding weights to enhance the recognition of damage targets. Finally, a 1 × 1 convolutional layer predicts the probabilities of each class. The DYFS structure is shown in Figure 8.

2.5. SATH–YOLO Detection

This paper proposes a lightweight improved model based on YOLOv8—SATH–YOLO (StarNet AIFI TDMDH-YOLO), whose structure is shown in Figure 9.
Based on the above components, we introduced the Star Block into the Backbone and Neck of YOLOv8, replacing the Bottleneck in the original C2f module with StarNet’s Star Blocks to form the STN-C2f module. This modification achieves the lightweight design of the Backbone and Neck networks, while ensuring detection accuracy. Next, we replaced the SPPF module at the end of YOLOv8’s Backbone with the AIFI (Attention-based Intrascale Feature Interaction) module from the RT-DETR model. This self-attention operation on high-level features not only avoids intrascale interaction of low-level features but also effectively reduces computational redundancy. Finally, the detection head of YOLOv8 is replaced with the TDMDH detection head. Traditional object detection heads typically have separate classification and localization branches, which lack interaction between the two tasks. TDMDH, by extracting features from multiple convolutional layers, learns task interaction characteristics, generates joint features, and significantly reduces model parameter count through shared convolution and dynamic feature selection.

3. Experiment-Related Preparation

3.1. Experimental Equipment and Hyperparameters

The equipment and some of the hyperparameters used in the experiments are listed in Table 1.

3.2. Experimental Dataset

Although the use of public benchmark datasets is crucial for model validation, there is currently a limited availability of public datasets for concrete bridge defects, particularly those that encompass a wide range of defect types and diverse environmental conditions. Therefore, the dataset utilized in this study was initially acquired through field photography of bridge crack defects under various environmental conditions, lighting variations, and shooting angles to ensure diversity and representativeness. The equipment used for image capture was an iPhone 13 Pro Max. To enhance the robustness of the dataset, data augmentation techniques were applied. These included operations such as rotation, scaling, translation, mirror flipping, and the addition of noise to the original images, generating a diverse set of crack image samples that simulate different environmental and operational conditions. This approach ultimately aimed to improve the generalization capability of the model.
After the image data collection was completed, manual annotation was performed using the semi-automatic labeling tool, labelImg. The final dataset comprised 4691 images. Using a dataset partitioning program, the dataset was randomly split into a training set (3801 images), a validation set (678 images), and a test set (212 images).

3.3. Evaluation Metrics

To objectively assess the effectiveness of the proposed network model in detecting concrete bridge defects, the following evaluation metrics were selected: precision (P), recall (R), mean average precision (mAP), giga floating point operations per second (GFLOPs), parameter count, and model size. Precision (P) measures the probability of correctly detecting the target, recall (R) determines whether the target can be identified within the complete dataset, and mAP represents the average precision (AP) value across all categories. The closer P, R, and mAP are to 1, and the smaller the computation volume, parameter count, and model size, the higher the model’s recognition accuracy and lightweight performance.
The specific formulas are as follows:
R = T P T P + F N
P = T P T P + F P
F A P = 0 1 p r     d r
F m A P = 1 n i = 1 n F A P i
In Equations (7) and (8), True Positive ( T P ) refers to the number of correctly predicted positive samples; False Positive ( F P ) indicates false positive cases, where the location of the target is detected but the target’s class is misidentified; False Negative ( F N ) refers to the number of undetected targets. Equation (9) defines the mean average precision ( F A P ), which is the average of precision values obtained at recall rates between 0 and 1. In Equation (10), n represents the number of classes in the dataset, and F A P i refers to the average precision of the i-th class. The final F m A P is the mean of the average precision (AP) across all classes. For the m A P evaluation in this study, we utilized mAP@50, which represents the average of   A P values when the Intersection over Union (IoU) threshold is set to 0.5.

4. Experiments and Discussion

4.1. Visual Comparative

To intuitively evaluate the detection performance of SATH–YOLO, a comparison was made with the baseline model, YOLOv8, in terms of crack detection results, as shown in Figure 10. Additionally, Grad-CAM [25] was employed to generate heatmaps that demonstrate the attention of both YOLO models on the crack regions and their respective detection performance. The visual results are presented in Figure 11, where the red and yellow regions in the heatmaps indicate the areas with higher attention from the models.
From Figure 10, the detection boxes of SATH–YOLO exhibited higher boundary precision and more stable box positioning across various crack shapes and lengths. Columns a and c showed that SATH–YOLO captured the full extent of the cracks with reduced background false detections and incomplete detections.
By comparing the visual results of different samples in Figure 11, the following observations can be made:
  • In rows a and b, the heatmap of YOLOv8 shows a slightly diffused attention range, failing to fully align with the edges of the cracks, resulting in less precise detection box boundaries. In contrast, the heatmap of SATH–YOLO better covers the entire crack region, demonstrating higher detection stability.
  • In row c, SATH–YOLO clearly focuses on the core area of the crack, while YOLOv8’s response is unevenly distributed, diminishing detection reliability.
  • In row d, the YOLOv8 heatmap exhibits significant background interference, with the detected area failing to clearly highlight the crack path. In contrast, the SATH–YOLO heatmap effectively reduces interference and accurately identifies crack features.
In summary, compared to YOLOv8, SATH–YOLO demonstrates superior visual attention concentration and boundary recognition precision in crack detection. The positioning capability of its detection boxes and its ability to focus on the target region in the heatmap highlight SATH–YOLO’s enhanced robustness and sensitivity in handling complex backgrounds, thereby providing higher reliability and accuracy for detection tasks.

4.2. Ablation Study

To validate the effectiveness of each improvement in the proposed model, an ablation study was conducted for each modified module. The experimental process strictly controlled model parameters to ensure consistency. The evaluation results of the ablation study are shown in Table 2.
As observed from Table 2, the introduction of each individual improvement leads to a certain degree of performance enhancement. Compared to the baseline model, the introduction of the STNC2f module reduces the model’s parameter count by 16.9%, the computation volume and model size by 14.8%, with only a 0.4% decrease in the mean average precision (mAP@50). Replacing the SPPF module in the backbone with the AIFI module enables AIFI to selectively perform self-attention on high-level features containing rich semantic information, preventing internal scale interactions of low-level features, thereby capturing the relationships between conceptual entities in the image. This approach aids in more efficient detection and recognition of objects by subsequent modules. As a result, the parameter count is reduced by 3.7%, computation volume by 1.2%, and model size by 2.2%, demonstrating that AIFI effectively reduces computational redundancy.
Introducing the Task-Dynamic Multi-Task Head into the detection head leads to information sharing and task decoupling mechanisms between the convolutional layers, regression, and classification branches. This enables the learning of joint features from multi-scale information interactions and dynamically selects relevant features to generate corresponding weights. This results in a significant reduction in both parameter count and model size, while also slightly improving mAP. Specifically, the parameter count decreases by 26.9%, model size by 24.7%, and mAP@50 increases by 0.3%, with computation volume increasing by just 0.6%.
When the STNC2f, AIFI, and TDMDH modules are combined, SATH–YOLO achieves significant improvements across all metrics compared to the baseline model. The parameter count decreases by 37.9%, computation volume by 8.6%, model size by 36.2%, and mAP@50 increases by 1%. These experimental results suggest that SATH–YOLO dynamically selects features and acquires rich semantic information, thereby improving performance while achieving model lightweighting.

4.3. Comparative Experiments

To further validate the performance of the SATH–YOLO algorithm in detecting concrete bridge crack defects, it was compared with other high-performance models in the YOLO series, including YOLOv3-tiny, YOLOv5n, YOLOv6, YOLOv8n, YOLOv9t, YOLOv10n, and YOLOv11n. The experimental environment and parameter configurations strictly followed the settings outlined in Section 2.1, with the experimental results shown in Table 2. The mAP@50 curves for all algorithms after training for 300 epochs are presented in Figure 12.
The experimental results indicate that YOLOv3-tiny, YOLOv6, and YOLOv8n have relatively high parameter counts, computation volumes, and model sizes, with subpar mAP@50 performance. While YOLOv5n and YOLOv9t perform well in terms of lightweight design, their parameter counts and model sizes are still higher than those of SATH–YOLO, and their detection mAP@50 is 0.8% and 1% lower, respectively. YOLOv10n shows similar mAP@50 performance to SATH–YOLO but has a 121.4% and 145.5% higher parameter count and model size, respectively. YOLOv11n, although having a computational advantage with a 14.9% reduction in computation volume compared to SATH–YOLO, has lower mAP@50 and a parameter count and model size 138% and 137.6% larger than SATH–YOLO’s. See Table 3.
As shown in Figure 12, SATH–YOLO’s mAP@50 curve remains consistently superior throughout the 300-epoch training process, with minimal fluctuations in the stable phase, indicating higher detection accuracy and stability. SATH–YOLO not only matches the convergence speed of other advanced models but also exhibits significant advantages in final detection performance. This demonstrates that the improvements in feature extraction and detection robustness effectively enhance the model’s overall performance.
Due to resource limitations, this study was unable to conduct testing on edge devices. However, future experiments will focus on deploying the StarNet SATH–YOLO model on platforms such as the NVIDIA Jetson Nano and Raspberry Pi. These tests will evaluate the model’s real-time performance and practicality in resource-constrained environments. The planned experiments will include assessing inference speed, memory usage, and power consumption, providing insights into the model’s suitability for edge deployment in practical applications.

4.4. Statistical Analysis and Significance

To assess the statistical significance of the performance improvement of the proposed SATH–YOLO model over the YOLOv8 baseline, an independent two-sample t-test was performed. The null hypothesis (H0) posits no significant difference between the models, while the alternative hypothesis (H1) asserts a significant difference in performance. A significance level of 0.05 was adopted. The t-test was conducted based on the accuracy values obtained from five independent training runs for each model.
The t-statistic is computed as follows:
t = X 1 ¯ X 2 ¯ s 1 2 n 1 + s 2 2 n 2
where X 1 ¯ , X 2 ¯ are the mean accuracies for YOLOv8 and SATH–YOLO, respectively, s 1 2 , s 2   2 are the sample variances, and n 1 , n 2 are the sample sizes.
The average accuracy of YOLOv8 was 97.2% (SD = 0.20%), while SATH–YOLO achieved 98.2% (SD = 0.16%). The t-test resulted in a t-statistic of 10.87 and a p-value of 4.54 ×   10 6 , indicating a statistically significant performance difference (p < 0.05). See Table 4.
The results demonstrate that the performance improvement observed in SATH–YOLO is statistically significant, as the p-value of 4.54 ×   10 6 is well below the 0.05 threshold. This confirms that the observed performance difference is not due to random chance, providing strong statistical evidence for the superiority of the proposed model.
These findings suggest that SATH–YOLO not only improves accuracy but also does so with statistical significance, further reinforcing its effectiveness and efficiency, particularly for deployment in resource-constrained environments.

5. Conclusions

With the widespread application of deep learning in practical scenarios, particularly in edge devices or resource-constrained environments, model lightweighting has become key to enhancing the applicability of models [26,27,28]. In order to address the issue that current mainstream object detection algorithms, such as YOLOv8, which struggle to fully exploit their performance advantages in resource-limited and edge-device environments, the concrete bridge disease detection model SATH–YOLO was proposed to improve the accuracy and efficiency of concrete bridge defect detection.
With the SATH–YOLO model, the C2f module in the Backbone and Neck networks was replaced with the STN-C2f module composed of Star Blocks from the lightweight StarNet network; the traditional SPPF feature fusion methods was replaced with AIFI module focusing on high-level semantic feature interaction; the lightweight TDMDH was utilized to minimize the parameter count and computation.
Compared to the baseline model YOLOv8, the SATH–YOLO model showed significant improvement in parameter count, computation volume, and model size, with no impact on precision. The experimental results from this study showed that SATH–YOLO is a promising model that allows the bridge detection to run efficiently with limited computational resources and low-power devices, accelerating inference speed, reducing power consumption and hardware costs, and enabling rapid deployment. Therefore, the proposed concrete bridge disease detection method based on SATH–YOLO has broad application prospects, especially in large-scale bridge monitoring tasks, where it can effectively reduce manual inspection workloads and improve the comprehensiveness and accuracy of defect identification. Despite the promising results of the SATH–YOLO model for crack detection in bridges, this study has several limitations. First, the model’s evaluation was limited to crack detection, and its generalizability to other types of structural damage remains untested. Additionally, the dataset used in this study may not fully represent real-world variability, and the lack of real-world testing on edge devices limits the practical assessment of the model’s deployment. Furthermore, the model’s stability and generalization ability in extreme environments need further optimization. Future research should focus on expanding the model to detect other types of damage, improving the dataset with more diverse and representative data, and testing the model on actual edge devices such as the Jetson Nano or Raspberry Pi to validate its feasibility for real-time deployment in resource-constrained environments. Moreover, exploring integrations with other detection technologies, such as drone inspections and laser scanning, could enhance the model’s robustness and advance the automation and intelligence of bridge defect detection. Overall, the proposed algorithm is of significant theoretical importance and practical feasibility, with strong potential for widespread application.

Author Contributions

Conceptualization, L.Z. and A.L.; methodology, L.Z.; software, L.Z.; validation, L.Z. and A.L.; formal analysis, A.L.; investigation, A.L.; resources, L.Z.; data curation, L.Z.; writing—original draft preparation, L.Z.; writing—review and editing, L.Z. and A.L.; visualization, L.Z.; supervision, A.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are not available upon request from the corresponding author due to the involvement of personal privacy in some of the monitoring footage.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yongliang, Z.; Dawei, Z. Durability assessment model of concrete bridge based on multi-evidence fusion under NRS. China Saf. Sci. J. 2023, 33, 169. [Google Scholar]
  2. Yang, G.; Qi, Y.; Du, Y.; Shi, X. Improved YOLOv7 and SeaFormer Bridge Cracks Identification and Measurement. J. Railw. Sci. Eng. 2025, 22, 429–442. (In Chinese) [Google Scholar]
  3. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  4. Zheng, X.; Zhang, S.; Li, X.; Li, G.; Li, X. Lightweight bridge crack detection method based on segnet and bottleneck depth-separable convolution with residuals. IEEE Access 2021, 9, 161649–161668. [Google Scholar] [CrossRef]
  5. Wang, J.; He, X.; Faming, S.; Lu, G.; Cong, H.; Jiang, Q. A real-time bridge crack detection method based on an improved inception-resnet-v2 structure. IEEE Access 2021, 9, 93209–93223. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Huang, J.; Cai, F. On bridge surface crack detection based on an improved YOLO v3 algorithm. IFAC-Pap. 2020, 53, 8205–8210. [Google Scholar] [CrossRef]
  7. Liao, Y.-n.; Li, W. Bridge crack detection method based on convolution neural network. Comput. Eng. Des. 2021, 42, 2366–2372. (In Chinese) [Google Scholar] [CrossRef]
  8. Huang, K.; Zhao, Y.; Hu, N.; Cao, J.; Xiang, Y. Study on Concrete Bridge Crack Recognition Using Attention-based YOLOv5 Network. Bull. Sci. Technol. 2024, 40, 71–76. (In Chinese) [Google Scholar] [CrossRef]
  9. Li, P.; Han, F.; Yang, K.; Li, Z. Automatic Detection and Classification of Bridge Cracks Based on Improved YOLOv5s Algorithm. Transp. Sci. Technol. 2024, 3, 53–58. [Google Scholar]
  10. Inam, H.; Islam, N.U.; Akram, M.U.; Ullah, F. Smart and automated infrastructure management: A deep learning approach for crack detection in bridge images. Sustainability 2023, 15, 1866. [Google Scholar] [CrossRef]
  11. Xu, Y.; Wei, S.; Bao, Y.; Li, H. Automatic seismic damage identification of reinforced concrete columns from images by a region-based deep convolutional neural network. Struct. Control Health Monit. 2019, 26, e2313. [Google Scholar] [CrossRef]
  12. Li, R.; Yuan, Y.; Zhang, W.; Yuan, Y. Unified vision-based methodology for simultaneous concrete defect detection and geolocalization. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 527–544. [Google Scholar] [CrossRef]
  13. Maeda, H.; Sekimoto, Y.; Seto, T.; Kashiyama, T.; Omata, H. Road damage detection and classification using deep neural networks with smartphone images. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 1127–1141. [Google Scholar] [CrossRef]
  14. Deng, L.; Chu, H.; Long, L.; Wang, W.; Kong, X.; Cao, R. Review of Deep Learning-based Crack Detection for Civil Infrastructures. China J. Highw. Transp. 2023, 36, 1–21. (In Chinese) [Google Scholar] [CrossRef]
  15. Luo, H.; Wang, W.; Ye, X.; Zhu, S.; Bai, Y. Research Progress of Directed Object Detection Based on Deep Learning. J. Image Signal Process. 2024, 13, 258. (In Chinese) [Google Scholar] [CrossRef]
  16. Redmon, J. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  17. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  18. Redmon, J. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  19. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  20. Ma, X.; Dai, X.; Bai, Y.; Wang, Y.; Fu, Y. Rewrite the Stars. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 5694–5703. [Google Scholar]
  21. Yi, L.; Huang, Z.; Yi, Y. Improved YOLOv8 foreign object detection method for transmission lines. Electron. Meas. Technol. 2024, 47, 125–134. (In Chinese) [Google Scholar] [CrossRef]
  22. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  23. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. Tood: Task-aligned one-stage object detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2017; IEEE Computer Society: Washington, DC, USA, 2021; pp. 3490–3499. [Google Scholar]
  24. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9308–9316. [Google Scholar]
  25. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  26. Xu, W.F.; Lei, Y.; Wang, H.T.; Zhang, X. Research on object detection models for edge devices. CAAI Trans. Intell. Syst. 2024, 1–11. Available online: https://link.cnki.net/urlid/23.1538.TP.20241211.1852.004 (accessed on 22 February 2025). (In Chinese).
  27. Peng, J.X.; Zhang, R.F.; Liu, Y.H. Corrupt fruit detection model based on improved YOLOv5n. Appl. Electron. Tech. 2024, 50, 55–60. (In Chinese) [Google Scholar] [CrossRef]
  28. Wang, M.; Dong, L.D.; Dong, W.; Zhang, H.X. An Elevator Passenger Abnormality Detection Method and Experimental. System Based on Edge Intelligence. J. Hangzhou Norm. Univ. 2024, 23, 439–445. (In Chinese) [Google Scholar] [CrossRef]
Figure 1. Structure diagram of the YOLOv8 algorithm.
Figure 1. Structure diagram of the YOLOv8 algorithm.
Sensors 25 01449 g001
Figure 2. StarNet Architecture Principles.
Figure 2. StarNet Architecture Principles.
Sensors 25 01449 g002
Figure 3. STNC2f module structure.
Figure 3. STNC2f module structure.
Sensors 25 01449 g003
Figure 4. AIFI module structure.
Figure 4. AIFI module structure.
Sensors 25 01449 g004
Figure 5. The feature fusion process of the AIFI module.
Figure 5. The feature fusion process of the AIFI module.
Sensors 25 01449 g005
Figure 6. Conv_GN module structure.
Figure 6. Conv_GN module structure.
Sensors 25 01449 g006
Figure 7. TDMDH structure diagram.
Figure 7. TDMDH structure diagram.
Sensors 25 01449 g007
Figure 8. DYFS structure diagram.
Figure 8. DYFS structure diagram.
Sensors 25 01449 g008
Figure 9. SATH–YOLO structure diagram.
Figure 9. SATH–YOLO structure diagram.
Sensors 25 01449 g009
Figure 10. Comparison of concrete bridge crack detection results. (ad) are four types of random cracks.
Figure 10. Comparison of concrete bridge crack detection results. (ad) are four types of random cracks.
Sensors 25 01449 g010
Figure 11. Visual comparison of focused feature areas. (ad) are four types of random cracks.
Figure 11. Visual comparison of focused feature areas. (ad) are four types of random cracks.
Sensors 25 01449 g011
Figure 12. Comparison of mAP@50 curves of YOLO series on bridge crack dataset.
Figure 12. Comparison of mAP@50 curves of YOLO series on bridge crack dataset.
Sensors 25 01449 g012
Table 1. The equipment and some of the hyperparameters used during the experiments.
Table 1. The equipment and some of the hyperparameters used during the experiments.
DeviceParameterHyperparameterValue
CPUIntel(R) Core(TM) i7-12700KFTraining StrategySGD
RAM32 GBMomentum0.937
Operating SystemWindows 10 ProInitial Learning Rate0.01
GPUNVIDIA GeForce RTX 4060TiFinal Learning Rate0.0001
GPU Memory8GBBatch Size16
Programming ToolVisual Studio Code 1.96.4Workers5
Programming LanguagePython 3.9.19Imgsz640
Deep Learning FrameworkPyTorch 2.3.1Total Epochs300
Table 2. Ablation experiment table.
Table 2. Ablation experiment table.
STNC2fAIFITDMDHParams/MFlops/GPrecision/%Recall/%mAP@50/%Model Size/MB
3.018.196.592.497.25.96
2.50↓6.9↓97.5↑92.8↑96.85.08↓
2.90↓8.0↓95.994.4↑96.95.83↓
2.20↓8.695.293.3↑97.5↑4.49↓
2.53↓7.1↓97.0↑94.0↑97.9↑5.09↓
1.85↓7.1↓96.494.7↑97.8↑3.75↓
2.19↓8.597.8↑91.797.6↑4.40↓
1.87↓7.4↓98.8↑94.1↑98.2↑3.80↓
“√” indicates that this module is introduced; “—” indicates that this module is not introduced; ↑ and ↓ both represent the parts that have been improved compared to the baseline model, where ↑ represents an increase and ↓ represents a decrease.
Table 3. Comparative Experiment Table.
Table 3. Comparative Experiment Table.
ModelParams/MFlops/GPrecision/%Recall/%mAP@50/%Model Size/MB
YOLOv3-tiny12.1318.979.982.288.923.2
YOLOv5n2.507.1↓98.293.497.45.05
YOLOv64.2311.897.690.397.38.30
YOLOv8n3.018.196.592.497.25.96
YOLOv9t1.977.695.794.6↑97.94.45
YOLOv10n2.276.5↓95.194.3↑98.25.53
YOLOv11n2.586.3↓97.195.3↑97.75.23
SATH–YOLO1.877.498.894.198.23.80
↑ and ↓ both represent the improvement compared to SATH–YOLO, where ↑ represents an increase and ↓ represents a decrease.
Table 4. Performance Comparison and Statistical Significance of YOLOv8 and SATH–YOLO.
Table 4. Performance Comparison and Statistical Significance of YOLOv8 and SATH–YOLO.
ModelAccuracy (%)Standard Deviation (%)t-Statisticp-ValueStatistical Significance
YOLOv897.20.20
SATH—YOLO98.20.1610.874.54 × 10 6 Significant
(p < 0.05)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zou, L.; Liu, A. Intelligent Detection Algorithm for Concrete Bridge Defects Based on SATH–YOLO Model. Sensors 2025, 25, 1449. https://doi.org/10.3390/s25051449

AMA Style

Zou L, Liu A. Intelligent Detection Algorithm for Concrete Bridge Defects Based on SATH–YOLO Model. Sensors. 2025; 25(5):1449. https://doi.org/10.3390/s25051449

Chicago/Turabian Style

Zou, Lanlin, and Ao Liu. 2025. "Intelligent Detection Algorithm for Concrete Bridge Defects Based on SATH–YOLO Model" Sensors 25, no. 5: 1449. https://doi.org/10.3390/s25051449

APA Style

Zou, L., & Liu, A. (2025). Intelligent Detection Algorithm for Concrete Bridge Defects Based on SATH–YOLO Model. Sensors, 25(5), 1449. https://doi.org/10.3390/s25051449

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop