Next Article in Journal
Detecting Excitations of Pipes, Ropes, and Bars Using Piezo Sensors and Collecting Information Remotely
Next Article in Special Issue
Vibration Analysis and Vehicle Detection by MEMS Acceleration Sensors Embedded in PCC Pavement
Previous Article in Journal
Top-k Shuffled Differential Privacy Federated Learning for Heterogeneous Data
Previous Article in Special Issue
Optimising Concrete Crack Detection: A Study of Transfer Learning with Application on Nvidia Jetson Nano
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLO-RD: A Road Damage Detection Method for Effective Pavement Maintenance

1
College of Computer Science and Technology, Changchun University, No. 6543, Satellite Road, Changchun 130022, China
2
School of Construction Engineering, Jilin University, No. 2699, Qianjin Street, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(5), 1442; https://doi.org/10.3390/s25051442
Submission received: 5 January 2025 / Revised: 4 February 2025 / Accepted: 6 February 2025 / Published: 27 February 2025
(This article belongs to the Special Issue Smart Sensors for Transportation Infrastructure Health Monitoring)

Abstract

:
Road damage detection is crucial for ensuring road safety and minimizing maintenance costs. However, detecting small damage, managing complex backgrounds, and identifying irregular damage shapes remain significant challenges. To address these issues, we propose YOLO-RD, an advanced detection framework that integrates innovative modules for feature enhancement, multi-scale robustness, and detail preservation. Specifically, the Star Operation Module (SOM) improves sensitivity to small-scale damage, the Multi-dimensional Auxiliary Fusion (MAF) module strengthens robustness in complex environments, and the Wavelet Transform Convolution (WTC) enables adaptive focus on irregular shapes. On the Japanese road dataset in RDD2022, YOLO-RD achieves a detection accuracy of 25.75%, with a notable 4.93% improvement in small object detection over the baseline YOLOv8. These results demonstrate the effectiveness and practicality of YOLO-RD in addressing diverse and challenging real-world scenarios, establishing it as a robust solution for automated road condition monitoring.

1. Introduction

Road damage, a common type of road degradation, severely impacts the lifespan of roads. With over 5.44 million kilometers of highways in China, which handle enormous traffic volumes each year, issues such as road cracks and potholes emerge frequently, exacerbated by changing climatic conditions [1]. These problems pose risks to road safety and increase maintenance burdens, making road damage detection essential [2]. Traditional manual inspection methods rely on on-site checks and data recording, which are both time-consuming and labor-intensive. Furthermore, the results are easily influenced by subjective factors, leading to unstable outcomes. With the development of computer vision and deep learning, automated detection technology offers an efficient, accurate, and cost-effective solution that enables the rapid processing of large-scale road damage images [3]. However, road damage detection faces three major technical challenges. First, small object detection [4] is challenging, as fine cracks are easily overlooked in large-scale images, leading to reduced recognition accuracy. Second, complex background [5] interference, including lighting changes, stains, and rain, can affect the discrimination of the model, causing false detections. Lastly, irregular damage shapes [6], such as irregular cracks and potholes, make it difficult to apply fixed rules for effective recognition. These issues complicate feature extraction and pattern recognition, increasing the difficulty of model optimization. Therefore, the discussion and research presented in this paper hold significant practical importance.
In the field of road damage detection, traditional methods such as laser detection [7] and geomagnetic detection [8] are overly reliant on manual labor, resulting in poor detection performance. These methods are unable to pinpoint the exact location of damage, and they are costly and require complex maintenance. Rule-based image processing methods [9], although they can improve detection efficiency to some extent, primarily rely on predefined feature extraction algorithms. They lack the ability to adapt to changes in environmental lighting and complex backgrounds, making it difficult to handle the diverse damage shapes encountered in real-world scenarios. In contrast, data-driven road damage detection methods [10], which use deep learning models to automatically learn and extract features from images, can accurately localize and identify damage. Zhang et al. [11] proposed a fast pavement damage detection network (FPDDN) for real-time and high-accuracy road damage detection. Ren et al. [12] proposed a novel road damage detection method based on an improved YOLOv5 network and street view images. This model can handle large-scale detection layers to enhance the detection accuracy of large damaged objects. These methods demonstrate significant advantages in terms of adaptability and robustness, reducing the influence of human subjectivity. Furthermore, through continuous model optimization and data updates, they effectively respond to varying damage shapes and environmental changes.
Object detection [13] is a fundamental task in computer vision, and significant progress has been made in research on object detection in recent years. Numerous studies have developed various efficient object detection methods, which not only continuously improve accuracy but also optimize processing speed and model lightweight design. For example, the You Only Look Once (YOLO) [14] series models achieve real-time detection with their single-stage structure, making them suitable for scenarios that require quick responses. Two-stage detection models, such as Mask R-CNN [15], offer better accuracy and are commonly used in applications that demand high detection precision. Additionally, models like Efficientnet [16] and Single Shot MultiBox Detector (SSD) [17] combine multi-scale feature fusion and lightweight design, balancing both detection efficiency and accuracy, thus laying a solid foundation for the wide application of object detection in practical scenarios. In this study, object detection methods are widely applied to road damage detection tasks. These methods can quickly and accurately identify and localize damage regions while effectively handling variations in damage shapes and sizes. Although segmentation methods provide pixel-level damage information, they have higher computational costs and slower processing speeds. Object detection, while maintaining high accuracy, offers faster processing speeds, significantly enhancing the practicality and reliability of detection systems.
The following details the three existing challenges and their impacts: as shown in Figure 1, there are four categories (class_0, class_1, class_2, class_3), which, respectively, represent longitudinal cracks, transverse cracks, alligator cracks, and potholes. These challenges include small object detection, complex background interference, and the damage of irregular shapes. Small object detection is particularly challenging in road damage recognition, especially for fine cracks. Due to their small size, crack features are often not prominent in images, especially in high-resolution images, where the edges of cracks may be difficult to distinguish due to blurring or noise. The low contrast between small cracks and the surrounding road surface further increases the difficulty for the model, leading to false positives or missed detections. Additionally, because small damage occupies a small percentage of the pixels, it can be overlooked during global feature extraction, which is a problem that is exacerbated in low-resolution or noisy images [18]. Complex background interference is another significant challenge in road damage detection. Common environmental factors, such as changes in lighting angles, tire tracks, and water coverage, often resemble or interfere with damage patterns, making them difficult to accurately identify. This background noise causes the model to misclassify the background as a damage site or ignore actual damage, thereby affecting detection accuracy. Furthermore, background interference increases the difficulty of feature extraction, further lowering the reliability and efficiency of the detection system [19]. Irregular damage shapes pose an additional challenge for feature extraction. Cracks are influenced by various factors during formation, resulting in discontinuous, branching, or mesh-like structures. These irregular shapes make it difficult for the model to extract consistent features, as feature information may be blurred or lost, especially during the deeper stages of feature extraction. The complexity of damage shapes further exacerbates the difficulty of model in capturing fine details, impacting the accuracy and robustness of detection [20].
To address the aforementioned challenges, we propose an improved detection framework based on YOLOv8 [21]. First, a Star Operation Module (SOM) [22] is integrated into the third and fourth stages of the backbone network. By projecting features into a high-dimensional space, this module retains the feature extraction capability for small objects (such as fine road cracks), thereby enhancing the performance of model in detecting small objects. Next, a Multi-dimensional Auxiliary Fusion (MAF) [23] module is introduced in the neck of the network. This module incorporates the multi-scale and Attention Refinement Module (ARM) to further fuse features from multiple dimensions, enhancing the representation ability of the features and improving the robustness of feature representation in complex backgrounds. Finally, Wavelet Transform Convolution (WTC) [24] replaces traditional convolution in the head of the network. By using wavelet transforms to decompose the input features into multi-frequency components, this method effectively expands the receptive field and focuses on low-frequency shape information, thereby improving the ability of model to capture irregular damage patterns.
To briefly summarize our contributions, we present them in five key aspects:
(1) We propose a novel road damage detection model that combines the Star Operation Module (SOM), Multi-dimensional Auxiliary Fusion (MAF) module, and Wavelet Transform Convolution (WTC) module to address the challenges of detecting road damage in complex scenarios.
(2) To improve detection of small-scale damage, such as fine cracks, we introduce the SOM. This module enhances the representation of small object features by utilizing high-dimensional feature projections.
(3) To tackle the challenges posed by complex backgrounds and feature interference, we propose the MAF module, which incorporates an Attention Refinement Module (ARM) to enhance feature robustness and discriminative power.
(4) To handle irregular damage shapes, we introduce the WTC, which decomposes input features into multi-frequency components, expanding the receptive field and improving the ability of model to capture low-frequency shape information.
(5) The proposed method was validated on public road damage datasets, achieving an average accuracy of 25.75%, making it a powerful and efficient solution for road damage detection in complex environments.

2. Related Work

2.1. One-Stage Detectors

One-stage object detectors, such as the YOLO and SSD series, prioritize real-time performance by predicting object classes and bounding box coordinates in a single forward pass, bypassing the region proposal step used in two-stage detectors. This architecture makes one-stage detectors highly efficient for real-time applications like autonomous driving and video surveillance. Single Shot MultiBox Detector (SSD) [17] leverages multi-scale feature maps and default anchor boxes to balance detection speed and accuracy, making it suitable for applications requiring efficient and lightweight solutions. YOLOv4 [25] introduced CSPNet for optimized feature learning. YOLOv6 [26] further improved performance by incorporating re-parameterizable backbones and Rep-PAN neck designs, boosting speed and accuracy. YOLOv8 brought additional advancements, including anchor-free detection mechanisms, dynamic task prioritization, and architectural refinements for robust and flexible real-time performance. YOLOv10 [27] took the field a step further by eliminating the non-maximum suppression (NMS) process, replacing it with a consistent dual assignment approach, ensuring both low latency and reliable detection. However, despite their speed advantage, one-stage detectors often face challenges in detecting small objects or handling complex backgrounds due to the absence of a refined region proposal mechanism.

2.2. Two-Stage Detectors

Two-stage detectors are designed to provide high accuracy, particularly in tasks that require precise object localization. R-CNN [28] introduced the use of selective search for region proposals and CNNs for classification and bounding box regression, but it was computationally inefficient due to redundant feature extraction. Fast R-CNN [29] improved upon this by using shared convolutional feature maps and RoI pooling, speeding up the process, though it still relied on external region proposal generation. Faster R-CNN [30] further advanced the model by introducing the Region Proposal Network (RPN), allowing the network to generate proposals internally and making the entire pipeline end-to-end, thus improving both speed and accuracy. Mask R-CNN [31] extended Faster R-CNN by adding a segmentation branch for pixel-level object masks, making it suitable for instance segmentation but increasing computational cost. Cascade R-CNN [32] refined proposals across multiple stages with progressively higher IoU thresholds, enhancing detection precision, especially for challenging objects, but at the cost of added computational complexity. While two-stage detectors excel in accuracy, their computational demands often make them less suitable for real-time applications compared to one-stage models.

2.3. Similar Works

Many researchers have already similar work on solving the problem of road damage detection [33,34,35]. Du et al. [36] propose a lightweight object detection algorithm with enhanced feature extraction based on the You Only Look Once (YOLO) algorithm. The Bidirectional Feature Pyramid Network (BIFPN) network structure is used for multi-scale feature fusion to enhance the feature extraction ability, and Varifocal Loss is used to optimize the sample imbalance problem, which improves the detection accuracy of road damage. Li et al. [37] introduce a novel approach for identifying and locating road damage based on an enhanced ML-YOLO algorithm. By refining the YOLOv8 object detection framework, they optimize both the traditional convolutional layers and the spatial pyramid pooling network. Despite significant advancements in road damage detection, many challenges remain, particularly in detecting small cracks and handling complex road backgrounds. These issues arise from the difficulty of detecting small objects and the need for robust feature representation to deal with diverse and irregular damage shapes. As a response, we propose the YOLO-RD model, which builds upon these existing approaches and introduces novel components to address the aforementioned challenges.
Starting with Feature Pyramid Networks (FPNs) [38], feature fusion strategies in object detection have undergone several advancements. An FPN uses a top-down pathway with lateral connections to pass high-level semantic information to lower layers, enhancing multi-scale detection. Later, the Asymptotic Feature Pyramid Network (AFPN) [39] introduced Asymptotic Feature Fusion (AFF), which progressively guides feature fusion to optimize multi-scale information interaction. Path Aggregation Networks (PANets) [40] extended FPNs by adding a bottom-up path augmentation, strengthening information flow from low-level features. CM-Net [41] focused on arbitrary-shaped text detection with a concentric feature fusion strategy for better local–global feature representation. Finally, EfficientDet [42] proposed the Bidirectional Feature Pyramid Network (BiFPN), which adaptively learns the importance of different feature levels using weighted connections, improving efficiency and detection accuracy. In contrast, our Multi-dimensional Auxiliary Fusion (MAF) module goes beyond traditional feature fusion strategies by introducing a more flexible and adaptive approach to integrate information across different feature dimensions. Unlike the static hierarchical fusion seen in FPN or BiFPN, the MAF module employs an Attention Refinement Module (ARM) to dynamically refine feature maps at different scales, ensuring that the most relevant features are emphasized while less important ones are suppressed. This adaptive fusion process enhances feature robustness and discriminability, which is particularly crucial for detecting small cracks and handling complex backgrounds in road damage detection.
Furthermore, the Star Operation Module (SOM) is designed to specifically address the challenge of small object detection. SOM enhances the representation of small object features by leveraging high-dimensional feature projections. It achieves this by utilizing star-shaped convolutions that capture fine-grained details, which are often overlooked in traditional convolutional layers. This module helps improve the detection of small road cracks, which are typically difficult for conventional methods to accurately detect. Additionally, to tackle the issue of irregular damage shapes, we introduce the Wavelet Transform Convolution (WTC) module, which decomposes input features into multi-frequency components, expanding the receptive field and improving the model’s ability to capture low-frequency shape information.

3. Methodology

3.1. Architecture of the Proposed Method

To address the challenges of road damage detection, including detection of small objects, complex background interference, and irregular damage shapes, we propose YOLO-RD, inspired by the architecture of YOLOv8. As shown in Figure 2, the overall architecture of YOLO-RD highlights the key components and modifications designed to tackle these challenges. While the original YOLOv8 model excels in general object detection tasks, it faces difficulties in meeting the specific demands of road damage detection. To address these limitations, YOLO-RD introduces several critical innovations, including the Star Operation Module (SOM) [22], Multi-dimensional Auxiliary Fusion (MAF), Wavelet Transform Convolution (WTC) [24], and Attention Refinement Module (ARM) [23]. These enhancements aim to improve the robustness of model and accuracy in road damage detection. The YOLO-RD framework comprises three main components: SOM-Backbone, MAF-PAFPN, and WT-Head.
In the YOLO-RD model, the features are first input into the backbone network, where they are subjected to four stages of feature extraction operations to obtain four output features. These four outputs are the inputs to the detection neck. In the detection neck, the features are first processed by MAF, and the four inputs are first processed by the ARM, and then they are spliced and adjusted from top to bottom through a series of channel dimensions, followed by a bottom-up and top-down feature fusion process. The detection neck outputs three features, which are also used as the input of the detection head. In the detection head, they are subjected to Wavelet Transform Convolution, and then classified and regressed, respectively.
The SOM-Backbone is an improved version of the YOLOv8 Backbone (CSPDarknet), which is designed to enhance small object detection in road damage detection tasks. CSPDarknet extracts low-level features such as edges, textures, and gradients through a split-and-merge strategy, reducing redundancy and improving computational efficiency. However, it has limitations in detecting small objects, as it struggles to retain fine-grained details, which affects the accuracy of road damage detection. To address this, we propose the SOM-Backbone. By applying element-wise multiplication, the Star Operation Module (SOM) enhances the representation of small-scale features, significantly improving the sensitivity of the network to minor road damage.
MAF-PAFPN, an enhanced version of PAFPN, addresses challenges in road damage detection by incorporating the Multi-dimensional Auxiliary Fusion (MAF) module. While PAFPN performs well in general object detection tasks, it faces significant challenges in road damage detection due to the complex and variable nature of road surfaces, which are often cluttered with occlusions and varying textures. MAF-PAFPN combines low-level spatial features with high-level semantic information through multi-scale feature aggregation and the Attention Refinement Module (ARM). The ARM refines feature maps by emphasizing critical regions, while multi-scale aggregation captures fine details across scales. This fusion improves robustness in cluttered environments and enhances detection accuracy for small road damage, such as cracks and potholes, in complex real-world conditions.
The Wavelet Transform Head (WT-Head) is the detection head of YOLO-RD, which is responsible for final object classification and bounding box regression. It processes features from the backbone and neck to predict object categories and locations. In road damage detection, especially for small and irregular cracks, the standard YOLOv8 head struggles with accurate detection and localization. To overcome this, WT-Head integrates Wavelet Transform Convolution (WTC), which enhances multi-scale feature extraction by decomposing feature maps into distinct frequency components. This enables the model to capture both fine details and broader structural patterns, improving detection accuracy for subtle road damage. In addition, WTC gradually reduces the receptive field of features in the process of using wavelet transform and inverse wavelet transform, and uses smaller convolution kernels for convolution operations, thereby reducing computational costs and promoting lightweight design of the model.

3.2. Star Operation Module

As illustrated in Figure 3, to address the challenge of detecting small cracks in road damage detection, we propose the Star Operation Module (SOM), a novel nonlinear feature fusion method. The SOM enhances the representation of small-scale features while preserving high-level semantic information, ensuring high detection accuracy. The module integrates depthwise convolutions, pointwise convolutions, element-wise multiplication, nonlinear activations, residual connections, and a final 1 × 1 convolution to refine feature maps effectively. The design concept of the SOM is analogous to stars in the universe, gradually bringing distant objects closer and focusing on small features. Imagine small cracks in an image as thin lines, easily drowned out by background noise. The * operation shown in Figure 3 represents element-by-element multiplication. The SOM extracts spatial details and semantic information of the cracks through two branches, and then dynamically fuses these features using element-wise multiplication, much like using a magnifying glass to focus on the crack area, making its features more prominent. Finally, through residual connections, the original feature information is preserved, ensuring that the details of the cracks are not lost in the deeper layers of the network. The transformation in SOM can be expressed as follows:
F fused = Conv 1 × 1 ( BN ( DWConv 1 ( F input ) ) ) ReLU 6 ( Conv 1 × 1 ( BN ( DWConv 1 ( F input ) ) ) )
F intermediate = DWConv 2 ( BN ( Conv 1 × 1 ( F fused ) ) ) + F input
F output = Conv 1 × 1 ( F intermediate )
where F input represents the input feature map. F fused represents the fused feature map, which is obtained by applying depthwise convolutions and pointwise convolutions, followed by element-wise multiplication and ReLU activation. This map enhances the spatial and semantic representation of small-scale features. F intermediate is the intermediate feature map after further refinement using depthwise convolution and residual addition with the original input features. Assume the input feature has dimensions H × W × C , where H and W denote the spatial dimensions and C is the number of channels. The input feature undergoes a series of transformations to enhance its representation while preserving spatial and semantic details. Firstly, the input feature map is processed by a Depthwise Convolution (DWConv1) to extract spatial features independently for each channel. The resulting feature map retains the same spatial dimensions H × W and channel count C. Next, Batch Normalization (BN) is applied to stabilize the mean and variance of the feature map, ensuring consistent training. The normalized feature map dimensions remain H × W × C . Following normalization, the feature map is divided into two branches. The first branch applies a Pointwise Convolution ( Conv 1 × 1 ), adjusting the channel count to 3 C . The second branch undergoes the same transformation, followed by a ReLU6 activation to introduce nonlinearity by clamping values to the range [ 0 ,   6 ] . The outputs of both branches are fused via element-wise multiplication (⊙), producing the enhanced feature map F fused with dimensions H × W × 3 C . Subsequently, the fused feature map passes through another Pointwise Convolution ( Conv 1 × 1 ), reducing the channel count back to C. This is followed by a Depthwise Convolution (DWConv2), which refines spatial features. The resulting feature map F intermediate retains dimensions H × W × C . To preserve the original spatial and semantic information, a residual connection adds F input to F intermediate . Finally, the enhanced feature map is refined through a last Pointwise Convolution ( Conv 1 × 1 ), adjusting the channel dimensions to C . This produces the final output feature map F output with dimensions H × W × C .

3.3. Multi-Dimensional Auxiliary Fusion

To tackle the challenge of small damage detection in road damage detection, especially in the presence of complex backgrounds, we propose the Multi-dimensional Auxiliary Fusion (MAF) module, integrated with an Attention Refinement Module (ARM). This innovation introduces an advanced strategy for feature fusion and dynamic attention adjustment, allowing the model to better focus on important features, such as small cracks, while suppressing irrelevant background information. As illustrated in Figure 2, the MAF module begins the fusion process early in the network, combining both low-level spatial features and high-level semantic features. By doing so, it ensures that crucial fine-grained details, which are essential for detecting small damage, are preserved throughout the network. The fusion process between different stages helps integrate multi-scale information, making it more effective for detecting subtle road damage.
The MAF module integrates features from multiple stages, dynamically adjusting the contribution of each stage based on its relevance to the detection task. The feature fusion process can be mathematically expressed as follows:
F MAF = i = 1 n α i F stage i , where α i R , i = 1 n α i = 1
where F stage i represents the feature map from each stage, and α i is the dynamic weight assigned to the feature map of each stage. These weights are learned during training, allowing the model to adaptively prioritize more relevant features for crack detection. This process ensures that low-level and high-level features are fused to enhance the sensitivity of model to small damage.
While MAF focuses on aggregating multi-scale features, the ARM fine-tunes the feature maps generated at different stages of the network by applying spatial refinement to enhance key regions. The ARM works specifically after the output of Stage 2, Stage 3, and Stage 4 feature maps, improving their representation by dynamically focusing on critical regions such as small cracks, and reducing the impact of irrelevant background features.
As shown in Figure 4, the Attention Refinement Module (ARM) plays a crucial role in refining feature maps through spatial feature adjustments to enhance the focus on critical regions such as small cracks, while suppressing irrelevant background features. ARM operates on feature maps generated from Stage 2, Stage 3, and Stage 4, where each stage contains feature maps with different levels of spatial and semantic information.
The ARM first generates spatial offsets Δ p k via a convolution operation, adjusting the receptive fields of the network to focus on important regions, such as small cracks. This process is mathematically represented as follows:
Δ p k = conv _ offset ( F stage i )
where F stage i represents the feature maps from different stages (Stage 2, Stage 3, Stage 4), and conv_offset is the convolution operation that generates the spatial offsets Δ p k .
Next, the ARM applies deformable convolution using the generated offset Δ p k , refining the feature maps to focus on discriminative regions while adjusting spatial locations. This process is mathematically expressed as follows:
F refined = deform _ conv ( F stage i + Δ p k )
where F refined is the feature map after deformable convolution, and Δ p k is the spatial offset generated in the previous step.
Following spatial refinement, the ARM further processes the adjusted feature map using a Sigmoid activation function to normalize the feature map, producing the final refined feature map F ARM . This step amplifies important regions (such as small cracks) and suppresses background noise, ensuring that key features are preserved. The final output is expressed as follows:
F ARM = sigmoid ( deform _ conv ( F stage i + Δ p k ) )
where F ARM represents the refined feature map, normalized using the Sigmoid function, with values constrained to the range [ 0 ,   1 ] .
Finally, the refined feature map F ARM is passed to the subsequent layers of the network for further processing. The MAF module effectively integrates multi-scale features from different stages of the network, ensuring the seamless fusion of low-level spatial features and high-level semantic features to retain critical details. Meanwhile, the ARM fine-tunes the feature maps at each stage, dynamically focusing on key regions associated with cracks and suppressing background noise. Together, these two modules address the challenges of detecting small cracks and handling diverse and irregular crack shapes in complex backgrounds. This synergy enhances the robustness and accuracy of model, providing an efficient and reliable solution for road crack detection under challenging real-world conditions.

3.4. Wavelet Transform Convolution

To improve the detection accuracy of irregular cracks in road damage detection, we propose the WT-Head, which modifies the original YOLOv8 detection head by replacing its standard convolution layers with the Wavelet Transform Convolution (WTC) module. As shown in Figure 2, the WT-Head consists of two main components: a classification head and a regression head. These components are supported by the WTC module, which replaces the standard convolutions in the detection head to enhance feature extraction by leveraging multi-frequency decomposition and processing.
As shown in Figure 5, The WTC module begins by applying the wavelet transform (WT) to the input feature map X, splitting it into four frequency bands: X L L (low-frequency, capturing global structural patterns) and X L H ,   X H L ,   and   X H H (high-frequency components capturing horizontal, vertical, and diagonal edge details), as shown in Figure 6. The low-frequency component (LL) contains the main structure and contour information of the image, which is suitable for extracting the overall contour and location information of the crack, but may be interfered with by background texture and shadow. The horizontal high-frequency component (LH) captures the horizontal edge and detail information of the image, which is suitable for detecting transverse cracks. The vertical high-frequency component (HL) reflects the vertical edge and detail information in the image, which is suitable for detecting longitudinal cracks. The diagonal high-frequency component (HH) reveals the diagonal edge and detail information in the image, which is suitable for detecting potholes or irregular crack networks. By decomposing the image into these components, the crack features in different directions can be analyzed in detail. Combining the information of multiple components can not only effectively enhance the expression of crack features, but also suppress interference factors such as background noise, shadows, and stains. This multi-scale and multi-directional information fusion method can significantly improve the accuracy and robustness of road crack detection, especially in the cases of complex backgrounds and irregular shapes, and can more effectively identify and locate cracks.
This process is mathematically represented as follows:
[ X L L , X L H , X H L , X H H ] = W T ( X ) = Conv ( [ f L L , f L H , f H L , f H H ] , X )
where the Haar wavelet filters f L L ,   f L H ,   f H L ,   f H H are used to perform the decomposition. After decomposition, each frequency band is processed independently by small-kernel convolutions to extract features specific to each band. For example, X L L provides a smoothed, global representation, while X L H ,   X H L ,   X H H focus on high-frequency edge and texture information. To further expand the receptive field, the WTC module employs a hierarchical decomposition approach. The low-frequency component X L L undergoes recursive decomposition into multiple levels:
X L L ( i ) , X L H ( i ) , X H L ( i ) , X H H ( i ) = W T ( X L L ( i 1 ) )
where i represents the decomposition level. This process progressively increases the receptive field while retaining the spatial resolution of features. After convolutional operations are applied at each level, the outputs are recombined through the inverse wavelet transform (IWT):
Z ( i ) = I W T Y L L ( i ) + Z ( i + 1 ) , Y L H ( i ) , Y H L ( i ) , Y H H ( i )
The final output Z ( 0 ) from the WTC module is seamlessly integrated into two distinct components for downstream processing. The Classification Head capitalizes on the low-frequency features ( X L L ) to effectively discern regions containing cracks by leveraging the global structural context encoded within these features. Conversely, the Regression Head harnesses the high-frequency components ( X L H ,   X H L ,   X H H ) to achieve precise delineation of crack boundaries, ensuring the accurate capture of intricate edge details and local textures. This dual-stream approach optimally exploits the hierarchical frequency information to balance global awareness with fine-grained localization.
Replacing the standard convolutions in YOLOv8 with WTC provides several advantages. The multi-frequency decomposition enables the model to efficiently separate and process both global and local information, improving detection performance on irregular cracks. The hierarchical decomposition significantly expands the receptive field without adding excessive computational overhead, ensuring both efficiency and accuracy. Moreover, the WTC module enhances the robustness of model against noise and varying background conditions, making it suitable for real-world road damage detection scenarios.

4. Experiment

4.1. Datasets and Evaluation Metrics

This paper utilizes the publicly available RDD2022 dataset [43], a multi-country collection of road images for automatic road damage detection, comprising 47,420 images from six countries: Japan, India, the Czech Republic, Norway, the United States, and China. After data cleaning, 7900 road images from Japan were selected for analysis. The Japanese dataset is particularly valuable for this study due to its diversity in both road types (e.g., urban and rural roads) and environmental conditions (e.g., sunny, rainy, and shaded conditions). These variations in rainfall and lighting angle present distinct challenges for road damage detection, making it an ideal representation of real-world scenarios. The dataset includes four types of damage: longitudinal cracks, transverse cracks, crocodile cracks, and potholes, with sample sizes of 4049, 3979, 6199, and 2243, respectively. In comparison, the Indian dataset is more limited, primarily featuring road images under consistent conditions such as clear weather and daytime lighting. This lack of environmental variety reduces its ability to reflect the complexities of road damage in different settings. As such, its usefulness in training models for more varied conditions, like those in Japan, is somewhat constrained. Additionally, the Norwegian dataset offers a unique challenge, with images taken in snowy, rainy, and low-light conditions, making it a valuable resource for testing the performance of model in harsh weather. Future work will prioritize improving detection accuracy in these challenging environments, particularly in low-light and snowy conditions, to further adapt the model to real-world road detection needs across different climates.
To demonstrate the effectiveness of YOLO-RD in road damage detection, we evaluate our model using several metrics, including mAP, mAP50, mAPs, mAPm, and mAPl. Mean Average Precision (mAP) measures the overall detection accuracy across all classes. mAP50 calculates the precision at an IoU threshold of 0.5. The metrics mAPs, mAPm, and mAPl assess detection precision for small, medium, and large objects, respectively, based on their area sizes: mAPs for objects smaller than 32 × 32 pixels, mAPm for those between 32 × 32 and 96 × 96 pixels, and mAPl for objects larger than 96 × 96 pixels. Giga Floating Point Operations Per Second (GFLOPS) evaluates the computational efficiency of model, while Parametersindicates the total number of parameters in the model.

4.2. Implementation Details

The proposed model and comparison models were implemented using the MMYOLO and MMDetection frameworks. YOLO-RD was trained on two RTX 3090 GPUs for 200 epochs, while the other models were trained using their default configurations. Both MMYOLO and MMDetection employed the SGD optimizer with a learning rate of 0.02. By default, the batch size was set to 16, and the confidence threshold for visualization was set to 0.5. The input and test image sizes for MMDetection were both 1333 × 800 pixels, while for MMYOLO, the input image size was set to 640 × 640 pixels.

4.3. Ablation Studies

As shown in Table 1, the ablation study highlights the significant improvements achieved by integrating the following three modules into the YOLOv8-based model, demonstrating clear advantages over the baseline YOLOv8-s [21].
Effect of the Star Operation Module (SOM). The introduction of the SOM significantly improves the detection performance of model, particularly in detecting small objects, which are often a critical challenge in road damage detection. Specifically, the overall mAP increases from 23.71 to 25.03, with small object detection (mAPs) improving from 12.73 to 14.32, medium object detection (mAPm) rising from 19.83 to 21.02, and large object detection (mAPl) increasing from 27.71 to 29.22. This improvement is primarily due to the high-dimensional feature projection capability of the SOM, which allows the model to better capture small-scale features like road cracks that are often difficult to detect. The SOM works by applying depthwise convolutions and pointwise convolutions, combined with element-wise multiplication, to enhance both spatial and semantic representations of small objects. By projecting features into a higher-dimensional space, the SOM enables the model to focus more effectively on subtle road damage, such as small cracks, which are often overshadowed by larger objects or background noise. This makes the model more sensitive to fine details, improving its ability to detect small cracks, which is a critical task in road surface analysis. At the same time, the SOM introduces a moderate increase in computational cost, with Flops growing from 14.27 G to 18.21 G and the number of parameters (Parameters) increasing from 11.14 M to 16.33 M. Despite this slight increase in complexity, the SOM offers a significant improvement in detection accuracy. The ability of the SOM to enhance small object detection, especially road cracks, proves to be highly valuable for road damage detection, where accurate identification of even the smallest defects plays a crucial role in road safety and maintenance.
Effect of the Multi-dimensional Auxiliary Fusion (MAF). The integration of the Multi-dimensional Auxiliary Fusion (MAF) module results in a significant enhancement in the performance of model. Specifically, the overall mAP increases from 23.71 to 25.26, with mAPs showing the most notable improvement, rising from 12.73 to 16.23. Additionally, mAPm increases from 19.83 to 21.34, and mAPl increases from 27.71 to 29.04. These improvements demonstrate MAF’s effectiveness in addressing the challenge of detecting small-scale objects like road cracks, especially in complex background scenarios where such features can be easily overlooked or overshadowed by larger objects or noise. MAF achieves this by aggregating multi-scale features and dynamically adjusting their contribution at different stages of the network. This fusion of low-level spatial features with high-level semantic features enables the model to retain critical details necessary for detecting small road damage while suppressing irrelevant background noise. The ability to adaptively prioritize important features, particularly those related to small cracks, results in significant improvements in mAPs. Regarding computational cost, the introduction of MAF leads to a moderate increase. Specifically, Flops rise from 14.27 G to 18.45 G, and the number of parameters increases from 11.14 M to 16.87 M. While this results in a slight increase in computational complexity, the trade-off between performance and complexity is well-balanced. The significant improvements in detection accuracy, particularly for small object detection, justify this increase. MAF’s ability to enhance model performance in challenging road crack detection scenarios makes it an essential component of YOLO-RD, significantly improving detection precision, especially in complex real-world environments.
Effect of the Wavelet Transform Convolution (WTC). The integration of the Wavelet Transform Convolution (WTC) module in the detection head results in notable improvements in the performance metrics of model. Specifically, the overall mAP increases from 23.71 to 24.04, with mAPs showing the most significant improvement, rising from 12.73 to 16.35, highlighting WTC’s effectiveness in enhancing the detection of small objects, such as small road cracks. This improvement is attributed to the ability of module to decompose feature maps into distinct frequency components, which allows the model to focus on both fine details and larger structures, improving its detection capabilities across multiple scales. Meanwhile, mAPm increases slightly from 19.83 to 20.16, and mAPl remains stable at 27.76, suggesting that while WTC excels at small object detection, its impact on medium and large objects is more modest. In terms of computational cost, the introduction of WTC leads to a reduction in both Flops and the number of parameters, making the model more computationally efficient. Specifically, Flops decrease from 14.27 G to 10.45 G, and the number of parameters drops from 11.14 M to 9.37 M. This reduction demonstrates that the WTC module not only maintains detection accuracy but also enhances computational efficiency by focusing on essential frequency components, rather than processing the entire feature map. The reduced computational overhead allows for faster inference, which is particularly beneficial for real-time road damage detection in resource-constrained environments. Overall, the integration of WTC provides a practical solution to the challenges of detecting irregular cracks in road surfaces, particularly small, subtle cracks that are often difficult to detect. The most noticeable improvements are observed in small object detection, confirming that the computational efficiency and performance enhancements achieved by WTC are well suited for real-world applications, making it an essential component in the YOLO-RD framework.

4.4. Comparisons

Comparison Experiments. Our proposed method was compared with several mainstream object detection models, including single-stage detectors (YOLOv5-s, YOLOv7-tiny [44], YOLOX-s, YOLOv8-s, YOLOv10-n [27]) and two-stage detectors (Faster R-CNN [30], Cascade R-CNN [32]). As shown in Table 2, among single-stage detectors, our mAP is 25.75, while YOLOv8-s achieves 23.71. Additionally, our mAPs is 17.66, whereas YOLOv8-s achieves 12.73. Although our model has a slight increase in computational cost, the significant improvements in mAP and mAPs demonstrate the clear advantage of our method in overall performance and small object detection. Compared to YOLOv10-n, our mAP is 25.75, while YOLOv10-n achieves 25.31. Similarly, our mAPs is 17.66, compared to YOLOv10-n’s 15.30. This indicates that our model not only outperforms YOLOv10-n in overall performance but also demonstrates superior performance in small object detection, with more balanced performance across objects of different scales. In comparison with the two-stage detector Cascade R-CNN, our mAP is 25.75, while Cascade R-CNN achieves 25.03. For mAPs, our model achieves 17.66, whereas Cascade R-CNN achieves only 12.82. Additionally, in terms of computational efficiency, our model has 20.89M parameters, which is significantly lower than the 69.16M parameters of Cascade R-CNN. This demonstrates that our method achieves comparable detection accuracy while being much more efficient, making it particularly suitable for practical deployment. These results indicate that our proposed method achieves an excellent balance between accuracy and efficiency, making it especially well suited for road crack detection tasks in complex and visually challenging environments. It can detect small, low-contrast cracks and subtle damage with high precision, ensuring reliable performance and establishing itself as a robust and practical solution for real-world road damage detection applications.

4.5. Visualization

As shown in Figure 7, the comparison between YOLOv8 (left) and the proposed YOLO-RD (right) highlights significant improvements in road damage detection. Three scenarios are analyzed: small damage detection (Left Group), complex background detection (Middle Group), and multi-scale damage detection (Right Group). Red boxes represent actual detected results, while green boxes indicate discrepancies between the two models.
The red boxes in Figure 7 represent road damage that can be detected by the YOLOv8 model, and the green boxes represent road damage that cannot be detected by YOLOv8 but can be detected by YOLO-RD. Note that the road damage marked by the red box can also be detected by YOLO-RD.
In the first row of the leftmost group, the cracks are thin and faint, making them difficult to detect due to their low visibility. YOLOv8 fails to recognize the full extent of the crack, as the red box shows partial detection. In contrast, YOLO-RD successfully captures the entire crack, as indicated by the complete bounding box (green), demonstrating its superior ability to detect low-contrast, small-scale cracks that are often overlooked. In the third row of the middle group, cracks are obscured by shadows cast across the road surface, significantly reducing their visual contrast. YOLOv8 struggles to distinguish the damages from the darkened regions, leading to missed detections or incomplete bounding boxes. In contrast, YOLO-RD accurately detects the cracks within the shadowed areas, as highlighted by the green boxes. This demonstrates the robustness of YOLO-RD in handling low-contrast conditions caused by environmental shadows, ensuring reliable detection even under visually challenging circumstances. In the first row of the rightmost group, cracks appear at varying scales, including wide, prominent fissures and thin lines. YOLOv8 detects the larger fissure but fails to identify the smaller crack, as shown by the absence of bounding boxes. YOLO-RD, on the other hand, successfully detects cracks of multiple scales, providing precise and tight bounding boxes. This highlights the capability of YOLO-RD to adapt to multi-scale variations, ensuring comprehensive detection.
The visualization results in Figure 6 demonstrate the clear advantages of YOLO-RD over YOLOv8 for road damage detection. Specifically, YOLO-RD excels in detecting fine cracks, resolving low-contrast interferences caused by shadows, and accurately identifying cracks of varying scales. These advancements position YOLO-RD as a robust and reliable solution for automated road condition monitoring, enabling accurate detection under diverse real-world conditions.

5. Discussion

This study introduces YOLO-RD, an innovative framework designed to address key challenges in road damage detection, such as small-scale damages, complex backgrounds, and irregularly shaped features. By integrating three core innovations—the Star Operation Module (SOM), Multi-dimensional Auxiliary Fusion (MAF), and Wavelet Transform Convolution (WTC)—YOLO-RD significantly enhances its capability to detect diverse road damages in challenging real-world conditions. These modules collectively improve the model’s ability to detect subtle, low-contrast, and multi-scale damages, as confirmed by extensive experiments. The model consistently outperforms baseline methods across critical metrics, including mAP and mAPs, demonstrating its potential as a reliable solution for automated road condition monitoring.
Despite the substantial success of YOLO-RD, several areas remain for further improvement. Future work will focus on the following directions:
1. Extending Data Collection and Analysis: To improve the generalization of the model, we will expand the dataset to include a wider range of damage types typical for both cement concrete and asphalt concrete pavements, addressing diverse geographical and material conditions. Specifically, we will collect data from different climatic regions and road types to include a broader variety of damage patterns and environmental contexts.
2. Adapting to Extreme Weather and Lighting Conditions: To increase the robustness of the model under harsh weather and low-light conditions, we plan to enhance the model’s performance in scenarios such as heavy rain, snowfall, and dim lighting, where road damage features become more difficult to detect. We will augment the dataset with images from these challenging environments and adjust the model’s training strategy to improve its capability to recognize low-contrast damage features, using techniques such as data augmentation to simulate rain and snow effects and improve low-light performance.
3. Improving Computational Efficiency: We aim to reduce computational complexity while maintaining model performance. This will involve applying various techniques, such as the following:
Model Pruning: Removing unnecessary connections or neurons to reduce computation and storage, enhancing inference speed.
Quantization: Converting floating-point computations to lower precision to reduce the computational load and memory usage.
Lightweight Design: Replacing the current architecture with more efficient models such as MobileNet or EfficientNet, ensuring high performance on resource-constrained devices without compromising detection accuracy.
These approaches will allow YOLO-RD to be deployed in real-time systems and on devices with limited computational resources, such as mobile devices and drones, without sacrificing performance.
4. Integrating Multi-Task Learning: Expanding the scope of YOLO-RD by integrating multi-task learning to handle additional road-related tasks, such as the following:
Crack Severity Assessment: Evaluating the severity of detected cracks based on features like width and depth to assess their potential impact on the road structure.
Road Type Classification: Incorporating a task to classify the road type (e.g., asphalt, concrete), enabling customized road maintenance strategies based on road material.
Repair Strategy Recommendation: Leveraging the detection results to automatically generate repair priority and maintenance strategies, facilitating more intelligent and effective road maintenance decisions.
Additionally, a key limitation of the current model is its performance in detecting extremely small cracks, particularly in low-contrast environments or when covered by significant background noise. Future improvements will focus on the following:
Enhancing Small Crack Detection: Adopting finer feature extraction techniques, such as local feature extraction and multi-scale convolutions, to improve the detection of microcracks that are often overlooked by current models.
Background Noise Suppression: Further optimizing the model’s performance in complex backgrounds by incorporating background noise suppression techniques or region-of-interest (ROI) extraction, ensuring that the model can better isolate and identify small cracks.
By addressing these areas, YOLO-RD can further enhance its performance and applicability in diverse real-world scenarios, making significant contributions to road safety and efficient pavement maintenance. This will result in a more reliable, efficient, and practical solution for automated road crack detection, even under challenging conditions.

Author Contributions

Conceptualization, W.W.; methodology, X.Y.; software, B.J.; validation, Z.T.; formal analysis, W.Z.; investigation, S.W.; resources, W.W.; data curation, Y.X.; writing—original draft preparation, X.Y.; writing—review and editing, W.W.; visualization, S.L.; supervision, L.Y.; project administration, L.Y.; funding acquisition, W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the following research projects: ‘Intelligent Pavement Defect Detection System Development’ (Grant No. 2024JBH05LU7) and ‘Development of an Intelligent Assisted Obstacle Avoidance System’ (Grant No. 2024JBH05LU5), both horizontal research projects; ‘Research on Key Technologies for Blind Guidance Based on Edge Computing’ (Grant No. 2024LY505L23), a key scientific research project of the Jilin Provincial Department of Education; and ‘Research on Key Technologies for Blind-Assisted Obstacle Avoidance Based on Deep Learning’ (Grant No. 2023RY014), a research project funded by the China University-Industry Collaboration Innovation Fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study, RDD2022, is publicly available at the following GitHub repository: https://github.com/sekilab/RoadDamageDetector.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tsubota, T.; Fernando, C.; Yoshii, T.; Shirayanagi, H. Effect of road pavement types and ages on traffic accident risks. Transp. Res. Procedia 2018, 34, 211–218. [Google Scholar] [CrossRef]
  2. Markowski, K. Road safety aspects in the management of road maintenance. Proc. Matec Web Conf. 2017, 122, 02009. [Google Scholar] [CrossRef]
  3. Guralnick, S.A.; Suen, E.S.; Smith, C. Automating inspection of highway pavement surfaces. J. Transp. Eng. 1993, 119, 1–12. [Google Scholar] [CrossRef]
  4. Liu, Y.; Sun, P.; Wergeles, N.; Shang, Y. A survey and performance evaluation of deep learning methods for small object detection. Expert Syst. Appl. 2021, 172, 114602. [Google Scholar] [CrossRef]
  5. Li, L.; Huang, W.; Gu, I.Y.H.; Tian, Q. Statistical modeling of complex backgrounds for foreground object detection. IEEE Trans. Image Process. 2004, 13, 1459–1472. [Google Scholar] [CrossRef] [PubMed]
  6. Iivarinen, J.; Visa, A.J. Shape recognition of irregular objects. In Proceedings of the Intelligent Robots and Computer Vision XV: Algorithms, Techniques, Active Vision, and Materials Handling; SPIE: Boston, MA, USA, 1996; Volume 2904, pp. 25–32. [Google Scholar]
  7. Tsai, Y.C.J.; Li, F. Critical assessment of detecting asphalt pavement cracks under different lighting and low intensity contrast conditions using emerging 3D laser technology. J. Transp. Eng. 2012, 138, 649–656. [Google Scholar] [CrossRef]
  8. Quan, W.; Wang, H.; Gai, Z. Spot vehicle speed detection method based on short-pitch dual-node geomagnetic detector. Measurement 2020, 158, 107661. [Google Scholar] [CrossRef]
  9. Cucchiara, R.; Piccardi, M.; Mello, P. Image analysis and rule-based reasoning for a traffic monitoring system. IEEE Trans. Intell. Transp. Syst. 2000, 1, 119–130. [Google Scholar] [CrossRef]
  10. Cavadas, F.; Smith, I.F.; Figueiras, J. Damage detection using data-driven methods applied to moving-load responses. Mech. Syst. Signal Process. 2013, 39, 409–425. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Liu, C. Real-Time Pavement Damage Detection with Damage Shape Adaptation. IEEE Trans. Intell. Transp. Syst. 2024, 25, 18954–18963. [Google Scholar] [CrossRef]
  12. Ren, M.; Zhang, X.; Chen, X.; Zhou, B.; Feng, Z. YOLOv5s-M: A deep learning network model for road pavement damage detection from urban street-view imagery. Int. J. Appl. Earth Obs. Geoinf. 2023, 120, 103335. [Google Scholar] [CrossRef]
  13. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  14. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo algorithm developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  15. Bharati, P.; Pramanik, A. Deep learning techniques—R-CNN to mask R-CNN: A survey. In Computational Intelligence in Pattern Recognition: Proceedings of CIPR 2019; Springer: Singapore, 2020; pp. 657–668. [Google Scholar]
  16. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  18. Wang, K.; Liu, M.; Ye, Z. An advanced YOLOv3 method for small-scale road object detection. Appl. Soft Comput. 2021, 112, 107846. [Google Scholar] [CrossRef]
  19. Mikhailenko, P.; Piao, Z.; Kakar, M.R.; Bueno, M.; Athari, S.; Pieren, R.; Heutschi, K.; Poulikakos, L. Low-Noise pavement technologies and evaluation techniques: A literature review. Int. J. Pavement Eng. 2022, 23, 1911–1934. [Google Scholar] [CrossRef]
  20. Hamidian, D.; Salajegheh, J.; Salajegheh, E. Damage detection of irregular plates and regular dams by wavelet transform combined adoptive neuro fuzzy inference system. Civ. Eng. J. 2018, 4, 305–319. [Google Scholar] [CrossRef]
  21. Sohan, M.; Sai Ram, T.; Reddy, R.; Venkata, C. A review on yolov8 and its advancements. In Proceedings of the International Conference on Data Intelligence and Cognitive Informatics, Tirunelveli, India, 18–20 November 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 529–545. [Google Scholar]
  22. Ma, X.; Dai, X.; Bai, Y.; Wang, Y.; Fu, Y. Rewrite the Stars. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–18 June 2024; pp. 5694–5703. [Google Scholar]
  23. Dai, X.; Chen, Y.; Xiao, B.; Chen, D.; Liu, M.; Yuan, L.; Zhang, L. Dynamic head: Unifying object detection heads with attentions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7373–7382. [Google Scholar]
  24. Finder, S.E.; Amoyal, R.; Treister, E.; Freifeld, O. Wavelet convolutions for large receptive fields. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 October 2024; Springer: Berlin/Heidelberg, Germany, 2025; pp. 363–380. [Google Scholar]
  25. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  26. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  27. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  28. Cheng, B.; Wei, Y.; Shi, H.; Feris, R.; Xiong, J.; Huang, T. Revisiting rcnn: On awakening the classification power of faster rcnn. In Proceedings of the European Conference on ComputerVision (ECCV), Munich, Germany, 8–14 September 2018; pp. 453–468. [Google Scholar]
  29. Girshick, R. Fast r-cnn. arXiv 2015, arXiv:1504.08083. [Google Scholar]
  30. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  31. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  32. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  33. Cao, W.; Liu, Q.; He, Z. Review of pavement defect detection methods. IEEE Access 2020, 8, 14531–14544. [Google Scholar] [CrossRef]
  34. Bhatt, P.M.; Malhan, R.K.; Rajendran, P.; Shah, B.C.; Thakar, S.; Yoon, Y.J.; Gupta, S.K. Image-based surface defect detection using deep learning: A review. J. Comput. Inf. Sci. Eng. 2021, 21, 040801. [Google Scholar] [CrossRef]
  35. Qi, Y.; Bai, M.; Li, Z.; Zhang, Z.; Wang, Q.; Tian, G. Study on intelligent recognition of urban road subgrade defect based on deep learning. Sci. Rep. 2024, 14, 28119. [Google Scholar] [CrossRef] [PubMed]
  36. Du, F.J.; Jiao, S.J. Improvement of lightweight convolutional neural network model based on YOLO algorithm and its research in pavement defect detection. Sensors 2022, 22, 3537. [Google Scholar] [CrossRef] [PubMed]
  37. Li, T.; Li, G. Road Defect Identification and Location Method Based on an Improved ML-YOLO Algorithm. Sensors 2024, 24, 6783. [Google Scholar] [CrossRef] [PubMed]
  38. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  39. Yang, G.; Lei, J.; Zhu, Z.; Cheng, S.; Feng, Z.; Liang, R. AFPN: Asymptotic feature pyramid network for object detection. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, HI, USA, 1–4 October 2023; pp. 2184–2189. [Google Scholar]
  40. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  41. Yang, C.; Chen, M.; Xiong, Z.; Yuan, Y.; Wang, Q. Cm-net: Concentric mask based arbitrary-shaped text detection. IEEE Trans. Image Process. 2022, 31, 2864–2877. [Google Scholar] [CrossRef] [PubMed]
  42. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10781–10790. [Google Scholar]
  43. Arya, D.; Maeda, H.; Ghosh, S.K.; Toshniwal, D.; Sekimoto, Y. RDD2022: A multi-national image dataset for automatic road damage detection. Geosci. Data J. 2024, 11, 846–862. [Google Scholar] [CrossRef]
  44. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
Figure 1. Visualization of road damage detection samples from the real world.
Figure 1. Visualization of road damage detection samples from the real world.
Sensors 25 01442 g001
Figure 2. Architecture of the proposed method.
Figure 2. Architecture of the proposed method.
Sensors 25 01442 g002
Figure 3. Star Operation Module.
Figure 3. Star Operation Module.
Sensors 25 01442 g003
Figure 4. Multi-dimensional Auxiliary Fusion.
Figure 4. Multi-dimensional Auxiliary Fusion.
Sensors 25 01442 g004
Figure 5. Wavelet Transform Convolution.
Figure 5. Wavelet Transform Convolution.
Sensors 25 01442 g005
Figure 6. Frequency decomposition.
Figure 6. Frequency decomposition.
Sensors 25 01442 g006
Figure 7. Visualization of YOLOv8 (left) and YOLO-RD (right) detection results.
Figure 7. Visualization of YOLOv8 (left) and YOLO-RD (right) detection results.
Sensors 25 01442 g007
Table 1. Ablation experiments.
Table 1. Ablation experiments.
ModelsmAPmAP50mAPsmAPmmAPlFlops/GParameters/M
YOLOv8-s (baseline)23.7152.9212.7319.8327.7114.2711.14
YOLOv8-s+SOM25.0355.2414.3221.0229.2218.2116.33
YOLOv8-s+MAF25.2655.4716.2321.3429.0420.1617.46
YOLOv8-s+WTC24.0453.6616.3520.1627.7610.459.37
YOLOv8-s+SOM+MAF25.6555.5316.2721.3129.5424.1022.65
YOLOv8-s+MAF+WTC24.6854.9616.2821.3530.0716.3315.70
YOLOv8-s+SOM+WTC25.1055.3216.4221.1929.3814.3914.57
YOLOv8-s+SOM+MAF+WTC25.7555.6217.6621.3729.5820.2820.89
Table 2. Comparison experiments.
Table 2. Comparison experiments.
ModelsBackbonemAPmAP50mAPsmAPmmAPlFlops/GParameters/M
YOLOv5-sYOLOv5CSPDarknet18.2344.0210.2515.6119.938.167.15
YOLOv7-tinyE-ELAN13.5136.216.9211.3615.226.786.15
YOLO X-sYOLOXCSPDarknet19.4346.1510.8616.0221.4713.549.06
YOLOv8-sYOLOv8CSPDarknet23.7152.9212.7319.8327.7114.2711.14
YOLOv10-nCSPDarknet25.153.713.7820.1429.538.252.69
Faster R-CNNResNet-5022.0152.19.817.928.1101.0028.29
Cascade R-CNNResNet-5025.0354.9512.8221.1631.44162.0069.16
OursCSPDarknet25.7555.6217.6621.3729.5820.2820.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Yu, X.; Jing, B.; Tang, Z.; Zhang, W.; Wang, S.; Xiao, Y.; Li, S.; Yang, L. YOLO-RD: A Road Damage Detection Method for Effective Pavement Maintenance. Sensors 2025, 25, 1442. https://doi.org/10.3390/s25051442

AMA Style

Wang W, Yu X, Jing B, Tang Z, Zhang W, Wang S, Xiao Y, Li S, Yang L. YOLO-RD: A Road Damage Detection Method for Effective Pavement Maintenance. Sensors. 2025; 25(5):1442. https://doi.org/10.3390/s25051442

Chicago/Turabian Style

Wang, Wei, Xiaoru Yu, Bin Jing, Ziqi Tang, Wei Zhang, Shengyu Wang, Yao Xiao, Shu Li, and Liping Yang. 2025. "YOLO-RD: A Road Damage Detection Method for Effective Pavement Maintenance" Sensors 25, no. 5: 1442. https://doi.org/10.3390/s25051442

APA Style

Wang, W., Yu, X., Jing, B., Tang, Z., Zhang, W., Wang, S., Xiao, Y., Li, S., & Yang, L. (2025). YOLO-RD: A Road Damage Detection Method for Effective Pavement Maintenance. Sensors, 25(5), 1442. https://doi.org/10.3390/s25051442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop