Next Article in Journal
Recent Advances in Polysaccharide-Based Electrospun Nanofibers for Food Safety Detection
Previous Article in Journal
An Enhanced Adaptive Kalman Filter for Multibody Model Observation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on Novel Surface Defect Detection Methods for Aeroengine Turbine Blades Based on the LFD-YOLO Framework

by
Wei Deng
1,
Guixiong Liu
1,* and
Jun Meng
2
1
School of Mechanical and Automotive Engineering, South China University of Technology, Guangzhou 510640, China
2
Jetech Technology Company, Shenzhen 518102, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(7), 2219; https://doi.org/10.3390/s25072219
Submission received: 14 February 2025 / Revised: 22 March 2025 / Accepted: 29 March 2025 / Published: 1 April 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
This study proposes a novel defect detection method to address the low accuracy and insufficient efficiency encountered during surface defect detection on aeroengine turbine blades (ATBs). The proposed approach employs the LDconv model to adjust the size and shape of convolutional kernels dynamically, integrates the deformable attention mechanism (DAT) to capture minute defect features effectively, and uses Focaler-CIoU to optimize the bounding box loss function of the detection network. Our approaches collectively provide precise detection of surface defects on ATBs. The results show that the proposed method achieves a mean average precision (mAP0.5) of 96.2%, an F-measure of 96.7%, and an identification rate (Ir) of 98.8%, while maintaining a detection speed of over 25 images per second. The proposed method meets the stringent requirements for accuracy and real-time performance in ATB surface defect detection.

1. Introduction

The aeroengine, a critical aircraft component, relies heavily on the quality of its turbine blades (ATBs) to ensure optimal performance and safety [1]. During operation, the engine is exposed to extreme conditions, including high temperatures, high pressures, and rotational speeds. In addition, the blades endure thermal stress, deformation forces, vibrational loads, aerodynamic forces, and centrifugal forces equivalent to 2000 times their weight during engine startup, speed changes during operation, and shutdown. These factors make ATBs highly susceptible to damage and failure, resulting in regular inspections to ensure their reliability and safety [2]. However, the large number and varying geometries of aeroengine blades pose significant challenges for defect detection. Traditional inspection methods are often time-consuming, labor-intensive, and not very reliable. Current visual inspection techniques for ATB defects can be broadly categorized into two different approaches: traditional manual inspection and deep learning-based visual inspection. Traditional methods strongly rely on human judgment, often guided by subjective assessments or referenced in aeroengine maintenance manuals. These approaches are inefficient [3,4] and susceptible to errors such as missed or false detections, compounded by inspector fatigue. In contrast, deep learning-based visual inspection methods leverage advanced endoscopic technology and artificial intelligence. By employing convolutional neural networks (CNNs) to extract features and using forward and backward propagation for defect detection and recognition, compared to traditional detection methods, the application of deep learning methods has significantly improved the efficiency and accuracy of ATB defect detection, while also enabling intelligent and automated detection processes [5,6,7]. In recent years, numerous researchers have continued to conduct in-depth studies on deep learning methods for ATB defect detection, making it a prominent research focus [8,9]. Meanwhile, in the field of surface defect detection, significant progress has also been made with deep learning methods based on various architectures. For example, Huang et al. [10] proposed SSA-YOLO, which significantly improved the detection accuracy of tiny defects on hot-rolled strip steel surfaces by introducing a channel attention mechanism. Cui et al. [11] developed SDDNet, combining feature retention blocks (FRB) and skip dense connection modules (SDCM) to address challenges such as large variations in defect textures and small defect sizes, achieving high-speed and high-precision surface defect detection. Usamentiaga et al. [12] systematically reviewed deep learning-based methods for metal surface defect detection, highlighting the advantages of the YOLO series algorithms in balancing real-time performance and accuracy.
However, the existing YOLO improvement methods are primarily designed for general industrial scenarios and face limitations when addressing the specific requirements of aeroengine turbine blade (ATB) surface defect detection, such as the diverse shapes and textures of damage and the presence of tiny defects. These limitations include the following: the insufficient adaptability of fixed-size convolution kernels, making it difficult to capture features of tiny defects on ATB surfaces; traditional IoU loss functions tend to optimize bounding boxes for larger targets, leading to reduced localization accuracy for tiny defects; the downsampling operations in YOLO reduce the resolution of deep feature maps, further exacerbating the dilution of features for tiny defects. These issues severely constrain the performance of YOLO in ATB surface defect detection and serve as the core motivation for the improvements proposed in this paper. To address the aforementioned challenges, we propose a novel surface defect detection method for ATBs based on the LFD (LDconv + Focaler-CIoU + DAT)-YOLO framework, which integrates YOLOv8, linear deformable convolution (LDconv), Focaler-CIoU, and the deformable attention mechanism (DAT). The proposed approach involves three innovative key approaches: replacement of the CBS module with Ldconv, optimization of the bounding box loss function, and integration of the deformable attention mechanism. Collectively, the proposed LFD-YOLO framework improves the level of accuracy and robustness of ATB defect detection, resulting in precise, in-situ, and automated inspection. It also ensures real-time performance, making it highly suitable for industrial applications.
The contributions of this paper are summarized as follows:
  • Proposal of an LFD-YOLO-based ATB surface defect detection method: a novel method for detecting surface defects on ATBs is proposed, employing the LFD-YOLO framework. This approach incorporates LDconv to adjust the size and shape of convolutional kernels dynamically, utilizes the deformable attention mechanism to learn minute defect features effectively, and includes the Focaler-CIoU module to optimize the bounding box loss function of the network. These innovations collectively provide precise and accurate detection of surface defects on ATBs.
  • Construction of an ATB detection framework: a specialized ATB detection framework is developed to capture defect images, including cracks, nicks, dents, and burns. The collected dataset is used to train the LFD-YOLO network, ensuring robust performance in defect identification.
  • Experimental validation: extensive experimental data demonstrate the effectiveness of the proposed method. The LFD-YOLO framework achieves a mean average precision (mAP0.5) of 96.2%, an F-measure of 96.7%, and an identification rate (Ir) of 98.8%. The system processes over 25 images per second, meeting the stringent requirements for accuracy and real-time performance in ATB surface defect detection.

2. Background Information

2.1. YOLO Series Network for ATB Surface Defect Detection

The YOLO (You Only Look Once) series of deep learning-based detection and recognition algorithms employ an independent convolutional neural network architecture to simultaneously classify objects and regress bounding box coordinates within the image. Known for their real-time performance, simplicity, efficiency, high accuracy, and strong generalization capabilities, these algorithms have been widely applied in turbine blade defect detection and other related areas [13,14,15,16,17,18,19,20,21]. For instance, Chen et al. [22] proposed a YOLOv4-based defect detection model for turbine blades. Their model utilizes spatial pyramid pooling (SPP) and the path aggregation Network (PANet) as their parameter aggregation methods [23]. By incorporating an SPP module, a CSPDarknet53 backbone, and a PAN, the YOLOv4 architecture can be optimized for rapid defect detection. Experimental data from a self-constructed blade defect dataset have demonstrated a mean average precision (mAP) of 0.9. Li et al. [24] introduced the intelligent defect detection model DDSC-YOLOv5s. This model integrates a deformable convolutional network to enhance feature extraction efficiency through depthwise separable convolution. K-means clustering is employed to optimize anchor box sizes, which improves the real-time detection capability for ATB defects. Experimental data from a self-built borescope blade defect dataset yield an mAP of 0.838, with only a 7.9% increase in computational load compared to the original YOLOv5. Liao et al. [25] developed an optimized YOLOv5 model with a BiFPN (Bidirectional feature pyramid network) structure for detecting and classifying defects on Si3N4 turbine blades. Experimental data from an enhanced dataset of blade X-ray defect images achieve an mAP of 0.986, with the detection speed as low as 16 milliseconds. In [26], a DBFF-YOLOv4-based algorithm is proposed to address the challenges of error-prone and time-consuming X-ray defect detection in turbine blades. This approach uses dual backbones for hierarchical defect feature extraction and introduces a novel connection method that incorporates all feature maps. The resulting defect detection system achieves an identification rate (Ir) of 96.7% on a self-constructed blade defect dataset. In [27], an optimized YOLOv7 model incorporating a convolutional block attention module and Alpha_GIOU (CBAM) is proposed. The dataset is preprocessed using gamma correction, and an improved CBAM is embedded into the model backbone. A branch is added to the CBAM channel attention module, and Alpha_GIOU replaces CIoU as the loss function. These modifications and additions significantly improve the real-time performance and visualization of ATB defects. Experimental work on a self-built blade defect dataset demonstrates an mAP of 0.961, representing a 1.0% improvement over the original YOLOv7 model.
As described above, the YOLO algorithm series employs a global convolutional neural network to process entire images, generating an efficient single-stage detection pipeline. These algorithms have been extensively applied in turbine blade defect detection due to their robustness and real-time performance. This study adopts and further optimizes the YOLO series framework to achieve precise and efficient surface defect detection for ATBs.

2.2. Other Common Networks Used for ATB Surface Defect Detection

In addition to the YOLO series networks, other deep learning networks such as Faster R-CNN [28,29], Mask R-CNN [30], and VFNet [31] have also been studied for ATB defect detection. In [32], a modified Faster R-CNN-based model was proposed to address the issues of tiny and sometimes discontinuous defects in blade surface defect detection. This model replaces RoI pooling with RoI align, incorporates feature pyramid networks (FPN) combined with ResNet50 for feature extraction, and improves the non-maximum suppression (NMS) algorithm. The model was tested on a self-constructed blade defect dataset, achieving a mean average precision (mAP) of 0.79, a 17.5% improvement over the original Faster R-CNN. Shang et al. [33] developed an enhanced Mask R-CNN-based method for ATB surface damage detection. This method constructs an enhanced Mask R-CNN network with three functionalities: damage pattern classification, damage localization, and damage region segmentation. It focuses on shallow texture information through a texture-focused multiscale feature fusion network (TFNet) [34] and introduces a balanced L1 loss (BL) to improve localization accuracy. Experimental data on a self-constructed simulated blade defect image dataset yielded an mAP of 0.604. Liu et al. [35] proposed a tiny defect detection network (MVFNet), which builds upon VFNet by using a spatial feature pyramid network (SFPN) for feature fusion. They introduced a spatial attention residual module (SAR) into the SFPN to capture global spatial information, thereby improving the detection accuracy of tiny defects on blade surfaces. Further tests on a blade defect dataset resulted in an mAP of 0.735, 7.7% higher than the original VFNet.
It is evident that networks such as Faster R-CNN, Mask R-CNN, and VFNet each have their unique characteristics and have been selected for use in research areas such as industrial defect detection. Compared to two-stage detection methods, the one-stage detection approach of the YOLO series significantly enhances detection speed by simplifying the object detection task into a single forward propagation, while maintaining high accuracy. This characteristic makes YOLO particularly suitable for ATB surface defect detection tasks, as this detection scenario places extremely high demands on real-time performance, requiring rapid response while ensuring detection accuracy.
The Experimental and Results sections are organized as follows: Section 3 describes the LFD-YOLO framework in detail. Section 4 presents the experimental setup, including the experimental data, to demonstrate the superior performance of the LFD-YOLO model, while Section 5 highlights the most significant advantages of the proposed model and its limitations.

3. Mechanism of ATB Surface Defect Detection Based on LFD-YOLO

3.1. Overall Framework for ATB Surface Defect Classification and Detection

Figure 1 illustrates the overall framework for ATB surface defect detection based on LFD-YOLO, which includes the following key modules: image preprocessing, feature extraction network, feature fusion network, loss function computation, defect classification, and defect information output. Specifically: (1) the image preprocessing module is designed to preprocess the input raw images by adjusting image attributes for enhanced visibility of defect features; (2) the DarkNet53 + C2f + LDconv + DAT feature extraction network is employed to extract morphological and positional defect features, as well as high-level semantic information, and outputs feature maps at different scale levels; (3) the Neck + LDconv + DAT feature fusion network utilizes the deformable attention mechanism to dynamically select sampling points in the image, focusing intensively on critical defect regions (this approach significantly reduces computational load while maintaining robust performance); (4) the Focaler-CIoU loss function computation module optimizes the IoU loss for multicategory regression samples using a linear interval mapping method, thereby improving the regression accuracy of the network; (5) the defect classification module classifies and displays the different types of ATB defects; (6) the defect information output module is used to output surface defect information related to ATBs.
Under harsh working conditions, ATBs are prone to various defects such as cracks, notches, dents, burns, tears, curled edges at the blade tip, material loss/chips, burn-through, bending, bulges, deposits, overlaps/laps, corrosion, carbon buildup, fractures, sulfidation, and coating loss. Among these, defects that require highly accurate detection include cracks, notches, dents, and burns [36]. Figure 2 presents typical surface defects in ATBs that require good detection methods; Figure 2a–d represent cracks, nicks, dents, and burns, respectively.

3.2. Design of an LFD-YOLO Network for ATB Surface Defect Detection

Figure 3 illustrates the network architecture of YOLOv8 [19], which consists of four main parts: input, backbone, neck, and head. Specifically, the input section employs mosaic-type data augmentation for images fed into the network, including adaptive image scaling, grayscale filling, and other preprocessing techniques; the backbone section comprises structures such as CBS, C2f, and SPPF, which extract image features through convolution and pooling operations; the Neck section utilizes the PAN–FPN structure, achieving feature map fusion at different scales via up-sampling, down-sampling, and feature concatenation; and the decoupled head structure of the head section decouples classification and regression processes. It includes positive and negative sample matching and loss computation and obtains the detection targets’ category and location information based on feature maps at different scales.
Based on YOLOv8 as the benchmark network, this paper proposes a novel network structure defined as LFD-YOLO (Figure 4). Unlike the classic YOLOv8 network, the LFD- YOLO network replaces the original CBS module with the LDconv module in the backbone section. It incorporates the deformable attention mechanism to improve the detection capability for tiny defects. In the neck section, the PAN–FPN structure is combined with the LDconv + DAT deformable attention mechanism to fuse multiscale features and improve the detection of various types of ATB surface defects. The target category and location information are obtained from feature maps at different scales. The head section employs Focaler-CIoU to optimize the bounding box loss function, which improves the accuracy of bounding box regression.
The following sections describe the mechanisms and characteristics of three core innovations: the introduction of LDconv to replace CBS, the addition of the DAT deformable attention mechanism, and the optimization of the bounding box loss function using Focaler-IoU.

3.2.1. Replacement of CBS by LDconv

Convolution operations are widely used in deep learning and other fields for extracting image features. However, standard convolution operations have several limitations: they are confined to local windows, making it difficult to capture information from other locations; moreover, the fixed shape of convolution kernels results in a parameter count that is proportional to the kernel size [37]. To address these issues, this paper employs a linear deformable convolution (LDConv) algorithm [38] that replaces the CBS module containing standard convolution in the YOLOv8 network (as shown in the light-yellow modules in the backbone and neck sections of Figure 4). LDConv provides an arbitrary number of parameters and sampling shapes for the convolution kernels, which enables adaptation to target variations, reduces network overhead, and improves the detection of tiny defects.
Figure 5 illustrates the structure of LDconv (using a kernel size of N = 5 as an example) [38]. First, the initial sampling coordinates P0 for the convolution kernel need to be generated. Algorithm 1 presents the Python 3.11 pseudocode for the P0 generation algorithm [38]. Regular or irregular coordinates are generated by calculating the base integers of the convolution kernel, followed by coordinate concatenation and other steps to obtain the corresponding sampling coordinates (P0 + Pi) for any point Pi on the feature map. For the sampling grid R and convolution parameters w, the corresponding convolution operation is expressed as follows:
C o n v P i = P 0 R w × P 0 + P i
Next, a Conv2d convolution operation is performed on the input image, from which the kernel offset Coffset is derived. This offset is then used to adjust the original coordinates, enabling dynamic reshaping of the convolution kernel. Based on the adjusted sampling shape, the feature map undergoes resampling (resample). The resampled feature map is then reshaped, convolved again, normalized, and finally passed through the SiLU activation function to output the resulting feature map.
Algorithm 1 Pseudocode for Initial Coordinate Generation for Convolution Kernels in a PyTorch-like Approach [38]
# func get_p_o(num_param, dtype)
# num_param:the kernel size of LDConv
# dtype:the type of data

####### function body ########
# get a base integer to define coordinate
base_int = round(math.sqrt(num_param))
row_number = num_param//base_int
mod_numer = num_param % base_int

# get the sampled coordinate of regular kernels
p_o_x,p_o_y = torch.meshigrid(
torch.meshgrid(0, row_numb)
torch.meshgird(0, base_int))

# flatten the sampled coordinate of regular kernels
p_o_x = torch.flatten(p_o_x)
P_o_y = torch.flatten(p_o_y)

# get the sampled coordinate of irregular kernels
If mod_number > 0:
mod_p_o_x, mod_p_o_y = torch.meshgird(
torch.arange(row_number,row_number+1,torch.arange(0,mod_number))

mod_p_o_x = torch.flatten(mod_p_o_x)
mod_p_o_y = torch.flatten(mod_p_o_y)
P_o_x,p_o_y = torch.cat((p_o_x,mod_p_o_x)),torch.cat((p_o_y,mod_p_o_y))

# get the completed sampled coordinate
p_o = torch.cat([p_o_x, p_o_y], 0)
p_o = p_o.view(1, 2 * num_param, 1, 1).type(dtype)
return p_o

3.2.2. Incorporation of the DAT Deformable Attention Mechanism

Some defects on the ATB surface are small but occur in high numbers. Using conventional attention mechanisms to process all pixels in the image would result in high computational costs. To address this, we introduce the deformable attention transformer mechanism to optimize the feature extraction and feature fusion networks of YOLOv8. DAT dynamically selects sampling points, which allows the network to focus on important features, thereby reducing computational load while improving the accuracy of ATB surface defect detection [39]. By adding DAT to the final layer of the backbone feature extraction network (as shown in the orange module in the backbone section of Figure 4), the network can flexibly control the intensity and scope of attention based on the input image features and better capture subtle features. In addition, the network’s ability to perceive and fuse deep discriminative features and high-level semantic information is significantly improved by incorporating DAT after the C2f module in the down-sampling stage of the neck feature fusion network (as shown in the orange module in the neck section of Figure 4).
Figure 6 illustrates the structure of the deformable attention transformer [34]. The DAT architecture adopts a 4-stage hierarchical framework. As the stages progress, the spatial dimensions of the feature maps are split in half, while the number of channels doubles. During the initialization phase, a 4 × 4 convolution is applied to reduce the image size to one-fourth of its original dimensions, thereby reducing the computational load. In stages 1–2, local information is integrated with shift-window attention to better capture local features. In stages 3–4, the deformable attention (DA) module is added to simulate the relationship between local and global contexts. This type of structure balances the need for local and global recognition, thus effectively reducing computational costs while improving both accuracy and efficiency.
Figure 7 illustrates the flowchart of the deformable attention (DA) module. The process includes the following steps:
(1)
For an input feature map x with dimensions H × W × C, the predefined scaling factor is α . Then, the down-sampling expression for the original grid can be represented as follows (as indicated by the process labeled ① in Figure 7):
H G = H α ,   W G = W α
From the down-sampled HG × WG × 2 uniform grid, the original reference points Pre are selected, where the values of the reference points are linearly spaced coordinates {(0, 0), …, (HG − 1, WG − 1)}.
(2)
If the projection weight for the query is defined as Wq, then by linearly projecting the feature map x onto the query tokens, the query vector q is obtained. Then, through the offset generation subnetwork (offset network), the deformation offset Δ p is generated (as indicated by the process labeled ② in Figure 7):
Δ p = θ offset ( q ) ,   q = x W q
The deformed reference points Pde are calculated by combining the original reference points Pre and the offset Δ p . Bilinear interpolation is then applied to Pde to obtain the sampled feature map x ˜ from x (as indicated by the process labeled ③ in Figure 7):
x ˜ = ϕ ( x ; P de ) ,   P de = P r e + Δ p
(3)
The vectors k ˜ and v ˜ are obtained by performing linear projections on x ˜ in the directions of the key tokens and value tokens, with projection weights Wk and Wv, respectively (as indicated by the process labeled ④ in Figure 7):
k ˜ = x ˜ W k ,   v ˜ = x ˜ W v
(4)
The multi-head attention module integrates the outputs. If the softmax function is denoted as σ(·), then the head size of the multi-head attention is dhead, and the query, key, and value vectors for the mth attention head are q ( m ) , k ˜ ( m ) , and v ˜ ( m ) , respectively, the bilinear interpolation operation for the relative position bias is denoted as ϕ ( B ^ ; R ) , and the output z(m) of the multi-head attention can be expressed as follows (as indicated by the process labeled ⑤ in Figure 7):
z ( m ) = σ q ( m ) k ˜ ( m ) / d head + ϕ ( B ^ ; R ) v ˜ ( m )

3.2.3. Optimization of the Focaler-IoU Bounding Box Loss Function

In defect detection, the intersection over union (IoU) is commonly used as a critical parameter to evaluate the performance of detection networks [13]. If the predicted bounding box and the ground-truth bounding box are denoted as B and Bgt, respectively, the IoU can be calculated using the following equation:
I o U = B B g t B B g t
The IoU serves as the criterion for distinguishing positive and negative samples during the training phase and as the loss function for the detection network. It is also utilized in non-maximum suppression (NMS) during the inference stage. However, when the IoU equals zero, it becomes difficult to accurately characterize the positional relationship between two bounding boxes, leading to stalled gradient updates and hindering further learning and optimization of the network.
The traditional CIoU loss function exhibits significant bias toward large targets, resulting in insufficient localization accuracy for small defects. This bias arises because CIoU assigns equal weight to all targets, neglecting the optimization for smaller ones. To address this issue, we introduce Focaler-CIoU. In the YOLOv8 network, the CIoU is employed as the bounding box regression loss. If the Euclidean distance between the center points of B and Bgt is defined as ρ 2 B , B g t , the weight parameter is ε , the aspect ratio correction factor is vAR, the diagonal distance of the minimum bounding box enclosing B and Bgt is cMBB, the width and height of B are wB and hB, and the width and height of Bgt are defined as wB−gt and hB−gt, respectively, the CIoU and its loss function LCIoU can be calculated using the following expressions:
CIoU = IoU ρ 2 B , B g t c M B B 2 ε v A R ; L C I o U = 1 CIoU ε = v A R ( 1 IoU ) + v A R ; v A R = 4 π 2 arctan w B g t h B g t arctan w B h B 2
To better focus on different regression samples and improve the performance of the detection network across various detection tasks, this study replaced CIoU in YOLOv8 with IoUFocaler [40]. The equation for IoUFocaler is as follows:
IoU F o c a l e r = 0 , IoU < d F o c a l e r IoU d F o c a l e r u F o c a l e r d F o c a l e r , d F o c a l e r IoU u F o c a l e r 1 , IoU > u F o c a l e r
where [ d F o c a l e r , u F o c a l e r ] [ 0 ,   1 ] . By adjusting the values of d F o c a l e r and u F o c a l e r , IoUFocaler is a better fit for different regression samples. Therefore, the Focaler-IoU bounding box optimization loss function LFocaler-CIoU proposed in this paper is expressed by Equation (10), which, after combining it with the DFL bounding box loss function LDFL, gives the optimized bounding box regression loss function LB_box as follows:
L Focaler - CIoU = L CIoU + IoU IoU Focaler ,   L B _ box = L D F L + L Focaler - CIoU
The effectiveness and extent of these improvements described in the above sections were validated in subsequent experiments, as presented in the following section.

4. Experimental Design and Results Analysis

4.1. Experimental Design

Figure 8 illustrates the ATB surface defect detection system developed in this study. The system hardware includes an endoscope (OLYMPUS Corporation, Beijing, China) and a deep learning computer equipped with an AMD 5950x CPU and an NVIDIA GeForce RTX 3090 24 GB GPU. The software is designed to display surface defect information of ATBs.
The experimental dataset comprises 600 images (with a resolution of 1024 × 768 pixels), captured under the lighting conditions provided by the endoscope’s built-in lighting system to minimize background noise. The images include four types of defects on ATB surfaces: cracks, nicks, dents, and burns. The distribution of these defects in the dataset is as follows: cracks (20%), nicks (20%), dents (30%), and burns (30%). To enhance the diversity and robustness of the dataset, data augmentation techniques were employed, including random rotation (±30°), scaling (0.8× to 1.2×), horizontal and vertical flipping, and brightness adjustment (±20%). The dataset is divided into a training set of 480 samples, and validation and test sets of 60 samples each. The initial learning rate is set to 0.01, the batch size to 8, and the number of training epochs to 500. The ATB surface defect detection network is trained using the PyTorch 2.0.1 framework to identify and localize defects.
The experimental work employs mAP0.5, F-measure, recognition rate (Ir), and FPS as evaluation metrics for detection performance, where:
(1)
mAP0.5 is defined as the mean average precision at an IoU threshold of 0.5 [41]:
m A P 0.5 = i = 1 k A P 0.5 i k ,
where k is the number of categories and A P 0.5 i represents the average precision of the ith category at an IoU threshold of 0.5.
(2)
F-measure is defined as the weighted harmonic mean of precision and recall [42]:
F - m e a s u r e = ( 1 + α p r ) × P r e c i s i o n × R e c a l l α p r × P r e c i s i o n + R e c a l l ,
where αpr represents the weighting coefficient between precision and recall. In this experiment, αpr is set to 0.8.
(3)
Ir denotes the recognition rate, defined as follows [41]:
I r = N defect M defect ,
where Ndefect and Mdefect represent the number of correctly detected defects and the total number of defects in the sample images, respectively.
(4)
FPS represents the number of images detected by the network per second [13]:
F P S = N u m frame T i m e
where Numframe is the total number of detected images and Time is the detection time. Real-time performance is one of the critical metrics for industrial inspection systems. In this study, based on the practical requirements of ATB defect detection, we define real-time performance as the system’s ability to process no fewer than 24 FPS.

4.2. Experimental Data Analysis of the LFD-YOLO Network Performance

4.2.1. Comparative Experimental Data on the Network Performance

To validate its effectiveness, we quantitatively compared the proposed LFD-YOLO network with other classical object detection networks, including SSD [43], Faster R-CNN [28], Mask R-CNN [30], YOLOv3 [44], YOLOv5 [16], YOLOv7 [18], YOLOv7-tiny [45], YOLOv8 [19], and YOLOv11 [46]. Among them, SSD is a classic model of single-stage detection methods, offering high detection speed and acceptable accuracy. Faster R-CNN and Mask R-CNN, as representatives of two-stage detection methods, are widely used in the field of object detection. Their high precision makes them important references for comparative experiments. The YOLO series models, with continuous iterations, have shown improved detection performance and are also included as references in comparative experiments. Table 1 presents the quantitative analysis results of these networks. Figure 9 illustrates a comparison of prediction results across different networks (with SSD and Faster R-CNN selected as representatives of single-stage and two-stage networks, respectively, and YOLOv11 chosen as a representative of the YOLO series, compared against our LFD-YOLO network).
The following information can be extracted from the data presented in Table 1:
(1)
The SSD network achieved an FPS of 35.6 frames per second (f·s−1), demonstrating good real-time performance. However, its mAP0.5 of 61.2% and F-measure of 60.9% indicate relatively low accuracy during ATB defect detection. Similarly, the YOLOv7-tiny network achieved an FPS of 47.8 f·s−1, showing excellent real-time performance, but its mAP0.5 of 77.1% and F-measure of 76.8% suggested suboptimal ATB defect detection accuracy.
(2)
The Faster R-CNN, Mask R-CNN, YOLOv3, YOLOv5, YOLOv7, YOLOv8, and YOLOv11 networks achieved mAP0.5 values of 73.7%, 76.3%, 77.5%, 86.8%, 89.1%, 89.7%, and 91.4%, respectively, and F-measure values of 74.1%, 78.2%, 76.7%, 86.9%, 88.7%, 90.3%, and 91.9%, respectively. These networks exhibited moderate to good real-time performance and reached a relatively high level of accuracy for ATB surface defect detection.
(3)
The LFD-YOLO network achieved an FPS of 25.4 f·s−1 and an mAP0.5 of 96.2%, demonstrating the highest accuracy for ATB defect detection while meeting real-time performance requirements, showcasing significant advantages.
In summary, the proposed LFD-YOLO network for ATB surface defect detection outperforms other classical networks in terms of accuracy.

4.2.2. Ablation Study

Ablation study I: to validate the effectiveness of Focaler-CIoU in ATB detection, an ablation study was conducted based on the improvements of YOLOv8 + LDconv + DAT, comparing it with EIoU [47], WIoU [48], and MPDIoU [49]. Table 2 presents the performance metrics comparison for ablation study I.
From Table 2, it can be observed that Focaler-CIoU performs exceptionally well in ATB defect detection. The LFD-YOLO network achieves superior mAP0.5, F-measure, and Ir metrics compared to EIoU, WIoU, and MPDIoU. In terms of real-time performance, the four IoU improvements—EIoU, WIoU, MPDIoU, and Focaler-CIoU—have varying degrees of impact on the network, but all still meet the real-time performance requirements for ATB defect detection.
Ablation study II: to verify the optimal placement of the DAT module, based on the improvements of YOLOv8 + LDconv + Focaler-CIoU, the DAT module was separately added to the end of the backbone and the neck, corresponding to YOLOv8 + LDconv + Focaler-CIoU + DAT (backbone) and YOLOv8 + LDconv + Focaler-CIoU + DAT (neck) in the table. An ablation study was conducted by comparing these configurations with the LFD-YOLO network. Table 3 presents the performance metrics comparison for ablation study II.
The proposed LFD-YOLO network achieves improvements in mAP0.5, F-measure, and Ir by 1.5%, 1.8%, and 3.6%, respectively, compared to adding the DAT module solely to the end of the backbone (94.7% → 96.2%, 94.9% → 96.7%, 95.2% → 98.8%). Similarly, improvements of 1.1%, 2.2%, and 2.8% are observed compared to adding the DAT module solely to the end of the neck (95.1% → 96.2%, 94.5% → 96.7%, 96.0% → 98.8%). These results demonstrate that adding the DAT module to both the backbone and the neck achieves the best performance, effectively enhancing the accuracy of ATB surface defect detection.
Ablation study III: to validate the effectiveness of each module in the proposed LFD-YOLO, an ablation study was conducted based on the YOLOv8 network. Table 4 presents the performance metrics of LFD-YOLO obtained from the ablation experiments.
From the results shown in Table 2, we obtained the following information:
(1)
The YOLOv8 + LDconv network shows improvements in mAP0.5, F-measure, and Ir by 3.1%, 2.9%, and 5.1%, respectively (89.7% → 92.8%, 90.3% → 93.2%, 87.9% → 93.0%), while the FPS decreases by 1.9 f·s−1 (29.1 f·s−1 → 27.2 f·s−1). This indicates that replacing CBS with LDconv effectively improved the accuracy of ATB surface defect detection, with a minor impact on real-time performance.
(2)
The YOLOv8 + DAT network demonstrates improvements in mAP0.5, F-measure, and Ir by 4.8%, 4.9%, and 8.4%, respectively (89.7% → 94.5%, 90.3% → 95.2%, 87.9% → 96.3%), while the FPS decreases by 1.5 f·s−1 (29.1 f·s−1 → 27.6 f·s−1). This suggests that incorporating the DAT deformable attention mechanism significantly increases the accuracy of ATB surface defect detection with a minimal impact on real-time performance.
(3)
The YOLOv8 + Focaler-CIoU network achieves improvements in mAP0.5, F-measure, and Ir by 2.4%, 1.2%, and 3.8%, respectively (89.7% → 92.1%, 90.3% → 91.5%, 87.9% → 91.7%), while the FPS decreases by 1.0 f·s−1 (29.1 f·s−1 → 28.1 f·s−1). This trend indicates that optimizing the bounding box loss function with Focaler-IoU moderately improved the accuracy of ATB surface defect detection with negligible impact on real-time performance.
(4)
The proposed LFD-YOLO network shows improvements in mAP0.5, F-measure, and Ir by 6.5%, 6.4%, and 8.9%, respectively (92.8% → 96.2%, 90.3% → 96.7%, 87.9% → 98.8%), while the FPS decreases by 3.7 f·s−1 (29.1 f·s−1 → 25.4 f·s−1). This demonstrates that LFD-YOLO significantly enhanced the accuracy of ATB surface defect detection, but moderately impacted real-time performance.
In summary, the substitution of LDconv for CBS, the addition of the DAT deformable attention mechanism, and the optimization of the bounding box loss function with Focaler-IoU in the proposed LFD-YOLO network all effectively improved the accuracy of ATB surface defect detection, with only minor impacts on real-time performance.

4.3. Effectiveness in Practical Applications

Figure 10 presents the ATB surface defect detection results for each ablation experiment and network. Figure 10a corresponds to the original image, while Figure 10b–f represent the detection results of the YOLOv8, YOLOv8 + LDconv, YOLOv8 + DAT, YOLOv8 + Focaler-CIoU, and LFD-YOLO networks, respectively. Table 5 presents the defect statistics for the number of defects identified by each ablation experiment network. Manual inspection reveals 11 instances of dents, 2 instances of nicks, 1 crack, and 1 burn.
From Figure 10 and Table 5, the following conclusions are made:
(1)
The YOLOv8, YOLOv8 + LDconv, and YOLOv8 + Focaler-CIoU networks can only identify three types of defects: dents, nicks, and cracks, detecting 8, 11, and 12 instances, respectively. These networks exhibit missed detections, with burns not being detected. The cumulative accuracy rates for the four types of defects are 53.3%, 73.3%, and 80%, respectively.
(2)
The YOLOv8 + DAT network identifies all four types of defects (dents, nicks, cracks, and burns), detecting a total of 13 instances. Although it missed some defects, the cumulative accuracy rate for the four types of defects is 86.7%.
(3)
The LFD-YOLO network also identifies all four types of defects (dents, nicks, cracks, and burns), detecting a total of 15 instances with no missed detections. The results align with manual inspections, achieving an accuracy rate of 100%.
In summary, as stepwise improvements to the networks are made, the recognition accuracy for ATB surface defects continues to increase, and the number of detectable ATB surface defects also rises. The proposed LFD-YOLO network was capable of identifying all the defects, demonstrating optimal defect recognition and localization accuracy and satisfactory real-time performance, thereby meeting the on-site detection requirements.

4.4. Model Generalization Experiment

To validate the generalizability of the proposed LFD-YOLO network, supplementary experiments were conducted on the public dataset NEU-DET [50]. The NEU-DET dataset, widely used for surface defect detection, includes six common types of steel surface defects: crazing, inclusion, patches, pitted surface, rolled-in scale, and scratches. The dataset consists of 1800 grayscale images, with 300 samples for each defect type, and each image has a resolution of 200 × 200 pixels. Table 6 presents the comparison of generalization experiment results, and Figure 11 visualizes the generalization experiment results.
It can be observed that the proposed LFD-YOLO achieves mAP0.5, F-measure, and Ir metrics of 92.2%, 93.7%, and 91.9%, respectively, on the NEU-DET dataset. Although these metrics show a slight decrease compared to those on the ATB dataset, they still reach a high level. The results demonstrate the generalizability and robustness of LFD-YOLO in detecting different types of surface defects.

5. Conclusions and Future Work

(1)
This paper proposes an ATB surface defect detection method based on LFD-YOLO, which meets the requirements of ATB surface defect detection. The proposed model effectively addresses the challenges of low accuracy and inadequate efficiency in ATB surface defect detection. The method builds upon YOLOv8, which was modified by substituting LDconv for the CBS module, incorporates the DAT deformable attention mechanism, and optimizes the bounding box loss function with Focaler-IoU. In the LFD-YOLO network, the LDconv module replaces the original CBS module in the backbone section, and the DAT deformable attention mechanism is added to improve the detection capability of tiny defects. In the neck section, the PAN–FPN structure is added and combined with the LDconv + DAT deformable attention mechanism to fuse multiscale features and improve the detection of various types of ATB surface defects. The head section employs Focaler-CIoU which optimizes the bounding box loss function, thus also contributing to improved bounding box regression accuracy.
(2)
The ablation experimental data confirm that replacing CBS with LDconv improves the network’s mAP0.5, F-measure, and Ir by 3.1%, 2.9%, and 5.1%, respectively, compared to YOLOv8, while the FPS decreases by 1.9 f·s−1. Incorporating the DAT deformable attention mechanism improves the network’s mAP0.5, F-measure, and Ir by 4.8%, 4.9%, and 8.4%, respectively, compared to YOLOv8, with the FPS decreasing by 1.5 f·s−1. Optimizing the bounding box loss function with Focaler-IoU increases the network’s mAP0.5, F-measure, and Ir by 2.4%, 1.2%, and 3.8%, respectively, compared to YOLOv8, while the FPS decreases by 1.0 f·s−1. The proposed LFD-YOLO network improved mAP0.5, F-measure, and Ir by 6.5%, 6.4%, and 8.9%, respectively, compared to YOLOv8, with the FPS decreasing by 3.7 f·s−1. Overall, the results indicate that LFD-YOLO significantly increases the accuracy of ATB surface defect detection, although with a moderate impact on real-time performance.
(3)
Comparative experiments and practical application results demonstrate that the proposed LFD-YOLO network outperforms SSD, Faster R-CNN, Mask R-CNN, YOLOv3, YOLOv5, YOLOv7, YOLOv7-tiny, and YOLOv8 in terms of accuracy for ATB surface defect detection. YOLOv8, YOLOv8 + LDconv, and YOLOv8 + Focaler-CIoU can only identify three types of defects (dents, nicks, and cracks), with some missed detections and burns not being detected. YOLOv8 + DAT can identify all four types of defects (dents, nicks, cracks, and burns), but still fails to detect some defects. In contrast, LFD-YOLO can identify all four types of defects—dents, nicks, cracks, and burns—with no missed detections. As network improvements are added incrementally, the recognition accuracy for ATB surface defects continues to increase. The LFD-YOLO network achieves optimal defect recognition accuracy, demonstrates satisfactory real-time performance, and meets the requirements for on-site detection.
Although the proposed method achieves better detection accuracy than the YOLOv8 network, its detection speed decreased. Future work could focus on optimizing the network through lightweight processing. Additionally, experiments on ATB defect detection under complex backgrounds are needed, along with further expansion of the dataset, to facilitate earlier industrialization and the identification of more defect types.

Author Contributions

W.D. conceived the research, implemented the methodology, conducted the experiments, and wrote the original draft of the manuscript. G.L. supervised the research, reviewed the manuscript, and provided critical revisions to improve the technical and theoretical content. J.M. contributed to data analysis and validation and assisted in refining the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Special Project for Research and Development in Key Areas of Guangdong Province (grant No. 2019B010154003).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

Author Jun Meng was employed by the company Jeet Technology Co., Ltd., Shenzhen, China. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhao, B.; Xie, L.; Li, H.; Zhang, S.; Wang, B.; Li, C. Reliability Analysis of Aero-Engine Compressor Rotor System Considering Cruise Characteristics. IEEE Trans. Reliab. 2020, 69, 245–259. [Google Scholar] [CrossRef]
  2. Ma, P.; Xu, C.; Xiao, D. Robotic Ultrasonic Testing Technology for Aero-Engine Blades. Sensors 2023, 23, 3729. [Google Scholar] [CrossRef] [PubMed]
  3. Aust, J.; Pons, D. Taxonomy of Gas Turbine Blade Defects. Aerospace 2019, 6, 58. [Google Scholar] [CrossRef]
  4. Le, H.F.; Zhang, L.J.; Liu, Y.X. Surface Defect Detection of Industrial Parts Based on YOLOv5. IEEE Access 2022, 10, 130784–130794. [Google Scholar] [CrossRef]
  5. Ding, P.; Song, D.; Shen, J.; Zhao, X.; Jia, M. A novel graph structure data-driven crack damage identification for compressor blade based on vibro-acoustic signal. Struct. Health Monit.-Int. J. 2024, 23, 3046–3062. [Google Scholar] [CrossRef]
  6. Abdulrahman, Y.; Eltoum, M.A.M.; Ayyad, A.; Moyo, B.; Zweiri, Y. Aero-Engine Blade Defect Detection: A Systematic Review of Deep Learning Models. IEEE Access 2023, 11, 53048–53061. [Google Scholar] [CrossRef]
  7. Aust, J.; Shankland, S.; Pons, D.; Mukundan, R.; Mitrovic, A. Automated Defect Detection and Decision-Support in Gas Turbine Blade Inspection. Aerospace 2021, 8, 30. [Google Scholar] [CrossRef]
  8. Zhang, H.-B.; Zhang, C.-Y.; Cheng, D.-J.; Zhou, K.-L.; Sun, Z.-Y. Detection Transformer with Multi-Scale Fusion Attention Mechanism for Aero-Engine Turbine Blade Cast Defect Detection Considering Comprehensive Features. Sensors 2024, 24, 1663. [Google Scholar] [CrossRef]
  9. Song, M.; Zhang, Y. Aviation-engine blade surface anomaly detection based on the deformable neural network. Signal Image Video Process. 2025, 19, 87. [Google Scholar] [CrossRef]
  10. Huang, X.; Zhu, J.; Huo, Y. SSA-YOLO: An Improved YOLO for Hot-Rolled Strip Steel Surface Defect Detection. IEEE Trans. Instrum. Meas. 2024, 73, 1–17. [Google Scholar] [CrossRef]
  11. Cui, L.; Jiang, X.; Xu, M.; Li, W.; Lv, P.; Zhou, B. SDDNet: A Fast and Accurate Network for Surface Defect Detection. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar] [CrossRef]
  12. Usamentiaga, R.; Lema, D.G.; Pedrayes, O.D.; Garcia, D.F. Automated Surface Defect Detection in Metals: A Comparative Review of Object Detection and Semantic Segmentation Using Deep Learning. IEEE Trans. Ind. Appl. 2022, 58, 4203–4213. [Google Scholar] [CrossRef]
  13. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A.; IEEE. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  14. Liu, G.; Yan, Y.; Meng, J. Study on the detection technology for inner-wall outer surface defects of the automotive ABS brake master cylinder based on BM-YOLOv8. Meas. Sci. Technol. 2024, 35, 055109. [Google Scholar] [CrossRef]
  15. Zhang, G.; Liu, G.; Zhong, F. Research on UAV Autonomous Recognition and Approach Method for Linear Target Splicing Sleeves Based on Deep Learning and Active Stereo Vision. Electronics 2024, 13, 4872. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Guo, Z.; Wu, J.; Tian, Y.; Tang, H.; Guo, X. Real-Time Vehicle Detection Based on Improved YOLO v5. Sustainability 2022, 14, 12274. [Google Scholar] [CrossRef]
  17. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  18. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M.; IEEE. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  19. Terven, J.; Córdova-Esparza, D.-M.; Romero-González, J.-A.J.M.L.; Extraction, K. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  20. Wang, C.-Y.; Yeh, J.; Liao, H.-Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. In Proceedings of the 18th European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–21. [Google Scholar]
  21. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  22. Chen, Z.-H.; Juang, J.-C. YOLOv4 Object Detection Model for Nondestructive Radiographic Testing in Aviation Maintenance Tasks. Aiaa J. 2022, 60, 526–531. [Google Scholar] [CrossRef]
  23. Kannadaguli, P. YOLO v4 Based Human Detection System Using Aerial Thermal Imaging for UAV Based Surveillance Applications. In Proceedings of the 2020 International Conference on Decision Aid Sciences and Application (DASA), Sakheer, Bahrain, 8–9 November 2020; pp. 1213–1219. [Google Scholar]
  24. Li, X.; Wang, W.; Sun, L.; Hu, B.; Zhu, L.; Zhang, J. Deep learning-based defects detection of certain aero-engine blades and vanes with DDSC-YOLOv5s. Sci. Rep. 2022, 12, 13067. [Google Scholar] [CrossRef]
  25. Liao, D.; Cui, Z.; Zhang, X.; Li, J.; Li, W.; Zhu, Z.; Wu, N. Surface defect detection and classification of Si3N4 turbine blades based on convolutional neural network and YOLOv5. Adv. Mech. Eng. 2022, 14, 1–13. [Google Scholar] [CrossRef]
  26. Wang, D.; Xiao, H.; Huang, S. Automatic Defect Recognition and Localization for Aeroengine Turbine Blades Based on Deep Learning. Aerospace 2023, 10, 178. [Google Scholar] [CrossRef]
  27. Li, S.; Yu, J.; Wang, H. Damages Detection of Aeroengine Blades via Deep Learning Algorithms. IEEE Trans. Instrum. Meas. 2023, 72, 1–11. [Google Scholar] [CrossRef]
  28. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  29. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J.; IEEE. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 27th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  30. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef]
  31. Ahmed, A.; Tangri, P.; Panda, A.; Ramani, D.; Nevronas, S.K.; IEEE. VFNet: A Convolutional Architecture for Accent Classification. In Proceedings of the 16th IEEE-India-Council International Conference (INDICON), Rajkot, India, 13–15 December 2019. [Google Scholar]
  32. Liu, Y.; Wu, D.; Liang, J.; Wang, H. Aeroengine Blade Surface Defect Detection System Based on Improved Faster RCNN. Int. J. Intell. Syst. 2023, 2023, 1992415. [Google Scholar] [CrossRef]
  33. Shang, H.; Sun, C.; Liu, J.; Chen, X.; Yan, R. Deep learning-based borescope image processing for aero-engine blade in-situ damage detection. Aerosp. Sci. Technol. 2022, 123, 107473. [Google Scholar] [CrossRef]
  34. Zhang, H.; Zu, K.; Lu, J.; Zou, Y.; Meng, D. EPSANet: An Efficient Pyramid Squeeze Attention Block on Convolutional Neural Network. In Proceedings of the 16th Asian Conference on Computer Vision (ACCV), Macao, China, 4–8 December 2022; pp. 541–557. [Google Scholar]
  35. Liu, P.; Yuan, X.; Han, Q.; Xing, B.; Hu, X.; Zhang, J. Micro-defect Varifocal Network: Channel attention and spatial feature fusion for turbine blade surface micro-defect detection. Eng. Appl. Artif. Intell. 2024, 133, 108075. [Google Scholar] [CrossRef]
  36. Aust, J.; Pons, D. Methodology for Evaluating Risk of Visual Inspection Tasks of Aircraft Engine Blades. Aerospace 2021, 8, 117. [Google Scholar] [CrossRef]
  37. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  38. Zhang, X.; Song, Y.; Song, T.; Yang, D.; Ye, Y.; Zhou, J.; Zhang, L. LDConv: Linear deformable convolution for improving convolutional neural networks. Image Vis. Comput. 2024, 149, 105190. [Google Scholar] [CrossRef]
  39. Xia, Z.; Pan, X.; Song, S.; Li, L.E.; Huang, G. Vision Transformer with Deformable Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4784–4793. [Google Scholar]
  40. Zhang, H.; Zhang, S. Focaler-IoU: More Focused Intersection over Union Loss. arXiv 2024, arXiv:2401.10525. [Google Scholar]
  41. Li, K.; Huang, Z.; Cheng, Y.-C.; Lee, C.-H.; IEEE. A maximal figure-of-merit learning approach to maximizing mean average precision with deep neural network based classifiers. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014. [Google Scholar]
  42. Nan, Y.; Chai, K.M.; Lee, W.S.; Chieu, H.L. Optimizing F-measure: A tale of two approaches. arXiv 2012, arXiv:1206.4625. [Google Scholar]
  43. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  44. Farhadi, A.; Redmon, J. Yolov3: An incremental improvement. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake, UT, USA, 18–22 June 2018; pp. 1–6. [Google Scholar]
  45. Ma, L.; Zhao, L.; Wang, Z.; Zhang, J.; Chen, G. Detection and Counting of Small Target Apples under Complicated Environments by Using Improved YOLOv7-tiny. Agronomy 2023, 13, 1419. [Google Scholar] [CrossRef]
  46. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  47. Zhang, Y.-F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  48. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  49. Siliang, M.; Yong, X. MPDIoU: A Loss for Efficient and Accurate Bounding Box Regression. arXiv 2023, arXiv:2307.07662. [Google Scholar]
  50. Song, K.; Yan, Y. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
Figure 1. Overall framework diagram of the ATB surface defect detection algorithm based on LFD-YOLO.
Figure 1. Overall framework diagram of the ATB surface defect detection algorithm based on LFD-YOLO.
Sensors 25 02219 g001
Figure 2. Images of typical surface defects in ATBs with high detection demands: (a) a crack, (b) nicks, (c) a dent, and (d) burns.
Figure 2. Images of typical surface defects in ATBs with high detection demands: (a) a crack, (b) nicks, (c) a dent, and (d) burns.
Sensors 25 02219 g002
Figure 3. Network architecture of YOLOv8 [19].
Figure 3. Network architecture of YOLOv8 [19].
Sensors 25 02219 g003
Figure 4. Network architecture of LFD-YOLO.
Figure 4. Network architecture of LFD-YOLO.
Sensors 25 02219 g004
Figure 5. Diagram of the LDconv structure [38].
Figure 5. Diagram of the LDconv structure [38].
Sensors 25 02219 g005
Figure 6. Structure of the DAT [39].
Figure 6. Structure of the DAT [39].
Sensors 25 02219 g006
Figure 7. Flowchart of the DA module [39].
Figure 7. Flowchart of the DA module [39].
Sensors 25 02219 g007
Figure 8. Diagram of the ATB surface defect detection system: (a) imaging process, (b) system interface.
Figure 8. Diagram of the ATB surface defect detection system: (a) imaging process, (b) system interface.
Sensors 25 02219 g008
Figure 9. Comparison of prediction results across different networks.
Figure 9. Comparison of prediction results across different networks.
Sensors 25 02219 g009
Figure 10. Images of the ATB surface defect detection using ablation experimental networks: (a) original image, (b) YOLOv8, (c) YOLOv8 + LDconv, (d) YOLOv8 + DAT, (e) YOLOv8 + Focaler-CIoU, and (f) LFD-YOLO.
Figure 10. Images of the ATB surface defect detection using ablation experimental networks: (a) original image, (b) YOLOv8, (c) YOLOv8 + LDconv, (d) YOLOv8 + DAT, (e) YOLOv8 + Focaler-CIoU, and (f) LFD-YOLO.
Sensors 25 02219 g010
Figure 11. Visualization of the generalization experiment results.
Figure 11. Visualization of the generalization experiment results.
Sensors 25 02219 g011
Table 1. Quantitative analysis of the different networks.
Table 1. Quantitative analysis of the different networks.
ModelBackbonemAP0.5, %F-Measure, %FPS, (f·s−1)
SSD [43]VGG-1661.260.935.6
Faster R-CNN [28]ResNet-5073.774.116.2
Mask R-CNN [30]ResNet-5076.378.219.3
YOLOv3 [44]DarkNet-5377.576.727.2
YOLOv5 [16]CSP-Darknet5386.886.927.7
YOLOv7 [18]DarkNet-53 + E-ELAN89.188.731.1
YOLOv7-tiny [45]DarkNet-53 + E-ELAN77.176.847.8
YOLOv8 [19]DarkNet53 + C2f89.790.329.1
YOLOv11 [46]DarkNet53 + C3k291.491.930.7
LFD-YOLODarkNet53 + C2f +DAT96.296.725.4
Table 2. Results of ablation study I.
Table 2. Results of ablation study I.
ModelmAP0.5, %F-Measure, %Ir, %FPS, (f·s−1)
YOLOv8 + LDconv + DAT93.092.492.327.0
YOLOv8 + LDconv + DAT + EIoU93.191.893.226.6
YOLOv8 + LDconv + DAT + WIoU93.592.492.726.3
YOLOv8 + LDconv + DAT + MPDIoU94.994.695.124.7
LFD-YOLO96.296.798.825.4
Table 3. Results of ablation study II.
Table 3. Results of ablation study II.
ModelmAP0.5, %F-Measure, %Ir, %FPS, (f·s−1)
YOLOv8 + LDconv + Focaler-CIoU93.293.794.027.1
YOLOv8 + LDconv + Focaler-CIoU+ DAT (backbone)94.794.995.226.9
YOLOv8 + LDconv + Focaler-CIoU+ DAT (neck)95.194.596.026.4
LFD-YOLO96.296.798.825.4
Table 4. Results of ablation study III.
Table 4. Results of ablation study III.
ModelmAP0.5, %F-Measure, %Ir, %FPS, (f·s−1)
YOLOv889.790.387.929.1
YOLOv8 + LDconv92.893.293.027.2
YOLOv8 + DAT94.595.296.327.6
YOLOv8 + Focaler-CIoU92.191.591.728.1
LFD-YOLO96.296.798.825.4
Table 5. Statistics on the identified surface defects using ablation experimental networks: YOLOv8, YOLOv8 + LDconv, and YOLOv8 + Focaler-CIoU.
Table 5. Statistics on the identified surface defects using ablation experimental networks: YOLOv8, YOLOv8 + LDconv, and YOLOv8 + Focaler-CIoU.
ModelDentsNicksCracksBurns
YOLOv86110
YOLOv8 + LDconv8210
YOLOv8 + Focaler-CIoU9210
YOLOv8 + DAT9211
LFD-YOLO11211
Table 6. Comparison of the generalization experiment results.
Table 6. Comparison of the generalization experiment results.
DatasetmAP0.5, %F-Measure, %Ir, %
NEU-DET [50]92.293.791.9
ATB96.296.795.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, W.; Liu, G.; Meng, J. Study on Novel Surface Defect Detection Methods for Aeroengine Turbine Blades Based on the LFD-YOLO Framework. Sensors 2025, 25, 2219. https://doi.org/10.3390/s25072219

AMA Style

Deng W, Liu G, Meng J. Study on Novel Surface Defect Detection Methods for Aeroengine Turbine Blades Based on the LFD-YOLO Framework. Sensors. 2025; 25(7):2219. https://doi.org/10.3390/s25072219

Chicago/Turabian Style

Deng, Wei, Guixiong Liu, and Jun Meng. 2025. "Study on Novel Surface Defect Detection Methods for Aeroengine Turbine Blades Based on the LFD-YOLO Framework" Sensors 25, no. 7: 2219. https://doi.org/10.3390/s25072219

APA Style

Deng, W., Liu, G., & Meng, J. (2025). Study on Novel Surface Defect Detection Methods for Aeroengine Turbine Blades Based on the LFD-YOLO Framework. Sensors, 25(7), 2219. https://doi.org/10.3390/s25072219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop