Next Article in Journal
Research on Displacement Tracking Device Inside Hybrid Materials Based on Electromagnetic Induction Principle
Previous Article in Journal
Small UAV Target Detection Algorithm Using the YOLOv8n-RFL Based on Radar Detection Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Deep Learning-Based Multi-Level Cross-Domain Foreign Object Detection in Power Transmission Lines

1
School of Mechanical and Electrical Engineering, Kunming University, Kunming 650214, China
2
Yunnan Key Laboratory of Intelligent Logistics Equipment and Systems, Kunming 650214, China
3
School of Fine Art and Design, Kunming University, Kunming 650214, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5141; https://doi.org/10.3390/s25165141
Submission received: 21 May 2025 / Revised: 7 August 2025 / Accepted: 8 August 2025 / Published: 19 August 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

With the rapid advancement of deep learning technology, deep learning-based methods have become the mainstream approach for detecting potential safety hazards in transmission lines, playing a crucial role in power grid safety monitoring. However, existing models are often overly complex and struggle with detecting small or occluded targets, limiting their effectiveness in edge-device deployment and real-time detection scenarios enhanced the YOLOv11 model by integrating it with the ConvNeXt network, a multi-level cross-domain analysis detection model (ConvNeXt-You Only Look Once) is proposed. Additionally, Bayesian optimization was employed to fine-tune the model’s hyperparameters and accelerate convergence. Experimental results demonstrate that CO-YOLO mAP@0.5 reached 98.4%, mAP@0.5:0.95 reached 66.1%, and FPS was 303, outperforming YOLOv11 and ETLSH-YOLO, in both accuracy and efficiency. Compared with the original model, CO-YOLO model improved by 1.9% in mAP@0.5 and 2.2% in mAP@0.5:0.95.

1. Introduction

As the core component of modern energy transmission, the stability and safety of power systems play a crucial role in ensuring the smooth operation of society [1,2,3] reducing—or even eliminating—grid faults has become an unwavering goal pursued by smart grid technology research teams worldwide. Among these challenges, foreign object intrusion is particularly noteworthy [4,5,6]. Foreign objects on transmission lines—such as bird nests, kites, balloons, plastic waste, and branches—often adhere to the lines due to aerodynamic forces [7,8,9]. These intrusions not only compromise the insulating properties of the line surfaces but also, through friction and collision with line components, can readily trigger short circuits, line breaks, electrical fires, or even large-scale power outages, thereby posing significant safety hazards to the reliable operation of power grids [10,11]. Consequently, the timely detection and elimination of potential safety risks on transmission lines is of paramount importance.
Traditional methods for detecting foreign objects on transmission lines primarily rely on periodic manual inspections and rudimentary image processing techniques to analyze transmission line images. These approaches are not only inefficient but also fail to achieve comprehensive coverage and real-time monitoring [12]. In recent years, the rapid development of deep learning technology has led to significant advances in image recognition and object detection. Research has predominantly focused on using convolutional neural networks (CNNs) to extract features from transmission line images and identify foreign objects [13,14]. Deep learning-based object detection methods can automatically learn image features, thereby markedly improving detection accuracy. These methods generally fall into two main types: two-stage networks that utilize candidate frames [15,16,17] and one-stage networks that rely on regression [18,19,20]. Notably, the YOLO model has become a research hotspot due to its fast detection speed and its ability to generate detection results for all objects with a single forward pass, thus enabling real-time detection. However, existing YOLO models rely on multi-scale feature maps to detect objects of various sizes and tend to emphasize overall object features. As a result, the relatively large scale of the detection layers makes it difficult to capture fine details of small objects, leading to lower recognition rates for occluded or variably sized images. Moreover, their high computational complexity when processing high-resolution images hampers real-time performance on detection devices [21,22]. To address these challenges, several approaches have been proposed. For example, Tu et al. [23] introduced a dual-branch downsampling module (DBD) into the neck network of YOLOv8 and integrated a mixed enhanced attention module (MIX) within the backbone to address issues arising from variable object scales and indistinct features; however, the large number of computational parameters restricts its deployment on smaller devices. Similarly, WANG et al. [24] incorporated deformable convolution (DCN) and the parameter-free SimAM attention mechanism into the ELAN-S module of YOLOv7-Tiny to enhance the model’s capability to extract features from irregularly shaped and intra-class diverse foreign objects (e.g., garbage bags and branches). They further optimized hyperparameters using genetic algorithms (GA) and employed space-to-depth (SPD) convolutions to improve the recognition of low-resolution and small objects while accelerating model convergence. Nonetheless, the extra parameters introduced by the DCN module and the insufficient extraction of critical features by SimAM—especially when processing complex textures and multi-scale features—have limited overall performance. Additionally, Li et al. [25] augmented the YOLOv7 model by incorporating a Large Selective Kernel Network (LSKNet) structure and introducing weighted spatial attention (WSA) after the final C2f module to enhance the detection of occluded objects; yet, WSA diminished the recognition rate for irregular targets (such as garbage bags and bird nests) by overemphasizing target regions. Moreover, Yu et al. [26] introduced a channel–spatial decoupling downsampling module (CSDovn) alongside a coordinate attention (CA) mechanism and the Mish activation function to significantly reduce the parameter count and computational complexity of the YOLOv7 model, although its performance in small object detection remains unsatisfactory. Despite these efforts to integrate attention mechanisms and feature fusion modules to bolster the extraction of multi-scale object features, current methods still face limitations by not fully leveraging multi-dimensional information—such as the appearance characteristics of objects and the spatial relationships among them.
Given this, we present a novel cross-domain detection model that integrates transmission line visual features—extracted via an improved YOLOv11—with spatial appearance cues derived from ConvNeXt-B. This design enables multi-level semantic fusion, effectively enhancing the recognition of small and occluded objects in complex scenes while maintaining high real-time performance. Additionally, we incorporate Bayesian optimization to fine-tune hyperparameters automatically, which significantly accelerates model convergence and improves generalization capability. Extensive experimental validation demonstrates the effectiveness of our method in practical transmission line inspection tasks.
The key contributions of this study are summarized as follows:
  • A structural enhancement of YOLOv11, used C3k2-Dual module, NonLocalBlockND module, C2PSA-DHSA module, DySample module, PIoU and combined ConvNeXT, enabling improved detection of small, dense, and occluded targets with faster inference, tailored for real-time transmission line scenarios.
  • A novel multi-level cross-domain fusion framework is proposed that combines object-level features and spatial semantic cues using ConvNeXt-B, thereby strengthening feature representation and detection robustness.
  • Integration of Bayesian optimization for automated hyperparameter tuning, leading to improved convergence speed and overall detection performance.
Section 2 introduces the original network model along with the proposed methods and their implementation. Section 3 presents a comparative analysis between the proposed approach and several existing methods. Section 4 concludes the paper with pertinent observations and future work direction.

2. Models and Methods

2.1. Object Detection Part

Object Detection Part consists of YOLOv11, which is mainly composed of three main parts: backbone, neck, and head, as shown in Figure 1.
The backbone is mainly used for data enhancement. The backbone is composed of multiple Conv and C3k2 modules and an SPPF module and a C2PSA module, which are mainly used to extract features from images. The neck network adopts the PAN structure and mainly consists of Upsample modules, C3k2 modules as well as Conv to achieve a multi-scale object detection effect. The head is mainly composed of multiple detection heads, which are responsible for predicting the object position and category based on the feature information refined by the neck. In this paper, C3k2 modules, C2PSA modules, Upsample modules, NonLocalBlockND modules, Loss functions are improved in five aspects concerning the characteristics of the transmission line foreign object intrusion dataset.

2.1.1. C3k2-Dual Module

The C3k2 module extracts feature information by calculating the interactions between all positions in the feature map, resulting in a quadratic increase in computational complexity. The structure is shown in Figure 2. To address this challenge, the C3k2-Dual module replaces traditional convolution (Conv) with Dual Conv, substantially enhancing computational efficiency. By incorporating grouped convolution, this design optimally arranges convolutional filters, reducing both computational cost and parameter count while improving model accuracy.
In detail, DualConv partitions N convolutional kernels into G groups (G is adjustable to balance the proportion of grouped convolutions, Process M parallel image channels simultaneously, thereby optimizing floating-point operations (FLOPs)), as shown in Figure 3. Each group processes the entire input feature map as follows: M/G input channels undergo parallel 3 × 3 and 1 × 1 convolutions to retain complete feature information, ensuring effective deep feature extraction, while the remaining (M-M/G) channels are processed exclusively by 1 × 1 convolutions to minimize the parameter burden. The resulting outputs from the 3 × 3 and 1 × 1 convolutions are summed to obtain the final feature map.

2.1.2. NonLocalBlockND Module

The NonLocalBlockND module improves model performance by capturing long-range dependencies across the feature map, as shown in Figure 4. Its key mechanism leverages three branches (θ, ϕ, and g) to derive downsampled feature representations, enabling similarity computation and weighted aggregation. Specifically, the θ branch generates query vectors, the ϕ branch produces key vectors for similarity estimation, and the g branch generates value vectors for weighted aggregation. The similarity between θ and ϕ produces an attention weight matrix, which, when multiplied by g, yields globally weighted feature representations. Finally, a 1 × 1 convolution restores the original channel dimensions, while a residual connection fuses the weighted and original features, effectively integrating local and global context information and significantly enhancing the model’s perceptual capabilities.

2.1.3. C2PSA-DHSA Module

The conventional C2PSA module enhances the model’s attention to critical image regions by stacking multiple PSA modules. However, this design overemphasizes spatial correlations within feature maps, leading to suboptimal generalization. To address this issue, this study replaces the attention mechanism in the C2PSA module with the DHSA module, significantly improving image restoration performance, as shown in Figure 5a. The structures of the C2PSA-DHSA and DHSA modules are illustrated in Figure 5b and Figure 6.
In the proposed C2PSA-DHSA module, we aim to overcome the limitations of conventional spatial attention mechanisms that often focus too narrowly on specific regions or suffer from overfitting due to excessive parameterization.
The module begins by splitting the input feature map F into two branches (F1 and F2). The F1 branch undergoes dynamic-range sorting along both horizontal and vertical axes, generating an ordered representation that preserves spatial distribution characteristics. The sorted feature map is then concatenated with F2, which retains the original unaltered spatial information, forming a new intermediate feature map F’. This combined representation is processed through depthwise separable convolution, enhancing the module’s ability to capture long-range dependencies while reducing computational overhead. Following this, histogram-based reshaping operations—Bilateral Histogram Recalibration (BHR) and Frequency Histogram Recalibration (FHR)—are applied to dynamically balance global structures and local details, adapting to the complexity of transmission line imagery.
Finally, a self-attention mechanism fuses these enriched features, enabling the network to better model semantic relevance and spatial dependencies. This refined attention fusion significantly improves the quality of feature extraction, particularly for irregular and occluded foreign objects.

2.1.4. DySample Module

In the YOLOv11 framework, the feature upsampler applies a fixed interpolation rule to convert low-resolution feature maps into high-resolution counterparts, as delineated in Equation (1). However, this method, which depends on adjacent pixel spacing, achieves mere spatial enlargement of the feature map without adequately preserving its detailed features or semantic content. Consequently, this limitation hampers the model’s performance in dense prediction tasks, such as foreign object detection in transmission lines. To overcome this shortfall, we introduce an advanced upsampling module, termed DySample. Departing from conventional fixed-rule interpolation, DySample employs a dynamic point-sampling strategy. Its operational mechanism is elucidated in Equation (2), with its architecture depicted in Figure 7. Within DySample, the F1 branch dynamically learns the coordinates of sampling points across the feature map. To regulate these points, a range factor—illustrated in Figure 8—is incorporated to limit the offset range of the sampling positions, yielding a refined sampling set, S. Subsequently, the feature map X is resampled using the grid_sample function based on the coordinates defined in S.
X high = Interp ( X low ) ,
X high = GridSample ( X low , S ) ,

2.1.5. Loss Function

Although YOLOv11 deviates from traditional approaches by employing the SIoU (Scale-Aware Intersection over Union) loss function instead of CIoU to improve the detection of small objects, this shift inadvertently introduces errors in computing overlapping regions between objects. As a result, the model’s localization precision and overall detection performance are markedly reduced. To mitigate this limitation, we propose replacing SIoU with the PIoU (Powerful Intersection over Union) loss function, the formulation of which is presented in Equation (3), B p represents the Predicted Box, and B g represents the Ground Truth Box. | B p B g | is the area of the intersection between the predicted box and the ground truth box, while | B p B g | is the area of their union. The PIoU loss function quantifies the overlap between predicted and ground-truth bounding boxes by calculating their IoU value. while additionally accounting for angular discrepancies between them. This enhancement refines the bounding box regression process, enabling superior adaptation to objects with high aspect ratios or those embedded in complex backgrounds.
L P I o U = 1 | B p B g | | B p B g | ,

2.1.6. Improved Network Model

In this paper, we mainly enhance the CO-YOLO backbone network by integrating three key modifications: the C3k2-Dual Module, the C2PSA-DHSA Module, and the NonLocalBlockND Module (added before the SPPF module) to refine feature extraction. Additionally, the DySample module is employed to enhance the upsampling capability of the neck, while Powerful-IoU Loss is incorporated into the detection head to improve detection performance. These improvements significantly enhance CO-YOLO’s ability to detect foreign objects in power transmission lines. The overall improved network structure is illustrated in Figure 9.

2.2. Relationship Detection

ConvNeXT-B

As shown in Figure 10, ConvNeXt extracts features by stacking multiple ConvNeXt blocks. In this study, we employ ConvNeXt to capture the feature relationships of foreign objects, enabling the recognition model to acquire richer appearance features and their spatial dependencies. This approach significantly enhances the network’s feature extraction capability, allowing it to better understand both the visual characteristics and spatial relationships of the detected objects.

2.3. CO-YOLO

As shown in Figure 11, ConvNeXt extracts features by stacking multiple ConvNeXt blocks. In this study, we utilize ConvNeXt to capture the feature relationships of foreign objects, enabling the recognition model to acquire richer appearance features and their spatial relationships, thereby enhancing the network’s feature extraction capability.

3. Experiments and Analyses

3.1. Experimental Settings

Table 1 summarizes the primary environment configurations used in our experiment, including the operating system, CPU, GPU, and the versions of Python, CUDA, and PyTorch. In order to optimize the hyperparameters of the model, we adopted Bayesian optimization and constructed the posterior distribution of the objective function by using Bayes theorem. By collecting functions (such as expected improvement EI, probability improvement PI, and upper confidence bound UCB), we selected the next combination of hyperparameters for evaluation based on the posterior distribution, so as to approximate the optimal value with fewer evaluations.
The final optimized settings were: optimizer = SGD, batch size = 16, initial learning rate (lr0) = 0.2, final learning rate (lrf) = 0.01, momentum = 0.937, weight decay = 0.0005, warmup epochs = 20, warmup momentum = 0.8, and total epochs = 350. All remaining hyperparameters were set to the default values provided by the official YOLOv11 implementation, which served as the baseline model.

3.2. DataSets

Due to the lack of publicly available datasets on foreign object intrusions in cross-domain transmission lines, we compiled a custom dataset of 979 images by integrating data from existing public datasets, media resources, and manually collected samples. This dataset includes four categories of foreign objects: bird nests, balloons, branches, and plastic debris. Given the diverse and complex nature of these intrusions, a larger volume of image data is essential for effective feature learning by the model. To address this, we applied a range of data augmentation techniques, such as random cropping, flipping, and rotation, to expand the dataset. Furthermore, we simulated images captured under adverse conditions by applying random grayscale transformations and adjusting hue, saturation, exposure, and brightness. As a result of this augmentation process, the dataset size increased to 2739 images. A detailed overview of the dataset is provided in Table 2 and Figure 12.
The dataset covers a variety of typical transmission line scenarios and fault types. During the data collection process, strict adherence to industry standards and practical application requirements is maintained to ensure the diversity and representativeness of the samples.
Data preprocessing includes image denoising, format unification, and removal of abnormal samples. The labeling process is conducted by a team of engineers with professional backgrounds, and a multi-round review mechanism is employed to ensure the accuracy of the labeling. The labeling content mainly includes the position boxes and category labels of the target objects.
For data quantification, we divide the dataset according to the following criteria:
-
Training set: 70% of the total samples, used for model parameter training;
-
Validation set: 15%, used for hyperparameter tuning and model selection;
-
Test set: 15%, used for final performance evaluation.

3.3. Evaluation Metrics

In our experiments, we used several metrics to evaluate the model’s performance, including accuracy, precision, recall, average precision (AP), mean average precision (mAP), confusion matrix, parameter count, GFLOPs, and frames per second (FPS). Among these, mAP serves as a key performance indicator, especially for object detection tasks.
The mean Average Precision (mAP) is often divided into two metrics: mAP@0.5 and mAP@0.5:0.95. Here, mAP@0.5 indicates the mean Average Precision (AP) across all categories when the Intersection over Union (IoU) threshold is fixed at 0.5. In contrast, mAP@0.5:0.95 represents the average AP calculated over a series of IoU thresholds ranging from 0.5 to 0.95, with an incremental step of 0.05. Given K categories where K > 1, the mAP can be formulated as Equation (4):
m A P = 1 K k = 1 K A P k ,
where A P k is the Average Precision for category K . The AP itself is calculated by integrating the area under the Precision-Recall curve, which reflects the balance between precision and recall across different decision thresholds.
In addition to mAP, the Confusion Matrix provides a detailed breakdown of the model’s predictions, helping to visualize true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). From this, important evaluation metrics like accuracy, precision, and recall are derived:
  • Accuracy measures the proportion of total predictions that are correct, defined as Equation (5) as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N ,
  • Precision reflects the proportion of predicted positive samples that are indeed positive shown as Equation (6) as follows:
P r e c i s i o n ( P ) = T P T P + F P ,
  • Recall (also called Sensitivity) captures the proportion of actual positive samples correctly identified, shown as Equation (7) as follows:
R e c a l l ( R ) = T P T P + F N
Furthermore, Parameter Count (parameters) and GFLOPs (giga floating point operations per second) evaluate the computational complexity of the model. A lower parameter count and fewer FLOPs generally indicate a more lightweight and efficient model. FPS (frames per second) measures the real-time inference speed of the model, an essential metric for deployment in time-sensitive applications.

3.4. Benchmark Comparison Experiment

Table 3 and Figure 13 present a comparison of key detection models, highlighting the superior parameter efficiency and FPS of single-stage detection models over their two-stage counterparts. The proposed CO-YOLO model demonstrates significant advantages over other YOLO series models (YOLOv5, YOLOv8, YOLOv10, and YOLOv11). Compared to YOLOv5, our model improves mAP@0.5 by 1.0% and mAP@0.5:0.95 by 2.8%, with a 12.7% increase in parameter count and a 1.4% rise in GFLOPs, while FPS only decreases by 6.0%. Similarly, compared to YOLOv11, CO-YOLO achieves a 1.9% boost in mAP@0.5 and a 2.2% increase in mAP@0.5:0.95, with a 9.2% rise in parameters, a 14.3% increase in GFLOPs, and a 12.1% drop in FPS. These results demonstrate that CO-YOLO maintains high real-time performance while significantly enhancing detection accuracy and robustness, particularly for challenging objects like bird nests and branches. This makes it a more efficient and reliable solution for foreign object intrusion detection in power transmission lines.
As shown in Table 4, we compared the performance of four different attention mechanisms to evaluate their impact on the model’s effectiveness. Notably, the NonLocalBlockND attention mechanism achieved the best results in both mAP@0.5:0.95 and FPS. Compared to the most parameter-efficient attention mechanism, NonLocalBlockND increased the parameter count by only 3.45%, while improving mAP@0.5:0.95 by 1.72%, mAP@0.5 by 0.61%, and FPS by 2.8%. These results demonstrate that NonLocalBlockND effectively integrates local and global contextual information, significantly enhancing the model’s perception capabilities. Considering both recall and precision, it is evident that the proposed algorithm outperforms other types of methods, thereby validating its effectiveness.
To comprehensively evaluate the impact of various improvements on model performance, we compared the effects of the C2PSA module, C3K2 module, and different loss functions, as shown in Table 5, Table 6 and Table 7.
By comparing different C2PSA variants, we found that the C2PSA-DHSA module outperforms the C2PSA-DAT module by enhancing target extraction capabilities while reducing parameter count. Specifically, the C2PSA-DHSA module decreases parameter count by 4.29%, while improving mAP@0.5 by 0.31% and mAP@0.5:0.95 by 0.63%. This demonstrates that the C2PSA-DHSA module significantly improves feature extraction efficiency while reducing computational cost. Further analysis of Table 6 shows that the C3K2-Dual module achieves the highest mAP@0.5 and mAP@0.5:0.95 among all variants, with improvements of 1.04% and 1.74%, respectively. This highlights the C3K2-Dual module’s outstanding performance in object detection, effectively enhancing model accuracy and robustness. Moreover, as shown in Table 7, the PIoU loss function achieves the best results in parameter efficiency, recall, and FPS. Notably, compared to the EIoU loss, PIoU reduces the parameter count by 5.28% while maintaining the same recall rate. This result indicates that the PIoU loss function reduces model complexity without compromising sensitivity, improving both efficiency and accuracy—aligning well with the principles of efficient module selection.

3.5. Subsection

As shown in Table 8 and Figure 14, the proposed CO-YOLO model was compared against several state-of-the-art algorithms for foreign object intrusion detection in power transmission lines. The experimental results demonstrate that CO-YOLO outperforms all other models in the key metrics of mAP@0.5 and mAP@0.5:0.95. Specifically, compared to ETLSH-YOLO, CO-YOLO improves mAP@0.5 by 3.7% and mAP@0.5:0.95 by 8.6%. Against DF-YOLO, the improvements are 2.3% and 12.2%, respectively. Compared to GEB-YOLO, CO-YOLO achieves a 2.0% gain in mAP@0.5 and a 12.2% increase in mAP@0.5:0.95. The model also surpasses TL-YOLO by 6.7% in mAP@0.5 and 2.2% in mAP@0.5:0.95. While CO-YOLO achieves an 8.7% higher mAP@0.5 compared to GCP-YOLO, its mAP@0.5:0.95 is 4.7% lower. Finally, compared to TFD-YOLOv8, CO-YOLO shows remarkable gains of 33.0% in mAP@0.5 and 80.4% in mAP@0.5:0.95. These results clearly highlight the superior detection accuracy and robustness of CO-YOLO, making it more effective at identifying foreign objects in power transmission lines and providing a more reliable safeguard for the safe operation of transmission infrastructure.

3.6. Ablation Study

To accurately evaluate the impact of different modules on model performance, validate the effectiveness of the design, and explore the relationships between the modules, we conducted a series of ablation experiments based on the YOLOv11 model. To ensure the scientific rigor and reliability of the results, all experimental settings were kept consistent, with the results summarized in Table 9.
Compared to the baseline YOLOv11 model, the integration of the C2PSA-DHSA module increased mAP@0.5 by 0.6% and decreased mAP@0.5:0.95 by 0.3%, while increasing the parameter count by 2.09% (from 2,582,932 to 2,637,080) and GFLOPs by 1.59% (from 6.3 to 6.4). The addition of the DySample module further boosted mAP@0.5 to 0.974 and mAP@0.5:0.95 to 0.638, while reducing the parameter count by 1.58% and lowering GFLOPs by 1.56%. When used independently, the C3K2-Dual module achieved the best mAP@0.5 (0.979) and mAP@0.5:0.95 (0.642), improving performance by 1.4% and 0.5%, respectively, compared to models without this module. The combination of the NonLocalBlockND, C2PSA-DHSA, and C3K2-Dual modules pushed mAP@0.5 to 0.982 and mAP@0.5:0.95 to 0.654, representing gains of 0.3% and 1.2%, respectively, over models without NonLocalBlockND. Finally, the introduction of a specific loss function resulted in the highest performance, with mAP@0.5 reaching 0.984 and mAP@0.5:0.95 reaching 0.641—improvements of 0.2% and 0.3% compared to models without this loss function. These results clearly demonstrate that the combination of these modules and the optimized loss function significantly enhances model performance while maintaining relatively low computational complexity, providing an efficient and reliable solution for foreign object intrusion detection in power transmission lines.
Through a comprehensive series of ablation experiments, we found that the combination of these modules and the optimized loss function significantly improves the performance of the YOLOv11 model while maintaining low computational complexity. Ultimately, the proposed CO-YOLO model achieved the best performance across all experiments, offering an efficient and accurate solution for foreign object detection in power transmission lines.

3.7. Attention Visualization Comparison Experiment

In the field of deep learning, attention mechanisms have been widely applied to various tasks to enhance a model’s ability to focus on key information, thereby significantly improving performance. In this study, we conducted a systematic visual comparison of different attention mechanisms by analyzing their focus regions and weight distributions, highlighting the strengths and weaknesses of each approach. As shown in Figure 15, the NonLocalBlockND attention mechanism evenly distributes attention across the target object, allowing the model to capture both local and global contextual information effectively. In contrast, the SCSA attention mechanism suffers from over-concentration, focusing excessively on a small portion of the target object, which limits the model’s ability to capture comprehensive features. This weakness is particularly evident when detecting objects like balloons, where both local and global information are crucial for accurate recognition. Similarly, the CA mechanism also exhibits over-concentration issues, often directing attention toward irrelevant objects, thereby weakening the model’s feature extraction capabilities. The results clearly demonstrate that NonLocalBlockND effectively distributes attention across transmission lines and foreign objects, integrating global information to help the model better understand the relationship between the target and its background—a key requirement for foreign object detection in power transmission lines.

3.8. Generalized Experiments

To verify the model’s performance in other fields or with different datasets and actual operation, we conducted validation experiments using FOTL (China Insulator Dataset, including insulator data from railway, communication, and power sectors) and CPLID (Southeast Asian Power Industry Dataset, including bird nests, branches, flying objects, and regional flora and fauna characteristics). The experimental results are shown in Table 10.
Table 10 states that the provided data shows the proposed improved model in this study demonstrates significant advantages over the FOTL and CPLID datasets on self-built datasets. Compared to the FOTL dataset, the Ours model improves accuracy (P) by 4.3%, recall rate (R) by 2.0%, mAP@0.5 by 4.1%, and mAP@0.5:0.95 by 6.4%. In comparison with the CPLID dataset, P improves by 6.9%, R by 9.8%, mAP@0.5 by 6.0%, and mAP@0.5:0.95 by 24.0%. These results indicate that the self-built dataset has advantages in data diversity and annotation quality, better supporting model training and optimization. Additionally, the model achieves an mAP@0.5 of 0.945 and 0.928 on the FOTL and CPLID datasets, respectively, demonstrating good adaptability; mAP@0.5:0.95 reaches 0.621 and 0.533, both superior to YOLOv8 and YOLOv10 models.

4. Conclusions

In this study, we proposed a multi-level cross-domain foreign object detection model tailored for power transmission lines. By integrating the YOLOv11 backbone with ConvNeXt-B and employing Bayesian optimization for hyperparameter tuning, the model effectively enhances detection accuracy and real-time performance, particularly for small and occluded targets, outperforming to of the classical models as in as follows
  • Compared to YOLOv11, CO-YOLO improves mAP@0.5 by 1.9% mAP@0.5:0.95 by 2.2%, with only a 12.1% reduction in FPS.
  • Compared to ETLSH-YOLO, it achieves improvements of 3.7% and 8.6% in mAP metrics.
  • When tested against TFD-YOLOv8, mAP@0.5 and mAP@0.5:0.95 increased by 33.0% and 80.4%, respectively.
CO-YOLO achieves 98.4% mAP@0.5, 66.1% mAP@0.5:0.95, and 303 FPS, outperforming all compared models.
The model also demonstrates excellent generalization on public datasets (FOTL, CPLID), showing gains of up to 6.4% mAP@0.5:0.95 compared to baseline models. These results highlight CO-YOLO’s high robustness, precision, and applicability in complex outdoor environments.
However, certain limitations persist. The model’s detection performance may still degrade under extreme environmental conditions (e.g., fog and snow) or for rare object types not present in the training set. To address this, we constructed a diverse dataset with simulated weather and resolution variations to enhance robustness.
Future research may also focus on: Lightweight models and acceleration, multimodal perception, and task-level generalization.

Author Contributions

Conceptualization, Q.L. and X.W.; methodology, X.W.; software, Y.S.; validation, Y.S., W.J. and Z.Z.; formal analysis, W.J.; investigation, Z.Z.; resources, F.S.; data curation, F.S.; writing—original draft preparation, L.Z.; writing—review and editing, L.Z.; visualization, Y.S.; supervision, X.W.; project administration, X.W.; funding acquisition, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Yunnan Key Laboratory of Intelligent Logistics Equipment and Systems under Grant 202449CE340008, in part by the Yunnan Shen Weiming Academician Workstation under Grant 202505AF350084, in part of by the College Student Innovation and Entrepreneurship Training Program under Grant 202411393012 and in part of by the Special Basic Cooperative Research Programs of Yunnan Provincial Undergraduate Universities’ Association under Grant 202301BA070001-003.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are not publicly available due to privacy and confidentiality restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chivunga, J.N.; Lin, Z.; Blanchard, R. Power Systems’ Resilience: A Comprehensive Literature Review. Energies 2023, 16, 7256. [Google Scholar] [CrossRef]
  2. Vazquez, D.A.Z.; Qiu, F.; Fan, N.; Sharp, K. Wildfire mitigation plans in power systems: A literature review. IEEE Trans. Power Syst. 2022, 37, 3540–3551. [Google Scholar] [CrossRef]
  3. Ferreira, V.H.; Zanghi, R.; Fortes, M.Z.; Sotelo, G.G.; Silva, R.D.B.M.; Souza, J.C.S.; Guimarães, C.H.C.; Gomes, S., Jr. A survey on intelligent system application to fault diagnosis in electric power system transmission lines. Electr. Power Syst. Res. 2016, 136, 135–153. [Google Scholar] [CrossRef]
  4. Cozza, A.; Pichon, L. Echo response of faults in transmission lines: Models and limitations to fault detection. IEEE Trans. Microw. Theory Tech. 2016, 64, 4155–4164. [Google Scholar] [CrossRef]
  5. Affijulla, S.; Tripathy, P. A robust fault detection and discrimination technique for transmission lines. IEEE Trans. Smart Grid 2017, 9, 6348–6358. [Google Scholar] [CrossRef]
  6. Kumar, B.R.; Mohapatra, A.; Chakrabarti, S.; Kumar, A. Phase angle-based fault detection and classification for protection of transmission lines. Int. J. Electr. Power Energy Syst. 2021, 133, 107258. [Google Scholar] [CrossRef]
  7. Chen, Z.; Yang, J.; Feng, Z.; Zhu, H. RailFOD23: A dataset for foreign object detection on railroad transmission lines. Sci. Data 2024, 11, 72. [Google Scholar] [CrossRef]
  8. Wu, Y.; Zhao, S.; Xing, Z.; Wei, Z.; Li, Y.; Li, Y. Detection of foreign objects intrusion into transmission lines using diverse generation model. IEEE Trans. Power Deliv. 2023, 38, 3551–3560. [Google Scholar] [CrossRef]
  9. Liu, C.; Ma, L.; Sui, X.; Guo, N.; Yang, F.; Yang, X.; Huang, Y.; Wang, X. YOLO-CSM-based component defect and foreign object detection in overhead transmission lines. Electronics 2023, 13, 123. [Google Scholar] [CrossRef]
  10. Tang, C.; Dong, H.; Huang, Y.; Han, T.; Fang, M.; Fu, J. Foreign object detection for transmission lines based on Swin Transformer V2 and YOLOX. Vis. Comput. 2024, 40, 3003–3021. [Google Scholar] [CrossRef]
  11. Peng, L.; Wang, K.; Zhou, H.; Li, H.; Yu, P. Detection of bolt defects on transmission lines based on multi-scale YOLOv7. IEEE Access 2024, 12, 156639–156650. [Google Scholar] [CrossRef]
  12. Fahim, S.R.; Sarker, Y.; Sarker, S.K.; Sheikh, M.R.I.; Das, S.K. Self attention convolutional neural network with time series imaging based feature extraction for transmission line fault detection and classification. Electr. Power Syst. Res. 2020, 187, 106437. [Google Scholar] [CrossRef]
  13. Kumar, V.R.; Jeyanthy, P.A.; Kesavamoorthy, R. Optimization-assisted CNN model for fault classification and site location in transmission lines. Int. J. Image Graph. 2024, 24, 2450008. [Google Scholar] [CrossRef]
  14. Zunair, H.; Khan, S.; Hamza, A.B. RSUD20K: A dataset for road scene understanding in autonomous driving. In Proceedings of the 2024 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 27–30 October 2024; pp. 708–714. [Google Scholar]
  15. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  16. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Boston, MA, USA, 7–12 June 2015; pp. 1440–1448. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer International Publishing: Cham, Germany, 2016; pp. 21–37. [Google Scholar]
  19. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  20. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  21. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  22. Nguyen, D.T.; Nguyen, T.N.; Kim, H.; Lee, H.J. A high-throughput and power-efficient FPGA implementation of YOLO CNN for object detection. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 2019, 27, 1861–1873. [Google Scholar] [CrossRef]
  23. Renwei, T.; Zhongjie, Z.; Yongqiang, B.; Ming, G.; Zhifeng, G. Key parts of transmission line detection using improved YOLO v3. Int. Arab J. Inf. Technol 2021, 18, 747–754. [Google Scholar] [CrossRef]
  24. Yaru, W.; Lilong, F.; Xiaoke, S.; Zhuo, Q.U.; Ke, Y.A.N.G.; Qianming, W.A.N.G.; Yongjie, Z.H.A.I. TFD-YOLOv8: A transmission line foreign object detection method. J. Graph. 2024, 45, 901. [Google Scholar]
  25. Li, S.J.; Liu, Y.X.; Li, M.; Ding, L. DF-YOLO: Highly accurate transmission line foreign object detection algorithm. IEEE Access 2023, 11, 108398–108406. [Google Scholar] [CrossRef]
  26. Yu, C.; Liu, Y.; Zhang, W.; Zhang, X.; Zhang, Y.; Jiang, X. Foreign objects identification of transmission line based on improved YOLOv7. IEEE Access 2023, 11, 51997–52008. [Google Scholar] [CrossRef]
  27. Zhao, L.; Zhang, Y.; Dou, Y.; Jiao, Y.; Liu, Q. ETLSH-YOLO: An Edge–Real-Time transmission line safety hazard detection method. Symmetry 2024, 16, 1378. [Google Scholar] [CrossRef]
  28. Zheng, J.; Liu, H.; He, Q.; Hu, J. GEB-YOLO: A novel algorithm for enhanced and efficient detection of foreign objects in power transmission lines. Sci. Rep. 2024, 14, 15769. [Google Scholar] [CrossRef] [PubMed]
  29. Shao, Y.; Zhang, R.; Lv, C.; Luo, Z.; Che, M. TL-YOLO: Foreign-Object detection on power transmission line based on improved Yolov8. Electronics 2024, 13, 1543. [Google Scholar] [CrossRef]
  30. Duan, P.; Liang, X. An Improved YOLOv8-Based Foreign Detection Algorithm for Transmission Lines. Sensors 2024, 24, 6468. [Google Scholar] [CrossRef]
  31. Xu, W.; Xiwen, C.; Haibin, C.; Yi, C.; Jun, Z. Foreign object detection method in transmission lines based on improved yolov8n. In Proceedings of the 10th International Symposium on System Security, Safety, and Reliability (ISSSR), IEEE, Xiamen, China, 30–31 March 2024; pp. 196–200. [Google Scholar]
Figure 1. YOLOv11 network structure.
Figure 1. YOLOv11 network structure.
Sensors 25 05141 g001
Figure 2. The C3 k2-Dual module structure.
Figure 2. The C3 k2-Dual module structure.
Sensors 25 05141 g002
Figure 3. The dual-core convolution principle.
Figure 3. The dual-core convolution principle.
Sensors 25 05141 g003
Figure 4. Principle of the non-local attention mechanism.
Figure 4. Principle of the non-local attention mechanism.
Sensors 25 05141 g004
Figure 5. C2PSA-DHSA module principle: (a) PSADHSA module structural; (b) Structure of the C2PSA-DHSA module.
Figure 5. C2PSA-DHSA module principle: (a) PSADHSA module structural; (b) Structure of the C2PSA-DHSA module.
Sensors 25 05141 g005
Figure 6. Principles of the attention mechanism of DHSA.
Figure 6. Principles of the attention mechanism of DHSA.
Sensors 25 05141 g006
Figure 7. The DySample module structure.
Figure 7. The DySample module structure.
Sensors 25 05141 g007
Figure 8. Two different range factors.
Figure 8. Two different range factors.
Sensors 25 05141 g008
Figure 9. The refined YOLOv11 structure.
Figure 9. The refined YOLOv11 structure.
Sensors 25 05141 g009
Figure 10. ConvNeXt network structure.
Figure 10. ConvNeXt network structure.
Sensors 25 05141 g010
Figure 11. The CO-YOLO network structure.
Figure 11. The CO-YOLO network structure.
Sensors 25 05141 g011
Figure 12. Part of the dataset.
Figure 12. Part of the dataset.
Sensors 25 05141 g012
Figure 13. Histogram of basic model experimental values.
Figure 13. Histogram of basic model experimental values.
Sensors 25 05141 g013
Figure 14. Other high-level modules compare data histograms.
Figure 14. Other high-level modules compare data histograms.
Sensors 25 05141 g014
Figure 15. Visual comparison of the attention mechanism.
Figure 15. Visual comparison of the attention mechanism.
Sensors 25 05141 g015
Table 1. Environment configuration.
Table 1. Environment configuration.
ItemConfiguration
Operation SystemWin11
CPU12th Gen Intel CoreTM i9-12900H
GPUNVIDIA GeForce RTX 3060 Laptop GPU (6 GB)
Python3.8
PyTorch2.1.2
Cuda10.0
Table 2. Detailed partitioning of datasets.
Table 2. Detailed partitioning of datasets.
TypeNestBranchBalloonPlasticAll
Number
Origin315132352180979
Enhanced5764986793282081
Table 3. Basic model comparison.
Table 3. Basic model comparison.
ModelTypeGFLOPsPramasP (Precision)R (Recall)mAP@0.5mAP@0.5:0.95FPS
Faster R-CNNTwo-stage1984287; 50000.5760.6010.9560.577192
Mask R-CNNTwo-stage2274439; 60000.5350.5980.9570.56184
Cascade R-CNNTwo-stage2016916; 10000.5670.6220.9620.591225
Fast R-CNNTwo-stage1314070; 00000.5960.6180.9660.591186
Libra R-CNNTwo-stage1804162; 70000.5670.6320.9780.585210
YOLOv5One-stage7.1250; 37240.9620.9580.9740.633323
YOLOv8One-stage8.1300; 64280.9610.9180.9660.628294
YOLOv10One-stage8.2269; 59760.9170.930.9520.629313
YOLOv11One-stage6.3258; 29320.9580.9460.9650.639345
OursOne-stage7.2282; 16960.960.9450.9840.661303
The bold part is the best value; GFLOPs = FLOPs/109.
Table 4. Contrasts of different attention mechanisms.
Table 4. Contrasts of different attention mechanisms.
ModelPramasPRmAP@0.5mAP@0.5:0.95FPS
CBAM2,648,8220.9670.9290.9740.634286
ECA2,624,5350.9610.9670.9780.638286
SCSA2,730,9080.9590.9620.9860.638278
Ours2,715,1560.9620.9630.9840.649294
The bold part is the best value.
Table 5. Comparison of different C2PSA modules.
Table 5. Comparison of different C2PSA modules.
ModelPramasPRmAP@0.5mAP@0.5:0.95
C2PSA-ACmix2,737,0440.9430.9470.9740.625
C2PSA-DAT2,755,2280.9710.9520.9680.633
C2PSA-EMA2,677,5640.9460.9440.9650.633
C2PSA-FL2,743,9000.8810.8560.9210.545
Ours2,637,0800.9550.9560.9710.637
The bold part is the best value.
Table 6. Comparison of different C3K2 modules.
Table 6. Comparison of different C3K2 modules.
ModelPramasPRmAP@0.5mAP@0.5:0.95FPS
C3K2-MLLAB2,729,5960.9370.960.9720.639203
C3K2-RFA2,953,8040.9520.9640.9690.643185
C3K2-SC2,775,8520.9610.950.9830.634250
Ours2,877,4200.9340.950.9750.644256
The bold part is the best value.
Table 7. Comparison of the different loss functions.
Table 7. Comparison of the different loss functions.
ModelPramasPRmAP@0.5mAP@0.5:0.95FPS
EIoU2,726,8120.9540.9560.9730.624313
WIoU2,726,8120.9610.950.9760.635294
DIoU2,726,8120.9510.9540.9720.65294
SIoU2,726,8120.9540.9560.9730.624303
The bold part is the best value.
Table 8. Comparison of the other advanced modules.
Table 8. Comparison of the other advanced modules.
ModelPRmAP@0.5mAP@0.5:0.95
ETLSH-YOLO [27]0.9350.9200.9390.595
DF-YOLO [25]0.9580.9150.9520.576
GEB-YOLO [28]0.9720.9340.955
TL-YOLO [29]0.9060.8860.9130.632
GCP-YOLO [30]0.8340.8120.8960.677
TFD-YOLOv8 [31]0.913 0.629 0.732 0.358
Ours0.9600.9450.9740.646
The bold part is the best value.
Table 9. Experimental data comparison of ablation.
Table 9. Experimental data comparison of ablation.
ModelC2PSA-DHSAC3k2-DualNonLocalBlockNDDySampleLossGFLOPsPramasPRmAP@0.5mAP@0.5:0.95FPS
YOLOv11 6.32,582,9320.9580.9460.9650.639345
YOLOv11 6.42,637,0800.9550.9560.9710.637294
YOLOv11 6.32,595,2840.9580.9560.9740.638270
YOLOv11 7.52,931,5680.970.9610.9790.642294
YOLOv11 6.52,769,3040.970.9440.9820.654270
YOLOv11 6.42,649,4320.960.9170.9650.633270
YOLOv11 7.63,063,7920.9610.9410.9730.646263
YOLOv11 7.52,943,9200.9640.9070.970.634263
YOLOv11 7.63,021,9960.9620.9260.9720.645286
YOLOv11 6.52,781,6560.9650.9410.9780.643263
YOLOv11 7.63,076,1440.9520.9460.9690.632263
Ours7.22,821,6960.960.9450.9840.641303
The bold part is the best value.
Table 10. Comparison experiment of different datasets.
Table 10. Comparison experiment of different datasets.
Date setsPRmAP@0.5mAP@0.5:0.95
FOTL0.9200.9260.9450.621
CPLID0.8980.8610.9280.533
Ours0.9600.9450.9840.661
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Q.; Wang, X.; Su, Y.; Jiang, W.; Zhang, Z.; Shen, F.; Zhu, L. Research on Deep Learning-Based Multi-Level Cross-Domain Foreign Object Detection in Power Transmission Lines. Sensors 2025, 25, 5141. https://doi.org/10.3390/s25165141

AMA Style

Liu Q, Wang X, Su Y, Jiang W, Zhang Z, Shen F, Zhu L. Research on Deep Learning-Based Multi-Level Cross-Domain Foreign Object Detection in Power Transmission Lines. Sensors. 2025; 25(16):5141. https://doi.org/10.3390/s25165141

Chicago/Turabian Style

Liu, Qingxue, Xia Wang, Yun Su, Wei Jiang, Zhe Zhang, Fuyu Shen, and Lizitong Zhu. 2025. "Research on Deep Learning-Based Multi-Level Cross-Domain Foreign Object Detection in Power Transmission Lines" Sensors 25, no. 16: 5141. https://doi.org/10.3390/s25165141

APA Style

Liu, Q., Wang, X., Su, Y., Jiang, W., Zhang, Z., Shen, F., & Zhu, L. (2025). Research on Deep Learning-Based Multi-Level Cross-Domain Foreign Object Detection in Power Transmission Lines. Sensors, 25(16), 5141. https://doi.org/10.3390/s25165141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop