You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

8 August 2025

YOLOv5s-TC: An Improved Intelligent Model for Insulator Fault Detection Based on YOLOv5s

,
,
,
and
College of Information Technology, Jilin Agricultural University, Changchun 130118, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Deep Power Vision Technology and Intelligent Vision Sensors: 2nd Edition

Abstract

Insulators play a pivotal role in power grid infrastructure, offering indispensable electrical insulation and mechanical support. Precise and efficient detection of insulator faults is of paramount importance for safeguarding grid reliability and ensuring operational safety. With the rapid advancements in UAV (unmanned aerial vehicle) technology and deep learning, there has been a notable transition from traditional manual inspections to automated UAV-based detection systems. To further enhance detection accuracy, this study conducts a series of systematic improvements to the YOLOv5s model and proposes an advanced intelligent insulator detection model, namely YOLOv5s-TC. Firstly, this new model replaces the C3 (Cross Stage Partial Bottleneck with 3 convolutions) module with Bottleneck Transformers to enhance feature learning ability. Secondly, the CBAM (Convolutional Block Attention Module) is introduced to make the model focus more on the key features of the images, thus improving the target localization ability. Finally, the improved loss function named OSIoU is adopted to further enhance detection accuracy. Comparative experiments demonstrate that YOLOv5s-TC achieves significant performance gains, with mean average precision improvements of 4.4%, 24.5%, and 13.9% over the original YOLOv5s, Faster R-CNN, and SSD models, respectively. The results indicate that YOLOv5s-TC offers superior detection performance and greater reliability for practical power grid inspection applications.

1. Introduction

As one of the most complex and critical human-engineered systems, the power grid serves as fundamental energy infrastructure underpinning socioeconomic development. Acting as the significant energy conduit for a nation, it plays a pivotal role in sustaining daily life, industrial operations, and social stability. For power utilities, ensuring secure and reliable grid operation constitutes the paramount operational objective.
Insulators, acting as crucial components in power transmission networks, perform dual core functions: providing mechanical support for conductors and establishing electrical insulation between transmission lines and tower structures. Under actual operating conditions, these components endure prolonged exposure to multiple environmental stressors, including solar radiation, ice accumulation, sandstorms, and high-voltage electrical stresses []. Such harsh operational environments accelerate material aging, leading to various failure modes such as string separation, mechanical fractures, and insulation flashover. Figure 1a and Figure 1b present a sound insulator and a damaged insulator, respectively. Compromised insulators significantly impair power transmission quality, with current statistics indicating that approximately 80% of transmission failures originate from insulator defects []. Consequently, implementing systematic inspection protocols, preventive maintenance programs, and timely defect remediation through cleaning or replacement emerges as a critical strategy for preventing transmission line failures and ensuring overall grid reliability.
Figure 1. Pictures of insulators. (a) Sound insulator. (b) Damaged insulator.
Currently, insulator fault detection is undergoing a significant transformation from traditional manual methods to intelligent UAV-based approaches. The intelligent detection system employs high-definition cameras as primary imaging devices coupled with deep learning algorithms for precise object detection and defect identification. Compared to conventional manual inspection, UAV-based intelligent detection offers substantial advantages by overcoming challenges such as high labor intensity, low efficiency, and significant safety risks. Furthermore, it demonstrates superior adaptability to diverse natural environments, including complex weather conditions and terrains, while reducing power outage frequencies, making it the optimal inspection solution.
To address the challenges of low recognition accuracy and difficult image recognition in intelligent insulator defect detection, this study proposes an enhanced YOLOv5s-TC detection model through improvements to the existing YOLOv5s framework. The proposed model incorporates three key innovations: (1) replacement of the C3 module with Bottleneck Transformers to enhance feature extraction capability, (2) integration of the CBAM to improve feature localization, and (3) implementation of an optimized Scaled Intersection over Union (OSIoU) loss function to boost detection accuracy. Additionally, this research establishes an insulator image dataset that not only expands existing data resources but also yields superior model parameters through training.

3. Improved Model: YOLOv5s-TC

3.1. The Architecture of YOLOv5s-TC

The overall structure of YOLOv5s-TC comprises an input layer, a backbone network, a neck network, and an output layer. Compared with YOLOv5s, the Focus module, CSP (Cross Stage Partial) module, and SPP (Spatial Pyramid Pooling) module remain unchanged. The main alterations lie in the number of network layers and parameters.
Based on YOLOv5s, the improvements in YOLOv5s-TC are as follows: a C3TR (C3 Transformer) module and a CCBL (Convolutional attention mechanism Convolutional Block Layer) module are newly introduced, and the loss function is adjusted.
In the new C3TR module, a Bottleneck Transformers structure is introduced to reduce the computational burden, shorten the model training time, and improve the detection accuracy. The new CCBL module adds a convolutional attention mechanism after the original CBL (Convolutional Block Layer) module, aiming to focus on the key information. The main adjustment to the loss function lies in the redefinition of the distance cost function. Figure 2 shows the network architecture of YOLOv5s-TC.
Figure 2. YOLOv5s-TC network architecture.

3.2. Replacement of C3 Module with C3TR

This improvement is inspired by BoTs (Bottleneck Transformers). The core of BoT is the BoT block, which is constructed by replacing the spatial 3 × 3 convolution in the ResNet (Residual Network) bottleneck block with an MHSA (Multi-Head Self-Attention) layer.
The MHSA module is a key component of the Transformer architecture, playing a significant role in sequential data processing, especially in natural language processing tasks. Its design enables the Transformer to effectively capture long-term dependencies. The MHSA module takes a feature map X of dimensions H × W × d as input, where H and W are the height and width of the feature map, and d is the dimensionality of the feature vector per point. Feature map X generates the query (q), key (k), and value (v) through three 1 × 1   convolutions. Relative position encodings, Rh and Rw, are introduced to account for the spatial positions of elements. Two attention scores, qkT and qrT, are calculated and summed to represent the importance of different positions in the input for output generation. These scores are normalized by the softmax function. The normalized weights are then used to weight-sum and aggregate v, generating the output feature map Z that encapsulates context-aware information for downstream tasks. The MHSA architecture is shown in Figure 3.
Figure 3. MHSA architecture.
In the YOLOv5s-TC model, the 3 × 3 convolutional layer in the ResNet structure is also replaced by the MHSA, and then it is encapsulated into C3TR. Subsequently, the C3TR is further used to replace the C3 module in YOLOv5s. This improvement combines the local perception ability of CNNs with the global perception ability of Transformers, enabling the extraction of useful features from the input images. As a result, the detection accuracy is significantly improved on the premise that the number of model layers, the number of network parameters, the computational load, and the training time have not changed. Therefore, this particular improvement enhances the detection performance of the model without consuming more computing power.

3.3. Addition of the Attention Mechanism

The attention mechanism in neural networks is an important technique for enhancing the performance of models in processing sequential data. In 2018, the CBAM attention mechanism model was proposed [], which aims to improve the performance of convolutional neural networks.

3.3.1. The Design of the Channel Attention Module

The design of the channel attention module in convolutional attention is illustrated in Figure 4.
Figure 4. Channel attention module.
For each channel, the GAP (Global Average Pooling) calculates the mean value of all spatial positions on that channel, thus compressing the spatial dimension of each channel into a single scalar. Subsequently, a lightweight FC (Fully Connected) network is employed to transform the compressed vector with the purpose of learning the non-linear relationships among channels and generating the channel weights. The FC network consists of two layers. One layer is responsible for dimensionality reduction to decrease the number of parameters and computational cost, while the other layer is tasked with dimensionality increase to restore the number of channels to the original value. The ReLU (Rectified Linear Unit) activation function is utilized between the two layers, and the sigmoid activation function is applied in the last layer to output the weight coefficient of each channel. The values of the weight coefficients all range between 0 and 1. Its expression is shown in Equation (1).
A c ( F ) = σ ( M L P ( F a v g c ) M L P ( F m a x c ) ) ,
where σ represents the sigmoid activation function, MLP represents the Multi-Layer Perceptron, F a v g c represents the feature value obtained after the sample goes through average pooling, F m a x c represents the feature value obtained after the sample goes through max pooling, and “⊕” represents the addition of features.
Finally, by multiplying the weight of each channel by the features of the corresponding channel in the original feature map, the channels that are more important for the current task are highlighted, while the unimportant channels are suppressed.

3.3.2. The Spatial Attention Module

In order to identify the important positions in the feature map, complement the channel attention mechanism, focus on the spatial distribution of the feature map, and suppress the unimportant regions, a spatial attention module is introduced, as shown in Figure 5.
Figure 5. Spatial attention module.
Firstly, we integrate different channel information of the feature map to generate a spatial attention map. This process starts with the compression of the input feature map. Specifically, we use the max- pooling and average pooling operations to aggregate the channel dimension, converting it into a one-dimensional vector feature map of size HW1. Subsequently, we concatenate the results of these two pooling operations, obtaining a two-channel feature map that contains both the maximum and average values. Then, a 7 × 7 convolution operation is applied to reduce the number of channels to 1. Next, a lightweight convolutional network is employed to process the fused feature map. This network is designed to learn spatial dependencies and generate the spatial attention map. For efficiency, it consists of one or more convolutional layers with small-sized kernels such as 3 × 3 or 1 × 1. The final convolutional layer incorporates a sigmoid activation function, which ensures that the values within the attention map range from 0 to 1, effectively representing the significance of each spatial position. Finally, we perform element-wise multiplication between the original feature map and the generated spatial attention map. This operation dynamically adjusts the feature responses at each position, enabling the model to concentrate on crucial spatial regions while mitigating the influence of less important areas. The weighting operation is
A s ( F ) = s i g m o i d ( f C o n v 7 7 ( [ F a v g S ; F m a x S ] ) )

3.3.3. The CBAM Structure Is Integrated into the Backbone Network of YOLOv5s

To more effectively enhance the feature extraction of distant targets by the backbone network, we integrate the CBAM structure into the backbone network of YOLOv5s, forming a new backbone feature extraction network, the CSP_CBAM_N structure. This reduces the impact of background noise on images and improves the accuracy of insulator defect detection. The structure diagram is shown in Figure 6.
Figure 6. CSP_CBAM_N architecture.
The specific locations for incorporating the CBAM (Convolutional Block Attention Module) mechanism within the YOLOv5s architecture will be elucidated based on experimental data.

3.4. Adjustment and Improvement of the Loss Function

The localization loss function is crucial in object detection. It evaluates the difference between predicted and ground truth bounding boxes, serving as a simple yet effective performance metric. Common localization loss functions include IoU (Intersection Over Union Loss), GIoU (Generalized Intersection Over Union Loss) [], DIoU (Distance Intersection Over Union Loss) [], CIoU (Complete Intersection Over Union Loss) [], and SIoU (Scale Invariant Over Union Loss). Based on their characteristics, we adopted the SIoU function and improved it.
The SIoU loss function contains four elements, namely the angular cost, the distance cost, the shape cost, and the IoU cost. The angular cost is shown in Equation (3), where α represents the horizontal angle between the center points of the predicted bounding box and the ground truth bounding box.
Λ = 1 2 × sin 2 ( a r c s i n ( s i n ( α ) π 4 ) ,
According to Equation (3), we redefine the distance cost function, as shown in Equation (4).
= t = x y ( 1 e γ ρ t ) ,
when γ = 2−Λ and t = x, ρ t represents the difference between the widths of the ground truth bounding box and the predicted bounding box divided by the square of the width of the minimum enclosing rectangle of the ground truth bounding box and the predicted bounding box. When t = y, ρ t represents the difference between the heights of the ground truth bounding box and the predicted bounding box divided by the square of the height of the minimum enclosing rectangle of the ground truth bounding box and the predicted bounding box.
The expressions of the improved shape cost are shown in Equations (5) and (6).
Ω =   t = x y ( 1 e ω t ) θ ,
W ω   =   | ω ω g t | m a x ( ω , ω g t ) ,   W h =   | h h g t | m a x ( h , h g t )
where (ω, h) represents the width and height of the predicted bounding box, and ( ω g t , h g t ) represents the width and height of the ground truth bounding box. θ is used to adjust the degree of attention to the shape loss to prevent excessive focus on the shape loss, which may reduce attention to the position of the predicted bounding box. In order to ensure appropriate adjustment, the value range of θ is limited to [2, 6]. The IoU cost function is still used to calculate the degree of overlap between the two bounding boxes.
The improved SIoU loss function is named OSIoU, and its final expression is
L O S S I S O o U   = 1 I o U + + Ω 2 ,
In this subsection, the improvements made to YOLOv5s are introduced. In the next subsection, based on the experimental and model training results, the regulatory effect of the loss function and the placement of the attention module will be explored. Additionally, the experimental data of YOLOv5s-TC will be compared with the experimental data of various other mainstream models to determine the superiority of YOLOv5s-TC.

4. Experimental Results and Analysis

4.1. Dataset Construction

The majority of insulator defect images were self-collected using UAVs, while a small portion were sourced from an electric power company. We also expanded the dataset through Mosaic data augmentation methods. The expansion ratio of data augmentation is 4:1. The entire dataset consists of 8700 images, and it is divided into three parts at a ratio of 7:2:1. Specifically, 6090 images are used as the training set, 1740 images as the validation set, and 870 images as the test set. A schematic diagram of some samples in the dataset is shown in Figure 7.
Figure 7. Some samples from the dataset.
The types of insulator defects in the dataset include four categories: damage, string drop, flashover, and stains. The materials of the insulators cover ceramic insulators, glass insulators, and composite insulators. There are also various suspension methods for the insulators. Additionally, the backgrounds of the dataset include diverse and complex terrains such as mountains, rivers, farmlands, grasslands, and deserts, making it highly representative.

4.2. Experimental Parameter Settings

We use Python 3.8 as the programming language, PyTorch1.11.0 as the deep learning framework, NVIDIA GeForce RTX 4060 as the GPU, and we also adopt CUDA 11.4 for acceleration. The hyperparameters used are shown in Table 1. The evaluation indicators include precision and recall rates, mean average precision (mAP), and F1 score (the harmonic mean of the precision and recall rates) [].
Table 1. Hyperparameter usage statistics.

4.3. Comparative Experiment of CBAM Location

In Section 3.3, the specific method for integrating CBAM into YOLOv5s is discussed. The specific insertion positions of CBAM are determined based on experimental data. As shown in Figure 8, where numbers 1–8 correspond to different CBL modules, a comparative analysis of the experimental results reveals that the optimal detection performance is achieved when the CBAM (Convolutional Block Attention Module) is simultaneously added after the convolutional layers of the first, second, sixth, seventh, and eighth CBL modules in YOLOv5s. The mAP value reaches 76.1%, representing a 4.4% improvement over the original YOLOv5s algorithm. This improvement enables the front-placed attention mechanism to not only focus on all key information in insulator images but also, through subsequent attention mechanisms, further emphasize critical information in feature maps, thereby enhancing the detection accuracy of defects in insulator images.
Figure 8. Experimental result of different CBAM adding positions.

4.4. Comparative Experiment of Different Loss Functions

We carried out recognition experiments on the CIoU, DIoU, GIoU, and OSIoU (Optimize SIoU loss function) by employing the YOLOv5s model on our self-constructed dataset. The experimental results are presented in Table 2. Notably, the mAP of the recognition process with the adoption of the OSIoU loss function was 74.4%. Both its precision and mAP surpass those of the other loss functions. Thus, it can be concluded that the modification of the loss function in YOLOv5s yielded certain positive outcomes. The improvements in YOLOv5s make the regression and positioning of defect bounding boxes more accurate, and the model becomes more robust to noise.
Table 2. Comparison of results with different loss functions.

4.5. Ablation Experiment

Further ablation experiments were conducted after replacing partial C3 modules in the backbone of YOLOv5s with Bottleneck Transformers, introducing the CBAM attention mechanism into the YOLOv5s model, and replacing the loss function with OSIoU. These experiments aimed to analyze the contribution of each individual improvement or pairwise combination among the three modifications to model performance, quantifying the independent effects and synergistic gains of each innovation. The results are presented in Table 3, where A denotes the replacement of partial C3 modules in the backbone with Bottleneck Transformers, B represents the introduced CBAM attention mechanism, and C indicates the adoption of OSIoU as the loss function. As shown in the results, simultaneous improvements in these three aspects lead to increases in precision, recall, and mean average precision, demonstrating that the three modifications exert a synergistic effect on model performance.
Table 3. Ablation experiment results.
The experimental results indicate that each of the three improvements to YOLOv5s contributes marginally independently. The detection performance of pairwise combinations is comparable to that of using only Bottleneck Transformers. However, when all three improvements act synergistically, all evaluation metrics show significant enhancements, demonstrating that their combined effect yields the strongest model gain. Thus, progress has been made in improving YOLOv5s. Subsequent work will compare the detection results of YOLOv5s-TC with other models to verify its superiority.

4.6. Comparative Experiment of YOLOv5s-TC

Based on the aforementioned experimental results, the optimization scheme for the loss function and the insertion positions of the CBAM were determined. With this, the construction of the YOLOv5s-TC model was completed. We trained the model using the dataset described in Section 4.1, and the results are shown in Figure 9.
Figure 9. The result of the YOLOv5s-TC model’s training process.
From the training results, it can be seen that various indicators of the YOLOv5s-TC model gradually start to converge after 200 training iterations. When the number of training iterations reaches 300, all the training indicators tend to stabilize. To further verify the superiority of YOLOv5s-TC, we selected YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x, along with four current mainstream algorithms, namely Faster R-CNN, SSD, YOLOv3, and YOLOv4, for comparative experiments. The comparison results demonstrate the performance of different insulator defect types and algorithms in terms of precision, recall rate, mAP, and F1 score. Table 4 and Table 5 present the comparative results of various models.
Table 4. Comparison of average detection precision for different types of defects.
Table 5. Comparison of models’ common indicators in insulator defect detection.
As can be seen from Table 3, the Faster R-CNN algorithm has the lowest average precision compared to several other detection algorithms, and situations such as missed detections and false detections may occur. In contrast, the improved YOLOv5s-TC algorithm exhibits the best detection performance for various types of defects as well as normal conditions. Compared with the original YOLOv5s algorithm, the mean Average Precision (mAP) of the YOLOv5s-TC algorithm increased by 4.1%, 9.4%, 10.8%, and 12% for normal insulators, string-missing insulators, damaged insulators, and flashover insulators, respectively.
Meanwhile, compared with YOLOv5m, which shows the best performance in detecting string-missing insulators among the other seven models, the mAP of the YOLOv5s-TC algorithm increased by 8.5%. When compared with YOLOv5x, which performs best in detecting damaged insulators among the other seven models, the mAP of the YOLOv5s-TC algorithm increased by 10.5%. Moreover, when compared with YOLOv5l, which has the best performance in detecting flashover insulators among the other seven models, the mAP of the YOLOv5s-TC algorithm increased by 10%.
These results indicate that the YOLOv5s-TC model significantly outperforms other algorithms in detecting the three types of insulator defects, namely string drop, damage, and flashover.
Upon examining Table 4, it is evident that the improved algorithm, YOLOv5s-TC, demonstrates superiority over other algorithms with respect to precision, recall rate, and F1 score. This further substantiates that the improved algorithm exhibits more favorable performance in the target detection task, and its detection outcomes are more dependable.

4.7. Comparison of Model Structure Loss Values Between YOLOv5s and YOLOv5s-TC

We compared the original YOLOv5s model and the improved YOLOv5s-TC model in terms of bounding box regression loss, confidence loss, and classification probability loss, and the results are shown in Figure 10.
Figure 10. Training loss curve.
It is clearly evident from Figure 10 that in the early stage of training (0 to 10 epochs), the losses of the two models were similar. During the subsequent training process, the three losses of YOLOv5s-TC converged faster, and the loss values were consistently lower than those of the original model. When training reached approximately 250 epochs, the losses of both models converged and stabilized, but the loss value of YOLOv5s-TC was still lower. This indicates that the improved model did not experience overfitting, and it demonstrated reliable performance during both the training and testing stages, once again confirming its reliability and effectiveness in object detection.

4.8. Comparison of Computational Efficiency Between YOLOv5s and YOLOv5s-TC

We compared the computational efficiency of the original YOLOv5s and improved YOLOv5s-TC, and the results are presented in Table 6. The data therein show that YOLOv5s-TC has significantly fewer parameters and lower computational costs. Combined with the mAP metrics in Table 5, it is clear that the optimized model reduces computational resources while enhancing the mAP. These findings demonstrate the practical value of the proposed optimization for insulator detection, highlighting the advantages of the improved model.
Table 6. Comparison of computational efficiency between models.

5. Conclusions

This study proposes an improved YOLOv5s-based model for defect detection of insulators in smart grids, with several key innovations: the C3 module in the YOLOv5s backbone is replaced with Bottleneck Transformers, the CBAM attention mechanism is introduced into the YOLOv5s model, and the more effective OSIoU loss function is adopted to replace the original loss function. These improvements significantly enhance detection accuracy and speed while reducing network parameters and computational load.
The experimental results show that the enhanced model outperforms traditional methods, including other earlier YOLOv5 versions and other object detection networks such as Faster R-CNN and SSD. Specifically, the model achieves an average detection precision of 96.5%. Comparative and ablation studies verify the contributions of each module, comprehensively confirming its significant superiority in the field of insulator defect detection. The research findings open up new ideas and lay a solid theoretical foundation for unmanned intelligent inspection of insulators, with high practical engineering application value.
In future research, it is advisable to further explore lightweight designs to adapt to deployment on edge devices. Integrating multimodal data (such as the fusion of infrared and visible light images) can be considered to improve detection capabilities in complex environments. Additionally, introducing distributed learning frameworks like federated learning into the model training process is expected to enable the application of large-scale intelligent power grid inspection while ensuring data privacy.

Author Contributions

Conceptualization, Y.Y. and Y.D.; methodology, Y.D.; software, X.W.; validation, S.H.; formal analysis, C.Z.; investigation, C.Z.; resources, S.H.; data curation, S.H.; writing—original draft preparation, X.W.; writing—review and editing, Y.Y. and C.Z.; funding acquisition, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Jilin Province, grant number 20230101175JC; the Research Project of the Education Department of Jilin Province, grant number JJKH20250568KJ; and the Provincial-level Innovation Training Program for College Students of Jilin Agricultural University in 2024, grant number 228.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pratibha, S.; Sumeet, S.; Sanjay, S. Object detection in power line infrastructure: A review of the challenges and solutions. Eng. Appl. Artif. Intell. 2024, 130, 107781. [Google Scholar] [CrossRef]
  2. Li, Y.; Yang, H.; Xu, Y.; Zhou, J.; Zhao, J. Research on UAV inspection technology for self-exploding insulators based on AlexNet network. Smart Power 2021, 49, 104–109. [Google Scholar]
  3. Yu, Z.; Lei, Y.; Shen, F.; Zhou, S.; Yuan, Y. Research on Identification and Detection of Transmission Line Insulator Defects Based on a Lightweight YOLOv5 Network. Remote Sens. 2023, 15, 4552. [Google Scholar] [CrossRef]
  4. Ortego, J.; Garnacho, F.; Alvarez, F.; Arcones, E.; Khamlichi, A. Locating Insulation Defects in HV Substations Using HFCT Sensors and AI Diagnostic Tools. Sensors 2024, 24, 5312. [Google Scholar] [CrossRef]
  5. Kumar, P.; Li, X.; Wang, Y. Deep learning-based crack detection in glass insulators. Eng. Appl. Artif. Intell. 2023, 120, 106789. [Google Scholar] [CrossRef]
  6. Liu, Y.; Huang, X. Efficient Cross-Modality Insulator Augmentation for Multi-Domain Insulator Defect Detection in UAV Images. Sensors 2024, 24, 428. [Google Scholar] [CrossRef]
  7. Zhang, T.; Zhang, Y.; Xin, M.; Liao, J.; Xie, Q. A Light-Weight Network for Small Insulator and Defect Detection Using UAV Imaging Based on Improved YOLOv5. Sensors 2023, 23, 5249. [Google Scholar] [CrossRef]
  8. Jeffrey, K.; Chen, S.; Huang, C. Automatic detection, classification and localization of defects in large photovoltaic plants using unmanned aerial vehicles (UAV) based infrared (IR) and RGB imaging. Energy Convers. Manag. 2023, 276, 116495. [Google Scholar] [CrossRef]
  9. Souza, B.J.; Stefenon, S.F.; Singh, G.; Freire, R.Z. Hybrid-YOLO for classification of insulators defects in transmission lines based on UAV. Int. J. Electr. Power Energy Syst. 2023, 148, 108982. [Google Scholar] [CrossRef]
  10. Patel, S.; Yang, F.; Johnson, E. Thermal anomaly detection in insulators using UAV and deep learning. Sustain. Energy Technol. Assess. 2022, 52, 102345. [Google Scholar] [CrossRef]
  11. Jae Jin, I.; Yeong Lim, D.; Cheol Bang, I. Development of fault diagnosis for nuclear power plant using deep learning and infrared sensor equipped UAV. Ann. Nucl. Energy 2023, 181, 109577. [Google Scholar] [CrossRef]
  12. Johnson, E.; Smith, A.; Wu, Q. Few-shot learning for insulator detection in low-resource scenarios. Neural Netw. 2022, 156, 234–245. [Google Scholar] [CrossRef]
  13. Chen, W.; Li, Y.; Zhao, Z. Missing-Sheds Granularity Estimation of Glass Insulators Using Deep Neural Networks Based on Optical Imaging. Sensors 2022, 22, 1737. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, X.; Miao, X.; Jiang, H.; Chen, J.; Wu, M.; Chen, Z. Tower Masking MIM: A Self-Supervised Pretraining Method for Power Line Inspection. IEEE Trans. Ind. Inform. 2023, 1, 20. [Google Scholar] [CrossRef]
  15. Wang, D.; Yang, F.; Kim, B. A data-augmentation method for small-sample insulator defect detection. Measurement 2023, 198, 113456. [Google Scholar] [CrossRef]
  16. Wu, J.; Deng, Q.; Xian, R.; Tao, X.; Zhou, Z. An Instance Segmentation Method for Insulator Defects Based on an Attention Mechanism and Feature Fusion Network. Appl. Sci. 2024, 14, 3623. [Google Scholar] [CrossRef]
  17. Qu, F.; Lin, Y.; Tian, L.; Du, Q.; Wu, H.; Liao, W. Lightweight Oriented Detector for Insulators in Drone Aerial Images. Drones 2024, 8, 294. [Google Scholar] [CrossRef]
  18. Sun, S.; Chen, C.; Yang, B.; Yan, Z.; Wang, Z.; He, Y.; Wu, S.; Li, L.; Fu, J. ID-Det: Insulator Burst Defect Detection from UAV Inspection Imagery of Power Transmission Facilities. Drones 2024, 8, 299. [Google Scholar] [CrossRef]
  19. Fang, G.; An, X.; Fang, Q.; Gao, S. Insulator Defect Detection Based on YOLOv5s-KE. Electronics 2024, 13, 3483. [Google Scholar] [CrossRef]
  20. Ji, Y.; Zhang, D.; He, Y.; Zhao, J.; Duan, X.; Zhang, T. Improved YOLO11 Algorithm for Insulator Defect Detection in Power Distribution Lines. Electronics 2025, 14, 1201. [Google Scholar] [CrossRef]
  21. Wang, T.; Zhai, Y.; Li, Y.; Wang, W.; Ye, G.; Jin, S. Insulator Defect Detection Based on ML-YOLOv5 Algorithm. Sensors 2024, 24, 204. [Google Scholar] [CrossRef] [PubMed]
  22. Zhao, J.; Miao, S.; Kang, R.; Cao, L.; Zhang, L.; Ren, Y. Insulator Defect Detection Algorithm Based on Improved YOLOv11n. Sensors 2025, 25, 1327. [Google Scholar] [CrossRef]
  23. Chai, Y.; Yao, X.; Chen, M.; Shan, S. FPFS-YOLO: An Insulator Defect Detection Model Integrating FasterNet and an Attention Mechanism. Sensors 2025, 25, 4165. [Google Scholar] [CrossRef]
  24. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar] [CrossRef]
  25. Rezatofighi, H.; Tsoi, N.; Gwak, J.Y.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019. [Google Scholar] [CrossRef]
  26. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar] [CrossRef]
  27. He, J.; Erfani, S.; Ma, X.; Bailey, J.; Chi, Y.; Hua, X. Alpha-IoU: A family of power intersection over union losses for bounding box regression. In Proceedings of the Advances in Neural Information Processing Systems, Virtual Conference, 6–14 December 2021. [Google Scholar] [CrossRef]
  28. Hamed, M.; Hossein, E.; Damon, H. A Concise Account for Challenges of Machine Learning in Seakeeping. In Proceedings of the 6th International Conference on Industry 4.0 and Smart Manufacturing, Prague, Czech Republic, 13–15 November 2024. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.