Next Article in Journal
A Generalization Performance Study Using Deep Learning Networks in Embedded Systems
Previous Article in Journal
Steganography and Steganalysis in Voice over IP: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Approaches on Defect Detection in High Resolution Aerial Images of Insulators

Institute of Human Factors and Ergonomics, College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1033; https://doi.org/10.3390/s21041033
Submission received: 31 December 2020 / Revised: 26 January 2021 / Accepted: 1 February 2021 / Published: 3 February 2021
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
By detecting the defect location in high-resolution insulator images collected by unmanned aerial vehicle (UAV) in various environments, the occurrence of power failure can be timely detected and the caused economic loss can be reduced. However, the accuracies of existing detection methods are greatly limited by the complex background interference and small target detection. To solve this problem, two deep learning methods based on Faster R-CNN (faster region-based convolutional neural network) are proposed in this paper, namely Exact R-CNN (exact region-based convolutional neural network) and CME-CNN (cascade the mask extraction and exact region-based convolutional neural network). Firstly, we proposed an Exact R-CNN based on a series of advanced techniques including FPN (feature pyramid network), cascade regression, and GIoU (generalized intersection over union). RoI Align (region of interest align) is introduced to replace RoI pooling (region of interest pooling) to address the misalignment problem, and the depthwise separable convolution and linear bottleneck are introduced to reduce the computational burden. Secondly, a new pipeline is innovatively proposed to improve the performance of insulator defect detection, namely CME-CNN. In our proposed CME-CNN, an insulator mask image is firstly generated to eliminate the complex background by using an encoder-decoder mask extraction network, and then the Exact R-CNN is used to detect the insulator defects. The experimental results show that our proposed method can effectively detect insulator defects, and its accuracy is better than the examined mainstream target detection algorithms.

1. Introduction

Electricity is one of the most essential elements to make the world go around, and the transmission of high-voltage electricity is very important for the practical use of it. In the transmission of high-voltage electricity, electrical insulators are used to support and separate electrical conductors without allowing current through themselves. Usually, insulators are exposed to the harsh environment of strong electric field with all the severe weather conditions such as blazing sunlight, typhoons or hurricanes, thunderstorms, freezing rain, and snowstorms. The harsh environment will make the insulators vulnerable damaged, which will threaten the safety of power grid systems and the use of electricity [1]. Therefore, it is necessary to develop effective methods for insulator defect detection to ensure the safe and reliable electric power transmission [2].
The current defect detection methods can be divided into three categories including physical methods [3,4], traditional vision-based methods [5,6,7,8,9,10,11], deep learning based methods [12,13,14].
The physical methods mainly include the ultrasonic wave method [3] and the ultraviolet pulse method [4] based on manual operation. In [3], the authors proposed an ultrasonic wave method for detection by receiving and processing ultrasonic signals between 20–40 kHz and 80–140 kHz because ultrasound at these frequencies mostly be emitted from the insulator failure location. In [4], the authors introduced an ultraviolet pulse method which identified the defect location by detecting the discharge of defective insulators. However, in practical application, the physical methods are easily affected by solar radiation, meteorological conditions, surrounding heat source, distance and other factors. In recent years, electric power inspection departments have begun to use UAV (unmanned aerial vehicle) instead of manual methods to detect insulator defect.
Among the vision-based insulator defect detection methods, the most frequently used ones are HOG + SVM [10] and Haar + AdaBoost [11]. As traditional detection algorithms, they mainly use sliding window to select the region of interest, extract the features of each window and then classify the feature samples to obtain the detection results. Besides, there are also methods based on contour features and gray similarity matching to classify intact and defective insulators [6]. For example, Zhang and Yang [2] proposed an insulators recognition method with curved smooth and weak textured (CSWT) surfaces. Li et al. [7] used the improved MPEG-7 EHD (edge histogram method) technique to recognize insulators. Wu et al. [5] used the global minimization active contour model (GMAC) for insulator segmentation. However, these methods are usually based accumulated experience to extract image features, so that they are inefficient, suffering from accuracy limitation and time-consuming problems.
With the continuous improvement of computer performance, detection methods based on deep learning frameworks have been widely used. It can effectively compensate the loss of feature information in the process of artificial image feature extraction and improve the efficiency of fault detection. A number of effective target detection algorithms have been proposed, such as Faster R-CNN (faster region-based convolutional neural network) [15], Yolo (you only look once) [16], SSD (single shot multibox detector) [17], DCNN (dynamic convolutional neural network) [18], etc. In reference [14], the authors proposed a method for insulator defect detection based on R-FCN (region-based fully convolutional network). This method is a two-stage object detection method based on RPN (regional recommendation network) and FCN [19] (fully convolutional network) with RPN generating the region proposals and FCN obtaining the feature maps. In reference [20], the proposed insulator defect detection method is based on the improved Faster R-CNN by using ResNet101 as the backbone. However, due to the fact that the high-resolution insulator images taken by UAV contain mostly irrelevant background information, the useful information of the insulator is not much in an image. In addition, the slender shape characteristics of insulators and different defect variations lead to the diverse and complex presentation of defects in images. Hence, the accuracy of the existing insulator detection methods still needs to be improved. The merits and demerits of the above-mentioned three method categories are summarized in Table 1.
Among these three categories, the physical methods are too labor work extensive, which makes it not practical for large-scale applications. The other two categories (i.e., traditional vision-based methods and deep learning based methods) mainly depend on camera systems. Most of the above-mentioned detection methods are based on monocular vision. They have a relatively simple hardware structure and are easily to be operated. However, these methods are disadvantaged by the instability of the available feature point extraction algorithms. Hence, extensive efforts based on multi vision had been conducted in displacement, crack, and deformation detection. In [21], the authors proposed a method for measuring the distance of a given target using deep learning and binocular vision methods, where target detection network methodology and geometric measurement theory were combined to obtain 3D (three-dimensional) target information. In [22], the authors presented a dynamic real-time method to detect surface deformation and full field strain in recycled aggregate concrete-filled steel tubular columns. In [23], the authors proposed a four-ocular vision system for 3D reconstruction of large-scale concrete-filled steel tube under complex testing conditions. Compared with the methods based on monocular vision, these technologies based on multi vision consistently demonstrate high accuracy, stability, and feasibility. However, the multi vision methods are usually based on more expensive equipment than monocular vision methods, and the weight of multi vision systems are usually heavier than monocular vision systems. These disadvantages make the multi vision methods not as feasible as monocular vision methods to be applied in UAV. Therefore, the methodology developed in this study is based on monocular vision.
To alleviate the low accuracy problems of the above-reviewed deep learning-based methods, two methods inspired by Faster R-CNN for insulator defect detection are proposed. A series of advanced techniques are used to improve the network performance including RoI Align (region of interest align), cascade regression, and FPN (feature pyramid network). Besides, a new detection pipeline is proposed to improve the detection accuracy, following the idea that reducing the inference of background information could improve the detection performance. The main contributions of this study are listed as follows.
  •  A new detection network is proposed for the detection of insulator defects, and the method is verified to be effective on a self-collected dataset that has been released to be publicly available online. Our proposed method addresses the misalignment and huge computational burden problems in Faster R-CNN, and advanced techniques including RoI Align, cascade regression and FPN are used to improve the network performance.
  •  A new two-stage defect detection pipeline is proposed. In the first stage, the insulator mask image is generated to reduce the interference of the background information in images by the encoder-decoder mask extraction network. The Exact R-CNN (exact region-based convolutional neural network) is then used to detect insulator defects in the second stage. Compared with the traditional target detection method, this two-stage pipeline can greatly eliminate the interference of background features on the target detection, and greatly improve the accuracy of target detection.

2. Related Work

The deep learning-based methods developed for object detection can be mainly divided into two categories: the one-stage method [16,17,24] and the two-stage method [15,20,25]. The one-stage method directly classifies the object and renders a regression of the object location from the raw image. The two-stage method divides the detection problem into two steps. The region of interest where the object may be located in is firstly extracted from the image, and then the candidate regions are corrected and identified. The one-stage method runs faster, but it has a relatively lower accuracy compared with the two-stage method. The methods developed in this study are based on a typical two-state method, namely Faster R-CNN. The following knowledge from the previous literature is used for the development of our proposed approaches.

2.1. Faster R-CNN

Faster R-CNN is an improvement of Fast R-CNN (fast region-based convolutional neural network) [26] and uses RPN instead of selective searching method to generate candidate regions. In the process of candidate regions generation, multiple bounding boxes are generated for each pixel in the feature map according to the offset of the corresponding anchors in different scales. When the candidate regions are selected, the feature maps are sent into two networks for classification and regression respectively. The classification in RPN is used to distinguish whether the corresponding anchor belongs to the foreground or the background, and the regression in RPN is to predict the offset coordinates between the anchor box and the ground truth. The schematic framework of RPN is shown in Figure 1.
A box can be described by a four-dimensional vector x , y , w , h where x and y are the coordinates of the center point in the box and w and h are the width and height of the box, respectively. Set the given anchor box as A = A x , A y , A w , A h and the ground truth as G = G x , G y , G w , G y . By learning the offsets of the dimensional vector, the predicted box G obtained from A after translation and scaling is closer to G . This process is shown in Figure 2.
After obtaining the candidate regions, the positive and negative samples are set by judging the threshold of intersection over union (IoU). If the IoU is higher than a given threshold, the anchor box is determined as the foreground, otherwise it belongs to the background. These candidate regions generated by RPN, namely RoI (region of interest), are further mapped into the feature maps which are then sent into the detection network. Due to the different size and length of the feature map, uniform size could be achieved through RoI pooling (region of interest pooling) to facilitate the subsequent re-classification and re-regression. The framework of the re-classification and re-regression network is shown in Figure 3.
The loss function of the Faster R-CNN is defined as shown in Formulas (1)–(5).
L p i , t i = 1 L c l s p i , p i * + λ 1 N r e g i p * L r e g t i , t i *
L r e g t i , t i * = i x , y , w , h s m o o t h L 1 t i , t i *
L c l s p i , p i * = i N p i log p i * + 1 p i log 1 p i *
s m o o t h L 1 x = 0.5 x 2           i f x < 1 x 0.5             o t h e r w i s e
p i = s o f t m a x x = e x i j e x i
where L p i , t i is the loss function of Faster R-CNN, L r e g t i , t i * is the loss function of regression, L c l s p i , p i * is the logarithmic loss of both categories (goals and non-goals), p i is the probability of the anchor box belonging to foreground, p i * is the ground truth label of the anchor box, t i is the predicted location coordinate of the anchor box, t i *   is the ground truth location coordinate of the anchor box, N c l s is the size of mini-batch, N r e g is the number of anchor boxes. In practical applications, there is always a big gap between N c l s and N r e g , and the gap would always cause the loss function to be difficult to converge. Therefore, the parameter λ is introduced to balance these two elements. For more details of Faster R-CNN, please refer to [15].

2.2. Cascade R-CNN

The detection models (e.g., SSD, Faster R-CNN) usually suffer the mismatch problem caused by the fixed IoU threshold (e.g., 0.5) to distinguish the positive and negative samples. Normally, a lower IoU threshold would generate more mismatch samples, while in contrast, a higher IoU threshold generates less. Therefore, in order to solve this mismatch problem, Cai et al. [27] proposed a multi-stage framework, namely cascade R-CNN. The network framework has multiple heads, and the threshold value of each network increases step by step. This network framework makes the output of each stage adapt to the threshold value of the next stage well, and it can reduce the probability of false positive samples and improve the effect of target detection. The framework of cascade R-CNN is shown in Figure 4.

2.3. Depthwise Separable Convolution

Depthwise separable convolution is a combination of depthwise and pointwise convolution. Its framework is shown in Figure 5. The depthwise convolution applies a filter to each input channel separately, while the pointwise convolution applies a 1 × 1 convolution to consolidate the output of depthwise convolution. In general, the computational cost of depthwise separable convolution is 8 to 9 times less than the standard convolution with decreased accuracy [28]. To alleviate this problem, a new architecture based on deeply separable convolution, namely linear bottlenecks and inverted residuals, was proposed.

2.4. Linear Bottleneck and Inverted Residuals

Chollet et al. [29] found that the performance of depthwise convolution after nonlinear activation function was worse than that after linear activation function. In another study [30], the researchers proved that in the low dimensional manifold of interest, the output would lose information through Relu function. The lower the dimension of the manifold of interest, the more information loss. To alleviate the problem mentioned above, linear bottleneck and inverted residual were proposed [30].
Its framework is shown in Figure 6, and it is implemented as follows: Conv1 × 1 (expansion) → Dwise3 × 3 (reduced computing costs) → Conv1 × 1 (projection). The purpose of the expansion layer is to reduce the lack of information because the use of nonlinear functions is inevitable. Linear activation function is used to minimize the information loss after dimension projection. Shortcuts are used directly between bottlenecks, which is different from ResNet [31].

2.5. U-Net

U-Net [32] is improved based on FCN. The encoder-decoder network framework of U-Net is shown in Figure 7. The encoder uses 3 × 3 convolution and pooling down sampling to extract image features and to reduce the spatial dimension. The decoder uses 3 × 3 convolution and pooling up sampling to repair the details and spatial dimensions of the image. In addition, the U-Net can combine the high-level features with the low-level features, so that more accurate segmentation results can be obtained.

3. Proposed Approaches

In this section, two models based on deep learning are proposed for insulator defect detection in high-resolution aerial images. The first model, namely Exact R-CNN, is inspired by the Faster R-CNN [15], but the network framework, the setting of super parameters, and the loss function are different from the Faster R-CNN. For simplicity, ERCN is used to represent Exact R-CNN. The second model, namely CME-CNN (cascade the mask extraction and exact region-based convolutional neural network), detects the defect of the insulator on the mask image of the information concentration through cascading an encoder-decoder mask extraction network and ERCN to improve the accuracy of the network. The depthwise separable convolution and linear bottlenecks are also used in CME-CNN to reduce the computational cost. Table 2 shows the comparison between the proposed methods and the traditional R-CNN methods. The source codes of our proposed approaches are available online [33].

3.1. Exact R-CNN

In the original article of Faster R-CNN, the backbone was inspired by Vgg16 [34]. Differently, Resnet50 is used in ERCN to replace the Vgg16. Compared with Vgg16, the deep framework of ResNet can not only extract more features, but also avoid the problem of gradient disappearing. This helps to promote convergence and model training. Considering that the features of insulator defect (small target) are easy to be ignored, the FPN framework [35] is adopted to extract both the semantic information of high-level features and the location information of the low-level features. Inspired by Mask R-CNN [36], the original RoI pooling can be changed to RoI Align to solve the misalignment problem caused by the twice quantization in the pooling operation and to improve the detection accuracy of small objects [37]. Figure 8 shows how RoI Align works. It uses bilinear interpolation to get the eigenvalues of floating-point coordinates. The improved backbone used in our method is shown in Figure 9, where C1~C5 are backbone ResNet50, and P1~P6 are the levels of feature maps after feature fusion.
To solve the mismatch problem, cascade R-CNN is used to increase the IoU threshold level step by step, so that the proposals of each stage can better adapt to the high threshold of the next level, reducing the probability of false positive samples and improving the accuracy of target detection. Its improved framework is shown in Figure 10. Compared with Figure 4, the improved framework in Figure 10 replaces the RoI pooling with RoI Align.
The original Faster R-CNN used 9 types of anchor boxes with ratios of 0.5, 1, 2, and sizes of 128, 256, 512. However, due to the different datasets used, these ratios and sizes of anchor boxes are not suitable for slender insulator defects, the K-means++ [38] algorithm is used to cluster the bounding boxes of the insulator defects in our dataset to generate 9 groups of anchor boxes of different ratios and sizes that are more suitable for the characteristics of the insulator defects. Then the network performs fine-tuning on the basis of these anchor boxes to generate more accurate bounding boxes. When distinguishing between positive samples and negative samples, the IoU filter is changed to GIoU [39] filter because GIoU takes into account that the two bounding boxes do not overlap. This can reflect the overlap mode, so that the extracted area is more consistent with the target area.
The GIoU diagram is shown in Figure 11. For the red box A and the green box B, GIoU is calculated as defined in Formulas (6) and (7).
I o U = A B A B
G I o U = I o U C A B C
The Smooth L1 is used as the loss function for the bounding box regression. It should be noted that the offset between the prediction and the target is regressed, hence the network output needs to be decoded to get the final frame coordinates. The decoding process is shown in Algorithm 1. In terms of the classification loss, we use the cross entropy function of the binary classes (i.e., insulator and background).
Considering that it is difficult to identify insulator defects on sample images with complex background environment, a cascaded network based on U-Net for improvement is proposed.

3.2. CME-CNN

In order to achieve more accurate detection of insulator defects, CME-CNN is to segment the insulators in the images to obtain the detailed information of the insulators without background information, and then use ERCN to detect the defect position. U-Net is used as the encoder-decoder mask extraction network to segment the insulators. The selection and arrangement of successive convolutional layers in U-Net are shown in Figure 12, which is consistent with [32]. Specifically, the encoder uses a 3 × 3 convolution and performs four down samplings. The decoder uses a 3 × 3 convolution and performs four up-sampling, splicing with low-level features in the up-sampling process. In this study, we use the method of padding zeros to make the size of feature map unchanged before and after convolution. Finally, a two-channel output including the background and target classification results is obtained. However, the complicated background of the insulator images and the various shapes and colors of the insulators make the segmentation task difficult. In order to further improve the encoder-decoder mask extraction network (U-Net) performance, the following improvements are proposed.
In general, deepening the network can make it extract more features and enhance the performance of the network. However, if the network is too deep, it will bring a series of problems such as over fitting, gradient disappearing, and gradient explosion. Therefore, as shown in Figure 13, three residual blocks are added between the encoding and decoding process of the U-Net. Because the deep residual network is easier to optimize than the ordinary deep neural network, it can effectively avoid the occurrence of gradient disappearance or gradient explosion caused by the network layers being too deep [31]. By introducing residual blocks, a deeper network can be trained.
Algorithm 1 Decode the output of the detection model
Input:
T h e   o u t p u t   o f   d e c t e c t i o n   m o d e l ,
T e n s o r = t 1 , , t 200 , t i R 6 ;
Output:
    A l l   b o u n d i n g   b o x e s   B = b 1 , , b 200 , b i R 4 ;
    A n d   t h e   c o r r e s p o n d i n g   s c o r e   S = s i , , s 200 , s i R 1 ;
1:  function Decode output T e n s o r
2:  for i 0 , 200 do
3:     x a , y a , h a , w a   a r e   p r i o r i   b o x e s   a t t r */
4:     x t , y t , w t , h t , b s c o r e , p s c o r e t i ;
5:     x c p   = x t × w a i + x a i ;
6:     y c p = y t × i + x a i ;
7:     h p = h a i × e h t ;
8:     w p = w a i × e w t ;
9:     x m i n = x c p w p / 2 ;  
10:     x m a x = x c p + w p / 2 ;
11:     y m i n = y c p h p / 2 ;
12:     y m a x = y c p + h p / 2 ;
13:     b i = x m i n , y m i n , x m a x , y m a x :
14:     s i = p s c o r e
15:    end for
16:  return B , S
17:  end function
Although the high-level network contains more semantic information [35], the receptive field is larger, and the loss of image details is more severe because of the extensive pooling operations. Differently, the low-level network contains more edge, texture and other details, but it lacks semantic information and cannot classify accurately. The U-Net method simply spliced high-level features and low-level features together and cannot give different attention to low-level features and high-level features [32], which was not conducive to the accurate classification of pixels by neural networks. These problems will cause inaccurate detection, indicating that it needed to pay different attention to these characteristics. Therefore, the squeeze and excitation (SE) block is used to assign different weights to each channel when the underlying features and advanced features are stitched together [40,41]. The squeeze and excitation block is divided into three parts including squeeze, explanation, and attention. Please refer to Figure 14 for the detailed framework of this block.
The squeeze part means that all the eigenvalues in each channel are summed up and averaged, namely global average pooling, which is defined in Formula (8) as,
z c   =   F s q u c 1 H * W i = 1 H j = 1 W u c i , j
where u c is a channel of the feature map, H is the height of the feature map, and W is the width of the feature map.
As defined in Formula (9), the excitation part is to generate the attention weights from zero to one between channels.
s = F e x z , W = σ g z , W = σ W 2 δ W 1 z
where δ means the Relu activation function, σ is the Sigmoid activation function, W 1 and W 2 are the weights that need to be learned. By learning these two weights, the one-dimensional activation weights can be obtained, and different channels of the feature maps are multiplied by the weights to enhance the focus on key channels.
Since the area occupied by an insulator is much less than the background, the imbalance of positive and negative samples during training would make the network difficult to learn the effective features of insulator defects. Therefore, the loss function of U-Net adopts focal loss [42] to reduce the weight of a large number of negative samples while training. The focal loss function is defined as,
L f l = a 1 γ γ log γ ,                                           γ = 1 1 a γ γ log 1 γ ,                   γ = 0
where α is used to balance the uneven proportion of positive and negative samples, γ adjusts the rate of weight reduction of simple samples. In reference [42], it performs well in target detection tasks when α = 0.25 and γ   = 2. Following its parameter setting,   α =0.25 and γ   = 2 are set in the experiment.
To decrease the computational cost caused by the cascading networks, depthwise separable convolution, linear bottlenecks and inverted residuals are adopted to reduce the number of model parameters while maintaining the accuracy [43]. By implementing all the above-introduced improvements and techniques, the general flowchart of our proposed detection framework is shown in Figure 15, and the corresponding pseudocodes are presented in Algorithms 2 and 3.
Algorithm 2 The pseudocode of our proposed CME-CNN
Input:
    D e c t i o n   m o d e l   i n d e x   u   a n d   e ;
N M S   t h r e s h o l d   N t ,   S c o r e   t h r e s h o l d   s ;  
Output:
    T h e   d e t e c t i o n   r e s u l t   i m g B u f f e r ;  
1: /* B u l i d   m o d e l   a n d   l o a d   m o d e l   p a r a m e t e r */
2: s e g a m e n t a t i o n   M o d e l   =   n e t s   F a c t o r y . g e t M o d e l u ;  
3. d e t e c t i o n   M o d e l   =   n e t s   F a c t o r y . g e t M o d e l e ;
4: i m g B u f f e r     ;
5: /* D e t e c t i o n   F r a m e w o r k   S t a r t */
6: while d e t e c t i o n   M o d e l   O n   do
7:    /* G e t   I m g   F r a m e * /
8:     i m g = c a m e r a . g e t I m g F r a m e ( ) ;
9:     / * P r e p r o c e s s   h i g h   r e s o l u t i o n   i m a g e * /
10:     i m g B l o c k s = P r e p r o c e s s i m g                 i m g B l o c k R 4 × 256 × 256 × 3
11.     / * M o r p h o l o g i c a l   f i l t e r   p r o c e s s i n g * /
12:     r e f i n e B l o c k s   =   F i l t e r m a s k B l o c k s ,   r e f i n e B l o c k s R 4 × 256 × 256 × 3
13:     / * G e t   t h e   i m g   w i t h o u t   b a c k g r o u n d   * /
14:     m a s k I m g   =   C o m b i n e r e f i n e B l o c k s . * i m g B l o c k s ,   m a s k I m g R 1024 × 1024 × 3
15.     o u t p u t = d e t e c t i o n   M o d e l . d e t e c t m a s k I m g ,
         o u t p u t R 200 × 6
16.    /* D e c o d e   t h e   o u t p u t */
17.     B o x e s , S c o r e = D e c o d e     o u t p u t o u t p u t
18.    /* N o n   M a x i m u m   S u p p r e s s i o n */
19.     B o x e s , S c o r e = N M S B o x e s , S c o r e , N t
20.    for e a c h   b o x , s c o r e B o x e s , S c o r e do
21.        /* D r a w   t h e   B o u n d i n g   B o x e s   i n */
22:        if s c o r e s then
23:          r e s u l t I m g = R e c t a n g l e r e s h a p e I m g , b o x
24:        /*Append the result in img buffer*/
25:            I m g b u f f e r i m g B u f f e r r e s u l t I m g
26:        end if
27:    end for
28:    if d e t e c t i o n   M o d e   o f f then
29:        break;
30:    end if
31: end while
Algorithm 3 Non-Maximum Suppression
Input:
    Bounding boxes B = { b 1 , , b 200 }   ,   b i R 4 ; N t
     S c o r e s = s 1 , , s 200 , s i R 1 ;
Output:
    Bounding boxes D , s c o r e   S ;
1: function N M S B , S , N t
2: D { }
3: while   B   #   e m p t y   do
4:     m a r g m a x S
5:     M b m :
6:     D D M :
7:     B B M :
8:    for b i   i n   B   do
9:        if IoU M , b i > N t then
10:             B B b i :
11:             S S S i :
12:        end if
13:    end for
14: end while
15: return D , S
16: end function
As illustrated in Figure 15, high-resolution images are firstly cut into several low-resolution images, and then semantic segmentation algorithms (e.g., U-Net) are used to segment the low-resolution images. Next, the segmentation results are merged to get the insulator image after removing the background. Finally, target detection algorithms (e.g., ERCN) are used to detect the defect of insulators. The detailed process to remove the background is shown in Figure 16. As illustrated in the extracted mask in Figure 16, a lot of noise can be observed, thus the morphological filtering is used to solve this problem. Firstly, an open operation is conducted to remove small black dots in the white area, following by a close operation to remove small white dots in the black area. Then, the connected domain analysis is carried out to remove the connected domains with too small area to obtain a more accurate insulator mask image. Finally, the insulator image after background removal is obtained by multiplying the extracted mask image with the original image.

4. Experiment

To examine the effectiveness of our proposed methods, 879 high-resolution insulator images taken by UAV were collected from the Internet, of which 660 were used as the training set and the other 219 were used as the test set. The images in the data set were high-resolution images, with a resolution ranging from 3936 × 2624 to 7360 × 4912. Figure 17 shows two examples of health insulators and insulators with defects. All the images were processed on intel-i7 6700 K (4.0 GHz) with GTX 1080 and were compressed to 1024 × 1024. However, if the image size was adjusted directly, the aspect ratio of the original image would be damaged, resulting in the loss of texture, edge and other information in the image. To avoid this problem, the image was filled with black edges of length L S / 2   (long side L, short side S) and used the bilinear interpolation method to reduce the images to 1024 × 1024. In this case, the aspect ratio of the original image could remain unchanged during the scaling process. For the encoder-decoder mask extraction network, four 256 × 256 image blocks containing all the image information are extracted from the 1024 × 1024 sample images.
However, the number of images in our dataset was obviously not enough for the training of a reliable deep learning-based model. Therefore, image augmentation technologies were used to expand the dataset, including flipping, random rotation, random scaling and reverse. Figure 18 shows an example of these extended images. Finally, 6600 training set images and 2,190 test set images were obtained. In fact, our dataset could get enough images after augmentation, which largely enhanced the generalization ability of our models. The augmented dataset used in this study is publicly available online [33]. The following strategies were also used for network training:
(1)
 Warmup strategy: In the beginning of training, the learning rate was not set as the initial value directly but increased it from a small value to the setting value by a constant amount at each iteration. This ramp avoids a sudden increase of the learning rate, allowing healthy convergence at the start of training. After the warmup, we went back to the original learning rate schedule [44].
(2)
 Hard negative mining strategy: In the training of RPN, the negative samples that were easy to be classified (GioU < 0.3) were not included in the loss because they could not contribute much to the loss. Only the positive samples and negative samples with 0.3 < GioU < 0.5 that were difficult to be distinguished were included.
(3)
 Sample balance strategy: The number of collected negative samples was far more than the number of positive samples. Hensman et al. [45] found that the distribution of the training data has a big impact on CNN performance and a balanced training set is optimal. Hence, the proportion of positive samples and negative samples for training was controlled to 1:1 to achieve better training performance.
The criteria defined in the PASCAL VOC 2012 competition was used to evaluate the predictions. When a prediction frames multiple insulator defects, the prediction will be assigned to the ground truth with the largest IoU. When multiple predictions frame the same insulator defect, they will be sorted according to the obtained confidences. All the predictions are compared with the ground truth for IoU calculation, and they will be considered as “match” when all IoU     0.5. Actually, to avoid the same insulator defect being detected by multiple predictions, only the result with the highest confidence was considered as the true positive.
When comparing the performance of different algorithms on our test set, the comparison indexes including average precision (AP), precision (P), recall rate (R) and frames per second (FPS) were used. P is the proportion of true positive predictions to the total number of positive predictions. R is the proportion of true positive predictions in all positive samples. AP is the cumulative mean sum of PR curve areas under different confidence thresholds. To show the performance of different algorithms more comprehensively, two additional metrics were added [46]: all true (AT) and all miss (AM). AT is the proportion of images in which all the insulator defects are detected correctly. AM is the proportion of images in which all the insulator defects are not detected. For the mask extraction network, Dice coefficient was used to evaluate its performance, which is defined as,
D i c e A , B = 2 A B A + B
where A is the ground truth area and B is the area obtained by the mask extraction network.

5. Results and Discussions

To verify the effectiveness of our proposed methods, their performance was compared with five commonly used target detection methods, i.e., HOG + SVM [10], Haar + AdaBoost [11], Faster R-CNN [15], YoloV3 [24] and YoloV4 [47]. Meanwhile, to show the performance of our proposed methods more comprehensively, the results under two different backbones, i.e., ResNet50 and MobileNetV2 were also compared. The results when using images with different resolutions as inputs and when using different network structures are also presented in this section for a deeper understanding on our methods.

5.1. Determination of the Number of Residual Blocks

To determine the number of residual blocks between the encoding and decoding process of the U-Net in Figure 13, an analysis was conducted to examine the performances when using different number of residual blocks. The results shown in Table 3 indicate that the Dice coefficient gradually increases with the number of residual blocks in the rage of 0~3, but the Dice increase speed greatly slows down when the number of residual blocks continuously increase. As presented in Table 3, the Dice coefficients are similar when the number of blocks is higher than 3. Therefore, as introduced in Figure 13, three residual blocks were used.

5.2. Qualitative Detection Results

Figure 19 shows the segmentation results of the mask extraction network mapped to the original image after denoising with morphological filter under the backbones of ResNet and MobileNetV2 respectively. We found that the improved mask extraction network based on U-Net has pretty good performance under different backbones. From our experimental data, the average Dice coefficient of the network is 0.97 on the backbone of ResNet and 0.94 on the backbone of MobileNetv2.
Figure 20 shows the qualitative results of all the methods examined on our dataset, where the yellow bounding box is the ground truth, and the red bounding box is the prediction. From Figure 18, a comparative summary can be concluded as follows.
Compared with the traditional vision methods, the method based on deep learning can detect more accurately and has less omissions and duplications. This is probably due to the fact that traditional vision-based methods usually use artificially designed features to represent objects and only consider some simple information (such as edge, color or texture) while the deep learning-based methods automatically extract both low-level and high-level features to represent the object. Therefore, the traditional vision method cannot effectively identify the insulator defects in situations with complex background.
Compared with Faster R-CNN and Yolov3, our proposed ERCN and CME-CNN have smaller prediction bias, higher positioning accuracy, and almost no omission and duplication. This is probably because our methods are based on multi-level classification and regression. Each cascade stage improves the detection quality by increasing the GIoU threshold, and introduce RoI Align to solve the mismatch problem, making the re-classification and re-regression network receive the correct features without position deviation. It makes the regression and classification effect better. In addition, in order to improve the detection performance of small targets, FPN is introduced which combines semantic features of different levels. These methods continue to strengthen the extraction and processing of image details, making the extracted target features completer and more accurate.
Compared with ERCN and Yolov4, there is almost no deviation between the prediction of CME-CNN and the ground truth. The detection performance of CME-CNN is better. This is probably because the complex background is simplified and the interference in detection is reduced by generating insulator mask image through the encoder-decoder mask extraction network. This makes the network more focused on the learning of target characteristics, making the network more efficient.
From the illustrated qualitative comparison results, we can conclude that the methods proposed in this paper are more accurate than the other methods mentioned above. There is basically no offset between the prediction and the ground truth, and the predictions are basically no omission or duplication. It can be seen that our method can better grasp the details of the image, can grasp more comprehensive and accurate features of the detected target.

5.3. Quantitative Detection Results

Table 4 gives the quantitative results of all the examined methods on our dataset. The results show that our proposed CME-CNN-ResNet50 has the best performance on five of the examined evaluation indexes (i.e., AP, P, R, AT and AM). Although with a not high FPS, it has great application value in practice because it does not require high real-time performance but requires high detection accuracy when using UAV to detect insulator defects. In addition, due to the introduction of depthwise separable convolution, ERCN-MobileNetV2 has the best FPS (i.e., 11.1). However, its detection performance is worse than CME-CNN-ResNet50. The FPS of CME-CNN-ResNet50 is 1.1, indicating that about one image can be processed in each second. This should be satisfactory for insulator defect detection using UAV because the detection using UAV does not have high requirements on the running time.
In general, the methods with higher detection accuracies usually need higher calculation costs [48]. For the insulator defect detection task in this study, the detection can be conducted on an expensive high-performance server in off-line practical applications. When an insulator defect is detected, it will be reported to the maintenance workers to manually replace the defect insulator. In this situation, the detection accuracy is the most important indicator, and the computation time is not as important as the detection accuracy. For example, the FPS of CME-CNN-ResNet50 is 1.1 and the FPS of YoloV4 is 10.6, which indicates that the computation time of CME-CNN-ResNet50 is 0.8 s slower than the number of YoloV4. In the above-mentioned application scenario, this 0.8 s doesn’t make any sense to speed up the maintenance that requires extensive human labor work. Therefore, in this study, we mainly focus on the detection accuracies of the examined methods.

5.4. Ablation Study Results

To examine how much the above-mentioned techniques contributed to the detection performance, an ablation study was conducted to compare the detection performance when using different skills. The results are given in Table 5, which shows that the use of cascade R-CNN increases the AP of the model by 1.9%, while the use of mask extraction increased the AP of the model by 4.2%. In addition, using both cascade R-CNN and mask extraction increases the R of the model by 5.8%.
Figure 21 shows the influence of different techniques on the model performance. We can find that the use of cascade R-CNN can reduce the offset between the prediction and ground truth and makes the model prediction more accurate. The use of mask extraction makes the model more sensitive to the capture of insulator defects, better detection accuracy for small insulators, and the prediction basically has no offset with the ground truth.

5.5. Detection Performance on Images with Different Resolutions

The above experiments clearly show that the methods proposed in this paper can obtain high detection performance on high-resolution images. However, high-resolution images cannot always be available in practical applications because of many factors such as cost limitation, environmental condition, etc. [49]. Therefore, an experiment was conducted to compare the influence of image resolutions on the performance of CME-CNN-ResNet50 which performs the best on all the examined indexes except FPS in Table 4.
The compression method mentioned in Section 4 was used to compress the image of our dataset to 512 × 512 and 256 × 256. Examples of the compressed images are shown in Figure 22. The compressed 512 × 512 images are regarded as medium-resolution images and the compressed 256 × 256 images are regarded as low-resolution images. The comparison results are shown in Table 6. The results show that the lower the resolution of the test images, the higher the FPS of the network. However, the lower resolution of the test images leads to lower values of the examined AP, P, R, AT, and AM. The AP on high-resolution images is 5.1% and 2.2% higher than the numbers on medium- and low-resolution images, respectively. This is mainly due to the information loss on small targets such as insulator defects in lower resolution images.

5.6. Detection Performance When Using Different ResNet Structures

Besides image resolution, the ResNet structure also has an impact on the detection performance. In [50], ResNet-v2 was proposed. Different from ResNet-v1, ResNet-v2 puts the activation function on the residual branch, which makes the computational speed faster in the back propagation and forward propagation. Here in this study, another experiment was conducted to compare the performance of different ResNet structures on the error rate of the model. Table 7 shows the error rates when using ResNet50-v1 or ResNet50-v2. The results show that the error rates of Faster R-CNN, ERCN, CME-CNN in ResNet50-v2 are reduced by 1.7%, 1.9%, 2.1%, respectively. This indicates that our proposed method can be further improved by alternatively using ResNet50-v2.

5.7. Novelty and Application of our Proposed Methods

Different from the traditional one-stage or two-stage detection methods, our proposed methods use multi-level classification and regression as well as information fusion and detect based on the mask image. On the one hand, multi-level classification and regression improves the quality of prediction box by increasing the GIoU threshold. Our methods also use RoI Align for allowing the selected RoI receive a correct feature without position bias, making a better regression and classification results. On the other hand, information fusion uses FPN network to combine different levels of semantic features to enhance the extraction of bottom details. When stitching the underlying and advanced features in U-Net, the SE block is used to assign different weights to each channel, and different attention is given to low-level features and high-level features. Residual block is also adopted to make the network learn deeper features. More importantly, the interference of complex background is eliminated after obtaining the mask image, which makes the detection more targeted and makes the network focus on training target features. Due to the introduction of depthwise separable convolution, our proposed methods even can be used for real-time detection.
Therefore, our model method CME-CNN-ResNet50 can be used for high-precision inspection tasks such as insulator defects. In addition, in the case of different backbones such as MobileNetv2 and ResNet50, our model can have a good trade-off between detection accuracy and running speed and can be applied to detection scenarios with different demands.

6. Conclusions

In this paper, two methods (i.e., ERCN and CME-CNN) for insulator defect detection based on Faster R-CNN are proposed to be applied to high-resolution aerial images. Our proposed methods in two different backbones (i.e., ResNet50 and MobileNetV2) are used to compare with five different detection methods (i.e., HOG+SVM, Haar+AdaBoost, Faster R-CNN, YoloV3, and YoloV4). The results show that the average precision (AP) of our proposed CME-CNN-ResNet50 achieves the best performance (88.7%) which is 52.9%, 58.2%, 6.5%, 10.8%, and 1.6% higher than the number of HOG+SVM, Haar+AdaBoost, Faster R-CNN-ResNet50, YoloV3, and YoloV4, respectively. The ablation study results show that the mask extraction proposed in CME-CNN contributes to a 4.2% increase on AP than without using this skill and using both cascade R-CNN and mask extraction increases the AP by 7.1%. Besides, the results also show that the performance of our proposed method can be further improved by alternatively using ResNet50-v2. However, the proposed methods are based on insulator images captured in daytime, while detecting insulator defects at night is still a challenging task. Therefore, one of our future work will focus on improving the night detection performance by using image enhancement technologies [51]. Moreover, researchers have introduced the concept of Corner Net [52] and Extreme Net [53] into target detection in the recent years, and satisfactory results have been reported. In our future work, we will also try to combine our proposed method with the detection of key points to see whether we can further improve the detection performance of our network.

Author Contributions

Conceptualization, Q.W., Z.L. and R.C.; formal analysis, Y.Y. and G.L.; funding acquisition, G.L. methodology, Q.W., Z.L. and R.C.; project administration, G.L.; resources, G.L.; software, Q.W., Z.L. and R.C.; supervision, G.L.; validation, Q.W., Z.L., R.C. and G.L.; visualization, Q.W., Z.L., R.C. and G.L.; writing—original draft, Q.W., Z.L. and R.C.; writing—review and editing, Y.Y. and G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shenzhen Fundamental Research Fund (grant number: JCYJ20190808142613246) and the Young Elite Scientists Sponsorship Program funded by the China Society of Automotive Engineers.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study is publicly available online [33].

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
R-CNNRegion-based Convolutional Neural Network
FPNFeature pyramid network
GIoUGeneralized Intersection over Union
RoIRegion of Interest
IoUIntersection over Union
ERCNExact R-CNN
RPNRegional recommendation network
FCNFully Convolutional Network
R-FCNRegion-based Fully Convolutional Network
YoloYou only look once
SSDSingle Shot multibox Detector
HOGHistogram of Oriented Gridients
SVMSupport Vector Machine
DCNNDynamic Convolutional Neural Network
RARoI Align
CRCascade R-CNN
MEMask Extraction
CME-CNNCascade the Mask Extraction and exact region-based Convolutional Neural Network

References

  1. Wang, J.; Liang, X.; Gao, Y. Failure analysis of decay-like fracture of composite insulator. IEEE Trans. Dielectr. Electr. Insul. 2014, 21, 2503–2511. [Google Scholar] [CrossRef]
  2. Tan, P.; Li, X.-F.; Xu, J.-M.; Ma, J.-E.; Wang, F.-J.; Ding, J.; Fang, Y.-T.; Ning, Y. Catenary insulator defect detection based on contour features and gray similarity matching. J. Zhejiang Univ. Sci. A 2020, 21, 64–73. [Google Scholar] [CrossRef]
  3. Yao, Z.; Yu, X.; Yao, J.; Sui, W.; Yu, X. Transmission line insulator fault detection based on ultrasonic technology. J. Phys. Conf. Ser. 2019, 1187, 022056. [Google Scholar] [CrossRef]
  4. Gao, Y.; Xu, A.; Ouyang, Q.; Zhang, L.; Wang, Y. Insulator Defect Detection Technology Research. In Proceedings of the 3rd International Conference on Material Mechanical and Manufacturing Engineering (IC3ME 2015), Guangzhou, China, 27–28 June 2015; Volume 1, p. 1. [Google Scholar]
  5. Wu, Q.; An, J.; Lin, B. A texture segmentation algorithm based on pca and global minimization active contour model for aerial insulator images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1509–1518. [Google Scholar]
  6. Zhang, J.; Yang, R. Insulators Recognition for 220kv/330kv High-voltage Live-line Cleaning Robot. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; pp. 630–633. [Google Scholar]
  7. Li, W.; Ye, G.; Huang, F.; Wang, S.; Chang, W. Recognition of Insulator Based on Developed MPEG-7 Texture Feature. In Proceedings of the 2010 3rd International Congress on Image and Signal Processing, Yantai, China, 16–18 October 2010; pp. 265–268. [Google Scholar]
  8. Liao, S.; An, J. A robust insulator detection algorithm based on local features and spatial orders for aerial images. IEEE Geosci. Remote Sens. Lett. 2014, 12, 963–967. [Google Scholar] [CrossRef]
  9. Li, D.; Zhang, M.; Tong, Z.B. Transmission line insulators detection based on KAZE algorithm. AMR 2014, 936, 2184–2189. [Google Scholar] [CrossRef]
  10. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR, San Diego, CA, USA, 20–25 June 2005; Volume 8. [Google Scholar]
  11. Zhang, S.; Bauckhage, C.; Cremers, A.B. Informed Haar-like Features Improve Pedestrian Detection. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 947–954. [Google Scholar]
  12. Huang, X.; Zhang, H.; Zhang, Y. Automatic identification and location technology of glass insulator self-shattering. J. Electron. Imaging 2017, 26, 1. [Google Scholar] [CrossRef] [Green Version]
  13. Zhao, Z.; Zhen, Z.; Zhang, L.; Qi, Y.; Kong, Y.; Zhang, K. Insulator detection method in inspection image based on improved faster R-CNN. Energies 2019, 12, 1204. [Google Scholar] [CrossRef] [Green Version]
  14. Li, S.; Zhou, H.; Wang, G.; Zhu, X.; Kong, L.; Hu, Z. Cracked insulator detection based on R-FCN. J. Phys. Conf. Ser. 2018, 1069, 012147. [Google Scholar] [CrossRef] [Green Version]
  15. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–20 June 2016; pp. 779–788. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S. SSD: Single shot multibox detector. Eur. Conf. Comput. Vision 2015, 12, 8. [Google Scholar]
  18. Kalchbrenner, N.; Grefenstette, E.; Blunsom, P. A convolutional neural network for modelling sentences. arXiv 2014, arXiv:1404.2188. [Google Scholar]
  19. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  20. Liao, G.-P.; Yang, G.-J.; Tong, W.-T.; Gao, W.; Lv, F.-L.; Gao, D. Study on Power Line Insulator Defect Detection via Improved Faster Region-based Convolutional Neural Network. In Proceedings of the 2019 IEEE 7th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 10–20 October 2019; pp. 262–266. [Google Scholar]
  21. Zhang, J.; Hu, S.; Shi, H. Deep learning based object distance measurement method for binocular stereo vision blind area. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 606–613. [Google Scholar] [CrossRef]
  22. Tang, Y.; Li, L.; Wang, C.; Chen, M.; Feng, W.; Zou, X.; Huang, K. Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision. Robot. Comput. Integr. Manuf. 2019, 59, 36–46. [Google Scholar] [CrossRef]
  23. Tang, Y.; Chen, M.; Lin, Y.; Huang, X.; Huang, K.; He, Y.; Li, L. Vision-based three-dimensional reconstruction and moni-toring of large-scale steel tubular structures. Adv. Civ. Eng. 2020, 1236021. [Google Scholar]
  24. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  25. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 CVPR, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  26. Girshick, R. Fast R-CNN. arXiv 2015, arXiv:1504.08083. [Google Scholar]
  27. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  28. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  29. Chollet, F. Xception: Deep learning with depthwise separable convolutions. arXiv 2017, arXiv:1610.02357. [Google Scholar]
  30. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 CVPR, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  33. GitHub Repository. Deep learning Approaches on Defect Detection in High Resolution Aerial Images of Insulators. Available online: https://github.com/TaoTao9/Deep-learning-approaches-on-defect-detection-in-high-resolu-tion-aerial-images-of-insulators (accessed on 1 February 2021).
  34. Qassim, H.; Verma, A.; Feinzimer, D. Compressed Residual-VGG16 CNN Model for Big Data Places Image Recognition. In Proceedings of the 2018 CCWC, Las Vegas, NV, USA, 8–10 January 2018; pp. 169–175. [Google Scholar]
  35. Lin, T.-Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  36. Li, G.; Xie, H.; Yan, W.; Chang, Y.; Qu, X. Detection of road objects with mall appearance in images for autonomous driving in various traffic situations using a deep learning based approach. IEEE Access 2020, 8, 211164–211172. [Google Scholar] [CrossRef]
  37. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. arXiv 2018, arXiv:1703.06870. [Google Scholar]
  38. Arthur, D.; Vassilvitskii, S. K-Means++: The Advantages of Careful Seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2007, New Orleans, LA, USA, 7–9 January 2007; pp. 1027–1035. [Google Scholar]
  39. Hou, Z.; Liu, X.; Chen, L. Object detection algorithm for improving non-maximum suppression using GIoU. IOP Conf. Ser. Mater. Sci. Eng. 2020, 790, 012062. [Google Scholar] [CrossRef]
  40. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  41. Hu, Y.; Wen, G.; Luo, M.; Dai, D.; Ma, J.; Yu, Z. competitive inner-imaging squeeze and excitation for residual network. arXiv 2018, arXiv:1807.08920. [Google Scholar]
  42. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. arXiv 2018, arXiv:1708.02002. [Google Scholar]
  43. Wang, G.; Yuan, G.; Li, T.; Lv, M. An multi-scale learning network with depthwise separable convolutions. IPSJ Trans. Comput. Vis. Appl. 2018, 10, 11. [Google Scholar] [CrossRef] [Green Version]
  44. Goyal, P.; Dollár, P.; Girshick, R.; Noordhuis, P.; Wesolowski, L.; Kyrola, A.; Tulloch, A.; Jia, Y.; He, K. Accurate, large min-ibatch SGD: Training imagenet in 1 hour. arXiv 2018, arXiv:1706.02677. [Google Scholar]
  45. Hensman, P.; Masko, D. The Impact of Imbalanced Training Data for Convolutional Neural Networks. Ph.D. Thesis, KTH Royal Institute of Technology, Stockholm, Sweden, 2015. [Google Scholar]
  46. Li, G.; Yang, Y.; Qu, X. Deep learning approaches on pedestrian detection in hazy weather. IEEE Trans. Ind. Electron. 2020, 67, 8889–8899. [Google Scholar] [CrossRef]
  47. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  48. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point set representation for object detection. IOP Conf. Ser. Mater. Sci. Eng. 2020, 790, 012062. [Google Scholar]
  49. Li, G.; Yang, Y.; Zhang, T.; Qu, X.; Cao, D.; Cheng, B.; Li, K. Risk assessment based collision avoidance decision-making for autonomous vehicles in multi-scenarios. Transp. Res. Part C Emerg. Technol. 2021, 122, 102820. [Google Scholar] [CrossRef]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. arXiv 2016, arXiv:1603.05027. [Google Scholar]
  51. Li, G.; Yang, Y.; Qu, X.; Cao, D.; Li, K. A deep learning based image enhancement approach for autonomous driving at night. Knowl. Based Syst. 2021, 213, 106617. [Google Scholar] [CrossRef]
  52. Law, H.; Deng, J. CornerNet: Detecting objects as paired keypoints. Int. J. Comput. Vis. 2020, 128, 642–656. [Google Scholar] [CrossRef] [Green Version]
  53. Zhou, X.; Zhuo, J.; Krahenbuhl, P. Bottom-up Object Detection by Grouping Extreme and Center Points. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–21 June 2019; pp. 850–859. [Google Scholar]
Figure 1. The schematic framework of RPN (regional recommendation network).
Figure 1. The schematic framework of RPN (regional recommendation network).
Sensors 21 01033 g001
Figure 2. Process of correcting the anchor box.
Figure 2. Process of correcting the anchor box.
Sensors 21 01033 g002
Figure 3. The framework of the re-regression and re-classification network.
Figure 3. The framework of the re-regression and re-classification network.
Sensors 21 01033 g003
Figure 4. Framework of cascade R-CNN.
Figure 4. Framework of cascade R-CNN.
Sensors 21 01033 g004
Figure 5. Framework of Depthwise Separable Convolution.
Figure 5. Framework of Depthwise Separable Convolution.
Sensors 21 01033 g005
Figure 6. Linear bottlenecks and inverted residuals.
Figure 6. Linear bottlenecks and inverted residuals.
Sensors 21 01033 g006
Figure 7. Framework of U-Net.
Figure 7. Framework of U-Net.
Sensors 21 01033 g007
Figure 8. How RoI (region of interest) Align works.
Figure 8. How RoI (region of interest) Align works.
Sensors 21 01033 g008
Figure 9. Framework of the improved backbone.
Figure 9. Framework of the improved backbone.
Sensors 21 01033 g009
Figure 10. The improved framework of cascade R-CNN.
Figure 10. The improved framework of cascade R-CNN.
Sensors 21 01033 g010
Figure 11. The calculation of GIoU.
Figure 11. The calculation of GIoU.
Sensors 21 01033 g011
Figure 12. Selection and arrangement of successive convolutional layers in U-Net. The first number in each box represents the resolution of the feature map, and the second number represents the number of feature map channels.
Figure 12. Selection and arrangement of successive convolutional layers in U-Net. The first number in each box represents the resolution of the feature map, and the second number represents the number of feature map channels.
Sensors 21 01033 g012
Figure 13. The residual blocks between the encoder and decoder.
Figure 13. The residual blocks between the encoder and decoder.
Sensors 21 01033 g013
Figure 14. The framework of the squeeze and excitation block.
Figure 14. The framework of the squeeze and excitation block.
Sensors 21 01033 g014
Figure 15. The general flowchart of our proposed CME-CNN (cascade the mask extraction and exact region-based convolutional neural network).
Figure 15. The general flowchart of our proposed CME-CNN (cascade the mask extraction and exact region-based convolutional neural network).
Sensors 21 01033 g015
Figure 16. The mask schematic diagram.
Figure 16. The mask schematic diagram.
Sensors 21 01033 g016
Figure 17. Examples of health insulators and insulators with defects. (a) Insulator with defects. (b) Health insulator.
Figure 17. Examples of health insulators and insulators with defects. (a) Insulator with defects. (b) Health insulator.
Sensors 21 01033 g017
Figure 18. Examples of augmented images. (a) flipping. (b) random rotation. (c) random scaling. (d) reverse.
Figure 18. Examples of augmented images. (a) flipping. (b) random rotation. (c) random scaling. (d) reverse.
Sensors 21 01033 g018aSensors 21 01033 g018b
Figure 19. Mask examples mapped to the raw image after filtering under different backbones. (a) MobileNetV2. (b) ResNet.
Figure 19. Mask examples mapped to the raw image after filtering under different backbones. (a) MobileNetV2. (b) ResNet.
Sensors 21 01033 g019aSensors 21 01033 g019b
Figure 20. Qualitative results of all the examined methods. (a) HOG+SVM. (b) Haar+Adaboost. (c) YoloV3. (d) YoloV4. (e) Faster R-CNN. (f) ERCN-MobileNetV2. (g) ERCN-ResNet50. (h) CME-CNN-MoblieNetV2. (i) CME-CNN-ResNet50.
Figure 20. Qualitative results of all the examined methods. (a) HOG+SVM. (b) Haar+Adaboost. (c) YoloV3. (d) YoloV4. (e) Faster R-CNN. (f) ERCN-MobileNetV2. (g) ERCN-ResNet50. (h) CME-CNN-MoblieNetV2. (i) CME-CNN-ResNet50.
Sensors 21 01033 g020aSensors 21 01033 g020b
Figure 21. Ablation study results when using different techniques. (a) FPN+RA. (b) FPN+RA+CR. (c) FPN+RA+CR+ME. (FPN: Feature Pyramid Network. RA: RoI Align. CR: Cascade R-CNN. ME: Mask Extraction).
Figure 21. Ablation study results when using different techniques. (a) FPN+RA. (b) FPN+RA+CR. (c) FPN+RA+CR+ME. (FPN: Feature Pyramid Network. RA: RoI Align. CR: Cascade R-CNN. ME: Mask Extraction).
Sensors 21 01033 g021aSensors 21 01033 g021b
Figure 22. Examples of high, medium and low resolution images. (a) High. (b) Medium. (c) Low.
Figure 22. Examples of high, medium and low resolution images. (a) High. (b) Medium. (c) Low.
Sensors 21 01033 g022
Table 1. A brief summary of the previous related methods.
Table 1. A brief summary of the previous related methods.
MethodMeritsDemeritsStudies
physical methodsnot affected by complex visual backgroundlaborious; easy to be affected by meteorological conditions, surrounding heat source, and other factors.[3,4]
traditional vision-based methodshigh efficiency,
low computational burden
easy to be affected by the complex background; depending on accumulated experience to extract image features[5,6,7,8,9,10,11]
deep learning-based methodshigh accuracy and feature extraction abilitydepending on large amounts of data.[12,13,14,15,16,17,18,19,20]
Table 2. Comparison Between Different Methods.
Table 2. Comparison Between Different Methods.
ModelBBBNFPNCRIRSConv method
R-CNNAlexNet××××Std-Conv
Fast R-CNNVgg16××××Std-Conv
Faster R-CNNVgg16×××Std-Conv
ERCNResNet50×Std-Conv
CME-CNNMobileNetV2Dwise-Conv
BB: Backbone. BN: Batch Normalization. IRS: Inverted Residuals. FPN: Feature Pyramid Network. CR: Cascade R-CNN. Dwise-Conv: depthwise separable convolution.
Table 3. The relationship between the number of residual blocks and Dice coefficient.
Table 3. The relationship between the number of residual blocks and Dice coefficient.
Number of Residual Blocks0123456
Dice Coefficient89.490.892.794.995.295.495.5
Table 4. Quantitative results of all the examined methods.
Table 4. Quantitative results of all the examined methods.
MethodAPPRATAMFPS
HOG+SVM35.856.545.133.213.4-
Haar+Adaboost30.549.242.237.530.5-
YoloV377.981.179.778.87.710.2
YoloV487.189.888.988.10.6710.6
Faster R-CNN-ResNet5082.284.482.981.62.23.7
ERCN-ResNet5084.587.786.886.21.73.2
CME-CNN-ResNet5088.791.290.188.60.621.1
ERCN-MobileNetV283.186.985.485.21.911.1
CME-CNN-MoblieNetV286.989.388.687.80.753.6
The bold number indicates the best method for each of the examined indexes.
Table 5. Detection results using different techniques in our network.
Table 5. Detection results using different techniques in our network.
FPNRACRMEPRAP
××85.984.381.6
×87.786.884.5
91.290.188.7
FPN: Feature Pyramid Network; RA: RoI Align; CR: Cascade R-CNN; ME: Mask Extraction. The bold number indicates the best method for each of the examined indexes.
Table 6. Detection results when using images with different resolutions.
Table 6. Detection results when using images with different resolutions.
ResolutionAPPRATAMFPS
High88.791.290.188.60.621.1
Medium86.588.287.185.90.752.5
Low83.686.284.482.31.153.6
Table 7. Detection error rates when using different ResNet structures.
Table 7. Detection error rates when using different ResNet structures.
StructureFaster R-CNNERCNCME-CNN
ResNet50-v115.612.38.8
ResNet50-v213.910.46.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wen, Q.; Luo, Z.; Chen, R.; Yang, Y.; Li, G. Deep Learning Approaches on Defect Detection in High Resolution Aerial Images of Insulators. Sensors 2021, 21, 1033. https://doi.org/10.3390/s21041033

AMA Style

Wen Q, Luo Z, Chen R, Yang Y, Li G. Deep Learning Approaches on Defect Detection in High Resolution Aerial Images of Insulators. Sensors. 2021; 21(4):1033. https://doi.org/10.3390/s21041033

Chicago/Turabian Style

Wen, Qiaodi, Ziqi Luo, Ruitao Chen, Yifan Yang, and Guofa Li. 2021. "Deep Learning Approaches on Defect Detection in High Resolution Aerial Images of Insulators" Sensors 21, no. 4: 1033. https://doi.org/10.3390/s21041033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop