Next Article in Journal
A Multi-User Public Key Encryption with Multi-Keyword Search out of Bilinear Pairings
Next Article in Special Issue
Detection-Based Object Tracking Applied to Remote Ship Inspection
Previous Article in Journal
Dynamic Properties of Microresonators with the Bionic Structure of Tympanic Membrane
Previous Article in Special Issue
A Binarized Segmented ResNet Based on Edge Computing for Re-Identification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ISSD: Improved SSD for Insulator and Spacer Online Detection Based on UAV System

College of Electrical Engineering, Guangxi University, Nanning 530000, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(23), 6961; https://doi.org/10.3390/s20236961
Submission received: 10 October 2020 / Revised: 25 November 2020 / Accepted: 3 December 2020 / Published: 5 December 2020
(This article belongs to the Special Issue Visual Sensors for Object Tracking and Recognition)

Abstract

:
In power inspection tasks, the insulator and spacer are important inspection objects. UAV (unmanned aerial vehicle) power inspection is becoming more and more popular. However, due to the limited computing resources carried by a UAV, a lighter model with small model size, high detection accuracy, and fast detection speed is needed to achieve online detection. In order to realize the online detection of power inspection objects, we propose an improved SSD (single shot multibox detector) insulator and spacer detection algorithm using the power inspection images collected by a UAV. In the proposed algorithm, the lightweight network MnasNet is used as the feature extraction network to generate feature maps. Then, two multiscale feature fusion methods are used to fuse multiple feature maps. Lastly, a power inspection object dataset containing insulators and spacers based on aerial images is built, and the performance of the proposed algorithm is tested on real aerial images and videos. Experimental results show that the proposed algorithm can efficiently detect insulators and spacers. Compared with existing algorithms, the proposed algorithm has the advantages of small model size and fast detection speed. The detection accuracy can achieve 93.8%. The detection time of a single image on TX2 (NVIDIA Jetson TX2) is 154 ms and the capture rate on TX2 is 8.27 fps, which allows realizing online detection.

1. Introduction

The insulator and spacer are important components of power transmission lines. The insulator can achieve electrical insulation and mechanical fixation. The spacer can fix the spacing between split wires, prevent whipping between wires, and suppress breeze vibration. The health status of insulators and spacers has an important impact on the stable and safe operation of high-voltage transmission lines [1]. Thus, it is necessary to periodically inspect the transmission line components, e.g., insulators and spacers. Figure 1 shows some electrical components on the transmission line. Figure 1a is the first type of insulator, named Insulator1. Figure 1b is the second type of insulator, named Insulator2. Figure 1c is the four-split spacer, named Spacer. At present, the common transmission line inspection methods mainly include manual inspection, robot-based inspection, and unmanned aerial vehicle (UAV)-based inspection [2,3]. Manual inspection of electric power exposes the inspection staff to certain dangers and the operation efficiency is not high. Robot-based inspection needs a specific mechanical structure to move on the wires and cross obstacles such as antivibration hammers [4,5,6]. A specific device is required for the robot to go up and down the tower, and the process is complicated, while the robot′s inspection efficiency is low. UAVs are flexible and have high inspection efficiency [7,8], and they are widely used in daily inspections.
At present, UAV power inspection usually uses the camera carried by the UAV to collect images of the transmission lines and their key components. For the inspection, the power components need to be detected in real time. Then, the state of the power component, i.e., normal or damaged, is judged. In the power inspection, accurate and effective inspection object detection is one of the key technologies. In recent years, with the rapid development of deep learning, researchers in the power field have also begun applying deep learning techniques into the detection of power inspection objects. Heng et al. [9] proposed a method for antivibration hammer detection. This method uses faster R-CNN (region-based convolutional neural network) [10] for antivibration hammer recognition and location extraction in aerial images. However, the accuracy of this method is relatively low and the detection speed is slow. To improve the detection accuracy and speed, Chen et al. [11] proposed a novel method to identify insulators and antivibration hammers. In this method, an SR-CNN (super-resolution convolutional neural network) [12] is used to reconstruct a blurred image with high resolution, and then the YOLOv3 (you only look once) [13] network is used to recognize the reconstructed image. For overhead distribution line insulator detection, Prates et al. [14] proposed a deep learning-based method, which uses the VGG [15] network to classify the insulator images. However, this method does not detect the position of insulators. After that, Jiang et al. [16] used MobileNet [17] as the backbone of SSD (single shot multibox detector) [18] to detect missing caps, which represent the most frequent fault of insulators. Here, low-level, middle-level, and high-level feature maps are used for insulator detection, which can improve the detection accuracy. Additionally, Hu et al. [19] used the improved faster R-CNN to detect railway catenary insulators. This method adds an angle of object during training, which makes the detection box closer to the object. However, this model has a larger weight, which can take up more memory space. Han et al. [20] proposed a cascaded model to detect multiple insulator faults in the aerial images. This method used the YOLOv3 tiny network to detect insulators. Most of these methods were tested on high-performance computers without testing on embedded devices. Thus, it is difficult to judge whether the methods meet the real-time requirements of UAVs for data processing.
The existing object detection algorithms for power inspection can be divided into two-stage detection frameworks (e.g., Faster R-CNN [10], R-FCN (region-based fully convolutional networks) [21]) and one-stage detection frameworks (e.g., SSD [18]). The detection speed of the two-stage detection framework is usually slower than that of the one-stage detection framework, and the model weight size is usually greater than the one-stage detection framework. As we know, the embedded device carried by the UAV has limited computing resources and small memory space. Thus, the two-stage detection framework is usually not suitable for deployment on UAVs equipped with embedded devices to achieve online inspection. In order to achieve effective online insulator and spacer detection, we first construct a new insulator and spacer detection dataset, named ISD-dataset. Then, we propose an improved SSD algorithm, named ISSD. The improved algorithm uses MnasNet [22] as the main feature extraction network to generate feature maps. Next, the multiscale feature maps are fused and then the detection results are generated through NMS (non-maximum suppression). Finally, the detection speed is tested on a TX2 (NVIDIA Jetson TX2) and server. The test results show that the proposed algorithm can meet the real-time requirements on a server and can be deployed on TX2 to realize online detection [23,24].
In summary, the main contributions of this paper are as follows:
(1)
A new dataset named ISD-dataset is constructed, which can be used for insulator and spacer detection through aerial images. The ISD-dataset contains three types of objects, i.e., Insulator1, Insulator2, and spacer, which allows effectively testing the performance of the object detection algorithm for UAV power inspection.
(2)
An improved SSD algorithm for UAV online detection of insulator and spacer is proposed. The improved algorithm uses a lightweight network, i.e., MnasNet, as the feature extraction network, which improves the feature extraction capability and reduces the amount of calculation and model size. This network can generate feature maps with rich features more quickly. Then, two multiscale feature fusion methods, i.e., Channel plus and Channel concatenate, are used to fuse low-level features and high-level features to improve the detection accuracy. The proposed algorithm is superior to previous algorithms in terms of accuracy and speed.
(3)
We test the performance of the proposed algorithm, faster R-CNN, R-FCN, and SSD on the ISD-dataset. The comparison of experimental results proves that the proposed algorithm has the advantages of being a small model with high accuracy and fast detection speed. The test results on TX2 show that the proposed algorithm is more suitable for deployment on mobile devices (e.g., TX2). The online detection results of video can be found at https://github.com/PromiseXuan/Insulator-detection.

2. Related Works

2.1. Faster R-CNN

Faster R-CNN [10] is a two-stage framework. This network constructs the region proposals first, and then classifies the selected regions. Faster R-CNN uses a set of basic convolutional networks to extract the feature map of the image. Two branches are formed on the feature map. One branch uses the feature map to generate region proposals through RPNs (region proposal networks). Another branch sends the feature map and region proposal to the ROI (region of interest) pooling layer to generate the proposal feature map. Finally, the object category and the location of the detection box on the proposal feature maps are calculated. Faster R-CNN uses RPNs to generate region proposals, which improves detection speed and accuracy. However, after ROI pooling, each region connects with multiple fully connected layers, which requires a large amount of parameter calculation. This leads to a relatively slow detection speed. Faster R-CNN needs to train RPNs and classification networks. The faster R-CNN model size is large, which takes up more computing resources. Furthermore, the performance of faster R-CNN for small object detection needs to be improved.

2.2. R-FCN

R-FCN [21] is also a two-stage framework. The R-FCN network consists of shared fully convolutional structures and does not contain a fully connected layer. In R-FCN, position-sensitive score maps are proposed to solve the contradiction problem of translation invariance and translation variance in object detection. There are two branches on the feature map obtained by the last convolutional layer of the feature extraction network. A branch is sent to the RPN to generate a region proposal. Another branch is sent to the convolutional layer to generate position-sensitive score maps. Finally, the corresponding category and location information can be obtained through ROI pooling. Compared with faster R-CNN, R-FCN removes the fully connected layer and reduces the amount of calculation, which increases the detection speed. However, because it is still a two-stage framework network, the detection time is also longer.

2.3. SSD

SSD [18] is a one-stage framework for object detection. The basic idea is that an image is input into the network, and then a reasonable output vector is set. Next, the regression approach is used to output the object′s bounding box and category confidence. SSD produces a fixed-size collection of bounding boxes and scores for those boxes, which exist as object class instances. Finally, NMS is used to produce the final result. SSD uses multiscale feature maps for object detection, which improves the detection accuracy of small objects.
As shown in Figure 2, the SSD network consists of four parts. The first part is the image input module. The input image size can be 300 × 300 or 512 × 512. The second part is the feature extraction network. SSD uses VGG16 [15], which is truncated before the classification layer, as the basic model. After that, new convolutional layers are added to obtain more feature maps for object detection. The third part is the detection layer. It uses a 3 × 3 convolution for each feature map to generate the confidence of each category and the offset of the predicted bounding box relative to the default box. The detection result contains two parts, i.e., category confidence and bounding box. The fourth part is NMS. The redundant detection boxes are removed and the detection boxes having the highest intersection over union with ground truth are retained to obtain the final detection result. However, the larger network model is not suitable for deployment on small mobile devices.

3. Proposed Method

The proposed network mainly includes five parts, i.e., input, backbone, neck, prediction, and NMS. Figure 3 shows the proposed network. Input denotes the input image, which normalizes the image size to 300 × 300 × 3. Backbone is a feature extraction network, which is used to generate feature maps to obtain image features. Neck is a multiscale feature fusion module, which combines low-level feature maps and high-level feature maps to generate new feature maps for object prediction. Prediction is the object prediction module, which performs a convolution operation on the feature map obtained by the neck to obtain the predicted bounding box and category confidence. NMS is used to remove redundant detection boxes.

3.1. Backbone

Backbone is a feature extraction network for feature map generation. As shown in Figure 3, the backbone consists of two parts, i.e., MnasNet and extra feature layers.
To run a model on mobile devices with limited computing resources and storage resources, we need to pay attention to three indicators of the model, i.e., model size, inference speed, and detection accuracy. MnasNet is a convolutional neural network model, which is suitable for mobile devices obtained using the neural architecture search approach. It takes the inference time of the model on the mobile device as the main optimization goal to optimize the model, allowing a good balance between accuracy and speed. The main network structure of MnasNet is shown in the upper left part of Figure 3. The objective function of MnasNet is shown in Equations (1) and (2), where n is the target model, Acc(n) represents the accuracy of the target model, Lat(n) represents the actual inference time of the target model on the mobile device, T is the target inference time, and y is a weight factor. When the actual inference time Lat(n) is less than the target inference time T, y = a; otherwise, y = b. When a = b = −0.07, MnasNet can achieve a good balance between accuracy and speed.
max n   A c c ( n ) × [ L a t ( n ) T ] y .
y = { a ,   i f   L a t ( n ) < T b ,   o t h e r w i s e .
In terms of model size, the parameter size of VGG16 is 138 M, while the parameter size of MnasNet is 3.9 M. On the ImageNet [25] benchmark, the accuracy of MnasNet is higher than that of VGG16. MnasNet has few parameters, fast speed, low delay, and high recognition rate, making it more suitable for use on embedded devices. In order to provide MnasNet with better feature extraction capability, transfer learning is employed. The pretrained model parameters are used as the initial parameters of the feature extraction network in the proposed method.
In order to strengthen the feature extraction capability of the network, extra feature layers are added after MnasNet. As shown in the upper right of Figure 3, the extra feature layers constitute three convolutional layers. The convolution calculation is shown in Equations (3) and (4), and the output feature map size is calculated according to Equation (5), where Gi is the feature map of the i-th layer, Wi and Bi are the weight vector and offset vector of the i-th layer convolution kernel, respectively, F() is the activation function, Hi and Oi are the height and width of the feature map, and k, q, and s are the size, padding, and step length of the convolution kernel, respectively. After the feature map undergoes a convolution with a step length of 2, the size of the feature map is reduced to half of the original size, and the receptive field can be expanded, which can provide low/intermediate information supplementation for the standard pretrained network. This contributes to the detection of small and medium objects.
G i = F ( W i G i 1 + B i ) .
F ( x ) = { x , x > 0 0 , otherwise .
H i = ( H i 1 k i + 2 q i s i + 1 ) O i = ( O i 1 k i + 2 q i s i + 1 ) .

3.2. Neck

In order to make the algorithm more adaptable to object detection of different sizes, low-level features and high-level features are usually simultaneously used. Since the characteristics of different objects have certain differences, shallow-level features can distinguish simple objects, and depth-level features are conducive to distinguishing complex objects. Figure 4 shows the modules of multiscale prediction in the SSD network. This structure can make predictions on feature maps of different sizes and improve the performance of the detection algorithm. However, the features extracted by this structure for object prediction are not robust enough. Most of the features are obtained from a shallower layer, and the used features are from the same layer.
In order to make the features used in object prediction more abundant and robust, we draw on the idea of FPN (Feature Pyramid Networks) [26] to propose two feature fusion methods, as shown in Figure 5. Figure 5a is the first method denoted as Channel plus. In this method, the high-level feature map is up-sampled so that the width O1, the height H1, and the number of channels C1 of the high-level feature map are the same size as those of the low-level feature map. Then, the values of the corresponding positions of the two feature maps are directly added to obtain the fused feature map. The calculation method is shown in Equation (6), where zl3 is the feature vector of the feature map l3 after fusion, fl1 is the feature vector of the high-level feature map l1, and fl2 is the feature vector of the low-level feature map l2. Here, the channel numbers of l1 and l2 are equal.
z l 3 = f l 1 + f l 2 .
Figure 5b shows the flowchart of the second method, which is named Channel concatenate. First, the high-level feature map is up-sampled to make the width and height consistent with the low-level feature map, and the number of channels is half that for the low-level feature map. Then the high-level feature map Cl1 can be generated. At the same time, a convolution operation is performed on the low-level feature map with a convolution kernel of 1 × 1, and the number of channels of the low-level feature map is changed to half of the original to obtain the low-level feature map Cl2. The two feature maps are concatenated on the basis of the channels to obtain the fused feature map Cl3. The calculation method is shown in Equation (7), where concatenate(·,·) is the channel concatenate function.
C l 3 = c o n c a t e n a t e ( C l 1 , C l 2 ) .
Through these two methods, high-level features and low-level features are fused to construct a deeper feature pyramid, which can integrate multilevel feature information and perform output prediction on different feature maps.

3.3. Loss Function and Optimization

The loss function of the network consists of two parts, i.e., the position error and the category confidence error. As shown in Equation (8), the total objective loss function is the weighted sum of the position error and the confidence error.
L ( t , p , u , v ) = 1 S ( L c l s ( t , p ) + λ L b o x ( t , u , v ) ) ,
where S is the number of matched default boxes, t = {0,1} (when the i-th default box matches with the j-th ground truth, t is 1; otherwise, t is 0), p is the category confidence, u is the predicted value of the bounding box offsets, v is the ground truth, and λ is a hyperparameter, which is the weight coefficient of position error loss and category error loss. Here, we set λ = 1.
The position error loss function Lbox is shown in Equations (9)–(12), where (o, r, w, h) represent the center point coordinate of the box and the length and width of the box. When the i-th default box and the j-th ground truth box can match, t = 1. We only calculate the position error loss of boxes which present object instances. Positive denotes that the detection box contains an object.
L b o x ( t , u , v ) = i P o s i t i v e S m { o , r , w , h } t i j F ( u i m v ˜ j m ) .
F ( e ) = { 0.5 e 2 i f   | e |   < 1 | e | 0.5 o t h e r w i s e .
v ˜ j o = ( v j o d i o ) / d i w v ˜ j r = ( v j r d i r ) / d i r .
v ˜ j w = log ( v j w / d i w ) v ˜ j h = log ( v j h / d i h ) .
As shown in Equations (13) and (14), Lcls(t,p) is the category confidence loss function. When calculating category confidence errors, both positive and negative samples must be calculated. k represents the k-th object. Negative denotes that there is no object in the detection box
L c l s ( t , p ) = ( i P o s i t i v e S t i j k log ( p ˜ i k ) + i Negative log ( p ˜ i 0 ) ) .
p ˜ i k = exp ( p i k ) k exp ( p i k ) .
The Adam [27] optimization algorithm is used to optimize the network, which can achieve the improvement of optimization quality and speed by obtaining an adaptive learning rate of each parameter during the model training optimization process.

4. Experiments

4.1. Offline Detection versus Online Detection

Most UAV-based transmission line inspection systems use offline inspection schemes. As shown in Figure 6, the UAV captures video with a camera, and then transmits it to a local computer or data processing center. Inspectors inspect the video and judge the status of the power components in the video.
We propose an online detection method. As shown in Figure 7, the UAV captures the image and transmits it to the onboard embedded processor TX2 for inspection. When an object is detected, the UAV adjusts the flight attitude according to the relative position of the object in the image so that object is in the center of the image. This clarifies the image of the detection object.
In offline detection, the UAV is only used to capture video. Offline detection results in a need to store the video during the entire inspection process. Online detection allows only storing images containing the detection object. Online inspection can improve the inspection efficiency and reduce required storage space.

4.2. UAV Platform

In order to realize online detection of high-voltage transmission lines, we built a custom UAV platform (shown in Figure 8). The UAV is based on an F450 quadrotor airframe, equipped with a Pixhawk running ArduPilot firmware as the autopilot and an XBee radio module for GCS (ground control station) telemetry. The UAV is powered by one 5000 mAh, 22.2 V, 25 C, 6 S1P lithium polymer (Li–po) battery pack, which allows it to fly up to 35 min and reach an average speed of 60 km/h.
The primary payload of the UAV is an embedded AI (artificial intelligence) system TX2, which is connected to a ZED Mini camera. The ZED Mini is a stereo camera. It is able to generate red/green/blue (RGB) images and corresponding three-dimensional (3D) point clouds. Since the coordinate origin of point clouds is set to the optical center of the left camera, we only used the left-view RGB image for target detection. In the future, we will combine 3D point clouds for UAV navigation and obstacle avoidance. The ZED Mini camera is rigidly mounted on the front part of UAV to capture forward-looking images.

4.3. Dataset

We kept a distance of 1.5 m to 2.0 m between the UAV and the power tower for shooting. In the insulator string, the diameter of each insulator was 280 mm. The length of the four-split spacer was 500 mm. The resolution of the captured image was 1920 × 1080, and a total of 1500 aerial images containing power components were selected.
To make the ground truth of the object detection dataset, we used Labelimg (https://github.com/tzutalin/labelImg) to label the insulators and spacers in the aerial images of various types of power poles and towers. Then, we built a Pascal VOC (Visual Object Classes) [28] format dataset named the ISD-dataset. In the ISD-dataset, insulator and spacer objects are included. According to the installation method of the insulator, the insulators are classified as Insulator1 or Insulator2. The number of various objects contained in the ISD-dataset is shown in Table 1. We randomly selected 70% of the images in the dataset as the training set and 30% of the images as the test set. In the training stage, random cropping, random left and right mirroring, etc. were used to enhance the data.

4.4. Experimental Setup

The experimental environment and configuration were as follows:
Server configuration: central processing unit (CPU) model, Intel(R) Xeon(R) Gold 6136; graphics processing unit (GPU) model, 8× TITAN V. CUDA version 10.2; operating system, Ubuntu 16.04; Pytorch [29] version 1.3.0.
Embedded device: GPU of TX2, NVIDIA Pascal™ architecture, equipped with 256 NVIDIA CUDA Cores; CUDA version 10.0; operating system, Ubuntu 18.04; Pytorch version 1.4.0.
In the experiment, the initial learning rate was set to 1 × 10−3. The learning rate was attenuated once every 20 k iterations, and the attenuation coefficient was 5 × 10−4. The batch size was 32, and the total number of iterations was 60 k.

4.5. Experimental Results and Analysis

In order to verify the effectiveness of the proposed algorithm, we selected faster R-CNN, R-FCN, SSDLite [30], and different backbones of SSD models for comparison. The backbones included VGG16, ResNet101 [31], and MobileNet_V2 [30]. The specific comparison models are shown in Table 2.

4.5.1. Qualitative Evaluation

In order to qualitatively analyze the performance of different detectors, we show the object detection results of the different algorithms using the ISD-dataset in Figure 9, Figure 10 and Figure 11. Figure 9, Figure 10 and Figure 11 show the visualization results of different algorithms for object detection in aerial images. The detected objects are marked with boxes. The images without box markings indicate that the algorithm did not detect the object. The last subfigures of Figure 9, Figure 10 and Figure 11 represent the ground truth.
In Figure 9, for Insulator1 object detection, the nine algorithms performed well. All methods could detect the Insulator1 object, and no missed detection occurred. However, R-FCN showed overlapping detection boxes after NMS. In Figure 10, for Insulator2 object detection, SSD_vgg, Ours2 and Faster R-CNN could successfully detect the Insulator2 object. R-FCN still showed overlapping detection boxes and the other algorithms failed to detect the Insulator2 object. In Figure 11, for Spacer object detection, SSD_mobilenet, SSDLite_mobile, and R-FCN did not detect the Spacer object. The other algorithms all detected the Spacer object. From the comparison results in Figure 9, Figure 10 and Figure 11, we can see that the proposed algorithm in this paper could effectively detect Insulator1, Insulator2, and Spacer, and the detection performance was closer to ground truth. Therefore, the proposed algorithm is better than the comparison algorithms.

4.5.2. Quantitative Evaluation

In order to quantitatively analyze the performance of compared algorithms, we use the mAP (mean Average Precision) [28] metric to evaluate the performance of the nine algorithms, and we analyzed the detection effect of different types of objects through the AP (Average Precision) metric. In addition, we counted the weight size of the models. In the detection speed experiment, we detected 300 images on the server and TX2, and we counted the detection time and total time for each image. We used the average time as the detection time for a single image, and we observed the stability of the algorithm by counting the variance of the detection time. Table 3 shows the detection accuracy of different algorithms. Table 4 shows the statistics of the detection speed of different algorithms, as well as the average detection time of a single image on the server and TX2 for each algorithm. In order to more directly compare the detection performance of different algorithms, we also drew the PR (Precision Recall) curves of each algorithm with an IoU (intersection over union) of 0.5. The specific detection results are shown in Figure 12.
The calculation methods of precision and recall are shown in Equation (15), where TP (true positive) means that the value of actual class is positive and the value of predicted class is also positive, FN (false negative) means that the value of actual class is positive and predicted class is negative, and FP (false positive) means that the value of actual class is negative and predicted class is positive. The PR curve reflects the relationship between precision and recall under different category confidence thresholds. In the PR figure, a larger area enclosed by the curve and the coordinate axis denotes better performance of the algorithm.
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N .
From the comparison analysis of Table 3 and Table 4 and Figure 12, the observations below can be made.
The accuracy of faster R-CNN is 93.4% and that of R-FCN is 88.1%, whereas that of Ours2 is 0.4% higher than that of faster R-CNN and 5.1% higher than that of R-FCN. The detection time test on TX2 showed that faster R-CNN had the longest detection time of 1036 ms, and R-FCN had a detection time of 576 ms. Faster R-CNN and R-FCN cannot meet real-time requirements, and they are not suitable for deployment on a UAV.
As shown in Figure 12, our proposed algorithm’s surrounding area of the PR curve was similar to that of SSD_vgg and SSD_resnet. The mAP values of SSD_vgg, SSD_resnet, Ours1, and Ours2 were very close, whereby that of Ours2 was 0.4% higher than that of SSD_vgg and 0.1% higher than that of SSD_resnet. However, the detection time on TX2 of Ours2 was 57 ms less than that of SSD_vgg and 46 ms less than that of SSD_resnet. In terms of model size, Ours2 was 167.44 MB smaller than SSD_resnet and 47.89 MB smaller than SSD_vgg. In terms of detection time and model size, Ours2 was better than SSD_vgg and SSD_resnet. In particular, the detection time of Ours2 was the shortest on both server (15 ms) and TX2 (154 ms).
In the lightweight model, the accuracy of Ours2 was 3.5% higher than that of SSD_mobilenet, 2.0% higher than that of SSDLite_mobile, and 1.1% higher than that of SSDLite_mnas. The model sizes of SSD_mobilenet, SSDLite_mobile, and SSDLite_mnas were smaller than that of ours2, but their detection time was longer than that of ours2. Although SSDLite_mobile and SSDLite_mnas use depthwise separable convolution to reduce the number of parameters, this did not increase their detection speed. The detection time of Ours2 on TX2 was 173 ms slower than SSD_mobilenet, 166 ms slower than SSDLite_mobile, and 84 ms slower than SSDLite_mnas. Ours2 could effectively improve the accuracy and detection speed with less of an increase in model size.
The biggest difference between Ours1 and Ours2 was the feature fusion method, and the method of channel concatenate could reduce the amount of calculation. The detection time on TX2 of Ours2 was 35 ms slower than Ours1, which proves that the fusion method of Channel concatenate is better than the method of Channel plus. We can see that the variance of detection time on TX2 of Ours2 was smaller compared to other algorithms, except for faster R-CNN, showing that the detection speed of Ours2 was relatively stable.
Considering the accuracy, speed, and model size, Ours2 could achieve the best performance. Ours2 is suitable for deployment on UAVs for power inspection.
In order to test the real-time performance of Ours1 and Ours2, we employed them for video detection. The resolution of the video was 1920 × 1080. The calculation method of fps is shown in Equation (16), where Framenum is the total number of frames and elapsedTime is the elapsed time. A higher fps denotes a smoother video. We counted the fps of Ours1 and Ours2 on the server and TX2 as shown in Table 5. In video detection, Ours2 could achieve 8.27 FPS on TX2.
F P S = f r a m e n u m e l a p s e d T i m e .
Figure 13 is a diagram of the detection accuracy and speed of our algorithm and the compared algorithms on the server and TX2. It can be seen that the two proposed algorithms are in a leading position in detection accuracy and speed, showing their superior performance.

5. Conclusions

This paper presented an insulator and spacer detection method based on improved SSD for UAV transmission lines. Firstly, feature extraction was performed through the MnasNet network, which can construct a lightweight network with smaller parameters and better performance for object detection. Then, a multiscale feature fusion method was used to fuse low-level features and high-level features to improve detection accuracy. The proposed object detection network was a lighter model. Finally, images of various scenes collected by a UAV were used to construct the insulator and spacer detection dataset, and experiments were conducted on the dataset to realize the identification and positioning of insulators and spacers. The experimental results verified the practicability and robustness of the improved SSD model in inspection image detection. Due to the small number of model training parameters and low model weight size, it is more suitable for deployment on embedded devices with limited computing resources. When the UAV is equipped with an embedded device, it can realize online identification without sending it back to the data terminal for detection, which can improve the inspection efficiency. Our UAV system has good scalability and can be used to detect more types of electrical components in the future.
In the future, we will collect more data to build a power inspection dataset containing more power components such as antivibration hammers, nests, connectors, and strain clamps. Furthermore, the current environment configured on the embedded device is the same as a high-performance computer environment. In the future, TensorRT will be used to optimize and accelerate the model, which can improve the detection speed and realize real-time detection.

Author Contributions

Conceptualization, X.L., Y.L., and F.S.; data curation, X.L., X.Z., and X.C.; formal analysis, X.L., Y.L., and F.G.; funding acquisition, F.S. and F.G.; investigation, X.L., X.Z., and X.C.; methodology, X.L.; project administration, F.S.; software, X.L.; supervision, Y.L. and F.S.; writing—original draft, X.L. and Y.L.; writing—review and editing, X.L. and Y.L. All authors read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61720106009 and 61773359.

Acknowledgments

The authors would like to thank Yanglian Huang, Zhiteng Li, Zhongjie Huang, and Yimin Luo for helping to build the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, S.; Hao, R.; Lee, J. Inspection of insulators on high-voltage power transmission lines. IEEE Trans. Power Deliv. 2009, 24, 2319–2327. [Google Scholar] [CrossRef]
  2. Katrasnik, J.; Pernus, F.; Likar, B. A survey of mobile robots for distribution power line inspection. IEEE Trans. Power Deliv. 2010, 25, 485–493. [Google Scholar] [CrossRef]
  3. Li, D.; Wang, X. The future application of transmission line automatic monitoring and deep learning technology based on vision. In Proceedings of the 2019 IEEE 4th International Conference on Cloud Computing and Big Data Analysis (ICCCBDA), Chengdu, China, 12–15 April 2019; pp. 131–137. [Google Scholar]
  4. Gao, Y.; Song, G.; Li, S.; Zhen, F.; Chen, D.; Song, A. LineSpyX: A power line inspection robot based on digital radiography. IEEE Robot. Autom. Lett. 2020, 5, 4759–4765. [Google Scholar] [CrossRef]
  5. Zhou, Q.; Zhou, X.-l.; Li, X.-p.; Xiao, J.; Zhou, T.; Wang, C.-j. Mechanical design and research of a novel power lines inspection robot. In Proceedings of the 2016 International Conference on Integrated Circuits and Microsystems (ICICM), Chengdu, China, 23–25 November 2016; pp. 363–366. [Google Scholar]
  6. Yue, X.; Wang, H.; Yang, Y.; Jiang, Y. Geometric design of an inspection robot for 110kV power transmission lines. In Proceedings of the 2016 4th International Conference on Applied Robotics for the Power Industry (CARPI), Jinan, China, 11–13 October 2016; pp. 1–5. [Google Scholar]
  7. Mao, T.; Huang, K.; Zeng, X.; Ren, L.; Wang, C.; Li, S.; Zhang, M.; Chen, Y. Development of power transmission line defects diagnosis system for UAV inspection based on binocular depth imaging technology. In Proceedings of the 2019 2nd International Conference on Electrical Materials and Power Equipment (ICEMPE), Guangzhou, China, 7–10 April 2019; pp. 478–481. [Google Scholar]
  8. Nguyen, V.N.; Jenssen, R.; Roverso, D. Automatic autonomous vision-based power line inspection: A review of current status and the potential role of deep learning. Int. J. Electr. Power Energy Syst. 2018, 99, 107–120. [Google Scholar] [CrossRef] [Green Version]
  9. Heng, Y.; Tao, G.; Ping, S.; Fengxiang, C.; Wei, W.; Xiaowei, L. Anti-vibration hammer detection in UAV image. In Proceedings of the 2017 2nd International Conference on Power and Renewable Energy (ICPRE), Chengdu, China, 20–23 September 2017; pp. 204–207. [Google Scholar]
  10. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. arXiv 2015, arXiv:1506.01497. [Google Scholar] [CrossRef] [Green Version]
  11. Chen, H.; He, Z.; Shi, B.; Zhong, T. Research on recognition method of electrical components based on YOLO V3. IEEE Access 2019, 7, 157818–157829. [Google Scholar] [CrossRef]
  12. Ward, C.M.; Harguess, J.; Crabb, B.; Parameswaran, S. Image quality assessment for determining efficacy and limitations of Super-Resolution Convolutional Neural Network (SRCNN). arXiv 2019, arXiv:1905.05373. [Google Scholar]
  13. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  14. Prates, R.M.; Cruz, R.; Marotta, A.P.; Ramos, R.P.; Filho, E.F.S.; Cardoso, J.S. Insulator visual non-conformity detection in overhead power distribution lines using deep learning. Comput. Electr Eng. 2019, 78, 343–355. [Google Scholar] [CrossRef]
  15. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  16. Jiang, H.; Qiu, X.; Chen, J.; Liu, X.; Miao, X.; Zhuang, S. Insulator fault detection in aerial images based on ensemble learning with multi-level perception. IEEE Access 2019, 7, 61797–61810. [Google Scholar] [CrossRef]
  17. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. arXiv 2015, arXiv:1512.02325. [Google Scholar]
  19. Hu, L.; Ma, J.; Fang, Y. Defect recognition of insulators on catenary via multi-oriented detection and deep metric learning. In Proceedings of the 2019 Chinese Control Conference (CCC), Guangzhou, China, 27–30 July 2019; pp. 7522–7527. [Google Scholar]
  20. Han, J.; Yang, Z.; Xu, H.; Hu, G.; Zhang, C.; Li, H.; Lai, S.; Zeng, H. Search like an eagle: A cascaded model for insulator missing faults detection in aerial images. Energies 2020, 13, 713. [Google Scholar] [CrossRef] [Green Version]
  21. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object detection via region-based fully convolutional networks. arXiv 2016, arXiv:1605.06409. [Google Scholar]
  22. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-aware neural architecture search for mobile. arXiv 2018, arXiv:1807.11626. [Google Scholar]
  23. Meng, L.; Peng, Z.; Zhou, J.; Zhang, J.; Lu, Z.; Baumann, A.; Du, Y. Real-time detection of ground objects based on unmanned aerial vehicle remote sensing with deep learning: Application in excavator detection for pipeline safety. Remote Sens. 2020, 12, 182. [Google Scholar] [CrossRef] [Green Version]
  24. Tijtgat, N.; Van Ranst, W.; Volckaert, B.; Goedemé, T.; De Turck, F. Embedded real-time object detection for a uav warning system. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 2110–2118. [Google Scholar]
  25. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  26. Lin, T.-Y.; Dollár, P.; Girshick, R.B.; He, K.; Hariharan, B.; Belongie, S.J. Feature pyramid networks for object detection. arXiv 2016, arXiv:1612.03144. [Google Scholar]
  27. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  28. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  29. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An imperative style, high-performance deep learning library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  30. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
Figure 1. Electrical components: (a) Insulator1; (b) Insulator2; (c) Spacer.
Figure 1. Electrical components: (a) Insulator1; (b) Insulator2; (c) Spacer.
Sensors 20 06961 g001
Figure 2. Single shot multibox detector (SSD) network.
Figure 2. Single shot multibox detector (SSD) network.
Sensors 20 06961 g002
Figure 3. Improved SSD network (ISSD).
Figure 3. Improved SSD network (ISSD).
Sensors 20 06961 g003
Figure 4. Multiscale prediction module in SSD.
Figure 4. Multiscale prediction module in SSD.
Sensors 20 06961 g004
Figure 5. Neck: (a) Channel plus; (b) Channel concatenate.
Figure 5. Neck: (a) Channel plus; (b) Channel concatenate.
Sensors 20 06961 g005
Figure 6. Offline detection.
Figure 6. Offline detection.
Sensors 20 06961 g006
Figure 7. Online detection.
Figure 7. Online detection.
Sensors 20 06961 g007
Figure 8. Unmanned aerial vehicle (UAV) platform. (a) Top view, (b) Front view.
Figure 8. Unmanned aerial vehicle (UAV) platform. (a) Top view, (b) Front view.
Sensors 20 06961 g008
Figure 9. Insulator1 detection results. The blue box is the ground truth of the position of insulator1 in the image.
Figure 9. Insulator1 detection results. The blue box is the ground truth of the position of insulator1 in the image.
Sensors 20 06961 g009
Figure 10. Insulator2 detection results. The blue box is the ground truth of the position of insulator2 in the image.
Figure 10. Insulator2 detection results. The blue box is the ground truth of the position of insulator2 in the image.
Sensors 20 06961 g010
Figure 11. Spacer detection results. The blue box is the ground truth of the position of spacer in the image.
Figure 11. Spacer detection results. The blue box is the ground truth of the position of spacer in the image.
Sensors 20 06961 g011
Figure 12. Performance comparison of different detectors: (a) faster R-CNN; (b) R-FCN; (c) SSD_vgg; (d) SSD_resnet; (e) SSD_mobilenet; (f) SSDLite_mobile; (g) SSDLite_mnas; (h) Ours1; (i) Ours2.
Figure 12. Performance comparison of different detectors: (a) faster R-CNN; (b) R-FCN; (c) SSD_vgg; (d) SSD_resnet; (e) SSD_mobilenet; (f) SSDLite_mobile; (g) SSDLite_mnas; (h) Ours1; (i) Ours2.
Sensors 20 06961 g012aSensors 20 06961 g012b
Figure 13. Speed versus accuracy: (a) test on the server; (b) test on TX2.
Figure 13. Speed versus accuracy: (a) test on the server; (b) test on TX2.
Sensors 20 06961 g013
Table 1. Number of objects in each category.
Table 1. Number of objects in each category.
LabelInsulator1Insulator2Spacer
Object numbers1485613199
Table 2. Models used in the experiment. Backbone: The basic network used for feature extraction. Main Modules: The basic framework, backbone and feature fusion method of the algorithm.
Table 2. Models used in the experiment. Backbone: The basic network used for feature extraction. Main Modules: The basic framework, backbone and feature fusion method of the algorithm.
Simplified NameBackboneMain Modules
Faster R-CNNResNet-101Faster R-CNN(ResNet-101) 1
R-FCNResNet-101R-FCN(ResNet-101)
SSD_vggVGG-16SSD(VGG-16)
SSD_resnetResNet-101SSD(ResNet-101)
SSD_mobilenetMobileNet_V2SSD(MobileNet_V2)
SSDLite_mobileMobileNet_V2SSDLite (MobileNet_V2)
SSDLite_mnasMnasNetSSDLite (MnasNet)
Ours1MnasNetSSD (MnasNet + Feature fusion + Channel plus) 2
Ours2MnasNetSSD (MnasNet + Feature fusion + Channel concatenate) 3
1 The basic framework of the algorithm is Faster R-CNN. The backbone is ResNet-101 and there is no feature fusion module. 2 The basic framework of the algorithm is SSD. The backbone is MnasNet and the feature fusion method of channel plus is used. 3 The basic framework of the algorithm is SSD. The backbone is MnasNet and the feature fusion method of channel concatenate is used.
Table 3. Detection accuracy of different algorithms. Note that the best results are bold-faced.
Table 3. Detection accuracy of different algorithms. Note that the best results are bold-faced.
MethodInsulator1 AP (%)Insulator2 AP (%)Spacer AP (%)mAP (%)
Faster R-CNN90.190.1100.093.4
R-FCN87.988.888.388.1
SSD_vgg90.589.7100.093.4
SSD_resnet90.990.1100.093.7
SSD_mobilenet90.981.199.190.3
SSDLite_mobile90.885.299.491.8
SSDLite_mnas90.988.299.192.7
Ours190.988.7100.093.2
Ours290.990.4100.093.8
Table 4. Statistics of detection time of different algorithms; server_time: the average detection time of a single image on the server; TX2_time: the average detection time of a single image on TX2; TX2_vartime: variance of the detection time of the image on TX2. Note that the best results are bold-faced.
Table 4. Statistics of detection time of different algorithms; server_time: the average detection time of a single image on the server; TX2_time: the average detection time of a single image on TX2; TX2_vartime: variance of the detection time of the image on TX2. Note that the best results are bold-faced.
MethodModel Size (MB)Server_Time (ms)TX2_Time (ms)TX2_Vartime (ms)
Faster R-CNN360.0782103640
R-FCN383.425757658
SSD_vgg91.622221166
SSD_resnet211.172620070
SSD_mobilenet40.383332776
SSDLite_mobile11.913232076
SSDLite_mnas19.772323870
Ours143.581718954
Ours243.731515445
Table 5. The fps of Ours1 and Ours2.
Table 5. The fps of Ours1 and Ours2.
ModelServer (fps)TX2 (fps)
Ours110.642.79
Ours236.858.27
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, X.; Li, Y.; Shuang, F.; Gao, F.; Zhou, X.; Chen, X. ISSD: Improved SSD for Insulator and Spacer Online Detection Based on UAV System. Sensors 2020, 20, 6961. https://doi.org/10.3390/s20236961

AMA Style

Liu X, Li Y, Shuang F, Gao F, Zhou X, Chen X. ISSD: Improved SSD for Insulator and Spacer Online Detection Based on UAV System. Sensors. 2020; 20(23):6961. https://doi.org/10.3390/s20236961

Chicago/Turabian Style

Liu, Xuan, Yong Li, Feng Shuang, Fang Gao, Xiang Zhou, and Xingzhi Chen. 2020. "ISSD: Improved SSD for Insulator and Spacer Online Detection Based on UAV System" Sensors 20, no. 23: 6961. https://doi.org/10.3390/s20236961

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop