Next Article in Journal
Comparative Analysis of Low-Cost Portable Spectrophotometers for Colorimetric Accuracy on the RAL Design System Plus Color Calibration Target
Next Article in Special Issue
Knowledge Reasoning- and Progressive Distillation-Integrated Detection of Electrical Construction Violations
Previous Article in Journal
Track Deflection Monitoring for Railway Construction Based on Dynamic Brillouin Optical Time-Domain Reflectometry
Previous Article in Special Issue
Lightweight Substation Equipment Defect Detection Algorithm for Small Targets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Scene Knowledge Integrating Network for Transmission Line Multi-Fitting Detection

1
College of Quality & Standardization, China Jiliang University, Hangzhou 310018, China
2
State Grid Huzhou Electric Power Supply Company, Huzhou 313000, China
3
Automation Department, North China Electric Power University, Baoding 071003, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8207; https://doi.org/10.3390/s24248207
Submission received: 25 October 2024 / Revised: 11 December 2024 / Accepted: 16 December 2024 / Published: 23 December 2024

Abstract

:
Aiming at the severe occlusion problem and the tiny-scale object problem in the multi-fitting detection task, the Scene Knowledge Integrating Network (SKIN), including the scene filter module (SFM) and scene structure information module (SSIM) is proposed. Firstly, the particularity of the scene in the multi-fitting detection task is analyzed. Hence, the aggregation of the fittings is defined as the scene according to the professional knowledge of the power field and the habit of the operators in identifying the fittings. So, the scene knowledge will include global context information, fitting fine-grained visual information and scene structure information. Then, a scene filter module is designed to learn the global context information and fitting fine-grained visual information, and a scene structure module is designed to learn the scene structure information. Finally, the scene semantic features are used as the carrier to integrate three categories of information into the relative scene features, which can assist in the recognition of the occluded fittings and the tiny-scale fittings after feature mining and feature integration. The experiments show that the proposed network can effectively improve the performance of the multi-fitting detection task compared with the Faster R-CNN and other state-of-the-art models. In particular, the detection performances of the occluded and tiny-scale fittings are significantly improved.

1. Introduction

The stable operation of transmission lines has a vital impact on the power system [1]. The fittings are the key components that can maintain the stability of transmission lines and are divided into protective fittings, connecting fittings, strain clamps and suspension clamps, etc., [2]. However, the fittings are susceptible to the interference of wind, rain and snow because transmission lines are usually distributed in harsh environments such as snowfields, plains and deserts. Therefore, the fittings are liable to defects such as corrosion, incline and damage [3,4,5]. So, it is necessary to inspect the fittings regularly to reduce the defects for the stable operation of transmission lines.
With the widespread application of unmanned aerial vehicle (UAV) inspection technology in transmission lines [6,7,8], the workload of operators has been significantly reduced. However, the large volume of aerial images captured by UAVs still requires manual inspection, which limits the overall efficiency Hence, it is essential to introduce intelligent detection technologies with transmission line components. The fittings, as one of the fundamental components of transmission lines [9], are responsible for greatly improving the efficiency of power maintenance [10], reducing the burden of operators and realizing the defect detection of the fittings.
At present, most state-of-the-art object detection methods [11,12,13] view partial regions of images superficially without using the domain knowledge of the object. These methods require high-quality feature representations of region proposals to achieve satisfactory detection results. However, in the multi-fitting detection task, the feature quality of region proposals is influenced by the following two issues:
(1)
Severe occlusion. Generally speaking, there are varying degrees of occlusion between the fittings due to the cameras’ shooting angles and the fittings’ connection modes. As shown in Figure 1a, the fittings such as yoke plates, u-type hanging rings and hanging boards are occluded by the shielded rings, which leads to a lack of the occluded region features. At the same time, the region proposal of shielding rings has noise features due to the existence of other fittings.
(2)
Tiny-scale object. As shown in Figure 1b, the tiny-scale fittings such as hanging boards and u-type hanging rings account for a miniature proportion in the whole image due to the influence of the camera range and the scale of the fittings, resulting in less information on the region proposal features.
Both of the problems above will cause a decline in the quality of the region proposal features. Accordingly, multi-fitting detection still has great potential for making further progress.
Most visual concepts in natural images consist of scenes, objects and relationships [14]. Psychology indicates that scene knowledge and object relationships play a vital role in the process of object recognition by humans. General scene knowledge has been validated to adequately improve the effect of object detection in the literature [14,15,16,17,18,19]. As shown in Figure 2a, if humans cannot directly identify the ship (object) in the image, they will tend to utilize the sea (scene) to infer the object. Obviously, the object at sea is more likely to be a ship than other transportation methods such as cars, trains, etc. In traditional object detection tasks, general scene knowledge often plays a vital role in improving detection accuracy by leveraging the relationships between objects and their surroundings. However, in the context of multi-fitting detection for transmission lines, the definition and application of scene knowledge present unique challenges. Unlike natural scenes, where the background (e.g., sea, forest) can provide strong contextual clues, the transmission line inspection scene lacks such explicit correlations. Here, the “scene” is redefined as the aggregation of fittings, emphasizing the domain-specific relationships and structural arrangements among fittings. Expert knowledge from the power industry, such as the typical combinations and spatial arrangements of fittings, can be integrated.
Hence, we define the scene as the aggregation of the fittings in the multi-fitting detection task. As shown in Figure 2b, the scene in the image is an anti-dancing scene. Experienced operators can infer the occluded fittings in the green box according to the visual features and structural characteristics. Therefore, in a multi-fitting detection task, the scene defined by this method will lead to scene knowledge, including global context information, fitting fine-grained visual information and structure information between fittings. As a result, unlike previous detection tasks [11,12,13], our method mainly considers the severe occlusion problem and tiny-scale object problem and utilizes the scene knowledge containing domain information to improve the effect of multi-fitting detection. Specifically, the proposed SKIN incorporates a scene filter module (SFM) to extract global context and fine-grained fitting information by filtering noise and highlighting relevant features. Additionally, the scene structure information module (SSIM) models the structural relationships among fittings using a scene-fitting prior matrix derived from professional knowledge. These innovations ensure that occluded and tiny fittings, often missed by general-purpose object detection methods, are effectively identified by leveraging scene-level domain knowledge. Compared with existing approaches that treat fitting detection as a generic object detection problem, the proposed method significantly enhances detection performance by embedding task-specific knowledge into the model architecture.
The main contributions of our papers can be summarized as follows:
(1)
We define the aggregation of fittings as a “scene,” incorporating electrical power industry knowledge and operators’ recognition habits to improve multi-fitting detection. Hence, as shown in Figure 3, we exemplify eleven common scenes to assist the multi-fitting detection. The Scene Knowledge Integrating Network (SKIN) integrates this knowledge through the scene filter and scene structure information modules.
(2)
The scene filter module collects global context and fine-grained visual information using a Gated Recurrent Unit (GRU), encoding it into scene features and passing it to the scene semantic features for further processing.
(3)
The scene structure information module encodes and learns scene structure information from a scene-fitting prior matrix, integrating this information with the scene semantic features to improve the detection of occluded and small fittings.

2. Related Work

2.1. Fitting Detection

The classical methods of image processing and machine learning were used in the early stage of the research on fitting detection technology. Some of these classical methods identified the fittings directly based on the morphology [20,21,22,23] of different categories of the fittings, and some of them used classical machine learning algorithms for recognition by extracting artificial features. For example, Liu et al. [24] applied the cascade classifier and the support vector machine (SVM) to complete fitting detection by extracting the Harr-Like features and the HOG features from images. Guo et al. [25] designed the multi-cycle and multi-class method based on SVM for fitting detection. However, the reliance on handcrafted features and basic models in these classical methods resulted in weak feature extraction capabilities and limited resistance to interference, leading to unsatisfactory fitting detection performance.
With the development of deep learning, more and more attention has been paid to the research of fitting detection based on deep learning. For example, in order to improve the detection effect of shockproof hammers and grading rings, Tang et al. [26] expanded the dataset by image rotation and adjusted the kernel size of the convolution neural network. Zhang et al. [27] proposed an improved training method based on transfer learning in the yolov3 model to reduce the dataset requirement of the model. As a result, the few-shot problem about clamps and shockproof hammers was solved and the accuracy and generalization of fitting detection were improved. Jiao et al. [28] enhanced the shockproof hammer dataset by the joint expansion of samples and label files to improve the recognition rate. Nevertheless, the above studies were mainly aimed at a few categories of fittings and focused on data augmentation. They did not integrate the overall rules of transmission lines and the characteristics of the fittings into the detection tasks.

2.2. Object Detection Model Integrating Knowledge

In recent years, some approaches based on deep learning have attempted to apply object–object or scene–object relationships in object detection tasks. For instance, in order to improve the detection network, Zheng et al. [18] modeled the object proposal features and the scene features as nodes and modeled the prior knowledge of object–object relationship and object-scene relationship as edges. Hang et al. [29] introduced global semantic pools by utilizing the weight of a classifier, and the high-level semantic representations of different categories in the global semantic pool were updated through attribute and relation. Then, the adaptive global reasoning module was used to obtain the enhanced features to assist the classification and regression. Chen et al. [30] introduced the concepts of image-level context and instance-level context and then modeled the relationship between different objects by proposing a spatial memory network (SMN). Liu et al. [14] proposed a structure inference net (SIN), in which the object–object relationship and scene knowledge were utilized by a GRU. Jiang et al. [31] proposed hybrid knowledge routed modules (HKRM) to integrate explicit knowledge and implicit knowledge into the model by using an explicit knowledge module and an implicit knowledge module. The networks mentioned above mainly modeled the relationships between objects to introduce knowledge, and a part of them included scene knowledge to assist detection. However, the common scene knowledge that cannot be applied to multi-fitting detection was utilized above.

3. Methods

3.1. Overview

The overall network architecture of this paper is shown in Figure 4. Firstly, the processed aerial image of the fittings is utilized as the input of the network. Then, the convolutional neural network resnet101 [32] is utilized as the backbone for feature extraction, and the original scene feature and the region proposal features are obtained by using the region proposal network (RPN) and the pooling layer. The original scene feature learns the global context information and fitting fine-grained visual information by the scene filter module in the supervised learning method and transmits the learned information to the scene semantic features. The scene structure information module supervises the region proposal features, learning the scene structure information from the scene-fitting prior matrix, and promotes the combination of the scene semantic features and the learning results to obtain the relative scene features of region proposals. Finally, these relative scene features are used for feature mining, and the results are integrated into the region proposal features. As a result, the scene knowledge from the relative scene features assists the network in inferring occluded and tiny-scale objects, thereby enhancing multi-fitting detection.

3.2. Scene Filter Module

3.2.1. The Generation of the Original Scene Feature

The scene filter module requires the original scene feature as input, so it is necessary to generate the original scene feature in the first place. Our model is based on Faster R-CNN, whose former part can be summarized as follows:
F C N N = Γ C N N ( Z i | | θ C N N )
K r o i s = Γ R P N ( F C N N | | θ R P N )
where Z i is the i-th input image, Γ C N N ( | | θ C N N ) is the process of feature extraction and θ C N N is the learnable weight in this process. The base features F C N N are utilized to obtain region proposals K r o i s by region proposal network Γ R P N ( | | θ R P N ) , in which θ R P N is the learnable weight.
Inspired by [14], we introduce the global region for generating the original scene feature. After that, the current region proposals can be represented as
K r o i s = { ( x 0 j , y 0 j , x 1 j , y 1 j ) } j = 1 N η + 1
where ( x 0 j , y 0 j ) means top-left coordinate of the j-th region and ( x 1 j , y 1 j ) means bottom-right coordinate of the j-th region. N η is the number of the region proposals and the extra one is the global region.
Then, the current region proposals are utilized to realize pooling operation on the base features, and the results are input into the fully connected layer to obtain the region proposal features and the original scene feature. This process can be expressed as
F t a i l = Γ t a i l ( [ F C N N , K r o i s ] | | θ t a i l ) R ( N η + 1 ) × C γ
where Γ t a i l ( | | θ t a i l ) represents the pooling layer and the fully connected layer and θ t a i l is the learnable weights. After that, F t a i l can be sliced into the region proposal features F p r o p R N η × C γ and the original scene feature F s c R 1 × C γ . In addition, C γ is the dimension of the features.

3.2.2. The Filtering of the Original Scene Feature

Although the original scene feature contains the global context information and the fitting fine-grained visual information, there is a lot of information noise due to the small proportion of effective fitting objects and large proportion of invalid background regions in UAV aerial images.
Thus, the scene filtering module is designed to filter the original scene features, removing redundant noise and emphasizing the fine-grained visual details of the fittings and the relevant global context information. The information used to filter the original scene features can be expressed as follows:
F p r o p = f p r o p m m = 1 N η F f o r e = f p r o p n n = 1 N μ
where represents the selecting process and N μ is the number of region proposals in the selecting process. It means that the features F f o r e , which contain the region proposal features of the fittings as possible, are selected from F p r o p . f p r o p m is the m-th feature in the F p r o p and f p r o p n is the n-th feature in the F f o r e .
From an individual perspective, a single region proposal feature represents the fine-grained visual information of a specific fitting within the image scene. In contrast, from a global perspective, the global context information is derived from the collective set of multiple region proposal features across the entire scene.
To effectively capture both the fine-grained visual details and the global context, a memory unit is needed. This memory unit sequentially processes each region proposal feature one by one, allowing it to retain the relevant fine-grained visual information of each fitting encountered. By sequentially aggregating these features, the memory unit can ultimately capture the global context composed of multiple region proposals within the scene.
However, in the process of aggregating these features, some background features may be present. These background features often contain irrelevant or invalid information that can introduce noise and reduce detection accuracy. Therefore, the memory unit must also have the capability to filter out or forget invalid background information while retaining the important fitting features.
To achieve this, we employ a mechanism that enables the memory unit to selectively remember useful information and discard irrelevant features, ensuring that the aggregated global context information remains accurate and meaningful for the multi-fitting detection task.
So, we select a Gated Recurrent Unit (GRU) [33], which has a simple structure, is easy to combine and has excellent long-term memory. Then, as shown in Figure 5, the original scene feature is filtered by utilizing GRU:
Firstly, the input is constructed as a sequence F f o r e = f p r o p t t = 1 N μ , in which t represents the t-th moment. f p r o p t is the single feature inputted at the t-th moment, and the original scene feature F s c is taken as the initial state of GRU. Then, the update gate u t at the t-th moment can be represented as
u t = σ ( W u [ F s c t 1 , f p r o p t ] )
where σ ( ~ ) is the sigmoid activation function, F s c t 1 is the state of the t 1 moment, [ ~ , ~ ] is the concatenation of matrices and W u is the learnable weights. Accordingly, using the update gate u t , the scene state F s c t at the t-th moment can be obtained:
F s c t = ( 1 u t ) F s c t 1 + u t F ˜ s c t
where represents the Hadamard product. Further, the formula of F ˜ s c t is
F ˜ s c t = φ ( W [ r t F s c t 1 , f p r o p t ] )
where φ ( ~ ) is the tanh activation function, r t is the reset gate at the t-th moment and W is the learnable weights. The reset gate determines how the new input information is combined with the previous memory, and its formula is as follows:
r t = σ ( W r [ F s c t 1 , f p r o p t ] )
where W r is the learnable weights.
After the above process, the filtered scene feature F s c = F s c t ( t = N μ ) can be obtained through the collaborative control of the update gate and reset gate.

3.2.3. The Constraint of the Scene Filtering

The scene filter module is designed in a supervised learning method to constrain the GRU’s memory direction (emphasizing the effective global context information and fitting fine-grained visual information and ignoring invalid background information). The process is as follows:
(1)
First, the scene label space is constructed. Specifically, the scene label Y i = y i 0 , y i 1 , , y i n s R n s + 1 is assigned to the i-th image Z i . In addition, n s is the number of scene categories and the extra dimension represents the situation of no scene. When the image contains the scenes, the values of their corresponding dimension are 1, and otherwise they are 0. And if the image does not contain the scenes, the y i 0 is 1.
(2)
Second, the scene classifier is constructed for completing the scene classification task. The filtered scene feature is mapped into the scene label space through the scene classifier:
Y ^ i = σ ( F s c W c a ) = y ^ i 0 , y ^ i 1 , , y ^ i n s R n s + 1
where W c a R C γ × ( n s + 1 ) is the weight of the scene classifier. And the distribution of Y ^ i in the scene label space is urged to approach Y i by the following loss function:
L s c = 1 N b 1 N s m = 1 N b n = 1 N s ( y m n log ( y ^ m n ) + ( 1 y m n ) log ( 1 y ^ m n ) )
where N s means n s + 1 , and N b is the number of images contained in a batch.
Hence, in order to complete the scene classification task, GRU will be constrained to remember the valid information and ignore the invalid information in the process of backpropagation. And the scene classifier weights will obtain the information during the mapping process. Therefore, referring to [29], we extract the scene classifier weight to form the scene semantic feature pool, which is similar to the memory when humans recall a category. So, the information can be transmitted into the semantic concepts of all scenes (not just the scenes that appear in the image) in the memory. Further, the scene filtering process is equivalent to the human viewing each image, and the semantic concept of the scene categories in memory is modified based on the difference in the image. Accordingly, the scene semantic features are also compatible with the specificity of each image.

3.3. Scene Structure Information Module

3.3.1. Scene-Fitting Prior Matrix Construction

The main function of the scene structure information module is to learn the structure information between the fittings and combine the information with the scene semantic features to generate the relative scene features. The structure information can be extracted in different ways: from the perspective of instance-level, it is the structure relation between the instance-level fittings and can be represented as the fitting-fitting co-occurrence matrix [34,35] or spatial location matrix [36]; from the perspective of scene-level, the structure information is the combination of the fittings included in the scene. Therefore, we construct the scene-fitting prior matrix using conditional probability to represent the scene-level structural information.
Hence, the conditional probability formula for constructing the scene-fitting prior matrix is
P ( f i t i | s c j ) = P ( f i t i , s c j ) P ( s c j )
This formula represents the probability of the i-th fitting when the j-th scene appears. There are n s + 1 categories of scenes, and no scene (consisting of scattered fittings that cannot be classified as any of the scenes in the definition) is also a category of scenes. The elements of the above formulate can be represented as
P ( f i t i , s c j ) = N i j N a l l
P ( s c j ) = N j N a l l
where N i j is the number of the j-th scene which contains the i-th fitting, N j is the number of the j-th scene and N a l l is the number of all scenes. Further, the conditional probability formula for constructing the scene-fitting prior matrix can be represented as
P ( f i t i | s c j ) = N i j N j
Therefore, the scene-fitting prior matrix M s f R ( n s + 1 ) × n f can be constructed after counting the N i j and N j . As shown in Figure 6, all eleven categories of scenes have obvious rules for the fitting combination (scene-level structure information), except no scene.

3.3.2. The Network Structure of SSIM

As shown in Figure 7, referring to [31], we propose the scene structure information module to learn the scene-level structure information. However, the scene-level structure information learned in this process depends on the statistical results and lacks compatibility with each image. Therefore, the specific information contained in the scene semantic features from the scene filtering module should be transmitted to the scene structure information module. This allows for the extraction of the relative scene features of the regions. Consequently, the relative scene features will encompass global context information, fine-grained visual details of the fittings and scene structure information, in line with the definition of scene knowledge presented in this paper. Then, the relative scene features are used for feature mining, and the results are integrated into the region proposal features for improving fitting detection. The process of the scene structure information module is as follows:
(1) Obtain the ground-truth scene vectors. There are the ground-truth categories G T c l s R N η × 1 corresponding to the region proposal features F p r o p R N η × C γ . Hence, the ground-truth scene vectors G T sv = { g t s v n } n = 1 N η R N η × ( n s + 1 ) can be obtained through the map of G T c l s on M s f and g t s v n is one of the ground-truth scene vectors. This explicit mapping using G T c l s will ensure the information extraction and provide a stable guarantee for the supervised learning of the scene-level structure information.
(2) Learn the scene-level structure information. The information is learned by constructing the multiple convolution layers:
F s v = C o n v s s v ( F p r o p )
where F s v is the predicted scene vectors and can be represented as { f s v n } n = 1 N η R N η × ( n s + 1 ) . f s v n is one of the predicted scene vectors, and C o n v s s v ( ~ ) means the multiple convolution layers. Further, the scene knowledge module will learn the scene-level structure information contained in the scene-fitting matrix by completing the task of urging F s v to approach G T sv . The loss function in this task is as follows:
L s v ( F s v , G T s v ) = 1 N b n = 1 N b m = 1 N η ( f s v m ) n ( g t s v m ) n 1
where 1 is the 1-norm of vector.
(3) Obtain the relative scene features. The scene structure information learned in the previous step depends on the statistical results, while ignoring the specific visual features of the fittings and the specific combination characteristics in each image (fitting fine-grained visual information and global context information). Therefore, it is necessary to transmit the specific information of the scene semantic features from the scene filter module to the results of the scene structure information module:
F r s = ψ ( F s v ) ( W c a ) T
where ψ ( ~ ) is the softmax activation function, ( ~ ) T is the transpose of matrix, F r s R N η × C γ is the relative scene features and W c a is the learnable parameters. Because the multiple fittings aggregate into a scene, the fitting will have the relative scene feature. The fitting’s relative scene feature is affected by the fitting fine-grained visual information of other fittings in the image, the global context information of the fitting’s location in the image and the structure information between the fitting and other fittings in the image, conforming to the definition of scene knowledge. The above three categories of information will affect each other in different images in order to make the relative scene features specific.
(4) Feature mining and knowledge integration. Finally, F r s is utilized for feature mining to obtain the scene knowledge features F s k :
F s k = F r s W s k
where W s k R C γ × C e is the weight of the transformation matrix. So, C e is the dimension of the scene knowledge features. The combined features obtained by combining F p r o p and F s k in cascade way are input into the classification detector and the regression detector to assist in the inference of the occluded fittings and the tiny-scale fittings.

4. Experiment

4.1. Experiment Settings

4.1.1. Dataset Description

The multi-fitting dataset, which is taken by UAV from the inspection site, is selected. It contains fourteen categories of fittings, including pre-twisted suspension clamp (PT), bag-type suspension clamp (BT), compression-type strain clamp (CT), wedge-type strain clamp (WT), hanging board (HB), u-type hanging ring (UT), yoke plate (YP), parallel groove clamp (PG), shockproof hammer (SH), spacer (SP), grading ring (GR), shielded ring (SR), weight (WE) and adjusting board (AB). As shown in Table 1, The train set has 1330 images, containing 16,358 fitting objects. The test set has 318 images, containing 2767 fitting objects.

4.1.2. Experiment Environment and Hyperparameter Setting

The Scene Knowledge Integrating Network is trained and tested on an NVIDIA RTX3090 professional accelerator card with 24 GB of VRAM, which is headquartered in Santa Clara, California, USA., ensuring sufficient computational power for large-scale deep learning experiments. The operating system is Ubuntu 18.04.5 LTS, with CUDA 11.2 used to optimize GPU-based training. The implementation is written in Python 3.8, utilizing the PyTorch framework (version 1.x), which provides a flexible and efficient platform for building and training deep learning models.
To verify the effectiveness of the proposed Scene Knowledge Integrating Network (SKIN), Faster R-CNN with ResNet101, pretrained on the ImageNet dataset, is used as the baseline model. ResNet101 serves as the backbone network for feature extraction, leveraging its deep architecture to extract high-quality features. The hyperparameters for the region proposal network (RPN) are set to 128 during training and 300 during testing, ensuring sufficient candidate proposals for accurate detection.
The dataset is augmented using image inversion, which enhances diversity in training samples and improves model robustness. The training process employs the stochastic gradient descent (SGD) algorithm with a momentum of 0.9 to stabilize convergence. The initial learning rate is set to 0.0034, which decays by 10% every 15 epochs to allow fine-tuning as training progresses. The total number of epochs is 20, balancing between convergence and computational cost.
The scene filter module (SFM) selects 40 region proposal features for filtering, which are processed to extract global context and fine-grained fitting visual information. In the scene structure information module (SSIM), the dimension of the scene knowledge features is set to 256, providing sufficient capacity to represent structural relationships and domain-specific scene knowledge effectively.
The experiments are conducted under controlled conditions to ensure reproducibility, with random seeds set for initialization and data splitting. Performance metrics, including mAP50, AP50, and AR, are used to evaluate the effectiveness of the proposed method in comparison to the baseline and other models. Our code is available at the following link: https://github.com/CharmingWang/SKIN (accessed on 15 December 2024).

4.2. Comparison with State-of-the-Art Models

To verify the improvements in model performance brought by the proposed scene filter module and scene structure information module, as well as the advantages over current general object detection algorithms, this paper compares the proposed method with commonly used object detection methods at present. The comparison methods include SSD [37], RetinaNet [38], YOLOv5 [39], YOLOv8 [40], R-FCN [41], EfficientNetv2 [42] and MobileNetv2 [43], as well as some object detection models based on the Transformer architecture [44], such as Swin Transformer [45], DETR [46], DINO [47] and CO-DETR [48]. The results are shown in Table 2. The model that incorporates the SFM and SSIM modules has higher AP50 values in categories such as BT, WT and SH compared to current object detection algorithms like YOLOv8, DETR and other mainstream methods. Additionally, the detection accuracy for categories like PT and SR is also close to the optimal value. Figure 8 presents a visual comparison between the proposed algorithm and other algorithms.
The experimental results show that our algorithm performs better than other general-purpose object detection algorithms on multiple indicators, especially with significant performance improvement in complex scenarios. Compared with the baseline model, the improved model has improved detection accuracy in most categories, and has achieved a 4.9% improvement in overall detection accuracy. This is because the introduction of the scene filtering module and the scene knowledge module enables the algorithm to have a deeper understanding of the special scene of transmission line fittings, thereby improving the detection performance.
Compared with traditional models, such as SSD and RetinaNet, and Transformer-based models, such as Swin Transformer and DETR, the method proposed in this paper has achieved significant advantages in the detection of most categories. Compared with advanced YOLOv8, Swin Transformer and DETR, mAP50 has increased by 0.9%, 1.1% and 3.7%, respectively. This shows that the scene filtering module effectively filters out large proportions of invalid areas, reduces noise interference and improves overall detection accuracy. Moreover, the scene knowledge module enhances the generalization ability of the model for different scenes and environments by learning the scene-fittings co-occurrence matrix and prior knowledge. The experimental results for testing speed are shown in Table 2. The proposed model cannot achieve the best performance. However, In current power line inspection workflows, unmanned aerial vehicles (UAVs) are typically used to capture large volumes of image data, which are then transmitted to local servers for processing. This approach leverages local computing resources to perform detailed and accurate analysis, ensuring that the detection results meet the stringent precision requirements of real-world applications. Although our method may not achieve the fastest detection speed, it satisfies the processing speed requirements for local server-based analysis and provides the high-accuracy results necessary for practical deployment in transmission line inspection tasks.
As shown in Figure 8, in each subfigure, the left side displays the test results of the baseline model, while the right side shows the test results of the proposed model. The yellow line indicates the change in the comparison of test results. From the figure, it can be seen that our proposed model significantly improves the missed detections and false detections of dense targets. The specific improvements are as follows:
(1)
In subfigure (a), our model detects two inverted bag-type suspension clamps.
(2)
In subfigure (b), our model detects small fitting targets connecting the grading ring and the yoke plate, specifically the u-type hanging ring.
(3)
In subfigures (c) and (d), our model detects the occluded link plates.
(4)
In subfigure (e), our model detects the previously missed yoke plate, and the misdetected wedge-type strain clamp bounding box is also correctly rectified.
(5)
In subfigure (f), our model corrects the false detection of a u-type hanging ring and accurately detects the previously missed yoke plate.
These improved detection results are attributed to our model’s ability to utilize scene information within the image to infer which fittings are more likely to appear, thereby producing more accurate results.

4.3. Ablation Analysis

Taking the Faster R-CNN as the baseline model, we conducted detailed experiments to prove the effectiveness of different modules and different hyperparameters in different modules in the multi-fitting detection task.
(1) Verify the effectiveness of the module. Based on Faster R-CNN, we added a scene filter module and scene structure information module in turn in order to verify the effectiveness of scene filter module and scene structure information module. As shown in Table 3, the AP50 value of adding SFM increases by 3.3% compared with the baseline mode. And the SSIM mainly supplements the scene structure information that the SFM has not learned, so it cannot be used alone. Therefore, the effectiveness of the SSIM is reflected by adding the SSIM on the basis of the SFM. So, as shown in Table 3, the AP50 value of the network with the SFM and the SSIM is increased by 4.8%, which is 1.5% higher than that of the network with the SFM. Further, the experiments above also explain the effectiveness of the global context information and fitting fine-grained visual information mined by the SFM and the effectiveness of the scene structure information mined by the SSIM.
(2) The effect of different prior matrices. In order to verify the validity of the scene-fitting prior matrix, we design the ones prior matrix and random prior matrix used in SSIM. As shown in Table 4, the effects of using the ones prior matrix and random prior matrix are obviously lower than the effect of using scene-fitting prior matrix and also lower than the effect of using SKIN only with SFM. Those show that the wrong prior matrix will mislead the learning direction of SKIN, make the SKIN learn the wrong scene structure information and lead to the poor effect.
(3) The effect of different N μ values in the SFM. In the scene filter module, we utilize GRU to filter the original scene feature. The N μ region proposal features are sequenced, and retained to learn fitting fine-grained visual information and global context information in the filtering process. As shown in Table 5, we tested the effect of different N μ values when the dimension of the scene knowledge features is 256. It can be seen that the SKIN obtains the best effect when N μ = 40 . This is because, when the N μ value is too small, the original scene feature only remembers part of the fitting fine-grained visual information and global context information in the image through GRU, and a lot of information is missed. When the N μ value is too large, there will be too much invalid background information which GRU cannot completely filter out, resulting in the noise information affecting the detection effect of the SKIN.
(4) The effect of different scene knowledge feature dimensions. In the scene structure information module, the relative scene features containing scene knowledge are utilized for feature mining and feature integration, and the dimension of the feature mining is determined by C e . As shown in Table 6, when N μ = 40 , we test the effects of different C e values. And the SKIN obtains the best effect when C e = 256 . This is because, when the C e value is too small, it is not enough to cover the scene knowledge obtained by feature mining. When the C e value is too large, there is much redundant information and the dimension of the combined features is too high. So, it is strenuous for the classifier and the regression to extract the vital information from high-dimension features, which affects the detection effect.

4.4. More Discussion

As shown in Figure 9, we provide new examples to discuss additional limitations of the model. The figure illustrates that our model still faces challenges related to severe occlusion and extreme scale variations:
(1)
Severe Occlusion: In some cases, when fittings are extensively obscured by other components, the model’s ability to infer their presence is reduced, even with the assistance of scene knowledge.
(2)
Extreme Scale Variations: Very small fittings, which occupy only a few pixels, pose challenges due to limited visual information, making them harder to detect accurately.
In future work, we plan to address these issues by incorporating higher-resolution imagery or multi-scale feature fusion techniques to improve detection performance.

5. Conclusions

In this study, we redefined the concept of a “scene” for transmission line fittings, distinguishing it from conventional object detection scenes due to the unique structural relationships between fittings. Building on this, we proposed the Scene Knowledge Integrating Network (SKIN) to address the challenges of severe occlusion and tiny-scale objects in multi-fitting detection.
The SKIN leverages the scene filter module (SFM) to capture fine-grained visual details and global context, while the scene structure information module (SSIM) models the structural relationships among fittings. These components aggregate information through a global semantic pool, enabling more effective feature mining and integration to enhance detection performance.
The experimental results demonstrate that the SKIN achieves a 4.8% increase in mAP compared to the baseline model. Additionally, the detection performance for tiny-scale fittings improves by 11.5%, and it improves by 9.9% for severely occluded fittings. These findings highlight the effectiveness of integrating scene knowledge and provide a new approach for applying domain-specific knowledge to solve detection challenges in the power industry.
In future research, we aim to further address the challenges of severe occlusion and extreme scale variations by exploring techniques such as higher-resolution imagery, multi-scale feature fusion and 3D spatial modeling to enhance detection performance under these conditions.

Author Contributions

Conceptualization, X.C.; Data curation, Q.W.; Investigation, J.X.; Methodology, X.C. and X.X.; Resources, J.X.; Software, X.C., J.X. and W.Z.; Validation, X.X.; Visualization, W.Z.; Writing—original draft, X.C.; Writing—review and editing, X.X. and Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by The Natural Science Foundation of China (52175257), The Active Design Projects of Key R&D Plans of Zhejiang Province (2021C01053), The National Key R&D Programs (2021YFC3340400) and The Special Fund for Basic Scientific Research of Central Universities (2023JC006, 2024MS136).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

We are sorry that the power data cannot be disclosed due to its particularity and confidentiality; further inquiries can be directed to the corresponding authors.

Conflicts of Interest

Author Jing Xu was employed by the company State Grid Huzhou Electric Power Supply Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Jenssen, R.; Roverso, D. Automatic autonomous vision-based power line inspection: A review of current status and the potential role of deep learning. Int. J. Electr. Power Energy Syst. 2018, 99, 107–120. [Google Scholar]
  2. Zhao, Z.; Qi, H.; Nie, L. A Review of Visual Inspection of Transmission Lines Based on Deep Learning. Guangdong Electr. Power 2019, 32, 13. [Google Scholar]
  3. Gao, R.; Cheng, X.; Fan, B. A Brief Discussion on the Necessity of Using X-ray Inspection for Defects in the Tension Lines of Transmission Lines. China Equip. Eng. 2020, 21, 181–182. [Google Scholar]
  4. Fang, Z.; Lin, W.; Fan, S.; Ma, Y.; Gao, X.; Wu, H. Defect Identification Method for Small Fittings of Transmission Line Towers Based on Hierarchical Recognition Model. Power Inf. Commun. Technol. 2020, 18, 16–24. [Google Scholar]
  5. Zhao, Z.; Zhang, W.; Qi, Y.; Zhai, J.; Zhao, Q. Causal Classification Method for Defects in Transmission Line Fittings by Integrating Deep Features. J. Beijing Univ. Aeronaut. Astronaut. 2021, 47, 461–468. [Google Scholar]
  6. Chen, R.; Xu, H. Research on UAV Power Inspection Technology for High-Voltage Transmission Lines. Electron. Test. 2021, 20, 92–94. [Google Scholar]
  7. Huang, Z.; Wang, H.; Zhai, X.; Wang, Y.Q.; Gao, C. Research and Application of Autonomous Inspection Methods for Transmission Lines Using Drones. J. Comput. Technol. Autom. 2021, 40, 157–161. [Google Scholar]
  8. Shen, J.; Zhang, X.; Chen, Y.; Wang, H.; Huang, Z.; Ji, Y. Drone Inspection Methods for Transmission Lines in Complex Scenarios. Eng. Surv. 2021, 49, 73–78. [Google Scholar]
  9. Liu, X.; Miao, X.; Jiang, H.; Chen, J. Data analysis in visual power line inspection: An in-depth review of deep learning for component detection and fault diagnosis. Annu. Rev. Control 2020, 50, 253–277. [Google Scholar] [CrossRef]
  10. Peng, X.; Qian, J.; Wu, G.; Mai, X.; Wei, L.; Rao, Z. Fully Autonomous Inspection System for Overhead Transmission Lines Using Robots and Demonstration Applications. High Volt. Eng. 2017, 43, 2582–2591. [Google Scholar]
  11. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision 2015, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  12. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  13. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  14. Liu, Y.; Wang, R.; Shan, S.; Chen, X. Structure inference net: Object detection using scene-level context and instance-level relationships. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6985–6994. [Google Scholar]
  15. Zhang, Z.; Hoai, M. Object detection with self-supervised scene adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 21589–21599. [Google Scholar]
  16. Sagar, A.S.; Chen, Y.; Xie, Y.; Kim, H.S. MSA R-CNN: A comprehensive approach to remote sensing object detection and scene understanding. Expert Syst. Appl. 2024, 241, 122788. [Google Scholar] [CrossRef]
  17. Xie, X.; Cheng, G.; Li, Q.; Miao, S.; Li, K.; Han, J. Fewer is more: Efficient object detection in large aerial images. Sci. China Inf. Sci. 2024, 67, 112106. [Google Scholar] [CrossRef]
  18. Li, Z.; Du, X.; Cao, Y. Gar: Graph assisted reasoning for object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1295–1304. [Google Scholar]
  19. Shu, X.; Liu, R.; Xu, J. A Semantic Relation Graph Reasoning Network for Object Detection. In Proceedings of the 2021 IEEE 10th Data Driven Control and Learning Systems Conference (DDCLS), Suzhou, China, 14–16 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1309–1314. [Google Scholar]
  20. Tan, L.; Wang, Y.; Shen, C. Obstacle Vision Detection and Recognition Algorithm for De-icing Robots on Transmission Lines. J. Instrum. Meas. 2011, 32, 8. [Google Scholar]
  21. Jin, L.; Hu, J.; Yan, S. Image-Based Fault Diagnosis Method for Spacers of High-Voltage Transmission Lines. High Volt. Eng. 2013, 39, 1040–1045. [Google Scholar]
  22. Wang, W.; Zhang, J.; Han, J.; Liu, L.; Zhu, M. Detection Method for Wire Breakage and Foreign Object Defects in Transmission Lines Based on UAV Images. Comput. Appl. 2015, 35, 2404–2408. [Google Scholar]
  23. Wan, L.; Wu, S.; Xie, F.; Liu, Q.; Dai, J.C. Monitoring System for Tension Splice Clamps of Transmission Lines Based on Image Processing. J. Wuhan Univ. (Eng. Ed.) 2020, 53, 1106–1111. [Google Scholar]
  24. Liu, H. Research on Visual Recognition Methods for Obstacles in High-Voltage Transmission Line; Harbin Institute of Technology: Harbin, China, 2017. [Google Scholar]
  25. Guo, S. Research on Obstacle Recognition and Localization for Line Inspection Robots Based on Binocular Vision; Shandong University of Science and Technology: Qingdao, China, 2020. [Google Scholar]
  26. Tang, Y.; Han, J.; Wei, W.; Ding, J.; Peng, X. Research on Component Recognition and Defect Detection in Transmission Lines Using Deep Learning. Electron. Meas. Technol. 2018, 41, 60–65. [Google Scholar]
  27. Zhang, Y.; Wu, G.; Liu, Z.; Yang, S.; Xu, W. Transfer Learning for Detection of Shock Absorbers and Clamps in Transmission Lines Based on YOLOv3 Network. Comput. Appl. 2020, 40, 188–194. [Google Scholar]
  28. Jiao, R.T.; Ni, H.; Wang, Z. Research on Identification of Shock Absorbers in Transmission Lines Based on Faster R-CNN Algorithm. J. Chang. Eng. Inst. (Nat. Sci. Ed.) 2021, 22, 38–43. [Google Scholar]
  29. Xu, H.; Jiang, C.; Liang, X.; Lin, L.; Li, Z. Reasoning-RCNN: Unifying Adaptive Global Reasoning into Large-Scale Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  30. Chen, X.; Gupta, A. Spatial memory for context reasoning in object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4086–4096. [Google Scholar]
  31. Jiang, C.; Xu, H.; Liang, X.; Lin, L. Hybrid knowledge routed modules for large-scale object detection. Adv. Neural Inf. Process. Syst. 2018, 31, 1559–1570. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Cho, K.; Van, M.B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, 1409, 1259. [Google Scholar]
  34. Galleguillos, C.; Rabinovich, A.; Belongie, S. Object categorization using co-occurrence, location and appearance. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8. [Google Scholar]
  35. Chen, Z.; Wei, X.S.; Wang, P.; Guo, Y. Multi-label image recognition with graph convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5177–5186. [Google Scholar]
  36. Zhai, Y.; Yang, X.; Wang, Q.; Zhao, Z.; Zhao, W. Hybrid Knowledge R-CNN for Transmission Line Multi-fitting Detection. IEEE Trans. Instrum. Meas. 2021, 70, 1–12. [Google Scholar]
  37. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in Pytorch. 2017. Available online: https://openreview.net/pdf?id=BJJsrmfCZ (accessed on 29 October 2017).
  38. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  39. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13039–13048. [Google Scholar]
  40. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  41. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  42. Kamal, M.R.M.; Shahbudin, S.; Rahman, F.Y.A. Photovoltaic (PV) Module Defect Image Classification Analysis Using EfficientNetV2 Architectures. In Proceedings of the 2023 IEEE 14th Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 5 August 2023; pp. 236–241. [Google Scholar]
  43. Kulkarni, U.; Gurlahosur, S.V.; Babar, P.; Muttagi, S.I.; Soumya, N.; Jadekar, P.A.; Meena, S.M. Facial Key points Detection using MobileNetV2 Architecture. In Proceedings of the 2023 IEEE 8th International Conference for Convergence in Technology (I2CT), Lonavla, India, 7–9 April 2023; pp. 1–6. [Google Scholar]
  44. Guo, M.; Xu, T.; Liu, J.; Liu, Z.; Jiang, P.; Mu, T.; Zhang, S.; Martin, R.R.; Cheng, M.; Hu, S. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  45. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  46. Ning, X.; Tian, W.J.; Yu, L.N.; Li, W. A Brain-Inspired CIRA-DETR Full Inference Method for Small and Occluded Object Detection. J. Comput. Sci. 2022, 45, 2080–2092. [Google Scholar]
  47. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  48. Zong, Z.; Song, G.; Liu, Y. Detrs with collaborative hybrid assignments training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6748–6758. [Google Scholar]
Figure 1. Fitting object detection problems (Objects are indicated by green boxes).
Figure 1. Fitting object detection problems (Objects are indicated by green boxes).
Sensors 24 08207 g001
Figure 2. Comparison chart of scene meanings.
Figure 2. Comparison chart of scene meanings.
Sensors 24 08207 g002
Figure 3. Scene definition diagram.
Figure 3. Scene definition diagram.
Sensors 24 08207 g003
Figure 4. SKIN model structure.
Figure 4. SKIN model structure.
Sensors 24 08207 g004
Figure 5. Scene filtering module structure diagram.
Figure 5. Scene filtering module structure diagram.
Sensors 24 08207 g005
Figure 6. Scene-fitting co-existence matrix.
Figure 6. Scene-fitting co-existence matrix.
Sensors 24 08207 g006
Figure 7. The network structure of SSIM.
Figure 7. The network structure of SSIM.
Sensors 24 08207 g007
Figure 8. Qualitative result comparison on fitting dataset.
Figure 8. Qualitative result comparison on fitting dataset.
Sensors 24 08207 g008
Figure 9. More test results.
Figure 9. More test results.
Sensors 24 08207 g009
Table 1. Fitting dataset.
Table 1. Fitting dataset.
Fitting
Name
Training SubsetTesting SubsetTotal
per Object
#Images#Objects#Images#Objects
PT56982750148
BT49717351504632198
CT254923321101033
WT24621242104
HB82538001465774377
UT70727671383573124
YP79415311612641795
PG5564202488
SH265924942601184
SP2895364264600
GR438701101153854
SR38195943971056
WE2462797783362
AB5061979662232202
total133016,358318276719,125
Table 2. Comparison results with other state-of-the-art models (Bold indicates the best value in the column).
Table 2. Comparison results with other state-of-the-art models (Bold indicates the best value in the column).
ModelsmAP50PTBTCTWTHBUTYPPGSHSPGRSRWEABTimems/i
SSD30051.478.385.540.811.029.223.258.35.082.969.592.671.597.754.28
SSD51274.391.290.153.941.963.659.274.951.890.474.892.976.499.679.236
RetinaNet69.881.293.244.971.360.058.368.97.788.174.091.462.999.376.650
YOLOv571.386.773.760.877.355.768.463.742.388.878.490.754.897.658.633
YOLOv875.485.777.469.674.960.374.571.258.690.763.991.679.598.659.7127
R-FCN67.076.335.459.373.357.648.778.452.772.562.487.769.994.768.4230
Efficientnetv268.747.564.769.274.758.943.859.756.187.669.390.677.893.068.720
Mobilenetv259.448.564.545.362.837.729.657.850.368.770.869.459.693.573.65
Swin Transformer75.287.479.486.573.469.976.374.232.989.876.376.151.299.880.2214
DETR72.674.773.662.867.573.863.673.744.274.387.595.269.797.657.7145
DINO75.891.287.55878.2637281.5348087.585.57290.579.8210
CO-DETR75.590.87357.577.567.57180.533.585.58783.271.59979.2175
Baseline71.481.689.256.064.749.649.678.833.381.186.489.762.710076.9158
Ours76.391.093.858.679.048.752.582.334.890.888.296.172.610080.4193
Table 3. Ablation experiments with modules.
Table 3. Ablation experiments with modules.
%SFMSSIMAP50−95AP50AR1AR100
Baseline 38.473.626.546.8
41.5+3.176.9+3.327.5+1.049.6+2.8
Ours42.0+3.678.4+4.827.4+0.949.9+3.1
Table 4. Results of different prior knowledge (Bold indicates the best value in the column).
Table 4. Results of different prior knowledge (Bold indicates the best value in the column).
Different MatrixAP50−95AP50AP75AR1AR100
Ones Prior Matrix41.176.440.427.349.7
Random Prior Matrix41.176.740.727.649.6
Scene-Fitting Prior Matrix42.078.441.227.449.9
Table 5. Results of different N μ values (Bold indicates the best value in the column).
Table 5. Results of different N μ values (Bold indicates the best value in the column).
ExperimentsAP50:95AP50AP75AR1AR10AR100
N μ = 8 41.676.941.327.549.449.6
N μ = 16 41.977.341.427.449.849.9
N μ = 24 41.677.241.027.149.249.4
N μ = 32 41.877.940.927.449.449.5
N μ = 40 42.078.441.227.449.849.9
N μ = 48 41.177.741.327.749.349.4
N μ = 56 41.777.642.727.349.549.7
N μ = 64 42.077.441.328.049.950.0
N μ = 128 41.277.140.027.749.349.5
Table 6. Results of different features C e (Bold indicates the best value in the column).
Table 6. Results of different features C e (Bold indicates the best value in the column).
ExperimentsAP50:95AP50AP75AR1AR10AR100
C e = 128 40.776.239.927.449.649.7
C e = 256 42.078.441.227.449.849.9
C e = 512 41.177.241.027.049.249.3
C e = 1024 41.177.340.427.449.249.3
C e = 2048 42.077.042.527.950.150.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, X.; Xu, X.; Xu, J.; Zheng, W.; Wang, Q. A Scene Knowledge Integrating Network for Transmission Line Multi-Fitting Detection. Sensors 2024, 24, 8207. https://doi.org/10.3390/s24248207

AMA Style

Chen X, Xu X, Xu J, Zheng W, Wang Q. A Scene Knowledge Integrating Network for Transmission Line Multi-Fitting Detection. Sensors. 2024; 24(24):8207. https://doi.org/10.3390/s24248207

Chicago/Turabian Style

Chen, Xinhang, Xinsheng Xu, Jing Xu, Wenjie Zheng, and Qianming Wang. 2024. "A Scene Knowledge Integrating Network for Transmission Line Multi-Fitting Detection" Sensors 24, no. 24: 8207. https://doi.org/10.3390/s24248207

APA Style

Chen, X., Xu, X., Xu, J., Zheng, W., & Wang, Q. (2024). A Scene Knowledge Integrating Network for Transmission Line Multi-Fitting Detection. Sensors, 24(24), 8207. https://doi.org/10.3390/s24248207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop