Next Article in Journal
A Fragile Image Watermarking Scheme in DWT Domain Using Chaotic Sequences and Error-Correcting Codes
Previous Article in Journal
Artificial Intelligence Techniques Used to Extract Relevant Information from Complex Social Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ST-CenterNet: Small Target Detection Algorithm with Adaptive Data Enhancement

College of Computer Science, Guangdong Polytechnic Normal University, Guangzhou 510665, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(3), 509; https://doi.org/10.3390/e25030509
Submission received: 12 February 2023 / Revised: 13 March 2023 / Accepted: 14 March 2023 / Published: 16 March 2023
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
General target detection with deep learning has made tremendous strides in the past few years. However, small target detection sometimes is associated with insufficient sample size and difficulty in extracting complete feature information. For safety during autonomous driving, remote signs and pedestrians need to be detected from driving scenes photographed by car cameras. In the early period of a medical lesion, because of the small area of the lesion, target detection is of great significance to detect masses and tumors for accurate diagnosis and treatment. To deal with these problems, we propose a novel deep learning model, named CenterNet for small targets (ST-CenterNet). First of all, due to the lack of visual information on small targets in the dataset, we extracted less discriminative features. To overcome this shortcoming, the proposed selective small target replication algorithm (SSTRA) was used to realize increasing numbers of small targets by selectively oversampling them. In addition, the difficulty of extracting shallow semantic information for small targets results in incomplete target feature information. Consequently, we developed a target adaptation feature extraction module (TAFEM), which was used to conduct bottom-up and top-down bidirectional feature extraction by combining ResNet with the adaptive feature pyramid network (AFPN). The improved new network model, AFPN, was added to solve the problem of the original feature extraction module, which can only extract the last layer of the feature information. The experimental results demonstrate that the proposed method can accurately detect the small-scale image of distributed targets and simultaneously, at the pixel level, classify whether a subject is wearing a safety helmet. Compared with the detection effect of the original algorithm on the safety helmet wearing dataset (SHWD), we achieved mean average precision (mAP) of 89.06% and frames per second (FPS) of 28.96, an improvement of 18.08% mAP over the previous method.

1. Introduction

The current progression in the fields of deep learning (DL), image processing, and computer vision (CV) technologies has changed the thinking about different features of day-to-day living [1]. The DL method has given a strong foundation for target detection with consistent accuracy [2]. Target detection has become one of the most essential topics in the CV community, requiring object classification and localization [3]. It has a variety of applications including autonomous driving [4], medical lesion detection [5], intelligent security [6], disaster management [7], agriculture surveys [8], urban planning [9], geographic information system updating [10], and many more. Small target detection task scenarios have important application value in various fields. In the field of manufacturing, target detection algorithms are needed to help people find defects [11] and deformation of small parts in the process of part assembly. In intelligent security, it is frequently necessary to detect remotely whether workers are wearing safely helmets correctly, while images of scattered workers present only dozens of pixels, or even fewer pixels. With these difficulties, small target detection demands to be solved for intelligent security.
However, detecting small targets is still a challenging problem. There is still a large performance gap between small and normal scale targets. Taking RetinaNet [12], one of the state-of the-art (SOTA) target detectors, as an example, it achieves 44.10 and 51.20 mAP on targets with medium and large sizes but only obtains 24.10 mAP on small targets on COCO [13] test-dev set. Such degradation is mainly caused by three factors: (1) The inherent scale of the target to be detected in some tasks is small [14], and the detailed information is difficult to find [15]. (2) The long shooting distance [16] leads to the small scale of the target to be detected, and the target occlusion or truncation caused by the movement of the device and jitter [17], the change of the shooting perspective [18], and the different brightness of the screen [19] also further increase the difficulty of small target detection. (3) The receptive field on low-resolution features may not match the size of small targets as pointed out in [20].
The definition of “small target” varies in different scenarios and there is not yet a unified standard. The existing small target definition methods are mainly divided into the two categories of relative scale and absolute scale. Relative scale refers to the definition of small targets from the perspective of the relative proportion of the targets and images [21]. However, these definitions based on relative scale have many drawbacks, which are affected by data preprocessing and model structure, and cannot effectively evaluate the detection performance of models for objects at different scales. The absolute scale refers to the definition of small objects in terms of their absolute pixel size. Since the selected data set only selects the head part of the target person for the box selection of the target box, it contains less information. Considering this factor, we adopt the definition based on absolute scale to determine the small target selection criterion, and define the image less than 32 × 32 as a small target [22].
CenterNet [23] recently has been reported as achieving SOTA performance in classification. Therefore, it is natural to attempt to extend its usage to detection tasks. CenterNet, with a simplified structure and strong expansibility, has a definite impact on the above problems. However, CenterNet has some flaws, such as predicting only through the last feature layer, resulting in the loss of target feature information and ineffectiveness in detecting small targets.
To address the above problems, we present a novel small target detection network with adaptive data enhancement (ST-CenterNet) for accurately increasing target samples directly and enhancing shallow semantic information. First, we design a novel selective small target replication algorithm (SSTRA) method to scale limited small targets. Then, shallow semantic information enhancement is performed by our proposed target adaptation feature extraction module (TAFEM), and this TAFEM uses the residual network ResNet and AFPN successively connect to retain more shallow features. Our main contributions of this paper include:
  • We propose a SSTRA algorithm to select all small targets by pixel filtering. By copying the identified small targets, it can effectively increase the number of small target samples.
  • We propose TAFEM to obtain feature maps of multiple receptive fields [24], by combining ResNet [25] with the adaptive feature pyramid network (AFPN) to retain more complete semantic information of shallow features.
  • A novel structure, AFPN, is proposed to enhance feature extraction and fusion, allowing TAFEM to detect small target regions more accurately.
The rest of this paper is arranged as follows: Section 2 covers the overview of the two main categories of current algorithms for deep learning-based target detection, and then introduces several main improvement directions of the small target detection based on deep learning. Section 3 is dedicated to the dataset and its details. The structure of the ST-CenterNet network described in the study is shown in Section 4, followed by a description of the network’s strategies. Section 5 explains the basic configuration in the experimental environment. Section 6 focuses on the analysis and results of the ablation experiments and comparison experiments using the ST-CenterNet network. The work results are discussed in Section 7, and conclusions are presented in Section 8.

2. Related Work

With the massive progress of convolutional neural networks (CNN), it has become mainstream in target detectors to adopt the modernized state-of-the-art CNN models as feature extractors [26]. The target detection methods are mainly divided into two-stage detection and one-stage detection algorithms. Two-stage detection algorithms, such as Faster R-CNN [27], Mask R-CNN [28], region-based fully convolutional networks (R-FCN) [29], and AAAI [30], use strategies such as selective search (SS) [31] to generate proposals that may contain targets of interest, and then match proposal features with the template. The matched proposals are obtained through classification and regression of proposal regions [32] and treated as the final target positions. One-stage detection algorithms (for example, single shot multibox detector (SSD) [33,34], YOLO series [35,36], RetinaNet, CornerNet [37], CenterNet), can directly obtain target position and category information by inputting images into the convolutional neural network.
Large numbers of detectors in deep learning have been designed for general targets. However, the insufficiency of small target detection and the extraction of feature information based on deep learning lead to unsatisfactory effectiveness. Small target detectability remains to be improved.
Data enhancement methods enhance the value of limited data. Mishra et al. [38] used the method of data transformation to increase the sampling rate of samples, but it may select poor samples and increase its detection error rate. Zhang et al. [39] added two sample-label data pairs proportionally, and generated new sample-label data. Chen et al. [40] proposed a dynamic scale training paradigm, which can dynamically guide data preparation through information feedback from the optimization process, and alleviate the challenge of scale transformation in object detection.
There have been some methods that have achieved some effects through feature enhancement. SSD detects targets with different scales and aspect ratios utilizing a multi-scale feature pyramid. The method uses convolution layers to predict the category fraction and the position offset of bounding boxes. Then, it obtains positioning bounding boxes by a non-maximum suppression post-processing strategy [41]. The module improves accuracy and efficiency to a certain degree but has poor detection performance for small targets. Therefore, improvements have been suggested to enhance the detection performance of SSD for small targets. The improved feature fusion single shot multibox detector (FSSD) [42], introduced by Li et al., added a light-weight feature fusion module based on SSD to generate a new feature pyramid detection model. Subsequently, the SSD was extended to the deconvolutional single shot detector (DSSD) [43], which up-samples low-resolution features, mainly through transposed deconvolution in the decoder [44].
Nonetheless, the deep convolutional networks are prone to lose the key position information of small targets when extracting feature information. Some algorithms determine the location region by the contextual information [45] of targets [46]. Lin et al. [47] introduced a multi-scale fusion strategy of a feature pyramid network (FPN) [48] to extract and fuse features at different scales, and obtained deep semantic information and shallow position information. Chen et al. [49] considered different feature extraction methods based on depth and shallow features to improve the detection effect of small targets. Ma et al. [50] proposed a new model that focuses on learning the deep features produced in the latter part of the network and made full use of the semantic and contextual information expressed by deep features. Huang et al. [51] presented a cross-scale feature fusion method that achieves an enhancement of contextual connection by using cross-scale feature mappings. However, direct cross-layer fusion may lead to a position offset and aliasing effect.
Aiming at these problems, some algorithms have proposed an attention mechanism [52] to improve feature information extraction. Attentional feature fusion (AFF) [53] adds local channel contextual information to the global channel through attentional feature fusion [54], to overcome semantic differences and scale inconsistencies among inputting features. Yu et al. [55] adopted dense connected convolutional networks to improve the ability of feature information extraction and to further enhance the contextual semantic information in shallow features. RFBNet [56] constructed receptive fields and a multi-branch convolution layer. The method used dilated convolution to expand the receptive fields of feature mappings, which improved the detectability of small targets to a certain extent.
Some algorithms replace the sparse coding labels with dense coding labels based on angle classification [57,58]. The arbitrary rotating rectangle is approximated into a two-dimensional Gaussian distribution [59]. Isokinetic rotation features were extracted by combining the isokinetic rotation network with detectors, aiming to solve the problem of multi-angle detection caused by the complex position information of small targets [60]. While these algorithms improve the performance of small target detection to some extent, they do not focus attention on increasing the number of small targets and simultaneously improving the capability to extract feature information. Related improvement methods are summarized in Table 1.
CenterNet is a one-stage target detection model with high accuracy. Target features are extracted from the inputting images by the baseline network and then introduced into the fully convolutional networks (FCN) to obtain a heatmap. Peak points of heatmaps are the centroids of targets. The position and category attributes of targets are obtained by using centroid location regression, which transforms the target detection problem into a key point prediction problem.
Motivated by these studies, we propose the ST-CenterNet model to expand small target samples based on the original dataset and to strengthen the extraction of small target feature information, improving the capability of small target detection.

3. Safety Helmet Wearing Dataset

3.1. Dataset Construction

Experiments are carried out on the open-source safety helmet wearing dataset (SHWD), which is used to provide head detection and safety helmet wearing. Referring to the standards of the PASCAL VOC dataset, an image dataset containing multiple scenes and targets was established, so that the target detection model has detection ability in different scenes, so as to facilitate the model for training, testing and analysis. The training and testing dataset contains 7581 images in which the targets are divided into two categories: hat and person. Among them, including 9044 people wearing helmets correctly (correctly worn, hat) and 111,514 people without helmets (not worn, person).
The images were acquired under complex lighting conditions, and there are complex backgrounds and small targets. Unlike the targets in natural scenes which are often taken from horizontal perspectives, images of the building site in SHWD are typically taken from several positions in diverse directions. The adopted dataset is inundated with low-resolution images with significant amounts of distractors and confusing target orientation. Moreover, our labeled frame only shows the head part of the workers, which further makes it difficult to detect whether a helmet is worn or not. These difficulties suggest that targets in SHWD usually are of diverse sizes and orientations with complex backgrounds, making them an excellent dataset for detection in complicated scenes.
The images in the dataset contain various angles, multiple and single targets, distances, occlusions, etc. The used public dataset is an xml file in PASCAL VOC format and was converted into a txt tag file in the CenterNet format. The dataset was labeled with labelImg for image targets, and the entire dataset was randomly divided into a training set and a test set, with the ratio of 8:2. The number of training set images in the 7581 images dataset is 6064, and the number of test set images is 1517.

3.2. Dataset Processing

The performance of detection results for small targets is weak due to the insufficient number of small targets in SHWD. Therefore, the network model will pay more attention to the training of medium and large targets during training, and ignore small targets, resulting in unsatisfactory detection effect of small targets.
In order to enrich the detection background and increase the number of small targets in SHWD, the proposed SSTRA is used to replicate and flip the small targets in the dataset preprocessing stage. Targets with image resolution less than or equal to 32 × 32 are considered small targets. Figure 1 demonstrates the comparison results of differences between the number of targets without SSTRA processing and the number of targets with SSTRA processing in the dataset. The blue rectangle shows that the SHWD dataset has included 9044 targets into hat, 11,514 targets into person. The orange rectangle shows that the SHWD processed by SSTRA has included 13,839 targets into hat, 112,966 targets into person. It can be found that adding the SSTRA selected processing function effectively optimized the model performance and increased the numbers of hat and person.
To further verify the effects of SSTRA, we represent the visual results of images processed by SSTRA in Figure 2. There are a total of two targets in the first test image in SHWD, including two positive samples (hat). SSTRA achieves the purpose of enriching training samples by resampling and flipping small targets. After the SSTRA processing mechanism, there are a total of six samples in the first image, including six positive samples (hat). It can be seen from the above that, the SSTRA processing mechanism can effectively increase the number of samples in the SHWD dataset.
In the experiment, our SSTRA detects and replicates only the header information of objects, because the specific category can be judged only through the detection of header information in the experimental detection stage. Our proposed SSTRA is implemented through the real box annotated by the dataset, then by finding the small target according to the size of the targets, and finally by copying the bounding box of the small target. Since the dataset is labeled with the header, subsequent operations are processed against the target header.

4. ST-CenterNet

4.1. Network Structure

Our proposed ST-CenterNet algorithm samples an inputting image of uniform size by the feature extraction network. The feature information extraction network subsequently extracts the target feature information. After that, the extracted information is passed into FCN. The target centroid, the width and height of targets, and the offset value of the centroid for heatmap prediction are obtained. The position, size and category of targets are obtained by centroid regression.
The SSTRA is proposed to address the shortage of the number of small target samples in the dataset. Targets with pixel values less than or equal to the number in the images are selected and considered as small targets. Then, all small targets are replicated and flipped to achieve the oversampling of small targets, thus increasing the sample size for them.
Utilization of low-level features is one way to pick up information about small targets. The backbone network uses ResNet-50 to down-sample the image and combines it with AFPN to obtain feature maps of different sizes. This allows more complete extraction of feature maps with strong deep semantic information and shallow position information during inference processing.
Compared with the conventional size targets, small targets have fewer available pixels, which makes it difficult to extract complete feature information. With the increase of network layers, their semantic information and position information are gradually lost and become difficult for the network to detect. The features cannot provide effective expressions of semantic information. To deal with that, multi-scale learning is used, as a strategy to effectively integrate two types of feature information and perform better semantic representation. According to [43,47], it is appropriate to use separate groups of features to model distinct factors. One concern is that the shallow feature information required by small target detection can be easily diluted in the extraction process. To prevent this, we propose the TAFEM. The combination of ResNet-50 and AFPN is adopted to further enhance the performance of extracting deep semantic feature information. The enhanced feature information allows a better fusion of deep and shallow feature information to obtain more complete small target feature information. Subsequently, the outputting fusion feature is used for prediction, which is divided into three parts for generating a heatmap of key points, scale prediction of the bounding box, and offset of central points. Finally, the position of the target is estimated by the predicted center point coordinates. The network structure diagram of ST-CenterNet is shown in Figure 3.
In the proposed algorithm, the loss function is divided into the following three parts: the loss of the heatmap, the loss of the width and height of the bounding box, and the offset loss of the central key point. The loss function is formulated as follows:
L S H L = L H + λ S I Z E L S I Z E + λ O F F L O F F
The notations L H , L W H , L O F F denote the loss value of the heatmap, the loss value of the width and height of the bounding boxes, and the offset loss value of the center point, respectively. We set λ W H as 0.1 and set λ O F F as 1 unless specified otherwise.
When calculating the loss value of the heatmap, the idea of focal loss [37] is used for reference. Focal loss is mainly to solve the problem of imbalanced classification in target detection. For the samples that may complete the classification, we take appropriate measures to reduce the proportion of training. We use L1 loss at the length and the width of bounding boxes. Spatial resolution of the feature map output by the backbone network becomes one-quarter of the original inputting image. It is equivalent to a pixel in the outputting feature image corresponding to the 4 × 4 region of the original image. Therefore, it is essential to introduce offset loss. We adopt the L1 loss function to predict the centroid offset loss for an improvement of the accuracy of centroid coordinate prediction.

4.2. Selective Small Target Replication Algorithm

The number of small target samples in SHWD is less than that of general targets. This leads us to pay more attention to the training of large and medium-sized targets, neglecting the training of small targets when using the network model for training. Because of the small size of small targets or the small proportion in the image, the position diversity of small targets in the image is insufficient.
Considering the above two problems, we utilize SSTRA to selectively oversample all targets in the image and to achieve separate replication of small targets to enrich their sample size. Extracted targets are used to obtain targets with pixel values less than or equal to 32 × 32 against the background of a construction site by screening. These are considered small targets. All of the original images that contain small targets are determined and then individually extracted. The extracted small targets are replicated twice and horizontally flipped at the second copy, and then are put back into the corresponding original images according to initial sizes of extraction. Related information about these small targets is added to the extensible markup language file corresponding to the image. The architecture of SSTRA is demonstrated in Figure 4.
Our SSTRA focuses on two replications of small targets, which individually match pixel value screening. Then, there is repeated sampling processing of small targets to obtain a larger number of samples and to further increase the diversity of positions to improve the detection performance of small targets.

4.3. Target Adaptation Feature Extraction Module

The detection task requires models to identify targets on more scales than the classification task requires, in order to preserve information from various layers. This is especially important for detection because each layer has different receptive fields. CenterNet [23] makes predictions based only on the last feature layer, which disregards details of the shallow features, resulting in poor performance in detecting small targets. As shown in Figure 5, our TAFEM uses ResNet-50 to down-sample images. The performance of the detector is degraded because of couplings of features at different scales and the mismatch between pyramidal layers and target sizes. The AFPN mainly adopts decoupling of detection of different sizes. In particular, it can be used for small targets with high-resolution feature mapping; it allocates more appropriate feature levels for them. It assigns small targets to the corresponding layer to obtain more comprehensive region information.
The AFPN constructs a four-layer feature pyramid by up-sampling the deep convolution feature map and fusing this with the shallow features through horizontal connection. The details of TAFEM are shown below.
For the network structure of the deeper layers, the parameters are generally set close to zero during initialization. This leads to problems such as gradient explosion as the network layer deepens during training to update the parameters of the shallow-layer network. The result may make it impossible to update the parameters of the shallow-level network. We set up a linear combination of the network layers based on the original network structure. With an increased number of network layers, the linear combination for the mapping is as follows:
W = φ γ , L i + α γ
The notation W denotes the output vector processed by linear combination of feature layers. The notation φ γ , L i denotes multiple convolutional layers, where φ is the sigmoid function, γ is the input vector of the feature layer, and L i is the weight layer. When both φ and γ have the same dimension, we perform an element-wise addition of the inputs and outputs between layers. The square matrix α is used for concrete linear operations. We set α as 1. However, when they have a different dimension, we perform a linear projection before addition.
The feature information of each predicted feature map is enhanced by boosting each feature map to the same number of channels by convolution with a convolution kernel size of 1. It is then fused backward and forward with the shallow features by lateral connections, which are similar to the jump connections in the residual structure. The features of each layer obtained after independent up and down-sampling are passed to an activation gate consisting of a combination of multiple convolutions and sigmoid activation functions. Thus, generated lateral features are used to fuse location features with semantic feature information:
A F = β S i P i , f , s
where β denotes the sigmoid function, S i denotes the convolution corresponding to the i t h layer, P i denotes the characteristic layer obtained by sampling corresponding to the i t h layer, and s = 2 i 1 denotes the convolution stride.
Deep and adjacent feature layers are added and element-wise multiplication fusion processing is conducted. Convolution is used to eliminate the aliasing effect of the up-sampling and to generate a new feature map according to the sizes of the targets:
P i + 1 = P i + l o g 2 w h 512
The notations P i represent each pyramid feature layer corresponding to the i t h layer; w and h are the width and the height of the bounding box of targets, respectively; and 512 represents the uniform size of the input image.
After the convolutional layers, the three feature maps (in Figure 3) are flattened into vectors and then concatenated together for recognition. By combining low-level, high-resolution information with high-level, strongly semantic information, AFPN allocates feature maps of different levels according to the targets at different sizes. It adapts to various receptive fields and improves the feature extraction ability of targets.

5. Experimental Settings

5.1. Experimental Configuration

In order to conduct model training and performance testing, we set up the experimental environment configuration. Detailed experimental settings are presented in Table 2.

5.2. Parameter Settings

We adopt a stochastic gradient descent (SGD) algorithm for model training to update and optimize the weight of the network model. The model training uses synchronous batch normalization [61]. According to the training parameters with good models and the memory capacity possessed, the number of iterations is set to 200. Model parameters are set to batch size of 10, weight decay of 5 × 10−4, momentum factor of 0.9, and a step learning rate decay [62]. In the training of the CONVOLUTIONAL neural network, the learning rate is set to be constant at 1 × 10−4. The updated formula of the weighting is as follows:
W i + 1 = W i ε η
The notations W i , W i + 1 denote the immediate parameter, and the updated parameter, respectively. The notation ε is influenced by an intermittently increasing parameter, which indicates the number of times that the performance of the tolerant network is without improvement. We set the intermittently increasing parameter as 2, and take the learning decay rate ε as 0.5. η denotes the learning rate.
To speed up the training, the backbone of the proposed model is initialized from a checkpoint pre-trained by ResNet-50 with self-training [63]. All other results are from models with random initialization, unless otherwise stated. We carry out sample shuffling and divide the dataset into 10 parts. The images are assigned to the training set and verification set according to the ratio of 8:2. During the training, the backbone network ResNet-50 that we used has pre-training weights. Some of the pre-training weights applied to the network are generic. It should be stressed that the backbone network is frozen first and more resources are placed in the network parameters in the later part of the training. This makes the training procedure more convenient to apply and brings better performance over time and resource utilization. After a period of network parameter training, the frozen part is thawed and all are trained together. In the frozen stage, we freeze the main stem of the model. The feature extraction network does not change, resulting in a small amount of video memory being occupied. The network is just fine-tuned at this time. In the unfrozen stage, the backbone of the model is not frozen. This changes the feature extraction network and takes up more video memory, resulting in a changeableness of parameters in the network. We take the iteration number of the frozen part as 100 times. Similarly, the iteration number of the unfrozen part is 100 times.

5.3. Evaluation Metrics

According to the standard protocol defined by the public dataset, we adopt the mean average precision (mAP), precision, recall, and f1-score to evaluate the detection accuracy of the algorithm in SHWD. We use frames per second (FPS) as the evaluation index of detection over speed. Because we expect to detect whether the remotely operating workers are wearing their helmets correctly, we also use the evaluation metric in pedestrian detection, log-average miss rate (MR−2), to evaluate the performance of the proposed detector.

6. Experimental Results

6.1. Comparative Experiment and Analysis of Results

When training and verifying the ST-CenterNet model, the change curve of the loss function was drawn through the training and verifying results information of each round, as shown in Figure 6. The change of the loss function of the CenterNet model is represented by the left bar. The change of the loss function of the ST-CenterNet model is represented by the right bar. It can be seen from Figure 6 that the initial loss value of the ST-CenterNet model was smaller than the initial loss value of the CenterNet model.
Most images in the dataset are small targets with disordered directions, complex backgrounds, and different scales. Specifically, Table 3 compares the detection performance of various target detection algorithms with the proposed algorithm on SHWD. The same experimental environment and dataset were used for comparative experiments. The experimental data of the algorithm in the following table are based on the enhancement of the algorithm using SSTRA. The experimental results on the safety helmet wearing dataset show that the proposed method has reasonable detection results compared with other state-of-the-art methods. As can be seen in Table 3, the highest detection accuracy of CenterNet algorithm is 70.98%, and the highest detection accuracy of the proposed ST-CenterNet is 89.06%, which is an improvement of 18.08% in detection accuracy.
The experimental results of ST-CenterNet show that adding data enhancement and feature enhancement modules can improve the accuracy of CenterNet’s target detection. However, this will inevitably increase the amount of calculation and model parameters, resulting in a decrease in the detection speed.
To more intuitively verify the target detection effect and model robustness of the proposed ST-CenterNet model, we present the visual results of CenterNet and ST-CenterNet for pairwise comparison in Figure 7. As shown in Figure 7, this demonstrates the comparison of the qualitative detection effect between the original algorithm and the proposed algorithm on SHWD. The selected images (in Figure 7) are the portion of the dataset that contains small targets that are difficult to detect. The left side of the two contrasted images shows the detection results obtained by the original algorithm, and the right side shows the detection results obtained by the proposed algorithm.
There are a total of six targets in the first test image, including six positive samples (hat). The CenterNet model detected a total of four targets, two missed detections, and one false detection. The ST-CenterNet model correctly detected a total of six targets, zero missed detections, and zero false detection. It can be seen from the above that, in the CenterNet model, there were serious false detection and missed detections for small targets and occluded targets. Compared with the original model, the qualitative detection results (in Figure 7) illustrate that, compared with CenterNet, visual results obtain a similar but significantly better performance for both SSTRA and TAFEM. The small targets that cause disorientation and are difficult to detect can be detected more accurately, which further proves that the adopted method effectively improves the detection performance of the model for them.

6.2. Ablation Experiment and Analysis of Results

According to the different improvement schemes, we conducted ablation experiments to explore the effects of the SSTRA module and TAFEM module on the model performance. The results of the ablation experiments are shown in Table 4.
In Table 4, the first line shows that CenterNet, the original algorithm, uses the original dataset. The second line ‘CenterNet+SSTRA’ illustrates the data enhancement using our SSTRA based on the original algorithm in the first line ‘CenterNet’. In the third line ‘CenterNet+TAFEM’, the feature extraction enhancement module TAFEM is added based on the original algorithm in the first line. In the fourth line ‘ST-CenterNet’, the TAFEM module is used for feature enhancement while SSTRA is used for data enhancement. Therefore, it can be seen from the table that SSTRA and TAFEM can each improve certain aspects of detection performance, and the combined effectiveness of SSTRA and TAFEM has more significant detection performance.
As can be seen from Table 4, the detection effect is 13.90% better than the original method on mAP. This indicates that the method of increasing the number of small targets, which replicates small targets, improves the target detection results. The TAFEM module makes deep and shallow features fusion, which introduces more accurate positioning information for each prediction layer. Compared with the original algorithm, the detection effect on mAP is improved by 13.22%, which proves that the proposed module can effectively fuse shallow details and deep semantic information. The performance of contrast images enhances the semantic information and feature expression ability of shallow feature images and improves the result of target detection. AFPN can focus on targets with specific sizes rather than wide backgrounds, effectively improving the detection performance. By adding the two proposed modules simultaneously, the detection effect of the proposed algorithm obtains a significant enhancement of 18.08% on mAP, compared to that of the original algorithm.
Four groups of models were trained, respectively, and each module was successively added on the basis of the CenterNet model. Four groups of models were tested in turn on the same test set, and the AP curves of hat and person detection are shown in Figure 8.
Compared with the results of “CenterNet”, the AP values of the “CenterNet+SSTRA” and “CenterNet+TAFEM” model were slightly improved. It was found that adding the SSTRA selected processing function effectively optimized the model performance and improved the detection accuracy of the model by increasing the number of small targets. The extraction of effective feature information and multi-scale feature fusion enabled the model to better detect targets. It can be found that adding the TAFEM feature enhancement function optimized the feature extraction efficiency and improved the detection performance. Compared with the “CenterNet+SSTRA” model, the AP of the “ST-CenterNet” was increased by 3.12% and 5.08%, respectively. Compared with the “CenterNet+TAFEM” model, the AP of the latter was increased by 3.22% and 6.36%, respectively. On the whole, each module in the proposed ST-CenterNet model effectively improved the detection accuracy, met the actual detection requirements, and verified the feasibility of the model.

6.3. Performance Validation

The confusion matrix is a situation analysis table that summarizes the prediction results of the classification model in deep learning. It summarizes the records in the data set in the form of a matrix according to the true category and the category judgment standard predicted by the classification model. The confusion matrix generated by the CenterNet and ST-CenterNet technique on the classification of hat and person under 200 epochs is in Figure 9. The figure (a) shows that the CenterNet model has categorized 1437 targets into hat, 11,293 targets into person. The figure (b) shows that the ST-CenterNet model has categorized 1645 targets into hat, 21,206 targets into person. It can be found that the proposed ST-CenterNet model effectively increases the number of correct recognitions of hat and person.
In addition, we adopt the log-average miss rate (MR−2) to evaluate the performance of the proposed detector. The purpose is mainly to verify the effectiveness of the proposed model in reducing the proportion of missed detection of the detection results. MR−2 represents the miss rate of the model, and the pedestrian detector performance is measured by the curve of the average miss rate versus False Positives Per Image (FPPI) as the abscissa, and Log (MR) as the ordinate. The expected result is a reduction of MR−2. As can be seen in Figure 10, the proposed method shows a reduction in MR−2 for both types of targets detected (hat and person), when compared to the original model.

7. Discussion

The experimental results on SHWD show that the mAP index of the model is improved by 13.9% after replicating and flipping small targets through the SSTRA. Besides, the TAFEM considerably improves the ability of feature extraction, and the mAP of the model is increased by 13.22% compared with the original model. Compared with the CenterNet, the mAP index of the proposed algorithm is increased by 18.08%, which proves the superiority of ST-CenterNet in small target detection. SOTA performance on various methods demonstrates the superiority of ST-CenterNet in small target detection. SSTRA and AFPN can be combined with diverse detectors and various backbones to strengthen small target detection. This ability is transferrable to more specific situations. During the experiment, we found that there are still some missing targets that have not been detected, so the future improvement needs to further strengthen the capture of shallow feature information. We may consider trying to add learning to strengthen the correlation between parts using a self-attention mechanism.

8. Conclusions

Our proposed algorithm for small target detection with adaptive data enhancement aims to solve the lack of shallow feature information extraction and insufficient numbers of small targets. SSTRA utilizes a screening and oversampling method, which brings great benefits to increasing the number of samples and to detection effectiveness, while maintaining the total quantity of images in the dataset. TAFEM adopts the combination of ResNet-50 and AFPN to enhance the performance of extracting shallow semantic information, so as to obtain more complete semantic feature information. For future work, we are committed to using a more lightweight network to reduce the computational complexity of the model parameters. We intend to study the combination with insights from the recently proposed transformer model to accurately deepen the self-attention mechanism and more perfectly explore the relationship between different levels of feature maps.

Author Contributions

Conceptualization, X.L. and Y.G.; methodology, Y.G.; validation, Y.G.; formal analysis, X.L. and Y.G.; investigation, Y.G.; resources, X.L.; data curation, Y.G.; writing—original draft preparation, Y.G.; writing—review and editing, X.L.; visualization, Y.G.; supervision, X.L.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China (62176067); Scientific and Technological Planning Project of Guangzhou (201903010041, 202103000040); Key Project of Guangdong Province Basic Research Foundation (2020B1515120095); Project Supported by Guangdong Province Universities and Colleges Pearl River Scholar Funded Scheme (2019).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The datasets analyzed during the current study are available in the Github repository, GitHub—tanypadam/Safety-Helmet-Wearing-Dataset: Safety helmet wearing detect dataset, with pretrained model.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiang, Q.; Tan, D.; Li, Y.; Ji, S.; Cai, C.; Zheng, Q. Object detection and classification of metal polishing shaft surface defects based on convolutional neural network deep learning. Appl. Sci. 2020, 10, 87. [Google Scholar] [CrossRef] [Green Version]
  2. Vaidya, B.; Paunwala, C. Deep learning architectures for object detection and classification. In Smart Techniques for a Smarter Planet; Springer: Cham, Switzerland, 2019; pp. 53–79. [Google Scholar]
  3. Peng, S.; Jiang, W.; Pi, H.; Li, X.; Bao, H.; Zhou, X. Deep snake for real-time instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8533–8542. [Google Scholar]
  4. Akyol, G.; Kantarcı, A.; Çelik, A.E.; Ak, A.C. Deep learning based, real-time object detection for autonomous driving. In Proceedings of the 2020 28th Signal Processing and Communications Applications Conference (SIU), Gaziantep, Turkey, 5–7 October 2020; pp. 1–4. [Google Scholar]
  5. Luan, B.; Sun, Y.; Tong, C.; Liu, Y.; Liu, H. R-FCN based laryngeal lesion detection. In Proceedings of the 2019 12th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 14–15 December 2019; pp. 128–131. [Google Scholar]
  6. Li, J.; Liu, H.; Wang, T.; Jiang, M.; Wang, S.; Li, K.; Zhao, X. Safety helmet wearing detection based on image processing and deep learning. In Proceedings of the 2020 International Conference on Communications, Information System and Computer Engineering (CISCE), Kuala Lumpur, Malaysia, 3–5 July 2020; pp. 343–347. [Google Scholar]
  7. Kaku, K. Satellite remote sensing for disaster management support: A holistic and staged approach based on case studies in Sentinel Asia. Int. J. Disaster Risk Reduct. 2019, 33, 417–432. [Google Scholar] [CrossRef]
  8. Chen, P.-C.; Chiang, Y.-C.; Weng, P.-Y. Imaging using unmanned aerial vehicles for agriculture land use classification. Agriculture 2020, 10, 416. [Google Scholar] [CrossRef]
  9. Wellmann, T.; Lausch, A.; Andersson, E.; Knapp, S.; Cortinovis, C.; Jache, J.; Scheuer, S.; Kremer, P.; Mascarenhas, A.; Kraemer, R.; et al. Remote sensing in urban planning: Contributions towards ecologically sound policies. Landsc. Urban Plan. 2020, 204, 103921. [Google Scholar] [CrossRef]
  10. Al-Quraishi, A.M.F.; Negm, A.M. Updates, conclusions, and recommendations for environmental remote sensing and GIS in Iraq. In Environmental Remote Sensing and GIS in Iraq; Springer: New York, NY, USA, 2020; pp. 517–529. [Google Scholar]
  11. Qiu, Z.; Wang, S.; Zeng, Z.; Yu, D. Automatic visual defects inspection of wind turbine blades via YOLO-based small object detection approach. J. Electron. Imaging 2019, 28, 043023. [Google Scholar] [CrossRef]
  12. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  13. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  14. Hu, G.X.; Yang, Z.; Hu, L.; Huang, L.; Han, J.M. Small object detection with multiscale features. Int. J. Digit. Multimed. Broadcast. 2018, 2018, 4546896. [Google Scholar] [CrossRef]
  15. Meng, Q.; Song, H.; Li, G.; Zhang, Y.; Zhang, X. A block object detection method based on feature fusion networks for autonomous vehicle. Complexity 2019, 2019, 4042624. [Google Scholar] [CrossRef]
  16. Zhao, X.; Pu, F.; Wang, Z.; Chen, H.; Xu, Z. Detection, tracking, and geolocation of moving vehicle from uav using monocular camera. IEEE Access 2019, 7, 101160–101170. [Google Scholar] [CrossRef]
  17. Zhou, T.; Li, Z.; Zhang, C. Enhance the recognition ability to occlusions and small objects with Robust Faster R-CNN. Int. J. Mach. Learn. Cybern. 2019, 10, 3155–3166. [Google Scholar] [CrossRef]
  18. Guo, H.; Wang, J.; Xu, M.; Zha, Z.-J.; Lu, H. Learning multi-view deep features for small object retrieval in surveillance scenarios. In Proceedings of the 23rd ACM international conference on Multimedia, Ottawa, ON, Canada, 23–26 June 2015; pp. 859–862. [Google Scholar]
  19. Liu, Y.; Yang, F.; Hu, P. Small-object detection in UAV-captured images via multi-branch parallel feature pyramid networks. IEEE Access 2020, 8, 145740–145750. [Google Scholar] [CrossRef]
  20. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z. Scale-aware trident networks for object detection. In Proceedings of the International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  21. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z. R-CNN for small object detection. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; Springer: Cham, Switzerland, 2016; pp. 214–230. [Google Scholar]
  22. Cheng, G.; Yuan, X.; Yao, X.; Yan, K.; Zeng, Q.; Han, J. Towards Large-Scale Small Object Detection: Survey and Benchmarks. arXiv 2022, arXiv:2207.14096. [Google Scholar]
  23. Cheng, G.; Yuan, X.; Yao, X.; Yan, K.; Zeng, Q.; Han, J. Objects as points. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 7263–7271. [Google Scholar]
  24. Miao, S.; Du, S.; Feng, R.; Zhang, Y.; Li, H.; Liu, T.; Zheng, L.; Fan, W. Balanced single-shot object detection using cross-context attention-guided network. Pattern Recognit. 2022, 122, 108258. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June –1 July 2016; pp. 770–778. [Google Scholar]
  26. Lee, Y.; Hwang, J.W.; Lee, S.; Bae, Y.; Park, J. An energy and GPU-computation efficient backbone network for real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 752–760. [Google Scholar]
  27. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  28. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  29. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object detection via region-based fully convolutional networks. In Proceedings of the Advance in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  30. Ming, Q.; Zhou, Z.; Miao, L.; Zhang, H.; Li, L. Dynamic anchor learning for arbitrary-oriented object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; pp. 2355–2363. [Google Scholar]
  31. Yin, S.; Li, H. Hot region selection based on selective search and modified fuzzy C-Means in remote sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5862–5871. [Google Scholar] [CrossRef]
  32. Dong, P.; Wang, W. Better region proposals for pedestrian detection with R-CNN. In Proceedings of the IEEE Vision Communication on Image Processing, Chengdu, China, 27–30 November 2016; pp. 1–4. [Google Scholar]
  33. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single shot multi-box detector. In Proceedings of the European Conference Computer Vision, Amsterdam, The Netherlands, Guangzhou, China, 11–14 October 2016; pp. 21–37. [Google Scholar]
  34. Liu, Z.; Fang, W.; Sun, J. SSD small object detection algorithm based on feature enhancement and sample selection. In Proceedings of the International Symposium on Distributed Computing and Applications for Business Engineering and Science, Nanning, China, 10–12 December 2021; pp. 96–99. [Google Scholar]
  35. Wu, T.H.; Wang, T.W.; Liu, Y.Q. Real-time vehicle and distance detection based on improved Yolov5 network. In Proceedings of the 2021 3rd World Symposium on Artificial Intelligence (WSAI), Guangzhou, China, 18–20 June 2021; pp. 24–28. [Google Scholar]
  36. Ge, Z.; Liu, S.; Wang, F. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  37. Law, H.; Deng, J. CornerNet: Detecting objects as paired keypoints. In Proceedings of the European Conference Computer Vision, Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  38. Mishra, S.; Shah, A.; Bansal, A.; Jagannatha, A.; Sharma, A.; Jacobs, D.; Krishnan, D. Object-aware cropping for self supervised learning. arXiv 2021, arXiv:2112.00319. [Google Scholar]
  39. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  40. Chen, Y.; Zhang, P.; Li, Z.; Li, Y.; Zhang, X.; Qi, L.; Sun, J.; Jia, J. Dynamic Scale Training for Object Detection. arXiv 2020, arXiv:2004.12432. [Google Scholar]
  41. Liu, G.; Wang, C.; Hu, Y. RPN with the attention-based multi-scale method and the adaptive non-maximum suppression for billboard detection. In Proceedings of the IEEE International Conference Computer Communication, Hangzhou, China, 15–19 April 2018; pp. 1541–1545. [Google Scholar]
  42. Belfodil, A.; Belfodil, A.; Bendimerad, A.; Lamarre, P.; Robardet, C.; Kaytoue, M.; Plantevit, M. FSSD-A Fast and Efficient Algorithm for Subgroup Set Discovery. In Proceedings of the 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Washington, DC, USA, 5–8 October 2019; pp. 91–99. [Google Scholar]
  43. Zhang, Y.; Zhou, W.; Wang, Y.; Xu, L. A real-time recognition method of static gesture based on DSSD. Multimed. Tools Appl. 2020, 79, 17445–17461. [Google Scholar] [CrossRef]
  44. Wu, M.; Huang, H. Small-Scale Aware SSD: Improvement for Small-scale Pedestrians Detection. In Proceedings of the International Conference on Electronic Communication and Artificial Intelligence (IWECAI), Zhuhai, China, 14–16 January 2022; pp. 419–424. [Google Scholar]
  45. Zhou, Q.; Wu, X.; Zhang, S.; Kang, B.; Ge, Z.; Latecki, L.J. Contextual ensemble network for semantic segmentation. Pattern Recognit. 2022, 122, 0031–3203. [Google Scholar] [CrossRef]
  46. Kaya, E.C.; Alatan, A.A. Improving proposal-based object detection using convolutional context features. In Proceedings of the IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018; pp. 1308–1312. [Google Scholar]
  47. Van, Q.T.; Young, K.M. Feature pyramid network with multi-scale prediction fusion for real-time semantic segmentation. Neurocomputing 2023, 519, 104–113. [Google Scholar]
  48. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  49. Chen, H.J.; Wang, Q.; Yang, G.; Han, J.; Yin, C.; Chen, J.; Wang, Y. SSD object detection algorithm with multi-scale convolution feature fusion. J. Front. Comput. Sci. Technol. 2019, 13, 1049–1061. [Google Scholar]
  50. Ma, W.; Wu, Y.; Cen, F.; Wang, G. MDFN: Multi-scale deep feature learning network for object detection. Pattern Recognit. 2020, 100, 107149. [Google Scholar] [CrossRef] [Green Version]
  51. Huang, W.; Li, G.; Chen, Q.; Ju, M.; Qu, J. CF2PN: A cross-scale feature fusion pyramid network based remote sensing target detection. Remote Sens. 2021, 13, 847. [Google Scholar] [CrossRef]
  52. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  53. Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional feature fusion. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 3559–3568. [Google Scholar]
  54. Luo, Z.; Li, J.; Zhu, Y. A deep feature fusion network based on multiple attention mechanisms for joint iris-periocular biometric recognition. IEEE Signal Process. Lett. 2021, 28, 1060–1064. [Google Scholar] [CrossRef]
  55. Yu, J.; Li, W.; Li, Z.; Wu, J.; Yang, H.; Yang, J. SAR image super-resolution base on weighted dense connected convolutional network. In Proceedings of the IEEE International Geoscience Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2101–2104. [Google Scholar]
  56. Liu, S.T.; Huang, D.; Wang, Y.H. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference Computer Vision, Munich, Germany, 8–14 September 2018; pp. 404–419. [Google Scholar]
  57. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align deep features for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  58. Yang, X.; Hou, L.; Zhou, Y.; Wang, W.; Yan, J. Dense label encoding for boundary discontinuity free rotation detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15819–15829. [Google Scholar]
  59. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with gaussian wasserstein distance loss. arXiv 2021, arXiv:2101.11952. [Google Scholar]
  60. Han, J.; Ding, J.; Xue, N.; Xia, G.S. Redet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2786–2795. [Google Scholar]
  61. Ting, Y.-S.; Teng, Y.-F.; Chiueh, T.-D. Batch normalization processor design for convolution neural network training and inference. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems (ISCAS), Daegu, Republic of Korea, 22–28 May 2021; pp. 1–4. [Google Scholar]
  62. He, K.; Girshick, R.; Dollar, P. Rethinking ImageNet pre-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–29 October 2019; pp. 4917–4926. [Google Scholar]
  63. Xie, Q.; Luong, M.T.; Hovy, E.; Le, Q.V. Self-training with noisy student improves imagenet classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 19 2020; pp. 10684–10695. [Google Scholar]
  64. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  65. Wu, T.; Zhu, H.; Fa, H.; Zhou, H. An improved target detection algorithm based on EfficientNet. J. Phys. Conf. Ser. 2021, 1983, 012017. [Google Scholar] [CrossRef]
  66. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
Figure 1. The quantities of target samples on SHWD and SHWD processed by SSTRA. In the legend, the histogram represents the proposed enhancement of SSTRA on SHWD. The figure demonstrates the quantities based on different categories between SHWD (the blue rectangle) and SHWD processed by SSTRA (the orange rectangle).
Figure 1. The quantities of target samples on SHWD and SHWD processed by SSTRA. In the legend, the histogram represents the proposed enhancement of SSTRA on SHWD. The figure demonstrates the quantities based on different categories between SHWD (the blue rectangle) and SHWD processed by SSTRA (the orange rectangle).
Entropy 25 00509 g001
Figure 2. Partial image processed by selective small target replication algorithm. The process involves adding the number of small targets by replication and horizontal flipping for training target detection models. We apply random position replication to all training images with small targets, then replicate and flip all small targets in the image and put them back in the original image according to the initial proportion, without blocking the original small target.
Figure 2. Partial image processed by selective small target replication algorithm. The process involves adding the number of small targets by replication and horizontal flipping for training target detection models. We apply random position replication to all training images with small targets, then replicate and flip all small targets in the image and put them back in the original image according to the initial proportion, without blocking the original small target.
Entropy 25 00509 g002
Figure 3. The network structure of ST-CenterNet. The dotted box on the left shows the execution process of the feature extraction network, in which the orange dotted box represents the processing method of the SSTRA (see Figure 4). After the feature extraction network is extracted, the feature is transmitted to the dotted box on the right for prediction (see Figure 5). The prediction is divided into three parts, which are the heatmap prediction, the width-height prediction, and the center point offset prediction.
Figure 3. The network structure of ST-CenterNet. The dotted box on the left shows the execution process of the feature extraction network, in which the orange dotted box represents the processing method of the SSTRA (see Figure 4). After the feature extraction network is extracted, the feature is transmitted to the dotted box on the right for prediction (see Figure 5). The prediction is divided into three parts, which are the heatmap prediction, the width-height prediction, and the center point offset prediction.
Entropy 25 00509 g003
Figure 4. Illustration of the proposed SSTRA. The algorithm calculates the size of all targets and then screens the pixel value. Selected processing mainly solves replication and horizontal flipping based on the boxes of small targets. Using the above operations, we expect to directly increase the sample size of small targets through oversampling.
Figure 4. Illustration of the proposed SSTRA. The algorithm calculates the size of all targets and then screens the pixel value. Selected processing mainly solves replication and horizontal flipping based on the boxes of small targets. Using the above operations, we expect to directly increase the sample size of small targets through oversampling.
Entropy 25 00509 g004
Figure 5. The structure of AFPN. To perform down-sampling, we use ResNet-50 to extract residual mapping and identity mapping ( C 2 , C 3 , C 4 , C 5 ) at each layer. We up-sample with AFPN and the deep features ( P 5 , P 4 ) and shallow features ( P 3 , P 2 ) are obtained by fusion of the horizontal connection.
Figure 5. The structure of AFPN. To perform down-sampling, we use ResNet-50 to extract residual mapping and identity mapping ( C 2 , C 3 , C 4 , C 5 ) at each layer. We up-sample with AFPN and the deep features ( P 5 , P 4 ) and shallow features ( P 3 , P 2 ) are obtained by fusion of the horizontal connection.
Entropy 25 00509 g005
Figure 6. Comparison of the loss function. The desired result is a convergence in the value of loss function when training and validating the model. Figure demonstrates the training and validation loss function values between CenterNet (a) and our method ST-CenterNet (b).
Figure 6. Comparison of the loss function. The desired result is a convergence in the value of loss function when training and validating the model. Figure demonstrates the training and validation loss function values between CenterNet (a) and our method ST-CenterNet (b).
Entropy 25 00509 g006
Figure 7. Visualized comparison of the proposed algorithm with the original algorithm on SHWD. We use pairwise comparison to demonstrate the effectiveness of the proposed algorithm for small target detection, in which the left image in each pair denotes the detection results of the original algorithm, while the image on the right -denotes the detection results of the proposed algorithm The figures with red boxes are the confidence of the detection result for correctly wearing the safety helmet, and the figures with blue boxes are the confidence of the detection result for not correctly wearing the safety helmet.
Figure 7. Visualized comparison of the proposed algorithm with the original algorithm on SHWD. We use pairwise comparison to demonstrate the effectiveness of the proposed algorithm for small target detection, in which the left image in each pair denotes the detection results of the original algorithm, while the image on the right -denotes the detection results of the proposed algorithm The figures with red boxes are the confidence of the detection result for correctly wearing the safety helmet, and the figures with blue boxes are the confidence of the detection result for not correctly wearing the safety helmet.
Entropy 25 00509 g007
Figure 8. AP curves for hat and person. The figure demonstrated the effectiveness of each module in the ST-CenterNet model on the detection accuracy. The desired result is to increase the area occupied under the Precision-Recall curve for each target class (hat and person).
Figure 8. AP curves for hat and person. The figure demonstrated the effectiveness of each module in the ST-CenterNet model on the detection accuracy. The desired result is to increase the area occupied under the Precision-Recall curve for each target class (hat and person).
Entropy 25 00509 g008
Figure 9. Confusion matrix of CenterNet and ST-CenterNet technique under 200 epochs. The desired result is an increase in the number of correctly identified hats and persons. Each figure demonstrates the prediction results based on different categories between CenterNet (a) and the proposed method ST-CenterNet (b).
Figure 9. Confusion matrix of CenterNet and ST-CenterNet technique under 200 epochs. The desired result is an increase in the number of correctly identified hats and persons. Each figure demonstrates the prediction results based on different categories between CenterNet (a) and the proposed method ST-CenterNet (b).
Entropy 25 00509 g009
Figure 10. MR−2 of CenterNet and our method ST-CenterNet on the SHWD. The desired result is a reduction in the value of MR−2 in each target category. Figures show MR−2 values of different categories. Each figure demonstrates the miss rate values based on different categories between CenterNet (a) and our method ST-CenterNet (b).
Figure 10. MR−2 of CenterNet and our method ST-CenterNet on the SHWD. The desired result is a reduction in the value of MR−2 in each target category. Figures show MR−2 values of different categories. Each figure demonstrates the miss rate values based on different categories between CenterNet (a) and our method ST-CenterNet (b).
Entropy 25 00509 g010
Table 1. Summary of related improvement methods.
Table 1. Summary of related improvement methods.
Improvement LiteratureBackboneMethodMeritsLimitations
Data
Enhancement
Mishra [38]ImageNetObject-aware cropping.Increases the sampling rate.Poor samples may be selected, increasing the prediction error.
Zhang [39]ResNetMix-up.Has good robustness to data with noisy labels and adversarial sample attacks.May crop blocks to nonimportant regions or occlude important regions.
Chen [40]ResNetScale transform.Information fed back through the optimization process dynamically guides data preparation.The uneven distribution of objects at different scales will greatly affect the detection quality.
Feature
Enhancement
Liu [33,34]VGG16Multi-scale feature pyramid.The multi-scale feature pyramid is used to detect objects of different scales and aspect ratios.Poor detection effect for small targets.
Belfodil [42]VGG16Lightweight feature fusion.Fast speed; merges feature maps of different layers.Having duplicate boxes; shallow feature maps have insufficient representation ability.
Zhang [43]ResNetA “wide-narrow-wide” structure.Has a large improvement compared to SSD for detection accuracy.Slow speed.
Attention MechanismDai [53]ResNetMulti-scale channel attention.Recognizes and detects objects under extreme scale changes.Limited scenario;
Not necessarily adaptively fusing the received features.
Yu [55]DenseNetDense connected convolutional networks.Enhance the semantic information of small targets in shallow features.Addition combined with identity functions hinder information propagation.
Liu [56]Deeplab-VGG16Dilated convolution; add inception structures.Considers the relationship between the size and eccentricity of the receptive field.Low speed and expensive computation.
Angle
Classfication
Han [57]RetinaNetReplacing the dire-ctional bounding box by the regression output.The aligned features are more beneficial for learning rotated targets.Requires heuristics to define anchors and complex RoI operations.
Yang [58]RetinaNetReplacing the sparse coding labels with dense coding labels.Eliminates periodicity of angle and adjust adaptively according to the aspect ratio.Needs a longer number of bits for encoding.
Yang [59]RetinaNetUsing the Gaussian Wasserstein distance to describe the rotated boxes distance.Avoids the rotation angle regression interval discontinuity and the square problem.Needs a longer number of bits for encoding, and the output is heavy.
Table 2. Experimental environment configuration.
Table 2. Experimental environment configuration.
NameConfiguration
CPUXeon E5-2690 V4 @2.20 GHz
Memory11GB
GPUNVIDIA GeForce GTX 2080Ti
GPU Accelerated LibrariesCUDA10.1.105, cuDNN7.6.5
Deep Learning FrameworkPython3.7.2, Pytorch1.4.0
Operating SystemUbuntu20.04
Table 3. Comparison of different algorithm configurations.
Table 3. Comparison of different algorithm configurations.
AlgorithmBackboneInput SizePrecisionRecallF1-ScoremAPFPS
SSD [33,34]VGG16512 × 51292.0952.550.6573.7723.39
YOLOv3 [64]DarkNet53512 × 51285.0553.310.6669.1732.54
YOLOv3_EfficientNet [65]EfficientNet512 × 51288.3429.250.4461.7920.57
YOLOv4 [66]CSPDarkNet53512 × 51249.8440.740.4540.847.39
YOLOv4_ResNetResNet50 [25]512 × 51268.4458.000.6359.4812.83
YOLOv5CSPDarkNet53512 × 51292.5466.930.7878.579.35
YOLOX [37]CSPDarkNet53512 × 51292.2178.100.8586.0915.10
CenterNet [23]ResNet50512 × 51288.9056.120.6970.9832.78
ST-CenterNetResNet50512 × 51294.1971.880.8289.0628.69
Table 4. Effect on component in ST-CenterNet.
Table 4. Effect on component in ST-CenterNet.
MethodPrecisionRecallF1-ScoremAPFPS
CenterNet [23]88.9056.120.6970.9832.78
CenterNet+SSTRA94.0875.150.8484.8815.25
CenterNet+TAFEM94.5672.900.8284.2016.64
ST-CenterNet94.1971.880.8289.0628.69
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Y.; Lu, X. ST-CenterNet: Small Target Detection Algorithm with Adaptive Data Enhancement. Entropy 2023, 25, 509. https://doi.org/10.3390/e25030509

AMA Style

Guo Y, Lu X. ST-CenterNet: Small Target Detection Algorithm with Adaptive Data Enhancement. Entropy. 2023; 25(3):509. https://doi.org/10.3390/e25030509

Chicago/Turabian Style

Guo, Yujie, and Xu Lu. 2023. "ST-CenterNet: Small Target Detection Algorithm with Adaptive Data Enhancement" Entropy 25, no. 3: 509. https://doi.org/10.3390/e25030509

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop