Next Article in Journal
SRTPN: Scale and Rotation Transform Prediction Net for Multimodal Remote Sensing Image Registration
Previous Article in Journal
Evaluation of Post-Tunneling Aging Buildings Using the InSAR Nonuniform Settlement Index
Previous Article in Special Issue
QMRNet: Quality Metric Regression for EO Image Quality Assessment and Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Scale Object Detector Based on Coordinate and Global Information Aggregation for UAV Aerial Images

1
Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China
2
School of Computer and Information Engineering, Henan University, Kaifeng 475004, China
3
Henan Province Engineering Research Center of Spatial Information Processing and Shenzhen Research Institute, Henan University, Kaifeng 475004, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(14), 3468; https://doi.org/10.3390/rs15143468
Submission received: 24 May 2023 / Revised: 3 July 2023 / Accepted: 6 July 2023 / Published: 9 July 2023
(This article belongs to the Special Issue Artificial Intelligence in Computational Remote Sensing)

Abstract

:
Unmanned aerial vehicle (UAV) image object detection has great application value in the military and civilian fields. However, the objects in the captured images from UAVs have problems of large-scale variation, complex backgrounds, and a large proportion of small objects. To resolve these problems, a multi-scale object detector based on coordinate and global information aggregation is proposed, named CGMDet. Firstly, a Coordinate and Global Information Aggregation Module (CGAM) is designed by aggregating local, coordinate, and global information, which can obtain features with richer context information. Secondly, a Feature Fusion Module (FFM) is proposed, which can better fuse features by learning the importance of different scale features and improve the representation ability of multi-scale features by reusing feature maps to help models better detect multi-scale objects. Moreover, more location information of low-level feature maps is integrated to improve the detection results of small targets. Furthermore, we modified the bounding box regression loss of the model to make the model more accurately regress the bounding box and faster convergence. Finally, we tested the CGMDet on VisDrone and UAVDT datasets. The proposed CGMDet improves mAP0.5 by 1.9% on the VisDrone dataset and 3.0% on the UAVDT dataset.

1. Introduction

UAV application technology has also made significant progress in recent years. Due to good mobility, convenient use, and low cost, UAVs have an extremely high application value in disaster monitoring [1], geological investigation [2], air traffic control [3], emergency relief [4], and other aspects. Therefore, UAV image object detection has been paid more attention by researchers. However, as the shooting angle and height of UAV are changeable, UAV object detection faces the two following challenges: (1) Due to the problem of UAV shooting perspective, there are considerable differences between the scales of targets of the same category or different categories, and there are many small objects, which greatly test the performance of the model to detect multi-scale targets and small targets. (2) In UAV images, there are usually many objects that are blocked, and weak light results in the poor visibility of boundary and features of the object, so it is hard to extract discriminant features from the model.
With the development of deep learning, traditional object detection methods, such as HOG [5] and SIFT [6], are gradually eliminated due to the need for a large amount of prior knowledge. However, the object detection method based on deep learning does not need the manual involvement and can dig deeper and more abstract features. Although the object detector based on deep learning has achieved great results in detecting natural images, there are still great challenges for the object detection of UAV images.
To resolve these problems, this paper proposed a multi-scale object detector based on coordinate and global information aggregation, named CGMDet. Firstly, we designed the CGAM, which can make the model focus on coordinate and global information to alleviate the interference brought by the background. Secondly, we proposed the FFM to better fuse the features of different scales and add the feature maps of larger sizes to feature fusion to more effectively detect multi-scale objects, especially small objects. In addition, we reduced the number of convolutional channels in the neck to decrease the number of parameters required by the network. Based on these works, we obtained an improved feature pyramid network called multi-feature fusion pyramid network (MF-FPN). Finally, we modified the bounding box regression loss to enable the model to more accurately regress bounding boxes. This modification allows high-quality anchors to contribute more gradients to the training process. Therefore, the model can achieve better detection accuracy and faster convergence speed.
In summary, the contributions of this study are as follows:
  • We propose a multi-scale object detector based on coordinate and global information aggregation for UAV aerial images, which can better detect targets with obscure features and targets with different scales;
  • To alleviate the problem of the non-apparent object features due to occlusion and low light, the CGAM is proposed. The module can capture local, coordinate, and global contextual information and fuse them to reduce the interference of background factors on the feature extraction process, thereby obtaining more robust feature information;
  • To make the model better detect multi-scale targets, the FFM is proposed. The module can learn the importance of different scale features in fusion and improve the representation ability of multi-scale features by reusing the features of different scales to improve the ability of model detection of multi-scale targets. At the same time, a larger size feature map is added to the feature fusion structure so that the model can detect small targets better. We named the improved feature pyramid network MF-FPN;
  • To more accurately regress the bounding boxes, we modified the bounding box loss to improve the positioning effect of bounding boxes and make the model converge faster;
  • We validated our CGMDet on two public UAV image datasets. The experimental results show that our model can better detect multi-scale targets and targets with less obvious features in UAV images.

2. Related Work

2.1. Object Detection

At present, the commonly used detectors are one-stage and two-stage detectors. Among them, the first step of the two-stage detector is to generate candidate regions, and the second step is to classify and regress each candidate region. The one-stage detector performs classification and regression directly. Classic two-stage detectors include R-CNN [7], Fast R-CNN [8], Faster R-CNN [9], SPP-Net [10], etc. The accuracy of the two-stage detector is higher than that of the one-stage detector, but the speed is slower than the one-stage detector. Commonly used one-stage detectors include SSD [11], RetinaNet [12], YOLO series [13,14,15,16,17,18], etc. Recently, some anchor-free detectors have been invented. The anchor-free method uses the features of object centers or key points to replace the complex anchor design. For example, FCOS [19] treats each pixel on the feature map as a training sample and uses a four-dimensional vector to regress the predicted box. CenterNet [20] represents objects using their center points and predicts bounding boxes by predicting the offset of the center point and the width and height of the object. The above detector has achieved good results in natural images. However, for UAV images, existing detection methods still face significant challenges. To date, many object detection methods for UAV images have been proposed. For example, Liu et al. [21] proposed an anchor-free detector, Edge YOLO, which can be detected in real-time on the edge computing platform. The data enhancement method is used to suppress overfitting in the training process, and a mixed random loss function is used to improve the detection accuracy of small targets. Jiang et al. [22] used the YOLO model to realize the transfer detection from ground thermal infrared video images to UAV thermal infrared videos. Aiming at the challenges of scale and the sparsity of object detection in aerial images, ClusDet [23] proposed that the CPNet is used to generate the clustering region of objects first and ScaleNet to predict the scale information for adjusting the clustering region. DMNet [24] determines the target area to be clipped through the generated density map and then fuses the detection results of the clipped area and the original image to obtain more accurate detection effects.

2.2. Attention Mechanism

Attention mechanisms are widely used because they allow the model to focus more on important information and ignore unimportant ones. The attention mechanism has played a significant role in object detection. The squeeze-and-excitation block (SE) [25] is a classic channel attention mechanism that can apply a weight to each feature channel, allowing the model to focus more on the important channel information. However, to save computation, SE performs a squeeze operation during the processing which can result in the loss of some channel information. To avoid losing channel information, the Efficient Channel Attention Module (ECA) [26] and Effective Squeeze-and-Excitation Block (ESE Block) [27] were proposed. To prevent channel information loss, the ECA uses a 1D convolution operation instead of the two fully connected layers in SE. The ESE Block removes the squeeze operation and uses one fully connected layer instead of the two fully connected layers in SE. In addition to channel attention mechanisms, coordinate attention (CA) [28] obtains feature maps integrated with spatial coordinate information by performing adaptive average pooling algorithms along the x and y directions of the feature map, respectively. Furthermore, the CBAM [29] introduces channel and spatial attention mechanisms to allocate different weights for different channels and spatial regions to obtain highly responsive feature information and improve the network performance. The adaptive attention fusion mechanism (AAFM) [30] adaptively fuses features within and between modules with learnable fusion factors to improve the feature representation.
To better extract contextual feature information and alleviate the interference caused by occlusion and weak light, we designed the CGAM to extract the local feature and continuously focus on the coordinate information. In addition, it combines global information to obtain features with richer contextual information.

2.3. Multi-Scale Feature Fusion

Targets in unmanned aerial vehicle (UAV) images have characteristics such as large-scale variations and a high proportion of small objects, which pose significant challenges to object detection tasks. In deep networks, low-level feature maps typically contain rich positional information, while high-level features typically contain rich semantic information. Therefore, for better performance, low-level and high-level features are usually fused. FPN [31] fuses the features of adjacent scales through a top–down path and horizontal connections. PANet [32] proposes a bidirectional fusion structure that combines features in both top–down and bottom–up directions. Zhao et al. [33] proposed MLFPN to extract more representative multi-level and multi-scale features through the TUM and FFM, and then integrate features through the SFAM to obtain features with rich contextual information. Tan et al. [34] proposed BiFPN, which adds lateral skip connections to the top–down and bottom–up pathways and assigns a learnable weight to each feature map during fusion to emphasize the importance of different feature maps for better feature fusion.
To improve the detection performance of multi-scale targets, we designed the FFM to learn the importance of different features in the fusion and improve the representation ability of different scale features through the repeated use of features of different scales to improve the model’s ability to detect multi-scale targets.

3. Methods

This paper proposes a multi-scale object detector based on coordinate and global information aggregation for UAV aerial images, named CGMDet. Firstly, we designed the CGAM, which allows the backbone to focus more on the coordinate information and global context information during the feature extraction process, to enhance the ability of the network to extract features. Then, the FFM was proposed to better fuse multi-scale features. Finally, we modified the bounding box loss to obtain better detection results. Figure 1 shows the overall architecture of CGMDet. Our CGMDet enables the better detection of targets under occluded and low-light conditions and multi-scale targets.
First, a 640 × 640 image is input into the backbone to extract features. The image passes through four CBS modules, which consist of convolution, batch normalization [35], and SiLU [36] activation function. In the first and third CBS modules, the convolution has a stride of 1, while in the second and fourth modules, it has a stride of 2. After obtaining the feature maps of four-fold downsampling, the CGAM extracts the feature. After that, the features of three scales were extracted by MP and CGAM. The MP module uses maximum pooling with the stride of 2 and 3 × 3 convolution with the stride of 2 to realize downsampling. SPPCSPC [18] aims to aggregate features with different receptive fields to obtain richer semantic contextual information. Then, the different scales of feature maps are fed into the neck, where our proposed FFM is used to fuse the features of different scales. Because most targets in UAV images are small, and low-level features are better for small target detection, we also input feature maps of size 160 × 160 into the feature fusion. Therefore, the feature map sizes that need to be fused in the neck are 160 × 160 , 80 × 80 , 40 × 40 , and 20 × 20 . ELAN-H [37] refined the fused features. RepConv can decouple the training and inference process, allowing the model to learn more knowledge during the training process without affecting the inference speed. Finally, the feature maps of size 80 × 80 , 40 × 40 , and 20 × 20 are used for object detection. Figure 2 shows the detailed network structure of our CGMDet.

3.1. Coordinate and Global Information Aggregation Module

Because of the camera angle of the UAV, there are often many cases where the targets are occluded. At the same time, the light is dimmer in nighttime scenes. The feature extraction process in these two scenarios is easily disturbed by background factors, which is easy to cause missed detections and false detections. To alleviate the interference of background factors, we designed the CGAM, which can integrate global information, coordinate information, and local information extracted by convolution, to obtain more robust features. The structure of CGAM is shown in Figure 3.
The CGAM consists of two branches. The first branch introduces the coordinate attention mechanism, which constantly focuses on the coordinate information when using convolution for feature extraction. The second branch obtains global information on the feature map through two pooling operations. By fusing the features extracted from the two branches, richer contextual features are obtained.
The first branch of CGAM first uses a 1 × 1 convolution to reduce the number of channels in the input feature X R C × H × W by half, obtaining the first intermediate feature map M 1 R C 2 × H × W . As shown in Formula (1):
M 1 = C o n v 1 × 1 F
The features map is then extracted using 3 × 3 convolution and the coordinate attention mechanism, obtaining the second and third intermediate output feature maps M 2 , M 3 R C 2 × H × W . As shown in Formulas (2) and (3):
M 2 = C A C o n v 3 × 3 C A C o n v 3 × 3 M 1
M 3 = C A C o n v 3 × 3 C A C o n v 3 × 3 M 2
where CA represents the coordinate attention mechanism. The coordinate attention is shown in Figure 4.
The coordinate attention mechanism first performs pooling operations on the input feature F R C × H × W along the horizontal and vertical directions, obtaining features f h R C × H × 1 and f w R C × 1 × W . As shown in Formulas (4) and (5):
f c h = 1 W 0 i W F c h , i
f c w = 1 H 0 j H F c j , w
where F c and f c represent the c-th channel of the input and output features, respectively. W and H represent the width and height of the input feature, respectively. Then, f h and f w are concatenated along the spatial dimension, and the number of channels is reduced using a 1 × 1 convolution. Furthermore, feature Q R C r × 1 × ( W + H ) is obtained by passing it through batch normalization and an activation function, where r is a scaling factor. Batch normalization is used to prevent gradient explosion or vanishing, making the model more stable during training, and the activation function introduces nonlinear factors to enhance the expression ability of the model. The formula is shown as (6):
Q = δ B N C o n v 1 × 1 [ f h , f w ]
where [ · ] denotes the channel concatenation operation. B N denotes batch normalization. δ represents a nonlinear activation function. Then, the feature tensor Q is split along the spatial dimension to obtain two feature tensors y h R C r × H × 1 and y w R C r × 1 × W . Increasing the number of channels for y h and y w to the same as the input feature map F by 1 × 1 convolution, and then the attention weights g h and g w are obtained by the sigmoid function. The formulas are shown as (7) and (8):
g h = σ ( C o n v 1 × 1 ( y h ) )
g w = σ ( C o n v 1 × 1 ( y w ) )
where σ denotes the sigmoid function. Finally, g h and g w are multiplied by the feature map F. As shown in Formula (9):
C A = F g h g w
The second branch of the CGAM module first uses global pooling operations to add the global contextual information of the backbone network. For the input feature X R C × H × W , perform global average pooling and global maximum pooling operations first, then add the results, and finally allocate weights for each channel through a fully connected layer and a sigmoid function, making the model focus on the highly responsive channel information. As shown in Formula (10):
O = σ F C G A P X G M P X
where G A P and G M P represent global average pooling and global maximum pooling, respectively. σ represents the sigmoid function. F C represents a fully connected layer. Then, multiply the result with the input feature X and use a 1 × 1 convolution to obtain the output M 4 R C 2 × H × W of the second branch. As shown in Formula (11):
M 4 = C o n v 1 × 1 X O
The CGAM module first performs channel concatenation on all intermediate output features M 1 , M 2 , M 3 , and M 4 from the two branches, and then uses a 1 × 1 convolution to obtain the output feature Z R 2 C × H × W of CGAM. As shown in Formula (12):
Z = C o n v 1 × 1 [ M 1 , M 2 , M 3 , M 4 ]
Our CGAM module can extract coordinate information, global information, and local information simultaneously and fuse them to obtain more robust features, thereby accurately locating the target, reducing the focus of the model on the background, and improving the detection ability of the model.

3.2. Multi-Feature Fusion Pyramid Network

The object scale changes greatly in UAV images, and there are many small objects. To enhance the ability of the network to detect multi-scale targets, we proposed the FFM, which can learn the importance of different features in fusion to integrate the features of different scales better. Moreover, reusing feature maps to enrich the context information of fused features and improve the expression ability of multi-scale features. In this way, the model can improve the detection ability of multi-scale targets. We added the feature map of size 160 × 160 to the MF-FPN for fusion to alleviate the difficulty in detecting small objects. The FFM is shown in Figure 5.
There are two main ways to fuse features. If only two feature maps need to be fused, such as the top–down path in the neck, the method shown in Figure 5a is used. This method assigns two learnable weights to the two feature maps to determine the importance of each feature map, as shown in Formula (13):
P = w 1 F 1 + w 2 F 2 w 1 + w 2 + Δ
where w 1 and w 2 are learnable parameters, and Δ is a small number to avoid numerical instability. For the case of fusing three feature maps, we use the method shown in Figure 5b. The calculation method is shown in Formula (14). We first use Formula (13) to fuse the three feature maps pairwise, where each feature map participates in two fusions, achieving the effect of reusing features. Then, we obtain three different intermediate feature maps and finally assign three learnable weights to fuse these three feature maps, obtaining the output feature map with rich contextual information for the final prediction.
N = w 1 P F 1 , F 2 + w 2 P F 1 , F 3 + w 3 P F 2 , F 3 w 1 + w 2 + w 3 + Δ
Since the fusion features contain feature maps with different scales and channel numbers, it is necessary to adjust the size and number of channels of the feature maps to be consistent before fusion. Our FFM integrates multi-scale features better by automatically learning the importance of different scale feature maps in the fusion process. Furthermore, the feature maps of different scales can be used to enrich the context information of fusion features and improve the representation ability of multi-scale features to enhance the detector’s perception of targets of different scales.
To preserve more feature information, the convolutional channels in the model are usually large, and larger channels bring more parameters to the model. The number of parameters of convolution can be calculated by the Formula (15).
P a r a m s = K h × K w × C i n × C o u t
where K w and K h represent the size of the convolutional kernel. C i n and C o u t represent the number of input and output channels of the convolution, respectively. Therefore, to decrease the parameters of the model, we modified the convolutional channel numbers in the neck of the model. First, the channel numbers of the 3 × 3 convolutions in ELAN-H_1, ELAN-H_2, and ELAN-H_3 to 32. Then, we adjusted the output channel numbers of the first two 1 × 1 convolutions in ELAN-H_4, ELAN-H_5, and ELAN-H_6 to 1 / 4 of the input channel numbers. ELAN-H_1, ELAN-H_2, ELAN-H_3, ELAN-H_4, ELAN-H_5, and ELAN-H_6 have the same structure. The changes in the number of parameters for all ELAN-H modules in the neck are shown in Table 1.
After the above improvements, we named the improved feature pyramid structure MF-FPN. The pseudo-code of the MF-FPN is shown in Algorithm 1. We first fuse two adjacent features in X = { x 1 , x 2 , x 3 , x 4 } from top to bottom to obtain four intermediate feature maps M = { m 1 , m 2 , m 3 , m 4 } . Then, the features from X and M are fused using the bottom–up path and skip connections to obtain three final features of different scales, denoted as Y = { y 1 , y 2 , y 3 } , which will be used for prediction.
Algorithm 1 The feature fusion method of MF-FPN.
  • Input:  X = { x 1 , x 2 , x 3 , x 4 } , X refers to four different scale feature maps of the backbone network output. The scale of x 1 is the smallest and x 4 is the largest.
  • Step 1:  M = { } , M refers to the intermediate feature map generated by the top–down branch of MF-FPN. C o n v ( ) represents a series of convolution operations required, R e s h a p e ( ) represents the upsampling and downsampling operations required, and F F M ( ) represents our feature fusion operation.
  •        for  i = 1 to 4 do
  •              if  i = 1  then
  •                   m i = C o n v x i
  •              else
  •                   m i = C o n v F F M x i , R e s h a p e x i 1
  •              end if
  •              M . a p p e n d m i
  •        end for
  • Step 2:  Y = { } , Y refers to the feature map generated by the bottom–up branch of MF-FPN for use in prediction.
  •        for  i = 1 to 3 do
  •              y i = C o n v F F M x i , M i , R e s h a p e M i + 1
  •              Y . a p p e n d y i
  •        end for
  • Output: Return Y .

3.3. Loss Function

The CIOU [38] loss is commonly used as the bounding box regression loss in existing models. The definition of CIOU is as follows:
L C I O U = 1 I O U + ρ 2 b , b g t c 2 + α v
I O U = | B B g t | | B B g t |
where ρ represents the Euclidean distance. b and b g t denote the center points of the predicted and ground truth box, respectively. c represents the diagonal length of the minimum bounding rectangle of the ground truth box and predicted box. v and α are defined as follows:
v = 4 π 2 a r c t a n w g t h g t a r c t a n w h 2
α = v 1 I O U + v
where h g t and w g t are the height and width of the ground truth box. h and w are the height and width of the predicted box.
However, accurately regressing the height and width of the bounding box cannot only be achieved through the aspect ratio. Because when w = k w g t and h = k h g t k R + , v = 0 . The EIOU loss [39] not only retains the advantages of the CIOU but also minimizes the differences between the height and width of the predicted and ground truth boxes, resulting in better localization performance. The definition of EIOU is shown in Formula (20).
L E I O U = 1 I O U + ρ 2 b , b g t c 2 + ρ 2 h , h g t h c + ρ 2 w , w g t w c
where h c and w c are the height and width of the minimum bounding box surrounding the ground truth and predicted box. EIOU can directly regress the height and width of the prediction box. Furthermore, to make the model converge faster, we use the focal EIOU loss [39], which combines the focal loss with the EIOU loss, as the bounding box regression loss for CGMDet. It allows high-quality anchors to contribute more gradients to the training process, thereby improving the convergence speed of the model. Its definition is as follows:
L F o c a l E I O U = I O U λ × L E I O U
where λ is an adjustable parameter, which we set to 0.5. Additionally, the confidence and classification losses of the model are calculated using binary cross entropy with Logits Loss (BCEWithLogitsLoss) [40]. The definition is as follows:
L B C E = n = 1 N y ^ i log σ y + 1 y ^ i log σ 1 y
where N is the number of input vectors. y ^ i and y are the predicted and truth vectors, respectively. σ is the sigmoid function. The overall loss of CGMDet can be obtained by combining the classification loss, confidence loss, and bounding box regression loss. The definition is as follows:
L o s s = λ 1 L b o x + λ 2 L o b j + λ 3 L c l s = λ 1 L F o c a l E I O U ( P b o x , T b o x ) + λ 2 L B C E ( P o b j , T o b j ) + λ 3 L B C E ( P c l s , T c l s )
where L b o x , L o b j , and L c l s represent the bounding box regression loss, confidence loss, and classification loss, respectively. P b o x and T b o x denote the predicted and ground truth box, respectively. P o b j and T o b j denote the predicted and truth confidence, respectively. Furthermore, P c l s and T c l s represent the predicted and truth class probability, respectively. The hyperparameters λ 1 , λ 2 , and λ 3 are set to 0.05, 0.7, and 0.3 by default.

4. Experiments

To verify the effectiveness of our detector, we conducted experiments on two publicly available UAV image datasets and compared it with other detectors.

4.1. Datasets

The VisDrone benchmark [41] contains 10,209 static images, among which 6471 images are used for training, 3190 for testing, and 548 for validation. The resolution of the images is approximately 2000 × 1500 pixels, collected by various drone platforms in different scenarios, as well as under different weather and lighting conditions. Figure 6 shows some images from this dataset.
The UAVDT benchmark [42] consists of 40,735 images, among which 24,206 images are used for training and 16,529 for validation. This dataset contains images with different weathers, flight heights, shooting angles, and occlusion scenes. The images in the dataset have a resolution of approximately 1080 × 540 pixels. The dataset includes three predefined classes: car, truck, and bus. Figure 7 shows some images from this dataset.

4.2. Implementation and Evaluation Criteria

4.2.1. Implementation

This paper validated the proposed object detector on the Ubuntu 18.04.6 LTS system, trained and tested on NVIDIA GeForce RTX 3090 (24 G) as the graphics processing unit, with an Intel(R) Xeon(R) Silver 4114 CPU @2.20 GHz and Python version 3.6. The CUDA version used was 11.7 and the PyTorch version used was 1.10.2.
During model training, the input image size was set to 640 × 640, and the Stochastic Gradient Descent (SGD) optimizer with momentum was used. The initial learning rate was set to 0.01, the momentum parameter was set to 0.937, the weight decay coefficient was set to 0.0005, and the batch size was set to 8. The total number of training iterations for the VisDrone dataset was 300, and for the UAVDT dataset was 200.

4.2.2. Evaluation Criteria

Precision P, recall R, average precision A P , and mean average precision m A P are used as metrics to evaluate the performance of our detector. P represents how many predicted positive samples are correct. R represents how many positive samples are predicted. The definitions of P and R are as follows:
P = T P T P + F P
R = T P T P + F N
where T P represents how many samples were correctly predicted to be positive. F P represents how many samples were incorrectly predicted to be positive. F N represents how many samples were incorrectly predicted to be negative. P and R are usually trade-offs between each other. Therefore, A P can better measure the detection capability of the network. The definition of A P is shown in Formula (26):
A P = 0 1 P ( R ) d R
where P ( R ) represents the precision value when the recall value on the P-R curve is R. m A P is the average of all class A P values, which can represent the average detection performance of the detector on the dataset. The definition is shown in Formula (27):
m A P = 1 K i = 1 K A P i
where A P i denotes the A P value of the i-th category. K denotes the number of target categories.
To better describe the ability of our detector to detect multi-scale objects, we also used COCO evaluation metrics [43], such as A P S , A P M , and A P L . A P S represents the A P value of small objects with an area less than 32 × 32 . A P M represents the A P value of medium objects with an area between 32 × 32 and 96 × 96 . A P L represents the A P value of large objects with an area greater than 96 × 96 .

4.3. Experimental Results

4.3.1. Experimental Results on the VisDrone Dataset

We evaluated our model using the VisDrone dataset and compared our model with other models to validate its effectiveness. As shown in Table 2, our CGMDet achieved the highest mAP0.5 and mAP. Compared with baseline YOLOv7 [18], mAP0.5 and mAP are increased by 1.9% and 1.2%, respectively. At the same time, CGMDet achieved the second-highest result on mAP0.75 and A P M , where it was only 0.1% lower than CDMNet’s A P M . Compared with YOLOv7, our CGMDet increases by 1.3% A P S and 1.2% A P M . Although the value of A P L is 0.4% lower than that of YOLOv7, on the whole, CGMDet is superior to YOLOv7 in the detection performance of multi-scale targets, which also indicates that our proposed FFM can better integrate the features of different scales and improve the model’s perception ability of features of different scales. Compared to Edge YOLO, our CGMDet is 5.7% lower on A P L , but our CGMDet is better than Edge YOLO on other metrics. Compared with NWD, CGMDet is 10.6% higher at mAP0.5 but 2.0% lower at A P S . Although our CGMDet is lower than ClusDet and DMNet on A P L , it is higher than them on other metrics. Our CGMDet was only 0.2% higher than CEASC on mAP0.5, but significantly higher than CEASC on mAP0.75 and mAP. At the same time, compared with CDMNet, CGMDet has slight disadvantages in mAP0.75, A P S , and A P M , but has obvious advantages in mAP0.5 and A P L . Compared to RetinaNet, Cascade-RCNN, Faster-RCNN, YOLOv3, YOLOX, YOLOv5l, HawkNet, and QueryDet, our CGMDet outperformed them on all metrics.
We also listed the mAP0.5 for each category to describe in more detail which categories our model has improved on. As shown in Table 3, our model has a higher mAP0.5 than other models for each category. In addition, except for the tricycle category, which has the same result as YOLOv7, all other categories have greatly improved, especially the bus and bicycle categories, which increased by 3.3% and 3.8%, respectively.
To make it more deployable on mobile devices, we also designed a tiny version of the model and conducted experiments. The inference time was obtained by calculating the average prediction time of all images in the test set. The results in Table 4 show that our CGMDet-tiny achieved the best results. Compared to YOLOv7-tiny, our CGMDet-tiny achieved a 4% improvement in mAP0.5, 3.6% in mAP0.75, and 3.1% in mAP. Meanwhile, A P S , A P M , and A P L increased by 2.8%, 4.1%, and 1%, respectively. Moreover, our model has only increased by 1.4 M parameters. Although our CGMDet-tiny is slower than other models in terms of inference time, it can still detect in real time. Moreover, judging from the results of detection performance, it is worth trading inference time for detection accuracy.
To better illustrate the advantages of our model, we provide the detection results of several images in different scenarios. As shown in Figure 8(a1–a4) are the results of YOLOv7, and Figure 8(b1–b4) are the detection results of our CGMDet. From the red dashed box in (a1,b1) of Figure 8, YOLOv7 recognized that text on the ground as a car, while our model can recognize it as the background. From Figure 8(a2,b2), our model can also distinguish between two objects with very similar features that are close together. Due to the indistinct features of small targets, it is difficult for the model to learn, and it is easy to recognize similar backgrounds as targets. However, our improved model is better able to detect small targets and can effectively distinguish the background, as shown in Figure 8(a3,b3). In addition, we also tested the detection performance in nighttime scenes, as shown in the red dashed box in Figure 8(a4,b4). YOLOv7 failed to detect it, while our model accurately marked it out.
As shown in Table 5, in order to verify the effectiveness of the proposed CGAM module, we compared it with other similar modules. Compared to ELAN, our CGAM improved the mAP0.5 by 0.7% and mAP by 0.5%. Although the mAP0.75 and mAP of CGAM have a slight disadvantage compared with the CSPRepResStage, the number of parameters brought to the model by CGAM is much smaller than that of the CSRepResStage. The results show that our CGAM can effectively enhance the model’s feature extraction ability without too many parameters.
We also compared our MF-FPN with other multi-scale fusion methods. We only incorporate the last three scale features generated by the YOLOv7 backbone. As shown in Table 6, our MF-FPN is similar to PAFPN on all metrics, but MF-FPN has fewer parameters and computations than PAFPN. Although the number of parameters and the calculation amount are the lowest among the three, the BiFPN also has a gap with the other two in performance. The results show that our MF-FPN can effectively integrate multi-scale features and enhance the model’s perception for multi-scale targets.
As shown in Figure 9, we present the visualized results of three different multi-scale fusion methods. As shown in Figure 9(b1), there are many false detections in the red dotted box, while the PAFPN and our MF-FPN reduce the number of false check targets. In the green dotted box, the MF-FPN accurately boxes the target, while the BiFPN and PAFPN do not fully box the target. As shown in the red and green dotted boxes in Figure 9(b2), BiFPN missed one of the two adjacent targets, while neither PAFPN nor MF-FPN missed it, but MF-FPN’s bounding box in the green dotted box was less accurate than PAFPN’s. As shown in the green dotted boxes in Figure 9(a3,b3), although PAFPN and BiFPN detected the truck, they generated a redundant detection box, and the detection box’s scope is not accurate. However, the MF-FPN does not generate redundant detection boxes, and the detection box can accurately enclose the target. It can also be seen from Figure 9 that our MF-FPN can also better detect targets of different scales.

4.3.2. Experimental Results on the UAVDT Dataset

We also evaluated our model using the UAVDT dataset and compared our model with others. As shown in Table 7, compared with YOLOv7, our model has increased mAP0.5 by 3.0%, mAP0.75 by 3.2%, and mAP by 2.3%. The detection performance of YOLOv7 on the UAVDT dataset is worse than that of YOLOv5l. Compared with YOLOv5l, the mAP0.5 is 1.2% lower, the mAP0.75 is 1.3% lower, and the mAP is 1.1% lower. However, our model outperforms YOLOv5l in terms of mAP0.5, mAP0.75, and mAP. We also listed the mAP0.5 results for each category in the table, which showed that our model improved by 2.9% over YOLOv7 in the car category, and the results for the truck and bus categories both improved by 3.4%. In addition, our model outperforms YOLOv5l by 3.3% in the truck category and is also 0.6% and 2.2% higher in the car and bus categories, respectively. At the same time, our model’s performance is also superior to YOLOv3 and YOLOX.
In Table 7, we also included the results of the tiny version for comparison. Our CGMDet-tiny improved the results for the car, truck, and bus categories by 2.3%, 1.0%, and 0.9%, respectively, compared to YOLOv7-tiny. It also increased the mAP0.5 by 1.4% and improved the mAP0.75 and mAP by 2.2% and 1.8%, respectively. In addition, our CGMDet-tiny outperforms YOLOX-tiny in all metrics. However, compared to YOLOv5s, our CGMDet-tiny only outperforms by 0.3% in terms of mAP.
To illustrate the superiority of our CGMDet, we present detection results for several images in different scenarios. As shown in Figure 10(a1–a3) are the results of YOLOv7, and Figure 10(b1–b3) are the results of our proposed model. From Figure 10(a1,b1), our model performs significantly better than YOLOv7 in terms of detecting small targets. From Figure 10(a2,b2), even under low light conditions at night, our model has a significant improvement over the baseline. In addition, as shown in Figure 10(a3,b3), YOLOv7 detected the left tree as a car, and the bus as a truck, and did not detect the objects that were truncated at the bottom of the image or slightly occluded on the right side of the image. In contrast, our model correctly recognized the tree as the background, accurately identified the object categories, and accurately detected the objects at the bottom and right of the image.

4.3.3. Ablation Experiments

We used the VisDrone dataset to conduct ablation experiments for our model to verify the effectiveness of our improved methods. For fairness, all experimental settings have the same parameters and are conducted in the same environment. As shown in Table 8, we use YOLOv7 as the baseline and achieved a 49% mAP. Furthermore, the results show that each improvement can enhance the detection ability of the model to some extent.
  • CGAM: To reflect the effectiveness of the CGAM, we replaced the ELAN [37] module in the YOLOv7 backbone with our CGAM module. Compared with YOLOv7, using CGAM increased the mAP0.5 by 0.7%. This is because our CGAM can simultaneously extract local, coordinate, and global information, making the extracted feature map richer in contextual information, and thereby improving the ability of the backbone network to extract features;
  • MF-FPN: To demonstrate the effectiveness of MF-FPN, we replaced the neck part of YOLOv7 with the proposed MF-FPN. Compared with YOLOv7, the improved model with MF-FPN increased mAP0.5 by 1%, and the parameters of the model also decreased by 2.2M. This shows that the MF-FPN can fully integrate multi-scale features with fewer parameters. This also proves that our FFM can fully integrate features of different scales and obtain multi-scale feature maps with stronger representation ability;
  • Focal-EIOU Loss: To reflect the effectiveness of Focal-EIOU Loss, we replaced the CIOU loss in YOLOv7 with Focal-EIOU Loss. Compared with CIOU Loss, Focal-EIOU Loss can more accurately regress the bounding box and allow high-quality anchor boxes to make more contributions during training, thereby improving detection performance. Compared with YOLOv7, the model’s mAP0.5 increased by 0.5%;
  • Proposed Method: When CGAM, MF-FPN, and Focal-EIOU loss were all incorporated into YOLOv7, our model was obtained. Compared with YOLOv7, the precision increased by 0.5%, the recall increased by 2.1%, the mAP0.5 increased by 1.9%, and the parameter size of our model was reduced by 0.7M compared to the baseline. The results show that our improvement methods are very effective, and each improvement can enhance the performance of the model;
We plotted the change process of mAP0.5 and mAP during training, as shown in Figure 11. It is evident from the figure that compared with YOLOv7, each of our improvement points significantly improved the model’s performance. The model we proposed by integrating all the improvement points undergoes an especially significant improvement compared to YOLOv7.
In addition, we also listed the changes in the convolutional parameter sizes of each ELAN-H module in the neck part of the model. As shown in Table 1, ELAN-H_3 and ELAN-H_4 are two additional modules that we added to our model.
To better demonstrate the effectiveness of CGMDet, we used Grad-CAM [54] to visualize the model’s execution results in the form of heatmaps. As shown in Figure 12, the first row of the image shows that, compared with YOLOv7, our model reduces the focus on similar objects around small targets and can more accurately detect small targets. The second row shows that our model alleviates the interference of background factors. The third row shows the heat map results generated by our model in low-light nighttime scenes, where we can observe that, even under low-light conditions, our model can accurately focus on the target while reducing the attention to the background.
We briefly tested the effects of λ 1 , λ 2 , and λ 3 in Formula (23) on the model’s performance. As shown in Table 9, when λ 1 = 0.07 , λ 2 = 0.7 , and λ 3 = 0.3 , mAP0.75 and mAP obtained the highest result, but mAP0.5 is 0.7% lower than the best result. When λ 1 = 0.05 , λ 2 = 0.6 , and λ 3 = 0.3 , the mAP is also the highest, but the mAP0.5 and mAP0.75 are not very good. When λ 1 = 0.05 , λ 2 = 0.7 , and λ 3 = 0.3 , the mAP0.5 is the highest, and the mAP0.75 and mAP are only 0.2% lower than the highest result. It can be seen from the results that, when λ 1 = 0.05 , λ 2 = 0.7 , and λ 3 = 0.3 , the mAP0.5, mAP0.75, and mAP can achieve good balance, so we choose them as the final values.
In addition, we explored the effects of different learning rates on the performance of our model. As shown in Table 10, the model performance gradually improves when the learning rate increases from 0.005 to 0.010. When it is higher than 0.010, the model’s overall performance declines slightly. When the learning rate is 0.011, mAP reaches the highest, but mAP0.5 and mAP0.75 are lower than when the learning rate is 0.010. Similarly, when the learning rate is 0.012, mAP0.75 reaches the highest, but mAP0.5 and mAP are relatively low. Therefore, we chose 0.010 as our final learning rate.

4.3.4. Extended Experiments

To verify the generalization ability of our model, we conducted experiments on the generic dataset VOC2012 [55]. We use VOC2012 train for training and VOC2012 val for validation. The training set contains 5717 images, and the validation set contains 5823 images. The value of mAP0.5 for each category is shown in Table 11. Our model is superior to the baseline model in some categories. However, some categories are worse than the baseline model. For further analysis, we list the detection results of different scales on the VOC2012 dataset in Table 12. Our model improves by 1.8% on A P S and 0.3% on A P M but decreases by 3.3% on A P L . This shows that our model is more suitable for detecting small and medium targets.
From the results of the extended experiment, our model’s performance on the VOC2012 dataset is not very good, which indicates that our model is more suitable for UAV images. However, it also shows that the generalization ability of our model needs to be stronger. The result of our analysis is that after adding the P2 layer to the model for feature fusion, the feature proportion of small and medium targets increases, resulting in the model paying more attention to small and medium targets while ignoring large targets. Therefore, our model is more suitable for detecting UAV images with more small- and medium-sized targets.

5. Discussion

In Table 6, although the PAFPN and MF-FPN are similar in performance, our CGMDet needs to integrate more features. If we use the PAFPN, the model’s number of parameters will increase. Therefore, the MF-FPN is more suitable for use in our CGMDet to avoid increasing the number of parameters in the model.
In Table 8, MF-FPN, although higher than Focal-EIOU on mAP0.5, is 0.4% lower on the F1-Score. This is because the precision of MF-FPN is increased by 2.8%, but the recall is reduced by 1.0%. However, Focal EIOU has improved both precision and recall. This shows that Focal-EIOU is better at balancing precision and recall. It can also be seen from the index F1-Score that the Focal EIOU has a better balance between precision and recall than MF-FPN, which helps improve the network’s quality.
In addition to the above successes, our CGMDet has certain limitations. The performance of CGMDet on natural images is not very good, which indicates that the model’s generalization ability needs to be stronger. In addition, in the UAV image, the distant target is usually densely arranged, and the target may be blocking another target. In this scenario, it is difficult for the model to determine the area and number of targets, resulting in the model generating many redundant detection boxes or detection boxes with inaccurate scopes. Therefore, the detection performance of CGMDet in this scenario needs to be improved. In addition, after the above improvements, the model complexity of CGMDet has increased, affecting the model’s inference time. Therefore, the model’s complexity needs to be further improved.

6. Conclusions

This study proposed a multi-scale object detector based on coordinate and global information aggregation for UAV aerial images. This detector can focus more on the features of the objects and better detect multi-scale objects. We designed the CGAM that integrates local, coordinate, and global information to obtain more robust features, effectively alleviating the interference of background factors such as occlusion and weak light in the feature extraction process. The proposed FFM can integrate the features of different scales more fully by automatically learning the importance of features of different scales in fusion. Furthermore, by reusing features, the expression ability of multi-scale features is enhanced, and the detector’s perception ability of different scale targets is improved. Finally, by modifying the loss function, the localization effect of the model on the target is improved. The experiments show that, compared with the baseline detector, the CGMDet improves mAP0.5 by 1.9% on the VisDrone dataset and 3.0% on the UAVDT dataset. Moreover, it can better detect multi-scale targets. In future work, we will focus on detecting dense objects and developing lightweight models while improving the generalization ability of models.

Author Contributions

Conceptualization, L.Z.; methodology, Z.L.; software, Z.L.; validation, L.Z., Z.L., H.Z. and Y.-E.H.; formal analysis, Y.L.; investigation, L.D.; resources, X.Z.; data curation, Y.-E.H.; writing—original draft preparation, L.Z. and Z.L.; writing—review and editing, L.Z. and Y.L.; visualization, H.Z.; supervision, H.Z.; project administration, Z.L.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Basic Research Program of China (Grant no. 2019YFE0126600); the Major Project of Science and Technology of Henan Province (Grant no. 201400210300); the Key Scientific and Technological Project of Henan Province (Grant no. 212102210496); the Key Research and Promotion Projects of Henan Province (Grants no. 212102210393 and 202102110121); Kaifeng Science and Technology Development Plan (Grant no. 2002001); National Natural Science Foundation of China (Grant no. 62176087); and Shenzhen Science and Technology Innovation Commission (SZSTI), Shenzhen Virtual University Park (SZVUP) Special Fund Project (Grant no. 2021Szvup032).

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

We sincerely thank the anonymous reviewers for the critical comments and suggestions for improving the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
HOGHistogram of Oriented Gradients
SIFTScale Invariant Feature Transform
SSDSingle Shot MultiBox Detector
YOLOYou Only Look Once
FCOSFully Convolutional One-Stage Object Detection
CGAMCoordinate and Global Information Module
MF-FPNMulti-Feature Fusion Pyramid Network
SESqueeze and Excitation
ECAEfficient Channel Attention
ESEEffective Squeeze and Excitation
CACoordinate Attention
CBAMConvolutional Block Attention Module
FPNFeature Pyramid Network
PANetPath Aggregation Network
MLFPNMulti-Level Feature Pyramid Network
TUMThinned U-shape Module
FFMFeature Fusion Module
BiFPNBidirectional Feature Pyramid Network
IOUIntersection over Union
P-R curvePrecision–Recall Curve
GFLOPsGiga Floating-point Operations per Second

References

  1. Wei, W.; Chao, F.; Ting, L. Multiperiod unmanned aerial vehicles path planning with dynamic emergency priorities for geohazards monitoring. IEEE Trans. Ind. Inform. 2022, 18, 8851–8859. [Google Scholar]
  2. Villarreal, C.A.; Garzón, C.G.; Mora, J.P.; Rojas, J.D.; Ríos, C.A. Workflow for capturing information and characterizing difficult-to-access geological outcrops using unmanned aerial vehicle-based digital photogrammetric data. J. Ind. Inf. Integr. 2022, 26, 100292. [Google Scholar] [CrossRef]
  3. Hailong, H.; Savkin, A.V.; Chao, H. Decentralized autonomous navigation of a uav network for road traffic monitoring. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 2558–2564. [Google Scholar]
  4. Do-Duy, T.; Nguyen, L.D.; Duong, T.Q.; Khosravirad, S.R.; Claussen, H. Joint optimisation of real-time deployment and resource allocation for uav-aided disaster emergency communications. IEEE J. Sel. Areas Commun. 2021, 39, 3411–3424. [Google Scholar] [CrossRef]
  5. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  6. Lowe, D.G. Object recognition from local scale-invariant features. Proc. Seventh IEEE Int. Conf. Comput. Vis. 1999, 2, 1150–1157. [Google Scholar]
  7. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 28 June 2014; pp. 580–587. [Google Scholar]
  8. Girshick, R. Fast r-cnn. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1907–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  12. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  13. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  14. Redmon, J.; Farhadi, A. Yolo9000: Better, faster, stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  15. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  16. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  17. Ultralytics. Yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 18 June 2022).
  18. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  19. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  20. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  21. Liu, S.; Zha, J.; Sun, J.; Li, Z.; Wang, G. EdgeYOLO: An Edge-Real-Time Object Detector. arXiv 2023, arXiv:2302.07483. [Google Scholar]
  22. Jiang, C.; Ren, H.; Ye, X.; Zhu, J.; Zeng, H.; Nan, Y.; Sun, M.; Ren, X.; Huo, H. Object detection from UAV thermal infrared images and videos using YOLO models. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102912. [Google Scholar] [CrossRef]
  23. Yang, F.; Fan, H.; Chu, P.; Blasch, E.; Ling, H. Clustered object detection in aerial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8311–8320. [Google Scholar]
  24. Li, C.; Yang, T.; Zhu, S.; Chen, C.; Guan, S. Density map guided object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 190–191. [Google Scholar]
  25. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  26. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  27. Lee, Y.; Park, J. CenterMask: Real-Time Anchor-Free Instance Segmentation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13903–13912. [Google Scholar]
  28. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  29. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  30. Ye, Y.; Ren, X.; Zhu, B.; Tang, T.; Tan, X.; Gui, Y.; Yao, Q. An Adaptive Attention Fusion Mechanism Convolutional Network for Object Detection in Remote Sensing Images. Remote Sens. 2022, 14, 516. [Google Scholar] [CrossRef]
  31. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  32. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  33. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2Det: A Single-Shot Object Detector Based on Multi-Level Feature Pyramid Network. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  34. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar]
  35. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the ICML’15: 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–11 July 2015; Volume 37, pp. 448–456. [Google Scholar]
  36. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, C.Y.; Liao, H.Y.M.; Yeh, I.H. Designing network design strategies through gradient path analysis. arXiv 2022, arXiv:2211.04800. [Google Scholar]
  38. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-iou loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  39. Zhang, Y.F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient iou loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  40. Sun, Z.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. BiFA-YOLO: A Novel YOLO-Based Method for Arbitrary-Oriented Ship Detection in High-Resolution SAR Images. Remote Sens. 2021, 13, 4209. [Google Scholar] [CrossRef]
  41. Zhu, P.; Wen, L.; Du, D.; Bian, X.; Fan, H.; Hu, Q.; Ling, H. Detection and tracking meet drones challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7380–7399. [Google Scholar] [CrossRef]
  42. Du, D.; Qi, Y.; Yu, H.; Yang, Y.; Duan, K.; Li, G.; Zhang, W.; Huang, Q.; Tian, Q. The unmanned aerial vehicle benchmark: Object detection and tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 370–386. [Google Scholar]
  43. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  44. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  45. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  46. Lin, H.; Zhou, J.; Gan, Y.; Vong, C.M.; Liu, Q. Novel up-scale feature aggregation for object detection in aerial images. Neurocomputing 2020, 411, 364–374. [Google Scholar] [CrossRef]
  47. Yang, C.; Huang, Z.; Wang, N. Querydet: Cascaded sparse query for accelerating high-resolution small object detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13668–13677. [Google Scholar]
  48. Wang, J.; Xu, C.; Yang, W.; Yu, L. A normalized Gaussian Wasserstein distance for tiny object detection. arXiv 2021, arXiv:2110.13389. [Google Scholar]
  49. Du, B.; Huang, Y.; Chen, J.; Huang, D. Adaptive Sparse Convolutional Networks with Global Context Enhancement for Faster Object Detection on Drone Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 13435–13444. [Google Scholar]
  50. Duan, C.; Wei, Z.; Zhang, C.; Qu, S.; Wang, H. Coarse-grained density map guided object detection in aerial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2789–2798. [Google Scholar]
  51. Liu, C.; Yang, D.; Tang, L.; Zhou, X.; Deng, Y. A Lightweight Object Detector Based on Spatial-Coordinate Self-Attention for UAV Aerial Images. Remote Sens. 2023, 15, 83. [Google Scholar] [CrossRef]
  52. Xu, S.; Wang, X.; Lv, W.; Chang, Q.; Cui, C.; Deng, K.; Wang, G.; Dang, Q.; Wei, S.; Du, Y. PP-YOLOE: An evolved version of YOLO. arXiv 2022, arXiv:2203.16250. [Google Scholar]
  53. Ultralytics. YOLOv8. Available online: https://github.com/ultralytics/ultralytics (accessed on 1 January 2023).
  54. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  55. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The overall architecture of CGMDet.
Figure 1. The overall architecture of CGMDet.
Remotesensing 15 03468 g001
Figure 2. The detailed architecture of CGMDet.
Figure 2. The detailed architecture of CGMDet.
Remotesensing 15 03468 g002
Figure 3. Coordinate and Global Information Aggregation Module.
Figure 3. Coordinate and Global Information Aggregation Module.
Remotesensing 15 03468 g003
Figure 4. Coordinate Attention.
Figure 4. Coordinate Attention.
Remotesensing 15 03468 g004
Figure 5. The structure of the Feature Fusion Module (FFM). (a) represents the feature fusion method in the top-down path; (b) represents the feature fusion method in the bottom-up path.
Figure 5. The structure of the Feature Fusion Module (FFM). (a) represents the feature fusion method in the top-down path; (b) represents the feature fusion method in the bottom-up path.
Remotesensing 15 03468 g005
Figure 6. Some examples in VisDrone.
Figure 6. Some examples in VisDrone.
Remotesensing 15 03468 g006
Figure 7. Some examples in UAVDT.
Figure 7. Some examples in UAVDT.
Remotesensing 15 03468 g007
Figure 8. The detection results of the VisDrone dataset under different scenes. (a1a4) are the detection results of YOLOv7; (b1b4) are the results of the proposed model.
Figure 8. The detection results of the VisDrone dataset under different scenes. (a1a4) are the detection results of YOLOv7; (b1b4) are the results of the proposed model.
Remotesensing 15 03468 g008
Figure 9. Visual results of different fusion methods. (a1a3) are the visualization results of PAFPN; (b1b3) are the visual results of BiFPN; and (c1c3) are visualization results for MF-FPN.
Figure 9. Visual results of different fusion methods. (a1a3) are the visualization results of PAFPN; (b1b3) are the visual results of BiFPN; and (c1c3) are visualization results for MF-FPN.
Remotesensing 15 03468 g009
Figure 10. Detection results of different scenes in the UAVDT dataset. (a1a3) are the detection results of YOLOv7; and (b1b3) are the detection results of our CGMDet.
Figure 10. Detection results of different scenes in the UAVDT dataset. (a1a3) are the detection results of YOLOv7; and (b1b3) are the detection results of our CGMDet.
Remotesensing 15 03468 g010
Figure 11. Comparison between mAP0.5 and mAP for different improvement points. (a) represents the result of mAP0.5; (b) represents the result of mAP.
Figure 11. Comparison between mAP0.5 and mAP for different improvement points. (a) represents the result of mAP0.5; (b) represents the result of mAP.
Remotesensing 15 03468 g011
Figure 12. Example of heatmap visualization. The first column is the original image, the second column is the visualization result of YOLOv7, and the third column is the visualization result of CGMDet.
Figure 12. Example of heatmap visualization. The first column is the original image, the second column is the visualization result of YOLOv7, and the third column is the visualization result of CGMDet.
Remotesensing 15 03468 g012
Table 1. Change of parameters in the neck part of the model.
Table 1. Change of parameters in the neck part of the model.
ModuleBaselineOurs
ELAN-H_11.26 M0.1 M
ELAN-H_20.32 M0.07 M
ELAN-H_3\0.07 M
ELAN-H_4\0.21 M
ELAN-H_51.26 M0.85 M
ELAN-H_65.05 M3.4 M
Table 2. Comparison with state-of-the-art detectors on the VisDrone dataset.
Table 2. Comparison with state-of-the-art detectors on the VisDrone dataset.
MethodmAP0.5mAP0.75mAP AP S AP M AP L
RetinaNet [12]35.918.519.414.129.533.7
Cascade R-CNN [44]39.923.423.216.536.839.4
Faster R-CNN [9]40.020.621.515.434.637.1
YOLOv3 [15]31.415.316.48.326.736.9
YOLOX [45]45.026.626.717.437.945.3
YOLOv5l [17]36.220.120.512.429.936.4
HawkNet [46]44.325.825.619.936.039.1
QueryDet [47]48.128.828.3\\\
Edge YOLO [21]44.826.226.416.338.753.1
NWD [48]40.3\\22.2\\
ClusDet [23]50.624.726.717.638.951.4
DMNet [24]47.628.928.219.939.655.8
CEASC [49]50.728.428.7\\\
CDMNet [50]49.529.829.220.840.741.6
YOLOv7 [18]49.027.828.118.939.447.8
CGMDet(Ours)50.929.429.320.240.647.4
Table 3. Detection results for each category on the VisDrone dataset.
Table 3. Detection results for each category on the VisDrone dataset.
MethodPedestrianPeopleBicycleCarVanTruckTricycleAwing-TricycleBusMotormAP0.5
YOLOv3 [15]12.87.84.043.023.516.59.55.129.012.531.4
YOLOv5l [17]44.436.815.673.939.236.222.611.950.542.837.4
YOLOv7 [18]57.648.721.685.451.945.837.918.363.060.049.0
CGMDet (ours)59.750.725.486.253.447.437.920.266.361.650.9
Table 4. Comparison of detection results of tiny version on VisDrone dataset.
Table 4. Comparison of detection results of tiny version on VisDrone dataset.
MethodmAP0.5mAP0.75mAP AP S AP M AP L Params (M)Inference Time (ms)
YOLOX-tiny [45]35.719.219.712.228.331.75.0\
YOLOv5s [17]28.714.015.19.222.231.87.010.8
YOLO-UAVlite [51]36.619.720.612.929.333.41.4\
YOLOv7-tiny [18]35.817.318.611.427.436.66.09.4
CGMDet-tiny (ours)39.820.921.714.231.537.67.421.2
Table 5. Comparison between CGAM and similar methods.
Table 5. Comparison between CGAM and similar methods.
MethodmAP0.5mAP0.75mAPParam (M)
YOLOv7 + ELAN [37]49.027.828.136.5
YOLOv7 + CSPRepResStage [52]49.728.228.741.4
YOLOv7 + C2f [53]48.026.827.536.7
YOLOv7 + CGAM (ours)49.728.028.638.0
Table 6. Comparison of different multi-scale fusion methods.
Table 6. Comparison of different multi-scale fusion methods.
MethodmAP0.5mAP0.75mAPParam (M)GFLOPs
YOLOv7 + PAFPN49.027.828.136.53103.3
YOLOv7 + BiFPN48.627.227.935.97102.5
YOLOv7 + MF-FPN (ours)49.027.628.236.50102.9
Table 7. Comparison with state-of-the-art detectors on the UAVDT dataset.
Table 7. Comparison with state-of-the-art detectors on the UAVDT dataset.
MethodCarTruckBusmAP0.5mAP0.75mAP
YOLOv3 [15]30.83.926.436.320.620.4
YOLOX [45]39.45.725.337.926.123.5
YOLOX-tiny [45]40.50.522.036.122.721.0
YOLOv5l [17]80.712.745.246.230.727.7
YOLOv5s [17]78.113.345.645.028.626.8
YOLOv7 [18]78.412.644.045.029.426.8
YOLOv7-tiny [18]75.56.946.743.024.325.0
CGMDet (ours)81.316.047.448.032.629.1
CGMDet-tiny (ours)77.87.947.644.426.526.8
Table 8. Ablation experiments.
Table 8. Ablation experiments.
MethodCGAMMF-FPNFocal-EIOUPrecisionRecallF1-ScoremAP0.5Params (M)GFLOPs
YOLOv7 58.349.253.449.036.5103.3
B1 55.851.253.449.738.0104.7
B2 61.148.253.950.034.3104.0
B3 59.050.354.349.536.5103.3
B458.851.354.850.935.8105.3
Table 9. Model’s performance changes of different values of λ i .
Table 9. Model’s performance changes of different values of λ i .
λ 1 λ 2 λ 3 mAP0.5mAP0.75mAP
0.030.70.349.827.728.5
0.070.70.350.229.629.5
0.050.60.350.529.329.5
0.050.80.350.629.229.4
0.050.70.249.728.929.0
0.050.70.450.628.929.3
0.050.70.350.929.429.3
Table 10. Model performance at different learning rates.
Table 10. Model performance at different learning rates.
Learning RatemAP0.5mAP0.75mAP
0.00548.527.928.0
0.00849.428.928.9
0.00950.329.029.3
0.01050.929.429.3
0.01150.729.129.5
0.01250.429.629.4
Table 11. Experimental results of different categories on the VOC2012 validation set.
Table 11. Experimental results of different categories on the VOC2012 validation set.
MethodHorsePersonBottleDogTVCarAeroplaneBicycleBoatChairTable
Baseline81.183.255.380.565.678.180.479.660.456.961.5
Ours80.183.856.981.768.077.579.277.060.557.362.9
PlantTrainCatSofaBirdSheepMbikeBusCowHeadFootHand
49.784.686.966.869.974.482.682.169.013.58.910.8
50.882.586.064.968.674.583.082.164.817.715.712.2
Table 12. Experimental results of different scales on the VOC2012 validation set.
Table 12. Experimental results of different scales on the VOC2012 validation set.
Method AP S AP M AP L
Baseline14.232.055.9
Ours16.032.352.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, L.; Liu, Z.; Zhao, H.; Hou, Y.-E.; Liu, Y.; Zuo, X.; Dang, L. A Multi-Scale Object Detector Based on Coordinate and Global Information Aggregation for UAV Aerial Images. Remote Sens. 2023, 15, 3468. https://doi.org/10.3390/rs15143468

AMA Style

Zhou L, Liu Z, Zhao H, Hou Y-E, Liu Y, Zuo X, Dang L. A Multi-Scale Object Detector Based on Coordinate and Global Information Aggregation for UAV Aerial Images. Remote Sensing. 2023; 15(14):3468. https://doi.org/10.3390/rs15143468

Chicago/Turabian Style

Zhou, Liming, Zhehao Liu, Hang Zhao, Yan-E Hou, Yang Liu, Xianyu Zuo, and Lanxue Dang. 2023. "A Multi-Scale Object Detector Based on Coordinate and Global Information Aggregation for UAV Aerial Images" Remote Sensing 15, no. 14: 3468. https://doi.org/10.3390/rs15143468

APA Style

Zhou, L., Liu, Z., Zhao, H., Hou, Y. -E., Liu, Y., Zuo, X., & Dang, L. (2023). A Multi-Scale Object Detector Based on Coordinate and Global Information Aggregation for UAV Aerial Images. Remote Sensing, 15(14), 3468. https://doi.org/10.3390/rs15143468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop