Next Article in Journal
Development of a Fixed-Wing Drone System for Aerial Insect Sampling
Previous Article in Journal
LAP-BFT: Lightweight Asynchronous Provable Byzantine Fault-Tolerant Consensus Mechanism for UAV Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Multi-Scale Transformer for Object Detection in Aerial Scenes

Department of Mechanical Engineering, College of Field Engineering, Army Engineering University, PLA, Nanjing 210007, China
*
Author to whom correspondence should be addressed.
Drones 2022, 6(8), 188; https://doi.org/10.3390/drones6080188
Submission received: 1 July 2022 / Revised: 20 July 2022 / Accepted: 25 July 2022 / Published: 27 July 2022

Abstract

:
Deep learning has promoted the research of object detection in aerial scenes. However, most of the existing networks are limited by the large-scale variation of objects and the confusion of category features. To overcome these limitations, this paper proposes a novel aerial object detection framework called DFCformer. DFCformer is mainly composed of three parts: the backbone network DMViT, which introduces deformation patch embedding and multi-scale adaptive self-attention to capture sufficient features of the objects; FRGC guides feature interaction layer by layer to break the barriers between feature layers and improve the information discrimination and processing ability of multi-scale critical features; CAIM adopts an attention mechanism to fuse multi-scale features to perform hierarchical reasoning on the relationship between different levels and fully utilize the complementary information in multi-scale features. Extensive experiments have been conducted on the FAIR1M dataset, and DFCformer shows its advantages by achieving the highest scores with stronger scene adaptability.

1. Introduction

In recent years, aerial images have become an essential data source in earth remote sensing because they can provide a large amount of information and are easy to access and real-time solid [1]. They can meet the requirements of practical tasks such as resource exploration [2], environmental monitoring [3], urban planning [4], and precision agriculture [5]. With the application of computer vision detection in remote sensing, aerial object detection has become a fundamental and active research topic. However, considering the characteristics of aerial images, effectively detecting objects in aerial scenes is still challenging.
The significant difference in the inherent scale of object instances mainly leads to the contrast of object scale in the image scene. Unlike the ground scene, the aerial scene has a long sight distance and a large field of vision, and the scale differences of many types of instances in the background are significant. In addition, especially for the same target, it is related to the change in the image acquisition distance [6]. The significant shift in image acquisition distance in a large-scale aerial scene will lead to a certain extent object scale variations. The farther the distance, the smaller the object. As shown in Figure 1.
Besides, aerial images are usually collected from a bird’s eye view. The interferences in the large-field scene and the complex spatial distribution of ground objects confuse the object features in the background. In this case, different objects may show similar semantic features, while the semantic features of the same category of objects may differ significantly, as illustrated in Figure 2.
The core of aerial object detection is to obtain the distinguishing features of objects in different scenes. Generally, there are three methods to obtain the features. In the early research, handcraft features are the core element of aerial object detection. Researchers directly use simple attributes of objects (color [7], contour [8], texture [9]) and some carefully designed feature descriptors (scale-invariant feature transformation and oriented gradient histogram) to distinguish objects. These handcraft features often perform well in simple scenes, but they cannot well depict objects in complex scenes. Meanwhile, researchers adopt feature coding methods [10,11,12] to encode color, contour, or texture features to deeply capture distinguishing features and mine semantic information, but the performance is limited and cannot meet the needs of detection in complex scenes.
In recent years, data-driven deep learning algorithms have made remarkable achievements in object detection [13,14,15], and a series of CNN-based aerial detection methods have been proposed. To overcome the challenges (i.e., large-scale variations and confused distinguishing features) in aerial object detection, Yang et al. [16] proposed a detection framework (ClusDet) for regional object clustering. It can significantly improve the detection performance of tiny objects, implicitly model a priori context information, and dramatically reduce the inference time. Further, to overcome the challenge of tiny scale and low signal-to-noise ratio of aerial objects, Wang et al. [17] presented a dataset (AI-TOD) for detecting tiny objects in aerial scenes. Based on this, a multiple center points learning-based detection network (M-CenterNet) was proposed. Different from the compound network structure of ClusDet, Li et al. [18] designed a detection framework (DMNet) based on the region clipping of the density map, which improved the detection accuracy. Meanwhile, Deng et al. [19] designed an end-to-end global-local adaptive network (GLSAN) to solve the problem of dense small objects and uneven distribution. The global-local fusion strategy was integrated into a progressive scale-change framework, and the global contour information and local detail information were utilized to enhance the robustness of the network against the scale variations of objects. Unlike the previous methods that mainly focus on improving the accuracy but ignore the memory and computation cost, Li et al. [20] proposed a detection framework called CorrNet. CorrNet adopts the lightweight VGG16 as the backbone network and adopts the coarse-to-fine strategy to promote the light development of the aerial object detection algorithm.
Although the CNN-based networks have made significant progress, the localization of convolution operation limits its ability to obtain the global information of large-scale and complex aerial scene images. The pooling operation with the receptive field expansion dilutes the feature details. Motivated by the excellent performance of the transformer [21] in natural language processing, researchers began to apply the transformer to visual recognition. Dosovitskiy et al. [22] transformed the image into patch sequence processing, designed the vision transformer (ViT), and achieved good performance in benchmark classification recognition. Compared with convolution in object detection, the transformer can emphasize the long-term dependencies between image patches and reserve abundant spatial information through the self-attention of multi-heads [23]. Therefore, researchers attempt to employ transformers to achieve accurate detection in complex aerial scenes. Li et al. [24] combined CNN with a transformer to design a network with an encoding decoding structure called TRD, which achieves good performances on NWPU VHR-10 [25] and DIOR [26] datasets. By combining convolution and transformer, Zhang et al. [27] designed a multi-scale network termed ViT-YOLO. It shows more substantial semantic resolution, effectively alleviates category confusion, and significantly enhances the detection performance of aerial objects. Similarly, Zhu et al. [28] proposed a detector called TPH-YOLOv5 based on transformer structure, which achieves good scene capture performance and impressive interpretability.
To sum up, attributed to the advantages of transformers in obtaining more context information and learning diversified feature representation, embedding transformers into the detection framework has the potential to overcome the challenges in aerial object detection.
This study proposes an effectively multi-scale detection framework called DFCformer based on the transformer. DFCformer is mainly composed of DMViT (Depatch Multi-scale Vision Transformer), FRGC (Feature Reconciliation Guidance Component), and CAIM (Cascaded Attentional Interactive Module). DMViT is the backbone network, which abandons the fixed-size patch embedding in the PVT [29] and introduces the DePatch embedding [30] to mitigate the semantic damage caused by image splitting and retain the complete semantic information in a patch. Specifically, DMViT adopts a scale adaptive attention mechanism to enhance the ability of different attention heads at the same layer to model objects with different scales, improve the ability to save fine-grained details, and reduce the computational power consumption. Meanwhile, FRGC helps to break the barriers between feature layers, filter the interference information on different feature layers, and improve the purity of crucial details. Besides, to overcome the limitation of category confusion on network accuracy, CAIM is employed to better integrate features with inconsistent semantics and scales and improve the ability of network fine-grained feature interpretation. The main contributions of this study are summarized as follows:
  • A backbone network combining deformable patch embedding and a multi-scale visual converter is proposed to improve the ability to capture the details of multi-scale aerial targets in complex scenes.
  • The cross-layer feature reconciliation guidance component enhances the semantic information of crucial features on different layers, which helps to alleviate category confusion and realize accurate classification and regression.
  • The attention-based feature fusion mechanism strengthens the information integration between multi-scale and multi-semantic features and overcomes the limitation of category confusion on network accuracy.

2. Related Work

Relevant prior work includes ConvNets for object detection, vision transformer for object detection, and aerial object detection.

2.1. ConvNets for Object Detection

As a mainstream and standard deep learning model, ConvNets is often used in object detection. According to the use of the region-of-interest proposal step, the existing ConvNets models can be divided into two categories: two-stage models and one-stage models. The two-stage models [14,31] represented by R-CNN [32] achieve good performances. Following the single-stage structural design strategy, the Yolo series [33,34,35] and SSD [36] have attracted wide attention because of their simple network structure and high reasoning performance.
The remarkable achievements of these models based on ConvNets in general object detection tasks inspire us to extend object detection to aerial object detection tasks, promote the intellectual development of aerial scene perception, and expand the scope of engineering applications of computer vision.

2.2. Vision Transformer for Object Detection

The multi-head self-attention mechanism is the core of the transformer, which enables the transformer to learn the complex dependencies from sequence to sequence [37]. Vision transformer [22] (ViT) divides the image into non-overlapping image patch sequences, thus producing a novel image classification model and creating a precedent for the transformer to migrate to object detection tasks. Since the pioneering ViT model was proposed, researchers have designed many excellent models by optimizing the model structure and the component’s function. For example, DETR [38] abstracts objects detection as a prediction model of transformer and loss function, thus eliminating the dependence on handcrafted modules and operations (i.e., RPN and NMS). To overcome the limitation of the slow convergence speed of DETR, Zhu et al. [39] proposed Deformable DETR. Deformable DETR adopts a deformable attention mechanism, which improves the aggregation of cross-scale feature maps under controllable computational cost and keeps the balance between performance and inference rate. Following the attention optimization strategy, Swin [40] and PVT [29] establish flexible global attention to effectively reduce the computational complexity of the model and improve the detection performance. Meanwhile, a plethora of studies on optimizing the patch embedding has made exemplary achievements. T2T-ViT [41] processes tokens in the way of aggregation and recursion to ensure the integrity expression of local information. MPViT [42] builds multi-path patch embedding to realize fine feature and coarse feature representation at the same feature level, which helps to detect dense objects. Considering the different information expressed by varying levels of features, LVT [43] introduces Convolutional Self-Attention (CSA) to deal with the underlying features by including dynamic kernels and learnable filters. The introduction of Recursive Atrous Self-Attention (RASA) is conducive to extracting multi-scale context information and increasing the representation ability of marginal additional parameter cost. DeepViT [44] proposes a mechanism to regenerate attention, which increases the diversity of attention maps at different layers to control the computational cost and alleviate the performance saturation problem caused by network deepening. RVT [45] introduces position-aware attention scaling mechanism and patch-wise enhancement mechanism to improve the robustness and generalization ability of the transformer-based network. These excellent studies enhance the expansion of the transformer and promote us adopt the transformer in object detection.

2.3. Aerial Object Detection

With the release of large annotation datasets for aerial object detection, many researchers attempt to apply deep learning models to aerial object detection. Pang et al. [46] proposed an end-to-end detection framework R2-CNN to improve the reasoning speed and reduce severe false alarms in aerial object detection. Then, Pan et al. [47] enhanced the generalization training process of the model based on maximizing the alignment of neurons and improved the detection performance for dense objects in aerial scenes. Ma et al. [48] employed the deep separable convolution in the transformer, which significantly reduces the memory and computation cost of multi-scale features. Ran et al. [49] maintained the small-scale object in the down sampling operation without losing features, and they adopted the strategy of enhancing the attention of the feature image channel and fusing the multi-scale context information of the network. Xu et al. [50] combined reinforcement learning and strategy gradient to build a scale zoom detection network called AdaZoom, which improves the robustness of the network to multi-scale aerial objects. These methods provide significant reference values for the future study of aerial object detection. However, how to overcome the challenges of aerial object detection still needs further investigation.

3. Methods

This section mainly describes the structure and action mechanism of the three main components in DFCformer.
Based on the above analysis and comparison, this paper adopts a four-stage progressive feature fusion detection strategy and proposes a multi-scale detection architecture DFCformer based on the transformer structure to overcome the challenges, (i.e., large-scale variation and confused distinguishing features) in aerial object detection. As shown in Figure 3, DFCformer is divided into three parts. The first part takes DMViT as the backbone and integrates deformation patch embedding and multi-scale adaptive self-attention to improve the ability to capture features of aerial objects. See Section 3.1 for the details of DMViT. The second part is the FRGC, which guides the feature interaction at different layers to handle the significant difference between aerial object classes and high similarity within classes and promotes the fine-grained detection of the framework. The discussion of FRGC is presented in Section 3.2. The third part is the CAIM, which adopts the cascade attentional fusion strategy to conduct hierarchical reasoning on the relationship between different levels of features. Also, CAIM mines the rich complementary information in multi-scale features and enhances the overall performance of the framework. The description of the CAIM is presented in Section 3.3.

3.1. DMViT

To improve semantic recognition ability and reduce category confusion, this paper collects and correlates scene information from more prominent neighborhoods to infer the correlation between objects, which is the key to capturing feature information from aerial images of a comprehensive visual range of complex scenes. However, for traditional ConvNets, the local filtering of the convolution kernel limits its ability to obtain global context information. In contrast, the transformer can construct the dependency between image feature blocks globally and retain sufficient spatial information for object detection through self-attention.
A series of detection frameworks have been proposed based on transformer structure. However, these detection frameworks still have limitations in accurate detection of multi-scale aerial objects with a large field of view and complex background, and this is mainly reflected in the following two aspects: (1) Figure 4a shows that using fixed-size patch embedding to segment images cannot accurately capture the critical information of objects of different scales and the consistent characteristics of the same object under different geometric changes. (2) Figure 4b shows that the model largely ignores the multi-scale of objects in the attention layer. The potential attention mechanism of the model only depends on the marked static receptive field and the unified information granularity in an attention layer, so it cannot capture the features of different scales simultaneously.
To overcome this limitation, this paper proposes a new construction called MViT (Multi-scale vision transformer), as shown in Figure 5. Different attention heads at the same layer of MViT can effectively model objects of various sizes and explain coarse-grained and fine-grained characteristics. Also, MViT has good computation efficiency and can retain fine-grained details and obtain more discriminate information.
Transformer originates from natural language processing, and its core task is to deal with the mapping relationship between sequences. DePatch (Deformable Patch Embedding) relieves the constraints in the original patch embedding and endows patch embedding with deformable ability, thus better locating essential structures and retaining rich semantic information. The details of Depatch are illustrated in Figure 6.
Given an input image X H × W × C , DePatch divides X into N patches with a size of t × t ( t = H × W N ), and the sequence of N patches is donated as { s m | 1 m N } . This paper defines the DePatch to better explain the process. First, the patch s m in the input image is approximated to a rectangular region. The center coordinate of the rectangular region is ( x c m , y c m ) , and its left-top and right-bottom corner coordinates are ( x l m , y l m ) and ( x r m , y r m ) , respectively. The mathematical derivation is shown as follows:
x l m = x c m + α x m t w m 2
y l m = y c m + α y m t h m 2
x r m = x c m + α x m + t w m 2
y r m = y c m + α y m + t h m 2
Specifically, ( α x , α y ) is the predicted offset; t w m and t h m are the scale of the patch s m .
This paper predicts ( α x m , α y m , t w m , t h m ) for all patches and then embeds them with the rectangular region. The offset and scale can be shown as follows:
α x , α y = T a n h ( W o f f s e t L ( X ) )
t w , t h = R e L U ( T a n h ( W s c a l e L ( A ) + b s c a l e ) )
where L ( ) is a single linear layer; W o f f s e t and W s c a l e are adopted to predict the offset and scale, Considering that the raw image contains little semantic information, the first module is difficult to predict the offset and scale beyond its region, so these weights (offset and scale) are initialized to zero at the beginning of training; b s c a l e corrects the deviation of the initial state to ensure that each patch focuses on the same rectangular area. Considering that the size of the generated region is different and the predicted coordinate forms are diverse, the computational power consumption is increased. In this paper, interpolation sampling is adopted to simplify the calculation. By taking n × n sampling points evenly in the region, each sampling coordinate is ( p x i , p y i ) ( 1 i n × n ) , and the features of all sampled points { p i } 1 i n × n are flattened to generate DePatch, as shown in Equation (7).
s m = W D e P a t c h c o n c a t ( p 1 , p 2 , , p n × n ) + b D e P a t c h
The feature at each sampling point is calculated by bilinear interpolation in Equation (8).
X ( p x , p y ) = q x , q y G ( p x , p y ; q x , q y ) X ( q x , q y )
G ( p x , p y ; q x , q y ) = m a x ( 0 , 1 | p x q x | ) m a x ( 1 , 1 | p y q y | )
After the above operation and processing, the constraint of the fixed scale of patch size can be released. Meanwhile, the location and scale of each patch can better adapt to the distribution of instances and reduce the semantic damage caused by image segmentation.
The flexible segmentation style retains rich semantic information. To improve the perception ability to the feature information of multi-scale objects, inspired by reference [51], this paper proposes MSSA (multi-scale self-attention) to replace the vanilla multi-head self-attention in the transformer to obtain MViT (multi-scale vision transformer).
MSSA divides the self-attention into several subsets with different attention granularities, realizes scale variations, and enhances the modeling ability for multi-scale objects. Specifically, in the fine-grained subset, MSSA focuses on the expression of local detail information and weakens the modeling of long-distance context information. In the coarse-grained subset, MSSA efficiently aggregates global information, focuses on sensing long-distance up and down information, fully captures features, and reduces the computational power consumption. MSSA first projects the input sequence A H × W × C (generated by DePatch) into Q (query), K (key), and V (value). As shown in Figure 7, unlike the prior vision transformer, the scales of K and V are sorted according to a certain proportion:
Q c = X W c Q
K c = M A ( X , r c ) W c K
V c = M A ( X , r c ) W c V
where W c Q , W c K , and W c V are the linear projection parameters in the c-th attention head, M A ( ) denotes the multi-scale aggregation of the tokens, and r c is the adjustment of the rate in the c-th attention head.
Then, the multi-scale self-attention is calculated in Equation (13)
M S S A c = Softmax ( Q c K c T d m )
where d m is the dimension.
The multi-scale of K and V is conducive to capturing multi-scale features. However, the proportion r c increases the computational power consumption. When r c is too large, K and V become smaller, and the computational power consumption decreases, but the feature perception ability is weakened. On the contrary, when r c is too small, K and V become larger, and the feature perception ability is enhanced, but the computational power consumption increases suddenly. Therefore, the value of r c should be set under the balance of computational power consumption and model performance.

3.2. FRGC

The complex and changeable distribution characteristics of ground objects and the unpredictable environmental interference in the field of view make it more difficult to interpret aerial scenes. Based on the analysis of the overall operation of object detection, this paper proposes an efficient feature extractor DMViT, which only captures external information. To filter redundant interference information, this paper designs FRGC to optimize the information processing process and alleviate category confusion. The details of FRGC are presented in Figure 8.
Considering the different proportions of information carried by different feature layers and the different critical features required by object detection tasks (a large proportion of low-level feature texture information is conducive to regression subtasks, while a large proportion of high-level feature semantic information is conducive to classification subtasks), the FRGC adopts a guiding cross-layer interaction mechanism to break the barriers between feature layers, filters the interference information on different levels of features, and improves the information discrimination and processing ability of multi-scale critical features of components.
In Figure 8, FRGC is composed of multiple FRGC-R and FRGC-C units. It refines the semantic perception and texture expression of aerial objects layer by layer, efficiently interacts with different semantic information and texture information, and generates the critical features required by regression and classification subtasks. Specifically, in FRGC-R, the texture features suppress the background noise and help to extract the critical features required by regression from the detailed information of the object. Meanwhile, in FRGC-C, the semantic guiding features effectively suppress the background noise, restrict the diffusion of semantic features, and help to obtain more classification features.
FRGC-R operates in two steps: first, the features f i from DMViT are interacted to suppress the background noise and improve the proportion of regression information in the enhanced features f i .
R = A c ( f i ) A s ( f i )
f i = R + λ r e g ( σ ( R ) ) f i + f i
λ r e g ( x ) = { x x 0.5 1 x else
where f i denotes the enhanced features; A c denotes the channel-wise attention operation; A s denotes the spatial attention operation; denotes the tensor product; denotes element-wise multiplication; σ denotes the sigmoid function; λ r e g ( ) denotes the depression function, and it mainly suppresses the regions with high response in the regression features to promote the model to find potential visual clues and realize accurate positioning.
Then, FRGC-R guides the enhanced features for interactive perception and generates the critical features f i r required by the regression subtask.
f 1 = C ( C ( f i ) ) C ( f i + 1 )
f i r = C ( C ( f 1 ) + C ( f i ) )
where f 1 denotes the intermediate variable, and C ( ) denotes the CBR operation (Convolution, batch normalization, and ReLu).
FRGC-C operates also in two steps: the features f i from DMViT are guided to suppress background noise and improve the proportion of classification information in the enhanced features f i .
R = A c ( f i ) A s ( f i )
f i = f i + R + φ c l s ( σ ( R ) ) f i
φ c l s = 1 1 + e θ ( x 0.5 )
where f i denotes the enhanced features; φ c l s ( ) denotes the excitation response function, and it mainly focuses on the high response part of the feature map and filters out the positioning clues used for interference noise or regression; θ denotes a factor used to regulate the intensity of feature activation (it is set to 18 in our experiment). Because the regional response of some critical features can stimulate the role of high-response key classification features and realize accurate classification, the irrelevant features with an attention weight of less than 0.5 cannot reach the activation threshold and are suppressed. In this way, the interference of irrelevant classification areas can be eliminated, and the fitting performance and anti-misjudgment mechanism of the model are improved.
Then, FRGC-C guides the enhanced features for interactive perception and generates the critical features f i c required by the regression subtask.
f 2 = C ( C ( f i ) ) C ( f i + 1 )
f i c = C ( C ( f 2 ) + C ( f i ) )
where f 2 denotes the intermediate variable, and C ( ) denotes the CBR operation (Convolution, batch normalization, and ReLu).

3.3. CAIM

In Section 3.2 and Section 3.3, the efficient multi-scale feature information detector and the powerful critical feature processing operations are introduced, respectively. This section focuses on organically integrating the features at different levels to obtain the final prediction result.
Feature fusion is an unreachable part of the object detection network, which combines features from different layers or branches: (1) The information in low layers can be further enhanced by feature fusion; (2) The middle layers consider both semantic information and detail information and can adaptively adjust the proportion of different abstract information to realize better utilization of flexible features; (3) At the top layers, richer semantic information can be mined when considering the adjacent resolution.
For aerial object detection, to overcome these challenges (i.e., large-scale variations and confused distinguishing features), an intuitive method is adopted to establish a multi-stage detection framework and fuse multi-scale features for object prediction [52]. However, the feature fusion adopted in most aerial object detection networks [53,54,55] ignores the modeling and reasoning of the relationship between different scale features, which is not conducive to the localization of aerial objects and the mining of object features.
The attention mechanism imitates the human’s cognitive awareness of specific information. It enlarges essential details and pays more attention to vital aspects. In recent years, the attention mechanism has widespread use in visual recognition. The advantages of the attention mechanism in long-distance context modeling contribute to the interactive clustering of global features. Inspired by the study [56], this paper designs a cascade attention interaction module (CAIM) to enhance the modeling and reasoning of the relationship between different scale features.
The core idea of CAIM is that by changing the size of the spatial pool, channel attention can be realized on multiple scales. Meanwhile, considering the emphasis on different levels of feature expression and computational power consumption, CAIM adopts a cascade way to integrate multi-scale features step by step, as shown in Figure 9.
Figure 9 demonstrates the details of CAIM.
First, the features f i r and f i + 1 r from FRGC-R are processed by broadcast addition to obtain the intermediate feature f ¯ r . Then the intermediate feature f ¯ r is input into a feature aggregation component to realize point-to-point channel interaction of spatial location.
f ¯ r = σ ( B ( PC ( γ ( B ( PC ( f r ) ) ) ) ) )
where B ( ) is denotes batch normalization, PC ( ) denotes the point-wise convolution, γ ( ) denotes the ReLu operation, and σ denotes the sigmoid function.
f l r = α f i r f ¯ r + ( 1 α ) f i + 1 r f ¯ r
where denotes the element-wise multiplication, and α is the adjustment coefficient used to adjust the weight of features of different layers in combination features.
Similarly, the features f i c and f i + 1 c are fused in the aggregation process.
Firstly, the features f i c and f i + 1 c from FRGC-C are processed by broadcast addition to obtain the intermediate feature f ¯ c . Then, the combined feature f ¯ c is input into a feature aggregation component to realize the point-to-point channel interaction of spatial location based on global average pooling operation.
f ¯ c = σ ( PC ( γ ( B ( PC ( GP ( f c ) ) ) ) ) )
where GP ( ) denotes the global average pooling operation.
Finally, the intermediate feature f ¯ c is combined with the original feature to obtain the classification combined feature f h c .
f h c = β f i c f ¯ c + ( 1 β ) f i + 1 c f ¯ c
where β is the adjustment coefficient used to adjust the weight of features of different layers in combination features.
Based on the above fusion branch interaction, the combined features f l r and f h c are processed in a similar way to obtain the final multi-scale fusion feature f m s c .
f m s c = δ f l r f ¯ + ( 1 δ ) f h c f ¯
f ¯ = σ ( f l r + f h c )
where f ¯ is the combined feature of f l r and f h c , and δ is the adjustment coefficient used to adjust the weight of features of different layers in the final multi-scale fusion feature.

3.4. Network Configuration

Several essential components of DFCformer are introduced in detail in Section 3.1, Sectiom 3.2 and Section 3.3 above. Designing an efficient network based on these three components is the core content of our discussion in this section. From the perspective of system theory, the whole is greater than the sum of parts, and each component should be organically connected to form a reasonable structure to maximize the overall performance. An appropriate design for a deep learning network is not the simple stacking of various elements but finding a balance between performance and computational power consumption [57].
The core component of the DFCformer is the backbone, and the remaining two parts are configured according to the structural characteristics of the backbone. Meanwhile, the DMViT in each stage of the backbone is composed of one Depatch and several MViTs. Therefore, this paper mainly analyzes the configuration of MViT in DMViT in different stages. Inspired by the reference [40], this paper focuses on comparing the DMViT under three configurations, as shown in Table 1.
In addition, this paper selects Top-1 accuracy to measure the detection accuracy of the three variants, i.e., DMViT-S, DMViT-M, and DMViT-L, and selects flops to measure the computational cost of the variants on the ImageNet-1k [58], as shown in Figure 10. Overall, the three variants perform well, and the worst accuracy is more than 77%, indicating the excellent detection performance of DMViT. Specifically, DMViT-S has the simplest structure and the lowest computational cost, but its accuracy is the lowest. On the contrary, DMViT-L achieves the highest accuracy of 82.2%, but its structure is the most complex, and the computational cost is 8.3 G FLOPs, which is 2 times that of DMViT-S. Compared with the DMViT-S, DMViT-M has a more complex structure and higher computational cost, but its accuracy is increased by 2.9%. Meanwhile, compared with DMViT-L, the accuracy of DMViT-M is decreased by 0.5%, but its computational power consumption is reduced by 22%. To sum up, DMViT-M achieves good performance under an acceptable structural complexity and computational cost. Therefore, this paper adopts DMViT-M as the backbone.

3.5. Multi-Task Learning

Aerial object detection is a multi-task learning problem, including regression and classification subtasks. The classification subtask is to learn the identification features of the key or prominent parts of the object [59]. The regression subtask is to locate the whole object and its boundary accurately. This paper further proposes multi-task learning, which includes (1) screening high-quality anchors, and (2) optimizing the calculation mode of the branch task loss.
Considering that the classification score and the IoU between the prediction box and the ground truth are the evaluation indicators of the two subtasks, this paper uses the high-order combination of classification score and IoU to screen the anchor box. Specifically, the following indicators are designed to calculate the quality of the anchors to encourage the network to dynamically focus on high-quality anchors from the perspective of joint optimization.
t = s η × o τ
where η and τ are used to adjust the influence of anchors in the classification score ( s ) and IoU value ( o ), respectively.
To improve the classification score of high-quality anchors and reduce the score of low-quality homogeneous anchors, this paper employs the maximum IoU value ( o ) in each instance corresponding to the maximum value of t . The binary cross-entropy (BCE) on the positive anchors for the classification task is defined as:
L b c e = i = 1 N p o s B C E ( s i , t max )
where i is the i-th positive anchor in an instance, and t max is the maximum value of t .
The final classification task loss function is defined as:
L c l s = i = 1 N p o s | s i t max | ω B C E ( s i , t max ) + j = 1 N n e g s j ω B C E ( s j , 0 )
where ω is the scale parameter, and j is the j-th negative anchor in an instance.
Similarly, to increase the proportion of high-quality anchor boxes in the regression task and reduce the proportion of low-quality anchor boxes in the regression task, this paper recalculates the GIoU loss [60] on each anchor box based on t max .
L r e g = i = 1 N p o s t max L G I o U ( b i , g i )
where g i and b i denote the ground-truth box and the predicted bounding box, respectively.
The total loss is the sum of L c l s and L r e g :
L = υ L c l s + ς L r e g
where υ and ς are the focusing parameters.

4. Experiment

4.1. Dataset

Object detection is a data-driven application in computer vision. The performance of the deep learning network depends on the quality and quantity of the given data. Different from raw image data sets such as ImageNet [58] and MSCOCO [61], a challenging and excellent aerial object detection data set should have the following properties:
  • The scale of the dataset is huge. With the increasing demand for aerial object detection, the detector needs higher critical generalization. Many algorithms have been proposed and perform well on small datasets, but their performance decreases rapidly on large datasets. Therefore, to train the detector more comprehensively, the corresponding dataset needs a large volume of object instances and images.
  • Data samples are rich in detail. The similarity within the aerial class increases the demand for fine-grained recognition. An excellent detection algorithm should be able to correctly identify objects belonging to specific subcategories. Most of the existing datasets contain coarse-grained information and lack detailed information, which makes it difficult to improve the detection performance of deep learning methods.
  • The image quality of data samples is high. Factors such as rain, fog, cloud, and jitter may interfere with the quality of aerial images. It is challenging to train excellent algorithms with low-quality samples to meet the requirements of aerial object detection.
Based on the above analysis, this paper comprehensively evaluates the standard aerial object datasets (DIOR [26], DOTA [62], VisDrone [63], VHR-10 [25], FAIR1M [64], RSOD [65], VEDAI [66], and HRSC2016 [67]) in terms of sample size (Images), the number of instances (Instances), type distribution (Categories), and sample quality (Quality) to select the best dataset. The results presented in Figure 11 show that these datasets perform differently in the four indicators, and FAIR1M has more significant advantages in all-around performance than other datasets. Thus, this paper selects FAIR1M as the data source.
FAIR1M is a new benchmark dataset for fine-grained object detection in high-resolution aerial scenes. Specifically, the training set and test set of the FAIR1M contain 16,488 images and 8137 images, respectively. The overall data set is composed of 5 categories and 37 subcategories. All classification and example statistics are illustrated in Figure 12.

4.2. Evaluations Metrics

In object detection, I o U is the most widely used index to measure the overlap between the prediction box (P) and the ground-truth box (G). However, the index that measures only from the geometric dimension cannot well reflect the quality of the information extracted by the prediction box. As shown in Figure 13, the I o U s of (a) and (b) are the same, but there is more interference in the information extracted by the prediction frame of a, which is not conducive to the identification of the instance category of the scene.
Referring to conclusions from the study [67], this paper selects F I o U as the evaluation index to penalize the exceptional results. Similar to the definition of I o U , T P , F P , T N , F N , A P F , and m A P F are defined for F I o U .
F I o U = P G G P P G G P G P 3
Precision F = F I o U T P s c o r e T P T P s c o r e T P + F P s c o r e F P
Recall F = F I o U T P T P + F N
where s c o r e denotes the classification score of the prediction box.

4.3. Implementation Details

To better fit the limited computing power of the drone platform, the computer equipped with NVIDIA rtx3090 GPU is selected as the experimental platform. Following the setting of the Swin Transformer [40], this paper uses ImageNet-1k to pre-train all transformer-based backbones and fine-tunes them on the training set of FAIR1M. Other convolution-based detection frameworks are only trained on the training set of FAIR1M. Meanwhile, this paper uses the AdamW optimizer to train DFCformer. The initial learning rate is set to 0.0001, and the weight attenuation is 0.005. First, 6 epochs are taken for warm-up training, and the learning rate is updated by one-dimensional linear interpolation. After warm-up training, the cosine annealing function updates the learning rate. The parameters of each part of the loss function are dynamically updated in the training process. Finally, the DFCformer becomes stable at 71 epochs with a batch size of 32, and the focusing parameters υ and ς are 0.6 and 0.5, respectively.

4.4. Comparison with the State-of-the-Art

Our method is compared with several counterparts, including CorrNet [20], ClusDet [16], GLSAN [19], M-CenterNet [17], DRN [44], TPH-YOLOv5 [28], and O2DETR [48]. Table 2 summarizes the detection sub-classification results. Overall, DFCformer has significant performance advantages over other networks, with a mAP of 40.14% and an mAPF of 13.12%, respectively. Although other networks obtain similar results on mAP, the results of mAPF show the difference in detection ability for similar objects.
In addition, the results of each subcategory have obvious discrimination, which meets the distribution of instances in each subcategory in Figure 12. Specifically, for objects with a large scale and apparent features, such as small cars and intersections, the networks perform well and can make accurate detections. On the contrary, the networks show poor performance and low feature classification scores for instances with a small volume, such as ARJ21 and C919. Further, due to the transformer’s intense need for data, the ability of the DFCformer to detect few-shot objects is lowest. Scarce objects and fine-grained objects are always challenging for visual recognition. In this case, the volume of data limits the evolution of network capacity, and the network is in a severe overfitting state. Meanwhile, the network’s sensitivity to subtle key features is too low, so it is difficult to capture the features of the objects and accurately identify the category of the objects.
Table 2 shows the test comparison results of the network on the FAIR1M dataset. On this basis, we compare the accuracy and speed of different networks in Figure 14. Regarding reasoning speed, CorrNet and TPH-YOLOv5 have significant advantages, reaching 31 FPS and 29 FPS, respectively. However, aerial target detection is an engineering application task. The balance between network reasoning speed and performance determines the ability of the network to solve the actual aerial target detection task. Therefore, we should make a comprehensive evaluation from two aspects: reasoning speed and detection accuracy. Although the reasoning speed of the DFCformer does not exceed that of CoorNet and TPH-YOLOv5, its 27 FPS reasoning speed still has certain advantages over the rest of the networks. More importantly, its detection accuracy has apparent advantages over other networks. By comprehensively analyzing the performance of reasoning speed and detection accuracy, the DFCformer has a solid ability to solve aerial target detection tasks.
Figure 15 shows the detection results of the networks in three aerial scenes. Figure 15a shows the aerial images of urban residential areas with complex backgrounds. In this scene, the automobile, truck, and other examples are distributed along the road. Due to their tiny pixel proportion, most networks in the experiment fail to capture compelling features to make accurate detections. The proposed DFCformer can only alleviate the limitation of minimal object feature extraction to a certain extent, and it is difficult to identify different types of vehicles accurately. The yellow dotted box in the figure indicates the complex area in the scene. The baseball field will be attached to the surrounding training area in the actual construction, which increases the inter-domain span between the natural setting and the training samples, and the network’s generalization ability is insufficient to overcome such an inter-domain span. All networks except DFCformer mistakenly detect the training field of the baseball field as a football field. DFCformer can overcome the inter-domain span migration, but it still fails to reflect the accurate edge distribution of the baseball field.
Figure 15b is an aerial image of the airport. This scene has a simple background, and the objects are distributed longitudinally along with the terminal. All networks can detect airplanes, but different networks have significant differences in the subdivision ability of similar objects. The convolution-based networks (M-CenterNet, CorrNet, ClusDet, and GLSAN) focus on acquiring local features but ignore the establishment of long-distance feature mapping. So, these networks cannot classify different objects accurately and even mistakenly detect the terminal as A220. On the contrary, the transformer-based networks focus on global feature information processing to capture rich features, and they can accurately identify similar objects.
Figure 15c shows an aerial image of the water area. It is challenging for the network to realize the fine classification of fishing boats, cruise ships, and motorboats. The similarity between the classes is still a challenge for the aerial detectors. Besides, the yellow dotted line area in the scene shows densely distributed objects with a small proportion of pixels, and almost all networks cannot accurately detect the objects in this area. In contrast, the proposed DFCformer is superior to other networks in terms of locating accuracy, classification accuracy, and prediction box quality.
The above analysis and the test results in Table 2 indicate that DFCformer has higher regression accuracy and classification scores than other networks, indicating that the transformer-based backbone network, feature reconciliation guidance component, and cascaded attentional interactive module in this paper can alleviate the limitations of current detection to a certain extent, In the follow-up work, we will further improve the fine classification ability of the network and improve the accuracy of similar object detection.

4.5. Ablation Study

The comparative experiment in Section 4.4 shows the excellent detection performance of the DFCformer. To further verify the effectiveness of DFCformer and the contribution of each component to the overall network, this section sets up ablation experiments under the same experimental conditions.
Analysis of the effectiveness of DMViT. To evaluate the effectiveness and the feature interpretation mechanism of DMViT, this paper chooses two standard backbones (ResNet101 [68] and PVT [29]) as substitutes for DMViT. As shown in Table 3, the experimental results of different combinations are different. Specifically, the mAP of taking DMViT as the backbone is 3.23% and 5.13% higher than that of taking PVT and ResNet101 as the backbone. Meanwhile, the mAPF of taking DMViT as the backbone is the best, which is about 3.1% higher than the worst.
Figure 16 shows the experimental results in different scenes. By comprehensively analyzing the results in the three scenes, due to the limited receptive field, ResNet101 only captures the local feature information. On the contrary, PVT only focuses on the global feature mapping and ignores the detailed information, thus reducing the separability of background and foreground. Due to the influence of deformable patches and multi-scale attention, DMViT considers multi-scale long-distance modeling and focuses on local detail information, so it can obtain fine-grained information of multi-scale instances at the global level and capture rich instance features.
A strong backbone plays an irreplaceable role in aerial object detection networks. DMViT makes an outstanding contribution to the overall performance of DFCformer with its comprehensive and accurate feature expression ability.
Analysis of the effectiveness of FRGC. To verify the effectiveness and the mechanism of FRGC, this paper replaces FRGC with the ordinary linear interpolation sampling operation (S) to process the features of adjacent layers. Considering the experimental cycle and the characteristics of FRGC’s feature processing, this paper separately takes the airplane sub-dataset in FAIR1M as the data source of this experiment. Then, two groups of experimental results are counted and presented in the form of a confusion matrix. As shown in Figure 17, after replacing FRGC with a linear interpolation sampling operation, the detection accuracy of specific examples decreases seriously, and the largest accuracy decrease reaches 23%. These degradations show that the simple sampling operation cannot efficiently deal with different feature layers. Also, the insufficient attention to fine-grained details and the imbalance of the sample number limit the network detection performance.
Analysis of the effectiveness of CAIM. To study the impact of the CAIM, this paper replaces the CAIM with standard feature fusion approaches (concatenation and addition) in the DFCformer. Figure 18 shows the comparison results. Comparing concatenation and addition, it can be found that our CAIM performs better in the two sub-datasets. The superior performances indicate that the CAIM helps to improve the multi-scale aerial object detection ability of the network.

5. Conclusions

To overcome the challenges in aerial object detection such as large-scale variations and confused distinguishing features, this paper proposes a novel and powerful aerial object detection framework called DFCformer. The whole framework includes three parts: DMViT, FRGC, and CAIM. DMViT is the backbone, which introduces deformation patch embedding and multi-scale adaptive self-attention to improve the ability to capture small-scale objects in aerial images with a large field of view. FRGC is the feature coordination and guidance component, which guides feature interaction layer by layer according to the focuses of feature semantic information expression at different levels. CAIM is the cascade attention interactive module, which adopts a cascade attention fusion strategy to perform hierarchical reasoning on the relationship between different levels of features. Finally, a serious of experiments are conducted on the FAIR1M, and the DFCformer achieves the best performance, highlighting its effectiveness. The disadvantage of DFCformer is that due to the use of a transformer-based network, the minimum size of configuration parameters reaches 65M, which is not easy for unmanned aerial vehicle platforms. Our future works will lightweight the network. We hope this attempt could promote the development of fine-grained object recognition in aerial scenes and broaden the application scope of the transformer.

Author Contributions

Conceptualization, X.H. and Q.W.; methodology, F.S.; software, H.W.; validation, G.L. and J.W.; formal analysis, G.L.; investigation, G.L.; resources, J.W.; data curation, G.L.; writing—original draft preparation, G.L.; writing—review and editing, G.L.; visualization, X.H.; supervision, F.S.; project administration, Q.W.; funding acquisition, F.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No.61671470), and the Key Research and Development Program of China (No. 2016YFC0802900).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data in the study comes from the public data set, and its link address is gaofen-challenge.com [11 May 2022].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bi, Q.; Qin, K.; Li, Z.; Zhang, H.; Xu, K.; Xia, G.S. A multiple-instance densely-connected ConvNet for aerial scene classification. IEEE Trans. Image Process. 2020, 29, 4911–4926. [Google Scholar] [CrossRef]
  2. Pérez-Álvarez, R.; Sedano-Cibrián, J.; de Luis-Ruiz, J.M.; Fernández-Maroto, G.; Pereda-García, R. Mining exploration with UAV, low-cost thermal cameras and GIS tools—application to the specific case of the complex sulfides hosted in Carbonates of Udías (Cantabria, Spain). Minerals 2022, 12, 140. [Google Scholar] [CrossRef]
  3. Padró, J.C.; Muñoz, F.J.; Planas, J.; Pons, X. Comparison of four UAV georeferencing methods for environmental monitoring purposes focusing on the combined use with airborne and satellite remote sensing platforms. Int. J. Appl. Earth Obs. Geoinf. 2019, 75, 130–140. [Google Scholar] [CrossRef]
  4. Latha, T.P.; Sundari, K.N.; Cherukuri, S.; Prasad, M.V.V.S.V. Remote sensing UAV/Drone technology as a tool for urban development measures in APCRDA. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 42, 525–529. [Google Scholar] [CrossRef] [Green Version]
  5. Delavarpour, N.; Koparan, C.; Nowatzki, J.; Bajwa, S.; Sun, X. A technical study on UAV characteristics for precision agriculture applications and associated practical challenges. Remote Sens. 2021, 13, 1204. [Google Scholar] [CrossRef]
  6. He, P.; Jiao, L.; Shang, R.; Wang, S.; Liu, X.; Quan, D.; Zhao, D. MANet: Multi-scale aware-relation network for semantic segmentation in aerial scenes. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5624615. [Google Scholar] [CrossRef]
  7. Swain, M.J.; Ballard, D.H. Color indexing. Int. J. Comput. Vis. 1991, 7, 11–32. [Google Scholar] [CrossRef]
  8. Gevers, T.; Smeulders, A.W.M. Pictoseek: Combining color and shape invariant features for image retrieval. IEEE Trans. Image Process. 2000, 9, 102–119. [Google Scholar] [CrossRef] [PubMed]
  9. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  10. Philbin, J.; Chum, O.; Isard, M.; Sivic, J.; Zisserman, A. Object Retrieval with Large Vocabularies and Fast Spatial Matching. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar]
  11. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent dirichlet allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  12. Hofmann, T. Unsupervised learning by probabilistic latent semantic analysis. Mach. Learn. 2001, 42, 177–196. [Google Scholar] [CrossRef]
  13. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  14. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  15. Xu, X.; Feng, Z.; Cao, C.; Li, M.; Wu, J.; Wu, Z.; Shang, Y.; Ye, S. An improved Swin Transformer-Based Model for Remote Sensing Object Detection and Instance Segmentation. Remote Sens. 2021, 13, 4779. [Google Scholar] [CrossRef]
  16. Yang, F.; Fan, H.; Chu, P.; Blasch, E.; Ling, H. Clustered Object Detection in Aerial Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  17. Li, C.; Yang, T.; Zhu, S.; Chen, C.; Guan, S. Density Map Guided Object Detection in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  18. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.S. Tiny Object Detection in Aerial Images. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar]
  19. Deng, S.; Li, S.; Xie, K.; Song, W.; Liao, X.; Hao, A.; Qin, H. A global-local self-adaptive network for drone-view object detection. IEEE Trans. Image Process. 2020, 30, 1556–1569. [Google Scholar] [CrossRef] [PubMed]
  20. Li, G.; Liu, Z.; Bai, Z.; Lin, W.; Ling, H. Lightweight salient object detection in optical remote sensing images via feature correlation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5617712. [Google Scholar] [CrossRef]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  22. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Hendria, W.F.; Phan, Q.T.; Adzaka, F.; Jeong, C. Combining transformer and CNN for object detection in UAV imagery. ICT Express 2021, in press. [CrossRef]
  24. Li, Q.; Chen, Y.; Zeng, Y. Transformer with transfer CNN for remote-sensing-image object detection. Remote Sens. 2022, 14, 984. [Google Scholar] [CrossRef]
  25. Cheng, G.; Zhou, P.; Han, J. Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  26. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Lu, X.; Cao, G.; Yang, Y.; Jiao, L.; Liu, F. ViT-YOLO: Transformer-Based YOLO for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  28. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 based on Transformer Prediction Head for Object Detection on Drone-Captured Scenarios. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  29. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  30. Chen, Z.; Zhu, Y.; Zhao, C.; Hu, G.; Zeng, W.; Wang, J.; Tang, M. Dpt: Deformable Patch-based Transformer for Visual Recognition. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Conference, 20–24 October 2021. [Google Scholar]
  31. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  32. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  33. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  34. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  35. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  36. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single Shot Multibox Detector. In Proceedings of the European Conference on Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  37. Wang, D.; Zhang, J.; Du, B.; Xia, G.S.; Tao, D. An empirical study of remote sensing pretraining. IEEE Trans. Geosci. Remote Sens. 2022; early access. [Google Scholar] [CrossRef]
  38. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. arXiv 2020, arXiv:2005.12872. [Google Scholar]
  39. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2021, arXiv:2010.04159. [Google Scholar]
  40. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  41. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.H.; Tay, F.E.; Feng, J.; Yan, S. Tokens-to-Token Vit: Training Vision Transformers from Scratch on Imagenet. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  42. Lee, Y.; Kim, J.; Willette, J.; Hwang, S.J. MPViT: Multi-path vision transformer for dense prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  43. Yang, C.; Wang, Y.; Zhang, J.; Zhang, H.; Wei, Z.; Lin, Z.; Yuille, A. Lite vision transformer with enhanced self-attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  44. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. Deepvit: Towards deeper vision transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  45. Mao, X.; Qi, G.; Chen, Y.; Li, X.; Duan, R.; Ye, S.; He, Y.; Xue, H. Towards robust vision transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  46. Pang, J.; Li, C.; Shi, J.; Xu, Z.; Feng, H. R2-CNN: Fast tiny object detection in large-scale remote sensing images. arXiv 2019, arXiv:1902.06042. [Google Scholar]
  47. Pan, X.; Ren, Y.; Sheng, K.; Dong, W.; Yuan, H.; Guo, X.; Ma, C.; Xu, C. Dynamic Refinement Network for Oriented and Densely Packed Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  48. Ma, T.; Mao, M.; Zheng, H.; Gao, P.; Wang, X.; Han, S.; Doermann, D. Oriented object detection with transformer. arXiv 2021, arXiv:2106.03146. [Google Scholar]
  49. Ran, Q.; Wang, Q.; Zhao, B.; Wu, Y.; Pu, S.; Li, Z. Lightweight oriented object detection using multiscale context and enhanced channel attention in remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 5786–5795. [Google Scholar] [CrossRef]
  50. Xu, J.; Li, Y.; Wang, S. AdaZoom: Adaptive zoom network for multi-scale object detection in large scenes. arXiv 2021, arXiv:2106.10409. [Google Scholar]
  51. Ren, S.; Zhou, D.; He, S.; Feng, J.; Wang, X. Shunted Self-Attention via Multi-Scale Token Aggregation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 21–24 June 2022. [Google Scholar]
  52. Zhang, K.; Shen, H. Multi-stage feature enhancement pyramid network for detecting objects in optical remote sensing images. Remote Sens. 2022, 14, 579. [Google Scholar] [CrossRef]
  53. Lee, C.; Park, S.; Song, H.; Ryu, J.; Kim, S.; Kim, H.; Pereira, S.; Yoo, D. Interactive Multi-Class Tiny-Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  54. Lu, X.; Ji, J.; Xing, Z.; Miao, Q. Attention and feature fusion SSD for remote sensing object detection. IEEE Trans. Instrum. Meas. 2021, 70, 5501309. [Google Scholar] [CrossRef]
  55. Yuan, W.; Xu, W. MSST-Net: A multi-scale adaptive network for building extraction from remote sensing images based on Swin transformer. Remote Sens. 2021, 13, 4743. [Google Scholar] [CrossRef]
  56. Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional Feature Fusion. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Virtual Conference, 5–9 January 2021. [Google Scholar]
  57. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jégou, H.; Douze, M. LeViT: A Vision Transformer in ConvNet’s Clothing for Faster Inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual Conference, 11–17 October 2021. [Google Scholar]
  58. Deng, J.; Dong, W.; Socher, R.; Li, J.L.; Li, K.; Li, F.F. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  59. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D Deep Learning Approach for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  60. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  61. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft Coco: Common Objects in Context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  62. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  63. Zhu, P.; Du, D.; Wen, L.; Bian, X.; Ling, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-VID2019: The Vision Meets Drone Object Detection in Video Challenge Results. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Seoul, Korea, 27–28 October 2019. [Google Scholar]
  64. Sun, X.; Wang, P.; Yan, Z.; Xu, F.; Wang, R.; Diao, W.; Fu, K. FAIR1M: A benchmark dataset for fine-grained object recognition in high-resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 116–130. [Google Scholar] [CrossRef]
  65. Xiao, Z.; Liu, Q.; Tang, G.; Zhai, X. Elliptic fourier transformation-based histograms of oriented gradients for rotationally invariant object detection in remote-sensing images. Int. J. Remote Sens. 2015, 36, 618–644. [Google Scholar] [CrossRef]
  66. Razakarivony, S.; Jurie, F. Vehicle detection in aerial imagery: A small target detection benchmark. J. Vis. Commun. Image Represent. 2016, 34, 187–203. [Google Scholar] [CrossRef] [Green Version]
  67. Liu, Z.; Yuan, L.; Weng, L.; Yang, Y. A High Resolution Optical Satellite Image Dataset for Ship Recognition and Some New Baselines. In Proceedings of the 6th International Conference on Pattern Recognition Applications and Methods, Porto, Portugal, 24–26 February 2017. [Google Scholar]
  68. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
Figure 1. Comparing the scale variation in nature scenes: (a) with those in aerial scenes (b,c). The scales of different instances in group; (b) differ significantly, while the same instance in group; (c) varies considerably.
Figure 1. Comparing the scale variation in nature scenes: (a) with those in aerial scenes (b,c). The scales of different instances in group; (b) differ significantly, while the same instance in group; (c) varies considerably.
Drones 06 00188 g001
Figure 2. The challenges of object detection in aerial scenes. The red dashed box in group: (a) shows that the same objects category is significantly different in different scenes. In contrast, the yellow dashed box in group; (b) shows that different object categories are similar in different scenes.
Figure 2. The challenges of object detection in aerial scenes. The red dashed box in group: (a) shows that the same objects category is significantly different in different scenes. In contrast, the yellow dashed box in group; (b) shows that different object categories are similar in different scenes.
Drones 06 00188 g002
Figure 3. The structure of the DFCformer. The DMViT consists of deformable patch embedding and a multi-scale visual transformer. FRGC-C and FRGC-R are the feature processing units of the classification task and regression task in FRGC, respectively. CAIM is a cascade attention interaction module to enhance the modeling and reasoning of the relationship between different scale features.
Figure 3. The structure of the DFCformer. The DMViT consists of deformable patch embedding and a multi-scale visual transformer. FRGC-C and FRGC-R are the feature processing units of the classification task and regression task in FRGC, respectively. CAIM is a cascade attention interaction module to enhance the modeling and reasoning of the relationship between different scale features.
Drones 06 00188 g003
Figure 4. The limitations of the transformer-based networks in aerial object detection. (a) The original patch embedding module divides the image in a fixed way. It sometimes destroys the semantics of objects. (b) The size of the circle represents the receptive field size of the token, and the number of circles represents the number of tokens in the self-attention calculation, both of which reflect the computation cost. ViT focuses on fine-grained objects, but it has an extremely heavy computation cost. PVT reduces the computation cost, but it only focuses on coarse-grained large objects and ignores fine-grained objects.
Figure 4. The limitations of the transformer-based networks in aerial object detection. (a) The original patch embedding module divides the image in a fixed way. It sometimes destroys the semantics of objects. (b) The size of the circle represents the receptive field size of the token, and the number of circles represents the number of tokens in the self-attention calculation, both of which reflect the computation cost. ViT focuses on fine-grained objects, but it has an extremely heavy computation cost. PVT reduces the computation cost, but it only focuses on coarse-grained large objects and ignores fine-grained objects.
Drones 06 00188 g004
Figure 5. The details of MViT, MA (Multi-scale aggregation) operations of the tokens enhance the multi-scale adaptability of MViT to capture multi-scale features.
Figure 5. The details of MViT, MA (Multi-scale aggregation) operations of the tokens enhance the multi-scale adaptability of MViT to capture multi-scale features.
Drones 06 00188 g005
Figure 6. Instruction of the depatch embedding.
Figure 6. Instruction of the depatch embedding.
Drones 06 00188 g006
Figure 7. Comparisons with different self-attention mechanisms in ViT, PVT, Swin, and DMViT. The global self-attention ViT focuses on small-scale feature maps. Swin focuses on the local region of large-scale feature maps. In the self-attention of PVT, Q focuses on the global while K and V focus on the local. Unlike the first three self-attention mechanisms, DMViT introduces multi-scale token aggregation in the self-attention to obtain the keys and values of different scales.
Figure 7. Comparisons with different self-attention mechanisms in ViT, PVT, Swin, and DMViT. The global self-attention ViT focuses on small-scale feature maps. Swin focuses on the local region of large-scale feature maps. In the self-attention of PVT, Q focuses on the global while K and V focus on the local. Unlike the first three self-attention mechanisms, DMViT introduces multi-scale token aggregation in the self-attention to obtain the keys and values of different scales.
Drones 06 00188 g007
Figure 8. Illustration of the FRGC. FRGC-C and FRGC-R are the feature processing units of the classification task and regression task in FRGC, respectively.
Figure 8. Illustration of the FRGC. FRGC-C and FRGC-R are the feature processing units of the classification task and regression task in FRGC, respectively.
Drones 06 00188 g008
Figure 9. Illustration of the proposed CAIM, which is a cascade attention interaction module to enhance the modeling and reasoning of the relationship between different scale features.
Figure 9. Illustration of the proposed CAIM, which is a cascade attention interaction module to enhance the modeling and reasoning of the relationship between different scale features.
Drones 06 00188 g009
Figure 10. Quantitative comparisons among the variants.
Figure 10. Quantitative comparisons among the variants.
Drones 06 00188 g010
Figure 11. Multi-dimensional comparisons of typical datasets in the field of aerial object detection.
Figure 11. Multi-dimensional comparisons of typical datasets in the field of aerial object detection.
Drones 06 00188 g011
Figure 12. The specific instance distribution of dataset categories.
Figure 12. The specific instance distribution of dataset categories.
Drones 06 00188 g012
Figure 13. Detection results in two samples (Red: ground truth, Blue: prediction). (a,b) are the same in IoU, but the predicted bounding box in (b) contains less confusing information.
Figure 13. Detection results in two samples (Red: ground truth, Blue: prediction). (a,b) are the same in IoU, but the predicted bounding box in (b) contains less confusing information.
Drones 06 00188 g013
Figure 14. Performance versus speed on the FAIR1M data set.
Figure 14. Performance versus speed on the FAIR1M data set.
Drones 06 00188 g014
Figure 15. Results of the networks in three aerial scenes. (a) aerial image of urban residential areas with complex backgrounds. (b) aerial image of the airport. (c) aerial image of the water area.
Figure 15. Results of the networks in three aerial scenes. (a) aerial image of urban residential areas with complex backgrounds. (b) aerial image of the airport. (c) aerial image of the water area.
Drones 06 00188 g015
Figure 16. Comparisons with different backbone networks.
Figure 16. Comparisons with different backbone networks.
Drones 06 00188 g016
Figure 17. The confusion matrices of the variants on the airplane sub-dataset.
Figure 17. The confusion matrices of the variants on the airplane sub-dataset.
Drones 06 00188 g017
Figure 18. Comparison results with PRF-curve on the two sub-datasets.
Figure 18. Comparison results with PRF-curve on the two sub-datasets.
Drones 06 00188 g018
Table 1. Different configurations of DMViT. C is the dimension, N denotes the number of MViT, h e a d represents the number of heads, and r c is the adjustment of the rate in the c-th attention head.
Table 1. Different configurations of DMViT. C is the dimension, N denotes the number of MViT, h e a d represents the number of heads, and r c is the adjustment of the rate in the c-th attention head.
DMViT-SDMViT-MDMViT-L
Stage1Depatch Embedding
r c = 8 N 1 = 1 C 1 = 64 h e a d = 2 r c = 8 N 1 = 1 C 1 = 64 h e a d = 2 r c = 8 N 1 = 1 C 1 = 64 h e a d = 2
Stage2Depatch Embedding
r c = { 2 c < h e a d 2 4 c h e a d 2 C 2 = 128 N 2 = 2 h e a d = 4 r c = { 2 c < h e a d 2 4 c h e a d 2 C 2 = 128 N 2 = 3 h e a d = 4 r c = { 2 c < h e a d 2 4 c h e a d 2 C 2 = 128 N 2 = 3 h e a d = 4
Stage3Depatch Embedding
r c = { 1 c < h e a d 2 2 c h e a d 2 C 2 = 128 N 3 = 4 h e a d = 8 r c = { 1 c < h e a d 2 2 c h e a d 2 C 2 = 128 N 3 = 6 h e a d = 8 r c = { 1 c < h e a d 2 2 c h e a d 2 C 2 = 128 N 3 = 8 h e a d = 8
Stage4Depatch Embedding
r c = 1 N 1 = 1 C 1 = 512 h e a d = 16 r c = 1 N 1 = 1 C 1 = 512 h e a d = 16 r c = 1 N 1 = 1 C 1 = 512 h e a d = 16
Table 2. The results on the FAIR1M dataset.
Table 2. The results on the FAIR1M dataset.
General CategoriesCategoriesCorrNet [20]ClusDet [16]GLSAN [19]M-CenterNet [17]TPH-YOLOv5 [28]O2DETR [48]DFCformer
AP/APF (%)AP/APF (%)AP/APF (%)AP/APF (%)AP/APF (%)AP/APF (%)AP/APF (%)
RoadBridge15.71/7.6514.65/8.2113.18/7.6615.16/8.3314.83/7.1217.61/7.9518.67/8.02
Roundabout21.15/6.8920.17/6.5322.33/7.9723.15/8.0920.15/8.4919.51/8.0524.51/8.37
Intersection63.86/11.267.45/15.6469.88/10.1570.13/16.6866.89/14.8665.31/14.6371.64/21.02
CourtBaseball Field69.5/11.566.8/13.4665.27/15.1165.33/14.4970.15/19.6869.94/19.3470.05/22.67
Tennis Court85.6/19.682.97/21.1183.39/23.1681.16/22.9888.64/27.6188.51/27.5389.15/28.15
Football Field57.64/11.755.86/10.6857.96/12.8758.84/13.3961.12/17.3862.28/17.4464.15/18.12
Basketball Field55.62/11.259.87/13.6360.12/18.6561.11/20.0963.48/19.8464.51/19.9365.77/20.15
VehicleBus25.16/10.823.12/9.0619.14/5.2320.07/9.5823.16/8.7322.02/8.6238.61/15.28
Van58.62/15.8159.64/18.6461.16/19.6258.87/17.7462.43/17.6261.15/16.5467.84/19.57
Trailer23.34/10.6822.99/11.6323.51/11.7721.08/9.8423.53/6.8823.49/6.5326.34/7.88
Tractor9.98/2.6810.08/3.5311.29/4.6412.21/4.0615.83/4.0316.77/4.587.15/1.17
Excavator21.08/10.6523.96/11.9422.03/10.9821.88/9.6722.15/3.1922.03/3.0227.98/8.12
Small Car71.35/18.1573.81/19.0871.19/20.0370.05/19.6376.62/20.0576.13/19.9578.65/25.33
Cargo Truck50.32/12.4251.21/15.8952.21/17.8350.56/18.3452.27/18.6250.06/18.1255.87/23.17
Dump Truck41.63/10.3545.19/12.3747.55/13.3843.15/11.8949.31/16.1547.35/15.4849.86/16.53
Truck Tractor33.68/12.9235.86/11.1434.19/10.2530.06/9.5937.14/12.2836.41/11.7339.97/13.08
Other-vehicle13.86/5.9115.12/7.2617.62/8.2115.31/8.8715.09/6.3415.98/6.8715.12/7.15
ShipWarship23.41/10.3820.17/9.6419.56/9.2520.06/8.9323.98/7.5722.08/7.2127.88/9.54
Tugboat24.89/10.5622.35/9.1220.11/8.6419.59/7.3224.64/7.6326.18/8.0530.54/9.35
Motorboat28.62/12.6225.64/11.0524.66/9.2523.65/8.7430.04/9.8331.25/9.9539.88/10.05
Passenger Ship13.62/6.179.52/2.5110.67/3.849.93/2.0815.09/6.3516.17/6.6819.97/7.41
Fishing Boat5.21/1.843.12/0.875.11/1.126.09/1.463.21/1.013.05/0.871.08/0.15
Engineering Ship23.31/10.9625.16/11.2622.86/9.5513.65/2.1812.05/2.9710.25/2.1828.97/12.15
Liquid Cargo Ship17.86/12.6115.64/8.7613.06/7.745.34/0.6312.75/3.1611.34/2.7515.35/3.45
Dry Cargo Ship29.86/12.3730.16/13.6228.62/12.1922.61/7.1525.09/5.9322.97/5.3624.15/4.51
other-ship6.11/2.865.11/1.293.26/0.254.88/1.014.35/0.614.18/0.862.86/0.48
AirplaneC9192.56/1.865.87/1.756.51/1.613.36/0.412.18/0.372.01/0.530.95/0.08
A22045.62/16.5544.12/18.3147.12/19.5849.62/15.3251.86/19.9650.54/19.1755.37/20.09
A32143.98/12.3545.86/14.1246.16/15.0241.15/13.6842.15/15.5243.66/15.9557.86/21.15
A33022.14/6.8527.85/7.8626.68/8.0521.67/8.8426.04/6.5325.37/6.2831.17/7.68
A35022.69/8.3122.83/9.3721.16/7.7823.36/7.4225.12/5.0126.12/5.4335.57/8.69
ARJ213.69/0.985.89/0.466.09/1.057.05/1.889.54/2.269.36/2.092.84/0.61
Boeing73742.67/15.5845.98/17.6744.89/18.5246.94/15.6148.68/16.8349.15/16.9748.97/17.02
Boeing74758.98/23.1961.31/22.1863.33/23.1467.95/24.5769.91/24.1670.18/24.8375.56/27.86
Boeing77727.98/7.1825.46/9.5628.96/6.8829.05/6.3427.58/7.5328.69/7.7431.25/11.37
Boeing78755.64/16.6753.88/19.9557.97/20.1859.91/18.8661.86/19.9463.87/20.2669.87/22.16
other-airplane65.32/20.1668.52/20.4769.88/21.8367.35/23.8167.82/24.3568.94/25.2373.69/27.68
m A P / m A P F 34.66/10.8234.95/11.3435.1/11.2434.09/10.8136.41/11.2536.34/10.9440.14/13.12
Table 3. The results of the ablation study on different component combinations.
Table 3. The results of the ablation study on different component combinations.
Combination m A P m A P F
ResNet101 + FRGC + CAIM34.8310.06
PVT + FRGC + CAIM36.8512.57
DMViT + FRGC + CAIM (DFCformer)40.0813.17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, G.; He, X.; Wang, Q.; Shao, F.; Wang, H.; Wang, J. A Novel Multi-Scale Transformer for Object Detection in Aerial Scenes. Drones 2022, 6, 188. https://doi.org/10.3390/drones6080188

AMA Style

Lu G, He X, Wang Q, Shao F, Wang H, Wang J. A Novel Multi-Scale Transformer for Object Detection in Aerial Scenes. Drones. 2022; 6(8):188. https://doi.org/10.3390/drones6080188

Chicago/Turabian Style

Lu, Guanlin, Xiaohui He, Qiang Wang, Faming Shao, Hongwei Wang, and Jinkang Wang. 2022. "A Novel Multi-Scale Transformer for Object Detection in Aerial Scenes" Drones 6, no. 8: 188. https://doi.org/10.3390/drones6080188

Article Metrics

Back to TopTop