Next Article in Journal
Revocable Identity-Based Matchmaking Encryption with Equality Test for Smart Healthcare
Previous Article in Journal
Fault Diagnosis Method of Plunger Pump Based on Meta-Learning and Improved Multi-Channel Convolutional Neural Network Under Small Sample Condition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UAV-DETR: An Enhanced RT-DETR Architecture for Efficient Small Object Detection in UAV Imagery

College of Computer and Information Science, Chongqing Normal University, Chongqing 401331, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4582; https://doi.org/10.3390/s25154582
Submission received: 15 June 2025 / Revised: 21 July 2025 / Accepted: 21 July 2025 / Published: 24 July 2025
(This article belongs to the Section Remote Sensors)

Abstract

To mitigate the technical challenges associated with small-object detection, feature degradation, and spatial-contextual misalignment in UAV-acquired imagery, this paper proposes UAV-DETR, an enhanced Transformer-based object detection model designed for aerial scenarios. Specifically, UAV imagery often suffers from feature degradation due to low resolution and complex backgrounds and from semantic-spatial misalignment caused by dynamic shooting conditions. This work addresses these challenges by enhancing feature perception, semantic representation, and spatial alignment. Architecturally extending the RT-DETR framework, UAV-DETR incorporates three novel modules: the Channel-Aware Sensing Module (CAS), the Scale-Optimized Enhancement Pyramid Module (SOEP), and the newly designed Context-Spatial Alignment Module (CSAM), which integrates the functionalities of contextual and spatial calibration. These components collaboratively strengthen multi-scale feature extraction, semantic representation, and spatial-contextual alignment. The CAS module refines the backbone to improve multi-scale feature perception, while SOEP enhances semantic richness in shallow layers through lightweight channel-weighted fusion. CSAM further optimizes the hybrid encoder by simultaneously correcting contextual inconsistencies and spatial misalignments during feature fusion, enabling more precise cross-scale integration. Comprehensive comparisons with mainstream detectors, including Faster R-CNN and YOLOv5, demonstrate that UAV-DETR achieves superior small-object detection performance in complex aerial scenarios. The performance is thoroughly evaluated in terms of mAP@0.5, parameter count, and computational complexity (GFLOPs). Experiments on the VisDrone2019 dataset benchmark demonstrate that UAV-DETR achieves an mAP@0.5 of 51.6%, surpassing RT-DETR by 3.5% while reducing the number of model parameters from 19.8 million to 16.8 million.

1. Introduction

The accelerated evolution of drone technologies and sustained improvements in data transmission capabilities have revealed significant utility for UAVs in multiple domains, including surveillance, emergency rescue, and supply chain management. The capacity for drones to acquire high-resolution remote sensing images with ease is facilitated by their advanced data acquisition capabilities, providing significant support for target detection tasks. Consequently, the precise identification of targets utilizing UAV-captured imagery has garnered substantial scholarly attention.
Since AlexNet [1] introduced Convolutional Neural Networks (CNNs) and achieved remarkable success in the ImageNet competition in 2012, CNN-based approaches have rapidly become the mainstream in the field of object detection. Subsequently, two-stage detectors, such as the R-CNN series [2,3], were proposed, which significantly improved both detection accuracy and speed. However, their high computational cost makes them unsuitable for deployment on UAV platforms. To address the limitations of two-stage detectors in terms of real-time performance, one-stage detectors such as YOLO [4] and SSD [5] were developed. Later studies focused on improving detection accuracy by addressing the specific challenges inherent in aerial imagery. For example, Sahin et al. [6] enhanced the detection of small objects in aerial images by expanding YOLOv3’s output layers to five, which improved accuracy but at the cost of significantly increased parameters. Similarly, Tan et al. [7] introduced the Receptive Field Block (RFB) module into YOLOv4 to address issues such as severe occlusion and uneven sample distribution in aerial scenarios. However, the use of dilated convolutions within the RFB module may introduce noise information. Although these modifications have effectively mitigated some challenges associated with aerial images, the intrinsic limitations of CNNs in capturing long-range dependencies render them suboptimal for aerial image-based tasks.
With the remarkable success of the Transformer architecture [8] in the field of natural language processing, researchers have begun exploring its innovative concepts for object detection tasks. In 2020, the Facebook team proposed DETR [9], an end-to-end object detection framework based on Transformers, which elegantly eliminates the need for the aforementioned hand-crafted processing steps, thereby streamlining the detection pipeline. Although DETR offers a more efficient and direct detection paradigm, its substantial parameter count poses a significant challenge, necessitating strategies for parameter optimization while maintaining model performance. As a result, numerous DETR variants have been proposed to address the issues of large parameter size and high computational cost. In 2021, Zhu et al. introduced Deformable DETR [10], which enhances the attention mechanism by focusing solely on key sampling points around reference locations to improve performance. In 2022, Roh et al. proposed Sparse DETR [11], which reduces the attention complexity within the encoder through a token sparsification strategy, thereby improving detection performance under the same computational budget. Also in 2022, Li et al. proposed DN-DETR [12], which leverages denoising training to address the instability of bipartite matching in DETR, significantly accelerating model convergence. Furthermore, Li et al. introduced DINO [13] in 2022, which achieves substantial model size reduction by improving both the encoder and decoder structures. To address the high computational overhead of DETR, Baidu proposed RT-DETR [14], a real-time detection framework based on DETR that removes threshold filtering and non-maximum suppression (NMS), enabling RT-DETR to achieve higher training accuracy with fewer iterations.
Although RT-DETR has achieved promising results, its neck network still relies on the traditional pyramid structure, which fails to fully exploit the correlations between feature maps at different scales, leading to suboptimal feature fusion. Moreover, UAV aerial images often suffer from uneven sample distribution and severe occlusions. The fixed receptive field of convolutional kernels in the neck network is insufficient to capture both global context and fine-grained details simultaneously. Therefore, based on RT-DETR, an improved UAV-DETR detector is proposed to achieve a balance between detection accuracy and computational efficiency. The main contributions of this work are as follows:
  • A lightweight and efficient backbone is developed by introducing the Channel-Aware Sensing (CAS) module, which replaces conventional convolutional blocks with a spatial-channel attention mechanism. This architecture not only enhances small-target detection capability but also substantially decreases parameter volume and computational demands. Experiments show that, compared with the original ResNet-18 backbone, CAS achieves better detection accuracy with lower GFLOPs and fewer parameters.
  • The Scale-Optimized Enhancement Pyramid (SOEP) is a proposed model that incorporates a small-object detection layer into the original multi-scale feature fusion network. It introduces the concept of Cross Stage Partial (CSP) and OmniKernel network and designs the CSP-OmniKernel module for feature fusion to enhance efficiency in feature representation learning. The CSP-OmniKernel module is engineered to facilitate feature fusion for enhanced detection performance of small targets, ranging from global to local perspectives.
  • A unified Context-Spatial Alignment Module (CSAM) is introduced, integrating Contextual Fusion and Calibration (CFC) and Spatial Feature Calibration (SFC) to jointly address semantic inconsistency and spatial misalignment in cross-scale feature fusion.
  • Extensive experiments on the VisDrone2019 dataset demonstrate that the proposed UAV-DETR achieves an mAP@0.5 of 51.6%, outperforming the baseline RT-DETR by 3.5%, while maintaining low computational complexity suitable for real-time UAV-based applications.

2. Related Work

2.1. CNN-Based Object Detection

Convolutional Neural Networks (CNNs) have laid the foundation for modern object detection. Girshick et al. [2] proposed R-CNN, which introduced region proposals and convolutional feature extraction, later improved by Fast R-CNN and Faster R-CNN. Redmon et al. [4] pioneered the YOLO (You Only Look Once), achieving real-time inference by integrating object localization and classification within an end-to-end unified network architecture. Liu et al. proposed SSD [5], which leveraged multi-scale features for efficient detection. These CNN-based models have been widely applied in drone detection tasks due to their balance between speed and accuracy.
However, UAV imagery presents unique challenges such as small, densely packed targets, large-scale variations, and complex backgrounds. To mitigate these, Lin et al. proposed Feature Pyramid Networks (FPN) [15] to enhance multi-scale feature extraction. Subsequent improvements like PANet [16] and BiFPN [17] further strengthened feature aggregation and semantic consistency across levels. Nonetheless, CNNs still struggle to capture global context due to their limited receptive fields.

2.2. Transformer-Based Object Detection

Transformer-based detectors offer significant potential for aerial analysis, where their global attention mechanisms effectively correlate spatially dispersed objects across wide-area UAV views, overcoming the limited receptive fields of CNNs. Carion et al. [9] introduced Detection Transformer (DETR). This approach reformulates object detection as a set prediction task, effectively eliminating the reliance on hand-crafted components such as anchor boxes and non-maximum suppression (NMS). Despite its architectural simplicity and end-to-end paradigm, DETR exhibits significant limitations: (1) slow convergence requiring 500 training epochs to achieve competitive accuracy; (2) suboptimal performance on small objects due to an information bottleneck in global attention mechanisms.
To address these limitations, Deformable DETR by Zhu et al. [10] employs sparse attention across a few key sampling points to speed up training and improve accuracy. Conditional DETR [18] and DAB-DETR [19] further refined the query representation and attention mechanisms. Li et al. proposed DN-DETR [12], incorporating denoising techniques to improve convergence stability.
In the real-time detection domain, RT-DETR was proposed by Baidu’s AI team, offering a hybrid encoder with intra-scale and cross-scale interaction modules, along with an IoU-aware query initialization mechanism. An effective trade-off between detection speed and accuracy is obtained, making it well-suited for edge deployment. However, when directly applied to UAV scenarios, RT-DETR still struggles with issues such as feature degradation from repeated downsampling, small object miss detection, and spatial-context misalignment.

2.3. UAV-Specific Detection Methods

Several works have specifically optimized detection for UAV scenarios. Yang et al. proposed UAV-YOLO [20], which improves small object sensitivity through multi-resolution feature fusion. Guo et al. introduced TinyDet [21], targeting low-computation platforms. Datasets such as VisDrone [22], UAVDT [23], and DOTA [24] have become benchmarks for evaluating UAV detection models. Nevertheless, many existing models either sacrifice accuracy for speed or fail to generalize across complex UAV scenes with scale, motion, and background variability.
While significant progress has been made from CNN-based to Transformer-based detection frameworks, challenges such as small object detection, feature misalignment, and efficient multi-scale fusion persist in UAV applications. Several recent DETR variants, such as Efficient-DETR and SMCA-DETR, have attempted to improve training convergence and localization accuracy. Moreover, recent approaches such as CFPT (Cross-Layer Feature Pyramid Transformer) [25] and LGI-DETR (Local–Global Interaction DETR) [26] have further explored cross-layer attention mechanisms and local-global feature fusion to mitigate spatial-contextual misalignment in UAV imagery. However, these methods often incur substantial architectural complexity, with parameter counts reaching 51.3 million and 21.1 million, respectively, and rely heavily on decoder-side enhancement modules. This results in significant computational overhead, thereby limiting their real-time performance and applicability in resource-constrained environments. Recent studies have also made notable progress in small object detection on the VisDrone2019-DET dataset. Zhao et al. [27] proposed an improved YOLOv7 model that achieves a high mAP@0.5 of 52.3%, but its relatively large model size (approximately 35.8 million parameters) restricts its suitability for lightweight deployment. In contrast, Wang et al. [28] designed the lightweight YOLO-PEL detector, which significantly reduces model size (only 2.23 million parameters), but its detection accuracy (mAP@0.5 of 32.5%) remains insufficient for practical applications. Among Transformer-based small object detection methods, Wei et al. [29] refined RT-DETR to develop DV-DETR, attaining an mAP@0.5 of 50.2% with 19.5 million parameters, demonstrating improved performance; however, limitations remain in feature alignment and contextual information fusion. Similarly, Wei and Wang [30] introduced RFAG-YOLO, which integrates receptive field attention and achieves 38.9% mAP@0.5 with only 5.9 million parameters, indicating that lightweight models still face challenges in accuracy. Wu et al. [31] combined HRFPN and EfficientVMamba techniques to achieve an mAP@0.5 of 38.9% with 33.5 million parameters under complex aerial scenarios, demonstrating strong detection capabilities, albeit with relatively heavy computational costs that restrict real-time applicability. Kong et al. [32] proposed Drone-DETR, an enhanced RT-DETR variant optimized for small object detection in remote sensing images, which achieves the highest mAP@0.5 of 53.9% and maintains an inference speed of 30 FPS. However, its computational complexity is as high as 128.3 GFLOPs, and it contains 28.7 million parameters, imposing stringent hardware demands that limit its practical deployment on UAV platforms. Baidya and Jeong [33] explored ConvMixer-based YOLOv5 heads to improve localization precision.
To address these limitations, the present study proposes UAV-DETR, which introduces three novel modules—CAS, SOEP, and CSAM—to enhance adaptive feature alignment and contextual perception. Under the premise of maintaining real-time inference speed (30 FPS), UAV-DETR achieves a competitive mAP@0.5 of 51.6%, with a significantly reduced parameter count of 16.8 million and computational complexity of 71.4 GFLOPs. The comprehensive performance demonstrates that UAV-DETR achieves an excellent balance among detection accuracy, model compactness, and computational efficiency, thus showing strong potential for practical application in resource-constrained UAV small object detection tasks.

2.4. RT-DETR Model

RT-DETR, developed by the Baidu research team, represents a major advancement in the real-time object detection field. In addressing the high computational cost issues of the previous DETR, which prevented its application in real-time detection, RT-DETR is designed with a computationally efficient hybrid encoder composed of the Attention-enhanced Intra-scale Feature Interaction (AIFI) module and the cross-scale feature fusion (CCFF) module. The design in question decouples within-scale feature interactions and cross-scale fusion, thereby enabling efficient processing of multi-scale features. Furthermore, RT-DETR employs an IoU-aware query selection mechanism that optimizes the initialization of decoder queries. These innovations enable RT-DETR to capitalize on its non-post-processing requirement, rendering it suitable for real-time object detection. The RT-DETR model is illustrated in Figure 1.
However, directly applying RT-DETR to object detection in drone imagery presents some challenges. Due to the multiple downsampling operations in the model’s backbone network, feature loss occurs during the downsampling process, leading to missed or incorrect detections of small objects. Furthermore, RT-DETR does not adequately address spatial misalignment and context mismatches during feature fusion, which negatively impacts the model’s overall performance.

3. Methodology

3.1. UAV-DETR Model

Given the shortcomings of RT-DETR in drone-based object detection, particularly its deficiencies in small object detection, feature information loss, and context misalignment. This paper proposes UAV-DETR, an enhanced multi-scale object detection framework specifically designed for drone-based imagery. Building upon the strengths of end-to-end detection, UAV-DETR proposes three key modules: CAS, SOEP, and CSAM, which include CFC and SFC. Collectively, these modules significantly elevate the model’s detection performance in complex aerial scenarios.
The CAS module improves the capability of the backbone network to detect multi-scale objects by refining the feature hierarchy and enhancing inter-channel interactions. The SOEP module strengthens the semantic representation of mid- and low-level features in the neck network, elevating small-target discrimination accuracy. The CFC and SFC modules calibrate and align features within the hybrid encoder, improving cross-scale feature integration. These improvements make UAV-DETR more accurate and real-time in its drone-based remote sensing object detection tasks. The UAV-DETR model is illustrated in Figure 2.

3.2. CasNet

In datasets like VisDrone-2019, which consist of low-altitude UAV imagery, targets are often extremely small—ranging from just a few to a few dozen pixels—and the background is filled with dense urban textures, glare, and motion blur. Traditional CNN backbones, relying on fixed receptive fields, struggle to balance global context modeling with fine-grained local detail extraction essential for identifying such minuscule objects. To address this, we replace the BasicBlock structures in the backbone with Convolutional Additive Self-attention (CAS) blocks [34]. Each CAS block first uses a stack of three 3 × 3 depthwise separable convolutions to integrate local neighborhood features while implicitly encoding positional information critical for precisely locating sub-20 px targets. This is followed by the Convolutional Additive Token Mixer (CATM), which generates attention weights along both spatial and channel dimensions using lightweight convolution-Sigmoid operations. The spatial component suppresses background clutter and highlights regions containing small objects, while the channel component amplifies discriminative features against complex backgrounds. These weights are fused additively to recalibrate subsequent features dynamically. The entire process avoids softmax normalization and large matrix multiplications, keeping inference complexity on par with standard convolutions while enabling long-range spatial and cross-channel interactions in a single forward pass. This long-range modeling leverages contextual cues to reinforce ambiguous small targets, and the additive fusion preserves subtle features often lost in Softmax-based mechanisms. As a result, the network produces more accurate responses to small and occluded targets while conforming to the computational constraints of mobile and edge computing platforms.
To meet the computational and deployment requirements, we can adopt simpler operations instead of relying on matrix multiplication—which introduces a significant inference overhead—or Softmax, which may pose challenges for deployment. Following this rationale, for a feature tensor of size H × W × C, information interaction can be decoupled into the spatial domain (N = H × W) and the channel domain (C). Notably, prior works on scale decomposition have also explored using multiple heads along the channel dimension to enable parallel computation, as well as dividing spatial patches to accelerate processing. However, these methods essentially operate on the fundamental dimensions of the feature tensor.
The concept of spatial interaction is demonstrated in Figure 3, where the input xRH×W×C. Initially, local token information is aggregated using a 3 × 3 convolutional layer. Subsequent to this, the reduction of the channel dimension to 1 is achieved by employing a 1 × 1 convolutional layer; subsequently, a Sigmoid activation function is applied to generate the feature-adaptive spatial attention map xsRH×W×1. The following formal description has been provided:
xs = Sigmoid(D1×1(D3×3,ReLU,BN (x))) ⊙ x
For channel-wise interaction, we draw inspiration from SENet [35]; however, rather than employing channel reduction, we utilize a 1 × 1 convolution to directly capture and integrate inter-channel information:
xc = Sigmoid(D1×1(P1×1(x))) ⊙ x
where P denotes the adaptive pooling and D is the group convolution layer, the group number is set to the channel number by default. Stacking these two operations yields the feature map after spatial and channel domain interactions, denoted as Φ(x).
This work introduces a novel similarity metric, defined as the summation of context scores between corresponding elements of Q, KRN×C:
Sim (Q, K) = Φ(Q) + Φ(K)
where Query, Key, and Value are generated via independent linear transformations, for example, Q = Wqx, K = Wkx, and V = Wvx, it can be demonstrated that the context mapping function, denoted by Φ(·), contains the essential information interactions. The advantage of this generalization is that it is not restricted to the domain of manual context design; rather, it extends to the realm of convolutional operations, thereby facilitating implementation. It is anticipated that the network will learn and identify more valuable tokens through the execution of multiple interactions on both Q and K. The integration of these components is achieved through additive operations, a linear approach that circumvents the high complexity of matrix multiplication while retaining the efficacy in preserving useful information. Therefore, the following formulation can be used to express the output of CATM:
O = Γ (Φ(Q) + Φ(K)) ⊙ V
where Γ(·) RN×C denotes the linear transformation for integrating the contextual information.

3.3. SOEP Module

Conventional detection layers (P3–P5) exhibit limited efficacy in identifying diminutive UAV targets. A strategic remediation involves incorporating a higher-resolution P2 detection tier, but this increases computation and post-processing time. It is imperative that novel multi-scale feature fusion structures are developed with a view to detecting targets on UAVs. This paper proposes a computationally efficient small-target feature-enhanced fusion network (Figure 4). The network integrates the P2 feature layer (small target information) with the P3 layer after SPDConv processing. The CSP-OmniKernel module fuses features based on the CSP and OmniKernel [36] method.

3.3.1. SPDConv

SPDConv [37] is a CNN module specifically designed to improve low-resolution image and small target detection. It reduces the information loss and enhances the capture of detailed features by replacing the traditional step-size convolution and pooling operations, enhancing model discrimination capability in processing small targets and low-resolution images. SPDConv is a CNN module specifically designed to improve low-resolution image and small target detection. It reduces the information loss and enhances the capture of detailed features by replacing the traditional step-size convolution and pooling operations, thereby enhancing the model’s performance in processing small targets and low-resolution images.
Specifically, SPDConv comprises a space-to-depth (SPD) transformation layer followed by a stride-1 convolutional layer. The SPD component partitions intermediate feature tensors of dimensions S × S × C through spatial decomposition, yielding sub-feature representations fx,y:
f 0,0 = X 0 : S : s , 0 : S : s , f 1,0 = X 1 : S : s , 0 : S : s , f s 1,0 = X s 1 : S : s , 0 : S : s ; f 0,1 = X 0 : S : s , 1 : S : , f 1,1 , , f s 1,1 = X s 1 : S : s , 1 : S : s , f 0 , s 1 = X 0 : S : s , s c a l e 1 : S : s , f 1 , s 1 , , f s 1 , s 1 = X s 1 : S : s , s 1 : S : s
Here, the s refers to the downsampling factor. The generation of each sub-feature map is achieved by means of downsampling the original feature map, with the process being executed for each map using a specific scaling factor. To mitigate feature degradation and channel explosion from aggressive downsampling, a scaling factor of 2 optimizes the information–computation trade-off.
The architecture subsequently concatenates all sub-feature maps along the channel dimension, thereby forming a new feature map X. This process reduces the spatial dimensions while concomitantly increasing the number of channels. Finally, the transformed feature map X is further refined by the application of a non-strided convolutional layer. This layer employs learnable parameters to reduce the channel dimensionality, with the aim of mitigating information redundancy caused by the increased number of channels, while preserving as much discriminative feature information as possible.

3.3.2. CSP-OmniKernel Module

The proposed CSP-OmniKernel module is designed based on the CSP architecture. The CSP architecture partitions the input feature map into dual pathways: a lightweight convolutional subnetwork processes one partition, while the complementary partition undergoes direct residual transmission to subsequent layers. The dual-path features are channel-wise concatenated, forming the input to the subsequent processing stage.
As illustrated in Figure 4, a convolutional layer partitions the input feature tensor into four distinct channels prior to OmniKernel processing. The transformed feature representation subsequently undergoes channel-axis concatenation with three complementary feature channels. The next step is a convolution to adjust the channel dimensionality. Parameters and computational costs are reduced, while the efficiency of feature extraction is enhanced. Consequently, both model training and inference are accelerated.
The OmniKernel network is comprised of three distinct branches: a global branch, a large-kernel branch, and a local branch. This architecture enables systematic acquisition of multi-scale feature representations spanning global to local contexts, thereby significantly improving detection efficiency for diminutive objects. The structural configuration of the OmniKernel network is schematically presented in Figure 5.
In the large-kernel branch, the input XLarge undergoes three convolution operations across three separate channels: a 31 × 31 depthwise separable convolution to capture a large receptive field and two depthwise separable convolutions with kernel sizes of 31 × 1 and 1 × 31, respectively, to extract elongated contextual information.
In the inference stage of the global branch, input images of substantially higher spatial resolution than those used in training are processed. Consequently, a 31 × 31 kernel is inadequate in covering the global receptive field. In order to address this issue, a dual-domain processing strategy is adopted in order to enhance the global modelling capability of the global branch. The global branch incorporates two modules, the Dual-Domain Channel Attention Module (DCAM) and the Frequency-Based Spatial Attention Module (FSAM). These modules contribute to improved global feature modeling through dual-domain enhancement.
Given the input feature XGlobal, the DCAM Module first applies Frequency Channel Attention (FCA) to XGlobal:
X F C A = I F ( F ( X G l o b a l ) W 1 × 1 ( G A P ( X G l o b a l ) )
where F and I F respectively denote the Fast Fourier Transform (FFT) and its inverse transform. The output of the FCA module is denoted by XFCA, the 1 × 1 convolutional layer by W1×1, and global average pooling by GAP. Element-wise multiplication is indicated by ⊗.
After undergoing Fourier transform processing, the global features, which have been globally modulated in the frequency domain, are subsequently input into the spatial channel attention (SCA) module:
X D C A M = X F C A W l × l ( G A P ( X F C A ) )
where XDCAM is the output of the DCAM module, which only enhances the dual-domain features at the coarse channel granularity. Subsequently, XDCAM is fed into the frequency-based spatial attention module to further refine the spectrum along the spatial dimension:
X F S A M = I F ( F ( W l × l ( X D C A M ) ) W l × l ( X D C A M )
where XFSAM represents the output produced by the FSAM module.
The local branch applies only a 1 × 1 depthwise separable convolution to the input XLocal, which is used for modulating local features.

3.4. CSAM Module

The hybrid encoder in RT-DETR falls short in cross-scale feature fusion: it does not adequately address spatial misalignment and fails to reconcile representational discrepancies across feature levels, which diminishes fusion effectiveness. Moreover, it overlooks the mismatch between pixel-level cues and their surrounding context, ignoring their distinct contextual dependencies and further constraining overall performance.
To address these shortcomings, we augment the CCFF module in RT-DETR’s hybrid encoder with two dedicated calibration mechanisms: Context Feature Calibration (CFC) and Spatial Feature Calibration (SFC) [38]. CFC harmonizes the semantic representations across different feature levels, while SFC rectifies the spatial misalignment introduced during multi-scale fusion, jointly improving feature alignment and overall information integration.

3.4.1. CFC Module

The contextual information feature calibration method can adaptively construct and optimize the semantic contextual representation for each pixel. Meanwhile, by applying sharpening operations to large-scale objects and conditionally preserving spatial details, this method effectively facilitates selective learning of local context, fundamentally alleviating the significant bias toward large objects in the captured contextual information.
For a given feature tensor XRC×H×W, a pyramid pooling operation is first applied to extract multi-scale contextual representations ZRC×M. Subsequently, the spatial affinity between pixels and contextual features is computed via a softmax function, generating a spatial attention map Θ ∈ RN×M. Here, C, H, and W respectively denote the channel depth, height, and width of the input feature, while N = H × W and M represent the total number of pixels and contextual elements.
Contextual feature calibration (CFC) implements the aggregation of contextual information for individual pixels under the guidance of a spatial attention map. Subsequently, the response characteristics of each semantic context vector undergo refinement to generate a fine-grained context representation, effecting a recalibration of the contextual information. The complete CFC process is formulated in Equation (9).
y i = α i j = 1 M f ( x i , z i ) z j + x i
In the formula, { x i , y i ,   α i , z j } ∈ RC×1 represents input, output, recalibration factor, and context information, respectively. The value range of i is 1 , , H × W , f(·) is a function to calculate the similarity between features. The architecture of the CFC module is depicted in Figure 6.
Where CPP stands for Cascading Pyramid Pooling Module and CRB stands for Context Recalibration Module.

3.4.2. SFC Module

To address the spatial detail degradation caused by repeated downsampling operations, existing methods [39] predominantly employ cross-level feature interaction mechanisms that enhance the semantic representation capacity of deep features by integrating fine-grained information from shallow high-resolution features. The feature fusion architecture adopted in the RT-DETR model also follows this technical paradigm. However, in-depth analysis reveals significant limitations in this coarse-grained cross-scale feature aggregation approach: First, the discrepancy in receptive fields among different hierarchical feature maps leads to mismatches in spatial coordinate systems, resulting in spatial heterogeneity of feature representations. Second, hierarchical features exhibit significant disparities in semantic abstraction levels, where direct feature superposition induces semantic confusion. Notably, conventional methods employing uniform channel-wise alignment [40] strategies neglect the differential semantic information distribution across channel dimensions. This homogenization process undermines the distinctive characteristics of feature channels [41,42], thereby constraining further performance improvements of models.
To mitigate these constraints, the SFC module partitions channel dimensions into multiple sub-features for individualized calibration while adaptively integrating gating mechanisms to fuse cross-scale features. As shown in Figure 7.
Given a low-resolution feature map F l R C l × H l × W l and a high-resolution feature map F h R C h × H h × W h , the SFC module first aligns their channel dimensions via convolutional layers. The low-resolution feature F l is then upsampled through bilinear interpolation and concatenated with F h . Subsequently, these are fed into convolutional blocks to predict two sets of offset feature maps, l R 2 × G × H × W , h R 2 × G × H × W for aligning the two feature levels, as well as two gating masks, βl and βh, which control the flow of feature information across the two scales. The calibrated features from both scales are ultimately fused via element-wise summation, with the complete operational procedure formalized in Equation (10).
O = βlT (V (Wl Fl), Δl) + βhT (Wh Fh, Δh)
Here, V(·) denotes the bilinear upsampling function, and T(·, Δ) represents a deformable sampling operator that adaptively adjusts spatial sampling locations based on learned offsets Δ. When Δ = 0, the operation becomes standard bilinear interpolation, resulting in an identity transformation under the gating mask β = 1 + tanh(0) = 1. When Δ ≠ 0, the offsets introduce spatial modulation, enabling the module to realign features across resolutions and correct spatial misalignments in a content-adaptive manner. Wl and Wh represent convolutional layers followed by Batch Normalization (BN) and ReLU activation. In addition, the SFC module adopts the 1 + tanh activation function for the gating masks. Thus, for the initial state T(F,0), the transformation corresponds to an identity mapping due to β = 1+ tanh(0) = 1. Under this condition, the SFC operation can be simplified and is expressed as shown in Equation (11).
O = V   ( W l   F l ) + W l F h

4. Results and Analysis

4.1. Experimental Dataset and Experimental Setup

4.1.1. Experimental Dataset

This study utilizes the publicly available VisDrone2019 dataset [43] for model training and evaluation. Containing drone-captured aerial footage, the dataset spans heterogeneous traffic environments (urban/rural) under diverse meteorological and photometric conditions. It contains 6471 training images, 548 validation images, and 1610 test images, with a significant prevalence of small-scale objects that facilitates rigorous evaluation of drone-based detection models. The proposed model demonstrates effectiveness in detecting ten object categories: van, tricycle, car, motor, awning-tricycle, bicycle, pedestrian, truck, bus, and people.

4.1.2. Experimental Setup

All architectural modifications and training experiments were performed under identical computational conditions, with detailed specifications provided in Table 1
The experimental network training parameters are enumerated in Table 2.

4.2. Evaluation Metrics

To comprehensively evaluate the efficacy of the proposed enhancements, three critical metrics were employed: parameter count, Floating Point Operations (FLOPs), and Mean Average Precision (mAP). The mathematical formulation is defined as
P = TP/(TP + FP)
R = TP/(TP + FN)
A P = 0 1 P ( R ) d R
m A P = i = 1 C A P i / C
where TP refers to the positive samples that have been predicted to belong to the positive category, FP denotes the negative samples that have been predicted to belong to the positive category, FN refers to the positive samples that have been predicted to belong to the negative category, P represents precision, and R denotes recall. AP (Average Precision) represents the average precision for a single class. The mean of the APs of the entire category, mAP, is influenced by the IoU threshold and is subsequently utilized at mAP@0.5 and mAP@0.5:0.95.

4.3. Results and Comparsion

4.3.1. Ablation Study

To quantify component-specific contributions, we conducted systematic ablation studies by sequentially integrating CasNet, SOEP, and CSAM modules into the baseline RT-DETR framework. Training and validation utilized the VisDrone2019 dataset, with experimental results demonstrating progressive performance gains at each integration stage. Quantitative ablation results are presented in Table 3.
The introduction of CasNet resulted in a substantial enhancement of mAP@0.5 (from 48.1% to 49.2%) and mAP@0.5:0.95 (from 29.5% to 30.3%), concurrently reducing both parameters and FLOPs. This outcome serves to substantiate the hypothesis that CasNet is characterized by a lightweight and efficient design. The incorporation of the SOEP module led to a substantial enhancement in recognition precision (mAP@0.5 to 50.7%, mAP@0.5:0.95 to 31.5%), primarily resulting in a significant boost in recall, though this was accompanied by an increase in computational load. The final integration of CSAM yielded the most optimal overall results (mAP@0.5 of 51.6%, mAP@0.5:0.95 of 32.1%). The total computational cost of UAV-DETR increases to 71.4 GFLOPs, which remains within an acceptable range and is suitable for real-time UAV applications.
These findings confirm that each module—CasNet, SOEP, and CSAM—contributes distinct and complementary improvements, jointly enhancing detection accuracy and robustness in UAV-based tiny object detection tasks.

4.3.2. Comparative Experiment

This paper selects several mainstream backbone networks in recent years—FasterNet [44], EfficientViT [45], and the original RT-DETR backbone ResNet18—for comparative experiments. By evaluating their overall floating-point operations (FLOPs) and detection accuracy on the VisDrone2019 dataset, we aim to identify the most suitable backbone network for object detection in drone imagery.
During the experiments, all models employing different backbone networks were trained and evaluated under identical hyperparameter settings. The training results for each backbone network are summarized in Table 4.
After conducting comparative experiments with mainstream feature-extraction networks, CasNet was selected as the backbone because it delivers superior feature-extraction performance on UAV imagery while requiring fewer floating-point operations than ResNet-18.
To rigorously benchmark UAV-DETR against state-of-the-art methods in drone-based detection, comparative experiments were conducted with mainstream detectors (Faster R-CNN, YOLO series) on the VisDrone2019 dataset. All models underwent consistent hyperparameter configuration during training and evaluation. Performance was quantified using standard metrics, including mAP@0.5 and GFLOPs, with comparative results detailed in Table 5.
In terms of detection accuracy (mAP@0.5), UAV-DETR achieves a score of 51.6%, outperforming most comparison models (including Faster R-CNN, the YOLO series, LGI-DETR, and the RT-DETR baseline) while approaching the top performers Enhanced YOLOv7 (52.3%) and Drone-DETR (53.9%). Specifically, it improves by 11.9 percentage points over the traditional two-stage detector Faster R-CNN (39.7%), indicating a significantly stronger capability in feature extraction and object localization for small targets. Compared with the YOLO series, UAV-DETR surpasses YOLOv5l (38.8%), YOLOv6m (37.2%), YOLOv7 (48.0%), and YOLOv8m (43.2%) by 12.8%, 14.4%, 3.6%, and 8.4%, respectively, though it trails Enhanced YOLOv7 (52.3%) by 1.3 percentage points. Among other recent detectors, UAV-DETR exceeds LGI-DETR (46.0%) by 5.6 points, CFPT (50.0%) by 1.6 points, and DV-DETR (50.2%) by 1.4 points, while being 2.3 points lower than Drone-DETR (53.9%).
Regarding model complexity, UAV-DETR has the smallest number of parameters among all models, at only 16.8 million, which is 2.7 million fewer than DV-DETR (19.5 M), the next lightest model. Although Enhanced YOLOv7 and Drone-DETR achieve slightly higher accuracy, they come at the cost of significantly increased model complexity, with Enhanced YOLOv7 requiring 19 M more parameters and Drone-DETR requiring 11.9 M more parameters compared with UAV-DETR. In terms of computational cost, UAV-DETR requires 71.4 GFLOPs, significantly lower than CFPT (297.6) and Drone-DETR (128.3) and only slightly higher than RT-DETR (57.0), while still achieving 3.5 percentage points higher detection accuracy, highlighting a superior trade-off between efficiency and performance.
In summary, the proposed UAV-DETR demonstrates superior detection accuracy across drone-based object detection tasks, even with a relatively low model size. These results indicate that UAV-DETR is highly suitable for real-time, accurate object detection in UAV imagery.

4.3.3. Generalization Experiments

To validate the generalization capability of the proposed UAV-DETR model, we conduct experiments on the DOTA dataset, which contains a wide variety of object categories and complex aerial scenes. Specifically, we compare UAV-DETR with several representative object detection models, including Faster R-CNN, YOLOv7, and RT-DETR. This evaluation helps assess the model’s performance in scenarios different from the VisDrone dataset, thereby demonstrating its robustness and adaptability across diverse UAV-based remote sensing environments.
Table 6 presents the experimental results of Faster R-CNN, YOLOv7, RT-DETR, and UAV-DETR on the DOTA dataset. As shown in the results, UAV-DETR achieves mAP@0.5 improvements of 7.0 and 2.1 percentage points over Faster R-CNN and RT-DETR, respectively, demonstrating the strong generalization capability of UAV-DETR.

4.3.4. Feature Fusion Visualization Comparison Experiment

For enhanced visual interpretability of the model’s decision mechanisms, Grad-CAM++ was utilized to generate class-discriminative activation heatmaps. In the heatmap, color intensity is used to represent the density of feature point activations, with red corresponding to the highest concentration. As illustrated in Figure 8, a noticeably denser concentration of feature points around the target objects—rather than the background—is observed in the UAV-DETR model when compared with the baseline. This demonstrates that the UAV-DETR model responds more effectively to small objects in complex scenes, resulting in improved detection performance with fewer missed or false detections, as well as enhanced suppression of background interference.

4.3.5. Comparison of Detection Results

Model detection performance can be qualitatively assessed through visualization. To conduct a targeted evaluation of small object detection capabilities, three challenging scenarios from the VisDrone2019 dataset were selected for comparative analysis between the UAV-DETR framework and its baseline RT-DETR counterpart.
As depicted in Figure 9, for the multi-scale object scenario depicted in Figure 9a, the UAV-DETR model demonstrates significantly better performance compared with the baseline model, particularly in densely populated regions with distant objects, where it exhibits notably higher detection accuracy. In the occlusion scenario shown in Figure 9b, the UAV-DETR model is able to accurately identify targets that are partially or completely obscured by trees, providing more complete and precise detection results. In the nighttime condition depicted in Figure 9c, the UAV-DETR model maintains stable performance and effectively identifies distant objects in low-light conditions.
From the detection results, it is evident that RT-DETR exhibits missed and false detections for small objects. In contrast, UAV-DETR is capable of detecting small targets that the original model fails to identify and also shows improved performance in detecting occluded objects. Moreover, it achieves this without compromising the detection of other targets, effectively balancing coarse and fine-grained information. The three proposed improvements significantly enhance the detection accuracy of RT-DETR in UAV aerial scenarios, thereby validating the effectiveness of the modifications.

5. Conclusions

This paper presents UAV-DETR, an enhanced Transformer-based object detection framework specifically designed for unmanned aerial vehicle (UAV) imagery. Built upon the foundation of RT-DETR, UAV-DETR integrates three well-established modules—Channel-Aware Sensing (CAS), Scale-Optimized Enhancement Pyramid (SOEP), and Context-Spatial Alignment Module (CSAM)—each improved to target specific challenges in UAV-based object detection. CAS is refined to enhance channel-wise feature sensitivity, thereby improving the detection of small and occluded objects. SOEP is optimized to mitigate feature degradation caused by repeated downsampling, enabling more robust multi-scale representation. CSAM is enhanced to address spatial and contextual misalignment across feature levels, ensuring better integration of spatial structure and semantic context. Through this targeted refinement and effective integration of the three modules, UAV-DETR significantly improves multi-scale perception and contextual understanding. Experimental evaluations on the VisDrone2019 dataset show that UAV-DETR obtains a mean average precision (mAP@0.5) of 51.6%, which represents a significant 3.5% improvement over the original RT-DETR model. Additionally, the model maintains low computational complexity and real-time inference speed, making it suitable for deployment in practical UAV-based applications. However, UAV-DETR also has some limitations. The detection performance is prone to degradation under certain conditions, such as fast object motion and adverse weather like haze. Future research will focus on further optimizing the model for diverse UAV scenarios and exploring adaptive real-time deployment strategies.

Author Contributions

Conceptualization, Y.Z. and Y.W.; Methodology, Y.W.; Software, Y.Z.; Validation, Y.Z. and Y.W.; Formal analysis, Y.Z.; Investigation, Y.Z.; Resources, Y.Z. and Y.W.; Data curation, Y.Z.; Writing—original draft preparation, Y.Z. and Y.W.; Writing—review and editing, Y.Z. and Y.W.; Visualization, Y.Z.; Supervision, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  2. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  3. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  4. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27 June–1 July 2016; pp. 779–788. [Google Scholar]
  5. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  6. Sahin, O.; Ozer, S. Yolodrone: Improved YOLO architecture for object detection in drone images. In Proceedings of the 2021 44th International Conference on Telecommunications and Signal Processing (TSP), Budapest, Hungary, 15–17 July 2021; pp. 361–365. [Google Scholar]
  7. Tan, L.; Lv, X.; Lian, X.; Wang, G. YOLOv4_Drone: UAV image target detection based on an improved YOLOv4 algorithm. Comput. Electr. Eng. 2021, 93, 107261. [Google Scholar] [CrossRef]
  8. Bujagouni, G.K.; Pradhan, S. Transformer-based deep learning hybrid architecture for phase unwrapping. Phys. Scr. 2024, 99, 7. [Google Scholar] [CrossRef]
  9. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 August 2020; pp. 213–229. [Google Scholar]
  10. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  11. Roh, B.; Shin, J.; Shin, W.; Kim, S. Sparse DETR: Efficient End-to-End Object Detection with Learnable Sparsity. In Proceedings of the 10th International Conference on Learning Representations (ICLR), Virtual Event, 25–29 April 2022. [Google Scholar]
  12. Li, F.; Zhang, H.; Liu, S.; Guo, J.; Ni, L.M.; Zhang, L. DN-DETR: Accelerate DETR training by introducing query denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–23 June 2022; pp. 13619–13627. [Google Scholar]
  13. Li, F.; Zhang, H.; Xu, H.; Liu, S.; Zhang, L.; Ni, L.M.; Shum, H.Y. Mask DINO: Towards a unified transformer-based framework for object detection and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–23 June 2023; pp. 3041–3050. [Google Scholar]
  14. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs beat YOLOs on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–22 June 2024; pp. 16965–16974. [Google Scholar]
  15. Lin, T.Y.; Dollár, P.; Girshick, R.; HE, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  16. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  17. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar]
  18. Meng, D.; Chen, X.; Fan, Z.; Zeng, G.; Li, H.; Yuan, Y.; Sun, L.; Wang, J. Conditional DETR for Fast Training Convergence. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 August 2021; pp. 3631–3640. [Google Scholar]
  19. Liu, S.; Li, F.; Zhang, H.; Yang, X.; Qi, X.; Su, H.; Zhu, J.; Zhang, L. DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR. In Proceedings of the International Conference on Learning Representations, Online, 25–29 April 2022. [Google Scholar]
  20. Liu, M.; Wang, X.; Zhou, A.; Fu, X.; Ma, Y.; Piao, C. UAV-YOLO: Small Object Detection on Unmanned Aerial Vehicle Perspective. Sensors 2022, 20, 2238. [Google Scholar] [CrossRef]
  21. Chen, S.; Cheng, T.; Fang, J.; Zhang, Q.; Li, Y.; Liu, W.; Wang, X. TinyDet: Accurate Small Object Detection in Lightweight Generic Detectors. arXiv 2023, arXiv:2304.03428. [Google Scholar]
  22. Zhu, P.; Wen, L.; Bian, X.; Ling, H.; Hu, Q. Vision Meets Drones: A Challenge. arXiv 2022, arXiv:1804.07437. [Google Scholar]
  23. Yu, H.; Li, G.; Zhang, W.; Huang, Q.; Du, D.; Tian, Q.; Sebe, N. The Unmanned Aerial Vehicle Benchmark: Object Detection, Tracking and Baseline. Int. J. Comput. Vis. 2020, 128, 1141–1159. [Google Scholar] [CrossRef]
  24. Li, K.; Wan, G.; Cheng, G.; Han, J.; Lu, X. Object Detection in Optical Remote Sensing Images: A Survey and a New Benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  25. Du, Z.; Hu, Z.; Zhao, G.; Jin, Y.; Ma, H. Cross-Layer Feature Pyramid Transformer for Small Object Detection in Aerial Images. arXiv 2024, arXiv:2407.19696. [Google Scholar] [CrossRef]
  26. Chen, Z. LGI-DETR: Local-Global Interaction for UAV Object Detection. arXiv 2025. [Google Scholar] [CrossRef]
  27. Zhao, D.; Shao, F.; Yang, L.; Luo, X.; Liu, Q.; Zhang, H.; Zhang, Z. Object Detection Based on an Improved YOLOv7 Model for Unmanned Aerial-Vehicle Patrol Tasks in Controlled Areas. Electronics 2023, 12, 4887. [Google Scholar] [CrossRef]
  28. Wang, Z.; Zhang, K.; Wu, F.; Lv, H. YOLO-PEL: The Efficient and Lightweight Vehicle Detection Method Based on YOLO Algorithm. Sensors 2025, 25, 1959. [Google Scholar] [CrossRef]
  29. Wei, X.; Yin, L.; Zhang, L.; Wu, F. DV-DETR: Improved UAV Aerial Small Target Detection Algorithm Based on RT-DETR. Sensors 2024, 24, 7376. [Google Scholar] [CrossRef]
  30. Wei, C.; Wang, W. RFAG-YOLO: A Receptive Field Attention-Guided YOLO Network for Small-Object Detection in UAV Images. Sensors 2025, 25, 2193. [Google Scholar] [CrossRef]
  31. Wu, S.; Lu, X.; Guo, C.; Guo, H. Accurate UAV Small Object Detection Based on HRFPN and EfficentVMamba. Sensors 2024, 24, 4966. [Google Scholar] [CrossRef]
  32. Kong, Y.; Shang, X.; Jia, S. Drone-DETR: Efficient Small Object Detection for Remote Sensing Image Using Enhanced RT-DETR Model. Sensors 2024, 24, 5496. [Google Scholar] [CrossRef] [PubMed]
  33. Baidya, R.; Jeong, H. YOLOv5 with ConvMixer Prediction Heads for Precise Object Detection in Drone Imagery. Sensors 2022, 22, 8424. [Google Scholar] [CrossRef] [PubMed]
  34. Zhang, T.; Li, L.; Zhou, Y.; Liu, W.; Qian, C.; Hwang, J.N.; Ji, X. Cas-ViT: Convolutional Additive Self-Attention Vision Transformers for Efficient Mobile Applications. arXiv 2024, arXiv:2408.03703. [Google Scholar]
  35. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  36. Cui, Y.; Ren, W.; Knoll, A. Omni-Kernel Network for Image Restoration. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; AAAI Press: Palo Alto, CA, USA, 2024; Volume 38, pp. 1426–1434. [Google Scholar]
  37. Sunkara, R.; Luo, T. No More Strided Convolutions or Pooling: A New CNN Building Block for Low Resolution Images and Small Objects. In Proceedings of the 22nd Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2022), Grenoble, France, 19–23 September 2022; pp. 443–459. [Google Scholar]
  38. Li, K.; Geng, Q.; Wan, M.; Zhou, Z. Context and Spatial Feature Calibration for Real-Time Semantic Segmentation. IEEE Trans. Image Process. 2023, 32, 5465–5477. [Google Scholar] [CrossRef]
  39. Chen, L.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  40. Lin, G.; Millan, A.; Shen, C.; Reid, I. RefineNet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 1925–1934. [Google Scholar]
  41. Li, X.; Hu, X.; Yang, J. Spatial group-wise enhance: Improving semantic feature learning in convolutional networks. arXiv 2019, arXiv:1905.09646. [Google Scholar]
  42. Huang, Z.; Wei, Y.; Wang, X.; Liu, W.; Huang, T.S.; Shi, H. Alignseg:Feature-aligned segmentation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 550–557. [Google Scholar]
  43. Du, D.; Zhu, P.; Wen, L.; Wen, L.; Bian, X.; Lin, H.; Hu, Q. VisDrone-DET2019: The vision meets drone object detection in image challenge results. In Proceedings of the IEEE/CVF international conference on computer vision workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 213–226. [Google Scholar]
  44. Chen, J.; Kao, S.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12023–12031. [Google Scholar]
  45. Liu, X.; Peng, H.; Zheng, N.; Yang, Y.; Hu, H.; Yuan, Y. Efficientvit: Memory efficient vision transformer with cascaded group attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14420–14430. [Google Scholar]
Figure 1. RT-DETR model structure.
Figure 1. RT-DETR model structure.
Sensors 25 04582 g001
Figure 2. UAV-DETR model structure.
Figure 2. UAV-DETR model structure.
Sensors 25 04582 g002
Figure 3. Architecture of CasNet.
Figure 3. Architecture of CasNet.
Sensors 25 04582 g003
Figure 4. Architecture of CSP-OmniKernel.
Figure 4. Architecture of CSP-OmniKernel.
Sensors 25 04582 g004
Figure 5. Architecture of OmniKernel.
Figure 5. Architecture of OmniKernel.
Sensors 25 04582 g005
Figure 6. Architecture of CFC module.
Figure 6. Architecture of CFC module.
Sensors 25 04582 g006
Figure 7. Architecture of SFC module.
Figure 7. Architecture of SFC module.
Sensors 25 04582 g007
Figure 8. Comparison of heatmaps of original image, RT-DETR, and UAV-DETR model. (a) Original image. (b) RT-DETR. (c) UAV-DETR.
Figure 8. Comparison of heatmaps of original image, RT-DETR, and UAV-DETR model. (a) Original image. (b) RT-DETR. (c) UAV-DETR.
Sensors 25 04582 g008
Figure 9. Comparative analysis of detection performance of original image, RT-DETR model, and UAV-DETR model: (a) multi-scale scenes; (b) occluded scenes; (c) nighttime scenes.
Figure 9. Comparative analysis of detection performance of original image, RT-DETR model, and UAV-DETR model: (a) multi-scale scenes; (b) occluded scenes; (c) nighttime scenes.
Sensors 25 04582 g009
Table 1. Experimental environment.
Table 1. Experimental environment.
ConfigurationsParameters
OSWindows 11
CPUIntel(R) Core(TM) i9-10920X@3.50 GHz
GPUNVIDIA GeForce RTX 3080 Ti 28 GB
CUDA VersionCUDA 12.1
Memory32 GB
Deep Learning FrameworkPyTorch 2.0.0
Table 2. Training parameters.
Table 2. Training parameters.
NameParameters
Epoch500
Batch Size4
Input Size640 × 640
Initial Learning Rate0.0001
Weight Decay Coefficient0.0001
OptimizerAdamW
Patience30
Table 3. Ablation study results (%).
Table 3. Ablation study results (%).
NumberModelsPRmAP@0.5mAP@0.5:0.95GFLOPsFPSParams/106
Exp1RT-DETR (Baseline)61.847.048.129.557.045.819.8
Exp2+CasNet62.947.549.230.349.931.814.9
Exp3+CasNet + SOEP Module62.849.050.731.567.632.515.8
Exp4+CasNet + SOEP Module + CSAM-Module63.250.551.632.171.433.016.8
Table 4. Comparative Backbone Network Performance.
Table 4. Comparative Backbone Network Performance.
BackbonemAP@0.5GFLOPs/G
ResNet1848.157.0
EfficientViT47.952.6
FasterNet47.654.9
CasNet49.249.9
Table 5. Comparison of experimental results.
Table 5. Comparison of experimental results.
ModelmAP@0.5Params/MGFLOPs
Faster RCNN39.741.4212.9
YOLOv5l38.846.1107.8
YOLOv6m37.234.885.6
YOLOv748.036.5103.3
YOLOv8m43.225.878.7
Enhanced YOLOv7 [26]52.335.8-
CFPT [24]50.051.3297.6
LGI-DETR [25]46.021.165.0
DV-DETR [28]50.219.584.6
Drone-DETR [31]53.928.7128.3
RT-DETR (Baseline)48.119.857.0
UAV-DETR51.616.871.4
Table 6. Generalization experiment results.
Table 6. Generalization experiment results.
ModelmAP@0.5Params/(M)GFLOPs/G
Faster R-CNN67.3136.9804.8
YOLOv774.536.6103.4
RT-DETR72.220.157.0
UAV-DETR74.316.871.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Wei, Y. UAV-DETR: An Enhanced RT-DETR Architecture for Efficient Small Object Detection in UAV Imagery. Sensors 2025, 25, 4582. https://doi.org/10.3390/s25154582

AMA Style

Zhou Y, Wei Y. UAV-DETR: An Enhanced RT-DETR Architecture for Efficient Small Object Detection in UAV Imagery. Sensors. 2025; 25(15):4582. https://doi.org/10.3390/s25154582

Chicago/Turabian Style

Zhou, Yu, and Yan Wei. 2025. "UAV-DETR: An Enhanced RT-DETR Architecture for Efficient Small Object Detection in UAV Imagery" Sensors 25, no. 15: 4582. https://doi.org/10.3390/s25154582

APA Style

Zhou, Y., & Wei, Y. (2025). UAV-DETR: An Enhanced RT-DETR Architecture for Efficient Small Object Detection in UAV Imagery. Sensors, 25(15), 4582. https://doi.org/10.3390/s25154582

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop