Next Article in Journal
Tone Mapping of HDR Images via Meta-Guided Bayesian Optimization and Virtual Diffraction Modeling
Previous Article in Journal
Deep Learning-Driven Automatic Segmentation of Weeds and Crops in UAV Imagery
Previous Article in Special Issue
GA-YOLOv11: A Lightweight Subway Foreign Object Detection Model Based on Improved YOLOv11
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RTUAV-YOLO: A Family of Efficient and Lightweight Models for Real-Time Object Detection in UAV Aerial Imagery

1
United Laboratories of TT&C and Communication, Jiuquan Satellite Launch Center, Lanzhou 732750, China
2
School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(21), 6573; https://doi.org/10.3390/s25216573
Submission received: 8 September 2025 / Revised: 16 October 2025 / Accepted: 24 October 2025 / Published: 25 October 2025
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection: 3rd Edition)

Abstract

Real-time object detection in Unmanned Aerial Vehicle (UAV) imagery is critical yet challenging, requiring high accuracy amidst complex scenes with multi-scale and small objects, under stringent onboard computational constraints. While existing methods struggle to balance accuracy and efficiency, we propose RTUAV-YOLO, a family of lightweight models based on YOLOv11 tailored for UAV real-time object detection. First, to mitigate the feature imbalance and progressive information degradation of small objects in current architectures multi-scale processing, we developed a Multi-Scale Feature Adaptive Modulation module (MSFAM) that enhances small-target feature extraction capabilities through adaptive weight generation mechanisms and dual-pathway heterogeneous feature aggregation. Second, to overcome the limitations in contextual information acquisition exhibited by current architectures in complex scene analysis, we propose a Progressive Dilated Separable Convolution Module (PDSCM) that achieves effective aggregation of multi-scale target contextual information through continuous receptive field expansion. Third, to preserve fine-grained spatial information of small objects during feature map downsampling operations, we engineered a Lightweight DownSampling Module (LDSM) to replace the traditional convolutional module. Finally, to rectify the insensitivity of current Intersection over Union (IoU) metrics toward small objects, we introduce the Minimum Point Distance Wise IoU (MPDWIoU) loss function, which enhances small-target localization precision through the integration of distance-aware penalty terms and adaptive weighting mechanisms. Comprehensive experiments on the VisDrone2019 dataset show that RTUAV-YOLO achieves an average improvement of 3.4% and 2.4% in mAP50 and mAP50-95, respectively, compared to the baseline model, while reducing the number of parameters by 65.3%. Its generalization capability for UAV object detection is further validated on the UAVDT and UAVVaste datasets. The proposed model is deployed on a typical airborne platform, Jetson Orin Nano, providing an effective solution for real-time object detection scenarios in actual UAVs.

1. Introduction

The rapid advancement and widespread deployment of Unmanned Aerial Vehicle (UAV) technology has positioned these platforms as critical infrastructure components across numerous application sectors, including intelligent urban infrastructure management [1,2], mineral and resource exploration [3], precision agricultural monitoring [4,5], disaster response and assessment [6], and perimeter security operations [7]. UAV-mounted imaging systems acquire substantial volumes of visual data from elevated vantage points, providing a crucial source of information for decision-making in these applications.
The task of real-time object detection in UAV aerial imagery presents multifaceted challenges that distinguish it from conventional object detection scenarios. Variations in flight altitude, viewing angle, and sensor focal length introduce substantial intra-class scale heterogeneity across acquired imagery, creating complex multi-scale object distributions that demand more efficient processing [2,3]. Targets in UAV aerial imagery are often extremely small [4]. In this study, small targets in UAV aerial imagery are defined based on their pixel area. Specifically, a target is considered small if its bounding box occupies fewer than 32 × 32 pixels. Small pixel areas lead to severe information sparsity, where discriminative features are highly compressed and prone to degradation during convolutional and downsampling operations. Additionally, unstructured backgrounds such as vegetation, urban areas, and water bodies introduce complex textures, illumination changes, and distractors, contributing to high false-positive and false-negative rates [6]. Furthermore, UAV platforms are subject to stringent hardware constraints—including limited computation, memory, and power—which necessitate highly efficient algorithms capable of balancing real-time performance with detection accuracy under embedded processing limitations.
Although traditional object detection architectures (including R-CNN [8], Faster R-CNN [9], SSD [10], and YOLO series [11]) have demonstrated robust performance on general scene benchmark datasets, their structures are mainly optimized for feature extraction of medium and large objects, and have limited representation capabilities for multi-scale and small object detection. The deep networks of general architectures bring huge memory burdens and computational requirements, making them difficult to deploy on resource-limited UAV platforms.
In recent years, a variety of real-time UAV detection architectures have been proposed. Weng et al. [12] introduced a multi-scale dilated attention mechanism and a GhostConv lightweight module to reduce algorithmic complexity while simultaneously enhancing multi-scale object feature extraction capabilities; Liu et al. [13] employed a gated single-head attention and large convolution kernel mechanism to achieve efficient extraction of local detail features; and Luo et al. [14] enhanced the model’s small object feature extraction capabilities while preserving computational efficiency by integrating multi-scale attention mechanisms and Shuffle block algorithms into the backbone network architecture. Although these methods have made some progress in improving the accuracy of real-time UAV target detection and reducing model complexity, there is still room for improvement in balancing detection accuracy and computational efficiency. Furthermore, these approaches fail to provide scalable model variants of varying computational complexity for practical deployment on UAV platforms with diverse computational resource constraints.
To address these limitations, we propose a family of real-time UAV detection models, termed RTUAV-YOLO, based on YOLOv11 architecture. This model further improves detection accuracy and computational efficiency, and constructs models of varying scales to accommodate the practical deployment requirements of various UAV platform hardware. The principal contributions of this work can be summarized as follows:
(1)
To address the loss of contextual information for small objects caused by multiple downsampling in YOLOv11, we introduced a Progressive Dilated Separable Convolution Module (PDSCM) in the P4–P5 phase of the backbone network. This module reduces feature sampling and uses expanding depth-wise separable convolutions to detect objects at different scales. It establishes contextual spatial relationships between these scales, enhancing the representation of small objects at multiple scales. A P2 detection head is introduced to the head, allowing the model to focus more on detailed features, thereby improving its sensitivity to small objects.
(2)
To address the issues of unbalanced features and decreased information content in small-scale object processing in traditional architectures, we designed a lightweight Multi-Scale Feature Adaptive Modulation module (MSFAM) and replaced the C3K2 module in the backbone network. This significantly enhanced the feature extraction capability of small objects through an adaptive weight generation mechanism and dual-channel heterogeneous feature aggregation. We also introduced a Lightweight DownSampling Module (LDSM) to replace the convolutional modules in the backbone and Neck, achieving efficient downsampling of feature maps while reducing computational complexity and preserving key small-object features.
(3)
To rectify the insensitivity of conventional Intersection over Union (IoU) metrics to-ward small objects, we develop a Minimum Point Distance Wise IoU (MPDWIoU) loss function. This loss function integrates minimum point distance metrics, dynamic anchor focusing strategies, and auxiliary bounding box supervision to specifically mitigate gradient imbalance and low precision in small-target localization while enhancing regression robustness in cluttered environments.
(4)
Comprehensive experiments on the VisDrone2019 dataset confirms that RTUAV-YOLO achieves an effective balance between computational efficiency and detection performance. RTUAV-YOLO achieves an average improvement of 3.4% and 2.4% in mAP50 and mAP50-95, respectively, compared to the baseline model YOLOv11, while reducing the number of parameters by 65.3%. Comprehensive comparison results across different model series are presented in Figure 1. Furthermore, cross-dataset evaluations on the UAVDT and UAVVaste datasets confirm RTUAV-YOLO’s generalization and robustness across diverse UAV aerial imagery conditions. The complete implementation code is publicly available at https://gitee.com/zhangruizhi0110/RTUAV-YOLO (accessed on 20 October 2025).

2. Related Works

2.1. General Object Detection Models and Limitations

Contemporary deep learning-based object detection methodologies are categorized into two principal paradigms: two-stage and single-stage architectures. Two-stage frameworks encompass R-CNN [8], SPPNet [15], Fast R-CNN [16], and Faster R-CNN [9]. These frameworks demonstrate superior accuracy and robustness when processing complex visual scenes; however, they exhibit substantial computational demands and prolonged inference latency. Single-stage architectures formulate object detection as a unified regression task, directly predicting bounding box coordinates and class probabilities from input imagery without intermediate region proposal generation. This paradigm substantially enhances inference velocity while concurrently reducing computational complexity. The YOLO family of architectures [11] represents the most prominent exemplar of this paradigm. YOLO architectures have achieved widespread adoption in practical applications due to their speed–accuracy trade-off characteristics. The most recent iteration, YOLOv11 [17], enhances feature extraction and processing pipelines through the integration of novel architectural components including C3k2 and C2PSA modules, thereby achieving improved detection accuracy while simultaneously reducing parametric complexity and computational overhead, ultimately demonstrating superior performance on standard benchmarks such as COCO [18].
Nevertheless, direct application of these general architectures to UAV aerial imagery reveals several critical limitations: (1) Multi-scale and small objects detection deficiencies: Successive downsampling operations in deep networks result in substantial information loss for diminutive objects and inadequate exploitation of fine-grained shallow features. (2) Background interference susceptibility: Conventional feature extraction mechanisms and detection head architectures exhibit limited robustness against the heterogeneous backgrounds characteristic of aerial imagery. (3) Computational resource incompatibility: High-precision deep architectures impose substantial computational burdens that exceed the processing capabilities of resource-constrained UAV platforms, thereby precluding actual deployment.
While general object detection frameworks establish foundational detection paradigms, their architectural designs and optimization objectives are difficult to adapt to the inherent complex characteristics of UAV aerial images, necessitating domain-specific architectural innovations.

2.2. Object Detection Model for UAV Aerial Images

To address the shortcomings of general target detection models, researchers has developed numerous object detection architectures specifically tailored for UAV aerial imagery. Numerous approaches focus on enhancing multi-scale feature extraction capabilities to address small-target detection challenges. Li et al. [19] achieved multi-scale feature fusion by introducing the receptive field convolution block attention module and the balanced spatial semantic information fusion pyramid network, effectively balancing spatial and semantic information; Zhou et al. [20] introduced the non-semantic sparse attention (NSSA) mechanism and the bidirectional multi-branch auxiliary feature pyramid network to achieve significant improvement in small object detection performance; Guo et al. [21] presented AugFPN to mitigate the semantic-detail information disparity inherent in feature representations, employing single-pass supervision during feature fusion to reduce this informational discrepancy. While these approaches substantially enhance small-target recall performance, the incorporation of sophisticated architectural components frequently results in considerable computational overhead.
To mitigate false negatives and false positives arising from background complexity, Cheng et al. [22] employed dual attention mechanisms for pre-fusion feature enhancement, enabling selective focus on discriminative target characteristics; Zhang et al. [23] introduced a feature enhancement module leveraging spatial and channel attention mechanisms to augment feature representations, thereby improving target region localization; Li et al. [24] incorporated large-kernel convolutions coupled with dual-channel attention to enhance complex small-target feature extraction while effectively suppressing background interference. These attention-based methodologies substantially improve model robustness against environmental noise and background clutter.
To ensure that the designed UAV real-time target detection model meets the actual deployment requirements, model lightweight technology must be adopted. Model lightweight technology typically encompass two primary paradigms. The first approach involves structural pruning [25,26,27,28], which systematically eliminates redundant parameters below predefined thresholds through algorithmic filtering mechanisms. Pruning techniques offer universal applicability across diverse architectural frameworks for parameter reduction. The alternative paradigm employs lightweight modular design for structural optimization. This approach centers on developing computationally efficient architectural innovations. Prominent examples include MobileNet [29], ShuffleNet [30], and GhostNet [31] leverage depth-wise separable convolutions (DWConv) and grouped convolutions for spatial feature extraction. DWConv architectures achieve substantial reductions in both parametric complexity and floating-point operations (FLOPs). Architectural optimization has emerged as the predominant strategy in UAV real-time object detection model lightweight owing to its superior potential for achieving optimal speed–accuracy trade-offs.
Despite advances in UAV-specific detection frameworks, approaches that enhance small-target detection or background robustness typically incur significant computational costs, whereas lightweight variants exhibit suboptimal detection performance [32,33]. Consequently, there exists a critical need for innovative lightweight architectures specifically engineered for UAV aerial imagery that can achieve state-of-the-art detection performance within stringent computational constraints.

2.3. Bounding Box Regression Loss Function

Bounding box regression accuracy constitutes a critical determinant of detection performance, particularly for small objects exhibiting challenging localization characteristics. Contemporary IoU-based loss functions including GIoU [34], DIoU [35], and CIoU [36] enhance convergence properties and localization precision through the incorporation of auxiliary geometric information such as centroid distances and aspect ratio constraints. The WIoU family [37] (encompassing variants v1, v2, and v3) employs adaptive weighting mechanisms to enhance regression robustness. Focal-IoU [38] identifies the insufficient gradient contribution of conventional IoU loss for small targets and addresses this limitation through adaptive loss weighting strategies during small-target detection.
Despite their individual merits, these loss formulations exhibit limitations when addressing the unique challenges inherent in UAV-based small-target detection [39], including substantial IoU sensitivity to minute positional perturbations and gradient imbalance induced by extensive background interference proximate to target boundaries. Specifically, existing works demonstrates insufficient investigation into precise minimum point distance modeling and adaptive focusing strategies tailored for small-target instances. This gap motivated our development of the MPDWIoU loss function, which integrates minimum point distance metrics, dynamic attention mechanisms, and auxiliary supervision strategies. This formulation specifically targets enhanced small-target localization optimization while mitigating the inherent deficiencies of conventional approaches in UAV aerial surveillance contexts.

3. Methods

3.1. RTUAV-YOLO Overall Framework

Figure 2 illustrates the RTUAV-YOLO architecture presented in this work, which incorporates systematic enhancements to the YOLOv11 baseline framework. Initially, we replace the traditional convolutional module with the PDSCM in the P4 stage of the backbone network, eliminating the one-time downsampling of feature maps. The PDSCM promotes efficient aggregation of multi-scale receptive fields by integrating depth-wise separable convolutions with gradually increasing dilation rates, fully utilizing contextual information. We integrate the MSFAM as a replacement for the C3K2 components within the backbone architecture. The MSFAM substantially augments the model’s representational capacity and multi-scale target adaptability via three fundamental mechanisms: dynamic weight synthesis, multi-scale feature extraction, and adaptive feature modulation. Dynamic weight synthesis generates dynamic weight coefficients through adaptive average pooling and convolutional layers, dynamically allocating attention resources based on target scale and complexity. Multi-scale feature extraction utilizes a dual-branch parallel architecture to simultaneously extract and complement features at different scales. Adaptive feature modulation adaptively enhances useful features and suppresses redundant information based on input content, significantly improving the accuracy and robustness of small-target detection. The LDSM is deployed across both backbone and neck components. This module reduces parametric complexity and computational overhead via grouped and pointwise convolutions, simultaneously accomplishing spatial downsampling, feature aggregation, and channel-wise transformations. We incorporate the MPDWIoU loss function, which integrates minimum point distance metrics, dynamic attention weighting, and auxiliary supervisory mechanisms. These mechanisms specifically target gradient imbalance challenges in small-target localization while enhancing bounding box regression robustness against complex background interference. These lightweight modules, along with the introduction of IoU specifically for small object detection, enable RTUAV-YOLO to achieve excellent accuracy while reducing computational complexity. This significantly improves the model’s detection performance and generalization capabilities for UAV aerial imagery detection. The following sections will fully describe the architectural design and operating principles of each module.

3.2. Multi-Scale Feature Adaptive Modulation Module

In object detection applications, feature representational quality constitutes a fundamental determinant of detection efficacy. Conventional convolutional architectures encounter several critical limitations when addressing multi-scale object detection: (1) Limited content adaptivity: Static weight parameters fail to dynamically modulate feature significance based on contextual image characteristics. (2) Inadequate scale accommodation: Multi-scale objects necessitate diverse receptive field configurations and scale-specific feature processing mechanisms. (3) Suboptimal feature integration: Conventional concatenation or summation operations inadequately exploit the complementary nature of multi-scale feature representations. To mitigate these limitations, we introduce the MSFAM, grounded in adaptive attention theory and multi-scale feature integration principles. This module facilitates adaptive feature importance calibration through input-dependent dynamic weight learning while capturing multi-scale feature information via parallel multi-branch architectures. The MSFAM serves as a direct replacement for C3K2 modules within the YOLOv11 backbone, implementing computationally efficient feature extraction and fusion through reasonable architectural design.
Compared to the baseline YOLO and various YOLO-based UAV object detection models, these variants use fixed convolution kernels for multi-scale feature extraction, and their network designs at different depths do not optimize for semantic and positional information. Furthermore, their feature extraction modules do not dynamically prioritize features based on object scale. To address these issues, MSFAM utilizes a dynamic weight generator to enhance the relevance of features for small objects, a dual-branch feature extractor for efficient extraction and fusion of multi-scale information, and adaptive feature fusion for the effective fusion of semantic and positional information. This enables RTUAV-YOLO to better capture fine-grained details and contextual information, significantly improving the accuracy of small object detection in UAV imagery.
The MSFAM employs a three-stage cascaded architecture comprising three fundamental components: the Dynamic Weight Generator (DWG), Dual-Branch Feature Extractor (DBFE), and Adaptive Feature Fusion (AFF) mechanism. Given an input feature tensor X R B × C × H × W , the forward inference process of MSFAM is formulated as:
Y = MSFAM ( X ) = AFF ( DBFE ( X ) , DWG ( X ) ) + X
where B, C, H, and W represent the batch size, number of channels, height, and width, respectively. Figure 3 shows the MSFAM used at different network depths.

3.2.1. Dynamic Weight Generator

The Dynamic Weight Generator (DWG) leverages global average pooling coupled with a two-layer multilayer perceptron (MLP) to synthesize adaptive modulation weights from global feature statistics, enabling dynamic feature importance calibration. The DWG initially aggregates global statistical representations from input features via global average pooling:
g = 1 H W i = 1 H j = 1 W X : , : , i , j
Subsequently, the two-layer MLP executes nonlinear transformations to accomplish dimensionality reduction and weight synthesis:
h = ReLU ( W 1 g + b 1 )
w = σ ( W 2 h + b 2 )
where W 1 R C / r × C , W 2 R C × C / r , and r denotes the dimensionality reduction ratio. The reduction ratio r is adaptively configured based on network hierarchy, embodying the principle that shallow layers necessitate enriched feature representations while deeper layers prioritize computational efficiency.
To ensure weight stability and optimization efficacy, we employ L2 normalization:
w n o r m = w | | w | | 2 + ϵ
where ϵ represents a numerical stability constant to prevent division by zero. L2 normalization projects the weight vector onto the unit hypersphere, ensuring | w n o r m | | 2 = 1 , thus avoiding gradient explosion and providing a stable training process.
The adaptive weight modulation mechanism enables content-aware feature weight adjustment while generating channel-specific modulation coefficients, thereby amplifying responses in feature maps corresponding to small-target regions.

3.2.2. Dual-Branch Feature Extractor

The dual-branch feature extractor employs a 3:1 channel allocation strategy coupled with an asymmetric dual-branch parallel architecture to facilitate efficient extraction and fusion of multi-scale information via distinct feature processing pathways. The specific implementation of the branch is differentiated according to the depth of the network. Shallow features exhibit high spatial resolution and rich detail information albeit with limited semantic content, whereas deep features demonstrate elevated semantic levels and abundant contextual information despite reduced spatial resolution. Within the shallow network layers P2 and P3, we implement a sophisticated feature extraction strategy to exploit the inherent richness of shallow features. Branch 1 employs standard 3 × 3 convolution to capture local detail features, while Branch 2 utilizes dilated convolution with a dilation rate of 2 to expand the receptive field and acquire broader contextual information. This architecture fully exploits the high-resolution characteristics inherent to shallow features. Its forward propagation process is as follows:
F 1 = Conv 3 × 3 ( Conv 1 × 1 ( X ) )
F 2 = DilatedConv 3 × 3 , d = 2 ( Conv 1 × 1 ( X ) )
Within the deep network layer P4, we adopt a streamlined feature extraction strategy to satisfy both semantic richness and computational efficiency requirements. Branch 1 employs two consecutive 1 × 1 convolutions for feature transformation while reducing computational complexity, whereas Branch 2 preserves the dilated convolution architecture to maintain contextual awareness. This architecture effectively balances semantic expressiveness with computational efficiency in deep network layers. Its forward propagation process is as follows:
F 1 = Conv 1 × 1 ( Conv 1 × 1 ( X ) )
F 2 = DilatedConv 3 × 3 , d = 2 ( Conv 1 × 1 ( X ) )

3.2.3. Adaptive Feature Fusion

The adaptive feature fusion module seamlessly integrates information from the primary feature branch and the contextual branch. The module adaptively modulates feature importance according to target scale during small object feature extraction, dynamically amplifying the weights of small object-related features while maintaining stable performance enhancements via residual connections. This fusion mechanism not only enhances feature representation quality but also preserves computational efficiency and training stability. Initially, the output features from both branches are consolidated via channel concatenation:
F concat = Concat ( [ F 1 , F 2 ] , dim = 1 )
Subsequently, a 1 × 1 convolution operation is applied to restore the original channel dimensionality:
F fused = Conv 1 × 1 ( F concat )
Incorporating the dynamic weights generated by the DWG, the final feature modulation process becomes:
Y = w norm F fused + X
where denotes element-wise multiplication, and the residual connection guarantees gradient flow stability and comprehensive information preservation.

3.3. Progressive Dilated Separable Convolution Module

Small objects in UAV aerial imagery are frequently obscured by complex background interference, yielding constrained informative content. To comprehensively exploit available feature information, we implement the PDSCM to detect multi-scale objects and establish inter-scale spatial relationships, thereby augmenting contextual and small object feature representations. This module accomplishes multi-scale feature extraction and progressive receptive field expansion via a cascade of depth-wise separable convolutions with incrementally increasing dilation rates. This module is integrated within the P4 stage of the backbone architecture. The comprehensive architecture of the PDSCM and corresponding layer-wise feature extraction heatmaps are illustrated in Figure 4.
The PDSCM employs a progressive dilation rate strategy to accomplish incremental receptive field expansion across five convolutional layers. For dilated convolution operations, the effective receptive field is computed as:
R F e f f e c t i v e = ( k 1 ) × d + 1
Considering the concatenation of convolutional layers, the cumulative receptive field becomes:
R F c u m u l a t i v e = i = 1 n R F i ( n 1 )
The resulting cumulative receptive field of the PDSCM yields R F t o t a l = 3 + 5 + 7 2 = 13 .
The PDSCM leverages progressive dilation to process multi-scale information hierarchically. For local detail extraction, dilated convolutions with unit dilation rate capture fine-grained local features. For intermediate-scale contextual analysis, dilated convolutions with dilation rate 2 extract mid-level contextual information. For global contextual modeling, dilated convolutions with dilation rate 3 acquire large-scale global information. The final output incorporates residual connections, preserving information flow while enhancing training stability. Through this meticulously engineered architecture, the PDSCM achieves optimal balance between computational efficiency and feature representational capacity.
In contrast to conventional YOLO variants, which rely on standard convolution operations with fixed receptive fields, the PDSCM model progressively increases the dilation rate, thereby facilitating enhanced multi-scale context extraction. This approach proves especially advantageous for detecting small objects in complex UAV environments, where small targets often require larger receptive fields to be distinguished from cluttered backgrounds. By more effectively expanding the receptive field, PDSCM significantly improves small-object localization and enhances overall detection accuracy.

3.4. Lightweight DownSampling Module

The LSDM constitutes a purpose-built efficient downsampling component. This module executes the essential functions of spatial resolution reduction and channel dimensionality modulation within the RTUAV-YOLO architecture. It accomplishes lightweight feature downsampling via a meticulously engineered dual-stage convolutional framework.
The initial stage concentrates on spatial downsampling, whereas the subsequent stage addresses channel dimensionality modulation. This approach circumvents complex channel transformations during downsampling operations and enhances gradient backpropagation stability. Grouped convolution preserves feature locality, whereas pointwise convolution facilitates efficient inter-channel information exchange. Conventional convolution modules typically employ a “convolution + batch normalization + activation” paradigm, performing direct downsampling with stride 2. This nonlinear processing can distort or alias detailed features prior to downsampling, rendering it inappropriate for subtle small object textures. Conversely, the LDSM initially executes linear grouped convolution for downsampling, effectively implementing a “sample-first, nonlinearity-second” strategy that aligns more closely with signal processing principles and superior detail preservation. Subsequently, it executes channel fusion and nonlinear transformation via 1×1 pointwise convolution. This lightweight downsampling module decomposes complex operations into elementary, independent sub-operations, substantially reducing parametric complexity while maintaining feature representational fidelity. It proves particularly advantageous for applications demanding high-performance small object detection under computational resource constraints. The comprehensive architecture of the LSDM is illustrated in Figure 5. Unlike conventional YOLO variants that typically rely on standard convolutional layers for downsampling, LDSM minimizes distortions during downsampling by adopting a dual-stage convolutional framework, ensuring that small-object textures remain intact. This design enables RTUAV-YOLO to achieve more precise detection of small targets in UAV imagery, where maintaining fine spatial details is essential for accurate detection in complex and cluttered environments.

3.5. Minimum Point Distance Wise IoU

Intersection over Union (IoU) constitutes a fundamental metric for object detection evaluation and bounding box regression loss computation. However, conventional IoU variants exhibit constrained sensitivity to spatial misalignment in small object scenarios, yielding suboptimal detection performance for objects with minimal pixel coverage.
Conventional IoU and its variants exhibit the following critical limitations when processing small objects: (1) gradient vanishing: when predicted and ground-truth boxes lack intersection, IoU equals zero, yielding vanishing gradients; (2) inadequate small object sensitivity: minor pixel displacement substantially affects small objects, yet IoU variation remains negligible; (3) scale dependency: identical absolute displacement disproportionately impacts objects across varying scales. To address these limitations, we introduce MPDWIoU, a novel metric specifically engineered to address small object detection challenges. It integrates multi-dimensional geometric relationships, an adaptive focusing mechanism, and auxiliary IoU computation to deliver efficient localization guidance for small objects while preserving robustness across diverse object scales.
Assume that the two bounding boxes are:
B 1 = ( x 1 1 , y 1 1 , x 2 1 , y 2 1 )
B 2 = ( x 1 2 , y 1 2 , x 2 2 , y 2 2 )
where ( x 1 , y 1 ) and ( x 2 , y 2 ) represent the coordinates of the upper left corner and the lower right corner, respectively.
The width and height are defined as follows:
w 1 = x 2 1 x 1 1 , h 1 = y 2 1 y 1 1
w 2 = x 2 2 x 1 2 , h 2 = y 2 2 y 1 2
MPDWIoU establishes direct geometric constraints through the minimum point distance (MPD) mechanism. The fundamental principle involves computing the minimum Euclidean distance among all corner point pairs between predicted and ground-truth boxes. This distance directly quantifies the closest spatial proximity between predicted and ground-truth boxes, furnishing pixel-level positional constraints for model optimization. Specifically, the MPD mechanism extracts the coordinates of the four corner vertices from both predicted and ground-truth boxes, computes the Euclidean distances between all corner pairs to identify the minimum value, and subsequently employs the diagonal length of the minimum circumscribed rectangle encompassing both boxes for normalization to ensure scale invariance. This architectural design enables MPD to deliver more precise positional gradient signals compared to traditional IoU metrics, particularly when predicted boxes exhibit slight spatial offsets. While IoU variations remain minimal, MPD can furnish substantially clearer gradient signals to guide model optimization. The detailed computational procedure for MPD is delineated as follows.
Initially, the Euclidean distances between the four corresponding corner vertices are computed:
d 1 = ( x 1 1 x 1 2 ) 2 + ( y 1 1 y 1 2 ) 2
d 2 = ( x 2 1 x 2 2 ) 2 + ( y 1 1 y 1 2 ) 2
d 3 = ( x 1 1 x 1 2 ) 2 + ( y 2 1 y 2 2 ) 2
d 4 = ( x 2 1 x 2 2 ) 2 + ( y 2 1 y 2 2 ) 2
Subsequently, MPD is defined and the diagonal length of the minimum bounding convex hull is determined:
MPD = min ( d 1 , d 2 , d 3 , d 4 )
c w = max ( x 2 1 , x 2 2 ) min ( x 1 1 , x 1 2 )
c h = max ( y 2 1 , y 2 2 ) min ( y 1 1 , y 1 2 )
c 2 = c w 2 + c h 2
The resultant MPD penalty term is formulated as:
P mpd = MP D 2 c 2
The aspect ratio penalty term is computed employing an adaptive weighting factor, leveraging the aspect ratio consistency component of CIoU:
v = 4 π 2 arctan w 2 h 2 arctan w 1 h 1 2
α = v v IoU + 1 + ϵ
P ar = α v
To ensure enhanced attention allocation to small objects while preventing excessive penalization of high-quality samples, scale-adaptive weight computation is executed utilizing a small object threshold of 32 × 32 = 1024 pixels. This mechanism is subsequently integrated with quality-aware suppression to derive a dynamic focus weight. The corresponding computational formulation is expressed as follows:
A 1 = w 1 h 1 ,     A 2 = w 2 h 2
A min = min ( A 1 , A 2 )
β = A min A min + 1024
γ = 2 2 IoU
W f = β exp ( γ )
The Inner-IoU enhancement mechanism augments the model’s sensitivity to positional variations in small objects through the construction of a scaled auxiliary bounding box. Through computing the IoU between the predicted box and the scaled ground-truth box, followed by weighted fusion with the original IoU, the precision of small object detection is substantially enhanced. The corresponding mathematical expression is delineated as follows, where the scaling parameter s = 1.25:
w ^ 1 = s w 1 ,     h ^ 1 = s h 1
w ^ 2 = s w 2 ,     h ^ 2 = s h 2
c 1 x = x 1 1 + x 2 1 2 , c 1 y = y 1 1 + y 2 1 2
c 2 x = x 1 2 + x 2 2 2 , c 2 y = y 1 2 + y 2 2 2
B 1 = c 1 x w ^ 1 2 , c 1 y h ^ 1 2 , c 1 x + w ^ 1 2 , c 1 y + h ^ 1 2
B 2 = c 2 x w ^ 2 2 , c 2 y h ^ 2 2 , c 2 x + w ^ 2 2 , c 2 y + h ^ 2 2
Io U aux = IoU ( B 1 , B 2 )
Integrating all constituent elements, MPDWIoU is formulated as:
MPDWIoU = IoU P mpd P ar + W f ( Io U aux IoU )
MPDWIoU addresses the fundamental gradient vanishing problem inherent in traditional IoU metrics under non-intersecting conditions via the minimum point distance mechanism. The framework incorporates dynamic focus weights to achieve adaptive attention allocation for small targets while augmenting boundary localization sensitivity through an integrated scaling mechanism. The method substantially enhances small-target detection accuracy and training stability while preserving computational efficiency, thereby furnishing a more precise and robust loss function solution for object detection in small-target-dense environments.

4. Experiments and Results

4.1. Datasets

To validate the efficacy of the proposed RTUAV-YOLO model, comprehensive comparative experiments were conducted utilizing the publicly available VisDrone2019 dataset [40], with model generalization capabilities further assessed on two additional datasets: UAVDT [41] and UAVVaste [42]. The VisDrone2019 dataset was developed by the AISKYEYE team within the Machine Learning and Data Mining Laboratory at Tianjin University. It is a large-scale UAV aerial image benchmark dataset collected under various low-altitude conditions using UAV. The dataset encompasses imagery acquired across varied real-world scenarios, diverse meteorological conditions, and multiple target scales. The dataset comprises ten distinct object categories: pedestrians, people, bicycles, cars, trucks, motorcycles, vans, tricycles, awning-tricycles, and buses. Specifically, the dataset contains 6471 training images, 548 validation images, and 3190 test images. Figure 6 illustrates the distribution of annotation categories and object size characteristics. The primary detection challenges inherent in this dataset manifest in several key aspects: (1) the substantial quantity of objects requiring detection, predominantly small-scale targets (defined as objects with bounding boxes smaller than 32 × 32 pixels); (2) significant inter-object occlusion and spatial overlap; (3) imbalanced class distribution. While these challenges substantially increase detection complexity, the dataset’s inherent richness and diversity establish a robust foundation for comprehensive model performance evaluation. The training and validation subsets of VisDrone2019 were employed for model training and performance assessment.
UAVDT constitutes a large-scale benchmark dataset specifically designed for unmanned aerial vehicle-based detection and tracking applications. This comprehensive dataset, comprising approximately 80,000 representative frames meticulously curated from 10 h of raw video footage, is purposefully constructed to address emerging challenges inherent in complex visual scenarios. The dataset incorporates comprehensive bounding box annotations alongside up to 14 attribute annotations (including meteorological conditions, flight altitude, camera perspective, vehicle classification, and occlusion characteristics), thereby supporting three fundamental computer vision tasks: object detection (DET), single object tracking (SOT), and multiple object tracking (MOT). The dataset encompasses annotations for approximately 2700 vehicles, totaling approximately 840,000 bounding boxes, thus establishing a crucial benchmark for aerial object detection and tracking research. This investigation utilizes the object detection subset of the dataset, which comprises 38,327 images with an average spatial resolution of 1080 × 540 pixels. The dataset encompasses three distinct object categories: passenger vehicles, buses, and trucks. The dataset comprises 23,258 training images and 15,069 evaluation images. UAVVaste constitutes a specialized dataset purposefully designed for aerial debris detection applications. This dataset contains 772 images and 3716 annotated landmarks depicting debris across diverse urban and natural environments, including streets, parks, and residential areas. For both datasets, the training were employed for model training, while the test were utilized for performance evaluation.

4.2. Implementation Details

All experiments were conducted on a high-performance computational server equipped with an Intel Core i7-14700K CPU and an NVIDIA GeForce RTX 4090 GPU. The software environment was built on Ubuntu 22.04, with Python 3.8, PyTorch 2.0, and CUDA 11.8.
Our model was trained from scratch for 300 epochs using a Stochastic Gradient Descent (SGD) optimizer. The batch size was set to 16, and the number of data loading workers was configured to 8. Input images were uniformly resized to 640 × 640 pixels to balance detection accuracy and computational load.
Key training hyperparameters were configured as follows: the optimizer momentum was set to 0.937 and the weight decay was 0.0005. The learning rate scheduling employed a linear warmup for the initial 3 epochs, followed by a cosine annealing strategy, starting with an initial learning rate of 0.01 and decaying to a final learning rate of 0.0001. An early stopping mechanism with a patience of 50 epochs was implemented to terminate the training if no improvement in validation metrics was observed, ensuring the selection of the best-performing model weights. For experimental consistency, all baseline models were trained under the identical hyperparameter configuration.
To enhance model robustness and prevent overfitting, we employed a comprehensive suite of data augmentation techniques. These included mosaic augmentation (enabled for the entire training duration except for the final 10 epochs), as well as random perspective shifts, hue, saturation, and value adjustments.
Taking the RTUAV-YOLO-X version as an example, Figure 7 shows its training results on the VisDrone2019 dataset. The maximum number of epochs configured for the entire training process was 300. However, due to the activation of the early stopping mechanism, training automatically terminated after 278 epochs. The total training time was 78,892.3 s, or approximately 21.92 h. This translates to an average training time of approximately 284 s per epoch. The model demonstrated stable convergence during training. The key performance metric, mAP50-95, showed a continuous upward trend, starting from 0.116 in the first epoch and peaking at approximately 0.324 in the 227th epoch. Thereafter, mAP50-95 began to stabilize with only minor fluctuations. Following the early stopping mechanism, validation set performance showed no improvement for 50 consecutive epochs, and training was terminated at epoch 278. This indicates that the model has reached its optimal convergence point, and further training may result in diminishing returns or potential overfitting. Analysis of the loss curves shows that the training process is stable and successful, without common problems such as gradient vanishing, and ultimately a detection model with strong generalization ability and excellent performance is constructed.

4.3. Evaluation Metrics

To comprehensively assess the efficacy of RTUAV-YOLO for small object detection tasks, this investigation employs precision (P), recall (R), mean average precision (mAP), parameter count (Params), and computational complexity as quantitative evaluation metrics.

4.3.1. Precision

Precision quantifies the proportion of correctly identified objects relative to the total number of predicted detections, thereby characterizing the model’s classification accuracy. The mathematical formulation is expressed as:
P = T P T P + F P
where TP denotes the count of true positive instances (correctly classified positive samples), and FP represents the count of false positive instances (incorrectly classified negative samples as positive).

4.3.2. Recall

Recall quantifies the proportion of correctly identified positive instances relative to the total number of actual positive samples, thereby indicating the model’s detection sensitivity and coverage capability. The mathematical expression is formulated as:
R = T P T P + F N
where FN is the number of samples predicted as negative but actually positive.

4.3.3. Mean Average Precision

Average Precision (AP) quantifies the detection performance for individual object categories by computing the area under the precision-recall curve, thereby capturing the trade-off relationship between precision and recall metrics. Mean Average Precision (mAP) provides a comprehensive assessment of the model’s overall detection capability by statistically aggregating the average precision values across all target categories. The computational framework is expressed as:
AP = 0 1 P RdR
m A P = 1 C c = 1 C A P c
where C denotes the total number of object categories within the detection framework, and A P c   represents the average precision metric for the C category target. To provide a more comprehensive performance evaluation, this investigation employs two distinct mAP metrics: mAP50 quantifies the mean average precision at a fixed IoU threshold of 0.5, while mAP50-95 represents the averaged precision across multiple IoU thresholds ranging from 0.5 to 0.95 with incremental steps of 0.05.

4.3.4. Model Parameter Scale

Model parameter count quantifies the total number of trainable parameters within the neural network architecture, serving as a fundamental metric for assessing model complexity, memory requirements, and computational overhead. Reduced parameter counts enable efficient model deployment in resource-limited computational environments, thereby enhancing practical applicability.

4.3.5. Floating Point Operations

Computational complexity, quantified through floating-point operations per second, represents the aggregate number of arithmetic operations executed by the neural network architecture during a single forward propagation cycle, typically expressed in giga-FLOPs (GFLOPs). Reduced GFLOPs requirements correspond to enhanced computational efficiency during inference, thereby facilitating deployment in resource-constrained real-time detection scenarios, particularly within UAV platforms where computational resources are inherently limited.

4.4. Comparative Experiments

4.4.1. Experimental Results on the VisDrone2019 Dataset

To rigorously assess the efficacy of the proposed methodology, comprehensive comparative evaluations were conducted across diverse state-of-the-art (SOTA) object detection architectures utilizing the VisDrone2019 dataset.
The quantitative comparison results with the baseline model, presented in Table 1, encompass performance metrics for the YOLOv8 [43] family, YOLOv11 [17] family, and the proposed RTUAV-YOLO series architectures.
With respect to mAP50 evaluation, RTUAV-YOLO-X demonstrated superior detection performance among all evaluated architectures, attaining 51.4% accuracy. This constitutes a substantial improvement of 4.5 percentage points (9.6% relative enhancement) over the baseline YOLOv11-X performance of 46.9%. Under the more rigorous mAP50-95 evaluation protocol, RTUAV-YOLO-X maintained exceptional performance, achieving 32.4%, corresponding to a significant improvement of 3.4 per-centage points (11.7% relative gain) over YOLOv11-X’s baseline of 29.0%. Notably, the RTUAV-YOLO architecture family exhibited consistent performance superiority across all model scales.
The RTUAV-YOLO architecture series exhibits an optimal precision-recall equilibrium. Specifically, RTUAV-YOLO-X attains precision and recall values of 61.3% and 48.1%, respectively, corresponding to improvements of 3.6 and 3.3 percentage points over YOLOv11-X’s baseline performance (57.7% precision, 44.8% recall). These results substantiate that the proposed methodology not only mitigates false positive occurrences but also significantly reduces false negative instances, thereby enhancing overall detection robustness.
Comprehensive complexity analysis demonstrates that the RTUAV-YOLO architecture achieves superior detection performance while maintaining exceptional parameter efficiency and computational optimization. Regarding computational overhead, except for RTU-AV-YOLO-N, which exhibits marginally elevated FLOP requirements compared to YOLOv11-N, all remaining RTUAV-YOLO variants demonstrate reduced computational complexity relative to their corresponding baseline architecture. Specifically, the RTUAV-YOLO-X variant comprises 22.86 M parameters with a computational complexity of 168.1 G FLOPs. Relative to the corresponding YOLOv8-X and YOLOv11-X architectures, our proposed model shows a 66.5% and 59.9% reduction in parameters and a 35.0% and 14.2% reduction in FLOPs, while simultaneously achieving marked performance enhancements across all evaluation metrics.
To provide intuitive validation of the differential detection capabilities, representative images from diverse scenarios were selected for comprehensive visual analysis. The comparative results are presented in Figure 8, with the baseline YOLOv11-S model displayed on the left and the proposed RTUAV-YOLO-S architecture on the right. These visualizations demonstrate that our proposed method achieves satisfactory detection accuracy. Notably, the proposed approach successfully identifies numerous small objects frequently undetected by conventional detection algorithms, as evidenced in the second column visualization.
Figure 9 shows heatmaps of small objects in the VisDron2019 dataset. Relative to the baseline architecture, RTUAV-YOLO exhibits substantially enhanced capabilities for localizing densely clustered small objects. In our model’s heatmaps, small objects display higher heat values, indicating that our model is more effective in capturing the characteristics of these small objects. Furthermore, empirical observations reveal that RTUAV-YOLO exhibits enhanced attention to contextual information surrounding small targets, reflecting superior contextual feature utilization during the detection process.
To provide comprehensive performance validation, extensive comparative analysis was conducted on the VisDrone2019 dataset, benchmarking the RTUAV-YOLO model family against state-of-the-art aerial object detection architectures. These comparative models primarily include PS-YOLO [44], MSUD-YOLO [45], LGFF-YOLO [46], FBRT-YOLO [47], RT-DETR [48], CSFPR-RTDETR [49] and UAV-DETR [50] model. The comparison results are shown in Table 2. “-” indicates that the model was not proposed in the original literature.
Regarding detection accuracy, the RTUAV-YOLO architectural family demonstrates substantial performance enhancements across all model scales. RTUAV-YOLO-X attained superior performance among all evaluated architectures, achieving a mAP50 of 51.4%, representing a 3.0 percentage point improvement over FBRT-YOLO-X (48.4%). Under the more rigorous mAP50-95 evaluation criterion, RTUAV-YOLO-X maintained superiority over all competing architectures with a score of 32.4%. Notably, the lightweight RTUAV-YOLO-N variant outperforms comparable PS-YOLO-N and FBRT-YOLO-N architectures with a mAP50 of 35.9%, exemplifying exceptional performance in compact model configurations.

4.4.2. Experimental Results on UAVDT and UAVVaste Datasets

To comprehensively assess the generalization performance of the proposed RTUAV-YOLO architecture for small object detection, extensive comparative evaluations were conducted across two distinct aerial imaging datasets: UAVDT and UAVVaste. These datasets exhibit substantial variations from VisDrone2019 regarding scene complexity, object scale distributions, and acquisition conditions, thereby providing rigorous validation of the model’s cross-domain adaptability and robustness. To maintain experimental consistency and ensure comparative validity, identical training protocols and hyper-parameter configurations were employed as established in the VisDrone2019 evaluation framework. Evaluation metrics included key performance indicators such as mAP50, mAP75, and mAP50-95. Across both datasets, the RTUAV-YOLO-X variant was employed for evaluation. Comparative performance results are presented in Table 3 and Table 4.
Evaluation on the UAVDT dataset demonstrates that RTUAV-YOLO-X attains superior performance across all assessment criteria. Specifically, RTUAV-YOLO-X attains 31.8% mAP50, representing a 0.7 percentage point enhancement over the second-best performing method, FBRT-YOLO (31.1%). Notably, under the more rigorous mAP75 criterion, which imposes stringent localization accuracy requirements, RTUAV-YOLO-X achieves 19.3%, substantially surpassing all comparative baselines and conclusively validating the proposed methodology’s efficacy. Under the comprehensive mAP50-95 evaluation framework, RTUAV-YOLO-X achieves 18.8%, representing a substantial 3.0 percentage point advancement over the YOLOv11-X baseline (15.8%), thereby establishing a marked performance superiority. Evaluation on the UAVVaste dataset provides additional validation of RTUAV-YOLO’s architectural superiority. RTUAV-YOLO-X attained 76.8% mAP50, demonstrating a 3.3 percentage point enhancement relative to RT-DETR-R50 (73.5%). This substantial performance gain substantiates the proposed methodology’s robustness across diverse aerial imaging scenarios. Under the comprehensive mAP50:95 assessment, RTUAV-YOLO-X achieved 38.2%, representing a remarkable 10.4% advancement over the YOLOv11-S baseline (27.8%). This substantial performance improvement fully demonstrates the technical advantages of the proposed method.
Comparative analysis across three distinct datasets reveals that RTUAV-YOLO exhibits exceptional cross-domain generalization capabilities. Primarily, RTUAV-YOLO maintains consistent performance superiority across all datasets, substantiating the architectural versatility of the proposed framework.
Furthermore, performance gains relative to baseline methods exhibit positive correlation with dataset complexity, thereby corroborating the methodology’s robustness in addressing challenging small object detection scenarios. Additionally, consistent performance across varying IoU thresholds demonstrates that RTUAV-YOLO not only achieves effective small object detection but also delivers enhanced localization precision, thereby establishing significant practical value for real-world aerial applications.

4.5. Ablation Experiments

To systematically validate the efficacy of individual architectural components within RTUAV-YOLO, comprehensive ablation studies were conducted on the VisDrone2019 dataset. These evaluations utilized YOLOv11-S as the baseline architecture and implemented a progressive module integration methodology. Initially, backbone network modifications were implemented, subsequently followed by the systematic integration of three novel modules: LDSM, MSFAM, and PDSCM. Subsequently, the bounding box regression loss function was substituted with the MPDWIoU formulation. Excluding the ablated components, training protocols and hyperparameter configurations remained consistent throughout the evaluation. Performance assessment employed mAP50 and mAP50-95 metrics, with parameter count and FLOPs serving as computational efficiency indicators. To verify the statistical significance of the ablation experiment results, we conducted multiple independent experiments on different combinations of each module (each experiment was run 3 times). The data in Table 5 are the average values of multiple independent experiments.
Empirical findings from our ablation studies show that architectural modifications to the backbone network significantly improve the detection of small objects by eliminating one downsampling operation, which preserves high-resolution feature maps. This adjustment also reduces the parameter count while only marginally increasing computational complexity. When progressively integrating modules, the LDSM contributes to the improved extraction of small target features by reducing computational complexity, thereby improving detection accuracy while maintaining efficiency. The MSFAM module further enhances feature extraction by fusing multi-scale features and dynamically adjusting feature importance. This leads to significant improvements in the detection of small objects across different scales.
The PDSCM module expands the spatial contextual relationships between small objects at multiple scales, further enhancing the model’s multi-scale representation capability and improving overall detection performance. When substituting the conventional IoU loss with MPDWIoU, we observe substantial improvements in bounding box regression, particularly for high-IoU regimes, without introducing additional parameters.
By analyzing the combined effects of each module, we found that while the introduction of a single module can bring a certain degree of performance improvement, it is the synergy of the modules that truly enables RTUAV-YOLO to achieve optimal performance. The results in Table 5 clearly demonstrate the modularity and synergistic complementarity of the various components. The independent contributions of each module are combined to significantly improve detection accuracy while maintaining the model’s computational efficiency. Our approach achieves an optimal balance between performance and efficiency, making it well-suited for resource-constrained drone deployment scenarios. The consistency of these results, supported by statistical analysis, further validates the robustness and reliability of the proposed method across various configurations.
Furthermore, to further validate the effectiveness of the proposed MPDWIoU loss, we conducted a comparative experiment with several widely used IoU-based loss functions, including GIoU, DIoU, CIoU, and WIoUv3. All models were trained under identical settings and datasets (VisDrone2019) using the RTUAV-YOLO-S configuration. The experimental results presented in Table 6 demonstrate that MPDWIoU achieves the highest detection accuracy on small targets, improving mAP50 by 0.6% and mAP50-95 by 0.4% compared to the GIoU loss. This improvement verifies that the proposed MPDWIoU effectively enhances localization precision and gradient stability for small objects through the introduction of minimum point distance constraints and adaptive weighting mechanisms.

4.6. Edge Computing Platform Deployment

To validate the deployment feasibility of the proposed RTUAV-YOLO architecture on edge computing platforms, comprehensive evaluation was conducted using the NVIDIA Jetson Orin Nano development board (NVIDIA, Santa Clara, CA, USA). Despite its compact form factor, the Jetson Orin Nano integrates 32 Tensor Cores and 1024 CUDA Cores, delivering substantial computational performance with optimized power efficiency. This configuration enables efficient computational processing for real-time object detection applications in UAV edge computing scenarios. The actual deployment and detection diagram of RTUAV-YOLO-S on NVIDIA Jetson Orin Nano is shown in Figure 10. The RTUAV-YOLO-S variant, when deployed on this edge computing platform, achieved a real-time inference rate of 37.8 FPS, satisfying real-time processing requirements while demonstrating an optimal balance between detection precision and computational efficiency.
To better evaluate the practical application of RTUAV-YOLO on a drone edge platform, we deployed several different real-time drone detection models on the device and quantitatively evaluated their detection accuracy and speed. The results are shown in Table 7. Compared with various advanced models, our proposed RTUAV-YOLO can better capture target features in complex scenes and achieve higher detection accuracy. Furthermore, its detection frame rate is better than that of most real-time drone models, meeting real-time detection requirements and being suitable for most practical drone application scenarios.

5. Discussion

5.1. Resource Suitability Under UAV Constraints

UAV onboard deployment is constrained by memory capacity, energy budget, and real-time requirements. To make these constraints explicit without relying on device-specific power probes, we use three hardware-agnostic proxies that are standard in efficient vision: parameter count (Params) to approximate the lower bound of model memory devoted to weights, floating-point operations (GFLOPs) to reflect per-image computational work and thus serve as an energy proxy under matched settings, and frames per second (FPS) to represent realized runtime on target hardware. We report these metrics for all comparison models and RTUAV-YOLO variants on a desktop GPU and a Jetson Orin Nano, and the results are shown in Table 1, Table 2 and Table 6.
In practical terms, fewer parameters indicate a smaller weight footprint and lighter data movement; lower GFLOPs indicate less arithmetic per image, which generally correlates with lower energy cost and latency; and higher FPS indicates stronger real-time feasibility under the same input resolution and batch size. Across our experiments, RTUAV-YOLO attains higher accuracy with markedly fewer parameters, lower or comparable GFLOPs, and higher FPS than comparable baselines on both platforms, indicating a favorable accuracy–efficiency balance for resource-constrained UAV scenarios.
We note that absolute power draw and peak activation memory depend on device configuration (e.g., precision, clocking, thermal limits) and deployment choices (e.g., batch size, resolution). The above proxies, together with on-device FPS, provide transparent and comparable evidence of suitability without over-claiming device-specific wattage. We plan to conduct rigorous memory and power analysis of RTUAV-YOLO deployed on actual drone hardware in subsequent research.

5.2. Resolution Choice and Small-Object Fidelity

Many remote-sensing scenes contain targets only a few pixels in size; when such imagery is resized to 640 × 640, fine-grained shape cues may be attenuated or lost. Our choice of a 640 × 640 input reflects the constraints of onboard UAV deployment—memory, energy, and strict real-time budgets—yet we explicitly consider the implications for small-object detection and through reasonable architectural design and appropriate data augmentation techniques, our proposed method mitigates the impact of this issue to a certain extent.
In addition to resource considerations, we adopt 640 × 640 to ensure apples-to-apples comparisons with prior art. The majority of recent YOLO-based detectors and widely used baselines for UAV/aerial benchmarks report results at this input size. Using the same resolution standardizes pre-processing, keeps mAP and runtime comparisons fair across methods, and aligns with community practice. Where relevant, we also provide results at a higher input size to quantify the accuracy–efficiency trade-off.
We quantify the fidelity–efficiency balance by comparing 640 × 640 versus 1280 × 1280 inputs on RTUAV-YOLO-S. At 640 × 640, training time per epoch was 126 s, throughput 162 FPS, with mAP50 was 42.9% and mAP50–95 was 25.9%. At 1280 × 1280, training time increased to 576 s, throughput decreased to 47 FPS, and accuracy rose modestly to mAP50 was 44.3% and mAP50-95 was 26.8%. Thus, larger inputs provide small gains in accuracy but impose substantial computational and latency costs that are incompatible with many real-time UAV missions.
Adopting 640 × 640 is a pragmatic operating point preserves essential target cues for most scenarios, matches the prevailing evaluation protocol for comparable methods, and meets real-time and resource constraints on typical UAV hardware. For missions where tiny-object fidelity dominates over latency or energy, higher input resolutions, tiling/windowing strategies, or adaptive-resolution policies are viable alternatives and constitute promising directions for future work.

5.3. Cross-Domain Generalization Capability Verification

To further evaluate the cross-domain generalization capability of the proposed RTUAV-YOLO architecture, we conducted additional experiments on the HIT-UAV [56] and AI-TOD [57] datasets, which are widely adopted benchmarks in UAV-based small object detection. Compared with the VisDrone2019, UAVDT, and UAVVaste datasets used in the main experiments, HIT-UAV and AI-TOD present markedly different imaging characteristics in terms of image resolution, viewing angle, background complexity, and detection targets. This enables a more comprehensive validation of the model’s robustness and adaptability to data distributions that differ significantly from the original training domain.
For consistency and fairness, we maintained identical training hyperparameters and evaluation metrics as described in Section 4.2. Table 8 and Table 9 report the quantitative detection results of RTUAV-YOLO and baseline YOLO series models on HIT-UAV and AI-TOD, respectively. The results show that RTUAV-YOLO maintains leading performance on the HIT-UAV and AI-TOD datasets. This demonstrates the strong generalization capabilities of our proposed model and its ability to effectively adapt to data from different distributions. In the future, we will pay more attention to the generalization ability verification of the model, which is crucial for building a general drone-view target detection model.

6. Conclusions

This study addresses the dual challenges of diminished small-target detection accuracy in UAV imagery and constrained computational resources on onboard platforms and present a comprehensive family of lightweight detection models, designated as RTUAV-YOLO. We designed three efficient lightweight modules: MSFAM, LDSM, and PDSCM. To address the loss of contextual information for small objects caused by multiple downsampling in YOLOv11, we used the PDSCM to reduce feature sampling from P4 to P5 and employed extended depth-wise separable convolutions to establish spatial contextual relationships between multiple scales, enhancing the representation of small objects at multiple scales. We introduced a P2 detection head within the detection head, enabling the model to focus more on detailed features, thereby improving sensitivity to small objects. To address the feature imbalance and information loss associated with traditional architectures when processing small objects, we replaced the C3K2 module in the backbone network with the MSFAM. This module significantly improves feature extraction for small objects through an adaptive weight generation mechanism and dual-channel heterogeneous feature aggregation. We also replaced the traditional convolutional modules in the backbone and neck with the LDSM, achieving efficient downsampling of feature maps while reducing computational complexity and preserving key small object features. To address the insensitivity of the traditional IoU metric to small objects, we designed the MPDWIoU loss function. This loss function integrates a minimum point distance metric, a dynamic anchor focus strategy, and auxiliary bounding box supervision to specifically alleviate gradient imbalance and low accuracy in small object localization, while enhancing regression robustness in cluttered environments.
Comprehensive experimental validation on the VisDrone2019 dataset confirms that RTUAV-YOLO achieves an effective balance between computational efficiency and detection performance. RTUAV-YOLO achieves an average improvement of 3.4% and 2.4% in mAP50 and mAP50-95, respectively, compared to the baseline model YOLOv11, while reducing the number of parameters by 65.3%. Furthermore, cross-dataset evaluations on the UAVDT and UAVVaste benchmarks confirm RTUAV-YOLO’s generalization and robustness across di-verse UAV aerial imagery conditions. When deployed on the Jetson Orin Nano, a common platform for UAVs, our method achieves a real-time detection frame rate of 37.8 FPS while maintaining excellent detection accuracy, providing an effective solution for real-time target detection on real-world UAVs.
Although this investigation has yielded promising outcomes, opportunities remain for further computational complexity reduction to facilitate deployment on ultra-low-power UAV platforms. Future research directions will encompass the development and implementation of advanced lightweight architectural components and optimized loss functions to further minimize computational requirements while enhancing detection accuracy and robustness in challenging scenarios involving high object density and severe occlusion conditions.

Author Contributions

Conceptualization, R.Z. and S.G.; methodology, R.Z.; software, R.Z.; validation, K.Z.; formal analysis, J.H.; investigation, L.Z.; data curation, L.L.; writing—original draft preparation, R.Z.; writing—review and editing, S.G.; visualization, R.Z.; project administration, L.Z.; funding acquisition, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request. The project page: https://gitee.com/zhangruizhi0110/RTUAV-YOLO.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, X.; He, J.; Yao, Y.; Zhang, J.; Liang, H.; Wang, H.; Hong, Y. Classifying Urban Land Use by Integrating Remote Sensing and Social Media Data. Int. J. Geogr. Inf. Sci. 2017, 31, 1675–1696. [Google Scholar] [CrossRef]
  2. Byun, S.; Shin, I.-K.; Moon, J.; Kang, J.; Choi, S.-I. Road Traffic Monitoring from UAV Images Using Deep Learning Networks. Remote Sens. 2021, 13, 4027. [Google Scholar] [CrossRef]
  3. Cao, D.; Zhang, B.; Zhang, X.; Yin, L.; Man, X. Optimization Methods on Dynamic Monitoring of Mineral Reserves for Open Pit Mine Based on UAV Oblique Photogrammetry. Measurement 2023, 207, 112364. [Google Scholar] [CrossRef]
  4. Albattah, W.; Masood, M.; Javed, A.; Nawaz, M.; Albahli, S. Custom CornerNet: A Drone-Based Improved Deep Learning Technique for Large-Scale Multiclass Pest Localization and Classification. Complex Intell. Syst. 2023, 9, 1299–1316. [Google Scholar] [CrossRef]
  5. Zhang, H.; Wang, L.; Tian, T.; Yin, J. A Review of Unmanned Aerial Vehicle Low-Altitude Remote Sensing (UAV-LARS) Use in Agricultural Monitoring in China. Remote Sens. 2021, 13, 1221. [Google Scholar] [CrossRef]
  6. Sun, G.; He, L.; Sun, Z.; Wu, Q.; Liang, S.; Li, J.; Niyato, D.; Leung, V.C.M. Joint Task Offloading and Resource Allocation in Aerial-Terrestrial UAV Networks With Edge and Fog Computing for Post-Disaster Rescue. IEEE Trans. Mob. Comput. 2024, 23, 8582–8600. [Google Scholar] [CrossRef]
  7. Teixidó, P.; Gómez-Galán, J.A.; Caballero, R.; Pérez-Grau, F.J.; Hinojo-Montero, J.M.; Muñoz-Chavero, F.; Aponte, J. Secured Perimeter with Electromagnetic Detection and Tracking with Drone Embedded and Static Cameras. Sensors 2021, 21, 7379. [Google Scholar] [CrossRef] [PubMed]
  8. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 3520–3529. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  10. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  11. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  12. Weng, S.; Wang, H.; Wang, J.; Xu, C.; Zhang, E. YOLO-SRMX: A Lightweight Model for Real-Time Object Detection on Unmanned Aerial Vehicles. Remote Sens. 2025, 17, 2313. [Google Scholar] [CrossRef]
  13. Liu, Y.; He, M.; Hui, B. ESO-DETR: An Improved Real-Time Detection Transformer Model for Enhanced Small Object Detection in UAV Imagery. Drones 2025, 9, 143. [Google Scholar] [CrossRef]
  14. Luo, X.; Zhu, X. YOLO-SMUG: An Efficient and Lightweight Infrared Object Detection Model for Unmanned Aerial Vehicles. Drones 2025, 9, 245. [Google Scholar] [CrossRef]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  16. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  17. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  18. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2016; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  19. Li, Y.; Li, Q.; Pan, J.; Zhou, Y.; Zhu, H.; Wei, H.; Liu, C. SOD-YOLO: Small-Object-Detection Algorithm Based on Improved YOLOv8 for UAV Images. Remote Sens. 2024, 16, 3057. [Google Scholar] [CrossRef]
  20. Zhou, S.; Zhou, H.; Qian, L. A Multi-Scale Small Object Detection Algorithm SMA-YOLO for UAV Remote Sensing Images. Sci. Rep. 2025, 15, 9255. [Google Scholar] [CrossRef] [PubMed]
  21. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AugFPN: Improving Multi-Scale Feature Learning for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12595–12604. [Google Scholar]
  22. Cheng, G.; Lang, C.; Wu, M.; Xie, X.; Yao, X.; Han, J. Feature Enhancement Network for Object Detection in Optical Remote Sensing Images. J. Remote Sens. 2021, 2021, 9805389. [Google Scholar] [CrossRef]
  23. Zhang, K.; Shen, H. Multi-Stage Feature Enhancement Pyramid Network for Detecting Objects in Optical Remote Sensing Images. Remote Sens. 2022, 14, 579. [Google Scholar] [CrossRef]
  24. Li, H.; Li, Y.; Xiao, L.; Zhang, Y.; Cao, L.; Wu, D. RLRD-YOLO: An Improved YOLOv8 Algorithm for Small Object Detection from an Unmanned Aerial Vehicle (UAV) Perspective. Drones 2025, 9, 293. [Google Scholar] [CrossRef]
  25. Chang, J.; Lu, Y.; Xue, P.; Xu, Y.; Wei, Z. Automatic Channel Pruning via Clustering and Swarm Intelligence Optimization for CNN. Appl. Intell. 2022, 52, 17751–17771. [Google Scholar] [CrossRef]
  26. Guo, S.; Wang, Y.; Li, Q.; Yan, J. DMCP: Differentiable Markov Channel Pruning for Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1539–1547. [Google Scholar]
  27. He, Y.; Zhang, X.; Sun, J. Channel Pruning for Accelerating Very Deep Neural Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1389–1397. [Google Scholar]
  28. Liu, Z.; Li, J.; Shen, Z.; Huang, G.; Yan, S.; Zhang, C. Learning Efficient Convolutional Networks Through Network Slimming. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2736–2744. [Google Scholar]
  29. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  30. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  31. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More Features From Cheap Operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
  32. Fan, Q.; Li, Y.; Deveci, M.; Zhong, K.; Kadry, S. LUD-YOLO: A Novel Lightweight Object Detection Network for Unmanned Aerial Vehicle. Inf. Sci. 2025, 686, 121366. [Google Scholar] [CrossRef]
  33. Cao, J.; Bao, W.; Shang, H.; Yuan, M.; Cheng, Q. GCL-YOLO: A GhostConv-Based Lightweight YOLO Network for UAV Small Object Detection. Remote Sens. 2023, 15, 4932. [Google Scholar] [CrossRef]
  34. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 658–666. [Google Scholar]
  35. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar] [CrossRef]
  36. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation. arXiv 2020, arXiv:2005.03572. [Google Scholar] [CrossRef]
  37. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  38. Zhang, Y.-F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and Efficient IOU Loss for Accurate Bounding Box Regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  39. Lu, S.; Lu, H.; Dong, J.; Wu, S. Object Detection for UAV Aerial Scenarios Based on Vectorized IOU. Sensors 2023, 23, 3061. [Google Scholar] [CrossRef] [PubMed]
  40. Du, D.; Zhu, P.; Wen, L.; Bian, X.; Lin, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-DET2019: The Vision Meets Drone Object Detection in Image Challenge Results. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  41. Du, D.; Qi, Y.; Yu, H.; Yang, Y.; Duan, K.; Li, G.; Zhang, W.; Huang, Q.; Tian, Q. The Unmanned Aerial Vehicle Benchmark: Object Detection and Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 370–386. [Google Scholar]
  42. Kraft, M.; Piechocki, M.; Ptak, B.; Walas, K. Autonomous, Onboard Vision-Based Trash and Litter Detection in Low Altitude Aerial Images Collected by an Unmanned Aerial Vehicle. Remote Sens. 2021, 13, 965. [Google Scholar] [CrossRef]
  43. Jocher, G.; Chaurasia, A.; Qiu, J. YOLOv8 by Ultralytics. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 20 August 2025).
  44. Zhong, H.; Zhang, Y.; Shi, Z.; Zhang, Y.; Zhao, L. PS-YOLO: A Lighter and Faster Network for UAV Object Detection. Remote Sens. 2025, 17, 1641. [Google Scholar] [CrossRef]
  45. Zhao, X.; Zhang, H.; Zhang, W.; Ma, J.; Li, C.; Ding, Y.; Zhang, Z. MSUD-YOLO: A Novel Multiscale Small Object Detection Model for UAV Aerial Images. Drones 2025, 9, 429. [Google Scholar] [CrossRef]
  46. Peng, H.; Xie, H.; Liu, H.; Guan, X. LGFF-YOLO: Small Object Detection Method of UAV Images Based on Efficient Local–Global Feature Fusion. J. Real-Time Image Proc. 2024, 21, 167. [Google Scholar] [CrossRef]
  47. Xiao, Y.; Xu, T.; Xin, Y.; Li, J. FBRT-YOLO: Faster and Better for Real-Time Aerial Image Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 8673–8681. [Google Scholar]
  48. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-Time Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 16965–16974. [Google Scholar]
  49. Hu, L.; Yuan, J.; Cheng, B.; Xu, Q. CSFPR-RTDETR: Real-Time Small Object Detection Network for UAV Images Based on Cross-Spatial-Frequency Domain and Position Relation. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5638219. [Google Scholar] [CrossRef]
  50. Zhang, H.; Liu, K.; Gan, Z.; Zhu, G.-N. UAV-DETR: Efficient End-to-End Object Detection for Unmanned Aerial Vehicle Imagery. arXiv 2025, arXiv:2501.01855. [Google Scholar]
  51. Yang, F.; Fan, H.; Chu, P.; Blasch, E.; Ling, H. Clustered Object Detection in Aerial Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8311–8320. [Google Scholar]
  52. Deng, S.; Li, S.; Xie, K.; Song, W.; Liao, X.; Hao, A.; Qin, H. A Global-Local Self-Adaptive Network for Drone-View Object Detection. IEEE Trans. Image Process. 2021, 30, 1556–1569. [Google Scholar] [CrossRef] [PubMed]
  53. Li, X.; Wang, W.; Wu, L.; Chen, S.; Hu, X.; Li, J.; Tang, J.; Yang, J. Generalized Focal Loss: Learning Qualified and Distributed Bounding Boxes for Dense Object Detection. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 21002–21012. [Google Scholar]
  54. Du, B.; Huang, Y.; Chen, J.; Huang, D. Adaptive Sparse Convolutional Networks With Global Context Enhancement for Faster Object Detection on Drone Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 13435–13444. [Google Scholar]
  55. Tang, S.; Zhang, S.; Fang, Y. HIC-YOLOv5: Improved YOLOv5 For Small Object Detection. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–16 May 2024; pp. 6614–6619. [Google Scholar]
  56. Suo, J.; Wang, T.; Zhang, X.; Chen, H.; Zhou, W.; Shi, W. HIT-UAV: A High-Altitude Infrared Thermal Dataset for Unmanned Aerial Vehicle-Based Object Detection. Sci. Data 2023, 10, 227. [Google Scholar] [CrossRef]
  57. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.-S. Tiny Object Detection in Aerial Images. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 3791–3798. [Google Scholar]
Figure 1. Comparison of RTUAV-YOLO with other models on the VisDrone2019 dataset. The circle radius represents GFlops.
Figure 1. Comparison of RTUAV-YOLO with other models on the VisDrone2019 dataset. The circle radius represents GFlops.
Sensors 25 06573 g001
Figure 2. Overall structure of the RTUAV-YOLO.
Figure 2. Overall structure of the RTUAV-YOLO.
Sensors 25 06573 g002
Figure 3. The specific architecture of MSFAM.
Figure 3. The specific architecture of MSFAM.
Sensors 25 06573 g003
Figure 4. The specific architecture of PDSCM.
Figure 4. The specific architecture of PDSCM.
Sensors 25 06573 g004
Figure 5. The specific architecture of LDSM.
Figure 5. The specific architecture of LDSM.
Sensors 25 06573 g005
Figure 6. Object type and size distribution of the VisDrone2019 dataset.
Figure 6. Object type and size distribution of the VisDrone2019 dataset.
Sensors 25 06573 g006
Figure 7. RTUAV-YOLO-X training results at VisDrone2019 dataset.
Figure 7. RTUAV-YOLO-X training results at VisDrone2019 dataset.
Sensors 25 06573 g007
Figure 8. Visual comparison of detection results. The left side shows the results of YOLOv11-S, the right side shows the results of RTUAV-YOLO-S, and the center shows a detailed comparison of detection details.
Figure 8. Visual comparison of detection results. The left side shows the results of YOLOv11-S, the right side shows the results of RTUAV-YOLO-S, and the center shows a detailed comparison of detection details.
Sensors 25 06573 g008
Figure 9. Visual heatmap analysis of YOLOv11-S and RTUAV-YOLO-S.
Figure 9. Visual heatmap analysis of YOLOv11-S and RTUAV-YOLO-S.
Sensors 25 06573 g009
Figure 10. Actual deployment diagram of RTUAV-YOLO on Jetson Orin Nano.
Figure 10. Actual deployment diagram of RTUAV-YOLO on Jetson Orin Nano.
Sensors 25 06573 g010
Table 1. Experimental comparison with YOLO series models on VisDrone2019.
Table 1. Experimental comparison with YOLO series models on VisDrone2019.
ModelP (%)R (%)mAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)FPS
YOLOv8-N42.533.933.319.63.168.9184
YOLOv8-S47.937.839.123.411.1728.8143
YOLOv8-M53.743.244.427.125.9079.389
YOLOv8-L56.744.145.928.443.69165.766
YOLOv8-X57.544.646.828.968.23258.548
YOLOv11-N42.333.633.119.32.626.6192
YOLOv11-S48.238.139.423.69.4621.7160
YOLOv11-M53.443.044.126.920.1168.597
YOLOv11-L56.944.346.028.525.3787.675
YOLOv11-X57.744.846.929.056.97196.056
RTUAV-YOLO-N44.735.635.921.30.798.3187
RTUAV-YOLO-S52.541.142.925.92.4921.4162
RTUAV-YOLO-M57.544.746.828.97.2356.1109
RTUAV-YOLO-L60.246.049.230.510.4379.188
RTUAV-YOLO-X61.348.151.432.422.86168.165
Table 2. Experimental comparison with SOTA UAV object detection models on VisDrone2019.
Table 2. Experimental comparison with SOTA UAV object detection models on VisDrone2019.
ModelmAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)FPS
PS-YOLO-N33.219.31.395.1213
PS-YOLO-S40.724.25.5320.0178
PS-YOLO-M46.028.314.8273.591
MUSD-YOLO43.4-6.77-82
LGFF-YOLO43.522.94.1512.4181
RT-DETR-R1844.626.720.0360.0101
RT-DETR-R3446.027.231.0492.078
RT-DETR-R5048.328.842.02136.069
FBRT-YOLO-N34.420.20.946.7192
FBRT-YOLO-S42.425.92.9222.9143
FBRT-YOLO-M45.928.47.2658.794
FBRT-YOLO-L47.729.714.63119.270
FBRT-YOLO-X48.430.122.89185.852
CSFPR-RTDETR42.324.914.0963.998
UAV-DETR-EV247.528.713.0243.1101
UAV-DETR-R1848.829.820.1970.291
UAV-DETR-R5051.131.542.07170.661
RTUAV-YOLO-N35.921.30.798.3187
RTUAV-YOLO-S42.925.92.4921.4162
RTUAV-YOLO-M46.828.97.2356.1109
RTUAV-YOLO-L49.230.510.4379.188
RTUAV-YOLO-X51.432.422.86168.165
Table 3. Experimental results on the UAVDT dataset.
Table 3. Experimental results on the UAVDT dataset.
ModelmAP50 (%)mAP75 (%)mAP50-95 (%)
YOLOv11-X [11]28.716.815.8
ClusDet [51]26.512.513.7
GLSAN [52]28.118.817.0
GFL [53]29.517.916.9
CEASC [54]30.917.817.1
FBRT-YOLO [47]31.118.918.4
RTUAV-YOLO-X31.819.318.8
Table 4. Experimental results on the UAVVaste dataset.
Table 4. Experimental results on the UAVVaste dataset.
ModelmAP50 (%)mAP75 (%)mAP50-95 (%)
YOLOv11-S [11]63.041.227.8
HIC-YOLOv5 [55]65.143.130.5
RT-DETR-R18 [48]72.645.736.3
RT-DETR-R50 [48]73.547.137.4
RTUAV-YOLO-X76.849.738.2
Table 5. Ablation experiment results. × indicates that the method was not used, and √ indicates that the method was used.
Table 5. Ablation experiment results. × indicates that the method was not used, and √ indicates that the method was used.
BackboneLDSMMSFAMPDSCMMPDWIoUmAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)
×××××39.423.69.4621.7
××××39.923.93.8727.0
×××40.124.03.2223.7
×××40.524.63.6225.7
×××40.424.33.5625.4
×××40.324.23.8727.0
××41.424.83.0124.2
××41.624.93.1223.9
××41.825.12.9422.9
×42.425.52.4921.4
42.925.92.4921.4
Table 6. Comparison of different IoU-based loss functions using RTUAV-YOLO-S.
Table 6. Comparison of different IoU-based loss functions using RTUAV-YOLO-S.
Loss FunctionmAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)
GIoU [34]42.325.52.4921.4
DIoU [35]42.225.22.4921.4
CIoU [36]42.425.52.4921.4
WIoUv3 [37]42.525.62.4921.4
MPDWIoU (Ours)42.925.92.4921.4
Table 7. Quantitative evaluation results on Jetson Orin Nano.
Table 7. Quantitative evaluation results on Jetson Orin Nano.
ModelmAP50 (%)mAP50-95 (%)Params (M)FPS
YOLOv8-N33.319.63.1635.2
YOLOv8-S39.123.411.1720.1
YOLOv11-N33.119.32.6245.1
YOLOv11-S39.423.69.4623.3
PS-YOLO-N33.219.31.3941.1
PS-YOLO-S40.724.25.5326.9
FBRT-YOLO-N34.420.20.9452.6
FBRT-YOLO-S42.425.92.9234.1
UAV-DETR-EV247.528.713.0219.4
RTUAV-YOLO-N35.921.30.7953.7
RTUAV-YOLO-S42.925.92.4937.8
Table 8. Comparative experimental results of the baseline model and RTUAV-YOLO on the HIT-UAV dataset.
Table 8. Comparative experimental results of the baseline model and RTUAV-YOLO on the HIT-UAV dataset.
ModelmAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)
YOLOv8-N71.443.23.168.9
YOLOv8-S80.149.311.1728.8
YOLOv8-M86.254.225.9079.3
YOLOv11-N71.743.22.626.6
YOLOv11-S80.649.59.4621.7
YOLOv11-M87.154.820.1168.5
RTUAV-YOLO-N76.943.20.798.3
RTUAV-YOLO-S82.349.32.4921.4
RTUAV-YOLO-M89.257.17.2356.1
Table 9. Comparative experimental results of the baseline model and RTUAV-YOLO on the AI-TOD dataset.
Table 9. Comparative experimental results of the baseline model and RTUAV-YOLO on the AI-TOD dataset.
ModelmAP50 (%)mAP50-95 (%)Params (M)FLOPs (G)
YOLOv8-N39.217.53.168.9
YOLOv8-S43.619.111.1728.8
YOLOv8-M46.122.325.9079.3
YOLOv11-N39.617.92.626.6
YOLOv11-S43.919.29.4621.7
YOLOv11-M46.722.820.1168.5
RTUAV-YOLO-N40.118.60.798.3
RTUAV-YOLO-S46.120.82.4921.4
RTUAV-YOLO-M49.225.37.2356.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, R.; Hou, J.; Li, L.; Zhang, K.; Zhao, L.; Gao, S. RTUAV-YOLO: A Family of Efficient and Lightweight Models for Real-Time Object Detection in UAV Aerial Imagery. Sensors 2025, 25, 6573. https://doi.org/10.3390/s25216573

AMA Style

Zhang R, Hou J, Li L, Zhang K, Zhao L, Gao S. RTUAV-YOLO: A Family of Efficient and Lightweight Models for Real-Time Object Detection in UAV Aerial Imagery. Sensors. 2025; 25(21):6573. https://doi.org/10.3390/s25216573

Chicago/Turabian Style

Zhang, Ruizhi, Jinghua Hou, Le Li, Ke Zhang, Li Zhao, and Shuo Gao. 2025. "RTUAV-YOLO: A Family of Efficient and Lightweight Models for Real-Time Object Detection in UAV Aerial Imagery" Sensors 25, no. 21: 6573. https://doi.org/10.3390/s25216573

APA Style

Zhang, R., Hou, J., Li, L., Zhang, K., Zhao, L., & Gao, S. (2025). RTUAV-YOLO: A Family of Efficient and Lightweight Models for Real-Time Object Detection in UAV Aerial Imagery. Sensors, 25(21), 6573. https://doi.org/10.3390/s25216573

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop