You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Review
  • Open Access

12 December 2025

Recent Real-Time Aerial Object Detection Approaches, Performance, Optimization, and Efficient Design Trends for Onboard Performance: A Survey

,
and
1
Institute for Intelligent Systems Research and Innovation, Deakin University, Waurn Ponds, VIC 3216, Australia
2
The Department of Mechatronics Engineering, School of Engineering, The University of Jordan, Amman 11942, Jordan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Image Processing and Analysis in Sensor-Based Object Detection

Abstract

The rising demand for real-time perception in aerial platforms has intensified the need for lightweight, hardware-efficient object detectors capable of reliable onboard operation. This survey provides a focused examination of real-time aerial object detection, emphasizing algorithms designed for edge devices and UAV onboard processors, where computation, memory, and power resources are severely constrained. We first review the major aerial and remote-sensing datasets and analyze the unique challenges they introduce, such as small objects, fine-grained variation, multiscale variation, and complex backgrounds, which directly shape detector design. Recent studies addressing these challenges are then grouped, covering advances in lightweight backbones, fine-grained feature representation, multi-scale fusion, and optimized Transformer modules adapted for embedded environments. The review further highlights hardware-aware optimization techniques, including quantization, pruning, and TensorRT acceleration, as well as emerging trends in automated NAS tailored to UAV constraints. We discuss the adaptation of large pretrained models, such as CLIP-based embeddings and compressed Transformers, to meet onboard real-time requirements. By unifying architectural strategies, model compression, and deployment-level optimization, this survey offers a comprehensive perspective on designing next-generation detectors that achieve both high accuracy and true real-time performance in aerial applications.

1. Introduction

A growing trend in the advancement of autonomous Unmanned Aerial Vehicles (UAVs) is their use for real-time mapping, alongside the implementation of deep learning techniques for the semantic analysis of data collected by UAVs [1]. Developing autonomous Unmanned Aerial Systems (UASs) primarily depends on their detection ability of static and moving objects [2], which makes UAV object detection essential during navigation for the aid of autonomous ground vehicle path planning [3,4]. UAVs have a broad range of applications that rely on object detection beyond navigation. They are employed in security and surveillance, including border patrol and the efficient monitoring of significant areas. UAVs provide valuable visual data to help responders plan rescues and locate survivors [3]. In the industry, UAVs play a vital role in inspecting infrastructure and monitoring environmental health. For example, UAVs are used for asphalt pavement cracking [5], wildlife conservation [6], and forest health checking (detecting dead trees or the incidence of disease) [7]. Similarly, UAVs are useful for mining exploration because they can provide relevant information on geological features, mineral deposits, and surface topography. In the same sense, UAVs are valuable tools in the construction industry. They help construction sites monitor safety regulations and compliance, as well as track construction progress.
RGB images are extensively used in aerial detection due to their rich colour and texture features. However, aerial detection poses various challenges. The distant areas of UAV aerial images frequently include numerous densely clustered small objects with few pixels of representation, making them difficult for standard detection algorithms to identify [8]. As a result, Information can be lost during the convolution process in feature extraction [9,10]. Previous studies have proposed several approaches to enhance the detection of small objects. One common strategy is replication augmentation, in which small objects are copied and randomly placed at different locations within the image, thereby increasing the number of small-object samples available for training [11]. Another approach involves scaling and splicing, in which larger objects in the original image are resized to generate additional small object instances. In addition, some methods focus on loss function adjustment, giving higher weight to the losses associated with small objects to encourage the model to pay more attention to them during training [11]. And some recent studies add an extra detection head to shallower layers for small object detection. However, improving the detection of small objects generally comes with an increase in model complexity and computational requirements, making it increasingly difficult to satisfy the requirements for real-time applications on memory- and processing-constrained platforms [12].
Fine-grained discrimination remains a major challenge in aerial detection, where visually similar classes often appear with only a few pixels and minimal texture. Although attention mechanisms, transformer-based feature alignment, and contrastive learning can significantly enhance fine-grained representations, feature map is another solution, in addition to spatial attention; these strategies typically increase computational burden and require more complex pipelines, limiting their suitability for real-time onboard deployment [13]. In low resolution, Spatially-adaptive Convolution SPDConv combined with Efficient Channel Attention ECA is used for fine-grained downsampling, and swin transformers for long-range dependancies [14]. This offers real-time performance but not on edge device.
High-resolution images facilitate the detection of small objects and the extraction of fine-grained features. However, increasing the resolution increases complexity. High-resolution feature extraction networks [15] for small object detection are complex; however, some emerging solutions include adding a super-resolution module that learns high-resolution features from low-resolution input, and this module is applied only during training to enhance detection while removed in inference for faster detection [16]. In [17] Fine-grained Information extraction module based on SPDConv with a multiscale feature fusion module based on BiFPN with skip connections to enhance fine-grained feature representation.
Another challenge is the scale and perspective differences: Objects may appear at different sizes and orientations depending on their distance from the camera and the angle of view. This challenge is addressed by using strategies such as Feature Pyramid Networks (FPNs), which have been widely applied to new detection tasks. These approaches advocate improved detection of small objects by leveraging features from multiple model layers. Nevertheless, they involve considerable computational requirements, potentially affecting real-time performance, especially on devices with limited resources [18].
Moreover, the orientation of objects presents another challenge [19]. Earlier datasets feature horizontal boundary boxes that are suitable for ground platforms. This can lead to inaccuracies in aerial detection, especially in detecting dense scenes. Addressing this issue depends on the algorithm; many studies adopt single-stage anchor-based algorithms with Horizontal Boundary Box (HBB) detections. Additionally, some studies employ Oriented Boundary Box (OBB) detections to handle objects at various angles. OBB detection can start with predefined anchors of different sizes and angles, or use horizontal anchors. After the initial detection, the model applies additional algorithms or layers to refine these anchors. This refinement process involves predicting offsets or adjustments to the position, size, and orientation of the anchors [20]. In anchor-free methods, the orientation is solved by different strategies, including regressing width, height, and angle at each center point using anchor-free feature maps [21]. Gaussian centerness and pseudo-domain angle encoding are used for OBB regression in [22], while ref. [23] uses channel expansion and dynamic label assignment is implemented to regress rotated boxes efficiently for real-time UAV imagery. These solutions increase the complexity of the models, affecting real-time inference.
In aerial remote sensing, a complex environment refers to a variety of scene conditions that make accurate object detection more difficult. Such conditions include background clutter, where targets are easily confused with surrounding structures, partial occlusions, in which objects are obscured by other elements in the scene, and changes in illumination, including shadows, bright sunlight, and low-light scenarios that alter visual appearance. Complex environments often contain unusual or previously unseen objects, which can significantly degrade object detection performance [19,24]. Context-Aware Compositional Networks address this challenge by modeling the relationships between object parts and their surrounding context, supplemented by additional contextual enhancements [25]. A solution to this includes self-attention in transformer models and dilated and cross-convolutions, which expand the receptive field but do not offer a lightweight solution for the process. Although methods such as attention mechanisms, Transformer-based feature alignment, and contrastive learning improve the representation of fine-grained local features, they are typically computationally demanding and rely on more complex processing pipelines [26].
Additionally, class imbalance in training datasets poses a problem. Some object classes are overrepresented, while others have limited samples, leading to models that perform well on common classes but poorly on rare ones. To mitigate this in aerial object detection, researchers adopt a variety of strategies, spanning data augmentation, loss design, feature representation, and learning paradigms. At the data level, techniques like K-means SMOTE generate synthetic minority class samples, while augmentation methods such as object copy paste and aggregated-mosaic increase the occurrence and contextual variation of rare objects [27]. For the algorithm loss, reweighting mechanisms, including Class Balanced Loss, based on the effective number of samples [28], focus training on difficult examples. For feature-level improvements, modules like Discriminative Feature Learning and Imbalanced Feature Semantic Enrichment [29], promote both semantic richness and better separation of rare classes. Finally, improved training protocols, such as duplicating rare-class objects (Auto Target Duplication) and ensuring that every augmentation patch contains at least one object (Assigned Stitch), help ensure that minority categories are sampled more effectively, thereby helping models learn from rare instances more reliably [30]. In [31], a hard chip mining strategy is designed to balance class distribution and produce challenging training examples. The approach begins by creating multi-scale image chips for training the detector. In parallel, object patches are extracted from the dataset to form an object pool, which is then used for data augmentation to alleviate class imbalance. The detector trained on this augmented data is subsequently run on the modified images to identify misclassified regions, from which hard chips are generated. The final detector is trained using a combination of regular chips and these hard examples. Altogether, these combined methods reduce sample scarcity, sharpen feature discrimination, and strengthen detection performance on long-tailed aerial datasets. Addressing these challenges is essential to improving the reliability and accuracy of object detection systems across diverse and complex settings.
Ultimately, real-time processing requirements for applications such as surveillance demand fast yet accurate object detection models, a difficult combination that researchers must balance to achieve both accuracy and speed [32]. Deep learning models for detecting small objects at low altitudes can be quite complex and computationally intensive. This complexity can be challenging to implement with edge computing systems that have limited processing and memory resources available [33].
For applications that demand faster detection, such as military operations and search-and-rescue missions, there is a need for onboard processing directly on the UAV. They make extensive use of object detection for secure navigation, to identify both dynamic and stationary objects, and for real-time decision making [3,4]. This is particularly relevant for applications such as path planning for unmanned autonomous ground vehicles, where UAVs assist with mapping and obstacle avoidance This enables real-time decision making without reliance on external networks, reducing communication delays and ensuring secure of information and rapid response in time sensitive situations.
Achieving real-time object detection processing on UAV platforms remains a significant challenge, particularly due to the limited computational resources. While improvements in GPU computing power have made object detection more accessible, optimizing performance while maintaining accuracy and efficiency remains an ongoing area of research. In object detection, Real-time processing initially relied on cloud-based computing, in which data is transmitted to remote servers for analysis. However, the need for lower latency and faster decision-making in various applications has led to the development of edge computing, making processing closer to the platform itself. In many applications, a hybrid approach, combining cloud and edge computing, is used, where edge processing is prioritized for speed-critical tasks, while cloud resources handle more complex computations and long-term data storage [34].
Real-time performance on UAVs with limited computational resources requires specialized algorithmic and architectural optimizations. Many approaches focus on designing efficient lightweight detection models to reduce computational overhead [33,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52]. In addition to architectural improvements, various optimization techniques are applied to enhance efficiency further. One of these approaches is quantization, which reduces numerical precision, lowering computational complexity and memory usage, making the model more suitable for edge devices [21,45,51,53,54,55,56,57]. Also, Pruning is another optimization method that eliminates low-importance weights, reducing model size while maintaining performance [5,45,58,59,60,61,62]. In addition to knowledge distillation, which allows a smaller model to learn from a larger, more complex model, retaining essential knowledge while significantly reducing computational demands [63,64]. These optimization approaches, whether used separately or together, play a critical role in achieving real-time object detection on UAVs.
Several recent surveys have examined object detection from different perspectives. For example, ref. [65] provides a broad 20-year overview of object detection algorithms and their evolution, while ref. [66] systematically categorizes detectors into traditional and deep learning pipelines, covering one-stage, two-stage, transformer-based, and lightweight methods. Their analysis highlights the strengths and weaknesses of classical approaches, summarizes performance gains achieved through architectural design and feature learning, and identifies gaps related to lightweight models through comparative experiments and evaluation metrics. However, although lightweight networks are mentioned, these works do not explain how such models are integrated and developed specifically for aerial object detection, as they have different challenges compared to general natural scene detectors, nor do they address real-time processing, compression techniques, edge deployment requirements, or suitable algorithms for embedded platforms.
Only a limited number of studies focus directly on aerial detection. Among them, ref. [9] summarizes the main challenges and recent advances in deep learning for aerial imagery and reviews frequently used datasets; however, it lacks a deeper discussion on how to handle these challenges effectively and does not address lightweight or real-time methods. Similarly, ref. [67] reviews techniques for small object localization in aerial images, and [68] surveys object-oriented detection, but both remain limited in scope. Real-time detection is considered in only a few works; for instance, ref. [69] focuses on FPGA-based implementations but does not cover other commonly used devices, nor does it discuss the unique challenges of aerial imagery that increase detection complexity.
The review in [39] provides a detailed analysis of real-time processing strategies, algorithm speed, and sensor usage in UAV-based object recognition. However, most of these studies primarily report overall performance metrics without examining architectural design, the evolution of each detector component, or the requirements for real-time onboard processing. They discuss lightweight network architectures used as full networks, without discussing the developed modules from these networks, which are integrated into the general detector’s design and structure to enable aerial detection and achieve real-time performance. Furthermore, they lack a detailed examination of quantization strategies and pruning methods, which are not detailed in those reviews. Deployment frameworks, lightweight design principles, and emerging trends in convolution methods, edge-transformers, efficient attention mechanisms, head design, and end-to-end lightweight pipelines suitable for real-time and edge-level deployment. And a lack of more details about emerging trends, including the emerging edge transformers, hardware-aware NAS, adapting large vision-language models or CLIP embeddings for aerial detection, or integrating NAS and compression methods with these models to improve performance, and emerging trends in this area towards real-time and edge performance for such models.
Although several surveys have examined aerial object detection, see Table 1, they exhibit important limitations that restrict their relevance to real-time UAV deployment. Existing reviews focus primarily on high-level algorithmic families and overlook the end-to-end onboard pipeline, including preprocessing, postprocessing, quantization, pruning, TensorRT acceleration, and cloud–edge latency. They also provide limited coverage, hardware-aware design, trained quantization rather than post-training quantization, and structured pruning methods. Transformer refinements are now essential for achieving fast and accurate models on edge devices. Critically, earlier surveys rarely analyze Neural Architecture Search (NAS) frameworks tailored for UAV constraints or the emerging use of CLIP-style semantic embeddings for domain adaptation, despite their growing role in improving robustness and cross-scene generalization. Dataset discussions also lack emphasis on fine-grained benchmarks and high-resolution scenarios, small-object definitions, and specific datasets for small objects. Finally, most reviews fail to provide quantitative comparisons mAP, FPS, parameter count, on real embedded devices such as Jetson Nano, Xavier NX, Orin, NPUs, or Raspberry Pi, and tarde offs with typical use of detectors.
Table 1. Summary of recent reviews and surveys on object detection.
By these gaps, this review concentrates on the architectural components of modern detectors, along with hardware-oriented considerations and adaptive strategies essential for achieving real-time aerial object detection. Our survey integrates lightweight architectural advances with the practical constraints of embedded UAV platforms, connecting efficient backbone design, small-object enhancement, and multi-scale fusion with deployment strategies such as quantization post-training and trained quantization, pruning types, re-parameterization, TensorRT optimization, and federated edge learning. By incorporating recent progress in UAV-oriented NAS, we highlight how automated architecture exploration produces models that better balance accuracy and latency under strict power budgets, which can be further integrated with other methods, such as pruning for further efficiency. Additionally, we address the emerging role of CLIP-based semantic and visual embeddings in domain adaptation, a role that prior reviews do not discuss. Figure 1 gives insight into the review structure and topics.
Figure 1. Overview of the survey’s main topics.
The survey contributions are summarized below:
  • Small object definition in aerial datasets and fine-grained datasets.
  • Specify the real-time processing approaches and analyses covering performance and the platforms used, general detectors, typical applications for each model, how to modify them to mitigate aerial detection challenges, and keep efficient, with more focus on real-time studies with limited resources.
  • Systematically review real-time aerial processing approaches, including platform-level constraints, performance analyses, and the typical adaptations required to modify general-purpose detectors for aerial challenges while maintaining efficiency on resource-limited hardware.
  • Analyze lightweight design strategies across the detection pipeline, covering backbone, neck, and attention mechanisms, and new research for the developed RT-DETR design for edge deployment.
  • Performance evaluation of recent edge research.
  • Explain additional optimization techniques, such as pruning and quantization, with details about the new method, including quantization-aware training.
  • Present emerging compression and hardware-aware optimization methods, including their integration with large vision–language models and multimodal distillation, to enable efficient deployment on UAV and edge devices.
  • We identify the key limitations in current real-time aerial object detection research and discuss open challenges, offering insights and future research directions to advance onboard, real-time UAV perception.
The remaining is organized as follows. Section 2 presents the commonly used datasets referenced in the surveyed studies, highlighting their characteristics and relevance to aerial detection tasks. Section 3 reviews real-time processing techniques and summarizes the performance of existing edge and onboard systems. Section 4 explains general object detection paradigms and general lightweight networks. Section 5 This section reviews lightweight optimization strategies for aerial object detection, covering efficient backbone and detector design, model compression techniques, and emerging hardware-aware methods. It also highlights recent trends in integrating large language models with compressed architectures to enhance aerial detection under real-time and resource-constrained conditions. Section 6 presents limitations in the current research, open challenges, and future directions. Finally, Section 7 presents the overall conclusions of this survey and summarizes the key insights derived from the reviewed literature.

2. Datasets and Recent Real-Time Research Applications

Aerial and remote-sensing datasets are collected from platforms such as UAVs, manned aircraft, and satellites. Therefore, they differ from natural-scene datasets, including COCO and Pascal VOC, in several measurable ways. UAV imaging is commonly grouped into three altitude bands: close-range eye-level flights at 0–5 m, low-to-medium operations between 5 and 120 m, and high-altitude aerial imaging conducted at heights exceeding 120 m [9]. Aerial imaging also includes remote sensing images collected from multiple sensors and platforms (e.g., Google Earth), including satellite images with multiple resolutions [70].
Aerial Images and natural scene ground-level images do not use the same definition of object scale. In the natural scene images, where objects appear at consistent scales, the term small object refers to objects measuring fewer than 32 × 32 pixels [71], medium objects 32 × 32 to 96 × 96 and large objects above 96 × 96 [72], while in aerial and UAV imagery, the height of the horizontal bounding box, referred to as the pixel size, is used as the measure of object scale. Based on this criterion, instances in the dataset are categorized into three groups: small (10–50 pixels), medium (50–300 pixels), and large (greater than 300 pixels) [70].
In aerial scenes, Fine-grained discrimination refers to the challenge of distinguishing visually similar classes that often occupy only a few pixels and exhibit minimal texture. Although several studies attempt to address this problem, they generally rely on aerial datasets not explicitly designed for fine-grained detection [14,15,17,18,19]. Among these, only ref. [50] employs a dataset explicitly created for fine-grained tasks, with the Mar-20 dataset [73] and SeaDronesee [74] serve as a benchmark curated specifically for this purpose. These datasets serve different applications, as described in Table 2, which offers a comprehensive overview of datasets commonly employed in aerial object detection, illustrating their practical applications across domains and corresponding references.
These datasets’ images are captured by different satellite and UAV platforms, which are commonly classified into three types: fixed-wing, rotary-wing, and hybrid designs. Fixed-wing drones operate at higher altitudes, cover large areas quickly, and offer long endurance due to their simple structure and efficient gliding capability, though they require runways for takeoff. Rotary-wing UAVs can hover, take off, and land vertically, and fly at low altitudes to capture detailed, high-resolution data, making them well-suited for close-range inspection and sensing tasks. Hybrid platforms combine the strengths of both systems [75].
Table 2. Summary of commonly used datasets in aerial object detection, including statistics and key attributes.
Table 2. Summary of commonly used datasets in aerial object detection, including statistics and key attributes.
DatasetDescriptionImagesInstances/ObjectsClassesSize/ResolutionAnnotation Type/NotesReferences
VisDrone [76,77]Drone-captured images and videos263 video clips with 179,264 frames and additional 10,209 static images540k10765 × 1360 to 1050 × 1400HBB; high proportion of small, occluded and truncated objects[11,12,33,34,35,36,39,41,44,45,46,47,51,52,57,61,62,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98]
DOTA [70,99]Aerial/satellite images2806 images (v1)188,282 instances (v1)15From 800 × 800 to about 4k × 4kOBB; rotated and multi-scale objects[14,38,40,82,84,100]
UAVDT [101]UAV-based vehicle detection,80k images (Video)841.5k vehicles41080 × 540HBB; low-altitude, high density, small object[79,87,88,95,102]
CARPK [103]Parking lot vehicle counting1448 images89,777 vehicles11280 × 720HBB; small-to-medium vehicles[6,91,104,105]
DroneVehicle [106]RGB-Infrared vehicle detection28,439 RGB-Infrared pairs953,087 both modalities5 Vehicles840 × 712OBB; multimodal RGB-IR[40,46,85]
NWPU VHR-10 [107]High-resolution remote sensing715 from google earth images 85 images from Vaihingen data set2.934k instances10from 533 × 597 px up to 1728 × 1028HBB; planes, ships, vehicles[46,105]
DIOR [108]Optical remote sensing images23,463 images192,472 instances20800 × 800, 0.5–3 0 mHBB; large-scale, cross-sensor variety[14,33,84]
UA-DETRACVehicle detection10 h of video (140,000 frames)1.21 million vehicles4960 × 540HBB; traffic videos[45]
UCAS-AODAerial vehicle detection1510 images2500 instances21000 × 1000HBB[84]
SODA-A [109]Small object detection in aerial images24,000 images338,000 small instances10800 × 800HBB; tiny objects[38]
SEVE [110]Small object detection17,992 pairs of images and labels-101920 × 1080HBB; Special vehicles in construction sites[110]
FSDsmall target scenes of fire and smoke7534 imagesN/AN/AN/AHBB; 3 Fire hazard scenario, 3 non-hazard scenario[42]
PVDPhotovolatic point defects (PDs) and line defects (LDs)1581 images27211N/AHBB[79]
SAR-SDShip Detection dataset1160 images24561resolutions (1 to 15 m).HBB[111]
AU-AirLow altitude traffic survellance32,823 images132,03481920 × 1080HBB[6]
Traffic-Net [112]Traffic sign detection4400 images15,000 signs1512 × 512HBB[52]
Rail-FOD23Railway infrastructure detection2000 images10,000 instances1512 × 512HBB[43]
Dead TreesForest dead tree detection3000 images15,000 instances1512 × 512HBB[7]
Mar-20Remote-sensing image dataset for fine-grained military aircraft recognition3842 images22,341 instances20800 × 800HBB; Fine grained[50]
VEDAI [113]Vehicle detection in parking lots1210 images3640 vehicles9(12.5 cm × 12.5 cm per pixel, 1024 × 1024HBB, multimodal, visible and Near infrared[114]
UAV ImageUAV-based detection10,000–50,000 images50,000+ instancesVarious1024 × 1024HBB[115]
PVEL-ADPhotovoltaic panel defect detection5000 images20,000 defects1512 × 512HBB; defect-focused[116]
RSODRemote sensing small object detection886 images5000 instances4600 × 600HBB; small objects[105]
SeaDronesSee [74]maritime search and rescue54,000400,00061280 × 960 to 5456 × 3632HBB; 91% small objects, finegrained[117]
AFOsmaritime search and rescue3647 images39,99111280 × 720 to 3840 × 2160HBB; small objects in open water[117]
  • Aerial Object Detection Applications

2.1. Surveillance and General Object Detection

One of the most prominent application areas is surveillance and general object detection from UAVs, where datasets like VisDrone [5,46,57,93] are extensively utilized due to their rich annotations, high scene variability, and suitability for real-time, low-altitude aerial scenarios. Another dataset used in surveillance is CARPK [91], which focuses on UAV-based vehicle detection and parking lot vehicle counting. Typical platforms are DJI Mavic, Phantom series (3, 3A, 3SE, 3P, 4, 4A, 4P) (DJI, Shenzhen, China), used in the VisDrone dataset and DJI M200 (DJI, Shenzhen, China) used in the DroneVehicle dataset, both of which are rotary UAVs.

2.2. Environmental Monitoring

Also, another vital domain of application is environmental monitoring. For example, outfall structure detection is conducted with tailored datasets created for certain environmental compliance cases [118]. Wildlife detection, which typically requires efficient models on UAV platforms, has datasets available, like WAID and AU-AIR [6], with annotations on aerial imagery of wildlife capturing various animal species and natural environments. Emergency response systems, such as UAV fire monitoring, use thermal and RGB imagery during and after fire events that have been annotated for real-time detection [119]. In addition to applications like dead tree detection in forestry, it enables health monitoring of vegetation through spectral analysis.

2.3. Remote Sensing and High-Resolution Geospatial Object Detection

An emerging area of research is focused on remote sensing and high-resolution geospatial object detection. For example, multi-class object detection in satellite imagery relies on datasets like DOTA, DOTA-CD and AIR-CD [100], which allow for object detection in different orientations and with multi-scale sizes, in both complex urban and rural settings. These datasets are important for accommodating scale and rotation variability prevalent in remote sensing images. The AERO system [34] likewise demonstrates an AI-enabled UAV observational platform that is trained on private aerial data as well as the VisDrone dataset. AERO seamlessly provides high-level object detection and detailed tracking of detected objects across its UAV deployments in real-world scenarios. Lastly, methods developed for ultra-high-resolution (UHR) imagery have employed frame-tiling approaches in conjunction with parallel YOLO processing [120], allowing for real-time inference without sacrificing detection accuracy. Remote sensing platforms include satellites and airborne digital monitoring systems.

2.4. Industrial and Public Safety

The applications of real-time aerial object detection are crucial for industrial and public safety monitoring and inspection across a variety of industries. For example, in the coal mining industry, the CMUOD dataset [48] is designed to replicate the visual conditions found in underground mining environments, such as low-light conditions and small spaces, when detecting objects. In urban mobility and traffic analysis, MultEYE has been developed for vehicle detection and tracking and uses a custom-built UAV-based traffic monitoring dataset to allow for real-time monitoring of crowded streets and flow analysis [120]. When considering industrial infrastructure monitoring more broadly, railway infrastructure inspection represents another key application. The use of the Rail-FOD23 dataset [43] is an example of how inspection is performed using high-resolution aerial images annotated to detect faults and foreign object debris along railway tracks to support the operational safety of trains. Likewise, in the energy space, UAV-based systems for aerial defect detection in photovoltaic (PV) panels have been used, as in [116], and can be utilized to identify surface defects, shading, and damage across large solar farms with minimal human interaction during inspection.

2.5. Search and Rescue (SAR) Operations

The use of UAV-based object detection has also been explored in search and rescue (SAR) operations. These tasks generally relate to spotting humans or objects in rough terrain, and researchers have evaluated algorithms with datasets such as DIVERSE TERRAIN, RIT QUADRANGLE, and AUVSI SUAS [121] that represent various environmental settings, such as urban, rural, and wooded. At the same time, the demand for real-time pedestrian detection in videos from UAVs has resulted in the development of lightweight detection models, which have been evaluated on a custom UAV-based pedestrian dataset [56]. In addition to marins SAR operations [117]. In SAR missions, fixed-wing UAVs are used for their long-range search [122].

3. Real-Time Processing Platforms

Numerous real-time object detection algorithms have been executed on desktop GPUs [33,38,39,40,41,42,45,46,47,78,79,81,82,83,84,86,87,89,91,105,110,123]. But for real-world applications, cloud computing, edge processing, or embedded processing are the most common. While cloud computing may provide substantial computational resources, it introduces latency when running methods such as UAV-based object detection or autonomous navigation in real time. Some studies have proposed a hybrid cloud-edge processor to optimize speed and cost effectiveness [124]. Real-time onboard performance is crucial, yet only a few studies have achieved near-real-time performance (<30 FPS), and fewer have successfully implemented real-time onboard detection at speed (≥30) FPS [125]. The standard video processing benchmark of object detection is 30 frames per second [126]. However, other studies record lower detection rates that are less than this level [47]. Despite the ability of high-performance desktop graphics cards, such as the GeForce GTX1080 Ti, GeForce RTX3090, and GeForce RTX4080, to achieve real-time performance [127], other computational systems remain difficult to match. Maintaining similar speeds across diverse computing environments, such as edge devices, remains challenging due to the computational demands of these algorithms, often resulting in severe speed degradation [49,50,90,115]. Table 3, which shows the performance of recent real-time studies on the VisDrone dataset with corresponding desktop platforms used.
Table 3. Real- time aerial studies tested on VisDrone2019/2021 with their corresponding performance and platforms used.
YOLOv8 models are evaluated across various versions, including YOLOv8n, YOLOv8s, and YOLOv5, on platforms such as NVIDIA GeForce RTX 3090 and NVIDIA GeForce RTX 4090. YOLOv8s achieves mAP50 of 42% at a speed of 126 FPS and an inference time of 7.8 ms evaluated on the VisDrone2019 dataset [123]. In contrast, YOLOv5 models (e.g., YOLOv5m and YOLOv5l) on NVIDIA GeForce RTX 3080Ti show similar performances but with varying speeds and inference times. YOLOv5 achieved 94.6% precision, with an inference speed of 17.8 FPS and an inference time of 70 ms on a TITAN RTX [84], while YOLOv8s achieved 45.8% precision with 77.6 FPS at 9.9 ms inference time on an A30 platform on the UAVDT dataset [119]. Other datasets, such as VisDrone2021-DET and DOTA, showcase similar trends, with the YOLO models offering varying trade-offs among precision, inference speed, and model size across different hardware platforms. In terms of real-time performance, YOLOv5s and YOLOv8s demonstrate competitive inference speeds, with YOLOv5s-obb showing higher mAP values and faster inference times on specific datasets such as DOTA and CARPK [86].
Despite these limitations, research on RGB-based object detection using edge computing has shown promising results on the NVIDIA Jetson family of devices, achieving speeds between 14 FPS and 47 FPS [6,39,49,50,51,57,102,115,119,128]. These findings highlight the potential of optimizing object detection models for real-time performance on resource-constrained platforms. Most studies report performance on desktop GPUs, whereas real-world applications employ three processing approaches: cloud, edge, and embedded computing. The following sections highlight key studies within each of these categories.

3.1. Cloud Computing

Cloud computing is a model in which end devices, such as mobile phones, autonomous vehicles, and sensors, are connected to central servers for data processing across a large network. Central servers provide computing, storage, and digital services to users through the internet. The main drawbacks of this method are the increasing volume of edge data (collected data from drones that is transferred to ground station for processing), which is restricted by network bandwidth, and the time required to transfer data to the server, along with the need for an internet connection [129]. To overcome these issues, edge computing is used. To evaluate the trade-offs between computational capacity and latency in edge and cloud settings, experiments in [130] were performed using the YOLOv8s (FP16) model on the Orin Nano. The assessment focused on round-trip time, inference latency, and communication delay to characterize real-time performance across both environments, see Table 4. Round-Trip Time (RTT) (ms) represents the total duration required for a complete inference loop, covering both network transmission delays and model processing time. It provides an overall evaluation of the system’s performance in real time.
Table 4. Latency results in the two different deployment scenarios, edge and cloud [130].
The cloud latency shows a low value of 6.82 ms compared to the edge at 32.59 ms; however, the cloud communication latency reduces its efficiency by 341.41 ms compared to the edge, 2.5 ms only.

3.2. Edge AI Object Detection

Edge AI applies AI processing near end users at the network’s edge. In edge computing data processing is performed directly on the device or node where the data is generated [131,132]. Devices capable of supporting edge computing include System-on-Chip (SoC), Field-Programmable Gate Array (FPGA), Application-Specific Integrated Circuit (ASIC), Central Processing Unit (CPU), and Graphics Processing Unit (GPU). FPGAs have been initially viewed as one of the most promising platforms for deploying AI models due to their low power consumption. However, their limitations in inference speed and the lack of broad support for deep-learning frameworks make them less competitive today. ASICs can deliver excellent performance for specialized applications, but their long development time and high cost restrict their ability to keep pace with the fast evolution of object detection methods. In contrast, ongoing advances in semiconductor manufacturing have significantly improved the performance of GPUs, CPUs, and memory. These developments have expanded the edge-level computational resources available for neural networks, making the deployment of AI models on edge devices increasingly practical [11]. GPU-based edge platforms are closely linked to the development of deep learning and have become widely adopted for object detection due to their enhanced computational power. Typical computing platforms designed for embedded applications generally consume no more than 15W when operating under load [102]. In video detection, the NVIDIA Jetson family are frequently combined with portable devices to facilitate online detection [133,134]. Four types of edge GPU-based platforms are commonly used in the literature: NVIDIA Jetson Nano [135], NVIDIA Jetson TX2 [115], and NVIDIA Jetson Xavier NX [136,137] NVIDIA Jetson AGX Xavier [127]. It is worth mentioning that GPUs have been first introduced for real-time aerial detection in 2019 [138].
These embedded GPU platforms have higher energy efficiency compared to laptop and desktop GPUs [139], where NVIDIA Jetson TX2 and Jetson Xavier NX models are mostly used due to their moderate power consumption compared to the other series [140]. Few studies adopt a cloud-edge collaborative framework, in which the cloud provides scalability with virtually unlimited resources for large-scale data or complex tasks. At the same time, edge devices can independently scale using lightweight models optimized for specific UAV tasks. In [124], It integrates an Edge-Embedded Lightweight (E2L) object detection algorithm with an attention mechanism that helps a model focus on the most relevant parts of the input when making decisions, enabling efficient detection on edge devices without sacrificing accuracy. A fuzzy neural network-based decision-making mechanism dynamically allocates tasks between edge and cloud systems. Experimental results demonstrate that the proposed framework outperforms YOLOv4 in edge-side processing speed (NVIDIA Jetson Xavier NX) and provides better overall performance than traditional edge or cloud computing methods in both speed and accuracy.

3.3. Embedded-Onboard Object Detection

Embedded systems such as FPGA and Raspberry Pi 4B are used in a few studies research [141,142]. These systems are typically deployed directly on the UAV, using embedded or edge GPUs to perform all processing, including object detection, onboard the platform itself [138]. Table 5 shows the performance of onboard detection algorithms on edge GPUs and embedded devices.
Table 5. Embedded/Edge Platforms with onboard aerial studies references.

Field-Programmable Gate Arrays (FPGAs) and Onboard Aerial Object Detection

Field-Programmable Gate Arrays (FPGAs) have gained increasing attention in aerial object detection because they provide true hardware-level parallelism and extremely low latency. Their large arrays of programmable logic units and DSP slices allow convolution, activation, and pooling operations to run concurrently, enabling fast per-frame processing even under tight compute budgets, which is a requirement typical of UAV-based video streams (Directory of Open Access Journals; Propulsion Tech Journal). Unlike fixed accelerators, FPGAs also offer full reconfigurability, allowing neural network operations or classical vision pipelines to be mapped directly into hardware. This flexibility supports aggressive quantization, pipelined dataflows, and resource-aware custom designs, particularly suitable for embedded and UAV platforms [69]. Another major advantage is energy efficiency. FPGA-based detectors generally consume far less power than GPU systems because of their optimized dataflow and parallel execution, which is essential for battery-powered aerial vehicles operating in resource-constrained environments.
Lightweight versions such as Tiny-YOLOv3, YOLOv11-Nano, and YOLOv11-S illustrate how reducing depth, channel width, and detection heads directly lowers Look-Up Table (LUT), A Look-Up Table (LUT)BRAM, and Digital Signal Processing Slice DSP consumption. Recent FPGA implementations of YOLOv11 use techniques such as depthwise-separable convolutions, CSP-based blocks, memory-efficient routing, loop tiling, and multi-PE parallelism to achieve real-time performance despite limited on-chip resources. Other optimization strategies, such as convolution lowering, increase PE utilization and reduce DRAM traffic by converting k × k kernels into expanded 1 × 1 operations. Deployment on specialized FPGA DPUs further boosts performance, as demonstrated by LCAM-YOLOX running at 195 FPS on ZCU102 and YOLOv5 achieving over 240 FPS on Versal platforms with minor layer modifications. Additional FPGA-friendly designs, such as Tiny DarkNet, also achieve high throughput and low power through fixed-point quantization and spatial parallelism. Comparisons with SSD reveal that although SSD can achieve high throughput, its deeper backbones and multi-scale feature processing require significantly more memory and power, making real-time FPGA deployment less common. Overall, YOLO-based models remain the preferred choice for embedded detection due to their predictable dataflow, lower resource demands, and suitability for highly parallel FPGA architectures, while the optimal FPGA device ultimately depends on whether the application prioritizes power, speed, or resource efficiency [143].

4. Aerial Real-Time Deep Learning Object Detection Algorithms

A typical object detector comprises two main components: a backbone for feature extraction and a head for making predictions about object categories and bounding boxes, see Figure 2 and Figure 3. The backbone, usually made up of convolutional layers, produces feature maps that capture key visual information such as edges, textures, and shapes from the input image. Common backbone architectures include the Visual Geometry Group (VGG), Residual Network (ResNet), its extended version ResNeXt, DenseNet, MobileNet, SqueezeNet, and ShuffleNet. Based on the configuration of the detection head, detectors are generally categorized as either two-stage or one-stage models [9]. In recent advancements, an intermediate component called the neck has been introduced to aggregate and enhance multi-scale feature maps from different backbone layers before passing them to the head.
Figure 2. Two-stage algorithm main diagram [144].
Figure 3. One-stage algorithm main diagram [144].

4.1. Two-Stage and Single Stage Detectors

The following Two-stage algorithms, such as R-CNN, SPP-Net, Fast R-CNN, and Faster R-CNN, achieve high precision but at the cost of slower inference, making real-time detection challenging. In contrast, one-stage algorithms like SSD, YOLO, and RetinaNet prioritize speed, making them suitable for real-time applications, though they often exhibit lower accuracy compared to two-stage methods [9].
Whether a detector uses a single-stage or two-stage framework, it still must choose between two main ways of producing bounding boxes: anchor-based or anchor-free detection. This distinction is especially important in aerial scenarios because it directly affects efficiency, small-object performance, and suitability for onboard real-time UAV deployment. Anchor-based methods rely on predefined bounding boxes during training to predict objects, whereas anchor-free approaches eliminate the need for predefined anchor boxes and directly predict. object locations from feature maps, offering greater flexibility and reduced computational overhead, see Figure 4.
Figure 4. (a) Standard Anchor assignment, the green anchor denotes a positive match (IoU > 0.7), the red anchor indicates a negative sample (IoU < 0.3), and the gray anchor represents an ignored case with an intermediate IoU. The blue box shows the ground-truth object, (b) anchor-free center-based detection, the center pixel is treated as the positive location, surrounding pixels receive reduced negative loss, and the object dimensions are predicted through regression [145].

4.1.1. Anchor Based Methods

In many object detection models, the image is split into a grid, and predefined anchor boxes of various sizes and shapes are placed at each grid cell. These anchors guide the model in predicting different objects in the image by providing a set of initial guesses for where objects could be. The model then learns to adjust the position, size, and shape of these anchor boxes to better match the actual objects, refining them into the final predicted bounding boxes. A process called Non-Maximum Suppression (NMS) is used to remove overlapping boxes, keeping only the best one. The model uses Intersection over Union (IoU), a metric that measures the overlap between the detected bounding box and the ground truth, and confidence scores to decide which boxes to keep. The confidence score indicates how sure the model is that a box contains an object and what type of object it is. This process ensures accurate and non-redundant detections. A prediction is regarded as a true positive if the Intersection over Union (IoU) between the prediction and its closest ground-truth annotation exceeds 0.5. In the post-processing step called Non-Maximum Suppression (NMS), these predictions are filtered to result in one bounding box for each object [146].
While these methods are effective, they present several challenges, including false positives, difficulties with varying aspect ratios, and high computational costs [147]. The reliance on anchor boxes increases redundancy and computational overhead, complicating Non-Maximum Suppression (NMS), a technique that removes duplicate detections by retaining only the bounding box with the highest confidence score [146]. Additionally, anchor-based methods often suffer from poor generalization across different object shapes and create an imbalance between positive and negative samples, which complicates training, particularly for edge AI systems [22,148].
Within anchor-based detection, different bounding box strategies exist. The Horizontal Bounding Box (HBB) approach is computationally efficient and offers faster inference due to the lower number of anchors required. However, it frequently produces false positives, especially in dense scenes, where it may fail to distinguish among multiple objects or include non-object regions within the bounding box. In contrast, the Oriented Bounding Box (OBB) approach enhances detection accuracy by incorporating angle parameters, making it particularly beneficial for objects with large aspect ratios, such as vehicles and ships see Figure 5. Despite this advantage, OBB-based methods significantly increase computational complexity, requiring approximately six times as many anchors as HBB, making them considerably slower [149].
Figure 5. (a) Horizontal Boundary Boxes (HBB) and (b) Oriented Boundary Boxes (OBB) [150].
For high recall rates, anchor-based detectors must densely place a large number of anchor boxes in the input image, such as over 180,000 anchor boxes in FPN networks for an image with a shorter side of 800 pixels [145]. Recall measures how well the model finds all the relevant objects in an image. It is the ratio between the number of actual objects that the model successfully detects-true positives-and the total number of actual objects, which is the sum of true positives and missed objects. A high recall means the model detects most of the objects but might include some false positives. Balancing recall with precision, which measures the accuracy of the detections, is essential for overall performance [151]. During training, the majority of these anchor boxes are marked as negative samples. The large quantity of negative samples exacerbates the imbalance between positive and negative samples in training. In addition, it can lead to potential latency bottlenecks when transferring predictions between devices in edge AI systems [152]. To address this issue, anchor-free methods have emerged.

4.1.2. Anchor Free Methods

Anchor-free detection eliminates the need for predefined anchor boxes and directly predicts object locations from feature maps. This approach improves adaptability to a range of object sizes and aspect ratios while significantly reducing computational complexity. Only a small number of anchor-free detectors are built on two-stage frameworks [21]. RepPoints [153] is a representative example that replaces anchors with learnable point sets that capture object extent and key semantic regions. SRAF-Net [154] also adopts an anchor-free strategy but embeds it within a Faster R-CNN structure, increasing complexity. Similarly, AOPG [152] introduces oriented proposals via feature alignment modules based on Faster R-CNN. Although these methods improve localization, especially for rotated objects, they remain computationally demanding and are not suitable for real-time UAV deployment. Popular anchor-free methods such as CenterNet [145], CornerNet [155], and FCOSR [21] adopt one-stage methods with different strategies, including predicting object centers, corners, or bounding box distances. Thereby avoiding the complexities associated with anchor box design and tuning [145]. These methods provide greater flexibility and efficiency in object detection by avoiding the constraints imposed by anchor boxes.

4.2. Lightweight Networks

In object detection models, the backbone is a critical component in lightweight design, often responsible for more than 50% of the total computational load during inference [156]. Many standard networks are too computationally heavy for real-time UAV or embedded applications, motivating the adoption of lightweight backbones in remote sensing image analysis [50]. Table 6 summarizes these backbones, highlighting their key convolution methods and structural strategies for efficient feature extraction.
Table 6. Summary of lightweight network architectures.
The evolution of lightweight backbones begins with SqueezeNet (2016) [157], which introduces the Fire module to replace standard 3 × 3 convolutions with 1 × 1 kernels, reducing parameters while maintaining channel flexibility. MobileNetV1 (2017) [158] adopts depthwise separable convolutions, significantly improving computational efficiency compared to standard VGG-style convolutions. Its successor, MobileNetV2 (2018) [159], integrates residual connections and linear bottlenecks, combining expansion and compression of feature dimensions to enhance feature extraction.
ShuffleNet (2018) [160] focuses on channel-level operations, using pointwise group convolutions and channel shuffle to maintain information flow across groups. ShuffleNetV2 (2018) [161] refines this design by splitting the input channels into two branches, processing one while leaving the other unchanged, followed by concatenation and channel shuffle, thereby balancing computational cost and simplicity.
GhostNet (2020) [164] emphasizes redundancy reduction through cost-efficient linear operations to generate extra feature maps, and GhostNetV2 (2022) [166] extends this with Dynamic Feature Convolution (DFC) attention for better local-global feature extraction in mobile-friendly applications.
Finally, FasterNet (2023) [165] addresses inefficiencies in depthwise convolutions using Partial Convolution (PConv) to limit memory access and redundant computations. This enables FasterNet to achieve faster inference across GPU, CPU, and ARM platforms while maintaining competitive accuracy. For instance, FasterNet-T0 surpasses MobileViT-XXS with 2.8× faster inference on GPU, 3.3× on CPU, and 2.4× on ARM, while improving ImageNet-1k top-1 accuracy by 2.9%. Moreover, the FasterNet-L variant achieves an 83.5% accuracy, comparable to Swin-B, while delivering 36% higher GPU throughput and reducing CPU compute time by 37% [165].

4.3. Neck Network

The neck in object detection models acts as a bridge between the backbone and the head, aggregating and refining multiscale features for final prediction. Early structures such as the Feature Pyramid Network (FPN) [167], enriched shallow features with deeper semantic information, but its top–down design struggled to preserve spatial detail. PANet [168], adopted in [43,47], addressed this by adding a bottom–up pathway to strengthen shallow feature representation. EfficientDet later introduced the Bi-Directional FPN (Bi-FPN) [169], used in [86], which employs weighted top–down and bottom-up fusion for improved efficiency. Fusion in FPN-style structures typically relies on element-wise addition for efficiency [167], although concatenation is sometimes used when a richer representation is required. Neural Architecture Search (NAS) further automated feature pyramid design in NAS-FPN [170], Figure 6 shows the general neck structures.
Figure 6. Different feature fusion networks [169].

4.4. Attention Modules

Attention modules play a critical role in enhancing feature representations by directing the model’s attention to the most informative regions of an image. In computer vision, attention mechanisms are predominantly categorized into two types: channel attention and spatial attention. Both types enhance original feature representations by aggregating information across positions, though they differ in their strategies, transformations, and strengthening functions [171]. Fundamental attention mechanisms and their enhanced variants have been employed throughout the research. Table 7 shows these basic mechanisms and their key features and limitations.
Table 7. Comparison of Attention Mechanisms.
When comparing attention modules in terms of computational overhead, ECA attention introduces less overhead than CBAM and SE. At the same time, SA has a similar overhead to ECA. The Triplet Attention mechanism also maintains lower overhead compared to both CBAM and SE. Similarly, NAM (Normalization-based Attention Module) and SimAM (Simple Attention Module) are also lightweight alternatives, with SimAM being parameter-free and NAM maintaining low complexity, both significantly lighter than the mentioned methods.
The Squeeze-and-Excitation (SE-2018) module [172] effectively strengthens channel-wise feature encoding but does not account for spatial dependencies.
To address this limitation, modules such as the Convolutional Block Attention Module (CBAM-2018) [173], Bottleneck Attention Module (BAM-2018) [174], sequentially derive attention map in two dimensions, channel and spatial dimension; however, they attempt to utilize positional information by decreasing the channel dimension and then generating spatial attention by convolutions, but convolutions are limited to capturing local relationships and struggle to model long range, which are crucial for vision tasks.
Global context network (GCNet-2019) [175] incorporates spatial attention, though their reliance on convolutional operations constrains their capacity to model long-range dependencies. However, these integrated CBAM–GCNet designs often face challenges, including slow convergence and increased computational complexity.
Efficient Channel Attention (ECA-Net) [176] streamlines the channel attention mechanism of the SE block by employing a 1D convolution to compute channel weights more efficiently. Similarly, Spatial Group-wise Enhance (SGE) [177] divides the channel dimension into multiple sub-groups, each intended to capture distinct semantic representations, and applies spatial attention within each group using attention masks that scale feature vectors across spatial locations, adding minimal computational overhead and virtually no additional parameters. Despite these innovations, such models often underutilize the joint relationship between spatial and channel attention. Moreover, they typically do not incorporate identity-mapping branches, thereby limiting their overall efficiency and effectiveness.
Coordinate Attention (CA) [178] introduces positional encoding tailored for lightweight networks, improving localization while maintaining efficiency. Meanwhile, the Normalization-based Attention Module (NAM-2021) [179] selectively suppresses less relevant features, contributing to computational efficiency. Triple Attention [180] achieves a balance between cost and performance by modeling both spatial and channel-wise interactions without significant overhead. While transformer-based self-attention mechanisms offer strong global feature modeling capabilities, their high computational cost poses challenges for real-time applications on resource-constrained UAV platforms.
The Shuffle Attention (SA) [181] module provides an efficient solution by integrating two types of attention mechanisms through the use of Shuffle Units. It begins by dividing the channel dimension into several sub-features, which are then processed in parallel. Each sub-feature is passed through a Shuffle Unit to capture dependencies across both spatial and channel dimensions. Finally, the outputs from all sub-features are combined, and a channel shuffle operation is applied to facilitate information exchange among them.
Simple Attention Module (SimAM) [182] offers a unique approach compared to traditional channel-wise or spatial-wise attention mechanisms by directly generating 3D attention weights for each neuron in a feature map—without introducing additional parameters. SimAM formulates an energy function to assess neuron importance and derives a fast, closed-form solution that can be implemented efficiently. Its simplicity avoids the need for complex structural design or tuning. Recent research has demonstrated the strategic integration of attention modules across different stages of detection architectures.

4.5. Real-Time Aerial Object Detectors

All algorithms mentioned previously in Section 4 are general-purpose detectors, mainly trained and evaluated on natural-scene datasets. Because aerial imagery has distinct properties (e.g., small targets, dense object distributions, complex backgrounds), it faces specific challenges that require dedicated treatment. The majority of recent studies reviewed in this survey primarily adopt YOLO detectors, particularly YOLOv8, followed by YOLOv5, and then YOLOv7. New versions of YOLO have been released: YOLOv11 and YOLOv12, and YOLO-Gold. A summary of the specifications for each YOLO version and its typical use is presented in Table 8.
Table 8. Summary of YOLO series models: architecture, features, and limitations.
Most aerial object detection studies develop YOLO algorithms to meet the requirements of aerial challenges or to make onboard detection more efficient. The new YOLO models (YOLOv5, YOLOv8) come in different sizes, each optimized for specific use cases, balancing speed, accuracy, and computational efficiency. The smallest version, YOLO nano version (YOLOvn) is designed for extreme efficiency, making it suitable for edge AI applications and low-power devices where real-time processing is essential with minimal hardware requirements used for lightweight aerial detection as in [35,36,39,48,50,51,89,90,116]. YOLO-Small (YOLOvs) provides a balance between speed and accuracy, making it an ideal choice for real-time applications on embedded systems and mobile devices like in [7,12,33,37,43,48,51,81,82,83,87,118]. YOLO-Medium (YOLOvm) provides a trade-off between computational cost and detection performance. YOLO-M is well-suited for applications that require improved accuracy while maintaining reasonable inference speeds. YOLO-Large (YOLOvl) is more performant, with more parameters and higher accuracy, but has higher computational requirements. YOLO-X-Large (YOLOvx) is more accurate but more rigorous in development, and most often operates in a high-performance computing environment where real-time requirements are less important [199].
Table 9 summarizes the developed aerial studies along with the basic models they employ. These models have been adapted to suit various aerial detection scenarios and different application requirements, with real-time performance on desktop GPUs.
Table 9. Developed aerial studies and their base detection model.
As shown in Table 9, for real-time applications, the smaller versions, such as YOLOvn and YOLOvs, are preferred because they achieve high-speed inference with lower latency, making them well-suited for tasks such as autonomous navigation, robotics, and surveillance. The ability to select a model based on the specific needs of an application allows YOLO to remain a versatile and widely used object detection framework in both edge and cloud-based AI systems, mainly the small and Nano ones. The performance evaluation of the lightweight detectors, YOLOv9t, YOLOv7tiny, YOLOv10n, and YOLOv10s has been conducted in [201], This study evaluated YOLO models on a Jetson Xavier-equipped drone, emphasizing speed, accuracy, and resource efficiency under TinyML principles. Key metrics included inference time, mAP, GPU usage, and power consumption. YOLOv10n achieved the best balance (10.34 ms, 0.657 mAP), while YOLOv9-tiny also performed well (13.55 ms, 0.688 mAP) with slightly higher resource use. YOLOv10s achieved the highest accuracy (0.77 mAP) but demanded more power and GPU resources, favoring accuracy scenarios. However, fewer approaches apply another detector, as in [119], in which SSD detector is used.
To achieve real time time performance diffrent methods have been adapted targeting the design of lightweight models aiming to reduce size and computational complexity [163,203,204], most of them achieve this by replacing the backbone with lightweight networks [119], or by changing some module in the network [7,33,36,38,40,43,47,61,62,78,79,81,82,83,84,86,89,104,110,111,200], with fewer studies have explored other methods such as parameter pruning [57,62], quantization [5,7,39,42,46,49,56,57,115] and Knowledge distillation [63,64]. The one-stage object detector comprises three main components: The backbone for feature extraction, the neck for multiscale feature fusion, and the head, which locates objects in images and assigns classes. Various optimization and design modifications have been applied to enhance the efficiency and speed of models, with the majority of these changes focused on the backbone [7,33,38,43,81,84,110,111] and neck structure [89,104]. Fewer studies modify the head [40,91]. Studies that change backbone and neck [36,40,47,61,62,78,79,81,82,83,86,200].
The Refined Anchor-Free Rotated YOLOX detector (R2YOLOX) [205], which is an enhanced version of YOLOX [188] for aerial imagery, is an example of an anchor-free YOLO version in addition to YOLOv8 and later versions, which have also adopted an anchor-free approach and been used in many research, further advancing real-time object detection performance [35,48,92]. In recent real-time research, various algorithms have been used for aerial object detection.

5. Optimization Methods

Aerial object detection systems deployed on UAV platforms face resource constraints in terms of computation, energy, and latency, making optimization a core requirement rather than an optional enhancement. Unlike ground-based detectors that can rely on powerful servers or desktop GPUs, onboard aerial systems must operate under limited computational resources while still providing real-time, reliable detection. Therefore, optimization methods in the literature aim to either reduce the computational burden, enhance detection accuracy under resource limitations, or balance both goals to meet real-time performance requirements. In [206], Model compression for object detection is generally categorized into five main approaches: lightweight network design, pruning, quantization, knowledge distillation, and neural architecture search (NAS), which are described in detail in the following subsections.

5.1. Lightweight Design for Real-Time Aerial Detection

The following subsections provide a detailed analysis of some studies, highlighting their contributions to onboard algorithm design and optimization. The following explains the design of the lightweight backbone, neck, and head.

5.1.1. Lightweight Backbone Networks

Since the backbone is the most computationally intensive component, most lightweight models focus on designing efficient backbone architectures, following optimization strategies commonly adopted in aerial detection algorithms.
A.
Convolution-Based Lightweight Designs
In real-time aerial detection, many studies replace the original backbone with lightweight architectures to improve efficiency under embedded constraints, as demonstrated in [40,41]. Recent backbone design trends focus on replacing standard convolutions with more efficient operators, particularly in YOLO-based models such as YOLOv5 and YOLOv8. Common approaches include Depthwise Separable Convolution (DWSeparableConv) [7,47,86], often coupled with channel shuffle operations [36,44], followed by broader adoption of Ghost Convolution [39,110], Partial Convolution (PConv) [38,43], and MobileNet Bottleneck Convolution (MBConv) [81,200]. Some models integrate hybrid dilated convolutions with partial convolutions to enrich receptive fields [38], while others employ advanced variants such as Omni-dimensional Dynamic Convolution (ODConv) [111] to introduce input-adaptive filtering.
Since the backbone is the most computationally demanding component of the YOLO architecture, it plays a central role in extracting multi-scale visual features before they are passed to the neck for further aggregation [207]. Therefore, optimizing the backbone directly affects both accuracy and inference speed. As illustrated in Table 10, which summarizes works achieving real-time performance on Jetson GPUs and embedded platforms, different studies target distinct architectural sections; however, nearly 85% apply lightweight modifications to the backbone, underscoring its dominant contribution to total computational cost.
Lightweight aerial detection backbones thus integrate a spectrum of strategies with efficient convolutions, compact attention modules, reparameterized multi-branch structures, and lightweight downsampling and fusion blocks to enable high-speed onboard inference.
A major category of these methods is convolution-based lightweight design, where computationally expensive operations are replaced with more efficient alternatives. ShuffleNetV2 is widely adopted in this context: YOLOv6 combines it with the Multi-Scale Dilated Attention (MSDA) module to enhance multi-scale processing [97]. At the same time, YOLOv5n incorporates ShuffleNetV2 together with Coordinate Attention to improve spatial sensitivity [50]. Rep-ShuffleNet continues this trend by removing the traditional 1 × 1 convolution used for dimensional alignment and introducing a depthwise convolution block (DwCB) within a dual-branch structure during training [48]. Ghost-based techniques are frequently used to reduce redundancy, as seen in WILD-YOLO, which replaces multiple YOLOv7 backbone modules with GhostConv [6]; MSGD-YOLO, which integrates Ghost and dynamic convolution for adaptive feature extraction [35]; and the MGC module, which merges MaxPooling, GhostConv, and PConv to minimize parameters [125]. A streamlined YOLOv8s variant similarly replaces parts of the C2f block with GhostBlockV2, reducing complexity while preserving representational capacity [90]. Complementing these approaches, the IRFM module combines multi-scale dilated convolutions with trainable weighted fusion to enhance contextual representation, especially for small-object detection in cluttered scenes [5].
B.
Attention-Enhanced Lightweight Modules
These methods maintain representational quality by integrating efficient channel- or coordinate-attention mechanisms. Coordinate Attention (CA) has been incorporated into the YOLOv5n backbone alongside ShuffleNetV2 to enhance spatial and orientation awareness [50]. Similarly, Rep-ShuffleNet integrates Efficient Channel Attention (ECANet), which employs adaptive k-nearest neighbor based kernel selection and a refined shortcut design to strengthen channel-wise feature interactions [48]. Beyond these explicit attention mechanisms, other models adopt attention implicitly. MSGD-YOLO utilizes dynamic convolution, which inherently provides input-adaptive attention [35], whereas the IRFM module introduces a trainable weighted fusion strategy that acts as a soft attention mechanism to emphasize multi-scale contextual responses [5].
C.
Reparameterization and Structural Re-Design
These approaches improve training expressiveness through multi-branch structures, then collapse them into efficient single-path inference architectures. RepVGG introduces a structural re-parameterization strategy in which multi-branch convolutional blocks used during training are merged into a single equivalent convolution at inference, enabling the model to maintain accuracy while substantially reducing MACs, an approach that forms the basis of the structural design used in Rep-ShuffleNet [48]. Building on this concept, Rep-ShuffleNet adopts a dual-branch architecture composed of a depthwise convolution branch (DwCB) and a shortcut pathway during training, which is then collapsed into a simplified structure for efficient inference [48]. Likewise, RTD-Net employs a Lightweight Feature Module (LFM) that distributes computation across homogeneous multi-branch pathways, reducing redundancy while preserving strong representational power [115].
D.
Lightweight Pooling, Downsampling and Feature Fusion Modules
These modules enhance feature hierarchy while reducing interpolation or element-wise overhead. Lightweight downsampling and fusion strategies are also employed to strengthen feature hierarchies with minimal computational overhead. The MGC module integrates MaxPooling with GhostConv to achieve efficient downsampling through a combination of pooling and lightweight convolution [125], while the DSDM block in Spatial Pyramid SOD-YOLO enhances deep–shallow hierarchical representation using a compact downsampling structure [46]. Additionally, the IRFM module incorporates a trainable weighted multi-scale fusion mechanism applied over dilated convolutions, improving contextual encoding and boosting small-object detection performance in complex environments [5].

5.1.2. Efficient Neck Networks

A.
Enhancements to Classical FPN Structures for UAV Detection
In UAV scenarios, where objects are small and spatially sparse, enhanced feature pyramids are widely adopted. Dense-FPN (D-FPN) [104], used in YOLOv8, improves shallow-feature extraction by repeatedly integrating shallow and deep features using downsampling, upsampling, and convolution. The SPPF module enhances multiscale perception by extracting deep semantic context before merging with shallow layers, while the Dense Attention Layer (DAL) strengthens channel- and spatial-wise focusing via average, max, and stochastic pooling.
YOLOv5’s neck is modified in [79] to include a Bidirectional FPN (BDFPN), which expands the FPN hierarchy and introduces skip connections for improved multi-scale fusion, particularly valuable in dense UAV scenes with large object-scale variation.
YOLOv8 adopts a lighter PAN-FPN variant that removes post-upsampling convolutions to reduce computational load. However, this limits small-object detection due to reduced high-resolution feature fusion. To improve this, [123] presents BSSI-FPN, extending the pyramid upward to preserve spatial detail and adding a micro-object detection head. Additional downsampling blocks between the backbone and the neck further enhance semantic–spatial integration.
B.
Attention- and Module-Based Feature Fusion Enhancements
Several designs enhance feature fusion via attention or cross-scale interactions. In [89], a PA-FPN is augmented with the Symmetric C2f (SCF) module for deep feature refinement and the Efficient Multiscale Attention (EMA) module for enhanced spatial–channel coupling. Its Feature Fusion (FF) block improves texture–semantic blending. Similarly, [12] integrates the AMCC module into PA-FPN to reduce redundancy and preserve small-object details via channel spatial attention.
The MLFF module in [62] merges shallow, middle, and deep features of identical spatial size but different channels to reduce computation while preserving multi-scale richness. The Scale Compensation Feature Pyramid Network (SCFPN) in [78] replaces the large-object detection layer with an ultra-small object layer to better target fine-scale features. Additional improvements include the C2f-EMBC module in [81], which replaces bottlenecks with EMBConv and adds SE-like attention for memory-efficient enhancement, and SCFPN’s adaptation in [82], which further strengthens the merging of low-level spatial cues with high-level semantics. Cross-layer and weighted fusion improvements also exist. CWFF [86] introduces cross-layer weighted integration while adding a high-resolution P2 layer to enhance micro-object detection.
C.
Lightweight Neck Variants for Embedded and UAV Platforms
To reduce computation on embedded platforms, Slim-FPN [36] replaces YOLOv5-N’s neck with depthwise separable designs (DSC, DSSconv, DSSCSP, SCSAconv, SCSACSP), combining efficient convolution and attention for lightweight fusion. In [89], the EMA and FF modules capture cross-scale dependencies for UAV imagery, while the SCF module optimizes bottleneck placement. The CSPPartialStage from [38] further strengthens spatial information aggregation in complex aerial contexts.
AFPN [61] mitigates feature degradation across scales by stabilizing hierarchical interactions, whereas [40] proposes Triple Cross-Criss FPN (TCFPN), introducing a bidirectional residual mechanism for richer multi-scale integration and a feature aggregation module that compresses and redistributes spatial features.
D.
Detection-Oriented Modifications for Small and Tiny Objects
In [50], deformable convolution replaces standard operators, improving regression accuracy on small aircraft. MSGD-YOLO [35] addresses information loss due to repeated downsampling by adding a small-target layer to GFPN and using MSGConv for lightweight multi-scale extraction. CSPStage modifications replace 3 × 3 convolutions with RepConv, which uses multi-branch training and single-branch inference via reparameterization for faster deployment. Triple Attention is added at the end of each fusion branch to refine spatial and channel features for small-target detection.
WILD-YOLO [6] partially replaces YOLOv7’s heavy ELAN module with FasterNet for efficiency, while [90] employs Bi-PAN-FPN for improved multi-scale merging throughout the detection pipeline.
E.
Specialized Lightweight Integration Modules
SOD-YOLO [46] introduces the Lightweight Feature Integration Module (LFI), forming a dual-branch structure that merges features while minimizing element-wise operations. The combined DSDM-LFIM architecture enhances deep–shallow fusion by retaining distinct feature streams, while a P2 detection layer explicitly targets small objects. RTD-Net’s Feature Fusion Module (FFM) [115] simplifies YOLOv5’s PANet using a BiFPN variant with jump connections and weighted fusion to balance contributions across feature resolutions.
Finally, ASFF [5] enhances YOLOv3 by generating an adaptive weighted sum of resized feature maps, improving multi-scale feature blending, while E-FPN [102] expands the pyramid to four levels using Enhance Blocks for robust cross-scale detection. While NAS-based methods offer automated architecture generation, many still prioritize accuracy over efficiency, limiting their suitability for onboard deployment. Across all enhancements, the overarching challenge remains balancing computational efficiency with the need for robust small-object detection in complex UAV imagery [208]. However, new NAS methods have emerged for hardware-aware design.

5.1.3. Head Optimization Strategies for Small Object Detection

For small object detection, several modifications have been made to the head. For instance, ref. [78] introduces a high-resolution detection branch for small objects, while in [79], YOLOv5 replaces the large object prediction head with one designed for tiny objects. In [40], feature mapping layers for large objects are pruned, and a shallow detection head with self-attention is added for small objects. Similarly, ref. [12] introduces an additional detection layer for small objects in YOLOv5s, while ref. [39] in YOLOv5-n replaces the original head with a new prediction head optimized for small UAV objects.
Additionally, ref. [91] in YOLOv5 introduces a new detection head with a stride of 4 for small vehicles. Furthermore, advanced detection heads like the RT-DETR head and IoU-aware query selection [200] have replaced traditional approaches, eliminating the need for Non-Maximum Suppression (NMS) and refining predictions for better small-object detection. Several studies have modified the original YOLOv8 and YOLOv5 loss functions to enhance object localization and detection accuracy, especially for small and occluded objects. The Normalized Wasserstein Distance (NWD) added with Generalized Intersection over Union (GIoU) to improve optimization weight and regression accuracy [12,89]. The Alpha-Complete Intersection over Union ( α -CIoU) addresses the imbalance between positive and negative samples in aerial image detection [33,38], while the Focal-EIOU loss enhances localization precision [39].
The Robust Intersection over Union (RIOU) loss prevents calculation failures and ensures better shape consistency between predicted and ground truth boxes [6,36,40,81]. Additionally, the Single Stage Headless (SSH) context structure improves feature extraction, aiding in the detection of small or occluded objects [110]. For embedded processing, several studies have focused on optimizing the detection head to improve both inference speed and detection accuracy.
In RTD-Net [115], for instance, introduces the Attention Prediction Head (APH) to enhance the model’s capability to focus on relevant objects. To achieve this, the Normalization-based Attention Module (NAM) is designed as a lightweight attention mechanism that is integrated before the detection heads, forming what are known as Attention-Preceding Heads (APHs). NAM draws inspiration from CBAM but incorporates a redesigned structure for channel and spatial attention submodules. Specifically, NAM generates attention maps separately for both channel and spatial dimensions, which are then applied to the input feature map via element-wise multiplication. This selective emphasis enables the model to concentrate more effectively on salient regions, thereby improving object detection performance. In parallel, efforts have also been made to reduce the computational complexity of detection heads. The Efficient Decoupled Head aims to minimize inference costs by decreasing the number of intermediate convolutional layers, all while preserving the spatial dimensions of the input feature maps. This structure builds on the concept of the decoupled head architecture, initially introduced in FCOS [148] and later adopted by anchor-free models such as YOLOX [188]. The decoupling of classification and regression branches facilitates better network convergence and enhances prediction precision. However, this improvement typically comes with increased inference overhead due to the parallel branch design. To address this issue, further refinements have led to the development of a more lightweight decoupled head [21], characterized by fewer convolutional layers and reduced channel widths, to be able to achieve real-time performance. Additionally, to maintain strong regression performance without compromising speed, implicit representation layers have been integrated. Re-parameterization techniques are also employed in this architecture to streamline the model during inference, reducing the computational burden while maintaining detection accuracy.

5.1.4. Loss Function

To improve detection accuracy, especially for small and challenging targets, a range of advanced loss functions have been proposed. In [50], EIoU replaces the traditional CIoU to enhance regression accuracy for small objects. Rep-ShuffleNet [48] incorporates BIoU Loss to improve localization precision, while refs. [35,97] adopts WiseIoU, which employs a dynamic, non-monotonic focusing mechanism to better balance loss contributions. SCAFPN [51] introduces Hybrid-Random Loss (HRL), specifically targeting the challenges of small object detection, and RTD-Net [115] applies L1 regularization to further refine localization accuracy.
Additionally, in [21], the training process is structured into three distinct stages to gradually enhance model performance. In the initial stage, conventional loss functions such as Generalized IoU (gIoU) for bounding box regression and Balanced Cross-Entropy for classification and objectness scoring are employed. In the second stage, HRL is introduced in combination with data augmentation approaches to improve the model’s robustness, particularly for small object detection. Finally, the third stage disables data augmentation to stabilize training, replaces gIoU with Complete IoU (CIoU) for improved localization precision, and applies L1 regularization to fine-tune the model. This progressive, multi-stage training strategy effectively balances accuracy, efficiency, and generalization across diverse detection scenarios.

5.1.5. Lightweight Transformers for Real-Time Aerial Detection

Transformer-based detectors are increasingly adapted for aerial object detection due to their strong ability to model long-range dependencies and global spatial context in UAV imagery. However, standard DETR and ViT architectures remain computationally heavy because the self-attention mechanism scales quadratically with sequence length. This creates substantial memory pressure and latency, making real-time deployment challenging both onboard and even on desktop GPUs [171,209]. The core issue stems from the need for self-attention to compute relationships between all token pairs, resulting in significant computational and memory overhead as the spatial resolution increases [210]. These limitations have motivated a surge in research into lightweight, scale-aware, and frequency-preserving transformer designs suitable for aerial platforms.
Despite these constraints, several enhanced transformer-based detectors have demonstrated real-time or near-real-time performance in UAV scenarios. For instance, UAV-DETR [211] employs an RT-DETR backbone and integrates a Multi-Scale Feature Fusion and Frequency Enhancement (MSFF-FE) module to retain high-frequency texture, a frequency-focused downsampling (FD) module to preserve spatial structure during downsampling, and a Semantic Alignment and Calibration (SAC) module to improve feature consistency across different scales. These adaptations enhance the detection of small and occluded objects while maintaining real-time speed. Likewise, ref. [117] further extends RT-DETR by incorporating partial convolutions in the backbone and introducing three specialized modules tailored for UAV open-water detection: the Small-Object Enhancement Module (SOEM), the Cross-Scale Feature Pyramid Interaction Module (CFPIM), and Multiscale Sensing Fusion (MSSF). SOEM strengthens fine-grained local perception, CFPIM improves hierarchical multi-scale fusion, and MSSF models correlations between global and local contexts to better represent small and sparse targets.
Parallel efforts investigate integrating lightweight transformer components into established detection pipelines to capture long-range dependencies while reducing computation. In [51,115], the authors propose SCAFPN to mitigate inter-layer interference during feature fusion and introduce CSL-MHSA to obtain long-range contextual cues for small objects. They further reduce overhead through depthwise separable convolutions (DSConv), SPD conv, and improved SimOTA label assignment, with FP16 quantization using NVIDIA TensorRT to accelerate inference. Their fusion strategy employs a progressive Mild module that merges adjacent pyramid levels (P2–P5), avoiding dense connections and reducing semantic gaps and computational burden.
The design in [115] expands on this direction through the Enhanced Contextual Transformer Block (ECTB), which incorporates Convolutional Multi-Head Self-Attention (CMHSA) inspired by Swin Transformer to strengthen global reasoning in challenging scenes with clutter or occlusion. Recognizing the prohibitive cost of traditional self-attention, the authors introduce an additive attention mechanism that replaces quadratic key–value interactions with linear element-wise operations, significantly reducing the computational load while maintaining accuracy. This lightweight formulation enables attention to be used across all network stages. Building on these ideas, they develop the SwiftFormer family, which achieves favorable accuracy–speed trade-offs and is reported to run approximately twice as fast as MobileViT-v2 [212]. These approaches collectively illustrate an emerging trend: replacing or augmenting classical self-attention with more efficient approximations to enable deployment in computation-constrained UAV systems.
In addition to these enhanced transformer modules, lightweight DETR-style detectors also achieve promising performance on aerial tasks. Another UAV-DETR [213] integrates Channel-Aware Sensing (CAS), a Scale-Optimized Enhancement Pyramid (SOEP), and a Context–Spatial Alignment Module (CSAM), achieving 51.6% mAP@0.5 on VisDrone2019 at 30 FPS with only 16.8M parameters. AUHF-DETR [214] reaches 68 FPS on AGX Xavier in UAV simulation. AUHF-DETR is a compact real-time detector built on the RT-DETR framework and tailored for UAV-based remote sensing, where targets are often very small and hardware resources are limited. The model uses a lightweight WTC-AdaResNet backbone to extract multi-scale features efficiently, and replaces standard global self-attention with a more economical PSA attention design to keep computation manageable on embedded GPUs. A specialized BDFPN module is introduced to strengthen small-object representation and ease the difficulties that arise in one-to-one matching during training. In addition, a loss term that adaptively emphasizes small targets further improves localization accuracy. With around ten million parameters and high inference speed on devices such as the AGX Xavier, AUHF-DETR improves its superiority over YOLOv8m and YOLOv11m in accuracy and inference. The comparison shows that AUHF-DETR delivers the most efficient Transformer-based performance on the Jetson AGX Xavier. AUHF-DETR-S achieves 68 FPS, far exceeding RT-DETR-r18 (39 FPS), while using half the parameters and less than half the GFLOPs. The medium variant (AUHF-DETR-M) also surpasses RT-DETR-r18 with higher speed and lower computation. Although YOLOv8-M and YOLOv11-M remain strong CNN baselines, AUHF-DETR-S provides competitive real-time speed with substantially lower complexity, demonstrating that the proposed design enables Transformers to operate efficiently on embedded UAV hardware, making it suitable for onboard real-time UAV applications. VMC-DETR [215] also improves small-object detection by integrating a frequency-domain VHeat C2f module, a large-kernel MFADM for multi-scale aggregation, and a CAGFM spatial–context fusion strategy, achieving 9.2 ms per image. Similarly, SF-DETR [216] introduces a lightweight dual-scale transformer backbone (ScaleFormerNet) with frequency-fused enhancement networks, reaching 51.0% mAP@0.5 while outperforming RT-DETR (R18) in the authors’ experiments.
Together, these models illustrate a consistent set of architectural principles observed in the recent literature: (1) hybrid CNN–ViT backbones that preserve edge details while enabling global reasoning; (2) explicit preservation of high-frequency information through wavelets, heat conduction, or frequency fusion; (3) simplified decoders and efficient attention approximations; and (4) progressive, rather than dense, multi-scale fusion to reduce semantic gaps and computation. While the original papers do not explicitly conclude with a unified design philosophy, the trends inferred across their architectures and results suggest that lightweight transformer-based approaches are becoming increasingly viable for real-time UAV detection when carefully engineered for efficiency, detail preservation, and hardware constraints.

5.2. Pruning

Pruning is an effective technique in convolutional neural networks (CNNs) that removes redundant parameters while maintaining overall model performance. It is typically categorized into two types: unstructured pruning and structured pruning.

5.2.1. Unstructured Pruning

Unstructured pruning operates at the individual weight level, removing less important weights by setting them to zero [58].

5.2.2. Structured Pruning

On the other hand, targets larger components of the network, such as entire filters [60] or channels [59,61]. By preserving the model’s original structure, it maintains compatibility with standard inference engines and leads to actual reductions in computation and inference time. This type of pruning evaluates the importance of filters or channels based on various criteria. For instance, one method ranks filters using their L1-norm values and prunes the least important ones [59]. Another approach evaluates the rank of output feature maps to identify and eliminate less critical filters [125]. Some techniques, such as adaptive batch normalization [60], rely on repeated cycles of pruning and fine-tuning to select optimal configurations, although this can be computationally expensive. To mitigate the overhead of multi-phase training, recent research has explored pruning techniques that require only a single training pass. One such method introduces Zero Invariant Groups (ZIGs), which deactivate unimportant channels and their associated weights, ensuring the output remains consistent with that of a smaller network [60]. It also employs Column Subset Selection (CSS) to detect redundancy among filters and applies a pruning-regrowth strategy that adjusts filter importance dynamically throughout training. DepGraph [58], used in [45]. introduces a group-level importance criterion, learning consistent sparsity across dependent parameters. By first Organizing parameters into sets, then modeling their interconnections through a dependency graph, and finally applying pruning at the group level. In [62], the feature extraction layers dedicated to detecting large objects in the YOLOv8 model are removed, leading to a substantial decrease in computational load. With some modifications to the structure, the accuracy is improved with fewer computational demands. The pruning process in [5] involves sparse training to determine the importance of channels, followed by the selective removal of channels based on importance scores. A local safety threshold is used to prevent excessive pruning, ensuring that the model retains its functionality. After pruning models are fine tuned to maintain its effectiveness.

5.3. Quantization

Model quantization involves transforming a neural network’s floating-point computations into fixed-point representations, typically using low-precision formats. This process often entails converting standard FP32 (32-bit single-precision floating-point) to either FP16 (16-bit half-precision floating-point) or INT8 (8-bit fixed-point integer). These compression techniques aim to enhance efficiency by minimizing model size and reducing storage requirements.

5.3.1. Post Training Quantization

TensorRT [51], deployed on GPUs, also helps accelerate performance through quantization and layer fusion [54,115,217]. TensorRT (NVIDIA Tensor Runtime) is a deep learning inference engine developed by NVIDIA that accelerates the execution of neural networks on NVIDIA GPUs. By optimizing models for inference, TensorRT delivers rapid performance, making it invaluable for applications requiring minimal latency and high throughput. Quantization strategies, including INT8 and FP16, have been widely explored to balance speed and accuracy. FP16 quantization has demonstrated significant improvements in inference speed, particularly on edge computing devices, while maintaining reasonable accuracy. However, despite its potential speed advantages, INT8 quantization can lead to substantial accuracy degradation in some cases [51,56]. FP32 quantization results are less favorable due to factors such as overhead from quantization/dequantization operations, unquantizable operators, and excessive memory usage on the shared hardware platform [51]. Additionally, techniques such as DSConv have been introduced to further optimize quantized models by utilizing low-cost integer operations while preserving kernel weights and output probability distributions. Replacing the standard convolution with DSConv optimized the model by quantizing the floating-point convolution kernel parameters and approximating their restoration, reducing computational resource requirements [45,55]. The impact of quantization on object detection models varies based on factors such as resolution. While reducing input resolution can enhance processing speed, it often comes at the cost of accuracy loss [127]. Methods like Selective Tile Processing (STP) have been shown to mitigate these drawbacks by maintaining localization quality while benefiting from quantization speed-ups. Automatic Mixed Precision (AMP) training, which combines FP32 and FP16 computations, has also been explored to reduce GPU memory usage and accelerate training with minimal accuracy trade-offs. Overall, FP16 quantization has proven to be a practical compromise between speed and accuracy, making it a practical choice for deployment.
In the context of aerial object detection, quantization is increasingly being explored to achieve real-time performance on resource-constrained hardware. For example, several recent UAV detection studies incorporate INT8 deployment pipelines or TensorRT-based FP16/INT8 engines to accelerate inference on Jetson-class processors [51,115,211].

5.3.2. Quantization-Aware Training (QAT)

QAT is designed to mitigate the accuracy degradation that often occurs when neural networks are quantized after training. Instead of applying low-bit quantization only at inference time, QAT introduces fake quantization operations during the forward pass of training. These operations simulate the behavior of low-precision arithmetic (e.g., 8-bit) while keeping the underlying weights in floating-point format for gradient computation. By exposing the network to quantization effects during training, the optimization process adapts the weights to be more resilient to numerical approximation, reducing the accuracy drop commonly observed with post-training quantization.
During training, fake quantization tracks the activation ranges that each layer encounters, encouraging the network to constrain its internal dynamic range. This results in more stable quantization boundaries and improves robustness once the model is converted to a fixed-point representation. Although representing weights and activations with limited bit-width inevitably introduces numerical approximation, deeper models and well-regularized architectures tend to tolerate this reduction with minimal accuracy loss [218]. However, this technique is used for natural scene datasets but has not been tested on aerial detectors yet.
These works indicate that quantization-aware optimization can make transformer-based aerial detectors more feasible for onboard execution, particularly when combined with lightweight modules or efficient attention mechanisms.

5.4. Knowledge Distillation

Knowledge distillation is a technique where a smaller, simpler model (the student) is trained to mimic the behavior of a larger, more complex model (the teacher). This approach helps improve the performance of the student model, especially on tasks with limited computational resources. In [63] knowledge distillation method is combined with a self-supervised distillation approach, with speed 311.5 FPS ON RTX 3090Ti. However, it is challenging for lightweight networks to learn effective visual features from large-scale complex datasets [64].

5.5. Neural Architecture Search (NAS) and Real Time

Neural Architecture Search (NAS) automates the design of neural networks and has become a central tool for producing architectures that satisfy both accuracy and efficiency. Traditional NAS approaches (RL or evolutionary controllers) are computationally expensive; more recent families focus on hardware-aware, efficient search and on one-shot, weight-sharing methods, which dramatically reduce the search cost while producing architectures that meet target latency and memory budgets [219].
A few practical NAS paradigms stand out for lightweight design:

5.5.1. Differentiable (Hardware-Aware) NAS

Differentiable NAS (FBNet style) [220] formulates architecture selection as a continuous optimization and incorporates measured operator latency into the search objective, enabling direct hardware-aware optimization for mobile targets. FBNet demonstrated that differentiable search with latency lookup tables can find ConvNet blocks that outperform hand-designed mobile nets on measured latency for target devices

5.5.2. Direct Hardware-Aware Search

ProxylessNAS [221], removes proxy datasets and allows searching directly on the target task and hardware, using path binarization and latency-aware regularization to produce specialized mobile networks with competitive accuracy and measured latency gains on mobile GPUs/CPUs.

5.5.3. One-Shot/Weight-Sharing and Specialized Deployment Once-for-All (OFA)

The OFA [222] network is designed to enable efficient deployment of neural networks across a wide range of devices and resource constraints without the need to train a separate model for each scenario. Instead of designing or training individual networks for different hardware or latency requirements, which is computationally expensive and environmentally costly, OFA decouples the training and search processes. A single, pre-trained OFA network can support a vast number of architectural configurations, allowing a specialized sub-network to be quickly selected to meet specific device or latency constraints without additional training. This flexibility is achieved through a progressive shrinking algorithm, which generalizes pruning across multiple dimensions, depth, width, kernel size, and input resolution, producing an enormous number of sub-networks that retain high accuracy.

5.5.4. Search Efficiency and Evaluation Methods

A separate but related strand of work focuses on efficient evaluation mechanisms (predictors, low-cost proxies, early stopping, and one-shot weight sharing) to make NAS feasible for real operational tasks. Surveys and methodological papers summarize how these techniques reduce GPU hours and make NAS practical for edge deployment [223].
The Inception-ResNet network has been improved with a Fast Architecture Search (FAS) module that automates the identification of optimal feature fusion paths and mitigates information loss during multi-layer extraction of dense, rotated targets. Anchor-free, decoupled detection heads have been proposed to separate bounding box regression from classification, enabling specialized feature learning and improved detection efficiency for densely packed or non-horizontal objects. Complementary techniques such as ellipse center sampling (ECS) expand the sampling area for rotated targets and convert rotated bounding boxes into a distance-based representation using an elliptical Gaussian distribution, facilitating gradient computation for arbitrary IoU angles. These approaches rely primarily on convolutional layers, reducing model complexity and enhancing deployability on edge platforms. Experimental evaluations on datasets such as DOTA and HRSC2016 demonstrate that these methods achieve improved accuracy and efficiency, highlighting their suitability for the unique challenges of aerial object detection [224].
YOLO-NAS has been specifically developed to enhance small-object detection, improve localization accuracy, and optimize the performance-to-compute ratio, making it particularly well-suited for real-time applications on edge devices. Its open-source design further supports research and experimentation. A key innovation of YOLO-NAS lies in its integration of quantization-aware modules, known as QSP and QCI, which employ re-parameterization techniques to enable 8-bit quantization while minimizing accuracy loss during post-training quantization. Additionally, the model incorporates a hybrid quantization strategy that selectively applies quantization to certain components, striking a balance between latency and accuracy rather than uniformly affecting all layers [225].
Neural Architecture Search (NAS) is regarded as essential for the development of efficient and robust perception models used in UAV and drone platforms, where real-time small-object detection and strict computational limitations remain major challenges.

5.5.5. Other Real-Time Research

In [120], A three-step approach has been developed and deployed within a Docker runtime environment on an Nvidia Jetson AGX Xavier board. To enable rapid object detection, the captured images are divided into K segments, each processed in parallel within separate containers running the YOLOv5 object detection algorithm. The final detection results are assembled from the individual K detections. Experimental findings strongly support the method’s efficiency, demonstrating near-real-time object detection for ultra-high-resolution (8K) images, with processing times of less than 1 s per frame.
In [57], The E3-UAV system includes three main components: the user space, decision module, and model running library. The user space allows configuration of UAV performance, model settings, and task metrics. The decision module optimizes flight parameters by balancing energy consumption with task needs, adjusting factors like altitude, speed, and detection model. It calculates energy use and selects the optimal configuration given trade-offs. The model running library supports YOLOv3 deployment on TensorRT, TensorFlow, and PyTorch for object detection. E3-UAV integrates with UAV control systems for real-time data processing and task management, focusing on energy efficiency and performance optimization.
In [118], the UAV-ODS system has three main modules: data acquisition, data processing, and data presentation. The acquisition module collects video, GPS, altitude, and speed data, which is sent to a ground control system for target detection. The processing module uses an RTMP server for real-time streaming and applies a YOLOv5s detection model to identify targets, with results sent to a server for display. The front-end module presents real-time detection and UAV flight positions. The system uses a dataset of 603 images from 16 locations in Wuhan for outfall detection, employing YOLOv5s for its speed, efficiency, and accuracy. The model’s backbone includes Focus and CSP modules, with FPN and PAN structures for multi-scale feature fusion. Post-processing, including thresholding and NMS, refines detection results.
Table 10. Summary of recent Object Detection Model architectures, tested on edge and onboard platforms.
Table 10. Summary of recent Object Detection Model architectures, tested on edge and onboard platforms.
Ref.Base ModelBackboneNeckHeadLossOpt.Platform
[46]yolov7The DSDM-LFIM backbone enhances small object detection by combining efficient deep-shallow feature extraction (DSD) with lightweight dual-branch feature fusion (LFI)Original multi-scale feature fusionAdds a high-resolution P2 branch to improve small-object detection. Uses K-means to optimize anchors per detection head for the VisDrone datasetSame as the Base modelScaling the number of channels by 0.21.4 M parameters, mAP50 33.4%, 36.6 FPS (VisDrone) inference on edge Devices (Atlas 200I DK A2)
[5]YOLOv3IRFM expands the receptive field using multi-scale dilated convolutions and fuses outputs via learnable weights and shortcut connections for better small object detectionAdaptively Spatial Feature Fusion (ASFF) to enhance multi-scale representationAnchor optimization integrated into training loop using dynamic anchor generation to maximize IoUN/ADeconvolution replaces nearest-neighbor upsampling. Coordinate decoding handled outside the model. Quantized to 8-bit using the uds710 tool. NMS and decoding are implemented in C++ for execution on the NPU.mAP50 89.7%, 35.7 FPS inference on T710 NPU (Neural Processing Unit) (UAV car custom datast)
[115]YOLOv5LFM: Uniform multi-branch design to reduce computation. ECTB: CMHSA for global feature capture and occlusion handlingEliminates redundant nodes (e.g., P1tm, P4tm) to simplify the structure. Adds shortcut connections for better feature propagation. Introduces learnable weighted fusion to adaptively emphasize informative features.The attention prediction head (APH) is designed based on the NAM attention mechanism to improve the ability of the model to extract attention regions in complex scenariosAs the base modelN/A21.7 M parameters, 33.4 inference FBS on NVIDIA Jeston TX2 (UAV air custom dataset)
[50]YOLOv5n– Combines ShuffleNet v2 with YOLOv5n.– Introduces a Coordinate Attention (CA) module at the end of the backbone to enhance spatial and orientation – Includes a custom CBRM module—composed of Conv, BatchNorm, ReLU, and MaxPool layers—for efficient feature extraction.Same as Base modelSame as Base modelReplaces CIoU with EIoU to improve bounding box regression and accelerate convergence, especially for small-scale aircraft targets.N/A0.9 M parameters, mAP50 84.8%, 22.6 FPS (post + preprocessing + inference) NVIDIA Jetson Xavier NX (MAR 20 dataset)
[6]YOLOv7– Replaces certain ELAN modules in YOLOv7 with the lightweight G-FasterNet, combining FasterNet and GhostNet to reduce parameters and memory usage. – GhostConv is used in place of standard Conv to preserve feature extraction efficiency with lower computational cost.– Replaces certain ELAN modules in YOLOv7 with the lightweight G-FasterNet, combining FasterNet and GhostNet to reduce parameters and memory usage.– GhostConv is used in place of standard Conv to preserve feature extraction efficiency with lower computational cost.SimAM attention in the head, faster net use partial convolution (PConv) scenariosSame as Base modelN/A13.38 M parameters mAP50 95.04% on Jetson Nano, Jetson Xavier NX NC S2 (WAID dataset)
[102]YOLOv4Same as Base model– E-FPN constructs a 4-level pyramid (F2–F5) for enhanced multiscale feature exchange. – An Enhance Block at the input improves semantic representation by splitting features into low- and high-resolution branches with depthwise conv and CBAM – A Refine Attention module at the output mitigates aliasing effects from repeated fusion, improving detection of small objects across scales.– PixED Head uses a spatial-channel encoder–decoder with pixel-encode (PE) and pixel-decode (PD) to boost tiny object detection efficiency. – A Feature Extraction Module (FEM) with depthwise conv, pointwise conv, and CBAM refines features. – An auxiliary head (Aux Head) aids sample assignment during training only, adding no inference cost.– Improved SimOTA label assigner with CIoU and Focal Loss addresses class and aspect-ratio imbalance. PLA loss aligns features between heads.N/A0.7 M parameters, mAP50 22.7%, 103 FPS on NVIDIA Jetson Xavier NX, and 24.3 FPS on Jetson Nano GPU
[49]YOLOv4MobileNetV3 is used to replace the original feature extraction network CSPDarkNet53 network– SPP+PAN+YOLO Head structure of YOLOv4 is still used in neck and head. – A portion of the original 3 × 3 standard convolutions in PANet is replaced with depthwise separable convolutions. This substitution reduces both computational cost and the number of parameters to approximately one-fourth of those in YOLOv4.MobileNetV3 and self-attention are integrated to enhance feature extraction, while Softer-NMS replaces DIoU-NMS to address the mismatch between classification confidence and localization accuracy. Rather than discarding overlapping boxes, Softer-NMS reduces their confidence scores and predicts localization confidence, resulting in more accurate detectionsSofter-NMS replaces DIoU-NMSN/A23.8 FPS Nvidia Jetson TX2, 9.6 FPS on Raspberry Pi 4B
[48]YOLOv8Use Rep-ShuffleNet, based on ShuffleNetv2 to improve the original backbone of YOLOv8s, add the lightweight channel attention mechanism ECANet,Same as Base modelSame as Base modelBinary Cross-Entropy (BCE) Loss is used for classification, while box regression combines Distribution Focal Loss (DFL) and CIoU Loss. – To enhance accuracy, CIoU is improved to BIoU Loss, which directly compares the actual aspect ratios of predicted and ground truth boxes, rather than relying on a relative aspect ratio similarity. This shift from approximate to precise comparison improves the model’s localization accuracy.Building on edge intelligence and federated learning concepts, the FI framework and the multilayer collaborative federated learning (MLC-FL) algorithm for efficient federated learning are introduced. – By using asynchronous communication and low-frequency data exchange, MLC-FL enables local models to optimize automatically and efficiently. This upgrades the traditional coal mine IoVT system into an intelligent, self-learning system.7.8M parameters, mAP50 94.6%, 21.6 FPS on NVIDIA Jeston AGX Xavier (CMUOD, survillance)
[92]YOLOv8Deformable Separable Convolution Block (DSCBlock), separating feature channels, a channel weighting module is proposed. This module calculates weights for the separated feature map, facilitating information exchange across channels and resolutions. Moreover, it compensates for the effect of point-wise (1 × 1) convolutions. A 3D channel weighting module is introduced to efficiently extract features by applying weighting operations along the channel dimension, avoiding the high cost of 1 × 1 convolutions, compensating for the accuracy loss with the efficient feature modeling capability of DCNv2The PA-FPN-CSPD framework introduces adaptive sampling and a novel channel weighting module to enhance feature interaction. Instead of using costly pointwise (1 × 1) convolutions, the channel weighting module operates along the third (channel) dimension, enabling efficient filtering of key features while reducing the impact of deformations. To strengthen information exchange across layers, it calculates adaptive weights for separated feature maps. Additionally, the newly designed lightweight network structure, named Cross-Stage Partially Deformable Network (CSPDBlock), built around the DSCBlock, further establishes multidimensional feature correlations, improving the representation and robustness of each layer.Same as Base modelSame as Base modelN/A8.6 Mparameter, mAP50 34.2%, 24.7 FPS, Jetson Xavier NX (VisDrone dataset)
[51]YOLOv5noptimization of the entire network using DSConv. Additionally, integrating SPD Conv, which is sensitive to small targets, using Cross-Space Learning Multi-Head Self-Attention mechanism, enhancing the C3 module by using (Distribution Shifting Convolution). DSConv achieves lower memory consumption and higher computational speed.Sparsely Connected Asymptotic Feature Pyramid Network (SCAFPN), introduces a sparse, asymptotic fusion strategy. It starts by merging adjacent low-level features and gradually incorporates higher-level features using an intermediate “Mild” module, which performs upsampling, weighted fusion, and 1 × 1 convolutions. This design limits fusion to neighboring layers, reduces parameter redundancy, preserves semantic integritylabel assignment strategy using SimOTAUse the SimOTA label assignment strategyemployed the Tensor Runtime engine (NVIDIA TensorRT) to perform FP16 quantization5.18 M parameter, mAP5034.8%, 35 FPS on NVIDIA Jetson Xavier Nx edge (VisDrone dataset)
[11]YOLOv7Same as Base modelSame as Base modelusing a decoupled regression detection headCombining Generalized IoU loss for precise localization, and balanced cross-entropy losses for objectness and classification to handle class imbalance. It also introduces a Hybrid Random Loss strategy during training to improve the detection of small objectsThis paper combines both lossy reduction and lossless reduction (re-parametrization). To enhance small object detection, a scaling and stitching approach is proposed in data augmentation and redesigns the loss function to focus more on small objects, FP16-precision with TensorRT40.5 M parameter, mAP50 44.8%, FPS inference on Nvidia Jetson AGX Xavier (VisDrone dataset)
[121]Mask R-CNN-ResNet18, YOLOv8, SSD-MobileNetSame as base modelSame as base modelSame as base modelPre-processing with Sobel–Feldman filter, Enhances contrast along object boundaries, emphasizes edges for better feature extraction, reduces background noise, and improves the visibility of small objects in complex scenesN/A83.3 FPS inference for YOLOv8, 47.6 FPS for mask R-CNN, 62.5 FPS inference for SSD-MobileNet on NVIDIA Jetson Nano, 2.56 FPS inference for YOLOv8, 17.86 FPS for mask R-CNN, 2.87 FPS inference for SSD-MobileNet on Rspbirry PI 4B
[34]YOLOv4 and YOLOv7Same as Base modelSame as Base modelSame as Base modelSame as Base model– A cloud–edge hybrid architecture in which AI tasks are handled locally at the edge, while the cloud is used for data storage, processing, and visualization. – TensorRT accelerator38–40 FPS on Jetson Xavier AGX edge (2688 × 1512 resolution) and 8–10 FPS for (3840 × 2160) resolution custom dataset
[100]YOLOv7-TinySame as Base modelSame as Base modeltruncated NMS (Non-Maximum Suppression) is usedManhattan Intersection over Union (MIOU) lossSatellite images are first divided into smaller tiles, and cloud-covered regions are filtered using the PID-Net method. The remaining clear tiles are then processed using a YOLOv7-Tiny model enhanced with MIOU loss to detect remote sensing objects. Finally, the results are mapped back to their original positions in the full image. Use TensorRTmAP50 76.9%, TensorRT-FP16 160 FPS on NVIDIA Jetson AGX Orin. Cloud (latency 8.3 ms object detection 6.3 ms, Post Processing 31.6 ms, Total 21.6 FPS (DOTA dataset)
[119]SSDmobilenet replaces VGG-16 or ResNetSame as Base modelSame as Base modelSame as Base modelN/AmAP50 92.7%, 26 FPS on NVIDIA Jetson Nano, 18 FPS on Raspberry Pi 3 B Fire detection
[93]YOLOv5, YOLOv6EfficientRep, (RepBlock, RepConv)Rep-PAN (RepBlock, RepConv)Same as Base modelSame as Base modelN/AN/A
[125]YOLOv8MaxPooling+Ghost ConvolutionPAFPN, CoordBlock includes coordinate attention and CoordConv to enhance features and reduce the loss of spatial information. Additionally, Partial Convolution (PConv) directly before the detection head.Extra head for small objectsN/Anano variant achives 1.2M parameters, mAP50 39.7%, 56 FPS with FP16 b = 1, 147FPS FP16 B = 16 inference on Jetson AGX Xavier. (VisDrone)
[57]YOLOv3Same as base modelSame as base modelThe anchor boxes are resized proportionally to match different input resolutions, ensuring optimal performance for each model configuration. For example, an anchor box of (2, 5) at a 416 × 416 resolution would be adjusted to (4, 10) for an 832 × 832 resolution.Same as base modelTensorRT inference engine with 16-bit quantization.less than 0.1 ms latency on NVIDIA Jetson Xavier NX. (VisDrone)
[56]YOLOv3Same as base modelSame as base modelSame as base modelSame as base model– Joint Quantization – Reduces the precision of weights and activations to lower bit-widths, thereby minimizing memory and computational requirements.– Tiling – Splits high-resolution images into smaller tiles to enable processing on limited hardware without sacrificing detection accuracy. Quantizationquantization speed-up baseline by 1.35 on Jetson TX2. NVIDIA Jetson TX2 (352 × 352 input)
[120]YOLOv4Shallower CPSDarkNet53, parameters are shared between the object detection and semantic segmentation tasksSame as base modeladding a segmentation head to an object detector backboneSame as base modelN/ANvidia Xavier NX
[35]YOLOv8– Integrating the Ghost module and dynamic convolution into the CSP Bottleneck with two convolutions (C2f). – Spatial Pyramid Pooling with Enhanced Local Attention Network (SPPELAN) replaces Spatial Pyramid Pooling Fast (SPPF) to expand the receptive fieldMulti-Scale Ghost Convolution (MSGConv) and Multi-Scale Generalized Feature Pyramid Network (MSGPFN). – Triple Attention39 is applied at the end of each information transfer branch to enhance the extraction of small-target information before sending the features to the network headDyHead enhances detection precision for small targets. By incorporating three self-attention mechanisms into the detection head, DyHead redefines the four-dimensional tensor L×H×W×C as a three-dimensional tensor L×S×C. This approach applies scale-aware, space-aware, and task-aware attention in the L, S, and C dimensions, respectively.Same as base modelN/A2.6 M parameters, mAP50 45.2%, 24.6 FPS on Nvidia Jetson Orin Nano
[118]YOLOv5The base model structure. use input Pixel-level and spatial-level augmentations, along with object mosaic and background fusion.Same as base modelSame as base modelSame as base modelUse TensorRT24-33 FPS inference on NVIDIA Jetson AGX Xavier (Outfall)
[214]RT-DETRPartition Split Spatial Attention (PSA) replaces global self-attention → local ROI attention with high/low-frequency decomposition.Bidirectional Dynamic Feature Fusion Pyramid Network (BDFPN) adds multi-scale bidirectional fusion with learnable dynamic weights, supplies denser supervision for small targetsSame as base modelremains RT-DETR-like but equipped with Inner-MPDIoU instead of L1 + GIoU (with a small-object penalty). The MPDIoU loss combines three components: an IoU-based overlap term, a minimum point–distance metric, and a normalization factor.same as base model10.29 M parameters, mAP50 53%, 68 FPS inference on NVIDIA Jetson AGX Xavier (VisDrone)
These studies collectively demonstrate that lightweight backbones, efficient feature fusion techniques, optimized detection heads, refined loss functions, and attention mechanisms contribute to improved object detection models, balancing accuracy and computational efficiency.
The models summarized in Table 10 collectively illustrate how recent research balances the competing demands of accuracy, computational cost, and real-time feasibility on resource-limited UAV platforms. A clear trend is the continual reduction in model parameters while maintaining acceptable detection performance. Many recent works adopt lightweight backbones such as ShuffleNetv2, GhostNet, Rep-ShuffleNet, or MobileNet, and replace standard convolutions with depthwise separable variants or Ghost modules. These strategies reduce FLOPs and memory consumption, yielding compact models (often under 2M parameters) that can achieve 20–100 FPS on devices such as Jetson Nano, Xavier NX, and AGX Xavier. However, this compression inevitably limits representational capacity, yielding lower mAP values than heavier detectors such as YOLOv7, YOLOv8-s, or Mask R-CNN, which achieve superior accuracy but require more powerful embedded hardware to sustain real-time inference. Another notable trade-off across the table concerns input image resolution. Higher resolutions enhance the visibility of small and dense objects, improving detection recall, yet they significantly increase computational burden and reduce throughput. Several studies mitigate this by adjusting anchor scales, adopting tiling strategies, or adding high-resolution branches such as P2 layers to preserve fine details without increasing input dimensions.
Because small-object detection remains one of the most significant challenges in UAV imagery, many architectures emphasize improved multi-scale representation and attention mechanisms. Enhanced feature pyramids (including E-FPN, SCAFPN, and MSGPFN), deformable convolution operators, pixel-encoder heads, weighted fusion layers, and lightweight attention modules such as NAM, SimAM, ECANet, or Triple Attention are frequently incorporated. These additions enhance the discrimination of small targets and robustness under occlusion or low contrast, though they often increase computational cost in the neck and head, leading to moderate FPS reductions when deployed on edge devices.
The influence of hardware constraints is also strongly reflected in the reported inference speeds. Devices such as Jetson Nano typically operate below 40 FPS unless the backbone is extremely lightweight, whereas Xavier NX and AGX Xavier offer sufficient capacity for mid-sized YOLOv5/YOLOv7 variants, achieving 20–103 FPS depending on the specific architecture and input resolution. The newer Orin series supports higher throughput, allowing the deployment of more complex heads and multiscale modules. Conversely, Raspberry Pi and low-power NPUs impose strict limits, making MobileNet-based SSD or highly compressed detectors the only viable options for real-time performance. These results demonstrate that model selection in UAV detection is inherently hardware-dependent.
Deployment-aware optimizations constitute another consistent theme. Many studies apply FP16 or INT8 quantization, TensorRT compilation, operator fusion, or reparameterization techniques to achieve substantial speed improvements, often doubling or tripling FPS with minimal or no loss of accuracy. Some works even redesign the entire decoding pipeline or integrate low-level C++ operations on NPUs, highlighting a shift toward co-optimizing the neural architecture and the deployment framework. As a consequence, the final achievable performance is increasingly determined not only by the network design but also by the compatibility between the model’s computational structure and the underlying hardware acceleration tools.
Based on the comparative analysis reported in [34], the inference-speed results show a consistent relationship between image resolution, model complexity, and hardware capability. On the RTX 8000, YOLOv7 in PyTorch maintains the highest throughput, reaching peak values of roughly 45–50 FPS at 1920 × 1080, demonstrating that desktop-class GPUs can sustain near–real-time performance even with heavier models. In contrast, Jetson platforms exhibit a sharp decline in speed as resolution increases. For example, YOLOv4-TRT on the Xavier AGX achieves noticeably higher FPS at 2688 × 1512 compared with 3840 × 2160, reflecting the increased computational load of processing more than twice the number of pixels. The Xavier NX shows the lowest speeds across all settings, with YOLOv7-PyTorch and YOLOv7-TRT generally operating below 10 FPS. The effect of cloud communication, sending detection results to the cloud, significantly reduces throughput, with the average speed dropping from about 12.3 FPS locally to around 5 FPS when cloud transmission is enabled. This demonstrates how network overhead can become a major bottleneck and highlights the need for efficient communication strategies in edge–cloud systems.

5.6. Leveraging CLIP Embeddings for Aerial Real-Time Visual Recognition

Contrastive Language–Image Pre-training (CLIP) is a vision-language model that learns a shared embedding space for images and natural language. The model employs two separate encoders, one for images and one for text, which are trained to project corresponding image–text pairs into a shared embedding space. During training, the similarity between matched pairs is maximized while that of mismatched pairs is minimized using a symmetric cross-entropy objective over the batch-wise similarity matrix [226]. This objective, closely related to the InfoNCE loss formulation [227], encourages semantic alignment between modalities. As a result, a pretrained CLIP model is capable of zero-shot classification, where candidate class labels are converted into text prompts, encoded by the text encoder, and the label whose embedding exhibits the highest cosine similarity with the image embedding is selected. This approach has been demonstrated to generalize effectively across diverse datasets without requiring additional task-specific fine-tuning [226].
Since CLIP is trained using natural language supervision rather than a fixed set of categorical labels, the resulting image–text embeddings capture a broad semantic space, enabling generalization to previously unseen visual concepts without requiring additional labeled data [226]. This characteristic renders CLIP and related vision–language models particularly suitable for open-vocabulary tasks, in which detectors or classifiers must accommodate categories that are not present during the original supervised training. Recent surveys and reviews have emphasized that large-scale vision language pretraining has emerged as a dominant approach for achieving open-vocabulary and zero-shot performance across multiple vision tasks, including classification, object detection, and semantic segmentation [228,229].
Researchers have adapted the CLIP paradigm for remote sensing either through domain-specific pretraining or by using CLIP as a semantic teacher. For example, RemoteCLIP [230] converts heterogeneous annotations (bounding boxes, segmentation masks) into caption style supervision and scales remote-sensing image–text datasets, producing embeddings better aligned with aerial scene semantics and improving downstream detection and segmentation tasks.
Building on RemoteCLIP, CastDet, an open vocabulary aerial object detection framework, uses a student teacher self learning paradigm in which a frozen RemoteCLIP model acts as a teacher to guide a detector during training. This external teacher provides semantic embeddings that refine both class-agnostic proposals and pseudo-labels for novel object classes unseen during supervised training. To maintain high-quality pseudo-labels, CastDet includes a dynamic label queue, ensuring that only reliable labels (as judged by RemoteCLIP) persist in training, which helps stabilize learning [231].
To enable real-time performance, several optimization strategies can be applied that reduce model size and computational demands without significantly affecting accuracy.

CLIP Optimization Techniques and Related Studies

  • Knowledge Distillation and efficient backbone replacement
    One approach is to train a smaller “student” model using knowledge from a larger pre-trained CLIP “teacher.” Techniques like those explored in CLIP-Knowledge Distillation (KD) CLIP-KD [232] show that guiding the student to replicate the teacher’s feature embeddings (feature-level mimicry) is particularly effective. This allows the student model to maintain strong cross-modal alignment for zero-shot tasks while requiring far fewer resources.
    CLIP’s original design uses large Vision Transformers or ResNets, which are computationally heavy. By distilling knowledge into a student model, it is possible to adopt a lighter backbone, such as MobileViT or compact convolutional networks [232]. This flexibility is critical for aerial deployment, as smaller backbones drastically reduce computational cost while preserving the model’s ability to align images and text.
  • Quantization and Pruning
    Further efficiency can be achieved by reducing the precision of model weights (quantization) and removing unnecessary parameters (pruning). For instance, TernaryCLIP [233] is a computationally efficient framework that reduces the precision of both the vision and text encoder weights in CLIP by representing them in ternary format, rather than using standard full-precision or floating-point values. It compresses both vision and text encoders to ternary weights while applying distillation-aware training to maintain performance. Combining pruning with quantization forms a multi-stage compression pipeline, yielding a compact, fast, and energy-efficient model capable of real-time operation.
In [116], while the model is still large, adapting CLIP for lightweight aerial systems is a promising and active research direction. The combination of domain-specific pre-training (e.g., RemoteCLIP), open-vocabulary detection frameworks (e.g., CastDet), and model compression techniques (distillation, quantization, pruning) creates a credible pathway for deploying vision-language models on resource-constrained platforms such as UAVs. However, further empirical research is needed, particularly on integrating lightweight networks and modules that have proven efficient in object detection models.
YOLO-World is a framework that has recently emerged, extending the traditional YOLO object detection paradigm to handle open-vocabulary detection. Unlike standard YOLO models, which are limited to a fixed set of categories, YOLO-World leverages vision-language modeling and pre-trained embeddings (e.g., CLIP embeddings) to recognize objects described by arbitrary text labels. This approach allows the detector to generalize to unseen categories without additional training (zero-shot capability). YOLO-World maintains the efficiency and speed characteristic of YOLO models while integrating language-guided supervision [234], achieving real-time performance on NVIDIA V100 platform with TensorRT.

5.7. Integrating NAS Methods with LLM Models for Edge Devices

In this study, a framework named PhaseNAS is proposed for detection tasks. Within PhaseNAS, LLM reasoning capacity is dynamically adjusted across exploration and refinement phases, and a structured template-based prompt language is introduced so that natural-language instructions can be translated into executable model configurations. A zero-shot detection metric is also developed to allow candidate YOLO-based architectures to be rapidly screened without the need for full training [235]. The search process is conducted within a constrained design space and is guided by three key principles. First, all architectural components are selected from predefined functional groups, such as convolutional layers or residual blocks, to ensure modularity. Second, consecutive modules must maintain dimensional consistency through strict channel alignment, enabling seamless integration throughout the network. Finally, the computational complexity of the generated architectures is restricted to predefined limits, ensuring that the resulting designs remain practical for deployment across a variety of hardware platforms. Through experiments conducted on different datasets, including VisDrone2019, it is demonstrated that superior architectures are consistently identified by PhaseNAS, with search time reduced by up to 86% while gains in accuracy and computational efficiency are achieved. On UAV detection benchmarks, YOLOv8 variants generated by PhaseNAS are shown to obtain higher mAP with reduced resource cost. PhaseNAS supports resource-adaptive model design, enabling UAVs and edge devices to perform real-time.

6. Limitation, Open Challenges and Future Directions in Onboard Real-Time Aerial Object Detection

6.1. Limitations of the Current Research

Current real-time aerial detection studies still face several structural and methodological limitations that restrict their suitability for UAV onboard deployment.
First, existing surveys and many recent studies focus primarily on detector families and high-level accuracy comparisons, yet provide only limited analysis of the end-to-end onboard pipeline, including preprocessing, postprocessing, quantization strategies, TensorRT deployment, and cloud–edge latency. This makes it difficult to translate algorithmic progress into deployable UAV solutions.
Second, lightweight and hardware-aware design remains underexplored; most works mention backbone efficiency but rarely examine trained quantization, structured pruning, re-parameterization, or on-device memory constraints, despite their direct effect on FPS and energy usage on Jetson-class devices.
Third, although Transformers are increasingly used to address fine-grained discrimination and long-range context, they often introduce heavy computation, and only a few studies redesign attention modules for edge devices. Similarly, real-time methods rarely evaluate performance on high-resolution datasets or fine-grained benchmarks such as Mar-20, thereby limiting understanding of robustness to complex scene variation.
Fourth, lack of consistent hardware-aware evaluation. Most studies benchmark FPS on desktop GPUs rather than on embedded platforms such as Jetson Nano, Xavier, Orin, EdgeTPU, or NPUs, making it difficult to translate these results to onboard UAV performance. Without measurements of pre/post-processing latency, the real-time suitability of many algorithms remains uncertain. Similarly, there is no standardized protocol for tiny-object evaluation: Definitions vary across datasets. This inconsistency hinders fair comparison and slows progress in developing reliable small-object detection methods.
Finally, modern topics essential for aerial settings like UAV-oriented NAS, adapting CLIP embeddings, domain generalization, or hardware model trained optimization are largely absent from previous surveys and only partially addressed in current literature. These gaps highlight the need for more comprehensive, deployment-oriented research.

6.2. Open Challenges

Real-time aerial detection must address a complex interaction between data characteristics, algorithm design, and hardware constraints. Several open challenges continue to limit consistent, reliable onboard performance:
  • Small-Object Detection and Fine-Grained Discrimination Small objects in aerial imagery often lose essential visual details due to high-altitude imaging, leading to blurred edges, weak textures, and low-resolution feature maps. As a result, detectors struggle to capture fine-grained cues and maintain high recall.
  • Multiscale Variation and Object Orientation Aerial scenes exhibit extreme scale variation and arbitrary object orientations, in which tiny targets lose texture during downsampling and large objects dominate feature learning. At the same time, rotated objects frequently misalign with horizontal bounding boxes, reducing localization accuracy. These factors jointly complicate feature extraction, multiscale fusion, and stable regression, making real-time onboard detection particularly challenging for lightweight models.
  • Loss Function Limitations for Tiny Objects Conventional detection losses (e.g., IoU-based and focal variants) tend to favor large objects, produce weak or unstable gradients for tiny targets, and often fail to assign reliable positives to small boxes. As a result, small objects are frequently overlooked or poorly localized.
  • Transformer complexity and limited edge adaptation Transformers provide strong global context but are computationally expensive, particularly for multi-scale features and high-resolution inputs. Only a limited number of works have redesigned attention mechanisms or used hybrid CNN–Transformer backbones that are efficient enough for real-time UAV deployment.
  • Lack of unified compression and adaptive modeling strategies Quantization, pruning, NAS, and CLIP-based distillation are often applied individually and post-hoc. However, unified compression-first pipelines are rarely investigated, such as quantization during training rather than post-training.
  • High-resolution and dense-scene processing
  • Multimodal datasets gap and onboard multimodal detection are still an open challenge.
  • Hardware and energy constraints on UAV processors. UAV onboard processors, such as Jetson Nano/Orin, NPUs, and embedded GPUs, face strict power limits (5–15 W) and limited memory bandwidth and cache sizes. Achieving real-time detection at ≥30 FPS under these constraints requires hardware-aware design, which remains insufficiently explored in many publications.
  • Limited aerial datasets for fine-grained, multimodal detection and general datasets like COCO for aerial detection.

6.3. Future Directions

The extensive range of object detection tasks in aerial imagery imposes considerable computational demands, significantly affecting the efficiency of detection models. Consequently, the development of lightweight network architectures has become a central focus in current research. The primary goal of these models is to reduce computational complexity and minimize the number of parameters without compromising detection accuracy. Several strategies have been proposed to achieve this objective:
  • Lightweight convolutional backbones. Backbones will continue moving toward operators such as depthwise convolution, Ghost modules, PConv, channel-split and shuffle, SPDConv and DSPConv.
  • Efficient multiscale and attention mechanisms. Developing more efficient attention modules (ECA, Coordinate Attention, NAM, SimAM) and lightweight multiscale fusion blocks will play a growing role in addressing dense and tiny objects. These modules preserve spatial–channel interactions with minimal computational overhead.
  • Lightweight and transformer detectors. Recent advances in Transformer-based object detection show a clear shift toward architectures that blend CNNs with ViTs to balance fine-grained edge information and global context, incorporate modules that retain crucial high-frequency details through techniques such as wavelet transforms or frequency-domain processing, streamline the decoding stage with lighter attention operations, and introduce gradual multi-scale feature fusion strategies to minimize semantic gaps while keeping computation manageable.The rise of lightweight and edge-oriented Vision Transformers (AUHF-DETR) is helping narrow the accuracy efficiency gap between CNN-based and attention-based models, and ongoing advances in structured pruning and quantization-aware training are expected to reduce this gap even further.
  • Improved training-time optimization. Methods such as SimOTA assignment, NMS-free training, balanced sample selection, and convergence-aware scheduling will become standard, as they enhance recall and robustness without adding to inference cost.
  • Structured pruning is an effective approach for removing redundant layers, filters, or channels without significantly degrading performance. In addition, compact transformer modules can be integrated into lightweight backbones to further strengthen global feature modeling.
  • Anchor-free detectors such as YOLOv8 and YOLOv11 offered for aerial scenarios due to their strong small-object sensitivity, but practical onboard deployment typically relies on their variants, which often need further optimization through pruning and quantization to meet real-time constraints.
  • Knowledge distillation for large models Distillation from large vision or multimodal models is used to transfer semantic richness into smaller models. This improves robustness and fine-grained discrimination while keeping inference lightweight for design. The CLIP-KD algorithm introduces a more efficient version compared to the original CLIP model.
  • Hardware-aware NAS integrated with compression. Neural Architecture Search will evolve toward pipelines that simultaneously account for quantization, pruning, operator fusion, and memory access patterns. This produces architectures that are inherently optimized for real-time edge execution rather than compressed post hoc.
  • Advanced hardware-aware deployment. Quantization-aware training, structured and channel-level pruning, and TensorRT/NPU acceleration. These methods jointly reduce latency, model size, and energy consumption for onboard UAV processors.
  • Cloud–edge tiling for high-resolution processing. High-resolution aerial and remote-sensing imagery will increasingly use tiling pipelines that split large frames into parallel patches. This enables cloud-side acceleration while the edge device maintains fast decision-making for navigation and safety-critical tasks.
  • Federated edge learning for adaptive UAV perception.
    UAV systems will benefit from federated learning, updating models locally and sharing only compressed gradients. This reduces bandwidth usage, improves privacy, and allows adaptation to new environments without transmitting sensitive aerial data.
  • Semantic enhancement via vision–language models. Integrating CLIP-based visual semantic embeddings and compressed multimodal backbones will improve cross-scene generalization and fine-grained recognition, even with limited aerial labels or domain shifts. It will need to integrate other compression techniques.
  • Unified compression-first model design. Combine NAS, pruning, quantization, and reparameterization from the start, producing architectures that naturally satisfy latency, memory, and power constraints rather than requiring heavy post-processing.
  • One-shot and weight-sharing architecture exploration. Future UAV detectors will increasingly rely on one-shot NAS and weight-sharing search spaces to rapidly evaluate large numbers of lightweight architectures. This avoids full training for each candidate and enables hardware-specific optimization for Jetson, NPU, or FPGA platforms.
  • Integrating CLIP embedding with NAS and optimization (quantization or pruning) helps in reducing the size, with the potential of increasing the performance.
  • Super-Resolution Branches Although not yet widely adopted, using a lightweight auxiliary super-resolution (SR) branch during training is emerging as a promising strategy for aerial small-object detection. The idea is to enhance high-resolution texture and edge information temporarily, lost during downsampling, without increasing inference cost, because the SR branch is removed during deployment (as in SuperYOLO). This approach improves feature quality for tiny objects while keeping the model efficient for onboard UAV hardware.

7. Conclusions

This survey goes beyond prior real-time studies by providing a tightly focused examination of onboard and edge-level aerial object detection. This survey emphasizes how modern lightweight design principles are applied across the entire detection pipeline. As feature extraction consumes more than 50% of computations compared to the remaining part of the object detector, it begins with lightweight convolutional feature extraction, showing how depthwise separable convolutions, partial convolutions, and efficient channel mixing reduce computation while preserving discriminative ability, which is an essential requirement for UAV platforms that operate under strict power and latency constraints. The discussion then moves to multiscale feature fusion in the neck, illustrating how streamlined FPN/PAN variants, cross-scale fusion blocks, and optimized aggregation modules generate rich spatial–semantic representations without introducing costly operations. Loss-function refinements, including adaptive IoU variants and task-aligned learning, are also analyzed for their role in stabilizing training and improving small-object sensitivity, which is often critical for aerial scenes.
The survey further investigates attention mechanisms, highlighting both lightweight fine-grained modules, such as channel or spatial attention, and soft pooling attention, and global modeling strategies inspired by Transformer designs. These modules enhance contextual understanding and long-range dependency modeling while being tailored for efficiency, making them viable for onboard processing. Building on this, the survey explores how vision–language models such as CLIP can be integrated into aerial detection pipelines to introduce semantic grounding, improve generalization to unseen categories, and support open-vocabulary detection. It describes strategies such as feature-space alignment, cross-modal distillation, or embedding projection so that CLIP’s strong semantic priors can be utilized without incurring the full computational overhead of large language models.
To address architectural exploration and hardware limitations simultaneously, the survey examines the role of Neural Architecture Search (NAS) in generating hardware-aware backbones and neck structures that match the resource budgets of UAV chips, edge GPUs, and FPGA fabrics. Complementary model compression techniques, such as structured pruning and mixed-precision quantization, are reviewed to demonstrate how they can reduce memory usage, execution latency, and energy consumption while preserving accuracy. The synthesis of these methods enables a path toward unifying lightweight design with larger pretrained models. CLIP or Transformer features can be selectively distilled into pruned, quantized, or NAS-optimized subnets that run efficiently on embedded GPUs or FPGA accelerators.
The survey also clarifies how hardware choice affects design decisions. GPUs typically deliver higher throughput and training flexibility, making them preferable for fog-node computation or heavier onboard processing when power budgets allow. FPGAs, although generally slower, offer significantly higher energy efficiency and deterministic low-latency execution, beneficial for endurance-critical UAV missions, long-flight platforms, or scenarios requiring tightly controlled timing. These distinctions motivate a forward-looking discussion of how emerging combinations of lightweight convolutional design, compact transformers, cross-modal embedding techniques, and hardware-aware NAS can drive progress toward accurate, fast, and resource-efficient aerial detection systems tailored for real-world onboard deployment.

Author Contributions

Conceptualization, N.H., A.A.A., and T.Z.; methodology, N.H., A.A.A. and T.Z.; literature review, N.H.; writing original draft preparation, N.H.; writing review and editing, N.H., A.A.A., and T.Z.; visualization, N.H., A.A.A. and T.Z.; supervision, A.A.A. and T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Nex, F.; Armenakis, C.; Cramer, M.; Cucci, D.A.; Gerke, M.; Honkavaara, E.; Kukko, A.; Persello, C.; Skaloud, J. UAV in the advent of the twenties: Where we stand and what is next. ISPRS J. Photogramm. Remote Sens. 2022, 184, 215–242. [Google Scholar] [CrossRef]
  2. Aposporis, P. Object detection methods for improving UAV autonomy and remote sensing applications. In Proceedings of the 2020 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), The Hague, The Netherlands, 7–10 December 2020; pp. 845–853. [Google Scholar] [CrossRef]
  3. Revathi, A.; Madhuvanthi, T.; Addakula, Y.; Maurya, A.; Ravi, V. Obstacle Detection in Path Planning for Unmanned Aerial Vehicles based on YOLO. In Proceedings of the 2024 International Conference on Communication, Computing and Internet of Things (IC3IoT), Chennai, India, 17–18 April 2024; pp. 1–6. [Google Scholar] [CrossRef]
  4. Chang, B.R.; Tsai, H.F.; Lyu, J.L. Drone-aided path planning for unmanned ground vehicle rapid traversing obstacle area. Electronics 2022, 11, 1228. [Google Scholar] [CrossRef]
  5. Liu, M.; Tang, L.; Li, Z. Real-Time Object Detection in UAV Vision Based on Neural Processing Units. In Proceedings of the 2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 4–6 March 2022; Volume 6, pp. 1951–1955. [Google Scholar] [CrossRef]
  6. Mou, C.; Zhu, C.; Liu, T.; Cui, X. A Novel Efficient Wildlife Detecting Method with Lightweight Deployment on UAVs Based on YOLOv7. IET Image Process. 2024, 18, 1296–1314. [Google Scholar] [CrossRef]
  7. Wang, X.; Zhao, Q.; Jiang, P.; Zheng, Y.; Yuan, L.; Yuan, P. LDS-YOLO: A Lightweight Small Object Detection Method for Dead Trees from Shelter Forest. Comput. Electron. Agric. 2022, 198, 107035. [Google Scholar] [CrossRef]
  8. Luo, J.; Chang, K.; Huang, J.; Sun, X.; Ji, Y. A UAV aerial image small object detection algorithm based on fine-grained feature preservation and multi-scale feature pyramid balancing. Complex Intell. Syst. 2026, 12, 12. [Google Scholar] [CrossRef]
  9. Tang, G.; Ni, J.; Zhao, Y.; Gu, Y.; Cao, W. A Survey of Object Detection for UAVs Based on Deep Learning. Remote Sens. 2023, 16, 149. [Google Scholar] [CrossRef]
  10. Nikouei, M.; Baroutian, B.; Nabavi, S.; Taraghi, F.; Aghaei, A.; Sajedi, A.; Moghaddam, M.E. Small object detection: A comprehensive survey on challenges, techniques and real-world applications. Intell. Syst. Appl. 2025, 27, 200561. [Google Scholar] [CrossRef]
  11. Liu, S.; Zha, J.; Sun, J.; Li, Z.; Wang, G. EdgeYOLO: An edge-real-time object detector. In Proceedings of the 2023 42nd Chinese Control Conference (CCC), Tianjin, China, 24–26 July 2023; pp. 7507–7512. [Google Scholar] [CrossRef]
  12. Qu, J.; Tang, Z.; Zhang, L.; Zhang, Y.; Zhang, Z. Remote sensing small object detection network based on attention mechanism and multi-scale feature fusion. Remote Sens. 2023, 15, 2728. [Google Scholar] [CrossRef]
  13. Liu, S.; Zhu, M.; Tao, R.; Ren, H. Fine-grained Feature Perception for Unmanned Aerial Vehicle Target Detection Algorithm. Drones 2024, 8, 181. [Google Scholar] [CrossRef]
  14. Xu, Z.; Wang, X.; Huang, K.; Chen, R. Low resolution remote sensing object detection with fine grained enhancement and swin transformer. Sci. Rep. 2025, 15, 24183. [Google Scholar] [CrossRef]
  15. Liu, Z.; Gao, G.; Sun, L.; Fang, Z. HRDNet: High-resolution detection network for small objects. arXiv 2020, arXiv:2006.07607. [Google Scholar] [CrossRef]
  16. Zhang, J.; Lei, J.; Xie, W.; Fang, Z.; Li, Y.; Du, Q. SuperYOLO: Super resolution assisted object detection in multimodal remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605415. [Google Scholar] [CrossRef]
  17. Bai, C.; Zhang, K.; Jin, H.; Qian, P.; Zhai, R.; Lu, K. SFFEF-YOLO: Small object detection network based on fine-grained feature extraction and fusion for unmanned aerial images. Image Vis. Comput. 2025, 156, 105469. [Google Scholar] [CrossRef]
  18. Ding, J.; Xue, N.; Xia, G.S.; Bai, X.; Yang, W.; Yang, M.Y.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; et al. Object Detection in Aerial Images: A Large-Scale Benchmark and Challenges. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7778–7796. [Google Scholar] [CrossRef] [PubMed]
  19. Muzammul, M.; Li, X. Comprehensive Review of Deep Learning-Based Tiny Object Detection: Challenges, Strategies, and Future Directions. Knowl. Inf. Syst. 2025, 67, 3825–3913. [Google Scholar] [CrossRef]
  20. Yu, X.; Lin, M.; Lu, J.; Ou, L. Oriented object detection in aerial images based on area ratio of parallelogram. J. Appl. Remote Sens. 2022, 16, 034510. [Google Scholar] [CrossRef]
  21. Li, Z.; Hou, B.; Wu, Z.; Ren, B.; Yang, C. FCOSR: A Simple Anchor-Free Rotated Detector for Aerial Object Detection. Remote Sens. 2023, 15, 5499. [Google Scholar] [CrossRef]
  22. Wang, Z.; Bao, C.; Cao, J.; Hao, Q. AOGC: Anchor-free oriented object detection based on Gaussian centerness. Remote Sens. 2023, 15, 4690. [Google Scholar] [CrossRef]
  23. Dong, Y.; Xie, X.; An, Z.; Qu, Z.; Miao, L.; Zhou, Z. NMS-free oriented object detection based on channel expansion and dynamic label assignment in UAV aerial images. Remote Sens. 2023, 15, 5079. [Google Scholar] [CrossRef]
  24. Liang, S.; Wang, W.; Chen, R.; Liu, A.; Wu, B.; Chang, E.C.; Cao, X.; Tao, D. Object Detectors in the Open Environment: Challenges, Solutions, and Outlook. arXiv 2024, arXiv:2403.16271. [Google Scholar] [CrossRef]
  25. Wang, A.; Sun, Y.; Kortylewski, A.; Yuille, A.L. Robust Object Detection under Occlusion with Context-Aware CompositionalNets. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12645–12654. [Google Scholar] [CrossRef]
  26. Xu, X.; Chen, Z.; Zhang, X.; Wang, G. Context-aware content interaction: Grasp subtle clues for fine-grained aircraft detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5641319. [Google Scholar] [CrossRef]
  27. Fonseca, J.; Douzas, G.; Bacao, F. Improving imbalanced land cover classification with k-means smote: Detecting and oversampling distinctive minority spectral signatures. Information 2021, 12, 266. [Google Scholar] [CrossRef]
  28. Cui, Y.; Jia, M.; Lin, T.Y.; Song, Y.; Belongie, S. Class-balanced loss based on effective number of samples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9268–9277. [Google Scholar] [CrossRef]
  29. An, Y.; Sun, Y.; Li, Y.; Yang, Y.; Yu, J.; Zhu, Z. Sample imbalance remote sensing small target detection based on discriminative feature learning and imbalanced feature semantic enrichment. Expert Syst. Appl. 2025, 273, 126753. [Google Scholar] [CrossRef]
  30. Zhao, B.; Wu, Y.; Guan, X.; Gao, L.; Zhang, B. An improved aggregated-mosaic method for the sparse object detection of remote sensing imagery. Remote Sens. 2021, 13, 2602. [Google Scholar] [CrossRef]
  31. Hong, S.; Kang, S.; Cho, D. Patch-level augmentation for object detection in aerial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 127–134. [Google Scholar] [CrossRef]
  32. Alsharabi, N. Real-time object detection overview: Advancements, challenges, and applications. J. Amran Univ. 2023, 3, 12. [Google Scholar] [CrossRef]
  33. Hu, M.; Li, Z.; Yu, J.; Wan, X.; Tan, H.; Lin, Z. Efficient-Lightweight YOLO: Improving Small Object Detection in YOLO for Aerial Images. Sensors 2023, 23, 6423. [Google Scholar] [CrossRef]
  34. Koubaa, A.; Ammar, A.; Abdelkader, M.; Alhabashi, Y.; Ghouti, L. AERO: AI-Enabled Remote Sensing Observation with Onboard Edge Computing in UAVs. Remote Sens. 2023, 15, 1873. [Google Scholar] [CrossRef]
  35. Xiao, L.; Li, W.; Yao, S.; Liu, H.; Ren, D. High-Precision and Lightweight Small-Target Detection Algorithm for Low-Cost Edge Intelligence. Sci. Rep. 2024, 14, 23542. [Google Scholar] [CrossRef]
  36. Liu, C.; Yang, D.; Tang, L.; Zhou, X.; Deng, Y. A Lightweight Object Detector Based on Spatial-Coordinate Self-Attention for UAV Aerial Images. Remote Sens. 2022, 15, 83. [Google Scholar] [CrossRef]
  37. Liu, S.; Cao, L.; Li, Y. Lightweight Pedestrian Detection Network for UAV Remote Sensing Images Based on Strideless Pooling. Remote Sens. 2024, 16, 2331. [Google Scholar] [CrossRef]
  38. Xie, S.; Zhou, M.; Wang, C.; Huang, S. CSPPartial-YOLO: A Lightweight YOLO-Based Method for Typical Objects Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 388–399. [Google Scholar] [CrossRef]
  39. Cao, J.; Bao, W.; Shang, H.; Yuan, M.; Cheng, Q. GCL-YOLO: A GhostConv-Based Lightweight YOLO Network for UAV Small Object Detection. Remote Sens. 2023, 15, 4932. [Google Scholar] [CrossRef]
  40. Yu, C.; Jiang, X.; Wu, F.; Fu, Y.; Pei, J.; Zhang, Y.; Li, X.; Fu, T. A Multi-Scale Feature Fusion Based Lightweight Vehicle Target Detection Network on Aerial Optical Images. Remote Sens. 2024, 16, 3637. [Google Scholar] [CrossRef]
  41. Chen, H.; Wang, D.; Fang, J.; Li, Y.; Xu, S.; Xu, Z. LW-YOLOv8: An Lightweight Object Detection Algorithm for UAV Aerial Imagery. In Proceedings of the 2024 6th International Conference on Natural Language Processing (ICNLP), Xi’an, China, 22–24 March 2024; pp. 446–450. [Google Scholar]
  42. He, Y.; Sahma, A.; He, X.; Wu, R.; Zhang, R. FireNet: A Lightweight and Efficient Multi-Scenario Fire Object Detector. Remote Sens. 2024, 16, 4112. [Google Scholar] [CrossRef]
  43. Xiang, Y.; Du, C.; Mei, Y.; Zhang, L.; Du, Y.; Liu, A. BN-YOLO: A Lightweight Method for Bird’s Nest Detection on Transmission Lines. J. Real Time Image Process. 2024, 21, 194. [Google Scholar] [CrossRef]
  44. Yue, M.; Zhang, L.; Huang, J.; Zhang, H. Lightweight and Efficient Tiny-Object Detection Based on Improved YOLOv8n for UAV Aerial Images. Drones 2024, 8, 276. [Google Scholar] [CrossRef]
  45. Dong, C.; Jiang, X.; Hu, Y.; Du, Y.; Pan, L. EL-Net: An Efficient and Lightweight Optimized Network for Object Detection in Remote Sensing Images. Expert Syst. Appl. 2024, 255, 124661. [Google Scholar] [CrossRef]
  46. Xiao, Y.; Di, N. SOD-YOLO: A Lightweight Small Object Detection Framework. Sci. Rep. 2024, 14, 25624. [Google Scholar] [CrossRef]
  47. Chen, N.; Li, Y.; Yang, Z.; Lu, Z.; Wang, S.; Wang, J. LODNU: Lightweight Object Detection Network in UAV Vision. J. Supercomput. 2023, 79, 10117–10138. [Google Scholar] [CrossRef]
  48. Wu, J.; Zheng, R.; Jiang, J.; Tian, Z.; Chen, W.; Wang, Z.; Yu, F.R.; Leung, V.C. A Lightweight Small Object Detection Method Based on Multilayer Coordination Federated Intelligence for Coal Mine IOVT. IEEE Internet Things J. 2024, 11, 20072–20087. [Google Scholar] [CrossRef]
  49. Liu, J.; Hu, C.; Zhou, J.; Ding, W. Object Detection Algorithm Based on Lightweight YOLOv4 for UAV. In Proceedings of the 2022 7th International Conference on Intelligent Computing and Signal Processing (ICSP), Xi’an, China, 15–17 April 2022; pp. 425–429. [Google Scholar] [CrossRef]
  50. Wang, J.; Bai, Z.; Zhang, X.; Qiu, Y. A Lightweight Remote Sensing Aircraft Target Detection Network Based on Improved YOLOv5n. In Proceedings of the 2023 9th International Conference on Computer and Communications (ICCC), Chengdu, China, 8–11 December 2023; pp. 1678–1683. [Google Scholar] [CrossRef]
  51. Xue, C.; Xia, Y.; Wu, M.; Chen, Z.; Cheng, F.; Yun, L. EL-YOLO: An Efficient and Lightweight Low-Altitude Aerial Objects Detector for Onboard Applications. Expert Syst. Appl. 2024, 256, 124848. [Google Scholar] [CrossRef]
  52. Liu, Z.; Chen, C.; Huang, Z.; Chang, Y.C.; Liu, L.; Pei, Q. A Low-Cost and Lightweight Real-Time Object-Detection Method Based on UAV Remote Sensing in Transportation Systems. Remote Sens. 2024, 16, 3712. [Google Scholar] [CrossRef]
  53. Vandersteegen, M.; Van Beeck, K.; Goedemé, T. Super Accurate Low Latency Object Detection on a Surveillance UAV. In Proceedings of the 2019 16th International Conference on Machine Vision Applications (MVA), Tokyo, Japan, 27–31 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  54. Golcarenarenji, G.; Martinez-Alpiste, I.; Wang, Q.; Alcaraz-Calero, J.M. Efficient Real-Time Human Detection Using Unmanned Aerial Vehicles Optical Imagery. Int. J. Remote Sens. 2021, 42, 2440–2462. [Google Scholar] [CrossRef]
  55. Nascimento, M.G.d.; Fawcett, R.; Prisacariu, V.A. DSConv: Efficient Convolution Operator. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5148–5157. [Google Scholar] [CrossRef]
  56. Plastiras, G.; Siddiqui, S.; Kyrkou, C.; Theocharides, T. Efficient Embedded Deep Neural-Network-Based Object Detection via Joint Quantization and Tiling. In Proceedings of the 2020 2nd IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS), Genova, Italy, 31 August–2 September 2020; pp. 6–10. [Google Scholar] [CrossRef]
  57. Suo, J.; Zhang, X.; Shi, W.; Zhou, W. E 3-UAV: An Edge-Based Energy-Efficient Object Detection System for Unmanned Aerial Vehicles. IEEE Internet Things J. 2023, 11, 4398–4413. [Google Scholar] [CrossRef]
  58. Fang, G.; Ma, X.; Song, M.; Mi, M.B.; Wang, X. DepGraph: Towards Any Structural Pruning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16091–16101. [Google Scholar] [CrossRef]
  59. Guan, T.; Zhang, C. Mixed Pruning Method for Vehicle Detection. In Proceedings of the 2020 IEEE 11th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 16–18 October 2020; pp. 232–235. [Google Scholar] [CrossRef]
  60. Ko, H.; Kang, J.K.; Kim, Y. An Efficient and Fast Filter Pruning Method for Object Detection in Embedded Systems. In Proceedings of the 2024 IEEE 6th International Conference on AI Circuits and Systems (AICAS), Abu Dhabi, UAE, 22–25 April 2024; pp. 204–207. [Google Scholar] [CrossRef]
  61. Fan, Q.; Li, Y.; Deveci, M.; Zhong, K.; Kadry, S. LUD-YOLO: A Novel Lightweight Object Detection Network for Unmanned Aerial Vehicle. Inf. Sci. 2025, 686, 121366. [Google Scholar] [CrossRef]
  62. Ning, T.; Wu, W.; Zhang, J. Small Object Detection Based on YOLOv8 in UAV Perspective. Pattern Anal. Appl. 2024, 27, 103. [Google Scholar] [CrossRef]
  63. Jobaer, S.; Tang, X.s.; Zhang, Y.; Li, G.; Ahmed, F. A Novel Knowledge Distillation Framework for Enhancing Small Object Detection in Blurry Environments with Unmanned Aerial Vehicle-Assisted Images. Complex Intell. Syst. 2025, 11, 63. [Google Scholar] [CrossRef]
  64. Li, Z.; Xu, P.; Chang, X.; Yang, L.; Zhang, Y.; Yao, L.; Chen, X. When Object Detection Meets Knowledge Distillation: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10555–10579. [Google Scholar] [CrossRef]
  65. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object Detection in 20 Years: A Survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  66. Pagire, V.; Chavali, M.; Kale, A. A Comprehensive Review of Object Detection with Traditional and Deep Learning Methods. Signal Process. 2025, 237, 110075. [Google Scholar] [CrossRef]
  67. Hua, W.; Chen, Q. A Survey of Small Object Detection Based on Deep Learning in Aerial Images. Artif. Intell. Rev. 2025, 58, 162. [Google Scholar] [CrossRef]
  68. Wen, L.; Cheng, Y.; Fang, Y.; Li, X. A Comprehensive Survey of Oriented Object Detection in Remote Sensing Images. Expert Syst. Appl. 2023, 224, 119960. [Google Scholar] [CrossRef]
  69. Hozhabr, S.H.; Giorgi, R. A Survey on Real-Time Object Detection on FPGAs. IEEE Access 2025, 13, 38195–38238. [Google Scholar] [CrossRef]
  70. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar] [CrossRef]
  71. Zhao, D.; Shao, F.; Liu, Q.; Yang, L.; Zhang, H.; Zhang, Z. A small object detection method for drone-captured images based on improved YOLOv7. Remote Sens. 2024, 16, 1002. [Google Scholar] [CrossRef]
  72. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar] [CrossRef]
  73. Wenqi, Y.; Gong, C.; Meijun, W.; Yanqing, Y.; Xingxing, X.; Xiwen, Y.; Junwei, H. MAR20: A benchmark for military aircraft recognition in remote sensing images. Natl. Remote Sens. Bull. 2024, 27, 2688–2696. [Google Scholar] [CrossRef]
  74. Varga, L.A.; Kiefer, B.; Messmer, M.; Zell, A. SeaDronesSee: A Maritime Benchmark for Detecting Humans in Open Water. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 2260–2270. [Google Scholar] [CrossRef]
  75. Partheepan, S.; Sanati, F.; Hassan, J. Autonomous unmanned aerial vehicles in bushfire management: Challenges and opportunities. Drones 2023, 7, 47. [Google Scholar] [CrossRef]
  76. Du, D.; Zhu, P.; Wen, L.; Bian, X.; Lin, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-DET2019: The vision meets drone object detection in image challenge results. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 213–226. [Google Scholar] [CrossRef]
  77. Cao, Y.; He, Z.; Wang, L.; Wang, W.; Yuan, Y.; Zhang, D.; Zhang, J.; Zhu, P.; Van Gool, L.; Han, J.; et al. VisDrone-DET2021: The Vision Meets Drone Object Detection Challenge Results. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2847–2854. [Google Scholar] [CrossRef]
  78. Chen, H.; Yang, W.; Zhou, G.; Zhang, G.; Nian, Z. MFRENet: Efficient Detection of Drone Image Based on Multiscale Feature Aggregation and Receptive Field Expanded. Pattern Anal. Appl. 2024, 27, 120. [Google Scholar] [CrossRef]
  79. Jiang, L.; Yuan, B.; Du, J.; Chen, B.; Xie, H.; Tian, J.; Yuan, Z. MFFSODNet: Multiscale feature fusion small object detection network for UAV aerial images. IEEE Trans. Instrum. Meas. 2024, 73, 5015214. [Google Scholar] [CrossRef]
  80. Muzammul, M.; Algarni, A.; Ghadi, Y.Y.; Assam, M. Enhancing UAV Aerial Image Analysis: Integrating Advanced SAHI Techniques with Real-Time Detection Models on the VisDrone Dataset. IEEE Access 2024, 12, 21621–21633. [Google Scholar] [CrossRef]
  81. Xue, H.; Wang, X.; Xia, Y.; Tang, Z.; Li, L.; Wang, L. Enhanced YOLOv8 for Small Object Detection in UAV Aerial Photography: YOLO-UAV. In Proceedings of the 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, Japan, 30 June–5 July 2024; pp. 1–8. [Google Scholar] [CrossRef]
  82. Ni, J.; Zhu, S.; Tang, G.; Ke, C.; Wang, T. A Small-Object Detection Model Based on Improved YOLOv8s for UAV Image Scenarios. Remote Sens. 2024, 16, 2465. [Google Scholar] [CrossRef]
  83. Niu, K.; Yan, Y. A Small-Object-Detection Model Based on Improved YOLOv8 for UAV Aerial Images. In Proceedings of the 2023 2nd International Conference on Artificial Intelligence and Intelligent Information Processing (AIIIP), Hangzhou, China, 27–29 October 2023; pp. 57–60. [Google Scholar] [CrossRef]
  84. Wu, S.; Lu, X.; Guo, C.; Guo, H. Accurate UAV Small Object Detection Based on HRFPN and EfficentVMamba. Sensors 2024, 24, 4966. [Google Scholar] [CrossRef] [PubMed]
  85. Ren, L.; Lei, H.; Li, Z.; Yang, W. AF-DETR: Efficient UAV small object detector via Assemble-and-Fusion mechanism. Pattern Anal. Appl. 2024, 27, 135. [Google Scholar] [CrossRef]
  86. Dong, D.; Li, J.; Liu, H.; Deng, L.; Gu, J.; Liu, L.; Li, S. EA-YOLO: An Efficient and Accurate UAV Image Object Detection Algorithm. IEEJ Trans. Electr. Electron. Eng. 2025, 20, 61–68. [Google Scholar] [CrossRef]
  87. Sang, M.; Tian, S.; Yu, L.; Wang, G.; Peng, Y. Environmentally adaptive fast object detection in UAV images. Image Vis. Comput. 2024, 148, 105103. [Google Scholar] [CrossRef]
  88. Xu, H.; Zheng, W.; Liu, F.; Li, P.; Wang, R. Unmanned aerial vehicle perspective small target recognition algorithm based on improved YOLOv5. Remote Sens. 2023, 15, 3583. [Google Scholar] [CrossRef]
  89. Qi, S.; Song, X.; Shang, T.; Hu, X.; Han, K. MSFE-YOLO: An improved YOLOv8 network for object detection on drone view. IEEE Geosci. Remote Sens. Lett. 2024, 21, 6013605. [Google Scholar] [CrossRef]
  90. Li, Y.; Fan, Q.; Huang, H.; Han, Z.; Gu, Q. A modified YOLOv8 detection network for UAV aerial image recognition. Drones 2023, 7, 304. [Google Scholar] [CrossRef]
  91. Hamzenejadi, M.H.; Mohseni, H. Real-time vehicle detection and classification in UAV imagery using improved YOLOv5. In Proceedings of the 2022 12th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 17–18 November 2022; pp. 231–236. [Google Scholar] [CrossRef]
  92. Hua, W.; Chen, Q.; Chen, W. A new lightweight network for efficient UAV object detection. Sci. Rep. 2024, 14, 13288. [Google Scholar] [CrossRef]
  93. Parkavi, A.; Alex, S.A.; Pushpalatha, M.; Shukla, P.K.; Pandey, A.; Sharma, S. Drone-based intelligent system for social distancing compliance using YOLOv5 and YOLOv6 with euclidean distance metric. SN Comput. Sci. 2024, 5, 972. [Google Scholar] [CrossRef]
  94. Wang, G.; Chen, Y.; An, P.; Hong, H.; Hu, J.; Huang, T. UAV-YOLOv8: A small-object-detection model based on improved YOLOv8 for UAV aerial photography scenarios. Sensors 2023, 23, 7190. [Google Scholar] [CrossRef]
  95. Xiao, Y.; Xu, T.; Xin, Y.; Li, J. FBRT-YOLO: Faster and Better for Real-Time Aerial Image Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 8673–8681. [Google Scholar] [CrossRef]
  96. Ben Rouighi, I.; Chtioui, H.; Jegham, I.; Alouani, I.; Ben Khalifa, A. FRD-YOLO: A faster real-time object detector for aerial imagery. J. Real Time Image Process. 2025, 22, 169. [Google Scholar] [CrossRef]
  97. Weng, S.; Wang, H.; Wang, J.; Xu, C.; Zhang, E. YOLO-SRMX: A Lightweight Model for Real-Time Object Detection on Unmanned Aerial Vehicles. Remote Sens. 2025, 17, 2313. [Google Scholar] [CrossRef]
  98. Song, Z.; Zhang, Y.; Abd Al Rahman, M. Ednet: Edge-optimized small target detection in UAV imagery—faster context attention, better feature fusion, and hardware acceleration. In Proceedings of the 2024 IEEE Smart World Congress (SWC), Denarau Island, Fiji, 2–7 December 2024; pp. 829–838. Available online: https://arxiv.org/html/2501.05885v1.
  99. Berner, C.; Brockman, G.; Chan, B.; Cheung, V.; Dębiak, P.; Dennison, C.; Farhi, D.; Fischer, Q.; Hashme, S.; Hesse, C.; et al. Dota 2 with Large Scale Deep Reinforcement Learning. arXiv 2019, arXiv:1912.06680. [Google Scholar] [CrossRef]
  100. Shen, Y.; Liu, D.; Chen, J.; Wang, Z.; Wang, Z.; Zhang, Q. On-board multi-class geospatial object detection based on convolutional neural network for High Resolution Remote Sensing Images. Remote Sens. 2023, 15, 3963. [Google Scholar] [CrossRef]
  101. Du, D.; Qi, Y.; Yu, H.; Yang, Y.; Duan, K.; Li, G.; Zhang, W.; Huang, Q.; Tian, Q. The Unmanned Aerial Vehicle Benchmark: Object Detection and Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 370–386. [Google Scholar]
  102. Min, X.; Zhou, W.; Hu, R.; Wu, Y.; Pang, Y.; Yi, J. Lwuavdet: A lightweight UAV object detection network on edge devices. IEEE Internet Things J. 2024, 11, 24013–24023. [Google Scholar] [CrossRef]
  103. Hsieh, M.R.; Lin, Y.L.; Hsu, W.H. Drone-Based Object Counting by Spatially Regularized Regional Proposal Network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4145–4153. [Google Scholar] [CrossRef]
  104. Wu, Z.; Wang, X.; Jia, M.; Liu, M.; Sun, C.; Wu, C.; Wang, J. Dense object detection methods in RAW UAV imagery based on YOLOv8. Sci. Rep. 2024, 14, 18019. [Google Scholar] [CrossRef]
  105. Yi, H.; Liu, B.; Zhao, B.; Liu, E. Small object detection algorithm based on improved YOLOv8 for remote sensing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 1734–1747. [Google Scholar] [CrossRef]
  106. Sun, Y.; Cao, B.; Zhu, P.; Hu, Q. Drone-Based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6700–6713. [Google Scholar] [CrossRef]
  107. Cheng, G.; Zhou, P.; Han, J. Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  108. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object Detection in Optical Remote Sensing Images: A Survey and a New Benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  109. Duan, R.; Deng, H.; Tian, M.; Deng, Y.; Lin, J. SODA: A Large-Scale Open Site Object Detection Dataset for Deep Learning in Construction. Autom. Constr. 2022, 142, 104499. [Google Scholar] [CrossRef]
  110. Qiu, Z.; Bai, H.; Chen, T. Special vehicle detection from UAV perspective via YOLO-GNS based deep learning network. Drones 2023, 7, 117. [Google Scholar] [CrossRef]
  111. Gautam, V.; Prasad, S.; Sinha, S. Joint-YODNet: A light-weight object detector for UAVs to achieve above 100 fps. In Computer Vision and Image (Proceedings of the 8th International Conference, CVIP 2023), Chandigarh, India, 16–18 December 2023; Communications in Computer and Information Science; Springer: Cham, Switzerland, 2024; Volume 2010, pp. 567–578. [Google Scholar] [CrossRef]
  112. Cao, F.; Chen, S.; Zhong, J.; Gao, Y. Traffic Condition Classification Model Based on Traffic-Net. Comput. Intell. Neurosci. 2023, 2023, 7812276. [Google Scholar] [CrossRef] [PubMed]
  113. Razakarivony, S.; Jurie, F. Vehicle detection in aerial imagery: A small target detection benchmark. J. Vis. Commun. Image Represent. 2016, 34, 187–203. [Google Scholar] [CrossRef]
  114. Zeng, H.; Li, J.; Qu, L. Lightweight low-altitude UAV object detection based on improved YOLOv5s. Int. J. Adv. Netw. Monit. Control. 2024, 9, 87–99. [Google Scholar] [CrossRef]
  115. Ye, T.; Qin, W.; Zhao, Z.; Gao, X.; Deng, X.; Ouyang, Y. Real-time object detection network in UAV-vision based on CNN and transformer. IEEE Trans. Instrum. Meas. 2023, 72, 2505713. [Google Scholar] [CrossRef]
  116. Saeed, F.; Aldera, S.; Al-Shamma’a, A.A.; Farh, H.M.H. Rapid adaptation in photovoltaic defect detection: Integrating CLIP with YOLOv8n for efficient learning. Energy Rep. 2024, 12, 5383–5395. [Google Scholar] [CrossRef]
  117. Ma, S.; Zhang, Y.; Peng, L.; Sun, C.; Ding, L.; Zhu, Y. OWRT-DETR: A novel real-time transformer network for small object detection in open water search and rescue from UAV aerial imagery. IEEE Trans. Geosci. Remote Sens. 2025, 63, 4205313. [Google Scholar] [CrossRef]
  118. Xu, H.; Huang, Q.; Yang, Y.; Li, J.; Chen, X.; Han, W.; Wang, L. UAV-ODS: A real-time outfall detection system based on UAV remote sensing and edge computing. In Proceedings of the 2022 IEEE International Conference on Unmanned Systems (ICUS), Guangzhou, China, 28–30 October 2022; pp. 1–9. [Google Scholar] [CrossRef]
  119. Nguyen, A.; Nguyen, H.; Tran, V.; Pham, H.X.; Pestana, J. A visual real-time fire detection using single shot multibox detector for UAV-based fire surveillance. In Proceedings of the 2020 IEEE Eighth International Conference on Communications and Electronics (ICCE), Phu Quoc Island, Vietnam, 13–15 January 2021; pp. 338–343. [Google Scholar] [CrossRef]
  120. Antonakakis, M.; Tzavaras, A.; Tsakos, K.; Spanakis, E.G.; Sakkalis, V.; Zervakis, M.; Petrakis, E.G. Real-time object detection using an ultra-high-resolution camera on embedded systems. In Proceedings of the 2022 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 21–23 June 2022; pp. 1–6. [Google Scholar] [CrossRef]
  121. Vasishta, M.S.; Amireddy, A.T.R.; Shrivastava, P.; Raghuram S; Talasila, V.; Shastry, P.N. Small Object Detection for UAVs Using Deep Learning Models on Edge Computing: A Comparative Analysis. In Proceedings of the 2024 5th International Conference on Circuits, Control, Communication and Computing (I4C), Bangalore, India, 4–5 October 2024; pp. 106–112. [Google Scholar] [CrossRef]
  122. Lyu, M.; Zhao, Y.; Huang, C.; Huang, H. Unmanned aerial vehicles for search and rescue: A survey. Remote Sens. 2023, 15, 3266. [Google Scholar] [CrossRef]
  123. Li, Y.; Li, Q.; Pan, J.; Zhou, Y.; Zhu, H.; Wei, H.; Liu, C. Sod-YOLO: Small-object-detection algorithm based on improved YOLOv8 for UAV images. Remote Sens. 2024, 16, 3057. [Google Scholar] [CrossRef]
  124. Yuan, Y.; Gao, S.; Zhang, Z.; Wang, W.; Xu, Z.; Liu, Z. Edge-cloud collaborative UAV object detection: Edge-embedded lightweight algorithm design and task offloading using fuzzy neural network. IEEE Trans. Cloud Comput. 2024, 12, 306–318. [Google Scholar] [CrossRef]
  125. Nghiem, V.Q.; Nguyen, H.H.; Hoang, M.S. LEAF-YOLO: Lightweight Edge-Real-Time Small Object Detection on Aerial Imagery. Intell. Syst. Appl. 2025, 25, 200484. [Google Scholar] [CrossRef]
  126. Guo, S.; Zhao, C.; Wang, G.; Yang, J.; Yang, S. Ec2detect: Real-time online video object detection in edge-cloud collaborative IoT. IEEE Internet Things J. 2022, 9, 20382–20392. [Google Scholar] [CrossRef]
  127. Ganesh, P.; Chen, Y.; Yang, Y.; Chen, D.; Winslett, M. YOLO-ReT: Towards high accuracy real-time object detection on edge GPUs. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 3267–3277. [Google Scholar] [CrossRef]
  128. Guo, Y.; Tong, X.; Xu, X.; Liu, S.; Feng, Y.; Xie, H. An anchor-free network with density map and attention mechanism for multiscale object detection in aerial images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6516705. [Google Scholar] [CrossRef]
  129. Khan, W.Z.; Ahmed, E.; Hakak, S.; Yaqoob, I.; Ahmed, A. Edge computing: A survey. Future Gener. Comput. Syst. 2019, 97, 219–235. [Google Scholar] [CrossRef]
  130. Rey, L.; Bernardos, A.M.; Dobrzycki, A.D.; Carramiñana, D.; Bergesio, L.; Besada, J.A.; Casar, J.R. A performance analysis of You Only Look Once models for deployment on constrained computational edge devices in drone applications. Electronics 2025, 14, 638. [Google Scholar] [CrossRef]
  131. Imran, H.A.; Mujahid, U.; Wazir, S.; Latif, U.; Mehmood, K. Embedded development boards for edge-AI: A comprehensive report. arXiv 2020, arXiv:2009.00803. [Google Scholar] [CrossRef]
  132. Chen, N.; Chen, Y. Anomalous vehicle recognition in smart urban traffic monitoring as an edge service. Future Internet 2022, 14, 54. [Google Scholar] [CrossRef]
  133. Ren, X.; Sun, M.; Zhang, X.; Liu, L.; Zhou, H.; Ren, X. An improved mask-RCNN algorithm for UAV TIR video stream target detection. Int. J. Appl. Earth Obs. Geoinf. 2022, 106, 102660. [Google Scholar] [CrossRef]
  134. Shin, D.J.; Kim, J.J. A deep learning framework performance evaluation to use YOLO in Nvidia Jetson platform. Appl. Sci. 2022, 12, 3734. [Google Scholar] [CrossRef]
  135. Menshchikov, A.; Shadrin, D.; Prutyanov, V.; Lopatkin, D.; Sosnin, S.; Tsykunov, E.; Iakovlev, E.; Somov, A. Real-time detection of hogweed: UAV platform empowered by deep learning. IEEE Trans. Comput. 2021, 70, 1175–1188. [Google Scholar] [CrossRef]
  136. Balamuralidhar, N.; Tilon, S.; Nex, F. MultEYE: Monitoring system for real-time vehicle detection, tracking and speed estimation from UAV imagery on edge-computing platforms. Remote Sens. 2021, 13, 573. [Google Scholar] [CrossRef]
  137. Xiong, Y.; Liu, H.; Gupta, S.; Akin, B.; Bender, G.; Wang, Y.; Kindermans, P.J.; Tan, M.; Singh, V.; Chen, B. MobileDets: Searching for Object Detection Architectures for Mobile Accelerators. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 19–25 June 2021; pp. 3825–3834. [Google Scholar]
  138. Hossain, S.; Lee, D.j. Deep learning-based real-time multiple-object detection and tracking from aerial imagery via a flying robot with GPU-based embedded devices. Sensors 2019, 19, 3371. [Google Scholar] [CrossRef] [PubMed]
  139. Nguyen, H.H.; Tran, D.N.N.; Jeon, J.W. Towards real-time vehicle detection on edge devices with nvidia jetson tx2. In Proceedings of the 2020 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Seoul, Republic of Korea, 1–3 November 2020; pp. 1–4. [Google Scholar] [CrossRef]
  140. Tijtgat, N.; Van Ranst, W.; Goedeme, T.; Volckaert, B.; De Turck, F. Embedded real-time object detection for a UAV warning system. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 2110–2118. [Google Scholar] [CrossRef]
  141. Chen, C.; Min, H.; Peng, Y.; Yang, Y.; Wang, Z. An intelligent real-time object detection system on drones. Appl. Sci. 2022, 12, 10227. [Google Scholar] [CrossRef]
  142. Sarkar, D.; Gunturi, S.K. Online health status monitoring of high voltage insulators using deep learning model. Vis. Comput. 2022, 38, 4457–4468. [Google Scholar] [CrossRef]
  143. Sali, S.M.; Meribout, M.; Majeed, A.A. Real Time FPGA Based CNNs for Detection, Classification, and Tracking in Autonomous Systems: State of the Art Designs and Optimizations. arXiv 2025, arXiv:2509.04153. [Google Scholar] [CrossRef]
  144. Carranza-García, M.; Torres-Mateo, J.; Lara-Benítez, P.; García-Gutiérrez, J. On the performance of one-stage and two-stage object detectors in autonomous vehicles using camera data. Remote Sens. 2020, 13, 89. [Google Scholar] [CrossRef]
  145. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar] [CrossRef]
  146. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  147. Chen, W.; Luo, J.; Zhang, F.; Tian, Z. A review of object detection: Datasets, performance evaluation, architecture, applications and current trends. Multimed. Tools Appl. 2024, 83, 65603–65661. [Google Scholar] [CrossRef]
  148. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A simple and strong anchor-free object detector. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1922–1933. [Google Scholar] [CrossRef] [PubMed]
  149. Yang, X.; Yan, J.; Feng, Z.; He, T. R3det: Refined single-stage detector with feature refinement for rotating object. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 3163–3171. [Google Scholar] [CrossRef]
  150. Wang, P.; Niu, Y.; Wang, J.; Ma, F.; Zhang, C. Arbitrarily oriented dense object detection based on center point network in remote sensing images. Remote Sens. 2022, 14, 1536. [Google Scholar] [CrossRef]
  151. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. Scrdet: Towards more robust detection for small, cluttered and rotated objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8232–8241. [Google Scholar] [CrossRef]
  152. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625411. [Google Scholar] [CrossRef]
  153. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. Reppoints: Point set representation for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9657–9666. [Google Scholar] [CrossRef]
  154. Liu, J.; Li, S.; Zhou, C.; Cao, X.; Gao, Y.; Wang, B. SRAF-Net: A scene-relevant anchor-free object detection network in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5405914. [Google Scholar] [CrossRef]
  155. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  156. Sun, Z.; Lin, M.; Sun, X.; Tan, Z.; Li, H.; Jin, R. MAE-DET: Revisiting maximum entropy principle in zero-shot NAS for efficient object detection. arXiv 2021, arXiv:2111.13336. [Google Scholar] [CrossRef]
  157. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar] [CrossRef]
  158. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  159. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  160. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  161. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet v2: Practical guidelines for efficient CNN architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  162. Gholami, A.; Kwon, K.; Wu, B.; Tai, Z.; Yue, X.; Jin, P.; Zhao, S.; Keutzer, K. SqueezeNext: Hardware-Aware Neural Network Design. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1638–1647. [Google Scholar] [CrossRef]
  163. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar] [CrossRef]
  164. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar] [CrossRef]
  165. Chen, J.; Kao, S.h.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, don’t walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12021–12031. [Google Scholar] [CrossRef]
  166. Tang, Y.; Han, K.; Guo, J.; Xu, C.; Xu, C.; Wang, Y. GhostNetv2: Enhance cheap operation with long-range attention. Adv. Neural Inf. Process. Syst. 2022, 35, 9969–9982. [Google Scholar]
  167. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  168. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef]
  169. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar] [CrossRef]
  170. Ghiasi, G.; Lin, T.Y.; Le, Q.V. NAS-FPN: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7036–7045. [Google Scholar] [CrossRef]
  171. Chen, Q.; Su, X.; Zhang, X.; Wang, J.; Chen, J.; Shen, Y.; Han, C.; Chen, Z.; Xu, W.; Li, F.; et al. Lw-detr: A transformer replacement to yolo for real-time detection. arXiv 2024, arXiv:2406.03459. [Google Scholar] [CrossRef]
  172. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  173. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  174. Park, J.; Woo, S.; Lee, J.Y.; Kweon, I.S. Bam: Bottleneck attention module. arXiv 2018, arXiv:1807.06514. [Google Scholar] [CrossRef]
  175. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 1971–1980. [Google Scholar] [CrossRef]
  176. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar] [CrossRef]
  177. Li, X.; Hu, X.; Yang, J. Spatial group-wise enhance: Improving semantic feature learning in convolutional networks. arXiv 2019, arXiv:1905.09646. [Google Scholar] [CrossRef]
  178. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar] [CrossRef]
  179. Liu, Y.; Shao, Z.; Teng, Y.; Hoffmann, N. NAM: Normalization-based attention module. arXiv 2021, arXiv:2111.12419. [Google Scholar] [CrossRef]
  180. Misra, D.; Nalamada, T.; Arasanipalai, A.U.; Hou, Q. Rotate to attend: Convolutional triplet attention module. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3139–3148. [Google Scholar] [CrossRef]
  181. Zhang, Q.L.; Yang, Y.B. Sa-net: Shuffle attention for deep convolutional neural networks. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2235–2239. [Google Scholar] [CrossRef]
  182. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  183. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar] [CrossRef]
  184. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef]
  185. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  186. Jocher, G. Ultralytics YOLOv5, Release v6.1. GitHub Repository, 2022. Available online: https://github.com/ultralytics/yolov5/releases/tag/v6.1 (accessed on 25 October 2025).
  187. Ultralytics. YOLOv5 vs YOLOv8: Evolution of Real Time Object Detection. 2023. Available online: https://docs.ultralytics.com/compare/yolov5-vs-yolov8/ (accessed on 29 November 2025).
  188. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  189. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  190. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar] [CrossRef]
  191. Jocher, G. Ultralytics YOLOv8. GitHub Repository. 2023. Available online: https://docs.ultralytics.com/models/yolov8/ (accessed on 25 October 2025).
  192. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning what you want to learn using programmable gradient information. In Computer Vision – ECCV 2024, Proceedings of the 18th European Conference, Milan, Italy, 29 September–4 October 2024; Springer: Cham, Switzerland, 2024; pp. 1–21. [Google Scholar] [CrossRef]
  193. Ultralytics. YOLOv10 vs. YOLOv9: A Comprehensive Technical Comparison. 2024. Available online: https://docs.ultralytics.com/compare/yolov10-vs-yolov9/ (accessed on 29 November 2025).
  194. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar] [CrossRef]
  195. Ultralytics. Ultralytics YOLO: GitHub Repository. GitHub. 2025. Available online: https://github.com/ultralytics/ultralytics (accessed on 25 October 2025).
  196. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  197. Vina, A. YOLO12 Explained: Real-World Applications and Use Cases. 2025. Available online: https://www.ultralytics.com/blog/yolo12-explained-real-world-applications-and-use-cases (accessed on 29 November 2025).
  198. Wang, C.; He, W.; Nie, Y.; Guo, J.; Liu, C.; Han, K.; Wang, Y. Gold YOLO: Efficient Object Detector via Gather and Distribute Mechanism. In Proceedings of the Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  199. Alif, M.A.R.; Hussain, M. YOLOv1 to YOLOv10: A comprehensive review of YOLO variants and their application in the agricultural domain. arXiv 2024, arXiv:2406.10139. [Google Scholar] [CrossRef]
  200. He, Z.; Cao, L. SOD-YOLO: Small Object Detection Network for UAV Aerial Images. IEEJ Trans. Electr. Electron. Eng. 2025, 20, 431–439. [Google Scholar] [CrossRef]
  201. Bakirci, M. Performance evaluation of low-power and lightweight object detectors for real-time monitoring in resource-constrained drone systems. Eng. Appl. Artif. Intell. 2025, 159, 111775. [Google Scholar] [CrossRef]
  202. Pudasaini, D.; Abhari, A. Scalable object detection, tracking and pattern recognition model using edge computing. In Proceedings of the 2020 Spring Simulation Conference (SpringSim), Fairfax, VA, USA, 18–21 May 2020; pp. 1–11. [Google Scholar] [CrossRef]
  203. Du, X.; Lin, T.Y.; Jin, P.; Ghiasi, G.; Tan, M.; Cui, Y.; Le, Q.V.; Song, X. Spinenet: Learning scale-permuted backbone for recognition and localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11592–11601. [Google Scholar] [CrossRef]
  204. Tan, M.; Le, Q. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  205. Liu, F.; Chen, R.; Zhang, J.; Xing, K.; Liu, H.; Qin, J. R2YOLOX: A lightweight refined anchor-free rotated detector for object detection in aerial images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5632715. [Google Scholar] [CrossRef]
  206. Lyu, Z.; Yu, T.; Pan, F.; Zhang, Y.; Luo, J.; Zhang, D.; Chen, Y.; Zhang, B.; Li, G. A survey of model compression strategies for object detection. Multimed. Tools Appl. 2024, 83, 48165–48236. [Google Scholar] [CrossRef]
  207. Han, S.; Jiang, X.; Wu, Z. An improved YOLOv5 algorithm for wood defect detection based on attention. IEEE Access 2023, 11, 71800–71810. [Google Scholar] [CrossRef]
  208. Thudumu, S.; Nguyen, H.; Du, H.; Duong, N.; Rasool, Z.; Logothetis, R.; Barnett, S.; Vasa, R.; Mouzakis, K. The M-factor: A Novel Metric for Evaluating Neural Architecture Search in Resource-Constrained Environments. arXiv 2025, arXiv:2501.17361. [Google Scholar] [CrossRef]
  209. Li, J.; Xie, C.; Wu, S.; Ren, Y. UAV-YOLOv5: A Swin-transformer-enabled small object detection model for long-range UAV images. Ann. Data Sci. 2024, 11, 1109–1138. [Google Scholar] [CrossRef]
  210. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar] [CrossRef]
  211. Zhang, H.; Liu, K.; Gan, Z.; Zhu, G.N. UAV-DETR: Efficient end-to-end object detection for unmanned aerial vehicle imagery. arXiv 2025, arXiv:2501.01855. [Google Scholar] [CrossRef]
  212. Mehta, S.; Rastegari, M. Separable self-attention for mobile vision transformers. arXiv 2022, arXiv:2206.02680. [Google Scholar] [CrossRef]
  213. Zhou, Y.; Wei, Y. UAV-DETR: An enhanced RT-DETR architecture for efficient small object detection in UAV imagery. Sensors 2025, 25, 4582. [Google Scholar] [CrossRef]
  214. Guo, H.; Wu, Q.; Wang, Y. AUHF-DETR: A Lightweight Transformer with Spatial Attention and Wavelet Convolution for Embedded UAV Small Object Detection. Remote Sens. 2025, 17, 1920. [Google Scholar] [CrossRef]
  215. Ge, X.; Qi, L.; Yan, Q.; Sun, J.; Zhu, Y.; Zhang, Y. Enhancing Real-Time Aerial Image Object Detection with High-Frequency Feature Learning and Context-Aware Fusion. Remote Sens. 2025, 17, 1994. [Google Scholar] [CrossRef]
  216. Wang, H.; Gao, J. Sf-detr: A scale-frequency detection transformer for drone-view object detection. Sensors 2025, 25, 2190. [Google Scholar] [CrossRef]
  217. Deng, J.; Shi, Z.; Zhuo, C. Energy-efficient real-time UAV object detection on embedded platforms. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst. 2019, 39, 3123–3127. [Google Scholar] [CrossRef]
  218. Galliera, R.; Suri, N. Object detection at the edge: Off-the-shelf deep learning capable devices and accelerators. Procedia Comput. Sci. 2022, 205, 239–248. [Google Scholar] [CrossRef]
  219. Elsken, T.; Metzen, J.H.; Hutter, F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019, 20, 1–21. [Google Scholar]
  220. Wu, B.; Dai, X.; Zhang, P.; Wang, Y.; Sun, F.; Wu, Y.; Tian, Y.; Vajda, P.; Jia, Y.; Keutzer, K. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10734–10742. [Google Scholar] [CrossRef]
  221. Cai, H.; Zhu, L.; Han, S. Proxylessnas: Direct neural architecture search on target task and hardware. arXiv 2018, arXiv:1812.00332. [Google Scholar] [CrossRef]
  222. Cai, H.; Gan, C.; Wang, T.; Zhang, Z.; Han, S. Once-for-all: Train one network and specialize it for efficient deployment. arXiv 2019, arXiv:1908.09791. [Google Scholar] [CrossRef]
  223. Song, X.; Xie, X.; Lv, Z.; Yen, G.G.; Ding, W.; Lv, J.; Sun, Y. Efficient evaluation methods for neural architecture search: A survey. IEEE Trans. Artif. Intell. 2024, 5, 5990–6011. [Google Scholar] [CrossRef]
  224. Kang, Y.; Zheng, B.; Shen, W. Research on Oriented Object Detection in Aerial Images Based on Architecture Search with Decoupled Detection Heads. Appl. Sci. 2025, 15, 8370. [Google Scholar] [CrossRef]
  225. Team, R. YOLO-NAS by deci achieves state-of-the-art performance on object detection using neural architecture search. Deci AI Blog 2023. [Google Scholar]
  226. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  227. van den Oord, A.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar] [CrossRef]
  228. Jia, C.; Yang, Y.; Xia, Y.; Chen, Y.T.; Parekh, Z.; Pham, H.; Le, Q.; Sung, Y.H.; Li, Z.; Duerig, T. Scaling up visual and vision-language representation learning with noisy text supervision. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 4904–4916. [Google Scholar]
  229. Gan, Z.; Li, L.; Li, C.; Wang, L.; Liu, Z.; Gao, J. Vision-language pre-training: Basics, recent advances, and future trends. Found. Trends® Comput. Graph. Vis. 2022, 14, 163–352. [Google Scholar] [CrossRef]
  230. Liu, F.; Chen, D.; Guan, Z.; Zhou, X.; Zhu, J.; Ye, Q.; Fu, L.; Zhou, J. Remoteclip: A vision language foundation model for remote sensing. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5622216. [Google Scholar] [CrossRef]
  231. Li, Y.; Guo, W.; Yang, X.; Liao, N.; He, D.; Zhou, J.; Yu, W. Toward open vocabulary aerial object detection with clip-activated student-teacher learning. In Computer Vision – ECCV 2024, Proceedings of the European Conference, Milan, Italy, 29 September–4 October 2024; Springer: Cham, Switzerland; pp. 431–448. [CrossRef]
  232. Yang, C.; An, Z.; Huang, L.; Bi, J.; Yu, X.; Yang, H.; Diao, B.; Xu, Y. Clip-kd: An empirical study of clip model distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16-22 June 2024; pp. 15952–15962. [Google Scholar] [CrossRef]
  233. Zhang, S.H.; Tang, W.C.; Wu, C.; Hu, P.; Li, N.; Zhang, L.J.; Zhang, Q.; Zhang, S.Q. TernaryCLIP: Efficiently Compressing Vision-Language Models with Ternary Weights and Distilled Knowledge. arXiv 2025, arXiv:2510.21879. [Google Scholar] [CrossRef]
  234. Cheng, T.; Song, L.; Ge, Y.; Liu, W.; Wang, X.; Shan, Y. Yolo-world: Real-time open-vocabulary object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16901–16911. [Google Scholar] [CrossRef]
  235. Kong, F.; Shan, X.; Hu, Y.; Li, J. Automated UAV Object Detector Design Using Large Language Model-Guided Architecture Search. Drones 2025, 9, 803. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.