Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (476)

Search Parameters:
Keywords = small drone detection

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 38551 KB  
Article
Tiny Object Detection via Normalized Gaussian Label Assignment and Multi-Scale Hybrid Attention
by Shihao Lin, Li Zhong, Si Chen and Da-Han Wang
Remote Sens. 2026, 18(3), 396; https://doi.org/10.3390/rs18030396 (registering DOI) - 24 Jan 2026
Abstract
The rapid development of Convolutional Neural Networks (CNNs) has markedly boosted the performance of object detection in remote sensing. Nevertheless, tiny objects typically account for an extremely small fraction of the total area in remote sensing images, rendering existing IoU-based or area-based evaluation [...] Read more.
The rapid development of Convolutional Neural Networks (CNNs) has markedly boosted the performance of object detection in remote sensing. Nevertheless, tiny objects typically account for an extremely small fraction of the total area in remote sensing images, rendering existing IoU-based or area-based evaluation metrics highly sensitive to minor pixel deviations. Meanwhile, classic detection models face inherent bottlenecks in efficiently mining discriminative features for tiny objects, leaving the task of tiny object detection in remote sensing images as an ongoing challenge in this field. To alleviate these issues, this paper proposes a tiny object detection method based on Normalized Gaussian Label Assignment and Multi-scale Hybrid Attention. Firstly, 2D Gaussian modeling is performed on the feature receptive field and the actual bounding box, using Normalized Bhattacharyya Distance for precise similarity measurement. Furthermore, a candidate sample quality ranking mechanism is constructed to select high-quality positive samples. Finally, a Multi-scale Hybrid Attention module is designed to enhance the discriminative feature extraction of tiny objects. The proposed method achieves 25.7% and 27.9% AP on the AI-TOD-v2 and VisDrone2019 datasets, respectively, significantly improving the detection capability of tiny objects in complex remote sensing scenarios. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 3rd Edition)
Show Figures

Figure 1

23 pages, 53605 KB  
Article
Multispectral Sparse Cross-Attention Guided Mamba Network for Small Object Detection in Remote Sensing
by Wen Xiang, Yamin Li, Liu Duan, Qifeng Wu, Jiaqi Ruan, Yucheng Wan and Sihan Wu
Remote Sens. 2026, 18(3), 381; https://doi.org/10.3390/rs18030381 - 23 Jan 2026
Viewed by 36
Abstract
Remote sensing small object detection remains a challenging task due to limited feature representation and interference from complex backgrounds. Existing methods that rely exclusively on either visible or infrared modalities often fail to achieve both accuracy and robustness in detection. Effectively integrating cross-modal [...] Read more.
Remote sensing small object detection remains a challenging task due to limited feature representation and interference from complex backgrounds. Existing methods that rely exclusively on either visible or infrared modalities often fail to achieve both accuracy and robustness in detection. Effectively integrating cross-modal information to enhance detection performance remains a critical challenge. To address this issue, we propose a novel Multispectral Sparse Cross-Attention Guided Mamba Network (MSCGMN) for small object detection in remote sensing. The proposed MSCGMN architecture comprises three key components: Multispectral Sparse Cross-Attention Guidance Module (MSCAG), Dynamic Grouped Mamba Block (DGMB), and Gated Enhanced Attention Module (GEAM). Specifically, the MSCAG module selectively fuses RGB and infrared (IR) features using sparse cross-modal attention, effectively capturing complementary information across modalities while suppressing redundancy. The DGMB introduces a dynamic grouping strategy to improve the computational efficiency of Mamba, enabling effective global context modeling. In remote sensing images, small objects occupy limited areas, making it difficult to capture their critical features. We design the GEAM module to enhance both global and local feature representations for small object detection. Experiments on the VEDAI and DroneVehicle datasets show that MSCGMN achieves mAP50 scores of 83.9% and 84.4%, outperforming existing state-of-the-art methods and demonstrating strong competitiveness in small object detection tasks. Full article
17 pages, 5027 KB  
Article
Symmetry-Enhanced YOLOv8s Algorithm for Small-Target Detection in UAV Aerial Photography
by Zhiyi Zhou, Chengyun Wei, Lubin Wang and Qiang Yu
Symmetry 2026, 18(1), 197; https://doi.org/10.3390/sym18010197 - 20 Jan 2026
Viewed by 150
Abstract
In order to solve the problems of small-target detection in UAV aerial photography, such as small scale, blurred features and complex background interference, this article proposes the ACS-YOLOv8s method to optimize the YOLOv8s network: notably, most small man-made targets in UAV aerial scenes [...] Read more.
In order to solve the problems of small-target detection in UAV aerial photography, such as small scale, blurred features and complex background interference, this article proposes the ACS-YOLOv8s method to optimize the YOLOv8s network: notably, most small man-made targets in UAV aerial scenes (e.g., small vehicles, micro-drones) inherently possess symmetry, a key geometric attribute that can significantly enhance the discriminability of blurred or incomplete target features, and thus symmetry-aware mechanisms are integrated into the aforementioned improved modules to further boost detection performance. The backbone network introduces an adaptive feature enhancement module, the edge and detail representation of small targets is enhanced by dynamically modulating the receptive field with deformable attention while also capturing symmetric contour features to strengthen the perception of target geometric structures; a cascaded multi-receptive field module is embedded at the end of the trunk to integrate multi-scale features in a hierarchical manner to take into account both expressive ability and computational efficiency with a focus on fusing symmetric multi-scale features to optimize feature representation; the neck is integrated with a spatially adaptive feature modulation network to achieve dynamic weighting of cross-layer features and detail fidelity and, meanwhile, models symmetric feature dependencies across channels to reduce the loss of discriminative information. Experimental results based on the VisDrone2019 data set show that ACS-YOLOv8s is superior to the baseline model in precision, recall, and mAP indicators, with mAP50 increased by 2.8% to 41.6% and mAP50:90 increased by 1.9% to 25.0%, verifying its effectiveness and robustness in small-target detection in complex drone aerial-photography scenarios. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

23 pages, 40307 KB  
Article
EFPNet: An Efficient Feature Perception Network for Real-Time Detection of Small UAV Targets
by Jiahao Huang, Wei Jin, Huifeng Tao, Yunsong Feng, Yuanxin Shang, Siyu Wang and Aibing Liu
Remote Sens. 2026, 18(2), 340; https://doi.org/10.3390/rs18020340 - 20 Jan 2026
Viewed by 118
Abstract
In recent years, unmanned aerial vehicles (UAVs) have become increasingly prevalent across diverse application scenarios due to their high maneuverability, compact size, and cost-effectiveness. However, these advantages also introduce significant challenges for UAV detection in complex environments. This paper proposes an efficient feature [...] Read more.
In recent years, unmanned aerial vehicles (UAVs) have become increasingly prevalent across diverse application scenarios due to their high maneuverability, compact size, and cost-effectiveness. However, these advantages also introduce significant challenges for UAV detection in complex environments. This paper proposes an efficient feature perception network (EFPNet) for UAV detection, developed on the foundation of the RT-DETR framework. Specifically, a dual-branch HiLo-ConvMix attention (HCM-Attn) mechanism and a pyramid sparse feature transformer network (PSFT-Net) are introduced, along with the integration of a DySample dynamic upsampling module. The HCM-Attn module facilitates interaction between high- and low-frequency information, effectively suppressing background noise interference. The PSFT-Net is designed to leverage deep-level features to guide the encoding and fusion of shallow features, thereby enhancing the model’s capability to perceive UAV texture characteristics. Furthermore, the integrated DySample dynamic upsampling module ensures efficient reconstruction and restoration of feature representations. On the TIB and Drone-vs-Bird datasets, the proposed EFPNet achieves mAP50 scores of 94.1% and 98.1%, representing improvements of 3.2% and 1.9% over the baseline models, respectively. Our experimental results demonstrate the effectiveness of the proposed method for small UAV detection. Full article
Show Figures

Figure 1

18 pages, 3987 KB  
Article
Low-Latency Autonomous Surveillance in Defense Environments: A Hybrid RTSP-WebRTC Architecture with YOLOv11
by Juan José Castro-Castaño, William Efrén Chirán-Alpala, Guillermo Alfonso Giraldo-Martínez, José David Ortega-Pabón, Edison Camilo Rodríguez-Amézquita, Diego Ferney Gallego-Franco and Yeison Alberto Garcés-Gómez
Computers 2026, 15(1), 62; https://doi.org/10.3390/computers15010062 - 16 Jan 2026
Viewed by 238
Abstract
This article presents the Intelligent Monitoring System (IMS), an AI-assisted, low-latency surveillance platform designed for defense environments. The study addresses the need for real-time autonomous situational awareness by integrating high-speed video transmission with advanced computer vision analytics in constrained network settings. The IMS [...] Read more.
This article presents the Intelligent Monitoring System (IMS), an AI-assisted, low-latency surveillance platform designed for defense environments. The study addresses the need for real-time autonomous situational awareness by integrating high-speed video transmission with advanced computer vision analytics in constrained network settings. The IMS employs a hybrid transmission architecture based on RTSP for ingestion and WHEP/WebRTC for distribution, orchestrated via MediaMTX, with the objective of achieving end-to-end latencies below one second. The methodology includes a comparative evaluation of video streaming protocols (JPEG-over-WebSocket, HLS, WebRTC, etc.) and AI frameworks, alongside the modular architectural design and prolonged experimental validation. The detection module integrates YOLOv11 models fine-tuned on the VisDrone dataset to optimize performance for small objects, aerial views, and dense scenes. Experimental results, obtained through over 300 h of operational tests using IP cameras and aerial platforms, confirmed the stability and performance of the chosen architecture, maintaining latencies close to 500 ms. The YOLOv11 family was adopted as the primary detection framework, providing an effective trade-off between accuracy and inference performance in real-time scenarios. The YOLOv11n model was trained and validated on a Tesla T4 GPU, and YOLOv11m will be validated on the target platform in subsequent experiments. The findings demonstrate the technical viability and operational relevance of the IMS as a core component for autonomous surveillance systems in defense, satisfying strict requirements for speed, stability, and robust detection of vehicles and pedestrians. Full article
Show Figures

Figure 1

24 pages, 6265 KB  
Article
On the Study of Performance Enhancement of 3D Printing and Industrial Application on Aviation Devices
by Hui-Pei Chang and Yung-Lan Yeh
Aerospace 2026, 13(1), 90; https://doi.org/10.3390/aerospace13010090 - 14 Jan 2026
Viewed by 154
Abstract
Three-dimensional printing is the most commonly used method for producing customized or mock-up products for industrial applications. In particular, aviation devices for drones usually require a high spatial resolution to satisfy the small size requirement. In practical applications of drones, the two main [...] Read more.
Three-dimensional printing is the most commonly used method for producing customized or mock-up products for industrial applications. In particular, aviation devices for drones usually require a high spatial resolution to satisfy the small size requirement. In practical applications of drones, the two main tasks are inspection and detection. However, the working environment is often filled with flammable gases, such as natural gas or petroleum gas. Thus, the parts of drones that can easily produce an electrical spark, such as electronic connectors, should be specially protected. In this study, atmosphere control was applied to enhance the printing performance and manufacture of anti-explosion devices. The results demonstrate that atmosphere control can efficiently improve the print quality and that the print resolution of a commercial 3D printer can be enhanced to reach the mm scale. In the anti-pressure testing via a high-pressure smoke experiment, the manufactured anti-explosion devices for drones showed an appropriate intrinsic safety level, suggesting that they can be used in drones used for daily inspections of pipelines in petrochemical plants. The two main contributions of this study are the development of a practical method for improving FDM 3D printers and an anti-explosion device for drones. Full article
Show Figures

Figure 1

19 pages, 2336 KB  
Article
A Lightweight Upsampling and Cross-Modal Feature Fusion-Based Algorithm for Small-Object Detection in UAV Imagery
by Jianglei Gong, Zhe Yuan, Wenxing Li, Weiwei Li, Yanjie Guo and Baolong Guo
Electronics 2026, 15(2), 298; https://doi.org/10.3390/electronics15020298 - 9 Jan 2026
Viewed by 206
Abstract
Small-object detection in UAV remote sensing faces common challenges such as tiny target size, blurred features, and severe background interference. Furthermore, single imaging modalities exhibit limited representation capability in complex environments. To address these issues, this paper proposes CTU-YOLO, a UAV-based small-object detection [...] Read more.
Small-object detection in UAV remote sensing faces common challenges such as tiny target size, blurred features, and severe background interference. Furthermore, single imaging modalities exhibit limited representation capability in complex environments. To address these issues, this paper proposes CTU-YOLO, a UAV-based small-object detection algorithm built upon cross-modal feature fusion and lightweight upsampling. The algorithm incorporates a dynamic and adaptive cross-modal feature fusion (DCFF) module, which achieves efficient feature alignment and fusion by combining frequency-domain analysis with convolutional operations. Additionally, a lightweight upsampling module (LUS) is introduced, integrating dynamic sampling and depthwise separable convolution to enhance the recovery of fine details for small objects. Experiments on the DroneVehicle and LLVIP datasets demonstrate that CTU-YOLO achieves 73.9% mAP on DroneVehicle and 96.9% AP on LLVIP, outperforming existing mainstream methods. Meanwhile, the model possesses only 4.2 MB parameters and 13.8 GFLOPs computational cost, with inference speeds reaching 129.9 FPS on DroneVehicle and 135.1 FPS on LLVIP. This exhibits an excellent lightweight design and real-time performance while maintaining high accuracy. Ablation studies confirm that both the DCFF and LUS modules contribute significantly to performance gains. Visualization analysis further indicates that the proposed method can accurately preserve the structure of small objects even under nighttime, low-light, and multi-scale background conditions, demonstrating strong robustness. Full article
(This article belongs to the Special Issue AI-Driven Image Processing: Theory, Methods, and Applications)
Show Figures

Figure 1

25 pages, 7611 KB  
Article
BFRI-YOLO: Harmonizing Multi-Scale Features for Precise Small Object Detection in Aerial Imagery
by Xue Zeng, Shenghong Fang and Qi Sun
Electronics 2026, 15(2), 297; https://doi.org/10.3390/electronics15020297 - 9 Jan 2026
Viewed by 220
Abstract
Identifying minute targets within UAV-acquired imagery continues to pose substantial technical hurdles, primarily due to blurred boundaries, scarce textural details, and drastic scale variations amidst complex backgrounds. In response to these limitations, this paper proposes BFRI-YOLO, an enhanced architecture based on the YOLOv11n [...] Read more.
Identifying minute targets within UAV-acquired imagery continues to pose substantial technical hurdles, primarily due to blurred boundaries, scarce textural details, and drastic scale variations amidst complex backgrounds. In response to these limitations, this paper proposes BFRI-YOLO, an enhanced architecture based on the YOLOv11n baseline. The framework is built upon four synergistic components designed to achieve high-precision localization and robust feature representation. First, we construct a Balanced Adaptive Feature Pyramid Network (BAFPN) that utilizes a resolution-aware attention mechanism to promote bidirectional interaction between deep and shallow features. This is complemented by incorporating the Receptive Field Convolutional Block Attention Module (RFCBAM) to refine the backbone network. By constructing the C3K2_RFCBAM block, we effectively enhance the feature representation of small objects across diverse receptive fields. To further refine the prediction phase, we develop a Four-Shared Detail Enhancement Detection Head (FSDED) to improve both efficiency and stability. Finally, regarding the loss function, we formulate the Inner-WIoU strategy by integrating auxiliary bounding boxes with dynamic focusing mechanisms to ensure precise target localization. The experimental results on the VisDrone2019 benchmark demonstrate that our method secures mAP@0.5 and mAP@0.5:0.95 scores of 42.1% and 25.6%, respectively, outperforming the baseline by 8.8% and 6.2%. Extensive tests on the TinyPerson and DOTA1.0 datasets further validate the robust generalization capability of our model, confirming that BFRI-Yolo strikes a superior balance between detection accuracy and computational overhead in aerial scenes. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

22 pages, 5463 KB  
Article
SRG-YOLO: Star Operation and Restormer-Based YOLOv11 via Global Context for Vehicle Object Detection
by Wei Song, Junying Min and Jiaqi Zhao
Automation 2026, 7(1), 15; https://doi.org/10.3390/automation7010015 - 7 Jan 2026
Viewed by 171
Abstract
Recently, these conventional object detection methods have certain defects that must be overcome, such as insufficient detection accuracy in complex scenes and low computational efficiency. Then, this paper proposes a Star operation and Restormer-based YOLOv11 model that leverages global context for vehicle detection [...] Read more.
Recently, these conventional object detection methods have certain defects that must be overcome, such as insufficient detection accuracy in complex scenes and low computational efficiency. Then, this paper proposes a Star operation and Restormer-based YOLOv11 model that leverages global context for vehicle detection (SRG-YOLO), which aims to enhance both detection accuracy and efficiency in complex environments. Firstly, during the optimization of YOLOv11n architecture, a Star block is introduced. By enhancing non-linear feature representation, this Star block improves the original C3K2 module, thereby strengthening multi-scale feature fusion and consequently boosting detection accuracy in complex scenarios. Secondly, for the detection heads of YOLOv11n, Restormer is incorporated via the improved C3K2 module to explicitly leverage spatial prior information, optimize the self-attention mechanism, and augment long-range pixel dependencies of YOLOv11n. This integration not only reduces computational complexity but also improves detection precision and overall efficiency through more refined feature modeling. Thirdly, a Context-guided module is integrated to enhance the ability to capture object details using global context. In complex backgrounds, it effectively combines local features with their contextual information, substantially improving the detection robustness of YOLOv11n. Finally, experiments on the VisDrone2019, KITTI, and UA-DETRAC datasets illustrate that SRG-YOLO achieves superior vehicle detection accuracy in complex scenes compared to conventional methods, with particular advantages in small object detection. Full article
(This article belongs to the Collection Automation in Intelligent Transportation Systems)
Show Figures

Figure 1

23 pages, 3153 KB  
Article
SSCW-YOLO: A Lightweight and High-Precision Model for Small Object Detection in UAV Scenarios
by Zhuolun He, Rui She, Bo Tan, Jiajian Li and Xiaolong Lei
Drones 2026, 10(1), 41; https://doi.org/10.3390/drones10010041 - 7 Jan 2026
Viewed by 473
Abstract
To address the problems of missed and false detections caused by insufficient feature quality in small object detection from UAV perspectives, this paper proposes a UAV small object detection algorithm based on YOLOv8 feature optimization. A spatial cosine convolution module is introduced into [...] Read more.
To address the problems of missed and false detections caused by insufficient feature quality in small object detection from UAV perspectives, this paper proposes a UAV small object detection algorithm based on YOLOv8 feature optimization. A spatial cosine convolution module is introduced into the backbone network to optimize spatial features, thereby alleviating the problem of small object feature loss and improving the detection accuracy and speed of the model. An improved C2f_SCConv feature fusion module is employed for feature integration, which effectively reduces feature redundancy in spatial and channel dimensions, thereby lowering model complexity and computational cost. Meanwhile, the WIoU loss function is used to replace the original CIoU loss function, reducing the interference of geometric factors in anchor box regression, enabling the model to focus more on low-quality anchor boxes, and enhancing its small object detection capability. Ablation and comparative experiments on the VisDrone dataset validate the effectiveness of the proposed algorithm for small object detection from UAV perspectives, while generalization experiments on the DOTA and SSDD datasets demonstrate that the algorithm possesses strong generalization performance. Full article
Show Figures

Figure 1

40 pages, 2728 KB  
Article
From Manned to Unmanned Helicopters: A Transformer-Driven Cross-Scale Transfer Learning Framework for Vibration-Based Anomaly Detection
by Geuncheol Jang and Yongjin Kwon
Actuators 2026, 15(1), 38; https://doi.org/10.3390/act15010038 - 6 Jan 2026
Viewed by 296
Abstract
Unmanned helicopters play a critical role in various fields including defense, disaster response, and infrastructure inspection. Military platforms such as the MQ-8C Fire Scout represent high-value assets exceeding $40 million per unit including development costs, particularly when compared to expendable multicopter drones costing [...] Read more.
Unmanned helicopters play a critical role in various fields including defense, disaster response, and infrastructure inspection. Military platforms such as the MQ-8C Fire Scout represent high-value assets exceeding $40 million per unit including development costs, particularly when compared to expendable multicopter drones costing approximately $500–2000 per unit. Unexpected failures of these high-value assets can lead to substantial economic losses and mission failures, making the implementation of Health and Usage Monitoring Systems (HUMS) essential. However, the scarcity of failure data in unmanned helicopters presents significant challenges for HUMS development, while the economic feasibility of investing resources comparable to manned helicopter programs remains questionable. This study presents a novel cross-scale transfer learning framework for vibration-based anomaly detection in unmanned helicopters. The framework successfully transfers knowledge from a source domain (Airbus large manned helicopter) using publicly available data to a target domain (Stanford small RC helicopter), achieving excellent anomaly detection performance without labeled target domain data. The approach consists of three key processes. First, we developed a multi-task learning transformer model achieving an F-β score of 0.963 (β = 0.3) using only Airbus vibration data. Second, we applied CORAL (Correlation Alignment) domain adaptation techniques to reduce the distribution discrepancy between source and target domains by 79.7%. Third, we developed a Control Effort Score (CES) based on control input data as a proxy labeling metric for 20 flight maneuvers in the target domain, achieving a Spearman correlation coefficient ρ of 0.903 between the CES and the Anomaly Index measured by the transfer-learned model. This represents a 95.5% improvement compared to the non-transfer learning baseline of 0.462. Full article
Show Figures

Figure 1

22 pages, 2359 KB  
Review
Airport Ground-Based Aerial Object Surveillance Technologies for Enhanced Safety: A Systematic Review
by Joel Samu and Chuyang Yang
Drones 2026, 10(1), 22; https://doi.org/10.3390/drones10010022 - 31 Dec 2025
Viewed by 408
Abstract
Airport airspace safety is increasingly threatened by small, unmanned aircraft systems and wildlife that traditional radar cannot detect. While earlier reviews addressed general counter-UAS techniques, individual sensors, or the detection of single objects such as birds or drones, none has systematically reviewed airport-based, [...] Read more.
Airport airspace safety is increasingly threatened by small, unmanned aircraft systems and wildlife that traditional radar cannot detect. While earlier reviews addressed general counter-UAS techniques, individual sensors, or the detection of single objects such as birds or drones, none has systematically reviewed airport-based, multi-sensor surveillance strategies through a safety-theoretical lens. A systematic review, performed according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 statement, synthesized recent research on fixed, ground-based aerial detection capabilities for small aerial hazards, specifically unmanned aircraft systems (sUAS) and avian targets, within operational airport environments. Searches targeted English-language, peer-reviewed articles from 2016 through 2025 in Web of Science and Scopus. Due to methodological heterogeneity across sensor technologies, a narrative synthesis was executed. The review of thirty-six studies, analyzed through Reason’s Swiss Cheese Model and Endsley’s Situational Awareness framework, found that only layered multi-sensor fusion architectures effectively address detection gaps for Low-Slow-Small (LSS) threats. Based on these findings, the review proposes seamless integration with Air Traffic Management (ATM) and UAS Traffic Management (UTM) systems through standardized data-exchange interfaces, complemented by theoretically grounded risk-based deployment strategies aligning surveillance technology tiers with operational risk profiles, from basic Remote ID receivers in low-risk rural environments to comprehensive multi-sensor fusion at high-density hubs, major airports, and urban vertiports. Full article
Show Figures

Figure 1

28 pages, 19143 KB  
Article
DAE-YOLO: Remote Sensing Small Object Detection Method Integrating YOLO and State Space Models
by Bing Li, Yongtao Kang, Yao Ding, Shaopeng Li, Zhili Zhang and Decao Ma
Remote Sens. 2026, 18(1), 109; https://doi.org/10.3390/rs18010109 - 28 Dec 2025
Viewed by 435
Abstract
Small object detection in remote sensing images provides significant value for urban monitoring, aerospace reconnaissance, and other fields. However, detection accuracy still faces multiple challenges including limited target information, weak feature representation, and complex backgrounds. This research aims to improve the performance of [...] Read more.
Small object detection in remote sensing images provides significant value for urban monitoring, aerospace reconnaissance, and other fields. However, detection accuracy still faces multiple challenges including limited target information, weak feature representation, and complex backgrounds. This research aims to improve the performance of the YOLO11 model for small object detection in remote sensing imagery by addressing key issues in long-distance spatial dependency modeling, multi-scale feature adaptive fusion, and computational efficiency. We constructed a specialized Remote Sensing Airport-Plane Detection (RS-APD) dataset and used the public VisDrone2019 dataset for generalization verification. Based on the YOLO11 architecture, we proposed the DAE-YOLO model with three innovative modules: Dynamic Spatial Sequence Module (DSSM) for enhanced long-distance spatial dependency capture; Adaptive Multi-scale Feature Enhancement (AMFE) for multi-scale feature adaptive receptive field adjustment; and Efficient Dual-level Attention Mechanism (EDAM) to reduce computational complexity while maintaining feature expression capability. Experimental results demonstrate that compared to the baseline YOLO11, our proposed model improved mAP50 and mAP50:95 on the RS-APD dataset by 2.1% and 2.5%, respectively, with APs increasing by 2.8%. This research provides an efficient and reliable small object detection solution for remote sensing applications. Full article
Show Figures

Graphical abstract

27 pages, 5048 KB  
Article
MCB-RT-DETR: A Real-Time Vessel Detection Method for UAV Maritime Operations
by Fang Liu, Yongpeng Wei, Aruhan Yan, Tiezhu Cao and Xinghai Xie
Drones 2026, 10(1), 13; https://doi.org/10.3390/drones10010013 - 27 Dec 2025
Viewed by 412
Abstract
Maritime UAV operations face challenges in real-time ship detection. Complex ocean backgrounds, drastic scale variations, and prevalent distant small targets create difficulties. We propose MCB-RT-DETR, a real-time detection transformer enhanced by multi-component boosting. This method builds upon the RT-DETR architecture. It significantly improves [...] Read more.
Maritime UAV operations face challenges in real-time ship detection. Complex ocean backgrounds, drastic scale variations, and prevalent distant small targets create difficulties. We propose MCB-RT-DETR, a real-time detection transformer enhanced by multi-component boosting. This method builds upon the RT-DETR architecture. It significantly improves detection under wave interference, lighting changes, and scale differences. Key innovations address these challenges. An Orthogonal Channel Attention (Ortho) mechanism preserves high-frequency edge details in the backbone network. Receptive Field Attention Convolution (RFAConv) enhances robustness against background clutter. A Small Object Detail Enhancement Pyramid (SOD-EPN) strengthens small-target representation. SOD-EPN combines SPDConv with multi-scale CSP-OmniKernel transformations. The neck network integrates ultra-lightweight DySample upsampling. This enables content-aware sampling for precise multi-scale localization. The method maintains high computational efficiency. Experiments on the SeaDronesSee dataset show significant improvements. MCB-RT-DETR achieves 82.9% mAP@0.5 and 49.7% mAP@0.5:0.95. These correspond to improvements of 4.5% and 3.4% relative to the baseline model. Inference speed maintains 50 FPS for real-time processing. The outstanding performance in cross-dataset tests further validates the algorithm’s strong generalization capability on DIOR remote sensing images and VisDrone2019 aerial scenes. The method provides a reliable visual perception solution for autonomous maritime UAV operations. Full article
Show Figures

Figure 1

23 pages, 4261 KB  
Article
Efficient Drone Detection Using Temporal Anomalies and Small Spatio-Temporal Networks
by Abhijit Mahalanobis and Amadou Tall
Sensors 2026, 26(1), 170; https://doi.org/10.3390/s26010170 - 26 Dec 2025
Viewed by 349
Abstract
Detecting small drones in Infrared (IR) sequences poses significant challenges due to their low visibility, low resolution, and complex cluttered backgrounds. These factors often lead to high false alarm and missed detection rates. This paper frames drone detection as a spatio-temporal anomaly detection [...] Read more.
Detecting small drones in Infrared (IR) sequences poses significant challenges due to their low visibility, low resolution, and complex cluttered backgrounds. These factors often lead to high false alarm and missed detection rates. This paper frames drone detection as a spatio-temporal anomaly detection problem and proposes a remarkably lightweight pipeline solution (well-suited for edge applications), by employing a statistical temporal anomaly detector (known as the temporal Reed Xiaoli (TRX) algorithm), in parallel with a light-weight convolutional neural network known as the TCRNet. While the TRX detector is unsupervised, the TCRNet is trained to discriminate between drones and clutter using spatio-temporal patches (or chips). The confidence maps from both modules are additively fused to localize drones in video imagery. We compare our method, dubbed TRX-TCRnet, to other state-of-the-art drone detection techniques using the Detection of Aircraft Under Background (DAUB) dataset. Our approach achieves exceptional computational efficiency with only 0.17 GFLOPs with 0.83 M parameters, outperforming methods that require 145–795 times more computational resources. At the same time, the TRX–TCRNet achieves one of the highest detection accuracies (mAP50 of 97.40) while requiring orders of magnitude fewer computational resources than competing methods, demonstrating unprecedented efficiency–performance trade-offs for real-time applications. Experimental results, including ROC and PR curves, confirm the framework’s exceptional suitability for resource-constrained environments and embedded systems. Full article
(This article belongs to the Special Issue Signal Processing and Machine Learning for Sensor Systems)
Show Figures

Figure 1

Back to TopTop