Next Article in Journal
Exploitation of OCO-3 Satellite Data to Analyse Carbon Dioxide Emissions from the Mt. Etna Volcano
Previous Article in Journal
An Innovative Approach for Calibrating Hydrological Surrogate Deep Learning Models
Previous Article in Special Issue
LocaLock: Enhancing Multi-Object Tracking in Satellite Videos via Local Feature Matching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SEMA-YOLO: Lightweight Small Object Detection in Remote Sensing Image via Shallow-Layer Enhancement and Multi-Scale Adaptation

1
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
2
The State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2025, 17(11), 1917; https://doi.org/10.3390/rs17111917
Submission received: 11 April 2025 / Revised: 22 May 2025 / Accepted: 30 May 2025 / Published: 31 May 2025

Abstract

:
Small object detection remains a challenge in the remote sensing field due to feature loss during downsampling and interference from complex backgrounds. A novel network, termed SEMA-YOLO, is proposed in this paper as an enhanced YOLOv11-based framework incorporating three technical advancements. By fundamentally reducing information loss and incorporating a cross-scale feature fusion mechanism, the proposed framework significantly enhances small object detection performance. First, the Shallow Layer Enhancement (SLE) strategy reduces backbone depth and introduces small-object detection heads, thereby increasing feature map size and improving small object detection performance. Then, the Global Context Pooling-enhanced Adaptively Spatial Feature Fusion (GCP-ASFF) architecture is designed to optimize cross-scale feature interaction across four detection heads. Finally, the RFA-C3k2 module, which integrates Receptive Field Adaptation (RFA) with the C3k2 structure, is introduced to achieve more refined feature extraction. SEMA-YOLO demonstrates significant advantages in complex urban environments and dense target areas, while its generalization capability meets the detection requirements across diverse scenarios. The experimental results show that SEMA-YOLO achieves mAP50 scores of 72.5% on the RS-STOD dataset and 61.5% on the AI-TOD dataset, surpassing state-of-the-art models.

1. Introduction

Object detection is a crucial research topic in the field of remote sensing and has been widely applied in various domains, including military applications [1], agriculture [2], urban planning [3], environmental monitoring [4], and traffic management [5]. These application scenarios cover a wide range of fields, but involve a variety of object types. According to the definition of small objects in MS-COCO [6], objects with a size smaller than 32 × 32 pixels are considered small objects. In particular, some small objects in remote sensing images are often difficult to identify accurately. The enhancement of the detection accuracy of small objects in remote sensing images has become a key research direction in the remote sensing field [7]. In recent years, the resolution of remote sensing images has improved significantly. In the optical domain, the rapid development of high-resolution (HR) satellites and unmanned aerial vehicle (UAV) technology has enabled remote sensing images to achieve meter-level or even sub-meter-level resolution. For example, the WorldView-3 satellite provides a panchromatic resolution of 0.31 m and a multispectral resolution of 1.24 m [8], while UAVs such as the DJI Phantom 4 RTK [9] can achieve centimeter-level resolution when flying at low altitudes. In the synthetic aperture radar (SAR) domain, the Gaofen-3 03 satellite offers a C-band resolution of up to 1 m [10], and the TerraSAR-X satellite can achieve a ground resolution of 1 m in SpotLight mode [11]. The widespread availability of high-resolution remote sensing imagery provides a robust basis for detecting and analyzing small objects.
However, small object detection in remote sensing images still encounters numerous challenges Figure 1. Firstly, it is the complexity of remote sensing imagery backgrounds. Due to imaging conditions, remote sensing images often suffer from geometric distortions, radiometric variations, shadows, and noise, which not only interfere with object recognition but also lead to false positives or missed detections [12]. Secondly, it is a limited number of pixels of small objects. This makes their feature representation inadequate for traditional feature extraction methods, thereby increasing detection difficulty [13]. Thirdly, the dense distribution of small objects. In scenarios where small objects are densely distributed, occlusion and overlap among objects can severely impact detection accuracy, making it challenging to distinguish and accurately locate objects [14]. These challenges hinder the performance of small object detection in HR remote sensing imagery, necessitating the development of more advanced algorithms for effective solutions.
To address these challenges, this paper implements small object detection in HR remote sensing images based on the YOLOv11 framework. Specifically, the proposed approach introduces the following contributions:
SEMA-YOLO, a lightweight and high-accuracy YOLOv11-based model, is proposed. This model demonstrates strong performance on both the RS-STOD dataset and the AI-TOD [15] dataset.
Shallow Layer Enhancement (SLE): To improve small object detection in YOLOv11, an extra tiny detection head is added, and the backbone output is adjusted from P5 to P4. This reduces parameters while enhancing localization and classification, especially in dense and complex scenes.
Global Context Pooling and Enhanced Adaptively Spatial Feature Fusion (GCP-ASFF): Traditional FPNs often lose fine details of small objects. To solve this, ASFF [16] dynamically learns spatial weights to optimize multiscale feature fusion and reduce conflicts. However, ASFF lacks global context awareness, so a Global Context Pooling (GCP) module is introduced to enhance semantic consistency and improve feature integration.
Receptive Field Attention Convolution Module (RFA): This lightweight module enhances detection under high IoU thresholds by dynamically generating receptive field attention weights, overcoming fixed kernel limitations [17]. Compared to YOLOv11’s C3k2 module, RFA improves fine-grained feature extraction and sensitivity to small objects.
The remainder of this paper is structured as follows. Section 2 reviews related works on small object detection in remote sensing images. Section 3 provides a detailed description of the proposed methodology. Section 4 presents the experimental setup, results, and analysis. Section 5 discusses further aspects of the proposed model. Section 6 concludes the paper.

2. Related Works

This section reviews recent progress in small object detection through two perspectives: task-specific methods designed to tackle the inherent challenges of such targets and mainstream detection frameworks. The core ideas of these approaches will be thoroughly examined, and their strengths and weaknesses in small object detection will be evaluated.

2.1. Task-Specific Methods

2.1.1. Multi-Scale Feature

Small objects, characterized by their limited resolution and low visibility, often exhibit insufficient feature representation in detection tasks. To mitigate this issue, multi-scale feature extraction and fusion have emerged as a critical strategy, facilitating the extraction of precise spatial features from the shallow layers of HR inputs, while concurrently acquiring abstract semantic representations from the deeper, lower-resolution (LR) layers.
In 2017, Lin et al. proposed FPN [18], which combines high-level semantic information with low-level spatial information to create multi-scale feature representations. This significantly improves the detection of multi-scale objects, especially small ones. FPN is a foundational structure in many modern object detection frameworks, and numerous variants have been proposed to further refine feature processing and improve detection performance. PANet [19] introduces a bottom-up pathway to strengthen feature propagation, improving localization and segmentation. NAS-FPN [20] leverages automated architecture search to optimize feature selection and fusion. BiFPN [21] refines this process by introducing weighted connections for more efficient multi-scale information flow. Recursive-FPN [22] incorporates extra feedback connections from FPN into the bottom-up backbone layers. AFPN [23] begins by fusing adjacent low-level features, followed by a progressive inclusion of higher-level features to ease the semantic divergence across non-adjacent levels. In the YOLO algorithm, YOLOv3 [24] performs multi-scale predictions by introducing parallel branches, where HR features specialize in detecting small objects, improving detection accuracy across different scales. YOLOv4 [25] further enhances this by incorporating Path Aggregation Network (PAN), which then became a standard component in the YOLO series for strengthening feature fusion. In Transformer-based detectors, FPT [26], as one of the earlier implementations, enables full feature interaction across space and scales, enhancing feature pyramids with richer contexts. Similarly, DNTR [27] integrates FPN through its DeNoising FPN (DNFPN) module, which utilizes contrastive learning to mitigate noise during the fusion of multi-scale features, improving tiny object detection. Most of these variants, such as PANet, NAS-FPN, and BiFPN, focus on improving multi-scale feature interactions to enhance model performance. However, this inevitably increases the number of parameters, and these methods often lack dedicated optimizations for small object detection.

2.1.2. Super-Resolution

Super-resolution (SR) techniques have been widely explored to enhance image quality by reconstructing HR images from LR counterparts. Given that small objects often occupy only a few pixels in an image, SR has been considered a promising approach to improve small object detection by enhancing fine-grained details. Many SR methods, such as GAN [28] and RCAN [29], have been applied to object detection.
Several studies have explored different ways to integrate SR with object detection. SuperYOLO [30] enhances small object detection in remote sensing images by combining multimodal fusion and SR-assisted learning while keeping computational cost low. Jin et al. [31] proposed an SRD network combined with an improved Faster R-CNN to address the challenges of pedestrian detection in low-quality images. SRCGAN-RFA-YOLO [32] integrates cyclic generative adversarial networks (CycleGAN) with residual feature aggregation for image SR and combines it with YOLOv3 to enhance small object detection in remote sensing images. Truong et al. [33] demonstrated that image SR significantly improves object detection performance on LR images by training a custom RCAN and evaluating its impact on detection tasks. Despite the remarkable success of detection methods combined with image SR, they often introduce significant model complexity and require additional computational overhead.

2.1.3. Context Based

Context-based methods in small object detection focus on utilizing the surrounding information of an object to improve detection accuracy. These approaches recognize that small objects often become more distinguishable when considering the spatial, semantic, and relational context within an image. By incorporating features from the local and global environment, context-based techniques enhance the recognition of small objects, addressing the limitations of traditional models that typically fail to capture subtle cues that aid in detection. As a result, context-based strategies have shown great promise in improving the robustness and reliability of small object detection systems. Several context-based methods have been proposed to further improve small object detection by utilizing different aspects of contextual information.
Lim et al. [34] proposed a context-based object detection method that improves small object detection by incorporating multi-scale features and an attention mechanism. PyramidBox [35] fully exploits multi-level contextual information to effectively detect blurry and indistinct faces that might otherwise be lost in the background. SCDNet [36] improves detection in densely populated scenes by isolating contextual cues using a specialized scene classification branch and integrating them via a streamlined fusion mechanism and a foreground refinement module. FS-SSD [37] makes use of implicit spatial relationships by analyzing the distances between objects, refining the detection of instances with initially low confidence scores. CAB Net [38] constructs HR semantic feature maps by introducing context-aware blocks with pyramidal dilated convolutions, and attaches them to a shallow backbone to preserve spatial detail while incorporating multi-level contextual information. Tong et al. [39] introduced a context decoupling strategy that separately feeds semantic and spatial features into the classification and localization branches, enabling more precise detection through task-specific contextual information.

2.2. Mainstream Frameworks

2.2.1. YOLO Frameworks

Currently, deep learning-based object detection methods can be categorized into two main types: one-stage and two-stage detection methods. YOLO (You Only Look Once) belongs to the former category. Among various one-stage detection approaches, YOLO maintains high accuracy while achieving real-time object detection.
Since the introduction of YOLOv1 [40], the YOLO algorithm has evolved significantly. YOLOv2 [41] introduced batch normalization and data augmentation. YOLOv3 [24] integrated FPN for better multi-scale detection. YOLOv4 [25] added Spatial Pyramid Pooling (SPP) and PAN to boost speed and accuracy. Subsequent versions, including YOLOv5 [42], adopted the PyTorch 2.7.0 framework, making the model more accessible and customizable. YOLOv6 [43] introduced the RepVGG module and a hybrid channel strategy. YOLOv7 [44] leveraged the Extended Efficient Layer Aggregation Network (E-ELAN) to enhance information flow between layers, improving efficiency and effectiveness. YOLOv8 [45], developed by Ultralytics, provided a scalable version to accommodate diverse applications. YOLOv9 [46] further improved performance by optimizing programmable gradient information (PGI) to enhance inter-layer information flow. YOLOv10 [47] eliminated the need for Non-Maximum Suppression (NMS) and introduced lightweight classification heads, spatial-channel decoupled scaling, and rank-based block designs. YOLOv11 replaced the C2F module in YOLOv8 with the more efficient C3K2 module and introduced the C2PSA module. YOLOv12 [48] introduced the attention-centric real-time object detector, incorporating an Attention-Aware (A2) module and a Residual Efficient Layer Aggregation Network (R-ELAN) to address optimization challenges associated with attention mechanisms, especially in large models.
Due to its real-time performance, high accuracy, and adaptability, the YOLO algorithm has been widely applied in the remote sensing field. For example, Qiu et al. [49] proposed ASFF-YOLOv5, an improved YOLOv5 network with multiscale feature fusion, designed for multi-element detection in UAV road traffic imagery. Lin et al. [50] introduced YOLO-DA, an efficient YOLO-based detector tailored to overcome detection challenges posed by high inter-class similarity and complex backgrounds in optical remote sensing images. Li et al. [51] developed RSI-YOLO, a YOLOv5-based remote sensing object detection model incorporating attention mechanisms, enhanced feature fusion structures, additional small object detection layers, and optimized loss functions, significantly improving detection precision and recall rates. Xie et al. [52] introduced CSPPartial-YOLO, a lightweight YOLO-based model that addresses the computational complexity and large model size issues commonly encountered in deep learning-based object detection.
Given YOLO’s outstanding performance in remote sensing object detection, this study selects YOLOv11 as the foundational framework for small object detection in HR remote sensing images.

2.2.2. Other Object Detection Frameworks

Object detection frameworks can be categorized from different perspectives. From the detection stage perspective, they can be broadly divided into one-stage and two-stage methods. One-stage methods directly extract features from images and predict object bounding boxes, prioritizing detection speed. In contrast, two-stage methods first generate candidate regions and then perform classification and regression, typically achieving higher accuracy. From the architectural perspective, object detection models can be classified into CNN-based and Transformer-based approaches. CNN-based models have been widely used in object detection, while Transformer-based models leverage self-attention mechanisms to capture global dependencies and enhance detection performance.
In two-stage methods, Faster R-CNN incorporates a Region Proposal Network (RPN) to improve detection speed, while Cascade R-CNN refines bounding box quality through multi-stage regression, leading to higher precision. Several approaches have also been proposed to enhance the detection of small objects in two-stage frameworks, such as [53,54,55]. However, two-stage object detectors generally suffer from high computational complexity and slow inference speed. Among one-stage methods, apart from YOLO, RetinaNet introduces focal loss to address class imbalance and improve small object detection. CenterNet reformulates object detection as a keypoint detection problem, reducing redundant bounding boxes and enhancing efficiency. Other one-stage approaches have also been developed to further improve accuracy and robustness, particularly for small object detection. FCOS [56] introduces an anchor-free and proposal-free fully convolutional one-stage object detection framework, simplifying the detection process while achieving superior accuracy. RFLA [57] introduces a Gaussian receptive field-based label assignment strategy to improve tiny object detection, addressing the limitations of anchor-based and anchor-free methods by reducing outlier tiny-sized ground truth samples and achieving more balanced learning. Kang et al. [58] proposed aligned matching to improve small object detection in SSD, achieving better performance without compromising large object detection or increasing model complexity.
The introduction of the Transformer [59] has established a unified framework bridging computer vision and natural language processing. DETR [60] stands as a landmark achievement in applying the Transformer to object detection within computer vision. However, compared to CNN-based models, DETR faces two significant challenges: slower convergence and weaker performance on small objects. Consequently, subsequent DETR-based models have focused their efforts on addressing these two critical issues. Deformable DETR [61] replaces the standard Transformer attention with deformable attention modules, achieving better performance (especially on small objects) and 10× faster convergence compared to DETR. DN-DETR [62] accelerates training by reconstructing noised ground-truth boxes, reducing bipartite graph matching difficulty, and improving performance. DINO [63] enhances DETR with contrastive denoising training, mixed query selection, and look-forward-twice box prediction, significantly boosting detection accuracy. Liu et al. proposed two models, DNTR [27] and DQ-DETR [64]. DNTR introduces the RCNN structure to DETR, while DQ-DETR optimizes the design of dynamic queries, both achieving significant improvements in tiny object detection. RT-DETR [65] designs an efficient hybrid encoder and proposes uncertainty-minimal query selection, achieving better speed and accuracy than contemporary YOLO and DETR-based models. Although DETR-based models have made strides in addressing real-time performance and small object detection, no existing model has successfully balanced both aspects. Specifically, RT-DETR exhibits suboptimal performance in small object detection, while other DETR variants still fall short of the real-time efficiency achieved by the YOLO series. Furthermore, the Transformer-based architecture of DETR models inherently results in larger model sizes, posing significant challenges for lightweight optimization.

3. Proposed Method

3.1. Overview

This study proposes SEMA-YOLO, an enhanced network based on YOLOv11, designed to improve small object detection in HR remote sensing images. It employs SLE to preserve low-level features and introduces a tiny detection head for better detail capture. Additionally, GCP-ASFF integrates global context information to reduce false and missed detections. Finally, RFA is applied to emphasize critical features, ensuring strong performance under high IoU thresholds. Figure 2 illustrates the complete structure of the proposed model.

3.2. Shallow Layer Enhancement

Increasing the feature map size is an effective method for improving small object detection performance, but it presents two main challenges. Firstly, reducing convolutional layers or downsampling operations to enlarge the feature map may diminish the model’s capacity to capture semantic features, potentially impairing detection accuracy. Secondly, increasing the feature map size significantly raises computational costs. To address these challenges, an SLE strategy is proposed, which is specifically designed for networks with a CNN + PAN backbone. This strategy improves detection accuracy with only a slight increase in computational costs and a reduction in the number of parameters, achieving a balance between model size and performance.
The original YOLOv11 adopts a PAN structure with three detection heads corresponding to P3 (1/8 of the input image), P4 (1/16), and P5 (1/32). However, for small object detection tasks, the P5 feature map struggles to capture small object features effectively due to its limited spatial resolution. To enhance the model’s perception capability, an additional P2 detection head is introduced, which has a feature map with a side length of 1/4 of the input image. This P2 feature map is obtained by applying two consecutive upsampling operations to the P4 feature map, while maintaining the standard detection branch design in YOLO. By directly utilizing shallow features for detection, this modification improves the recognition of small objects.
To further enhance detection, the backbone network is modified by terminating convolutional operations at the P4 layer and eliminating unnecessary downsampling operations. This adjustment preserves more low-level features and minimizes the loss of semantic information. Despite the reduction in backbone depth, the P5 detection head is retained to ensure multi-scale detection capability. Specifically, the P4 feature map undergoes processing by a C2PSA module and is separately downsampled to achieve the same spatial resolution as the P5-level feature map. The two downsampled feature maps are then fused using a Concat operation to create an enhanced P5 feature map. This modification reduces computational costs while maintaining detection performance across multiple scales. Compared to the original YOLOv11n network, the SLE-enhanced model reduces the number of parameters from 2.583 M to 2.075 M, achieving a 19.7% reduction. On the RS-STOD dataset, this model also improves mAP50:95 by 0.052, demonstrating the effectiveness of the proposed modifications.

3.3. GCP-ASFF Module

In traditional FPN, the direct fusion of feature maps at different levels often results in inconsistent gradient propagation due to differences in spatial features and semantic information. For example, high-level features (e.g., those subject to 32× downsampling) are more suitable for detecting large objects, whereas low-level features (e.g., those subject to 8× downsampling) are more sensitive to small objects. However, when both large and small objects coexist in an image, the fusion of features across different levels may cause interference, leading to inconsistent gradient propagation and subsequently reducing detection accuracy. In other words, small objects tend to lose significant details after multiple downsampling steps (e.g., a 4 × 4-pixel object is reduced to a 1 × 1 effective area after 4× downsampling), and the weak signals from low-level features can easily be overwhelmed by high-level features. To overcome the shortcomings of multi-scale feature fusion, Adaptively Spatial Feature Fusion (ASFF) is introduced.
Improving the feature fusion capability in FPN enables the network to more effectively handle the challenges posed by intricate environments and objects at varying scales, ultimately enhancing detection accuracy. Let x 1 , x 2 , and x 3 denote the feature vectors corresponding to levels 1, 2, and 3, respectively. The fused feature at hierarchical level k can be expressed through the following parametric combination:
z i j k = α i j k · x i j 1 k + β i j k · x i j 2 k + γ i j k · x i j 3 k
where z i j k denotes the synthesized feature vector at spatial position ( i , j ) of the k-th level pyramid. The term x i j n k represents the transformed feature vector from pyramid level n to k at coordinate ( i , j ) , while α i j k , β i j k , and γ i j k correspond to learnable spatial weights for feature integration. The weight derivation follows a three-stage computational framework. Level-specific transformation is achieved through depth-wise separable convolutions:
Channel projection: Apply 1 × 1 convolutional kernels to x 1 k , x 2 k , and x 3 k , generating intermediate tensors λ α k , λ β k , and λ γ k , respectively. The 1 × 1 convolution reduces channel dimensions while preserving spatial information, where λ φ k R H k × W k × 1 ( φ { α , β , γ } ) contains unnormalized attention signals for each feature level.
Attention generation: These intermediate results undergo softmax normalization across channels:
φ ^ i j k = e λ φ i j k ψ { α , β , γ } e λ ψ i j k , φ { α , β , γ }
where φ ^ i j k represents the normalized attention weight at location ( i , j ) for level k.
Parameter stabilization: The attention weights are adjusted through linear scaling to maintain numerical stability within [0, 1]:
φ i j k = φ ^ i j k ψ { α , β , γ } ψ ^ i j k
This final normalization step guarantees φ i j k = 1 while preventing gradient saturation during backpropagation.
By introducing learnable spatial weight parameters, ASFF dynamically adjusts the contribution of features from different levels during the fusion process. For instance, when detecting small objects, the model assigns higher weights to low-level features while suppressing redundant information in high-level features. This effectively reduces feature conflicts and enhances scale invariance. However, ASFF only fuses multi-scale features at the same spatial location through spatial weights, lacking the utilization of global contextual information. This constraint limits the receptive field in intricate situations like detecting small objects. To overcome this challenge, we propose GCP-ASFF, which is an enhancement of ASFF. The introduced GCP enhances the semantic consistency of multi-scale features by extracting channel-level global features, thereby improving the quality of feature fusion. The overall structure of GCP-ASFF is illustrated in Figure 3.
The SA (Spatial Alignment), WF (Weight Fusion), and FA (Feature Aggregation) modules in the diagram retain the ASFF structure. Figure 4 illustrates the comprehensive structure of the GCP module, and the subsequent algorithm flow is outlined below:
1. Multi-scale Feature Alignment and Global Context Extraction. Resize multi-level features to a unified resolution and extract channel-wise global descriptors via GAP:
F i resized = Resize ( F i ) , i { 1 , 2 , 3 } , g i = 1 H × W h = 1 H w = 1 W F i resized ( h , w , c ) , g i R C ,
where F i R C × H i × W i denotes the input feature maps from different pyramid levels.
2. Lightweight Context Integration. Concatenate the global descriptor g i with the resized features and compress channels using a 1 × 1 convolution:
F ^ i = Conv 1 × 1 ( Concat ( F i resized , g i 1 H × W ) ) , F ^ i R C × H × W ,
where ⊗ denotes broadcasting and 1 H × W is an all-ones matrix.
3. Adaptive Spatial Weighting and Fusion. Generate spatial attention weights α i and fuse context-enhanced features
α i = Softmax Conv F ^ i , α i R 1 × H × W , F fused = i = 1 3 α i F ^ i ,
where ⊙ denotes element-wise multiplication.
The GCP module is pivotal in enhancing the feature representation through two core processes: channel compression and global descriptor generation.
Channel Compression: The GCP module utilizes a 1 × 1 convolution to reduce the concatenated input channels (which consists of the original feature map and its global average pooled representation) back to the original dimensionality, ensuring efficient processing and maintaining essential features.
Global Descriptor Generation: It generates global descriptors by applying an adaptive average pooling operation to the input feature map. This results in a compact representation per channel, effectively capturing the global context.
These mechanisms enable the model to leverage both detailed local features and broader contextual information, thus enhancing the overall detection performance.

3.4. RFA-C3k2 Module

In response to the characteristics of remote sensing images, such as small target sizes, dense distributions, and complex backgrounds, the RFA-C3k2 module has been developed, and its specific implementation is illustrated in Figure 5. The core improvement lies in the introduction of Receptive Field Attention Convolution (RFAConv), which dynamically adjusts the attention weights of convolutional kernels, effectively enhancing the model’s ability to perceive multi-scale features.
In small object detection tasks, the fixed receptive field of traditional convolution is insufficient for effectively capturing the local details of small objects in remote sensing images. RFAConv addresses this issue by employing a multi-branch convolutional structure to generate attention weights, enabling the dynamic fusion of features across different receptive field windows. For example, the large receptive field branch captures the contextual relationships between targets and their surrounding environment, while the small receptive field branch focuses on local texture features. These complementary capabilities enhance the model’s sensitivity to small objects. In most cases, the computation of RFA can be reformulated as:
F = S o f t m a x ( g 1 × 1 ( A v g P o o l ( X ) ) ) × R e L U ( N o r m ( g k × k ( X ) ) ) = A r f × F r f
where g i × i corresponds to a grouped convolution with a kernel size of i × i , k denotes the size of the convolutional kernel, N o r m refers to the normalization operation, and X represents the input feature maps. The resulting output F is derived by multiplying the attention map A r f with the spatial features F r f transformed by the receptive field.
The RFA module dynamically adjusts receptive field weights through a two-stage process:
Receptive Field Attention Generation: Global spatial context is first captured via an average pooling layer (AvgPool2d), followed by a 1 × 1 grouped convolution to produce unnormalized attention weights ( dynamic _ weights R b × c × k 2 × h × w ).
Adaptive Feature Weighting: The attention weights are normalized channel-wise using Softmax, ensuring that the sum of weights for each spatial position across the k 2 receptive field windows equals 1. These normalized weights are then multiplied element-wise with multi-scale convolutional features (from the generate_feature branch), enabling adaptive fusion where “larger receptive fields focus on semantic context, while smaller ones emphasize local details”. Mathematically, given an input feature map X, the RFAConv first computes the attention weights A r f and receptive field features F r f through dual branches. The final output F is formulated as:
F = i = 1 k 2 A r f ( i ) F r f ( i )
Here, i indexes the spatial positions of the k × k convolutional kernel. This mechanism allows the model to dynamically allocate attention weights based on target scales—enhancing local details for densely packed small objects while reinforcing global context modeling for isolated large targets. Compared to the original C3k2 module, which extracts features using fixed-size convolutional kernels, the RFA-C3k2 module dynamically generates receptive field attention weights, breaking the limitation of parameter sharing in convolutional kernels. This improves the model’s capacity to identify local features in small targets.
In the redesigned bottleneck structure, RFAConv is incorporated to improve the model’s capability to sample input feature maps more effectively, thereby facilitating the learning of fine details in small target objects as well as background information. Following the convolutional operations, residual connections are employed to merge the input feature maps with the output feature maps.

4. Experimental Results

4.1. Datasets

The performance of the SEMA-YOLO model was evaluated on the RS-STOD dataset and the AI-TOD dataset. The descriptions of these two datasets are as follows:
RS-STOD is an HR remote sensing dataset for small object detection, with a training-to-testing ratio of approximately 4:1. This dataset contains a total of 50,854 instances across 2354 aerial images, covering five categories: small vehicle, large vehicle, ship, airplane, and oil tank. Based on the definition of small objects in the MS-COCO, small objects account for 93% of the RS-STOD dataset. The images are sourced from the New Zealand land use website and Microsoft Bing Maps imagery tiles, with resolutions ranging from 0.4 m to 2 m.
AI-TOD is an aerial imagery dataset used for small object detection. It contains 28,036 images and 700,621 object instances, divided into 8 categories. The AI-TOD dataset has been re-partitioned based on the following considerations: Firstly, the original dataset exhibits a highly imbalanced sample distribution, which may affect the final training performance. Secondly, the dataset size has been reduced to accommodate limited computational resources. Additionally, the partitioning process strictly follows a random principle to ensure objectivity and fairness, while guaranteeing that all comparative methods are evaluated under the same conditions. Ultimately, 2700 images from the AI-TOD dataset are allocated to the training set, while 300 images are assigned to the test set.

4.2. Implementation Details

In this study, all experiments were conducted on a single NVIDIA RTX 4090 GPU, implemented using the PyTorch framework. The Stochastic Gradient Descent (SGD) optimizer was employed for training. Experiments were performed on both the RS-STOD and AI-TOD datasets. For the RS-STOD dataset, the input image size was uniformly resized to 512 × 512 pixels (consistent with the original resolution of the dataset), with a batch size set to 16. For the AI-TOD dataset, the input image size was set to 640 × 640 pixels, and the batch size was adjusted to 4. Unless otherwise specified, all experiments followed the aforementioned configurations.

4.3. Evaluation Metrics

A set of standard metrics is adopted to evaluate the detection performance of the model, including Precision, Recall, F1-score, and mean Average Precision (mAP) at IoU thresholds of 0.5 and 0.5–0.95. The calculation formulas for some of the evaluation metrics are as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
m A P = i = 1 k j = 1 n 1 ( R j + 1 R j ) P inter ( R j + 1 ) i k
Here, T P , F P , and F N denote the number of true positives, false positives, and false negatives, respectively. Precision measures the correctness of the predicted positive samples, while recall indicates the proportion of ground-truth objects that are successfully detected. The F1-score provides a harmonic mean of them. The mAP is computed as the mean of the A P i over all N object categories.
In addition, the number of neural network parameters (M), Giga Floating-point Operations Per Second (GFLOPs), Model size (MB) and frames per second (FPS) are reported to measure the model’s complexity and computational cost. These reflect, respectively, the model’s complexity, computational workload, storage footprint, and inference speed. A rough estimation of FPS is given by:
F P S = f r a m e N u m e l a p s e d T i m e
f r a m e N u m refers to the number of images (or video frames) processed, and e l a p s e d T i m e is the total time taken by the model to perform detection on these frames.

4.4. Comparison with State-of-the-Art Methods

To evaluate the performance of SEMA-YOLO, we conducted comparisons with the RT-DETR series and several lightweight YOLO models such as YOLOv8n and YOLOv9t. The results are shown in Table 1.
The SEMA-YOLO model obtained scores of 0.725 for mAP50 and 0.468 for mAP50:95. Compared to RT-DETR-R50, SEMA-YOLO improves mAP50 by 0.225 and mAP50:95 by 0.197. Among the YOLO models, SEMA-YOLO outperforms the best-performing variant, YOLOv8n, with an mAP50 improvement of 0.054 and an mAP50:95 improvement of 0.053. The strong performance in mAP50:95 indicates that the model maintains reliable localization accuracy even at high IoU thresholds, demonstrating its adaptability to varying object scales and complex detection scenarios.
As shown in Table 2, SEMA-YOLO achieved the highest AP50 across all five object categories. For small objects, it improved AP50 for small vehicles to 0.722 and large vehicles to 0.196, representing increases of 20.3 % and 39.0 % compared to the second-best model YOLOv11n. For large-scale objects, SEMA-YOLO maintained the highest AP50 for airplanes at 0.987 and oil tanks at 0.926, demonstrating the effectiveness of retaining the large-object detection head.
Further analysis of the RS-STOD dataset reveals significant variation in the difficulty of detecting different object categories. Large vehicles are the most challenging due to their low representation, which limits the model’s ability to learn their features effectively. Aircraft are the easiest to detect, as they have distinctive visual characteristics and appear against relatively simple backgrounds. Although small vehicles are challenging due to their size, their high frequency in the dataset enables the model to learn them well, leading to reasonable detection performance.
Quantitative metrics in Table 2 demonstrate the overall superiority of SEMA-YOLO, while the visual analysis of representative remote sensing scenarios intuitively highlights the detection performance of different models. Four representative scenarios are selected, including small vehicles in complex backgrounds, densely clustered harbor vessels, aircraft in airport scenes, and scale-variant oil tanks. The detection results are presented in Figure 6.
In order to evaluate the detection performance on different datasets, we conducted comparative experiments using the publicly available AI-TOD dataset. As shown in Table 3, the proposed method demonstrates significant improvements over existing approaches, particularly outperforming the best-performing YOLO model (YOLOv11) in key metrics.
SEMA-YOLO achieves an mAP50 of 0.615, representing a 0.482 improvement over the RT-DETR-L. Under the more stringent mAP50:95 evaluation metric, it attains a detection accuracy of 0.284 while maintaining superior precision (0.740) and recall (0.557). The comparison of visualization results among different models is shown in Figure 7. These results demonstrate robust performance in tiny object detection and indicate that the proposed approach not only performs well on custom RS-STOD datasets but also exhibits strong competitiveness on standard public benchmarks.
The model demonstrates strong generalization performance, as evidenced by its excellent results on both the RS-STOD and AI-TOD datasets. This robust generalization is primarily attributed to the integration of two key modules in the SEMA-YOLO architecture: the SLE module, which enhances the representation of small objects by enlarging the feature maps, and the GCP-ASFF module, which mitigates inconsistencies across features of different scales. The combination of these modules allows the model to effectively handle small objects across diverse datasets, thus showcasing its superior generalization capability and adaptability to various remote sensing small object detection tasks.

4.5. Ablation Experiments

To evaluate the effectiveness of the proposed improvements, an ablation experiment was conducted on our custom-built RS-STOD dataset, as shown in Table 4. The goal is to verify the contribution of the SLE strategy, GCP-ASFF module, and RFA module to the overall detection performance. The baseline model, YOLOv11n, serves as a reference, providing a starting point for assessing the impact of these modifications.
The results demonstrate that the SLE strategy significantly enhances detection performance. Compared to the baseline, SLE improves precision from 0.709 to 0.752 and recall from 0.643 to 0.681, leading to a notable increase in mAP. This suggests that SLE effectively strengthens feature extraction, optimizing both object localization and classification. Moreover, when combined with ASFF, the performance further improves, highlighting the strong compatibility between these two components. It should be emphasized that, in the ablation study presented in Table 4, the “ASFF” module corresponds to the original Adaptively Spatial Feature Fusion without the Global Context Pooling (GCP) enhancement. The progression from “+SLE+ASFF+RFA” to the final “SEMA-YOLO” model involves replacing the ASFF module with the proposed GCP-ASFF module, which integrates the GCP mechanism. This substitution directly demonstrates the effectiveness of the GCP module in enhancing multi-scale feature fusion and overall detection performance. Hence, the GCP module’s contribution has been verified within this experimental framework. When the RFA module is solely integrated with SLE, the model’s performance metrics experience a slight decline, indicating that RFA alone has limited effectiveness in enhancing feature representation. However, with the introduction of ASFF, the incorporation of RFA significantly improves the model’s recall rate and achieves superior performance under the more stringent mAP50:95 metric. It is hypothesized that this is because ASFF enables the model to learn spatial weights for multi-scale feature map fusion, thereby increasing the contribution of large-scale feature maps. This allows RFA to more effectively capture fine-grained details, leading to enhanced overall performance. It is worth noting that while the SLE+ASFF combination achieves the highest precision, the complete SEMA-YOLO network including the RFA module shows a slight decrease in precision. This is mainly because the RFA module enhances the capture of fine-grained and multi-scale features, significantly improving recall and enabling detection of more challenging targets. The increase in recall typically accompanies a slight increase in false positives, thus resulting in a minor precision decrease. This precision-recall trade-off reflects the model’s balanced optimization strategy. Overall, by maintaining high precision and improving recall, SEMA-YOLO achieves superior overall detection performance, validating the advantage of integrating these modules.
Grad-CAM [66], as an advanced visualization technique, precisely maps the feature map regions that make decisive contributions to prediction results in deep neural networks. By computing gradients of target classes with respect to convolutional feature maps and performing subsequent weighting operations, Grad-CAM effectively localizes critical regions in the network decision-making process, significantly enhancing the transparency and interpretability of the SEMA-YOLO model. The Grad-CAM heatmap visualization results presented in Figure 8 clearly demonstrate the superior performance of SEMA-YOLO in small object detection tasks: the model can effectively distinguish between targets and background, with our proposed GCP-ASFF module successfully suppressing activation responses in non-target regions, while the SLE and RFA modules enhance feature representation for small objects. This high concentration of activation areas and accurate recognition directly validates the effectiveness of our proposed SEMA-YOLO model in addressing small object detection problems under complex backgrounds. The degree of highlighting in target regions in the activation maps is highly consistent with actual detection performance, providing intuitive visual evidence for the rationality of the model design.
These findings confirm the effectiveness of our design choices. By strategically integrating SLE, GCP-ASFF, and RFA, a well-balanced model is achieved that improves detection accuracy while maintaining computational efficiency. The final model demonstrates superior performance, validating our approach and providing valuable insights for future research in object detection.

5. Discussion

5.1. Model Potential

As summarized in Table 5, the scalability of SEMA-YOLO is further demonstrated by evaluating the s- and m-size variants. When model capacity is increased from n to s, the mAP50 improves from 72.5% to 75.3%, and continuing to m-size yields 76.8%, indicating that our architectural design can effectively leverage additional parameters for more accurate detections. Specifically, the small vehicle and ship categories show particularly strong gains with deeper models, suggesting that the multi-scale feature fusion in GCP-ASFF complements RFA’s fine-scale feature extraction.
A notable advantage of our approach is its consistently favorable trade-off between accuracy and computational overhead. Although SEMA-YOLO-m approaches the complexity of RT-DETR-R50 with comparable GFLOPs (138.1 vs. 136), it achieves superior detection performance across most categories (e.g., 24.5% vs. 4.1% AP for large vehicle) and a substantial margin (+26.8%) in overall mAP50. These results illustrate that our design priorities—particularly the combination of SLE for preserving spatial granularity and the GCP-ASFF mechanism for robust cross-scale fusion—enable the model to more efficiently utilize its computational budget.
Although larger variants of SEMA-YOLO deliver higher detection performance, the lightweight SEMA-YOLO-n model still presents a compelling balance of accuracy and real-time capabilities (14.2 GFLOPs) for practical deployment on embedded platforms (e.g., Jetson Orin NX). In future work, efforts could focus on refining the GCP module for even more efficient global context extraction and exploring adaptive training strategies for further enhancing small object detection under extreme scale variations or low-contrast environments.
Overall, the marked improvements across diverse object classes, combined with favorable parameter efficiency and robust generalization on challenging remote sensing scenes, affirm the potential of SEMA-YOLO as a state-of-the-art framework for small object detection tasks.

5.2. Computational Efficiency

To further analyze the computational efficiency and deployment potential of SEMA-YOLO, we compared its parameter count and computational cost with those of the RT-DETR series and lightweight YOLO models. The results are shown in Table 6.
The increase in model complexity and size in SEMA-YOLO is relatively modest, yet it significantly improves detection accuracy. With 3.6 million parameters and a model size of 7.43 MB, SEMA-YOLO is larger than several lightweight YOLO models, such as YOLOv9t (2 million parameters, 4.44 MB) and YOLOv10n (2.7 million parameters, 5.90 MB). This slight increase in complexity contributes to a notable enhancement in performance, demonstrating the effectiveness of the architecture.
SEMA-YOLO’s computational cost of 14.2 GFLOPs and frame rate of 185 FPS place it between the lightweight YOLO models and the more complex RT-DETR series. While its performance is slightly lower than that of YOLOv10n (277 FPS) and YOLOv8n (244 FPS), it outperforms the RT-DETR series, such as RT-DETR-L (110 GFLOPs, 182 FPS) and RT-DETR-R50 (136 GFLOPs, 156 FPS). The frame rate of 185 FPS is sufficient to meet real-time requirements, while the 14.2 GFLOPs offer strong deployment potential.
In summary, SEMA-YOLO offers a strong balance between computational efficiency and performance. Its higher parameter count and computational cost compared to lightweight YOLO models are justified by its substantial improvements in detection accuracy. For small object detection, increasing model complexity to enhance precision is a highly reasonable approach. While its FPS is slightly lower than some of the other YOLO variants, it still maintains real-time performance, making it suitable for applications where higher detection precision is critical.

6. Conclusions

This study introduces SEMA-YOLO, an effective framework for small object detection in HR remote sensing imagery. By incorporating novel modules like SLE, GCP-ASFF, and RFA on the YOLOv11n baseline, multi-scale feature fusion is enhanced, improving detection accuracy under complex conditions. Extensive experiments on the RS-STOD and AI-TOD datasets show that SEMA-YOLO delivers competitive accuracy, real-time inference speed, and a compact model size. Future work will focus on exploring Transformer-based architectures, advanced cross-scale fusion techniques, and practical deployment strategies.

Author Contributions

Methodology, Z.W., H.Z., X.Z., X.B. and X.L.; Software, Z.W., H.Z., X.Z. and X.B.; Writing—original draft, Z.W., H.Z. and X.Z.; Writing—review & editing, X.B. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Undergraduate Training Programs for Innovation of Wuhan University (Grant No.S202510486303).

Data Availability Statement

In this study, the open remote sensing datasets that support the findings are as follows: The RS-STOD dataset is available at https://github.com/lixinghua5540/STOD, (accessed on 11 April 2025), and the AI-TOD dataset is available at https://github.com/jwwangchn/AI-TOD, (accessed on 11 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, W.; Cheng, G.; Wang, M.; Yao, Y.; Xie, X.; Yao, X.; Han, J. MAR20: A benchmark for military aircraft recognition in remote sensing images. Natl. Remote Sens. Bull. 2024, 27, 2688–2696. [Google Scholar] [CrossRef]
  2. Ariza-Sentís, M.; Vélez, S.; Martínez-Peña, R.; Baja, H.; Valente, J. Object detection and tracking in Precision Farming: A systematic review. Comput. Electron. Agric. 2024, 219, 108757. [Google Scholar] [CrossRef]
  3. Hoalst-Pullen, N.; Patterson, M.W. Applications and Trends of Remote Sensing in Professional Urban Planning: Remote sensing in professional urban planning. Geogr. Compass 2011, 5, 249–261. [Google Scholar] [CrossRef]
  4. Li, J.; Pei, Y.; Zhao, S.; Xiao, R.; Sang, X.; Zhang, C. A review of remote sensing for environmental monitoring in China. Remote Sens. 2020, 12, 1130. [Google Scholar] [CrossRef]
  5. Tan, Q.; Ling, J.; Hu, J.; Qin, X.; Hu, J. Vehicle detection in high resolution satellite remote sensing images based on deep learning. IEEE Access 2020, 8, 153394–153402. [Google Scholar] [CrossRef]
  6. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the 13th European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar] [CrossRef]
  7. Wang, X.; Wang, A.; Yi, J.; Song, Y.; Chehri, A. Small object detection based on deep learning for remote sensing: A comprehensive review. Remote Sens. 2023, 15, 3265. [Google Scholar] [CrossRef]
  8. European Space Agency. WorldView-3 Mission. 2025. Available online: https://earth.esa.int/eogateway/missions/worldview-3 (accessed on 11 March 2025).
  9. DJI. P4RTK System Specifications. 2019. Available online: https://www.dji.com/cn/support/product/phantom-4-rtk (accessed on 11 March 2025).
  10. Chinese Academy of Sciences. Gaofen-3 03 Satellite SAR Payload Achieves 1-Meter Resolution with World-Leading Performance. 2022. Available online: http://aircas.ac.cn/dtxw/kydt/202204/t20220407_6420970.html (accessed on 11 March 2025).
  11. Wikipedia Contributors. TerraSAR-X. 2025. Available online: https://en.wikipedia.org/wiki/TerraSAR-X (accessed on 11 March 2025).
  12. Wang, C.; Bai, X.; Wang, S.; Zhou, J.; Ren, P. Multiscale visual attention networks for object detection in VHR remote sensing images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 310–314. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Ye, M.; Zhu, G.; Liu, Y.; Guo, P.; Yan, J. FFCA-YOLO for small object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5611215. [Google Scholar] [CrossRef]
  14. Deng, Z.; Sun, H.; Zhou, S.; Zhao, J.; Lei, L.; Zou, H. Multi-scale object detection in remote sensing imagery with convolutional neural networks. ISPRS J. Photogramm. Remote Sens. 2018, 145, 3–22. [Google Scholar] [CrossRef]
  15. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.S. Tiny object detection in aerial images. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 3791–3798. [Google Scholar] [CrossRef]
  16. Liu, S.; Huang, D.; Wang, Y. Learning Spatial Fusion for Single-Shot Object Detection. arXiv 2019, arXiv:1911.09516. [Google Scholar] [CrossRef]
  17. Zhang, X.; Liu, C.; Yang, D.; Song, T.; Ye, Y.; Li, K.; Song, Y. RFAConv: Innovating Spatial Attention and Standard Convolutional Operation. arXiv 2023, arXiv:2304.03198. [Google Scholar] [CrossRef]
  18. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. arXiv 2016, arXiv:1612.03144. [Google Scholar] [CrossRef]
  19. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef]
  20. Ghiasi, G.; Lin, T.Y.; Le, Q.V. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7029–7038. [Google Scholar] [CrossRef]
  21. Zhu, L.; Deng, Z.; Hu, X.; Fu, C.W.; Xu, X.; Qin, J.; Heng, P.A. Bidirectional Feature Pyramid Network with Recurrent Attention Residual Modules for Shadow Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 122–137. [Google Scholar] [CrossRef]
  22. Qiao, S.; Chen, L.C.; Yuille, A. DetectoRS: Detecting Objects with Recursive Feature Pyramid and Switchable Atrous Convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10213–10224. [Google Scholar] [CrossRef]
  23. Yang, G.; Lei, J.; Zhu, Z.; Cheng, S.; Feng, Z.; Liang, R. AFPN: Asymptotic Feature Pyramid Network for Object Detection. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, HI, USA, 1–4 October 2023; pp. 2184–2189. [Google Scholar] [CrossRef]
  24. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  25. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  26. Zhang, D.; Zhang, H.; Tang, J.; Wang, M.; Hua, X.; Sun, Q. Feature Pyramid Transformer. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 323–339. [Google Scholar] [CrossRef]
  27. Liu, H.I.; Tseng, Y.W.; Chang, K.C.; Wang, P.J.; Shuai, H.H.; Cheng, W.H. A DeNoising FPN with Transformer R-CNN for Tiny Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4704415. [Google Scholar] [CrossRef]
  28. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 73–76. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar] [CrossRef]
  30. Zhang, J.; Lei, J.; Xie, W.; Fang, Z.; Li, Y.; Du, Q. SuperYOLO: Super Resolution Assisted Object Detection in Multimodal Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605415. [Google Scholar] [CrossRef]
  31. Jin, Y.; Zhang, Y.; Cen, Y.; Li, Y.; Mladenovic, V.; Voronin, V. Pedestrian detection with super-resolution reconstruction for low-quality image. Pattern Recognit. 2021, 115, 107846. [Google Scholar] [CrossRef]
  32. Bashir, S.M.A.; Wang, Y. Small Object Detection in Remote Sensing Images with Residual Feature Aggregation-Based Super-Resolution and Object Detector Network. Remote Sens. 2021, 13, 1854. [Google Scholar] [CrossRef]
  33. Truong, N.T.; Vo, N.D.; Nguyen, K. The effects of super-resolution on object detection performance in an aerial image. In Proceedings of the 7th NAFOSTED Conference on Information and Computer Science (NICS), Ho Chi Minh City, Vietnam, 26–27 November 2020; pp. 256–260. [Google Scholar] [CrossRef]
  34. Lim, J.S.; Astrid, M.; Yoon, H.J.; Lee, S.I. Small Object Detection using Context and Attention. In Proceedings of the International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 13–16 April 2021; pp. 181–186. [Google Scholar] [CrossRef]
  35. Tang, X.; Du, D.K.; He, Z.; Liu, J. PyramidBox: A Context-assisted Single Shot Face Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 797–813. [Google Scholar] [CrossRef]
  36. Zhao, Z.; Du, J.; Li, C.; Fang, X.; Xiao, Y.; Tang, J. Dense Tiny Object Detection: A Scene Context Guided Approach and a Unified Benchmark. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5606913. [Google Scholar] [CrossRef]
  37. Liang, X.; Zhang, J.; Zhuo, L.; Li, Y.; Tian, Q. Small Object Detection in Unmanned Aerial Vehicle Images Using Feature Fusion and Scaling-Based Single Shot Detector with Spatial Context Analysis. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1758–1770. [Google Scholar] [CrossRef]
  38. Cui, L.; Lv, P.; Jiang, X.; Gao, Z.; Zhou, B.; Zhang, L.; Shao, L.; Xu, M. Context-Aware Block Net for Small Object Detection. IEEE Trans. Cybern. 2022, 52, 2300–2313. [Google Scholar] [CrossRef] [PubMed]
  39. Tong, K.; Wu, Y. Small object detection using hybrid evaluation metric with context decoupling. Multimed. Syst. 2025, 31, 141. [Google Scholar] [CrossRef]
  40. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  41. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef]
  42. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Xie, T.; Fang, J.; Yifu, Z.; Wong, C.; et al. ultralytics/yolov5: V7.0—YOLOv5 SOTA Realtime Instance Segmentation; Zenodo: Geneve, Switzerland, 2022. [Google Scholar] [CrossRef]
  43. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  44. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar] [CrossRef]
  45. Jocher, G.; Qiu, J.; Chaurasia, A. Ultralytics YOLO, Version 8.0.0; Github: San Francisco, CA, USA, 2023; Available online: https://github.com/ultralytics/ultralytics, (accessed on 2 March 2025).
  46. Wang, C.Y.; Yeh, I.H.; Mark Liao, H.Y. Yolov9: Learning what you want to learn using programmable gradient information. In Proceedings of the European Conference on Computer Vision (ECCV), Milan, Italy, 29 September–4 October 2024; pp. 1–21. [Google Scholar] [CrossRef]
  47. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar] [CrossRef]
  48. Tian, Y.; Ye, Q.; Doermann, D. YOLOv12: Attention-Centric Real-Time Object Detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar] [CrossRef]
  49. Qiu, M.; Huang, L.; Tang, B.H. ASFF-YOLOv5: Multielement detection method for road traffic in UAV images based on multiscale feature fusion. Remote Sens. 2022, 14, 3498. [Google Scholar] [CrossRef]
  50. Lin, J.; Zhao, Y.; Wang, S.; Tang, Y. YOLO-DA: An efficient YOLO-based detector for remote sensing object detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 6008705. [Google Scholar] [CrossRef]
  51. Li, Z.; Yuan, J.; Li, G.; Wang, H.; Li, X.; Li, D.; Wang, X. RSI-YOLO: Object detection method for remote sensing images based on improved YOLO. Sensors 2023, 23, 6414. [Google Scholar] [CrossRef]
  52. Xie, S.; Zhou, M.; Wang, C.; Huang, S. CSPPartial-YOLO: A Lightweight YOLO-Based Method for Typical Objects Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 388–399. [Google Scholar] [CrossRef]
  53. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z.X. Scale-Aware Trident Networks for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6053–6062. [Google Scholar] [CrossRef]
  54. Cao, C.; Wang, B.; Zhang, W.; Zeng, X.; Yan, X.; Feng, Z.; Liu, Y.; Wu, Z. An Improved Faster R-CNN for Small Object Detection. IEEE Access 2019, 7, 106838–106846. [Google Scholar] [CrossRef]
  55. Ren, Y.; Zhu, C.; Xiao, S. Small object detection in optical remote sensing images via modified faster R-CNN. Appl. Sci. 2018, 8, 813. [Google Scholar] [CrossRef]
  56. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. arXiv 2019, arXiv:1904.01355. [Google Scholar] [CrossRef]
  57. Xu, C.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G.S. RFLA: Gaussian receptive field based label assignment for tiny object detection. In Proceedings of the European Conference on Computer Vision(ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 526–543. [Google Scholar] [CrossRef]
  58. Kang, S.H.; Park, J.S. Aligned Matching: Improving Small Object Detection in SSD. Sensors 2023, 23, 2589. [Google Scholar] [CrossRef] [PubMed]
  59. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar] [CrossRef]
  60. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar] [CrossRef]
  61. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar] [CrossRef]
  62. Li, F.; Zhang, H.; Liu, S.; Guo, J.; Ni, L.M.; Zhang, L. DN-DETR: Accelerate DETR Training by Introducing Query DeNoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13609–13617. [Google Scholar] [CrossRef]
  63. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection. arXiv 2022, arXiv:2203.03605. [Google Scholar] [CrossRef]
  64. Huang, Y.X.; Liu, H.I.; Shuai, H.H.; Cheng, W.H. DQ-DETR: DETR with Dynamic Query for Tiny Object Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Paris, France, 26–27 March 2025; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer: Cham, Swizterland, 2025; pp. 290–305. [Google Scholar] [CrossRef]
  65. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-time Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar] [CrossRef]
  66. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
Figure 1. The challenges of small object detection in remote sensing images. (Left): the shadows. (Middle): limited number of pixels of vehicles. (Right): dense distribution of ships.
Figure 1. The challenges of small object detection in remote sensing images. (Left): the shadows. (Middle): limited number of pixels of vehicles. (Right): dense distribution of ships.
Remotesensing 17 01917 g001
Figure 2. The overall framework of SEMA-YOLO.
Figure 2. The overall framework of SEMA-YOLO.
Remotesensing 17 01917 g002
Figure 3. The overview of GCP-ASFF. Before feature fusion, global average pooling is applied to the feature map of each level to generate channel descriptor vectors. These vectors are then concatenated with the original features to enhance the representational capacity of the features.
Figure 3. The overview of GCP-ASFF. Before feature fusion, global average pooling is applied to the feature map of each level to generate channel descriptor vectors. These vectors are then concatenated with the original features to enhance the representational capacity of the features.
Remotesensing 17 01917 g003
Figure 4. The overview of GCP. The GCP module acts as a global semantic filter, where channel-wise statistics guide local feature enhancement. High global activations amplify target-related patterns, while low activations suppress background noise.
Figure 4. The overview of GCP. The GCP module acts as a global semantic filter, where channel-wise statistics guide local feature enhancement. High global activations amplify target-related patterns, while low activations suppress background noise.
Remotesensing 17 01917 g004
Figure 5. The overview of RFA-C3k2.
Figure 5. The overview of RFA-C3k2.
Remotesensing 17 01917 g005
Figure 6. Detection results on RS-STOD datasets. SEMA-YOLO demonstrates more complete bounding boxes and fewer false positives across all scenarios.
Figure 6. Detection results on RS-STOD datasets. SEMA-YOLO demonstrates more complete bounding boxes and fewer false positives across all scenarios.
Remotesensing 17 01917 g006
Figure 7. Detection results on AI-TOD datasets.
Figure 7. Detection results on AI-TOD datasets.
Remotesensing 17 01917 g007
Figure 8. Heatmap of ablation experiment.
Figure 8. Heatmap of ablation experiment.
Remotesensing 17 01917 g008
Table 1. Comparison experiments for SEMA-YOLO in RS-STOD.
Table 1. Comparison experiments for SEMA-YOLO in RS-STOD.
MethodP↑R↑F1-Score↑mAP50mAP50:95
RT-DETR-L0.4250.4430.4340.4290.230
RT-DETR-R500.4900.5070.4980.5000.271
YOLOv8n0.7320.6300.6770.6710.416
YOLOv9t0.7070.6320.6670.6610.411
YOLOv10n0.7130.6240.6660.6620.411
YOLOv11n0.7080.6430.6740.6720.412
YOLOv12n0.7130.6190.6630.6610.406
SEMA-YOLO0.7480.6990.7220.7250.468
Note: P represents precision, and R represents recall.
Table 2. Comparison experiments for SEMA-YOLO in RS-STOD. (AP50).
Table 2. Comparison experiments for SEMA-YOLO in RS-STOD. (AP50).
MethodSVLVSHAPOT
RT-DETR-L0.3080.0250.3560.8560.602
RT-DETR-R500.3790.0410.4790.9080.691
YOLOv8n0.6000.1360.7440.9830.894
YOLOv9t0.5870.1330.7160.9840.881
YOLOv10n0.5880.1480.7120.9790.884
YOLOv11n0.6000.1410.7460.9840.887
YOLOv12n0.5800.1340.7300.9840.878
SEMA-YOLO0.7220.1960.7930.9870.926
Note: Table shows the AP50 results for each object category. SV corresponds to small vehicle, LV to large vehicle, SH to ship, AP to airplane, and OT to oil tank.
Table 3. Comparison experiments on AI-TOD dataset.
Table 3. Comparison experiments on AI-TOD dataset.
MethodP↑R↑F1-Score↑mAP50mAP50:95
RT-DETR-L0.5370.1990.2900.1330.043
RT-DETR-R500.3940.3080.3460.2420.080
YOLOv8n0.6620.5480.6000.5570.235
YOLOv9t0.670.5220.5870.5470.232
YOLOv10n0.6190.5140.5620.5320.229
YOLOv11n0.7030.5240.6000.5630.239
YOLOv12n0.6750.5190.5870.5390.228
SEMA-YOLO0.7400.5570.6360.6150.284
Note: P represents precision, and R represents recall.
Table 4. Ablation experiment for SLE, GCP-ASFF, RFA in RS-STOD.
Table 4. Ablation experiment for SLE, GCP-ASFF, RFA in RS-STOD.
MethodP↑R↑mAP50mAP50:95Para (M)↓GFLOPs↓
Baseline0.7090.6430.6710.4122.5836.3
+SLE0.7520.6810.7170.4642.0759.7
+SLE+RFA0.7540.6800.7150.4562.0789.8
+SLE+ASFF0.7650.6870.7220.4653.46812.8
+SLE+ASFF+RFA0.7490.6980.7210.4663.47112.9
SEMA-YOLO0.7460.7000.7250.4683.64514.2
Note: P represents precision, and R represents recall.
Table 5. Comprehensive Model Performance Comparison.
Table 5. Comprehensive Model Performance Comparison.
MethodmAP50mAP50:95Para (M)↓GFLOPs↓
RT-DETR-R500.5000.27141.9136
YOLOv11n0.6710.4122.66.3
SEMA-YOLOn0.7250.4683.614.2
SEMA-YOLOs0.7530.49913.443.6
SEMA-YOLOm0.7680.52129.7138.1
Table 6. Model Complexity and Efficiency.
Table 6. Model Complexity and Efficiency.
MethodPara (M)↓GFLOPs↓Size (MB)↓FPS↑
RT-DETR-L32.011063.1182
RT-DETR-R5041.913684.0156
YOLOv8n3.08.75.97244
YOLOv9t2.07.74.44227
YOLOv10n2.76.75.90277
YOLOv11n2.66.35.23232
YOLOv12n2.66.35.29227
SEMA-YOLO3.614.27.43185
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Z.; Zhen, H.; Zhang, X.; Bai, X.; Li, X. SEMA-YOLO: Lightweight Small Object Detection in Remote Sensing Image via Shallow-Layer Enhancement and Multi-Scale Adaptation. Remote Sens. 2025, 17, 1917. https://doi.org/10.3390/rs17111917

AMA Style

Wu Z, Zhen H, Zhang X, Bai X, Li X. SEMA-YOLO: Lightweight Small Object Detection in Remote Sensing Image via Shallow-Layer Enhancement and Multi-Scale Adaptation. Remote Sensing. 2025; 17(11):1917. https://doi.org/10.3390/rs17111917

Chicago/Turabian Style

Wu, Zhenchuan, Hang Zhen, Xiaoxinxi Zhang, Xuechen Bai, and Xinghua Li. 2025. "SEMA-YOLO: Lightweight Small Object Detection in Remote Sensing Image via Shallow-Layer Enhancement and Multi-Scale Adaptation" Remote Sensing 17, no. 11: 1917. https://doi.org/10.3390/rs17111917

APA Style

Wu, Z., Zhen, H., Zhang, X., Bai, X., & Li, X. (2025). SEMA-YOLO: Lightweight Small Object Detection in Remote Sensing Image via Shallow-Layer Enhancement and Multi-Scale Adaptation. Remote Sensing, 17(11), 1917. https://doi.org/10.3390/rs17111917

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop