Next Article in Journal
Image Motion and Quality in Polar Imaging with a Large Wide-Space TDI Camera
Previous Article in Journal
ICT-Net: A Framework for Multi-Domain Cross-View Geo-Localization with Multi-Source Remote Sensing Fusion
Previous Article in Special Issue
SSFAN: A Compact and Efficient Spectral-Spatial Feature Extraction and Attention-Based Neural Network for Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLO-SBA: A Multi-Scale and Complex Background Aware Framework for Remote Sensing Target Detection

1
Laboratory for Big Data and Decision, College of Systems Engineering, National University of Defense Technology, Deya Road, Changsha 410073, China
2
State Key Laboratory of Complex Electromagnetic Environment Effects on Electronics and Information Systems, College of Electronic Science, National University of Defense Technology, Deya Road, Changsha 410073, China
3
Center for Machine Vision and Signal Analysis, Faculty of Information Technology and Electrical Engineering, University of Oulu, 90014 Oulu, Finland
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(12), 1989; https://doi.org/10.3390/rs17121989
Submission received: 21 April 2025 / Revised: 27 May 2025 / Accepted: 3 June 2025 / Published: 9 June 2025

Abstract

Remote sensing target detection faces significant challenges in handling multi-scale targets, with the high similarity in color and shape between targets and backgrounds in complex scenes further complicating the detection task. To address this challenge, we propose a multi-Scale and complex Background Aware network for remote sensing target detection, named YOLO-SBA. Our proposed YOLO-SBA first processes the input through the Multi-Branch Attention Feature Fusion Module (MBAFF) to extract global contextual dependencies and local detail features. It then integrates these features using the Bilateral Attention Feature Mixer (BAFM) for efficient fusion, enhancing the saliency of multi-scale target features to tackle target scale variations. Next, we utilize the Gated Multi-scale Attention Pyramid (GMAP) to perform channel–spatial dual reconstruction and gating fusion encoding on multi-scale feature maps. This enhances target features while finely suppressing spectral redundancy. Additionally, to prevent the loss of effective information extracted by key modules during inference, we improve the downsampling method using Asymmetric Dynamic Downsampling (ADDown), maximizing the retention of image detail information. We achieve the best performance on the DIOR, DOTA, and RSOD datasets. On the DIOR dataset, YOLO-SBA improves mAP by 16.6% and single-category detection AP by 0.8–23.8% compared to the existing state-of-the-art algorithm.

1. Introduction

Remote sensing target detection is pivotal in various domains, including defense, agriculture, and urban development. Its core value lies in the ability to acquire information about the Earth’s surface in a non-contact, large-scale, and timely manner, providing important data for scientific decision-making and social development. In recent years, the development of deep learning has significantly advanced target detection technology [1,2,3,4]. The YOLO series [5,6,7,8,9,10,11] of detectors, while continuously improving detection accuracy, have effectively reduced model computational costs, providing efficient solutions for real-time applications and edge deployment. Sun et al.’s research [12] has demonstrated the great potential of YOLO in remote sensing target detection. However, the diversity of target scales and complex backgrounds in remote sensing images pose significant challenges to general target detection algorithms.
In remote sensing scenarios [13], it is common to have targets of different categories and scales or targets of the same category but different sizes within the same field of view. As shown in Figure 1, we can intuitively see that current target detection algorithms face significant challenges in handling multi-scale targets. These algorithms struggle to balance different sizes of targets, and the interference from complex backgrounds exacerbates the difficulty of detection, leading to frequent missed or false detections. In Figure 1a, where golf courses and running tracks are present, the algorithm exhibits insufficient sensitivity for small targets and excessive sensitivity for large targets, leading to false and missed detections. In Figure 1b, featuring dense urban structures, the intricate background of the remote sensing image creates a visually chaotic scene, where the color and shape similarities between targets and backgrounds make differentiation very challenging. In Figure 1c, under the coexistence of multi-scale variations and complex backgrounds, the scale differences between complex urban backgrounds and targets cause the algorithm to lose its ability to discriminate between targets and backgrounds, leading to severe false and missed detections. Hence, there is a pressing requirement to develop an advanced target detection algorithm capable of tackling the challenges arising from diverse target scales and complex backgrounds in remote sensing.
To tackle the challenge of detecting targets at multiple scales, researchers have explored various approaches. Zhang et al. [15] introduced a label assignment approach centered on the central region, aiming to provide additional positive samples for small targets, thereby enhancing the instance and feature level representations of these targets. Zhong et al. [16] introduced the PSWP-DETR detector, which integrates multi-scale features through a scale-aware difference module to capture edge features and local textures. However, these methods often overlook the interplay between context information extraction and feature fusion when handling multi-scale targets, thus not fully exploiting the processing advantages of advanced attention mechanisms for multi-scale information fusion. In terms of complex background processing, Yang et al. [17] proposed the Frequency Self-Attention Refinement module to eliminate redundant information across various frequency bands. Zhang et al. [18] proposed the Balanced Decoupled Loss for handling foreground and background information losses. However, these algorithms often ignore the extraction and enhancement of detail information while suppressing background redundancy. The suppression of complex backgrounds and the enhancement of multi-scale target feature saliency should complement each other, but existing methods have not adequately addressed their relationship.
To address the aforementioned challenges, this paper proposes a multi-scale and complex background-aware network for remote sensing target detection, named YOLO-SBA. We adopt the network structure of YOLO for efficient feature extraction, achieving channel-level enhancement for multi-scale targets and fine-grained suppression of complex backgrounds, while maximizing the retention of image detail information to significantly improve the capability of remote sensing target detection. Specifically, we proposed the Multi-Branch Attention Feature Fusion Module (MBAFF) to efficiently fuse multi-scale information and improve feature weakening issues by combining multi-branch dilated convolutions with the Bilateral Attention Feature Mixer (BAFM). Next, the Gated Multi-scale Attention Pyramid (GMAP) suppresses spectral redundancy through channel–spatial dual reconstruction mechanisms and enhances target saliency expression while suppressing background noise through adaptive gating fusion. Additionally, we improved the downsampling method for global network inference, Asymmetric Dynamic Downsampling (ADDown), by combining decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism to enhance edge sensing and prevent information loss. These innovations enable the model to fully learn key information during global network inference, enhancing its effectiveness and robustness, thereby improving the performance of remote sensing target detection.
Enhancing the saliency of multi-scale targets effectively reduces background interference, while complex background suppression improves the distinguishability of targets at all scales. These two aspects complement each other, jointly improving the effectiveness of remote sensing target detection.
Breakthroughs and Contributions of This Paper:
1. This study introduces a multi-scale and complex background-aware network for remote sensing target detection, termed YOLO-SBA. This algorithm is designed to tackle the inherent difficulties posed by variations in target scales and complex backgrounds in the context of remote sensing target detection. It extends the model’s receptive field, enhances the feature representation of targets at multiple scales, and reduces the noise redundancy in complex backgrounds.
2. We proposed the Multi-Branch Attention Feature Fusion Module (MBAFF). It combines a multi-branch dilated convolution structure with the Bilateral Attention Feature Mixer (BAFM) to capture contextual information of feature maps at different spatial scales. By incorporating an efficient attention mechanism, it enhances the expression of target features, mitigating the feature weakening issue caused by target multi-scale variations.
3. We proposed the Gated Multi-scale Attention Pyramid (GMAP). By feeding the feature maps derived from the multi-level receptive field pyramid into a channel–spatial dual reconstruction mechanism and a learnable gating mechanism for fusion attention, it achieves adaptive background noise suppression while enhancing features.
4. We proposed Asymmetric Dynamic Downsampling (ADDown). By combining decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism, this mechanism fully retains the key information extracted by each module during downsampling, significantly enhancing the model’s robustness.
5. We performed comprehensive experiments on publicly available remote sensing target detection datasets, including DIOR [14], DOTA-v1.0 [19], and RSOD [20], demonstrating that the proposed algorithm YOLO-SBA outperforms existing target detection algorithms. On the DIOR [14] dataset, YOLO-SBA achieved a 16.6% improvement in mAP and a 0.8–23.8% improvement in single-category detection AP compared to the current state-of-the-art algorithm. Additionally, ablation studies were conducted to verify the efficacy of the proposed modules.

2. Related Work

This section primarily explores the three issues that this paper focuses on: target size variability, complex background interference, and the improvement of downsampling algorithms. It provides a detailed introduction to the research background and development history of each issue.

2.1. Remote Sensing Image Target Detection Algorithms for Multi-Scale Targets

Compared to conventional targets, remote sensing targets exhibit greater variations in scale. This variability is not only evident between different categories of targets, such as aircraft and vehicles, which are typically smaller, and buildings and playgrounds, which are typically larger. It also occurs within the same category of targets, where significant differences in size can be observed [21,22]. To address this, some researchers focus on learning multi-scale features.
Li et al. [23] introduced an enhanced R 3 D e t algorithm, replacing the traditional feature pyramid network with a search-based architecture. This enables the network to adaptively learn and choose feature combinations for updates, thereby enriching multi-scale feature information. Moreover, they integrated shallow features into the context information enhancement module to enhance the semantic information of small targets, facilitating the detection of water surface vessels. Teng et al. [24] introduced a Multi-Scale Perception (MSP) module, combining the global contextual cues it extracts with the local spatial contextual correlations encoded by Clip Long Short-Term Memory. They further utilized these rich semantic features to develop an Adaptive Anchor module, alleviating the scale variations of remote sensing image targets. Huang et al. [25] employed a feature pyramid structure with multi-scale feature representations, integrating deformable convolution modules to learn multi-scale features of targets. Zhou et al. [26] divided each convolutional layer within the multi-scale structure into sub-layers of equal size. By obtaining super-scale features through multi-scale sub-layers and enhancing information transmission within convolutional layers, they improved multi-scale target detection. Li et al. [27] introduced the enhanced feature pyramid network architecture, enabling multi-scale detection through the creation of top-down feature maps. Zhong et al. [16] handle multi-scale information by designing scale-aware PdConv and multi-scale feature extraction SDM. Guo et al. [28] achieve multi-scale feature fusion by calculating cross-attention between low-resolution feature maps and high-resolution feature maps, transferring the semantic information of high-resolution feature maps to low-resolution feature maps.
Existing algorithms have achieved remarkable advancements in multi-scale feature extraction and enhancement. Nevertheless, they have paid less attention to the interplay between context information extraction and feature fusion during the feature extraction process. The effective integration of advanced attention mechanisms with multi-scale feature extraction remains largely unexplored. Therefore, in this paper, we introduce MBAFF, which combines a multi-branch dilated convolution structure with a bilateral attention mechanism. This method incorporates effective attention mechanisms into multi-scale feature integration, comprehensively comprehending image context and enhancing detection performance for targets of different scales.

2.2. Remote Sensing Image Target Detection Algorithms for Complex Backgrounds

Remote sensing images, with their extensive coverage, encompass numerous targets across diverse categories. Yet, only a limited number of these categories are annotated for detection, leading to the background dominating the image [29,30,31,32,33,34]. Targets in remote sensing are frequently embedded in intricate backgrounds, presenting difficulties for detection.
To address this, researchers typically focus on reducing background interference. Yang et al. [35] introduced an automatic aircraft target detection algorithm that fuses multi-scale circular frequency filters with convolutional neural networks. This algorithm utilizes multi-scale circular frequency filters to remove noise from images and extract candidate regions of aircraft targets at different scales. Ye et al. [36] aimed to suppress background clutter interference by fusing features from different layers in a hierarchical manner. Subsequently, the combined features were mapped into a reduced-dimensionality subspace via a feature selection mechanism to capture more pertinent contextual details. Wang et al. [37] introduced a background filtering module grounded in the principle of attention mechanisms to mitigate background interference in optical remote sensing images. Gao et al. [38] handle complex background interference by designing a spatial bias module (SBM) and a multi-task enhancement structure (MES). Ma et al. [39] reduce mutual interference between different categories by dividing the feature maps into different channels and separating features according to categories, effectively suppressing background interference and improving detection accuracy. Yang et al. [17] propose BAC-FSAR, which achieves filtering of different frequency information by introducing feature processing in the frequency domain, using channel-adaptive pooling as a low-frequency feature filter, and using deep convolution to extract high-frequency features.
Similar to the research problem addressed in this paper, TMAFNet [33], when handling complex background tasks, does not design an independent background noise suppression algorithm. Instead, it indirectly suppresses redundant information through target enhancement and selective fusion. In contrast, the method proposed in this paper for dealing with complex backgrounds is significantly different and has been experimentally proven to outperform TMAFNet [33] in terms of performance on the DIOR [14] dataset.
Existing background suppression algorithms often focus solely on eliminating background noise or merely suppressing noise by enhancing target features. However, if both target feature saliency enhancement and background suppression can be performed simultaneously, it would be more conducive to suppressing redundant information and thus improving overall performance. In this study, we propose GMAP, which inputs multi-receptive field feature maps extracted by a cascaded pooling structure into a channel–spatial dual reconstruction mechanism. Through a gating fusion mechanism, GMAP enhances the saliency of target features while better suppressing redundant information from complex backgrounds.

2.3. Evolution of Downsampling Algorithm Improvements

With the widespread application of deep learning, numerous researchers have focused on enhancing downsampling techniques to retain valuable information during network inference and improve the accuracy of subsequent tasks.
Son et al. [40] introduced a degradation modeling approach using adaptive downsampling techniques. This method utilizes adversarial training strategies and adaptive data loss to dynamically capture the low-frequency distribution characteristics of real low-resolution images. This allows the super-resolution network to adaptively adjust downsampling parameters to match real complex degradation patterns. Yang et al. [41] compressed the original high-resolution images to extremely low resolutions through a multi-scale residual structure. They also utilized a joint feature representation of directional coding features and local binary patterns to retain texture information under extreme dimensionality reduction conditions.
However, these methods overlook the need for fine-grained information in remote sensing images and are prone to losing key information extracted during network inference, which in turn affects the accuracy of target detection. Therefore, this paper proposes Asymmetric Dynamic Downsampling (ADDown), which combines decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism. This enhances independent sensing of horizontal and vertical edges, achieving efficient utilization of spatial information and adaptive enhancement of channel features. During the downsampling phase of global network inference, it fully retains the detail information of the image.
This study tackles the difficulties posed by variations in target scales and intricate backgrounds in remote sensing target detection by incorporating algorithm designs that enhance the prominence of multi-scale targets and mitigate complex background interference. It constructs a multi-scale and complex background-aware network called YOLO-SBA, significantly improving target detection accuracy.

3. Method

This part offers an in-depth explanation of the YOLO-SBA algorithm, tackling the two inherent significant issues in remote sensing target detection: multi-scale target distribution and complex background interference. We adopt the network structure of YOLOv10 for efficient feature extraction, achieving channel-level enhancement for multi-scale targets and fine-grained suppression of complex backgrounds, while preserving image detail information to substantially enhance remote sensing target detection capabilities.
As shown in Figure 2, YOLO-SBA first processes the extracted feature maps through the Multi-Branch Attention Feature Fusion Module (MBAFF), combining the multi-branch information extracted by dilated convolutions with the Bilateral Attention Feature Mixer (BAFM) to facilitate contextual feature integration and improve the feature representation for targets at various scales. Then, it utilizes the Gated Multi-scale Attention Pyramid (GMAP) to process the differentiated multi-level receptive field features through a channel–spatial dual reconstruction mechanism, finely suppressing spectral redundancy. Through the integration of a gating fusion mechanism, the model’s capability to differentiate targets from backgrounds is improved, thereby reducing background interference. Additionally, Asymmetric Dynamic Downsampling (ADDown) achieves efficient utilization of spatial information and adaptive enhancement of channel features by combining decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism, ensuring that the fine-grained effective information extracted by key modules is not lost during downsampling. Ultimately, the detection task is executed via the YOLOv10 [6] detection head. Thanks to the careful design of the network architecture and feature processing, YOLO-SBA demonstrates superior ability in differentiating target information from redundant backgrounds, resulting in improved performance in remote sensing target detection.
In addition, we retained two classic modules of YOLOv10 [6], namely PSA (Partial Self-Attention) and C2fCIB, which are briefly introduced below. PSA is an efficient self-attention mechanism introduced in YOLOv10. It divides the feature map extracted from the backbone into two parts and only inputs one part into a submodule consisting of multi-head self-attention and a feed-forward network. Then, the two parts of the feature map are connected and fused through convolution operations to efficiently capture global features. C2fCIB is an efficient feature extraction module proposed in YOLOv10 [6]. It uses cheap depthwise convolution for spatial mixing and efficient pointwise convolution for channel mixing to achieve more efficient feature extraction. C2fCIB divides the input feature map into two parts, one of which is processed by CIB, and the other is directly output. Finally, the two parts are fused through convolution to improve model performance and efficiency, saving computing resources while enhancing feature representation capabilities.

3.1. Multi-Branch Attention Feature Fusion Module (MBAFF)

The MBAFF (Multi-Branch Attention Feature Fusion) module is an innovative feature fusion structure designed for tasks involving multi-scale variations in targets. The core of this structure consists of two primary components: the Multi-Branch Dilated Convolution Structure and the Bilateral Attention Feature Mixer (BAFM). Figure 3 illustrates that the operational procedure of this module is divided into two crucial phases. Initially, the input feature map is processed through a parallel multi-branch dilated convolution mechanism to capture features across multiple scales. Subsequently, the BAFM component adaptively fuses the heterogeneous features through a bilateral attention mechanism. During the feature extraction phase, the multi-branch structure utilizes dilated convolution operations with varying dilation rates. Through the parallel computation of differentiated receptive fields, it effectively captures the contextual information of the feature map at various spatial scales. This architectural design overcomes the scale constraints of conventional single-branch convolutions, substantially improving the feature responsiveness of targets and boosting the expressiveness of multi-scale target features.
The multi-branch dilated convolution structure substantially improves the feature representation ability of targets across various scales via the combined approach of multi-scale feature extraction and dynamic receptive field enlargement. This structure is composed of multiple parallel branches, where each branch processes through D-Bottleneck as shown in Figure 4 and cascaded dilated convolutions. Its core is manifested in three aspects:
1. By adaptively adjusting the dilation rate (d) parameter, the structure accomplishes an exponential increase in the receptive field while preserving the spatial resolution of the feature map. For instance, with a 3 × 3 convolution kernel and a default configuration of d = 2 (illustrated in Figure 5), the effective receptive field of a single convolutional layer expands from 3 × 3 to 5 × 5. As shown in Figure 4, after passing through three D-Bottleneck modules, the accumulated receptive field reaches 25 × 25. This progressive expansion mechanism effectively captures the cross-regional correlation features of multi-scale targets.
2. The multi-branch structure realizes heterogeneous multi-scale feature representation. The base branch (Branch_1) retains the detailed texture at the original resolution, while the deeper branches adopt a step-wise increasing dilation rate to encode global information. This design enables the network to simultaneously perceive local features and global contextual information.
3. By fusing the outputs of branches with different receptive fields through concatenation operations and then utilizing feature reorganization along the channel dimension, the multi-scale representation capability is enhanced.
Specifically, the feature map X R B × C × H × W input into the MBAFF first undergoes channel expansion via a 1 × 1 convolution. It is then split into two base branches, B r a n c h _ 1 and B r a n c h _ 2 , and iteratively generates new branches through n cascaded D-bottleneck modules. Here, the D-bottleneck denotes a residual convolutional block with a dilation rate. Subsequently, all branches { B r a n c h 1 , B r a n c h 2 , , B r a n c h n + 2 } are concatenated along the channel dimension. In each D-bottleneck unit, we expand the receptive field through dilated convolutions, with the computational process as follows:
B r a n c h k = Conv 3 × 3 ( B r a n c h k , dilation = d )
B r a n c h k + 1 = Conv 3 × 3 ( B r a n c h k , dilation = d )
The expansion rate d controls the receptive field range, and in this paper, the expansion rate d = 4 . The effective receptive field is expanded to ( d + 1 ) ( c 1 ) + 1 (c is the size of the convolutional kernel), and the basis for selecting the expansion rate is detailed in Section 4.5 through ablation experiments. Lastly, a convolutional layer with reduced filters restores the initial channel dimension.
The multi-branch dilated convolution structure substantially enhances the detection capability of targets at different scales in remote sensing images by simultaneously optimizing feature resolution and receptive field. This architecture, while preserving the spatial resolution of the input feature map, improves the model’s adaptability to adapt to targets of varying sizes via the use of different dilation rates in parallel branches. It addresses the issue of feature weakening in the distribution of multiple targets. The tensor obtained after concatenating multi-scale features is then input into the Bilateral Attention Feature Mixer (BAFM) for cross-dimensional feature refinement, as shown in Figure 6.
The Bilateral Attention Feature Mixer (BAFM) accomplishes accurate integration of features across layers using a dual-path attention mechanism, as depicted in Figure 6.
This module receives two input feature maps: Feature Map X R B × C × H × W (original input feature) and Residual Feature Map Y R B × C × H × W (residual feature extracted by the Multi-Branch Dilated Convolution Structure). The feature fusion process consists of two core steps: initially, the input features undergo cross-modal initialization, where feature map X is element-wise added to feature map Y to generate the baseline feature matrix. The baseline feature matrix, by preserving the common information between feature map X and feature map Y, constructs a feature base with rich feature information. Subsequently, we input the baseline feature matrix into both the local attention path and the global attention path to compute local and global attention weights, respectively.
In the local attention path, local feature encoding is first performed. The local feature encoding process, designed in this study, includes four processing stages: Initially, channel compression reduces the original number of channels to C / r , extracting local inter-channel correlations. Subsequently, a ReLU activation function introduces spatial non-linearity through non-linear mapping, enhancing the expression capability of local patterns. Then, convolutional computation restores the original channel dimension, preserving the integrity of local features. Finally, the processed result is input into the Sigmoid function to produce a spatial attention weight matrix, thus completing the calculation of local attention weights.
In the network structure of the global attention computation path, the hierarchical compression and adaptive weighting processes are closely integrated and performed continuously. The main steps are as follows: Firstly, the input feature map is compressed in the spatial dimension through a global average pooling (GAP) layer, resulting in a global average feature for each channel. This converts the two-dimensional feature map into a scalar value for each channel, achieving preliminary extraction and compression of global contextual information.
The tensor output after global average pooling encodes the global spatial response intensity of each channel. Subsequently, through dimensionality reduction, ReLU activation function for non-linear transformation, and channel dimension reconstruction, it captures the cross-channel dependencies, realizes non-linear interaction between channels, and generates channel attention weights. Finally, the global attention weights are combined with the local attention weights, and a Sigmoid function is utilized to produce the global–local dual-path combined attention map.
By adaptively fusing the input features Feature map X and Residual Feature map Y based on the computed fusion weights from the dual pathways, the network can dynamically adjust the fusion ratio according to the feature importance, thereby further enhancing the feature expression capability for targets at different scales.

3.2. Gated Multi-Scale Attention Pyramid (GMAP)

The Gated Multi-scale Attention Pyramid (GMAP) module is developed to tackle the challenge of complex backgrounds in remote sensing images. It proposes an innovative architecture that combines cascaded pooling with a dual attention reconstruction mechanism. As shown in Figure 7, GMAP consists of three core components: 1. Cascaded Pooling Multi-scale Feature Extraction. This component constructs a multi-level receptive field feature pyramid. 2. Channel–Spatial Dual Reconstruction Mechanism. This mechanism suppresses spectral redundancy, weakens the interference of complex backgrounds, and enhances the saliency of targets. 3. Adaptive Gating Fusion. This component dynamically balances the weights of multi-scale features. The module’s computational procedure can be succinctly described as follows: initially, the Feature Map is fed into a cascaded pooling structure to perform multi-scale feature extraction.The extracted feature tensors undergo channel reconstruction and spatial reconstruction, respectively. Finally, a gating mechanism computes fusion weights to direct the integration of the enhanced multi-scale features. Following this, we will offer an in-depth explanation of the cascaded pooling architecture, the dual reconstruction approach, and the gating fusion method.

3.2.1. Cascaded Pooling Structure

Given the input feature map { F i n R B × C × H × W } , the GMAP employs three cascaded max pooling operations to respectively obtain the outputs y 1 , y 2 , and y 3 . Generate a three-level feature pyramid { F in , y 1 , y 2 , y 3 } that includes the original input features, with its effective receptive field expanding exponentially:
R = { r , 3 r 2 , 5 r 4 , 7 r 6 } ,
where r is the base receptive field. When the kernel size k = 5 , the sequence of receptive fields is 5 × 5 9 × 9 13 × 13 17 × 17 , covering multi-scale context from global to local. The multi-scale feature tensor obtained through the cascaded pooling structure contains more extensive and rich information compared to the original feature map.

3.2.2. Dual Reconstruction Mechanism

Channel Reconstruction
We enhance the feature representation capability related to the target by suppressing redundant channels in multispectral remote sensing images through channel reconstruction. Firstly, we split the input X along the channel dimension into two sub-features, X 1 and X 2 . Subsequently, we apply grouped depthwise convolutions to each branch to obtain X ^ i . The X ^ i is then input into the SE (Squeeze-and-Excitation) mechanism to adaptively modify the channel weights:
z = 1 H W h = 1 H w = 1 W X ^ i ( h , w )
s = σ ( W 2 δ ( W 1 z ) )
X i att = s X ^ i
Equations (4)–(6) describe the core computational steps of the SE (Squeeze-and-Excitation) attention mechanism. First, through global average pooling (GAP) as described in Equation (4), the input feature map X ^ i is averaged across the spatial dimensions (height H and width W), resulting in a global response z for each channel. Then, using the channel weighting computation in Equation (5), a channel weight vector s is calculated through two fully connected layers (with weight matrices W 1 and W 2 for the first and second layers, respectively) and ReLU and Sigmoid activation functions. This vector s is employed to rescale each channel of the input feature map. Eventually, via the feature rescaling in Equation (6), the channel weight vector s undergoes element-wise multiplication (denoted by ⊙) with the input feature map X ^ i , yielding the attention-refined feature map X i att . This mechanism adaptively modulates the response of each channel, allowing the network to concentrate on the most task-relevant features, thereby improving model performance. The channel reconstruction weights X Channel   Reconstruction are obtained by adding the attention weights calculated based on X 1 and X 2 .
Space Reconstruction
We enhance the spatial saliency of small targets and suppress background clutter through spatial reconstruction.
Here, we compute the spatial attention map through cross-channel aggregation A R B × 1 × H × W .
A average = 1 C c = 1 C X Channel   Reconstruction ( c )
A max = max c X Channel   Reconstruction ( c )
A = σ ( Conv 7 × 7 ( [ A average ; A max ] ) )
We then apply secondary weighting to the attention map obtained from channel reconstruction:
X Space   Reconstruction = A X Channel   Reconstruction
By utilizing the dual-channel reconstruction mechanism, we amplify the prominence of target feature representations while mitigating the impact of non-target areas.

3.2.3. Gated Fusion

We facilitate the integration of the enhanced multi-scale feature tensor via the creation of gating signals.
For each scale feature f i { f 0 , f 1 , f 2 , f 3 } : Initially, we compute the weight values through Global Average Pooling (GAP). Subsequently, we introduce a learnable method for generating gating weights:
g i = sigmoid ( W g e i + b g )
Here, both W g and b g are learnable parameters of the network, utilizing the sigmoid function as the activation mechanism to produce the gating weights.
Finally, the weights derived from the gating process direct the weighted integration of the multi-scale feature tensor:
w i = exp ( g i ) j = 0 3 exp ( g j ) ,         F Fusion = i = 0 3 w i · f i
Based on the previous description, we know that here, f 0 dominates small target detection, and f 3 dominates large area target recognition. Utilizing the Gated Multi-scale Attention Pyramid (GMAP), we amplify the prominent features of targets while effectively reducing the superfluous details from intricate backgrounds.

3.3. Asymmetric Dynamic Downsampling (ADDown)

To ensure that the key modules better fulfill their roles and that the effective information they extract is not lost during downsampling, this paper proposes ADDown, as shown in Figure 8. ADDown combines asymmetric decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism. This enhances the independent sensing of horizontal and vertical edges, effectively prevents information loss, and achieves efficient utilization of spatial information and adaptive enhancement of channel features. During the downsampling phase of global network inference, this mechanism fully retains the extracted key information, significantly enhancing the model’s robustness.
Figure 8 demonstrates that the ADDown module utilizes a two-step feature transformation framework: Initially, channel dimension alignment is achieved via a 1 × 1 convolutional layer; subsequently, the input undergoes downsampling through asymmetric depthwise separable convolution; finally, feature calibration is accomplished via dynamic compression-ratio channel attention. Below, we detail the computational workflow of ADDown.
The input feature map X R B × C 1 × H × W undergoes a learnable 1 × 1 convolution to accurately modify the number of feature channels, yielding X R B × C 2 × H × W .
Following the adjustment of the channel count to obtain X , it is fed into a decomposed depthwise separable convolution to improve the response ability towards target features while performing spatial downsampling.
X h = f 3 × 1 ( X ) = i = 1 1 W h [ i , 0 ] · X [ : , : , h + i , w ]
X v = f 1 × 3 ( X h ) = j = 1 1 W v [ 0 , j ] · X h [ : , : , h , w + j ]
In the equation, W h R C 2 × 3 and W v R C 2 × 3 represent the depthwise convolution kernels for the horizontal and vertical directions, respectively. Through separable convolution, these kernels achieve an equivalent receptive field of 5 × 5 .
R e f f = R 3 × 1 R 1 × 3 = 3 + ( 3 1 ) = 5
Why Choose Decomposed Depthwise Separable Convolution? The square structure of a traditional 3 × 3 convolutional kernel is susceptible to variations in shape orientation. When the target shape has varied directions, its edge response will diminish due to the fixed geometric layout of the convolutional kernel. Decomposed convolution achieves spatial decoupling through independent horizontal and vertical convolutions, as shown in Formulas (16)–(18).
R ( θ ) = i = 1 1 j = 1 1 W i , j · X x + i , y + j
R h = i = 1 1 W h , i · X x + i , y
R v = j = 1 1 W v , j · X x , y + j
This separable operation enables the convolution kernel to respond independently to horizontal/vertical features. When the target direction or shape changes by θ degrees, the response intensity of the original 3 × 3 convolution decreases due to the angle deviation, following a cos θ attenuation pattern. In contrast, decomposed convolution maintains rotational sensitivity through dynamic weight combination ( W h · cos θ + W v · sin θ ).
Finally, through feature enhancement with an attention mechanism of adaptive compression ratio, the response to effective information is further enhanced.
z = 1 H W h = 1 H w = 1 W X [ : , : , h , w ]
q = δ ( W 2 ( δ ( W 1 z ) ) )
A = σ ( q )
Here, W 1 R max ( C 2 / 16 , 4 ) × C 2 and W 2 R C 2 × max ( C 2 / 16 , 4 ) form the dynamic bottleneck layer. When C 2 < 64 , the minimum intermediate dimension of 4 channels is maintained to prevent excessive compression and potential information loss.

4. Experiments

In this section, we perform comprehensive experiments on three publicly available remote sensing target detection benchmark datasets: DIOR [14], DOTAv1.0 [19], and RSOD [20]. These experiments demonstrate the superiority of the YOLO-SBA introduced in this paper and the efficacy of the three modules: Multi-Branch Attention Feature Fusion Module (MBAFF), Gated Multi-scale Attention Pyramid (GMAP), and Asymmetric Dynamic Downsampling (ADDown). Section 4.1 presents the three datasets widly used in remote sensing object detection; Section 4.2 outlines the evaluation criteria used for experimental validation; in Section 4.3, we elaborate on the experimental equipment information and parameter settings; in Section 4.4, we compare our approach with state-of-the-art target detection algorithms used in remote sensing, demonstrating that the YOLO-SBA algorithm presented in this study is at the forefront of research; in Section 4.5, we performed ablation studies to validate the efficacy of the proposed modules; and in Section 4.6, we visually analyze the superiority of the algorithm and its detection capability on remote sensing images.

4.1. Datasets

This study performs experiments on three prominent remote sensing object detection datasets: DIOR [14], DOTAv1.0 [19], and RSOD [20].
DIOR [14] datasets. The DIOR [14] dataset, released by China’s Northwestern Polytechnical University in 2020, serves as a comprehensive benchmark for object detection in optical remote sensing images. It comprises 23,463 images with an 800 × 800 pixel resolution, spanning 20 categories including airplane (AL), airport (AT), baseball field (BF), basketball court (BC), bridge (B), chimney (C), dam (D), expressway service area (ESA), expressway toll station (ETS), harbor (HB), golf course (GC), ground track field (GTF), overpass (O), ship (S), stadium (SD), storage tank (ST), tennis court (TC), train station (TS), vehicle (V), and windmill (WM). The dataset includes 190,288 annotated horizontal bounding box instances across various spatial resolutions (0.5 to 30 m), encompassing urban, rural, mountainous, and coastal scenes. With its diverse categories and numerous small targets, DIOR is widely employed for assessing target detection algorithm performance.
DOTAv1.0 [19] datasets. The DOTAv1.0 [19] dataset, released by Wuhan University in 2018, serves as a benchmark for target detection in aerial images. It includes 2806 remote sensing images with a resolution of 4000 × 4000 pixels (some varying from 800 × 800 to 20,000 × 20,000 pixels). The dataset includes 15 common target categories: Plane (PL), Baseball Diamond (BD), Bridge (BR), Ground Track Field (GTF), Small Vehicle (SV), Large Vehicle (LV), Ship (S), Tennis Court (TC), Basketball Court (BC), Storage Tank (ST), Soccer-ball Field (SBF), Roundabout (R), Harbor (HB), Swimming Pool (SP), and Helicopter (HC). It contains a total of 188,282 annotated bounding boxes. In this paper, the DOTA [19] dataset is processed by cropping and stitching to ensure that the image resolution does not exceed 1024 × 1024. Then, 21,046 images are randomly selected from the processed dataset for experiments.
RSOD [20] datasets. The RSOD [20] (Remote Sensing Object Detection) dataset, released by Wuhan University in 2015, functions as an open benchmark for target detection in remote sensing imagery. It includes four primary target categories: aircraft (446 images with 4993 annotations), oil tanks (165 images with 1586 annotations), playgrounds (189 images with 191 annotations), and overpasses (176 images with 180 annotations). The dataset contains a total of 976 images and 19,950 annotated instances, each with bounding box coordinates and category labels. The images have an 800 × 800 pixel resolution and depict various scenes including urban areas, suburban regions, and transportation hubs. The target sizes vary considerably, from small oil tanks around 10 pixels to large airport facilities spanning several hundred pixels.

4.2. Evaluation Metrics

We adopt mAP [42] (Mean Average Precision) as the primary evaluation metric to assess the model’s performance. The computation of this metric depends on the equilibrium between Precision and Recall, defined as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
Here, TP (True Positive) represents the number of correct detections where the Intersection over Union (IoU) between the predicted bounding box and the ground truth exceeds a threshold (commonly 0.5). FP (False Positive) indicates the number of incorrect detections with insufficient or excessive IoU, whereas FN (False Negative) denotes the number of missed true targets. The IoU is calculated as the ratio of the overlapping area to the total area of the predicted and actual bounding boxes:
I o U = A B A B
For each category, the predicted bounding boxes are sorted in descending order according to their confidence levels. Then, Precision and Recall values are computed progressively at various confidence thresholds to generate the P-R curve (with Recall on the x-axis and Precision on the y-axis). The area beneath this curve represents the AP (Average Precision) for that category:
A P = 0 1 P ( r )   d r
We also introduce size-based evaluation metrics: APs (Average Precision for small targets), APm (Average Precision for medium targets), and APl (Average Precision for large targets), to comprehensively assess the model’s ability to detect targets across various scales. APs is for small targets with a pixel area less than 32 × 32; APm evaluates the detection performance for medium targets with a pixel area between 32 × 32 and 96 × 96; APl evaluates the detection performance for large targets with a pixel area greater than 96 × 96.

4.3. Training Details

This study follows the efficient design principles of the Ultralytics [10] framework for experimental setup and parameter configuration. The SGD optimizer is used with an initial learning rate of l r 0 = 0.01 , a momentum of 0.937, and a weight decay of 5 × 10 4 . The batch size ranges from 4 to 16 to balance memory usage and convergence stability. The input size is set to 1024 × 1024 . Data augmentation techniques include Mosaic augmentation followed by RandomPerspective, which involves rotation, shearing, and perspective transformations. The probability parameters, such as scale, are set to 0.5, while shear is set to 0.0. Experiments are performed on an Ubuntu system equipped with an NVIDIA GeForce RTX 4090 GPU and 64 GB RAM.

4.4. Comparison with State-of-the-Art Algorithms

In this part, we evaluate our approach against traditional target detection methods and cutting-edge algorithms across three datasets.
Comparison on DIOR [14] dataset. On the DIOR [14] dataset, we evaluated our approach against 18 sophisticated algorithms, as detailed in Table 1. The YOLO-SBA algorithm introduced in this paper achieved an mAP of 89.8% on the DIOR [14] dataset, marking a 16.6% improvement over the current leading AGMF-Net [38] algorithm. Among the 20 categories in the DIOR [14] dataset, we achieved the best detection accuracy in all 20 categories, respectively: 97.5% (Airplane, AL), 93.3% (Airport, AT), 96.5% (Baseball, BF), 92.1% (Basketball Court, BC), 71.2% (Bridge, B), 89.9% (Chimney, C), 85.6% (Dam, D), 96.9% (Expressway Track Field, ESA), 91.8% (Expressway Toll Station, ETS), 90% (Golf Course, GC), 93.2% (Ground Track Field, GTF), 77.9% (Harbor, HB), 76.7% (Overpass, O), 95.6% (Ship, S), 97.5% (Stadium, SD), 93.1% (Storage Tank, ST), 97.2% (Tennis Court, TC), 82.3% (Train Station, TS), 81.8% (Vehicle, V), and 96.8% (Windmill, WM).
Comparison on DOTA [19] dataset. We evaluated our approach against 18 cutting-edge algorithms on the DOTA [19] dataset, as detailed in Table 2. The YOLO-SBA algorithm presented in this paper achieved a mean Average Precision (mAP) of 76.14% on the DOTA [19] dataset, demonstrating a 0.81% improvement over the current leading FSoD-Net [43] algorithm. Among the 15 categories in the DOTA [19] dataset, we achieved the best detection accuracy in 4 categories, respectively: 86.8% (Harbor, HB), 88.7% (Large Vehicle, LV), 65.1% (Soccer-ball field, SBF), and 95.6% (Tennis Court, TC).
Table 1. Comparison with state-of-the-art algorithms on the DIOR [14] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 1. Comparison with state-of-the-art algorithms on the DIOR [14] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPALATBFBCBCDESAETSGCGTFHBOSSDSTTCTSVWM
YOLT [44]60.364.769.032.887.832.371.545.854.955.865.666.349.949.987.730.373.382.029.952.474.0
Faster-RCNN [45]53.651.461.662.280.727.074.237.353.445.169.661.843.749.056.141.839.673.944.734.065.3
Cascade R-CNN [46]67.481.281.490.181.146.981.568.184.265.374.581.637.860.468.988.860.581.057.547.780.7
HSF-Net [47]54.951.062.160.280.727.472.843.852.145.267.558.734.048.970.845.553.572.946.837.467.0
SSD [42]60.992.948.377.977.235.766.646.547.958.537.353.747.939.959.677.580.888.355.638.886.7
YOLOv3 [5]59.967.554.765.886.834.273.534.355.749.667.368.954.351.786.840.367.883.932.349.173.6
YOLOv4 [7]65.978.369.769.688.338.776.948.657.657.974.170.652.856.889.450.477.583.346.055.775.9
RFB-Net [48]62.693.848.380.277.740.767.547.448.570.037.354.948.243.166.178.080.089.155.640.488.3
RetinaNet [49]65.263.447.883.378.448.267.949.453.273.666.359.147.645.772.082.480.790.155.047.792.9
CenterNet [50]61.664.066.365.786.334.873.141.160.854.273.066.045.353.381.353.563.780.944.146.378.8
EfficientDet D0 [51]59.991.947.379.276.036.066.546.865.647.537.053.345.839.854.977.276.387.354.431.985.1
EfficientDet D2 [51]61.572.566.264.387.033.274.043.171.854.355.652.737.447.086.166.870.981.150.243.373.0
EfficientDet D4 [51]66.186.557.475.785.233.575.465.680.167.458.371.435.650.678.890.361.882.954.630.081.5
S3FD [52]54.771.746.372.775.531.071.947.951.749.667.270.923.045.144.361.343.980.030.034.374.8
M2Det [53]57.663.367.671.180.232.172.558.062.352.569.863.541.352.035.966.335.972.252.425.078.3
FSoD-Net [43]71.888.966.986.890.245.579.648.286.975.567.077.353.659.778.369.975.091.452.352.090.6
TMAPNet [33]72.992.277.775.091.347.178.653.667.166.278.576.364.961.490.572.275.490.762.155.283.2
AGMF-Net [38]73.290.972.879.389.744.781.459.366.062.773.879.265.061.791.778.675.890.760.058.083.1
YOLO-SBA (ours)89.897.593.396.592.171.289.985.696.991.890.093.277.976.795.697.593.197.282.381.896.8
Table 2. Comparison with state-of-the-art algorithms on the DOTA [19] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 2. Comparison with state-of-the-art algorithms on the DOTA [19] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPBDBCBGTFHBHLVPRSSVSBFSTSPTC
YOLT [44]61.4460.570.9322.0937.5456.5822.1670.4887.2057.6284.4370.7552.7078.9655.7693.94
Faster-RCNN [45]56.1363.7050.1029.6054.8665.7148.3762.5976.1540.9167.728.5955.8467.3343.6486.89
HSF-Net [47]62.3669.8869.4037.7057.9866.8441.7864.2479.9652.8171.8466.7647.4761.7959.1587.87
SSD [42]50.1932.8339.4023.0238.8265.2122.1662.8485.3734.8352.3133.4452.2967.3355.7387.30
YOLOv3 [5]65.6768.7866.8245.9351.9274.0356.6860.6793.9145.0185.5650.1252.4583.4755.8593.88
YOLOv4 [7]67.9772.8260.0470.0072.2764.5360.2167.3977.0979.2978.9749.7661.4560.9660.9783.81
RFB-Net [48]64.0349.7663.8335.7436.3563.2120.5377.5387.7567.9390.0578.0033.5789.5472.7893.84
RetinaNet [49]63.2768.1762.3365.9676.2265.9648.5062.7872.9980.9068.5925.5157.7851.3157.8184.20
CenterNet [50]73.9478.5666.1145.3953.3978.8666.8280.2497.3769.0290.3062.1664.8685.7575.6394.58
EfficientDet D0 [51]61.9664.6364.6344.2272.2551.1455.0552.3993.0942.7481.5738.8752.1278.2744.5492.18
EfficientDet D2 [51]68.6173.6060.5172.1474.6264.5760.5067.3877.3780.2878.6450.4661.8261.5962.1183.58
EfficientDet D4 [51]70.7272.2364.0173.2673.2365.3562.0263.1593.2779.9485.6159.7040.9275.9459.5692.66
S3FD [52]30.2230.6036.4323.7125.1233.919.1126.9236.3316.5127.0125.1128.5227.2026.2381.01
M2Det [53]34.4859.3443.9127.4240.6335.2218.2126.9136.1349.8418.119.1135.3227.3317.3172.70
ICN [54]68.1674.3079.0647.7070.3267.0250.2367.8281.3662.9069.9864.8953.6478.264.1790.76
Rol Transformer [55]69.5678.5277.2743.4475.9262.8347.6773.6888.6453.5483.5968.8158.3981.4658.9390.74
Rotated-YOLOX [56]66.3265.1776.5845.2454.9756.0344.8872.7188.7857.8879.6072.9741.3682.2365.2390.87
FSoD-Net [43]75.3371.1378.9351.3457.4671.3464.8682.1593.1170.6492.3978.1853.3293.4878.6992.87
YOLO-SBA (ours)76.1475.4068.5052.4067.8086.8066.4088.7095.9065.4090.8074.4065.1083.0065.9095.60
Comparison on RSOD [20] dataset. On the RSOD [20] dataset, we evaluated our method against eight sophisticated algorithms, as detailed in Table 3. The YOLO-SBA algorithm introduced in this study achieved an mAP of 94.7% on the RSOD [20] dataset, representing a 0.4% improvement over the current leading AGMF-Net [38] algorithm. In the RSOD [20] dataset, across its 4 categories, we obtained the top detection accuracy of 97.56% for the Aircraft class.

4.5. Ablation Study

To demonstrate the efficacy of the YOLO-SBA algorithm introduced in this paper, as well as the MBAFF module, GMAP module, and ADDown module, this section showcases three ablation experiments performed on the DIOR [14], DOTA [19], and RSOD [20] datasets. The results are presented in Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9. Specifically, Table 4, Table 5 and Table 6 present the comparison of the YOLO-SBA or modules with the baseline in terms of mAP and detection accuracy for each category on each dataset. Table 7, Table 8 and Table 9 display the APs, APm, and APl values for each dataset, comparing the proposed algorithm or modules with the baseline. These comparisons offer a more precise evaluation of the individual impacts of the modules and the overall effectiveness of the algorithm.
Baseline: The baseline for this section is YOLOv10 [6], which extracts features from remote sensing images using an enhanced version of CSPNet [62]. Subsequently, it employs SPPF and PSA for feature integration and augmentation. Ultimately, the refined feature maps are fed into the detection head to complete the object detection task. The baseline achieves mAP values of 85.9%, 72.00%, and 88.6% on the DIOR [14], DOTA [19], and RSOD [20] datasets, respectively.
MBAFF: The MBAFF module captures local details and global semantic information through a multi-branch structure with differentiated receptive fields. It also implements adaptive feature enhancement through a dual-channel attention mechanism, strengthening the feature response of targets with multi-scale variations. As shown in Table 4, Table 5 and Table 6, after adding the MBAFF module, the model’s mAP value increased by 2.5%, 3.0%, and 4.7% on the DIOR [14], DOTA [19], and RSOD [20] datasets, respectively. In the evaluation of single-category target detection, the single-category AP value on the DIOR [14] dataset increased by 0.35–7.17%, as shown in Table 4, fully demonstrating the module’s capability to handle target detection at different scales. From Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9, we can see that after adding the MBAFF module, both APl and mAP values showed significant improvements, with APl and mAP increasing by 2.5–4.7% on the three datasets. On the RSOD [20] dataset, the AP value for the overpass category increased by 16.20%.
GMAP: The GMAP module integrates multi-scale feature extraction with a gating fusion mechanism. It extracts multi-scale representations through a cascaded pooling structure, utilizes a dual-channel reconstruction mechanism to finely suppress background noise interference, and adaptively modulates the fusion weights of different scale features using a gating fusion mechanism. This amplifies the prominence of target features and more effectively reduces the noise interference from complex backgrounds. As shown in Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9, after adding the GMAP module to the baseline, the AP values for target detection in various categories improved significantly. On the DIOR [14] dataset, the AP values for bridge (B), dam (D), expressway toll station (ETS), overpass (O), storage tank (ST), train station (TS), and vehicle (V) categories increased by 7.01%, 4%, 6.31%, 5.17%, 3.29%, 6.3%, and 7.17%, respectively. As shown in Table 7, Table 8 and Table 9, after adding the GMAP module, the mAP values on the three datasets improved, with significant increases in APs and APm values. The APs value on the DIOR [14] dataset increased by 4.7%, while on the RSOD [20] and DOTA [19] datasets, the APs increased by 0.8%, demonstrating the effective suppression of complex background influence and the improvement in detecting small targets.
ADDown: Integrating asymmetric decompositional depthwise separable convolution downsampling with a dynamic channel attention mechanism for feature calibration effectively retains more key information and enhances the model’s robustness. As shown in Table 7, Table 8 and Table 9, after adding the ADDown module to the baseline, the mAP values on the DIOR [14], DOTA [19], and RSOD [20] datasets increased by 1.0%, 1.31%, and 0.8%, respectively. On the RSOD [20] dataset, APm increased by 1.1%. On the DIOR [14] dataset, APs increased by 4.9% and APm by 2.5%. On the DOTA [19] dataset, APs improved by 0.9%, showcasing its prowess in detecting medium and small targets and highlighting the module’s ability to preserve crucial detail information.
YOLO-SBA: As shown in Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9, after integrating the three modules into the baseline, the YOLO-SBA achieved significant improvements in mAP. On the DIOR [14], DOTA [19], and RSOD [20] datasets, the mAP values increased by 3.9%, 4.14%, and 6.1%, respectively, compared to the baseline. The single-category detection AP values for targets increased by 0.19–29.94%. This indicates that the synergistic utilization of the three modules can lead to improved performance, thereby validating the effectiveness of the algorithm introduced in this study.
To analyze the single module and the interaction between modules in more detail, this paper conducted ablation experiments on the RSOD [20] dataset for the three module combinations: MBAFF+GMAP, MBAFF+ADDown, and GMAP+ADDown, as shown in Table 10.
MBAFF+GMAP. Combining Table 6 and Table 10, the key roles of each module become clearer. After combining the baseline with MBAFF+GMAP, the accuracy is significantly improved, but compared to using only MBAFF or GMAP, the accuracy on aircraft and oil tank decreases, and the mAP also decreases. This is because after densely inserting MBAFF in the backbone, the model’s perception of multi-scale information in the input image is enhanced, and then it is input into GMAP for redundant background information suppression. However, during the network transmission process, such as in the downsampling stage, there is no guidance for key information. Ordinary downsampling operations may weaken the transmission of target information. The perceived detail information is weakened after the original downsampling operation and is directly input into GMAP for redundancy suppression. It is easy to mistake the detail target information in the multi-scale information as noise, leading to a decrease in mAP compared to using only MBAFF.
MBAFF+ADDown. After combining the baseline with MBAFF+ADDown, the network performance is improved, but it is not as good as the result of using only MBAFF. The reason is that after densely inserting MBAFF into the backbone, the network’s perception of multi-scale information is enhanced, and with the effective information transmission of ADDown, the model learns a lot of image information, including redundant background information. Since there is no processing of redundant background information at the end of the backbone, the unfiltered complex information interferes with the detection effect. At this point, effective downsampling information transmission is not conducive to the model’s judgment of targets, so the performance is worse than using only MBAFF.
GMAP+ADDown. After combining the baseline with GMAP+ADDown, the model performance is significantly improved compared to using only GMAP or ADDown. This is because after adding GMAP+ADDown, the model’s perception capability remains unchanged, capturing the same amount of image information. However, through GMAP+ADDown, the model’s information transmission and filtering capabilities are both enhanced, so the model performance is improved after adding GMAP+ADDown.
As shown in Table 11, we evaluated the parameter quantity, computational complexity, and inference time of the modules and algorithms proposed in this paper. The results show that compared to the baseline, the proposed YOLO-SBA does not significantly increase the model complexity and computational load. Although the inference speed is slightly slower, the detection performance is significantly improved. Moreover, the inference speed remains in the milliseconds range, meeting the requirements for real-time detection. greenThis indicates that in practical application scenarios, such as autonomous driving, security surveillance, and human-computer interaction, YOLO-SBA can quickly respond and provide accurate detection results, ensuring the efficient operation and timely decision-making of the system.
To illustrate the basis for the value of the expansion rate d in the MBAFF module proposed in this paper, as shown in Figure 9, we conducted ablation experiments on the RSOD [20] dataset. We chose different expansion rates for the MBAFF module to determine the best option. The experimental results clearly show that when the expansion rate d = 4 , the model achieves the best performance. Therefore, the expansion rate used in this paper is 4.

4.6. Visual Analysis

In Figure 10, we provide a visual demonstration of the YOLO-SBA algorithm’s performance across various tasks on the DIOR [14], RSOD [20], and DOTA [19] datasets. The tasks include detecting small, medium, and large objects, identifying targets in complex scenarios, and handling multi-scale detection within the same background. It is clear that the YOLO-SBA algorithm exhibits excellent performance in all five scenarios.
As shown in Figure 11, for different scenes (such as mountains and cities), compared to using only YOLOv10 [6], the heatmaps generated by YOLO-SBA more accurately highlight the target regions and reduce background noise interference. This indicates that YOLO-SBA has significant advantages in handling the dual challenges of complex backgrounds and multi-scale targets, providing more precise and robust detection results.

5. Discussion

This study introduces a multi-scale and complex background-aware network named YOLO-SBA, tailored for remote sensing target detection. The efficacy of the algorithm and its individual modules is validated through experimental evaluation. From the experimental results, it is evident that each module exhibits different advantages in the detection task. In this section, we analyze the proposed algorithm and the roles of each module, and provide explanations for the anomalous performance in the experiments.
MBAFF. Incorporating the MBAFF module into the baseline results in a notable enhancement of multi-scale target detection capabilities. On the DIOR [14] dataset, the single-category AP values increased by 0.35–7.17%. In other datasets, the detection accuracy for various categories also showed significant improvements, demonstrating the efficacy of this module in multi-scale target detection.
GMAP. Integrating the GMAP module into the baseline enhances detection accuracy across all categories, particularly for medium and small targets. The APs and APm metrics on the three datasets showed significant improvements. We believe this is primarily due to better suppression of complex backgrounds, which allows medium and small targets previously hidden in complex backgrounds to be detected, thus improving accuracy.
ADDown. After adding the ADDown module, the detection accuracy for all categories improved uniformly, with significant advantages in detecting medium and small targets. This indicates that ADDown effectively retains critical detail information during network inference. However, on the RSOD [20] dataset, after adding ADDown, the APs value is lower than the baseline, but the mAP value still increases. This is because the model’s weight update rule is designed based on mAP, which is a result of the model’s trade-off after training to achieve better mAP performance. It only indicates that the performance improvement of ADDown on the RSOD [20] dataset is not particularly significant, but it is still effective.
YOLO-SBA. YOLO-SBA, by integrating the advanced characteristics of three modules, exhibits excellent performance in detecting multi-scale targets and targets in complex backgrounds, significantly surpassing the baseline model. Specifically, the input remote sensing images, after target saliency enhancement by MBAFF, provide good guidance for background suppression by GMAP. The collaboration of these two modules improves the model’s capability to differentiate targets from backgrounds. Simultaneously, ADDown retains detail information during the downsampling process of network inference, ensuring that the effective information extracted by key modules is preserved. Its design optimizes the functionality of MBAFF and GMAP. The synergistic integration of these three modules improves the model’s capability to detect targets across multiple scales and mitigate complex backgrounds, thereby increasing the accuracy of remote sensing target detection.
The different performances of YOLO-SBA on the three datasets. In this paper, we conducted extensive testing on the proposed YOLO-SBA algorithm. For the DIOR [14] and RSOD [20] datasets, YOLO-SBA demonstrated excellent detection accuracy, achieving an mAP value of around 90 and significant improvements in detection accuracy for most categories, surpassing the current state-of-the-art algorithms. This indicates that YOLO-SBA performs exceptionally well in handling these datasets and can accurately detect and classify various targets.
However, on the DOTA [19] dataset, although YOLO-SBA improved the mAP value and achieved the best performance, the gap in accuracy compared to the DIOR [14] and RSOD [20] datasets was significant. This is mainly due to the characteristics of the DOTA [19] dataset itself and some limitations of the algorithm. The image resolution in the DOTA [19] dataset ranges from 800 × 800 to 20,000 × 20,000, which is much larger than other datasets. This not only increases the demand for computing resources but also affects the speed and efficiency of detection. In addition, the DOTA [19] dataset is large in scale, so researchers usually crop the original dataset to generate new datasets with resolutions below 1024 × 1024 for training and testing target detection algorithms. The original DOTA [19] dataset with 2806 images was expanded to 140,000 images after cropping and conversion, which undoubtedly increased the challenge of computing power. For the convenience of training and testing, this paper randomly selected 21,046 images, including 7492 background images (images without targets), from the expanded dataset. The total number of target instances in all selected images is close to 200,000, which is a huge challenge for the algorithm. Moreover, after cropping, some targets may be located at the edge of the image, only showing a corner or half of the target. In the process of stitching, there may be cases where corner pieces are stitched together to form an image, resulting in discontinuous backgrounds. These factors further increase the complexity of detection. Therefore, these factors together may result in the algorithm’s performance on the DOTA [19] dataset being less excellent than on other datasets, including the baseline detection results. This also provides us with directions for further improvement and optimization of the algorithm.

6. Conclusions

This paper proposes a multi-scale and complex background-aware network named YOLO-SBA for remote sensing target detection, focusing on addressing the challenges of multi-scale target diversity and complex background coexistence. Firstly, we designed the MBAFF module, which is tailored for multi-scale variation target detection tasks. It employs a multi-branch dilated convolution structure with differentiated receptive fields and utilizes a dual-path attention mechanism to adaptively fuse heterogeneous features, enhancing the model’s ability to express features of multi-scale variation targets. Next, to address the challenges of complex backgrounds, we proposed the GMAP module, which adopts an innovative architecture combining cascaded pooling with a dual attention reconstruction mechanism. It integrates the attention generated by a gating mechanism to guide the fusion process, achieving feature enhancement and background suppression, and significantly improving the model’s target detection accuracy in complex backgrounds. Additionally, we designed the ADDown module to ensure that the effective information extracted by key modules is not lost during network inference. It performs downsampling through decompositional depthwise separable convolution, better preserving the key information in the feature maps and enhancing the quality of the downsampled feature maps, collaboratively addressing the challenges of remote sensing image target detection with key modules.
The proposed YOLO-SBA tackles the challenges of multi-scale variations and complex backgrounds in remote sensing images, significantly improving the accuracy of remote sensing target detection. Experimental results on the DIOR [14], DOTA [19], and RSOD [20] datasets indicate that the algorithm achieves state-of-the-art detection accuracy, and ablation experiments demonstrate the effectiveness of the proposed modules, driving the development of remote sensing target detection tasks.

Author Contributions

Conceptualization: Y.Y. and Y.G.; Methodology: Y.Y. and X.Z.; Validation: Y.W., Y.G. and T.J.; Formal Analysis: Y.Y. and J.C.; Investigation: Y.W. and J.C.; Resources: Y.G. and X.Z.; Data Curation: T.J.; Writing—Original Draft Preparation: Y.Y.; Writing—Review & Editing: Y.W. and J.C.; Visualization: Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Natural Science Foundation of China (NSFC) under grant number 61873274.

Data Availability Statement

The data presented in this study are openly available at https://drive.google.com/drive/folders/1UdlgHk49iu6WpcJ5467iT-UqNPpx__CC (accessed on 7 July 2022) (DetectIon in Optical Remote sensing images, DIOR), https://captain-whu.github.io/DOTA/dataset.html (accessed on 26 January 2018) (Dataset for Object deTection in Aerial images, DOTA) and https://github.com/RSIA-LIESMARS-WHU/RSOD-Dataset- (accessed on 30 April 2019) (Remote Sensing Object Detection, RSOD).

Acknowledgments

I would like to express my gratitude to the National Natural Science Foundation of China (NSFC) for providing the financial support that made this research possible.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. LMSD-YOLO: A lightweight YOLO algorithm for multi-scale SAR ship detection. Remote Sens. 2022, 14, 4801. [Google Scholar] [CrossRef]
  2. Chen, J.; Liu, L.; Deng, W.; Liu, Z.; Liu, Y.; Wei, Y.; Liu, Y. Refining Pseudo Labeling via Multi-Granularity Confidence Alignment for Unsupervised Cross Domain Object Detection. IEEE Trans. Image Process. 2025, 34, 279–294. [Google Scholar] [CrossRef]
  3. Ruan, Z.; Wei, Y.; Guo, Y.; Xie, Y. Hybrid attentive prototypical network for few-shot action recognition. Complex Intell. Syst. 2024, 10, 8249–8272. [Google Scholar] [CrossRef]
  4. Pu, N.; Chen, W.; Liu, Y.; Bakker, E.M.; Lew, M.S. Dual gaussian-based variational subspace disentanglement for visible-infrared person re-identification. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2149–2158. [Google Scholar] [CrossRef]
  5. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  6. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  7. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  8. Li, C.; Li, L.; Geng, Y.; Jiang, H.; Cheng, M.; Zhang, B.; Ke, Z.; Xu, X.; Chu, X. Yolov6 v3. 0: A full-scale reloading. arXiv 2023, arXiv:2301.05586. [Google Scholar] [CrossRef]
  9. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  10. Jocher, G.; Stoken, A.; Borovec, J.; Changyu, L.; Hogan, A.; Diaconu, L.; Poznanski, J.; Yu, L.; Rai, P.; Ferriday, R.; et al. ultralytics/yolov5: v3. 0. Zenodo 2020. [Google Scholar] [CrossRef]
  11. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  12. Sun, H.; Yao, G.; Zhu, S.; Zhang, L.; Xu, H.; Kong, J. SOD-YOLOv10: Small Object Detection in Remote Sensing Images Based on YOLOv10. IEEE Geosci. Remote Sens. Lett. 2025, 22, 8000705. [Google Scholar] [CrossRef]
  13. Gui, S.; Song, S.; Qin, R.; Tang, Y. Remote sensing object detection in the deep learning era—A review. Remote Sens. 2024, 16, 327. [Google Scholar] [CrossRef]
  14. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  15. Zhang, T.; Zhang, X.; Zhu, X.; Wang, G.; Han, X.; Tang, X.; Jiao, L. Multistage enhancement network for tiny object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5611512. [Google Scholar] [CrossRef]
  16. Zhong, X.; Zhan, J.; Xie, Y.; Zhang, L.; Zhou, G.; Liang, M.; Yang, K.; Guo, Z.; Li, L. Adaptive Deformation-Learning and Multiscale-Integrated Network for Remote Sensing Object Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5611619. [Google Scholar] [CrossRef]
  17. Yang, X.; Zhang, S.; Duan, S.; Yang, W. An effective and lightweight hybrid network for object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5600711. [Google Scholar] [CrossRef]
  18. Zhang, Z.; Mei, S.; Ma, M.; Han, Z. Adaptive composite feature generation for object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5631716. [Google Scholar] [CrossRef]
  19. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar] [CrossRef]
  20. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate object localization in remote sensing images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  21. Dong, Z.; Wang, M.; Wang, Y.; Zhu, Y.; Zhang, Z. Object detection in high resolution remote sensing imagery based on convolutional neural networks with suitable object scale features. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2104–2114. [Google Scholar] [CrossRef]
  22. Li, G.; Liu, Z.; Zeng, D.; Lin, W.; Ling, H. Adjacent context coordination network for salient object detection in optical remote sensing images. IEEE Trans. Cybern. 2022, 53, 526–538. [Google Scholar] [CrossRef]
  23. Li, J.; Li, Z.; Chen, M.; Wang, Y.; Luo, Q. A new ship detection algorithm in optical remote sensing images based on improved R3Det. Remote Sens. 2022, 14, 5048. [Google Scholar] [CrossRef]
  24. Teng, Z.; Duan, Y.; Liu, Y.; Zhang, B.; Fan, J. Global to local: Clip-LSTM-based object detection from remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5603113. [Google Scholar] [CrossRef]
  25. Huang, H.; Huo, C.; Wei, F.; Pan, C. Rotation and scale-invariant object detector for high resolution optical remote sensing images. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 1386–1389. [Google Scholar] [CrossRef]
  26. Zhou, L.; Zheng, C.; Yan, H.; Zuo, X.; Liu, Y.; Qiao, B.; Yang, Y. RepDarkNet: A multi-branched detector for small-target detection in remote sensing images. ISPRS Int. J. Geo-Inf. 2022, 11, 158. [Google Scholar] [CrossRef]
  27. Li, Y.; Huang, Q.; Pei, X.; Jiao, L.; Shang, R. RADet: Refine feature pyramid network and multi-layer attention network for arbitrary-oriented object detection of remote sensing images. Remote Sens. 2020, 12, 389. [Google Scholar] [CrossRef]
  28. Guo, Y.; Wu, H.; Yang, S.; Cai, Z. Crater-DETR: A Novel Transformer Network for Crater Detection Based on Dense Supervision and Multiscale Fusion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5614112. [Google Scholar] [CrossRef]
  29. Zhang, L.; Wang, Y.; Huo, Y. Object detection in high-resolution remote sensing images based on a hard-example-mining network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 8768–8780. [Google Scholar] [CrossRef]
  30. Han, J.; Zhang, D.; Cheng, G.; Guo, L.; Ren, J. Object detection in optical remote sensing images based on weakly supervised learning and high-level feature learning. IEEE Trans. Geosci. Remote Sens. 2014, 53, 3325–3337. [Google Scholar] [CrossRef]
  31. Wang, B.; Zhao, Y.; Li, X. Multiple instance graph learning for weakly supervised remote sensing object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5613112. [Google Scholar] [CrossRef]
  32. Cheng, B.; Li, Z.; Xu, B.; Yao, X.; Ding, Z.; Qin, T. Structured object-level relational reasoning CNN-based target detection algorithm in a remote sensing image. Remote Sens. 2021, 13, 281. [Google Scholar] [CrossRef]
  33. Gao, T.; Liu, Z.; Zhang, J.; Wu, G.; Chen, T. A task-balanced multiscale adaptive fusion network for object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5613515. [Google Scholar] [CrossRef]
  34. Bai, Z.; Li, G.; Liu, Z. Global–local–global context-aware network for salient object detection in optical remote sensing images. ISPRS J. Photogramm. Remote Sens. 2023, 198, 184–196. [Google Scholar] [CrossRef]
  35. Yang, X.; Hou, L.; Zhou, Y.; Wang, W.; Yan, J. Dense label encoding for boundary discontinuity free rotation detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 15819–15829. [Google Scholar]
  36. Ye, X.; Xiong, F.; Lu, J.; Zhou, J.; Qian, Y. F3-Net: Feature Fusion and Filtration Network for Object Detection in Optical Remote Sensing Images. Remote Sens. 2020, 12, 4027. [Google Scholar] [CrossRef]
  37. Wang, H.; Li, H.; Qian, W.; Diao, W.; Zhao, L.; Zhang, J.; Zhang, D. Dynamic pseudo-label generation for weakly supervised object detection in remote sensing images. Remote Sens. 2021, 13, 1461. [Google Scholar] [CrossRef]
  38. Gao, T.; Li, Z.; Wen, Y.; Chen, T.; Niu, Q.; Liu, Z. Attention-free global multiscale fusion network for remote sensing object detection. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5603214. [Google Scholar] [CrossRef]
  39. Ma, W.; Wu, Y.; Zhu, H.; Zhao, W.; Wu, Y.; Hou, B.; Jiao, L. Adaptive Feature Separation Network for Remote Sensing Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5639717. [Google Scholar] [CrossRef]
  40. Son, S.; Kim, J.; Lai, W.S.; Yang, M.H.; Lee, K.M. Toward real-world super-resolution via adaptive downsampling models. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8657–8670. [Google Scholar] [CrossRef]
  41. Yang, Z.; Leng, L.; Min, W. Extreme downsampling and joint feature for coding-based palmprint recognition. IEEE Trans. Instrum. Meas. 2020, 70, 5005112. [Google Scholar] [CrossRef]
  42. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  43. Wang, G.; Zhuang, Y.; Chen, H.; Liu, X.; Zhang, T.; Li, L.; Dong, S.; Sang, Q. FSoD-Net: Full-scale object detection from optical remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602918. [Google Scholar] [CrossRef]
  44. Van Etten, A. You only look twice: Rapid multi-scale object detection in satellite imagery. arXiv 2018, arXiv:1805.09512. [Google Scholar] [CrossRef]
  45. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  46. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  47. Li, Q.; Mou, L.; Liu, Q.; Wang, Y.; Zhu, X.X. HSF-Net: Multiscale deep feature embedding for ship detection in optical remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 7147–7161. [Google Scholar] [CrossRef]
  48. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  49. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  50. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar] [CrossRef]
  51. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  52. Zhang, S.; Zhu, X.; Lei, Z.; Shi, H.; Wang, X.; Li, S.Z. S3fd: Single shot scale-invariant face detector. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 192–201. [Google Scholar]
  53. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  54. Azimi, S.M.; Vig, E.; Bahmanyar, R.; Körner, M.; Reinartz, P. Towards multi-class object detection in unconstrained remote sensing imagery. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 150–165. [Google Scholar]
  55. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  56. Liu, F.; Chen, R.; Zhang, J.; Xing, K.; Liu, H.; Qin, J. R2YOLOX: A lightweight refined anchor-free rotated detector for object detection in aerial images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5632715. [Google Scholar] [CrossRef]
  57. Xu, Z.; Xu, X.; Wang, L.; Yang, R.; Pu, F. Deformable convnet with aspect ratio constrained nms for object detection in remote sensing imagery. Remote Sens. 2017, 9, 1312. [Google Scholar] [CrossRef]
  58. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS–improving object detection with one line of code. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5561–5569. [Google Scholar]
  59. Dong, R.; Xu, D.; Zhao, J.; Jiao, L.; An, J. Sig-NMS-based faster R-CNN combining transfer learning for small target detection in VHR optical remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8534–8545. [Google Scholar] [CrossRef]
  60. Guo, Y.; Ji, J.; Lu, X.; Xie, H.; Tong, X. Geospatial object detection with single shot anchor-free network. In Proceedings of the IGARSS 2020-2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 280–283. [Google Scholar] [CrossRef]
  61. Zhang, J.; Zhu, H.; Wang, P.; Ling, X. ATT squeeze U-Net: A lightweight network for forest fire detection and recognition. IEEE Access 2021, 9, 10858–10870. [Google Scholar] [CrossRef]
  62. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
Figure 1. The heatmap visualization results of YOLOv10 [6] for remote sensing image target detection, where the source image is from the DIOR [14] dataset. The first row shows the Ground Truth, with yellow boxes and text labels indicating the location and category information of the targets. The second row shows the heatmap visualization of model attention during the YOLOv10 [6] detection process. (ac) show the detection results of YOLOv10 [6] in multi-scale target scenarios, complex background scenarios, and scenarios where multi-scale targets and complex backgrounds coexist. Red boxes highlight the targets not captured in the heatmap, indicating that the general object detection model YOLOv10 [6] does not perform well in remote sensing target detection tasks.
Figure 1. The heatmap visualization results of YOLOv10 [6] for remote sensing image target detection, where the source image is from the DIOR [14] dataset. The first row shows the Ground Truth, with yellow boxes and text labels indicating the location and category information of the targets. The second row shows the heatmap visualization of model attention during the YOLOv10 [6] detection process. (ac) show the detection results of YOLOv10 [6] in multi-scale target scenarios, complex background scenarios, and scenarios where multi-scale targets and complex backgrounds coexist. Red boxes highlight the targets not captured in the heatmap, indicating that the general object detection model YOLOv10 [6] does not perform well in remote sensing target detection tasks.
Remotesensing 17 01989 g001
Figure 2. Network Structure Diagram of the YOLO-SBA. Among them, MBAFF, GMAP, and ADDown are modules proposed in this paper. MBAFF and ADDown are extensively integrated into both the backbone and neck components, whereas GMAP is incorporated solely into the backbone section.
Figure 2. Network Structure Diagram of the YOLO-SBA. Among them, MBAFF, GMAP, and ADDown are modules proposed in this paper. MBAFF and ADDown are extensively integrated into both the backbone and neck components, whereas GMAP is incorporated solely into the backbone section.
Remotesensing 17 01989 g002
Figure 3. Network Structure Diagram of the MBAFF Module.
Figure 3. Network Structure Diagram of the MBAFF Module.
Remotesensing 17 01989 g003
Figure 4. Diagram of the D-Bottleneck Architecture.
Figure 4. Diagram of the D-Bottleneck Architecture.
Remotesensing 17 01989 g004
Figure 5. Visual understanding of dilated convolutions enhancing the convolutional receptive field.
Figure 5. Visual understanding of dilated convolutions enhancing the convolutional receptive field.
Remotesensing 17 01989 g005
Figure 6. Schematic diagram of the BAFM Module.
Figure 6. Schematic diagram of the BAFM Module.
Remotesensing 17 01989 g006
Figure 7. Network structure diagram of Gated Multi-scale Attention Pyramid (GMAP).
Figure 7. Network structure diagram of Gated Multi-scale Attention Pyramid (GMAP).
Remotesensing 17 01989 g007
Figure 8. ADDown Network Architecture Diagram.
Figure 8. ADDown Network Architecture Diagram.
Remotesensing 17 01989 g008
Figure 9. Performance of the MBAFF module added on the RSOD [20] dataset under different expansion rates d.
Figure 9. Performance of the MBAFF module added on the RSOD [20] dataset under different expansion rates d.
Remotesensing 17 01989 g009
Figure 10. Remote sensing object detection results based on YOLO-SBA in various scenarios. (a) dense small object detection; (b) medium object detection; (c) large object detection; (d) object detection in complex environments; (e) multi-scale object detection within the same background.
Figure 10. Remote sensing object detection results based on YOLO-SBA in various scenarios. (a) dense small object detection; (b) medium object detection; (c) large object detection; (d) object detection in complex environments; (e) multi-scale object detection within the same background.
Remotesensing 17 01989 g010
Figure 11. The heatmap visualization of model attention during the target detection process of YOLOv10 [6] and YOLO-SBA on common remote sensing images, corresponding to the images in Section 1. To more clearly perceive the changes in the heatmaps that reflect the performance of the model in target detection, we visualize the true annotations and the detection results of YOLOv10 [6] and YOLO-SBA together. The yellow boxes from top to bottom represent the true positions of the targets in the dataset, the detection results of YOLOv10 [6], and the detection results of YOLO-SBA.
Figure 11. The heatmap visualization of model attention during the target detection process of YOLOv10 [6] and YOLO-SBA on common remote sensing images, corresponding to the images in Section 1. To more clearly perceive the changes in the heatmaps that reflect the performance of the model in target detection, we visualize the true annotations and the detection results of YOLOv10 [6] and YOLO-SBA together. The yellow boxes from top to bottom represent the true positions of the targets in the dataset, the detection results of YOLOv10 [6], and the detection results of YOLO-SBA.
Remotesensing 17 01989 g011
Table 3. Comparison with state-of-the-art algorithms on the RSOD [20] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 3. Comparison with state-of-the-art algorithms on the RSOD [20] dataset. The metric is mAP(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPAircraftOil TankOverpassPlayground
Faster RCNN [45]84.5070.8490.1978.7498.09
Cascade R-CNN [46]91.3094.2096.1083.2099.00
Deformable R-FCN [57]87.9071.8790.3589.5999.88
Soft-NMS [58]86.6076.1090.3081.3098.80
Sig-NMS [59]89.4080.6090.6087.4099.10
SSAFNet [60]92.8095.7598.3984.6692.50
ATTS [61]91.3094.5096.0074.9099.90
AGMF-Net [38]94.3096.0299.0282.4399.70
YOLO-SBA (ours)94.7097.5696.2285.8199.31
Table 4. Ablation study of different YOLO-SBA modules on the DIOR [14] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
Table 4. Ablation study of different YOLO-SBA modules on the DIOR [14] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPALATBFBCBCDESAETSGCGTFHBOSSDSTTCTSVWM
Baseline85.9095.8890.2196.0490.5461.7086.9578.4196.7185.4186.8890.9773.5770.5695.0494.4089.1296.7272.8172.6894.20
+MBAFF88.4096.9092.1596.3591.0568.4888.2181.6796.6288.1188.6293.2775.5275.6195.3597.1592.2197.4577.1279.8596.68
+GMAP88.7097.2091.0596.2791.5168.7188.7982.4197.1591.7288.2692.8775.7575.7395.5597.4392.4197.0779.1179.8596.14
+ADDown86.9095.6290.2294.6089.8065.6786.8678.6795.4389.8186.8191.2774.8073.6294.1195.1291.1895.5676.5878.4595.16
YOLO-SBA89.8097.5093.3096.5092.1071.2089.9085.6096.9091.8090.0093.2077.9076.7095.6097.5093.1097.2082.3081.8098.80
Table 5. Ablation study of different YOLO-SBA modules on the DOTA [19] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
Table 5. Ablation study of different YOLO-SBA modules on the DOTA [19] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPBDBCBGTFHBHLVPRSSVSBFSTSPTC
Baseline72.0077.8768.5950.6858.3983.8936.4688.2594.8857.0590.2076.4254.4683.1163.0496.11
+MBAFF75.0076.0066.7452.0070.0086.8053.9088.6095.6065.3090.6075.5059.5082.5066.2095.10
+GMAP73.4473.4071.5050.5063.4085.3042.7089.1094.8060.7090.5075.8061.3082.0064.9095.70
+ADDown73.3173.6071.3050.1063.3085.1042.5089.1094.8059.5090.4075.8061.2082.0065.2095.70
YOLO-SBA76.1475.4068.5052.4067.8086.8066.4088.7095.9065.4090.8074.4065.1083.0065.9095.60
Table 6. Ablation study of different YOLO-SBA modules on the RSOD [20] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
Table 6. Ablation study of different YOLO-SBA modules on the RSOD [20] dataset. Metrics are mAP(%) and AP(%) for each category, and the bold values denote the top performance, while blue values signify the second-best results.
MethodmAPAircraftOiltankOverpassPlayground
baseline88.696.1095.4368.8694.03
+MBAFF93.395.9795.1585.0697.09
+GMAP89.195.8794.6870.7695.00
+ADDown89.495.6595.3771.3195.08
YOLO-SBA94.797.5696.2285.8199.31
Table 7. Ablation study of different YOLO-SBA modules on the DIOR [14] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 7. Ablation study of different YOLO-SBA modules on the DIOR [14] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodAPsAPmAPl
baseline26.252.376.4
+MBAFF30.254.780.0
+GMAP30.954.580.0
+ADDown31.154.877.7
YOLO-SBA32.756.781.2
Table 8. Ablation study of different YOLO-SBA modules on the DOTA [19] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 8. Ablation study of different YOLO-SBA modules on the DOTA [19] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodAPsAPmAPl
baseline27.549.455.0
+MBAFF27.849.759.5
+GMAP28.349.757.2
+ADDown28.449.656.5
YOLO-SBA27.650.061.1
Table 9. Ablation study of different YOLO-SBA modules on the RSOD [20] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
Table 9. Ablation study of different YOLO-SBA modules on the RSOD [20] dataset. Metrics include APs(%), APm(%), and APl(%), and the bold values denote the top performance, while blue values signify the second-best results.
MethodAPsAPmAPl
baseline35.567.066.0
+MBAFF35.667.968.8
+GMAP36.368.367.5
+ADDown34.968.166.1
YOLO-SBA38.469.272.0
Table 10. Ablation experiments on the RSOD [20] dataset with two modules combined.
Table 10. Ablation experiments on the RSOD [20] dataset with two modules combined.
MethodmAPAircraftOiltankOverpassPlayground
Baseline88.696.1095.4385.0697.09
+MBAFF+GMAP93.1594.9295.8182.8699.00
+MBAFF+ADDown90.3195.5996.0476.3193.30
+GMAP+ADDown90.6695.9896.6174.295.86
Table 11. Evaluation of the proposed algorithm in terms of parameter quantity, GFLOPS, and inference speed.
Table 11. Evaluation of the proposed algorithm in terms of parameter quantity, GFLOPS, and inference speed.
MethodParametersGFLOPSInference Time
DIOR [14]DOTA [19]RSOD [20]
Baseline8.1 M24.64.0 ms4.0 ms3.9 ms
+MBAFF10.0 M25.55.4 ms5.4 ms5.0 ms
+GMAP8.0 M24.94.2 ms4.2 ms3.7 ms
+ADDown7.4 M20.24.8 ms4.8 ms4.0 ms
YOLO-SBA10.1 M25.85.6 ms5.6 ms4.8 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, Y.; Wei, Y.; Zhou, X.; Guo, Y.; Chen, J.; Jiang, T. YOLO-SBA: A Multi-Scale and Complex Background Aware Framework for Remote Sensing Target Detection. Remote Sens. 2025, 17, 1989. https://doi.org/10.3390/rs17121989

AMA Style

Yuan Y, Wei Y, Zhou X, Guo Y, Chen J, Jiang T. YOLO-SBA: A Multi-Scale and Complex Background Aware Framework for Remote Sensing Target Detection. Remote Sensing. 2025; 17(12):1989. https://doi.org/10.3390/rs17121989

Chicago/Turabian Style

Yuan, Yifei, Yingmei Wei, Xiaoyan Zhou, Yanming Guo, Jiangming Chen, and Tingshuai Jiang. 2025. "YOLO-SBA: A Multi-Scale and Complex Background Aware Framework for Remote Sensing Target Detection" Remote Sensing 17, no. 12: 1989. https://doi.org/10.3390/rs17121989

APA Style

Yuan, Y., Wei, Y., Zhou, X., Guo, Y., Chen, J., & Jiang, T. (2025). YOLO-SBA: A Multi-Scale and Complex Background Aware Framework for Remote Sensing Target Detection. Remote Sensing, 17(12), 1989. https://doi.org/10.3390/rs17121989

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop