Next Article in Journal
Sensor-Based Yield Prediction in Durum Wheat Under Semi-Arid Conditions Using Machine Learning Across Zadoks Growth Stages
Next Article in Special Issue
Eye in the Sky for Sub-Tidal Seagrass Mapping: Leveraging Unsupervised Domain Adaptation with SegFormer for Multi-Source and Multi-Resolution Aerial Imagery
Previous Article in Journal
Spatiotemporal Analysis of Eco-Geological Environment Using the RAGA-PP Model in Zigui County, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Dynamic Context Branch Attention Network for Detecting Small Objects in Remote Sensing Images

1
School of Computer Science, Hubei University of Technology, Wuhan 430068, China
2
Hubei Provincial Key Laboratory of Green Intelligent Computing Power Network, Wuhan 430068, China
3
Hubei Provincial Engineering Research Center for Digital & Intelligent Manufacturing Technologies and Applications, Wuhan 430068, China
4
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430010, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(14), 2415; https://doi.org/10.3390/rs17142415
Submission received: 7 May 2025 / Revised: 30 June 2025 / Accepted: 10 July 2025 / Published: 12 July 2025
(This article belongs to the Special Issue High-Resolution Remote Sensing Image Processing and Applications)

Abstract

Detecting small objects in remote sensing images is challenging due to their size, which results in limited distinctive features. This limitation necessitates the effective use of contextual information for accurate identification. Many existing methods often struggle because they do not dynamically adjust the contextual scope based on the specific characteristics of each target. To address this issue and improve the detection performance of small objects (typically defined as objects with a bounding box area of less than 1024 pixels), we propose a novel backbone network called the Dynamic Context Branch Attention Network (DCBANet). We present the Dynamic Context Scale-Aware (DCSA) Block, which utilizes a multi-branch architecture to generate features with diverse receptive fields. Within each branch, a Context Adaptive Selection Module (CASM) dynamically weights information, allowing the model to focus on the most relevant context. To further enhance performance, we introduce an Efficient Branch Attention (EBA) module that adaptively reweights the parallel branches, prioritizing the most discriminative ones. Finally, to ensure computational efficiency, we design a Dual-Gated Feedforward Network (DGFFN), a lightweight yet powerful replacement for standard FFNs. Extensive experiments conducted on four public remote sensing datasets demonstrate that the DCBANet achieves impressive mAP@0.5 scores of 80.79% on DOTA, 89.17% on NWPU VHR-10, 80.27% on SIMD, and a remarkable 42.4% mAP@0.5:0.95 on the specialized small object benchmark AI-TOD. These results surpass RetinaNet, YOLOF, FCOS, Faster R-CNN, Dynamic R-CNN, SKNet, and Cascade R-CNN, highlighting its effectiveness in detecting small objects in remote sensing images. However, there remains potential for further improvement in multi-scale and weak target detection. Future work will integrate local and global context to enhance multi-scale object detection performance.

1. Introduction

Remote Sensing Object Detection (RSOD) focuses on accurately and efficiently identifying specific object classes within optical remote sensing images (RSIs) [1]. RSOD has long been a fundamental and challenging task. With the rapid advancements in deep learning, RSOD methods have made significant progress. For instance, novel frameworks have enhanced both accuracy and performance, including one-stage detectors like CFA [2] and RTMDet [3], and two-stage approaches such as RoI Transformer [4] and Oriented R-CNN [5].
However, remote sensing images contain a high prevalence of small objects. These objects are typically defined as having a bounding box area of fewer than 1024 pixels. Due to their limited size, these objects yield low-quality features, posing significant challenges for accurate recognition [6]. These small objects are susceptible to information loss during the convolution and pooling operations in neural networks. Furthermore, small objects are easily obscured by or confused with complex backgrounds, posing significant challenges for accurate localization and classification. Notably, for such objects, even slight perturbations in the bounding box can dramatically impact the Intersection over Union (IoU) metric, which in turn increases the learning difficulty for the regression branch.
To tackle the inherent difficulties in detecting small objects within RSIs, numerous advanced approaches have been developed [7]. These methods primarily employ strategies such as super-resolution reconstruction [8], multi-scale feature fusion [9,10], data augmentation [11], anchor optimization [12], and loss function refinement [13] to enhance the detection accuracy of small objects in RSIs. However, these methods still exhibit several critical limitations. While super-resolution modules can enhance the detailed features of small objects, they not only significantly increase computational complexity but may also degrade detection performance due to the generated artificial textures or edge artifacts. Multi-scale feature fusion methods that rely solely on simple concatenation or summation operations often suffer from feature incompatibility caused by the semantic gap between high-level and low-level features, leading to the suppression of crucial small object features by larger objects or background noise. Furthermore, other improvement strategies like anchor design optimization or loss function adjustments typically demonstrate only marginal gains in small object detection performance.
Recent studies [14,15] have demonstrated that contextual information plays a pivotal role in small object detection within remote sensing images. This is primarily because different categories of small objects may exhibit highly similar visual features, while environmental context can provide more discriminative feature representations for target identification [16]. In recent years, context-learning-based approaches for small object detection have emerged as a research focus, with a series of related works dedicated to exploring effective methods for leveraging scene context to enhance detection performance. YOLOX-CA [17] improves small object detection performance by incorporating convolutional kernels with varying dilation rates, effectively aggregating multi-scale contextual information to enhance the model’s sensitivity towards small targets. Li et al. [18] proposed a receptive field expansive deformable convolution that adaptively adjusts the context extraction scope and covers irregular contextual regions, thereby capturing appropriate contextual features to enhance the discriminability of small objects. DFCE [19] effectively detects small targets by integrating local and global contextual information and dynamically enhancing key features through a crossover mechanism.
Although existing context-learning-based methods have significantly improved the detection performance of small objects in remote sensing images, current approaches [20,21] still suffer from notable limitations. Primarily, most studies simply concatenate multi-scale contextual features without effectively highlighting crucial information. Furthermore, these methods fail to establish a truly adaptive context extraction mechanism, consequently limiting the model’s capability to dynamically adjust the utilization of contextual information based on target characteristics. To address these challenges, we propose a Dynamic Context Scale-Aware Block (DCSA Block) featuring multiple parallel branches that generate feature maps with varying receptive fields, coupled with a selective mechanism to adaptively determine the optimal contextual scope for each target. Furthermore, we introduce an Efficient Branch Attention (EBA) to dynamically reweight different branches, thereby enhancing those containing critical information. To compensate for the computational overhead induced by the multi-branch architecture, we devise a Dual-Gated Feedforward Network (DGFFN) that significantly reduces both parameters and computational complexity while improving inference efficiency. The main contributions of this work are threefold:
  • We propose a DCSA Block with a multi-branch architecture, which can adaptively capture the optimal range of contextual information for small objects in RSIs. Furthermore, we introduce an EBA module that employs dynamic weight adjustment to significantly enhance the network’s capability to utilize branches containing critical information.
  • We propose a DGFFN that effectively reduces the number of parameters and complexity while improving the detection speed of the model by introducing gating mechanisms into the feedforward network.
  • We construct a novel lightweight backbone network called the Dynamic Context Branch Attention Network (DCBANet) by repeatedly stacking the proposed components. The network demonstrates outstanding performance on public remote sensing datasets, validating that our components effectively and dynamically adapt to small targets of various sizes. This highlights the model’s universal applicability in detecting small objects across the entire scale range (<1024 pixels).

2. Related Works

2.1. RSOD Frameworks

RSOD frameworks based on deep learning can be categorized into one-stage methods and two-stage methods. One-stage methods, such as R3Det [22] and S2A-Net [23], demonstrate superior computational efficiency and real-time processing capabilities by directly predicting the bounding box, class, and confidence of the objects in the input image, albeit with relatively lower detection accuracy. In contrast, two-stage methods, exemplified by advanced frameworks like Cascade R-CNN [24] and Dynamic R-CNN [25], are based on the classical Regional Convolutional Neural Network [26] (RCNN) architecture, which first generates region proposals and subsequently performs a more fine-grained object categorization and bounding box regression on these proposals. Therefore, they have a distinct advantage in terms of object detection accuracy. Inspired by the DETR [27] model, some recent studies [28,29] have proposed end-to-end oriented detectors, which effectively resolve the inconsistency between object classification and localization in traditional models.
While these one-stage and two-stage frameworks have achieved remarkable success, their direct application to small object detection in RSIs is often sub-optimal. The core challenge stems from the feature representation of small objects. The hierarchical structure of deep networks, with its repeated down-sampling operations, inherently leads to the loss of crucial high-resolution spatial information. This makes it exceedingly difficult for subsequent detection heads to accurately localize and classify tiny targets, regardless of whether they are anchor-based or anchor-free. Therefore, designing a backbone network that can effectively preserve and enhance discriminative features for small objects is of paramount importance.

2.2. Small Object Detection

In recent years, small object detection has emerged as a prominent research direction in the field of computer vision. QueryDet [30] proposed a detector based on coarse location queries. The method first predicts regions in the image where small objects may exist and then utilizes high-resolution features in these specific regions for object localization and classification. EFPN [31] introduces a super-resolution module into the traditional feature pyramid network to apply super-resolution operations only to small-sized feature maps. It improves detail representation for small objects, but at the expense of increased computational effort and model complexity. ABFPN [32] expands the receptive field of the feature maps by employing skip-ASPP blocks with varying dilation rates and applies jump-joining to adequately fuse the multi-scale features, which effectively avoids the problem of potentially weakening the semantic information when performing lateral joining. ABNet [33] proposed an adaptive FPN capable of dynamically extracting multi-scale features across different channels and spatial locations. SCRDet++ [34] applies an instance-level denoising method to small object detection for the first time and improves the Smooth-L1 loss function to address challenges in boundary detection.
In summary, existing small object detection methods primarily focus on strategies like multi-scale feature fusion, super-resolution, and context enhancement. However, many context-based approaches simply aggregate contextual information without a mechanism to dynamically adapt the context range based on the target’s specific characteristics. This static context aggregation can be sub-optimal, as different small objects may require different contextual cues for accurate identification. This gap motivates our work on a dynamic context-aware module.

2.3. Attention Mechanisms

Attention mechanisms [35] have become a powerful tool for enhancing feature representation in computer vision, and they are particularly relevant for small object detection. By allowing the model to focus on the most informative features, attention can help distinguish tiny objects from a cluttered background. [36]. SENet [37] uses global average pooling to capture global information, followed by fully connected and activation layers in the excitation module to capture channel relationships to output channel weights. ECANet [38] considers only local cross-channel interactions, where each channel interacts with its K-nearest neighboring channels rather than capturing all inter-channel relationships, while using 1 × 1 convolution instead of fully connected layers to generate channel weights. GENet [39] captures the broader context by aggregating feature responses from spatial neighborhoods and then redistributes the aggregated information to local features to adjust the original feature responses, ultimately yielding a spatial attention map. PSANet [40] proposes a point-wise spatial attention that achieves the aggregation of long-range contextual information by learning the correlation between each pixel pair and thus enabling the dynamic adjustment of weights between them. ViT [41] splits the image into small patches and employs a self-attention mechanism to learn inter-region relationships, which can better aggregate the image’s global context. ResNeSt [42] enhances the model’s feature representation capability by incorporating multiple branches into the ResNet architecture and introduces a new attention mechanism, Split-Attention, which can dynamically adjust the weights of different branches.
However, most channel attention mechanisms like SENet or ECANet treat all branches or feature groups equally or competitively, which may not be ideal for our multi-branch architecture, where different branches capture distinct types of contextual information. A mechanism that can independently weight and preserve critical information from each branch is therefore desirable for our specific task of small object detection.

2.4. Feedforward Network

Feedforward networks (FFNs) are an important part of transformer models [41,43] and are primarily composed of a multilayer perceptron (MLP). The MLP enhances the feature representation by a series of linear transformations and nonlinear activation functions. However, the parameters in the MLP account for more than 60% of the entire model, making it the main source of the computational bottleneck issue. To address this problem, SparseFFN [44] reduces the parameters of the FFN through channel-wise sparsification using grouped linear layers and spatial-wise sparsification via feature pooling to capture regional semantics. Compact FFN [45] factorizes the projection matrix at the second linear layer to reduce redundancy and further compresses the FFN using reparameterization techniques. Some studies have focused on leveraging gating mechanisms [46] to simplify FFNs, as these mechanisms are capable of controlling which information should be retained, enhanced, or discarded. MAN [47] introduces simple spatial attention in a linear gating unit to remove unnecessary linear layers, thus drastically reducing the number of parameters in the FFN. MogaNet [48] reduces the large number of redundant features caused by conventional FFNs by introducing a channel aggregation module based on gating mechanisms to redistribute the multi-order features in the feedforward network. Although these methods reduce the number of parameters in the FFN, they also lead to a reduction in the model’s overall performance.
While these methods successfully reduce the parameters and complexity of the FFN, they often do so at the cost of performance, which can be detrimental for the already challenging task of small object detection, where every bit of feature representation capability matters. Therefore, designing a lightweight FFN that minimizes performance degradation is a key challenge. This motivates our proposed Dual-Gated Feedforward Network (DGFFN), which aims to strike a better balance between efficiency and accuracy, a critical consideration for deploying detectors on large-scale remote sensing data.

3. Method

3.1. DCBANet Backbone Architecture

As shown in Figure 1, the overall architecture of our proposed DCBANet backbone network adopts a multi-stage design, where each stage outputs a feature map of different sizes. DCBANet is inspired by LSKNet [14], ECANet [38], ResNeSt [42], GCT [46], MAN [47], and VAN [49].
Each stage in the backbone network consists of a repetitive stack of DCBA-Former based on the MetaFormer [50] architecture. As illustrated in Figure 2, each DCBA-Former comprises a DCSA residual sub-block and a DGFFN residual sub-block. The DCSA sub-block serves to dynamically extract spatial context information for small objects. The DGFFN sub-block primarily functions to enhance the feature representation at each location and introduce nonlinearity.

3.2. Dynamic Context Scale-Aware Block

To enable adaptive adjustment of contextual information, we propose the DCSA Block (illustrated in Figure 3). It is a block-level multi-branch design, which differs from conventional network-level multi-scale feature extraction. While traditional multi-scale approaches generate feature maps of different spatial resolutions at different stages of the network to handle variations in object size, our DCSA Block operates on a single feature map from a specific stage. It incorporates a multi-branch architecture with varying kernel sizes to generate feature maps with diverse receptive fields at the same scale and resolution. A selective mechanism is then employed to dynamically choose the most appropriate contextual range for each target. This approach is particularly effective for small objects, as their recognition often depends on fine-grained local context rather than broad semantic information from different scales. This design effectively overcomes the limitation of fixed contextual ranges inherent to single-kernel approaches.
Specifically, given an input image X R C × H × W , we divide it into multiple parts in terms of channel dimensions. These parts are then processed by distinct branches, each employing a different large convolution kernel. To mitigate the substantial parameter overhead induced by large convolutional kernels, we decompose each of them into two smaller depthwise convolutional kernels. This decomposition process can be formulated as follows:
C o n v K ( X ) = C o n v 1 × 1 ( D W C o n v k 1 ( D W D C o n v k 2 , d ( X ) ) )
where K denotes a large convolution kernel; C o n v 1 × 1 denotes a 1 × 1 convolution; D W C o n v k 1 denotes a depthwise convolution with a kernel size of k 1 ; and D W D C o n v k 2 , d denotes a depthwise dilated convolution with a kernel size of k 2 and a dilation rate of d . The input in each branch sequentially undergoes both a depthwise convolution and a depthwise dilated convolution to generate two feature maps with distinct receptive fields: one with a smaller receptive field (defined as U s ) and the other with a larger one (defined as U l ):
U s = C o n v 1 × 1 D W C o n v k 1 X
U l = C o n v 1 × 1 D W D C o n v k 2 , d U s
U s and U l are first compressed along the channel dimension to reduce redundant features and then concatenated. The combined feature map is subsequently subjected to average pooling and maximum pooling to extract spatial relationships. The pooled feature map is defined as U ~ :
U = C o n c a t ( C o n v 1 × 1 ( U s ) , C o n v 1 × 1 ( U l ) )
U ~ = C o n c a t ( P a v g ( U ) , P m a x ( U ) )
where P a v g and P m a x denote average pooling and maximum pooling operations, respectively. To extract appropriate context information for different objects in U s and U l , the weights of different regions are first calculated:
W = σ ( C o n v 1 × 1 ( U ~ ) )
where σ represents the activation function and W denotes the spatial weight coefficient with two channels (defined as W s and W l ). Weighting and fusing the weight coefficients with the corresponding feature maps realize the selection of appropriate contextual information for small objects. X i denotes the final obtained spatial feature map:
X i = W s · U s + W l · U l

3.3. Efficient Branch Attention

To highlight critical branches and enhance the utilization of discriminative features, we introduce an Effective Branch Attention (EBA) module that dynamically adjusts branch-wise weights. The key innovation lies in employing 1 × 1 convolutions (instead of fully connected layers) to generate channel-wise weights, which effectively prevents information loss during channel compression while reducing parameter overhead. Furthermore, the application of sigmoid activation to channel weights ensures more stable training dynamics. Notably, the sigmoid activation allows each channel weight to be activated independently without competition among branches, thereby better preserving discriminative feature representations. Specific details are shown in Figure 4.
First, we concatenate the results obtained from all branches along the channel dimension to obtain Y , which is assumed to have the shape of C × H × W :
Y = i = 1 b X i
where X i is the result of different branches and b is the number of branches. Then, a global average pooling operation is applied to it:
Y ~ i = 1 H W h = 1 H   w = 1 W   y i , h , w     f o r   i = 1,2 , , C
where y i , h , w represents the value located at spatial coordinates   ( h , w ) within the i -th feature map Y , while Y ~ i corresponds to the aggregated channel-wise descriptor for the i -th channel. The pooled feature vector Y ~ = Y ~ 1 , Y ~ 2 , , Y ~ c R C thus consists of C elements, corresponding to the number of channels in Y . Subsequently, we employ a lightweight 1 × 1 convolution to generate channel-wise weights, which are then activated via a sigmoid function:
w = σ C o n v 1 × 1 ( Y ~ )
where σ denotes the sigmoid activation function. In contrast, Split Attention [42] employs softmax activation, which creates zero-sum competition among branches and can potentially lead to the loss of effective features. Conversely, the sigmoid function permits independent multi-channel activation, unaffected by other branches, thereby enabling more effective feature fusion. The final output of EBA X E B A is computed as a weighted sum of the input branches:
X E B A = i = 1 b X i · w i
In the final step, a 1 × 1 convolution is applied to X E B A . The result is then added to the original input X through a residual connection to produce the final output of the DCSA Block:
X o u t p u t = C o n v 1 × 1 X E B A + X
The algorithmic details of the DCSA Block are summarized in Algorithm 1.
Algorithm 1: Forward Propagation Process of the DCSA Block
 Input: Feature map X_in
 Output: Feature map X_out
1: Split X_in into n branches: X_1, X_2, …, X_n
2: Initialize an empty list of branch outputs: Outputs = []
3: for i = 1 to n do
4: //Decomposed large kernel convolution
5:  U_s_i = Conv1×1(DWConv_k1(X_i))
6:  U_l_i = Conv1×1(DWDConv_k2,d(U_s_i))
7: //Context Adaptive Selection Module (CASM)
8:  U_combined_i = Concat(Conv1×1(U_s_i), Conv1×1(U_l_i))
9:  W_i = σ(Conv1×1(Concat(P_avg(U_combined_i), P_max(U_combined_i))))
10:  X_branch_out_i = W_s_i × U_s_i + W_l_i × U_l_i
11:  Append X_branch_out_i to Outputs
12: end for
13://Efficient Branch Attention (EBA)
14: X_cat = Concat(Outputs)
15: Y_pooled = GlobalAvgPool(X_cat)
16: w_eba = σ(Conv1 × 1(Y_pooled))
17: X_eba = Σ (w_eba_i × X_branch_out_i)
18://Residual Connection
19: X_out = Conv1×1(X_eba) + X_in
20: return X_out

3.4. Dual-Gated Feedforward Network

The MLP typically maps input features to hidden features with 4× or 8× higher dimensions, leading to a substantial parameter count in the fully connected layers [44]. Directly reducing the hidden feature dimensions or performing grouping operations may drastically reduce model performance. To reduce parameters while maintaining superior performance, we design a gating mechanism-based feedforward network named the Dual-Gated Feedforward Network (DGFFN).
The DGFFN (Figure 5) comprises two components: a Spatial Gating Unit (SGU) and a Channel Gating Unit (CGU). The SGU enhances local feature representation by generating spatial gating weights. The CGU introduces three learnable parameters to enhance its contextual information modeling capability by adjusting the competitive and complementary relationships between channels.
In the DGFFN, the input tensor is assumed to be R C × H × W , and the hidden layer expansion ratio is m . The input features are first extended to the high-dimensional hidden feature tensor X h via a 1 × 1 convolution, which is in line with the traditional MLP.
X h = G E L U ( B N ( C o n v 1 × 1 ( X ) ) )         X h R m C × H × W
Spatial Gating Unit (SGU). The SGU is built on the basis of a gated linear unit, which receives the high-dimensional hidden feature X h from the previous step as input, and then divides it equally into two parts along the channel dimension:
g a t e , X ˙ = c h u n k ( X h , 2 , d i m = 1 )
where g a t e is the gating feature used to generate the gating weights and X ˙ is the other feature part, which undergoes the subsequent weighting operation. At this point, the channel dimension of both parts is m C / 2 . The gating feature g a t e is transformed into gating weights via a depthwise convolution. The gating weights are then element-wise multiplied by the feature part X ˙ to implement the gating effect:
X S G U = X ˙ · D W C o n v g a t e
where X S G U is the feature after spatial gating with the same channel dimension m C / 2 . After the SGU, the channel dimension is again restored to C via 1 × 1 convolution.
X = G E L U ( B N ( C o n v 1 × 1 ( X S G U ) ) )         X R C × H × W
Channel Gating Unit (CGU). In the CGU, three learnable parameters α , β , and γ are introduced, where α is used to control the scaling factor of the feature channel, β controls the offset of the gating value, and γ controls the channel normalization.
First, for the input X R C × H × W , compute a spatial embedding e d for each channel using its L2-norm scaled by a learnable parameter:
e d = h = 1 H     w = 1 W     x c , h , w 2 + ϵ 1 2 α
where x c , h , w is the feature value at spatial coordinates ( h , w ) of the c -th channel; ϵ indicates a small constant added to prevent numerical instability; and α is a learnable parameter used to scale the embedding values. The resulting e d value represents the scaled spatial intensity of each channel. Next, each channel is normalized:
n o r m = γ R M S ( e d ) = γ ( 1 C i = 1 C e d i 2 + ε ) 1 2
where R M S denotes the root mean square value; C is the number of channels; and γ denotes a learnable scaling factor that modulates the gating strength of each channel by controlling the degree of channel normalization. The gating factor g a t e is then calculated as
g a t e = 1 + t a n h e d · n o r m + β
where β represents a bias and t a n h · refers to the hyperbolic tangent function. This function first compresses the input value to [ 1 ,   1 ] , and the addition of the constant 1 subsequently adjusts the range to [ 0 ,   2 ] , yielding a suitable gating value for weighting. Finally, the gated weighting operation is performed via element-wise multiplication:
X = X · g a t e
The DGFFN primarily reduces the computational complexity by decreasing the input channel dimensions of the subsequent depthwise convolutional layer and the second linear (e.g., 1 × 1 convolutional) layer, enabled by the split operation in the SGU. Since the CGU does not involve computationally intensive convolution or fully connected operations, its computational complexity is generally considered negligible. The DGFFN can replace the feedforward networks in various transformer models to achieve model lightweighting while enhancing performance.

4. Experiments

4.1. Datasets

DOTA [51] is a leading benchmark dataset in remote sensing image analysis, consisting of 2806 high-resolution aerial images. The dataset is split into training, validation, and test sets at a 1411:458:937 ratio. Image resolutions range from 800 × 800 to 4000 × 4000 pixels, highlighting significant multi-scale properties. This dataset includes 15 common classes annotated with oriented bounding boxes (OBBs), with a total of 188,282 annotated instances. The average width of these instances is 27.60 pixels, with approximately 60% of instances measuring less than 20 pixels in width.
NWPU VHR-10 [52] is a dataset comprising 650 remote sensing images with very high resolutions. The images feature spatial resolutions between 0.5 m and 2 m and pixel dimensions varying between 533 × 597 and 1728 × 1028. The dataset is annotated with 10 object categories, totaling 3651 annotated instances. For this study, we apply a random split approach to partition the dataset into training and test sets equally, resulting in 325 images allocated to training and 325 images to testing.
SIMD [53] is a specialized remote sensing dataset for detecting vehicles and aircrafts. It contains 5000 images at a resolution of 1024 × 768 pixels, divided into 4000 for training and 1000 for testing. It covers 15 object categories, which the original authors classified into three groups based on object scale characteristics: small-scale objects (including Cars, Vans, Stair Trucks, Pushback Trucks, Others, Boats, and Helicopters); medium-scale objects (including Trucks, Buses, Trainer Aircrafts, and Fighter Aircrafts); and large-scale objects (including Long Vehicles, Airliners, Propeller Aircrafts, and Chartered Aircrafts). Among these, small-scale targets account for over 80% of the total instances.
AI-TOD [54] is a dataset specifically constructed for remote sensing small target detection, which contains 28,036 remote sensing images with 800 × 800 pixels. The dataset contains 700,621 annotated object instances across eight categories. Notably, the average object size is merely 12.8 pixels, which is significantly smaller than in other datasets. This characteristic makes AI-TOD particularly challenging for evaluating the performance of small object detection models.
The AI-TOD dataset further divides small objects (<1024 pixels) into three scale ranges based on the area of the bounding box: generally small (256–1024 pixels), tiny (64–256 pixels), and very tiny (8–64 pixels). According to this classification criterion, Figure 6 shows the scale distribution of small- and medium-sized objects in DOTA, NWPU VHR-10, SIMD, and AI-TOD. Both DOTA and AI-TOD contain many small objects, especially AI-TOD, which contains 255,227 tiny objects and 43,519 very tiny objects.

4.2. Experimental Setup and Evaluation Metrics

4.2.1. Experimental Setup

For our experiments, we integrated the proposed DCBANet backbone into the MMDetection framework (version 3.3.0). We consistently used a standard Feature Pyramid Network (FPN) as the neck. For the detection head, unless otherwise specified, the default detectors used in the experiment are Faster-RCNN (HBB task) or Oriented-RCNN (OBB task).
Our proposed method was implemented using the PyTorch framework (version 1.10.0), and the experiments were carried out on an NVIDIA RTX 4090 GPU with 24 GB (NVIDIA Corporation; Santa Clara, CA, USA) of memory. We employed the AdamW optimizer for network optimization, with hyperparameters including an initial learning rate of 1 × 10 4 , a weight decay coefficient of 0.05, and beta parameters ( β 1 , β 2 ) set to (0.9, 0.999). During the inference phase, the key parameters were set as follows: the maximum number of proposals was limited to 2000, the confidence threshold was 0.05, and the IoU threshold for Non-Maximum Suppression (NMS) was 0.5. For consistent calculation of model complexity, the input tensor size was fixed at (1, 3, 1024, 1024) when calculating the parameter count (Param) and Floating-point Operations (FLOPs).
For the DOTA dataset, during single-scale training, the original images were cropped into 1024 × 1024-pixel patches with a stride of 824 pixels and a 200-pixel overlap. During multi-scale training, the images were resized with scaling factors of 0.5, 1.0, and 1.5, followed by cropping with a stride of 524 pixels and a 500-pixel overlap. The batch size was set to 2 during training. Consistent with mainstream practices, the model was trained using both the training and validation sets, and the final test results were generated through the official DOTA evaluation server.
For the SIMD, NWPU VHR-10, and AI-TOD datasets, we standardized the resolution of the original images to 1024 × 1024 pixels for both training and testing. Random flipping was applied as a data augmentation strategy, and the batch size was set to 4.

4.2.2. Evaluation Metrics

This study employs Average Precision (AP) and mean Average Precision (mAP) to evaluate detection accuracy, Frames Per Second (FPS) to measure detection speed, and Floating-point Operations (FLOPs) to assess computational complexity.
AP measures the overall performance of the model at different confidence thresholds by calculating the area under the Precision–Recall curve. By default, the Intersection over Union (IoU) threshold is 0.5. The calculation formula is as follows:
A P = 0 1 P R d R
where P (Precision) represents the proportion of samples predicted by the model to be positive that are actually positive, and R (Recall) represents the proportion of positive samples correctly detected by the model out of all actual positive samples.
The calculation of mAP involves first computing the AP for each category and then taking the average of the AP values across all categories:
m A P = 1 k i = 1 k A P i
where k stands for the total number of object categories.
When conducting experiments on the AI-TOD dataset, we evaluate our results based on the COCO benchmark. The evaluation relies on the Average Precision (AP) metric, where the primary metric, mAP@0.5:0.95, is calculated by averaging AP values over ten IoU thresholds from 0.5 to 0.95, providing a comprehensive assessment of localization accuracy. We also report mAP@0.5 (IoU = 0.5) and mAP@0.75 (IoU = 0.75). Furthermore, to analyze performance across different object sizes, we utilize m A P s , m A P m , and m A P l , which correspond to mAP@0.5:0.95 for small (<1024 pixels), medium (1024–9216 pixels), and large (>9216 pixels) objects. Given the abundance of small targets in AI-TOD, the m A P s metric is particularly critical for our analysis.

4.3. Ablation Study

Ablation Study on Branch Number and Large Kernel Decomposition Sequences. In the DCSA Block, different large kernel decomposition sequences generate feature maps with varying receptive fields, which directly influence subsequent context-adaptive performance. Meanwhile, the number of branches significantly impacts the model’s detection speed. A well-designed combination of branch numbers and large kernel decomposition sequences can achieve an optimal trade-off between speed and accuracy. LSKNet [14] was selected to be the baseline model. We replaced its LK Selection module with the multi-branch-based DCSA Block and evaluated the performance of different branch numbers and large kernel decomposition sequences. The selection of kernel decomposition sequences is guided by the following principles:
1. Diverse Receptive Fields Coverage: The primary objective is to cover a wide spectrum of receptive fields. A larger RF, achieved through larger kernels ( k ) or dilation rates ( d ), enables the model to capture broader contextual cues, which are vital for distinguishing small objects from complex backgrounds. A smaller receptive field focuses on fine-grained local details.
2. Computational Efficiency: Our decomposition strategy, breaking a large kernel into two smaller depthwise convolutions (e.g., ( k 1 , d 1 ) and ( k 2 , d 2 )), significantly reduces the parameter count and computational load compared to a single large dense kernel. The choice of k 1 and k 2 is kept small (e.g., 3, 5, 7) to maintain this efficiency.
3. Avoiding Gridding Artifacts: While a large dilation rate d can rapidly expand the RF, it can also lead to the “gridding effect [47],” where the convolution kernel samples the input feature map sparsely, potentially missing important local information. Our sequential decomposition mitigates this by combining a standard convolution ( d = 1) with a dilated one, ensuring both dense local coverage and expanded contextual view.
As shown in Table 1, when using a three-branch configuration, the sequence producing RFs of {15, 23, 31} (derived from kernel sequences {(3, 1), (7, 2)}, {(5, 1), (7, 3)}, and {(7, 1), (9, 3)}) yielded the best mAP of 75.05%. This configuration provides a balanced and well-spaced distribution of receptive fields, covering small, medium, and large contextual areas without significant overlap. In contrast, increasing the number of branches to four (e.g., with RFs {15, 23, 29, 39}) did not lead to further improvement and decreased the inference speed (FPS), suggesting that three distinct contextual ranges are sufficient for this task and additional branches may introduce redundancy. The baseline model, with a single RF of 23, lacks this flexibility, resulting in a lower mAP of 74.72%. This confirms that our multi-branch decomposition strategy, guided by the principles above, is effective at enhancing feature extraction capabilities in an efficient manner.
Ablation study on the impact of DGFFN on parameter count, complexity, and performance. To assess the efficacy and robustness of the introduced DGFFN module, ablation studies were performed across three distinct backbones: LSKNet [14], Swin-t [43], and DCBANet (Ours). Table 2 shows the number of parameters, FLOPs, mAP, and FPS for models incorporating either the standard MLP or the proposed DGFFN within these backbone networks. Additionally, we also evaluate the performance of GSAU [47] and CAFFN [48], which are based on gating mechanisms.
The experimental results clearly demonstrate the lightweight advantages of our proposed DGFFN across various backbone architectures. Specifically, when integrated into LSKNet, DGFFN reduces the parameter count by 15.3% and FLOPs by 17.2%. A similar trend is observed with Swin-t, where the parameters and FLOPs are reduced by 15.4% and 14.0%, respectively. Within our own DCBANet, the reduction is even more pronounced, with a 15.7% decrease in parameters and a 17.8% drop in FLOPs. Importantly, this significant reduction in computational cost does not compromise performance; on the contrary, it leads to a notable improvement in inference speed (FPS) across all tested backbones, while maintaining or even slightly improving the mAP. This consistently demonstrates the effectiveness and generalizability of DGFFN as a lightweight yet powerful replacement for standard MLP modules. In contrast, although GSAU and CAFFN also achieve reductions in parameters and FLOPs to varying degrees, they result in a loss of model accuracy. Particularly, GSAU significantly reduces the parameter count by drastically decreasing the hidden feature dimensions, but this severely compromises model performance. This elucidates the rationale behind DGFFN’s retention of high-dimensional hidden features to strike an optimal trade-off between computational efficiency and model performance.
Ablation study on the effectiveness of the DCSA Block, EBA, and DGFFN. To evaluate the effectiveness of each component in DCBANet, we adopt LSKNet as the baseline model and gradually replace or integrate our proposed DCSA Block, EBA, and DGFFN modules. Experiments are then conducted on the DOTA, NWPU VHR-10, SIMD, and AI-TOD dataset to observe their impact on performance. The experimental results are presented in Table 3.
The first step was to replace the baseline’s standard large kernel attention with our proposed DCSA Block. As shown by comparing the first and second rows of Table 3, this single change brings consistent performance gains across all datasets. The mAP improved by 0.24% on DOTA, 0.63% on NWPU VHR-10, and 1.96% on SIMD, and the mAP@0.5 on AI-TOD increased by 1.6%. Notably, the m A P s on SIMD saw a significant rise of 2.60%, and the m A P s on AI-TOD also improved from 38.2% to 38.7%. This demonstrates that the DCSA Block’s ability to dynamically capture multi-range contextual information is a robust and effective strategy for enhancing feature representation, especially for small objects.
Next, we introduced the EBA module to the model already equipped with the DCSA Block. The results show another wave of consistent improvements. The mAP further increased on DOTA (+0.21%), NWPU VHR-10 (+0.39%), and SIMD (+0.74%). On the challenging AI-TOD dataset, the m A P s score was further boosted to 39.7%. This consistent improvement across diverse datasets validates the effectiveness of the EBA. By adaptively reweighting the outputs of the parallel branches within the DCSA Block, the EBA successfully enhances critical features while suppressing less informative ones, leading to a more powerful and discriminative feature fusion.
Finally, we replaced the standard FFN with our lightweight DGFFN. The last row of Table 3 shows the performance of our full model. This change not only significantly reduced the model’s complexity (parameters from 13.99 M to 11.28 M, a 19.4% reduction; FLOPs from 51.33 G to 43.14 G, a 15.9% reduction) but also further improved or maintained the detection accuracy. The final model achieves the highest mAP on all four datasets: 77.33% on DOTA, 89.17% on NWPU VHR-10, 80.27% on SIMD, and a state-of-the-art mAP@0.5 of 65.0% on AI-TOD. This demonstrates that DGFFN is a highly effective module that successfully reduces computational overhead without sacrificing, and in some cases even enhancing, the model’s detection performance.
In summary, the comprehensive ablation studies across multiple datasets with varying characteristics robustly demonstrate that each component of our proposed DCBANet contributes positively and universally to the final performance, validating the soundness of our overall design.

4.4. Main Result

To achieve optimal performance, our backbone network was first pre-trained on the MS COCO dataset for 100 epochs and then fine-tuned with a 1× schedule on the DOTA dataset and a 3× schedule on the NWPU VHR-10 and SIMD datasets.
Results on the DOTA Dataset. A performance comparison between DCBANet and other leading methods on the DOTA dataset is presented in Table 4. Our method achieves optimal performance under the RoI Transformer [4] detection framework, obtaining mAP scores of 77.33% and 80.91% under single-scale and multi-scale training strategies, respectively. In the single-scale training results, the DCBANet ranks first in mAP among all compared methods. After multi-scale training, the model achieved significant performance improvements, attaining state-of-the-art accuracy in the majority of object categories.
Results on the NWPU-VHR10 Dataset. Table 5 shows the results on the NWPU VHR-10 dataset, alongside comparisons with other state-of-the-art methods. Our method achieves a mAP of 89.17%, outperforming all competing methods, with a 1.14% improvement over the baseline model. However, our method exhibits relatively lower detection accuracy on the “ship” category, mainly because images containing ships in this dataset provide limited contextual information for detection. A comprehensive analysis will be provided in Section 4.5.
Results on the SIMD Dataset. We measured the performance of the DCBANet against seven other state-of-the-art methods on the SIMD dataset under the same experimental settings, as shown in Table 6. The DCBANet achieves state-of-the-art performance with minimal parameters and lowest computational complexity, attaining 80.27% mAP and surpassing all competing methods. Notably, the DCBANet reaches a state-of-the-art A P s of 67.45%, outperforming the second-best method by 1.40% and significantly exceeding other methods. This emphasizes the crucial role of dynamic contextual information in detecting small-scale objects. Nevertheless, the detection results of the DCBANet on medium- and large-sized objects are marginally lower than those of certain comparative methods, indicating its current shortcomings in handling multi-scale targets.
Results on the AI-TOD Dataset. To rigorously evaluate the performance of our proposed method on a dedicated small object detection benchmark, we conducted comprehensive experiments on the AI-TOD dataset. The comparative results against several state-of-the-art methods are presented in Table 7.
The results clearly demonstrate the superior performance of DCBANet, particularly in detecting small objects. Our method achieves a mean Average Precision (mAP@0.5:0.95) of 42.4%, surpassing all other competing methods. Notably, this represents a 2.0% improvement over the strong baseline LSKNet [14] and a significant 7.7% lead over ResNeSt [42].
The most compelling evidence of our model’s effectiveness lies in the m A P s metric, which specifically evaluates performance on small objects. DCBANet achieves an m A P s score of 39.9%, outperforming the second-best method, LSKNet, by 1.7%. This result is particularly significant as it directly validates our core design philosophy of enhancing feature representation for small targets through dynamic contextual information. When compared to other well-established backbones like Swin-t [43] and ResNeSt [42], our method shows an impressive improvement of 4.6% and 4.8% in m A P s , respectively, highlighting the substantial advantage of our proposed architecture in this challenging scenario.

4.5. Visual Analytics

The detection results across three datasets are visualized in Figure 7. Our method demonstrates effective detection of objects at various scales, particularly in complex environments where the DCBANet leverages surrounding contextual information to assist small target detection. However, as shown in the first and second images of the second row, certain ships remain undetected. In these cases, the ships exhibit severe background fusion (characterized by low contrast and texture similarity) while lacking any meaningful contextual information in the entire image, making them particularly challenging to identify. In contrast, the ships in the first row, which are located near shorelines with the key contextual cue of “harbor” present, achieve significantly higher detection accuracy. This comparison reveals a critical limitation of our method in detecting weak targets under conditions of both visual ambiguity and contextual scarcity.
To better illustrate the advantages of the DCBANet over other methods, Figure 8 compares the detection results of the DCBANet and other advanced methods in challenging detection scenes. As shown, the proposed DCBANet generally achieves superior detection results compared to the other methods presented. Specifically, in the complex environment of Figure 8a, the DCBANet is the only method that can identify all small objects. Furthermore, for scenes with dense targets (Figure 8b) and shadow occlusion (Figure 8c), the DCBANet exhibits fewer missed detections and classification errors than the other advanced methods.

5. Conclusions

In order to address the key challenges of small target detection in remote sensing images, we introduced DCBANet, a novel backbone network designed to dynamically perceive and adapt to contextual cues. Our contributions are threefold: (1) the DCSA Block, featuring a multi-branch architecture and internal CASM units, which adaptively capture contextual information across various receptive fields; (2) the EBA module, which intelligently fuses features from different branches by prioritizing the most informative ones; and (3) the lightweight DGFFN, which significantly reduces computational complexity while maintaining high performance.
Our comprehensive experiments on four challenging public datasets validate the superiority of our approach. The DCBANet achieves state-of-the-art mAP@0.5 scores of 80.79% on DOTA, 89.17% on NWPU VHR-10, and 80.27% on SIMD. More importantly, on the dedicated small object benchmark AI-TOD, our model achieves a remarkable 42.4% mAP@0.5:0.95, with significant improvements in small object-specific metrics ( m A P s ) over previous methods. This robustly demonstrates that our dynamic context-aware design is highly effective for tackling the small object detection problem.
Despite these promising results, there remains room for improvement. Our current model focuses primarily on local context and could be enhanced by better integrating global scene-level information to improve performance on multi-scale and weakly defined targets. Future work will explore more advanced feature fusion techniques and aim to develop a more holistic context-aware system for Remote Sensing Object Detection.

Author Contributions

Conceptualization, H.J.; methodology, Y.S. and T.B.; software, H.J. and Y.S.; validation, H.J. and T.B.; writing—original draft preparation, Y.S. and T.B.; writing—review and editing, H.J. and T.B.; visualization, Y.C.; supervision, K.S. and T.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Hubei Provincial Natural Science Foundation (Grant No. 2025AFB061), the Hubei University of Technology Doctoral Research Startup Project (Grant No. XJ2024004101), and the National Natural Science Foundation of China (Grant No. 42301457).

Data Availability Statement

The datasets presented in this study can be downloaded here: https://captain-whu.github.io/DOTA/dataset.html (DOTA). https://gcheng-nwpu.github.io/\#Datasets (NWPU VHR-10). https://github.com/ihians/simd (SIMD). https://github.com/jwwangchn/AI-TOD (AI-TOD).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RSODRemote Sensing Object Detection
RSIsRemote Sensing Images
DCBANetDynamic Context Branch Attention Network
DCSA BlockDynamic Context Scale-Aware Block
EBAEfficient Branch Attention
FFNFeedforward Network
DGFFNDual-Gated Feedforward Network
SGUSpatial Gating Unit
CGUChannel Gating Unit
MLPMultilayer Perceptron
FLOPsFloating-point Operations
APAverage Precision
mAPMean Average Precision
FPSFrames Per Second

References

  1. Gui, S.; Song, S.; Qin, R.; Tang, Y. Remote Sensing Object Detection in the Deep Learning Era—A Review. Remote Sens. 2024, 16, 327. [Google Scholar] [CrossRef]
  2. Guo, Z.; Liu, C.; Zhang, X.; Jiao, J.; Ji, X.; Ye, Q. Beyond Bounding-Box: Convex-hull Feature Adaptation for Oriented and Densely Packed Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8788–8797. [Google Scholar] [CrossRef]
  3. Lyu, C.; Zhang, W.; Huang, H.; Zhou, Y.; Wang, Y.; Liu, Y.; Zhang, S.; Chen, K. RTMDet: An Empirical Study of Designing Real-Time Object Detectors. arXiv 2022, arXiv:2212.07784. [Google Scholar]
  4. Ding, J.; Xue, N.; Long, Y.; Xia, G.-S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2844–2853. [Google Scholar] [CrossRef]
  5. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3500–3509. [Google Scholar] [CrossRef]
  6. Cheng, G.; Yuan, X.; Han, X.J. Towards Large Scale Small Object Detection: Survey and Benchmarks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13467–13488. [Google Scholar] [CrossRef]
  7. Chen, G.; Wang, H.; Chen, K.; Li, Z.; Song, Z. A Survey of the Four Pillars for Small Object Detection: Multiscale Representation, Contextual Information, Super-Resolution, and Region Proposal. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 936–953. [Google Scholar] [CrossRef]
  8. Hua, X.; Cui, X.; Xu, X. Weakly Supervised Underwater Object Real-time Detection Based on High-resolution Attention Class Activation Mapping and Category Hierarchy. Pattern Recognit. 2025, 159, 111111. [Google Scholar] [CrossRef]
  9. Gao, T.; Xia, S.; Liu, M. MSNet: Multi-Scale Network for Object Detection in Remote Sensing Images. Pattern Recognit. 2025, 158, 110983. [Google Scholar] [CrossRef]
  10. Chen, C.; Zeng, W.; Zhang, X. HFPNet: Super Feature Aggregation Pyramid Network for Maritime Remote Sensing Small-Object Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 5973–5989. [Google Scholar] [CrossRef]
  11. Wei, C.; Bai, L.; Chen, X.; Han, J. Cross-Modality Data Augmentation for Aerial Object Detection with Representation Learning. Remote Sens. 2024, 16, 4649. [Google Scholar] [CrossRef]
  12. Cheng, G.; Wang, J.; Li, K.; Xie, X. Anchor-Free Oriented Proposal Generator for Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625411. [Google Scholar] [CrossRef]
  13. Song, G.; Du, H.; Zhang, X. Small Object Detection in Unmanned Aerial Vehicle Images Using Multi-Scale Hybrid Attention. Eng. Appl. Artif. Intell. 2024, 128, 107455. [Google Scholar] [CrossRef]
  14. Li, Y.; Hou, Q.; Zheng, Z.; Cheng, M.-M.; Yang, J.; Li, X. Large Selective Kernel Network for Remote Sensing Object Detection. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 16748–16759. [Google Scholar] [CrossRef]
  15. Cui, L.; Lv, P.; Jiang, X.; Gao, Z. Context-Aware Block Net for Small Object Detection. IEEE Trans. Cybern. 2022, 52, 2300–2313. [Google Scholar] [CrossRef]
  16. He, X.; Zheng, X.; Hao, X.; Jin, H.; Zhou, X.; Shao, L. Improving Small Object Detection via Context-Aware and Feature-Enhanced Plug-and-Play Modules. J. Real-Time Image Process. 2024, 21, 44. [Google Scholar] [CrossRef]
  17. Wu, C.; Zeng, Z. YOLOX-CA: A Remote Sensing Object Detection Model Based on Contextual Feature Enhancement and Attention Mechanism. IEEE Access 2024, 12, 84632. [Google Scholar] [CrossRef]
  18. Li, Z.; Wang, Y.; Zhang, Y. Context Feature Integration and Balanced Sampling Strategy for Small Weak Object Detection in Remote Sensing Imagery. IEEE Geosci. Remote Sens. Lett. 2024, 21, 6009105. [Google Scholar] [CrossRef]
  19. Ding, S.; Xiong, M.; Wang, X. Dynamic feature and context enhancement network for faster detection of small objects. Expert Syst. Appl. 2025, 265, 125732. [Google Scholar] [CrossRef]
  20. Yang, H.; Qiu, S. A Novel Dynamic Contextual Feature Fusion Model for Small Object Detection in Satellite Remote-Sensing Images. Information 2024, 15, 230. [Google Scholar] [CrossRef]
  21. Wang, B.; Ji, R.; Zhang, L. Learning to zoom: Exploiting mixed-scale contextual information for object detection. Expert Syst. Appl. 2025, 264, 125871. [Google Scholar] [CrossRef]
  22. Yang, X.; Liu, Q.; Yan, J.; Li, A. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. arXiv 2019, arXiv:1908.05612. [Google Scholar] [CrossRef]
  23. Han, J.; Ding, J.; Li, J.; Xia, G. Align Deep Features for Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5602511. [Google Scholar] [CrossRef]
  24. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar] [CrossRef]
  25. Zhang, H.; Chang, H.; Ma, B. Dynamic R-CNN: Towards High Quality Object Detection via Dynamic Training. In Computer Vision–ECCV 2020; Lecture Notes in Computer Science, Volume 12360; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Springer: Cham, Switzerland, 2020; pp. 260–275. [Google Scholar] [CrossRef]
  26. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  27. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Computer Vision–ECCV 2020; Lecture Notes in Computer Science, Volume 12346; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Springer: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar] [CrossRef]
  28. Zhao, J.; Ding, Z.; Zhou, Y.; Zhu, H.; Du, W. OrientedFormer: An End-to-End Transformer-Based Oriented Object Detector in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5640816. [Google Scholar] [CrossRef]
  29. Zhu, X.; Su, W.; Lu, L. Deformable DETR: Deformable Transformers for End-to-End Object Detection. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 3–7 May 2021; pp. 1–16. [Google Scholar]
  30. Yang, C.; Huang, Z.; Wang, N. QueryDet: Cascaded Sparse Query for Accelerating High-Resolution Small Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13658–13667. [Google Scholar] [CrossRef]
  31. Deng, C.; Wang, M.; Liu, L. Extended Feature Pyramid Network for Small Object Detection. IEEE Trans. Multimed. 2022, 24, 1968–1979. [Google Scholar] [CrossRef]
  32. Zeng, N.; Wu, P.; Wang, Z. A Small-Sized Object Detection Oriented Multi-Scale Feature Fusion Approach with Application to Defect Detection. IEEE Trans. Instrum. Meas. 2022, 71, 3507014. [Google Scholar] [CrossRef]
  33. Liu, Y.; Li, Q.; Yuan, Y.; Du, Q.; Wang, Q. ABNet: Adaptive Balanced Network for Multiscale Object Detection in Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5614914. [Google Scholar] [CrossRef]
  34. Yang, X.; Yan, J.; Liao, W.; Yang, X.; Tang, J.; He, T. SCRDet++: Detecting Small, Cluttered and Rotated Objects via Instance-Level Feature Denoising and Rotation Loss Smoothing. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 2384–2399. [Google Scholar] [CrossRef]
  35. Vaswani, A.; Shazeer, N.M.; Parmar, N. Attention is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  36. Guo, M.; Xu, T.; Liu, J. Attention Mechanisms in Computer Vision: A Survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
  38. Wang, Q.; Wu, B.; Zhu, P. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar] [CrossRef]
  39. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Vedaldi, A. Gather-Excite: Exploiting Feature Context in Convolutional Neural Networks. arXiv 2018, arXiv:1810.12348. [Google Scholar] [CrossRef]
  40. Zhao, H.; Zhang, Y.; Liu, S.; Shi, J. PSANet: Point-wise Spatial Attention Network for Scene Parsing. In Computer Vision–ECCV 2018; Lecture Notes in Computer Science, Vol. 11213; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; pp. 270–286. [Google Scholar] [CrossRef]
  41. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  42. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y. ResNeSt: Split-Attention Networks. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 2735–2745. [Google Scholar] [CrossRef]
  43. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  44. Chen, Z.; Zhu, Y.; Li, Z.; Yang, F.; Zhao, C.; Wang, J.; Tang, M. The Devil is in Details: Delving Into Lite FFN Design for Vision Transformers. In Proceedings of the ICASSP 2024–2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 4130–4134. [Google Scholar] [CrossRef]
  45. Xu, H.; Zhou, Z.; He, D. Vision Transformer with Attention Map Hallucination and FFN Compaction. arXiv 2023, arXiv:2306.10875. [Google Scholar]
  46. Yang, Z.; Zhu, L.; Wu, Y. Gated Channel Transformation for Visual Recognition. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11791–11800. [Google Scholar] [CrossRef]
  47. Wang, Y.; Li, Y.; Wang, G. Multi-Scale Attention Network for Single Image Super-Resolution. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 5950–5960. [Google Scholar] [CrossRef]
  48. Li, S.; Wang, Z.; Liu, Z.; Tan, C.; Lin, H.; Wu, D.; Chen, Z.; Zheng, J.; Li, S. MogaNet: Multi-Order Gated Aggregation Network. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 25–29 April 2022. [Google Scholar]
  49. Guo, M.H.; Lu, C.Z.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Visual Attention Network. Comput. Vis. Media 2023, 9, 733–752. [Google Scholar] [CrossRef]
  50. Yu, W.; Si, C.; Zhou, P.; Luo, M. MetaFormer Baselines for Vision. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 896–912. [Google Scholar] [CrossRef]
  51. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar] [CrossRef]
  52. Haroon, M.; Shahzad, M.; Fraz, M. Multisized Object Detection Using Spaceborne Optical Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3032–3046. [Google Scholar] [CrossRef]
  53. Cheng, G.; Han, J.; Zhou, P. Multi-Class Geospatial Object Detection and Geographic Image Classification Based on Collection of Part Detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
  54. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G. Tiny Object Detection in Aerial Images. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 3791–3798. [Google Scholar] [CrossRef]
  55. Zeng, Y.; Chen, Y.; Yang, X. ARS-DETR: Aspect Ratio-Sensitive Detection Transformer for Aerial Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5610315. [Google Scholar] [CrossRef]
  56. Xie, X.; Cheng, G.; Rao, C. Oriented Object Detection via Contextual Dependence Mining and Penalty-Incentive Allocation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5618010. [Google Scholar] [CrossRef]
  57. Hou, L.; Lu, K.; Xue, J. Shape-Adaptive Selection and Measurement for Oriented Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence; 2022; Volume 36, pp. 923–932. [Google Scholar] [CrossRef]
  58. Cheng, G.; Yao, Y.; Li, S.; Li, K. Dual-Aligned Oriented Detector. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618111. [Google Scholar] [CrossRef]
  59. Han, J.; Ding, J.; Xue, N.; Xia, G. ReDet: A Rotation-Equivariant Detector for Aerial Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2785–2794. [Google Scholar] [CrossRef]
  60. Lin, T.; Goyal, P.; Girshick, R. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef]
  61. Tian, Z.; Shen, C.; Chen, H. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar] [CrossRef]
  62. Chen, Q.; Wang, Y.; Yang, T. You Only Look One-Level Feature. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 13034–13043. [Google Scholar] [CrossRef]
  63. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the DCBANet backbone network. It is divided into four stages, each of which consists of a series of serially stacked DCBA-Formers. The embedding dimensions C i at each stage are set to {64, 128, 320, 512}, and the number of DCBA-Formers. N i is configured as {2, 2, 4, 2}.
Figure 1. The overall architecture of the DCBANet backbone network. It is divided into four stages, each of which consists of a series of serially stacked DCBA-Formers. The embedding dimensions C i at each stage are set to {64, 128, 320, 512}, and the number of DCBA-Formers. N i is configured as {2, 2, 4, 2}.
Remotesensing 17 02415 g001
Figure 2. The structure of a DCBA-Former.
Figure 2. The structure of a DCBA-Former.
Remotesensing 17 02415 g002
Figure 3. The architecture of the proposed DCSA Block. It consists of multiple parallel branches to generate features with diverse receptive fields. Each branch contains a Context Adaptive Selection Module (CASM), shown in detail on the left, which dynamically weights features from smaller ( U s ) and larger ( U l ) receptive fields. The outputs from all branches are subsequently processed by an EBA. The detailed structure of the EBA is depicted in Figure 4.
Figure 3. The architecture of the proposed DCSA Block. It consists of multiple parallel branches to generate features with diverse receptive fields. Each branch contains a Context Adaptive Selection Module (CASM), shown in detail on the left, which dynamically weights features from smaller ( U s ) and larger ( U l ) receptive fields. The outputs from all branches are subsequently processed by an EBA. The detailed structure of the EBA is depicted in Figure 4.
Remotesensing 17 02415 g003
Figure 4. The flowchart of the EBA module. c i denotes the number of channels in the feature maps of each branch. c represents the total number of channels across all branches.
Figure 4. The flowchart of the EBA module. c i denotes the number of channels in the feature maps of each branch. c represents the total number of channels across all branches.
Remotesensing 17 02415 g004
Figure 5. The structure of the Dual-Gated FFN. The SGU and CGU based on gating mechanisms are introduced into the traditional MLP.
Figure 5. The structure of the Dual-Gated FFN. The SGU and CGU based on gating mechanisms are introduced into the traditional MLP.
Remotesensing 17 02415 g005
Figure 6. Distribution of object scales in the DOTA, SIMD, NWPU VHR-10, and AI-TOD datasets. The scales are categorized as: normal (1000–4096 pixels), general small (256–1024 pixels), tiny (64–256 pixels), and very tiny (4–64 pixels).
Figure 6. Distribution of object scales in the DOTA, SIMD, NWPU VHR-10, and AI-TOD datasets. The scales are categorized as: normal (1000–4096 pixels), general small (256–1024 pixels), tiny (64–256 pixels), and very tiny (4–64 pixels).
Remotesensing 17 02415 g006
Figure 7. Visualization of detection performance across four challenging aerial object detection datasets. Each row showcases examples from a different dataset. From top to bottom: DOTA, VHR-10, SIMD, and AI-TOD. Yellow ellipses indicate missed detections.
Figure 7. Visualization of detection performance across four challenging aerial object detection datasets. Each row showcases examples from a different dataset. From top to bottom: DOTA, VHR-10, SIMD, and AI-TOD. Yellow ellipses indicate missed detections.
Remotesensing 17 02415 g007
Figure 8. Comparison of the performance of the DCBANet and other advanced methods in detecting small targets in challenging scenarios. (a) Complex environment. (b) Dense distribution. (c) Shadow occlusion. The first column represents the input image, and the second column represents the ground truth within red dashed bounding boxes. The last four columns are the detection results of DCBANet, LSKNet [14], Cascade-RCNN [24], and FCOS [62], respectively. The DCBANet achieves the best results.
Figure 8. Comparison of the performance of the DCBANet and other advanced methods in detecting small targets in challenging scenarios. (a) Complex environment. (b) Dense distribution. (c) Shadow occlusion. The first column represents the input image, and the second column represents the ground truth within red dashed bounding boxes. The last four columns are the detection results of DCBANet, LSKNet [14], Cascade-RCNN [24], and FCOS [62], respectively. The DCBANet achieves the best results.
Remotesensing 17 02415 g008
Table 1. Branch number and large kernel decomposition sequence partial configurations and results, where RF denotes large kernel receptive fields, k denotes the convolution kernel, and d denotes dilation rate. Experiments were conducted on the DOTA dataset with a training schedule of 3×. The initial weights of all models in this experiment were randomly initialized.
Table 1. Branch number and large kernel decomposition sequence partial configurations and results, where RF denotes large kernel receptive fields, k denotes the convolution kernel, and d denotes dilation rate. Experiments were conducted on the DOTA dataset with a training schedule of 3×. The initial weights of all models in this experiment were randomly initialized.
GroupBranch NumberRF Kernel   Seq   ( k , d )mAPFPS
baseline123(5,1), (7,3)74.7220.7
DCSA Block215
27
(3,1), (7,2)
(3,1), (9,3)
74.1620.2
321
23
27
(3,1), (7,3)
(5,1), (7,3)
(3,1), (9,3)
74.5518.9
315
23
31
(3,1), (7,2)
(5,1), (7,3)
(7,1), (9,3)
75.0519.6
415
23
29
39
(3,1), (7,2)
(5,1), (7,3)
(5,1), (7,4)
(7,1), (9,4)
74.6815.2
Table 2. Performance comparison between standard MLP and DGFFN. The proposed DGFFN reduces both parameters (Param) and FLOPs while improving detection performance (mAP) and inference speed (FPS) across all architectures. Experiments were conducted on the DOTA dataset with a training schedule of 3×. The initial weights of all models in this experiment were randomly initialized. ‘↑’ and ‘↓’ indicate the rise or fall of each metric, respectively.
Table 2. Performance comparison between standard MLP and DGFFN. The proposed DGFFN reduces both parameters (Param) and FLOPs while improving detection performance (mAP) and inference speed (FPS) across all architectures. Experiments were conducted on the DOTA dataset with a training schedule of 3×. The initial weights of all models in this experiment were randomly initialized. ‘↑’ and ‘↓’ indicate the rise or fall of each metric, respectively.
BackboneFFNParam (M)FLOPs (G)mAP (%)FPS
LSKNet [14]MLP13.8454.3174.7220.7
DGFFN11.72(↓15.3%)44.96(↓17.2%)74.89(↑0.17%)24.1(↑16.4%)
Swin-t [43]MLP27.4995.0373.9627.7
DGFFN23.27(↓15.4%)81.73(↓14.0%)74.12(↑0.16%)29.6(↑6.4%)
DCBANetMLP13.3852.5075.0519.6
DGFFN11.28(↓15.7%)43.14(↓17.8%)75.16(↑0.11%)23.4(↑16.2%)
GSAU [47]8.18 (↓38.8%)27.06(↓48.4%)74.02(↓1.03%)24.4(↑24.5%)
CAFFN [48]12.39(↓7.3%)51.62(↓1.6%)74.32(↓0.73%)19.2(↓1.0%)
Table 3. Ablation study on the effectiveness of the DCSA Block, EBA, and DGFFN across four datasets. The baseline model (first row) is LSKNet [14]. ‘×’ and ‘√’ indicate exclusion or inclusion of the corresponding module, respectively. Since there are no objects with a large scale in the AITOD dataset, the m A P l metric values are all 0.
Table 3. Ablation study on the effectiveness of the DCSA Block, EBA, and DGFFN across four datasets. The baseline model (first row) is LSKNet [14]. ‘×’ and ‘√’ indicate exclusion or inclusion of the corresponding module, respectively. Since there are no objects with a large scale in the AITOD dataset, the m A P l metric values are all 0.
DCSA BlockEBADGFFNParamFLOPsDOTANWPU
VHR-10
SIMDAI-TOD
mAPmAPmAP m A P s m A P m m A P l mAP@0.5mAP@0.75 m A P s m A P m
×××13.84 m54.31 G76.7988.0377.1063.7586.5391.0363.244.938.276.3
××13.28 m49.96 G77.0388.6679.0666.3589.2791.1064.346.038.777.7
×13.99 m51.33 G77.2489.0579.8067.1790.3691.3264.846.839.778.9
11.28 m43.14 G77.3389.1780.2767.4591.1491.8565.047.139.979.3
Table 4. Comparison results between the DCBANet and other state-of-the-art methods on the DOTA dataset, where * denotes multi-scale training. The best results for each category are highlighted in bold, while the second-best results are underlined. The training schedule was set to 1×.
Table 4. Comparison results between the DCBANet and other state-of-the-art methods on the DOTA dataset, where * denotes multi-scale training. The best results for each category are highlighted in bold, while the second-best results are underlined. The training schedule was set to 1×.
MethodBackbonemAPPLBDBRGTFSVLVSHTCBCSTSBFRAHASPHC
End-to-end
D.DETR-O [29]R5069.4884.8970.7146.0461.9273.9978.8387.7190.0777.9778.4147.0754.4866.8767.6655.62
ARS-DETR [55]R5074.1686.9775.5648.3269.2077.9277.9487.6990.5077.3182.8660.2864.5874.8871.7666.62
OrientedFormer [28]R5075.3788.1479.1351.9667.3481.0283.2688.2990.9085.5786.2560.8466.3673.8171.2356.49
One-stage
R3Det [22]R10173.7988.7683.0950.9167.2776.2380.3986.7290.7884.6883.2461.9861.3566.9170.6353.94
S2Anet [23]R5074.1289.1182.8448.3771.1178.1178.3887.2590.8384.9085.6460.3662.6065.2663.1957.94
CFA [2]R5074.5188.3483.0951.9272.2379.9578.6887.2590.9085.3885.7159.6363.0573.3370.3647.86
DFDet [56]R5074.7188.9279.2548.4070.0080.2278.8587.2190.9083.1383.9260.0766.4968.2776.7858.11
SASM [57]R5074.9286.4278.9752.4769.8477.3075.9986.7290.8982.6385.6660.1368.2573.9872.2262.37
RTMDet [3]CSPNeXt-t75.3689.2180.0347.8869.7382.0583.3388.6390.9186.3186.8559.9452.3074.2371.9757.03
Two-stage
RoI Transformer [4]R5074.6188.6582.6052.5370.8777.9376.6786.8790.7183.8352.8153.9567.6174.6768.7561.03
AOPG [12]R5075.2489.2783.4952.5069.9773.5182.3187.9590.8987.6484.7160.0166.1274.1968.3057.80
DODet [58]R5075.4989.3484.3151.3971.0479.0482.8688.1590.9086.8884.9162.6967.6375.4772.2245.54
Oriented RCNN [5]R5075.8789.4682.1254.7870.8678.9383.0088.2090.9087.5084.6863.9767.6374.9468.8452.28
SCRDet++ [34]R10176.2089.7783.9056.3073.9872.6075.6382.8290.7687.8986.1465.2463.1776.0568.0670.24
ReDet [59]ReR5076.2588.7982.6453.9774.0078.1684.0688.0490.8987.7885.7561.7660.3975.9668.0763.59
DCBANet -77.3389.7882.3453.5976.5580.3084.8988.4590.9087.6585.2458.4464.8076.2477.8262.90
DCBANet *-80.9189.0485.1160.6681.9081.2185.4888.4490.8788.3187.9069.7866.9179.2081.5877.25
Table 5. Comparison results of the DCBANet with other state-of-the-art methods on the NWPU-VHR10 dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 3×.
Table 5. Comparison results of the DCBANet with other state-of-the-art methods on the NWPU-VHR10 dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 3×.
MethodmAPPLSHSTBDTCBCGTFHABRVE
RetinaNet [60]81.8098.5077.4995.1896.6672.0573.9099.1773.3363.6068.11
FCOS [61]82.8799.7775.8795.2897.7371.0577.3198.6778.2566.3568.48
YOLOF [62]85.3999.3282.9882.7898.3172.8987.3099.1683.7980.2067.20
Faster-RCNN [26]85.4399.9276.6889.7197.7784.6283.5499.7882.6676.1063.56
Cascade-RCNN [24] 86.1199.3673.0093.0198.1084.4379.4799.9591.9877.0964.69
Dynamic-RCNN [25] 87.4099.6973.4592.9396.6788.8386.8799.8887.3882.2266.03
LSKNet [14]88.0399.7882.0995.5497.2479.5473.7999.9888.1988.7975.40
DCBANet89.1799.5274.9790.4897.4191.7686.3499.8888.8889.9172.55
Table 6. Performance comparison of the DCBANet with other state-of-the-art methods on the SIMD dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 3×.
Table 6. Performance comparison of the DCBANet with other state-of-the-art methods on the SIMD dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 3×.
MethodParam(M)FLOPs(G)mAP m A P s m A P m m A P l
FCOS [61]32.15206.7270.8853.4384.2488.04
Cascade-RCNN [24]41.42225.0675.9559.7588.9291.43
LSKNet [14]31.03187.1677.1063.7586.5391.03
ResNeXt [63]40.89229.0577.6362.4390.6191.26
Dynamic-RCNN [25]41.75225.7278.3162.8591.3692.32
ResNeSt [42]44.66481.3079.5766.0591.3491.38
Swin-t [43]44.81233.0080.0865.7892.6592.63
DCBANet28.49176.0580.2767.4591.1491.85
Table 7. Performance comparison of the DCBANet with other state-of-the-art methods on the AI-TOD dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 2×.
Table 7. Performance comparison of the DCBANet with other state-of-the-art methods on the AI-TOD dataset. Optimal results are denoted in boldface, and sub-optimal results are underlined. The training schedule was set to 2×.
MethodmAP@0.5:0.95mAP@0.5mAP@0.75 m A P s m A P m
YOLOF [62]8.520.06.28.625.9
Faster-RCNN [26]32.260.733.831.666.7
Dynamic-RCNN [25]34.560.134.933.562.2
Swin-t [43]34.757.835.535.170.5
ResNeSt [42]36.861.137.635.365.9
ResNeXt [63]37.763.140.535.972.7
LSKNet [14]40.463.244.938.276.6
DCBANet42.465.047.139.979.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, H.; Song, Y.; Bai, T.; Sun, K.; Chen, Y. A Novel Dynamic Context Branch Attention Network for Detecting Small Objects in Remote Sensing Images. Remote Sens. 2025, 17, 2415. https://doi.org/10.3390/rs17142415

AMA Style

Jin H, Song Y, Bai T, Sun K, Chen Y. A Novel Dynamic Context Branch Attention Network for Detecting Small Objects in Remote Sensing Images. Remote Sensing. 2025; 17(14):2415. https://doi.org/10.3390/rs17142415

Chicago/Turabian Style

Jin, Huazhong, Yizhuo Song, Ting Bai, Kaimin Sun, and Yepei Chen. 2025. "A Novel Dynamic Context Branch Attention Network for Detecting Small Objects in Remote Sensing Images" Remote Sensing 17, no. 14: 2415. https://doi.org/10.3390/rs17142415

APA Style

Jin, H., Song, Y., Bai, T., Sun, K., & Chen, Y. (2025). A Novel Dynamic Context Branch Attention Network for Detecting Small Objects in Remote Sensing Images. Remote Sensing, 17(14), 2415. https://doi.org/10.3390/rs17142415

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop