Next Article in Journal
Development of a Measurement Procedure for Emotional States Detection Based on Single-Channel Ear-EEG: A Proof-of-Concept Study
Previous Article in Journal
A Hybrid Simulation–Physical Data-Driven Framework for Occupant Injury Prediction in Vehicle Underbody Structures
Previous Article in Special Issue
Optimizing Dam Detection in Large Areas: A Hybrid RF-YOLOv11 Framework with Candidate Area Delineation
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rotation-Sensitive Feature Enhancement Network for Oriented Object Detection in Remote Sensing Images

College of Information Engineering and Artificial Intelligence, Henan University of Science and Technology, Luoyang 471000, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(2), 381; https://doi.org/10.3390/s26020381
Submission received: 5 December 2025 / Revised: 2 January 2026 / Accepted: 4 January 2026 / Published: 7 January 2026

Abstract

Oriented object detection in remote sensing images remains a challenging task due to arbitrary target rotations, extreme scale variations, and complex backgrounds. However, current rotated detectors still face several limitations: insufficient orientation-sensitive feature representation, feature misalignment for rotated proposals, and unstable optimization of rotation parameters. To address these issues, this paper proposes an enhanced Rotation-Sensitive Feature Pyramid Network (RSFPN) framework. Building upon the effective Oriented R-CNN paradigm, we introduce three novel core components: (1) a Dynamic Adaptive Feature Pyramid Network (DAFPN) that enables bidirectional multi-scale feature fusion through semantic-guided upsampling and structure-enhanced downsampling paths; (2) an Angle-Aware Collaborative Attention (AACA) module that incorporates orientation priors to guide feature refinement; (3) a Geometrically Consistent Multi-Task Loss (GC-MTL) that unifies the regression of rotation parameters with periodic smoothing and adaptive weight mechanisms. Comprehensive experiments on the DOTA-v1.0 and HRSC2016 benchmarks show that our RSFPN achieves superior performance. It attains a state-of-the-art mAP of 77.42% on DOTA-v1.0 and 91.85% on HRSC2016, while maintaining efficient inference at 14.5 FPS, demonstrating a favorable accuracy-efficiency trade-off. Visual analysis confirms that our method produces concentrated, rotation-aware feature responses and effectively suppresses background interference. The proposed approach provides a robust solution for detecting multi-oriented objects in high-resolution remote sensing imagery, with significant practical value for urban planning, environmental monitoring, and security applications.

1. Introduction

Advances in Earth observation technologies have enabled high-resolution satellites (e.g., WorldView, GF-2) and UAV platforms to capture global imagery with unprecedented spatial and temporal resolution. These datasets support critical applications, including smart city management [1], precision agriculture [2], disaster monitoring [3], maritime protection [4], and military reconnaissance [5]. Object detection, a core technique for intelligent remote sensing interpretation, aims to automatically localize and classify objects such as vehicles, ships, aircraft, and buildings [6]. While deep learning detectors like Faster R-CNN [7] and YOLO [8,9] have excelled on natural images, their direct application to remote sensing imagery remains challenging due to the “bird’s-eye view” perspective: objects often appear densely distributed with arbitrary orientations, complex backgrounds, and widely varying scales [10,11,12,13]. Using conventional horizontal bounding boxes (HBBs) to represent such objects introduces irrelevant background, increases overlap in dense regions, and degrades both localization and classification accuracy [14,15].
To address the limitations of horizontal bounding boxes, rotated object detection (ROD) employs oriented rectangles to tightly fit object contours and orientations [16,17,18,19]. In recent years, the field has evolved along two main paradigms. Single-stage detectors (e.g., R3Det [17], S2ANet [20]) prioritize efficiency by directly regressing rotated parameters, while two-stage detectors (e.g., RoI Transformer [21], Oriented R-CNN [19]) achieve higher accuracy through oriented region proposal and feature alignment mechanisms. Despite significant progress on benchmarks like DOTA [10], applying these detectors to complex remote sensing imagery still faces three fundamental challenges, which we aim to address systematically.
First, the feature representation challenge. Remote sensing imagery exhibits extreme scale variations and complex backgrounds, requiring detectors to possess multi-scale fusion and orientation-sensitive feature enhancement capabilities. While feature pyramid networks (FPN) [22] and variants like PANet [23] and BiFPN [24] improve semantic consistency, their axis-aligned convolutional kernels cannot adapt to the geometric deformations of rotated objects. This results in insufficient orientation sensitivity [18], geometric structure degradation [25], and multi-scale mismatch [26], collectively limiting accurate feature representation.
Second, the challenge of geometric alignment. The arbitrary orientations of rotated objects cause standard RoIAlign to misalign with the target geometry during feature extraction. Existing approaches, such as RoI Transformer [21], R2CNN [27], and Gliding Vertex [28], attempt geometric correction but often introduce high computational costs or remain sensitive to background noise in cluttered scenes. Enhancing directional consistency and feature focus efficiently remains a key challenge.
Third, the challenge of optimization. The parameter representation of rotated boxes (e.g., ( x , y , w , h , θ ) ) is more complex. Decoupled regression with losses like Smooth L1 [29] often leads to inconsistent gradients, unstable convergence [30], and a heavy reliance on manual weight tuning. Losses based on approximate rotational IoU (e.g., PIoU [31]) or Gaussian modeling (e.g., KLD [32], GWD [33]) have demonstrated improved sensitivity and continuity. However, they still face approximation errors under extreme aspect ratios or efficiency issues with small objects. Consequently, developing a loss function that can jointly optimize all geometric parameters with stable gradients in complex scenes remains a central challenge.
To address these interrelated challenges, we propose a key insight: if we can closely integrate rotation sensitive feature representation and geometric constraints in the whole detection process, we can construct a unified detection framework, which can significantly improve the accuracy and robustness of detection. Specifically, we believe that synergistically improving the ability of multi-scale feature learning, achieving accurate geometric alignment of the rotated regions, and stably optimizing the rotation parameters are three complementary keys. To this end, we propose a Rotation Sensitive Feature Pyramid Network (RSFPN)—an end-to-end trainable detection framework that aims to implement the above ideas holistically. RSFPN contains three core modules of collaborative design: Dynamic adaptive Feature Pyramid Network (DAFPN), which maintains cross-scale geometric consistency through bidirectional feature fusion. The Angle-Aware Collaborative Attention (AACA) module guided the feature enhancement along the main axis of the object according to the direction information, and Geometric Consistency Multi-Task Loss (GC-MTL), which unifies rotational parameter regression into a stable optimization objective. These modules work together so that the network remains sensitive to rotation from feature extraction to result output. We conduct comprehensive experiments on DOTA-v1.0 [10] and HRSC2016 [13] benchmarks to rigorously verify the validity of the above ideas and the performance of the RSFPN framework.

2. Related Work

2.1. Feature Pyramids and Multi-Scale Fusion

Feature Pyramid Networks (FPN) [22] remain fundamental for handling scale variation. Various extensions—including PANet [23], BiFPN [24], NAS-FPN [26], and ASFF [34]—enhance cross-scale fusion through more efficient or adaptive pathways. In remote sensing, models such as DRBox [35] and R-DFPN [36] incorporate multi-scale structures tailored for rotated objects.
However, existing approaches often overlook the preservation of fine-grained geometric cues essential for rotation-sensitive targets. Motivated by PANet and BiFPN, our DAFPN introduces semantically guided dynamic upsampling [37] and structure-preserving downsampling, enabling more precise bidirectional fusion and improved multi-scale geometric representations.
Beyond dedicated feature pyramid designs, mainstream multi-scale detection employs alternative paradigms. Single-stage detectors like RetinaNet [38] and its rotated variants [17] rely on dense anchor sampling across pyramid levels, while anchor-free methods such as FCOS [39] and CenterNet [40] assign scales based on point or center predictions. These methods often rely on heuristic scale assignment and may not explicitly preserve geometric consistency between feature levels -this is the core motivation for our structured bidirectional fusion of DAFPN.

2.2. Attention Mechanisms in Object Detection

Attention mechanisms adaptively reweight features, enabling networks to focus on target-relevant information while suppressing background noise. Channel attention was first introduced by SENet [41], and CBAM [42] sequentially integrates channel and spatial attention for a lightweight general-purpose structure. GCNet [43] and Non-Local Networks [44] capture long-range dependencies, while Transformers [45] have inspired detectors like DETR [46] to model global relationships effectively.
Nevertheless, these methods largely ignore explicit geometric priors such as object orientation, limiting their effectiveness in rotated detection. To address this gap, our AACA module generates angle-guided direction-aware attention maps, enabling collaborative enhancement of spatial and channel responses along target-aligned axes.
To enhance features for rotated objects, common strategies include: rotation-equivariant networks (e.g., ReDet [47]) that build in symmetry; deformable convolutions [48] that adapt sampling grids; explicit feature alignment modules like RoI Transformer [21] and Rotated RoIAlign [19]. In contrast, our AACA module adopts a complementary, lightweight attention paradigm that directly injects the predicted orientation as a spatial-channel prior, enhancing discriminability without altering the core feature alignment process.

2.3. Loss Functions for Rotated Bounding Boxes

Rotated bounding boxes are typically represented by either the eight-point polygon format or the five-parameter ( x , y , w , h , θ ) formulation, and their regression introduces challenges such as angle discontinuity and parameter coupling. Prior works attempt to address these issues via improved losses or encodings, such as IoU-Smooth L1 in SCRDet [30], CSL’s [49] circular label smoothing, DCL’s dense angle encoding [50], and Gaussian-based metrics including GWD and KLD [32,33]. PSC [51] further stabilizes angle regression through phase-based modeling.
However, most existing approaches optimize center, scale, and angle parameters independently with manually tuned loss weights, often resulting in unstable gradients and inconsistent convergence [17,30]. To overcome these limitations, our GC-MTL unifies all rotated box parameters into a geometrically consistent multi-task framework with angle-period smoothing and adaptive weighting to improve optimization stability.
Alongside advanced loss functions, alternative regression paradigms have emerged. Gaussian modeling methods (e.g., KLD [32], GWD [33]) represent boxes as distributions for differentiable loss. Another line of work uses progressive refinement [20] or repoints [18] to decouple optimization. While effective, they can add complexity. Our GC-MTL differentiates itself by offering a unified, parameter-efficient loss that enforces geometric consistency among all parameters within a standard regression head, ensuring stable training without architectural modification.

2.4. Comparative Analysis

To clearly position our work and highlight its systematic advancements, Table 1 contrasts the technical designs and inherent limitations of key representative methods. Unlike prior works that address these challenges in isolation, our RSFPN framework provides a unified solution: it simultaneously enhances rotation-sensitive feature representation (via DAFPN), geometric alignment and focus (via AACA), and regression stability (via GC-MTL). The synergistic effect of these modules, as evidenced by our ablation studies, is what enables RSFPN to effectively overcome the combined limitations outlined in the table.

3. Methods

3.1. Overall Framework

We propose a novel Rotation-Sensitive Feature Pyramid Network (RSFPN) to effectively address the challenges of rotated object detection in remote sensing imagery. As shown in Figure 1, RSFPN adopts a hierarchical and modular design that decomposes the detection pipeline into three core components: the feature extraction and enhancement module, the rotated proposal generation and alignment module, and the feature refinement and detection module. This structured design not only improves the interpretability of the network architecture but also strengthens the synergistic interactions among modules.
The input image first passes through a ResNet-50 backbone to extract fundamental feature representations. These features are then fed into the Dynamic Adaptive Feature Pyramid Network (DAFPN). Unlike conventional feature pyramids, DAFPN incorporates a semantics-guided upsampling path and a structure-enhanced downsampling path to achieve bidirectional interaction and deep fusion across multiple feature levels, thereby providing high-quality and multi-scale feature representations for subsequent processing. In the rotated proposal generation and alignment module, the Oriented Region Proposal Network (Oriented RPN) takes the enhanced features produced by DAFPN and efficiently generates a set of rotation-aligned proposals using a midpoint-offset parameterization. Each proposal is subsequently processed by the Rotated RoIAlign operation, which extracts geometrically aligned features from the corresponding feature maps and effectively mitigates feature misalignment issues commonly encountered in rotated object detection. In the feature refinement and detection module, the aligned features obtained via Rotated RoIAlign are forwarded to the Angle-Aware Cooperative Attention (AACA) module. This module treats the rotation angle as a strong geometric prior and generates a direction-aware mask through a lightweight mapping network. The mask guides the cooperative interaction of spatial and channel attention mechanisms, thereby enhancing discriminative target features while suppressing background noise in cluttered scenes. Finally, the AACA-enhanced features are sent to the detection head for classification and refined bounding-box regression.
The entire network is trained in an end-to-end manner and is jointly optimized using the proposed Geometrically Consistent Multi-Task Loss (GC-MTL). This loss function unifies the regression of all rotated bounding box parameters within a single geometrically consistent framework. By incorporating angle-period handling, shape constraints, and an adaptive weighting mechanism, GC-MTL ensures stable training and reliable convergence, directly addressing the core optimization challenges in rotated object detection.

3.2. Feature Extraction and Enhancement Module

3.2.1. Backbone Feature Extraction

We employ a ResNet-50 backbone pretrained on the ImageNet dataset to extract four hierarchical feature maps, denoted as { C 2 , C 3 , C 4 , C 5 } (see Figure 2), from the input image. These feature maps have spatial downsampling ratios of 4 × , 8 × , 16 × , and 32 × , respectively. They provide the foundation for multi-scale feature representation, with gradually decreasing spatial resolution and progressively increasing channel capacity. The shallow feature maps (e.g., C 2 and C 3 ) retain abundant spatial details, which are beneficial for localizing small objects, while the deeper feature maps (e.g., C 4 and C 5 ) contain richer semantic information, which is advantageous for object classification.

3.2.2. Dynamic Adaptive Feature Pyramid Network

To enhance the multi-scale representation capability required for complex objects in remote sensing imagery, we propose a Dynamic Adaptive Feature Pyramid Network (DAFPN). Unlike conventional FPNs that simply perform a top-down lateral aggregation, DAFPN adopts a bidirectional feature fusion architecture composed of a semantically guided upsampling path and a structurally enhanced downsampling path. This design enables a deep coupling between spatial details and semantic information across scales. The overall architecture is illustrated in Figure 2, where the left branch represents the upsampling path and the right branch denotes the downsampling enhancement pathway.
The upsampling path enables the effective propagation of high-level semantic information to lower layers (corresponding to the left branch in Figure 2). Specifically, the high-level feature map F high first undergoes global average pooling to extract global contextual information, producing a channel vector F global . Meanwhile, the low-level feature map F low is processed by a 3 × 3 convolution to capture local edge and texture details. The resulting feature is then modulated by F global through channel-wise multiplication, achieving semantically guided adaptive weighting and enhancing the sensitivity of low-level features to high-level semantics. Subsequently, the high-level feature is upsampled using the CARAFE module, which generates content-aware dynamic kernels to better recover structural boundaries and alleviate aliasing artifacts commonly observed in conventional upsampling. Finally, the enhanced high-level feature and the adaptively weighted low-level feature are fused via element-wise addition to produce direction-sensitive feature maps. This guided upsampling procedure is applied recursively across all pyramid levels ( C 5 C 2 ), forming a multi-scale set of direction-sensitive features { P 5 , P 4 , P 3 , P 2 } . The design substantially strengthens the ability of shallow layers to interpret high-level contextual information, thereby improving the representation capability for small objects and arbitrarily oriented targets in remote sensing imagery.
The downsampling path is designed to supplement the spatial structural information of high-level features (corresponding to the right side of Figure 2). This module begins with the low-level features produced by the upsampling path (e.g., P 2 ) and progressively propagates spatial details upward. First, a 3 × 3 convolution with a step size of 2 is used to down-sample the low-level features. While this operation reduces the spatial resolution, it also expands the receptive field, which is able to capture richer spatial context information. Subsequently, a 1 × 1 convolution is applied to the high-level semantic features for channel alignment, followed by element-wise addition with the downsampled low-level features to achieve scale-consistent semantic–structural fusion. Through this design, high-level semantic information is stably transmitted during the downsampling process while avoiding additional spatial misalignment. The final multi-scale feature set { N 2 , N 3 , N 4 , N 5 } integrates the semantic and structural information from both the upsampling and downsampling paths, providing stronger capabilities for object localization and shape characterization.
Overall, the Dynamic Adaptive Feature Pyramid Network (DAFPN) achieves bidirectional feature fusion through semantic guidance in the upsampling path and structural compensation in the downsampling path. The upsampling path enhances the semantic representations of low-level features, while the downsampling path strengthens the spatial structure of high-level features. Through the coordinated interaction of these two paths across multiple scales, the network gains improved rotation awareness and geometric modeling capability. This design effectively boosts the detection accuracy and robustness for small objects and dense rotated targets in complex remote sensing scenes.

3.3. Rotated Proposal Generation and Alignment

3.3.1. Oriented Region Proposal Network

Following the oriented region proposal design in Oriented R-CNN, our framework employs a lightweight fully convolutional network to generate high-quality rotated proposals (see Figure 3). The module takes the multi-scale feature maps enhanced by DAFPN, namely { N 2 , N 3 , N 4 , N 5 } , as input. Each feature level corresponds to a different receptive field, enabling the network to accommodate targets of varying scales. At each feature level, a 3 × 3 convolution layer is first applied to transform the input features and produce a 256-dimensional shared feature representation. This shared representation is then fed into two parallel 1 × 1 convolutional branches. The classification branch estimates the objectness score at each spatial location, while the regression branch predicts the geometric parameters of the rotated bounding boxes.
At each spatial location of the feature maps, three horizontal anchor boxes are predefined with aspect ratios of {1:2, 1:1, 2:1}, which correspond to slender objects (e.g., ships), approximately square objects (e.g., oil tanks), and wide rectangular objects (e.g., sports fields), respectively. The base scales of the anchors are dynamically assigned according to the feature pyramid hierarchy, specifically { N 2 : 32 2 , N 3 : 64 2 , N 4 : 128 2 , N 5 : 256 2 } pixels. The regression branch outputs a 6-dimensional offset vector ( δ x , δ y , δ w , δ h , δ α , δ β ) , which is decoded into the six-parameter representation of a rotated proposal box through the following transformation:
w = a w · e δ w , h = a h · e δ h x = δ x · a w + a x , y = δ y · a h + a y Δ α = δ α · w , Δ β = δ β · h
The variables ( a x , a y , a w , a h ) denote the parameters of the anchor box, while ( x , y , w , h , Δ α , Δ β ) represent the predicted parameters of the rotated proposal. This midpoint-offset parameterization inherits the regression formulation used for horizontal bounding boxes, while the additional terms Δ α and Δ β capture the rotational characteristics of the target (see Table 2 for a notation summary). In this way, the rotational prediction is performed under a stable and bounded constraint, leading to more reliable modeling of orientation variations.
During training, we adopt strict rotated IoU thresholds to assign positive and negative samples by matching each anchor box with the corresponding ground-truth box. Specifically, an anchor box is regarded as a positive sample only when the rotated IoU 0.7 ; it is treated as a negative sample when the rotated IoU 0.3 ; Anchor boxes whose rotated IoU falls between 0.3 and 0.7 are marked as ignore samples. This strategy ensures high-quality sample selection while preventing ambiguous samples from misleading the training process.

3.3.2. Rotated RoIAlign Feature Extraction

To extract orientation-consistent features from rotated proposals, our framework employs the Rotated RoIAlign operation (see Figure 4). This operation resolves feature misalignment through precise geometric transformations and serves as a crucial step to ensure that the subsequent angle-aware collaborative attention module receives high-quality feature inputs. Given the six-parameter representation of a rotated proposal ( x , y , w , h , Δ α , Δ β ) , the four vertex coordinates are first computed based on the midpoint offset formulation:
v 1 = x , y h / 2 + Δ α , 0 v 2 = x + w / 2 , y + 0 , Δ β v 3 = x , y + h / 2 + Δ α , 0 v 4 = x w / 2 , y + 0 , Δ β
Subsequently, the parallelogram proposal is rectified into an oriented rectangle to obtain the standard rotated representation ( x , y , w , h , θ ) , where θ [ π / 2 , π / 2 ] denotes the rotation angle defined as the inclination between the long side of the rectangle and the horizontal axis. In the local coordinate system of the rotated box, a regular P × P sampling grid (typically P = 7 ) is constructed. The local coordinates ( r i , s j ) of each grid point ( i , j ) are computed as follows:
r i = ( i + 0.5 ) × w / P w / 2 , s j = ( j + 0.5 ) × h / P h / 2
where i , j = 0 , 1 , , P 1 . The local coordinates are then mapped to the global feature space through a rotation matrix:
x i j y i j = cos θ sin θ sin θ cos θ r i · w s j · h + x y
Since the transformed sampling coordinates ( x i j , y i j ) are generally fractional, bilinear interpolation is employed to obtain continuous feature representations. For each sampling location, the four nearest integer-coordinate points on the feature map, Q 11 , Q 12 , Q 21 , Q 22 , are identified, and the interpolation weights are computed based on the relative offsets Δ x and Δ y :
f x , y = 1 Δ x 1 Δ y · f Q 11 + 1 Δ x Δ y · f Q 12 + Δ x 1 Δ y · f Q 21 + Δ x Δ y · f Q 22
This operation ensures that smooth and continuous feature representations are obtained even when the sampling locations fall between pixels, effectively avoiding the feature distortion commonly caused by orientation inconsistencies in conventional methods. The resulting P × P × C feature tensor provides a geometrically aligned and direction-consistent input for the subsequent Angle-Aware Synergistic Attention module. This step plays a crucial role in dense small-object detection on DOTA-v1.0 and in recognizing extremely high–aspect-ratio targets on HRSC2016.

3.4. Feature Refinement and Detection Module

3.4.1. Angle-Aware Collaborative Attention

To address the feature degradation of rotated objects under complex backgrounds and significant orientation variations in remote sensing imagery, we propose an Angle-Aware Collaborative Attention (AACA) module. The module takes as input the geometrically aligned features F in produced by Rotated RoIAlign, together with their corresponding rotation angle θ , and achieves direction-aware feature enhancement through a collaborative modeling strategy across the angle, spatial, and channel dimensions. As illustrated in Figure 5, AACA consists of three parallel branches—angle attention, spatial attention, and channel attention—whose outputs are lightly fused to produce the final enhanced RoI features.
To generate the direction-aware mask A orient , the rotation angle θ is first encoded and concatenated with the normalized spatial coordinates of each position in the P × P grid. This joint representation is then processed by a two-layer MLP to produce the final mask.
A orient = Sigmoid ( W 2 · ReLU ( W 1 θ ^ + b 1 ) + b 2 ) R 1 × P × P
where W 1 , b 1 , W 2 , b 2 are learnable parameters, and θ ^ represents a feature vector that combines Angle and position information. The intermediate layer employs the ReLU activation, while the output layer uses the Sigmoid function. This mask provides a spatial prior for the main orientation of the object, complementing the traditional attention mechanisms that lack explicit directional constraints.
To capture local structural cues within the RoI, the input feature F in is first aggregated along the channel dimension using average pooling and max pooling, producing two spatial response maps. These maps are concatenated and passed through a 7 × 7 convolution, followed by a Sigmoid function, to generate the spatial attention A spatial .
A spatial = Sigmoid ( f 7 × 7 [ AvgPool ( F in ) MaxPool ( F in ) ] ) R 1 × P × P
Here, ⊕ indicates channel concatenation. To model global semantic dependencies, channel statistics are extracted via global average pooling and max pooling, which are then input into a shared two-layer MLP to obtain channel-wise weights A channel .
A channel = Sigmoid ( MLP GAP ( F in ) + GMP ( F in ) ) R 1 × 1 × C
The orientation mask and spatial attention are concatenated along the channel dimension and fused using a 3 × 3 convolution to produce the final direction-aware spatial modulation mask M s . This mask implements the core design of “orientation prior-guided spatial attention,” enabling spatial focus that aligns with the primary axis of the object.
M s = Sigmoid ( f 3 × 3 A orient A spatial ] R 1 × P × P
The input features are modulated separately by the channel attention and direction-aware spatial mask, then concatenated and fused via a 1 × 1 convolution to produce the final output.
F out = f 1 × 1 ( F in A channel F in M s ) R P × P × C
Here, ⊙ denotes element-wise multiplication (utilizing the broadcasting mechanism). The AACA structure is lightweight and can be seamlessly integrated into two-stage rotated detection frameworks, significantly enhancing the discriminative capability of RoI features in complex remote sensing scenarios.

3.4.2. Detection Head

Receiving the enhanced RoI features output by the Angle-Aware Collaborative Attention (AACA) module, with dimensions P × P × C , the framework employs a lightweight detection head to perform the final classification and localization tasks. The detection head consists of two main components: a shared base feature encoder and two separate branches dedicated to classification and regression, respectively.
First, the RoI features of dimensions P × P × C are flattened into a feature vector and fed into two fully connected layers (e.g., FC1 and FC2, each with a dimension of 1024). These fully connected layers serve as a shared backbone, responsible for further encoding the high-dimensional features into more discriminative representations, thereby providing an informative basis for the subsequent task-specific branches.
The classification branch receives the aforementioned encoded features and outputs a probability distribution over ( K + 1 ) classes for each candidate region through a fully connected layer, where K denotes the number of object categories (e.g., 15 in DOTA-v1.0), and the additional class represents the background. The regression branch shares the same input features as the classification branch and outputs, via another fully connected layer, the bounding box refinement parameters corresponding to the K object categories for each candidate region. This framework follows the midpoint offset representation used in Oriented R-CNN; thus, for each category, the regression branch predicts a 6-dimensional vector ( δ x , δ y , δ w , δ h , δ α , δ β ) . These parameters are employed to refine the initial proposal boxes generated by the rotated region proposal network, yielding the final precise rotated detection boxes.

3.4.3. Loss Function

In rotated object detection, a bounding box is represented as ( x , y , w , h , θ ) , where ( x , y ) denotes the center coordinates, w and h represent the width and height, and θ is the rotation angle, ranging within [ π / 2 , π / 2 ) . Conventional rotated object detection methods typically employ multiple independent loss functions to optimize these parameters, which may lead to training instability and difficulties in hyperparameter tuning. To address these issues, we propose a Geometric Consistency Multi-Task Loss (GC-MTL), which unifies the regression of all rotated bounding box parameters within a single loss function, incorporates angle periodicity handling and shape constraints, and employs an adaptive weighting mechanism to balance the contributions of different loss components. The overall loss function consists of two parts: a classification loss and a regression loss.
L t o t a l = λ c l s L c l s + λ r e g L r e g
Here, L cls denotes the classification loss, for which Focal Loss [38] is employed to address class imbalance; L reg denotes the regression loss, which jointly optimizes all parameters of the rotated bounding box; λ reg represents the weighting coefficient for the regression loss, with a default value of 1.0.
L c l s = 1 N i = 1 N c = 1 C α c ( 1 p i , c ) γ y i , c l o g ( p i , c )
Here, N denotes the total number of samples in a batch, C represents the total number of classes, p i , c is the predicted probability of the i-th sample belonging to class c, and y i , c is the ground-truth label of the i-th sample (one-hot encoded). α c is the class weight factor for class c, used to address class imbalance, with a default value of 0.25. γ is the focusing parameter, employed to reduce the weight of easily classified samples, with a default value of 2.0.
The primary innovation proposed in this work lies in unifying the regression of all geometric parameters of rotated bounding boxes within a single loss function:
L r e g = 1 N p o s i = 1 N p o s [ L b a s e ( i ) + λ θ L θ ( i ) + λ s h a p e L s h a p e ( i ) ]
Here, N pos denotes the number of positive samples; L base ( i ) is the basic regression loss; L θ ( i ) is the angle regression loss; and L shape ( i ) represents the shape constraint loss. λ base , λ θ , and λ shape are the corresponding weighting coefficients.
The basic regression loss employs the Smooth L1 loss to optimize the center coordinates and the scale parameters of the bounding box:
L b a s e ( i ) = j { x , y , w , h } s m o o t h L 1 ( Δ i j )
Smooth L 1 ( x ) = 0.5 x 2 / β if | x | < β | x | 0.5 β otherwise
The term Δ i j denotes the difference between the predicted value and the ground truth, with the smoothing coefficient set to β = 1.0 by default. This loss applies a quadratic penalty to small errors and a linear penalty to large errors, effectively balancing convergence speed and training stability.
To address the half-period property of the rotation angle θ [ π / 2 , π / 2 ) , we design an angle regression loss that maps the angular discrepancy into a normalized range through the arctan function, while automatically handling the periodicity issue:
L θ ( i ) = 2 π · arctan ( p θ i g θ i ε )
where p θ i and g θ i denote the predicted angle and the ground-truth angle, respectively, and ϵ is a smoothing coefficient (set to 0.1 by default). The proposed loss function offers the following advantages: (1) Since the angular range follows a half-period definition, the maximum error does not exceed π / 2 . By applying the arctan function, the angular discrepancy is mapped into the interval [ 0 , 1 ) , which effectively avoids gradient instability caused by periodic discontinuities. (2) When the angular error approaches zero, the gradient becomes larger, facilitating faster convergence. Conversely, when the angular error is large, the gradient saturates, preventing gradient explosion.
The shape constraint loss is designed to preserve the aspect-ratio consistency of the bounding box and to prevent the generation of geometrically implausible shapes. This loss measures the relative discrepancy between the predicted aspect ratio and the ground-truth aspect ratio:
L s h a p e ( i ) = log ( p w i / p h i g w i / g h i ) · 2 p w i p h i + g w i g h i
Here, p w i and p h i denote the width and height of the predicted bounding box for the i-th sample, respectively, g w i and g h i denote the width and height of the corresponding ground-truth bounding box. The ratios p w i / p h i and g w i / g h i represent the aspect ratios of the predicted and ground-truth boxes, respectively. This loss computes the discrepancy in the logarithmic space of the aspect ratio and assigns smaller weights to objects with large aspect ratios, thereby improving its adaptability across different object scales.
To automatically balance the base regression loss, angle loss, and shape constraint loss, we introduce an adaptive weight adjustment mechanism based on gradient statistics. We monitor the gradient statistics of each loss component over the most recent K iterations. We set σ base , σ θ , and σ shape to represent the gradient standard deviation of the three types of loss in the latest K = 100 iterations, respectively, and the weight update formula is as follows:
λ θ = σ base σ θ , λ shape = σ base σ shape
This mechanism ensures that the gradient contribution of each loss term is roughly balanced, and there is no need to manually adjust the parameters, which improves the training stability.
Based on theoretical analysis, the proposed loss function is expected to enhance training stability, detection accuracy, and generalization capability for rotated object detection. In the subsequent experimental section, we conduct comprehensive ablation studies to verify the contribution of each component and compare our method with mainstream loss functions, thereby demonstrating the effectiveness of the proposed approach.

3.4.4. Training Stability and Implementation Considerations

The design of GC-MTL prioritizes training stability and ease of use. The adaptive gradient-balancing mechanism (Equation (18)) automatically equilibrates the contributions of the base, angle, and shape losses, preventing gradient conflict and ensuring stable convergence. The arctan-based angle loss (Equation (16)) provides smooth, bounded gradients that inherently avoid discontinuity issues at angular boundaries. The hyperparameter K (sliding-window size) is set to 100 as a robust default; performance is not sensitive to moderate variations around this value, reducing the need for manual tuning. Collectively, these design choices yield a stable, reproducible training process that requires minimal hyperparameter adjustment.

4. Experiments and Results

4.1. Datasets and Evaluation Metrics

4.1.1. Experimental Datasets

To comprehensively evaluate the performance of the proposed Rotation-Sensitive Feature Pyramid Network (RSFPN) under various challenging scenarios, we conducted experiments on two widely recognized benchmark datasets in remote sensing rotated object detection: DOTA-v1.0 and HRSC2016. These datasets focus on multi-scale object detection and the recognition of objects with extreme aspect ratios, enabling a systematic assessment of the effectiveness and robustness of the proposed method.
(1)
DOTA-v1.0
DOTA-v1.0 is a large-scale and high-quality publicly available benchmark widely used in aerial image object detection. DOTA-v1.0 is a large-scale benchmark for oriented object detection, containing 15 categories across 2806 images with significant scale and orientation variations. The official split consists of 1411 training images, 458 validation images, and 937 testing images.
In this study, we used the officially released training and validation sets, which contain 208,261 valid annotated instances. The detailed category- and scale-wise distributions are summarized in Table 3.
(2)
HRSC2016
HRSC2016 is a ship detection dataset characterized by instances with extreme aspect ratios (often >4:1), presenting a challenge for oriented bounding box regression. The dataset contains 1061 images with spatial resolutions ranging from 0.4 m to 2 m. The official split consists of 626 training images, 244 validation images, and 191 testing images. All ship instances in this dataset are precisely annotated using rotated bounding boxes. Most objects have aspect ratios greater than 4:1, with some exceeding 8:1, making HRSC2016 an effective benchmark for assessing the regression accuracy of detectors on slender targets.
To ensure fairness and reproducibility in the experiments, as well as to enhance the generalization capability of the proposed model, we applied a unified and rigorous data preprocessing and augmentation pipeline to all training data. For DOTA-v1.0, we cropped each large-scale aerial image into 1024 × 1024 sub-images with an overlap of 200 pixels to alleviate boundary truncation of objects. Following the dataset’s common practice, objects truncated by the cropping boundaries were retained in the sub-images with their original annotations. During training and validation, these truncated instances were fully utilized. During testing, potential duplicate detections of the same object from overlapping sub-images were merged via non-maximum suppression to ensure each object was evaluated only once. For the HRSC2016 dataset, all images are resized while preserving their original aspect ratios. The shorter side of each image is scaled to 800 pixels, with the longer side capped at 1333 pixels. Data augmentation included random flipping, multi-scale resizing, and random rotation perturbations to improve robustness against scale variation and orientation diversity.

4.1.2. Evaluation Metrics

We adopt the mean Average Precision (mAP) with an Intersection over Union (IoU) threshold of 0.5 as the primary metric for detection accuracy, following the standard protocol for DOTA-v1.0 [10]. To comprehensively assess practicality, we also report the number of parameters (Params), floating-point operations (FLOPs), and inference speed in frames per second (FPS) on two NVIDIA Tesla T4 GPUs. This equipment was purchased from Dell Server supplier in China.

4.1.3. Implementation Details

All experiments were implemented with PyTorch 1.10.0 and MMDetection 2.25.1. The network was trained in an end-to-end manner using the stochastic gradient descent (SGD) optimizer, with a momentum of 0.9 and a weight decay of 0.0001. The initial learning rate was set to 0.02 and decayed according to a Cosine Annealing schedule during training. The batch size was fixed at 16. The backbone for feature extraction was a ResNet-50 pretrained on ImageNet. Considering the characteristics and scale of each dataset, the model was trained for 40 epochs on DOTA-v1.0, while 100 epochs were used for HRSC2016.

4.2. Experimental Results

To objectively evaluate the effectiveness of the proposed RSFPN method, we compare it with a series of representative state-of-the-art approaches. These methods encompass the main technical directions in the field of rotated object detection, ensuring a comprehensive comparison. All comparative experiments are conducted under identical experimental environments, dataset splits, and evaluation protocols to guarantee fairness and reproducibility. To systematically validate the overall advantages of the Rotated Sensitive Feature Pyramid Network (RSFPN), this section presents a detailed comparison on two benchmark datasets, DOTA-v1.0 and HRSC2016, evaluating multiple aspects including detection accuracy, model complexity, and inference speed.

4.2.1. Comparison of Detection Accuracy

On the DOTA-v1.0 test set, RSFPN is compared with 18 state-of-the-art methods. Table 4 provides a detailed breakdown of the AP@0.5 for each category as well as the overall mAP.
RSFPN achieves the highest overall mAP of 77.42%, outperforming all compared methods. It surpasses the strong baseline Oriented R-CNN by 1.55% and the closely ranked Rotated Faster R-CNN by 0.27%. Notably, RSFPN ranks first in 11 out of 15 categories, demonstrating excellent generalization. Significant improvements are observed on challenging dense small objects: for Small Vehicles (SV) and Large Vehicles (LV), RSFPN attains 77.15% and 82.45% AP, corresponding to gains of +1.91% and +2.06% over Oriented R-CNN, respectively. This validates the effectiveness of the DAFPN module in enhancing multi-scale feature representation. Furthermore, for targets with complex geometries such as Bridges (BD) and Storage Tanks (ST), RSFPN also achieves top performance (74.25% and 88.03% AP), benefiting from the orientation-aware capability of the AACA module.
The experimental results on the HRSC2016 dataset further validate the comprehensive advantages of RSFPN in handling targets with extreme aspect ratios. As shown in Table 5, our method achieves an AP50 of 91.85%, demonstrating a significant improvement over other compared methods. Compared with recent methods, the performance gains of RSFPN indicate that our approach offers competitive advantages in feature extraction, geometric modeling, and loss optimization.

4.2.2. Model Complexity and Inference Speed

In terms of model complexity, RSFPN demonstrates a well-balanced design. As shown in Table 6, RSFPN has 42.3M parameters and requires 218.9 GFLOPs, placing it at a moderate level compared with other state-of-the-art methods. Compared to Oriented R-CNN, RSFPN introduces the DAFPN and AACA modules, incurring an additional 0.8M parameters and 3.2 GFLOPs, which yields a 1.55% improvement in mAP, reflecting a reasonable trade-off between performance and computational complexity.
Regarding inference speed, RSFPN achieves 14.5 FPS, which is comparable to S2A-Net and slightly lower than Oriented R-CNN at 15.2 FPS. This minor reduction in speed mainly results from the bidirectional feature fusion in DAFPN and the attention computation overhead in AACA. However, compared with similarly accurate methods such as Rotated Faster R-CNN (13.5 FPS) and Oriented RepPoints (13.9 FPS), RSFPN attains faster inference while maintaining higher precision, demonstrating a superior balance between accuracy and efficiency.
It is noteworthy that RSFPN achieves excellent performance based on the standard ResNet-50 backbone without relying on specialized network designs, enhancing the method’s practicality and scalability. Compared to other approaches that employ specialized network architectures, such as ReDet (using ReResNet-50), RSFPN delivers a significant accuracy improvement while maintaining competitive inference speed.

4.2.3. Overall Performance Analysis

Synthesizing the above experimental results, RSFPN demonstrates superior overall performance across detection accuracy, model complexity, and inference speed. From the perspective of the accuracy-efficiency trade-off, RSFPN achieves a high mAP of 77.42% while maintaining an inference speed of 14.5 FPS, indicating strong practical applicability. Compared with the fastest inference method, Oriented R-CNN, RSFPN trades 4.6% of speed for a 1.55% improvement in accuracy. In comparison with the similarly accurate Rotated Faster R-CNN, RSFPN achieves higher precision while also increasing inference speed by 7.4%.
In terms of per-class performance, RSFPN exhibits notable advantages on challenging dense small objects (e.g., small vehicles) and extreme aspect ratio targets (e.g., bridges and ships). This improvement is primarily attributed to the rich multi-scale features provided by DAFPN and the precise orientation-aware capability of the AACA module. Particularly in ship detection, RSFPN attains an AP of 90.12%, significantly outperforming other methods, which validates the effectiveness of the proposed approach in handling rotation-sensitive targets.
We note that RSFPN shows slightly lower performance on a few regularly shaped categories (e.g., airplanes, tennis courts) compared to methods specifically optimized for such targets. This reflects the inherent trade-off in our design, which prioritizes challenging rotation-sensitive detection tasks over specialized architectures for regular shapes.

4.2.4. Stability and Robustness Analysis

We acknowledge that deep learning training involves stochastic processes. To assess the stability of RSFPN, we evaluated the model’s performance stability. While the overall mAP improvement over the nearest competitor (Rotated Faster R-CNN) is 0.27%, our method demonstrates consistent superiority in handling specific difficult categories.
Specifically, for categories with extreme aspect ratios (e.g., Bridges) and dense small objects (e.g., Small Vehicles), RSFPN consistently outperforms competing methods by a margin of 1.0–3.0% (as shown in Table 4). This suggests that the gains are driven by the structural advantages of the DAFPN and AACA modules rather than random noise. However, we also acknowledge that for simpler categories (e.g., Planes), the performance saturation leads to marginal variance. Future work will involve more extensive cross-validation to further quantify these statistical bounds.

4.3. Ablation Studies

4.3.1. Ablation Analysis of Core Modules

To validate the effectiveness of each core module in RSFPN, systematic ablation experiments are conducted on the DOTA-v1.0 dataset. All experiments adopt the same training settings and evaluation protocols, with ResNet-50 as the backbone and input images resized to 1024 × 1024 pixels.
The results in Table 7 demonstrate the individual and synergistic contributions of each module. Adding DAFPN brings a +1.05% mAP gain, with notable improvements on small vehicles (SV, +1.61%) and bridges (BD, +1.62%), confirming its strength in multi-scale feature fusion.
Building on the features from DAFPN, the Angle-Aware Collaborative Attention (AACA) module contributes a further 0.64% mAP gain. Notably, the bridge (BD) category shows the largest single-stage improvement here (+1.00%). The cumulative gain from DAFPN and AACA on bridges reaches 2.62%, demonstrating the importance of the orientation-aware mechanism for detecting directional, linear targets.
The Geometric Consistency Multi-Task Loss (GC-MTL) is then incorporated, yielding the final RSFPN model with 77.42% mAP. From the “+AACA” stage to the full model, a slight mAP adjustment (−0.14%) occurs, attributable to GC-MTL’s strict shape and angle constraints correcting some inaccurate localizations. Nevertheless, GC-MTL provides substantial gains for large vehicles (LV), small vehicles (SV), and ships (SH), underscoring its advantage in optimizing regression for targets with extreme aspect ratios and high rotation sensitivity, ensuring higher geometric accuracy.
The ablation studies systematically confirm that each core module in RSFPN is indispensable: DAFPN serves as the foundation for performance improvement, AACA enables precise enhancement of direction-sensitive targets, and GC-MTL optimizes the training process to ensure high-precision and robust model outputs. The synergistic interaction of these three components forms the cornerstone of RSFPN’s superior performance. The cumulative 3.02% gain observed in the bridge (BR) category provides the most compelling evidence of this collaborative effect.

4.3.2. Verification of Module Combination Effectiveness

Analysis of the results in Table 8 reveals a significant synergistic effect pattern: the combination of DAFPN and AACA led to a 1.69% increase in mAP, which exceeds the sum of their individual contributions (1.05% + 0.48% = 1.53%). This indicates a pronounced synergistic enhancement between the two modules. The high-quality multi-scale features provided by DAFPN establish a solid foundation for the attention mechanism in AACA, while the direction-aware capability of AACA further reinforces the effective utilization of DAFPN features.
Considering all ablation study results, several conclusions can be drawn: each core module is not only independently effective but also exhibits significant synergistic effects; the internal design of each module has been thoroughly validated, reflecting the central principles of feature fusion and geometric priors; and the final model achieves a favorable balance between accuracy and efficiency. These findings provide strong experimental support for the overall design of RSFPN.

4.4. Visualization Analysis

4.4.1. Detection Result Visualization

Figure 6 illustrates the comprehensive detection performance of the RSFPN model across five representative scenarios. In dense small-object scenes, the model achieves high recall and low miss rates for tiny targets such as vehicles in parking lots (size < 32 × 32 pixels), demonstrating the multi-scale feature fusion capability of the Dynamic Adaptive Feature Pyramid Network (DAFPN). For medium and large objects, the detection boxes accurately align with the contours of airplanes, vehicles, and other targets, reflecting the model’s stable regression of conventional object geometric features. In extreme aspect-ratio object detection tasks (based on the HRSC2016 ship dataset), the model generates highly fitting rotated bounding boxes for ships with varying aspect ratios (2:1 to 15:1), validating the effectiveness of the Rotated RoIAlign operation and the angle-aware mechanism. In complex background scenarios, such as ports and urban streets, the model maintains a low false positive rate, indicating that the Angle-Aware Collaborative Attention (AACA) module effectively suppresses background interference. Finally, multi-scale detection results within the same scene show that the model can accurately detect objects of different sizes simultaneously, without exhibiting scale bias, highlighting the algorithm’s practicality and robustness in real-world settings.
Overall, the visualization results qualitatively confirm the robustness of the RSFPN model in handling key challenges such as scale variation, rotational diversity, and background interference, providing strong support for its practical application in complex remote sensing imagery.

4.4.2. Analysis of Heatmap Comparison Results

To further investigate RSFPN’s feature focusing capability and its orientation sensitivity for rotated objects, class activation maps (CAMs) were generated using the Grad-CAM [56] technique (see Figure 7). A comparative analysis was conducted between the baseline model (Oriented R-CNN) and RSFPN regarding high-level feature responses.
In the heatmap comparison across typical remote sensing scenarios, the baseline model generally exhibited dispersed activation regions, susceptibility to background texture interference, and imprecise focus on object bodies. In contrast, RSFPN consistently demonstrated structured and stable feature responses across different scenes. Its heatmaps in complex scenarios, such as dense ship clusters, multi-directional airplanes at airports, and urban traffic scenes, showed highly concentrated activations aligned with the principal orientation of target structures while effectively suppressing irrelevant activations from background regions such as water surfaces, runways, roads, and vegetation. Overall, the feature patterns generated by RSFPN align more closely with the geometric structure and principal orientation of objects, exhibiting stronger target focus, rotation-consistent geometry, and background suppression. These observations indicate that through the collaborative optimization of its modules, RSFPN maintains stable and discriminative feature representations across diverse remote sensing scenes, significantly enhancing the interpretability and reliability of rotated object detection.

4.4.3. Visualization Comparison Case

In order to clearly show the advantages and improvements in our method, we select typical scenes and visually compare RSFPN with two representative baseline methods: Oriented R-CNN (an efficient two-stage rotation detection framework) and ReDet (an advanced method based on rotation equivariant). Figure 8 presents two categories of key cases.
The comparison shows that in scenes with dense targets and arbitrary directions, RSFPN has high detection rate and accurate frame, while Oriented R-CNN has missed detection, indicating that RSFPN has obvious advantages in dealing with complex and irregular rotating targets, and its dynamic feature enhancement and attention mechanism can better adapt to the diversity of real remote sensing scenes. In the scene with more orientation to the aircraft, RSFPN has sporadic missed detection, while ReDet has completed all detection. It shows that the equivariant backbone network of ReDet provides it with the inherent advantage of dealing with rotationally symmetric targets. However, the design of RSFPN focuses more on learning and adaptation to cope with irregular rotation and complex background.

5. Conclusions

In this work, we propose RSFPN, a rotation-sensitive feature enhancement framework for rotated object detection in remote sensing images. By incorporating a Dynamic Adaptive Feature Pyramid Network (DAFPN), Angle-Aware Collaborative Attention (AACA), and Geometric Consistency Multi-Task Loss (GC-MTL), RSFPN achieves significant improvements in multi-scale feature representation, orientation-sensitive modeling, and rotated bounding box regression stability.
Experiments on the DOTA-v1.0 and HRSC2016 datasets demonstrate that RSFPN substantially outperforms baseline methods in overall accuracy and across multiple orientation-sensitive object categories, particularly in scenarios involving densely distributed small objects, extreme aspect ratios, and multi-directional target distributions. Grad-CAM visualization further shows that RSFPN generates more focused, cleaner, and orientation-consistent feature responses, effectively suppressing complex background interference and enhancing model interpretability.
While the evaluation is conducted on aerial benchmarks, the design principles of RSFPN are general and potentially transferable to other rotated detection scenarios. The DAFPN module, with its dynamic bidirectional fusion, is well suited to handle severe scale variations in UAV imagery. The orientation prior in AACA can enhance feature discriminability in SAR image interpretation, where targets exhibit diverse azimuth angles. Moreover, the geometry-aware formulation of GC-MTL is not domain-specific and could benefit rotated object detection in broader contexts, such as scene text detection or industrial inspection. These characteristics suggest that RSFPN provides a versatile framework that may generalize beyond optical remote sensing data.
Overall, RSFPN provides an efficient, robust, and practically applicable solution for rotated object detection in complex remote sensing scenarios. Future work will explore lightweight design, enhanced rotation modeling capabilities, and extensions to challenging tasks such as optical–SAR multimodal fusion and large-scale real-time monitoring.

Author Contributions

Conceptualization, J.X. and H.H.; methodology, J.X. and A.M.; software, J.X. and C.Z.; validation, J.X., S.K. and A.M.; formal analysis, A.M.; investigation, S.K.; resources, A.M.; data curation, J.X.; writing—original draft preparation, J.X.; writing—review and editing, J.X., S.K. and A.M.; visualization, A.M.; supervision, H.H.; project administration, H.H.; funding acquisition, H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under Grant No. 61672210, the Major Science and Technology Program of Henan Province under Grant No. 221100210500, the Central Government Guiding Local Science and Technology Development Fund Program of Henan Province under Grant No. Z20221343032, and the Provincial Science and Technology Research and Development Plan Joint Fund Program of Henan Province under Grant No. 245200810016.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

We clarify that our research findings are based on the analysis of publicly available datasets: DOTA-v1.0: https://captain-whu.github.io/DOTA/dataset.html (accessed on 15 September 2025). HRSC2016: https://ieee-dataport.org/documents/hrsc2016-0 (accessed on 16 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, C.; Fang, H.; Guo, S.; Tang, P.; Xia, Z.; Zhang, X.; Du, P. Detecting urban functional zones changes via multi-source temporal fusion of street view and remote sensing imagery. Int. J. Appl. Earth Obs. Geoinf. 2025, 144, 104933. [Google Scholar] [CrossRef]
  2. Xue, B.; Kong, Y.; Zarco-Tejada, P.J.; Tian, L.; Poblete, T.; Wang, X.; Zheng, H.; Jiang, C.; Yao, X.; Zhu, Y.; et al. Mitigating the phenological influence on spectroscopic quantification of rice blast disease severity with extended PROSAIL simulations. Remote Sens. Environ. 2026, 332, 115063. [Google Scholar] [CrossRef]
  3. Misra, A.; White, K.; Nsutezo, S.F.; Straka, W., III; Lavista, J. Mapping global floods with 10 years of satellite radar data. Nat. Commun. 2025, 16, 5762. [Google Scholar] [CrossRef] [PubMed]
  4. Jiang, D.; Marino, A.; Ionescu, M.; Gvilava, M.; Savaneli, Z.; Loureiro, C.; Spyrakos, E.; Tyler, A.; Stanica, A. Combining optical and SAR satellite data to monitor coastline changes in the Black Sea. ISPRS J. Photogramm. Remote Sens. 2025, 226, 102–115. [Google Scholar] [CrossRef]
  5. Toumi, A.; Cexus, J.C.; Khenchaf, A.; Abid, M. A Combined CNN-LSTM Network for Ship Classification on SAR Images. Sensors 2024, 24, 7954. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, X.; Zhang, T.; Wang, G.; Zhu, P.; Tang, X.; Jia, X.; Jiao, L. Remote sensing object detection meets deep learning: A metareview of challenges and advances. IEEE Geosci. Remote Sens. Mag. 2023, 11, 8–44. [Google Scholar] [CrossRef]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  8. Ali, M.L.; Zhang, Z. The YOLO framework: A comprehensive review of evolution, applications, and benchmarks in object detection. Computers 2024, 13, 336. [Google Scholar] [CrossRef]
  9. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  10. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3974–3983. [Google Scholar]
  11. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  12. Sun, X.; Wang, P.; Yan, Z.; Xu, F.; Wang, R.; Diao, W.; Chen, J.; Li, J.; Feng, Y.; Xu, T.; et al. FAIR1M: A benchmark dataset for fine-grained object recognition in high-resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 116–130. [Google Scholar] [CrossRef]
  13. Liu, Z.; Yuan, L.; Weng, L.; Yang, Y. A high resolution optical satellite image dataset for ship recognition and some new baselines. In Proceedings of the International Conference on Pattern Recognition Applications and Methods, Porto, Portugal, 24–26 February 2017; Volume 2, pp. 324–331. [Google Scholar]
  14. Wang, K.; Wang, Z.; Li, Z.; Su, A.; Teng, X.; Pan, E.; Liu, M.; Yu, Q. Oriented object detection in optical remote sensing images using deep learning: A survey. Artif. Intell. Rev. 2025, 58, 350. [Google Scholar] [CrossRef]
  15. Wen, L.; Cheng, Y.; Fang, Y.; Li, X. A comprehensive survey of oriented object detection in remote sensing images. Expert Syst. Appl. 2023, 224, 119960. [Google Scholar] [CrossRef]
  16. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-oriented scene text detection via rotation proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef]
  17. Yang, X.; Yan, J.; Feng, Z.; He, T. R3det: Refined single-stage detector with feature refinement for rotating object. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 3163–3171. [Google Scholar]
  18. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1829–1838. [Google Scholar]
  19. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3520–3529. [Google Scholar]
  20. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align deep features for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602511. [Google Scholar] [CrossRef]
  21. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–10 June 2019; pp. 2849–2858. [Google Scholar]
  22. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  23. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  24. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  25. Liu, K.; Zou, J.; Zhang, W.; Li, Q.; Wang, Q. MSDP-Net: Multi-scale Distribution Perception Network for Rotating Object Detection in Remote Sensing. Pattern Recognit. 2025, 172, 112740. [Google Scholar] [CrossRef]
  26. Ghiasi, G.; Lin, T.Y.; Le, Q.V. Nas-fpn: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7036–7045. [Google Scholar]
  27. Jiang, Y.; Zhu, X.; Wang, X.; Yang, S.; Li, W.; Wang, H.; Fu, P.; Luo, Z. R2CNN: Rotational region CNN for orientation robust scene text detection. arXiv 2017, arXiv:1706.09579. [Google Scholar] [CrossRef]
  28. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.S.; Bai, X. Gliding vertex on the horizontal bounding box for multi-oriented object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 1452–1459. [Google Scholar] [CrossRef]
  29. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  30. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. Scrdet: Towards more robust detection for small, cluttered and rotated objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8232–8241. [Google Scholar]
  31. Chen, Z.; Chen, K.; Lin, W.; See, J.; Yu, H.; Ke, Y.; Yang, C. Piou loss: Towards accurate oriented object detection in complex environments. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 195–211. [Google Scholar]
  32. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning high-precision bounding box for rotated object detection via kullback-leibler divergence. Adv. Neural Inf. Process. Syst. 2021, 34, 18381–18394. [Google Scholar]
  33. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with gaussian wasserstein distance loss. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 11830–11841. [Google Scholar]
  34. Liu, S.; Huang, D.; Wang, Y. Learning spatial fusion for single-shot object detection. arXiv 2019, arXiv:1911.09516. [Google Scholar] [CrossRef]
  35. Liu, L.; Pan, Z.; Lei, B. Learning a rotation invariant detector with rotatable bounding box. arXiv 2017, arXiv:1711.09405. [Google Scholar] [CrossRef]
  36. Yang, X.; Sun, H.; Fu, K.; Yang, J.; Sun, X.; Yan, M.; Guo, Z. Automatic ship detection in remote sensing images from google earth of complex scenes based on multiscale rotation dense feature pyramid networks. Remote Sens. 2018, 10, 132. [Google Scholar] [CrossRef]
  37. Wang, J.; Chen, K.; Xu, R.; Liu, Z.; Loy, C.C.; Lin, D. Carafe: Content-aware reassembly of features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3007–3016. [Google Scholar]
  38. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  39. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  40. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  41. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  43. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  44. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  45. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the NIPS’17: 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  46. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  47. Han, J.; Ding, J.; Xue, N.; Xia, G.S. Redet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 2786–2795. [Google Scholar]
  48. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  49. Yang, X.; Yan, J. Arbitrary-oriented object detection with circular smooth label. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 677–694. [Google Scholar]
  50. Yang, X.; Hou, L.; Zhou, Y.; Wang, W.; Yan, J. Dense label encoding for boundary discontinuity free rotation detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15819–15829. [Google Scholar]
  51. Yu, Y.; Da, F. Phase-shifting coder: Predicting accurate orientation in oriented object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 13354–13363. [Google Scholar]
  52. Liao, M.; Zhu, Z.; Shi, B.; Xia, G.s.; Bai, X. Rotation-sensitive regression for oriented scene text detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5909–5918. [Google Scholar]
  53. Yi, J.; Wu, P.; Liu, B.; Huang, Q.; Qu, H.; Metaxas, D. Oriented object detection in aerial images with box boundary-aware vectors. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 2150–2159. [Google Scholar]
  54. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  55. Yang, S.; Pei, Z.; Zhou, F.; Wang, G. Rotated faster R-CNN for oriented object detection in aerial images. In Proceedings of the 2020 3rd International Conference on Robot Systems and Applications, Chengdu, China, 14–16 June 2020; pp. 35–39. [Google Scholar]
  56. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. The overall architecture of the proposed Rotation-Sensitive Feature Pyramid Network (RSFPN).
Figure 1. The overall architecture of the proposed Rotation-Sensitive Feature Pyramid Network (RSFPN).
Sensors 26 00381 g001
Figure 2. Overall architecture of the proposed Dual-Path Adaptive Feature Pyramid Network (DAFPN): (a) Backbone; (b) Semantic-guided upsample path; (c) Structure-enhanced downsample path.
Figure 2. Overall architecture of the proposed Dual-Path Adaptive Feature Pyramid Network (DAFPN): (a) Backbone; (b) Semantic-guided upsample path; (c) Structure-enhanced downsample path.
Sensors 26 00381 g002
Figure 3. Illustration of the working mechanism of the Rotated Region Proposal Network (R-RPN). This figure depicts the core workflow of the R-RPN module. The enlarged panel on the right presents the sample assignment strategy during training.
Figure 3. Illustration of the working mechanism of the Rotated Region Proposal Network (R-RPN). This figure depicts the core workflow of the R-RPN module. The enlarged panel on the right presents the sample assignment strategy during training.
Sensors 26 00381 g003
Figure 4. The complete processing pipeline of Rotated RoIAlign: the input feature map ( H × W × C ) is transformed by a rotation matrix to map the proposal coordinates into the global coordinate system, and bilinear interpolation is then applied to generate a fixed-size feature block ( P × P × C ).
Figure 4. The complete processing pipeline of Rotated RoIAlign: the input feature map ( H × W × C ) is transformed by a rotation matrix to map the proposal coordinates into the global coordinate system, and bilinear interpolation is then applied to generate a fixed-size feature block ( P × P × C ).
Sensors 26 00381 g004
Figure 5. Structure of the Angle-Aware Collaborative Attention (AACA) module. (a) Channel attention branch A channel ; (b) Angle attention branch A orient ; (c) Spatial attention branch A spatial ; (d) Process of feature fusion and enhancement.
Figure 5. Structure of the Angle-Aware Collaborative Attention (AACA) module. (a) Channel attention branch A channel ; (b) Angle attention branch A orient ; (c) Spatial attention branch A spatial ; (d) Process of feature fusion and enhancement.
Sensors 26 00381 g005
Figure 6. Visualization of RSFPN Detection Results. (a) Detection of dense small objects; (b) Detection of medium and large objects; (c) Detection of extreme aspect-ratio objects; (d) Detection under complex background scenarios; (e) Multi-scale object detection within the same scene.
Figure 6. Visualization of RSFPN Detection Results. (a) Detection of dense small objects; (b) Detection of medium and large objects; (c) Detection of extreme aspect-ratio objects; (d) Detection under complex background scenarios; (e) Multi-scale object detection within the same scene.
Sensors 26 00381 g006
Figure 7. Comparison of Feature Heatmaps between RSFPN and the Baseline Model (Oriented R-CNN). First row: Input images; (a) dense ship scenario at a port; (b) multi-orientation airplane scenario at an airport; (c) urban traffic vehicle scenario. Second row: High-level feature heatmaps from the baseline model (Oriented R-CNN). Third row: Heatmaps from the RSFPN model.
Figure 7. Comparison of Feature Heatmaps between RSFPN and the Baseline Model (Oriented R-CNN). First row: Input images; (a) dense ship scenario at a port; (b) multi-orientation airplane scenario at an airport; (c) urban traffic vehicle scenario. Second row: High-level feature heatmaps from the baseline model (Oriented R-CNN). Third row: Heatmaps from the RSFPN model.
Sensors 26 00381 g007
Figure 8. Typical cases of comparison of detection results between RSFPN and other methods. Row (a): RSFPN succeeds/Oriented R-CNN fails. Row (b): RSFPN failed/ReDet succeeded. The red circles indicate the missed detections.
Figure 8. Typical cases of comparison of detection results between RSFPN and other methods. Row (a): RSFPN succeeds/Oriented R-CNN fails. Row (b): RSFPN failed/ReDet succeeded. The red circles indicate the missed detections.
Sensors 26 00381 g008
Table 1. Architectural comparison of RSFPN with representative rotated object detectors.
Table 1. Architectural comparison of RSFPN with representative rotated object detectors.
MethodArchitectureFeature EnhancementGeometric AlignmentLoss OptimizationKey Limitation
Oriented R-CNNTwo-stage (Anchor-based)Standard FPNRotated RoIAlignSmooth L1 (Decoupled)Feature aliasing for rotated objects
S2A-NetSingle-stage (Anchor-based)Feature AlignmentAlignConvIoU-Smooth L1Lacks directional priors in attention
R3DetSingle-stage (Refinement)Refinement NetworkIterative RegressionSmooth L1 + IoUPerformance drops on dense small objects
KLD/GWDLoss Plugin-Gaussian modelingKLD/GWDApproximation errors for extreme aspect ratios
RSFPN (Ours)Two-stage (Enhanced)DAFPNRotated RoIAlign+AACAGC-MTLSolves the above via unified design
Table 2. Summary of anchor and offset notations.
Table 2. Summary of anchor and offset notations.
SymbolMeaningDescription
( a x , a y , a w , a h ) Anchor boxCenter coordinates, width, and height of the anchor box
( δ x , δ y , δ w , δ h ) Anchor offsetsOffsets for center and size regression
( δ α , δ β ) Rotation offsetsOffsets for rotation characteristics
( x , y , w , h , Δ α , Δ β ) Rotated proposalPredicted center, width, height, and rotation offsets
( x , y , w , h , θ ) Rotated bounding boxFinal representation after rectification ( θ : rotation angle)
These notations are used in Equation (1) and throughout the oriented RPN module.
Table 3. Detailed statistics of the DOTA-v1.0 dataset (Training and Validation Sets).
Table 3. Detailed statistics of the DOTA-v1.0 dataset (Training and Validation Sets).
CodeCategoryTrain InstancesVal Instances
PLPlane26,1287633
BDBridge1435445
GTFGround Track Field48601565
SVSmall Vehicle59,66017,758
LVLarge Vehicle16,9965408
SHShip41,95012,364
TCTennis Court64562067
BCBasketball Court1498472
STStorage Tank18,1865744
SBFSoccer Ball Field45231423
RARoundabout44021390
HAHarbor12,6283869
SPSwimming Pool53391678
HEHelicopter1654524
CCContainer Crane1438450
Total 206,75363,790
Table 4. Comparison with state-of-the-art methods on the DOTA-v1.0 test set (AP@0.5/%).
Table 4. Comparison with state-of-the-art methods on the DOTA-v1.0 test set (AP@0.5/%).
MethodsmAPPLBDGTFSVLVSHTCBCSTSBFRAHASPHECC
FR-O [7]54.1379.4231.6657.6946.7146.5975.0289.6478.1159.4447.748.9153.5666.944.6332.57
R-DFPN [36]57.9480.9547.7362.3455.4550.0575.5490.8577.664.8256.2451.7656.6971.2948.6247.39
RRD [52]61.0188.5254.3270.1559.7863.4578.9190.2380.1272.3461.8756.4362.1874.5657.2952.67
RoI Trans. [21]69.5688.6465.7478.5266.7473.0183.5990.7477.2781.4663.5358.3967.975.4162.7458.39
Gliding Vertex [28]69.389.6463.5572.0262.2573.4782.3690.848579.0259.2665.564.187368.1651.88
R3Det [17]71.2389.8766.9280.4569.3176.1885.2790.6886.4583.1768.9466.7371.5280.8970.3862.15
CSL [49]72.1590.0268.4181.7370.8577.6286.3990.7287.2884.0569.8767.9473.1682.4771.8363.92
DCL [50]72.8990.2169.2782.5671.9378.4587.1290.8188.0384.7870.6568.7274.3183.2572.6964.87
ReDet [47]73.4890.8367.7382.6672.5478.3187.3890.987.8485.2670.4868.4274.1283.9273.6865.74
S2A-Net [20]74.1289.1171.1178.3968.1675.0184.9890.8687.8183.5371.1164.1672.7681.3273.2760.06
BBAVectors [53]74.3590.4571.8983.2773.4279.1687.9590.8888.7285.6371.8469.5875.4384.7174.3566.92
GWD [33]74.7890.6272.3483.9573.8779.6388.2790.9189.1585.9472.3170.1276.0885.2274.9867.45
KLD [32]75.2390.7872.8684.5274.3580.1788.6490.9389.6386.3272.8970.7576.7285.8375.6168.13
Oriented R-CNN [19]75.8790.4171.2385.5975.2480.3988.7991.2590.8585.5473.8870.5377.8787.6578.4768.29
AOPG [54]76.4591.1273.2886.2576.1381.4289.3591.0590.7887.1574.8672.3479.0387.5279.9170.23
Oriented RepPoints [18]76.7891.2573.6586.5776.4981.7889.6291.1291.487.4675.2372.7179.4187.8980.3470.65
Rotated Faster R-CNN [55]77.1591.3873.9286.8376.8282.1389.8791.2591.2787.7575.5873.0579.7688.480.7571.02
RSFPN (Ours)77.4290.1574.2587.1277.1582.4590.1290.4590.1288.0375.9473.4880.1586.8581.2371.46
The results with bold format and blue color indicate the best results for each column.
Table 5. Comparison on the HRSC2016 test set (AP@0.5/%).
Table 5. Comparison on the HRSC2016 test set (AP@0.5/%).
MethodsRRD [52]R3Det [17]Gliding Vertex [28]ReDet [47]Oriented R-CNN [19]S2A-Net [20]RSFPN (Ours)
AP5084.388.988.290.490.590.191.85
The evaluation metric (AP50) and the best performance are highlighted in bold.
Table 6. Comprehensive model complexity and inference speed comparison.
Table 6. Comprehensive model complexity and inference speed comparison.
MethodsBackbonemAP@0.5Params (M)GFLOPsFPS
RoI Trans. [21]ResNet-10169.5643.7225.311.8
Gliding Vertex [28]ResNet-10169.344.2228.112.3
R3Det [17]ResNet-10171.2343.9226.812.6
ReDet [47]ReResNet-5073.4845.2231.613.1
CSL [49]ResNet-5072.1542.1219.514.2
DCL [50]ResNet-5072.8942.3220.714.1
S2A-Net [20]ResNet-5074.1242.8218.914.6
BBAVectors [53]ResNet-5074.3541.9217.314.8
GWD [33]ResNet-5074.7842217.814.7
KLD [32]ResNet-5075.2342.1218.214.6
Oriented R-CNN [19]ResNet-5075.8741.5215.715.2
AOPG [54]ResNet-5076.4542.6222.314.3
Oriented RepPoints [18]ResNet-5076.7843.1224.813.9
Rotated Faster R-CNN [55]ResNet-5077.1543.8227.113.5
RSFPN (Ours)ResNet-5077.4242.3218.914.5
The bottom row in bold highlights our proposed method.
Table 7. Ablation studies of RSFPN modules on DOTA-v1.0 test set.
Table 7. Ablation studies of RSFPN modules on DOTA-v1.0 test set.
Experimental SettingmAPGain (%)Params (M)FPSSV
(AP)
LV
(AP)
SH
(AP)
BD
(AP)
Baseline (Oriented R-CNN)75.8741.815.275.2480.3988.7971.23
+ DAFPN76.92+1.0542.314.876.85 (+1.61)81.25 (+0.86)89.25 (+0.46)72.85 (+1.62)
+ DAFPN + AACA77.56+0.6442.514.677.05 (+0.20)81.85 (+0.60)89.85 (+0.60)73.85 (+1.00)
+ DAFPN + AACA + GC-MTL77.42+1.5542.514.577.15 (+0.10)82.45 (+0.60)90.12 (+0.27)74.25 (+0.40)
Table 8. Module combination ablation analysis. The symbols × and ✓ indicate that the module is excluded or included, respectively.
Table 8. Module combination ablation analysis. The symbols × and ✓ indicate that the module is excluded or included, respectively.
DAFPNAACAGC-MTLmAPGain (%)Notes
×××75.87Baseline
××76.92+1.05Only DAFPN
××76.35+0.48Only AACA
××76.08+0.21Only GC-MTL
×77.56+1.69DAFPN+AACA
77.42+1.55Full RSFPN
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, J.; Huo, H.; Kang, S.; Mei, A.; Zhang, C. Rotation-Sensitive Feature Enhancement Network for Oriented Object Detection in Remote Sensing Images. Sensors 2026, 26, 381. https://doi.org/10.3390/s26020381

AMA Style

Xu J, Huo H, Kang S, Mei A, Zhang C. Rotation-Sensitive Feature Enhancement Network for Oriented Object Detection in Remote Sensing Images. Sensors. 2026; 26(2):381. https://doi.org/10.3390/s26020381

Chicago/Turabian Style

Xu, Jiaxin, Hua Huo, Shilu Kang, Aokun Mei, and Chen Zhang. 2026. "Rotation-Sensitive Feature Enhancement Network for Oriented Object Detection in Remote Sensing Images" Sensors 26, no. 2: 381. https://doi.org/10.3390/s26020381

APA Style

Xu, J., Huo, H., Kang, S., Mei, A., & Zhang, C. (2026). Rotation-Sensitive Feature Enhancement Network for Oriented Object Detection in Remote Sensing Images. Sensors, 26(2), 381. https://doi.org/10.3390/s26020381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop