Next Article in Journal
STM-Net: A Multiscale Spectral–Spatial Representation Hybrid CNN–Transformer Model for Hyperspectral Image Classification
Previous Article in Journal
Scalable Context-Preserving Model-Aware Deep Clustering for Hyperspectral Images
Previous Article in Special Issue
FI-MambaNet: A Feature Integration Network with Mamba and Multi-Head Self-Attention for Remote Sensing Object Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Graph Synchronization to Policy Learning: Angle-Synchronized Graph and Bilevel Policy Network for Remote Sensing Object Detection

by
Jie Yan
1,
Jialang Liu
2,
Lixing Tang
3,
Xiaoxiang Wang
1 and
Yanming Guo
2,*
1
School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Haidian Compus, Beijing 100876, China
2
The Laboratory for Big Data and Decision, National University of Defense Technology (NUDT), Changsha 410004, China
3
College of Computer Science and Mathematics, Central South University of Forestry and Technology, Changsha 410004, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(24), 4029; https://doi.org/10.3390/rs17244029 (registering DOI)
Submission received: 24 September 2025 / Revised: 22 November 2025 / Accepted: 25 November 2025 / Published: 14 December 2025
(This article belongs to the Special Issue Efficient Object Detection Based on Remote Sensing Images)

Highlights

What are the main findings?
  • We proposed the ASBPNet framework that significantly improves oriented object detection through geometric alignment and policy adaptation.
  • We achieved breakthrough performance: 68.10% mAP on DIOR-R, 98.20% mAP (12) on HRSC2016, and 79.60% mAP on DOTA-v1.0.
What is the implication of the main finding?
  • The framework improves geometric stability and localization accuracy while remaining lightweight, achieving a balanced, efficient solution for high-density small object detection.

Abstract

Detection of rotating targets in complex remote sensing scenarios often suffers from angular inconsistencies and boundary jitter, especially for small-to-medium objects with rapid pose changes or indistinct boundaries in dense environments. To address this, we propose ASBPNet, a unified framework coupling geometric alignment with policy adaptation. It features the following: (1) Angle-Synchronized Graph (ASG), which injects angle–alignment relationships and residual-based boundary refinement to improve rotational consistency and reduce boundary errors for small objects; (2) Bilevel Policy Optimization (BPO), which unifies control over rotation enhancement, sample allocation, block scanning, and rotational NMS for cross-stage policy coordination and improved recall. Together, ASG and BPO form a tightly coupled pipeline in which geometric alignment directly reinforces policy optimization, yielding mutually enhanced rotation robustness, boundary stability, and detection recall across densely distributed remote sensing scenes. We conducted systematic evaluations on datasets including DIOR-R, HRSC2016, and DOTAv1.0: compared to baselines, overall accuracy achieved significant improvement on DIOR-R, with performance reaching 98.2% on HRSC2016. Simultaneously, enhanced robustness and boundary stability were demonstrated in complex backgrounds and dense small-object scenarios, validating the synergistic value of geometric alignment and policy adaptation.

1. Introduction

Remote sensing object detection refers to the automatic identification, localization, and extraction of targets of interest using observation images from satellites, drones, airborne platforms, or ground sensors [1,2,3]. Such images usually present a large-scale span, complex backgrounds, and diverse target morphology [4]. There are significant differences in spatial distribution and texture features between targets, and the background is often messy and dynamic, making it more difficult to distinguish the foreground from the background [5,6]. The targets themselves are also rich in morphological variations, from geometrically regular ships to complex industrial facilities, all of which put forward higher requirements on the feature characterization and generalization ability of the detection model [7,8]. Consequently, remote sensing object detection demands that models not only capture high-resolution details but also achieve robust contextual modeling and discrimination in complex environments [9]. In order to cope with these challenges, existing remote sensing object detection methods have played an important role in remote sensing interpretation and intelligent monitoring by means of feature pyramids [10,11,12], attention mechanisms [13,14], and multi-task learning [15,16] to continuously improve accuracy and robustness. However, the current methods are still prone to the problems of angular inconsistency and boundary jitter in directional target detection, and the strategies of rotational enhancement, sample allocation, block scanning, and Non-Maximum Suppression (NMS) involved in the training and inference process are often fragmented and optimized, which leads to the difficulty of balancing accuracy and delay. Thus, current remote sensing object detection models face the following limitations:
  • The dual constraints of limited training efficiency and lack of small-target recall. Although multi-strategy combinations can improve the performance of the model in different scenarios, the traditional detection process often splits the execution of enhancement, assignment, cropping, and post-processing strategies, and the lack of global synergy and dynamic adaptation mechanisms between strategies leads to slow convergence of training and weak generalization ability of strategy migration. At the same time, the existing strategies are often insufficient to model the distribution and scale sensitivity of small targets, which makes it difficult to effectively improve the edge-detection capability and small-target recall rate, and ultimately affects the overall robustness and consistency of precision.
  • Boundary fitting and angular consistency bottlenecks for rotating small targets. Small- and medium-scale targets in remote sensing images often show the characteristics of tiny size, changing attitude, and fuzzy boundary, which makes the rotation detection task face double challenges in returning to the frame boundary and maintaining the angular stability. Traditional methods often ignore the rotational correlation between candidate frames, and it is difficult to achieve the synergistic optimization of angular alignment and boundary correction, resulting in target edge drift and rotational angle jitter, which is especially obvious in multi-target dense scenarios.
In order to improve the accuracy of rotation detection and boundary fitting for small targets in remote sensing images, existing methods commonly use strategies such as rotating frame regression [17,18], angle classification [19,20], and direction-sensitive feature extraction [21] for modeling [22], while boundary refinement is performed through post-processing modules. These methods mitigate the localization error caused by target rotation to a certain extent, but due to the lack of modeling constraints on spatial geometric isotropy, it is often difficult to maintain stable consistency between the angle prediction and the edge contour. In addition, the boundary optimization process mostly relies on a fixed structure or heuristic design, with limited generalization capability, and is prone to problems such as jitter and trailing in high-density or complex backgrounds, leading to unstable performance in small-target detection.
To overcome these issues, we propose ASBPNet, a unified framework that concurrently addresses geometric instability and strategic fragmentation through two synergistic components: the Angle-Synchronized Graph (ASG) and the Bilevel Policy Optimization (BPO). The ASG establishes explicit geometric constraints among candidate boxes via graph message passing with SE(2)-equivariance, directly reducing angular jitter and improving boundary fitting. The BPO introduces a key innovation through its structured bilevel architecture, which formally separates strategy learning from execution. In this design, an upper-level reinforcement learning agent learns to coordinate all strategies, while a lower-level detector executes under these optimized strategies. This separation can uniformly optimize various non-differentiable strategies, including expansion, label assignment, and NMS, so as to jointly adapt under delay constraints.
A critical aspect of this design is the closed-loop interaction between these components: the stable geometric features from ASG provide a reliable basis for BPO’s policy evaluation, while the strategies optimized by BPO (like sample assignment) directly govern the learning process of ASG. This ensures that gains in one component consistently benefit the other, leading to aligned and robust performance across both training and inference.
The innovative contributions of this study are summarized as follows:
  • We propose a deeply integrated approach that unifies graph-based geometric modeling and reinforcement learning-based policy optimization. This is achieved by synergizing an Angle-Synchronized Graph Head (ASG), which solves angular inconsistency and boundary jitter through equivariant message passing, and a Bilevel Policy Optimization (BPO) module, which overcomes the fragmentation of non-differentiable strategies in training and inference.
  • To solve the problems of boundary instability and poor rotational robustness in the detection of remotely sensed targets at arbitrary angles, we propose the Angle-Synchronized Graph Head (ASG). Inspired by geometric equivariance and residual learning, this module enhances the angular consistency by introducing synapse-level equivariance constraints, and achieves the boundary refinement of candidate frames with the help of the micro-residual gating mechanism. It significantly improves the angle prediction accuracy and boundary alignment effect for small targets and dense structures.
  • To alleviate the problem of training–inference inconsistency and delayed feedback during multi-strategy decision-making, we propose the Bilevel Policy Optimization (BPO). This module achieves consistent modeling of training and inference policies through a two-layer optimization framework combining microagentable metrics with a delayed pairing mechanism to improve the generalization and model robustness of detection policies.
  • We conducted systematic evaluations on DIOR-R, HRSC2016, and DOTAv1.0. On the DIOR-R dataset, ASBPNet improves the mAP of the baseline from 63.40% to 68.10% with only a slight increase in parameters and computation; on the HRSC2016 dataset, it achieves 98.2% mAP (12), which is a new performance for similar models. Meanwhile, the heatmap visualization demonstrates that this method has stronger target-focusing and interference-suppression ability, which significantly reduces the misdetection rate in complex backgrounds and is especially suitable for high-density and small-target scenarios in the field of remote sensing.

2. Related Work

2.1. Graph Neural Network in Remote Sensing Object Detection

With the continuous development of graph representation learning, graph neural networks (GNN) have been explored in remote sensing object detection, with the advantage of being able to model complex relationships between targets and between targets and scenes [23]. By explicitly capturing topologies such as road networks, building layouts, or target co-occurrences, GNN methods effectively improve the detection performance of small, dense, and occluded targets, and make up for the shortcomings of traditional convolutional networks or Transformers in spatial relationship modeling [24,25]. According to different research focuses, GNN methods in remote sensing object detection can be broadly classified into three categories: relational inference [26,27], context-aware [28], and hierarchical graph modeling [29]; these show unique advantages in terms of accuracy and robustness, but still face the challenges of relying on the a priori for the composition of the graph, over-smoothing of the features, insufficient adaptation to multi-scale, and limited integration with the detection framework, which constrain the further development of GNNs in remote sensing object detection. The further development of GNN in remote sensing object detection is constrained.
Among them, relational inference-based models have been widely used in remote sensing image processing in recent years. These models mainly rely on the structure of graph neural networks to capture spatial and semantic relationships by constructing topological maps between targets, which is especially suitable for small-target detection tasks in complex scenes. The RelaDet [30] and GCN-based Detector [31] are the most representative models among them. RelaDet can model the contextual dependency between targets through relational embedding, which is especially suitable for the detection of densely distributed aircraft and ships; while GCN-based Detector introduces the mechanism of adjacency matrix propagation, which greatly enhances the structural sensing ability between targets, and is especially suitable for the scenes with occlusion or background interference. Although these models can effectively improve the detection accuracy and context-modeling capability, they still have limitations in applications. Relational embedding and neighbor propagation mechanisms usually rely on predefined topological graphs, lack adaptivity to dynamic scenes, and are prone to introduce noise when the distribution of targets is complex or the structure is ambiguous; meanwhile, graph convolution may lead to over-smoothing of features in multi-layer propagation, making it difficult to distinguish between neighboring but semantically significantly different targets, which can lead to confusing or redundant detection. In addition, relational modeling puts more emphasis on global consistency, with insufficient discriminative support for small-scale targets, limiting its further use in high-resolution remote sensing scenarios.
Complementary to this direction, context-aware models focus on utilising global and local contextual information in remote sensing images to aid target detection. Classical context-aware models, such as Contextual Graph Detector [26], extract image features through graph convolutional propagation and mask-attention mechanisms, and are able to effectively capture the contextual dependencies between multi-scale targets and explicitly model them. Hierarchical GNN [32], by introducing a hierarchical context-modeling mechanism, can solve the common problems of background interference and target obfuscation in remote sensing images, and is especially suitable for dealing with scenes with dense targets and large-scale differences. The advantage of these context-aware models is that they can significantly enhance the robustness of detection and adaptability to complex backgrounds, especially in large-scale remote sensing images. However, such approaches still have limitations in the modeling process. The graph convolution and attention mechanism mostly relies on neighborhood or super-pixel aggregation, which can capture the context but lacks the constraint of rotational invariance, and is prone to cause angular inconsistency and boundary jitter in directional targets; at the same time, context propagation emphasizes global consistency but may weaken the discriminative features of individual targets, leading to over-smoothing and redundancy detection; the hierarchical graph structure is also prone to introducing noise when aggregating across scales, making targets with large-scale differences underperform in boundary refinement.
In remote sensing object detection, hierarchical graph modeling approaches are gradually becoming an important direction by fusing local and global features by constructing graph structures at different levels. Representative models include PolyRoof [33] and HNS-GNN [34]. PolyRoof adopts the Attention U-Net backbone and vertex-level graph modeling in the roof polygonization task, which upgrades the pixel features into a vector structure, but relies on fixed adjacency connections in vertex relationship reasoning, which is prone to introducing misconnections in complex geometrical patterns, leading to discontinuities in local boundaries. HNS-GNN introduces structure-aware graph reasoning in road detection. GNN introduces a structure-aware graph inference module in road detection with simultaneous multi-task learning of regions and boundaries, but its hierarchical aggregation process overemphasizes boundary consistency, which may weaken the discriminative nature of fine-grained textures and easily cause pseudo-edge or false detection in scenes with strong background noise. Overall, this type of method effectively enhances multi-scale context perception capability through hierarchical graph structure, but there are still problems of noise propagation and insufficient feature refinement in the graph construction and hierarchical aggregation process, which limits its performance in large-scale complex remote sensing scenarios.
Compared with the existing methods, the ASBPNet framework proposed in this paper is more targeted at the graph neural network level. By constructing the graph structure between candidate frames through ASG and introducing SE(2) equivariant constraints in the message-passing process, we are able to maintain the feature consistency under rotation and translation, so as to realize the angular alignment and boundary refinement of candidate frames, and effectively alleviate the problems of angular inconsistency and boundary jitter that are prone to occur in previous methods for directional targeting. On this basis, we further design a residual boundary optimization mechanism, so that the nodes can predict the boundary residuals based on the neighborhood information, which significantly improves the geometric stability and positioning accuracy of the oriented frames. Meanwhile, this paper combines BPO to incorporate multiple policy decisions in training and inference into a unified optimization framework, which achieves a dynamic balance between detection accuracy and efficiency under the delay constraint. Overall, this design of combining the geometric modeling capability of graph neural networks and the adaptive optimization of reinforcement learning enables this paper’s method to achieve better accuracy, robustness, and efficiency than existing methods in remote sensing object detection.

2.2. Intensive Learning in Remote Sensing Object Detection

As the scale and complexity of remote sensing imagery increase, reinforcement learning (RL) has been progressively introduced into remote sensing object detection tasks to address challenges posed by small-scale targets, dense distributions, and multi-directional rotations [35,36,37]. RL models adaptively optimize the detection process through a state-action-reward mechanism, demonstrating advantages in candidate region selection, scale adjustment, rotation angle prediction, and redundancy suppression. Depending on methodological focus and application scenarios, RL approaches in remote sensing object detection can be categorized into region search and bounding box localization [38,39], rotation and scale modeling [40], and policy and fusion optimization [41], each exhibiting distinct capabilities. However, they also face challenges such as complex state spaces, reward function design sensitivity, and insufficient deep coupling with detectors.
Among them, reinforcement learning-based region search and candidate box localization have been widely used in remote sensing image processing in recent years. These methods model the detection problem of large-size remote sensing images as a sequential decision-making process of states, actions, and rewards, which enables the agent to gradually focus on potential target regions in a huge candidate space, thus significantly reducing the dependence on the whole high-resolution image. DeepRL-Detector [42] and RL-CNN [43] are the most representative models among them. The former learns whether to sample candidate regions at high resolution through a dual-layer network of coarse-grained and fine-grained strategies, which reduces redundant computation while maintaining detection accuracy, but is prone to missing fine-grained targets at the boundary due to the reliance on fixed grids for region delineation. The latter uses a reinforcement learning agent to gradually locate aircraft targets and verifies the detection results with a CNN, breaking through the limitation that traditional methods can only detect a fixed number of targets, but its action space is designed to gradually reduce the search area, which may trigger excessive search and localization conflicts in dense multi-target scenarios, thus affecting efficiency and accuracy. These methods have significant advantages in improving the computational efficiency and positioning accuracy of remote sensing object detection, but the inherent limitations of the candidate region delineation and action design make them still face certain challenges in scenes with complex target distribution or dense multi-targets.
Complementary to this direction, rotation and scale modeling approaches use reinforcement learning to dynamically adjust the angle and scale of candidate frames in remote sensing object detection to cope with multi-directional and multi-scale targets. Representative models such as FFPN-RL [44] achieve ship orientation prediction through feature pyramid fusion and iterative rotational actions, but the discrete action design is prone to slow convergence or error accumulation. These methods have advantages in rotational modeling and scale adaptation, but the limitations of action design and feature structure still affect their performance in complex scenes.
The strategy and fusion optimization approach emphasizes modeling the target detection process as a strategy selection and fusion learning problem, and realizing multi-link adaptive regulation by means of reinforcement learning. Representative models include Reinforcement Learning of Object Detection Strategies [45] and the HRL Saliency Model [46], which learn fusion strategies from different perspectives and information sources through Q-learning, combine PCA and RBF networks for appearance modeling, and then use Bayesian methods to improve detection confidence. The former uses Q-learning to learn the fusion strategy of different views and information sources, combines PCA and RBF network for appearance modeling, and then uses Bayesian method to improve the detection confidence, which significantly improves the detection efficiency in specific tasks, but due to the dependence of strategy learning on a predefined appearance-modeling module, it is easy to suffer from insufficient generalization in complex and diverse remote sensing images. The latter uses hierarchical reinforcement learning to generate saliency maps, fuses the bottom-layer color features with the top-layer line density features step by step, and judges the target saliency with the help of the thematic model to adaptively control the learning depth, which is able to effectively inhibit the background interference in airport detection. However, the inverse transfer of saliency maps between the layers is prone to introducing erroneous feedback when the features are weakened or the interference is enhanced, which leads to the insufficiently fine boundary delineation. Overall, such methods enhance the adaptability and robustness of detection through strategy optimization and fusion learning, but there are still structural limitations in appearance-modeling dependency and hierarchical feedback mechanisms.
In contrast, our ASBPNet framework introduces a structured bilevel optimization architecture. An upper-level policy learns to coordinate all key strategies including rotation augmentation, sample assignment, multi-tile scanning, and rotated NMS within a unified space. A lower-level detector then executes the actual detection task under these optimized strategies. This clear separation enables direct optimization of accuracy–efficiency trade-offs under latency constraints through dual variable updates, while differentiable performance proxies ensure stable policy learning. By addressing fundamental limitations in existing reinforcement learning methods, particularly in state space complexity and reward sensitivity, our bilevel design achieves significant improvements in small-target recall, angular consistency, and inference stability.
The analyses in Section 2.1 and Section 2.2 reveal complementary yet unaddressed challenges in existing approaches. While GNN-based methods improve relational modeling, they typically fail to enforce rotational equivariance, leading to persistent angular inconsistency in oriented object detection. Concurrently, RL-based methods provide adaptive policy optimization but remain insufficiently coupled with the geometric refinement process, resulting in fragmented strategy coordination. To address these interconnected limitations, we propose ASBPNet—a unified framework that systematically integrates equivariant graph modeling with Bilevel Policy Optimization. The methodology detailed next builds upon this coordinated design to simultaneously advance geometric stability and strategic coherence.
Relevant work summaries are presented in Table 1.

3. Methodology

ASBPNet is a unified framework where the Angle-Synchronized Graph (ASG) and Bilevel Policy Optimization (BPO) modules interact in a tightly coupled manner. Figure 1 illustrates the overall architecture and data flow. The pipeline commences with the backbone and neck networks extracting multi-scale features from the input image, subsequently generating an initial set of oriented candidate boxes. These candidate boxes and their associated features form the primary input to the ASG module.The code will be released at https://github.com/westneverlost/Angle-Synchronized-Graph-and-Bilevel-Policy-Network (accessed on 30 August 2025).
The ASG module then constructs a graph over these candidates and performs angle-synchronized message passing. Its output is a refined set of candidate boxes with significantly improved angular consistency and boundary precision. This geometrically stable set of detections is crucial for the subsequent stages.
Concurrently, the BPO module governs the strategic flow throughout the pipeline. During training, BPO observes performance proxies (e.g., IoU, recall) from the detector’s current state and outputs a set of optimized strategy parameters. These parameters control the data augmentation and sample assignment for the entire model, thereby influencing the feature learning in the backbone and the candidate boxes presented to the ASG. During inference, BPO’s learned policies directly configure the multi-tile scanning and rotated NMS procedures, which take the refined boxes from ASG as their input. The final output is produced by applying these optimized post-processing strategies to the stable geometries generated by ASG.

3.1. Baseline: YOLOv5-OBB

YOLOv5 with Oriented Bounding Box is an important detection framework in the field of remote sensing object detection, which is centered on extending the classical YOLOv5 architecture to the rotating box detection task in order to cope with the orientation diversity and scale inhomogeneity that are prevalent in remote sensing images. Architecturally, YOLOv5-OBB retains the end-to-end detection paradigm of YOLOv5, including input pre-processing, backbone feature extraction, feature fusion necking, and multi-scale prediction header.
Unlike the traditional horizontal frame detection pipeline, YOLOv5-OBB introduces a rotated frame parametric design in the prediction header section. Specifically, each candidate frame additionally outputs the rotation angle of the target in addition to the predicted center coordinates and width–height, enabling the model to directly regress to bounding boxes of arbitrary directions. This enables the network to perform more robustly when facing targets with significant directionality such as ships, aircraft, and buildings.
YOLOv5-OBB relies on the CSPDarknet backbone for multi-layer convolutional feature extraction in feature learning, and enhances multi-scale feature interactions through Path Aggregation Network (PANet) to ensure consistent modeling of large and small targets in high-resolution remote sensing scenes. Residual concatenation and cross-layer fusion further alleviate the gradient-vanishing problem and accelerate the convergence process under the rotation detection task.
In addition, YOLOv5-OBB adopts a rotating frame-specific label assignment and loss function design. To ensure the accuracy of detection, it combines IoU-based regression loss with angular regression loss to effectively suppress problems such as angular blurring and unstable bounding box regression.
Practical remote sensing missions often face challenges such as target orientation diversity, boundary jitter, and scale differences. Meanwhile, remote sensing images not only have a large number of invalid regions but also often contain complex texture patterns and non-homogeneous backgrounds, which require the model to have good geometric adaptability and contextual modeling capabilities while guaranteeing accuracy. YOLOv5-OBB adopts an end-to-end rotated frame detection architecture, which is equipped with fast prediction and rotationally invariant modeling capabilities, and is able to more adequately retain the orientation information in remote sensing images. For information in remote sensing images, in the detection head, the parametric design of the rotating frame of YOLOv5-OBB facilitates the capture of angular changes, which helps to achieve stable regression and fine alignment of candidate frames, and provides a natural advantage and strong support for mitigating the problems of orientation inconsistency and boundary instability.
Therefore, YOLOv5-OBB is chosen as the structural baseline in this paper, which not only considers target orientation diversity and boundary refinement but also has the ability to maintain detection stability under complex strategy space. YOLOv5-OBB has good generalization ability in remote sensing tasks, and it is especially suitable for multi-directional and multi-scale target-detection scenarios. However, YOLOv5-OBB is based on the rotating frame structure of horizontal frame extension, which is prone to boundary jitter in the dense target region, and has the problem where the strategy selection relies on manual experience, which is prone to the contradiction of delay increase and accuracy decrease in practical reasoning. If it can be optimized by combining the geometric constraints of remote sensing images with the task-adaptive structure, the performance of the model under orientation-consistent modeling and delay-constrained optimization will be significantly improved.

3.2. Angle-Synchronized Graph Head (ASG)

Accurate oriented object detection requires the predicted bounding boxes to maintain consistent angles and precise boundaries with the underlying targets. However, in dense or complex scenes, candidate boxes generated by a standard detector are often processed in isolation, lacking explicit constraints on their mutual rotational relationships. This isolation is a fundamental cause of angle inconsistency and boundary jitter, as the regression of each box fails to benefit from the collective geometric context. To address this core limitation, we introduce the Angle-Synchronized Graph Head (ASG). This module explicitly models the interdependencies between candidate boxes by constructing a graph where nodes represent boxes and edges encode their geometric relations. It then performs synchronized message passing that inherently respects rotational equivariance, allowing each node to refine its parameters based on coherent information from its neighbors. This process directly enforces angular alignment and enables collaborative boundary refinement, thereby overcoming the inherent instability of processing candidate boxes independently.
Each candidate box is regarded as a node i, whose feature is defined as
v i = ( f i , b i , s i ) ,
where f i R C is the RoI feature of candidate i extracted from the detector backbone, b i = ( x i , y i , w i , h i , θ i ) represents the oriented bounding box parameters with ( x i , y i ) as the center coordinates, ( w i , h i ) as width and height, θ i as the rotation angle, and s i [ 0 , 1 ] denotes the confidence score of the candidate. The edge feature between nodes i and j is defined as
e i j = c i c j 2 , log min ( w i , h i ) min ( w j , h j ) , cos ( θ i θ j ) , sin ( θ i θ j ) ,
where c i = ( x i , y i ) and c j = ( x j , y j ) are the centers of nodes i and j, the first two terms represent the distance and scale ratio between nodes, and the last two terms encode the cosine and sine of the relative rotation angle for geometric constraints during message passing.
To ensure isometry under rotational transformations during message passing, we explicitly align features from neighboring nodes prior to aggregation. Since candidate boxes from different orientations encode features within their respective local coordinate systems, directly aggregating these unaligned features would compromise angular consistency. Therefore, we employ the relative rotation angle Δ θ i j = θ j θ i to rotate features from neighboring node j into the coordinate system of central node i. This rotation is achieved via the SE(2) rotation operator Γ ( Δ θ i j ) , which applies a 2D rotation matrix to the spatial dimensions of the features while preserving other feature channels. This operation ensures all features participating in message passing are geometrically aligned, enabling the network to learn transformation-invariant representations. The specific procedure is as follows:
f ˜ j = Γ ( Δ θ i j ) f j ,
with the SE(2) rotation operator defined as
Γ ( Δ θ ) = cos Δ θ sin Δ θ sin Δ θ cos Δ θ I C 2 ,
where f ˜ j is the rotated feature used for message passing and I C 2 indicates that the remaining feature channels are kept unchanged.
The node feature x i ( l ) is updated through multiple layers of message passing:
x i ( l + 1 ) = x i ( l ) + j N ( i ) α i j Φ [ x i ( l ) , f ˜ j , e i j ] ,
where x i ( l ) is the feature of node i at layer l; initialized as x i ( 0 ) = f i , N ( i ) is the neighborhood of node i, Φ ( · ) is a learnable message function, and α i j is the attention coefficient:
α i j = softmax a σ W [ x i ( l ) , f ˜ j , e i j ] ,
This aggregated message, refined by the attention gate, is used to predict a set of micro-residuals Δ b i . These residuals represent precise corrections to the original box parameters, computed through a linear layer that transforms the node’s updated feature. The key advantage of this residual formulation is that it decomposes the complex task of direct coordinate regression into learning smaller, more stable adjustments. This allows the model to specialize in refining initial detections by making targeted modifications to existing box coordinates. The final refined box is then obtained by applying these residuals:
Δ b i = [ Δ x i , Δ y i , Δ w i , Δ h i , Δ sin θ i , Δ cos θ i ] ,
which is applied to update the box as
b i = Apply ( b i , Δ b i ) , θ i = atan 2 ( sin θ i + Δ sin θ i , cos θ i + Δ cos θ i ) ,
where b i is the refined oriented bounding box, θ i is the refined rotation angle, and Apply ( · ) denotes applying the residual to the original box, with residuals derived from neighbor-informed message aggregation.
The final output set of ASG is
B = { ( b i , s i ) i = 1 , , N } ,
where b i is the refined OBB computed from the original b i and residual Δ b i , and s i is the updated confidence score from the message-passed features. Compared to the original set { b i , s i } , B achieves better angular consistency, boundary precision, and redundancy suppression.

3.3. Bilevel Policy Optimization (BPO)

In rotated-object detection, multiple non-differentiable and coupled strategy decisions exist in both training and inference, including rotation augmentation, positive or negative sample assignment, multi-tile scanning, and rotated NMS. To jointly optimize these strategies under latency constraints, we propose a Bilevel Policy Optimization (BPO) framework. BPO formulates a bilevel optimization problem, where an upper-level policy decides strategy parameters and a lower-level detector evaluates the performance under those strategies.
The Bilevel Policy Optimization (BPO) framework is formally structured into two distinct layers with clearly separated responsibilities:
  • Upper-Level Policy: This acts as the strategic planner, implemented as a reinforcement learning agent. It observes the detection system’s performance metrics and latency status, then outputs optimized parameters for the four core strategies: rotation augmentation, sample assignment, multi-tile scanning, and rotated NMS. This policy is trained to maximize a compound reward that balances accuracy improvements against computational costs, with an adaptive dual variable enforcing the latency constraint through projected ascent updates.
  • Lower-Level Detector: This serves as the tactical executor, comprising the core detection network including the ASG module. It receives strategy configurations from the upper level as fixed hyperparameters, executes the detection task, and returns performance measurements. A key component is its differentiable performance surrogate that enables effective policy learning by approximating non-differentiable evaluation metrics.
This formal separation creates a coherent optimization cycle where the upper level learns what strategies to apply while the lower level specializes in executing detection under those strategies, with performance feedback closing the loop.
We define the environment state at time step t as
s t = ρ ¯ 0.5 , t , ρ ¯ 0.75 , t , r t small , p t , h t 1 , τ t 1 ,
where ρ ¯ 0.5 , t and ρ ¯ 0.75 , t denote the mean rotated IoU at thresholds 0.5 and 0.75 for the mini-batch, r t small is the recall for small objects, p t is the precision, h t 1 encodes the previous strategy action in one-hot form, and τ t 1 is the latency observed in the previous step.
The LC-BPO policy outputs an action vector that configures the four strategy knobs:
a t = a t R , a t D , a t M , a t N ,
where a t R selects the rotation augmentation angle and probability, a t D sets thresholds for positive/negative assignment, a t M chooses multi-tile scanning parameters (tile size, stride, overlap, rotation switch), and a t N determines rotated NMS thresholds.
Given a t , the lower-level detector evaluates its effect on the current mini-batch and returns proxy statistics:
m t = f detector ( X t ; a t ) ,
where X t is the mini-batch input and m t summarizes approximated rotated IoU, recall, and precision.
The standard average precision (AP) metric relies on a non-differentiable Heaviside step function for determining true positives based on IoU thresholds:
H ( x ) = 1 if x 0 0 if x < 0
where x = rioU i τ i . This discontinuity prevents gradient-based optimization. Therefore, to obtain a differentiable surrogate, we adopt a smoothed AP proxy via Heaviside relaxation by approximating H ( x ) with a scaled sigmoid function:
AP ^ t = 1 M i = 1 M σ rIoU i τ i ϵ p i ,
where M is the number of detections, rIoU i is the rotated IoU of detection i, τ i is the threshold, p i is the confidence score, ϵ is a smoothing factor, and σ ( · ) is the sigmoid function.
The step-wise reward balances accuracy gains and latency penalties under a budget:
r t = AP ^ t AP ^ t 1 + β r t small r t 1 small λ max ( 0 , τ t B ) ,
where β tunes the emphasis on small-object recall, λ is a Lagrangian multiplier, and B is the latency budget.
The upper-level policy π θ is optimized using a policy gradient method (e.g., PPO):
θ θ + η θ E π θ t r t ,
where η is the learning rate and θ denotes policy parameters.
To enforce the latency constraint, the dual variable λ is updated by projected ascent. This update is central to our framework’s ability to balance accuracy and efficiency, as it dynamically adjusts the penalty weight based on constraint violation, steering policy learning toward latency-compliant solutions.
λ max 0 , λ + γ ( τ t B ) ,
where γ is the dual ascent step size.
After convergence, the framework yields an optimized strategy set for training and inference:
A * = { a 1 * , a 2 * , , a T * } ,
which, when deployed, produces the final latency-compliant rotated detection results:
D = { ( d k , s k ) k = 1 , , N } ,
where d k are the refined oriented bounding boxes and s k are the updated confidence scores.

4. Experiments

4.1. Datasets

DOTAv1.0 is one of the largest rotating frame detection datasets currently available, containing 2806 frames with resolutions up to 4000 × 4000 remote sensing images with 188,282 instances of 15 types of targets: aircraft (PL), baseball diamonds (BD), bridges (BR), surface runways (GTF), small vehicles (SV), large vehicles (LV), ships (SH), tennis courts (TC), basketball courts (BC), storage tanks (ST), football fields (SBF), roundabouts (RA), harbors (HA), swimming pools (SP), and helicopters (HC).
HRSC2016 focuses on ship detection and contains 1061 high-resolution shoreline images with a total of 2976 labeled ship instances. With extreme target aspect ratios and large-scale spans, HRSC2016 is a classic dataset for evaluating the detection performance of elongated targets.
DIOR-R is a rotating frame extension of the DIOR dataset, providing 23,463 multi-source remote sensing images covering 20 common categories (e.g., aircraft, vehicles, ports, tanks, etc.), with a total of about 190 k directionally labeled instances.

4.2. Experiments Settings

To ensure comparability with mainstream methods, we maintained consistency in data pre-processing: DOTA-v1.0 employed single-scale training and testing; FAIR1M-v1.0 employs a multi-scale approach, where images are first scaled to three scales (0.5×, 1.0×, 1.5×), then cropped into 1024 × 1024 patches with 500-pixel overlap; HRSC2016 maintains aspect ratios while uniformly resizing the longer side to 800 pixels. All remote sensing object detection experiments were conducted on four RTX 4090 GPUs with a batch size of 8, learning rate of 0.00005, and 12 training epochs. Results are reported across the DOTA-v1.0, HRSC2016, and FAIR1M-v1.0 datasets.

5. Results in Classic Datasets

Figure 2 provides a visual comparison of feature attribution heatmaps between the baseline method and our proposed approach, offering analytical insights into the geometric stability of our framework. In the baseline heatmap, the elliptical region highlights critical issues with small rotating objects: the activation areas exhibit diffuse boundaries and significant background interference. This visual pattern correlates directly with the baseline’s tendency for boundary jitter and angular inconsistency, as the model fails to focus precisely on target contours.
In contrast, our ASBPNet generates markedly more concentrated activations with sharply defined boundaries in the same region. This cleaner visual manifestation demonstrates the effectiveness of our geometric modeling: the Angle-Synchronized Graph (ASG) head successfully enforces spatial coherence through its equivariant message passing, while the Bilevel Policy Optimization (BPO) module contributes by optimizing feature learning through improved sample assignment. The significantly reduced background activation further confirms our model’s enhanced discrimination capability, validating that our unified framework effectively addresses both boundary instability and background distraction in rotated object detection.
As shown in Table 2 and Figure 3, on the DIOR-R dataset, ASBPNet achieves a mAP of 68.10, which is 8.56 and 6.12 percentage points higher compared to the mainstream two-stage methods Faster RCNN-O and TIOE-Det, respectively, and also outperforms the Transformer architectures ARS-DETR and PKINet-S, demonstrating a stronger generalized detection capability. In particular, compared with our baseline YOLOv5m-OBB, mAP improves by 4.7 percentage points with only 0.9M increase in the number of parameters and 5G increase in the number of FLOPs, which fully demonstrates the significant performance gain of this architecture in efficiency-constrained scenarios.
Table 3 shows that in the HRSC2016 dataset, ASBPNet achieves 90.60 on mAP (07), which is 2.1 percentage points higher than the 88.50 of the baseline YOLOv5m-OBB. This effect comes from the angle-aware adaptive grouping mechanism we introduce, which enables the model to be more accurate in feature extraction for rotating targets. On mAP (12), ASBPNet further achieves 98.20, an improvement of 1.4 from the baseline 96.80, thanks to the stable enhancement of the bounding box regression process by the delayed pairwise optimization strategy, which effectively reduces the localization error caused by the angular point offset of the long bar target. In terms of the number of parameters, ASBPNet is controlled at 22.1 M, which is only 0.9 M more than the baseline, but achieves higher detection accuracy, and this coordination between accuracy improvement and lightweight parameter scale indicates that the boundary perception module we designed has extremely strong feature enhancement capability and computational utilization. FLOPs increase slightly from 153 G to 158 G, but while maintaining very high inference efficiency, the FLOPs increase slightly from 153 G to 158 G, but maintain the high inference efficiency while significantly enhancing the sensitivity to the target’s rotational state, which is especially suitable for the vessel targets with high aspect ratios and variable orientations in HRSC2016. These results demonstrate that ASBPNet achieves high-precision modeling of angle-sensitive regions without substantially increasing model complexity, and its rotational robustness and boundary regression stability provide a reliable guarantee for the detection of complex attitude remote sensing targets.
Table 4 and Figure 4 shows that ASBPNet achieves an mAP of 79.60 at a single scale, significantly outperforming YOLOv5m OBB’s 77.30 and surpassing the strongest competitor in the same group, MTP, with 79.03. At the multi-scale level, ASBPNet achieves an mAP of 81.50, slightly below LSKNet S’s 81.64 but securing top or second-place performance across most categories with greater overall stability. Category-by-category comparisons reveal improvements primarily in direction-sensitive and densely populated small-target categories like bridges, helicopters, and cars. This demonstrates the method’s effectiveness in suppressing angular inconsistencies and boundary jitter, yielding cleaner suppression results. This effect stems from enhanced consistency through geometric alignment at the candidate level and graph message passing, while also benefiting from the strategy layer’s joint adaptation of rotation enhancement, positive–negative assignment, tile scanning, and rotation NMS. Categories with relatively smaller gains, such as trucks and swimming pools, are more constrained by scale and texture representation, tending to rely on stronger backbone or multi-scale feature supplementation. Overall, ASBPNet achieves steady-state gains across all classes without modifying the backbone or neck architecture.

Ablation Study

Validation of the effectiveness of geometric modeling and strategy optimization for collaborative enhancement. Table 5 and Figure 5 show that the introduction of the SE2-ASG module alone leads to a +0.8 mAP improvement, mainly due to the fact that the module maintains the angular consistency of the candidate frames through rotated iso-graph messaging, which significantly mitigates the boundary jitter problem. LC-BPO alone also delivers +0.7 mAP, indicating that the two-layer policy optimization under the delay constraint can stabilize the consistency in the training–inference phase and improve the recall of small targets. The largest improvement (+1.3) is achieved when combining the two, suggesting that geometric modeling and strategy optimization are complementary: the former improves the spatial consistency of the candidate frames, while the latter improves the adaptability of the multi-stage strategy, and the synergistic effect is significantly better than that of the separate modules.
The synergy between ASG and BPO originates from their interaction in feature regularization and adaptive optimization. ASG enhances spatial coherence and structural stability within feature maps, producing cleaner geometric priors for detection. BPO then leverages these priors to adaptively adjust anchor assignment and suppression strategies, improving the balance between precision and recall. In turn, BPO’s adaptive weighting further emphasizes features refined by ASG, forming a mutually reinforcing loop. This cooperative mechanism enables the combined model to achieve a +1.3 mAP gain, demonstrating that geometry-aware feature alignment and adaptive strategy optimization jointly contribute to more stable and discriminative detection performance.
While ASBPNet employs COCO pre-training compared to the ImageNet initialization of YOLOv5m-OBB baseline, this difference does not compromise the fairness of comparison. In remote sensing detection, both COCO and ImageNet serve primarily as general visual priors, whose influence is largely normalized after fine-tuning on large-scale domain-specific datasets like DIOR-R. Importantly, our experimental setup aligns with standard practices in the field, where numerous state-of-the-art methods, such as LSKNet [61] and PKINet [62], typically employ ImageNet pre-training before fine-tuning on remote sensing benchmarks including DIOR and DOTA. The +4.7 mAP improvement achieved with minimal parameter and FLOPs increase (+0.9M, +5G) strongly suggests that the performance gain originates primarily from our proposed ASG and BPO modules rather than pre-training bias. The architectural innovations enable more effective utilization of the feature representations regardless of the initial pre-training source.
Comparative analysis of angular synchronization and graphical information modeling. Table 6 and Figure 6 compare the different messaging and synchronization strategies. The results show that the Angle Sync constraint is indispensable, and the mAP decreases to 80.78 (−0.22) after removal, suggesting that angle alignment is crucial for rotational target stability. Meanwhile, the angle encoding embedded in von Mises performs the best (81.05), which is better than using the original angle directly (80.60), proving that the periodic modeling is more suitable for the rotational geometry property. In terms of message-passing methods, GATv2 maintains the best balance, and GraphSAGE and EdgeConv are slightly degraded but the difference is limited, indicating that the module is robust to message functions.
Structural perceptual analysis for boundary modeling. Table 7 and Figure 7 shows that there is a “moderate optimality” between the neighborhood size and the number of layers on the performance; k = 8 and three layers have the best message delivery. Too small a neighborhood results in insufficient information (80.82), and too large a neighborhood or too deep a layer has limited benefits (81.04/81.03), indicating that excessive aggregation introduces noise. In terms of feature levels, P3 alone decreases significantly (−0.3), while multi-layer fusion (P3–P6) can be slightly improved, reflecting the role of cross-layer feature integration for boundary modeling of complex targets.
Sensitivity analysis of weights and steps. Table 8 and Figure 8 show that both excessively small and large values of the synchronization loss of weight λ sync degrade performance, with the optimal range falling between 1 and 2, reflecting the moderate constraining effect of angle regularization. Adaptive prediction (81.00) and linear search (81.06) for the step size η yield the best results, while a fixed step size leads to degradation, indicating that this module benefits from adaptive strategies during iterative updates.
An analysis of the advantages of strategy collaboration in reinforcement learning. Table 9 shows that a single action space cannot steadily improve the precision, but instead both show a decrease (−0.34 to −0.58). In particular, DPA alone optimization tends to lead to an imbalance between recall and precision. In contrast, the joint optimization (C) obtains the best mAP (80.90) while ensuring the delay is controllable, indicating that multi-strategy coupling can compensate the limitations with each other, which is the key to RL optimization.
DNMS threshold adjustment and scale sensitivity. Table 10 demonstrates the performance of delay-constrained non-maximal suppression (DNMS) under different scale thresholds. The results show that moderately increasing the IoU thresholds (0.45/0.55/0.60) for medium and large targets can effectively suppress the redundant frames, and the mAP improves to 81.00. The reason is that the adaptive strategy of BPO retains the recall of small targets at higher IoUs and enhances the discriminative power of large targets at the same time. And when the threshold is too high (0.50/0.60/0.65), the discrimination of small targets is weakened, resulting in the overall precision falling back. This indicates that DNMS needs to dynamically balance overlap suppression and small-target retention in multi-scale targets, which is tightly coupled with the delayed budget-scheduling mechanism of BPO.
Interaction of slicing strategy with rotation enhancement. Table 11 compares the effect of different slicing configurations on accuracy and latency. A smaller tile (768/256/33%) leads to a slight accuracy improvement (80.96) with denser spatial coverage, but a significant increase in inference latency (+6%). In contrast, large tiles (1280/512/20%) reduce redundant computation but result in a slight decrease in mAP due to an increase in misdetection. Turning off rotational enhancement (OFF) resulted in a performance degradation (−0.05), verifying that rotational enhancement is critical for the angular alignment of aerial-type targets (e.g., aircraft, ships). Overall, the design of the AMTS plays a key role in the accuracy–delay trade-off by dynamically selecting the tile scale and rotation switch in the BPO framework.
Rotation enhancement distributions and a priori matching. Table 12 shows that different rotation enhancement distributions significantly affect model performance. The uniform distribution (Uniform) introduced too many invalid rotations with full angular coverage, leading to a decrease in mAP (80.82). In contrast, the Bimodal distribution (±15°/±30°) has the highest accuracy (81.02) because it is more consistent with the orientation prior (e.g., runway, road, boat direction) in the remote sensing scene. This suggests that LC-BPO can better optimize the angular inconsistency problem and improve the rotational robustness of the model if it is combined with the data distribution prior when learning the enhancement strategy.
Impact of positive and negative sample segmentation on boundary learning. Table 13 explores the impact of positive and negative thresholds in Dynamic Sample Allocation (DPA) on the results. Looser thresholds (0.4, 0.2) lead to too many mis-matched samples and a significant decrease in mAP (80.60). When the positive threshold is increased to 0.6, the mAP reaches its highest (80.96), suggesting that a stricter definition of positive samples can improve the boundary fitting accuracy. However, if the negative sample threshold is increased to 0.4 at the same time, some of the fuzzy boundary targets are misclassified as negative examples, recall decreases, and the overall performance is not as good as the default setting. This verifies the key role of BPO in threshold scheduling, i.e., dynamic balance between precision and recall.
Robustness of policy optimization with delay budget effect. Table 14 shows that both PPO and SAC converge stably, verifying the robustness of the reinforcement learning framework. More importantly, the delay budget B has a significant effect on the performance: when relaxed to 15 ms, the mAP improves to 81.05, indicating that the policy space can be more fully explored when resources allow. The mAP reaches 81.00 for small-target weight β = 0.5 , indicating that increasing the reward signal for small targets is effective in mitigating the lack of recall. Experiments with the AP agent smoothing factor ϵ show that values either too small or too large can lead to instability, while the default value of 0.1 is optimal. These results demonstrate that BPO can maintain a balance between computational resources, reward design, and convergence stability.
Advantages of module complementarity on complex categories:Figure 9 compares the results for the three most challenging target categories (bridge BR, vessel SV, and helicopter HC). The results show that SE2-ASG alone (B) improves significantly on the geometrically sensitive BR category (+1.8), while LC-BPO alone (C) improves even more on the dynamically distributed SV category (+1.4). The complete scheme (D) combines the two and significantly improves all three classes to 64.1/82.3/67.6. This suggests that angular equivariance modeling with graph neural networks and strategy optimization with reinforcement learning complement each other, with the former addressing angular consistency and boundary stability, and the latter improving the robustness of small targets and complex scenarios, and the synergy is particularly prominent in the complex target classes.

6. Discussion

To qualitatively evaluate ASBPNet’s boundary stability and feature-focusing capability in rotating-object detection, we selected eight representative remote sensing scenes from the DOTA dataset for visual analysis. We selected both conventional conditions (Figure 10) and complex conditions (Figure 11) to discuss the detection performance of both methods. For each scenario, we present the detection results alongside their corresponding attention heatmaps.
In both regular and complex scenes, the activation range of the baseline model YOLOv5-OBB thermogram is relatively diffuse, and many background areas such as water surface and open space show large highlight responses, resulting in the boundaries and contours of the foreground targets not being clear enough, and the detection frame Angular drift and boundary jitter occur, reflecting the model’s deficiencies in rotational consistency and feature focusing. In contrast, the activation distribution of the ASBPNet heatmap is obviously more focused on the actual buildings and dock structures, with clear boundaries, target contours, and geometries being better portrayed, while the responses from the water surface and other irrelevant areas are effectively suppressed. The effectiveness of the proposed BPO method in suppressing irrelevant surface texture and background shadows is verified.
The resulting detection frames are highly consistent with the orientation and geometry of the real target, which fully demonstrates the robust detection and focusing capability of the model on regular small targets and boundary-sensitive targets. Comparing the detection result plots of the two methods, some of the detection frames generated by the baseline method fail to achieve accurate wrapping of the target, and there is a slight boundary jitter. Taking the aircraft in the first row of Figure 10 as an example, the detection frame of the baseline method has a large offset from the real boundary in the left wing region, while ASBPNet is able to achieve a highly consistent fit with the contour of the airframe. In the second and third rows of vehicle detection, the detection frame generated by the baseline method does not fit well with the vehicle boundary, and the tail of some vehicles still maintains a certain distance from the detection frame, while ASBPNet is able to achieve complete coverage of the target. In the tennis court detection task in the fourth row, the detection frame of the baseline method has obvious deviation between the left boundary and the real field, while ASBPNet shows significant advantages in boundary fitting and achieves a high degree of consistency with the target contour.

7. Conclusions

In this paper, we propose a unified detection framework that combines geometric perception and policy optimization to address the key problems of remote sensing object detection, such as insufficient small-target recall, unstable boundary fitting, and inference delay due to multi-policy coupling. The core of the method includes two innovative modules, Angle-Synchronized Graph Head (ASG) and Bilevel Policy Optimization (BPO). The former achieves angular consistency modeling and boundary residual optimization by constructing graph structures between candidate frames and introducing rotational equivariant messaging, thus significantly mitigating the boundary jitter and angular deviation problems of rotating small targets. The latter unifies multiple non-differentiation strategies in the training and inference phases into a reinforcement learning optimization framework through the modules of rotational augmentation, sample allocation, multi-block scanning, and rotational NMS, which enables the detector to achieve dynamic accuracy–efficiency trade-offs under delay constraints.
In the experimental part, this paper performs systematic validation on several remote sensing rotating frame detection datasets such as DIOR-R, HRSC2016, and DOTA-v1.0. The results show that the method outperforms the baseline YOLOv5-OBB in terms of overall accuracy and achieves significant performance improvement on complex categories such as bridges, ships and helicopters. Meanwhile, ablation experiments further validate the complementary nature of the two modules: while ASG mainly improves angular consistency and geometric stability, BPO effectively improves small-target recall and multi-strategy adaptivity, and the combination of the two can achieve optimal performance.
Overall, the framework proposed in this paper breaks through the bottleneck of existing methods in rotational consistency modeling and delay constraint optimization while maintaining lightweight and real-time performance. It not only provides a new solution for rotating target detection in remote sensing under complex background but also provides a useful exploration for the deep integration of graph neural networks and reinforcement learning in remote sensing visual tasks.
For future work, we aim to explore several research directions that further leverage the unique synergy between geometry-aware representation and policy optimization in ASBPNet. One direction is cross-modal adaptation. We will extend ASG to align geometric cues across optical, SAR, and multispectral inputs and couple this with a policy module that selects modality-specific scanning and suppression strategies. Another direction is efficient large-model migration. We will study adapter-style modules, low-rank adaptation and policy-conditioned fine-tuning to transfer pretrained backbones while keeping ASG’s structural priors intact and maintaining low FLOPs. We will also investigate temporal and multi-resolution designs so that ASG can aggregate spatio-temporal geometry and BPO can schedule computation dynamically across frames and scales. We will validate these extensions through transfer experiments, per-class robustness tests, and deployment-aware latency–accuracy trade-off analysis.

Author Contributions

Conceptualization, J.Y., X.W. and Y.G.; methodology, J.Y. and L.T.; software, J.L. and L.T.; validation, J.Y., J.L. and L.T.; formal analysis, J.Y. and L.T.; investigation, J.Y. and L.T.; resources, X.W. and Y.G.; data curation, J.L. and L.T.; writing—original draft preparation, J.Y.; writing—review and editing, J.Y., L.T., J.L., X.W. and Y.G.; visualization, J.L. and L.T.; supervision, X.W. and Y.G.; project administration, X.W. and Y.G.; funding acquisition, X.W. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Training Program for Excellent Young Innovators of Changsha (Grant No. kq2209001), Hunan Excellent Young Scientists Fund (Grant No. 2025JJ40066).

Data Availability Statement

The datasets DIOR-R, HRSC2016, and DOTA-v1.0 arepublicly available from their respective sources.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASBPNetEdge Feature and Particle Focusing Network
ASGAngle-Synchronized Graph Head
BPOBilevel Policy Optimization

References

  1. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  2. Deng, Z.; Sun, H.; Zhou, S.; Zhao, J.; Lei, L.; Zou, H. Multi-scale object detection in remote sensing imagery with convolutional neural networks. ISPRS J. Photogramm. Remote Sens. 2018, 145, 3–22. [Google Scholar] [CrossRef]
  3. Cheng, G.; Han, J. A survey on object detection in optical remote sensing images. ISPRS J. Photogramm. Remote Sens. 2016, 117, 11–28. [Google Scholar] [CrossRef]
  4. Li, X.; Deng, J.; Fang, Y. Few-shot object detection on remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5601614. [Google Scholar] [CrossRef]
  5. Liu, Y.; Li, Q.; Yuan, Y.; Du, Q.; Wang, Q. ABNet: Adaptive balanced network for multiscale object detection in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5614914. [Google Scholar] [CrossRef]
  6. Mohammadpour, P.; Viegas, D.X.; Viegas, C. Vegetation mapping with random forest using sentinel 2 and GLCM texture feature—A case study for Lousã region, Portugal. Remote Sens. 2022, 14, 4585. [Google Scholar] [CrossRef]
  7. Zhang, S.; He, G.; Chen, H.B.; Jing, N.; Wang, Q. Scale adaptive proposal network for object detection in remote sensing images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 864–868. [Google Scholar] [CrossRef]
  8. Wu, T.; Dong, Y. YOLO-SE: Improved YOLOv8 for remote sensing object detection and recognition. Appl. Sci. 2023, 13, 12977. [Google Scholar] [CrossRef]
  9. Wang, X.; Wang, A.; Yi, J.; Song, Y.; Chehri, A. Small object detection based on deep learning for remote sensing: A comprehensive review. Remote Sens. 2023, 15, 3265. [Google Scholar] [CrossRef]
  10. Chen, C.; Gong, W.; Chen, Y.; Li, W. Object detection in remote sensing images based on a scene-contextual feature pyramid network. Remote Sens. 2019, 11, 339. [Google Scholar] [CrossRef]
  11. Zhang, X.; Zhu, K.; Chen, G.; Tan, X.; Zhang, L.; Dai, F.; Liao, P.; Gong, Y. Geospatial object detection on high resolution remote sensing imagery based on double multi-scale feature pyramid network. Remote Sens. 2019, 11, 755. [Google Scholar] [CrossRef]
  12. Chen, J.; Wang, S.; Chen, L.; Cai, H.; Qian, Y. Incremental detection of remote sensing objects with feature pyramid and knowledge distillation. IEEE Trans. Geosci. Remote Sens. 2020, 60, 5600413. [Google Scholar] [CrossRef]
  13. Du, Z.; Liang, Y. Object detection of remote sensing image based on multi-scale feature fusion and attention mechanism. IEEE Access 2024, 12, 8619–8632. [Google Scholar] [CrossRef]
  14. Ghaffarian, S.; Valente, J.; Van Der Voort, M.; Tekinerdogan, B. Effect of attention mechanism in deep learning-based remote sensing image processing: A systematic literature review. Remote Sens. 2021, 13, 2965. [Google Scholar] [CrossRef]
  15. Cui, F.; Jiang, J. MTSCD-Net: A network based on multi-task learning for semantic change detection of bitemporal remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2023, 118, 103294. [Google Scholar] [CrossRef]
  16. Niu, Y.; Guo, H.; Lu, J.; Ding, L.; Yu, D. SMNet: Symmetric multi-task network for semantic change detection in remote sensing images based on CNN and transformer. Remote Sens. 2023, 15, 949. [Google Scholar] [CrossRef]
  17. Wu, F.; He, J.; Zhou, G.; Li, H.; Liu, Y.; Sui, X. Improved oriented object detection in remote sensing images based on a three-point regression method. Remote Sens. 2021, 13, 4517. [Google Scholar] [CrossRef]
  18. Zhou, K.; Zhang, Z.; Gao, C.; Liu, J. Rotated feature network for multiorientation object detection of remote-sensing images. IEEE Geosci. Remote Sens. Lett. 2020, 18, 33–37. [Google Scholar] [CrossRef]
  19. Shi, P.; Zhao, Z.; Fan, X.; Yan, X.; Yan, W.; Xin, Y. Remote sensing image object detection based on angle classification. IEEE Access 2021, 9, 118696–118707. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Guo, W.; Wu, C.; Li, W.; Tao, R. FANet: An arbitrary direction remote sensing object detection network based on feature fusion and angle classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5608811. [Google Scholar] [CrossRef]
  21. Ding, L.; Bruzzone, L. DiResNet: Direction-aware residual network for road extraction in VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 10243–10254. [Google Scholar] [CrossRef]
  22. Chen, Y.; Wang, Z.; Xiong, Z.; Zhang, Y.; Xu, X. SOAM Block: A scale–orientation-aware module for efficient object detection in remote sensing imagery. Symmetry 2025, 17, 1251. [Google Scholar] [CrossRef]
  23. Li, Y.; Chen, R.; Zhang, Y.; Zhang, M.; Chen, L. Multi-label remote sensing image scene classification by combining a convolutional neural network and a graph neural network. Remote Sens. 2020, 12, 4003. [Google Scholar] [CrossRef]
  24. Li, Q.; Chen, Y.; Zeng, Y. Transformer with transfer CNN for remote-sensing-image object detection. Remote Sens. 2022, 14, 984. [Google Scholar] [CrossRef]
  25. Zhang, C.; Su, J.; Ju, Y.; Lam, K.M.; Wang, Q. Efficient inductive vision transformer for oriented object detection in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5616320. [Google Scholar] [CrossRef]
  26. Liu, B.; Xu, C.; Cui, Z.; Yang, J. Progressive context-dependent inference for object detection in remote sensing imagery. IEEE Trans. Image Process. 2022, 32, 580–590. [Google Scholar] [CrossRef]
  27. Cong, R.; Zhang, Y.; Fang, L.; Li, J.; Zhao, Y.; Kwong, S. RRNet: Relational reasoning network with parallel multiscale attention for salient object detection in optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5613311. [Google Scholar] [CrossRef]
  28. Feng, X.; Han, J.; Yao, X.; Cheng, G. TCANet: Triple context-aware network for weakly supervised object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 6946–6955. [Google Scholar] [CrossRef]
  29. Zhu, Z.; Sun, X.; Diao, W.; Chen, K.; Xu, G.; Fu, K. Invariant structure representation for remote sensing object detection based on graph modeling. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625217. [Google Scholar] [CrossRef]
  30. Liu, N.; Celik, T.; Zhao, T.; Zhang, C.; Li, H.C. AFDet: Toward more accurate and faster object detection in remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12557–12568. [Google Scholar] [CrossRef]
  31. Zhang, X.; Tan, X.; Chen, G.; Zhu, K.; Liao, P.; Wang, T. Object-based classification framework of remote sensing images with graph convolutional networks. IEEE Geosci. Remote Sens. Lett. 2021, 19, 8010905. [Google Scholar] [CrossRef]
  32. Chen, B.; Gao, Z.; Li, Z.; Liu, S.; Hu, A.; Song, W.; Zhang, Y.; Wang, Q. Hierarchical GNN framework for earth’s surface anomaly detection in single satellite imagery. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5627314. [Google Scholar] [CrossRef]
  33. Amrullah, C.; Panangian, D.; Bittner, K. PolyRoof: Precision roof polygonization in urban residential building with graph neural networks. arXiv 2025, arXiv:2503.10913. [Google Scholar] [CrossRef]
  34. Wang, T.; Wang, G.; Tan, K.E. Holistically-nested structure-aware graph neural network for road extraction. In Proceedings of the International Symposium on Visual Computing, Virtual Event, 4–6 October 2021; pp. 144–156. [Google Scholar]
  35. Cui, Y.; Hou, B.; Wu, Q.; Ren, B.; Wang, S.; Jiao, L. Remote sensing object tracking with deep reinforcement learning under occlusion. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5605213. [Google Scholar] [CrossRef]
  36. Uzkent, B.; Yeh, C.; Ermon, S. Efficient object detection in large images using deep reinforcement learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 1824–1833. [Google Scholar]
  37. Shen, X.; Liu, B.; Zhou, Y.; Zhao, J. Remote sensing image caption generation via transformer and reinforcement learning. Multimed. Tools Appl. 2020, 79, 26661–26682. [Google Scholar] [CrossRef]
  38. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate object localization in remote sensing images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  39. Qian, X.; Lin, S.; Cheng, G.; Yao, X.; Ren, H.; Wang, W. Object detection in remote sensing images based on improved bounding box regression and multi-level features fusion. Remote Sens. 2020, 12, 143. [Google Scholar] [CrossRef]
  40. Cheng, G.; Zhou, P.; Han, J. Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  41. Ma, W.; Guo, Q.; Wu, Y.; Zhao, W.; Zhang, X.; Jiao, L. A novel multi-model decision fusion network for object detection in remote sensing images. Remote Sens. 2019, 11, 737. [Google Scholar] [CrossRef]
  42. Prashant, M.; Easwaran, A.; Das, S.; Yuhas, M. Guaranteeing out-of-distribution detection in deep RL via transition estimation. arXiv 2025, arXiv:2503.05238. [Google Scholar] [CrossRef]
  43. Karimzadeh, M.; Esposito, A.; Zhao, Z.; Braun, T.; Sargento, S. RL-CNN: Reinforcement learning-designed convolutional neural network for urban traffic flow estimation. In Proceedings of the 2021 International Wireless Communications and Mobile Computing (IWCMC), Harbin, China, 28 June–2 July 2021; pp. 29–34. [Google Scholar]
  44. Fu, K.; Li, Y.; Sun, H.; Yang, X.; Xu, G.; Li, Y.; Sun, X. A ship rotation detection model in remote sensing images based on feature fusion pyramid network and deep reinforcement learning. Remote Sens. 2018, 10, 1922. [Google Scholar] [CrossRef]
  45. Paletta, L.; Rome, E. Reinforcement learning of object detection strategies. In Proceedings of the 8th International Symposium on Intelligent Robotic Systems (SIRS), Reading, UK, 18–20 July 2000. [Google Scholar]
  46. Zhao, D.; Ma, Y.; Jiang, Z.; Shi, Z. Multiresolution airport detection via hierarchical reinforcement learning saliency model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2855–2866. [Google Scholar] [CrossRef]
  47. Hu, Z.; Gao, K.; Zhang, X.; Wang, J.; Wang, H.; Yang, Z.; Li, C.; Li, W. EMO2-DETR: Efficient-matching oriented object detection with transformers. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5616814. [Google Scholar] [CrossRef]
  48. Wang, J.; Yang, W.; Li, H.C.; Zhang, H.; Xia, G.S. Learning center probability map for detecting objects in aerial images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4307–4323. [Google Scholar] [CrossRef]
  49. Dai, L.; Liu, H.; Tang, H.; Wu, Z.; Song, P. AO2-DETR: Arbitrary-oriented object detection transformer. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 2342–2356. [Google Scholar] [CrossRef]
  50. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards more robust detection for small, cluttered and rotated objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8232–8241. [Google Scholar]
  51. Yang, X.; Yan, J.; Feng, Z.; He, T. R3Det: Refined single-stage detector with feature refinement for rotating object. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3163–3171. [Google Scholar] [CrossRef]
  52. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  53. Yujie, L.; Xiaorui, S.; Wenbin, S.; Yafu, Y. S2ANet: Combining local spectral and spatial point grouping for point cloud processing. Virtual Real. Intell. Hardw. 2024, 6, 267–279. [Google Scholar] [CrossRef]
  54. Chen, M.; Xu, K.; Chen, E.; Zhang, Y.; Xie, Y.; Hu, Y.; Pan, Z. Semantic attention and structured model for weakly supervised instance segmentation in optical and SAR remote sensing imagery. Remote Sens. 2023, 15, 5201. [Google Scholar] [CrossRef]
  55. Liu, N.; Han, J. DHSNet: Deep hierarchical saliency network for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 678–686. [Google Scholar]
  56. Bharati, P.; Pramanik, A. Deep learning techniques—R-CNN to Mask R-CNN: A survey. In Computational Intelligence in Pattern Recognition: Proceedings of CIPR 2019; Springer: Singapore, 2019; pp. 657–668. [Google Scholar]
  57. Han, J.; Ding, J.; Xue, N.; Xia, G.S. ReDet: A rotation-equivariant detector for aerial object detection. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2786–2795. [Google Scholar]
  58. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with Gaussian Wasserstein distance loss. In Proceedings of the International Conference on Machine Learning (ICML), PMLR, Virtual, 18–24 July 2021; pp. 11830–11841. [Google Scholar]
  59. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning high-precision bounding box for rotated object detection via Kullback–Leibler divergence. Adv. Neural Inf. Process. Syst. 2021, 34, 18381–18394. [Google Scholar]
  60. Li, S.; Yan, F.; Liu, Y.; Shen, Y.; Liu, L.; Wang, K. A multi-scale rotated ship targets detection network for remote sensing images in complex scenarios. Sci. Rep. 2025, 15, 2510. [Google Scholar] [CrossRef]
  61. Li, Y.; Li, X.; Dai, Y.; Hou, Q.; Liu, L.; Liu, Y.; Cheng, M.M.; Yang, J. LSKNet: A foundation lightweight backbone for remote sensing. Int. J. Comput. Vis. 2025, 133, 1410–1431. [Google Scholar] [CrossRef]
  62. Cai, X.; Lai, Q.; Wang, Y.; Wang, W.; Sun, Z.; Yao, Y. Poly kernel inception network for remote sensing detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 27706–27716. [Google Scholar]
  63. Wei, H.; Zhang, Y.; Chang, Z.; Li, H.; Wang, H.; Sun, X. Oriented objects as pairs of middle lines. ISPRS J. Photogramm. Remote Sens. 2020, 169, 268–279. [Google Scholar] [CrossRef]
  64. Ming, Q.; Miao, L.; Zhou, Z.; Dong, Y. CFC-Net: A critical feature capturing network for arbitrary-oriented object detection in remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5605814. [Google Scholar] [CrossRef]
  65. Liu, S.; Zhang, L.; Lu, H.; He, Y. Center-boundary dual attention for oriented object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5603914. [Google Scholar] [CrossRef]
  66. Zhang, C.; Lam, K.M.; Wang, Q. CoF-Net: A progressive coarse-to-fine framework for object detection in remote-sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5600617. [Google Scholar] [CrossRef]
  67. Wang, D.; Zhang, J.; Xu, M.; Liu, L.; Wang, D.; Gao, E.; Han, C.; Guo, H.; Du, B.; Tao, D.; et al. MTP: Advancing remote sensing foundation model via multitask pretraining. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11632–11654. [Google Scholar] [CrossRef]
  68. Li, X.; Chen, L.; Wang, D.; Yang, H. Detection of ship targets in remote sensing image based on improved YOLOv5. In Proceedings of the International Conference on Electronic Information Engineering and Computer Science (EIECS), Changchun, China, 16–18 September 2022; SPIE: Bellingham, WA, USA, 2023; Volume 12602, pp. 308–314. [Google Scholar]
  69. Chen, L.; Luo, C.; Li, X.; Xiao, J. Rotating target detection algorithm in remote sensing images based on improved YOLOv5s. In Proceedings of the 2023 4th International Conference on Computer Vision, Image and Deep Learning (CVIDL), Zhuhai, China, 12–14 May 2023; pp. 180–184. [Google Scholar]
  70. Bao, M.; Chala Urgessa, G.; Xing, M.; Han, L.; Chen, R. Toward more robust and real-time unmanned aerial vehicle detection and tracking via cross-scale feature aggregation based on the center keypoint. Remote Sens. 2021, 13, 1416. [Google Scholar] [CrossRef]
  71. Ma, C.; Yin, H.; Weng, L.; Xia, M.; Lin, H. DAFNet: A novel change-detection model for high-resolution remote-sensing imagery based on feature difference and attention mechanism. Remote Sens. 2023, 15, 3896. [Google Scholar] [CrossRef]
  72. Cheng, G.; Yao, Y.; Li, S.; Li, K.; Xie, X.; Wang, J.; Yao, X.; Han, J. Dual-aligned oriented detector. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618111. [Google Scholar] [CrossRef]
  73. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625411. [Google Scholar] [CrossRef]
  74. Deng, L.; Tan, Y.; Zhao, D.; Liu, S. Research on object detection in remote sensing images based on improved horizontal target detection algorithm. Earth Sci. Inform. 2025, 18, 304. [Google Scholar] [CrossRef]
  75. Xiang, H.; Jing, N.; Jiang, J.; Guo, H.; Sheng, W.; Mao, Z.; Wang, Q. RTMDet-R2: An improved real-time rotated object detector. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Xiamen, China, 13–15 October 2023; pp. 352–364. [Google Scholar]
  76. Wang, K.; Wang, Z.; Li, Z.; Su, A.; Teng, X.; Pan, E.; Liu, M.; Yu, Q. Oriented object detection in optical remote sensing images using deep learning: A survey. Artif. Intell. Rev. 2025, 58, 350. [Google Scholar] [CrossRef]
  77. Wang, Z.; Wan, S.; Ma, X. Remote sensing image dense target detection based on rotating frame. J. Phys. Conf. Ser. 2021, 2006, 012049. [Google Scholar] [CrossRef]
  78. Zhang, G.; Lu, S.; Zhang, W. CAD-Net: A context-aware detection network for objects in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10015–10024. [Google Scholar] [CrossRef]
  79. Varotto, L.; Cenedese, A.; Cavallaro, A. Active sensing for search and tracking: A review. arXiv 2021, arXiv:2112.02381. [Google Scholar] [CrossRef]
  80. Chen, H.; Wang, L.; Zhang, L.; Li, Y.; Wu, Y.; Qi, J. Research on remote sensing image target detection methods based on convolutional neural networks. J. Phys. Conf. Ser. 2021, 2025, 012068. [Google Scholar] [CrossRef]
Figure 1. Architectureof ASBPNet and its core components. (a) The structure of ASBPNet. (b) The structure of Angle-Synchronized Graph Head (ASG). (c) The structure of Bilevel Policy Optimization (BPO).
Figure 1. Architectureof ASBPNet and its core components. (a) The structure of ASBPNet. (b) The structure of Angle-Synchronized Graph Head (ASG). (c) The structure of Bilevel Policy Optimization (BPO).
Remotesensing 17 04029 g001
Figure 2. Visual contrast of feature attribution. (a) Drone RGB input image. (b) Baseline method heatmap. (c) Our method’s heatmap.
Figure 2. Visual contrast of feature attribution. (a) Drone RGB input image. (b) Baseline method heatmap. (c) Our method’s heatmap.
Remotesensing 17 04029 g002
Figure 3. Comparison with SOTA methods on DIOR-R datasets.
Figure 3. Comparison with SOTA methods on DIOR-R datasets.
Remotesensing 17 04029 g003
Figure 4. Comparison of YOLOV5m and ASBPNet on DOTAv1.0 dataset.
Figure 4. Comparison of YOLOV5m and ASBPNet on DOTAv1.0 dataset.
Remotesensing 17 04029 g004
Figure 5. Ablation study of ASG and BPO on YOLOv5m-OBB.
Figure 5. Ablation study of ASG and BPO on YOLOv5m-OBB.
Remotesensing 17 04029 g005
Figure 6. Variants with different message-passing, alignment, and synchronization strategies.
Figure 6. Variants with different message-passing, alignment, and synchronization strategies.
Remotesensing 17 04029 g006
Figure 7. Variants on neighborhood size, message-passing depth, Top-N selection, and feature hierarchy.
Figure 7. Variants on neighborhood size, message-passing depth, Top-N selection, and feature hierarchy.
Remotesensing 17 04029 g007
Figure 8. B-3 Variants on loss weighting l a m b d a t e x t s y n c and step-size e t a .
Figure 8. B-3 Variants on loss weighting l a m b d a t e x t s y n c and step-size e t a .
Remotesensing 17 04029 g008
Figure 9. Comparison of key difficult categories (multi-scale).
Figure 9. Comparison of key difficult categories (multi-scale).
Remotesensing 17 04029 g009
Figure 10. Comparison between baseline and our method under conventional scenarios.
Figure 10. Comparison between baseline and our method under conventional scenarios.
Remotesensing 17 04029 g010
Figure 11. Comparison of baseline and our method in complex scenarios.
Figure 11. Comparison of baseline and our method in complex scenarios.
Remotesensing 17 04029 g011
Table 1. Summary of related methods.
Table 1. Summary of related methods.
Method CategoryCore IdeaRepresentative ModelsAdvantagesLimitations
GNN Methods
Relational InferenceModel contextual dependencies by constructing topological graphs between objects.RelaDet; GCN-based DetectorStrong for dense and occluded objects; enhanced contextual reasoning.Requires predefined graph structures; potential over-smoothing.
Context-Aware ModelingCapture multi-scale contextual cues using graph convolution or attention.Contextual Graph Detector; Hierarchical GNNRobust under complex backgrounds; suitable for large-scale RS scenes.Lacks rotational invariance; may introduce angular inconsistency.
Hierarchical Graph FusionFuse local and global information using multi-level graph structures.PolyRoof; HNS-GNNImproved multi-scale representation.Cross-level noise propagation; insufficient refinement.
RL Methods
Region SearchModel detection as a sequential decision process progressively focusing on candidate regions.DeepRL-Detector; RLCNNEfficient localization; reduced HR image computation.Fixed search grids; action conflicts in dense scenes.
Rotation and Scale AdjustmentLearn actions that dynamically adjust bounding box orientation and scale.FFPN-RLEffective for multi-oriented, multi-scale objects.Discrete actions cause error accumulation; slow convergence.
Policy Fusion and OptimizationFuse multiple decision strategies to enhance adaptability and robustness.Strategy Learning; HRL Saliency ModelStrong adaptability across scenes.Depends on predefined modules; hierarchical feedback errors.
Table 2. Comparison with SOTA methods on DIOR-R datasets.
Table 2. Comparison with SOTA methods on DIOR-R datasets.
MethodPre-Training#PFLOPsFPSmAP (%)
RetinaNet-OIN23.057.55
Faster RCNN-OIN41.1 M198 G19.059.54
TIOE-DetIN41.1 M198 G61.98
ARS-DETRIN41.1 M198 G1266.12
O-RepPointsIN36.6 M2966.71
DCFLIN2966.80
LSKNet-SIN31.0 M161 G65.90
PKINet-SIN30.8 M190 G5.267.03
YOLOv5m-OBB (baseline)IN21.2 M153 G62.363.40
ASBPNetCO22.1 M158 G61.168.10
The red values represent the best (optimal) performance, while the blue values indicate the second-best (suboptimal) performance.
Table 3. Comparison with SOTA methods on HRSC2016 datasets.
Table 3. Comparison with SOTA methods on HRSC2016 datasets.
MethodPre-TrainingmAP (07)mAP (12)#PFLOPsFPS
DRNIN92.70
CenterMapIN92.8041.1 M198 G
RoI Trans.IN86.2055.1 M200G6.0
G.V.IN88.2041.1 M198 G
R3DetIN89.2696.0141.9 M336 G12.0
DALIN89.7736.4 M216 G
GWDIN89.8597.3747.4 M456 G
S2ANetIN90.1795.0138.6 M198 G14.3
AOPGIN90.3496.22
ReDetIN90.4697.6331.6 M24.7
O-RCNNIN90.5097.6041.1 M199 G32.0
RTMDetCO90.6097.1052.3 M205 G
YOLOv5m-OBB (baseline)IN88.5096.8021.2 M153 G62.6
ASBPNetCO90.6098.2022.1 M158 G61.7
The red values represent the best (optimal) performance, while the blue values indicate the second-best (suboptimal) performance.
Table 4. Per-class mAP comparison with mAP moved before class columns.
Table 4. Per-class mAP comparison with mAP moved before class columns.
MethodmAP↑PLBDBRGTFSVLVSHTCBCSTSBFRAHASPHC
Single-Scale
EMO2-DETR [47]70.9187.9979.4645.7466.6478.9073.9073.3090.4080.5585.8955.1963.6251.8370.1560.04
CenterMap [48]71.5989.0280.5649.4161.9877.9974.1983.7489.4478.0183.5247.6465.9363.6867.0761.59
AO2-DETR [49]72.1586.0175.9246.0266.6579.7079.9389.1790.4481.1976.0056.9162.4564.2265.8058.96
SCRDet [50]72.6189.9880.6552.0968.3668.3660.3272.4190.8587.9486.8665.0266.6866.2568.2465.21
R3Det [51]73.7089.5081.2050.5066.1070.9078.7078.2090.8085.3084.2061.8063.8068.2069.8067.20
RoI Trans. [52]74.0589.0177.4851.6472.0774.4377.5587.7690.8179.7185.2758.3664.1176.5071.9954.06
S2ANet [53]74.1289.1182.8448.3771.1178.1178.3987.2590.8384.9085.6460.3662.6065.2669.1357.94
SASM [54]74.9286.4278.9752.4769.8477.3075.9986.7290.8982.6385.6660.1368.2573.9872.2262.37
G.V. [55]75.0289.6485.0052.2677.3473.0173.1486.8290.7479.0286.8159.5570.9172.9470.8657.32
O-RCNN [56]75.8789.4682.1254.7870.8678.9383.0088.2090.9087.5084.6863.9767.6974.6484.9352.28
ReDet [57]76.2588.7982.6453.9774.0078.1384.0688.0490.8987.7885.7561.7660.3975.9668.0763.59
R3Det-GWD [58]76.3488.8282.9455.6372.7578.5283.1087.4690.2186.3685.4464.7061.4173.4676.9457.38
R3Det-KLD [59]77.3688.9084.1755.8069.3578.7284.0887.0089.7584.3285.7364.7461.8076.6278.4970.89
ARC [60]77.3589.4082.4855.3373.8879.3784.0588.0690.9086.4484.8363.6370.3274.2971.9165.43
LSKNet-S [61]77.4989.6685.5257.7275.7074.9578.6988.2490.8886.7986.3866.9263.7777.7774.4764.82
PKINet-S [62]78.3989.7284.2055.8177.6380.2584.4588.1290.8887.5786.0766.8670.2377.4773.6262.94
O2DNet [63]71.0189.3682.1847.3161.2471.3374.0278.6490.8182.2381.4260.9060.2258.2167.0261.04
CFC-Net [64]73.5289.1380.4452.4170.0476.3178.1487.2390.9284.5485.6160.5161.5367.8268.0150.15
CBDA-Net [65]75.7489.2485.9450.3265.0177.7182.3387.9390.5286.5185.9066.9266.5167.4371.3262.91
CoF-Net [66]77.2189.6083.1248.3173.6478.2383.0486.7290.2482.3286.6167.6164.6374.7071.3278.42
MTP [67]79.0389.8785.0958.2771.7081.7087.1088.9891.4485.4186.4557.4468.4778.4282.9771.71
YOLOv5m-OBB (baseline) [68]77.3089.9081.5058.8075.5076.9079.3081.3089.7083.0087.1067.1072.0079.5077.0060.90
ASBPNet79.6091.5083.0064.4077.7080.8081.3083.1090.9084.6089.1069.1073.9081.0078.7064.90
Multi-Scale
CSL [69]76.1790.2585.5354.6475.3170.4473.5177.6290.8486.1586.6969.6068.0473.8371.1068.93
CFA [70]76.6789.0883.2054.3766.8781.2380.9687.1790.2184.3286.0952.3469.9475.5280.7667.96
DAFNeT [71]76.9589.4086.2753.7060.5182.0481.1788.6690.3783.8187.2753.9369.3875.6181.2670.86
DODet [72]80.6289.9685.5258.0181.2278.7185.4688.5990.8987.1287.8070.5071.5482.0677.4374.47
AOPG [73]80.6689.8885.5760.9081.5178.7085.2988.8590.8987.6087.6571.6668.6982.3177.3273.10
KFIoU [74]80.9389.4484.4162.2282.5180.1086.0788.6890.9087.3288.3872.8071.9578.9674.9575.27
RTMDet-RKFIoU [75]81.3388.0186.1758.5482.4481.3084.8288.7190.8988.7787.3771.9671.1881.2381.4077.13
Multi-Scale
RVSA [76]81.2488.9785.7661.4681.2779.9885.3188.3090.8485.0687.5066.7773.1184.7581.8877.58
LSKNet-S [61]81.6489.5786.3463.1383.6782.2086.1088.6690.8988.4187.4271.7269.5878.8881.7776.52
FR-O [77]54.1179.4277.1317.7264.1435.3238.0137.2689.4369.6259.3450.3052.9347.9147.4346.31
CAD-Net [78]69.9687.8282.4149.4373.5171.1363.5276.6290.9179.2373.3348.4460.9262.0367.0162.24
APE [79]75.8190.0183.6553.4376.0274.0177.2779.5390.8287.2784.5167.7260.3174.6071.8465.62
SCRDet [80]72.6390.0180.7152.1468.4368.4260.3272.4490.9287.9486.9665.0266.7366.3168.2365.21
YOLOv5m-OBB (baseline)80.2092.5084.0061.5078.5080.5082.0084.5093.0086.0090.5070.0074.0082.0079.0065.00
ASBPNet81.5093.4084.9064.1079.9082.3083.2085.6093.3086.8091.5071.8075.3082.9079.9067.60
The red values represent the best (optimal) performance, while the blue values indicate the second-best (suboptimal) performance.
Table 5. Ablation study of ASG and BPO on YOLOv5m-OBB.
Table 5. Ablation study of ASG and BPO on YOLOv5m-OBB.
SettingDescriptionmAPΔmAP#ParamsFLOPs (G)
AYOLOv5m-OBB80.2021.2 M153
B+ASG81.00+0.8021.9 M156
C+BPO80.90+0.7021.2 M153
DASG + BPO (Ours)81.50+1.3022.1 M158
Table 6. B-1: Variants with different message-passing, alignment, and synchronization strategies.
Table 6. B-1: Variants with different message-passing, alignment, and synchronization strategies.
VariantDescriptionmAP
B (default)GATv2 message + SE(2) equivariant alignment + Angle Sync81.00
B-msg-sageGraphSAGE message passing80.86
B-msg-edgeEdgeConv-style message passing80.90
B-sync-offDisable Angle Sync regularization80.78
B-sync-pcaSync prior: local PCA principal direction80.98
B-sync-houghSync prior: Hough line direction81.02
B-enc-rawθAngle encoding: raw θ (no sin/cos)80.60
B-enc-vmAngle encoding: sin/cos + von Mises embedding81.05
Table 7. B-2: Variants on neighborhood size, message-passing depth, Top-N selection, and feature hierarchy.
Table 7. B-2: Variants on neighborhood size, message-passing depth, Top-N selection, and feature hierarchy.
VariantConfigurationmAPΔmAP
k-4kNN neighbors k = 4 80.82−0.18
k-8 (default) k = 8 81.00
k-12 k = 12 81.04+0.04
k-16 k = 16 80.95−0.05
depth-2Message layers = 280.89−0.11
depth-3 (default)Message layers = 381.00
depth-4Message layers = 481.03+0.03
TopN-200Node number N = 200 80.90−0.10
TopN-300 (default) N = 300 81.00
TopN-400 N = 400 81.01+0.01
P3-onlyP3 RoI only80.70−0.30
P3–P5 (default)Fusion of P3–P581.00
P3–P6Fusion of P3–P681.02+0.02
Table 8. B-3: Variants on loss weighting λ sync and step-size η .
Table 8. B-3: Variants on loss weighting λ sync and step-size η .
VariantConfigurationmAPΔmAP
λ sync = 0 Remove synchronization term80.78−0.22
λ sync = 0.5 Sync weight = 0.580.94−0.06
λ sync = 1 . 0 (default)Default sync weight81.00
λ sync = 2.0 Sync weight = 2.081.02+0.02
λ sync = 4.0 Sync weight = 4.080.88−0.12
η -fixedFixed step size η = 1.0 80.96−0.04
η -learn (default)Lightweight MLP predicts η ( 0 , 1 ] 81.00
η -linesearchLine-search single-step backtracking81.06+0.06
Table 9. B-3: RL refinement ablation on top of setting C.
Table 9. B-3: RL refinement ablation on top of setting C.
VariantAction SpacemAPLatency Change
C1 R-Aug onlyLearn rotation augmentation angle distribution80.560%
C2 DNMS onlyLearn rNMS threshold (class/scale adaptive)80.480%
C3 AMTS onlyLearn slice size/stride/overlap/rotation80.42+4%
C4 DPA onlyLearn rIoU positive/negative thresholds80.320%
C (default)R-Aug + DNMS + AMTS (with budget)80.90+2%
Table 10. C-1: DNMS ablation with scale-specific rNMS IoU thresholds.
Table 10. C-1: DNMS ablation with scale-specific rNMS IoU thresholds.
SmallMediumLargemAPΔmAPLatency
0.350.450.5580.72−0.180%
0.400.500.55 (default)80.90+2%
0.450.550.6081.00+0.10+2%
0.500.600.6580.84−0.06+2%
Table 11. C-2: AMTS ablation with different tile/stride/overlap/rotation settings.
Table 11. C-2: AMTS ablation with different tile/stride/overlap/rotation settings.
Tile/Stride/Overlap/RotationmAPΔmAPLatency
768/256/33%/ON80.96+0.06+6%
1024/384/25%/ON (default)80.90+2%
1024/384/25%/OFF80.85−0.05+1%
1280/512/20%/ON80.88−0.02−3%
Table 12. C-3: R-Aug ablation with different rotation augmentation distributions.
Table 12. C-3: R-Aug ablation with different rotation augmentation distributions.
DistributionAngle/Probability SchememAPΔmAP
Uniform U [ 0 , 180 ] 80.82−0.08
BimodalHigh weights at ± 15 / ± 30 81.02+0.12
Tri-modal (default)Mix of 0 , ± 30 , and uniform80.90
Table 13. C-4: DPA ablation with different positive/negative rIoU thresholds.
Table 13. C-4: DPA ablation with different positive/negative rIoU thresholds.
(IoUpos, IoUneg)mAPΔmAP
(0.4, 0.2)80.60−0.30
(0.5, 0.3) (default)80.90
(0.6, 0.3)80.96+0.06
(0.6, 0.4)80.88−0.02
Table 14. C-5: Sensitivity analysis of training algorithm, reward weights, and latency budget.
Table 14. C-5: Sensitivity analysis of training algorithm, reward weights, and latency budget.
HyperparameterSettingmAPLatency
AlgorithmPPO (default)80.90+2%
SAC80.88+2%
Budget B10.0 ms80.70≤10 ms
12.5 ms (default)80.90≤12.5 ms
15.0 ms81.05≤15 ms
Small-object weight β 0.080.78+1%
0.2 (default)80.90+2%
0.581.00+3%
AP surrogate ϵ 0.0580.82+2%
0.10 (default)80.90+2%
0.2080.84+2%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, J.; Liu, J.; Tang, L.; Wang, X.; Guo, Y. From Graph Synchronization to Policy Learning: Angle-Synchronized Graph and Bilevel Policy Network for Remote Sensing Object Detection. Remote Sens. 2025, 17, 4029. https://doi.org/10.3390/rs17244029

AMA Style

Yan J, Liu J, Tang L, Wang X, Guo Y. From Graph Synchronization to Policy Learning: Angle-Synchronized Graph and Bilevel Policy Network for Remote Sensing Object Detection. Remote Sensing. 2025; 17(24):4029. https://doi.org/10.3390/rs17244029

Chicago/Turabian Style

Yan, Jie, Jialang Liu, Lixing Tang, Xiaoxiang Wang, and Yanming Guo. 2025. "From Graph Synchronization to Policy Learning: Angle-Synchronized Graph and Bilevel Policy Network for Remote Sensing Object Detection" Remote Sensing 17, no. 24: 4029. https://doi.org/10.3390/rs17244029

APA Style

Yan, J., Liu, J., Tang, L., Wang, X., & Guo, Y. (2025). From Graph Synchronization to Policy Learning: Angle-Synchronized Graph and Bilevel Policy Network for Remote Sensing Object Detection. Remote Sensing, 17(24), 4029. https://doi.org/10.3390/rs17244029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop