Next Article in Journal
Design of a High-Gain Multi-Input LNA with 16.4 Degree Phase Shift Within the 32 dB Gain Range
Previous Article in Journal
Modeling and Parameter Analysis of Basic Single Channel Neuron Mass Model for SSVEP
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mission Sequence Model and Deep Reinforcement Learning-Based Replanning Method for Multi-Satellite Observation

School of Aeronautics and Astronautics, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(6), 1707; https://doi.org/10.3390/s25061707
Submission received: 2 February 2025 / Revised: 28 February 2025 / Accepted: 6 March 2025 / Published: 10 March 2025
(This article belongs to the Section Remote Sensors)

Abstract

:
With the rapid increase in the number of Earth Observation Satellites (EOSs), research on autonomous mission scheduling has become increasingly critical for optimizing satellite sensor operations. While most existing studies focus on static environments or initial planning states, few address the challenge of dynamic request replanning for real-time sensor management. In this paper, we tackle the problem of multi-satellite rapid mission replanning under dynamic batch-arrival observation requests. The objective is to maximize overall observation revenue while minimizing disruptions to the original scheme. We propose a framework that integrates stochastic master-satellite mission allocation with single-satellite replanning, supported by reactive scheduling policies trained via deep reinforcement learning. Our approach leverages mission sequence modeling with attention mechanisms and time-attitude-aware rotary positional encoding to guide replanning. Additionally, scalable embeddings are employed to handle varying volumes of dynamic requests. The mission allocation phase efficiently generates assignment solutions using a pointer network, while the replanning phase introduces a hybrid action space for direct task insertion. Both phases are formulated as Markov Decision Processes (MDPs) and optimized using the PPO algorithm. Extensive simulations demonstrate that our method significantly outperforms state-of-the-art approaches, achieving a 15.27% higher request insertion revenue rate and a 3.05% improvement in overall mission revenue rate, while maintaining a 1.17% lower modification rate and achieving faster computational speeds. This demonstrates the effectiveness of our approach in real-world satellite sensor applications.

1. Introduction

The past few decades have seen a rapid increase in Earth Observation Satellites (EOSs) in orbit. With advancements in on-board computation capabilities, communication systems, and attitude maneuvering abilities, EOSs have become increasingly important in various societal fields, such as environmental monitoring, public safety, and disaster prevention and response [1,2].
The essence of satellite observation scheduling lies in matching tasks to satellites and determining the precise observation time for each task [3]. Most existing studies focus on static scheduling scenarios, where all observation tasks are assumed to be predefined at once, and the scheduling results remain fixed throughout the process [4]. The diverse demands for Earth observation and the complexity of observation tasks have led to the need for efficient satellite scheduling [5]. Most current studies focus on static scheduling scenarios, where all observation tasks are assumed to be deterministic and remain fixed throughout the scheduling process. However, in real-world satellite mission planning scenarios, there are both internal uncertainties (e.g., partial satellite equipment failures) and external uncertainties (e.g., changes in observation opportunities or user demand), which make the planning process highly dynamic [6]. These factors lead to the episodic and unpredictable nature of observation requests, highlighting the need for continuous adjustment and adaptation of mission replanning in response to unexpected emergencies [7].
Dynamic requests are typically delivered in batches with uncertain scales, requiring coordinated efforts among multiple satellites to meet rescheduling demands. In urgent scenarios, an efficient planning mechanism is essential to reduce problem complexity and alleviate the computational burden on satellites [8]. For dynamic scheduling in emergency tasks, many studies suggest that it is important to insert new tasks while minimizing changes to the original plan, which can benefit subsequent routine operations [9]. Minimizing disturbances to the original planning scheme is significant for three main reasons: First, it ensures the stability and continuity of the satellite observation system [10], which is especially important when task arrivals become regular. Without this consideration, the planning scheme may require frequent revisions [11]. Second, for periodic observation targets, such as tracking or astronomical observation tasks, replanning may undermine the beneficial features of the original scheme. Tasks from the original plan could be removed, making it difficult to stably meet user requirements and leading to resource wastage [12]. Lastly, due to the characteristics of Agile Earth Observation Satellites (AEOSs), the execution time of observation tasks is closely linked to imaging quality. Even minor perturbations to the original plan could affect the actual observation outcomes [8].
The operations of AEOSs inherently involve time-dependent mission sequences due to attitude transitions and the associated resource consumption between observations. To improve the quality of replanning solutions, it is essential to ensure maneuver continuity and efficiency [13]. This requires accurate modeling and real-time awareness of satellite scheme states, effectively integrated with rapid task replanning.
With the rise of onboard computing capabilities, autonomous satellite scheduling is becoming a key trend, enabling rapid responses to dynamic requests while reducing reliance on ground control [14]. However, existing static scheduling algorithms often suffer from high computational overhead and poor performance when handling batch arriving requests. Recently, deep reinforcement learning (DRL) has been widely adopted for solving combinatorial optimization problems [15], demonstrating advantages in fast inference and strong generalization. We believe that well-trained, end-to-end inference models are well suited to address the challenge of efficiently responding to large-scale observation requests.
Therefore, the main goal of this study is to investigate the mission replanning problem for multiple AEOS on dynamic arrival of observation requests. We propose a mission sequence-based replanning (MSRP) algorithm tailored for multi-satellite systems. MSRP introduces a temporary master–slave architecture for task allocation and single-satellite-level replanning. By a customized attention mechanism, the algorithm uses time-attitude information to enhance the modeling of sequence dependencies. Mission allocation and replanning decisions are made sequentially in two stages based on satellite states, with both phases jointly optimized to ensure high-quality strategy generation.
The contributions of this paper are as follows.
  • To the best of our knowledge, this is the first work to directly model satellite observation sequences using deep learning, incorporating time and attitude information. We utilize task timing and side-swing angles as positional encodings for self-attention modeling and introduce a gated global pooling method based on interval information to generate a comprehensive satellite state representation. Ablation studies and comparisons with classical methods demonstrate that our approach significantly improves mission replanning performance.
  • We propose a DRL-based mission replanning algorithm for a single satellite, designed to handle discrete-continuous hybrid decision variables. This algorithm achieves dual-objective optimization by maximizing task revenues and minimizing changes to the original plan. Experimental results validate the effectiveness of the algorithm, demonstrating its generalization ability across different-scale scenarios.
  • We propose an on-site mission allocation algorithm for request-receiving satellites, integrating cross-attention mechanisms and pointer networks to provide end-to-end allocation solutions. Combined with the replanning method, this two-stage optimization approach demonstrates the advantages of our algorithm over classical methods.
The rest of this article is organized as follows. Section 2 provides a review of previous research on satellite task replanning techniques. Section 3 presents our solution framework in the scenario, along with the optimization objectives and constraints of the problem. Section 4 introduces the mission sequence modeling and dynamic request modeling approach, along with the DRL-based mission allocation and replanning algorithms. In Section 5, a series of experiments and analyses are conducted to verify the effectiveness of the proposed methods. Finally, Section 6 concludes the paper.

2. Literature Review

2.1. Dynamic Scheduling of AEOSs

An increasing amount of research focuses on the dynamic scheduling of Agile Earth Observation Satellites (AEOSs). Pemberton and Greenwald [16] described the dynamic scheduling of EOS and analyzed the various emergency conditions associated with it. Liang et al. [17] developed a scalable modeling method based on predicate logic and proposed a knowledge-based, priority rule-driven heuristic approach for rapid responses. Wen et al. [18] introduced a multi-scenario fusion-based onboard scheduling and coordination approach, which transforms complex onboard rescheduling problems into a solution selection problem from ground-generated solutions. He et al. [13] proposed a hierarchical scheduling method for real-time scheduling, which dynamically adjusts scheduling precision based on task observation times, thereby improving the overall algorithm efficiency. Han et al. [19] utilized a chance-constrained programming model to describe uncertainty and introduced an improved simulated annealing-based heuristic, along with a fast insertion strategy for large-scale observation tasks. Li et al. [14] proposed a low-complexity onboard periodic rescheduling algorithm based on a rolling horizon optimization strategy, integrating greedy-based allocation algorithms, pointer network-based scheduling algorithms, and local iterative search algorithms to provide high-quality solutions. Liu et al. [20] combined task planning with transient scheduling, proposing a simultaneous planning and execution strategy along with a heuristic rolling schedule to address dynamic scheduling without considering initial plans.
In emergency situations, various task information must be comprehensively considered, including users’ expected completion times, time windows, inter-task intervals, and dynamic environmental factors, adding more complexity to task scheduling [8]. Wei et al. [21] considered the transient transition time between requests and proposed a multi-objective memetic approach that accounts for time-dependent transition times. Peng et al. [22] modeled the dependency of transition times between consecutive observations and developed a greedy random iterative local search method for fast feasibility checks and task allocation. Du et al. [23] employed a task clustering preprocessing method to enhance task scheduling efficiency by combining potential objectives. Du et al. [24] trained a probability prediction model using historical satellite scheduling data to improve task allocation strategies. Song et al. [25] proposed a graph-structured mixed-integer mathematical model, introducing satellite scheduling knowledge to develop a genetic algorithm, which experimentally demonstrated its effectiveness in improving system operation efficiency.
However, in the research on the rescheduling of observation tasks, few studies directly model the original task sequence. These studies often overlook the influence of original task execution time and resource demands on the dynamic insertion of new requests.

2.2. Multi-Satellite Scheduling Architecture

Multi-satellite cooperative autonomous mission planning can typically be decomposed into a top-level multi-satellite task allocation problem and a bottom-level single-satellite task scheduling problem [26,27]. Bianchessi et al. [28] proposed a three-phase scheduling method to reduce the problem’s complexity, which includes request analysis, task allocation, and a distributed optimization approach. Li et al. [29] constructed a dedicated experimental platform for coordinating tasks between geostationary and low Earth orbit satellites and proposed a task allocation algorithm tailored for this platform. Qi et al. [30] designed an evolutionary ant colony optimization method and an interactive rescheduling approach for the task pre-planning and re-planning phases, respectively, to maximize observation profit and achieve load balancing. Liu et al. [31] addressed the single-satellite scheduling problem based on Q-networks and introduced a profit-based competitive strategy to resolve scheduling conflicts in multi-satellite systems, thereby reducing the burden on cooperative communications. Most of the studies above focus on static tasks, with few designs addressing task rescheduling. Du et al. [23] proposed a new multi-dimensional multi-agent collaborative model and a secondary allocation strategy based on contract network protocols (CNPs) to reduce task allocation conflicts and efficiently minimize losses in case of failure by reinserting tasks. Chen et al. [32] introduced a hierarchical task planning framework, where the global planner divides the clusters, and the local planner searches for the optimal solution using Discrete Particle Swarm Optimization (DPSO) algorithms. In emergency situations, task priorities are used to reschedule part of the cluster tasks.
Some multi-agent system (MAS) optimization algorithms have been applied to satellite task allocation. Yang et al. [33] proposed a dynamic distributed architecture using an improved contract net protocol and blackboard model for task coordination but did not address constraints on revising original plans. Li et al. [34] introduced an iterative coordination model based on Nash Equilibrium (NE) theory to solve joint space observation problems but overlooked task allocation under emergency conditions. MAS-based methods often require multiple coordination rounds, limiting their suitability for rapid-response scenarios.
Recent advancements in multi-agent reinforcement learning (MARL) have inspired new approaches to dynamic task allocation. Many studies on dynamic task coordination focus on unmanned aerial vehicles (UAVs). For example, Liu et al. [35] addressed dynamic task allocation for heterogeneous UAVs by proposing a proposer-responder mechanism and implementing a Q-network for response generation. This mechanism effectively models the dynamic process of task request arrival, allocation, and processing.
In satellite observation mission allocation, Wang et al. [9] applied transfer learning to adapt historical strategies into current initial strategies and developed a hybrid MARL algorithm for dynamic task planning. This approach focuses on action learning but lacks knowledge transfer, limiting its effectiveness with heterogeneous agents. Li et al. [36] replaced the contract net protocol with a multi-agent deep deterministic policy gradient (MADDPG) algorithm to solve real-time multi-satellite cooperative observation scheduling, reducing communication overhead. Saeed et al. [37] proposed a multi-agent, multi-target reinforcement learning framework for dynamic task and sensor resource matching but did not address satellite maneuvering constraints. Zhang et al. [38] used the multi-agent proximal policy optimization (MAPPO) algorithm for task scheduling under satellite state constraints. However, it does not adapt to dynamic changes in satellite resources, environmental conditions, or observation tasks. Essentially, it offers a rapid solution for static scenarios.
In summary, most MARL-based satellite mission planning methods focus on solving the initial allocation problem, with limited attention to the replanning problem. Existing approaches rely on fast neural network inference to handle dynamic situations but lack comprehensive modeling of the environment state. Additionally, this paper focuses on task allocation by individual satellites based on received requests, where each satellite directly assigns tasks according to its strategy. The issue of collaborative allocation, as typically explored in MARL research, is not considered in this study.

2.3. General Models and Algorithms

The scheduling problem of multiple Earth Observation Satellites is a complex combinatorial optimization problem and has been proven to be NP-hard [39].
Task planning models for multi-satellite observation missions include integer programming models [40], constraint satisfaction models [41], and graph-based models [42]. The classical solution algorithms include exact algorithms [43,44], heuristic methods [45,46] and meta-heuristics methods [47]. For large-scale problems, constructing an exact model is practically impossible. Heuristic or metaheuristic methods can be used to optimize the search strategy for solving the problem. Li et al. [48] designed a genetic-evolutionary bi-level solution framework, using an improved genetic algorithm to separately solve the task allocation and resource scheduling problems. Yang et al. proposed a hybrid local replanning strategy for multi-satellite imaging mission planning in uncertain environments, enhancing the adaptive differential evolution algorithm. This method integrates effective task insertion rules based on task observation order [7]. Zheng et al. [49] employed a multi-objective hybrid dynamic mutation genetic algorithm combined with periodic and near-real-time replanning techniques for onboard replanning in disruptive scenarios.
These methods rely on specifically designed heuristic strategies, may lead to local optima [50], and are unable to meet the real-time computation requirements of large-scale dynamic scenarios. Recently, due to the powerful modeling and reasoning capabilities of neural networks, they have also been applied to problem-solving.
Neural networks are used as predictive models to assist in real-time decision making. Chen et al. [51] introduced a real-time multi-satellite scheduling approach that combines a hierarchical prediction model based on a stacked multi-channel transformer network with a heuristic local search algorithm, achieving high-quality initial solutions and rapid computation. Gu et al. [52] enhanced the replanning algorithm’s effectiveness by utilizing satellite cloud image forecasting information from predictive recurrent neural networks.
Another approach is to directly use neural networks to build decision-making strategies. Deep reinforcement learning (DRL) continuously learns and optimizes scheduling strategies through feedback from the environment, enabling it to make optimal decisions based on relevant states. This method has shown great potential in solving multiple AEOS scheduling problems. For instance, Wang et al. [53] developed a satellite scheduling model based on the dynamic stochastic knapsack problem and optimized it using a deep reinforcement learning framework to maximize total expected profit, achieving real-time scheduling for imaging satellites. However, this model focuses solely on immediate request responses, neglecting the updates to long-term mission plans.
Many graph neural network (GNN) models have been used to enhance the representation of task planning problems [54]. Wang et al. [55] modeled the multi-satellite scheduling problem as a heterogeneous graph subject to various constraints, using a GNN and Transformer based encoder to enhance information representation. They trained a two-stage dynamic decoder using reinforcement learning. Feng et al. [56] utilized GNNs to extract feature knowledge of large-scale tasks and their constraint relationships while employing a metaheuristic optimization module to resolve dynamic constraint conflicts.
However, graph-based models often represent edges as various constraints to avoid conflicts, which is not suitable for replanning scenarios. This modeling approach struggles to capture the original sequence and handle task replacement, potentially leading to inflexibility in task insertion. Additionally, for dynamic requests arrival and insertion in our study, GNNs require frequent updates to the graph structure, resulting in significant computational overhead.
The sequence-to-sequence (seq2seq) model is an advanced combinatorial optimization framework that provides an encode-decode strategy learning paradigm. Pointer networks (Ptr-Net) [57] and Transformer architectures [58] have been proven effective in solving problems like the vehicle routing problem (VRP) and traveling salesman problem (TSP). Liu et al. [59] proposed a model with a local attention mechanism to reduce the search space in the autoregressive process, improving decision quality. Chen et al. [60] used attention layers to model dynamic time dependencies between satellite tasks, training an encoder–decoder neural network with the REINFORCE algorithm. Li et al. [14] used a trained pointer network to generate single EOS scheduling solutions in an end-to-end manner for rapid onboard replanning. Liang et al. [61] implemented rapid rescheduling using an improved pointer network with a masking mechanism to maximize the completion profit of dynamic tasks while minimizing the impact on scheduled tasks. Long et al. [62] proposed a scalable task planning method based on a Transformer model with time encoding, which we consider as one of the baselines in this study.
Multi-satellite task allocation involves specific constraints and is a structured sequential decision-making process. The attention-based Ptr-Net uses a softmax probability distribution as a pointer of the input sequence to gradually select input elements to construct the output sequence, making it effective for handling long sequences and combinatorial optimization problems of varying scales. Particularly for replanning problems, where the number of incoming requests and satellites is uncertain, our method employs a dynamic encoder and a pointer network-based decoder to handle dynamic task allocation. Compared to the Transformer model, the pointer network has a lower computational complexity, directly outputs the strategy probability distribution, and is better suited for discrete decision problems like task allocation.

3. Problem Statement

This section first introduces the scenario and the replanning framework, followed by a description of the definitions and assumptions. Next, the objective function and constraints of the problem are presented.

3.1. Scenario and Replanning Framework

During orbital operations, satellites may encounter various environmental changes that could render original observation tasks invalid or require modifications. Internal factors include satellite payload or platform failures that lead to task cancellations, which in our scenario are handled as dynamic requests directed to the nearest available satellite. External factors include changes in observation conditions, such as cloud cover or lighting, which may result in task cancellations or suspensions, as well as shifts in user demand or emergency observation requests triggered by natural disasters. As shown in Figure 1, dynamic requests originate from ground user demands, autonomous satellite discoveries, or unforeseen events. We assume that these requests initially reach individual satellite nodes.
For scenarios involving batch arrivals of dynamic observation requests and their timely scheduling, we propose a framework combined with mission allocation and single-satellite replanning. Based on the assumption of a distributed constellation where each node has equal intelligence, we adopt a temporary master–slave architecture to minimize the demands of communication resources for mission allocation. Specifically, the satellite that first receives the requests acts as the temporary master, while other satellites serve as slaves.
By modeling the original plans of each satellite, the master satellite can directly generate high-quality allocation schemes for slave satellites. Once the allocation is received, each satellite executes its own single-satellite replanning. This approach reduces the overall scale of problem solving, minimizes communication and negotiation overhead, and provides precise replanning solutions tailored to the actual state of each satellite. We believe that it achieves more accurate responses while minimizing modifications to the original plans.

3.2. Assumptions and Definitions

Our scenario can be described as follows.
  • Assuming that a group of satellites S a t = { s a t j | j = 1 , 2 , , N S } . Each satellite is represented as s a t j = { v j , a m a x j , C m e m j , C p o w e r j , a j , h m e m j , h p o w e r j } , where v j , a m a x j , C m e m j , C p o w e r j are static parameters that remain constant during the scheduling process. These parameters, which reflect inherent task characteristics, are pre-determined and include the attitude transition rate ( v j ), maximum side swing angle ( a m a x j ), memory capacity ( C m e m j ), and power capacity ( C p o w e r j ). Conversely, a j , h m e m j , h p o w e r j are dynamic parameters, which are updated at each scheduling step to represent the current side swing angle ( a j ), remaining memory ( h m e m j ), and remaining power ( h p o w e r j ) at any given time. It is important to note that, in our model, s a t j operates continuously on orbit O r b i t j and does not perform any orbit-raising maneuvers.
  • Dynamic requests can be represented as R e q = { m i | i = 1 , 2 , , N R } . Each request is defined as m i = { ( s i j , e i j , d i j , a i j , p i , m e m i , p o w e r i ) | j = 1 , 2 , , N S } . It contains a quadruple, s i j , e i j are the start time and end time of the visible time window of s a t j to request m i ; d i j is the execution duration required for s a t j ; a i j represents the side swing angle required for s a t j observation to m i . Additionally, the request includes the priority p i , required memory consumption m e m i , and required power consumption p o w e r i .
  • When a request is scheduled into the mission sequence of a designated satellite s a t j , it is represented as M i = { m k , r k , ( T s i , T e i ) , s a t j } , where ( T s i , T e i ) represent the start and end time of the mission execution, and r k indicates the reward associated with the task. Each satellite has an optimal initial plan I n i t P l a n j = { M 1 j M 2 j M L j j } , that contains L j missions; The updated mission plan is represented as N e w P l a n j = { M 1 j M 2 j M L j j } , with a length of L j .
  • We define that new requests arrive in batches, and at each time step, only one satellite will be designated as the master satellite for task allocation.
  • The integration of new requests into the mission sequence must satisfy the constraints of the satellite imaging tasks.

3.3. Objective Function and Constraints

The objective function in this work consists of two main components: mission revenues and disturbance costs. We define mission priority as its revenue, aiming to maximize the total revenue generated by the satellite from executing the target observation in one episode. Disturbance costs represent the impact of inserting new tasks into the initial plan throughout the replanning process, including time changes and task substitutions. The objective function we design is as follows:
m a x f = μ · f 1 ( 1 μ ) · f 2
f 1 = j N S i L j p i , f 2 = j N S i L j p i · τ c h a n g e
τ c h a n g e = ω 1 , i f m i i s r e m o v e d , ω 2 , i f m i i s s h i f t e d , 0 , i f m i i s u n c h a n g e d ,
where μ is the scale scaling factor, which can be adjusted based on the user preferences. f 1 represents the total revenue of the new plan, and f 2 is the penalty for changes to the initial plan. The changes are classified into three categories, each assigned a different penalty factor 0 < ω 2 < ω 1 1 : removed, execution time shifted, and no change occurred.
The following constraints should be satisfied.
  • Visible Time Window Constraint: The start and end of the execution time in the satellite’s mission sequence must satisfy
    T s i j s i j , T e i j = T s i j + d i j e i j ,
  • Transition Time Constraint: Continuous missions must satisfy the attitude transition requirement. After completion of the previous task M i , the satellite must undergo a specified maneuver, ensuring that M i + 1 can still complete the observation before the end of its visible time window.
    T e i + | a i a i + 1 | v + d i + 1 e i + 1
  • Adjustment Time Range Constraint: The time window within which M i can be moved is determined by its visible time window ( s i , e i ) , as well as the execution times of the preceding task M i 1 and the following task M i + 1 . The maximum advancement time is
    T f o r w a r d = T s i m a x { s i , T e i 1 + | a i a i 1 | v }
    The maximum delay time is
    T b a c k w a r d = m i n { e i , T s i + 1 | a i + 1 a i | v } T e i
  • Uniqueness constraints: Each request has a different visible time window for each satellite, but it can only be executed by one satellite at a time.
    N e w P l a n j N e w P l a n k = , j , k = 1 , 2 , , N S
    I n i t P l a n j I n i t P l a n k = , j , k = 1 , 2 , , N S
    Each task can only be observed once. For any given satellite, each selected task can have at most one preceding task and one succeeding task.
  • Resource Constraint: The complete N e w P l a n j is managed through the removal of resource-exceeding tasks from its terminal end.
    i L j m e m i C m e m j , j = 1 , 2 , , N S
    i L j p o w e r i C p o w e r j , j = 1 , 2 , , N S

4. Solution Method

This section presents a mission sequence-based replanning (MSRP) algorithm for multi-satellite observation. To address the complexity of original plans and the uncertainty of request scales, we propose mission sequence modeling and dynamic request embedding to enhance state representation. As described in Section 3.1, the MSRP framework consists of two components. In mission allocation, we design a pointer network-based algorithm to generate assignment solutions. In the satellite replanning part, we introduce a neural network-based task insertion strategy to select tasks and determine execution times for received requests. Algorithm 1 outlines the MSRP framework. Both components are modeled as Markov Decision Processes (MDPs) and trained using the proximal policy optimization (PPO) algorithm.
Algorithm 1 The framework of MSRP
Input: Satellite set S a t , Request set R e q arrive at s a t * , original plan { I n i t P l a n j | j = 1 , 2 , , N S }
Output: New plan { N e w P l a n j | j = 1 , 2 , , N S }

Enhance State Representation:
E M S , E G S = M i s s i o n S e q u e n c e M o d e l ( S a t , I n i t P l a n )
E m , E R = D y n a m i c R e q u e s t M o d e l ( R e q )

Mission Allocation in  s a t * :
for  m i R e q  do
  Select s a t j S a t through A P o l i c y ( E G S , E R , m i )
  Assign m i to R e q j
end for
return  R e q j for each s a t j

 /*proceed in parallel*/
Mission Replanning in  s a t j :
while not terminated do
  Get ( m i , T s i ) though R P o l i c y ( E M , E m , s a t j )
  Insert m i into N e w p l a n j
end while

Output Optimal Scheme  { N e w P l a n j }

4.1. Mission Sequence Modeling

We introduce a spatiotemporal embedding method for modeling mission sequences, capturing both time and space-related factors. Temporal aspects focus on task execution times (start times and intervals), while spatial factors address the required observation side swing angles. By employing attention blocks, we encode both timestamp and maneuver information into embedding vectors. A gating mechanism integrates the mission intervals, followed by global pooling to produce a unified representation of the current satellite state. This approach effectively combines spatiotemporal dynamics for enhanced task sequence modeling. The complete structure of the method is fully displayed in Figure 2a.
To ensure the timeliness of request responses, we fix the length of the mission sequence to L. If the sequence length exceeds L, we truncate it to the L tasks closest to the decision time. If length is less than L, we append virtual missions at the end, with their execution time set to the orbit end time and revenues set to 0.
First, the input satellite planning schemes I n i t P l a n = { M 1 M 2 M L } are transformed into high-dimensional representations by embedding each mission M i into a vector e i . These individual embeddings are then concatenated to form the vector E = [ e 1 , e 2 , , e L ] R L × d , which represents the entire mission sequence in a continuous vector space.
Inspired by the multilevel temporal information rotational position encoding method in [63], we embed the timestamp information from the original missions and the side swing angles for request–satellite pairs, modeling the temporal correlations between mission sequence. By incorporating temporal and attitude information, we enhance the attention mechanism to explicitly capture the absolute timing information of the tasks and calculate the relative transition span between tasks. Figure 3 illustrates the schematic diagram of the embedding process.
We standardize the time granularity for all tasks using the midpoint of each execution time as the timestamp, expressed as T i m e S t a m p s = { t i = s t i + e t i 2 , i = 1 , 2 , , L } .
Satellites require attitude adjustments between consecutive observations. In our model, transmission time is proportional to the side swing angle difference between adjacent tasks. Therefore, we utilize A t t i t u d e = { a 1 , a 2 , , a L } to effectively enhance the learning of the transition interval information between tasks.
For a mission vector e i E , it encodes the timestamp and attitude positions into a d-dim space with the following function.
e i R = e i · P i
where P i = [ P i 1 P i 2 P i d / 4 ] R d × d is a position embedding matrix. Each sub-matrix P i j R 4 × 4 is
P i j = c o s ( t i θ j ) s i n ( t i θ j ) 0 0 s i n ( t i θ j ) c o s ( t i θ j ) 0 0 0 0 c o s ( a i θ j ) s i n ( a i θ j ) 0 0 s i n ( a i θ j ) c o s ( a i θ j )
where { θ j = 10000 4 ( j 1 ) / d , j = 1 , 2 , , d / 4 } are pre-defined. Due to the sparsity of P i , the computation can be accelerated by performing element-wise multiplication and addition operations.
The modeling of time-attitude dependencies is achieved through the computation of self-attention mechanisms. As illustrated in Figure 3, the pairwise inner product calculations effectively capture the relative time differences and relative side swing angle variations between tasks. It is evident that the sinusoidal and cosine functions in the encoding enable e k R to be represented as a function of e i R for any given execution time and attitude offset Δ t , Δ a (as shown in Equation (14)), allowing the model to efficiently learn and focus on relative time-attitude differences.
e i R · e k R = e i P i P k e k = g ( e i , e k , Δ t i k , Δ a i k )
Moreover, integrating attitude information provides richer insights than relying solely on task execution times, as the relative side swing angle difference is proportional to the transition time between tasks: t i i + 1 = | a i + 1 a i | v , enabling more precise modeling.
This encoding adapts to variations in visible time windows induced by different task distributions and satellite configurations. Unlike conventional positional encoding, it remains unaffected by task sequence order changes, making it more robust for replanning.
After adding rotational position encodings, we obtain the mission vector E P , which then is sent to a multi-head self-attention block. First, E P is linearly mapped through query, key, and value components to obtain three feature matrices:
Q , K , V = E P W Q , E P W K , E P W V
where Q , K , V R n × d and W Q , K , V R d × d . We add a low-rank decomposition projection to generate a more compact contextual representation. The corresponding K , V are reduced in size by a trainable mapping function h : E R L × d D R l × d . Here, l is an optional size, representing the range of selected correlations.
D K = h ( K ) = ( S o f t m a x ( K W l ) ) K
D V = h ( V ) = ( S o f t m a x ( V W l ) ) V
where D K , D V R l × d and W l R l × d is trainable.
Afterward, the input Q is passed through the attention mechanism to compute the embeddings.
A = S o f t m a x ( Q D K d / h ) D V
The final output is E M R L × d which merged with multi-head and through linear layers. The embedding of each satellite is then concatenated and reshaped back into E M S R N S × L × d .
In this module, we design a global pooling method for mission sequences to capture the satellite’s overall schedule. It combines a virtual node-based attention mechanism to weight E M and a gating mechanism that uses execution durations and task intervals. These are then combined and weighted-summed with E M to obtain each satellite’s state representation.
We use the average of the task embeddings e m e a n = 1 L i = 1 L E M as the virtual node. Then, a single-head attention layer is utilized to calculate the weight α i for each mission.
α i = e x p ( e m e a n e i ) k e x p ( e m e a n e k )
During observation schedule adjustments, requests may need to be inserted into or replace existing missions, taking into account the preceding and following intervals. Additionally, since task durations vary and cannot be treated as single time points, their execution times must also be considered.
In P l a n = { M 1 M 2 M L } , for each M i , we consider a triplet I i = ( I i l e f t , d i , I i r i g h t ) , where I i l e f t = T s i T e i 1 , I i r i g h t = T s i + 1 T e i , and d i is the execution duration of M i .
We incorporate mission interval information E I by calculating intermediate features and assigning weights through a gating mechanism, allowing the model to identify important task intervals and adaptively fuse the corresponding features.
E I * = T a n h ( W P E I )
G = S i g m o i d ( W G E I )
where W P , W G are learnable weights of a linear layer. The gating parameters are multiplied by the normalized weights and summed to yield the time interval weight β .
β = G E I *
Ultimately, the two weights are summed to compute the global state of the mission sequence.
E G = ( α + β ) · E M
The collection of states for S a t is represented as E G S R N S × d .

4.2. Dynamic Request Modeling

Due to the uncertain number of dynamic requests, directly concatenating all information would increase the number of trainable parameters. Inspired by [64], we use a scalable mechanism to embed incoming requests. The structure of the embedding method is shown in Figure 2b.
Input requests R e q = m 1 , m 2 , , m N R are first processed through a linear layer to obtain high-dimensional representations, followed by ReLU activation and another linear layer to generate request embeddings. For each m i ,
e m i = W R 2 R e l u ( W R 1 m i )
where W R 1 , W R 2 are learnable weights of linear layer. The vectors of all requests are denoted as E m .
A pooling module is then used to generate the global state of the requests. Each request vector e m i is passed through a linear layer with Tanh activation, followed by a Sigmoid-activated output layer to compute request weights.
a m i = S i g m o i d ( W R 4 T a n h ( W R 3 e m i ) )
where W R 3 , W R 4 are learnable. We compute the weighted average of E m as the global state E R of the input requests.
E R = i = 1 N R a m i e m i

4.3. Multi-Satellite Mission Allocation

In our framework, mission allocation is modeled as a sequential decision problem, formulated as a Markov Decision Process (MDP) to apply DRL algorithms.
  • State: In mission allocation, the state at step t comprises two parts: the mission sequence of satellites from t 1 and the dynamic request set. The satellites’ immediate state and resource changes are recorded within schedule. Using the mission sequence modeling method, we derive satellites’ global embedding E G S R N S × d . For dynamic requests, we generate the global embedding E R and individual request vectors as e m i E m .
  • Action: The action space consists of satellites awaiting task allocation, defined as a discrete set. We employ a pointer mechanism to generate the policy function, with the detailed procedure outlined in the following steps and visually represented in Figure 4a.
    (a)
    Apply a cross-attention layer, where the query is the satellite state E G S R N S × d , and the key and value are the requests E m R N R × d . The output is the mission-satellite association vector { h 1 , h 2 , , h N S } , which aims to capture the relationships between the satellite and all requests.
    (b)
    Concatenating each request vector e m i with global embedding E R to construct vector z = ( e m i , E R ) .
    (c)
    We use a pointer mechanism to generate the probability distribution of the allocation policy π ( a t | s t ) .
    u j = v 1 t a n h ( W P 1 h j + W P 2 z )
    π ( a t | s t ) = S o f t m a x ( u j ) , j = 1 , 2 , , N S
  • Reward: The objective of mission allocation is to maximize the overall reward, which is determined after replanning by each satellite. Therefore, mission allocation is jointly optimized with the single-satellite replanning algorithm. If a request is successfully inserted into the plan, the step reward is the task’s revenue; otherwise, it returns zero. The revenue of each request is normalized by the sum of total revenues.

4.4. Single-Satellite Mission Replanning

To meet the mission replanning requirements in dynamic environments, a sequential decision-making algorithm for uncertain request arrivals has been developed. The objective is to maximize the task revenues while minimizing the changes to the original schedule. Single-satellite mission replanning can also be modeled as a Markov Decision Process.
  • State: In this replanning phase, the satellite state is represented by its mission sequence embedding E M R L × d , which incorporates resource consumption. The global state E G R 1 × d is given by the pooling method. The dynamically arriving requests, with an uncertain scale N R j , are encoded and represented as E m R N R j × d .
  • Action: The replanning actions are composite, using a hybrid action space. Each action at step t is represented as a t = ( m i , T s i ) where m i denotes the task selected from the allocated set, and T s i refers to the execution time of the selected task. We introduce a termination action to end the replanning process. The action generation is handled by two branches of the actor network, which output discrete and continuous actions, respectively. The specific actor network is illustrated in Figure 4b.
    (a)
    State Extraction: Two consecutive cross-attention blocks are used to build a compressed representation of the current state. In the first block, q = E m ; k , v = E M , which generates the task-satellite affinity vector h m i d R N R j × d . In the second block, q = E G ; k , v = h m i d , where the satellite’s global state is used as the query to capture the current global state h g .
    (b)
    Discrete Action: The mission selection policy distribution π d ( m t | s t ) is generated using the pointer mechanism.
    u i = v 2 t a n h ( W P 3 h m i d + W P 4 h g )
    π d ( m t | s t ) = S o f t m a x ( u i ) , i = 1 , 2 , , N R j
    (c)
    Continuous Action: Add the selected task information into the actor network to generate the mean μ , and variance σ . After sampling from N ( μ , σ ) , we scale the action to the range ( 1 , 1 ) using the Tanh function. The output is then re-normalized to the specific mission’s start time using the formula.
    T s i = y c s c a l e c + b i a s c , s c a l e c = s i + e i d i 2 , b i a s c = e i s i d i 2
    (d)
    Mission Insertion: The mission m i is inserted based on T s i . In case of conflicts with existing missions, a fast insertion approach (FIA) like principle [19] is applied to resolve the conflict by shifting existing task or removing it if necessary.
  • Reward: Our optimization objective consists of two parts: mission revenues and the cost function associated with changes to the initial mission sequence. Thus, we design the reward as follows:
    r t = α · l e n ( t ) l e n ( t 1 ) + 1 · p s e l β · p i · τ c h a n g e
    τ c h a n g e = 0.8 , i f r e m o v e d , T s i ( t ) T s i ( t 1 ) e i s i , i f s h i f t e d , 0 , i f u n c h a n g e d ,
    where α , β are adjustable parameters that satisfy 0 < α , β < 1 . l e n ( t ) represents the length of the satellite mission sequence at step t, p i denotes the revenue of m i , and τ c h a n g e is the penalty factor for different levels of modifications. The goal of our reward design is to maximize both the task plan length and revenues while minimizing the disruption to the original plan during urgent request scheduling.

4.5. Training

We use proximal policy optimization (PPO) to sequentially train mission replanning and allocation policy. PPO is a policy-based reinforcement learning algorithm that estimates the performance of the policy using a novel objective with clipped probability ratios. The clipping function limits the magnitude of policy updates, ensuring stability and reliability during training.
We implement the DRL algorithm using the actor–critic architecture. The actor network outputs the policy π ( a t | s t , θ ) , while a critic network V ϕ ( s ) is used as an estimator of the state-value function V π ( s ) . The training process is illustrated as Algorithm 2.
Algorithm 2 Procedure for training policy based on PPO
  1:
Input: Initial policy parameters θ 0 , initial value function parameters ϕ 0 , clipping parameter ϵ , discount factor γ , number of updates N updates , number of steps N steps , number of epochs K, learning rate α .
  2:
Output: Optimized policy parameters θ *
  3:
Initialize θ θ 0 , ϕ ϕ 0 . Initialize buffer B
  4:
for update = 1 N updates  do
  5:
   envs.reset()
  6:
   for step = 1 N steps  do
  7:
     Mission Sequence Modeling and Dynamic Request Modeling s t
  8:
     Using current policy to generate a t π θ ( a t | s t )
  9:
     /*Do mission allocation or replanning*/
10:
      s t + 1 , r t envs . step ( a t )
11:
     Store { s t , a t , r t } to buffer B
12:
   end for
13:
   for epoch = 1 K  do
14:
     Compute advantages using Generalized Advantage Estimation (GAE):
15:
      A ^ t = l = 0 T t ( γ λ ) l δ t + l , where δ t = r t + γ V ϕ ( s t + 1 ) V ϕ ( s t )
16:
     Compute returns R ^ t = A ^ t + V ϕ ( s t )
17:
     Update policy, value function using PPO objective:
18:
      θ θ + α θ E t min r t ( θ ) A ^ t , clip ( r t ( θ ) , 1 ϵ , 1 + ϵ ) A ^ t
19:
      ϕ ϕ α ϕ E t V ϕ ( s t ) R ^ t 2
20:
   end for
21:
end for
22:
Return Optimized policy parameters θ * θ
The training process for both mission allocation and replanning agents is similar. The main difference is that mission allocation involves assigning requests to satellites sequentially, with the reward coming from the result of each satellite replanning. The general training flow is to first train the single-satellite replanning network and then fix the optimal replanning policy and integrate it into the mission allocation environment.

5. Computational Experiments

This section outlines the experimental setup and presents three simulations to validate our algorithm. Experiment 1 tests single-satellite replanning, Experiment 2 evaluates the mission allocation and replanning framework, and Experiment 3 conducts ablation studies to verify the method’s effectiveness.

5.1. Scenario Settings

The simulation scenario is based on the multi-satellite observation requirements and constellation parameters derived from the Walker Delta configuration. The constellation was created on 1 April 2020, with six orbital planes, each containing one satellite. The simulation time spans from 00:00:00 to 24:00:00 (UTC), aligning with an orbital period of 6080 seconds. Table 1 shows the specific orbital parameters of the satellites and Table 2 shows the setting of simulated observation requests.
The observation missions are generated as real-world application scenarios, distributed randomly within the geographic region from 3 N to 53 N and from 74 E to 133 E. The initial mission plan for each satellite consists of 100, 200, or 300 missions. The proportion of new dynamic requests per satellite relative to the initial plan size is set to 10%, 20%, 30%. In mission allocation, the number of requests is equal to the single-satellite capacity multiplied by the satellite count.
Training Setup. Based on the ADSMP algorithm from [65], we generate the optimal mission plans for a multi-satellite system. During training, we randomly select a certain proportion from the original plan as requests, with the remaining used as I n i t P l a n . For validation, once I n i t P l a n is determined, new observation requests are randomly generated each time.
The algorithm was implemented in Python 3.9.16, and the experiments were conducted on a system with an Intel(R) Xeon(R) 8375C 2.90 GHz processor, running Ubuntu 20.04, and equipped with a single NVIDIA GeForce RTX 4090 GPU.
For the single-satellite mission replanning algorithm, training is performed with 100 task size and 20% dynamic request ratio. For the mission allocation algorithm, the number of satellites is set to 4, with 100 task size and 20% dynamic request ratio, resulting in a total of 80 requests for training. The network parameters are shown in Table 3.
The convergence curve is shown in Figure 5. The fluctuations during the training process are mainly due to the randomness of the scenario’s input mission sequences.
The metrics considered in the experiment include:
  • Insert rate of new requests: N I R = N i n s N R ;
  • Revenue rate of new request: N R R = i N i n s p i k N R p k ;
  • Execution rate of missions: E R = l e n ( N e w P l a n ) N R + l e n ( I n i t P l a n ) ;
  • Total Revenue rate: R R = i N e w P l a n p i j N R p j + k I n i t P l a n p k ;
  • Modification rate: M R = i I n i t P l a n p i · τ c h a n g e i I n i t P l a n p i ;
  • Computation time: T i m e ( s ) .
It is worth clarifying that M R is a metric to quantify the level of disturbance to the original plan. Based on Equations (3) and (33), we set the weight ω 1 for removed tasks to −1, since these tasks are completely removed. The degree of shift is defined by the proportion of the change in execution time relative to the visible time window, which gives the weight ω 2 . We define M R as the ratio of the revenues before and after replanning.

5.2. Comparison Algorithms

We selected the following algorithms for comparison:
  • Multiple Strategies Local Replanning (MSLR) Algorithm: This algorithm integrates multiple insertion principles from [66]: direct insertion, move insertion, replace insertion; along with a hybrid insertion strategy from [7]: direct insertion, iterative insertion, conflict replacement insertion. Since our scenario focuses on mission-satellite visible time windows, we combined the merging insertion into direct insertion.
  • Transformer-based single-satellite replanning (TSR) Algorithm: We adopt the Transformer-based architecture with temporal encoding from [62] as the task scheduling method for single-satellite replanning. After obtaining discrete decisions, we use a neural network with the same architecture as MSRP to compute continuous actions.
  • Fast Insertion Approach (FIA) Algorithm: Based on the FIA principles from [19], this approach prioritizes tasks and determines insertion positions based on their feasibility between adjacent tasks in the satellite sequence, optimizing overall gain.
  • Multi-Satellite Replanning with adaptive Differential Evolution (RDE) Algorithm: We implemented an adaptive differential evolution algorithm [7] combined with the MSLR local replanning rule. The fitness function is defined as (1) with μ = 0.5 .
  • Transformer-based multi-satellite replanning (TMR) Algorithm: We combine the multi-satellite task allocation framework based on the attention mechanism from [67] with TSR as the single-satellite replanning method, forming our comparative algorithm.
  • Plan Regenerate (PRG) method: Based on the ADSMP method, we regenerate the task sequence by inputting both dynamic requests and the original missions.
  • Ablation-MLP: To validate the effectiveness of our mission sequence modeling module, we developed a baseline embedding network for comparison, represented by a three-layer MLP with 128 nodes. This network processes satellite mission sequence information and matches the input for the subsequent policy network. We used the PPO algorithm for training.
  • Ablation-pooling: To verify the effectiveness of our global pooling module, we replaced it with a mean pooling method for ablation. The PPO algorithm was used to train this version.
  • Ablation-Transformer-PE: Replace mission sequence modeling method by standard attention with classic position encoding (PE) for ablation.
  • Ablation-Transformer-RoPE: Replace mission sequence modeling method by standard attention with Rotary Position Embedding (RoPE) for ablation.

5.3. Experiment 1 on Single-Satellite Replanning

We compared the replanning component of the MSRP algorithm with the MSLR algorithm, FIA algorithm and TSR algorithm. Figure 6 and Figure 7 present the performance evaluation of the four algorithms on varying scales of missions and request arrivals for single-satellite mission replanning.
When the scenario scale is relatively small, particularly 100 missions or 10% new requests, all algorithms respond quickly and achieve replanning with a low modification rate. The main reason is that smaller tasks and request scales leave ample idle time in the mission sequence, allowing for direct insertion. As task volume increases, all algorithms show a gradual decline in insertion rates and total revenue. However, our proposed algorithm consistently outperforms others across all scenarios. When the number of tasks reaches 300 or a new ratio of 50%, the average NIR and RR of MSRP remain around 95%, which is significantly higher than the other algorithms.
As more tasks are inserted, FIA’s reliance on simple interval assessments with a fixed 0.5 threshold for execution timing becomes ineffective, lowering the new task insertion rate. While FIA’s modification rate remains low without a replacement strategy, its limited insertion capacity leads to reduced overall revenue as requests increase.
The MSLR method uses complex insertion rules to handle overlapping conflicts that FIA cannot address. However, MSLR has the longest computation time due to iterative rule matching. Its limited flexibility in task insertion leads to frequent replacements, reducing overall revenues and increasing modifications to the original scheme, which can be proved in Figure 6b.
Compared to the TSR method, MSRP demonstrates significant performance improvements when handling 100 to 300 tasks. When the scenario scale is small, TSR and our model demonstrate similar performance, with metrics such as NIR, NRR, ER, and RR exceeding 95%. However, when the tasks increases to 300 or the proportion of new requests rises to 50%, MSPR maintains its metrics above 95%, while TSR’s performance drops significantly, with NRR and other metrics falling to around 90%. The downward trend suggests that as the number of tasks and the proportion of new requests increase, the performance gap between TSR and MSPR is likely to widen further, as illustrated in Figure 6a and Figure 7a.
We believe that the cumulative effect of replanning intensifies with each task insertion. As the proportion of new requests increases, the performance disparity between TSR and MSRP becomes more pronounced, primarily due to our algorithm’s enhanced precision in determining insertion positions with successive insertions. Compared to MSRP, TSR does not incorporate attitude information and can only perceive relative time differences between tasks, failing to address the effects of time required for attitude adjustments between tasks. As a result, it cannot effectively reduce the search space for execution times within the feasible insertion window, ultimately limiting its ability to improve replanning performance. Furthermore, our task sequence modeling approach introduces information about task intervals and durations through a global pooling module, effectively guiding new requests to be inserted into idle time slots of the sequence. This is evidenced by our method’s lower MR compared to TSR. As mentioned earlier, the pointer network has lower computational complexity than the Transformer, resulting in significantly reduced time consumption for MSRP. In scenarios with 300 tasks, the time required by MSRP is only 50% of that needed by TSR.
Our method leverages mission sequence modeling to directly determine the insertion approach. It effectively considers the relationships between the satellite state, the original plan, and incoming requests, providing an optimized replanning solution. In addition, it reduces computation time, enabling faster response.
The performance of our algorithm and MSLR is illustrated through three insertion cases in Figure 8. The number in the task block indicates its duration or priority. The dashed box marks the visible time window. Scheduled tasks are aligned on the timeline, with the numbers above indicating their start times.
Case 1 demonstrates direct task insertion. Our replanning algorithm assigns a specific start time for each task. m t is scheduled at 2017 s with a preceding task shifted forward. This creates a gap for subsequent tasks like m k to be inserted directly, which MSLR cannot achieve.
Case 2 shows the task insertion process. We directly insert ( m i , T s i ) , with gaps on both sides to allow future adjustments. In contrast, MSLR requires three rounds of checks to refuse Rule 1. Then, Rule 2 inserts the task after the preceding one, limiting flexibility for future adjustments. This case highlights that our method offers more flexibility and saves significant search time.
Case 3 shows a replacement insertion. MSLR searches three steps, confirms the new request’s reward (0.8) exceeds the combined reward of two original tasks (0.7) and proceeds with replacement. However, our method ensures the subsequent tasks meet the T b a c k w a r d constraint, enabling both replacement and shift insertion. This demonstrates that our algorithm achieves higher overall revenues in many cases.
Under different metrics, we conducted a “paired sample Wilcoxon signed-rank test” on the results of our model compared to other models, and calculated the confidence interval for the differences in results between the models. The results indicate that there is a significant difference between the outcomes of our model and those of the other models as shown in Table 4.
In summary, across experiments of varying scales, our proposed algorithm outperforms others in terms of insertion rate and revenue rate, while minimizing modifications to the original plans. As a result, it provides higher quality single-satellite replanning solutions.

5.4. Experiment 2 on Multi-Satellite Replanning Framework

In this section, we validate the effectiveness of multi-satellite mission replanning across different scenarios. As shown in Figure 9, MSRP performs consistently well with 100–300 tasks, while other algorithms experience varying levels of decline, highlighting its strong replanning ability.
Our method outperforms RDE across various task volumes. Specifically, MSRP’s average NIR is 27.3% higher, and its NRR is 13.6% higher than RDE, demonstrating better performance in responding to dynamic requests. The efficient replanning ability of MSRP increases overall revenue, which is reflected in a high insertion rate accompanied by a high execution rate and a total revenue rate. As shown in Figure 9e, the MR of MSRP and RDE is similar at around 10%, but MSRP handles nearly 30% more new requests than RDE. This advantage stems from MSRP’s enhanced global state modeling and its ability to capture the relationship between new requests and the original mission sequence.
Compared to PRG, the results show that our replanning algorithm achieves a total revenue nearly identical to PRG, with only a 0.8% difference. However, as shown in Figure 9e, our MR is significantly lower than PRG, and its growth rate is slower as the scale increases. This demonstrates that our approach can achieve high-revenue task insertion while minimizing changes to the original sequence.
Compared to the learning-based TMR method, our approach outperforms it across task volumes ranging from 100 to 300, with less sensitivity to scale variations. Specifically, under the condition of consuming almost the same amount of time, NIR and NRR are 13.8% and 15.3% higher, while ER and RR are 3.2% and 3.1% higher, respectively, with a reduction of 1.2% in MR. By incorporating relative time differences and side swing angle differences into the replanning model, mission sequence modeling significantly enhances the model’s dynamic adaptability, enabling it to handle varying scales of input requests without being constrained by sequence order or frequent task insertions. Additionally, the proposed allocation method enhances the capability to perceive task insertion effects by capturing the correlation between satellite states and dynamic requests. Since our single-satellite replanning performance already exceeds TSR, combined with the flexibility of the pointer network in task allocation, our approach achieves more efficient single-satellite replanning, significantly improving both new request benefits and overall task benefits.
Figure 9 shows that our algorithm is the fastest, with a slow increase in computation time as the task scale grows. In contrast, the RDE algorithm involves multiple evolution iterations and rule matching during task insertion, leading to high time consumption, with solving times reaching hundreds of seconds. While PRG also uses neural network inference to speed up computation, it must plan all tasks in the scenario, which increases computational load and leads to higher time consumption than our approach. The results of the significance difference analysis and confidence interval of this experiment are shown in Table 5, indicating that there are significant differences between our model and other models.

5.5. Experiment 3 for Ablation

The effectiveness of our method is validated through comparisons with ablation models. Results in Figure 10a show that MSRP outperforms Ablation-MLP by 23.3% in NIR and 17.2% in NRR. This highlights the importance of our proposed mission sequence modeling.
Mission sequence modeling also improves the model’s generalization. Trained in a 100–20% environment, MSRP shows stable performance across all test scenarios, with NIR and NRR standard deviations of 0.0401 and 0.088, respectively. In contrast, while ablation-MLP performs well in training-scale environments, its performance varies significantly across different scenarios, with standard deviations of 0.200 for NIR and 0.192 for NRR.
The comparison with the ablation-pooling confirms the effectiveness of the proposed gated global pooling mechanism. Results show that MSRP consistently outperforms across multiple metrics, especially as the task volume increases. We believe that the pooling mechanism uses mission time intervals to enhance the accuracy of the global state representation, facilitating better task allocation.
Compared to the standard Transformer with PE, MSRP improves NIR and NRR by 14.6% and 13.5%, respectively, and increases ER and RR by 3.5% each. Compared further to the ablation model with RoPE, MSRP achieves improvements of 8.9% and 7.5% in NIR and NRR, respectively, and 2.5% and 2.7% in ER and RR, with a 1.7% reduction in MR. Both methods use sequential numbering for positional information, limiting their effectiveness in replanning scenarios with frequent task insertion. In contrast to the standard attention model, RoPE calculates relative positional information, which we believe better preserves and enhances the modeling of relative differences between tasks. As a result, it outperforms the standard attention model using absolute positional information.
In our model, we replace the sequential numbering in rotational positional encoding with task execution times and side-swing angles. This relative positional encoding approach directly models the transition and maneuvering times between tasks, further enhancing the representation of mission sequences. Additionally, changes in sequence numbering during the replanning process do not affect this modeling. Ablation results demonstrate that our approach effectively supports replanning, improving task insertion rates and overall revenue while reducing MR. Table 6 presents a statistical analysis, showing significant differences between the results of this model and the ablated version.

5.6. Sensitivity Analysis

In this section, we analyze the sensitivity of several hyperparameters in MSRP.
Regarding Equation (33), α , β can be determined based on the real scheduling scenario. A larger α reflects a higher priority for urgent incoming requests, while a larger β prioritizes maintaining the original user requirements. We can adjust the weights through experimentation, and select the values that yield the optimal solution. In this paper, we set up imaging tasks through simulation, without a real mission scheduling context. Therefore, we assign both α , β as 0.5.
As ω 1 and ω 2 in the reward function, we believe that adjusting ω 1 allows us to control the model’s mission insertion strategy and influence the conservativeness of task replacement. For ω 2 , we use the ratio of the changed time to VTW length as the weight, ensuring that 0 < ω 2 < ω 1 .
We conducted corresponding experiments by setting ω 1 to 0.5, 0.8, and 1 to test the effects. The experimental results in Figure 11 show that when ω 1 = 0.5 , the NIR is relatively high, but the MR also increases, resulting in suboptimal overall task revenue. When set ω 1 to 1, the insertion strategy is relatively conservative. However, overall revenue does not improve compared to ω 1 = 0.5 . We decided to set ω 1 to 0.8, as it strikes a better balance between insertion rate and modification rate, leading to the highest overall revenue.
We trained the model with PPO learning rates of 1 × 10−4, 2.5 × 10−4, and 5 × 10−4, with the training curves shown in Figure 5. The convergence speed is slower with a smaller learning rate, and faster with a larger one. As seen in Figure 12, the models with learning rates of 1 × 10−4 and 2.5 × 10−4 perform similarly, while the model with lr = 5 × 10−4 shows a slight drop in performance as the task scale increases. The model with lr = 2.5 × 10−4 also has the smallest standard deviation across metrics in different scenarios, indicating better strategy adaptation. Therefore, we chose 2.5 × 10−4 as our optimal learning rate.
Figure 13 illustrates the performance of MSRP under different task volumes and satellite counts. Here, we use sat 1 as the seed satellite and use the Walker constellation architecture, with each satellite occupying a separate orbital plane to generate the test scenarios.
We generate dynamic requests as a fixed proportion of each satellite’s original task volume. As the number of satellites increases, the total number of requests also grows. As the satellite count rises from five to eight, the performance of replanning gradually declines. Specifically, NIR and NRR decrease notably, while ER and RR show smaller declines. This is due to the increased task volume with more satellites, reducing the available space for new insertion. Despite this, overall revenue remains around 92%, demonstrating the adaptability and effectiveness of our method across varying scales.
Furthermore, when the task volume per satellite is 100 or 200, our method shows stable performance across all metrics. When the volume reaches 300, there is a slight performance decline, with NRR and RR still maintaining above 82% and 92%, respectively. This indicates that our model performs well in larger-scale scenarios.

6. Conclusions

The multi-satellite observation mission replanning problem under dynamic arrival of requests is addressed in this work. Our goal is to maximize the revenues after replanning while minimizing changes to the original plan. Currently, no studies directly model the original task sequence. To address this, we introduce a mission sequence modeling method based on time-attitude rotational position encoding, which enhances the capture of correlations between tasks. Additionally, we design a global pooling module that incorporates interval information to generate a compact representation for satellite states. We propose a multi-satellite mission replanning framework comprising two components: mission allocation and replanning. For each module, we design neural network-based policy functions and optimize them using the PPO algorithm. Computational experiments demonstrate that our approach effectively generates solutions for arriving observation requests, outperforming state-of-the-art methods in terms of new request insertion revenue rate (15.27% improvement), overall revenue rate (3.05% increase) and modification rate (1.17% reduction), while achieving shorter computation times. Furthermore, ablation studies validate the effectiveness of the proposed techniques.
Although the decoder uses a pointer network, the encoder must recalculate the mission sequence modeling at each step based on the replanning scheme. This becomes computationally expensive as the sequence length L increases. Future work could explore incremental computation techniques to improve efficiency. Furthermore, our task allocation method depends on the quality of the single-satellite replanning strategy, which may limit the generalization of the task allocation policy. A possible improvement would be to consider joint optimization across the entire system. Lastly, the hybrid action space used to generate replanning strategies requires careful design of the reward function. Future research on multi-satellite observation mission replanning may be oriented in two key directions. While our study focuses on fast mission allocation by individual satellites, simultaneous allocation by multiple satellites will require effective conflict resolution strategies when the satellite number increases. In addition, incorporating environmental factors such as cloud cover and light conditions into the replanning process could improve image quality and mission benefits, making it a valuable direction for further exploration.

Author Contributions

Conceptualization, P.L. and H.W.; methodology, P.L.; software, P.L. and P.C.; validation, P.L. and P.C.; writing—original draft preparation, P.L. and P.C.; writing—review and editing, H.W.; visualization, P.L. and P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Shanghai Aerospace Science and Technology Innovation Fund under Grant SAST2022-037.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are all contained within this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lemaître, M.; Verfaillie, G.; Jouhaud, F.; Lachiver, J.; Bataille, N. Selecting and scheduling observations of agile satellites. Aerosp. Sci. Technol. 2002, 6, 367–381. [Google Scholar] [CrossRef]
  2. Crisp, N.H.; Roberts, P.C.E.; Livadiotti, S.; Oiko, V.T.A.; Edmondson, S.; Haigh, S.J.; Huyton, C.; Sinpetru, L.A.; Smith, K.L.; Worrall, S.D.; et al. The benefits of very low earth orbit for earth observation missions. Prog. Aerosp. Sci. 2020, 117, 100619. [Google Scholar] [CrossRef]
  3. Zheng, Z. Autonomous Onboard Mission Planning for Multiple Satellite Systems. Ph.D. Thesis, Delft University of Technology, Delft, The Netherlands, 2019. [Google Scholar] [CrossRef]
  4. Cui, J.; Zhang, X. Application of a Multi-Satellite Dynamic Mission Scheduling Model Based on Mission Priority in Emergency Response. Sensors 2019, 19, 1430. [Google Scholar] [CrossRef] [PubMed]
  5. Lu, Z.; Shen, X.; Li, D.; Cheng, S.; Wang, J.; Yao, W. Super-agile satellites imaging mission planning method considering degradation of image MTF in dynamic imaging. Int. J. Appl. Earth Obs. Geoinf. 2024, 131, 103968. [Google Scholar] [CrossRef]
  6. Xiao, Y.; Zhang, S.; Yang, P.; You, M.; Huang, J. A two-stage flow-shop scheme for the multi-satellite observation and data downlink scheduling problem considering weather uncertainties. Reliab. Eng. Syst. Saf. 2019, 188, 263–275. [Google Scholar] [CrossRef]
  7. Yang, X.; Hu, M.; Huang, G.; Li, A. A Hybrid Local Replanning Strategy for Multi-Satellite Imaging Mission Planning in Uncertain Environments. IEEE Access 2023, 11, 120780–120804. [Google Scholar] [CrossRef]
  8. Wang, J.; Zhu, X.; Qiu, D.; Yang, L.T. Dynamic Scheduling for Emergency Tasks on Distributed Imaging Satellites with Task Merging. IEEE Trans. Parallel Distrib. Syst. 2014, 25, 2275–2285. [Google Scholar] [CrossRef]
  9. Chong, W.; Jun, L.; Ning, J.; Jun, W.; Hao, C. A Distributed Cooperative Dynamic Task Planning Algorithm for Multiple Satellites Based on Multi-agent Hybrid Learning. Chin. J. Aeronaut. 2011, 24, 493–505. [Google Scholar] [CrossRef]
  10. Li, H.; Li, Y.; Liu, Y.; Zhang, K.; Li, X.; Li, Y.; Zhao, S. A Multi-Objective Dynamic Mission-Scheduling Algorithm Considering Perturbations for Earth Observation Satellites. Aerospace 2024, 11, 643. [Google Scholar] [CrossRef]
  11. Yang, Y.; Liu, D. Distributed Imaging Satellite Mission Planning Based on Multi-Agent. IEEE Access 2023, 11, 65530–65545. [Google Scholar] [CrossRef]
  12. Sun, H.; Xia, W.; Hu, X.; Xu, C. Earth observation satellite scheduling for emergency tasks. J. Syst. Eng. Electron. 2019, 30, 931–945. [Google Scholar] [CrossRef]
  13. He, L.; Liu, X.L.; Chen, Y.W.; Xing, L.N.; Liu, K. Hierarchical scheduling for real-time agile satellite task scheduling in a dynamic environment. Adv. Space Res. 2019, 63, 897–912. [Google Scholar] [CrossRef]
  14. Li, H.; Li, Y.; Meng, Q.Q.; Li, X.; Shao, L.; Zhao, S. An onboard periodic rescheduling algorithm for satellite observation scheduling problem with common dynamic tasks. Adv. Space Res. 2024, 73, 5242–5253. [Google Scholar] [CrossRef]
  15. Li, K.; Zhang, T.; Wang, R. Deep Reinforcement Learning for Multiobjective Optimization. IEEE Trans. Cybern. 2021, 51, 3103–3114. [Google Scholar] [CrossRef] [PubMed]
  16. Pemberton, J.C.; Greenwald, L. On the need for dynamic scheduling of imaging satellites. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2002, 34, 165–171. [Google Scholar]
  17. Liang, J.; Zhu, Y.H.; Luo, Y.Z.; Zhang, J.C.; Zhu, H. A precedence-rule-based heuristic for satellite onboard activity planning. Acta Astronaut. 2021, 178, 757–772. [Google Scholar] [CrossRef]
  18. Wen, J.; Liu, X.; He, L. Real-time online rescheduling for multiple agile satellites with emergent tasks. J. Syst. Eng. Electron. 2021, 32, 1407–1420. [Google Scholar] [CrossRef]
  19. Han, C.; Gu, Y.; Wu, G.; Wang, X. Simulated Annealing-Based Heuristic for Multiple Agile Satellites Scheduling Under Cloud Coverage Uncertainty. IEEE Trans. Syst. Man Cybern.-Syst. 2023, 53, 2863–2874. [Google Scholar] [CrossRef]
  20. Liu, S.; Chen, Y.; Xing, L.; Guo, X. Time-dependent autonomous task planning of agile imaging satellites. J. Intell. Fuzzy Syst. 2016, 31, 1365–1375. [Google Scholar] [CrossRef]
  21. Wei, L.; Xing, L.; Wan, Q.; Song, Y.; Chen, Y. A Multi-objective Memetic Approach for Time-dependent Agile Earth Observation Satellite Scheduling Problem. Comput. Ind. Eng. 2021, 159, 107530. [Google Scholar] [CrossRef]
  22. Peng, G.; Song, G.; He, Y.; Yu, J.; Xiang, S.; Xing, L.; Vansteenwegen, P. Solving the Agile Earth Observation Satellite Scheduling Problem With Time-Dependent Transition Times. IEEE Trans. Syst. Man Cybern.-Syst. 2022, 52, 1614–1625. [Google Scholar] [CrossRef]
  23. Du, B.; Li, S. A new multi-satellite autonomous mission allocation and planning method. Acta Astronaut. 2019, 163, 287–298. [Google Scholar] [CrossRef]
  24. Du, Y.; Wang, T.; Xin, B.; Wang, L.; Chen, Y.; Xing, L. A Data-Driven Parallel Scheduling Approach for Multiple Agile Earth Observation Satellites. IEEE Trans. Evol. Comput. 2020, 24, 679–693. [Google Scholar] [CrossRef]
  25. Song, Y.; Xing, L.; Wang, M.; Yi, Y.; Xiang, W.; Zhang, Z. A knowledge-based evolutionary algorithm for relay satellite system mission scheduling problem. Comput. Ind. Eng. 2020, 150, 106830. [Google Scholar] [CrossRef]
  26. Zhu, W.; Hu, X.; Xia, W.; Jin, P. A two-phase genetic annealing method for integrated Earth observation satellite scheduling problems. Soft Comput. 2019, 23, 181–196. [Google Scholar] [CrossRef]
  27. Wang, Y.; Liu, D.; Liu, J. A Bilevel Programming Model for Multi-Satellite Cooperative Observation Mission Planning. IEEE Access 2024, 12, 145439–145449. [Google Scholar] [CrossRef]
  28. Bianchessi, N.; Cordeau, J.F.; Desrosiers, J.; Laporte, G.; Raymond, V. A heuristic for the multi-satellite, multi-orbit and multi-user management of Earth observation satellites. Eur. J. Oper. Res. 2007, 177, 750–762. [Google Scholar] [CrossRef]
  29. Li, J.; Chen, Y.; Liu, X.; He, R. JADE implemented multi-agent based platform for multiple autonomous satellite system. In Proceedings of the 2018 SpaceOps Conference, Marseille, France, 28 May–1 June 2018; p. 2349. [Google Scholar] [CrossRef]
  30. Qi, J.; Guo, J.; Wang, M.; Wu, C. A Cooperative Autonomous Scheduling Approach for Multiple Earth Observation Satellites With Intensive Missions. IEEE Access 2021, 9, 61646–61661. [Google Scholar] [CrossRef]
  31. Liu, Y.; Chen, Q.; Li, C.; Wang, F. Mission planning for Earth observation satellite with competitive learning strategy. Aerosp. Sci. Technol. 2021, 118, 107047. [Google Scholar] [CrossRef]
  32. Chen, Y.; Tian, G.; Guo, J.; Huang, J. Task Planning for Multiple-Satellite Space-Situational-Awareness Systems. Aerospace 2021, 8, 73. [Google Scholar] [CrossRef]
  33. Yang, W.; He, L.; Liu, X.; Chen, Y. Onboard coordination and scheduling of multiple autonomous satellites in an uncertain environment. Adv. Space Res. 2021, 68, 4505–4524. [Google Scholar] [CrossRef]
  34. Li, J.; Yang, A.; Jing, N.; Hu, W. Coordinated planning of space-aeronautics earth-observing based on CSP theory. In Proceedings of the 2013 21st International Conference on Geoinformatics, Kaifeng, China, 20–22 June 2013. [Google Scholar] [CrossRef]
  35. Liu, D.; Dou, L.; Zhang, R.; Zhang, X.; Zong, Q. Multi-Agent Reinforcement Learning-Based Coordinated Dynamic Task Allocation for Heterogenous UAVs. IEEE Trans. Veh. Technol. 2023, 72, 4372–4383. [Google Scholar] [CrossRef]
  36. Dalin, L.; Haijiao, W.; Zhen, Y.; Yanfeng, G.; Shi, S. An Online Distributed Satellite Cooperative Observation Scheduling Algorithm Based on Multiagent Deep Reinforcement Learning. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1901–1905. [Google Scholar] [CrossRef]
  37. Saeed, A.K.; Holguin, F.; Yasin, A.S.; Johnson, B.A.; Rodriguez, B.M. Multi-Agent and Multi-Target Reinforcement Learning for Satellite Sensor Tasking. In Proceedings of the 2024 IEEE Aerospace Conference, Big Sky, MT, USA, 2–9 March 2024; pp. 1–13. [Google Scholar] [CrossRef]
  38. Zhang, G.; Li, X.; Hu, G.; Li, Y.; Wang, X.; Zhang, Z. MARL-Based Multi-Satellite Intelligent Task Planning Method. IEEE Access 2023, 11, 135517–135528. [Google Scholar] [CrossRef]
  39. Xhafa, F.; Ip, A.W. Optimisation problems and resolution methods in satellite scheduling and space-craft operation: A survey. Enterp. Inf. Syst. 2021, 15, 1022–1045. [Google Scholar] [CrossRef]
  40. Cho, D.H.; Kim, J.H.; Choi, H.L.; Ahn, J. Optimization-Based Scheduling Method for Agile Earth-Observing Satellite Constellation. J. Aerosp. Inf. Syst. 2018, 15, 611–626. [Google Scholar] [CrossRef]
  41. Kim, J.; Ahn, J.; Choi, H.L.; Cho, D.H. Task Scheduling of Agile Satellites with Transition Time and Stereoscopic Imaging Constraints. J. Aerosp. Inf. Syst. 2020, 17, 285–293. [Google Scholar] [CrossRef]
  42. Li, P.; Li, J.; Li, H.; Zhang, S.; Yang, G. Graph Based Task Scheduling Algorithm for Earth Observation Satellites. In Proceedings of the 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 1–7. [Google Scholar] [CrossRef]
  43. Gabrel, V.; Moulet, A.; Murat, C.; Paschos, V.T. A new single model and derived algorithms for the satellite shot planning problem using graph theory concepts. Ann. Oper. Res. 1997, 69, 115–134. [Google Scholar] [CrossRef]
  44. Valicka, C.G.; Garcia, D.; Staid, A.; Watson, J.P.; Hackebeil, G.; Rathinam, S.; Ntaimo, L. Mixed-integer programming models for optimal constellation scheduling given cloud cover uncertainty. Eur. J. Oper. Res. 2019, 275, 431–445. [Google Scholar] [CrossRef]
  45. Song, Y.; Xing, L.; Chen, Y. Two-stage hybrid planning method for multi-satellite joint observation planning problem considering task splitting. Comput. Ind. Eng. 2022, 174, 108795. [Google Scholar] [CrossRef]
  46. Chen, Y.; Lu, J.; He, R.; Ou, J. An Efficient Local Search Heuristic for Earth Observation Satellite Integrated Scheduling. Appl. Sci. 2020, 10, 5616. [Google Scholar] [CrossRef]
  47. Luo, Q.; Peng, W.; Wu, G.; Xiao, Y. Orbital Maneuver Optimization of Earth Observation Satellites Using an Adaptive Differential Evolution Algorithm. Remote Sens. 2022, 14, 1966. [Google Scholar] [CrossRef]
  48. Li, Y.; Luo, J.; Zhang, W.; Xiang, F. Genetic-evolutionary bi-level mission planning algorithm for multi-satellite cooperative observation. Syst. Eng. Electron. 2024, 46, 2044. [Google Scholar]
  49. Zheng, Z.; Guo, J.; Gill, E. Onboard autonomous mission re-planning for multi-satellite system. Acta Astronaut. 2018, 145, 28–43. [Google Scholar] [CrossRef]
  50. Mao, H.; Alizadeh, M.; Menache, I.; Kandula, S. Resource Management with Deep Reinforcement Learning. In Proceedings of the 15th ACM Workshop on Hot Topics in Networks, Atlanta GA USA, 9–10 November 2016; Association for Computing Machinery: New York, NY, USA, 2016. HotNets ’16. pp. 50–56. [Google Scholar] [CrossRef]
  51. Chen, H.; Luo, Z.; Peng, S.; Wu, J.; Li, J. HiPGen: An approach for fast generation of multi-satellite observation plans via a hierarchical multi-channel transformer network. Adv. Space Res. 2022, 69, 3103–3116. [Google Scholar] [CrossRef]
  52. Gu, Y.; Han, C.; Chen, Y.; Xing, W.W. Mission Replanning for Multiple Agile Earth Observation Satellites Based on Cloud Coverage Forecasting. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 594–608. [Google Scholar] [CrossRef]
  53. Wang, H.; Yang, Z.; Zhou, W.; Li, D. Online scheduling of image satellites based on neural networks and deep reinforcement learning. Chin. J. Aeronaut. 2019, 32, 1011–1019. [Google Scholar] [CrossRef]
  54. Schuetz, M.J.; Brubaker, J.K.; Katzgraber, H.G. Combinatorial optimization with physics-inspired graph neural networks. Nat. Mach. Intell. 2022, 4, 367–377. [Google Scholar] [CrossRef]
  55. Wang, Z.; Hu, X.; Ma, H.; Xia, W. Learning multi-satellite scheduling policy with heterogeneous graph neural network. Adv. Space Res. 2024, 73, 2921–2938. [Google Scholar] [CrossRef]
  56. Feng, X.; Li, Y.; Xu, M. Multi-satellite cooperative scheduling method for large-scale tasks based on hybrid graph neural network and metaheuristic algorithm. Adv. Eng. Inform. 2024, 60, 102362. [Google Scholar] [CrossRef]
  57. Nazari, M.; Oroojlooy, A.; Takáč, M.; Snyder, L.V. Reinforcement learning for solving the vehicle routing problem. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018. NIPS’18. pp. 9861–9871. [Google Scholar]
  58. Kool, W.; van Hoof, H.; Welling, M. Attention, Learn to Solve Routing Problems! In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019.
  59. Liu, Z.; Xiong, W.; Han, C.; Yu, X. Deep Reinforcement Learning with Local Attention for Single Agile Optical Satellite Scheduling Problem. Sensors 2024, 24, 6396. [Google Scholar] [CrossRef]
  60. Chen, M.; Du, Y.; Tang, K.; Xing, L.; Chen, Y.; Chen, Y. Learning to Construct a Solution for the Agile Satellite Scheduling Problem With Time-Dependent Transition Times. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 5949–5963. [Google Scholar] [CrossRef]
  61. Liang, J.; Liu, J.P.; Sun, Q.; Zhu, Y.H.; Zhang, Y.C.; Song, J.G.; He, B.Y. A Fast Approach to Satellite Range Rescheduling Using Deep Reinforcement Learning. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 9390–9403. [Google Scholar] [CrossRef]
  62. Long, Y.; Shan, C.; Shang, W.; Li, J.; Wang, Y. Deep Reinforcement Learning-Based Approach With Varying-Scale Generalization for the Earth Observation Satellite Scheduling Problem Considering Resource Consumptions and Supplements. IEEE Trans. Aerosp. Electron. Syst. 2024, 60, 2572–2585. [Google Scholar] [CrossRef]
  63. Jiang, Y.; Yang, Y.; Xu, Y.; Wang, E. Spatial-Temporal Interval Aware Individual Future Trajectory Prediction. IEEE Trans. Knowl. Data Eng. 2024, 36, 5374–5387. [Google Scholar] [CrossRef]
  64. Agrawal, A.; Bedi, A.S.; Manocha, D. RTAW: An Attention Inspired Reinforcement Learning Method for Multi-Robot Task Allocation in Warehouse Environments. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 1393–1399. [Google Scholar] [CrossRef]
  65. Li, P.; Wang, H.; Zhang, Y.; Pan, R. Mission planning for distributed multiple agile Earth observing satellites by attention-based deep reinforcement learning method. Adv. Space Res. 2024, 74, 2388–2404. [Google Scholar] [CrossRef]
  66. Sun, H.; Xia, W.; Wang, Z.; Hu, X. Agile earth observation satellite scheduling algorithm for emergency tasks based on multiple strategies. J. Syst. Sci. Syst. Eng. 2021, 30, 626–646. [Google Scholar] [CrossRef]
  67. Ou, J.; Xing, L.; Yao, F.; Li, M.; Lv, J.; He, Y.; Song, Y.; Wu, J.; Zhang, G. Deep reinforcement learning method for satellite range scheduling problem. Swarm Evol. Comput. 2023, 77, 101233. [Google Scholar] [CrossRef]
Figure 1. Scenario and framework of multi-satellite mission replanning.
Figure 1. Scenario and framework of multi-satellite mission replanning.
Sensors 25 01707 g001
Figure 2. The structure of mission sequence modeling and dynamic request modeling method. (a) Mission sequence modeling. (b) Dynamic request modeling.
Figure 2. The structure of mission sequence modeling and dynamic request modeling method. (a) Mission sequence modeling. (b) Dynamic request modeling.
Sensors 25 01707 g002
Figure 3. The schematic diagram of the rotational positional encoding mechanism.
Figure 3. The schematic diagram of the rotational positional encoding mechanism.
Sensors 25 01707 g003
Figure 4. The actor network of MSRP. (a) Mission allocation. (b) Mission replanning.
Figure 4. The actor network of MSRP. (a) Mission allocation. (b) Mission replanning.
Sensors 25 01707 g004
Figure 5. Training curves of MSRP allocation and replanning.
Figure 5. Training curves of MSRP allocation and replanning.
Sensors 25 01707 g005
Figure 6. Experiment results with different single-satellite replanning algorithms under different numbers of missions. (a) Comparison of 4 metrics. (b) Comparison of Modification rate and running time.
Figure 6. Experiment results with different single-satellite replanning algorithms under different numbers of missions. (a) Comparison of 4 metrics. (b) Comparison of Modification rate and running time.
Sensors 25 01707 g006
Figure 7. Experiment results with different single-satellite replanning algorithms under different new request ratio. (a) Comparison of of 4 metrics. (b) Comparison of Modification rate and running time.
Figure 7. Experiment results with different single-satellite replanning algorithms under different new request ratio. (a) Comparison of of 4 metrics. (b) Comparison of Modification rate and running time.
Sensors 25 01707 g007
Figure 8. Three mission insertion cases.
Figure 8. Three mission insertion cases.
Sensors 25 01707 g008
Figure 9. Experiment results with multi-satellite replanning algorithms under various number of missions. Subfigures (ad) respectively present the comparison of NIR, NRR, ER, and RR results among the four algorithms under varying task quantities. Subfigure (e) displays the MR performance and running time of the different algorithms.
Figure 9. Experiment results with multi-satellite replanning algorithms under various number of missions. Subfigures (ad) respectively present the comparison of NIR, NRR, ER, and RR results among the four algorithms under varying task quantities. Subfigure (e) displays the MR performance and running time of the different algorithms.
Sensors 25 01707 g009
Figure 10. Ablation experiment results under different number of missions. Subfigures (ad) respectively present the comparison of NIR, NRR, ER, and RR result under varying task quantities. Subfigure (e) displays the comparison of MR performance and running time.
Figure 10. Ablation experiment results under different number of missions. Subfigures (ad) respectively present the comparison of NIR, NRR, ER, and RR result under varying task quantities. Subfigure (e) displays the comparison of MR performance and running time.
Sensors 25 01707 g010
Figure 11. Sensitivity analysis of ω 1 with different number of missions.
Figure 11. Sensitivity analysis of ω 1 with different number of missions.
Sensors 25 01707 g011
Figure 12. Sensitivity analysis of learning rate.
Figure 12. Sensitivity analysis of learning rate.
Sensors 25 01707 g012
Figure 13. Sensitivity analysis of the number of satellite with different number of missions.
Figure 13. Sensitivity analysis of the number of satellite with different number of missions.
Sensors 25 01707 g013
Table 1. The orbital parameters of 6 satellites.
Table 1. The orbital parameters of 6 satellites.
Satellitea (km)eI (°) ω (°) Ω (°)M (°)
Sat17200.00.00062796.57600175.720.0750
Sat27200.00.00062796.57600145.7230.0750
Sat37200.00.00062796.57600115.7260.0750
Sat47200.00.00062796.5760085.7290.0750
Sat57200.00.00062796.5760055.72120.0750
Sat67200.00.00062796.5760025.72150.0750
Table 2. The parameters of simulated observation requests.
Table 2. The parameters of simulated observation requests.
ParametersValueUnitDescription
d i j [ 5 , 25 ) sRequired execution duration
p i [0.1, 0.9]1Revenue (priority) of request
a i j [−50, 50]1Side swing angle range
C m e m j 1000GBMemory capacity
m e m i 1 × d i j GBMemory consumption
C p o w e r j 512W· hPower capacity
p o w e r i [ 0.3 , 0.8 ) × d i j W· hPower consumption
v j 1s/degTime per degree of transition
Table 3. The hyperparameters of algorithms.
Table 3. The hyperparameters of algorithms.
HyperparametersReplanningAllocation
Number of envs12832
Total timesteps64 × 10564 × 105
Mini-batches328
Learning rate2.5 × 10−42.5 × 10−4
Learning rate decayCosineAnnealingStepAnnealing
γ 0.990.99
GAE λ 0.950.95
Update epochs84
Cliping coefficient0.20.1
Entropy coefficiant0.010.01
Embedding dimension128128
Gate embedding dimension3232
Table 4. Statistical significance and confidence intervals for single-satellite replanning algorithm performance.
Table 4. Statistical significance and confidence intervals for single-satellite replanning algorithm performance.
Algorithm
Comparison
Statistical
Indicators
NIRNRRERRRMRTimeSignificant
Difference
(y/n)
MSRP|MSLRWilcoxon signed
ranks test p-value
0.00330.00330.00330.00330.0010.0005y
Confidence
Interval (X1–X2)
[0.0425,
0.0902]
[0.0489,
0.1031]
[0.0359,
0.0830]
[0.0265,
0.0642]
[−0.0934,
−0.0312]
[−5.7909,
−1.9408]
MSRP|FIAWilcoxon signed
ranks test p-value
0.00050.00050.00050.00050.00150.0005y
Confidence
Interval (X1–X2)
[0.1791,
0.2486]
[0.1392,
0.2057]
[0.1017,
0.1769]
[0.0557,
0.1114]
[0.0076,
0.0298]
[−5.8088,
−1.9696]
MSRP|TSRWilcoxon signed
ranks test p-value
0.00770.00770.00760.00770.00050.0005y
Confidence
Interval (X1–X2)
[0.0077,
0.0390]
[0.0090,
0.0538]
[0.0032,
0.0321]
[0.0032,
0.0299]
[−0.0316,
−0.0116]
[−2.2554,
−0.7101]
Table 5. Statistical significance and confidence intervals for multi-satellite replanning algorithm performance.
Table 5. Statistical significance and confidence intervals for multi-satellite replanning algorithm performance.
Algorithm
Comparison
Statistical
Indicators
NIRNRRERRRMRTimeSignificant
Difference (y/n)
MSRP|TMRWilcoxon signed
ranks test p-value
0.00040.00030.00040.00030.00760y
Confidence
Interval (X1–X2)
[0.0986,
0.1787]
[0.1111,
0.1915]
[0.0177,
0.0474]
[0.0174,
0.0412]
[−0.0200,
−0.0035]
[−1.3985,
−0.7713]
MSRP|RDEWilcoxon signed
ranks test p-value
000.00040.00030.63970y
Confidence
Interval (X1–X2)
[0.2239,
0.3071]
[0.1527,
0.2239]
[0.0177,
0.0474]
[0.0174,
0.0412]
[−0.0339,
0.0213]
[−686.6162,
−452.3476]
MSRP|PRGWilcoxon signed
ranks test p-value
0.00230.001300.039400y
Confidence
Interval (X1–X2)
[−0.0407,
−0.0134]
[−0.0444,
−0.0145]
[0.0484,
0.0850]
[−0.0096,
0.0022]
[−0.2271,
−0.1573]
[−40.1716,
−24.7170]
Table 6. Statistical significance and confidence intervals for ablation experiments.
Table 6. Statistical significance and confidence intervals for ablation experiments.
Algorithm
Comparison
Statistical
Indicators
NIRNRRERRRMRTimeSignificant
Difference (y/n)
MSRP|
Ablation-MLP
Wilcoxon signed
ranks test p-value
0.00020.00010.00070.00020.12970.1084y
Confidence
Interval (X1–X2)
[0.1291,
0.3426]
[0.1290,
0.3346]
[0.0170,
0.0433]
[0.0201,
0.0463]
[−0.0002,
0.0232]
[−0.0273,
1.2680]
MSRP|
Ablation-pooling
Wilcoxon signed
ranks test p-value
0.00040000.39270.0001y
Confidence
Interval (X1–X2)
[0.0113,
0.0411]
[0.0355,
0.0817]
[0.0098,
0.0228]
[0.0103,
0.0241]
[−0.0096,
0.0252]
[0.3827,
1.5095]
MSRP|Ablation-
Transformer-RoPE
Wilcoxon signed
ranks test p-value
0.00130.0010.00230.00160.00010.0001y
Confidence
Interval (X1–X2)
[0.0470,
0.1266]
[0.0375,
0.1196]
[0.0113,
0.0363]
[0.0121,
0.0371]
[−0.0242,
−0.0105]
[0.2066,
0.8139]
MSRP|Ablation-
Transformer-PE
Wilcoxon signed
ranks test p-value
000.00010.00010.02680y
Confidence
Interval (X1–X2)
[0.1077,
0.1899]
[0.0976,
0.1791]
[0.0211,
0.0476]
[0.0208,
0.0450]
[−0.0092,
0.0024]
[0.4075,
1.0628]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, P.; Cui, P.; Wang, H. Mission Sequence Model and Deep Reinforcement Learning-Based Replanning Method for Multi-Satellite Observation. Sensors 2025, 25, 1707. https://doi.org/10.3390/s25061707

AMA Style

Li P, Cui P, Wang H. Mission Sequence Model and Deep Reinforcement Learning-Based Replanning Method for Multi-Satellite Observation. Sensors. 2025; 25(6):1707. https://doi.org/10.3390/s25061707

Chicago/Turabian Style

Li, Peiyan, Peixing Cui, and Huiquan Wang. 2025. "Mission Sequence Model and Deep Reinforcement Learning-Based Replanning Method for Multi-Satellite Observation" Sensors 25, no. 6: 1707. https://doi.org/10.3390/s25061707

APA Style

Li, P., Cui, P., & Wang, H. (2025). Mission Sequence Model and Deep Reinforcement Learning-Based Replanning Method for Multi-Satellite Observation. Sensors, 25(6), 1707. https://doi.org/10.3390/s25061707

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop