Abstract
This paper investigates the single agile optical satellite scheduling problem, which has received increasing attention due to the rapid growth in earth observation requirements. Owing to the complicated constraints and considerable solution space of this problem, the conventional exact methods and heuristic methods, which are sensitive to the problem scale, demand high computational expenses. Thus, an efficient approach is demanded to solve this problem, and this paper proposes a deep reinforcement learning algorithm with a local attention mechanism. A mathematical model is first established to describe this problem, which considers a series of complex constraints and takes the profit ratio of completed tasks as the optimization objective. Then, a neural network framework with an encoder–decoder structure is adopted to generate high-quality solutions, and a local attention mechanism is designed to improve the generation of solutions. In addition, an adaptive learning rate strategy is proposed to guide the actor–critic training algorithm to dynamically adjust the learning rate in the training process to enhance the training effectiveness of the proposed network. Finally, extensive experiments verify that the proposed algorithm outperforms the comparison algorithms in terms of solution quality, generalization performance, and computation efficiency.
1. Introduction
Agile optical satellites (AOSs), as a new generation of optical satellites, have superior attitude maneuverability, which can adjust attitude flexibly and quickly on the three axes of pitch, roll and yaw. Over past years, they have played increasingly significant roles in many fields, such as environmental monitoring, marine development, agricultural production, city planning, and military reconnaissance [1,2]. The agile characteristics improve the observation capability of AOSs but also lead to a more complex attitude maneuver process. With the surge in earth observation demand, scheduling an agile optical satellite has become a challenging problem.
Compared with traditional optical satellites, AOSs can observe the target during a much longer observable period due to its ability to rotate around the pitch axis, as shown in Figure 1. The period during which the satellite can observe the target is called the visible time window (VTW), and the duration of the actual observation event is called the observation window (OW). In this study, the AOS adopts a common push-broom observation mode [3] and adjusts its attitude only by rolling and pitching. In the observation process of an AOS, its pitch angle determines the start time of the OW, and its roll angle is determined by the relative position of the AOS and the target. Thus, in a VTW, the pitch angle is variable but the roll angle can be regarded as a fixed value. When an AOS observes two adjacent targets, a short period is needed for the attitude adjustment, which is called the transition time. The time interval between two successive observation actions must not be less than their required transition time [4]. Only if the observation time of the previous task is determined can the observation time of the next one be determined. Therefore, the observation time of multiple tasks must be determined one by one, which is a sequential decision process.
Figure 1.
Observation process of the traditional optical satellite and AOS.
The purpose of the single AOS scheduling problem (SAOSSP) is to reasonably arrange the observation sequence and observation time of tasks [3]. The SAOSSP has been proven as an NP-hard problem [5], exhibiting the characteristics of a combination explosion, and its solving time grows exponentially with the problem scale [6]. The complexity of this problem is mainly reflected in three aspects. First, each target may have several VTWs in different orbits, causing multiple possibilities of the VTW selection, and the observation event can start at any time within the selected VTW, causing various optional possibilities of the OW. Second, owing to the complicated constraints and the considerable solution space, an enormous number of constraint-checking steps and a thorough search in the solution space are needed [7], demanding high computational expenses. Third, this problem is typically an oversubscribed scheduling problem where only partial tasks have the chance to be accomplished [8,9], and the increase of the problem scale can further expand its computational complexity. Therefore, the SAOSSP is hard and complex to solve, and an efficient scheduling method is badly required to accomplish the more observation tasks and reach the maximum observation profit.
Over the past few decades, extensive research has been carried out, and many contributions have been made to solve the agile satellite scheduling problem. Existing approaches can be roughly distributed into three categories: exact algorithms, heuristic algorithms and deep reinforcement learning (DRL) algorithms. Some researchers adopt exact algorithms to solve this problem. Lemaître et al. [5] gave the general description of the agile satellite scheduling problem for the first time and proposed a dynamic programming algorithm. Chu et al. [10] presented a branch and bound algorithm with a look-ahead method and three pruning strategies to tackle the simplified problem in a small-size instance. Peng et al. [11] considered the time-dependent profits and presented an adaptive-directional dynamic programming algorithm with decremental state space relaxation for the single-orbit scheduling of the single agile satellite. The above studies summarily prove that exact algorithms can explore the whole search space and obtain the optimal solution, and they are applicable to the single-orbit or small-scale satellite scheduling. However, as the problem scale expands, the computational cost of the exact algorithms tends to be unacceptable because of the NP-hard characteristic and the complex constraints.
Differing from exact algorithms, heuristic algorithms can iteratively search for a good solution with relatively lower computational cost, and they have been extensively applied to the agile satellite scheduling due to the excellent abilities of exploration and exploitation, such as genetic algorithms (GAs) [12,13], particle swarm optimization (PSO) algorithms [14,15], ant colony optimization (ACO) algorithms [16,17], artificial bee colony (ABC) algorithms [18] and so on. Chatterjee and Tharmarasa [19] formulated a mixed-integer non-linear optimization model and proposed an elitist mixed coded genetic algorithm algorithm to solve the agile satellite scheduling problem. Du et al. [20] considered the drift angle constraint of the observation instrument and developed an improved ant colony algorithm based on a sensational and consciousness strategy to solve the area target observation problem of agile satellites. Yang et al. [21] improved the three search phases of the basic ABC algorithm and presented a hybrid discrete artificial bee colony (HDABC) algorithm to address the satellite scheduling problem. These heuristic algorithms can optimize solutions through superior search mechanisms and iterative updates of population, but the issue of considerable computational time and the convergence problem still exist.
To avoid the above shortcomings, researchers demand for a novel, efficient, and non-iterative method to solve the agile satellite scheduling problem. With the development of artificial intelligence technology, deep reinforcement learning (DRL) methods have achieved excellent results in many fields [22,23,24,25], especially in large-scale combination optimization problems [26,27]. Because the agile satellite scheduling problem can be formulated as a sequential decision problem [28,29], DRL algorithms can be adopted to replace heuristic algorithms, which can directly generate solutions without an iterative search. Chen et al. [30] developed an end-to-end network model with a convolution neural network (CNN) as the encoder, a gated recurrent unit (GRU) as the decoder, and an attention mechanism to obtain the relevance between the output task and each input task, and this model achieved good results in the small-scale scheduling instances. Zhao et al. [31] proposed a two-phase neural combination optimization method, which adopted a Pointer Network (PtrNet) to generate a permutation of executable tasks and a deep deterministic policy gradient algorithm to determine the observation time of selected tasks. Wei et al. [32] proposed a DRL and parameter transfer-based approach (RLPT), consisting of a task encoder composed of a GRU, a feature encoder composed of a CNN, and a decoder composed of an attention mechanism. These studies have preliminarily verified the feasibility and effectiveness of DRL, but there are still some weaknesses. First, the neural networks in these methods can only deal with the inputs of fixed dimensions, and no effective approach is provided to handle the inputs of variable dimensions, especially VTWs, the number of which may be different for different tasks. Second, the above methods based on the encoder–decoder structure all adopt the global attention mechanism whose computation is a bit large, and they do not pay attention to the internal connections among tasks. Third, the above studies only train and validate the proposed methods on the instances of a certain scale, and they do not really apply them to the large-scale problems. Therefore, it is necessary to further explore the improvement of DRL to overcome these deficiencies and efficiently solve the SAOSSP.
In this study, we propose a deep reinforcement learning algorithm with a local attention mechanism (DRLLA) to solve the SAOSP in a non-iterative manner. The primary contributions are summarized as follows:
- (1)
- A neural network framework based on an encoder–decoder structure is designed to construct a high-quality scheduling solution in an auto-regression manner. Furthermore, a local attention mechanism is proposed to guide the network model to focus on the more promising tasks in the solution construction process by narrowing the candidate task range into a local one, which greatly improves the solution quality.
- (2)
- An adaptive learning rate strategy is developed to guide the actor–critic training algorithm to train the proposed network model. In the training process, the adaptive learning rate strategy appropriately reduces or increases the learning rate of the training algorithm according to the reward curve trend, obtaining better training effectiveness.
- (3)
- Extensive experiments are conducted to verify the proposed algorithm with our created SAOSSP datasets, which contain numerous various-scale instances and can be publicly available for related studies. The experimental results validate that the proposed algorithm has superior performance in solution quality, generalization performance, and computational efficiency over the comparison algorithms.
The remainder of this paper is organized as follows. In Section 2, we describe the SAOSSP in detail and build the corresponding mathematical model with complicated constraints and an optimization objective. Section 3 presents a neural network with a local attention mechanism for the SAOSSP and an actor–critic training algorithm with an adaptive learning rate for the training of the proposed network. Section 4 presents the experimental results, and Section 5 presents the conclusions of the study with a summary and directions for future work.
2. Problem Description and Model
Based on the previous literature, the SAOSP can be described as a combination optimization problem that aims to arrange the observation sequence of candidate tasks and determine their observation time in the selected VTWs to gain the maximum observation profit. In this section, a detailed description is given and a mathematical model of SAOSP is built with complex constraints and an optimization objective.
2.1. Assumptions
In practice, AOS scheduling is a complicated process with a series of practical constraints. In order to simplify and standardize the SAOSSP, it is necessary to make some reasonable assumptions based on the actual engineering background and the previous literature [33,34,35,36,37], which are listed as follows:
- (1)
- The AOS observes only one target every time;
- (2)
- All tasks are point targets or small strip targets which can be observed in one pass;
- (3)
- All tasks are executed at most once;
- (4)
- At the initial moment of each orbital cycle, the AOS completely releases the memory space for the storage of observation data, and its energy for the observation and attitude maneuver is full;
- (5)
- The influence of cloud cover is not considered.
2.2. Notations and Variables
The inputs of the SAOSSP contain task requirements and satellite parameters, and the parameters of VTWs are calculated according to the target location and the orbital parameters of the AOS. The output solutions contain the permutation of executable tasks and the corresponding observation action sequence of the AOS. The variables which are used to describe the SAOSSP are defined in Table 1.
Table 1.
Variable definition.
In addition to the above variables, several other concepts and variables are involved in this problem. Firstly, the attitude adjustment on each axis can be simplified as “constant acceleration—constant speed—constant deceleration” or “constant acceleration—constant deceleration” [38], and the transition time is formulated as below:
where is the adjusted angle value on one axis, a is the angular acceleration value, ω is the maximum angular velocity value, and ST is the stability time after adjustment. If the roll angle changes by and the pitch angle changes by , the total transition time will be .
Secondly, in the VTW , the relationship between the pitch angle and the observation start time can be formulated as a linear function [39]:
where , and .
Thirdly, the image quality is determined by the observation start time [4], so we respectively define and as the earliest start time and the latest start time meeting the minimum image quality requirement in the VTW . In addition, the earliest start time corresponds to the maximum pitch angle meeting the minimum image quality requirement, and the latest start time corresponds to the minimum one . All of them can be formulated as below:
Finally, three decision variables are defined to establish the mathematical model of the SAOSSP, which is shown in the following formulae:
2.3. Mathematical Formulation
The mathematical formulation of the SAOSSP is defined as follows, where the optimization objective is to maximize the profit ratio of the completed tasks and complicated constraints are taken into consideration.
subject to
Equation (11) denotes the execution uniqueness constraint that every task is executed at most once. Equation (12) denotes the lighting constraint that every candidate VTW must meet the lighting condition. Equation (13) denotes the image quality constraint that the observation start time must satisfy the minimum image quality requirement. Equation (14), where u and v separately denote the VTW indexes of the two tasks, indicates the transition time constraint that the time interval between two observation actions must be sufficient for attitude adjustment. Equation (15) denotes the memory constraint whereby the total memory consumption for observation cannot exceed the maximum in each orbit cycle. Equation (16) denotes the energy constraint whereby the total energy consumption for attitude adjustment and observation cannot exceed the maximum in each orbit cycle.
3. Method
In this section, we propose a deep reinforcement learning algorithm with a local attention mechanism (DRLLA) to address the SAOSSP. The architecture of the designed neural network, its crucial components, and a training approach are described in turn.
3.1. Architecture of the Proposed Neural Network
The architecture of the proposed neural network is depicted in Figure 2, which is an end-to-end framework with an encoder–decoder structure. The neural network is composed of five components: a static embedding layer, a static encoder, a dynamic embedding layer, a dynamic encoder and a decoder. In addition, a local attention mechanism is proposed to improve the generation of solutions.
Figure 2.
Architecture of the proposed neural network. For task , is its static embedding form, and is its static encoding form. At step n, is the dynamic embedding form mapped by , and is its dynamic encoding form. is the output of the dynamic encoder. is the hidden state of the LSTM cell in the decoder.
In the general SAOSSP, given a set of tasks , the neural network is used as the policy network to extract features from inputs and generate a permutation of the executable tasks and a corresponding observation action sequence in an auto-regression way, both of which compose the solution of the problem. The inputs of this problem comprise two parts: task information and satellite state information. Task information is static and consists of time window information and requirement information. For the task , its time window information is denoted by a set , and its requirement information is denoted by a vector . Satellite state information is dynamic and changes every time a task is executed. After the AOS executes the task through the observation action , its state information is denoted as a vector , where n is the step number and the observation end time is also the start time of the free state. The output permutation of the executable tasks is denoted as , where is the task index and N is the length of output permutation. Accordingly, the observation action sequence of the AOS is . Obviously, the elements of can be obtained in . To better utilize the above input information, the designed neural network must be able to process both static and dynamic information simultaneously. Therefore, the proposed neural network adopts a static embedding layer and a static encoder to handle the static inputs and a dynamic embedding layer and a dynamic encoder to deal with the dynamic inputs on the basis of the basic encoder–decoder structure. The above inputs have been normalized before they are formally inputted.
Formally, the set of inputs is denoted as . Each input is denoted as a sequence of tuples , where is the static element, and is the dynamic one at the decoding step n. In addition, denotes the state set of all inputs at the decoding step n, and is the initial state of the inputs. is the final permutation, where is an initial virtual tuple composed of the virtual static elements and the initial satellite state vector. is the decoded permutation up to the step n.
At every decoding step , the neural network generates the probability distribution of inputs in at first. Then, points to an available input with the highest probability which is chosen as the input of the next decoding step and added into , and is updated to according to the constraints of the actual problem, as formulated in the following equations:
where is the state transition function to update the state set of inputs, and θ are learnable parameters.
The above process continues until all the available tasks are completed, and the probability chain rule is adopted to factorize the probability of generating the sequence Y as
The reward of the final scheduling results is the profit ratio of the completed tasks, which is denoted by .
3.2. Local Attention Mechanism
The decoding process of the neural network is dynamic, and every time, one task is selected through an attention mechanism according to the static encoding information and the current dynamic encoding information, which utilizes auto-regression. The attention mechanism is adopted to calculate the probability distribution of candidate tasks, and that with the highest probability is selected. In this process, the selection order of tasks represents their execution order. Notably, the unreasonable sorting order can form a low-quality solution.
In the previous research [30,31,32], a global attention mechanism is adopted to calculate the probability distribution of all unselected tasks and make neural networks focus on the essential parts of global task information. However, as the scale of the problem expands, the length of the task sequence also becomes longer, making it more difficult for the global attention mechanism to find out the best candidate task quickly and accurately at every decoding step. In the early training process, the neural network has not formed the optimal decision policy, which is highly likely to generate unreasonable sorting orders. Once the task that should have been executed last is selected first, others will lose the opportunity to be executed. For example, Figure 3 shows several VTWs of five tasks in the four orbits. The orbit is not considered since no VTW exists in this orbit. As shown in Figure 4a, the global attention mechanism is applied to this example, and all the unselected tasks are included in the candidate range at every decoding step. If is selected to be executed in the VTW , the tasks without later VTWs will have no chance to be completed, and and will be abandoned, which could have been executed in the previous orbits. Owing to the tendency to generate unseasonable sorting orders, it is hard for the global attention mechanism to generate great solutions in the training process, making it difficult for the neural network to learn excellent experiences and further influencing the training effect.
Figure 3.
VTWs of five tasks in the four orbits.
Figure 4.
Comparison between global attention and local attention. Under the same circumstances, the global attention mechanism needs more computation. However, it tends to generate unreasonable task orders, leading some tasks to be abandoned, so its decoding step number is less. (a) Global attention. At the first step, is selected from all the tasks to be executed. and are deleted since they have no VTWs in the following orbits, and and are the remaining tasks. At the second step, is selected, and is deleted. The total step number is 2. (b) Local attention. At the first step, , , , and have VTWs in the orbits and , and is selected from them to be executed. At the second step, , , and are still available in these two orbits, and is selected from them, causing the scheduling in to be finished. At the third step, , , and with VTWs in the next two orbits need to be scheduled, and is selected. At the last step, only and have VTWs in the last orbit . is selected, and is deleted. The total step number is 4.
To avoid these deficiencies, we propose a local attention mechanism that reduces the range of candidate tasks and focuses on the local task information. An ideal scheduling solution can fully utilize orbit resources and arrange as many feasible tasks as possible in an orbit without causing conflicts. Thus, the range of candidate tasks is limited to the current several orbital cycles at every decoding step, and the local attention mechanism only considers the unselected tasks that satisfy the condition of having VTWs in this local range. The local attention is set to only consider the unselected tasks with VTWs in the current orbit or the next one at every decoding step. As illustrated in Figure 4b, the local attention is employed for the above example. Only , , , and , which have VTWs in or , are taken into account at the first decoding step. If no unselected tasks have VTWs in the current orbit, those with VTWs in the following two orbits will be considered. If there is only one orbit left, the local attention will only need to focus on the left tasks in this orbit. In this way, the scheduling solutions can be significantly improved. For one thing, it is easier to find the best candidate task within a local range. For another, the observation time of the next task can be arranged close to that of the previous one, preventing some tasks from losing their observation opportunities.
In the proposed local attention mechanism, only the conditional probabilities of partial tasks need to be calculated. Let be the set of encoded static elements of tasks meeting the above condition at the decoding step n, and is the number of these tasks. The conditional probabilities can be calculated as follows:
3.3. Compositions of the Neural Network
The neural network consists of a static embedding layer, a static encoder, a dynamic embedding layer, a dynamic encoder, and a decoder, which are elaborated in detail below.
3.3.1. Static Embedding Layer
The static embedding layer is used to embed the static elements, including time window information and requirement information, into a high-dimensional vector space. The static elements have the following characteristics: first, the number of tasks is variable; second, the input task sequence does not have an apparent sorting order. However, the features of the time window information and the requirement information are distinct. The VTW number of different tasks is not a fixed value, and the time window information of a task is a time-related sequence. The requirement information of every task is a vector with fixed dimensions. Therefore, the static embedding layer must be able to process task window information and requirement information separately.
The structure of the static embedding layer is shown in Figure 5. For , its time window set is embedded through a fully connected network () and an LSTM network () successively, and a vector is obtained. Meanwhile, its requirement vector is embedded through a fully connected network (), and a vector is obtained. Then, the concatenated form of and is further processed through a fully connected network (). The concatenating operator is denoted by . The final embedding output of T is denoted as . The calculating process of is formulated as follows:
where , , and are learnable parameters of the corresponding networks. With the static embedding layer, the proposed neural network can handle the inputs of variable dimensions.
Figure 5.
Structure of the static embedding layer.
3.3.2. Static Encoder
The static encoder is used to extract features from the static inputs for the subsequent solution construction. Because the input order of tasks is meaningless for this problem, position coding and recurrent neural networks are inapplicable to encoding. Thus, a multi-head self-attention sub-layer () and a fully connected feed-forward sub-layer () are adopted in the static encoder. The multi-head self-attention mechanism [27] can divide the input sequence into multiple sub-sequences and perform self-attention on each sub-sequence. The self-attention extracts features by calculating the correlation among tasks, and the extracted features of a task contain not only its own feature information but also global feature information. Each sub-sequence corresponds to a head, and different heads focus on different aspects of the input sequence. This mechanism allows the network model to capture different relationships between tasks in the input sequence and efficiently extract features from a long sequence.
To improve the training efficiency and performance, each sub-layer adds a skip-connection structure [40] to alleviate the gradient disappearance problem and a layer-normalization operator [40] to stabilize the training process. The layer-normalization operator is denoted by , the output of the multi-head self-attention layer is denoted by , and the final encoding output of is denoted by . The static encoding process is formulated as below:
where is the learnable parameter set of the multi-head attention layer, and is the learnable parameter set of the fully connected feed-forward layer.
3.3.3. Dynamic Embedding Layer
In the dynamic embedding layer, a fully connected network () is adopted to embed the dynamic elements into a high-dimensional vector space, as formulated in the following equation:
where is the embedding form of at step n, and is the learnable parameter set of the fully connected network.
3.3.4. Dynamic Encoder
The dynamic encoder is used for the further extraction of dynamic features. At every step, its inputs comprise the embedding form of the current state and the encoding form of the task that is selected at the previous step. The states of the AOS have the following characteristics: first, it is dynamically changing; second, the current state is affected by the previous one; third, the state sequence is time related. Hence, the dynamic encoder is composed of an LSTM cell (), a concatenating operator, and a fully connected network (). At step n, converts to a feature vector through the LSTM cell at first. Then, the concatenation of and is transformed to by a fully connected network. This process is formulated as follows:
where is the learnable parameter set of the LSTM cell, and is the learnable parameter set of the fully connected network.
3.3.5. Decoder
The decoder makes decisions based on the outputs of the static encoder and the dynamic encoder, and it contains three components: an LSTM cell, a local attention layer, and a single-step scheduler.
At decoding step n, the LSTM cell () is adopted to process the outputs of the dynamic encoder at first owing to it appearing as the sequential characteristic, which is formulated as below:
where is the learnable parameter set of the LSTM cell.
Then, the designed local attention layer calculates the probability distribution of the candidate tasks according to the hidden state of the LSTM cell and the outputs of the static encoder, which are formulated in Equations (20) and (21), and the task with the highest probability will be selected to be scheduled. In this process, the range of the candidate tasks is narrowed, and the local attention mechanism only calculates the probability distribution of the tasks in this range. Significantly, the static encoding results of each task contain global feature information owing to the multi-head self-attention mechanism so that the task selection can tend toward global optimization.
Finally, the single-step scheduler selects the task with the highest probability and sets the earliest feasible time as its observation start time, and the single-step scheduling result is generated according to the complex constraints formulated in Section 2.3. Once the task is successfully executed, the rest of the unselected tasks will be judged on whether they have VTWs in the remaining period, and those without VTWs will be deleted, whose probability will be recorded.
3.4. Training Method
An actor–critic algorithm with an adaptive learning rate is adopted to train the proposed neural network, as presented in Algorithm 1. The actor–critic algorithm and the adaptive learning rate strategy are elaborated in detail below.
| Algorithm 1 Training algorithm |
|
3.4.1. Actor–Critic Algorithm
The actor–critic training algorithm is a common training algorithm for the optimization of neural networks [32], and it comprises two neural networks: an actor network and a critic network. The actor network is the proposed neural network for generating the scheduling result, and the profit ratio of the scheduling result is the reward of the actor network. The critic network is a separate network for evaluating the reward of inputs. Both of these networks need to be trained. In this study, the critic network is composed of an embedding layer, an encoder, and a decoder. Its embedding layer and encoder employ the same architecture as the actor network, while its decoder consists of two one-dimensional convolutional layers.
Algorithm 1 shows the pseudo-code of the training algorithm. The neural networks are trained on four kinds of samples with different task sizes Epoch times, which are introduced in Section 4.1. Each kind of sample contains BN batches. At the training step κ, given a batch of samples whose size is BS, the optimization process of the neural networks can be divided into five steps: (1) the actor network generates solutions through step-by-step construction; (2) the reward of every solution is calculated, and is the reward of the solution b; (3) the critic network generates the evaluation value of every sample, and is the evaluation value of the sample b; (4) the policy gradient of the actor network is calculated according to Equation (31), and that of the critic network is formulated in Equation (32); (5) Adam optimizer is adopted to optimize the parameters of the two networks according to the corresponding policy gradients and learning rates. The learning rate of the actor network is adjusted dynamically through the adaptive learning rate strategy, and that of the critic network is fixed.
3.4.2. Adaptive Learning Rate Strategy
In the training process of the neural network, the learning rate is a crucial hyperparameter that determines the update stride of network parameters. A large learning rate may lead to easy divergence, while a small learning rate may lead to slow convergence [41]. In the prior experiments, the conventional fixed learning rates and exponential learning rates failed to balance the early exploration and the later convergence, and a slightly higher learning rate could bring the reward curve to a higher level but make it difficult to converge. Therefore, the learning rate should be adjusted dynamically, and its appropriate decay and increase can make the reward curve converge to a higher level.
To improve the training effect of the actor network, an adaptive learning rate strategy is proposed that dynamically adjusts the learning rate according to the training performance. The reward curve can intuitively reflect the training performance. However, the neural network is trained on the four kinds of samples with significantly different optimal rewards, resulting in the reward curve appearing as a step shape. Hence, the learning rate cannot be adjusted according to the reward value, but the trend of the reward curve can be selected as the indicator for adjusting the learning rate. The adjustment process of the adaptive learning rate contains two steps: first, the learning rate overall decays by a fixed decay rate to ensure the convergence of the reward curve; then, for every ten training steps, if the trend of the reward curve does not reach the expected value, the learning rate will increase slightly to help the reward curve reach a higher level. After the learning rate is initialized or increases, the reward curve may be at a lower level in the early stage, but it is expected to rise rapidly in the middle stage of the following training and converge stably in the later stage, similar to a cosine wave of half a cycle. Thus, a cosine curve is used as the reference curve, and the expected value is calculated according to its slope. Details of the adaptive learning rate strategy are elaborated below.
Firstly, some variables need to be defined to describe the adaptive learning rate: at training step κ, given a batch of samples, (1) is the learning rate of the actor network; (2) is the average reward of this batch of solutions, denoting the reward at the current step; (3) given which is the reward set of the previous ten steps, is its mean value; (4) is the average evaluation value of this batch of samples, denoting the evaluation value at the current step; (5) is the mean value of ; (6) is the slope of the reward curve; (7) represents the expected slope; (8) if , κ will be recorded as whose initial value is zero. (9) K is the total training steps; and (10) γ is the decay rate.
Secondly, the trend of the reward curve is expected to be close to that of a cosine curve, as shown in Figure 6. After ten training steps, the slope of the reward curve is compared with the expected value, which is gained according to the slope of the reference cosine curve. If the slope of the reward curve is less than the expected value, the learning rate of the actor network will increase slightly; otherwise, the learning rate will keep decaying at the following ten training steps. Once the learning rate increases, the reward curve is expected to rise like a new cosine curve at the following training steps.
Figure 6.
Diagram of the reward curve and the reference cosine curves. In this diagram, a fold line is used to represent the reward curve. However, the actual reward curve is a fluctuating curve. The appropriate increase in the learning rate can accelerate the rise of the reward curve in the mid-term of training. A new cosine curve is set as the reference cosine curve whenever the learning rate increases.
Thirdly, the reference cosine curve function is formulated in Equation (33), and its derivative is formulated in Equation (34). The derivative function cannot be directly set as the expected slope for two reasons: (1) is just an evaluation value and not the actual optimal value, and it may be inaccurate and larger than 1; (2) and are the values at one step and cannot accurately represent training results. Thus, some improvements are made based on the derivative function. First, is replaced with to ensure the outcome is between 0 and 1. Then, and are, respectively, replaced with the mean values and . In addition, is replaced with K, resulting in that the expected slope can be larger in the early stage of training and be smaller in the later stage. In this way, the learning rate can have more chances to rise in the early stage but can keep decaying in the later stage. The final expected slope is formulated in Equation (35) through these improvements.
Fourthly, the reward curve slope is calculated according to the reward values of the previous ten steps through the least square method, which is formulated as outlined below:
where is the mean value of one to ten.
Eventually, the adaptive learning rate of the next step is calculated as follows:
where
As formulated in Equations (37) and (38), if the step number κ is the multiple of 10 and the reward curve slope is less than the expected value , the learning rate will increase slightly; otherwise, the learning rate will decay by the decay rate γ. As formulated in Equation (38), the increasing extent of the learning rate is determined by the difference between the evaluation value and the reward value and the difference between the last increased learning rate and the current one, and the learning rate cannot exceed the last increased learning rate. Through the adaptive learning rate strategy, the learning rate shows a downward trend but rises slightly at a few training steps, improving the exploration and convergence of the training algorithm.
3.5. Complexity Analysis
In the actor network, the time complexity of the static embedding layer is , and that of the static encoder is . At the decoding step n, the time complexity of the dynamic embedding layer and that of the dynamic encoder are both , and that of the decoder is . The time complexity of the N-step decoding process is formulated as follows:
where is the computation time of the dynamical embedding layer at a step, is the computation time of the dynamical encoder at a step, and is the computation time of the decoder at a step. , , and are constants. Since and , the time complexity of the actor network is , which is formulated in Equation (40). The time complexity of generating a solution through the well-trained actor network is also .
In the critic network, the time complexity of its embedding layer is , that of its encoder is , and that of its decoder is . Thus, the time complexity of the critic network is , which is formulated in Equation (41).
In the training process, the time complexity of training once for a batch of samples is formulated in Equation (42). And solving an SAOSSP only needs the actor network to run once, so the time complexity of solving is also .
From the above analysis, the training time and solving time of DRLLA rise polynomially as the task size increases. In addition, the training time is also directly affected by the training times, the number of sample types, and the batch size.
4. Computational Experiments
In this section, a dataset of the SAOSSP with different task scales is created to train and test the proposed neural network. Then, sufficient experiments are carried out to verify the effectiveness of the proposed algorithm. The main aspects of validation include the training performance, the adaptability to various-scale instances, and the effect of the local attention mechanism and the adaptive learning rate strategy. The experiments are conducted on a laptop computer with Intel(R) Core (TM) i7-7700HQ CPU @ 2.80 GHz and 40 GB RAM. The DRL framework embedded in Pytorch 1.5.1 in Python 3.8 is adopted in this study.
4.1. Dataset
Due to the lack of public datasets, a large number of instances are designed referring to Reference [4]. In these instances, the orbital parameters of the AOS are listed in Table 2, and its other attribute parameters are listed in Table 3. Numerous tasks are randomly distributed in different areas. For each task, the requested observing duration is a random integer between 5 and 10, and the minimum image quality and priority are both random integers between 1 and 10, which are listed in Table 4. The scheduling time horizon is from 1 January 2023, 00:00:00, to 1 January 2023, 24:00:00. With the growth in the task number, the tasks are distributed more densely in the region, aggravating conflicts among tasks, and only partial tasks have a chance to be completed.
Table 2.
Orbital parameters of the AOS.
Table 3.
Attribute parameters of the AOS.
Table 4.
Attribute parameters of tasks.
In order to enhance and verify the adaptability of the proposed algorithm, it is necessary to ensure the diversity of the training and testing datasets, whose details are shown in Table 5. The training dataset contains four kinds of training samples, each with a quantity of 2560. Tasks in the training samples are distributed in a region with the latitude from 3° N to 53° N and the longitude from 73° E to 133° E. Four testing datasets with tasks located in different areas are created to verify the effectiveness of the proposed algorithm, including , , , and . Each of them contains five types of testing samples, each with a quantity of 128. The datasets can be publicly available at https://github.com/neverlinever/Dataset_AOSSP.git (20 August 2024).
Table 5.
Details of the training and testing datasets.
4.2. Training Process
The parameter settings of the DRLLA are listed in Table 6, containing the network parameters and the training parameters. The total number of training epochs is 10. In every epoch, the network is trained on the training dataset in the order of 200-task, 150-task, 100-task, and 50-task samples, resulting in its reward curve in the training process appearing in a stepped shape. One training epoch contains 80 training steps because of the training parameter settings.
Table 6.
Parameter settings of DRLLA.
The training process is shown in Figure 7. As depicted in Figure 7a, the reward curve fluctuates sharply in the first three training epochs but trends toward stability and convergence in the last three training epochs. The learning rate curve overall shows a downward trend but rises multiple times.
Figure 7.
Training process of DRLLA. (a) The reward curve and the learning rate curve in the whole training process. (b) The reward curve and the learning rate curve in the second training epoch.
Figure 7b shows the details of the training process in the second training epoch (from the 80th training step to the 160th training step), and the learning rate rises obviously five times in this epoch. It can be seen from these two figures that the adaptive learning rate strategy can balance the early exploration and the later convergence of the actor network and the actor network is trained well through the actor–critic training algorithm with the adaptive learning rate.
4.3. Comparison with the State-of-the-Art Algorithms
To verify the effect of DRLLA, we compare it with seven state-of-the-art algorithms for satellite scheduling problems or other combination optimization problems, including Transformer (TRFM) [42], PtrNet [31], RLPT [32], a reinforcement learning-based genetic algorithm (RLGA) [43], a hybrid discrete artificial bee colony algorithm (HDABC) [21], an improved simulated annealing algorithm (ISA) [37], and an improved ant colony optimization algorithm (IACO) [44]. TRFM, PtrNet, and RLPT are neural network models with encoder–decoder structures, and RLGA, HDABC, ISA, and IACO are heuristic algorithms with improved search mechanisms. The heuristic algorithms maintain their original settings. The comparison experiments contain two parts: first, these algorithms are tested on the dataset to validate the effectiveness of DRLLA under different problem scales; then, they are tested on the other three datasets , , and , whose task distribution is different from , to demonstrate the generalization ability of DRLLA in various instances.
Figure 8 and Figure 9 and Table 7 show the detailed testing results of the first comparison experiment on the testing dataset . The boxplot graphs in Figure 8 illustrate the profit ratio distribution of solutions obtained by different algorithms. In these boxplot graphs, the top lines denote maximum values, the bottom lines denote minimum values, the middle boxes denote interquartile ranges, the red lines denote medians, the blue dashed lines denote mean values, and the red circles denote outliers. As shown in these boxplot graphs, the profit ratio distributions of DRLLA are highly concentrated, and its median lines are higher than the comparison algorithms, demonstrating that DRLLA can obtain solutions with the highest profit ratio when handling various AOS scheduling instances. The profit ratio distributions of TRFM are also concentrated, and its median lines are higher than other comparison algorithms but lower than DRLLA. The profit ratio distributions of the remaining algorithms are relatively scattered. In general, the stability and generalization of DRLLA and TRFM are superior to the remaining comparison algorithms, and DRLLA can obtain solutions with higher profit ratios than the comparison algorithms.
Figure 8.
Profit ratio distributions of different algorithms on the dataset .
Figure 9.
Curves of mean profit and mean completion number. (a) Mean profit. (b) Mean completion number.
Table 7.
Detailed testing results of different algorithms on the dataset .
Table 7 presents the detailed testing results of the first comparison experiment on the dataset . The indicators include the mean profit ratio (MPR), mean profit (MP), mean completion number (MCN), and mean computational time (MCT), and the best values are bolded. Figure 9 depicts the MP curves and MCN curves of different algorithms. The testing results are presented as follows:
- (1)
- As for the DRL-based algorithms, DRLLA can obtain better solutions with the higher profit and completion number than these comparison algorithms in a relatively small amount of computational time under different task scales; TRFM, only second to DRLLA, can also gain acceptable solutions with a little more computational time; although RLPT has the shortest computational time in these testing cases, it is hard to achieve satisfactory solutions.
- (2)
- Among the heuristic algorithms, ISA and HDABC can obtain relatively acceptable solutions, but the computational time of HDABC far exceeds other algorithms; RLGA and IACO perform worse than the other algorithms in terms of solution quality. Comparing with the DRL-based algorithms, these heuristic algorithms need more computational time due to the huge computation burden of population iteration.
- (3)
- As the task number increases, the MPR values of all algorithms continuously decrease, while the MP and MCN values of most algorithms keep growing. The main reason is that the increases in the total profit and the total task number are more significant than in the gained profit and the completion number, indicating that the problem scales exceed the observation capability of the AOS.
- (4)
- In general, DRLLA shows excellent performance in solution quality over the comparison algorithms. With the increase of task scale, the gap of the MPR values between DRLLA and the comparison algorithms becomes significant. In addition, the computational time of DRLLA is not less than that of RLPT, but its computation efficiency is superior to other comparison algorithms, especially heuristic algorithms. Thus, the comparison results on the dataset fully demonstrate that DRLLA is capable of obtaining high-quality solutions within acceptable computational time.
To further validate the generalization ability of the proposed algorithm on various instances, we conduct comparison experiments of DRLLA and the comparison algorithms on three other testing datasets , , and , and the testing results are presented in Table 8, where the best values are bolded. HDABC is not used as the comparison algorithm because of its high time consumption. According to the comparison results, DRLLA exhibits significant advantages over the comparison algorithms in terms of solution quality on these three testing datasets. In terms of computational time, the DRL-based algorithms outperform the heuristic algorithms due to the different solution constructions, and the computational time of DRLLA is far less than that of the heuristic algorithms. The superior testing performance of DRLLA on the different datasets fully demonstrates that it has excellent generalization ability to be extended for various SAOSSP instances.
Table 8.
Testing results on the datasets , , and .
4.4. Ablation Experiments
DRLLA consists of two main parts: the neural network and the training algorithm. In the proposed neural network, a multi-head self-attention mechanism is introduced to the static encoder for extracting task features, and the local attention mechanism is applied to decoding for constructing the scheduling solutions. In the training algorithm, the adaptive learning rate strategy is adopted to improve the training effect. In this subsection, ablation experiments are performed to further validate their effectiveness.
4.4.1. Effectiveness Verification of the Multi-Head Self-Attention Mechanism
To demonstrate the contribution of the multi-head self-attention (MHSA), we, respectively, adopt a convolutional neural network (CNN) and LSTM to extract features in the static encoder to build comparison neural network models. The kernel size of CNN is set as 5. The comparison network models are trained first on the dataset and then tested on the dataset .
Table 9 presents the testing results of the proposed network with MHSA and the comparison networks with CNN and LSTM, and the best values are bolded. It can be seen that MHSA performs better than CNN and LSTM in terms of MPR, and it takes the least computational time in the instances with 150, 200, and 250 tasks, indicating that the features extracted by MHSA are more beneficial for the subsequent solution construction. The main reason is that their feature extraction techniques are different. CNN uses a convolution kernel to scan the task information to extract local features, and the extracted features of a task are related to its own and adjacent tasks. LSTM processes task information in a sequencing order, and the extracted features of a task are relevant to its own and the previous tasks. As for CNN and LSTM, the sequencing order of the input tasks can influence the feature extraction results. In comparison, MHSA adopts multiple independent self-attention heads to extract diverse features, and each head focuses on different parts of the task sequence that are not related to the task order. Thus, MHSA has significant advantages over CNN and LSTM in feature extraction.
Table 9.
Testing results of MHSA, CNN and LSTM on .
4.4.2. Effectiveness Verification of the Local Attention Mechanism
To demonstrate the contribution of the local attention mechanism, we replace it with the global attention mechanism to build a comparison network, train the comparison network with the same training algorithm, and compare the testing results of the two algorithms on the dataset . This deep reinforcement learning algorithm with the global attention mechanism is denoted by DRLGA.
Table 10 lists the testing results of DRLLA and DRLGA, where the best values are bold. As for MPR, DRLLA is significantly better than DRLGA, indicating that the local attention mechanism is superior to the global attention mechanism in improving solution quality. The main reason is that the local attention mechanism, in comparison with the global attention mechanism, can guide the network model to focus on the more promising candidate task range, preventing it from generating unreasonable task permutations. However, the MCT values of DRLGA are much lower than those of DRLLA. The main reason is that the global attention mechanism tends to generate inappropriate task permutations, leading some unscheduled tasks to lose their observation opportunities and reducing the number of decoding steps. This mechanism shortens the computation time but seriously sacrifices the solution quality. In comparison, the local attention mechanism takes more computational time, but it dramatically improves the solution quality.
Table 10.
Testing results of DRLLA and DRLGA on .
4.4.3. Effectiveness Verification of the Adaptive Learning Rate Strategy
To validate the feasibility and effectiveness of the adaptive learning rate strategy, we compare the proposed DRLLA with the adaptive learning rate (DRLLA-a) to two compared algorithms: DRLLA with a fixed learning rate of 0.001 (DRLLA-f), and DRLLA with an exponential learning rate (DRLLA-e). In DRLLA-e, the initial learning rate is 0.01, and the decay rate is 0.001. The three algorithms are trained on the dataset first and then tested on the dataset . The comparison results of training and testing are presented as follows.
Figure 10 depicts the training process of these three algorithms, including their profit ratio curves and learning rate curves. It can be observed that DRLLA-a performs best among these algorithms in the training process. Its profit ratio curve is similar to those of the comparison algorithms in the first training epoch but quickly reaches a higher level in the following training epochs owing to the appropriate rise of the learning rate. In the later stage of training, its learning rate decreases to a low level, and its profit ratio curve tends to be stable and convergent. DRLLA-e can achieve a little better profit ratio than DRLLA-f in the early stage of training, while the opposite is true in the later stage of training. The profit ratio curve of DRLLA-e is more stable than that of DRLLA-f, especially in the later stage. The training results indicate that (1) a high learning rate can improve the exploration ability of the network model in the early stage of training, while a low learning rate can improve its stability and convergence in the later stage of training; (2) according to such characteristics, the proposed adaptive learning rate strategy appropriately increases the learning rate on the basis of the exponential learning rate, enhancing the early exploration ability and the later convergence ability of the training algorithm.

Figure 10.
Training process of DRLLA-a, DRLLA-f, and DRLLA-e. (a) Profit ratio curves of DRLLA-a, DRLLA-f, and DRLLA-e in the training process. (b) Learning rate curves of DRLLA-a, DRLLA-f, and DRLLA-e in the training process.
Table 11 lists the testing results of the three algorithms, where the best values are bolded. The standard deviation of profit ratios (STD) is set as an additional comparison indicator, and the lower STD value indicates the more densely distributed solutions and further demonstrates the better stability and generalization of the algorithm. DRLLA-a can reach the highest MPR value within the acceptable computational time in most testing instances, and its STD value, close to that of DRLLA-e, is significantly better than that of DRLLA-f. As for DRLLA-f, its MPR value is slightly lower than that of DRLLA-a in most testing instances and even exceeds that of DRLLA-a when the task number is 250, while it performs the worst in terms of STD. DRLLA-e has superior stability ability and takes the least computational time, but it fails to generate solutions with high profit ratios. The testing results demonstrate that the proposed neural network trained with the adaptive learning rate training algorithm has an excellent solving ability, and the comprehensive testing performance of DRLLA-a is superior to the compared algorithms.
Table 11.
Testing results of the neural networks trained by three training algorithms on .
5. Conclusions and Future Work
In this paper, a deep reinforcement learning algorithm with a local attention mechanism is proposed to address the scheduling problem of a single agile optical satellite with different task scales. Two techniques are effectively adopted to improve the performance of the algorithm: (1) local attention mechanism and (2) adaptive learning rate strategy. The local attention mechanism narrows the range of candidate tasks and selects the next-scheduled task from a more promising range, significantly improving the quality of the generated solution. The adaptive learning rate strategy dynamically decreases or increases the learning rate according to the reward curve trend in the training process, enhancing the early exploration and later convergence abilities. Based on these techniques, the proposed algorithm exhibits superior performance in solution quality, generalization, and efficiency in comparison with the state-of-the-art algorithms. The experimental results also validate the effectiveness of the local attention mechanism for generating high-quality solutions and the adaptive learning rate strategy for improving the training effect.
This paper can provide an efficient and effective approach for agile optical satellite scheduling, while there are still some areas for improvement. Some practical factors are not fully considered, such as the influence of cloud cover, lighting conditions, and observation angles on the imaging results. For future work, we will give a comprehensive consideration of various practical constraints and establish a more realistic AOS scheduling model. In addition, we will further improve the proposed algorithm to make it more available for practical application, and the algorithm will be extended to more complex instances, such as multi-AOS scheduling, emergency scheduling of AOSs, and uncertainty scheduling of AOSs.
Author Contributions
Conceptualization, Z.L. and W.X.; Data curation, X.Y.; Formal analysis, Z.L.; Funding acquisition, W.X.; Investigation, Z.L. and C.H.; Methodology, Z.L. and W.X.; Project administration, W.X.; Resources, C.H. and X.Y.; Software, Z.L. and C.H.; Supervision, W.X.; Validation, Z.L.; Visualization, Z.L. and C.H.; Writing—original draft, Z.L.; Writing—review and editing, Z.L. and W.X. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Ministry’s and National Commission’s Scientific Research Project (Grant No. TJ2022B020200-003-020).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Wu, G.; Pedrycz, W.; Li, H.; Ma, M.; Liu, J. Coordinated planning of heterogeneous earth observation resources. IEEE Trans. Syst. Man Cybern. Syst. 2015, 46, 109–125. [Google Scholar] [CrossRef]
- Berger, J.; Lo, N.; Noël, M.; Noutegne, L. DynaQUEST: A New Approach to the Dynamic Multi-satellite Scheduling Problem. In Proceedings of the ICORES; SCITEPRESS—Science and Technology Publications, Lda.: Setúbal, Portugal, 2020; pp. 194–201. [Google Scholar]
- Lu, Z.; Shen, X.; Li, D.; Chen, Y.; Li, D. A Mission Planning Modeling Method of Multipoint Target Imaging Within a Single Pass for Super-Agile Earth Observation Satellite. IEEE Syst. J. 2021, 16, 1921–1932. [Google Scholar] [CrossRef]
- Liu, X.; Laporte, G.; Chen, Y.; He, R. An adaptive large neighborhood search metaheuristic for agile satellite scheduling with time-dependent transition time. Comput. Oper. Res. 2017, 86, 41–53. [Google Scholar] [CrossRef]
- Lemaître, M.; Verfaillie, G.; Jouhaud, F.; Lachiver, J.M.; Bataille, N. Selecting and scheduling observations of agile satellites. Aerosp. Sci. Technol. 2002, 6, 367–381. [Google Scholar] [CrossRef]
- Stollenwerk, T.; Michaud, V.; Lobe, E.; Picard, M.; Basermann, A.; Botter, T. Agile earth observation satellite scheduling with a quantum annealer. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 3520–3528. [Google Scholar] [CrossRef]
- Wei, L.; Xing, L.; Wan, Q.; Song, Y.; Chen, Y. A multi-objective memetic approach for time-dependent agile earth observation satellite scheduling problem. Comput. Ind. Eng. 2021, 159, 107530. [Google Scholar] [CrossRef]
- Wang, J.; Song, G.; Liang, Z.; Demeulemeester, E.; Hu, X.; Liu, J. Unrelated parallel machine scheduling with multiple time windows: An application to earth observation satellite scheduling. Comput. Oper. Res. 2023, 149, 106010. [Google Scholar] [CrossRef]
- Peng, G.; Song, G.; He, Y.; Yu, J.; Xiang, S.; Xing, L.; Vansteenwegen, P. Solving the agile earth observation satellite scheduling problem with time-dependent transition times. IEEE Trans. Syst. Man Cybern. Syst. 2020, 52, 1614–1625. [Google Scholar] [CrossRef]
- Chu, X.; Chen, Y.; Xing, L. A branch and bound algorithm for agile earth observation satellite scheduling. Discret. Dyn. Nat. Soc. 2017, 2017, 1–15. [Google Scholar] [CrossRef]
- Peng, G.; Song, G.; Xing, L.; Gunawan, A.; Vansteenwegen, P. An exact algorithm for agile earth observation satellite scheduling with time-dependent profits. Comput. Oper. Res. 2020, 120, 104946. [Google Scholar] [CrossRef]
- Barkaoui, M.; Berger, J. A new hybrid genetic algorithm for the collection scheduling problem for a satellite constellation. J. Oper. Res. Soc. 2020, 71, 1390–1410. [Google Scholar] [CrossRef]
- Zheng, Z.; Guo, J.; Gill, E. Swarm satellite mission scheduling & planning using hybrid dynamic mutation genetic algorithm. Acta Astronaut. 2017, 137, 243–253. [Google Scholar]
- Yan, B.; Wang, Y.; Xia, W.; Hu, X.; Ma, H.; Jin, P. An improved method for satellite emergency mission scheduling scheme group decision-making incorporating PSO and MULTIMOORA. J. Intell. Fuzzy Syst. 2022, 42, 3837–3853. [Google Scholar] [CrossRef]
- Wu, X.; Yang, Y.; Sun, Y.; Xie, Y.; Song, X.; Huang, B. Dynamic regional splitting planning of remote sensing satellite swarm using parallel genetic PSO algorithm. Acta Astronaut. 2023, 204, 531–551. [Google Scholar] [CrossRef]
- Cui, K.; Xiang, J.; Zhang, Y. Mission planning optimization of video satellite for ground multi-object staring imaging. Adv. Space Res. 2018, 61, 1476–1489. [Google Scholar] [CrossRef]
- He, L.; Liu, X.L.; Chen, Y.W.; Xing, L.N.; Liu, K. Hierarchical scheduling for real-time agile satellite task scheduling in a dynamic environment. Adv. Space Res. 2019, 63, 897–912. [Google Scholar] [CrossRef]
- Luo, K. A hybrid binary artificial bee colony algorithm for the satellite photograph scheduling problem. Eng. Optim. 2020, 52, 1421–1440. [Google Scholar] [CrossRef]
- Chatterjee, A.; Tharmarasa, R. Reward factor-based multiple agile satellites scheduling with energy and memory constraints. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3090–3103. [Google Scholar] [CrossRef]
- Du, B.; Li, S.; She, Y.; Li, W.; Liao, H.; Wang, H. Area targets observation mission planning of agile satellite considering the drift angle constraint. J. Astron. Telesc. Instrum. Syst. 2018, 4, 047002. [Google Scholar]
- Yang, Y.; Liu, D. A Hybrid Discrete Artificial Bee Colony Algorithm for Imaging Satellite Mission Planning. IEEE Access 2023, 11, 40006–40017. [Google Scholar] [CrossRef]
- Gu, Z.; Ma, Q.; Gao, X.; You, J.W.; Cui, T.J. Direct electromagnetic information processing with planar diffractive neural network. Sci. Adv. 2024, 10, eado3937. [Google Scholar] [CrossRef] [PubMed]
- Gao, X.; Ma, Q.; Gu, Z.; Cui, W.Y.; Liu, C.; Zhang, J.; Cui, T.J. Programmable surface plasmonic neural networks for microwave detection and processing. Nat. Electron. 2023, 6, 319–328. [Google Scholar] [CrossRef]
- Gu, Z.; Ma, Q.; Gao, X.; You, J.W.; Cui, T.J. Classification of metal handwritten digits based on microwave diffractive deep neural network. Adv. Opt. Mater. 2024, 12, 2301938. [Google Scholar] [CrossRef]
- Ma, Q.; Gao, X.; Gu, Z.; Liu, C.; Li, L.; You, J.W.; Cui, T.J. Intelligent neuromorphic computing based on nanophotonics and metamaterials. MRS Commun. 2024, 1–20. [Google Scholar] [CrossRef]
- Nazari, M.; Oroojlooy, A.; Snyder, L.; Takác, M. Reinforcement Learning for Solving the Vehicle Routing Problem. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018); Neural Information Processing Systems Foundation: Montreal, QC, Canada, 2018. [Google Scholar]
- Kool, W.; Van Hoof, H.; Welling, M. Attention, learn to solve routing problems! arXiv 2018, arXiv:1803.08475. [Google Scholar]
- He, Y.; Xing, L.; Chen, Y.; Pedrycz, W.; Wang, L.; Wu, G. A generic Markov decision process model and reinforcement learning method for scheduling agile earth observation satellites. IEEE Trans. Syst. Man Cybern. Syst. 2020, 52, 1463–1474. [Google Scholar] [CrossRef]
- Ren, L.; Ning, X.; Wang, Z. A competitive Markov decision process model and a recursive reinforcement-learning algorithm for fairness scheduling of agile satellites. Comput. Ind. Eng. 2022, 169, 108242. [Google Scholar] [CrossRef]
- Chen, M.; Chen, Y.; Chen, Y.; Qi, W. Deep Reinforcement Learning for Agile Satellite Scheduling Problem. In Proceedings of the 2019 IEEE Symposium Series on Computational Intelligence (SSCI), Xiamen, China, 6–9 December 2019; pp. 126–132. [Google Scholar]
- Zhao, X.; Wang, Z.; Zheng, G. Two-phase neural combinatorial optimization with reinforcement learning for agile satellite scheduling. J. Aerosp. Inf. Syst. 2020, 17, 346–357. [Google Scholar] [CrossRef]
- Wei, L.; Chen, Y.; Chen, M.; Chen, Y. Deep reinforcement learning and parameter transfer based approach for the multi-objective agile earth observation satellite scheduling problem. Appl. Soft Comput. 2021, 110, 107607. [Google Scholar] [CrossRef]
- Berger, J.; Lo, N.; Barkaoui, M. QUEST—A new quadratic decision model for the multi-satellite scheduling problem. Comput. Oper. Res. 2020, 115, 104822. [Google Scholar] [CrossRef]
- Han, C.; Gu, Y.; Wu, G.; Wang, X. Simulated annealing-based heuristic for multiple agile satellites scheduling under cloud coverage uncertainty. IEEE Trans. Syst. Man Cybern. Syst. 2022, 53, 2863–2874. [Google Scholar] [CrossRef]
- Zhang, J.; Xing, L. An improved genetic algorithm for the integrated satellite imaging and data transmission scheduling problem. Comput. Oper. Res. 2022, 139, 105626. [Google Scholar] [CrossRef]
- Song, Y.; Ou, J.; Pedrycz, W.; Suganthan, P.N.; Wang, X.; Xing, L.; Zhang, Y. Generalized model and deep reinforcement learning-based evolutionary method for multitype satellite observation scheduling. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 2576–2589. [Google Scholar] [CrossRef]
- Wu, X.; Yang, Y.; Xie, Y.; Ma, Q.; Zhang, Z. Multi-region Mission Planning by Satellite Swarm Using Simulated Annealing and Neighborhood Search. IEEE Trans. Aerosp. Electron. Syst. 2023, 60, 1416–1439. [Google Scholar] [CrossRef]
- Yue, A.M.; Feng, B.W.; Xueqin, C.C.; Yanjun, D.Y.; Chaoyong, E.L. Preparation of Papers for IFAC Conferences & Symposia: Autonomous Distribution Algorithm for Formation Satellites Under Emergent Imaging Requests. IFAC-Pap. Online 2019, 52, 460–465. [Google Scholar]
- He, L.; Liu, X.; Laporte, G.; Chen, Y.; Chen, Y. An improved adaptive large neighborhood search algorithm for multiple agile satellites scheduling. Comput. Oper. Res. 2018, 100, 12–25. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Xu, C.; Liu, S.; Yang, Z.; Huang, Y.; Wong, K.K. Learning rate optimization for federated learning exploiting over-the-air computation. IEEE J. Sel. Areas Commun. 2021, 39, 3742–3756. [Google Scholar] [CrossRef]
- Bresson, X.; Laurent, T. The transformer network for the traveling salesman problem. arXiv 2021, arXiv:2103.03012. [Google Scholar]
- Song, Y.; Wei, L.; Yang, Q.; Wu, J.; Xing, L.; Chen, Y. RL-GA: A reinforcement learning-based genetic algorithm for electromagnetic detection satellite scheduling problem. Swarm Evol. Comput. 2023, 77, 101236. [Google Scholar] [CrossRef]
- Zhou, Z.; Chen, E.; Wu, F.; Chang, Z.; Xing, L. Multi-satellite scheduling problem with marginal decreasing imaging duration: An improved adaptive ant colony algorithm. Comput. Ind. Eng. 2023, 176, 108890. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).