Next Article in Journal
Deep Learning in Image Processing and Pattern Recognition
Next Article in Special Issue
Dual-Domain Superposition for Maritime Relay Communications: A Flexible-Coded Transmission Design Towards Spectrum–Reliability Synergy
Previous Article in Journal
Robust 12-Lead ECG Classification with Lightweight ResNet: An Adaptive Second-Order Learning Rate Optimization Approach
Previous Article in Special Issue
IoT-NTN with VLEO and LEO Satellite Constellations and LPWAN: A Comparative Study of LoRa, NB-IoT, and Mioty
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Requirements in Advanced Orbiting Systems

1
School of Electronics & Information Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
Marketing Department, The 28th Research Institute of China Electronic Technology Group Co., Ltd., Nanjing 210000, China
3
Key Laboratory of Intelligent Support Technology for Complex Environments, Ministry of Education, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(10), 1939; https://doi.org/10.3390/electronics14101939
Submission received: 9 April 2025 / Revised: 2 May 2025 / Accepted: 6 May 2025 / Published: 9 May 2025
(This article belongs to the Special Issue Future Generation Non-Terrestrial Networks)

Abstract

:
In order to efficiently transmit spatial data with diversified service types, this paper proposes an Advanced Intelligent Framing Algorithm (AIFG) for Advanced Orbiting System (AOS), based on the virtual channel multiplexing AOS technology, aiming to meet differentiated quality of service (QoS) requirements. To enable the timely transmission of delay-sensitive services, the optimal time threshold for framing such services is modeled as a Markov decision process (MDP). The algorithm utilizes Proximal Policy Optimization (PPO), achieving an optimal solution based on performance metrics such as frame multiplexing efficiency, average framing time, and average packet delay. To improve the multiplexing efficiency of non-delay-sensitive services, this type of traffic is transmitted only after completely filling a frame. Simulation results show that, compared to traditional frame generation algorithms, the AIFG algorithm reduces the average queuing delay of services by an average of 32%. Moreover, the AIFG algorithm increases throughput by an average of 32%, and improves multiplexing efficiency by an average of 61%. Thus, the AIFG algorithm balances the transmission requirements of both real-time and non-real-time services, and enhances the QoS of the AOS.

1. Introduction

In recent years, space exploration activities such as deep space exploration, manned spaceflight, and satellite development have become increasingly active globally. The returned data cover various services, including engineering telemetry, physiological telemetry, engineering control, scientific observation, scientific experiments, delayed playback, audio, video, etc. [1]. These data exhibit significant heterogeneity in service types, transmission rates, and QoS requirements. These developments have placed higher demands on the real-time performance and reliability of space data systems. In response, the Consultative Committee for Space Data Systems (CCSDS) established the Advanced Orbiting System (AOS) protocol. The AOS protocol allows multiple data packets with a similar QoS to be multiplexed within the same virtual channel. These data packets are segmented and encapsulated into fixed-length Multiplexing Protocol Data Units (MPDUs), which are then assembled into complete frames with added control information to improve the spatial link multiplexing efficiency.
To provide a comprehensive background and highlight the motivation for this study, we first review the existing research on AOS framing algorithms, which can be broadly categorized into three types as follows: isochronous, high-efficiency, and adaptive framing algorithms. The principle of isochronous frame generation is to set a fixed framing threshold. Regardless of the number of incoming packets, the framing occurs once the set framing time is reached. If the total number of incoming packets does not meet the MPDU length, idle packets are used for padding [2,3,4]. For example, for the ON/OFF source arrival model obeying the Pareto distribution in an AOS, Zhao et al. established the isochronous frame generation method. They obtained the calculation formulas of average packet delay and MPDU average multiplexing efficiency, which verified the feasibility of the isochronous frame generation algorithm in AOS [3]. However, for non-delay-sensitive packets, using the isochronous frame generation algorithm may result in a low frame multiplexing efficiency due to the less stringent timing requirements. As a result, high-efficiency frame generation algorithms were proposed. The principle of high-efficiency frame generation is that, in continuous packet arrival scenarios, a frame is generated and transmitted only when the accumulated packet length reaches the length of an MPDU [5,6,7]. For example, Zhang et al. studied the simplified calculation method of the average frame generation time and the average packet delay in the efficient frame generation algorithm. For the case where the packet arrival rate is the Poisson distribution, they used the Erlang distribution to obtain a clearer correlation between performance indicators [7]. However, the standards for the framing algorithms in the aforementioned two methods are too rigid. In contrast, the adaptive frame generation algorithm can be controlled by pre-setting the framing threshold and MPDU length. Specifically, when the frame waiting time is less than the set threshold value or the MPDU area has been filled with arriving packets, the frame is encapsulated and sent immediately. When the frame waiting time reaches the maximum threshold and the MPDU buffer is not filled, the idle packet is used to complete the entire transmission frame [8,9,10,11]. Tian et al. studied the adaptive frame generation algorithm in the AOS protocol and provided an effective method of calculating the information transmission rate [9]. Dai Changhao et al. proposed an adaptive frame generation algorithm based on wavelet neural network traffic prediction. Under certain delay constraints, the algorithm can adaptively adjust the frame time according to the traffic prediction results [11]. However, most of these algorithms need to set relevant parameters in advance, such as framing time threshold, number of iterations, etc., which cannot interact with the environment in real time. These algorithms result in a low application efficiency in space systems with various types of services and time-varying network states.
Reinforcement learning (RL) is an intelligent method used to achieve the iterative optimization of a policy through a continuous interaction with the environment. This method ensures that the agent selects the best action that is suitable for task requirements under different network conditions. RL does not need to predefine a large number of rules, and it has strong adaptability. Typical RL algorithms include value-based Q-learning and SARSA [12,13,14], policy-based REINFORCE and Actor–Critic methods, and Proximal Policy Optimization (PPO), which is popular recently [15,16,17]. PPO introduces restrictive optimization when updating strategies to ensure stability and efficiency. It is an advanced algorithm in deep reinforcement learning, which is widely used in robot control, autonomous driving, intelligent recommendation systems and other fields [18,19,20,21].
To this end, in this paper, we introduce the PPO algorithm in reinforcement learning into the framing process. We propose the advanced intelligent frame generation (AIFG) algorithm in Advanced Orbiting Systems for differentiated QoS requirements. The main contributions are as follows:
  • For massive heterogeneous services in AOS, we establish a service classification framework for delay-sensitive services and non-delay-sensitive services. Among them, delay-sensitive services include audio data, telemetry data, telecommand data, etc., while non-delay-sensitive services include video data, remote control data, observation data, experimental data, etc. Based on this, we design real-time framing algorithms for delay-sensitive services and efficient framing algorithms for non-delay-sensitive services, which are discussed in detail in Section 2 and Section 5.
  • In order to effectively reduce the queuing delay of delay-sensitive services, we model the problem of calculating the optimal time threshold for framing such services as a Markov decision process. This model captures the sequential decision-making nature of the problem. Then, we employ the PPO algorithm to obtain the optimal solution based on performance indicators such as frame multiplexing efficiency, average framing time, and average packet delay. This modeling and optimization process is elaborated in Section 4.
  • We develop a joint simulation platform based on NS-3 and PyTorch, which enables a comprehensive performance evaluation of the proposed AIFG algorithm under realistic network conditions. The results show that compared with the traditional frame generation algorithm, the average queuing delay of the service under the AIFG algorithm is reduced by an average of 31.69%, the average multiplexing efficiency is increased by an average of 60.88%, and the average throughput is increased by an average of 31.92%. The AIFG algorithm accommodates different transmission needs of real-time and non-real-time services, enhancing the network QoS of the AOS. The detailed simulation setup and performance analysis are presented in Section 6.
This article is organized as follows. In Section 2, the system model is established, including the data arrival model, service classification, and queuing structures for different service types. Section 3 describes the materials and methods of this paper. Then, Section 4 formulates the optimal frame generation time threshold calculation as a Markov decision process and presents a solution based on the Proximal Policy Optimization algorithm. Section 5 introduces the design of the advanced intelligent frame generation (AIFG) algorithm to address differentiated QoS requirements. Section 6 details the simulation setup, parameter settings, and performance evaluation of the proposed algorithm compared with baseline methods. The results of the simulation and theoretical analysis are given in Section 6. Finally, Section 7 concludes this article and highlights potential future research directions.

2. System Model

In this section, we establish the system model for an AOS environment, which provides the foundation for subsequent algorithm design and performance evaluations. The model describes the integration of heterogeneous data sources, the classification of services based on their delay sensitivity, and the differentiated frame generation processes to meet various QoS requirements.
In the AOS, various types of data are received from spacecraft payloads. As shown in Figure 1, the system input integrates six types of heterogeneous data sources. Among them, telemetry data monitor device states, while telecommand data issue control instructions with the highest operational priority. Voice data are real-time streaming media, requiring stringent low-latency transmission. Non-real-time bulk data, such as video, observation and experimental data generated from scientific tasks, allow for flexible delay tolerance. After classification, data packets are separated into two service queues. Voice, telemetry, and telecommand data, which are highly delay-sensitive, are directed into the delay-sensitive service queue and processed using the real-time framing algorithm to generate complete frames. The real-time frame generation process operates within a time-slotted framework. Each frame period collects the arrived packets and supplements idle packets if necessary. Each frame carries both the arrived packets during each time window and dynamically generated idle packets to maintain frame structure integrity under bandwidth fluctuations. In contrast, video, observation, and experimental data, which are less delay-sensitive, enter the non-delay-sensitive service queue and are processed using the high-efficiency framing algorithm. After framing, all data frames are scheduled and transmitted through the virtual channel scheduling module for subsequent transmission within the AOS network. The overall structure of the system model is illustrated in Figure 1, showing the classification of incoming data, the queuing and the framing mechanisms.
We construct a dual-queue framing model to accommodate both delay-sensitive and non-delay-sensitive traffic in AOS. Two distinct traffic models are employed to more accurately reflect real-world service characteristics. First, we define two types of data: real-time data U h = { u h 0 , u h 1 , , u h H } and non-real-time data U l = { u l 0 , u l 1 , , u l L } . Among them, real-time data include audio data, telemetry data, etc. Non-real-time data include video data, remote control data, observation data, experimental data, etc. After the arrival of all data packets, real-time data enter the delay-sensitive service queue, and non-real-time data enter the non-delay-sensitive service queue.
On the one hand, the framing task of packets in the delay-sensitive service queue is primarily handled by the real-time framing (RTFG) module. Delay-sensitive data such as telemetry, telecommand, and voice traffic typically consist of short-duration flows with limited burstiness and minimal long-range dependence. These characteristics have been extensively validated in the literature, which supports modeling such traffic using Poisson arrival processes. We establish a packet arrival model. We assume that the packet arrival process follows a Poisson distribution with arrival rate parameter λ . In this case, the probability of receiving exactly k packets during a time interval t is characterized by the Poisson distribution, as given by Equation (1):
P ( X = k ) = ( λ t ) k k ! e λ t k = 0 , 1 , 2 , 3 ,
where X denotes the number of packet arrivals within time duration t, and λ is the average arrival rate.
Assuming that each packet has a fixed size l s , the number of packets which can be accommodated in an MPDU is n, and the length of the MPDU is l m s = n × l s . Thus, incoming packets can exactly fill the MPDU data area, avoiding the issue of incomplete packet filling.
The maximum number of packets that can be packed into the MPDU data area is set to be N, and the time required to generate a frame is denoted as t s w . As shown in Equation (2), the probability density function for framing time t s w within the interval 0 , T is
f s w ( t ) = λ e λ t ( λ t ) N 1 ( N 1 ) !
The probability that the framing time t s w is below the threshold value T t h = T is
P ( t s w = T ) = T + λ e λ t λ t N 1 N 1 ! d t
The average framing time for the RTFG module is
Φ ( S w ) = 0 T t f s w ( t ) d t + T P ( t s w = T ) = 0 T t λ e λ t λ t N 1 N 1 d t + T T + λ e λ t λ t N 1 N 1 d t
Let l s be the packet length and l m s be the packet length of the MPDUs. The average multiplexing efficiency for the RTFG module is
E ( e f f m ) = P ( t S W < T ) + n = 0 N 1 n l s l m s P ( A ( T ) = n ) = 0 T t λ e λ t λ t N 1 N 1 d t + n = 0 N 1 n l s l m s e λ t λ T n n !
Let T d be the packet delay, P ( A ( t ) = n ) be the probability of reaching n packets in 0 , T time, and P ( A ( t ) = N + x ) be the probability of reaching N + x packets in 0 , T time, with the average packet delay of the RTFG module when i packets arrive being
ϑ ( T d ) = 1 N · E ( e f f m ) · n = 1 N 1 P ( A ( t ) = n ) · T 2 · n + x = 0 P ( A ( t ) = N + x ) · i = 1 N ( N i ) · T N + x + 1
On the other hand, the framing task of packets in the non-delay-sensitive service queue is handled by the high-efficiency framing (HEFG) module. The non-real-time packets are framed according to the maximum number of packets in the MPDU. Non-delay-sensitive traffic, such as observation or scientific payload data, often exhibits bursty behavior, long-range dependence, and variable flow sizes. To better model such phenomena, we introduce a self-similar ON/OFF traffic model, where each of the M sources alternates between ON (sending packets) and OFF (idle) periods.
Both ON durations τ l and OFF durations θ l follow Pareto distributions, with tail behavior:
P ( τ > t ) t α , 1 < α < 2
The aggregate active source count ξ at time t follows a binomial distribution with success probability p = E [ τ ] E [ τ ] + E [ θ ] , and it can be approximated by a Poisson process with intensity λ = M · p as M . In the queue model, we set S w as the time required to generate a frame. Then, the average frame time of the HEFG module is
Φ ( S w ) = 0 t · f ON ( t ) d t
where f ON ( t ) is the empirically derived arrival distribution during ON periods. The average queuing delay ϑ ( T d ) for non-delay-sensitive services under self-similar ON/OFF traffic is obtained in a manner analogous to that for Poisson arrivals, but it is adapted to capture long-range dependence and burstiness. Assuming an aggregated arrival rate λ eff , the average frame generation time in this case is
Φ ( S w ) = N λ eff
Then, the average queuing delay can be estimated as
ϑ ( T d ) ρ · σ 2 2 ( 1 ρ ) · 1 Φ ( S w )
or, in simplified form,
ϑ ( T d ) N · l s 2 · λ eff · ( 1 ρ )
where ρ is the queue utilization factor. This formulation captures the impact of long-term correlations in bursty traffic on the frame queuing behavior.
Finally, all the encapsulated transmission frames subsequently enter the virtual channel scheduling process and carry out the next transmission.

3. Materials and Methods

In this section, we detail the research methodology adopted to develop and validate the proposed advanced intelligent frame generation (AIFG) algorithm. The complete process is structured into several key steps, as illustrated in Figure 2.
  • Queuing System Modeling: We establish the system model for the AOS environment, integrating heterogeneous data sources. The system architecture includes service classification (delay-sensitive and non-delay-sensitive), queuing structures, and differentiated framing processes, as described in Section 2.
  • Problem Formulation: The framing problem for delay-sensitive services is formulated as a Markov decision process, where the agent selects optimal frame generation times to minimize queuing delay and maximize the transmission efficiency. Key mathematical models include the Poisson arrival model (Equation (1)), average frame generation time and packet delay equations (Equations (2)–(6)), and the self-similar traffic model formula (Equations (7)–(11)). The above content is displayed in Section 2.
  • Algorithm Design: First, a model algorithm for Optimal Frame Generation Time Threshold Calculation was developed, based on which the RTFG was subsequently derived. A real-time framing algorithm (RTFG) is designed for delay-sensitive queues to satisfy strict latency constraints. A high-efficiency framing algorithm (HEFG) is applied to non-delay-sensitive queues for maximizing frame utilization. Then, an advanced intelligent frame generation algorithm (AIFG) is composed of the RTFG and HEFG. Reinforcement learning-based adaptive framing using the PPO algorithm optimizes frame generation thresholds dynamically based on real-time traffic conditions. The above content is displayed in Section 2 and Section 4.
  • Simulation Setup: We implement the AIFG algorithm using NS-3 and PyTorch. The simulation experiments are conducted under varying traffic intensities and service mixes. The key simulation parameters include a fixed frame capacity, delay threshold for real-time services, and maximum tolerable delay for non-delay-sensitive services. More than 300-round simulation runs are performed to ensure statistical reliability, and average values along with standard deviations are reported. The above content is shown in Section 5.
  • Performance Evaluation: The performance metrics used for evaluation include the following: average queuing delay, frame multiplexing efficiency and system throughput. Comparative analyses are performed against conventional isochronous framing and high-efficiency framing baselines. Additionally, the proposed AIFG algorithm is evaluated alongside a DQN+HEFG algorithm to further demonstrate its advantages in learning stability, convergence efficiency, and QoS adaptability.
The complete research process is summarized in Figure 2, showing the sequence from system modeling, problem formulation, algorithm design, and simulation implementation to the performance evaluation.

4. Optimal Frame Generation Time Threshold Calculation

Real-time frame generation in AOS faces a fundamental trade-off between latency and transmission efficiency. Fixed threshold-based strategies are unable to adapt to dynamic traffic patterns and often fail to meet delay constraints for delay-sensitive services such as voice and control commands. Therefore, this section investigates a dynamic threshold adjustment method that minimizes queuing delay while maintaining high frame utilization. Among them, the determination of the optimal framing time threshold is the basic condition for the implementation of the real-time framing algorithm. Therefore, we first explain the framing time threshold calculation algorithm. To this end, we formulate the frame generation problem as a Markov decision process and propose a reinforcement learning-based solution using the Proximal Policy Optimization algorithm. The agent learns to select the optimal frame generation threshold dynamically by interacting with the queuing environment and receiving feedback based on performance metrics, such as delay and frame efficiency.

4.1. Overall Structure

In the real-time frame generation (RTFG) module, under the condition that the system buffer capacity is fixed, if the frame generation waiting time threshold T t h is set too small, the average packet delay will decrease. However, this will require the addition of a large number of idle packets in the MPDU, leading to a reduction in the frame multiplexing efficiency. On the other hand, if T t h is set too large, the frame multiplexing efficiency will improve, but the average packet delay will increase, which will negatively impact real-time data with stringent delay requirements. Therefore, it is necessary to calculate the optimal frame generation time threshold T t h .
In this paper, the problem is modeled as a Markov decision process M = ( S , A , P , R ) and solved with the Proximal Policy Optimization algorithm. The overall structure of the optimal frame generation time threshold calculation model is shown in Figure 3. The model involves three main components: state space, action space, and reward.
As illustrated in Figure 3, the proposed framework consists of three main interacting modules: the agent, the environment, and the reward computation module.
  • Agent: The agent module contains the actor network, critic network, and the weight parameter vector of the policy function. The agent implements a Proximal Policy Optimization algorithm with an actor–critic structure. The actor network generates actions a t , representing the selected frame generation threshold based on the current state s t , while the critic network evaluates the expected return of each action. These networks are trained using a clipped surrogate loss function, value function regression, and policy entropy regularization.
  • Environment: The environment models the delay-sensitive queuing system and real-time frame generation process in the AOS scenario. It includes idle packet filling and transmission scheduling based on traffic load and queue conditions. At each time step, the environment produces a new state s t , capturing features such as packet arrival rate, queue length, frame generation time, and frame length.
  • Reward: The reward module quantifies the QoS outcome of each action by combining parameters such as frame multiplexing efficiency and average delay into a composite reward signal r t . This reward is used to reinforce beneficial actions and suppress suboptimal ones, guiding the agent to converge toward optimal framing policies.
Moreover, the buffer stores critical transition data collected during agent–environment interactions. Specifically, it logs tuples of ( s k , a k , π o l d ( a k s k ) , r k ), including the current state, selected action, action probability under the old policy, and corresponding reward. These stored experiences are used for batch policy updates, enabling efficient and stable training.
These components operate in a closed feedback loop where the agent continuously observes the environment, takes action, receives rewards, stores transitions in a buffer, and updates its policy through backpropagation. The integrated design ensures adaptive, real-time optimization of framing behavior in delay-sensitive service transmission.

4.1.1. State Space

In this paper, the training cycle is divided into several continuous time slots. At time slot t, the agent observes and collects the environment state s t , and the state space for all time slots is represented as S = [ s 1 , s 2 , , s t ] , s t , defined as Equation (12):
s t = { R s , L s , T f , L f }
where R s represents the average arrival rate of packets in time slot t, and L s represents the average queue length in time slot t, which is used to determine the current load status. Specifically, if the queue length is large, it indicates that the current queue is in a state of congestion. Thus, the next round should be framed as soon as possible, and the frame threshold should be reduced. T f represents the average frame generation time in time slot t, and L f represents the average frame length in time slot t.
This state space representation is independently constructed by us, taking into account the queuing dynamics and framing requirements specific to the AOS environment.

4.1.2. Action Space

After each time slot, the agent makes decisions based on the environment state and the reward function. The action space in the current time slot is A = [ a 1 , a 2 , , a t ] , where the action a t represents the frame generation threshold for the RTFG algorithm. The decision on the threshold value determines when the real-time data packets in the queue will be framed. The action impacts the environment and interacts with the environment based on the chosen threshold. The range of actions is a t [ 0 , T m ] , where T m is the maximum allowed average queue delay for real-time business. In time slot t, after the algorithm’s decision, a fixed frame generation threshold T t h is set, and a fixed frame length L t h is also established. According to the rules of the real-time frame generation algorithm, the data packets in the real-time data service queue are combined into a data frame and sent only when at least one of these two conditions is met. Obviously, the threshold value of one round will affect the decision of the next round.
In continuous action space, the standard deviation σ of the action decays gradually to a minimum value σ min . This parameter controls the randomness of action selection. The update rule for the standard deviation is as shown in Equation (13):
σ t + 1 = max ( σ t Υ , σ min )
where σ t is the standard deviation of the action after the t time interval update, Υ = α t is the decay rate, α is the proportional coefficient of the decay rate, and σ min is the minimum standard deviation of the action. The constraint condition for the standard deviation of the action is
σ t σ min
This ensures that the standard deviation of the action will not be lower than the set minimum value, thus ensuring continuous exploration of the agent.
The definition of the action space, including the mapping between the selected threshold and frame generation behavior, is tailored to reflect the control objectives of our system model.

4.1.3. Reward Function

During each training step, the agent receives the current reward value r t based on the composite function R ( s t , a t ) , which indicates the reward for the action a t taken at state s t of the current slot t. The reward function value r t is obtained by feedback after setting the threshold value. Moreover, r t provides an effective signal for the agent to guide its decision making in the next training step. The reward function defined in this paper is
R ( s t , a t ) = α L f β T f γ T h T f δ T h β L f α T f o t h e r w i s e
where L f is the normalized average effective frame length (excluding idle packets), T f  is the normalized average frame generation time, and α and β are the adjustable weights for the average effective frame length and average frame generation time, respectively, with α + β = 1 .
The reward function in Equation (15) is specifically formulated in this work, combining delay and frame utilization in a custom composite formulation to reflect the QoS needs of real-time services in the AOS.

4.2. Algorithm Design

The optimal frame generation time threshold is derived based on the policy. The policy is typically represented by π θ . When the state is s t , the probability of the agent executing action a t based on the parameters θ is given by Equation (16):
π θ ( a s ) = p [ a t = a s t = s , θ ]
where θ is the weight parameter vector of the policy function; πθ (as) is the policy function obtained by function fitting of the parameter vector θ . When the policy gradient is E ^ ( θ t ) and the learning rate is α , the formula for updating the policy parameter θ of the objective function by using the gradient rise method is as shown in Equation (17):
θ t + 1 = θ t + α E ^ ( θ t )
In this paper, the transition function P ( s t + 1 | s t , a t ) is used to represent the probability that the environment state is transferred from s t to s t + 1 . The actions made by the agent will affect the state of the environment and the related reward values. For example, the frame time threshold calculated by the time slot t 1 will affect the average queue length, average frame time and average frame length in the time slot t. If the threshold value is large, the number of packets combined in the data frame will increase, and the average queue length will decrease. At the same time, the average frame time and average frame length will also increase.
The policy optimization and loss function structure presented in this section is formulated by the authors to support stable training in the context of real-time adaptive framing. The specific expressions for the surrogate objective, critic loss, and entropy term reflect our design choices.

4.2.1. Clipped Surrogate Objective

The clipped surrogate objective function κ C L I P ( θ ) defined in this paper is
κ C L I P ( θ ) = E t min r t ( θ ) A ^ t , c l i p r t ( θ ) , 1 ϵ , 1 + ϵ A ^ t
where r t ( θ ) = π θ ( a t s t ) π θ old ( a t s t ) represents the probability ratio of the current policy and the old policy. A ^ t represents the advantage estimation of the time interval t, and the advantage function represents the advantage of taking action a t under the state s t . ϵ is a clipping coefficient used to limit the range of policy updates, which is used to limit the range of policy updates. c l i p r t ( θ ) , 1 ϵ , 1 + ϵ limits r t ( θ ) to the range of [ 1 ϵ , 1 + ϵ ] . This shearing mechanism ensures that the policy update does not cause the r t ( θ ) to be too large or too small, thereby avoiding excessively deviating from the old policy. At the same time, the clipping constraint of the policy update range is
1 ϵ r t ( θ ) 1 + ϵ
The difference between the new policy and the old policy is limited to a small range, avoiding the instability caused by excessive policy updates.

4.2.2. Mean Squared Error (MSE) of the State Value Function

The mean squared error (MSE) of the state value function κ critic ( θ ) is defined as
κ critic ( θ ) = 1 N N ( G t V θ ( s t ) ) 2
This formula shows the mean square error between the predicted state value function V θ ( s t ) and the actual return G t . V θ ( s t ) is the estimated state value of the state s i evaluation network under the current policy, and G t is the return estimated by the Monte Carlo method. The specific formula for the total return with the time step of t is given by Equation (21):
G t = R t + γ R t + 1 + γ 2 R t + 2 + . . .
where γ is the discount factor, typically set between [ 0 , 1 ] , representing the weight of future rewards. The agent learns more accurate state value estimation by minimizing the mean square error (MSE).

4.2.3. Policy Entropy Reward Function

The policy entropy reward function κ S ( θ ) defined in this paper is
κ S ( θ ) = E t H ( π θ ( · | s t ) )
The policy entropy reward function can improve the degree of divergence of reinforcement learning policy exploration and achieve optimal decision generation through a more comprehensive decision test. H ( π θ ( · | s t ) ) represents the entropy of the policy π θ under the state s t , which represents the randomness of the policy. The larger the entropy, the higher the randomness of the policy and the stronger the exploration.
Finally, the loss function combines these three parts of functions to obtain
κ θ = κ C L I P θ + c 1 κ critic θ c 2 κ S θ
In this formula, c 1 and c 2 are hyperparameters that weigh the value function loss and entropy reward, respectively, and c 1 and c 2 control the weight of the value function loss and entropy penalty term, respectively. The overall goal of the algorithm is to maximize the expected return by optimizing the proximal policy objective and the entropy term, while minimizing the value function loss. At the same time, the policy updates are constrained to ensure stable training.
The entropy-based regularization and its integration into the total loss function are designed by the authors to balance exploration and stability in the decision-making process.
The pseudo-code of the optimal frame generation time threshold calculation algorithm is as shown in Algorithm 1.
Algorithm 1 Optimal Frame Generation Time Threshold Calculation
  • Require: PPO hyperparameters, number of iterations
  • Ensure: Optimal frame generation time threshold T h
  • Initialize network parameters θ , and buffer D
  • for each episode do
  •    Initialize PPO hyperparameters
  •    for each Delay Sensitive Queue do
  •      Construct the state space s k for current time slot
  •      According to s k , the action a k is selected by π o l d , and save action probability P ( π o l d ( a k s k ) )
  •      if Execute a k , r t ( θ ) > 1  then
  •         Abandon the threshold value result derived from this action
  •         Break
  •         Execute a k , s k s k + 1 , obtain the reward function
  •         Store the current dataset { s k , a k , P ( π o l d ( a k s k ) ) , r k } in buffer D
  •         if  a k a 1 (not the first action) then
  •           for  n u p d a t e  do
  •              if time step reaches u p d a t e _ timestep  then
  •                Calculate objective function κ CLIP ( θ ) using Equation (18), update PPO policy loss function
  •                Calculate the mean squared error using Equations (20) and (21), update the critic network
  •              end if
  •              if action space is sufficient & time step = action std decay freq (action standard deviation decay frequency) then
  •                Update action standard deviation according to Equation σ t + 1 = max ( σ t Υ , σ min )
  •              end if
  •              Complete a series of updates of the agent
  •           end for
  •           Calculate cumulative reward s u m ( r k ) and store it
  •         end if
  •         if the amount of data in buffer D exceeds maximum capacity > K  then
  •           Update network parameters θ old θ , clear buffer D
  •         end if
  •          s k s k + 1
  •      end if
  •    end for
  • end for

5. Advanced Intelligent Frame Generation Algorithm

As described in the system model in Section 2 of this paper, for the RTFG module, when the frame generation waiting time for the delay-sensitive queue reaches the threshold T h , the data packets arriving during this period are encapsulated into frames. On the contrary, if the number of packets in the delay-sensitive queue has filled the MPDU packet area before the threshold value T h arrives, it is also encapsulated into a frame. For the HEFG module, if the packet size of the non-delay sensitive queue reaches the maximum length L h of an MPDU packet area, it is encapsulated into a frame without filling the idle packet. The pseudo-code of the advanced intelligent framing generation algorithm in Advanced Orbiting Systems designed for differentiated QoS requirements is as shown in Algorithm 2.
Algorithm 2: Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Demands in Advanced Orbiting Systems
  • Require: Average packet arrival rate λ ; number of iterations G ; decay frequency α ; policy update period T c ; discount factor γ ; clipping parameter c ; number of policy updates per training episode n u p d a t e ; policy optimization network learning rate ρ ; training sample size N
  • Ensure: Average frame generation time Φ ( S w ) , frame multiplexing efficiency E ( e f f m ) , average packet delay ϑ ( T d ) , throughput
  • Initialize network simulation environment based on network parameters and create simulation topology
  • if  U integerValue = 1 (indicating real-time business) then
  •    Generate a data stream that follows a Poisson distribution based on the real-time business requirements
  •    Initialize network parameters θ
  •    Use Algorithm 1 to obtain the optimal frame generation time threshold T h
  •    Set the real-time frame generation threshold T t h = T h and set maximum MPDU length for the transmission frame L t h = L h
  •    if the number of packets arriving within T 1 exceeds the threshold n = N  then
  •      Encapsulate n packets into the MPDU packet area
  •      Set T t h = T 1
  •      Calculate Φ ( S w ) , E ( e f f m ) and ϑ ( T d ) by using Equations (4), (5) and (6), respectively
  •    else if the number of packets arriving in T n < N then
  •      Encapsulate packets into the MPDU packet area
  •      Fill the idle packets, and set T t h = T
  •      Calculate Φ ( S w ) , E ( e f f m ) and ϑ ( T d ) by using Equations (4), (5) and (6) respectively
  •      End the time slot, get the reward r t and the new state s t + 1
  •      Release the current frame and send it to the virtual channel scheduler
  •    end if
  • else if   U integerValue = 2 (indicating non-real-time business) then
  •    Generate a data stream that conforms to a self-similar traffic model, in accordance with the characteristics of non-real-time service requirements
  •    Initialize θ
  •    Set L t h = L h , and observe the MPDU packet area status
  •    if the arriving packet length L = L h , MPDU packet area is filled then
  •      Encapsulate the data into the transmission frame and calculate Φ ( S w ) and ϑ ( T d ) for the HEFG module by using Equations (8)–(11)
  •    end if
  • end if
  • The iteration is completed. Return Φ ( S w ) for both delay-sensitive and non-delay-sensitive queues, E ( e f f m ) of the delay-sensitive queue, and ϑ ( T d ) for both delay-sensitive and non-delay-sensitive queues
The advanced intelligent frame generation algorithm described in this section, including the modeling of action, reward, and policy structure, is proposed by the authors as an original design for the specific requirements of AOS.

6. Simulation Results and Analysis

This section presents the simulation results that evaluate the effectiveness of the proposed advanced intelligent frame generation (AIFG) algorithm under different delay constraints. The main goals of the simulation are to validate the adaptability of AIFG in varying network conditions, and to compare its performance against three benchmark algorithms: AFG + HEFG, IFG + HEFG, and FIFO + FIFO. Additionally, to strengthen the evaluation from the perspective of reinforcement learning methodologies, we introduce a Deep Q-Network-combined framing policy (DQN + HEFG) as an additional comparative baseline. This inclusion supplements the analysis with a machine learning-based benchmark, allowing for a more comprehensive performance assessment of AIFG against both traditional and contemporary intelligent algorithms. The simulation scenario emulates data traffic in an AOS, incorporating delay-sensitive and non-delay-sensitive services with dynamic arrival patterns.
The evaluation focuses on three critical performance indicators: average queue delay, frame multiplexing efficiency, and throughput. To explore the robustness of AIFG, the simulations are conducted under four different delay threshold settings: 5 ms, 10 ms, 15 ms, and 20 ms.
Beyond the standard evaluations, this section further investigates the scalability and practical applicability of AIFG. We extend the simulation environment by varying the number of network nodes (from 25 up to 200), altering the MPDU payload length (ranging from 800 to 1800 bytes), and adjusting satellite link bandwidth (from 3 Mbps to 20 Mbps). These additional experiments serve to analyze the computational overhead, convergence behavior, and performance resilience of AIFG under more realistic and heterogeneous system conditions.
The results demonstrate that AIFG can dynamically adjust its frame generation policy based on real-time feedback from the system environment, effectively balancing delay and efficiency. The subsequent subsections provide detailed discussions of the simulation setup, algorithm configurations, and quantitative comparisons under each experimental scenario.

6.1. Simulation Scenario and Parameter Settings

The topology of the simulation system designed for this study is shown in Figure 4. The simulation system consists of 50 transmitting nodes representing spacecraft probe loads and 50 receiving nodes representing simulated ground terminal devices. After collecting various types of data and encapsulating and processing the data accordingly, the payloads divide the data packets into real-time data packets and non-real-time data packets. Then, the payloads send them to the communication satellite, respectively, entering the time-sensitive and non-time-sensitive service queues. Afterwards, corresponding to different service type queues, different frame generation algorithms are applied to complete the data transmission between communication satellites. Ultimately, based on the business type, the satellite transmits the data to multiple ground terminal devices for further analysis and processing. In order to simulate the complex space links and various services in the AOS, the communication satellite nodes transmit typical acquisition data and communication information collected by aerospace payloads to the receiving terminal nodes. The simulation focuses on the frame generation module in the system. Based on the random data generated by the ON/OFF source model, the link is simulated as the link type of the communication satellite. At the same time, AIFG in the AOS for differentiated QoS requirements is deployed on the transmitting communication satellite to perform data transmission between two satellite nodes.
To evaluate the performance of the proposed AIFG algorithm, a simulation environment was configured to emulate the data transmission scenario of Advanced Orbiting Systems. The specific simulation parameters settings are shown in Table 1, and they are summarized as follows. The simulation runs for 50 s, during which 100 real-time data streams and 110 non-real-time data streams are simultaneously transmitted. Real-time data include delay-sensitive services, such as voice or telemetry data, while non-real-time data include observation and video information. The bottleneck link bandwidth is set to 1 Mbps to model constrained access segments, while the satellite link bandwidth is configured as 6 Mbps, simulating higher-capacity backbone transmission. The simulation operates in discrete time steps, with each time slot spanning 50 ms, during which the agent observes system states and decides frame generation thresholds. The maximum MPDU size for all framing algorithms (RTFG, IFG, HEFG) is unified at 1500 bytes, corresponding to a standard Ethernet MTU, ensuring consistency across different service types.
The main simulation parameters of the PPO algorithm are set as shown in Table 2.
In the simulation case, to demonstrate the performance advantages of the proposed algorithm in terms of network multiplexing efficiency, average queue delay, and throughput, the proposed algorithm is compared with the following four algorithms:
  • AFG + HEFG: The adaptive frame generation (AFG) algorithm is used for time-sensitive service queues, and the high-efficiency frame generation (HEFG) algorithm is used for non-time-sensitive service queues. Because of the requirements of AOS transmission frames, the frame generation length for the HEFG algorithm is fixed at 1500 bytes.
  • IFG + HEFG: For delay-sensitive business queues, the isochronous frame generation (IFG) algorithm is applied. For non-delay-sensitive business queues, the HEFG algorithm is used. The frame generation length for HEFG is also fixed at 1500 bytes.
  • FIFO + FIFO: No frame generation algorithm is applied for either the delay-sensitive or non-delay-sensitive business queues. Instead, the standard First-In First-Out (FIFO) algorithm is used.
  • DQN + HEFG: The Deep Q-Network (DQN) algorithm dynamically selects the delay threshold for real-time service queues, while HEFG continues to handle non-delay-sensitive services.

6.2. Simulation Results and Analysis

Under different delay requirements, the AOS automatically adjusts and produces varying network performance results, leading to changes in parameters such as the final threshold values, delay, multiplexing efficiency, and throughput. To achieve more accurate results from the advanced intelligent frame generation algorithm, the simulation sets four different delay requirements T h : 5 ms, 10 ms, 15 ms, and 20 ms. At the same time, according to the delay requirements, iteration parameters, reward function parameters, and final reward values of the old and new policy are flexibly adjusted to train the agent to obtain a more optimized delay threshold.
In the simulation of this algorithm, the change in the delay requirement combined with adjustments to the reward function parameters can guide the correct training and learning of the agent, which makes the agent complete. Then, the agent is able to make the optimal policy. The difference in the four parameter settings in the reward function will greatly affect the performance of the algorithm. Therefore, the numerical selection of the parameter group is particularly important. This simulation takes the setting delay requirement of 5 ms as an example. Multiple parameter sets of reward functions are as shown in Table 3. For each case, the reward function, throughput, average delay, and multiplexing efficiency are observed. The optimal parameter group value is obtained via a comparison, and the reward value convergence curve changes as shown in Figure 5.
Different parameter settings will bring different changes in the value of the reward function. Figure 5 clearly shows the mainstream trend of the reward function, which gradually rises in the early stage of training. When the number of iterations is about 40, it begins to stabilize and enters the platform period, indicating the convergence of the reward function. Among the four parameter groups selected, the growth rate of the convergence function is the fastest in the case of Parameter Group 3, and the growth rate of the Parameter Group 2 is slower. At the same time, the convergence effect of Parameter Groups 1, 3 and 4 is better, but the convergence effect of Parameter Group 2 is poor.
From Figure 6, it can be observed that the average delay and throughput are optimal for Parameter Group 3, while the multiplexing efficiency is highest for Parameter Group 4. Overall, selecting Parameter Group 3 for AIFG algorithm training is the most optimal choice.
For the four different network processing strategies, different simulation performances are observed under varying delay threshold requirements. Except for the FIFO + FIFO algorithm, which cannot accurately set delay requirements, the other strategies produce a single result under each delay requirement. In this simulation, the parameter groups of the optimal reward function are set for the four delay requirements, and the specific values are shown in Table 4. These sets determine the weights for delay and efficiency trade-offs in the agent’s learning process. The detailed values of α , β , γ , and δ for each delay condition are summarized in Table 4, covering thresholds of 5 ms, 10 ms, 15 ms, and 20 ms, respectively. The following series of figures provides the frame generation delay, throughput, and multiplexing efficiency for the four strategies under different delay requirements.
From Figure 7, it can be seen that the throughput of the AIFG algorithm is generally higher compared to other frame generation algorithms. When the delay requirement is 5 ms, the throughput of the AIFG algorithm is 466 Kbps, which is an increase of 78.05% over both the AFG + HEFG and IFG + HEFG algorithms. With a delay requirement of 10 ms, the throughput of AIFG is 464 Kbps, which is an increase of 25.37% compared to that of AFG + HEFG and IFG + HEFG. When the delay requirement is 15 ms, the AIFG algorithm achieves the maximum throughput of 606.456 Kbps, which is 28.66% higher than that of the AFG + HEFG and IFG + HEFG algorithms. At a delay requirement of 20 ms, the throughput of AIFG also improves, with an increase of 17.18% compared to both AFG + HEFG and IFG + HEFG. The AIFG consistently exhibits superior throughput, especially under 5 ms and 10 ms delay thresholds, where it outperforms DQN + HEFG by margin of 49.9% and 12.7% respectively. Under 15 ms and 20 ms delay thresholds, the throughput of AIFG is 15.8% and 6.1% higher than that of DQN + HEFG, respectively. When the delay requirement is low, AIFG shows a higher throughput improvement rate, and its performance change is optimal. Conversely, when the delay requirement is high, AIFG is able to achieve a larger throughput value. This highlights the effectiveness of PPO in guiding reward shaping and decision making over longer time horizons.
Due to the FIFO transmission mechanism, the queue’s multiplexing efficiency is always 100%. As shown in Figure 8, the multiplexing efficiency of the AIFG algorithm is higher compared to the IFG + HEFG, AFG + HEFG and DQN + HEFG algorithms under smaller delay requirements. With a delay requirement of 5 ms, AIFG’s multiplexing efficiency is 69.25%, which is an increase of 92.36% over both AFG + HEFG and IFG + HEFG, and an increase of 63.2% over DQN + HEFG. Under a delay requirement of 10 ms, the multiplexing efficiency of the AIFG algorithm is 69.1%, which is 52.47% higher than that of the AFG + HEFG and IFG + HEFG algorithms, and an increase of 35.5% over DQN + HEFG. At a 15 ms delay requirement, AIFG’s multiplexing efficiency is 12.26% higher than that of AFG + HEFG and IFG + HEFG, while it is only 3.2% higher than that of DQN + HEFG. When the delay requirement is 20 ms, the multiplexing efficiency of AIFG reaches its maximum value of 94.7%, with an increase of 1.4% compared to DQN + HEFG, but with a slight decrease compared to AFG + HEFG and IFG + HEFG. When the delay requirement is high, the multiplexing efficiency improvement rate of AIFG decreases. This is constrained by the time threshold and influenced by the fixed AOS transmission frame size. Nevertheless, the overall trend in multiplexing efficiency remains upward. Among learning-based methods, AIFG shows better multiplexing efficiency than DQN + HEFG, especially under strict delay conditions (5 ms, 10 ms), indicating a more refined policy learned through PPO.
Figure 9 shows that AIFG consistently exhibits the lowest delay across all time requirements, outperforming all other methods. This is especially noticeable under higher time requirements (15 ms and 20 ms), where the delay is substantially smaller. Meanwhile, the FIFO-based method exhibits slower processing times due to its inherent characteristics. With a delay requirement of 5 ms, the average queue delay for AIFG, DQN + HEFG, AFG + HEFG, and IFG + HEFG is almost the same, ranging from 6 to 7 ms. However, at a 10 ms delay requirement, AIFG’s average queue delay is reduced by 19.39% compared to AFG + HEFG and by 40.33% compared to IFG + HEFG and by 8.9% compared to DQN + HEFG. At a 15 ms delay requirement, AIFG’s average queue delay decreases by 37.30% compared to AFG + HEFG and by 39.62% compared to IFG + HEFG and by 22.1% compared to DQN + HEFG. At a 20 ms delay requirement, AIFG’s average queue delay is 43.90% lower than AFG + HEFG and 43.45% lower than IFG + HEFG and 30.3% lower than DQN + HEFG. When the delay requirement is low (e.g., 5 ms), the queue length typically does not exceed the AFG threshold, so the algorithm’s performance is similar to that of AFG, and the advantages of AIFG are not fully manifested. However, when the delay requirement is high, AIFG dynamically optimizes the gap between frames, reducing delays and improving overall performance by adapting to changing conditions in the transmission environment. In this part, DQN + HEFG performs better than traditional rule-based approaches (AFG/IFG + HEFG) but does not reach the stability and optimization level of PPO-based AIFG. The simulation results validate the adaptability of the AIFG algorithm under varying delay requirements and the advantage of AIFG in reducing average queue delay.
Based on the analysis of the above results, in the face of time-sensitive service queues, the AIFG algorithm proposed in this paper can effectively adjust the framing threshold according to the queue state. Compared with the traditional algorithm, it achieves a better multiplexing efficiency and a lower queuing delay while ensuring excellent throughput, which is an effective performance improvement. In terms of throughput performance, the AIFG algorithm has an average increase of 31.92% compared with the AFG + HEFG, IFG + HEFG and DQN + HEFG algorithms. In terms of the multiplexing efficiency, the AIFG algorithm has an average increase of 74.93% compared with the AFG + HEFG, IFG + HEFG and DQN + HEFG algorithms when the delay requirement is below 5 ms. In terms of the average queuing delay, the AIFG algorithm is 33.53% lower than the AFG + HEFG algorithm and 41.13% lower than the IFG + HEFG algorithm and 20.43% lower than the DQN + HEFG algorithm when the delay requirement is above 5 ms. In reinforcement learning applications, while DQN provides improvements over heuristic baselines, the PPO-based AIFG exhibits higher adaptability, stability, and better global optimization performance, making it more suitable for diverse and dynamic space network environments.
To evaluate the scalability and robustness of the proposed AIFG algorithm under varying network sizes, we expand the simulation to assess performance across different network sizes, including 25–25, 50–50, 100–100, and 200–200 transmitter–receiver pairs. We maintain the same satellite relay architecture. These different simulation situations are all performed under the condition that the delay requirement is 20 ms. To further validate the scalability of the proposed AIFG algorithm, we specifically focus on the PPO training time, training memory consumption, convergence iterations, and resulting performance metrics such as delay, multiplexing efficiency, and throughput. The key performance metrics observed include the following:
  • Training computational overhead, measured as the cumulative time and training memory consumption required for PPO training to reach a stable policy.
  • Convergence time, quantified as the number of iterations needed for the reward function to stabilize within a defined tolerance range.
As shown in Figure 10, the reward sum becomes negative in large-scale cases due to an increased queuing delay. In contrast, for smaller networks (25 and 50 nodes), the algorithm achieves convergence within fewer episodes and yields positive reward values. Notably, the algorithm’s accumulated reward converges well even in large-scale environments.
As summarized in Table 5, the training time increases with the node scale, reaching approximately 798 s in the 200-node configuration, compared to just 41.5 s for 50 nodes. Similarly, the number of PPO training iterations required for reward convergence increases from 40 to 80. Nonetheless, AIFG still delivers high efficiency and throughput across all scenarios. Although the computational cost increases as the network grows, the performance in delay remains acceptable.
These results, summarized in Figure 10 and Table 5, demonstrate that AIFG remains computationally viable and robust under scale expansion, offering consistent performance and QoS adaptability even in dense satellite constellations. In summary, the AIFG algorithm demonstrates strong scalability with a predictable computational cost growth and robust convergence behavior, indicating its feasibility in large-scale deployment.
To further investigate the robustness and adaptability of the proposed AIFG algorithm under varying network configurations, two additional simulation scenarios were conducted. The first scenario explores the influence of different MPDU sizes on system performance. Specifically, MPDU sizes were varied across 800, 1200, 1500 (default), and 1800 bytes. As shown in Table 6, the AIFG algorithm maintains consistently high throughput and multiplexing efficiency across different MPDU sizes. The results show that while increasing MPDU size improves multiplexing efficiency up to a point, overly large frame sizes (e.g., 1800 bytes) lead to degraded throughput, likely due to transmission and fragmentation inefficiencies. The default size of 1500 bytes strikes a good balance among the average delay (10.77 ms), efficiency (99.49%), and throughput (566.45 Kbps), confirming the appropriateness of the default setting. Interestingly, even when reducing the frame size to 800 bytes, the algorithm achieves strong performance with an average delay of only 7.23 ms and a throughput of 559.61 Kbps. These results indicate that the AIFG algorithm is capable of dynamically adjusting the frame threshold to maintain high performance, even under less efficient framing conditions.
In the second scenario, the satellite link bandwidth was adjusted across 3, 6 (default), 9, 15, and 20 Mbps to simulate the variability caused by orbital distance, channel fading, or dynamic scheduling. As summarized in Table 7, the AIFG algorithm continues to demonstrate resilient performance, with stable average delay and throughput across all bandwidth conditions. Although slight improvements in throughput are observed as bandwidth increases, the average delay and efficiency remain within a narrow and acceptable range. The default configuration of 6 Mbps achieved the highest throughput at 575.18 Kbps and the best multiplexing efficiency at 99.52%. Even in constrained link scenarios (e.g., 3 Mbps), the algorithm sustains efficient operation, with a throughput of 566.86 Kbps. These findings confirm the algorithm’s robustness under dynamic bandwidth constraints and validate its practical applicability in real-world spaceborne communication environments.
These extended experiments further support that the AIFG algorithm not only provides adaptive framing under differentiated QoS demands but also maintains high levels of delay control, bandwidth efficiency, and throughput under fluctuating MPDU configurations and varying relay link conditions. This reinforces the algorithm’s suitability for deployment in dynamic space networking environments with heterogeneous service profiles.

6.3. Results

Based on the simulation results described in Section 6.2, the proposed AIFG algorithm demonstrates significant advantages over traditional framing algorithms across multiple dimensions:
  • Reduction in Queue Delay: As claimed in the abstract, AIFG effectively reduces the average queuing delay for delay-sensitive services. Simulation results show a reduction of at least 8.90%, and up to 43.90%, compared to benchmark methods such as AFG + HEFG, IFG + HEFG, and DQN + HEFG. It demonstrated effective delay control under diverse QoS constraints.
  • Improvement in Throughput: The throughput achieved by AIFG increases by at least 6.10%, with improvements up to 78.05% under low delay constraints, confirming its efficiency in data delivery across time-sensitive and non-sensitive flows.
  • Enhancement of Multiplexing Efficiency: AIFG demonstrates improved frame utilization, with the multiplexing efficiency increasing by, at most, 92.36% in comparison with traditional methods, especially under stricter time thresholds.
These quantitative results fully support the core contributions presented in the abstract and validate the effectiveness of AIFG algorithm in adapting to differentiated QoS requirements in Advanced Orbiting Systems.
It is important to note that in the proposed AIFG algorithm, the framing time threshold is not statically set. Instead, it is dynamically adjusted at runtime by the reinforcement learning agent. The PPO algorithm continuously updates the threshold value based on the observed system state, including packet arrival rate, queue length, and frame generation time. As such, the threshold evolves differently under each delay requirement scenario (e.g., 5 ms, 10 ms, 15 ms, 20 ms), and there is no single fixed optimal value applicable across all conditions. During simulation, different delay requirement environments are used to train the agent, and reward parameters are tuned accordingly.
The performance improvements reported in this section—such as higher throughput, reduced average queue delay, and increased multiplexing efficiency—are the result of adaptive threshold selection throughout the simulation. The reward function parameters used to guide this adaptive process are provided in Table 3 and Table 4. These results demonstrate the effectiveness of using policy-based learning to dynamically regulate the frame generation process. AIFG achieves superior performance not through a static configuration but via a policy-driven dynamic adjustment under varying QoS conditions.

6.4. Discussion

Although the proposed AIFG algorithm has demonstrated substantial improvements in reducing queue delay, increasing throughput, and enhancing the multiplexing efficiency, several limitations and open challenges remain:
  • Simulation-Based Validation: The current evaluation is based solely on a simulation platform using NS-3 (version 3.33) and PyTorch (version 2.2.2). Real-world satellite communication environments introduce complexities, such as hardware constraints, channel fading, and protocol interoperability, which are not modeled here.
  • Static Traffic Model: The number of service streams and their classifications were fixed during the simulations. In practical deployments, service types and loads are often time-varying and mission-dependent, which may impact the algorithm’s adaptability.
  • Energy Efficiency Not Considered: This study focuses on QoS metrics without evaluating energy consumption, which is a critical constraint in space systems. Future work should explore low-power optimization and computational efficiency.
  • Algorithm Scalability and Convergence: Although PPO provides stable learning, its convergence speed and computational cost may become significant in larger-scale AOS. Exploring more lightweight or distributed RL techniques would be beneficial.
  • Cross-Layer Optimization: Integrating AIFG with transport-layer protocols or application-layer protocols, or co-designing with physical-layer resource adaptation, could further enhance end-to-end QoS.
These limitations provide a roadmap for future research and guide the further refinement of the AIFG framework toward deployment in practical aerospace communication systems.

7. Conclusions

To further enhance the service quality of the AOS for handling massive heterogeneous services, this paper proposes an advanced intelligent frame generation (AIFG) algorithm designed to meet differentiated QoS requirements. Based on the classification framework for delay-sensitive and non-delay-sensitive services, real-time frame generation and high-efficiency frame generation algorithms are designed for each type of service. In the real-time framing algorithm, this paper models the optimal framing time threshold calculation problem of delay-sensitive services as a Markov decision process and uses the reinforcement learning algorithm to obtain the optimal solution. For non-delay-sensitive services, the high-efficiency frame generation algorithm ensures that the frame is filled before transmission and it is adopted to maximize frame utilization. Simulation results show that compared to traditional frame generation algorithms and the learning-based DQN algorithm, the proposed algorithm accommodates the different transmission needs of real-time and non-real-time services, reduces the average queue delay for real-time services, increases the average multiplexing efficiency for services, and improves the overall network service quality of the AOS. For future work, efforts can focus on extending AIFG to real-world scenarios with more complex environmental dynamics and hardware constraints. The algorithm’s reinforcement learning-based structure and modular framing design also provide a flexible foundation for further research in intelligent communications across other delay-constrained or resource-limited network environments.

Author Contributions

Conceptualization, J.P. and J.C.; methodology, J.P.; software, J.P.; validation, J.P., H.S. and J.C.; formal analysis, J.P.; investigation, J.P., H.S. and Y.F.; resources, J.P.; data curation, J.P.; writing—original draft preparation, J.P.; writing—review and editing, J.P. and H.S.; visualization, J.P.; supervision, J.P. and J.C.; project administration, J.P.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Key Laboratory of Intelligent Support Technology for Complex Environments, Ministry of Education (No. B2202401).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within this article.

Acknowledgments

We would like to thank for the technical support provided to us by the Key Laboratory of Intelligent Support Technology for Complex Environments, Ministry of Education.

Conflicts of Interest

Author Jun Chen was employed by The 28th Research Institute of China Electronic Technology Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AFGAdaptive Frame Generation
IFGIsochronous Frame Generation
AIFGAdvanced Intelligent Frame Generation Algorithm
RTFGReal-time Framing
HEFGHigh-efficiency Framing
FIFOFirst-In First-Out
PPOProximal Policy Optimization
DQNDeep Q-Network
RLReinforcement Learning
AOSAdvanced Orbiting System
MPDUMultiplexing Protocol Data Unit
NS-3Network Simulator 3
QoSQuality of Service

References

  1. Bie, Y.; Zhang, X.; Wang, Y.; Hu, Z. Research on Virtual Channel Hybrid Scheduling Algorithm in Advanced Onorbit Systems. J. Electron. Inf. Technol. 2021, 43, 1913–1921. [Google Scholar]
  2. Liu, R.; Pan, W.; Su, J.; Fan, Y. Simulation of Multiplexing Efficiency of AOS Isochronous Frame Generation Algorithm Based on Self-similar Traffic Flow. Digit. Technol. Appl. 2015, 10, 32. [Google Scholar]
  3. Zhao, Y.; Liu, L. A Scheduling Model of Isochronous Frame Generation on Self-similar Traffic of AOS Multiplexing. Int. J. Online Eng. 2018, 14, 15. [Google Scholar] [CrossRef]
  4. Zhao, Y.; Bo, B.; Feng, Y. A Scheduling Method of Cross-Layers Optimization of Polling Weight for AOS Multiplexing. Int. J. Digit. Multimed. Broadcast. 2019, 2019, 2560623. [Google Scholar] [CrossRef]
  5. Li, Q.; Wang, R.; Tian, Y.; Liu, L.; Guo, Y. Research and simulation on high efficient frame generation model of AOS considering packet extracting time under finite buffer. In Proceedings of the 2014 IEEE 5th International Conference on Software Engineering and Service Science, Beijing, China, 27–29 June 2014; IEEE: New York, NY, USA, 2014; pp. 613–616. [Google Scholar]
  6. Bie, Y.; Zhang, Z.; Srivastava, G.; Hu, Z.; Laghari, A.A.; Sampedro, G.A.; Abbas, S. Adaptive Framing and Virtual Channel Scheduling Algorithm Based on Advanced Orbiting System for Consumer Sustainability in Industry 5.0. IEEE Trans. Consum. Electron. 2024, 70, 4882–4893. [Google Scholar] [CrossRef]
  7. Zhang, M.; Xia, Y. Simplified calculation on the time performance of high efficiency frame generation algorithm in advanced orbiting systems. In Proceedings of the 2013 5th IEEE International Symposium on Microwave, Antenna, Propagation and EMC Technologies for Wireless Communications, Chengdu, China, 29–31 October 2013; IEEE: New York, NY, USA, 2013; pp. 492–495. [Google Scholar]
  8. Ma, X.; Fu, Y.; Liu, T. An adaptive channel multiplexing method for spatial data based on AOS protocol. In Proceedings of the 2013 International Conference on Mechatronic Sciences, Electric Engineering and Computer (MEC), Shengyang, China, 20–22 December 2013; IEEE: New York, NY, USA, 2013; pp. 2870–2874. [Google Scholar]
  9. Tian, Y.; Na, X.; Xia, Y.; Gao, X.; Liu, L. Performance Analysis and Application Research of AOS Adaptive Frame Generation Algorithm. J. Astronaut. 2012, 2, 242–248. [Google Scholar]
  10. Bi, M.; Pan, C.; Zhao, Y. Throughput analysis and frame length adaptive for AOS packet service. In Proceedings of the 2009 International Conference on Networks Security, Wireless Communications and Trusted Computing, Wuhan, China, 25–26 April 2009; IEEE: New York, NY, USA, 2009; Volume 1, pp. 38–41. [Google Scholar]
  11. Dai, C.; Ceng, G.; Liang, J.; Zhang, D. AOS Adaptive Frame Generation Algorithm Based on Traffic Flow Prediction. Comput. Meas. Control 2017, 25, 176–178. [Google Scholar]
  12. Zhu, H. A Single-point Signal Timing Scheme Selection Algorithm Based on Q-learning. Highw. Automot. Appl. 2022, 1, 44–47. [Google Scholar]
  13. Clifton, J.; Laber, E. Q-learning: Theory and applications. Annu. Rev. Stat. Its Appl. 2020, 7, 279–301. [Google Scholar] [CrossRef]
  14. Zhao, D.; Wang, H.; Shao, K.; Zhu, Y. Deep reinforcement learning with experience replay based on SARSA. In Proceedings of the 2016 IEEE Symposium Series on Computational Intelligence (SSCI), Athens, Greece, 6–9 December 2016; IEEE: New York, NY, USA, 2016; pp. 1–6. [Google Scholar]
  15. Zhang, J.; Kim, J.; O’Donoghue, B.; Boyd, S. Sample efficient reinforcement learning with REINFORCE. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 10887–10895. [Google Scholar]
  16. Haarnoja, T.; Zhou, A.; Hartikainen, K.; Tucker, G.; Ha, S.; Tan, J.; Kumar, V.; Zhu, H.; Gupta, A.; Abbeel, P.; et al. Soft actor-critic algorithms and applications. arXiv 2018, arXiv:1812.05905. [Google Scholar]
  17. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  18. Lopes, G.C.; Ferreira, M.; da Silva Simões, A.; Colombini, E.L. Intelligent control of a quadrotor with proximal policy optimization reinforcement learning. In Proceedings of the 2018 Latin American Robotic Symposium, 2018 Brazilian Symposium on Robotics (SBR) and 2018 Workshop on Robotics in Education (WRE), João Pessoa, Brazil, 6–10 November 2018; IEEE: New York, NY, USA, 2018; pp. 503–508. [Google Scholar]
  19. Guan, Y.; Ren, Y.; Li, S.E.; Sun, Q.; Luo, L.; Li, K. Centralized cooperation for connected and automated vehicles at intersections by proximal policy optimization. IEEE Trans. Veh. Technol. 2020, 69, 12597–12608. [Google Scholar] [CrossRef]
  20. Padhye, V.; Lakshmanan, K.; Chaturvedi, A. Proximal policy optimization based hybrid recommender systems for large scale recommendations. Multimed. Tools Appl. 2023, 82, 20079–20100. [Google Scholar] [CrossRef]
  21. Luong, N.C.; Hoang, D.T.; Gong, S.; Niyato, D.; Wang, P.; Liang, Y.C.; Kim, D.I. Applications of Deep Reinforcement Learning in Communications and Networking: A Survey. IEEE Commun. Surv. Tutor. 2019, 21, 3133–3174. [Google Scholar] [CrossRef]
Figure 1. System model.
Figure 1. System model.
Electronics 14 01939 g001
Figure 2. Research methodology flowchart.
Figure 2. Research methodology flowchart.
Electronics 14 01939 g002
Figure 3. Overall structure of the optimal frame generation time threshold calculation model.
Figure 3. Overall structure of the optimal frame generation time threshold calculation model.
Electronics 14 01939 g003
Figure 4. Topology diagram.
Figure 4. Topology diagram.
Electronics 14 01939 g004
Figure 5. Comparison of reward convergence curves for four parameter settings (delay requirement: 5 ms).
Figure 5. Comparison of reward convergence curves for four parameter settings (delay requirement: 5 ms).
Electronics 14 01939 g005
Figure 6. Simulation Performance comparison of AIFG algorithm for four parameter settings (delay requirement: 5 ms).
Figure 6. Simulation Performance comparison of AIFG algorithm for four parameter settings (delay requirement: 5 ms).
Electronics 14 01939 g006
Figure 7. Throughput simulation results (delay-sensitive service queue).
Figure 7. Throughput simulation results (delay-sensitive service queue).
Electronics 14 01939 g007
Figure 8. Multiplexing efficiency simulation results (delay-sensitive service queue).
Figure 8. Multiplexing efficiency simulation results (delay-sensitive service queue).
Electronics 14 01939 g008
Figure 9. Average queue delay simulation results (delay-sensitive service queue).
Figure 9. Average queue delay simulation results (delay-sensitive service queue).
Electronics 14 01939 g009
Figure 10. Reward convergence curves under different network scales (delay requirement: 20 ms).
Figure 10. Reward convergence curves under different network scales (delay requirement: 20 ms).
Electronics 14 01939 g010
Table 1. Simulation parameters.
Table 1. Simulation parameters.
Simulation ParameterValue
Simulation Time50 s
Real-time Data Traffic100 streams
Non-real-time Data Traffic110 streams
Bottleneck Link Bandwidth1 Mbps
Satellite Link Bandwidth6 Mbps
Time Slot Size50 ms
Maximum MPDU Size in RTFG/IFG/HEFG1500 bytes
Table 2. PPO algorithm parameters.
Table 2. PPO algorithm parameters.
Algorithm ParameterDefinition
Standard Deviation of Initial Action Distribution σ 2.0
Linear Decay Frequency ξ 0.2 Hz
Minimum Action Standard Deviation σ min 0.4
Action Standard Deviation Decay Frequency (per time step)5000 Hz
Discount Factor γ 0.99
Clipping Parameter c 0.2
Policy Update Frequency100 Hz
Model Saving Frequency 1 × 10 5 Hz
Number of Running Rounds per PPO Update Cycle1
Evaluation Network Learning Rate 0.001
Proximal Policy Action Network Learning Rate ρ 0.0003
Table 3. Reward function parameters (delay requirement: 5 ms).
Table 3. Reward function parameters (delay requirement: 5 ms).
Parameter Set Number α β γ δ
10.70.30.21.20
20.40.60.40.75
30.90.50.81.50
40.50.50.51.20
Table 4. Reward function parameters (different delay threshold requirements).
Table 4. Reward function parameters (different delay threshold requirements).
Parameter α β γ δ
5 ms0.90.50.81.50
10 ms0.80.20.51.10
15 ms0.60.40.50.90
20 ms0.80.20.40.75
Table 5. Performance metrics and computational overhead of AIFG under different network scales (delay requirement: 20 ms).
Table 5. Performance metrics and computational overhead of AIFG under different network scales (delay requirement: 20 ms).
Network Size (T-R Pairs)PPO Training Time (s)PPO Training Memory Consumption (MB)Convergence IterationsAverage Delay (ms)Multiplexing Efficiency (%)Throughput (Kbps)
25–25 Nodes98.52260.64458.6492386.3845389.417
50–50 Nodes41.46261.614010.2825098.6160554.862
100–100 Nodes365.48261.096852.4370095.4074643.059
200–200 Nodes797.71259.978070.7490095.4598643.448
Table 6. Performance comparison under different MPDU sizes (delay requirement: 20 ms).
Table 6. Performance comparison under different MPDU sizes (delay requirement: 20 ms).
MPDU Size (Bytes)Average Delay (ms)Multiplexing Efficiency (%)Throughput (Kbps)
8007.2345796.110559.608
12009.2789698.992566.066
1500 (Default)10.7710099.494566.453
180012.0586097.593525.173
Table 7. Performance comparison under different satellite link bandwidth (delay requirement: 20 ms).
Table 7. Performance comparison under different satellite link bandwidth (delay requirement: 20 ms).
Relay Link Bandwidth (Between Satellites) (Mbps)Average Delay (ms)Multiplexing Efficiency (%)Throughput (Kbps)
311.013096.588566.859
6 (Default)11.337699.523575.183
911.025597.558569.503
1511.031996.951567.063
2011.276399.044574.157
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, J.; Chen, J.; Shi, H.; Feng, Y. Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Requirements in Advanced Orbiting Systems. Electronics 2025, 14, 1939. https://doi.org/10.3390/electronics14101939

AMA Style

Peng J, Chen J, Shi H, Feng Y. Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Requirements in Advanced Orbiting Systems. Electronics. 2025; 14(10):1939. https://doi.org/10.3390/electronics14101939

Chicago/Turabian Style

Peng, Jiahui, Jun Chen, Huaifeng Shi, and Yibing Feng. 2025. "Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Requirements in Advanced Orbiting Systems" Electronics 14, no. 10: 1939. https://doi.org/10.3390/electronics14101939

APA Style

Peng, J., Chen, J., Shi, H., & Feng, Y. (2025). Advanced Intelligent Frame Generation Algorithm for Differentiated QoS Requirements in Advanced Orbiting Systems. Electronics, 14(10), 1939. https://doi.org/10.3390/electronics14101939

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop