Next Article in Journal
CTAB-Assisted Formation of Hierarchical Porosity in Cu-BDC-NH2 Metal–Organic Frameworks and Its Enhanced Peroxidase-like Catalysis for Xanthine Sensing
Previous Article in Journal
Sustainable Production of Porous Activated Carbon from Hydrothermally Carbonized Jamoya Fruit Seeds and Its Potential for Adsorbing the Azo Dye Carmoisine B
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Solving Dynamic Multi-Objective Flexible Job Shop Scheduling Problems Using a Dual-Level Integrated Deep Q-Network Approach

School of Artificial Intelligence and Computer Science, Jiangnan University, Li Hu Avenue, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Processes 2025, 13(2), 386; https://doi.org/10.3390/pr13020386
Submission received: 29 November 2024 / Revised: 30 December 2024 / Accepted: 29 January 2025 / Published: 31 January 2025
(This article belongs to the Section Automation Control Systems)

Abstract

:
Economic performance in modern manufacturing enterprises is often influenced by random dynamic events, requiring real-time scheduling to manage multiple conflicting production objectives simultaneously. However, traditional scheduling methods often fall short due to their limited responsiveness in dynamic environments. To address this challenge, this paper proposes an innovative online rescheduling framework called the Dual-Level Integrated Deep Q-Network (DLIDQN). This framework is designed to solve the dynamic multi-objective flexible job shop scheduling problem (DMOFJSP), which is affected by six types of dynamic events: new job insertion, job operation modification, job deletion, machine addition, machine tool replacement, and machine breakdown. The optimization focuses on three key objectives: minimizing makespan, maximizing average machine utilization ( U a v e ), and minimizing average job tardiness rate ( T R a v e ). The DLIDQN framework leverages a hierarchical reinforcement learning approach and consists of two integrated IDQN-based agents. The high-level IDQN serves as the decision-maker during rescheduling, implementing dual-level decision-making by dynamically selecting optimization objectives based on the current system state and guiding the low-level IDQN’s actions. To meet diverse optimization requirements, two reward mechanisms are designed, focusing on job tardiness and machine utilization, respectively. The low-level IDQN acts as the executor, selecting the best scheduling rules to achieve the optimization goals determined by the high-level agent. To improve scheduling adaptability, nine composite scheduling rules are introduced, enabling the low-level IDQN to flexibly choose strategies for job sequencing and machine assignment, effectively addressing both sub-tasks to achieve optimal scheduling performance. Additionally, a local search algorithm is incorporated to further enhance efficiency by optimizing idle time between jobs. The numerical experimental results show that in 27 test scenarios, the DLIDQN framework consistently outperforms all proposed composite scheduling rules in terms of makespan, surpasses the widely used single scheduling rules in 26 instances, and always exceeds other reinforcement learning-based methods. Regarding the U a v e metric, the framework demonstrates superiority in 21 instances over all composite scheduling rules and maintains a consistent advantage over single scheduling rules and other RL-based strategies. For the T R a v e metric, DLIDQN outperforms composite and single scheduling rules in 20 instances and surpasses other RL-based methods in 25 instances. Specifically, compared to the baseline methods, our model achieves maximum performance improvements of approximately 37%, 34%, and 30% for the three objectives, respectively. These results validate the robustness and adaptability of the proposed framework in dynamic manufacturing environments and highlight its significant potential to enhance scheduling efficiency and economic benefits.

1. Introduction

The flexible job shop scheduling problem (FJSP) is a common challenge in production scheduling and extends the classical job shop scheduling problem (JSP). JSP is a combinatorial optimization problem (COP) that is NP-hard [1]. In JSP, each job consists of a series of operations, with each operation assigned to a specific machine. In contrast, FJSP introduces a machine selection mechanism, allowing each operation to be assigned to any machine from a set of compatible options, increasing scheduling flexibility and complexity.
Most research on FJSP focuses on static production environments. However, real-world production is often uncertain, with dynamic events such as order changes, machine breakdowns, and task adjustments being unavoidable [2]. These disruptions can cause deviations from the original plan, rendering existing schedules ineffective and negatively impacting production efficiency and resource utilization. As a result, developing dynamic scheduling methods for multi-objective FJSP is crucial. These methods not only enhance theoretical models and algorithms but also offer significant practical value.
In practical manufacturing environments, dynamic real-time scheduling methods help manage uncertainty by improving system flexibility and robustness. They allow quick adjustments to job sequences and resource allocations, minimizing the impact of disruptions like machine breakdowns or order changes. Real-time scheduling optimizes multiple objectives, such as minimizing makespan, maximizing machine utilization, and reducing job tardiness. This multi-objective optimization improves efficiency, balances resource use, and ensures smooth operation despite dynamic disturbances.
Moreover, real-time scheduling supports continuous decision-making in highly automated production environments. It allows systems to quickly respond to changing conditions, reducing the need for human intervention and decision delays. This enhances operational efficiency and stability. In smart manufacturing, real-time scheduling helps businesses quickly recover from disruptions, improving customer satisfaction while boosting competitiveness, efficiency, and resource utilization.
The dynamic multi-objective flexible job shop scheduling problem (DMOFJSP) has been extensively studied in recent years, with numerous solutions emerging from years of exploration. Traditional methods for solving DMOFJSP generally fall into two categories: meta-heuristic algorithms and dispatching rules [3]. Meta-heuristic methods encompass a variety of algorithms, such as genetic algorithms (GA) [4,5], particle swarm optimization (PSO) [6], simulated annealing (SA) [7], grey wolf optimization (GWO) [8], ant colony optimization (ACO) [9,10], and tabu search (TS) [11]. These methods typically decompose dynamic scheduling problems into a series of static sub-problems, which are then solved sequentially. Although these methods can generate solutions close to the global optimum, their high computational complexity necessitates recalculating an optimal solution whenever a dynamic event occurs. This often results in prolonged computation times, hindering quick responses to dynamic changes and limiting their suitability for real-time scheduling.
Dispatching rules are widely used in dynamic job shop scheduling, where they prioritize jobs and machines based on specific criteria to allocate and process jobs. Known for their computational efficiency and simplicity, these rules are highly responsive in dynamic environments. However, many dispatching rules have limitations, often being too myopic to yield optimal long-term results [12]. In different shop configurations, no single dispatching rule consistently outperforms others [13,14]. A promising approach to balance time efficiency and solution quality involves dynamically selecting the most appropriate rule at each decision point. This enables short-term optimization of multiple objectives, ensuring timely scheduling while improving overall shop efficiency.
To select the most appropriate rule at each scheduling decision point, the DMOFJSP can be modeled as a Markov Decision Process (MDP), providing a formal framework for decision-making under uncertainty. In this framework, an intelligent agent chooses the optimal action from the action space based on the current state features of the environment. This involves selecting from predefined scheduling rules to optimize the scheduling objectives. Recently, reinforcement learning (RL), a proven method for solving MDP problems [15], has made significant strides in dynamic real-time scheduling. For instance, Aydin and Oztemel [16] developed an enhanced Q-learning algorithm that trains an agent based on real-time shop floor conditions. The agent selects the most appropriate dispatching rule from three candidates, aiming to minimize mean tardiness in a dynamic job shop with new job insertions. Bouazza et al. [17] applied Q-learning to solve the dynamic flexible job shop scheduling problem with new job insertions. Two Q-matrices were used to store the probabilities for machine selection and scheduling rules, with the goal of minimizing the weighted average waiting time. More relevant research findings will be discussed in detail in Section 2.
Traditional reinforcement learning (RL) has made strides in dynamic scheduling, yet two critical challenges persist. Firstly, early studies primarily depended on Q-learning or SARSA algorithms [15,18,19], which employ Q-tables to record Q-values for state–action pairs. However, in real-world production scenarios, state features are usually continuous, leading to a rapid expansion—or even an infinite growth—of the state space. For intricate scheduling problems, the sheer volume of state–action combinations renders Q-table storage and computation unfeasible. Secondly, classical RL faces difficulties in managing continuous or high-dimensional state spaces. A common workaround is discretizing these spaces into finite intervals [20], but this approach sacrifices precision by overlooking minor state variations. Furthermore, identifying the optimal discretization strategy is complex, as it can significantly influence outcomes. Additionally, traditional RL is constrained in handling multi-objective optimization and adaptability, limiting its efficacy in complex production settings. These methods often struggle to reconcile conflicting objectives and lack the agility to respond to unforeseen disruptions, such as equipment failures or urgent task insertions.
In recent years, deep reinforcement learning (DRL), capable of directly handling continuous state spaces, has emerged as a promising alternative to traditional RL methods [21,22]. Among DRL methods, the Deep Q-Network (DQN) is one of the most recognized. It maps continuous states to the maximum Q-values for each action, enabling the selection of optimal actions at decision points, and has been widely applied in real-time dynamic scheduling. However, designing a single RL agent to optimize multiple objectives in multi-objective rescheduling problems remains highly challenging. Each objective corresponds to a distinct reward function and strategy, making it difficult for a single agent to balance them effectively. To address this issue, hierarchical reinforcement learning (HRL) provides a promising solution [23,24]. HRL leverages hierarchical temporal and spatial abstraction, where a high-level controller learns strategies for different objectives over extended time scales, while a low-level executor determines specific actions in real-time to achieve the high-level goals. By allowing the high-level controller to adaptively adjust optimization objectives (i.e., reward functions) and the low-level executor to select suitable scheduling rules, HRL facilitates balanced optimization across multiple objectives throughout the scheduling process.
Given the reasons outlined above, this study proposes a real-time dynamic scheduling framework based on hierarchical deep reinforcement learning (HDRL) to effectively address the complex scheduling challenges in real-world production environments. The framework integrates multiple optimization techniques and employs an integrated DQN structure, specifically designed to solve the dynamic multi-objective flexible job shop scheduling problem (DMOFJSP), which involves six types of dynamic disturbances: new job insertions, job cancellation, job operation modification, machine addition, machine tool replacement and machine breakdown. The model simultaneously optimizes three key objectives: makespan, average machine utilization, and average job tardiness rate, providing more accurate and efficient scheduling solutions.
Building upon the motivations discussed above, this paper introduces a model called the Dual-Level Integrated Deep Q-Network (DLIDQN) to solve the dynamic multi-objective flexible job shop scheduling problem (DMOFJSP). The key contributions of this study are as follows:
  • State Feature Extraction: Six state features are extracted and normalized to the range [0, 1] to accurately represent the system’s state at each rescheduling point. These features are specifically designed to account for six types of dynamic events: new job insertions, job cancellation, job operation modification, machine addition, machine tool replacement and machine breakdown. By normalizing state features, the model’s adaptability and convergence speed can be significantly improved.
  • Composite Dispatching Rule Design: Three job selection rules and three machine assignment rules are formulated, resulting in nine composite dispatching rules through pairwise combinations. These rules constitute the agent’s action space, addressing both job sequencing and machine allocation tasks. They exhibit strong generalization capabilities, making them suitable for scheduling problems across various production configurations.
  • Dual-Level Integrated Deep Q-Network (DLIDQN): This paper proposes a Dual-Level Integrated Deep Q-Network (DLIDQN) framework, which combines various customized network architectures and employs a dual-level structure to enhance the efficiency and accuracy of scheduling decisions. The framework consists of high-level and low-level IDQN agents. The high-level IDQN is responsible for strategy decision-making, analyzing state features to set optimization goals for the low-level agent (using different reward functions). The low-level IDQN operates at the execution layer, calculating the Q-values for each dispatching rule and selecting the optimal scheduling action based on the optimization goals set by the high-level agent. At each rescheduling point, the framework dynamically adjusts strategies through an efficient collaboration mechanism, maximizing cumulative rewards, thereby significantly improving decision-making efficiency and enhancing adaptability to dynamic environmental changes.
  • Multi-Objective Optimization and Local Search Algorithm: This paper optimizes three key indicators: makespan, average machine utilization, and average job tardiness rate. Two high-level optimization goals are defined to balance these objectives, each with its own reward function during training. To further improve scheduling performance and stability, a specialized local search algorithm is introduced to ensure the quality and stability of the scheduling solution.
Finally, extensive experiments conducted on the dataset developed in this study validate the effectiveness and advantages of the proposed model, highlighting its capability and practicality in addressing dynamic scheduling problems across various scenarios.
The remainder of this paper is organized as follows: Section 2 reviews dynamic scheduling methods based on classical RL and DRL. Section 3 introduces the theoretical foundations of RL, including Q-learning and deep Q-learning, and defines various architectures of IDQN. Section 4 establishes the mathematical model of DMOFJSP. Section 5 details the key elements of MDP, including the definition of states, actions, and rewards, describes the implementation of the local search algorithm, and provides the specific implementation details of DLIDQN. Section 6 presents and analyzes the experimental results. Finally, Section 7 summarizes this research and suggests directions for future work.

2. Literature Review

In research on the application of reinforcement learning (RL) to dynamic job shop scheduling problems, several effective models have been developed, though methodological challenges remain. Early studies predominantly employed traditional RL algorithms (such as Q-learning) to tackle these problems. For instance, Wei and Mingyang [25] proposed a set of composite scheduling rules encompassing both job and machine selection, specifically for dynamic job shop scheduling with random job arrivals. Using the Q-learning algorithm, they developed agents capable of intelligently identifying and selecting appropriate scheduling rules during each rescheduling process. Zhang et al. [26] modeled the scheduling problem as a semi-Markov Decision Process (semi-MDP) and applied Q-learning to learn the optimal choice among five heuristic rules. Chen et al. [27] introduced a rule-driven approach for multi-objective dynamic scheduling, where Q-learning was employed to learn optimal weights within a discrete value range to form composite scheduling rules. Shahrabi et al. [28] applied Q-learning to optimize the parameters of the variable neighborhood search (VNS) algorithm in dynamic job shop environments, addressing challenges such as random job arrivals and machine failures. Shiue et al. [29] proposed a reinforcement learning (RL)-based real-time scheduling mechanism, where agents were trained via Q-learning to select the most appropriate scheduling rule from a predefined set. Targeting the minimization of earliness and tardiness penalties in a job shop with new job insertions, Wang [30] developed a dynamic multi-agent scheduling model and employed a weighted Q-learning algorithm to train the agents. Both studies predominantly relied on Q-learning, utilizing Q-tables to determine the optimal scheduling rule for each state. However, in real-world production environments, the state space is often continuous or infinite, rendering the construction and maintenance of large Q-tables impractical. A typical workaround is to discretize the continuous state space, though this approach often compromises accuracy to some degree.
Deep reinforcement learning (DRL), which combines the strengths of deep learning and reinforcement learning, has become an effective approach for addressing complex dynamic job shop scheduling problems. Among DRL-based models, Deep Q-Network (DQN) methods have been widely applied in this area. Waschneck et al. [31] proposed a dynamic scheduling strategy using multiple DQN agents, each focused on optimizing the scheduling rules for a specific work center while also considering interactions with other agents to improve machine utilization. Altenmüller et al. [32] developed a DQN-based framework to address job scheduling problems with strict deadlines. Luo [33] addressed dynamic scheduling problems involving new job insertions, aiming to minimize delays. His method utilized DQN to train intelligent agents capable of selecting the optimal rule from six predefined composite scheduling rules at each decision point, using seven continuous state features. Luo et al. [34] proposed a two-level Deep Q-Network (THDQN) algorithm to optimize both total weighted tardiness and average machine utilization. According to Luo et al. [35], they developed a hierarchical deep reinforcement learning framework to address the multi-objective flexible job shop scheduling problem with partial no-wait constraints, training it using a multi-agent PPO algorithm. Li et al. [36] introduced a hybrid Deep Q-Network (HDQN) algorithm that effectively solves multi-objective problems in dynamic flexible job shop scheduling with insufficient transportation resources. In the flexible job shop scheduling problem with random job arrivals, Zhao et al. [37] enhanced the DRL method by incorporating an attention mechanism to optimize the total tardiness objective. Wu et al. [38] proposed a dual-layer DDQN method that utilizes decision points to select scheduling rules, aiming to reduce both maximum completion time and total tardiness in dynamic flexible job shop scheduling with random workpiece arrivals. Table 1 summarizes and compares the results of these Q-learning and DQN-based studies with those of the current work.

3. Preliminaries

3.1. Definition of RL and Q-Learning

In reinforcement learning (RL), common problems are modeled as a Markov Decision Process (MDP), represented by the quintuple ( S , A , P , γ , R ) . In this model, S denotes the set of states s, A the set of actions a, P represents the transition probabilities between states, R is the reward received when taking action a in state s, and  γ is the discount factor that balances the importance of future rewards. Within the MDP framework, an agent interacts with its environment based on a policy π , aiming to maximize the total expected return over time. At each time step t, the agent observes the current state s t S , selects an action a t A according to the policy π ( S A ) , transitions to a new state s t + 1 based on the state transition probability p ( s t + 1 | s t , a t ) P ( S × A S ) , and receives an immediate reward r t R .
The core objective of reinforcement learning is to find the optimal policy π * that maximizes the expected value of future discounted rewards when selecting action a in state s and acting according to the policy π , as demonstrated in Equation (1).
Q π * ( s , a ) = max π Q π ( s , a ) = max π E R t + γ R t + 1 + γ 2 R t + 2 + s t = s , a t = a , π
where the discount factor γ ( 0 , 1 ] decays over time, balancing the importance of short-term and long-term rewards.
Q π ( s , a ) , also known as the Q-function or action-value function, was proven by Bellman [39] to satisfy the Bellman optimality equation under the optimal policy π * , as shown in Equation (2). This equation forms the theoretical foundation of the Q-learning algorithm.
Q π * ( s , a ) = s p ( s s , a ) r ( s , a , s ) + γ max a Q π * ( s , a )

3.2. Definition of Deep Q-Learning

Q-learning requires continuous updates and maintenance of a Q-table that stores the Q-values for each state–action pair when addressing decision-making problems. Q-learning performs well when the state space is small. However, as the state space grows, the size of the Q-table increases, consuming significant resources and causing a sharp decline in decision-making efficiency. To overcome the curse of dimensionality in standard Q-learning, Mnih et al. [40] introduced the concept of a Deep Q-Network (DQN). DQN uses the state features at time step t (typically continuous values) as input to a deep neural network, which outputs the Q-values corresponding to each action in state s t . The size of the output layer corresponds to the number of actions. This enables DQN to efficiently handle complex decision-making processes in continuous state spaces.
To overcome the limitations of standard Q-learning, deep Q-learning (DQL) introduces two key improvements. First, DQL incorporates an experience replay mechanism, which stores the agent’s experiences at each time step t ( s t , a t , r t , s t + 1 ) in an experience replay pool. When updating the neural network, a random batch of past experiences is sampled from this buffer for training. This method reduces the correlation between consecutive data, enhancing the stability of the training process. Second, DQL introduces an independent target network Q ^ ( s , a ; θ ) . During training, the weights θ of the target network remain fixed and are only updated by copying the evaluation network’s weights θ every C steps. The loss between the target network’s output y t and the evaluation network’s output y is used to update the evaluation network’s parameters, with the calculation of y provided in Equation (3).
y = max Q ( s t , a ; θ )
The target value y t , as given in Equation (4), is calculated where max Q ^ ( s t + 1 , a ; θ ) represents the maximum Q value the target network can obtain in state s t + 1 .
y t = r t + γ max Q ^ ( s t + 1 , a ; θ )

3.3. Definition of IDQN

The Integrated Deep Q-Network (IDQN) algorithm combines the strengths of DQN, double DQN (DDQN), noisy DQN (NDQN), prioritized experience replay (PER), and dueling DQN (D3QN) frameworks. It models states using neural networks, mitigates Q-value overestimation through DDQN, improves exploration diversity by introducing noise, accelerates training via prioritized experience replay, and enhances Q-value estimation accuracy with the dueling structure. Crucially, these optimization techniques are complementary, effectively enhancing overall performance while mitigating individual weaknesses.
The standard DQN tends to overestimate Q-values because it uses the same Q-values for both action selection and evaluation [41]. To address this issue, Van Hasselt et al. [42] proposed the double DQN (DDQN) algorithm. This algorithm employs two DQN networks: the online Q network is responsible for selecting the action with the highest Q-value, while the target network Q ^ is used to estimate the state–action value, which is then used to compute the target value y t , as defined in Equation (5). By decoupling action selection from evaluation, DDQN effectively mitigates the overestimation problem, leading to a more stable learning process.
y t = r t + γ Q ^ ( s t + 1 , arg max a Q ( s t + 1 , a ; θ ) ; θ )
Noisy DQN (NDQN) enhances exploration diversity by introducing noise into the parameters of the Q-network. At the beginning of each episode, Gaussian noise is added to each parameter of the Q-network Q ( s , a ; w ) , resulting in Q ˜ ( s , a , ξ ; μ , σ ) . Here, μ and σ represent the mean and standard deviation, respectively, which are parameters to be learned, while ξ denotes random noise, with each element independently sampled from a standard normal distribution N ( 0 , 1 ) . The next action a t is selected using Q ˜ ( s , a , ξ ; μ , σ ) , as shown in Equation (6).
a t = arg max a A Q ˜ ( s , a , ξ ; μ , σ )
In classic DQN, experience replay samples randomly from the experience pool. However, the TD errors of different samples may vary. The TD error in the Q-network is defined as the difference between the target value provided by the target network and the estimated value produced by the online network. Priority Experience Replay (PER) DQN assigns a priority to each sample based on its TD error. The loss function, which incorporates the sample priorities, is defined in Equation (7), where ω i represents the weight.
L i ( θ i ) = E ω i y t Q ( s , a ; θ i ) 2
In the DQN algorithm, Q ( s , a ) is influenced by both the state and the action, though the impact of these two factors is not equal. Dueling DQN decomposes the Q-network into two separate components: one calculates the action advantage function A ( s , a ) , while the other computes the state value function V ( s ) . These components are then combined to form the state–action value function. The ’advantage’ refers to how one action compares to others within the current state. If the advantage is greater than zero, it indicates that the action outperforms the average; if it is less than zero, it suggests that the action under performs the average. This approach assigns higher values to actions with greater advantages, which facilitates faster convergence. The  Q ( s , a ) function is then expressed as shown in Equation (8).
Q ( s , a ) = V ( s ) + A ( s , a )

4. DMOFJSP Formulation

The dynamic multi-objective flexible job shop scheduling problem (DMOFJSP) addressed in this paper, which involves six disturbance events, is defined as follows: There are n jobs, J = J 1 , J 2 , J 3 , , J n , to be processed on m machines, M = M 1 , M 2 , M 3 , , M m . Each job J i consists of n i operations, with the j-th operation of job J i denoted as O i , j . Each operation O i , j corresponds to a set of compatible machines, M i , j ( M i , j M ), from which any machine M k can be selected for processing the operation. The processing time of operation O i , j on machine M k is denoted as t i , j , k , and the actual completion time is represented by C i , j . The relevant notations are presented in Table 2.
This study chooses makespan, average machine utilization rate ( U a v e ), and average job tardiness rate ( T R a v e ) as optimization goals because they are critical in production scheduling. These goals also account for the impact of dynamic events and the capabilities of the DLIDQN model. Makespan is important for measuring production efficiency, as disruptions like new job insertion and machine breakdown can extend production time. Minimizing makespan helps DLIDQN reduce delays caused by these events and improves system responsiveness. U a v e shows how well resources are allocated. Events like machine addition or replacement can cause fluctuations, but DLIDQN dynamically adjusts resource use to avoid both under-utilization and overload. T R a v e affects delivery times and is influenced by events like job modification or deletion. By focusing on these goals, DLIDQN effectively manages dynamic environments, improving the robustness and practical value of the scheduling system.
To simplify the problem, the following predefined constraints must be satisfied.
  • Each machine is capable of processing only one operation at a given time.
  • All operations within each job are required to adhere to a predetermined priority sequence and must be processed without interruption.
  • Job transportation times and machine setup time are disregarded in this study.
  • Each job is assigned a specific delivery deadline. Jobs not completed by their deadlines are classified as late, with their delays duly recorded.
  • The arrival times of jobs may differ. When a new job is introduced, all ongoing operations are allowed to finish before the remaining operations and the newly added job proceed to the scheduling stage to formulate a revised production plan.
  • If a task must be canceled, the system ceases the current process and shifts all remaining tasks to the scheduling stage in order to devise a fresh production plan.
  • Job operation modification involves expanding the existing job by introducing new operations. Once the current operation is completed, the additional operations will be integrated into the job, and all remaining tasks will be transferred to the rescheduling phase to develop a new production plan.
  • The process of machine addition refers to the system’s ability to deploy additional machines. After the completion of current tasks by all machines, the system will assess the average machine utilization U a v e . If  U a v e is higher than a given threshold, new machines will be introduced and integrated into the new scheduling plan.
  • Machines are capable of switching between different processing types by replacing tools. During the tool replacement process, machines enter a conversion state and temporarily halt operations. The processing time for a machine accounts for the additional time required for tool changes.
  • When a machine experiences a breakdown, it transitions into a repair state. If alternative machines are unavailable, the affected job enters a blocked state. Upon completion of the machine repair, the job exits the blocked state, and all jobs re-enter the scheduling phase with an updated production plan.
  • If no available machine is able to process the current operation, the job enters a blocked state. Once the final operation of the job is completed, the job transitions into a finished state, and no further processing is carried out.
When the operation has not been assigned to a specific machine, its processing time cannot be represented by t i , j , k . Instead, the average processing time of operation O i , j across all machines in its compatible machine set is used, denoted as t i , j ¯ , as shown in Equation (9).
t i , j ¯ = k M i , j t i , j , k | M i , j |
The estimated completion time E T i for the remaining operations of job J i is the theoretically calculated time required to complete all unfinished operations, which is the sum of the average processing times of all remaining operations. The details are shown in Equation (10).
E T i = j = O P i + 1 n i t i , j ¯
The due date D i of job J i is defined as the job’s arrival time A i plus the product of the job’s due date tightness ( D D T ) and the estimated completion time of all remaining operations. The formula is given in Equation (11).
D i = A i + ( D D T · U r i ) · E T i = A i + ( 0.2 + 0.5 · U r i ) · E T i
At the current decision point, O P T i indicates the total processing time of the completed operations for job J i , while O P i refers to the number of operations that have been completed for job J i up to this point. This formula can be found in Equation (12).
O P T i = A i + j = 1 O P i k M i , j X i , j , k t i , j , k
The completion rate C R J i of job J i is used to describe the job’s completion status, defined as the ratio of the processing time of completed operations to the estimated total processing time. The formula is provided in Equation (13).
C R J i = O P T i O P T i + E T i
C T k denotes the completion time of the most recently scheduled operation on machine M k . It represents the time at which the latest operation assigned to machine M k finishes. The corresponding formula is given in Equation (14).
C T k = max C i , j X i , j , k = 1
The machine utilization U k represents the operational efficiency of machine M k , indicating the proportion of machine M k ’s working time in the total running time. Its mathematical definition is illustrated in Equation (15).
U k = i = 1 n j = 1 O P i X i , j , k · t i , j , k C T k
The processing tardiness rate T R i of job J i represents the difference between the estimated completion time of all operations and the job’s deadline D i , divided by the total theoretical processing time of the job. It is defined in Equation (16).
T R i = O P T i + E T i D i O P T i + E T i
This study optimizes three metrics: minimizing the maximum job processing time (makespan) C m a x , minimizing the inverse of the average utilization of all machines 1 / U a v e , and minimizing the average job processing tardiness rate T R a v e . The mathematical formulation of the objective functions is as follows:
Minimize C m a x = max 1 i n C i , n i 1 / U a v e = k = 1 m U k m T R a v e = i = 1 n T R i n

5. Methods for DMOFJSP

This section begins by defining the key attributes of the MDP, including state features, candidate composite scheduling rules (i.e., actions), and the reward function configuration. Next, it details the DLIDQN network architecture, which incorporates two deep neural networks for hierarchical decision-making and integrates a real-time scheduling framework to address dynamic events. Finally, a local search algorithm leveraging machine idle time is introduced to refine the scheduling results further.

5.1. Definition of State Features

Reinforcement learning requires the agent to select the next action based on the current environmental state, highlighting the importance of accurately representing the environment through state features. In intricate real-world scenarios, this study identifies six normalized ratios, each within the range [0, 1], as inputs to the network. By scaling features to the same range, the model becomes more robust to dynamic events like machine breakdowns and job insertions. It is less affected by changes in scheduling scale or event frequency, which improves adaptability. This approach creates a uniform data distribution and avoids gradient instability caused by varying feature scales. It also ensures fair treatment of all features, preventing bias toward specific objectives. Additionally, reducing input variation helps the model explore the policy space more efficiently, especially in complex scenarios with dynamic disturbances. Detailed definitions of these features are provided below.

5.1.1. Detailed Definitions and Formula Explanations

(1) The average utilization rate of machines, denoted as U a v e , is defined in Equation (18).
U a v e = k = 1 m U k m
(2) The standard deviation of machine utilization rate U s t d is outlined in Equation (19).
U s t d = k = 1 m ( U k U a v e ) 2 m
(3) The average completion rate of jobs C R J a v e is given by Equation (20).
C R J a v e = i = 1 n C R J i n
(4) The standard deviation of job completion rate C R J s t d is provided in Equation (21).
C R J s t d = i = 1 n ( C R J i C R J a v e ) 2 n
(5) The average processing tardiness rate of jobs T R a v e is expressed in Equation (22).
T R a v e = j = 1 n T R i n
(6) The standard deviation of job processing tardiness rate T R s t d is shown in Equation (23).
T R s t d = i = 1 n ( T R i T R a v e ) 2 n

5.1.2. Impact of State Feature Changes on Scheduling Decisions

To illustrate more intuitively how changes in the above state features impact scheduling decisions, we provide three examples. These examples focus on the variations in average machine utilization rate ( U a v e ), average job completion rate ( C R J a v e ), and average job processing tardiness rate ( T R a v e ) and explore how these changes drive the dynamic adjustment of scheduling strategies.
  • Changes in U a v e : If the U a v e increases from 50% to 80%, the scheduling system may adjust accordingly. For example, the system may determine that machine resources are being well-utilized and decide to increase the workload by scheduling more jobs. Additionally, the system may assign new jobs to machines with lower utilization rates to balance the workload and avoid overloading specific machines.
  • Changes in C R J a v e : When the C R J a v e increases, it means most jobs are nearly finished. In this case, the scheduling system may shift focus to newly inserted or unstarted jobs, prioritizing them over those that are almost complete. The system may also delay the start of some new jobs to prevent congestion, which helps improve resource allocation and overall efficiency.
  • Changes in T R a v e : If the T R a v e increases, it indicates potential delays, possibly due to machine breakdowns or resource shortages. The scheduling system may respond by prioritizing jobs with higher tardiness rates, ensuring that critical tasks are completed on time. It may also reassign resources, such as moving delayed jobs to idle machines or prioritizing repairs on affected machines, to reduce overall delays and improve performance.

5.2. Definition of Proposed Composite Dispatching Rules

Two critical sub-tasks in scheduling are operation sequencing and machine allocation. The agent addresses these sub-tasks based on state features. Single dispatching rules, due to their myopic nature (Nie et al., 2013), often fail to perform well across diverse configurations, necessitating the use of composite dispatching rules for action selection. Accordingly, this study designs three dispatching rules for sub-task 1 and three additional rules for sub-task 2. By pairing these two types of rules, a total of nine composite dispatching rules are formed. Detailed descriptions of each dispatching rule are provided below.

5.2.1. Scheduling Rule 1 for Sub-Task 1

First, let U C j o b represent the set of unfinished jobs. Sort the jobs J i in this set based on their completion rate C R J i , prioritizing jobs with the lowest completion rate for execution. Simultaneously, the calculated job completion rate C R J i is weighted by an urgency coefficient. This rule preferentially selects jobs with relatively low completion rates but high urgency, effectively reducing the average job processing tardiness rate T R a v e . The specific procedure is detailed in Algorithm 1.
Algorithm 1: Procedure for Dispatching Rule 1 in Sub-task 1
Processes 13 00386 i001

5.2.2. Scheduling Rule 2 for Sub-Task 1

For rule 2 of sub-task 1, begin by defining the set of unfinished jobs, U C j o b . From this set, extract the delayed jobs J i , where the delivery deadline D i has been exceeded, to form a set of tardy jobs, T a r d j o b . If delayed jobs exist, prioritize those with the longest overdue time and highest urgency. If no delayed jobs are present, select jobs from the U C j o b set which exhibit a smaller difference between their overdue time and estimated completion time, combined with high urgency. The detailed procedure is outlined in Algorithm 2.
Algorithm 2: Procedure for Dispatching Rule 2 in Sub-task 1
Processes 13 00386 i002

5.2.3. Scheduling Rule 3 for Sub-Task 1

The third dispatching rule for sub-task 1 involves randomly selecting a job J i and processing its next operation. The procedure is expressed in Algorithm 3.
Algorithm 3: Procedure for Dispatching Rule 3 in Sub-task 1
    Input: Current time T c u r
             Set of unfinished jobs U C j o b
1
Randomly select a job i from U C j o b ;
2
Set the next operation of job i as the current processing operation;
3
Update the operation number for job i:
O P i = O P i + 1
    Output: Selected job i and updated operation number O P i

5.2.4. Scheduling Rule 1 for Sub-Task 2

After selecting the operation to be processed, a set of available machines from its compatible machine set M i , j , which are neither broken down nor undergoing tool changes, can be identified. The goal of the second sub-task is to allocate the operation O i , j to a suitable machine M k . Rule 1 selects the machine that can provide service the earliest for operation O i , j , thereby enhancing the average machine utilization rate U a v e . The detailed procedure is illustrated in Algorithm 4.
Algorithm 4: Procedure for Dispatching Rule 1 in Sub-task 2
Processes 13 00386 i003

5.2.5. Scheduling Rule 2 for Sub-Task 2

The second scheduling rule of sub-task 2 builds on the first rule by combining the processing time t i , j , k of operation O i , j on machine M k with the earliest available time. This rule takes both the availability time and processing capability into account. The process is outlined in Algorithm 5.
Algorithm 5: Procedure for Dispatching Rule 2 in Sub-task 2
Processes 13 00386 i004

5.2.6. Scheduling Rule 3 for Sub-Task 2

To enhance the robustness of the scheduling process and avoid local optima, a random selection rule is recommended as the third rule for sub-task 2. The step-by-step procedure can be found in Algorithm 6.
Algorithm 6: Procedure for Dispatching Rule 3 in Sub-task 2
Input: Current time T c u r
           Operation O i , j
           Set of compatible machines M i , j
Output: Selected machine M k
By combining the rules of sub-task 1 and sub-task 2, nine composite scheduling rules are generated. This design enhances the flexibility and diversity of the scheduling strategies. The job sequencing and machine allocation rules address multiple scheduling needs, such as job delays, machine availability, and urgency, while also handling the impact of six dynamic events on the system. These combined rules form various scheduling strategies, allowing the system to adapt to complex production scenarios and improving its overall adaptability.
Although the rule set contains only nine composite rules, they effectively cover the main production scheduling needs. First, rules 1 and 2 prioritize jobs with higher delays and urgency, ensuring timely scheduling and proper prioritization. They address the conflict between job completion rate and deadlines, especially when dynamic events, such as job insertion or modification occur, enabling quick strategy adjustments. Second, rules 1 and 2 consider machine availability and job processing time to select the most suitable machines. This improves machine utilization ( U a v e ) and reduces production bottlenecks, especially when dynamic events like machine breakdowns or tool changes affect the system, ensuring optimal resource allocation.

5.3. Definition of Reward Function

In the deep Q-learning algorithm, when different actions a t are executed in state s t , the agent transitions to a new state s t + 1 , and this process generates a corresponding reward r t . The magnitude of the reward r t directly influences the agent’s strategy selection, determining how it will choose the next action in subsequent states. To achieve multi-objective optimization, this study uses a Dual-Level Integrated Deep Q-Network structure and switches between different reward mechanisms. However, using makespan as the optimization goal leads to a sparse reward problem in reinforcement learning. In DMOFJSP, makespan is only determined once all operations for each job are completed. This means that the agent cannot obtain immediate feedback on makespan during the early stages of scheduling. The agent only receives a final reward based on the maximum completion time when all jobs are finished. Before that, there is little reward signal related to makespan optimization, making it difficult to learn effective strategies. To address this, the study uses job tardiness rate ( T R a v e ) and machine utilization rate ( U a v e ) as the main reward mechanisms. These two objectives provide timely feedback after each decision, helping the agent make better choices faster and improving the scheduling strategy, which indirectly helps reduce makespan.
Specifically, the high-level IDQN first receives the state features as input, and based on this input, it outputs different goal values g 1 , 2 . When the high-level IDQN selects a goal value g, the low-level IDQN chooses distinct reward algorithms depending on the selected goal. If  g = 1 , the reward algorithm is based on the average machine utilization rate U a v e , as defined in Equations (13) and (16). If  g = 2 , the reward algorithm focuses on the average job processing tardiness rate T R a v e , defined in Equations (14) and (20). The detailed execution process of these two distinct reward algorithms is outlined in the pseudocode in Algorithm 7.
Algorithm 7: Reward Calculation Based on g = 1 or g = 2
Processes 13 00386 i005
This multi-objective optimization approach enables the agent to flexibly choose different reward mechanisms, allowing it to balance machine utilization and job tardiness rates in complex dispatching tasks, ultimately optimizing overall scheduling performance.

5.4. Local Search Algorithm

This study uses a unique local search algorithm to determine the processing time interval for operation O i , j on machine M k . The algorithm focuses on selecting idle time intervals for local optimization, which is essential for solving the DMOFJSP. This problem involves dynamic events, such as job insertions and machine breakdowns, making the scheduling process complex and uncertain. Idle time intervals, which represent machine downtime, provide a flexible optimization space. By scheduling jobs within these intervals, machine utilization is maximized, resource wastage is avoided, and scheduling conflicts caused by unexpected events are reduced. The local optimization targets the best idle time intervals, ensuring proper job sequencing and resource allocation. This reduces computational load and speeds up the optimization process. Overall, this method enhances the system’s adaptability and stability in dynamic environments, effectively addressing varying scheduling needs while optimizing objectives like minimizing makespan, maximizing U a v e , and minimizing T R a v e .
The search process is detailed in Algorithm 8. We create an idle time interval list I T I L k for machine M k , consisting of I k idle time intervals, denoted as I T I 1 , I T I 2 , , I T I I k . Next, we search through the list I T I L k for idle time intervals that satisfy the following conditions: we search for intervals where the end time E I T I is later than the end time E O i , j 1 of operation O i , j 1 , and the duration is long enough to accommodate the processing time of O i , j . Once a suitable idle time interval is found, we schedule operation O i , j within that interval. During scheduling, we ensure that the start time S I T I of the idle time interval is no earlier than the availability time A T k of machine M k , and the end time E O i , j 1 of the previous operation O i , j 1 is later than the start time S T i of the job. These time constraints are crucial for ensuring the feasibility and effectiveness of the scheduling plan.
Algorithm 8: Local Search Process
Processes 13 00386 i006

5.5. The Network of the DLIDQN

The proposed DLIDQN framework integrates multiple network structures, consisting of two IDQNs. Each IDQN is primarily composed of a dueling DQN model and includes noisy layers to enhance exploration capabilities. The higher-level IDQN model consists of three hidden layers, each with ten nodes. It takes a 6-dimensional state feature as input and outputs a Q-value that combines the state value and advantage function, where the state value branch outputs one node, and the advantage function branch outputs two nodes, corresponding to the two high-level optimization goals (i.e., the two forms of reward functions). The lower-level IDQN model is more complex, containing seven hidden layers, each with 50 nodes. With a 7-dimensional state feature (six state features and one high-level goal) as input, it calculates the final Q-value by combining the state value and the advantage function. The advantage function branch produces nine nodes, each corresponding to one of the nine composite scheduling rules. Both models use the ReLU activation function and are optimized with the mean square error (MSE) loss function and the Adam optimizer. At each rescheduling event, scheduling rules are executed to generate a new state, and a reward algorithm is selected based on the goal values to obtain corresponding rewards. The state, action, reward, new state, reward function targets, and completion indicator are stored for prioritized experience replay to update the network parameters. Figure 1 illustrates the structure of the proposed DLIDQN and the overall algorithm process.

6. Numerical Experiments

A detailed account of the numerical experiments used to assess the effectiveness and generalization of the DLIDQN framework is provided in this section, along with an analysis of the results. We first introduce the parameter settings of the dataset and algorithms used during the training and testing process. After completing the training, we evaluate the DLIDQN model against each composite dispatching rule introduced in this study, as well as commonly adopted classic dispatching rules. To further highlight the strengths of the proposed model, we compare DLIDQN with three other deep reinforcement learning algorithms that utilize the same action space of nine composite dispatching rules. Through experiments and visualizations, the high performance and strong generalization capability of the DLIDQN model are confirmed.

6.1. Training Process

This section explains the model training process, which is carried out over 200 runs using randomly generated data. To ensure the model can adapt to different workshop configurations, the initial number of jobs, machines, operations per job, and new job insertions are varied randomly during each training session. New jobs arrive according to a Poisson distribution, with the time between consecutive job insertions following an exponential distribution exp ( 1 / λ ) . Other environmental parameters are also randomly updated within a specified range. The instance parameters used during training are listed in Table 3. The number of machines available for each operation is randomly generated, with a maximum of ( M n u m 2 ) to ensure at least two machines are always available.
Figure 2 illustrates the training logic flow of the DLIDQN model. During the training process, the dual IDQN models collaborate to optimize performance. First, the high-level model receives the state feature S t and generates the target value g, then combines S t and g to create a new state S t , which is passed to the low-level model. The low-level model selects action A t based on S t , applies the scheduling rules, and uses a local search algorithm to improve the results. Next, the new state S t + 1 is calculated, and the reward R t + 1 is obtained. If all jobs are scheduled, the final results are output, marking the end of the training. Otherwise, the tuple ( S t , A t , g , R t + 1 , S t + 1 , d o n e ) is saved in the replay buffer. After the defined step count is reached, the prioritized experience replay mechanism is used to update the network parameters, continuing the process until all jobs are scheduled.
Table 4 lists the parameter settings for the training algorithm. Most of these parameters are based on pre-experimental results, while some use common default values from deep reinforcement learning algorithms. Specifically, the buffer capacity is set to 2000. When the capacity is exceeded, old data will be overwritten to make space for new data. The learning rate ( η ) is set to 0.001, balancing convergence speed and model stability. Pre-experimental results show that a high learning rate can cause instability, while a low learning rate slows convergence, making it hard to achieve good results in a limited number of iterations. The target network update frequency (C) is set to 200 steps. Regular updates help stabilize the training process and reduce fluctuations. The experiments also found that too low of an update frequency causes significant fluctuations, while too high of a frequency can slow down convergence.
A test instance containing 16 machines, with  E a v e set to 70 and 40 new job insertions, was prepared to validate the DLIDQN model obtained after each training round. During testing, the curve of makespan for each training round and the average makespan over every 20 training rounds are shown in Figure 3. The graph shows the trend of average makespan achieved by DLIDQN during each training epoch. The green “Makespan” curve represents the makespan for each epoch, with some fluctuations. The red “Average Makespan Obtained Every 20 Training Epochs” line demonstrates a steady decline as training progresses.
In the first 100 epochs, the makespan decreases quickly, indicating that the model is learning the key features of the task. Between 100 and 200 epochs, the decline slows, suggesting that the model is entering a fine-tuning phase. After 200 epochs, the makespan stabilizes, and further training does not significantly improve performance.
Comparing the performance across different epochs shows that there is still room for improvement after 100 epochs. However, after 200 epochs, the makespan reaches its lowest point. The small performance gap between 150 and 200 epochs indicates that 200 epochs are enough to cover the task’s complexity and achieve effective learning. Moreover, the use of diverse training instances ensures the model’s ability to generalize and perform well in different production environments. Therefore, 200 training epochs are sufficient for the model to reach optimal training and provide strong scheduling performance.

6.2. Testing Settings

After completing model training, the evaluation phase begins, with all test samples generated randomly. Table 5 shows the specific parameters for each test sample. The initial number of jobs is set to 20, 30, or 40, with the number of operations per job assigned arbitrarily and processing times for each operation determined randomly. The number of machines is initialized to 8, 12, or 16 units, and the number of dynamic events is selected randomly from 30, 50, or 70, with event types also allocated at random. Other parameters related to jobs and machines follow the same setup as in the training stage. A total of 27 distinct test configurations were established, each tested 20 times independently. These test configurations cover different task scales, machine numbers, and dynamic event distributions, validating the robustness of the DLIDQN model in complex dynamic environments. The results show that despite significant randomness and uncertainty in the input parameters, the model consistently produces high-quality scheduling results. This demonstrates that the DLIDQN model can adapt to changes in task scale, processing time fluctuations, and disruptions from dynamic events. This robustness highlights the model’s wide applicability and reliability in real-world production environments.
The testing was implemented in Python, running in a Python 3.8 environment on a computer with an Intel i7-9750H processor, 24 GB of RAM, and a GTX 1650Ti graphics card. In this setup, we compared the performance of our DLIDQN algorithm against the baseline algorithms using the same set of test cases.

6.3. Comparisons with the Proposed Composite Dispatching Rules

To assess the performance of DLIDQN, this study conducted 20 independent experiments on each of 27 test instances with various parameter configurations, comparing DLIDQN with the proposed nine composite scheduling rules. During the experiments, we calculated the mean and standard deviation of makespan, U a v e , and  T R a v e across 20 runs for each instance and summarized the results in Table 6, Table 7 and Table 8, where the best performance values are highlighted in bold. To more intuitively present the data in the tables, we created corresponding line charts (Figure 4, Figure 5 and Figure 6) to compare the performance of single composite rules and DLIDQN across different scenarios. These line charts clearly illustrate DLIDQN’s advantages in key metrics such as average makespan, average utilization, and average turnaround time, further demonstrating its superior performance under various parameter configurations.
The results show that DLIDQN, with its action space covering nine scheduling rules, can flexibly select the best rules for composite scheduling. This allows it to outperform most single scheduling rules in various production environments. In comparisons with the random strategy (Rule 9), DLIDQN performs better in nearly all test cases. However, in some scenarios, its performance is slightly lower than certain single rules. For example, in the U a v e metric, DLIDQN falls short of Rule 1 in six instances. Rule 1 prioritizes urgent jobs by calculating their completion rate C R J i and urgency coefficient. It also selects machines that can start jobs earlier, minimizing idle time. Similarly, DLIDQN’s T R a v e is slightly worse than Rules 4 and 7 in some cases, as these rules also prioritize urgent jobs. Single scheduling rules are simple and clear, requiring fewer calculations. This makes them faster and more efficient in certain test cases, avoiding the complexity and computational costs of DLIDQN while achieving better performance in specific scenarios.
DLIDQN may require more computation and strategy exploration during learning, which can lead to suboptimal performance in some cases. However, no single scheduling rule can consistently achieve optimal results across various production environments. This demonstrates DLIDQN’s strong adaptability and excellent performance when handling untrained production configurations. Figure 7 shows the three-dimensional Pareto front for DLIDQN and composite scheduling rules in three test cases, further proving the superiority of the DLIDQN model.

6.4. Comparisons with Classic Dispatching Rules

We compared DLIDQN with five widely recognized classic dispatching rules to more comprehensively validate its superiority, followed by a detailed explanation of each rule.
  • First In First Out (FIFO): Decide the next operation to start based on the arrival time of jobs, selecting the next operation for the job that arrived first.
  • Earliest Due Date (EDD): Sort jobs by their due dates and choose the next operation for the job with the earliest due date to ensure timely delivery.
  • Shortest Processing Time (SPT): Schedule based on the estimated shortest time required to complete each job, selecting the next operation for the job with the shortest processing time.
  • Longest Processing Time (LPT): Opposite to SPT, choose the job with the longest processing time, typically used for optimizing resource allocation to expedite the processing of tasks that have been waiting longer.
  • Most Remaining Processing Time (MRT): Consider jobs with the most remaining processing time (those with the longest remaining time) to balance the overall schedule.
It should be noted that the aforementioned classic scheduling strategies lack explicit machine assignments when handling job allocations, making them difficult to adapt to the flexible job shop scheduling problem (FJSP) addressed in this paper. To overcome this shortcoming, we established a supplementary machine assignment principle: for each job, we select the earliest available machine from its compatible set for job allocation, allowing for a fair performance comparison between these classic strategies and our method.
We independently conducted 20 experiments for each test instance, and the results are summarized in Table 9, Table 10 and Table 11. These tables provide the averages and standard deviations for three key performance indicators: makespan, average machine utilization ( U a v e ), and average tardiness rate ( T R a v e ). The optimal values are highlighted in bold. The line charts corresponding to the tables are shown in Figure 8, Figure 9 and Figure 10. After analyzing the data, we found that our algorithm performs better than traditional scheduling strategies in most test cases across all three performance metrics. Although in some scenarios, the DLIDQN framework performs slightly worse than EDD in terms of T R a v e , this is because EDD prioritizes jobs with the earliest due dates, effectively reducing tardiness risk. In cases where due dates are closely distributed or many jobs are urgent, EDD’s focus on due dates helps lower the average tardiness rate. In comparison, the framework uses dynamic rule selection to optimize multiple objectives, so it does not always prioritize the most urgent jobs. However, it achieves the best performance in most instances. In certain specific cases, the difference in T R a v e between the framework and the optimal value is small. Therefore, the experimental results show that the proposed composite scheduling rules are effective in reducing tardiness and improving machine utilization.
Furthermore, as shown in Figure 11, the Pareto optimal front clearly highlights DLIDQN’s advantage over other well-known dispatching rules in a range of representative instances.

6.5. Comparisons with Other RL-Based Scheduling Algorithms

In order to demonstrate the superiority of our method for multi-objective optimization tasks, we carried out a comprehensive comparison, evaluating DLIDQN against three mainstream reinforcement learning-based training algorithms. These three algorithms are DQN (the classic Deep Q-Network), dueling double DQN (D3QN, an improved architecture of double DQN), and HDMDDQN, a novel two-layer deep double Q-network training method proposed by Wang et al. (2022) [43].
DQN, as the control group, features a simple design, relying on a single RL agent for rule selection without higher-level scheduling objectives. Details of its reward mechanism are provided in Algorithm 9. For fairness, all methods share the same action space as DLIDQN, consisting of nine composite scheduling rules. We compared the means and standard deviations of makespan, U a v e , and  T R a v e across various production environments. The results are presented in Table 12, Table 13 and Table 14. The optimal values are highlighted in bold. The line charts corresponding to the tables are shown in Figure 12, Figure 13 and Figure 14.
Algorithm 9: DQN Reward Algorithm
Processes 13 00386 i007
The comparison results show that DLIDQN demonstrates a significant advantage over other RL training methods in the vast majority of test instances. Through performance comparisons with the single-layer DQN, we convincingly prove that the proposed two-layer structure is not only indispensable but also exhibits high effectiveness in practical applications. The high-level DQN component of DLIDQN intelligently selects the optimal optimization objective at different rescheduling points, thereby achieving an excellent balance among makespan, U a v e , and T R a v e . Additionally, compared to the other two advanced algorithms with a dual-layer structure, our proposed DLIDQN model, which integrates multiple network structures and a prioritized experience replay mechanism, also displays significant performance advantages, further highlighting the exceptional capabilities of DLIDQN. In Figure 15, the performance of DLIDQN and various RL-based methods is evaluated across multiple representative instances for three objectives. It is clear that DLIDQN outperforms the others in all three objectives. Moreover, a view of Figure 16 indicates that our method remains at the forefront of the Pareto optimal front. The results from the Pareto chart and other graphs show that DLIDQN performs well in reducing completion time, improving resource utilization, and lowering delay rates. This proves its advantage in single-objective optimization and also demonstrates its strong performance in multi-objective optimization, highlighting its efficiency and broad applicability in practical scenarios.

7. Conclusions

This paper presents the Dual-Level Integrated Deep Q-Network (DLIDQN) algorithm, which provides an effective solution to the dynamic multi-objective flexible job shop scheduling problem (DMOFJSP), involving six dynamic events: new job insertions, job cancellation, job operation modification, machine addition, machine tool replacement and machine breakdown. The algorithm implements goal-oriented scheduling through dual IDQN agents. At rescheduling moments, the high-level IDQN dynamically generates an optimization target for the low-level IDQN based on the current state features. The low-level IDQN, guided by this goal and considering the state features, executes the most suitable scheduling rules to ensure the achievement of the specified goals. In addition, DLIDQN integrates the benefits of diverse network architectures, resulting in powerful optimization capabilities. By designing two reward algorithms and six essential state features, as well as developing nine composite scheduling rules, the algorithm demonstrates outstanding performance in optimizing makespan, average machine utilization ( U a v e ), and average job processing tardiness rate ( T R a v e ).
To comprehensively assess the performance and generalization ability of DLIDQN, a series of extensive numerical experiments were conducted across various shop configurations. The results indicate that DLIDQN not only outperforms all the proposed composite scheduling rules but also surpasses existing classical scheduling rules and other reinforcement learning-based scheduling methods. Notably, DLIDQN performs exceptionally well in untrained production environments, demonstrating its remarkable algorithmic performance and adaptability.
Future research will focus on exploring the application of advanced policy-based algorithms, such as actor–critic (AC) and Proximal Policy Optimization (PPO), to dynamic multi-objective flexible job shop scheduling problems (DMOFJSP). Additionally, we plan to investigate multi-agent reinforcement learning (MARL) to analyze how multiple agents can achieve their goals through interactive learning in a shared environment. By integrating these insights with policy-based training methods, we aim to further enhance the performance of scheduling algorithms in complex dynamic environments.

Author Contributions

Conceptualization, H.X., J.Z., L.H. and J.T.; methodology, H.X., J.Z. and C.Z.; software, H.X., J.Z. and J.T.; validation, H.X., J.Z. and L.H.; formal analysis, J.Z. and C.Z.; writing—original draft preparation, H.X. and J.Z.; writing—review and editing, L.H., J.T. and C.Z.; visualization, H.X. and J.Z.; supervision, L.H., J.T. and C.Z.; project administration, H.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Garey, M.R.; Johnson, D.S.; Sethi, R. The Complexity of Flowshop and Jobshop Scheduling. Math. Oper. Res. 1976, 1, 117–129. [Google Scholar] [CrossRef]
  2. Ouelhadj, D.; Petrovic, S. A survey of dynamic scheduling in manufacturing systems. J. Sched. 2009, 12, 417–431. [Google Scholar] [CrossRef]
  3. Lou, P.; Liu, Q.; Zhou, Z.; Wang, H.; Sun, S.X. Multi-agent-based proactive–reactive scheduling for a job shop. Int. J. Adv. Manuf. Technol. 2012, 59, 311–324. [Google Scholar] [CrossRef]
  4. Pezzella, F.; Morganti, G.; Ciaschetti, G. A genetic algorithm for the Flexible Job-shop Scheduling Problem. Comput. Oper. Res. 2008, 35, 3202–3212. [Google Scholar] [CrossRef]
  5. Kundakcı, N.; Kulak, O. Hybrid genetic algorithms for minimizing makespan in dynamic job shop scheduling problem. Comput. Ind. Eng. 2016, 96, 31–51. [Google Scholar] [CrossRef]
  6. Ning, T.; Huang, M.; Liang, X.; Jin, H. A novel dynamic scheduling strategy for solving flexible job-shop problems. J. Ambient Intell. Humaniz. Comput. 2016, 7, 721–729. [Google Scholar] [CrossRef]
  7. Cruz-Chávez, M.A.; Martínez-Rangel, M.G.; Cruz-Rosales, M.H. Accelerated simulated annealing algorithm applied to the flexible job shop scheduling problem. Int. Trans. Oper. Res. 2017, 24, 1119–1137. [Google Scholar] [CrossRef]
  8. Zhu, Z.; Zhou, X. An efficient evolutionary grey wolf optimizer for multi-objective flexible job shop scheduling problem with hierarchical job precedence constraints. Comput. Ind. Eng. 2020, 140, 106280. [Google Scholar] [CrossRef]
  9. El Khoukhi, F.; Boukachour, J.; Alaoui, A.E.H. The “Dual-Ants Colony”: A novel hybrid approach for the flexible job shop scheduling problem with preventive maintenance. Comput. Ind. Eng. 2017, 106, 236–255. [Google Scholar] [CrossRef]
  10. Zhang, S.; Li, X.; Zhang, B.; Wang, S. Multi-objective optimisation in flexible assembly job shop scheduling using a distributed ant colony system. Eur. J. Oper. Res. 2020, 283, 441–460. [Google Scholar] [CrossRef]
  11. Poppenborg, J.; Knust, S.; Hertzberg, J. Online scheduling of flexible job-shops with blocking and transportation. Eur. Ind. Eng. 2012, 6, 497–518. [Google Scholar] [CrossRef]
  12. Mohan, J.; Lanka, K.; Rao, A.N. A Review of Dynamic Job Shop Scheduling Techniques. Procedia Manuf. 2019, 30, 34–39. [Google Scholar] [CrossRef]
  13. Baker, K.R. Sequencing Rules and Due-Date Assignments in a Job Shop. Manag. Sci. 1984, 30, 1093–1104. [Google Scholar] [CrossRef]
  14. Nie, L.; Gao, L.; Li, P.; Li, X. A GEP-based reactive scheduling policies constructing approach for dynamic flexible job shop scheduling problem with job release dates. J. Intell. Manuf. 2013, 24, 763–774. [Google Scholar] [CrossRef]
  15. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  16. Aydin, M.; Öztemel, E. Dynamic job-shop scheduling using reinforcement learning agents. Robot. Auton. Syst. 2000, 33, 169–178. [Google Scholar] [CrossRef]
  17. Bouazza, W.; Sallez, Y.; Beldjilali, B. A distributed approach solving partially flexible job-shop scheduling problem with a Q-learning effect. IFAC-PapersOnLine 2017, 50, 15890–15895. [Google Scholar] [CrossRef]
  18. Wei, Y.; Zhao, M. Composite rules selection using reinforcement learning for dynamic job-shop scheduling. In Proceedings of the IEEE Conference on Robotics, Automation and Mechatronics, Singapore, 1–3 December 2004; Volume 2, pp. 1083–1088. [Google Scholar]
  19. Wang, Y.C.; Usher, J.M. Learning policies for single machine job dispatching. Robot. Comput.-Integr. Manuf. 2004, 20, 553–562. [Google Scholar] [CrossRef]
  20. Chen, R.; Yang, B.; Li, S.; Wang, S. A self-learning genetic algorithm based on reinforcement learning for flexible job-shop scheduling problem. Comput. Ind. Eng. 2020, 149, 106778. [Google Scholar] [CrossRef]
  21. Li, Y. Deep Reinforcement Learning: An Overview. arXiv 2017, arXiv:1701.07274. [Google Scholar]
  22. Wang, L.; Pan, Z.; Wang, J. A Review of Reinforcement Learning Based Intelligent Optimization for Manufacturing Scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  23. Rafati, J.; Noelle, D.C. Learning representations in model-free hierarchical reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 10009–10010. [Google Scholar]
  24. Li, A.C.; Florensa, C.; Clavera, I.; Abbeel, P. Sub-policy adaptation for hierarchical reinforcement learning. arXiv 2019, arXiv:1906.05862. [Google Scholar]
  25. Wei, Y.; Zhao, M. A reinforcement learning-based approach to dynamic job-shop scheduling. Acta Autom. Sin. 2005, 31, 765. [Google Scholar]
  26. Zhang, Z.; Zheng, L.; Weng, M.X. Dynamic parallel machine scheduling with mean weighted tardiness objective by Q-Learning. Int. J. Adv. Manuf. Technol. 2007, 34, 968–980. [Google Scholar] [CrossRef]
  27. Chen, X.; Hao, X.; Lin, H.W.; Murata, T. Rule driven multi objective dynamic scheduling by data envelopment analysis and reinforcement learning. In Proceedings of the 2010 IEEE International Conference on Automation and Logistics, Hong Kong, China, 16–20 August 2010; pp. 396–401. [Google Scholar] [CrossRef]
  28. Shahrabi, J.; Adibi, M.A.; Mahootchi, M. A reinforcement learning approach to parameter estimation in dynamic job shop scheduling. Comput. Ind. Eng. 2017, 110, 75–82. [Google Scholar] [CrossRef]
  29. Shiue, Y.R.; Lee, K.C.; Su, C.T. Real-time scheduling for a smart factory using a reinforcement learning approach. Comput. Ind. Eng. 2018, 125, 604–614. [Google Scholar] [CrossRef]
  30. Wang, Y.F. Adaptive job shop scheduling strategy based on weighted Q-learning algorithm. J. Intell. Manuf. 2020, 31, 417–432. [Google Scholar] [CrossRef]
  31. Waschneck, B.; Reichstaller, A.; Belzner, L.; Altenmüller, T.; Bauernhansl, T.; Knapp, A.; Kyek, A. Optimization of global production scheduling with deep reinforcement learning. Procedia CIRP 2018, 72, 1264–1269. [Google Scholar] [CrossRef]
  32. Altenmüller, T.; Stüker, T.; Waschneck, B.; Kuhnle, A.; Lanza, G. Reinforcement learning for an intelligent and autonomous production control of complex job-shops under time constraints. Prod. Eng. 2020, 14, 319–328. [Google Scholar] [CrossRef]
  33. Luo, S. Dynamic scheduling for flexible job shop with new job insertions by deep reinforcement learning. Appl. Soft Comput. 2020, 91, 106208. [Google Scholar] [CrossRef]
  34. Luo, S.; Zhang, L.; Fan, Y. Dynamic multi-objective scheduling for flexible job shop by deep reinforcement learning. Comput. Ind. Eng. 2021, 159, 107489. [Google Scholar] [CrossRef]
  35. Luo, S.; Zhang, L.; Fan, Y. Real-time scheduling for dynamic partial-no-wait multiobjective flexible job shop by deep reinforcement learning. IEEE Trans. Autom. Sci. Eng. 2022, 19, 3020–3038. [Google Scholar] [CrossRef]
  36. Li, Y.; Gu, W.; Yuan, M.; Tang, Y. Real-time data-driven dynamic scheduling for flexible job shop with insufficient transportation resources using hybrid deep Q network. Robot. Comput.-Integr. Manuf. 2022, 74, 102283. [Google Scholar] [CrossRef]
  37. Zhao, L.; Fan, J.; Zhang, C.; Shen, W.; Zhuang, J. A DRL-Based Reactive Scheduling Policy for Flexible Job Shops with Random Job Arrivals. IEEE Trans. Autom. Sci. Eng. 2023, 21, 2912–2923. [Google Scholar] [CrossRef]
  38. Wu, Z.; Fan, H.; Sun, Y.; Peng, M. Efficient Multi-Objective Optimization on Dynamic Flexible Job Shop Scheduling Using Deep Reinforcement Learning Approach. Processes 2023, 11, 2018. [Google Scholar] [CrossRef]
  39. Bellman, R. A Markovian decision process. J. Math. Mech. 1957, 6, 679–684. [Google Scholar] [CrossRef]
  40. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing Atari with Deep Reinforcement Learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  41. Hasselt, H. Double Q-learning. In Advances in Neural Information Processing Systems 23, Proceedings of the 24th Annual Conference on Neural Information Processing Systems 2010, Vancouver, BC, Canada, 6–9 December 2010; Lafferty, J., Williams, C., Shawe-Taylor, J., Zemel, R., Culotta, A., Eds.; Neural Information Processing Systems Foundation, Inc. (NeurIPS): San Diego, CA, USA, 2010. [Google Scholar]
  42. Van Hasselt, H.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar] [CrossRef]
  43. Wang, H.; Cheng, J.; Liu, C.; Zhang, Y.; Hu, S.; Chen, L. Multi-objective reinforcement learning framework for dynamic flexible job shop scheduling problem with uncertain events. Appl. Soft Comput. 2022, 131, 109717. [Google Scholar] [CrossRef]
Figure 1. The structure of DLIDQN and the overall algorithm process.
Figure 1. The structure of DLIDQN and the overall algorithm process.
Processes 13 00386 g001
Figure 2. Dynamic multi-objective scheduling workflow of DLIDQN.
Figure 2. Dynamic multi-objective scheduling workflow of DLIDQN.
Processes 13 00386 g002
Figure 3. Average makespan achieved by DLIDQN at each training epoch. The green “Makespan” curve represents the makespan recorded for each epoch. The red “Average makespan obtained every 20 training epochs” line indicates the average makespan over every 20 epochs.
Figure 3. Average makespan achieved by DLIDQN at each training epoch. The green “Makespan” curve represents the makespan recorded for each epoch. The red “Average makespan obtained every 20 training epochs” line indicates the average makespan over every 20 epochs.
Processes 13 00386 g003
Figure 4. Average makespan comparison across scenarios with single composite rules vs. DLIDQN.
Figure 4. Average makespan comparison across scenarios with single composite rules vs. DLIDQN.
Processes 13 00386 g004
Figure 5. Average machine utilization rate comparison across scenarios with single composite rules vs. DLIDQN.
Figure 5. Average machine utilization rate comparison across scenarios with single composite rules vs. DLIDQN.
Processes 13 00386 g005
Figure 6. Average job tardiness rate comparison across scenarios with single composite rules vs. DLIDQN.
Figure 6. Average job tardiness rate comparison across scenarios with single composite rules vs. DLIDQN.
Processes 13 00386 g006
Figure 7. Pareto fronts obtained by DLIDQN and the proposed composite dispatching rules after 20 runs on three representative instances.
Figure 7. Pareto fronts obtained by DLIDQN and the proposed composite dispatching rules after 20 runs on three representative instances.
Processes 13 00386 g007
Figure 8. Average makespan comparison across scenarios with classic dispatching rules vs. DLIDQN.
Figure 8. Average makespan comparison across scenarios with classic dispatching rules vs. DLIDQN.
Processes 13 00386 g008
Figure 9. Average machine utilization rate comparison with classic dispatching rules vs. DLIDQN.
Figure 9. Average machine utilization rate comparison with classic dispatching rules vs. DLIDQN.
Processes 13 00386 g009
Figure 10. Average job tardiness rate comparison across scenarios with classic dispatching rules vs. DLIDQN.
Figure 10. Average job tardiness rate comparison across scenarios with classic dispatching rules vs. DLIDQN.
Processes 13 00386 g010
Figure 11. Pareto fronts obtained by DLIDQN and well-known classic dispatching rules after 20 runs on three representative instances.
Figure 11. Pareto fronts obtained by DLIDQN and well-known classic dispatching rules after 20 runs on three representative instances.
Processes 13 00386 g011
Figure 12. Average makespan comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Figure 12. Average makespan comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Processes 13 00386 g012
Figure 13. Average machine utilization rate comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Figure 13. Average machine utilization rate comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Processes 13 00386 g013
Figure 14. Average job tardiness rate comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Figure 14. Average job tardiness rate comparison across scenarios with other RL-based algorithms vs. DLIDQN.
Processes 13 00386 g014
Figure 15. The average performance of DLIDQN and other RL-based algorithms across three objectives was calculated for three representative instances.
Figure 15. The average performance of DLIDQN and other RL-based algorithms across three objectives was calculated for three representative instances.
Processes 13 00386 g015
Figure 16. Pareto fronts obtained by DLIDQN and other RL-based algorithms after 20 runs on three representative instances.
Figure 16. Pareto fronts obtained by DLIDQN and other RL-based algorithms after 20 runs on three representative instances.
Processes 13 00386 g016
Table 1. Existing reinforcement learning-based research on the dynamic job shop scheduling problem.
Table 1. Existing reinforcement learning-based research on the dynamic job shop scheduling problem.
WorkAlgorithmProblemObjectiveDynamic Events
Aydin and Oztemel (2000) [16]Q-learningDynamic job schedulingMean tardinessNew job insertions
Bouazza et al. (2017) [17]Q-learningDynamic multi-objective flexible job schedulingMakespan;
Total weighted completion time;
Weighted average waiting time
New job insertions
Yingzi and Mingyang (2004) [18]Q-learningDynamic job schedulingMean tardinessNew job insertions
Wang (2020) [30]Q-learningDynamic job schedulingEarliness punishment;
Tardiness punishment
New job insertions
Zhang et al. (2007) [26]Q-learningDynamic parallel-machine schedulingMean weighted tardinessSequence-dependent setup times;
Machine job qualification
Chen et al. (2010) [27]Q-learningDynamic multi-objective job schedulingMean tardiness;
Mean flow time
Fluctuation of work in process
Shahrabi et al. (2017) [28]Q-learningDynamic job schedulingMean flow timeNew job insertions;
Machine breakdowns
Waschneck et al. (2018) [31]Deep Q-learningDynamic flexible job schedulingUptime utilizationMachine breakdowns
Altenmüller et al. (2020) [32]Deep Q-learningDynamic job schedulingCount of violation eventsNew job insertions;
Machine breakdowns
Luo (2020) [33]DQNDynamic multi-objective flexible job schedulingTotal tardinessNew job insertions
Luo et al. (2021) [34]THDQNDynamic flexible job schedulingTotal weighted tardiness rate;
Average machine utilization
New job insertions
Luo et al. (2021b) [35]HMAPPODynamic partial-no-wait multi-objective flexible job shop scheduling (DMOFJSP-PNW)Total weighted tardiness;
Average machine utilization rate;
Variance of machine workload
New job insertions;
Machine breakdowns
Li et al. (2022) [36]HDQNDynamic multi-objective flexible job schedulingCompletion time;
Total energy
Insufficient transportation resources
Wu et al. (2023) [38]Dual layer DDQNDynamic multi-objective flexible job schedulingMaximum completion time;
Total tardiness
Random workpiece arrivals
Table 2. List of relevant notations.
Table 2. List of relevant notations.
ParameterDefinition
nThe total number of jobs
mThe total number of machines
J i The i-th job
O i , j The j-th operation of job J i
n i The number of operations for job J i
M k The k-th machine
M i , j The set of compatible machines for operation O i , j
t i , j , k The processing time of operation O i , j on machine M k
A i The arrival time of job J i
U r i The urgency of job J i is classified into three levels: level 1 being the most urgent, and level 3 being the least. Jobs with higher urgency are assigned shorter due dates
D D T Due date tightness (DDT) is an indicator that measures the relationship between a job’s due date and its estimated completion time. A lower DDT value indicates a tighter due date, leaving less time to complete the task, whereas a higher DDT value suggests that the job has more flexibility in completing the task
D i The due date of job J i
O P i The number of operations completed for the job J i at the current time
X i , j , k Determine whether the operation O i , j is being processed on machine M k . If  O i , j is assigned to machine M k , then X i , j , k = 1 ; otherwise, X i , j , k = 0
Y i , j , h , g Determine the relative priority between operations O i , j and O h , g . If  O i , j is a predecessor of O h , g , then Y i , j , h , g = 1 ; if O i , j is a successor of O h , g , then Y i , j , h , g = 1
C i , j The completion time of operation O i , j
U k The utilization of machine M k
T r i The processing tardiness rate of job J i
Table 3. Parameter settings for various production configurations in the training process.
Table 3. Parameter settings for various production configurations in the training process.
ParameterValue
The total number of initial jobsUnif [1, 20]
The total number of machinesUnif [8, 16]
Number of operations belonging to a jobUnif [1, 20]
Available machines per operationUnif [1, m−2]
Operation time on an available machineUnif [1, 50]
Urgency degree U r i of each job[1, 2, 3]
Due date tightness ( D D T )[0.7, 1.2, 1.7]
Mean value ( E a v e ) of the interarrival time exp ( 1 / λ ) [30, 50, 70]
Machine repair timeUnif [1, 99]
Tool replacement timeUnif [1, 50]
Table 4. Hyperparameter settings for the training algorithm.
Table 4. Hyperparameter settings for the training algorithm.
HyperparameterValue
Number of training epochs (L)20
Priority experience replay buffer size (N)2000
Minibatch size to perform gradient descent (batch_size)32
Controls the influence of sample prioritization ( α )0.6
Adjusts the correction level of importance sampling ( β )0.4
Epsilon in ϵ -greedy action policy ( ϵ )Linear drop from 0.6 to 0.01
Target network update frequency (C)200
Discount rate ( γ )0.95
Learning rate ( η )0.001
OptimizerAdam
Table 5. Parameter settings for various production configurations in the testing process.
Table 5. Parameter settings for various production configurations in the testing process.
ParameterValue
The total number of initial jobsUnif [1, 20]
The total number of machines[8, 12, 16]
Number of operations belonging to a jobUnif [1, 20]
Number of available machines of each operationUnif [1, m−2]
Processing time of an operation on an available machineUnif [1, 50]
Urgency degree U r i of each job[1, 2, 3]
Due date tightness ( D D T )[0.7, 1.2, 1.7]
Mean value ( E a v e ) of the interarrival time exp ( 1 / λ ) [30, 50, 70]
The total number of new insert jobs[20, 30, 40]
Machine repair timeUnif [1, 99]
Tool replacement timeUnif [1, 50]
Table 6. Analysis of the averages and standard deviations of makespan across 20 runs for single composite rules vs. DLIDQN.
Table 6. Analysis of the averages and standard deviations of makespan across 20 runs for single composite rules vs. DLIDQN.
nmeOursrule1rule2rule3rule4rule5rule6rule7rule8rule9 (Random)
20830 1.635 × 10 3 / 7.285 × 10 1 1.760 × 10 3 / 8.863 × 10 1 1.956 × 10 3 / 1.091 × 10 2 3.170 × 10 3 / 2.472 × 10 2 1.947 × 10 3 / 9.476 × 10 1 2.258 × 10 3 / 1.192 × 10 2 4.365 × 10 3 / 3.926 × 10 2 3.300 × 10 3 / 2.255 × 10 2 3.528 × 10 3 / 1.585 × 10 2 1.074 × 10 4 / 5.153 × 10 2
50 1.867 × 10 3 / 4.618 × 10 1 1.989 × 10 3 / 5.903 × 10 1 2.204 × 10 3 / 1.403 × 10 2 3.283 × 10 3 / 2.414 × 10 2 2.015 × 10 3 / 7.731 × 10 1 2.409 × 10 3 / 1.035 × 10 2 4.428 × 10 3 / 3.635 × 10 2 3.184 × 10 3 / 1.990 × 10 2 4.090 × 10 3 / 2.830 × 10 2 1.099 × 10 4 / 5.016 × 10 2
70 1.937 × 10 3 / 5.959 × 10 1 2.033 × 10 3 / 6.116 × 10 1 2.295 × 10 3 / 1.025 × 10 2 3.644 × 10 3 / 3.290 × 10 2 2.215 × 10 3 / 1.224 × 10 2 2.563 × 10 3 / 1.619 × 10 2 4.803 × 10 3 / 3.720 × 10 2 4.026 × 10 3 / 2.684 × 10 2 4.210 × 10 3 / 2.481 × 10 2 1.282 × 10 4 / 6.665 × 10 2
1230 1.087 × 10 3 / 6.275 × 10 1 1.253 × 10 3 / 3.809 × 10 1 1.431 × 10 3 / 7.658 × 10 1 2.570 × 10 3 / 2.763 × 10 2 1.358 × 10 3 / 6.320 × 10 1 1.668 × 10 3 / 1.043 × 10 2 3.526 × 10 3 / 2.826 × 10 2 2.196 × 10 3 / 8.715 × 10 1 2.765 × 10 3 / 1.205 × 10 2 9.697 × 10 3 / 5.170 × 10 2
50 1.220 × 10 3 / 4.861 × 10 1 1.397 × 10 3 / 8.655 × 10 1 1.699 × 10 3 / 1.828 × 10 2 2.684 × 10 3 / 2.016 × 10 2 1.563 × 10 3 / 1.071 × 10 2 1.886 × 10 3 / 1.116 × 10 2 3.824 × 10 3 / 3.419 × 10 2 2.244 × 10 3 / 2.145 × 10 2 2.885 × 10 3 / 1.982 × 10 2 1.051 × 10 4 / 6.544 × 10 2
70 9.584 × 10 2 / 3.765 × 10 1 1.220 × 10 3 / 7.093 × 10 1 1.382 × 10 3 / 1.150 × 10 2 2.198 × 10 3 / 1.623 × 10 2 1.263 × 10 3 / 8.941 × 10 1 1.601 × 10 3 / 1.058 × 10 2 3.278 × 10 3 / 2.886 × 10 2 1.954 × 10 3 / 8.076 × 10 1 2.316 × 10 3 / 1.332 × 10 2 8.610 × 10 3 / 5.791 × 10 2
1630 7.884 × 10 2 / 5.579 × 10 1 1.046 × 10 3 / 7.275 × 10 1 1.191 × 10 3 / 7.147 × 10 1 2.031 × 10 3 / 2.084 × 10 2 9.994 × 10 2 / 6.505 × 10 1 1.311 × 10 3 / 9.132 × 10 1 2.951 × 10 3 / 3.290 × 10 2 1.430 × 10 3 / 7.051 × 10 1 1.902 × 10 3 / 1.376 × 10 2 8.592 × 10 3 / 6.833 × 10 2
50 7.910 × 10 2 / 2.519 × 10 1 1.072 × 10 3 / 6.334 × 10 1 1.309 × 10 3 / 1.137 × 10 2 2.211 × 10 3 / 2.079 × 10 2 1.045 × 10 3 / 8.777 × 10 1 1.394 × 10 3 / 5.468 × 10 1 3.098 × 10 3 / 2.996 × 10 2 3.206 × 10 3 / 2.254 × 10 2 2.112 × 10 3 / 1.028 × 10 2 8.832 × 10 3 / 7.015 × 10 2
70 7.568 × 10 2 / 4.819 × 10 1 1.064 × 10 3 / 7.385 × 10 1 1.241 × 10 3 / 1.171 × 10 2 1.889 × 10 3 / 2.262 × 10 2 1.004 × 10 3 / 6.775 × 10 1 1.386 × 10 3 / 1.198 × 10 2 2.762 × 10 3 / 2.020 × 10 2 1.428 × 10 3 / 1.060 × 10 2 1.723 × 10 3 / 8.687 × 10 1 8.073 × 10 3 / 3.702 × 10 2
30830 1.841 × 10 3 / 4.900 × 10 1 1.912 × 10 3 / 5.429 × 10 1 2.111 × 10 3 / 1.031 × 10 2 3.596 × 10 3 / 2.777 × 10 2 1.986 × 10 3 / 8.544 × 10 1 2.326 × 10 3 / 9.872 × 10 1 4.684 × 10 3 / 3.704 × 10 2 3.664 × 10 3 / 9.822 × 10 1 4.016 × 10 3 / 2.115 × 10 2 1.204 × 10 4 / 5.869 × 10 2
50 1.841 × 10 3 / 6.275 × 10 1 1.939 × 10 3 / 5.414 × 10 1 2.185 × 10 3 / 1.616 × 10 2 3.954 × 10 3 / 4.840 × 10 2 2.069 × 10 3 / 1.001 × 10 2 2.378 × 10 3 / 1.120 × 10 2 4.990 × 10 3 / 4.393 × 10 2 3.631 × 10 3 / 2.182 × 10 2 4.195 × 10 3 / 1.729 × 10 2 1.218 × 10 4 / 4.072 × 10 2
70 2.547 × 10 3 / 4.407 × 10 1 2.624 × 10 3 / 4.412 × 10 1 2.826 × 10 3 / 1.044 × 10 2 4.143 × 10 3 / 3.045 × 10 2 2.644 × 10 3 / 9.337 × 10 1 3.064 × 10 3 / 1.507 × 10 2 5.462 × 10 3 / 4.116 × 10 2 4.995 × 10 3 / 2.614 × 10 2 5.369 × 10 3 / 1.531 × 10 2 1.580 × 10 4 / 5.752 × 10 2
1230 1.785 × 10 3 / 6.856 × 10 1 1.941 × 10 3 / 3.315 × 10 1 2.285 × 10 3 / 1.889 × 10 2 3.569 × 10 3 / 2.648 × 10 2 2.127 × 10 3 / 1.225 × 10 1 2.535 × 10 3 / 1.649 × 10 2 4.919 × 10 3 / 4.702 × 10 2 3.877 × 10 3 / 2.003 × 10 2 5.016 × 10 3 / 1.578 × 10 2 1.551 × 10 4 / 6.152 × 10 2
50 1.491 × 10 3 / 5.342 × 10 1 1.651 × 10 3 / 4.640 × 10 1 1.897 × 10 3 / 1.000 × 10 2 3.003 × 10 3 / 2.336 × 10 2 1.742 × 10 3 / 8.557 × 10 1 2.102 × 10 3 / 1.125 × 10 2 4.318 × 10 3 / 2.883 × 10 2 3.403 × 10 3 / 2.269 × 10 2 3.954 × 10 3 / 2.443 × 10 2 1.401 × 10 4 / 5.852 × 10 2
70 1.366 × 10 3 / 4.562 × 10 1 1.539 × 10 3 / 3.842 × 10 1 1.766 × 10 3 / 1.054 × 10 2 3.119 × 10 3 / 2.519 × 10 2 1.570 × 10 3 / 7.571 × 10 1 1.986 × 10 3 / 1.050 × 10 2 4.137 × 10 3 / 3.324 × 10 2 2.753 × 10 3 / 1.771 × 10 2 3.059 × 10 3 / 1.326 × 10 2 1.320 × 10 4 / 6.762 × 10 2
1630 8.613 × 10 2 / 5.363 × 10 1 1.090 × 10 3 / 5.304 × 10 1 1.288 × 10 3 / 8.674 × 10 1 2.358 × 10 3 / 1.419 × 10 2 1.087 × 10 3 / 7.024 × 10 1 1.453 × 10 3 / 7.323 × 10 1 3.313 × 10 3 / 3.232 × 10 2 1.883 × 10 3 / 1.270 × 10 2 2.423 × 10 3 / 1.872 × 10 2 8.405 × 10 3 / 4.059 × 10 2
50 1.226 × 10 3 / 4.746 × 10 1 1.411 × 10 3 / 6.309 × 10 1 1.629 × 10 3 / 1.817 × 10 2 2.822 × 10 3 / 2.216 × 10 2 1.452 × 10 3 / 7.994 × 10 1 1.949 × 10 3 / 1.152 × 10 2 4.038 × 10 3 / 3.789 × 10 2 2.550 × 10 3 / 1.941 × 10 2 3.278 × 10 3 / 2.100 × 10 2 1.397 × 10 4 / 8.220 × 10 2
70 1.027 × 10 3 / 4.635 × 10 1 1.230 × 10 3 / 6.183 × 10 1 1.458 × 10 3 / 1.564 × 10 2 2.711 × 10 3 / 2.615 × 10 2 1.295 × 10 3 / 9.345 × 10 1 1.688 × 10 3 / 1.083 × 10 2 3.625 × 10 3 / 3.254 × 10 2 2.364 × 10 3 / 1.891 × 10 2 2.581 × 10 3 / 1.516 × 10 2 1.204 × 10 4 / 6.382 × 10 2
40830 2.068 × 10 3 / 6.852 × 10 1 2.153 × 10 3 / 5.187 × 10 1 2.396 × 10 3 / 1.644 × 10 2 4.473 × 10 3 / 4.886 × 10 2 2.219 × 10 3 / 8.323 × 10 1 2.618 × 10 3 / 9.296 × 10 1 5.361 × 10 3 / 3.166 × 10 2 4.267 × 10 3 / 2.126 × 10 2 6.022 × 10 3 / 2.332 × 10 2 1.420 × 10 4 / 8.249 × 10 2
50 3.585 × 10 3 / 4.697 × 10 1 3.630 × 10 3 / 5.702 × 10 1 4.011 × 10 3 / 2.587 × 10 2 6.367 × 10 3 / 5.365 × 10 2 3.760 × 10 3 / 1.398 × 10 2 4.133 × 10 3 / 1.037 × 10 2 8.084 × 10 3 / 4.942 × 10 2 7.740 × 10 3 / 3.701 × 10 2 8.743 × 10 3 / 3.495 × 10 2 2.401 × 10 4 / 7.333 × 10 2
70 3.578 × 10 3 / 5.479 × 10 1 3.651 × 10 3 / 6.753 × 10 1 3.859 × 10 3 / 1.471 × 10 2 5.369 × 10 3 / 2.846 × 10 2 3.624 × 10 3 / 1.232 × 10 2 4.108 × 10 3 / 1.300 × 10 2 7.084 × 10 3 / 4.547 × 10 2 6.700 × 10 3 / 2.265 × 10 2 7.399 × 10 3 / 2.355 × 10 2 2.181 × 10 4 / 7.874 × 10 2
1230 1.757 × 10 3 / 5.697 × 10 1 1.865 × 10 3 / 4.361 × 10 1 2.139 × 10 3 / 1.564 × 10 2 3.729 × 10 3 / 4.427 × 10 2 1.929 × 10 3 / 8.429 × 10 1 2.407 × 10 3 / 8.586 × 10 1 4.866 × 10 3 / 2.835 × 10 2 4.057 × 10 3 / 2.235 × 10 2 4.137 × 10 3 / 1.524 × 10 2 1.634 × 10 4 / 8.712 × 10 2
50 1.797 × 10 3 / 6.543 × 10 1 1.925 × 10 3 / 6.132 × 10 1 2.181 × 10 3 / 1.151 × 10 2 3.812 × 10 3 / 3.176 × 10 2 2.022 × 10 3 / 9.291 × 10 1 2.437 × 10 3 / 1.527 × 10 2 5.150 × 10 3 / 3.098 × 10 2 3.868 × 10 3 / 2.097 × 10 2 4.454 × 10 3 / 2.197 × 10 2 1.706 × 10 4 / 9.795 × 10 2
70 1.893 × 10 3 / 4.507 × 10 1 1.988 × 10 3 / 4.965 × 10 1 2.301 × 10 3 / 1.727 × 10 2 3.902 × 10 3 / 4.383 × 10 2 2.049 × 10 3 / 9.024 × 10 1 2.467 × 10 3 / 1.061 × 10 2 4.598 × 10 3 / 3.579 × 10 2 3.975 × 10 3 / 2.067 × 10 2 4.642 × 10 3 / 2.020 × 10 2 1.761 × 10 4 / 8.169 × 10 2
1630 1.410 × 10 3 / 4.413 × 10 1 1.610 × 10 3 / 5.994 × 10 1 1.888 × 10 3 / 1.368 × 10 2 3.466 × 10 3 / 3.161 × 10 2 1.639 × 10 3 / 1.052 × 10 2 2.011 × 10 3 / 8.084 × 10 1 4.702 × 10 3 / 3.209 × 10 2 3.171 × 10 3 / 1.935 × 10 2 3.863 × 10 3 / 2.113 × 10 2 1.664 × 10 4 / 9.417 × 10 2
50 1.477 × 10 3 / 3.263 × 10 1 1.643 × 10 3 / 3.908 × 10 1 1.891 × 10 3 / 9.949 × 10 1 3.622 × 10 3 / 2.864 × 10 2 1.643 × 10 3 / 5.604 × 10 1 2.141 × 10 3 / 1.384 × 10 2 4.856 × 10 3 / 3.314 × 10 2 2.682 × 10 3 / 1.192 × 10 2 3.792 × 10 3 / 2.014 × 10 2 1.857 × 10 4 / 7.537 × 10 2
70 1.357 × 10 3 / 6.114 × 10 1 1.581 × 10 3 / 3.710 × 10 1 1.842 × 10 3 / 2.554 × 10 2 3.147 × 10 3 / 2.013 × 10 2 1.644 × 10 3 / 8.888 × 10 1 2.095 × 10 3 / 1.163 × 10 2 4.598 × 10 3 / 3.579 × 10 2 3.599 × 10 3 / 1.606 × 10 2 3.528 × 10 3 / 1.585 × 10 2 1.676 × 10 4 / 9.295 × 10 2
Table 7. Analysis of the averages and standard deviations of U a v e across 20 runs for single composite rules vs. DLIDQN.
Table 7. Analysis of the averages and standard deviations of U a v e across 20 runs for single composite rules vs. DLIDQN.
nmeOursrule1rule2rule3rule4rule5rule6rule7rule8rule9
20830 9.819 × 10 1 / 7.412 × 10 3 9.786 × 10 1 / 8.309 × 10 3 9.126 × 10 1 / 4.266 × 10 2 6.194 × 10 1 / 3.469 × 10 2 8.157 × 10 1 / 2.368 × 10 2 8.014 × 10 1 / 2.618 × 10 2 4.408 × 10 1 / 3.713 × 10 2 4.303 × 10 1 / 2.308 × 10 2 4.716 × 10 1 / 1.629 × 10 2 1.700 × 10 1 / 7.108 × 10 3
50 9.884 × 10 1 / 5.444 × 10 3 9.856 × 10 1 / 4.914 × 10 3 9.026 × 10 1 / 6.047 × 10 2 6.401 × 10 1 / 3.804 × 10 2 8.391 × 10 1 / 2.246 × 10 2 8.176 × 10 1 / 2.058 × 10 2 4.714 × 10 1 / 3.564 × 10 2 4.876 × 10 1 / 2.612 × 10 2 4.621 × 10 1 / 2.803 × 10 2 1.772 × 10 1 / 6.924 × 10 3
70 9.910 × 10 1 / 4.426 × 10 3 9.889 × 10 1 / 4.923 × 10 3 9.117 × 10 1 / 5.952 × 10 2 6.360 × 10 1 / 5.165 × 10 2 8.411 × 10 1 / 2.252 × 10 2 8.160 × 10 1 / 2.760 × 10 2 4.676 × 10 1 / 3.738 × 10 2 3.988 × 10 1 / 2.379 × 10 2 4.761 × 10 1 / 2.182 × 10 2 1.683 × 10 1 / 6.905 × 10 3
1230 9.244 × 10 1 / 7.699 × 10 3 9.349 × 10 1 / 1.177 × 10 2 8.570 × 10 1 / 5.049 × 10 2 5.355 × 10 1 / 4.538 × 10 2 7.627 × 10 1 / 2.820 × 10 2 7.701 × 10 1 / 2.083 × 10 2 3.759 × 10 1 / 2.850 × 10 2 4.180 × 10 1 / 1.998 × 10 2 4.493 × 10 1 / 1.883 × 10 2 1.320 × 10 1 / 7.380 × 10 3
50 9.204 × 10 1 / 1.044 × 10 2 9.031 × 10 1 / 2.308 × 10 2 8.002 × 10 1 / 7.483 × 10 2 5.320 × 10 1 / 3.260 × 10 2 7.474 × 10 1 / 2.953 × 10 2 7.371 × 10 1 / 2.061 × 10 2 3.818 × 10 1 / 3.268 × 10 2 4.955 × 10 1 / 2.253 × 10 2 4.963 × 10 1 / 2.350 × 10 2 1.285 × 10 1 / 7.699 × 10 3
70 9.278 × 10 1 / 1.432 × 10 2 8.977 × 10 1 / 2.094 × 10 2 8.189 × 10 1 / 6.008 × 10 2 5.540 × 10 1 / 4.414 × 10 2 7.420 × 10 1 / 2.454 × 10 2 7.382 × 10 1 / 2.469 × 10 2 3.720 × 10 1 / 3.083 × 10 2 4.439 × 10 1 / 1.414 × 10 2 4.769 × 10 1 / 2.009 × 10 2 1.348 × 10 1 / 8.405 × 10 3
1630 8.353 × 10 1 / 1.646 × 10 2 8.424 × 10 1 / 1.960 × 10 2 7.469 × 10 1 / 4.256 × 10 2 4.671 × 10 1 / 3.732 × 10 2 6.820 × 10 1 / 2.214 × 10 2 7.263 × 10 1 / 1.677 × 10 2 3.269 × 10 1 / 3.158 × 10 2 4.659 × 10 1 / 2.116 × 10 2 5.027 × 10 1 / 2.083 × 10 2 1.055 × 10 1 / 8.867 × 10 3
50 8.436 × 10 1 / 1.386 × 10 2 8.635 × 10 1 / 1.786 × 10 2 7.527 × 10 1 / 4.805 × 10 2 4.765 × 10 1 / 3.658 × 10 2 7.282 × 10 1 / 2.635 × 10 2 7.160 × 10 1 / 1.896 × 10 2 3.395 × 10 1 / 2.480 × 10 2 3.778 × 10 1 / 2.153 × 10 2 4.818 × 10 1 / 2.864 × 10 2 1.110 × 10 1 / 8.002 × 10 3
70 8.654 × 10 1 / 1.417 × 10 2 8.540 × 10 1 / 1.844 × 10 2 7.686 × 10 1 / 5.307 × 10 2 5.165 × 10 1 / 5.378 × 10 2 7.100 × 10 1 / 2.767 × 10 2 7.311 × 10 1 / 1.813 × 10 2 3.536 × 10 1 / 2.611 × 10 2 4.712 × 10 1 / 1.760 × 10 2 5.431 × 10 1 / 1.854 × 10 2 1.154 × 10 1 / 6.833 × 10 3
30830 9.945 × 10 1 / 4.542 × 10 3 9.944 × 10 1 / 2.196 × 10 3 9.397 × 10 1 / 3.625 × 10 2 6.126 × 10 1 / 3.676 × 10 2 8.624 × 10 1 / 2.012 × 10 2 8.460 × 10 1 / 2.231 × 10 2 4.657 × 10 1 / 3.859 × 10 2 4.319 × 10 1 / 1.412 × 10 2 4.567 × 10 1 / 1.861 × 10 2 1.697 × 10 1 / 8.199 × 10 3
50 9.942 × 10 1 / 4.098 × 10 3 9.886 × 10 1 / 5.555 × 10 3 9.163 × 10 1 / 5.653 × 10 2 5.773 × 10 1 / 6.025 × 10 2 8.345 × 10 1 / 1.971 × 10 2 8.352 × 10 1 / 2.325 × 10 2 4.316 × 10 1 / 3.053 × 10 2 4.297 × 10 1 / 1.486 × 10 2 4.521 × 10 1 / 1.653 × 10 2 1.684 × 10 1 / 5.152 × 10 3
70 9.985 × 10 1 / 1.511 × 10 3 9.984 × 10 1 / 1.219 × 10 3 9.563 × 10 1 / 3.517 × 10 2 7.039 × 10 1 / 4.722 × 10 2 8.841 × 10 1 / 2.262 × 10 2 8.786 × 10 1 / 1.950 × 10 2 5.089 × 10 1 / 3.653 × 10 2 3.882 × 10 1 / 1.385 × 10 2 4.792 × 10 1 / 1.180 × 10 2 1.673 × 10 1 / 5.804 × 10 3
1230 9.744 × 10 1 / 8.648 × 10 3 9.724 × 10 1 / 5.246 × 10 3 8.892 × 10 1 / 7.667 × 10 2 5.838 × 10 1 / 3.681 × 10 2 8.022 × 10 1 / 2.833 × 10 2 7.809 × 10 1 / 2.531 × 10 2 4.218 × 10 1 / 3.261 × 10 2 4.124 × 10 1 / 2.179 × 10 2 4.215 × 10 1 / 1.135 × 10 2 1.251 × 10 1 / 4.692 × 10 3
50 9.853 × 10 1 / 5.352 × 10 3 9.753 × 10 1 / 6.076 × 10 3 9.044 × 10 1 / 6.014 × 10 2 6.064 × 10 1 / 4.969 × 10 2 8.177 × 10 1 / 2.110 × 10 2 8.009 × 10 1 / 1.651 × 10 2 4.182 × 10 1 / 2.874 × 10 2 3.680 × 10 1 / 2.715 × 10 2 4.075 × 10 1 / 2.342 × 10 2 1.233 × 10 1 / 5.013 × 10 3
70 9.689 × 10 1 / 7.467 × 10 3 9.658 × 10 1 / 7.288 × 10 3 9.015 × 10 1 / 4.312 × 10 2 5.416 × 10 1 / 3.870 × 10 2 8.102 × 10 1 / 2.006 × 10 2 8.034 × 10 1 / 1.722 × 10 2 4.026 × 10 1 / 3.295 × 10 2 4.342 × 10 1 / 2.216 × 10 2 4.929 × 10 1 / 1.353 × 10 2 1.197 × 10 1 / 5.064 × 10 3
1630 9.234 × 10 1 / 1.387 × 10 2 9.278 × 10 1 / 8.972 × 10 3 8.542 × 10 1 / 5.236 × 10 2 5.089 × 10 1 / 4.835 × 10 2 7.481 × 10 1 / 2.249 × 10 2 7.656 × 10 1 / 1.815 × 10 2 3.525 × 10 1 / 3.206 × 10 2 4.167 × 10 1 / 1.526 × 10 2 4.703 × 10 1 / 3.045 × 10 2 1.066 × 10 1 / 5.622 × 10 3
50 9.202 × 10 1 / 1.081 × 10 2 9.309 × 10 1 / 1.089 × 10 2 8.409 × 10 1 / 7.432 × 10 2 5.202 × 10 1 / 4.224 × 10 2 7.576 × 10 1 / 2.185 × 10 2 7.436 × 10 1 / 2.586 × 10 2 3.687 × 10 1 / 2.584 × 10 2 4.203 × 10 1 / 1.529 × 10 2 4.516 × 10 1 / 2.002 × 10 2 1.005 × 10 1 / 5.314 × 10 3
70 9.194 × 10 1 / 1.180 × 10 2 9.156 × 10 1 / 1.311 × 10 2 8.334 × 10 1 / 7.675 × 10 2 4.941 × 10 1 / 3.913 × 10 2 7.551 × 10 1 / 2.314 × 10 2 7.598 × 10 1 / 2.102 × 10 2 3.540 × 10 1 / 2.470 × 10 2 3.903 × 10 1 / 1.986 × 10 2 4.704 × 10 1 / 2.615 × 10 2 1.034 × 10 1 / 4.783 × 10 3
40830 9.942 × 10 1 / 5.093 × 10 3 9.927 × 10 1 / 3.991 × 10 3 9.348 × 10 1 / 5.970 × 10 2 5.885 × 10 1 / 4.101 × 10 2 8.675 × 10 1 / 1.696 × 10 2 8.648 × 10 1 / 1.646 × 10 2 4.630 × 10 1 / 2.494 × 10 2 3.948 × 10 1 / 1.510 × 10 2 3.919 × 10 1 / 1.124 × 10 2 1.692 × 10 1 / 7.472 × 10 3
50 9.996 × 10 1 / 5.870 × 10 4 9.994 × 10 1 / 7.161 × 10 4 9.422 × 10 1 / 5.475 × 10 2 6.686 × 10 1 / 4.412 × 10 2 9.016 × 10 1 / 1.741 × 10 2 8.867 × 10 1 / 1.208 × 10 2 5.025 × 10 1 / 2.854 × 10 2 3.613 × 10 1 / 1.458 × 10 2 3.811 × 10 1 / 1.664 × 10 2 1.583 × 10 1 / 3.981 × 10 3
70 9.998 × 10 1 / 4.882 × 10 4 9.992 × 10 1 / 1.274 × 10 3 9.630 × 10 1 / 3.884 × 10 2 7.334 × 10 1 / 3.733 × 10 2 9.147 × 10 1 / 1.560 × 10 2 8.896 × 10 1 / 1.367 × 10 2 5.386 × 10 1 / 3.623 × 10 2 3.943 × 10 1 / 1.182 × 10 2 4.280 × 10 1 / 1.049 × 10 2 1.637 × 10 1 / 4.873 × 10 3
1230 9.928 × 10 1 / 3.390 × 10 3 9.905 × 10 1 / 3.948 × 10 3 9.284 × 10 1 / 5.917 × 10 2 5.807 × 10 1 / 5.659 × 10 2 8.411 × 10 1 / 1.804 × 10 2 8.244 × 10 1 / 1.774 × 10 2 4.330 × 10 1 / 2.074 × 10 2 3.627 × 10 1 / 1.503 × 10 2 4.491 × 10 1 / 1.010 × 10 2 1.217 × 10 1 / 5.432 × 10 3
50 9.880 × 10 1 / 5.327 × 10 3 9.857 × 10 1 / 6.843 × 10 3 9.235 × 10 1 / 5.124 × 10 2 5.820 × 10 1 / 4.482 × 10 2 8.427 × 10 1 / 2.015 × 10 2 8.336 × 10 1 / 1.317 × 10 2 4.200 × 10 1 / 2.054 × 10 2 4.051 × 10 1 / 1.549 × 10 2 4.427 × 10 1 / 1.612 × 10 2 1.200 × 10 1 / 5.070 × 10 3
70 9.911 × 10 1 / 3.814 × 10 3 9.917 × 10 1 / 3.356 × 10 3 9.308 × 10 1 / 6.190 × 10 2 5.911 × 10 1 / 5.269 × 10 2 8.510 × 10 1 / 2.217 × 10 2 8.485 × 10 1 / 1.387 × 10 2 4.265 × 10 1 / 3.274 × 10 2 3.847 × 10 1 / 1.434 × 10 2 4.242 × 10 1 / 1.317 × 10 2 1.193 × 10 1 / 3.164 × 10 3
1630 9.781 × 10 1 / 6.968 × 10 3 9.738 × 10 1 / 7.856 × 10 3 9.030 × 10 1 / 6.627 × 10 2 5.299 × 10 1 / 4.800 × 10 2 8.122 × 10 1 / 2.140 × 10 2 8.144 × 10 1 / 1.328 × 10 2 3.816 × 10 1 / 2.113 × 10 2 3.894 × 10 1 / 2.641 × 10 2 4.332 × 10 1 / 2.197 × 10 2 1.003 × 10 1 / 4.763 × 10 3
50 9.917 × 10 1 / 2.110 × 10 3 9.885 × 10 1 / 3.234 × 10 3 9.287 × 10 1 / 5.615 × 10 2 5.478 × 10 1 / 4.127 × 10 2 8.505 × 10 1 / 1.465 × 10 2 8.177 × 10 1 / 2.072 × 10 2 3.947 × 10 1 / 2.433 × 10 2 4.850 × 10 1 / 1.660 × 10 2 4.555 × 10 1 / 2.156 × 10 2 9.741 × 10 2 / 4.147 × 10 3
70 9.734 × 10 1 / 4.744 × 10 3 9.596 × 10 1 / 8.007 × 10 3 8.883 × 10 1 / 8.667 × 10 2 5.546 × 10 1 / 3.013 × 10 2 8.050 × 10 1 / 1.758 × 10 2 7.884 × 10 1 / 1.969 × 10 2 3.758 × 10 1 / 2.630 × 10 2 4.297 × 10 1 / 1.436 × 10 2 4.584 × 10 1 / 1.905 × 10 2 9.760 × 10 2 / 4.974 × 10 3
Table 8. Analysis of the averages and standard deviations of T R a v e across 20 runs for single composite rules vs. DLIDQN.
Table 8. Analysis of the averages and standard deviations of T R a v e across 20 runs for single composite rules vs. DLIDQN.
nmeOursrule1rule2rule3rule4rule5rule6rule7rule8rule9
20830 6.657 × 10 1 / 1.125 × 10 2 7.832 × 10 1 / 8.543 × 10 3 7.542 × 10 1 / 2.043 × 10 2 8.764 × 10 1 / 9.013 × 10 3 6.841 × 10 1 / 1.915 × 10 2 7.350 × 10 1 / 1.453 × 10 2 8.503 × 10 1 / 2.299 × 10 2 6.034 × 10 1 / 2.664 × 10 2 6.514 × 10 1 / 1.870 × 10 2 8.283 × 10 1 / 1.975 × 10 2
50 5.858 × 10 1 / 4.284 × 10 2 8.034 × 10 1 / 5.551 × 10 3 7.714 × 10 1 / 1.725 × 10 2 8.605 × 10 1 / 1.192 × 10 2 6.726 × 10 1 / 2.832 × 10 2 7.232 × 10 1 / 2.591 × 10 2 8.303 × 10 1 / 2.657 × 10 2 6.420 × 10 1 / 1.405 × 10 2 6.528 × 10 1 / 2.448 × 10 2 8.418 × 10 1 / 1.999 × 10 2
70 5.745 × 10 1 / 8.057 × 10 3 7.676 × 10 1 / 9.679 × 10 3 7.526 × 10 1 / 1.843 × 10 2 8.666 × 10 1 / 2.041 × 10 2 7.142 × 10 1 / 1.690 × 10 2 7.375 × 10 1 / 1.885 × 10 2 8.573 × 10 1 / 1.898 × 10 2 6.548 × 10 1 / 1.204 × 10 2 7.035 × 10 1 / 1.244 × 10 2 8.479 × 10 1 / 1.955 × 10 2
1230 5.259 × 10 1 / 1.194 × 10 2 6.843 × 10 1 / 8.545 × 10 3 6.401 × 10 1 / 3.337 × 10 2 8.304 × 10 1 / 1.944 × 10 2 5.307 × 10 1 / 2.691 × 10 2 6.092 × 10 1 / 2.733 × 10 2 7.862 × 10 1 / 2.724 × 10 2 5.505 × 10 1 / 1.377 × 10 2 6.101 × 10 1 / 1.634 × 10 2 7.823 × 10 1 / 1.725 × 10 2
50 5.152 × 10 1 / 1.306 × 10 2 6.553 × 10 1 / 1.023 × 10 2 6.494 × 10 1 / 2.327 × 10 2 8.080 × 10 1 / 1.189 × 10 2 5.021 × 10 1 / 2.541 × 10 2 5.817 × 10 1 / 2.486 × 10 2 7.782 × 10 1 / 3.133 × 10 2 5.035 × 10 1 / 2.125 × 10 2 5.533 × 10 1 / 1.918 × 10 2 7.883 × 10 1 / 2.352 × 10 2
70 5.041 × 10 1 / 1.302 × 10 2 6.729 × 10 1 / 1.072 × 10 2 6.346 × 10 1 / 1.987 × 10 2 8.024 × 10 1 / 1.510 × 10 2 5.066 × 10 1 / 3.302 × 10 2 5.918 × 10 1 / 1.869 × 10 2 7.741 × 10 1 / 2.459 × 10 2 5.187 × 10 1 / 1.339 × 10 2 5.816 × 10 1 / 1.273 × 10 2 8.071 × 10 1 / 2.278 × 10 2
1630 3.798 × 10 1 / 1.751 × 10 2 5.851 × 10 1 / 1.400 × 10 2 5.469 × 10 1 / 3.217 × 10 2 7.957 × 10 1 / 2.129 × 10 2 3.279 × 10 1 / 3.796 × 10 2 4.734 × 10 1 / 3.284 × 10 2 7.425 × 10 1 / 4.493 × 10 2 4.288 × 10 1 / 2.154 × 10 2 4.975 × 10 1 / 1.638 × 10 2 7.832 × 10 1 / 2.746 × 10 2
50 3.117 × 10 1 / 1.502 × 10 2 5.208 × 10 1 / 1.873 × 10 2 5.592 × 10 1 / 2.547 × 10 2 7.418 × 10 1 / 2.109 × 10 2 3.593 × 10 1 / 3.608 × 10 2 4.829 × 10 1 / 2.416 × 10 2 7.414 × 10 1 / 3.213 × 10 2 5.275 × 10 1 / 1.583 × 10 2 5.249 × 10 1 / 1.912 × 10 2 7.938 × 10 1 / 1.999 × 10 2
70 2.826 × 10 1 / 2.583 × 10 2 5.474 × 10 1 / 1.682 × 10 2 5.201 × 10 1 / 3.392 × 10 2 7.427 × 10 1 / 2.913 × 10 2 3.074 × 10 1 / 3.159 × 10 2 4.413 × 10 1 / 2.971 × 10 2 7.040 × 10 1 / 2.999 × 10 2 3.872 × 10 1 / 2.245 × 10 2 4.350 × 10 1 / 1.553 × 10 2 7.708 × 10 1 / 2.326 × 10 2
30830 5.900 × 10 1 / 6.196 × 10 3 8.032 × 10 1 / 5.038 × 10 3 7.687 × 10 1 / 2.268 × 10 2 8.764 × 10 1 / 9.259 × 10 3 6.842 × 10 1 / 2.928 × 10 2 7.354 × 10 1 / 2.018 × 10 2 8.453 × 10 1 / 2.357 × 10 2 6.097 × 10 1 / 1.229 × 10 2 6.385 × 10 1 / 1.695 × 10 2 8.101 × 10 1 / 1.665 × 10 2
50 6.151 × 10 1 / 7.524 × 10 3 7.355 × 10 1 / 5.036 × 10 3 7.089 × 10 1 / 2.229 × 10 2 8.221 × 10 1 / 1.666 × 10 2 6.556 × 10 1 / 2.681 × 10 2 6.975 × 10 1 / 3.707 × 10 2 8.094 × 10 1 / 2.447 × 10 2 6.042 × 10 1 / 1.592 × 10 2 6.185 × 10 1 / 2.045 × 10 2 8.104 × 10 1 / 1.848 × 10 2
70 7.271 × 10 1 / 3.618 × 10 3 8.214 × 10 1 / 4.351 × 10 3 8.106 × 10 1 / 2.450 × 10 2 8.837 × 10 1 / 1.221 × 10 2 7.650 × 10 1 / 1.684 × 10 2 7.934 × 10 1 / 1.677 × 10 2 8.753 × 10 1 / 1.580 × 10 2 7.071 × 10 1 / 7.209 × 10 3 7.209 × 10 1 / 1.023 × 10 2 8.625 × 10 1 / 1.411 × 10 2
1230 6.067 × 10 1 / 1.041 × 10 2 7.675 × 10 1 / 6.345 × 10 3 7.436 × 10 1 / 2.310 × 10 2 8.585 × 10 1 / 2.219 × 10 2 6.876 × 10 1 / 2.909 × 10 2 7.250 × 10 1 / 2.246 × 10 2 8.290 × 10 1 / 1.992 × 10 2 6.173 × 10 1 / 1.566 × 10 2 6.855 × 10 1 / 1.132 × 10 2 8.328 × 10 1 / 2.167 × 10 2
50 5.031 × 10 1 / 1.188 × 10 2 7.320 × 10 1 / 6.748 × 10 3 6.955 × 10 1 / 3.014 × 10 2 8.389 × 10 1 / 1.318 × 10 2 4.998 × 10 1 / 2.011 × 10 2 6.685 × 10 1 / 2.613 × 10 2 8.207 × 10 1 / 2.097 × 10 2 5.415 × 10 1 / 1.543 × 10 2 5.870 × 10 1 / 2.085 × 10 2 8.102 × 10 1 / 2.667 × 10 2
70 4.854 × 10 1 / 9.758 × 10 3 7.245 × 10 1 / 7.092 × 10 3 6.894 × 10 1 / 3.044 × 10 2 8.380 × 10 1 / 1.556 × 10 2 5.731 × 10 1 / 2.510 × 10 2 6.456 × 10 1 / 2.422 × 10 2 8.089 × 10 1 / 2.429 × 10 2 5.351 × 10 1 / 1.923 × 10 2 5.945 × 10 1 / 1.437 × 10 2 8.408 × 10 1 / 2.010 × 10 2
1630 3.849 × 10 1 / 1.738 × 10 2 4.799 × 10 1 / 1.155 × 10 2 5.667 × 10 1 / 3.163 × 10 2 7.589 × 10 1 / 2.592 × 10 2 4.151 × 10 1 / 3.330 × 10 2 5.391 × 10 1 / 3.453 × 10 2 7.642 × 10 1 / 2.348 × 10 2 4.459 × 10 1 / 1.750 × 10 2 4.972 × 10 1 / 1.525 × 10 2 7.757 × 10 1 / 2.054 × 10 2
50 5.003 × 10 1 / 1.339 × 10 2 6.464 × 10 1 / 8.000 × 10 3 6.290 × 10 1 / 2.658 × 10 2 7.973 × 10 1 / 1.475 × 10 2 4.863 × 10 1 / 2.805 × 10 2 5.814 × 10 1 / 1.995 × 10 2 7.799 × 10 1 / 1.546 × 10 2 4.865 × 10 1 / 1.665 × 10 2 5.323 × 10 1 / 1.698 × 10 2 8.059 × 10 1 / 2.170 × 10 2
70 4.882 × 10 1 / 1.172 × 10 2 6.515 × 10 1 / 1.430 × 10 2 6.128 × 10 1 / 3.036 × 10 2 8.034 × 10 1 / 1.295 × 10 2 4.908 × 10 1 / 2.071 × 10 2 5.886 × 10 1 / 2.801 × 10 2 7.775 × 10 1 / 2.789 × 10 2 4.995 × 10 1 / 1.641 × 10 2 5.632 × 10 1 / 1.975 × 10 2 8.163 × 10 1 / 2.442 × 10 2
40830 6.549 × 10 1 / 6.776 × 10 3 7.711 × 10 1 / 7.521 × 10 3 7.587 × 10 1 / 2.163 × 10 2 8.540 × 10 1 / 1.405 × 10 2 7.324 × 10 1 / 1.626 × 10 2 7.584 × 10 1 / 1.859 × 10 2 8.638 × 10 1 / 1.712 × 10 2 6.068 × 10 1 / 1.397 × 10 2 6.588 × 10 1 / 1.286 × 10 2 8.055 × 10 1 / 1.637 × 10 2
50 6.082 × 10 1 / 1.018 × 10 2 8.140 × 10 1 / 6.494 × 10 3 7.945 × 10 1 / 1.812 × 10 2 8.771 × 10 1 / 1.183 × 10 2 7.705 × 10 1 / 2.021 × 10 2 7.987 × 10 1 / 1.879 × 10 2 8.750 × 10 1 / 1.531 × 10 2 6.682 × 10 1 / 1.317 × 10 2 7.062 × 10 1 / 9.001 × 10 3 8.169 × 10 1 / 2.249 × 10 2
70 6.572 × 10 1 / 7.475 × 10 3 8.558 × 10 1 / 5.269 × 10 3 8.517 × 10 1 / 2.017 × 10 2 8.998 × 10 1 / 8.636 × 10 3 8.094 × 10 1 / 1.775 × 10 2 8.322 × 10 1 / 1.402 × 10 2 8.874 × 10 1 / 1.757 × 10 2 7.155 × 10 1 / 1.138 × 10 2 7.452 × 10 1 / 1.101 × 10 2 8.772 × 10 1 / 1.300 × 10 2
1230 5.907 × 10 1 / 8.579 × 10 3 7.243 × 10 1 / 5.455 × 10 3 6.928 × 10 1 / 2.974 × 10 2 8.051 × 10 1 / 1.613 × 10 2 6.350 × 10 1 / 2.210 × 10 2 6.920 × 10 1 / 2.874 × 10 2 8.188 × 10 1 / 2.338 × 10 2 5.491 × 10 1 / 1.459 × 10 2 5.961 × 10 1 / 1.937 × 10 2 7.802 × 10 1 / 1.869 × 10 2
50 5.464 × 10 1 / 1.071 × 10 2 7.632 × 10 1 / 1.083 × 10 2 7.269 × 10 1 / 3.742 × 10 2 8.616 × 10 1 / 1.230 × 10 2 6.737 × 10 1 / 2.455 × 10 2 6.972 × 10 1 / 2.913 × 10 2 8.501 × 10 1 / 2.112 × 10 2 5.753 × 10 1 / 1.518 × 10 2 5.941 × 10 1 / 1.654 × 10 2 8.135 × 10 1 / 2.164 × 10 2
70 5.796 × 10 1 / 3.508 × 10 3 7.912 × 10 1 / 5.226 × 10 3 7.634 × 10 1 / 2.454 × 10 2 8.767 × 10 1 / 1.225 × 10 2 6.814 × 10 1 / 2.769 × 10 2 7.313 × 10 1 / 1.620 × 10 2 8.521 × 10 1 / 2.389 × 10 2 5.829 × 10 1 / 1.085 × 10 2 6.444 × 10 1 / 1.244 × 10 2 8.589 × 10 1 / 1.831 × 10 2
1630 5.018 × 10 1 / 9.645 × 10 3 6.895 × 10 1 / 9.719 × 10 3 6.709 × 10 1 / 2.979 × 10 2 8.264 × 10 1 / 1.478 × 10 2 5.808 × 10 1 / 3.893 × 10 2 6.529 × 10 1 / 2.370 × 10 2 8.185 × 10 1 / 2.219 × 10 2 5.243 × 10 1 / 1.347 × 10 2 5.772 × 10 1 / 8.974 × 10 3 8.046 × 10 1 / 1.731 × 10 2
50 6.226 × 10 1 / 4.583 × 10 3 7.461 × 10 1 / 5.823 × 10 3 7.317 × 10 1 / 2.471 × 10 2 8.695 × 10 1 / 1.195 × 10 2 6.303 × 10 1 / 2.238 × 10 2 7.009 × 10 1 / 1.846 × 10 2 8.568 × 10 1 / 1.549 × 10 2 6.522 × 10 1 / 1.267 × 10 2 6.601 × 10 1 / 6.936 × 10 3 8.777 × 10 1 / 9.884 × 10 3
70 4.869 × 10 1 / 1.310 × 10 2 7.273 × 10 1 / 6.786 × 10 3 7.062 × 10 1 / 2.400 × 10 2 8.495 × 10 1 / 1.294 × 10 2 5.799 × 10 1 / 2.923 × 10 2 6.641 × 10 1 / 2.204 × 10 2 8.189 × 10 1 / 1.948 × 10 2 5.362 × 10 1 / 1.454 × 10 2 5.810 × 10 1 / 1.630 × 10 2 8.276 × 10 1 / 2.197 × 10 2
Table 9. Analysis of the averages and standard deviations of makespan across 20 runs for classic dispatching rules vs. DLIDQN.
Table 9. Analysis of the averages and standard deviations of makespan across 20 runs for classic dispatching rules vs. DLIDQN.
nmeOursFIFOEDDSPTLPTMRT
20830 1.635 × 10 3 / 7.285 × 10 1 3.808 × 10 3 / 1.586 × 10 2 3.432 × 10 3 / 1.683 × 10 2 3.023 × 10 3 / 2.343 × 10 2 2.590 × 10 3 / 1.685 × 10 2 1.844 × 10 3 / 7.232 × 10 1
50 1.867 × 10 3 / 4.618 × 10 1 3.503 × 10 3 / 2.867 × 10 2 3.781 × 10 3 / 2.280 × 10 2 3.039 × 10 3 / 1.589 × 10 2 2.776 × 10 3 / 1.404 × 10 2 1.877 × 10 3 / 6.817 × 10 1
70 1.937 × 10 3 / 5.959 × 10 1 4.015 × 10 3 / 2.511 × 10 2 4.229 × 10 3 / 1.651 × 10 2 3.183 × 10 3 / 1.393 × 10 2 3.301 × 10 3 / 1.578 × 10 2 2.043 × 10 3 / 6.586 × 10 1
1230 1.087 × 10 3 / 6.275 × 10 1 2.616 × 10 3 / 1.811 × 10 2 2.638 × 10 3 / 1.328 × 10 2 2.371 × 10 3 / 1.630 × 10 2 1.997 × 10 3 / 9.397 × 10 1 1.173 × 10 3 / 6.041 × 10 1
50 1.220 × 10 3 / 4.861 × 10 1 2.920 × 10 3 / 1.441 × 10 2 3.208 × 10 3 / 1.562 × 10 2 2.623 × 10 3 / 1.271 × 10 2 2.400 × 10 3 / 1.614 × 10 2 1.355 × 10 3 / 6.211 × 10 1
70 9.584 × 10 3 / 3.765 × 10 1 2.240 × 10 3 / 7.706 × 10 1 1.982 × 10 3 / 1.558 × 10 2 1.617 × 10 3 / 9.245 × 10 1 1.563 × 10 3 / 6.518 × 10 1 1.103 × 10 3 / 5.697 × 10 1
1630 7.884 × 10 2 / 5.579 × 10 1 1.888 × 10 3 / 1.519 × 10 2 1.799 × 10 3 / 1.051 × 10 2 1.615 × 10 3 / 1.384 × 10 2 1.421 × 10 3 / 7.434 × 10 1 8.478 × 10 2 / 4.950 × 10 1
50 7.910 × 10 2 / 2.519 × 10 1 1.827 × 10 3 / 6.531 × 10 1 1.793 × 10 3 / 8.798 × 10 1 1.639 × 10 3 / 1.109 × 10 2 1.468 × 10 3 / 7.821 × 10 1 9.243 × 10 2 / 3.926 × 10 1
70 7.568 × 10 2 / 4.819 × 10 1 1.702 × 10 3 / 8.798 × 10 1 1.736 × 10 3 / 8.543 × 10 1 1.609 × 10 3 / 9.762 × 10 1 1.484 × 10 3 / 1.021 × 10 2 9.083 × 10 2 / 3.925 × 10 1
30830 1.841 × 10 3 / 4.900 × 10 1 3.625 × 10 3 / 1.499 × 10 2 3.587 × 10 3 / 1.073 × 10 2 2.983 × 10 3 / 1.195 × 10 2 2.995 × 10 3 / 1.340 × 10 2 1.878 × 10 3 / 5.872 × 10 1
50 1.841 × 10 3 / 6.275 × 10 1 4.059 × 10 3 / 1.659 × 10 2 4.354 × 10 3 / 1.886 × 10 2 3.487 × 10 3 / 2.077 × 10 2 3.183 × 10 3 / 1.261 × 10 2 1.961 × 10 3 / 9.982 × 10 1
70 2.547 × 10 3 / 4.407 × 10 1 5.664 × 10 3 / 1.865 × 10 2 5.733 × 10 3 / 1.833 × 10 2 4.469 × 10 3 / 1.619 × 10 2 4.804 × 10 3 / 1.626 × 10 2 2.610 × 10 3 / 6.262 × 10 1
1230 1.785 × 10 3 / 6.856 × 10 1 4.109 × 10 3 / 2.138 × 10 2 4.350 × 10 3 / 1.920 × 10 2 4.032 × 10 3 / 2.308 × 10 2 3.756 × 10 3 / 2.272 × 10 2 1.902 × 10 3 / 4.610 × 10 1
50 1.491 × 10 3 / 5.342 × 10 1 3.782 × 10 3 / 1.717 × 10 2 3.653 × 10 3 / 1.777 × 10 2 2.970 × 10 3 / 1.827 × 10 2 3.081 × 10 3 / 1.579 × 10 2 1.641 × 10 3 / 5.171 × 10 1
70 1.366 × 10 3 / 4.562 × 10 1 2.955 × 10 3 / 1.452 × 10 2 2.941 × 10 3 / 1.397 × 10 2 2.375 × 10 3 / 1.445 × 10 2 2.209 × 10 3 / 1.390 × 10 2 1.484 × 10 3 / 6.512 × 10 1
1630 8.613 × 10 2 / 5.363 × 10 1 2.213 × 10 3 / 9.350 × 10 1 2.491 × 10 3 / 1.188 × 10 2 1.926 × 10 3 / 7.350 × 10 1 1.888 × 10 3 / 1.060 × 10 2 9.702 × 10 2 / 4.338 × 10 1
50 1.226 × 10 3 / 4.746 × 10 1 2.932 × 10 3 / 1.363 × 10 2 3.035 × 10 3 / 1.794 × 10 2 2.566 × 10 3 / 1.275 × 10 2 2.253 × 10 3 / 1.470 × 10 2 1.328 × 10 3 / 7.583 × 10 1
70 1.027 × 10 3 / 4.635 × 10 1 2.639 × 10 3 / 9.533 × 10 1 2.475 × 10 3 / 1.503 × 10 2 2.253 × 10 3 / 1.292 × 10 2 2.173 × 10 3 / 1.194 × 10 2 1.134 × 10 3 / 3.871 × 10 1
40830 2.068 × 10 3 / 6.852 × 10 1 4.521 × 10 3 / 2.399 × 10 2 4.716 × 10 3 / 1.659 × 10 2 3.325 × 10 3 / 1.293 × 10 2 3.552 × 10 3 / 1.246 × 10 2 2.195 × 10 3 / 7.037 × 10 1
50 3.585 × 10 3 / 4.697 × 10 1 8.473 × 10 3 / 2.913 × 10 2 7.601 × 10 3 / 2.736 × 10 2 5.469 × 10 3 / 1.052 × 10 2 7.083 × 10 3 / 2.537 × 10 2 3.588 × 10 3 / 8.020 × 10 1
70 3.578 × 10 3 / 5.479 × 10 1 6.688 × 10 3 / 1.646 × 10 2 6.967 × 10 3 / 3.112 × 10 2 5.103 × 10 3 / 1.575 × 10 2 5.731 × 10 3 / 3.152 × 10 2 3.552 × 10 3 / 8.742 × 10 1
1230 1.757 × 10 3 / 5.697 × 10 1 4.648 × 10 3 / 2.678 × 10 2 4.663 × 10 3 / 2.821 × 10 2 3.200 × 10 3 / 1.990 × 10 2 3.157 × 10 3 / 1.523 × 10 2 1.846 × 10 3 / 4.708 × 10 1
50 1.797 × 10 3 / 6.543 × 10 1 4.416 × 10 3 / 2.563 × 10 2 4.196 × 10 3 / 1.621 × 10 2 3.375 × 10 3 / 1.981 × 10 2 3.525 × 10 3 / 2.258 × 10 2 1.930 × 10 3 / 6.738 × 10 1
70 1.893 × 10 3 / 4.507 × 10 1 4.633 × 10 3 / 2.002 × 10 2 4.523 × 10 3 / 2.069 × 10 2 3.488 × 10 3 / 1.515 × 10 2 3.305 × 10 3 / 1.456 × 10 2 1.978 × 10 3 / 5.715 × 10 1
1630 1.410 × 10 3 / 4.413 × 10 1 3.567 × 10 3 / 1.703 × 10 2 3.621 × 10 3 / 1.192 × 10 2 2.687 × 10 3 / 1.494 × 10 2 2.677 × 10 3 / 1.324 × 10 2 1.491 × 10 3 / 6.010 × 10 1
50 1.477 × 10 3 / 3.263 × 10 1 2.831 × 10 3 / 9.829 × 10 1 2.998 × 10 3 / 9.805 × 10 1 2.458 × 10 3 / 1.281 × 10 2 2.488 × 10 3 / 8.177 × 10 1 1.570 × 10 3 / 3.971 × 10 1
70 1.357 × 10 3 / 6.114 × 10 1 3.200 × 10 3 / 2.024 × 10 2 3.267 × 10 3 / 1.511 × 10 2 2.465 × 10 3 / 1.021 × 10 2 2.513 × 10 3 / 1.200 × 10 2 1.511 × 10 3 / 6.568 × 10 1
Table 10. Analysis of the averages and standard deviations of U a v e across 20 runs for classic dispatching rules vs. DLIDQN.
Table 10. Analysis of the averages and standard deviations of U a v e across 20 runs for classic dispatching rules vs. DLIDQN.
nmeOursFIFOEDDSPTLPTMRT
20830 9.819 × 10 1 / 7.412 × 10 3 3.667 × 10 1 / 1.372 × 10 2 3.932 × 10 1 / 1.277 × 10 2 4.747 × 10 1 / 2.344 × 10 2 5.361 × 10 1 / 2.003 × 10 2 8.372 × 10 1 / 1.774 × 10 2
50 9.884 × 10 1 / 5.444 × 10 3 4.225 × 10 1 / 2.985 × 10 2 4.074 × 10 1 / 1.898 × 10 2 5.313 × 10 1 / 1.919 × 10 2 6.092 × 10 1 / 1.586 × 10 2 8.800 × 10 1 / 1.091 × 10 2
70 9.910 × 10 1 / 4.426 × 10 3 4.018 × 10 1 / 1.988 × 10 2 3.609 × 10 1 / 9.805 × 10 3 5.197 × 10 1 / 2.275 × 10 2 4.604 × 10 1 / 1.325 × 10 2 8.933 × 10 1 / 1.272 × 10 2
1230 9.244 × 10 1 / 7.699 × 10 3 3.799 × 10 1 / 2.046 × 10 2 3.499 × 10 1 / 1.366 × 10 2 4.081 × 10 1 / 2.113 × 10 2 4.745 × 10 1 / 2.597 × 10 2 8.120 × 10 1 / 1.247 × 10 2
50 9.204 × 10 1 / 1.044 × 10 2 3.797 × 10 1 / 1.946 × 10 2 3.705 × 10 1 / 1.374 × 10 2 4.917 × 10 1 / 2.388 × 10 2 4.562 × 10 1 / 1.739 × 10 2 7.800 × 10 1 / 1.787 × 10 2
70 9.278 × 10 1 / 1.432 × 10 2 3.824 × 10 1 / 1.289 × 10 2 4.174 × 10 1 / 2.680 × 10 2 4.909 × 10 1 / 2.053 × 10 2 5.414 × 10 1 / 1.991 × 10 2 7.794 × 10 1 / 1.452 × 10 2
1630 8.353 × 10 1 / 1.646 × 10 2 3.604 × 10 1 / 2.208 × 10 2 3.669 × 10 1 / 1.664 × 10 2 3.896 × 10 1 / 2.598 × 10 2 4.567 × 10 1 / 1.633 × 10 2 7.375 × 10 1 / 1.336 × 10 2
50 8.436 × 10 1 / 1.386 × 10 2 3.811 × 10 1 / 1.319 × 10 2 4.075 × 10 1 / 1.741 × 10 2 4.896 × 10 1 / 2.840 × 10 2 5.437 × 10 1 / 1.819 × 10 2 7.497 × 10 1 / 1.271 × 10 2
70 8.654 × 10 1 / 1.417 × 10 2 4.033 × 10 1 / 1.052 × 10 2 3.805 × 10 1 / 1.103 × 10 2 4.677 × 10 1 / 1.728 × 10 2 4.591 × 10 1 / 1.437 × 10 2 7.043 × 10 1 / 2.031 × 10 2
30830 9.945 × 10 1 / 4.542 × 10 3 4.184 × 10 1 / 1.349 × 10 2 4.044 × 10 1 / 1.234 × 10 2 5.112 × 10 1 / 1.459 × 10 2 5.763 × 10 1 / 1.589 × 10 2 9.019 × 10 1 / 1.457 × 10 2
50 9.942 × 10 1 / 4.098 × 10 3 3.863 × 10 1 / 1.011 × 10 2 3.668 × 10 1 / 1.242 × 10 2 4.507 × 10 1 / 1.783 × 10 2 5.021 × 10 1 / 1.668 × 10 2 8.905 × 10 1 / 1.632 × 10 2
70 9.985 × 10 1 / 1.511 × 10 3 3.346 × 10 1 / 8.075 × 10 3 3.352 × 10 1 / 7.793 × 10 3 4.253 × 10 1 / 9.587 × 10 3 4.128 × 10 1 / 1.227 × 10 2 9.267 × 10 1 / 7.287 × 10 3
1230 9.744 × 10 1 / 8.648 × 10 3 3.686 × 10 1 / 1.305 × 10 2 3.433 × 10 1 / 1.266 × 10 2 4.193 × 10 1 / 1.832 × 10 2 4.190 × 10 1 / 1.831 × 10 2 8.435 × 10 1 / 6.755 × 10 3
50 9.853 × 10 1 / 5.352 × 10 3 3.577 × 10 1 / 1.684 × 10 2 3.547 × 10 1 / 1.428 × 10 2 4.739 × 10 1 / 1.850 × 10 2 4.311 × 10 1 / 2.077 × 10 2 8.536 × 10 1 / 1.252 × 10 2
70 9.689 × 10 1 / 7.467 × 10 3 4.008 × 10 1 / 1.274 × 10 2 3.910 × 10 1 / 1.808 × 10 2 5.039 × 10 1 / 1.979 × 10 2 5.364 × 10 1 / 1.916 × 10 2 8.379 × 10 1 / 1.527 × 10 2
1630 9.234 × 10 1 / 1.387 × 10 2 3.535 × 10 1 / 1.053 × 10 2 2.994 × 10 1 / 1.379 × 10 2 4.116 × 10 1 / 1.323 × 10 2 4.171 × 10 1 / 1.737 × 10 2 7.898 × 10 1 / 1.461 × 10 2
50 9.202 × 10 1 / 1.081 × 10 2 3.441 × 10 1 / 1.619 × 10 2 3.411 × 10 1 / 1.137 × 10 2 4.623 × 10 1 / 1.690 × 10 2 4.693 × 10 1 / 1.559 × 10 2 7.725 × 10 1 / 1.865 × 10 2
70 9.194 × 10 1 / 1.180 × 10 2 3.890 × 10 1 / 1.124 × 10 2 3.546 × 10 1 / 1.906 × 10 2 4.243 × 10 1 / 1.519 × 10 2 4.491 × 10 1 / 2.340 × 10 2 8.049 × 10 1 / 1.116 × 10 2
40830 9.942 × 10 1 / 5.093 × 10 3 3.783 × 10 1 / 1.741 × 10 2 3.834 × 10 1 / 1.348 × 10 2 5.328 × 10 1 / 1.160 × 10 2 5.129 × 10 1 / 1.228 × 10 2 9.211 × 10 1 / 1.105 × 10 2
50 9.996 × 10 1 / 5.870 × 10 4 3.209 × 10 1 / 7.069 × 10 3 3.616 × 10 1 / 1.039 × 10 2 5.301 × 10 1 / 9.666 × 10 3 3.973 × 10 1 / 1.206 × 10 2 9.398 × 10 1 / 4.763 × 10 3
70 9.998 × 10 1 / 4.882 × 10 4 3.866 × 10 1 / 7.138 × 10 3 3.756 × 10 1 / 1.354 × 10 2 5.461 × 10 1 / 1.414 × 10 2 4.915 × 10 1 / 1.504 × 10 2 9.328 × 10 1 / 9.328 × 10 3
1230 9.928 × 10 1 / 3.390 × 10 3 3.142 × 10 1 / 1.127 × 10 2 3.023 × 10 1 / 1.253 × 10 2 4.543 × 10 1 / 1.875 × 10 2 4.542 × 10 1 / 1.716 × 10 2 8.752 × 10 1 / 1.112 × 10 2
50 9.880 × 10 1 / 5.327 × 10 3 3.665 × 10 1 / 1.492 × 10 2 3.590 × 10 1 / 1.209 × 10 2 4.519 × 10 1 / 2.199 × 10 2 4.780 × 10 1 / 1.799 × 10 2 8.699 × 10 1 / 1.328 × 10 2
70 9.911 × 10 1 / 3.814 × 10 3 3.425 × 10 1 / 1.191 × 10 2 3.447 × 10 1 / 1.468 × 10 2 4.507 × 10 1 / 1.780 × 10 2 4.790 × 10 1 / 1.640 × 10 2 8.842 × 10 1 / 1.098 × 10 2
1630 9.781 × 10 1 / 6.968 × 10 3 3.528 × 10 1 / 1.503 × 10 2 3.313 × 10 1 / 7.206 × 10 3 4.649 × 10 1 / 1.482 × 10 2 4.536 × 10 1 / 1.654 × 10 2 8.578 × 10 1 / 1.574 × 10 2
50 9.917 × 10 1 / 2.110 × 10 3 4.355 × 10 1 / 1.325 × 10 2 4.208 × 10 1 / 1.292 × 10 2 5.324 × 10 1 / 1.991 × 10 2 5.246 × 10 1 / 1.120 × 10 2 8.820 × 10 1 / 6.211 × 10 3
70 9.734 × 10 1 / 4.744 × 10 3 3.848 × 10 1 / 1.938 × 10 2 3.549 × 10 1 / 1.093 × 10 2 4.949 × 10 1 / 1.732 × 10 2 4.580 × 10 1 / 1.992 × 10 2 8.281 × 10 1 / 1.517 × 10 2
Table 11. Analysis of the averages and standard deviations of T R a v e across 20 runs for classic dispatching rules vs. DLIDQN.
Table 11. Analysis of the averages and standard deviations of T R a v e across 20 runs for classic dispatching rules vs. DLIDQN.
nmeOursFIFOEDDSPTLPTMRT
20830 6.657 × 10 1 / 1.125 × 10 2 7.383 × 10 1 / 1.928 × 10 2 6.505 × 10 1 / 2.400 × 10 2 7.000 × 10 1 / 2.139 × 10 2 7.442 × 10 1 / 1.251 × 10 2 7.871 × 10 1 / 8.480 × 10 3
50 5.858 × 10 1 / 4.284 × 10 2 7.271 × 10 1 / 1.623 × 10 2 6.169 × 10 1 / 1.952 × 10 2 7.855 × 10 1 / 1.444 × 10 2 7.605 × 10 1 / 1.455 × 10 2 8.078 × 10 1 / 7.737 × 10 3
70 5.745 × 10 1 / 8.057 × 10 3 7.768 × 10 1 / 1.403 × 10 2 6.952 × 10 1 / 1.888 × 10 2 7.475 × 10 1 / 1.843 × 10 2 6.730 × 10 1 / 1.800 × 10 2 8.036 × 10 1 / 5.971 × 10 3
1230 5.259 × 10 1 / 1.194 × 10 2 7.322 × 10 1 / 1.308 × 10 2 5.539 × 10 1 / 9.554 × 10 3 5.981 × 10 1 / 1.249 × 10 2 6.266 × 10 1 / 1.210 × 10 2 6.926 × 10 1 / 1.468 × 10 2
50 5.152 × 10 1 / 1.306 × 10 2 6.651 × 10 1 / 1.583 × 10 2 5.063 × 10 1 / 1.379 × 10 2 5.431 × 10 1 / 1.787 × 10 2 5.005 × 10 1 / 2.423 × 10 2 6.875 × 10 1 / 1.433 × 10 2
70 5.041 × 10 1 / 1.302 × 10 2 7.337 × 10 1 / 8.501 × 10 3 5.386 × 10 1 / 2.020 × 10 2 6.473 × 10 1 / 2.360 × 10 2 5.732 × 10 1 / 1.480 × 10 2 6.744 × 10 1 / 1.008 × 10 2
1630 3.798 × 10 1 / 1.751 × 10 2 5.850 × 10 1 / 3.235 × 10 2 4.617 × 10 1 / 1.963 × 10 2 4.737 × 10 1 / 3.067 × 10 2 5.019 × 10 1 / 2.351 × 10 2 5.337 × 10 1 / 1.823 × 10 2
50 3.117 × 10 1 / 1.502 × 10 2 5.132 × 10 1 / 1.626 × 10 2 3.693 × 10 1 / 1.873 × 10 2 4.621 × 10 1 / 1.855 × 10 2 5.109 × 10 1 / 1.779 × 10 2 5.372 × 10 1 / 1.648 × 10 2
70 2.826 × 10 1 / 2.583 × 10 2 5.035 × 10 1 / 1.458 × 10 2 3.160 × 10 1 / 2.550 × 10 2 4.051 × 10 1 / 2.101 × 10 2 3.898 × 10 1 / 1.543 × 10 2 4.808 × 10 1 / 2.423 × 10 2
30830 5.900 × 10 1 / 6.196 × 10 3 5.494 × 10 1 / 1.016 × 10 2 5.722 × 10 1 / 1.563 × 10 2 7.238 × 10 1 / 1.158 × 10 2 7.240 × 10 1 / 1.346 × 10 2 8.114 × 10 1 / 5.384 × 10 3
50 6.151 × 10 1 / 7.524 × 10 3 7.827 × 10 1 / 8.962 × 10 3 5.814 × 10 1 / 2.205 × 10 2 6.426 × 10 1 / 2.322 × 10 2 7.011 × 10 1 / 1.406 × 10 2 8.222 × 10 1 / 8.668 × 10 3
70 7.271 × 10 1 / 3.618 × 10 3 8.343 × 10 1 / 6.147 × 10 3 7.445 × 10 1 / 9.601 × 10 3 8.055 × 10 1 / 8.791 × 10 3 8.374 × 10 1 / 7.061 × 10 3 8.446 × 10 1 / 3.724 × 10 3
1230 6.067 × 10 1 / 1.041 × 10 2 7.969 × 10 1 / 8.965 × 10 3 5.756 × 10 1 / 2.526 × 10 2 7.246 × 10 1 / 1.155 × 10 2 7.618 × 10 1 / 9.478 × 10 3 8.246 × 10 1 / 3.205 × 10 3
50 5.031 × 10 1 / 1.188 × 10 2 7.584 × 10 1 / 1.830 × 10 2 5.732 × 10 1 / 1.475 × 10 2 6.355 × 10 1 / 1.894 × 10 2 6.663 × 10 1 / 1.268 × 10 2 7.560 × 10 1 / 8.467 × 10 3
70 4.854 × 10 1 / 9.758 × 10 3 7.090 × 10 1 / 1.846 × 10 2 5.102 × 10 1 / 2.223 × 10 2 6.546 × 10 1 / 1.319 × 10 2 6.212 × 10 1 / 1.267 × 10 2 7.514 × 10 1 / 1.039 × 10 2
1630 3.849 × 10 1 / 1.738 × 10 2 6.171 × 10 1 / 1.479 × 10 2 3.958 × 10 1 / 1.420 × 10 2 5.046 × 10 1 / 2.517 × 10 2 5.075 × 10 1 / 2.122 × 10 2 6.028 × 10 1 / 1.644 × 10 2
50 5.003 × 10 1 / 1.339 × 10 2 6.893 × 10 1 / 1.354 × 10 2 5.511 × 10 1 / 1.653 × 10 2 5.887 × 10 1 / 1.757 × 10 2 6.251 × 10 1 / 1.991 × 10 2 6.825 × 10 1 / 1.926 × 10 2
70 4.882 × 10 1 / 1.172 × 10 2 4.504 × 10 1 / 1.608 × 10 2 4.593 × 10 1 / 2.556 × 10 2 5.773 × 10 1 / 2.425 × 10 2 6.330 × 10 1 / 1.228 × 10 2 6.730 × 10 1 / 1.014 × 10 2
40830 6.549 × 10 1 / 6.776 × 10 3 7.838 × 10 1 / 1.384 × 10 2 6.078 × 10 1 / 2.138 × 10 2 7.419 × 10 1 / 1.499 × 10 2 7.929 × 10 1 / 7.281 × 10 3 8.493 × 10 1 / 4.188 × 10 3
50 6.082 × 10 1 / 1.018 × 10 2 8.310 × 10 1 / 9.457 × 10 3 6.225 × 10 1 / 1.657 × 10 2 7.624 × 10 1 / 7.822 × 10 3 7.880 × 10 1 / 5.875 × 10 3 8.689 × 10 1 / 2.150 × 10 3
70 6.572 × 10 1 / 7.475 × 10 3 8.561 × 10 1 / 7.765 × 10 3 6.904 × 10 1 / 1.192 × 10 2 8.479 × 10 1 / 8.289 × 10 3 8.327 × 10 1 / 7.667 × 10 3 8.913 × 10 1 / 2.529 × 10 3
1230 5.907 × 10 1 / 8.579 × 10 3 7.605 × 10 1 / 1.137 × 10 2 6.441 × 10 1 / 1.454 × 10 2 6.911 × 10 1 / 1.527 × 10 2 7.254 × 10 1 / 1.331 × 10 2 8.083 × 10 1 / 3.797 × 10 3
50 5.464 × 10 1 / 1.071 × 10 2 7.303 × 10 1 / 1.628 × 10 2 5.606 × 10 1 / 2.521 × 10 2 7.541 × 10 1 / 1.266 × 10 2 7.200 × 10 1 / 1.988 × 10 2 7.975 × 10 1 / 6.598 × 10 3
70 5.796 × 10 1 / 3.508 × 10 3 7.846 × 10 1 / 1.147 × 10 2 5.573 × 10 1 / 2.382 × 10 2 5.363 × 10 1 / 8.257 × 10 3 7.318 × 10 1 / 1.081 × 10 2 8.055 × 10 1 / 5.346 × 10 3
1630 5.018 × 10 1 / 9.645 × 10 3 7.147 × 10 1 / 1.287 × 10 2 5.164 × 10 1 / 1.410 × 10 2 6.386 × 10 1 / 1.373 × 10 2 6.178 × 10 1 / 1.306 × 10 2 7.101 × 10 1 / 1.165 × 10 2
50 6.226 × 10 1 / 4.583 × 10 3 7.045 × 10 1 / 1.149 × 10 2 5.864 × 10 1 / 1.181 × 10 2 6.289 × 10 1 / 1.453 × 10 2 6.574 × 10 1 / 9.230 × 10 3 7.528 × 10 1 / 4.476 × 10 3
70 4.869 × 10 1 / 1.310 × 10 2 7.245 × 10 1 / 1.797 × 10 2 4.990 × 10 1 / 2.615 × 10 2 6.492 × 10 1 / 1.149 × 10 2 6.864 × 10 1 / 1.459 × 10 2 7.545 × 10 1 / 8.700 × 10 3
Table 12. Analysis of the averages and standard deviations of makespan across 20 runs for other RL-based algorithms vs. DLIDQN.
Table 12. Analysis of the averages and standard deviations of makespan across 20 runs for other RL-based algorithms vs. DLIDQN.
nmeOursDQND3QNHDMDDQN
20830 1.635 × 10 3 / 7.285 × 10 1 2.386 × 10 3 / 1.656 × 10 2 2.912 × 10 3 / 2.174 × 10 2 2.093 × 10 3 / 1.262 × 10 2
50 1.867 × 10 3 / 4.618 × 10 1 2.677 × 10 3 / 1.970 × 10 2 3.137 × 10 3 / 3.069 × 10 2 2.248 × 10 3 / 1.448 × 10 2
70 1.937 × 10 3 / 5.959 × 10 1 3.003 × 10 3 / 2.023 × 10 2 3.430 × 10 3 / 2.934 × 10 2 2.422 × 10 3 / 1.964 × 10 2
1230 1.087 × 10 3 / 6.275 × 10 1 1.867 × 10 3 / 1.270 × 10 2 2.304 × 10 3 / 2.598 × 10 2 1.526 × 10 3 / 1.470 × 10 2
50 1.220 × 10 3 / 4.861 × 10 1 2.219 × 10 3 / 2.468 × 10 2 2.598 × 10 3 / 2.130 × 10 2 1.721 × 10 3 / 1.758 × 10 2
70 9.584 × 10 2 / 3.765 × 10 1 1.924 × 10 3 / 1.660 × 10 2 2.043 × 10 3 / 1.930 × 10 2 1.451 × 10 3 / 1.253 × 10 2
1630 7.884 × 10 2 / 5.579 × 10 1 1.547 × 10 3 / 1.635 × 10 2 1.814 × 10 3 / 1.770 × 10 2 1.069 × 10 3 / 9.941 × 10 1
50 7.910 × 10 2 / 2.519 × 10 1 1.858 × 10 3 / 1.646 × 10 2 1.913 × 10 3 / 2.086 × 10 2 1.195 × 10 3 / 1.102 × 10 2
70 7.568 × 10 2 / 4.819 × 10 1 1.857 × 10 3 / 1.354 × 10 2 1.788 × 10 3 / 1.554 × 10 2 1.204 × 10 3 / 1.783 × 10 2
30830 1.841 × 10 3 / 4.900 × 10 1 2.548 × 10 3 / 1.202 × 10 2 3.083 × 10 3 / 2.833 × 10 2 2.151 × 10 3 / 1.307 × 10 2
50 1.841 × 10 3 / 6.275 × 10 1 2.835 × 10 3 / 1.983 × 10 2 3.439 × 10 3 / 3.073 × 10 2 2.206 × 10 3 / 1.792 × 10 2
70 2.547 × 10 3 / 4.407 × 10 1 3.385 × 10 3 / 2.226 × 10 2 3.987 × 10 3 / 3.447 × 10 2 3.017 × 10 3 / 2.216 × 10 2
1230 1.785 × 10 3 / 6.856 × 10 1 3.005 × 10 3 / 3.705 × 10 2 3.378 × 10 3 / 2.679 × 10 2 2.331 × 10 3 / 2.154 × 10 2
50 1.491 × 10 3 / 5.342 × 10 1 2.746 × 10 3 / 2.367 × 10 2 2.941 × 10 3 / 2.804 × 10 2 1.880 × 10 3 / 1.998 × 10 2
70 1.366 × 10 3 / 4.562 × 10 1 2.391 × 10 3 / 2.133 × 10 2 2.692 × 10 3 / 3.133 × 10 2 1.741 × 10 3 / 1.170 × 10 2
1630 8.613 × 10 2 / 5.363 × 10 1 1.769 × 10 3 / 1.492 × 10 2 2.248 × 10 3 / 2.555 × 10 2 1.225 × 10 3 / 1.641 × 10 2
50 1.226 × 10 3 / 4.746 × 10 1 2.451 × 10 3 / 2.362 × 10 2 2.727 × 10 3 / 2.052 × 10 2 1.651 × 10 3 / 2.411 × 10 2
70 1.027 × 10 3 / 4.635 × 10 1 2.416 × 10 3 / 2.551 × 10 2 2.431 × 10 3 / 3.394 × 10 2 1.497 × 10 3 / 2.043 × 10 2
40830 2.068 × 10 3 / 6.852 × 10 1 3.060 × 10 3 / 2.890 × 10 2 4.111 × 10 3 / 4.236 × 10 2 2.419 × 10 3 / 2.065 × 10 2
50 3.585 × 10 3 / 4.697 × 10 1 4.887 × 10 3 / 3.161 × 10 2 6.079 × 10 3 / 6.604 × 10 2 3.997 × 10 3 / 3.871 × 10 2
70 3.578 × 10 3 / 5.479 × 10 1 4.518 × 10 3 / 1.816 × 10 2 5.134 × 10 3 / 3.955 × 10 2 4.008 × 10 3 / 2.101 × 10 2
1230 1.757 × 10 3 / 5.697 × 10 1 2.762 × 10 3 / 1.791 × 10 2 3.438 × 10 3 / 2.914 × 10 2 2.119 × 10 3 / 2.098 × 10 2
50 1.797 × 10 3 / 6.543 × 10 1 2.871 × 10 3 / 2.330 × 10 2 3.683 × 10 3 / 3.346 × 10 2 2.158 × 10 3 / 2.101 × 10 2
70 1.893 × 10 3 / 4.507 × 10 1 2.934 × 10 3 / 1.419 × 10 2 3.571 × 10 3 / 3.300 × 10 2 2.262 × 10 3 / 1.898 × 10 2
1630 1.410 × 10 3 / 4.413 × 10 1 2.854 × 10 3 / 2.830 × 10 2 3.159 × 10 3 / 2.557 × 10 2 1.798 × 10 3 / 2.171 × 10 2
50 1.477 × 10 3 / 3.263 × 10 1 2.764 × 10 3 / 2.313 × 10 2 3.550 × 10 3 / 3.488 × 10 2 1.860 × 10 3 / 2.026 × 10 2
70 1.357 × 10 3 / 6.114 × 10 1 2.669 × 10 3 / 2.094 × 10 2 2.978 × 10 3 / 2.564 × 10 2 1.824 × 10 3 / 2.414 × 10 2
Table 13. Analysis of the averages and standard deviations of U a v e across 20 runs for other RL-based algorithms vs. DLIDQN.
Table 13. Analysis of the averages and standard deviations of U a v e across 20 runs for other RL-based algorithms vs. DLIDQN.
nmeOursDQND3QNHDMDDQN
20830 9.819 × 10 1 / 7.412 × 10 3 7.424 × 10 1 / 4.370 × 10 2 6.461 × 10 1 / 4.627 × 10 2 7.937 × 10 1 / 3.099 × 10 2
50 9.884 × 10 1 / 5.444 × 10 3 7.305 × 10 1 / 4.664 × 10 2 6.542 × 10 1 / 5.372 × 10 2 7.968 × 10 1 / 4.610 × 10 2
70 9.910 × 10 1 / 4.426 × 10 3 6.962 × 10 1 / 3.636 × 10 2 6.493 × 10 1 / 4.626 × 10 2 7.980 × 10 1 / 4.454 × 10 2
1230 9.244 × 10 1 / 7.699 × 10 3 6.709 × 10 1 / 3.878 × 10 2 5.591 × 10 1 / 4.300 × 10 2 7.173 × 10 1 / 4.426 × 10 2
50 9.204 × 10 1 / 1.044 × 10 2 6.006 × 10 1 / 5.249 × 10 2 5.266 × 10 1 / 3.982 × 10 2 6.925 × 10 1 / 4.896 × 10 2
70 9.278 × 10 1 / 1.432 × 10 2 5.847 × 10 1 / 4.656 × 10 2 5.619 × 10 1 / 3.518 × 10 2 7.009 × 10 1 / 3.631 × 10 2
1630 8.353 × 10 1 / 1.646 × 10 2 5.889 × 10 1 / 4.527 × 10 2 4.952 × 10 1 / 3.488 × 10 2 6.659 × 10 1 / 2.753 × 10 2
50 8.436 × 10 1 / 1.386 × 10 2 5.142 × 10 1 / 3.624 × 10 2 5.093 × 10 1 / 4.067 × 10 2 6.640 × 10 1 / 3.389 × 10 2
70 8.654 × 10 1 / 1.417 × 10 2 4.868 × 10 1 / 3.722 × 10 2 5.154 × 10 1 / 4.283 × 10 2 6.460 × 10 1 / 4.527 × 10 2
30830 9.945 × 10 1 / 4.542 × 10 3 7.676 × 10 1 / 2.806 × 10 2 6.845 × 10 1 / 4.141 × 10 2 8.336 × 10 1 / 3.115 × 10 2
50 9.942 × 10 1 / 4.098 × 10 3 7.174 × 10 1 / 4.945 × 10 2 6.108 × 10 1 / 4.225 × 10 2 8.070 × 10 1 / 5.105 × 10 2
70 9.985 × 10 1 / 1.511 × 10 3 7.832 × 10 1 / 3.734 × 10 2 7.055 × 10 1 / 5.408 × 10 2 8.276 × 10 1 / 4.155 × 10 2
1230 9.744 × 10 1 / 8.648 × 10 3 6.593 × 10 1 / 6.212 × 10 2 5.972 × 10 1 / 4.095 × 10 2 7.584 × 10 1 / 4.801 × 10 2
50 9.853 × 10 1 / 5.352 × 10 3 6.246 × 10 1 / 3.722 × 10 2 6.035 × 10 1 / 4.394 × 10 2 7.853 × 10 1 / 4.864 × 10 2
70 9.689 × 10 1 / 7.467 × 10 3 6.544 × 10 1 / 4.112 × 10 2 5.951 × 10 1 / 5.339 × 10 2 7.668 × 10 1 / 3.496 × 10 2
1630 9.234 × 10 1 / 1.387 × 10 2 6.106 × 10 1 / 4.021 × 10 2 5.038 × 10 1 / 3.474 × 10 2 7.014 × 10 1 / 4.341 × 10 2
50 9.202 × 10 1 / 1.081 × 10 2 5.664 × 10 1 / 4.694 × 10 2 5.196 × 10 1 / 3.620 × 10 2 7.099 × 10 1 / 6.776 × 10 2
70 9.194 × 10 1 / 1.180 × 10 2 5.235 × 10 1 / 4.360 × 10 2 5.228 × 10 1 / 5.052 × 10 2 7.013 × 10 1 / 4.678 × 10 2
40830 9.942 × 10 1 / 5.093 × 10 3 7.429 × 10 1 / 4.121 × 10 2 6.117 × 10 1 / 4.876 × 10 2 8.243 × 10 1 / 3.908 × 10 2
50 9.996 × 10 1 / 5.870 × 10 4 7.731 × 10 1 / 4.184 × 10 2 6.752 × 10 1 / 5.806 × 10 2 8.638 × 10 1 / 5.852 × 10 2
70 9.998 × 10 1 / 4.882 × 10 4 8.117 × 10 1 / 2.664 × 10 2 7.429 × 10 1 / 4.296 × 10 2 8.517 × 10 1 / 3.509 × 10 2
1230 9.928 × 10 1 / 3.390 × 10 3 7.197 × 10 1 / 3.915 × 10 2 6.157 × 10 1 / 4.371 × 10 2 8.024 × 10 1 / 4.583 × 10 2
50 9.880 × 10 1 / 5.327 × 10 3 7.052 × 10 1 / 4.750 × 10 2 5.832 × 10 1 / 4.088 × 10 2 7.995 × 10 1 / 4.937 × 10 2
70 9.911 × 10 1 / 3.814 × 10 3 7.121 × 10 1 / 2.417 × 10 2 6.207 × 10 1 / 3.849 × 10 2 8.072 × 10 1 / 4.488 × 10 2
1630 9.781 × 10 1 / 6.968 × 10 3 5.944 × 10 1 / 4.672 × 10 2 5.521 × 10 1 / 4.119 × 10 2 7.739 × 10 1 / 5.489 × 10 2
50 9.917 × 10 1 / 2.110 × 10 3 6.278 × 10 1 / 5.165 × 10 2 5.393 × 10 1 / 4.234 × 10 2 7.956 × 10 1 / 4.841 × 10 2
70 9.734 × 10 1 / 4.744 × 10 3 6.096 × 10 1 / 4.650 × 10 2 5.647 × 10 1 / 3.668 × 10 2 7.621 × 10 1 / 5.192 × 10 2
Table 14. Analysis of the averages and standard deviations of T R a v e across 20 runs for other RL-based algorithms vs. DLIDQN.
Table 14. Analysis of the averages and standard deviations of T R a v e across 20 runs for other RL-based algorithms vs. DLIDQN.
nmeOursDQND3QNHDMDDQN
20830 6.657 × 10 1 / 1.125 × 10 2 7.293 × 10 1 / 2.358 × 10 2 8.239 × 10 1 / 2.777 × 10 2 7.086 × 10 1 / 2.652 × 10 2
50 5.858 × 10 1 / 4.284 × 10 2 7.441 × 10 1 / 1.567 × 10 2 8.098 × 10 1 / 3.450 × 10 2 7.040 × 10 1 / 2.595 × 10 2
70 5.745 × 10 1 / 8.057 × 10 3 7.575 × 10 1 / 1.474 × 10 2 8.212 × 10 1 / 3.412 × 10 2 7.341 × 10 1 / 1.825 × 10 2
1230 5.259 × 10 1 / 1.194 × 10 2 6.583 × 10 1 / 2.002 × 10 2 7.243 × 10 1 / 5.078 × 10 2 5.805 × 10 1 / 3.818 × 10 2
50 5.152 × 10 1 / 1.306 × 10 2 6.506 × 10 1 / 1.821 × 10 2 7.440 × 10 1 / 2.920 × 10 2 5.445 × 10 1 / 3.458 × 10 2
70 5.041 × 10 1 / 1.302 × 10 2 7.029 × 10 1 / 3.741 × 10 2 7.170 × 10 1 / 3.485 × 10 2 5.738 × 10 1 / 2.395 × 10 2
1630 3.798 × 10 1 / 1.751 × 10 2 5.707 × 10 1 / 3.973 × 10 2 6.859 × 10 1 / 5.074 × 10 2 4.028 × 10 1 / 4.550 × 10 2
50 3.117 × 10 1 / 1.502 × 10 2 6.460 × 10 1 / 3.907 × 10 2 6.622 × 10 1 / 4.081 × 10 2 4.444 × 10 1 / 4.880 × 10 2
70 2.826 × 10 1 / 2.583 × 10 2 6.417 × 10 1 / 3.933 × 10 2 6.463 × 10 1 / 4.617 × 10 2 3.828 × 10 1 / 5.359 × 10 2
30830 5.900 × 10 1 / 6.196 × 10 3 7.405 × 10 1 / 1.665 × 10 2 8.105 × 10 1 / 3.604 × 10 2 7.039 × 10 1 / 2.302 × 10 2
50 6.151 × 10 1 / 7.524 × 10 3 7.114 × 10 1 / 2.346 × 10 2 7.691 × 10 1 / 3.471 × 10 2 6.612 × 10 1 / 2.547 × 10 2
70 7.271 × 10 1 / 3.618 × 10 3 7.756 × 10 1 / 1.138 × 10 2 8.503 × 10 1 / 3.213 × 10 2 6.954 × 10 1 / 4.078 × 10 2
1230 6.067 × 10 1 / 1.041 × 10 2 7.242 × 10 1 / 1.797 × 10 2 8.091 × 10 1 / 3.695 × 10 2 6.804 × 10 1 / 2.036 × 10 2
50 5.031 × 10 1 / 1.188 × 10 2 7.263 × 10 1 / 2.682 × 10 2 7.756 × 10 1 / 5.029 × 10 2 6.253 × 10 1 / 2.306 × 10 2
70 4.854 × 10 1 / 9.758 × 10 3 7.002 × 10 1 / 2.594 × 10 2 7.589 × 10 1 / 4.567 × 10 2 6.057 × 10 1 / 2.173 × 10 2
1630 3.849 × 10 1 / 1.738 × 10 2 6.253 × 10 1 / 4.275 × 10 2 7.025 × 10 1 / 4.501 × 10 2 4.435 × 10 1 / 4.732 × 10 2
50 5.003 × 10 1 / 1.339 × 10 2 6.871 × 10 1 / 3.377 × 10 2 7.539 × 10 1 / 4.799 × 10 2 5.307 × 10 1 / 4.648 × 10 2
70 4.882 × 10 1 / 1.172 × 10 2 7.246 × 10 1 / 5.090 × 10 2 7.234 × 10 1 / 5.064 × 10 2 5.372 × 10 1 / 3.008 × 10 2
40830 6.549 × 10 1 / 6.776 × 10 3 7.456 × 10 1 / 2.167 × 10 2 8.141 × 10 1 / 3.439 × 10 2 7.105 × 10 1 / 2.174 × 10 2
50 6.082 × 10 1 / 1.018 × 10 2 7.882 × 10 1 / 7.882 × 10 2 8.416 × 10 1 / 8.416 × 10 2 7.495 × 10 1 / 7.495 × 10 2
70 6.572 × 10 1 / 7.475 × 10 3 8.138 × 10 1 / 2.135 × 10 2 8.766 × 10 1 / 2.664 × 10 2 7.994 × 10 1 / 1.607 × 10 2
1230 5.907 × 10 1 / 8.579 × 10 3 7.117 × 10 1 / 4.097 × 10 2 7.590 × 10 1 / 5.410 × 10 2 5.404 × 10 1 / 2.853 × 10 2
50 5.464 × 10 1 / 1.071 × 10 2 7.219 × 10 1 / 2.530 × 10 2 8.121 × 10 1 / 4.172 × 10 2 6.486 × 10 1 / 2.902 × 10 2
70 5.796 × 10 1 / 3.508 × 10 3 7.472 × 10 1 / 1.971 × 10 2 8.307 × 10 1 / 4.075 × 10 2 6.855 × 10 1 / 2.770 × 10 2
1630 5.018 × 10 1 / 9.645 × 10 3 7.341 × 10 1 / 3.631 × 10 2 7.766 × 10 1 / 4.318 × 10 2 5.820 × 10 1 / 3.263 × 10 2
50 6.226 × 10 1 / 4.583 × 10 3 7.594 × 10 1 / 1.817 × 10 2 8.385 × 10 1 / 3.791 × 10 2 6.554 × 10 1 / 3.236 × 10 2
70 4.869 × 10 1 / 1.310 × 10 2 7.241 × 10 1 / 3.470 × 10 2 7.975 × 10 1 / 4.295 × 10 2 5.989 × 10 1 / 3.449 × 10 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, H.; Zheng, J.; Huang, L.; Tao, J.; Zhang, C. Solving Dynamic Multi-Objective Flexible Job Shop Scheduling Problems Using a Dual-Level Integrated Deep Q-Network Approach. Processes 2025, 13, 386. https://doi.org/10.3390/pr13020386

AMA Style

Xu H, Zheng J, Huang L, Tao J, Zhang C. Solving Dynamic Multi-Objective Flexible Job Shop Scheduling Problems Using a Dual-Level Integrated Deep Q-Network Approach. Processes. 2025; 13(2):386. https://doi.org/10.3390/pr13020386

Chicago/Turabian Style

Xu, Hua, Jianlu Zheng, Lingxiang Huang, Juntai Tao, and Chenjie Zhang. 2025. "Solving Dynamic Multi-Objective Flexible Job Shop Scheduling Problems Using a Dual-Level Integrated Deep Q-Network Approach" Processes 13, no. 2: 386. https://doi.org/10.3390/pr13020386

APA Style

Xu, H., Zheng, J., Huang, L., Tao, J., & Zhang, C. (2025). Solving Dynamic Multi-Objective Flexible Job Shop Scheduling Problems Using a Dual-Level Integrated Deep Q-Network Approach. Processes, 13(2), 386. https://doi.org/10.3390/pr13020386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop