Next Article in Journal
Two-Speed Transmission Structure and Optimization Design for Electric Vehicles
Previous Article in Journal
Optimization of the Voltage Total Harmonic Distortion in Multilevel Inverters by Using the Taguchi Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Agent Reinforcement Learning for Extended Flexible Job Shop Scheduling

1
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
2
State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
3
Beijing Engineering Research Center of Intelligent Systems and Technology, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
4
Guangdong Engineering Research Center of 3D Printing and Intelligent Manufacturing, Cloud Computing Center, Chinese Academy of Sciences, Dongguan 523808, China
5
Intelligent Manufacturing Center, Qingdao Academy of Intelligent Industries, Qingdao 266109, China
6
State Key Laboratory of Precision Electronic Manufacturing Technology and Equipment, Guangdong University of Technology, Guangzhou 510006, China
7
School of Electrical and Computer Engineering, Institute of Technology, Debremarkos University, Debremarkos 269, Ethiopia
8
State Key Laboratory for Management and Control of Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Gang Xiong is the co-first author.
Machines 2024, 12(1), 8; https://doi.org/10.3390/machines12010008
Submission received: 25 November 2023 / Revised: 8 December 2023 / Accepted: 11 December 2023 / Published: 22 December 2023

Abstract

:
An extended flexible job scheduling problem is presented with characteristics of technology and path flexibility (dual flexibility), varied transportation time, and an uncertain environment. The scheduling can greatly increase efficiency and security in complex scenarios, e.g., distributed vehicle manufacturing, and multiple aircraft maintenance. However, optimizing the scheduling puts forward higher requirements on accuracy, real time, and generalization, while subject to the curse of dimension and usually incomplete information. Various coupling relations among operations, stations, and resources aggravate the problem. To deal with the above challenges, we propose a multi-agent reinforcement learning algorithm where the scheduling environment is modeled as a decentralized partially observable Markov decision process. Each job is regarded as an agent that decides the next triplet, i.e., operation, station, and employed resource. This paper is novel in addressing the flexible job shop scheduling problem with dual flexibility and varied transportation time in consideration and proposing a double Q-value mixing (DQMIX) optimization algorithm under a multi-agent reinforcement learning framework. The experiments of our case study show that the DQMIX algorithm outperforms existing multi-agent reinforcement learning algorithms in terms of solution accuracy, stability, and generalization. In addition, it achieves better solution quality for larger-scale cases than traditional intelligent optimization algorithms.

1. Introduction

Flexible job shop scheduling (FJSP) is regarded as an effective measure to deal with the challenge of mass personalized and customized manufacturing in the era of Industry 4.0, and is widely extended to many real applications [1]. In this paper, we focus on a kind of extended FJSP, which considers both technology and path flexibility (dual flexibility), varying transportation times among stations, and an uncertain environment. This extended FJSP is termed FJSP-DT (FJSP with dual flexibility and transportation time variation). FJSP-DT occurs in real complex scenarios, such as distributed automobile assembly scheduling, multiple aircraft maintenance scheduling, and robot scheduling in a warehouse sorting center, etc. FJSP-DT takes the following characters into consideration and is abstractly depicted in Figure 1.
  • Dual flexibity. FJSP-DT considers both technology and path flexibility. As shown in Figure 1, the blue operation may be processed before the red operation, or vice versa, which results in technology flexibility. The same operations can be processed by different robots, e.g., the orange operation can be processed by Robot m 1 or Robot m 4 , hence leading to path flexibility.
  • Varied transportation time. One job needs to be processed through multiple stations, and the transportation time between operations varies depending on the distance between stations.
  • Cooperative gaming and uncertain environment. Multiple jobs coordinate their operations to achieve shorter makespan employing preempt resources, such as robots, stations, etc. The environment also suffers from uncertain factors, e.g., robot failures.
Based on the thorough survey of [2] and papers related to JSP published after 2022 (e.g., [3,4,5,6,7,8,9,10]), we classified the works related to FJSP into 23 categories by the factors considered (shown in Table 1), such as machine availability, routing, varied transportation time, etc. It can be concluded that there is a lack of FJSP studies that take all the characteristics above into account. It is necessary to conduct in-depth research due to the importance of FJSP-DT applications. However, solving FJSP-DT faces significant challenges.
  • Generalization. The scheduling algorithm should be adaptable to changing conditions, such as altered operation sequences, modified stations, robot distributions, or varying job quantities.
  • Real-time. Due to the uncertainty of the environment, the scheduler needs to respond in real time (in seconds) to robot failures and make real-time scheduling.
  • Accuracy. Significant operational risks may arise in the application scenario corresponding to FJSP-DT. In automobile assembly scheduling, poor coordination between operations may lead to collisions. Therefore, the algorithm needs to ensure operational safety.
  • Curse of dimensionality. The solution space expands as the number of agents and resources (such as robots, and stations) increases, resulting in a computational dilemma. Here, M, S, and T are used to represent the sets of robots, stations, and time respectively. N and O represent the sets of agents and operations respectively. The solution space of FJSP-DT is A | M | × | S | × | T | | N | × | O | which causes an explosion in combinations.
The methods of solving FJSP-related problems fall under three categories: mathematical programming, meta-heuristic algorithms, and reinforcement learning. Mathematical programming mainly applies to small-scale problems due to its long computation time [11]. Therefore, more than 60% of the publications applied evolutionary algorithms (meta-heuristic) to overcome the challenge [12]. However, these methods require high computational costs and result in unsuitability for large-scale, real-time scheduling problems [13].
Reinforcement learning has been widely applied to solve numerous optimization problems in the real world [14], and was regarded as one of the top five most promising algorithms [15] to solve JSP due to its strong self-learning ability. Since the Deep Q-Network was proposed, multi-agent reinforcement learning (MARL) has been a focused piece of research in solving FJSP. To map FJSP to the MARL-solving architecture, each job or robot is regarded as an agent in the scheduling environment, where multiple agents cooperate and learn their policies. After offline training, the MARL algorithm can be deployed online to make decisions based on real-time observations. Consequently, MARL has the great potential to model complex workshop scheduling problems [16], and can achieve real-time computation and adaptability [17]. Therefore, we adapt MARL to solve the FJSP-DT problem with several adaptive improvements in architecture design, value function calibration, etc. The main contributions of our paper can be summarized as follows.
  • The FJSP-DT abstracted from real and complex scenarios is proposed. As far as our literature review is concerned, the problem is new and needs to be fully investigated. The FJSP-DT is modeled as a decentralized partially observable Markov decision process in that an agent may only observe the information around it in real applications. Then, the MARL-based method can be applied to achieve high adaptability and real-time decisions.
  • We build an adaptive and stable multi-agent learning framework by combining a graph convolutional network (GCN) and the actor–critic structure. GCN extracts embedding graph’s structural features and represents system states in non-Euclidean space, thereby degrading dimension curses and adapting to various scheduling environments. The actor–critic structure can update the network parameters in a single step without running an episode, making it faster than the policy gradient algorithm.
  • A double Q-value mixing algorithm (DQMIX) under the above framework is proposed to address the challenge of fast convergence and high adaptability. The algorithm combines an unrestricted optimal network with a monotonic mixing network to improve exploration and exploitation capabilities. It also integrates mechanistic constraints into data-based learning, mitigating the curse of dimensionality by eliminating invalid actions. In addition, the reward function is designed as a function of reduction in makespan estimation to mitigate the learning challenges caused by sparse feedback.

2. Related Works

In this section, we will focus our review on the research about reinforcement learning in applications to the optimization of JSP, including single-agent reinforcement learning and multi-agent reinforcement learning.
  • Single-Agent Reinforcement Learning (SARL): The algorithm only contains one agent that makes all the decisions for a control system.
  • Multi-Agent Reinforcement Learning (MARL): The algorithm comprises multiple agents that interact with the environment through their respective policies.

2.1. SARL for Scheduling

SARL virtualizes an agent interacting with the scheduling environment, learning a scheduling policy, and then making decisions. The early paper applying SARL to JSP may be traced back to Zhang and Dietterich (1995) to learn a heuristic evaluation function over states [18]. Subsequently, Aydin and Öztemel (2000) [19] applied reinforcement learning to choose dispatching rules depending on the current state of a production system. Since the proposal of Deep Q-Network (DQN), using SARL to solve JSP has attracted more and more attention.

2.1.1. SARL with Value Iteration

Waschneck et al. (2018) [20] applied the DQN algorithm to solve a dynamic and flexible production problem with the objective of maximizing plant throughput. The proposed model took machine availability and processing characteristics as states and mapped the states to the station selection. Luo (2020) [21] developed an optimization algorithm based on Double DQN (DDQN) for dynamic FJSP with order insertion. The algorithm can select appropriate scheduling rules according to the job state and obtain a plan better than the general scheduling rules. Lang et al. (2020) [22] combined the DQN algorithm with discrete event simulation to solve a flexible job shop problem with process planning. Two independent DQN agents are trained. One agent selects operation sequences, while the other assigns jobs to machines. Du et al. (2021) [6] considered an FJSP with time-of-use electricity price constraint and dual-objective optimization for the makespan and total price and proposed a hybrid multi-objective optimization algorithm of estimation of distribution algorithm and DQN to solve the problem. Li et al. (2022) [5] presented dynamic FJSPs with insufficient traffic resources (DFJSP-ITR). They proposed a hybrid DQN (HDQN) that includes double Q-learning, prioritized replay, and a soft target network update policy to minimize the maximum duration and total energy consumption. Gu et al. (2023) [23] integrated DQN method into a scalp swarm algorithm (SWA) framework to dynamically tune the population parameters of SWA for JSP solving.

2.1.2. SARL with Policy Iteration

Wang et al. (2021) [24] considered the uncertainties, such as the mechanical failure of the job shop, and proposed a dynamic scheduling method based on the proximal policy optimization (PPO) to find the optimal scheduling policy. The states are defined by the job processing state matrix, the designated machine matrix, and the processing time matrix of the operations. An action set is defined as the operation selected from the candidate operation set, and the reward is related to machine utilization. The results showed that the proposed approach based on reinforcement learning obtained comparative solutions and achieved adaptive and real-time scheduling.
The application of SARL to solve the scheduling problem has some limitations. Firstly, FJSP-DT occurs in an uncertain environment, and the information obtained by the agent is likely to be incomplete, which is difficult for SARL to handle since SARL depends on global information. Secondly, SARL does not tackle communication and collaboration between jobs or machines, resulting in the loss of important scheduling information [16]. Thirdly, the action space of SARL expands with the number of jobs or machines [25], and a high action dimension can pose a challenge to policy learning. Research shows that the performance of policy gradient methods gradually lags as the action dimension increases [26].

2.2. MARL for Scheduling

MARL aims to model complex worlds where each agent can make adaptive decisions, realizing competition and cooperation with humans and other agents, and is attracting more and more attention in academia and industry [27].
From the perspective of the multi-agent system’s training paradigm, agents’ training can be broadly divided into distributed and centralized schemes.
  • Distributed Training Paradigm (DTP): In the distributed Paradigm, agents learn independently of other agents and do not rely on explicit information exchange.
  • Centralized Training Paradigm (CTP): The centralized paradigm allows agents to exchange additional information during training, which is then abandoned during tests. Agents receive only the locally observable information and independently determine actions according to their policies during execution.

2.2.1. MARL with DTP

Regarding works on MARL with DTP, Aissani et al. (2009) [28] applied MARL for adaptive scheduling in multi-site companies. The supervisor agent sends requests to inventory agents and resource agents at different sites for a solution in the company’s multi-agent system. The inventory agent asks the resource agent to propose a solution. The resource agent then starts its decision-making algorithm based on the SARSA (state–action–reward–state-action) algorithm using the data system (the resource state, task duration, etc.) and sends back a solution. Martínez et al. (2020) [29] proposed a MARL tool for JSPs where machines are regarded as agents. This tool allows the user to either keep the best schedule obtained by a Q-learning algorithm or modify it by fixing some operations to satisfy certain constraints. The tool then optimizes the modified solution, taking into account the user’s preferences and using the possible alternatives. Hameed et al. (2020) [30] presented a distributed reinforcement learning approach for JSPs. The innovation of this paper is that the authors modeled various relationships within the manufacturing environments (robot manufacturing cells) as graph neural networks (GNN). Zhou et al. (2021) [31] proposed a new distributed architecture with multiple artificial intelligence (AI) schedulers for the online scheduling of orders in smart factories. Each machine is taken as a scheduler agent, which collects the scheduling states of all machines as input for training separately and executes the scheduling policy, respectively. Popper et al. (2021) [32] proposed a distributed MARL scheduling method for the multi-objective optimization problem of minimizing energy consumption and delivery delay in the production process. The basic algorithm is solved by PPO, which regulates the joint behavior of each agent through a common reward function. The algorithm can schedule any number of operations. Burggräf et al. (2022) [7] presented a deep MARL approach with distributed actor-critic architecture to solve dynamic FJSPs. The novelty of this work lies in parameterizing the state and action spaces. Zhang et al. (2022) [8] constructed a multi-agent manufacturing system (MAMS) with the capability of online scheduling and policy optimization for the dynamic FJSP. Various machine tools are modeled as agents capable of environmental perception, information sharing, and autonomous decision making.

2.2.2. MARL with CTP

Wang et al. (2021) [33] proposed a flexible and hybrid production scheduling problem and introduced a multi-agent deep reinforcement learning (MADRL) scheduling method called Multi-Agent Deep Deterministic Policy Gradient (MADDPG). Similarly, Wang et al. (2022) [9] introduced a modeling method of the decentralized partially Markov decision process (Dec-POMDP) for the resource preemption working environment (PRE), and applied QMIX to solve the PRE scheduling problem, where each job is an agent who selects its action according to current observation. Jing (2022) [10] designed a MARL scheduling framework based on GCN, namely a graph-based multi-agent system (GMAS), to solve the FJSP. First, a probabilistic model of the directed acyclic graph of the FJSP is constructed from the product processing network and workshop environment. Then, the author modeled the FJSP as a topological graph prediction process and adjusted the scheduling policy by predicting the connection probabilities between the edges.
In contrast to DTP, CTP shares information among agents through a centralized evaluation function, making learning more stable and leading to fast convergence. Therefore, regarding the solving method for FJSP-DT, MARL integrated with the CTP training paradigm bears the great potential of agility, adaptability, and accuracy, and is the focus of study therewith.

3. Problem Description and Model Formulation

This section describes the FJSP-DT problem and its modeling as a decentralized partially observable Markov decision process.

3.1. Problem Description

There are | M | robots in the FJSP-DT workshop and | S | stations. In the workshop, a total of | N | jobs need to be processed and job n i has | O i | operations. The relationships between two operations may be sequential, parallel, or conflicted. The processing time t j k of the operation o i j depends on the robot m k . There exists a subset of robots M j to satisfy the processing requirements of the operation o i j . Each robot covers a subset of stations. Multiple robots can share the same station. The transportation time τ p q from the station s p to the station s q for the next operation needs to be considered. The notations formulating the problem are given in Table 2.
There are some assumptions as below.
  • All jobs are available after release dates.
  • A robot can only process one job at a time.
  • Each operation can only be processed on one robot at a time.
  • Each operation cannot be interrupted during processing.
  • There is no buffer in the station, and each station can only accommodate one job at a time.

3.2. Dec-POMDP Framework

FJSP-DT can be regarded as a cooperative multi-agent Markov game due to the jobs’ cooperative scheduling objective. A fully cooperative multi-agent task can be formulated as a decentralized partially observable Markov decision process (Dec-POMDP), described by a tuple N , S , A , P , R , O , Ω , γ , where:
  • N = { 1 , , n } is the set of n > 1 agents;
  • S is a set of states denoting all joint states possible by the multi-agents;
  • A = × i N A i is the set of joint actions, where A i denotes the set of actions for agent i;
  • P is the transition probability function;
  • R = R ( s , a ) is the reward function, mapping states and joint actions to real numbers;
  • O = × i N O i is the set of joint observations, where O i is the set of observations available to agent i;
  • Ω is the observation probability function;
  • γ 0 , 1 is a discount factor.
The agents and environment interact at each time step. t = 0 , 1 , 2 , . s S describes the true state of the environment, and the state at time step t is denoted by s t . The action for agent i is denoted by a i A i . At every time step t, each agent i takes an action a t i , leading to one joint action a = a 1 , a 2 , , a n T , a A at every time step. In the fully cooperative setting, all agents share the same reward function R 1 = R 2 = = R n = R ( s , a ) . The state transition function P = P ( s | s , a ) : S × A × S [ 0 , 1 ] describes the dynamics of the environment. The observation function Ω ( s , a ) : S × A O specifies the probabilities p ( o | a , s ) of joint observations. Due to the partial observability, each agent only draws individual observations o i O i , defined by the observation function Ω ( s , a ) . The observations for agent i at time step t are denoted as o t i . At every time step, the environment emits one joint observation o = o 1 , , o n T , o O . The number of joint observations is | O | = | O 1 × O 2 × × O n | = Π i | O i | , and the number of joint actions is | A | = | A 1 × A 2 × × A n | = Π i | A i | . And then, both state and action explosions occur.
Each agent has an action–observation history τ i H ( O × A ) * . At time step t, the action–observation history for agent i is denoted by τ t i = [ a 0 i , o 1 i , , a t 1 i , o t i ] T . A joint action–observation history τ = τ 1 , , τ n T denotes the action–observation histories of all agents. We refer to the state, action, observation, reward, and action–observation variables at time step t as S t , A t , O t , R t , and H t , respectively.
The stochastic policy for agent i is denoted by π i ( a i | τ i ) : H × A 0 , 1 , then a multi-agent policy will be denoted by π ( a | τ ) = ( π 1 , , π n ) .
A centralized training and decentralized execution paradigm [25] is adopted in our learning algorithm. Each job is taken as an agent. During training, each agent can access the environment state s and all agent’s observations o and their histories τ . However, during execution, each agent can only access its own observations o i and action–observation history τ i to determine its action according to its policy. The modeling for FJSP-DT under the Dec-POMDP framework will be detailed in Section 3.3Section 3.5.

3.3. State and Observations

The states and observations, including four 4 factors: job, operation, station, and robot, are high-dimension data. Hence, graph-embedding technology is applied to represent states and observations. Then, GCN is applied to compress states and observation dimensions, since it is effective in reducing the computational cost of large and sparse embedding graphs.
The graph can be defined as G = { V , E , C , X } with | V | = | N | + | O | + | S | + | M | nodes and | E | edges. v i V denotes the ith node of the graph; e ( v i , v j ) E denotes the edge between the node v i V and the node v j V . C R | V | × | V | denotes the adjacency matrix, and the element c i j C represents the weight of the edge e ( v i , v j ) . The feature matrix of the nodes is represented by X R | V | × p , where p denotes the dimension of the feature vector. The schematic diagram of G is shown in the lower left part of Figure 2.
The types of nodes for our problem include four categories: job, operation, station, and robot. We utilize the following form to represent the feature matrix.
X = j 11 j | N | 1 m 11 m | M | 1 s 11 s | S | 1 o 11 o | O | 1 j 1 p j | N | p m 1 p m | M | p s 1 p s | S | p o 1 p o | O | p ,
where j, m, s, and o represent job, robot, station, and operation, respectively, and p is the number of features. The features of the job include the job type, the number of operations remaining, and the total remaining processing time of all unfinished operations. The features of the operation include the operation type and the processing time. The features of the station and the robot include the node type and the remaining occupancy time.
The edges between operations, robots, or stations do not change during processing. However, the edges between jobs and operations, jobs and robots, and jobs and stations change as machining progresses. Hence, the weights of those dynamic edges are updated during machining, which reflects the current state of the availability of robots and stations, each job’s location, and the machining progress.
The adjacency matrix C, combining the feature matrix X of the embedding graph, is of high dimension, and introduces difficulty for network parameter optimization therewith. Hence, GCN [34] is adopted to extract the structural features of the graph which is aggregated with node features to generate new node representations. Then, those new representations can be input into the network for learning. This GCN-based approach can not only improve the representation ability of node features but also compress the network’s parameter size and speed up optimization.
A GCN layer can be described as
H h + 1 = ρ D ˜ 1 2 C ˜ D ˜ 1 2 H ( h ) W ( h ) ,
where C ˜ = C + I | V | is the self-ring adjacency matrix of the graph G, and I is an identity matrix with vertices. The element d ˜ i , i in the matrix D ˜ is equal to j c ˜ i , j . H ( h ) is the input matrix of the h layer of the GCN, and the input of the first layer ( h = 0 ) is the feature matrix X of the nodes. W ( h ) is a learnable parameter matrix in the hth layer.
Equation (2) represents the multi-layer GCN. In this equation, we take the normalized self-ring adjacency matrix D ˜ 1 2 C ˜ D ˜ 1 2 as the operator of the feature matrix. Therefore, the graph embedding matrix is denoted as
X ˜ = D ˜ 1 2 C ˜ D ˜ 1 2 X ,
where X is the feature matrix of the nodes, and X ˜ is the embedding graph matrix as input of the learning model.

3.3.1. State

The state includes information about all jobs. According to Equations (1) and (3), at time step t, state S t can be represented by the following form,
S t = D ˜ t 1 2 C ˜ t D ˜ t 1 2 X t ,
which is a matrix that can be flattened into a vector before inputting into the network. The state is utilized during algorithm training.

3.3.2. Observations

In Dec-POMDP, each agent cannot fully observe other agents’ information. Thus, each agent’s original observations are defined as a subgraph G i = { V i , E i , C i , X i } , where we exclude other jobs’ nodes and edges except job n i . Its feature matrix can be rewritten to
X i = j i 1 m 11 m | M | 1 s 11 s | S | 1 o 11 o | O | 1 j i p m 1 p m | M | p s 1 p s | S | p o 1 p o | O | p .
Then, the transformed observation matrix of agent i can be represented as
O t i = D t i ˜ 1 2 C t i ˜ D t i ˜ 1 2 X t i .
The observation defined by (6) is required during algorithm execution besides the training period.

3.4. Actions

At each decision point, the agent must choose a station and a robot as an action, called station–robot pair. Each agent needs a matrix to represent its action space, and the joint action space of multiple agents should be represented with a three-dimensional tensor. This high-dimensional action space will cause difficulties in learning policies.
Some mechanistic constraints are introduced to reduce the action space. In general, a robot can only reach some of the stations, so the number of accessible station–robot pairs is cut down. For example, if the workshop has 20 robots and 20 stations, and a robot can only reach 4 stations, then the number of station–robot pairs can be reduced to 80 ( 20 × 4 ), far less than 400 ( 20 × 20 ). Furthermore, fewer feasible station–robot pairs satisfy the constraints of operation precedence and the availability of robots and stations in the machining process. Those constraints will further reduce the number of alternative actions during model training.
Therefore, we adopt mechanic constraints to pick out feasible station–robot pairs as each agent’s action space. Suppose that the number of possible combinations is K; then, the action set is represented as
A i = { a 1 , a 2 , , a K , a K + 1 , a K + 2 , a K + 3 , a K + 4 }
See Table 3 for the details. At time step t, the possible actions for agent i are denoted A t i A i .

3.5. Reward

The reward function guides the agent to minimize C m a x or maximize C m a x , or c C m a x , where C m a x denotes the maximum completion time of all jobs, which determines the makespan, c is a constant. However, the scheduling in FJSP-DT is a sparse reward task, and we cannot accurately compute the actual makespan C m a x until all jobs’ tasks are completed. Thus, the agents cannot receive real rewards at each decision epoch. This brings forward learning challenges. To address this challenge, we estimate the makespan ( E C m a x ) at each step that is utilized to evaluate the quality of each agent’s action [35].
E C m a x = max E C j i 1 i | N | , 1 j | O | ,
where E C j i denotes the estimated completion time of the operation o i , j .
We employ E C m a x ( t ) to symbolize the E C m a x at time step t. The immediate reward function is designed as a function of reduction in E C m a x .
r t ( s , a ) = E C m a x ( t ) E C m a x ( t + 1 ) .
At t = 0 , a constant c is set as an initial value, namely E C m a x 0 = c . Maximizing the immediate reward function (9) at each time step is equivalent to maximizing the accumulated reward c C m a x .
The estimated completion time of a certain operation is equal to the processing time of that operation plus the estimated completion time of its preceding operation, which is represented as follows:
E C j i = E C j 1 i + t j ,
where E C j 1 i denotes the estimated completion time of the operation o i , j 1 , which is the previous operation of the current operation o i j . The variable t j denotes the processing time of the current operation o i j . We estimate t j by averaging a set of processing times { t j 1 , t j 2 , , t j k } , where 1 , , k are indices of the robots that can support the operation o j .
If C m a x exceeds C u p which is the upper bound set by the human planner, the solution is poor, and a negative reward should be given. On the contrary, C m a x being less than C u p indicates the solution is acceptable, and a positive reward should be given. Therefore, the modified reward function is represented as
r t ( s , a ) = E C m a x ( t ) E C m a x ( t + 1 ) if t < | T | and E C m a x ( t ) C u p , C u p C m a x × α if t = | T | and E C m a x ( t ) C u p , Z if E C m a x ( t ) > C u p .
where T = 1 , 2 , , C u p , Z > 0 , α > 0 , C u p C m a x × α denotes a non-negative reward between the upper bound of the makespan and C m a x . It encourages the agents to find good policies to reduce C m a x as much as possible.

4. Algorithm

This section describes the DQMIX algorithm optimizing the above Dec-POMDP model. We expound it from four aspects: algorithm overview, agent network, critic network, and loss function.

4.1. Algorithm Overview

DQMIX addresses the challenges of FJSP-DT scheduling through three designs.
  • We utilize a learning algorithm to satisfy the real-time and adaptive requirements of FJSP-DT scheduling. The scheduling procedure of FJSP-DT is modeled as a Dec-POMDP model, and a real-time scheduling policy that can adapt to an uncertain environment is obtained through reinforcement learning.
  • Regarding accuracy, we propose double critic networks to assist agent training, enhancing the stability and quality of learning. In addition, a MASK layer is added in the DQMIX output, which outputs a set of feasible actions based on the current state and conflicted relationships, ensuring that the output actions are accurate.
  • The MARL architecture is adopted to reduce the dimensionality of action space. Each job agent only needs to choose actions based on their own observations, and each agent model only needs to output a 1 × | A i | dimensional vector instead of n × | A i | dimensional vectors. The GCN network layer is designed to compress the number of input parameters from | V | × | V | × p to | V | × p , which further alleviates the dimensionality curse.
As shown in Figure 2, the DQMIX scheduling framework includes the FJSP-DT workshop environment and MARL networks.
The FJSP-DT workshop environment is depicted in Figure 1, consisting of stations, robots, and jobs. We view the workshop state as a graph, with stations, robots, jobs, and operations as the nodes of the graph, and the state of the nodes and edges representing the processing state. We treat the job as an agent, and it determines the station and robot required for each job.
The MARL network comprises n agent networks and two critic networks, namely the mixing network and the optimal network.
The agent networks of DQMIX refer to the agents’ policy models. The critic networks provide global guidance for the collaboration of the agents. It assists in optimizing the parameters of the agent networks during training.
The learning process of DQMIX is described as follows.
  • Firstly, an environment instance is generated according to FJSP-DT.
  • Secondly, starting at time t = 0 , each agent chooses an action a t i based on its own observation o t i and its own policy. The actions of all agents form the joint action vector a t .
  • Thirdly, according to the processing state s t and scheduling action a t at time t, the environment updates the processing state s t + 1 and observations o t + 1 at time t + 1 . The environment outputs a reward r according to the reward function, and stores the tuple ( s t , o t , a t , s t + 1 , o t + 1 ) as a POMDP instance in the experience memory.
  • Next, the agents continue to interact with the environment until all jobs’ tasks are completed, or the makespan exceeds C u p .
  • Finally, when the scheduling ends, a complete POMDP chain is formed, and the environment gives the final reward r.
The agents repeat the above steps to obtain samples and update network parameters.

4.2. Agent Network

For each agent, there is one agent network that represents its action-value function Q i ( τ i , a i ) . The agent network includes a graph convolutional network (GCN) layer, a recurrent neural network (RNN) layer, a fully connected network (FCN) layer, and a MASK layer.
The GCN layer aggregates the state features of the job graph and adjacency matrix information into a transformed state matrix.
The RNN network receives the GCN output and the action–observation history τ t 1 i at time step t and outputs a hidden vector.
The FCN layer takes the hidden vector as input and outputs an action-value vector Q t i ( o t i , τ t i , · ) of dimension | A i | .
The MASK layer is applied to constrain the agent network output to a valid action vector. The valid action vectors A i , a v a i l for agent i at each step are given by the scheduling environment, and they are represented as a vector consisting of 0 and 1, A i , a v a i l = , 0 , , 1 , T , where 1 denotes the available action and 0 denotes the unavailable action. Then, the action value vector output by the MASK layer can be represented as
Q t i , a v a i l ( o t i , τ t i , · ) = Q t i ( o t i , τ t i , · ) + ( 1 A i , a v a i l ) × M ,
where M is a large negative number (e.g., M = 1 × 10 8 ).
All agent networks share parameters for speeding up networks’ training.
Each agent utilizes the agent network to choose its action at each decision point by estimating action values for a given observation.
The action values chosen from the agent network outputs also are taken as input for the mixing network. The joint action is calculated by the argmax operator and the joint hidden feature vector τ is calculated by the sum of agent networks’ hidden features. The joint action and the joint hidden feature vector are taken as the input of the optimal network.

4.3. Critic Network

DQMIX integrates two critic networks: the mixing network and the optimal network, improving the fitting ability of QMIX [36].

4.3.1. Motivation of Proposing Double Critic Networks

QMIX is a MARL algorithm proposed in 2018 [36]. Although widely used, it may not achieve optimal solutions in solving non-monotonic problems due to its structural constraint of monotonicity [37]. To avoid the limitations of QMIX, QTRAN [37] introduces a joint action-value function to replace the mixing function in QMIX, which improves the fitting performance. However, it may not perform well in some complex multi-agent tasks [38]. WQMIX [39] adds an optimal value function to narrow the bias caused by QMIX, but the optimal value function in WQMIX contains a hyperparameter, which increases the learning uncertainty during training.
DQMIX integrates the monotonic function Q m i x while introducing a new joint action-value function Q * without any restrictions. Q * is an optimal joint action-value function of the state, the action–observation history, and the joint action. Q m i x is a mixing value function, the same as the monotonic value function in QMIX. They are represented as follows:
Q m i x = f ( s , [ Q i ] ) ,
Q * = f ( s , τ , a ) .
where τ represents the joint action–observation history and [ Q i ] represents the agent utilities.
The above design mainly lies in two considerations.
  • The non-negative monotonicity of the mixing value function should be kept when the optimal policy has been recovered, consistent with the goal pursued in cooperative games. Each agent’s marginal return with the optimal policy is non-negative, i.e.,
    Q m i x s , [ Q i ] Q i τ i , a i 0 .
  • To overcome the limitations of Q m i x when the optimal policy has not yet been recovered, a joint action-value function Q * ( s , τ , a ) is introduced. Q * ( s , τ , a ) is a function of state s, action–observation history τ and joint action a. It reduces the dependencies on agents’ utilities.
According to the above analysis, we introduce Theorem 1 [37,40] to improve QMIX and reformulate the decomposition condition of the joint action-value function as the value decomposition theorem (VDT).
The VDT says if and only if there is a joint action-value function Q * ( s , τ , a ) satisfying the conditions (16)–(19), the real action-value function Q * ( s , τ , a ) can be decomposed into a set of individual action-value functions { Q i } i N .
Q m i x s , τ , a ^ = Q * s , τ , a ^ ,
Q m i x s , τ , a Q * s , τ , a ,
Q m i x s , τ , a Q i τ i , a i 0 , i N ,
a ^ = argmax a i Q i τ i , a i i N .
The literature [40] provides detailed proof. The following conditions can be deduced by substitution based on the above conditions:
Q * s , τ , a ^ = Q m i x s , τ , a ^ Q m i x s , τ , a Q * s , τ , a .
Equation (20) satisfies the individual global maximization (IGM) principle [37].

4.3.2. Mixing Network

The mixing network is a feedforward neural network that takes the agent networks’ outputs as input and mixes them monotonically. To enforce the monotonicity constraint of (15), the weights (but not the biases) of the mixing network are restricted to be non-negative. The non-negative weights are produced by a set of hypernetworks [41] with the state as input.

4.3.3. Optimal Network

The true joint action-value function is approximated by the optimal network Q * . The optimal network is not restricted to be monotonic by using non-negative weights. Therefore, we can simplify the architecture by inputting agents’ observations (represented as a hidden feature vector τ ), the joint action a t , and the state (if necessary) s t into the network.
We construct the network above to improve scalability and sample efficiency as follows. Since the joint action space is × i N A i , it requires high complexity to find an optimal action as the number of agents n grows. The optimal joint action is sampled from all the agent networks by decentralized policies with linear-time individual argmax operations. The optimal network shares parameters from the agent networks, which enhances good sample efficiency.

4.4. Loss Function

We design the loss functions according to VDT as below.
L ( s , τ , a , r , s , τ ; θ ) = L s t a r + λ o p t L o p t + λ n o p t L n o p t .
L s t a r ( ; θ ) = Q * ( s , τ , a ) y 2 .
L o p t ( ; θ ) = Q * ( s , τ , a ^ ) Q m i x ( s , τ , a ^ ) 2 .
L o p t ( ; θ ) = Q m i x ( s , τ , a ) y 2           i f   c 1 , (24a) Q m i x ( s , τ , a ) Q c l i p ( s , τ , a ) 2           i f   c 2 , (24b) Q m i x ( s , τ , a ) Q * ( s , τ , a ) 2           i f   c 3 , (24c)
where Q * denotes the output of the optimal network; Q m i x denotes the output of the mixing network; Q c l i p denotes a clip function; a denotes the joint action vector performed by the agents; a ^ denotes the optimal joint action vector calculated by the argmax operator with the maximum value of the agent networks; c 1 denotes y > Q * ( s , τ , a ) ; c 2 denotes y Q * ( s , τ , a ) and Q * ( s , τ , a ) < Q * ( s , τ , a ^ ) ; c 3 denotes y Q * ( s , τ , a ) and Q * ( s , τ , a ) Q * ( s , τ , a ^ ) .
y = r + γ Q * ( s , τ , a ^ ; θ ) ,
a ^ = argmax a i Q i ( τ i , a i ; θ ) i = 1 n ,
a ^ = argmax a i Q i ( τ i , a i ; θ ) i = 1 n ,
where y denotes the output of the target network of the optimal network, θ denotes the parameter of the target network, and a ^ is the optimal actions maximizing the agent network output Q i ( τ i , a i ) for i N .
Q c l i p ( s , τ , a ) = clip Q m i x ( s , τ , a ) , Q * ( s , τ , a ) , Q * ( s , τ , a ^ ) .
Equation (28) is used to enforce the output of the mixing network between the evaluation action value Q * ( s , τ , a ) and the optimal action value Q * ( s , τ , a ^ ) .
Equation (22) makes the optimal network maximize the reward as far as possible, and Equation (23) ensures that the optimal network and the mixing network are optimal at the same time. The equations in (24) make two networks track each other when performing non-optimal action. If the joint action value is underestimated, Equation (24a) makes the mixing network track the true value. If the joint action value is overestimated, Equations (24b) and (24c) keep the gap between the mixing network output and the optimal network output to the extent that satisfies the necessary conditions of VDT.
The training scheme consists of two loops: the outer loop is used to complete sufficient training steps, while the inner loop is used to complete an MDP instance, i.e., an episode. The details are presented in Algorithm 1.
Algorithm 1 DQMIX Algorithm.
1:
Initialize network parameters θ , learning rate α , discount rate γ , exploration rate ϵ , replay buffer D, batch size b, the max number of episodes M, the current episode e = 0 , the max number of steps in one episode C u p , the number of jobs n. Create environment model E.
2:
while  e < M  do
3:
    Reset environment E, initialize state s t , observations o 0 = { o 0 i } i = 1 n , action–observation history        τ 1 = { τ 1 i } i = 1 n , available action set A 0 a v a i l = { A 0 i , a v a i l } i = 1 n , episode data d, the current step t = 0 ;       done = False.
4:
    while  t < C u p  do
5:
        Receive the state s t and observations o t available actions A t a v a i l from E, set the agent i = 0 .
6:
        while  i < n  do
7:
           Choose action by
a t i = randomly choose an action from A t i , a v a i l if random ( ) < ϵ ; argmax Q i , a v a i l ( o t i , τ t 1 i , · ) otherwise ;
8:
           put a t i to action vector a t , i i + 1
9:
        end while
10:
        Execute actions a t in E to get r , s t + 1 , o t + 1 , d o n e ;
11:
        Update available action set A t + 1 a v a i l ;
12:
        Store ( s t , o t , a t , A t a v a i l , r , s t + 1 , o t + 1 , A t + 1 a v a i l , d o n e ) to d;
13:
        if done then
14:
           break
15:
        end if
16:
         s t s t + 1 , o t o t + 1 , A t a v a i l A t + 1 a v a i l , t t + 1
17:
    end while
18:
    Store episode data d to replay buffer D;
19:
    if the size of D more than the least size then
20:
        Sample a batch of b episodes from D;
21:
        Compute [ Q i ] , y , Q * , Q m i x ;
22:
        Update network parameters θ by minimizing loss with loss function (21).
23:
    end if
24:
     e e + 1
25:
end while

5. Case Study

The experiments evaluate the performance of the proposed algorithm in terms of solution quality, convergence, scalability, and generalization. The time efficiency of the experiments is not measured because the well-trained DQMIX can solve problems in real time. The ablation experiments are performed to investigate the effectiveness of GCN and the reward function. The following experiments run on the computer with the GPU model of Tesla V100. Tesla V100 is manufactured by NVIDIA Corporation, an American technology company based in Santa Clara, CA, USA.

5.1. Case Description and Algorithm Settings

5.1.1. Case Description

A typical application for the FJSP-DT is aircraft maintenance scheduling. In the FJSP-DT, we consider aircraft as jobs. Therefore, we set up four typical cases, namely N3, N6, N8, and N12, representing three, six, eight, and twelve jobs, respectively. N3 is a small case, and N6, N8, and N12 are medium-to-large-sized cases. The base settings are given in Table 4. The detailed settings of N3 are given in Figure 3 and Table 5, Table 6 and Table 7.

5.1.2. Experimental Settings

To validate the effectiveness of the proposed algorithm, we selected four representative algorithms that can solve the FJSP and its extended problems, namely hGATS [42], hGAVNS [43], MGA [44], and MILP. We make adaptive modifications to the original implementation of hGATS, hGAVNS, and MGA. The initialization parameters for each algorithm are set as follows: population size is 200, crossover probability is 0.85, mutation probability is 0.15, maximum iteration is 1e4, and maximum stagnation generation is 50. Each algorithm utilizes a different random seed to run each case three times, and the average of the results from each iteration is taken as the evaluation result. The MILP model is solved using the open-source solver CBC programming. Solver CBC stands for the “constrained binary optimization” solver. It is an open-source software library used for solving optimization problems with linear and nonlinear constraints. The CBC solver is widely used in various fields such as operations research, engineering, finance, and logistics to find the optimal solution to complex problems.
To evaluate the convergence, scalability, and generalization, we utilize two well-known MARL (MARL) algorithms as baselines: centrally-weighted QMIX (CWQMIX) [39] and QTRAN [37]. CWQMIX has an unrestricted joint action-value network but takes the action values of all agents as its input, which is different from DQMIX. QTRAN also has an unrestricted joint action-value network but does not import a mixing network like DQMIX. The primary hyperparameters of the MARL algorithms are determined through several trials. For the N3 case, the algorithms converge better with a learning rate of 5 × 10 4 and a batch size of 32, but for the medium and large cases, the learning rate decreases to 5 × 10 5 , and batch size increases to 128, then the algorithms perform better. The number of training steps is 1 × 10 5 for the N3 case and 3 × 10 5 for others. The hidden layer parameters for each network are shown in Table 8. We considered the matching between the hidden and input layer dimensions and determined the settings of these parameters through multiple tests.
The reward function of all MARL models is set to
r t ( s , a ) = E C m a x ( t ) E C m a x ( t + 1 ) if t < | T | and E C m a x ( t ) C u p , C u p C m a x × 10 if t = | T | and E C m a x ( t ) C u p , C u p if E C m a x ( t ) > C u p .
where T is set to the total number of operations in the cases.

5.2. Solution Quality

The solutions of the comparative algorithms for the case of N3, N6, N8, and N12 are shown in Table 9.
As shown in Table 9, the optimal solution for the case of N3 given by MILP is 13. hGATS, hGAVNS, and MGA obtain the optimal solution. While it does not reach the optimal solution, the DQMIX algorithm converges to a good local optimum that is very close to the optimal solution. For case N8, the performance of DQMIX in terms of makespan is better than that of hGATS and MGA (shorter makespan is better). For case N12, the solution quality of DQMIX is better than that of hGATS and hGAVNS. In other cases, the solution quality of DQMIX is comparable to that of other traditional algorithms. Compared with the MARL algorithms, the average of the solutions generated by DQMIX is superior to those of CWQMIX and QTRAN.
From the R P D (29) indicators (shown in Table 10), the deviation of the average solution obtained by DQMIX from the best-known solution ( B K S ) does not exceed 10%. Moreover, as the problem size increases, the deviation becomes smaller, which to some extent indicates that DQMIX is increasingly superior in solving large-scale problems.
R P D is a common indicator in workshop schedules, represented as
R P D = u U ( m a k e s p a n u B K S ) / B K S U ,
where m a k e s p a n u is the makespan of the uth experiment, and U is the number of experimental runs.
Previous studies have demonstrated that when faced with large-scale problems that are beyond the capabilities of mathematical programming methods (such as MILP), intelligent search algorithms (e.g., GA) can still yield near-optimal solutions. While machine learning algorithms (e.g., reinforcement learning), exhibit strong real-time performance, their solution quality often requires enhancement. The findings from this section’s experiments demonstrate that the solution quality of the DQMIX model proposed is comparable to that of advanced intelligent search algorithms (hGAVNS, hGATS, and MGA). Notably, as the problem scale expands, DQMIX’s solution quality surpasses that of search-based algorithms (such as those in N12), a trend that warrants further exploration. Secondly, while DQMIX, CWQMIX, and QTRAN are all reinforcement learning algorithms, the solution quality of DQMIX surpasses that of CWQMIX and QTRAN. This superiority can be attributed to the fact that DQMIX employs a double Q-value-function optimization structure, which distinguishes it from CWQMIX and QTRAN.

5.3. Computation Time

Compared to traditional algorithms, such as genetic algorithms, machine learning models have shorter execution times and can satisfy the requirements of scenarios with high real-time requirements, such as the scenarios mentioned in this paper.
Table 11 presents the experimental data on training duration and execution duration of DQMIX model and three intelligent search algorithms (MGA, hGAVNS, hGATS) in the N8 case. According to the experimental results, the execution duration of DQMIX is in seconds, while intelligent search algorithms require more than half an hour, and hGAVNS even nearly an hour. Despite the longer training time required by DQMIX, this is somewhat compensated by the adaptability of DQMIX (see Section 5.6). A well-trained DQMIX model can adapt to changes in dynamic environments to some extent, thus meeting the needs of real-time scheduling in dynamic environments. In contrast, search algorithms must recalculate for any instance, with the same training time and execution time, being unable to satisfy the real-time requirements under the dynamic environment.

5.4. Convergence

Convergence analysis is mainly applicable to the learning algorithms DQMIX, CWQMIX, and QTRAN.
As shown in Figure 4, the three algorithms show good convergence in the case of N3. However, while QTRAN shows quick convergence, DQMIX outperforms CWQMIX and QTRAN in terms of convergence along the tail. In the case of N6, DQMIX demonstrates faster convergence compared to the other two algorithms, resulting in a shorter test makespan than QTRAN and CWQMIX. These results indicate that DQMIX is superior to the other two algorithms in terms of convergence.

5.5. Scalability

Scalability refers to the ability of models to maintain performance as the number of agents increases. The experimental results are presented in Figure 4c,d.
When the number of agents increases from 3 and 6 to 8 and 12, DQMIX exhibits stability, while QTRAN and CWQMIX show reduced stability. Additionally, DQMIX consistently produces superior solutions compared to the other two algorithms, as shown in Table 9. These findings suggest that DQMIX is more scalable and reliable, while also giving better solutions than the other MARL algorithms across varying numbers of agents.
Our ablation experiments reveal that DQMIX’s particular network benefits the learning policies. The optimal network and the mixing network of CWQMIX takes the individual action values as input, which may restrict the networks’ fitting ability and thus converge to local optima. On the other hand, QTRAIN does not have a mixing network, which may cause the network to be unstable. Therefore, compared to the first two algorithms, DQMIX maintains the mixing network and takes state and action as inputs for the optimal network, resulting in better solution quality and stability.

5.6. Generalization

During the generalization test, the algorithms that are well trained in a static environment are evaluated in a dynamic environment. The test scenarios incorporate factors such as robot breakdowns, variations in the number of jobs, and changes in operations. The well-trained models from the N8 case are employed as the test models.

5.6.1. Generalization to Robot Breakdowns

To simulate robot breakdowns, a simulator is deployed where the probability of robot failure follows the exponential distribution with parameter λ . In this study, we simulate 150 instances of robot breakdowns. The test solution is computed as an average of 150 solutions.
Table 12 presents the results of the robot breakdown generalization test for the three algorithms. Overall, the results show that all three algorithms have good generalization under robot breakdowns. The average makespan of DQMIX is significantly lower than that of the other two algorithms, which suggests that DQMIX’s solution quality is superior. However, the solving success rate of DQMIX is slightly lower than that of CWQMIX. This is because some solutions generated by DQMIX exceed the upper bound of the makespan set by the worker. In terms of stability, the M I R of QTRAN is the smallest, and it seems that QTRAN is the most stable. However, when we take into account the deviation of the makespan from the optimal solution, as captured by the N M I R metric, we observe that DQMIX performs better. In fact, DQMIX has both higher solution quality and greater stability than QTRAN when we consider N M I R .
M I R (makespan increment rate) is to measure the dynamic scheduling performance [10], as shown in Equation (31).
M I R = D y n a m i c M a k e s p a n S t a t i c M a k e s p a n S t a t i c M a k e s p a n .
N M I R (normalized M I R ), represented as Equation (32), reflects the deviation of the makespan solved under the dynamic environment from the best-known makespan.
N M I R i = M I R × S t a t i c M a k e s p a n i L B B e s t S t a t i c M a k e s p a n L B .
where B e s t S t a t i c M a k e s p a n is the best-known makespan solved by algorithms in a static scheduling environment. L B is the lower bound of the solution, which may be lower than the optimal solution.

5.6.2. Generalization to Varied Quantity of Jobs

This test investigates the algorithms’ generalization to varying job quantities. The experiments are performed on the job set of { 4 , 5 , 6 , 7 , 8 } , and the results are shown in Table 13.
According to Table 13, the three algorithms still have generalization for the varied number of jobs. Specifically, the average of the solutions of CWQMIX and DQMIX are significantly lower than QTRAN, indicating that CWQMIX and DQMIX have better generalization.

5.6.3. Generalization to Changes in Operations

This test verifies the generalization of the algorithms to different operation settings. We constructed 10 cases, each containing 8 jobs, with randomly assigned operations for each job. The test results are shown in Table 14. According to Table 14, the average solution of DQMIX is lower than QTRAN and CWQMIX.
The generalization tests conducted on the three algorithms reveal that DQMIX has the best generalization.

5.7. GCN Effectiveness Verification

Compared with GCN, we artificially design a state vector and an observation vector, called state vector representations. DQMIX and DQMIX-NO represent the model based on the GCN embedding and the state vector representations, respectively. We test the two models in the cases of N3 and N8 three times on three random seeds, and the results are shown in Figure 5.
According to Figure 5a, we observe that DQMIX-NO outperforms DQMIX in terms of convergence speed and final solution quality. However, in the N8 case (Figure 5b), DQMIX converges faster and achieves better results than DQMIX-NO. Additionally, DQMIX demonstrates greater stability in both cases.
Our experiments show that DQMIX can be as effective as DQMIX-NO in small-scale cases. However, for larger-scale cases, DQMIX shows better convergence and stability, which improves the overall training effect. Another benefit of DQMIX is that it requires no domain expertise for state modeling.

5.8. Ablation on Reward Function

The reward function based on makespan increment estimation is proposed to aid model training. To verify the effectiveness of the reward function, we make a comparison with a sparse reward function.
The sparse reward function is defined as follows.
r t ( s , a ) = 0 if t < | T | and E C m a x ( t ) C u p , Z if E C m a x ( t ) > C u p , C u p C m a x × α if t = | T | and E C m a x ( t ) C u p .
The DQMIX model with the sparse reward function is labeled as DQMIX-SPARSE.
We perform ablation with the reward functions in the cases of N3 and N8, and the results are shown in Figure 6.
Figure 6a shows that in the case of N3, DQMIX achieves better convergence and solution quality than DQMIX-SPARSE, while in the case of N8 (Figure 6b), the difference between the two is not significant in terms of convergence. It is worth noting that our training policy of sampling complete trajectories in batches and using them for learning may alleviate the negative impact of sparse feedback on model training.

6. Conclusions and Future Work

The paper introduces a complex scheduling problem called FJSP-DT, which has path and operation sequence flexibility and varied transportation time. FJSP-DT is modeled as Dec-POMDP, and GCN is applied to compress its states and observations. Mechanistic constraints are employed to exclude invalid actions and reduce the action space. We propose a MARL-based DQMIX algorithm to solve FJSP-DT. This algorithm incorporates double critic networks: the optimal network and the mixing network, enhancing its exploration and exploitation capabilities. The experimental results demonstrate that DQMIX can obtain comparable solutions to traditional algorithms. DQMIX shows high convergence and good generalization in the presence of robot breakdowns, varying quantities of jobs, and changes in operations.
Our future works will mainly focus on three aspects: (1) improving the scalability of algorithms through data and knowledge fusion-driven methods, (2) refining the reward function through hierarchical reinforcement learning, and (3) constructing a virtual–physical integrated system to support aircraft maintenance operations.

Author Contributions

Conceptualization, S.P., G.X., Z.S. and Y.H.; methodology, S.P.; software, S.P.; validation, S.P.; formal analysis, S.P.; investigation, S.P.; resources, Z.S. and Y.H.; data curation, S.P.; writing—original draft preparation, S.P.; writing—review and editing, G.X., J.Y., T.S.T., Z.T., Z.S., S.P. and Y.H.; visualization, S.P.; supervision, G.X.; project administration, F.-Y.W.; funding acquisition, G.X., Z.S. and Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2021YFB3301504, in part by the National Natural Science Foundation of China under Grant U19B2029, Grant U1909204 and Grant 92267103, in part by the Guangdong Basic and Applied Basic Research Foundation under Grant 2021B1515140034, and in part by the CAS STS Dongguan Joint Project under Grant 20211600200022 and Grant 20201600200072.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Chaudhry, I.A.; Khan, A.A. A research survey: Review of flexible job shop scheduling techniques. Int. Trans. Oper. Res. 2016, 23, 551–591. [Google Scholar] [CrossRef]
  2. Xiong, H.; Shi, S.; Ren, D.; Hu, J. A survey of job shop scheduling problem: The types and models. Comput. Oper. Res. 2022, 142, 105731. [Google Scholar] [CrossRef]
  3. Luo, Q.; Deng, Q.; Xie, G.; Gong, G. A Pareto-based two-stage evolutionary algorithm for flexible job shop scheduling problem with worker cooperation flexibility. Robot. Comput.-Integr. Manuf. 2023, 82, 102534. [Google Scholar] [CrossRef]
  4. Wei, Z.; Liao, W.; Zhang, L. Hybrid energy-efficient scheduling measures for flexible job-shop problem with variable machining speeds. Expert Syst. Appl. 2022, 197, 116785. [Google Scholar] [CrossRef]
  5. Li, Y.; Gu, W.; Yuan, M.; Tang, Y. Real-time data-driven dynamic scheduling for flexible job shop with insufficient transportation resources using hybrid deep Q network. Robot. Comput.-Integr. Manuf. 2022, 74, 102283. [Google Scholar] [CrossRef]
  6. Du, Y.; Li, J.Q.; Chen, X.L.; Duan, P.Y.; Pan, Q.K. Knowledge-based reinforcement learning and estimation of distribution algorithm for flexible job shop scheduling problem. IEEE Trans. Emerg. Top. Comput. Intell. 2023, 7, 1036–1050. [Google Scholar] [CrossRef]
  7. Burggräf, P.; Wagner, J.; Saßmannshausen, T.; Ohrndorf, D.; Subramani, K. Multi-agent-based deep reinforcement learning for dynamic flexible job shop scheduling. Procedia CIRP 2022, 112, 57–62. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Zhu, H.; Tang, D.; Zhou, T.; Gui, Y. Dynamic job shop scheduling based on deep reinforcement learning for multi-agent manufacturing systems. Robot. Comput.-Integr. Manuf. 2022, 78, 102412. [Google Scholar] [CrossRef]
  9. Wang, X.; Zhang, L.; Lin, T.; Zhao, C.; Wang, K.; Chen, Z. Solving job scheduling problems in a resource preemption environment with multi-agent reinforcement learning. Robot. Comput.-Integr. Manuf. 2022, 77, 102324. [Google Scholar] [CrossRef]
  10. Jing, X.; Yao, X.; Liu, M.; Zhou, J. Multi-agent reinforcement learning based on graph convolutional network for flexible job shop scheduling. J. Intell. Manuf. 2022, 1–19. [Google Scholar] [CrossRef]
  11. Ku, W.Y.; Beck, J.C. Mixed integer programming models for job shop scheduling: A computational analysis. Comput. Oper. Res. 2016, 73, 165–173. [Google Scholar] [CrossRef]
  12. Gao, K.; Cao, Z.; Zhang, L.; Chen, Z.; Han, Y.; Pan, Q. A review on swarm intelligence and evolutionary algorithms for solving flexible job shop scheduling problems. IEEE/CAA J. Autom. Sin. 2019, 6, 904–916. [Google Scholar] [CrossRef]
  13. Tian, Y.; Si, L.; Zhang, X.; Cheng, R.; He, C.; Tan, K.C.; Jin, Y. Evolutionary large-scale multi-objective optimization: A survey. ACM Comput. Surv. 2021, 54, 174. [Google Scholar] [CrossRef]
  14. Afshin, O.; Davood, H. A review of cooperative multi-agent deep reinforcement learning. Appl. Intell. 2023, 53, 13677–13722. [Google Scholar]
  15. Lihu, A.; Holban, S. Top five most promising algorithms in scheduling. In Proceedings of the 2009 5th International Symposium on Applied Computational Intelligence and Informatics, Timisoara, Romania, 28–29 May 2009; pp. 397–404. [Google Scholar]
  16. Wang, X.; Zhang, L.; Ren, L.; Xie, K.; Wang, K.; Ye, F.; Chen, Z. Brief review on applying reinforcement learning to job shop scheduling problems. J. Syst. Simul. 2021, 33, 2782. [Google Scholar]
  17. Liu, Y.K.; Zhang, X.S.; Zhang, L.; Tao, F.; Wang, L.H. A multi-agent architecture for scheduling in platform-based smart manufacturing systems. Front. Inf. Technol. Electron. Eng. 2019, 20, 1465–1492. [Google Scholar] [CrossRef]
  18. Zhang, W.; Dietterich, T.G. A reinforcement learning approach to job-shop scheduling. In Proceedings of the IJCAI, Citeseer, Montreal, QU, Canada, 20–25 August 1995; Volume 95, pp. 1114–1120. [Google Scholar]
  19. Aydin, M.E.; Öztemel, E. Dynamic job-shop scheduling using reinforcement learning agents. Robot. Auton. Syst. 2000, 33, 169–178. [Google Scholar] [CrossRef]
  20. Waschneck, B.; Reichstaller, A.; Belzner, L.; Altenmüller, T.; Bauernhansl, T.; Knapp, A.; Kyek, A. Optimization of global production scheduling with deep reinforcement learning. Procedia CIRP 2018, 72, 1264–1269. [Google Scholar] [CrossRef]
  21. Luo, S. Dynamic scheduling for flexible job shop with new job insertions by deep reinforcement learning. Appl. Soft Comput. 2020, 91, 106208. [Google Scholar] [CrossRef]
  22. Lang, S.; Behrendt, F.; Lanzerath, N.; Reggelin, T.; Müller, M. Integration of deep reinforcement learning and discrete-event simulation for real-time scheduling of a flexible job shop production. In Proceedings of the 2020 Winter Simulation Conference (WSC), Orlando, FL, USA, 14–18 December 2020; pp. 3057–3068. [Google Scholar]
  23. Gu, Y.; Chen, M.; Wang, L. A self-learning discrete salp swarm algorithm based on deep reinforcement learning for dynamic job shop scheduling problem. Appl. Intell. 2023, 53, 18925–18958. [Google Scholar] [CrossRef]
  24. Wang, L.; Hu, X.; Wang, Y.; Xu, S.; Ma, S.; Yang, K.; Liu, Z.; Wang, W. Dynamic job-shop scheduling in smart manufacturing using deep reinforcement learning. Comput. Netw. 2021, 190, 107969. [Google Scholar] [CrossRef]
  25. Gronauer, S.; Diepold, K. Multi-agent deep reinforcement learning: A survey. Artif. Intell. Rev. 2022, 55, 895–943. [Google Scholar] [CrossRef]
  26. Chandak, Y.; Theocharous, G.; Kostas, J.; Jordan, S.; Thomas, P. Learning action representations for reinforcement learning. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 941–950. [Google Scholar]
  27. Liu, Y.; Li, Z.; Jiang, Z.; He, Y. Prospects for multi-agent collaboration and gaming: Challenge, technology, and application. Front. Inf. Technol. Electron. Eng. 2022, 23, 1002–1009. [Google Scholar] [CrossRef]
  28. Aissani, N.; Trentesaux, D.; Beldjilali, B. Multi-agent reinforcement learning for adaptive scheduling: Application to multi-site company. IFAC Proc. Vol. 2009, 42, 1102–1107. [Google Scholar] [CrossRef]
  29. Martínez Jiménez, Y.; Coto Palacio, J.; Nowé, A. Multi-agent reinforcement learning tool for job shop scheduling problems. In Proceedings of the International Conference on Optimization and Learning, Cadiz, Spain, 17–19 February 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 3–12. [Google Scholar]
  30. Hameed, M.S.A.; Schwung, A. Reinforcement learning on job shop scheduling problems using graph networks. arXiv 2020, arXiv:2009.03836. [Google Scholar]
  31. Zhou, T.; Tang, D.; Zhu, H.; Zhang, Z. Multi-agent reinforcement learning for online scheduling in smart factories. Robot. Comput.-Integr. Manuf. 2021, 72, 102202. [Google Scholar] [CrossRef]
  32. Popper, J.; Motsch, W.; David, A.; Petzsche, T.; Ruskowski, M. Utilizing multi-agent deep reinforcement learning for flexible job shop scheduling under sustainable viewpoints. In Proceedings of the 2021 International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME), Mauritius, 7–8 October 2021; pp. 1–6. [Google Scholar]
  33. Wang, S.; Li, J.; Luo, Y. Smart scheduling for flexible and hybrid production with multi-agent deep reinforcement learning. In Proceedings of the 2021 IEEE 2nd International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 17–19 December 2021; Volume 2, pp. 288–294. [Google Scholar]
  34. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  35. Zhang, C.; Song, W.; Cao, Z.; Zhang, J.; Tan, P.S.; Chi, X. Learning to dispatch for job shop scheduling via deep reinforcement learning. Adv. Neural Inf. Process. Syst. 2020, 33, 1621–1632. [Google Scholar]
  36. Rashid, T.; Samvelyan, M.; Schroeder, C.; Farquhar, G.; Foerster, J.; Whiteson, S. QMIX: Monotonic value function factorisation for deep multi-agent reinforcement learning. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 4295–4304. [Google Scholar]
  37. Son, K.; Kim, D.; Kang, W.J.; Hostallero, D.E.; Yi, Y. QTRAN: Learning to factorize with transformation for cooperative multi-agent reinforcement learning. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 5887–5896. [Google Scholar]
  38. Mahajan, A.; Rashid, T.; Samvelyan, M.; Whiteson, S. MAVEN: Multi-agent variational exploration. Adv. Neural Inf. Process. Syst. 2019, 32, 7611–7622. [Google Scholar]
  39. Rashid, T.; Farquhar, G.; Peng, B.; Whiteson, S. Weighted QMIX: Expanding monotonic value function factorisation for deep multi-agent reinforcement learning. Adv. Neural Inf. Process. Syst. 2020, 33, 10199–10210. [Google Scholar]
  40. Son, K.; Ahn, S.; Reyes, R.D.; Shin, J.; Yi, Y. QTRAN++: Improved value transformation for cooperative multi-agent reinforcement learning. arXiv 2020, arXiv:2006.12010. [Google Scholar]
  41. Ha, D.; Dai, A.M.; Le, Q.V. HyperNetworks. In Proceedings of the 5th International Conference on Learning Representations, ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  42. Li, X.; Gao, L. An effective hybrid genetic algorithm and tabu search for flexible job shop scheduling problem. Int. J. Prod. Econ. 2016, 174, 93–110. [Google Scholar] [CrossRef]
  43. Li, X.; Gao, L.; Pan, Q.; Wan, L.; Chao, K.M. An effective hybrid genetic algorithm and variable neighborhood search for integrated process planning and scheduling in a packaging machine workshop. IEEE Trans. Syst. Man, Cybern. Syst. 2018, 49, 1933–1945. [Google Scholar] [CrossRef]
  44. Liu, Q.; Li, X.; Gao, L.; Li, Y. A modified genetic algorithm with new encoding and decoding methods for integrated process planning and scheduling problem. IEEE Trans. Cybern. 2020, 51, 4429–4438. [Google Scholar] [CrossRef]
Figure 1. Illustration of FJSP characterized by dual flexibility and varied transportation time in an uncertain environment.
Figure 1. Illustration of FJSP characterized by dual flexibility and varied transportation time in an uncertain environment.
Machines 12 00008 g001
Figure 2. The scheduling algorithm architecture of solving the FJSP-DT using the DQMIX.
Figure 2. The scheduling algorithm architecture of solving the FJSP-DT using the DQMIX.
Machines 12 00008 g002
Figure 3. Process diagram.
Figure 3. Process diagram.
Machines 12 00008 g003
Figure 4. Test performance of the MARL algorithms. T in the x-axis represents the number of training time steps, the same in the following.
Figure 4. Test performance of the MARL algorithms. T in the x-axis represents the number of training time steps, the same in the following.
Machines 12 00008 g004
Figure 5. Test performance for GCN.
Figure 5. Test performance for GCN.
Machines 12 00008 g005
Figure 6. Test performance for reward functions.
Figure 6. Test performance for reward functions.
Machines 12 00008 g006
Table 1. Literature related to job shop scheduling problems (JSP).
Table 1. Literature related to job shop scheduling problems (JSP).
TypeNumber 1
Classical JSP68
Dynamic JSP48
JSP considering the machine availability52
Flexible JSP with alternative machines (FJSP)225
JSP with alternative routings32
JSP considering batches39
JSP considering setup times69
JSP considering transportation time1
JSP with nondeterministic or non-constant processing time32
Distributed JSP (DSJSP)25
JSP with dual-resource constraints (DRJSP)30
JSP considering energy and pro-environment132
JSP with a prior job5
JSP with dependent jobs2
JSP with no-wait constraint for operations on the same job18
JSP with blocking constraint for capacities of buffer8
JSP with reentrancy9
JSP with preemption2
JSP considering overtime work2
JSP with limited buffer capacity2
JSP considering outsourcing (subcontracting)6
JSP considering robot or automated guided vehicle (AGV)28
FJSP with worker cooperation flexibility1
FJSP with technology and path flexibility (FJSP-DT)-
1 The statistics in the table refer to papers published between 2016 and 2023. Specifically, data before 2022 is referenced from [2], while the rest of the data is obtained through Google Scholar search, https://scholar.google.com/ (accessed on 10 December 2023).
Table 2. Notations.
Table 2. Notations.
SymbolDefinition
NA set of | N | jobs, N = { n 1 , n 2 , , n | N | } .
n i The ith job in the set N.
OA set of all jobs’ operations, O = { o 1 , o 2 , , o | O | } .
SA set of all stations for processing, S = { s 1 , s 2 , , s | S | } .
s k The kth station in the set S.
MA set of | M | robots, M = { m 1 , m 2 , , m | M | } .
m k The kth robot in the set M.
O i A set of operations for job n i , O i = { o i 1 , o i 2 , , o | O i | } .
o i j The jth operation of job n i .
M j A set of robots that can support operation o j , M j = { m j 1 , m j 2 , , m | M j | } .
m j k The kth robot that can support operation o j .
M l A set of robots that the station s l can cover.
m l k The kth robot that the station s l can cover.
S k A set of stations that the robot m k can reach.
s k l The lth station that the robot m k can reach.
Table 3. The description of actions.
Table 3. The description of actions.
ActionDescription
a i ( i K ) Specifying a station and a robot are available for the agent to perform the next operation.
a K + 1 Performing the transportation action. If the current station cannot support the next operation, the agent must move.
a K + 2 Staying on processing. The agent must not be interrupted to carry out other things.
a K + 3 Taking a wait action. The agent performs nothing.
a K + 4 Representing stop. At this point, the agent has completed all operations.
Table 4. Case parameters.
Table 4. Case parameters.
CaseNumber of JobsNumber of OperationsNumber of StationsNumber of Robots C up LB 1
N3312551813
N664122359054
N8855223510058
N121283223512066
1 The lower bound of the makespan.
Table 5. Alternative machines.
Table 5. Alternative machines.
OperationAlternative Robots (Processing Time)
1M1(1), M4(1)
2M3(5), M4(5), M5(5)
3M1(4), M2(4)
4M2(1), M3(1), M5(1)
Table 6. Robots to stations.
Table 6. Robots to stations.
RobotAvailable Stations
M1S1, S4, S5
M2S2
M3S2, S3, S5
M4S4
M5S2, S3, S5
Table 7. Transportation time.
Table 7. Transportation time.
StationS1S2S3S4S5
S101122
S210112
S311011
S421101
S522110
Table 8. Model parameters.
Table 8. Model parameters.
NetworkParameters
GCNOne layer, with the hidden dimension of 512
RNNTwo layers, with the hidden dimension of 512 in each layer.
Mixing NetworkTwo layers, with the hidden dimension of 512 in each layer.
Optimal NetworkThree layers, with the hidden dimension of 512 in each layer.
Table 9. Comparison of the makespan between the comparative algorithms.
Table 9. Comparison of the makespan between the comparative algorithms.
CaseDQMIXhGATShGAVNSMGACWQMIXQTRANMILP
BEST MEAN BEST MEAN BEST MEAN BEST MEAN BEST MEAN BEST MEAN
N31414131313131314.014.017.114.016.213
N66364.56163.661626263.368.083.672.084.6-
N86870.37171.6676870717375.77678.7-
N128184.78585.68585.683848185.78289.7-
Table 10. Comparison of the R P D for the comparative algorithms.
Table 10. Comparison of the R P D for the comparative algorithms.
CaseDQMIXhGATShGAVNSMGACWQMIXQTRANMILP
BEST MEAN BEST MEAN BEST MEAN BEST MEAN BEST MEAN BEST MEAN
N37.7%7.7%0.0%0.0%0.0%0.0%0.0%0.0%7.7%31.5%7.7%24.5%13
N63.3%5.7%0.0%4.3%0.0%1.6%1.6%3.8%11.5%37.0%18.0%38.7%61
N81.5%4.9%6.0%6.9%0.0%1.5%4.5%6.0%9.0%13.0%13.4%17.5%67
N120.0%4.6%4.9%5.7%4.9%5.7%2.5%3.7%0.0%8.8%1.2%10.7%81
Table 11. Training and execution time of each model in N8 case.
Table 11. Training and execution time of each model in N8 case.
ModelTraining DurationExecution Duration
DQMIX5.0 h 47.08 min4.59 s
hGATS47.77 min47.77 min
hGAVNS51.58 min51.58 min
MGA32.86 min32.86 min
Table 12. Test with robot breakdowns.
Table 12. Test with robot breakdowns.
ItemsDQMIXCWQMIXQTRAN
StaticMakespan 1697376
Number of instances 2150150150
Makespan mean 373.078.579.3
Makespan variance6.03.73.6
M I R 5.7%7.5%4.4%
N M I R 5.7%10.2%7.2%
Success rate99.3%100%98.7%
1 The makespan solved in the static environment. 2 Cases produced by random seeds on the same exponential distribution. 3 The makespans solved under robot breakdowns.
Table 13. Test with varied job quantity.
Table 13. Test with varied job quantity.
CasesNumber of JobsDQMIXCWQMIXQTRAN
N44686581
N55686679
N66706881
N77717976
N88817983
Mean-71.671.480
Variance-29.349.37
Table 14. Test with the variation in operations.
Table 14. Test with the variation in operations.
CaseDQMIXCWQMIXQTRAN
1777374
2897372
3717288
4797274
5686969
67510072
76610072
8696678
9686771
106710072
Mean72.979.274.2
Success rate100%70%100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, S.; Xiong, G.; Yang, J.; Shen, Z.; Tamir, T.S.; Tao, Z.; Han, Y.; Wang, F.-Y. Multi-Agent Reinforcement Learning for Extended Flexible Job Shop Scheduling. Machines 2024, 12, 8. https://doi.org/10.3390/machines12010008

AMA Style

Peng S, Xiong G, Yang J, Shen Z, Tamir TS, Tao Z, Han Y, Wang F-Y. Multi-Agent Reinforcement Learning for Extended Flexible Job Shop Scheduling. Machines. 2024; 12(1):8. https://doi.org/10.3390/machines12010008

Chicago/Turabian Style

Peng, Shaoming, Gang Xiong, Jing Yang, Zhen Shen, Tariku Sinshaw Tamir, Zhikun Tao, Yunjun Han, and Fei-Yue Wang. 2024. "Multi-Agent Reinforcement Learning for Extended Flexible Job Shop Scheduling" Machines 12, no. 1: 8. https://doi.org/10.3390/machines12010008

APA Style

Peng, S., Xiong, G., Yang, J., Shen, Z., Tamir, T. S., Tao, Z., Han, Y., & Wang, F. -Y. (2024). Multi-Agent Reinforcement Learning for Extended Flexible Job Shop Scheduling. Machines, 12(1), 8. https://doi.org/10.3390/machines12010008

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop