Next Article in Journal
Exact Solution for the Production Planning Problem with Several Regimes Switching over an Infinite Horizon Time
Next Article in Special Issue
Demand Prediction of Shared Bicycles Based on Graph Convolutional Network-Gated Recurrent Unit-Attention Mechanism
Previous Article in Journal
Solving the Flying Sidekick Traveling Salesman Problem by a Simulated Annealing Heuristic
Previous Article in Special Issue
Resilient Supply Chain Optimization Considering Alternative Supplier Selection and Temporary Distribution Center Location
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Q-Learning-Based Brain Storm Optimization for Integrated Distributed Flow Shop and Distribution Scheduling Problems

College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(20), 4306; https://doi.org/10.3390/math11204306
Submission received: 26 September 2023 / Revised: 12 October 2023 / Accepted: 13 October 2023 / Published: 16 October 2023

Abstract

:
In recent years, integrated production and distribution scheduling (IPDS) has become an important subject in supply chain management. However, IPDS considering distributed manufacturing environments is rarely researched. Moreover, reinforcement learning is seldom combined with metaheuristics to deal with IPDS problems. In this work, an integrated distributed flow shop and distribution scheduling problem is studied, and a mathematical model is provided. Owing to the problem’s NP-hard nature, a multi-objective Q-learning-based brain storm optimization is designed to minimize makespan and total weighted earliness and tardiness. In the presented approach, a double-string representation method is utilized, and a dynamic clustering method is developed in the clustering phase. In the generating phase, a global search strategy, a local search strategy, and a simulated annealing strategy are introduced. A Q-learning process is performed to dynamically choose the generation strategy. It consists of four actions defined as the combinations of these strategies, four states described by convergence and uniformity metrics, a reward function, and an improved ε-greedy method. In the selecting phase, a newly defined selection method is adopted. To assess the effectiveness of the proposed approach, a comparison pool consisting of four prevalent metaheuristics and a CPLEX optimizer is applied to conduct numerical experiments and statistical tests. The results suggest that the designed approach outperforms its competitors in acquiring promising solutions when handling the considered problem.

1. Introduction

Facing the competitive market environments, an increasing number of enterprises are adjusting their production and distribution activities in an integrated manner in order to quickly deliver orders to meet customer expectations. Production and distribution are two core components which have a crucial impact in driving the business performance of the supply chain [1]. Traditionally, these two components are individually organized and managed, resulting in poor efficiency from an economic and customer satisfaction point of view [2]. Furthermore, researchers have revealed that the integrated scheduling of production and distribution can reduce total operation costs by 3% to 20% [3]. Hence, integrated production and distribution scheduling (IPDS) is necessary to achieve overall optimization. In the real world, applications involving IPDS appear in many fields, e.g., newspapers, medicine, and furniture manufacturing [4]. In the IPDS process, orders are initially manufactured on machines and, subsequently, timely transported by vehicles to customers in diverse locations. In this situation, managers must make production and distribution decisions concurrently, leading to strong competitiveness and high service quality. Thus, IPDS has attracted much attention from scholars and practitioners [5].
With the development of the globalized economy and the intensive cooperation among enterprises, an emerging distributed manufacturing mode containing multiple factories is replacing the traditional centralized manufacturing mode with a single factory [6,7]. The distributed flow shop is highly utilized as a distributed manufacturing system [8], with its versatile functionality being beneficial in enhancing productivity and lowering costs across many industries, e.g., cell manufacturing, petrochemical, and automotive engines [9]. Motivated by this, distributed flow shop scheduling problems considering various constraints, including blocking, lot streaming, no wait, no idle, and limited buffers, have been addressed in the last decade [10,11,12,13,14]. Moreover, as these problems are recognized as being NP-hard, a variety of metaheuristics have been put forward as a viable means of solving them [13]. Despite the extensive efforts that have been made on distributed flow shop scheduling in terms of problem investigation and method development, few researchers have addressed its integration with distribution planning. However, their integration scheduling has far-reaching applications in reality, e.g., house customization, catering, healthcare, and e-commerce businesses. Inspired by this prospect, this work presents an integrated distributed flow shop and distribution scheduling problem.
Over the past few years, metaheuristics have become commonly used methods for tackling IPDS problems, and their search efficiency has been rigorously confirmed [5]. Nevertheless, these techniques still exhibit certain limitations, such as a lack of self-learning ability and less use of historical information. Hence, to identify more effective solutions, many investigations have integrated advanced tools, e.g., reinforcement learning (RL), into the framework of metaheuristics [15]. The combination of metaheuristics and RL allows for an adaptive parameter adjustment and a dynamic search operator selection, resulting in a significant improvement in optimization performance [16]. With its powerful capability to address optimization problems in diverse application areas (e.g., control, manufacturing, and routing), the combination of metaheuristics and RL has been increasingly recognized among researchers as a suitable approach [17,18,19]. However, it has not been investigated in the context of IPDS problems by prior studies. Accordingly, this work combines a brain storm optimization (BSO) [20] with a RL algorithm, namely, Q-learning, to construct a solution approach. Against the previous literature, threefold contributions are summarized below:
  • A multi-objective integrated distributed flow shop and distribution scheduling problem is addressed. To clearly describe it, this work formulates a mathematical model with makespan and total weighted earliness and tardiness minimization;
  • A multi-objective Q-learning-based brain storm optimization (MQBSO) is designed to handle the addressed problem. In MQBSO, a double-string representation approach is used to denote a solution, and a random method is employed to initialize the population. In the clustering phase, a dynamic clustering method is adopted to create clusters. In the generating phase, a Q-learning process is performed to guide MQBSO in choosing the generation strategy, where four actions, four states, a reward function, and an improved ε-greedy method are included. In the selecting phase, a new selection method is applied to obtain a better population;
  • To examine the performance of MQBSO, experiments are implemented on a group of instances in comparison with four metaheuristics and a CPLEX solver. The experimental results reveal that MQBSO exhibits excellent performance.
We organize the rest of this work as follows: in Section 2, the relative literature is reviewed; in Section 3, the problem under study is presented and formulated; in Section 4, BSO and Q-learning are described; in Section 5, the introduced method is given in detail; numerical experiments and statistical tests are provided in Section 6; and finally, Section 7 ends this work with some conclusions and points out research directions for the future.

2. Literature Review

In this section, three parts are included. First, we reviewed the relevant literature on IPDS problems with respect to production scheduling models, distribution models, objective functions, and solution methods. Second, we reviewed the studies involving the combination of metaheuristics with Q-learning in scheduling problems. Finally, the related studies about the BSO were reviewed.

2.1. Relevant Literature on IPDS

The research of IPDS problems can be traced back to the 1970s and was introduced by Potts [21]. This research addressed a scheduling problem with a single machine, job release time, and job ship time but not the loading capacity or vehicle routing. In recent years, a significant amount of work on IPDS under different environments has been reported [5,22]. Liu and Liu [4] explored an IPDS regarding perishable items, where multiple orders, multiple vehicles, multiple customers, and one machine were involved. They employed CPLEX and an improved large neighborhood search mechanism to minimize total weighted delivery time. Roberto and Marcelo [23] focused on an IPDS to reach total system makespan minimization. They adopted a set of equivalent parallel machines to process jobs and employed a single vehicle to perform multiple route tasks. To address this problem, they developed an iterated greedy algorithm. Jia et al. [24] studied an IPDS aiming at minimizing total weighted tardiness. In this problem, non-identical parallel batch machines were used to process jobs, and multiple vehicles were adopted to execute distribution tasks. They presented an ant colony optimization as a solution approach. In light of the advancements made in the economy and technology, manufacturing systems have become more complex. Yagmur and Kesen [25] put forward an integrated flow shop and distribution scheduling aimed at minimizing the sum of total traveling time and total tardiness. In this problem, jobs were processed on machines in a flow shop and delivered to customers in batches. They proposed a memetic algorithm to deal with it. Mohammadi et al. [26] investigated an integrated flexible job shop and distribution scheduling to minimize the sum of scheduling costs and the weighted sum of earliness and tardiness. To tackle it, a hybrid particle swarm optimization was given.
Unlike the above IPDS problems that dealt with one factory and a vehicle routing problem during production and distribution stages, respectively, research concerning distributed production environments and multi-depot vehicle routing problems have been reported. Gharaei and Jolai [27] addressed an IPDS with minimizing total tardiness and total delivery cost. This problem involved the fabrication of jobs across several non-identical factories, followed by transportation to customers via vehicles. They used a multi-objective evolutionary algorithm combined with a bee method to solve it. Fu et al. [28] introduced an integrated distributed flow shop and a vehicle routing problem with the aim of minimizing the makespan. To figure it out, they designed an enhanced black widow optimization. Additionally, they extended this IPDS problem with time window constraints [29]. To confront the problem, they presented an enhanced brain storm optimization. Qin et al. [30] put forward an IPDS in a distributed hybrid flow shop environment to reach a minimal sum of earliness, tardiness, and delivery cost. Given the NP-hard nature of the problem, they developed an adaptive human-learning-based genetic algorithm.
By analyzing the IPDS problems in the literature, we observed that they have the following characteristics:
  • Most studies considered a single factory at the production stage, while the distributed manufacturing system was not fully considered;
  • Most research concentrated on a single-objective optimization which usually involved time and cost criteria, while not enough consideration was given to multi-objective optimization;
  • Metaheuristics have become the mainstream method to cope with IPDS problems, and their outstanding performance has been verified.

2.2. Q-Learning Applied to Scheduling

Recently, RL has provided an effective approach to handling optimization problems, and has attracted increasing concern from researchers and engineers [31]. Q-learning is a typical RL method, which guides agents to decide the optimum behavior through the trial-and-error method [32]. Motivated by this, many researchers have combined it with metaheuristics to tackle production scheduling problems in recent years. Li et al. [16] applied a Q-learning algorithm to an artificial bee colony algorithm for solving a flow shop scheduling. Zhao et al. [32] put forward a hyper-heuristic combining with a Q-Learning process to deal with a distributed blocking flow shop scheduling. Li et al. [15] developed a new shuffled frog-leaping algorithm with a RL algorithm named Q-learning to work out a distributed assembly hybrid flow shop scheduling. Li et al. [33] introduced a multi-objective evolutionary algorithm based on decomposition with Q-learning for a fuzzy, flexible job shop scheduling. Wang et al. [34] devised a dual Q-learning method to handle an assembly job shop scheduling. Cheng et al. [18] designed a multi-objective Q-learning-based hyper-heuristic method to figure out a mixed shop scheduling. In the aforementioned literature, the combination of metaheuristics with Q-Learning has been studied in various production scheduling problems, e.g., flow shops, job shops, and mix shops. Nevertheless, research on IPDS in a distributed flow shop environment remains scarce. To this end, this work presents a combination of BSO and Q-Learning to address an integrated distributed flow shop and distribution scheduling.

2.3. Relevant Literature on BSO

The BSO algorithm is a promising metaheuristic method, introduced by Shi in 2011 [20]. It has many advantages, such as easy implementation, high stability, and a fine search ability. In past years, BSO and its various variants have been widely developed to settle all kinds of complex and intractable optimization problems. Xu et al. [35] introduced an improved BSO to handle a real-parameter numerical optimization problem. Cheng et al. [36] proposed a modified BSO for solving a knowledge spillover problem. Hao et al. [37] designed a hybrid BSO to tackle a distributed hybrid flow shop scheduling problem. Zhao et al. [38] developed a Q-learning-based BSO to cope with an energy-efficient distributed assembly no-wait flow shop scheduling problem. Ma et al. [39] presented a BSO method with multi-objective search mechanisms to handle a home health care scheduling and routing problem. Ke [40] devised an enhanced BSO with new convergent and divergent operations to deal with a cumulative capacitated vehicle routing problem. Many existing studies have verified its powerful exploration and exploitation abilities. However, to the best of our knowledge, it has rarely been applied to IPDS problems. Thus, this work considers it as a basic optimizer.

3. Proposed Problem and Model

3.1. Problem Description

This work delves into a multi-objective IPDS with minimizing makespan and total weighted earliness and tardiness, consisting of two stages, namely the production stage and the distribution stage. The former comprises multiple identical factories, and each one is a flow shop. A specific set of jobs must be allocated across the flow shops, and each job must follow a fixed route on the machines at their respective flow shop. After processing, the jobs need to be shipped to customers located in different geographical areas via vehicles within time windows during the distribution stage. Consequently, to handle this problem, four decisions are made simultaneously, i.e., factory assignment of jobs, job processing sequence, vehicle allocation of jobs, and the job delivery sequence.
Note that a job is an order from a unique customer. Thus, the terms “job”, “order”, and “customer” are interchangeable as they are directly related to each other on a one-to-one basis. To clarify the problem being investigated, a visual example is shown in Figure 1 along with a mathematical model. A comprehensive listing of the related parameters and variables involved in constructing this model is provided in Table 1.

3.2. Mathematical Model

For a solution to be considered viable, it must adhere to the following restrictions:
  • At time zero, all machines must be available for use;
  • Once a machine processes a job, it cannot process another one;
  • At no time can a single job be worked on by multiple machines;
  • Interrupting a job that has already started on a machine is not allowed;
  • Each customer is visited only once;
  • Vehicles must obey capacity limits.
Using the definitions provided above, we constructed the mathematical model below.
Minimize   C m a x
Minimize   T W E T
s.t.
g F k N X k j g = 1 , j N , j 0 , k j
g F j N X k j g = 1 , k N , k j
g F Y j g = 1 , j N , j 0
k N X k j g + l N X j l g 2 · Y j g , j N , j 0 , k j l , g F
g F X k j g + X j k g 1 , k , j N , k j 0
C i j C i 1 j + α i j , j N , j 0 , i M
C i j C i k + α i j + G · ( g F X k j g 1 ) , k , j N , j 0 , k j , i M
h V j N Z k j h · Y k g V , g F , j 0
h V k N Z k j h = 1 , j N , j 0
h V j N Z k j h = 1 , k N , k 0
k N Z k j h l N Z j l h = 0 , j N , h V
k N j N Z k j h · ψ j ϕ , h V , k j , j 0
j N Z 0 j h · Y j g 1 , h V , g F ,   j 0
S h C m j + G · Z k j h 1 , k , j N , k j , h V
A j S h + γ g j + G · Z 0 j h + Y j g 2 , j N , j 0 , h V , g F
A j L k + β k j + G · Z k j h 1 , k , j N , k j 0 , h V , g F
L j = A j + φ j , j N , j 0
C m a x C m j , j N , j 0
T W E T = j = 1 n E j × m a x d j A j , 0 + T j × m a x L j d j , 0
C i j 0 , i M , j N , j 0
X k j g 0 , 1 , k , j N , k j , g F
Y j g 0 , 1 , j N , j 0 , g F
Z k j h 0 , 1 , k , j N , k j , h V
A j , L j 0 , j N , j 0
The objective functions represented by Equations (1) and (2) are to minimize makespan and total weighted earliness and tardiness, respectively. The constraints laid out in Equations (3) and (4) dictate that there is precisely one preceding job and one succeeding job for each job on the machine. The constraint outlined in Equation (5) states that each job cannot be allocated to multiple factories simultaneously. Equation (6) signifies that each job owns at most one predecessor and one successor on a machine concurrently. Equation (7) mandates that there exists a strict ordering between two consecutive jobs. Equation (8) denotes that each job is worked on by only one machine at any given time. According to Equation (9), each machine must be dedicated to processing a single job at any given time. Equation (10) limits the number of vehicles. Equations (11) and (12) guarantee that no customer is visited multiple times. Equation (13) shows the constraint of maintaining continuity in vehicle routes. Equation (14) ensures that a vehicle’s maximum capacity limit is strictly observed. Equation (15) stipulates that no vehicle can be employed more than once. Equation (16) demonstrates the start time for vehicle delivery. Equations (17) and (18) provide information regarding when the vehicles arrive at customers, whereas Equation (19) determines when the vehicles depart from customers. Equations (20) and (21) define the makespan and total weighted earliness and tardiness, respectively. Equations (22)–(26) show the variable’s domain.

4. BSO and Q-Learning

4.1. BSO

Vitalized by the human inventive idea creation process, i.e., brainstorming process, Shi initially proposed a promising swarm-based metaheuristic: BSO [20]. The original BSO is straightforward in its design and can be easily implemented. Over the past years, BSO and its derivatives have achieved impressive results in tackling a range of complex problems, including real-parameter numerical optimization, distributed flow shops, and knowledge spillover problems [35,36,37,38,39,40]. Comprehensive experiments have verified that BSO possesses powerful performance to provide an outstanding compromise between exploration and exploitation abilities.
BSO begins with a population composed of multiple candidate individuals and each one means a solution to the optimization problem. Then, it performs three phases called the clustering phase, generating phase, and selecting phase to search solutions. The clustering phase involves using a clustering method to partition the population into several distinct clusters. For each cluster, the best individual in it is denoted as the center individual and the remaining are regarded as the normal ones. In the generating phase, new individuals are produced by employing one or two individuals from clusters. Each newly created individual is paired with an individual from the existing population. In the selecting phase, a selection method is adopted to store the better one from the paired individuals and save it for the next population. Finally, the three phases are iterated until a termination criterion is reached and the optimal solution is returned.

4.2. Q-Learning

Q-learning, as a typical model-free algorithm in RL, was originally introduced by Watkins and Dayan in 1992 [41]. Through a trial-and-error process, the method guides agents towards selecting the most beneficial behavior, and it comprises an environment, an agent, an action set, a state set, and a reward function. The main procedure of Q-learning is shown below. An agent selects an action a t in terms of its state s t at time t in the environment. Then, it will obtain an immediate reward r t + 1 and its state will be transferred to a new state s t + 1 . Meanwhile, the Q-table is updated according to Equation (27), where Q s t , a t stands for the Q-value of conducting an action a t at state s t , θ represents the learning rate, ϑ refers to the discount rate, and max a Q s t + 1 , a means the biggest Q-value in the Q-table at state s t + 1 .
Q s t , a t Q s t , a t + θ · r t + 1 + ϑ · max a Q s t + 1 , a Q s t , a t
The action selection is carried out in line with the Q-table. The values in an initial Q-table are equal to 0, in which the quantity of rows and columns matches the quantity of states and actions, respectively. The ϵ -greedy approach is commonly applied, and its details are described below. If a random value r v 0 , 1 is less than ϵ , then choose an action a randomly. Otherwise, the action a with a maximum Q-value is selected.

5. Presented MQBSO Algorithm

This section introduces the framework of MQBSO as follows: Step 1, initialize the population and parameters; Step 2, implement a dynamic clustering method to construct multiple clusters; Step 3, employ a Q-learning process to determine a search strategy for generating new individuals; Step 4, apply a selection approach to create the next population; and Step 5, update the population state and Q-table.

5.1. Solution Representation and Initial Population

A double-string representation method is employed to represent solutions, that is, individuals. The solution is denoted as a job scheduling sequence string π 1 , π 2 , , π n and a factory assignment string ω 1 , ω 2 , , ω n , where π j indicates a job index, j 1 , 2 , , n , π j 1 , 2 , , n , ω j means a factory index assigned to the job located at the j -th position in the job scheduling sequence string, ω j 1 , 2 , , f . To better clarify this method, Figure 2 provides a visual example using a scenario where eight jobs are assigned to three factories. In it, jobs 5 and 3 are handled in factory 1, jobs 7, 4, and 2 are assigned to factory 2, and the rest belong to factory 3.
Note that the aforementioned approach only focuses on assigning jobs to factories without obtaining other decisions, namely, the sequence in which jobs are processed on machines, allocation of vehicles for jobs, and delivery sequence of jobs. To convert the solution into an actionable plan, we present the following rules:
  • Jobs that have been allocated to a factory are processed in a sequence that matches their assigned order on the machines;
  • After the production of jobs, they are assigned in a sequential manner to a vehicle. The earlier the processing is completed, the sooner the vehicle is assigned. Meanwhile, vehicles are required not to exceed their maximum load capacity.
  • Jobs are transported in the exact sequence in which they are loaded into the vehicle.
According to the above method and rules, we can successfully transform a solution into practicable decisions.
Ensuring both the stability and diversity of MQBSO, this work adopts a random method to initialize p s feasible solutions for constructing the initial population.

5.2. Clustering

Along with the evolution of BSO, the population should be gathered into various numbers of clusters during the iteration. Thus, this work develops a dynamic clustering approach driven by a nondominated sorting method [42]. This work sorts all individuals in the population and assigns them sorted values based on their dominated relationships. The smaller the sorted value, the better the individual. Hence, this work sets the best individuals with the smallest sorted value of 1 as the center individuals, while the rest as normal individuals. For each center individual, this work constructs a cluster. To conduct such work, the normal ones are randomly assigned to these clusters. Notice that when only one individual has the sorted value of 1, this work selects the individuals with the sorted value of 2 as center individuals.

5.3. Generating

Generating is the core process of BSO, which controls the global search and local search abilities of an algorithm through applying one individual or two individuals in clusters. In this process, three vital parameters are used, i.e., r g , r o , r t . Let r g decide whether a new individual is constructed by one individual or two individuals. r o determines whether a new individual is created by utilizing one center individual or one normal individual. r t decides whether a new individual is produced via employing two center individuals or two normal ones. Based on them, this work introduces a global search strategy and a local search strategy. The details about them are provided as follows:
  • A global search strategy is designed when utilizing two randomly chosen center or normal individuals to construct new individuals, where the sequence-based [43] and the two-point crossover [44] methods are adopted. The former is applied to update the job scheduling sequence string and the latter is utilized to update the factory assignment string. The main contents of these two methods are described below.
  • Sequence-based crossover method: First, randomly have two job position indexes p 1 , p 2 . Then, jobs between them in the individual x 1 are copied to the same positions of a new individual x . Finally, the missing jobs in x are added as their appearance in the individual x 2 . Hence, the job scheduling sequence string of x is obtained. Two-point crossover method: First, two factory position indexes p 3 , p 4 are generated at random. Then, extract factory assignments beyond p 3 and p 4 in x 1 and place them into the same positions of x . Finally, the rest factory assignments in x are filled with the elements at the same positions of x 2 . Thus, the factory assignment string of x is acquired. By employing these two methods, a new individual is successfully produced. To clearly exhibit the generation process, Figure 3 depicts an illustrative example.
  • A local search strategy is developed when using one randomly selected center or normal individual to construct new individuals, where five kinds of neighborhood structures, named NS1, NS2, NS3, NS4, and NS5 are introduced. The main idea of NS1 is as follows: randomly produce two job positions for a chosen individual, then swap these two jobs and exchange their respective factory assignments.
Different from NS1, the problem-dependent information is considered for the final four neighborhood structures. Given that one of the primary objectives of the problem investigated is to optimize the makespan criterion, it is essential to focus on improving the performance of the key factory that has a maximum completion time. To reach this, it is vital to adjust both the factory assignment and job processing sequence associated with this factory. According to the key factory theory [45] and the characteristics of the presented problem, this work designs four problem-dependent neighborhood structures. The details about them are given as follows: NS2 randomly swaps two jobs in the key factory; NS3 employs an insert operator, where two jobs in the key factory are randomly chosen, and one of which is inserted into the front of the other, along with corresponding changes are made to their factory assignments; NS4 is similar to NS2 except that the factory assignment of the two jobs is different, and one must be the key factory; and NS5 changes the factory assignment of a job in the key factory.
It is worth mentioning that for each chosen center or normal individual, only one of five neighborhood structures is selected at random to produce a new individual. This work stores the better one between the chosen and new individuals in accordance with their dominated relationship. To intuitively understand the five neighborhood structures, Figure 4 shows five examples of these structures, and the key factory is denoted as “*”.
Furthermore, to avoid premature convergence and enhance the local search ability, we adopt a simulated annealing strategy [46]. The steps of this strategy are outlined below.
Step 1: Set an initial temperature T 0 = 0.5 · i = 1 m j = 1 n α i j / 10 · n · m , a final temperature T m i n = 0.5 · T 0 , and update the current temperature T : = T 0 .
Step 2: Select a center or normal individual x at random and set an intermediary individual x 3 : = x .
Step 3: Randomly perform one of the five neighborhood structures on x 3 to generate a new individual x .
Step 4: If x dominates x 3 , x 3 : = x ; otherwise, x 3 : = x 3 .
Step 5: Adjust the temperature according to the formula: T : = T T · 0.1 , and repeat Steps 3–5 until T T m i n .
Step 6: Generate the new individual x : = x 3 and save it.
In the generating phase, Q-learning is implemented to assist MQBSO in choosing a generation strategy, the introduction of this process is described in Section 5.5.

5.4. Selecting

In the original BSO, the selecting phase is performed to store better individuals to enter the next population. It usually saves the better ones between new individuals and their connected individuals in the population. This method may discard some better individuals. In order to strengthen MQBSO’s capability to explore promising directions, this work employs an effective selection method [44] by utilizing the nondominated sorting and crowding distance approaches. The basic idea of this method is shown as follows: First, merge the existing population with new individuals; second, sort all individuals and calculate their crowding distance values; and finally, extract the best p s individuals as the next population based on their sorted and crowding distance values.

5.5. Q-Learning Process

In this work, Q-learning is composed of four actions, four states, a reward function, and an action selection method. Four actions are defined as the combinations of a global search strategy, a local search strategy, and a simulated annealing strategy. Four states are described by the convergence and uniformity metrics. A reward function is designed based on the states and an improved ε -greedy method is newly developed.

5.5.1. Action Set

The action set consists of four actions, and each one is a combination of a global search strategy, a local search strategy, and a simulated annealing strategy. Action a 1 contains a global search strategy. Action a 2 includes a local search strategy. Action a 3 is composed of a global search strategy and a local search strategy. Action a 4 is made up of a global search strategy and a simulated annealing strategy. According to the above designs, Figure 5 gives the framework of an action set in MQBSO, where r n denotes a random number between 0 , 1 .
This work uses an improved ε -greedy method [32] to conduct the selection of an action, where ε is calculated by Equation (28). In Equation (28), ρ c denotes the current number of fitness function evaluations, and ρ m a x represents a stopping condition equaling to 150 · n · m fitness function evaluations. Randomly generate a float number λ 0 , 1 , if λ is bigger than 1 ε , one of four actions is chosen at random; otherwise, the agent selects the action with the highest Q-value as its preferred option. This method ensures that the selection probability of the two cases is about half in the beginning, along with the agent maintaining a certain ability to explore during the early stage. As the algorithm evolves, the action with the highest Q-value is favored. Thus, MQBSO can effectively intensify the exploration ability and reduce the possibility of dropping into local optima.
ε = 0.5 / 1 + e 10 · ρ c 0.6 · ρ max / ρ max ,

5.5.2. State Set

To clearly depict the state, a convergence index (C-metric) [47] and a uniformity index (D-metric) [33] are used, which are computed based on the formulas defined below.
C U , W = w W | u U : u w / W
As per Equation (29), U and W refer to the sets of non-dominated solutions acquired in the μ -th and μ 1 -th iterations. C ( U , W ) quantifies the percentage of solutions in set W which are dominated by solutions within set U , and W stands for the size of W . Generally, a bigger C U , W suggests that the performance of set U is better.
D = e = 1 E z e z ¯ / E · z ¯ ,
z ¯ = e = 1 E z e / E ,
As per Equation (30), E is a non-dominated solution set and E indicates its size. z e refers to the shortest Euclidean distance from the e -th solution and any other solution in E , and z ¯ means the average distance as measured using the formula presented in Equation (31). Usually, a higher D-metric value implies a better performance of uniformity.
D ( U , W ) = D U D W
The occurrence percentage of C ( U , W ) > 0 , C ( U , W ) 0 , D ( U , W ) > 0 , and D ( U , W ) 0 in the entire search process of MQBSO on four instances are displayed in Figure 6. We see that each of the four cases appear. As a result, it is rational to consider four combinations of C ( U , W ) and D ( U , W ) as four states, i.e., s ( 1 ) : C ( U , W ) > 0 and D ( U , W ) > 0 , s ( 2 ) : C ( U , W ) > 0 and D ( U , W ) 0 , s ( 3 ) : C ( U , W ) 0 and D ( U , W ) > 0 , and s ( 4 ) : C ( U , W ) 0 and D ( U , W ) 0 .

5.5.3. Reward

In this work, four states are labeled as s 1 , s 2 , s 3 , and s 4 . The bigger the label value of the state, the worse the state. s 1 means the best state while s 4 is the worst one. Thus, rewards are set based on the label value of states. Four reward values r v are 5, 3, 3, and 1 for these four states, respectively.
After a detailed discussion, Algorithm 1 provides the main procedure of MQBSO.
Algorithm 1: MQBSO
Input: Population size p s , three generation parameters r g , r o , r t , learning rate θ , discount rate ϑ , and a stopping condition ρ m a x .
Output: A non-dominated solution set.
Initialize the population and parameters (c.f. Section 5.1)
While the stopping condition is not reached do
Implement a dynamic clustering method in the clustering phase (c.f. Section 5.2).
Employ a Q-learning process in the generating phase (c.f. Section 5.3 and Section 5.5).
Apply a selection method in the selecting phase (c.f. Section 5.4).
Update the population state and Q-table.
End while

6. Experimental Evaluations

In this section, comparison experiments and statistical tests are conducted to study MQBSO’s performance. MQBSO and comparison algorithms are programmed in C++ and performed on a personal computer equipped with HUAWEI Core i5-8265U CPU and 8 GB RAM in Qingdao, China. Four comparison methods are well-known metaheuristics: multi-objective whale swarm algorithm (MOWSA) [43], multi-objective brain storm optimization (MOBSO) [44], nondominated sorting genetic algorithm II (NSGA-II) [42], and multi-objective evolutionary algorithm based on decomposition (MOEA/D) [48]. In fairness, this work sets 150 · n · m fitness function evaluations as a termination criterion, and all methods independently solve each instance 20 times.

6.1. Test Instances

Since the previous work does not address the presented problem, we cannot find available benchmarks. Therefore, we produce a set of instances based on the problems’ features. The benchmarks regarding flow shop scheduling and vehicle routing problems are employed to construct instances, i.e., the VFR benchmark [49] for the former, and the Gehring & Homberger benchmark [50] for the latter.
In the VFR benchmark, “VFRa_b_c” means the c -th instance of a test problem having a jobs and b machines. Four instances are adopted, including VFR30_5_1, VFR30_10_1, VFR60_5_1, and VFR60_10_1, respectively. The instances VFR90_5_1, VFR90_10_1, VFR120_5_1, and VFR120_10_1 do not exist in the VFR benchmark. Thus, this work creates them by applying the information of the first 90, 120 jobs and the first 5, 10 machines regarding the VFR100_20_1 and VFR200_20_1.
In the Gehring & Homberger benchmark, an instance named C1_2_1 is applied, where coordinates, time windows, and service times are provided. Through choosing the first 30, 60, 90, and 120 customers from it, four instances, namely, C1_2_1, C1_2_1, C1_2_1, and C1_2_1 are obtained. The chosen instances contain eight combinations of n × m : 30,60,90,120 × ( 5,10 ) , where n and m indicate the number of jobs and machines, respectively. This work sets the number of factories f 2,3 , 4 . Hence, a total of 24 test instances are successfully generated.
For convenience, Table 2 provides abbreviations for the 24 test instances. In it, the first instance called “2-5-30” consists of 2 factories, 5 machines, and 30 jobs, in which the production and distribution data are from VFR30_5_1 and C1_2_1, respectively. The remaining can be deduced by analogy. Furthermore, this work sets the load of jobs from an interval [10,40], the maximum load capacity of vehicles is 100, 125, 150, and 200 when the quantity of jobs is 30, 60, 90, and 120, respectively, and the unit earliness and tardiness weights of jobs are generated at random from 0.1,0.2,0.3,0.4,0.5 .

6.2. Parameters Setting

An orthogonal experiment method [51] is employed in this research to implement the parameter experiment on a moderate-size instance named “3-5-60” with 3 factories, 5 machines and 60 jobs. This method is viewed as a modern technique for analyzing and optimizing parameters across various research fields. Generally, practical engineering design and optimization tasks require the consideration of three or more influential factors, leading to the application of full factorial design analysis. MQBSO considers four parameters, i.e., p s , r g , r o , and r t . Each parameter has four levels: p s { 20,40,60,80 } , r g { 0.20,0.40,0.60,0.80 } , r o { 0.20,0.40,0.60,0.80 } , r t { 0.20,0.40,0.60,0.80 } , and an array L 16 4 4 containing 16 combinations is created. To promote fairness in the comparison of different algorithms, this work sets the same number of fitness function evaluations as a stopping condition, i.e., 150 · n · m , where n and m are the quantity of jobs and machines, respectively. Each combination is executed 20 times individually and the results are analyzed by applying the IGD-metric. The average IGD (AIGD) results are calculated and provided in Table 3. Through analyzing the obtained results, a hopeful parameter combination is produced, i.e., p s = 40 , r g = 0.40 , r o = 0.20 , r t = 0.80 .
Similarly, the parameters of four peers are achieved via using the same method, and their parameters are displayed in Table 4. In addition, the learning rate θ and discount rate ϑ are set by drawing on existing research [32], which are 0.5 and 0.8, respectively.

6.3. Performance Metrics

Convergence and uniformity are two significant measures for examining the efficiency of multi-objective algorithms. Hence, this work utilizes three common performance metrics to carry out algorithmic measurements, i.e., C-metric [47], Inverted Generational Distance-metric (IGD-metric) [47], and Hypervolume-metric (HV-metric) [52]. The C-metric is defined in Equation (29), and the details of the IGD- and HV-metrics are offered below.
The IGD-metric is applied to reflect both convergence and uniformity of a method and its calculation formula is shown as:
I G D P * , P = p P * d i s t p , P / P *
where P and P * mean an obtained solution set by an algorithm and the Pareto optimal front, respectively. In reality, the Pareto optimal front is difficult to achieve for the examined problem. Hence, this work generates P * by merging the obtained solution sets of all employed algorithms and discarding the dominated solutions. d i s t p , P represents the shortest Euclid distance from a solution p in P * and any other solution in P , and P * is the quantity of solutions in P * . In this work, all objective values are normalized within 0,1 , and an algorithm with lower IGD-metric values is ideal.
The HV-metric is also used to account for the convergence and uniformity of an algorithm. It estimates the volume of the area covered by the obtained nondominated solutions of an algorithm and a known reference point H * . It is calculated as:
H V = δ o = 1 O s v o
where the Lebesgue measure δ is implemented to determine the volume of a set. O denotes the quantity of solutions in the solution set O , and s v o is the super volume of the area coved by the o -th solution in O and P * . As in the IGD-metric, the obtained objective values of all algorithms are transformed into the range of 0,1 , and H * is 1,1 . Generally, a larger HV-metric value shows a better performance.

6.4. Experimental Results and Analysis

This section compares MQBSO with its four rivals on the C-metric, IGD-metric, and HV-metric. The four peers are, respectively, MOWSA, MOBSO, NSGA-II, and MOEA/D, and the results are presented in Table 5, Table 6 and Table 7.
Table 5 exhibits the comparison results concerning C-metric, where the symbols “QB”, “WS”, “BS”, “GA”, and “EA” indicate the nondominated solution sets obtained by MQBSO, MOWSA, MOBSO, NSGA-II, and MOEA/D, respectively. A higher value of C Q B , R suggests a superior performance of MQBSO, R W S , B S , G A , E A . As shown in Table 5, the values of C Q B , W S and C Q B , B S are bigger than those of C W S , Q B and C B S , Q B in 22 instances, along with the values of C Q B , G A and C Q B , E A which are larger than those of C G A , Q B and C E A , Q B in 21 instances. Thus, MQBSO is significantly better than its peer methods. Furthermore, we find that MQBSO obtains better average results than MOWSA, MOBSO, NSGA-II, and MOEA/D. In light of the comparative analysis results, it is evident that MQBSO is a valid method.
Table 6 reports the comparison results with respect to the IGD-metric. Usually, a lower IGD-metric value means a better performance of an algorithm. As displayed in Table 6, MQBSO produces the best results in 22 out of 24 test instances compared with MOWSA, MOBSO, NSGA-II, and MOEA/D. Furthermore, the average values of all instances are shown at the bottom of Table 6. The average values of MQBSO, MOWSA, MOBSO, NSGA-II, and MOEA/D are 0.1077, 0.3497, 0.3714, 0.4178, and 0.4655, respectively. Clearly, MQBSO achieves a minimum average result. Thereby, we verify that MQBSO can obtain more promising solutions than its competitors in settling the studied problem.
Table 7 provides the comparison results regarding the HV-metric, where an algorithm with higher HV-metric values exhibits better than those with lower values. The results demonstrated in Table 7 ensure that the solutions obtained by MQBSO can cover a large volume on most test instances, meaning that MQBSO is superior. MQBSO is better than MOWSA and MOBSO in 22 instances, and it performs better than NSGA-II and MOEA/D in 20 out of 24 test instances. Thereby, we verify that MQBSO is competent at generating better solutions. In addition, by analyzing the average values acquired by MQBSO, MOWSA, MOBSO, NSGA-II, and MOEA/D, we draw a similar conclusion, the average values are, respectively, 0.8440, 0.5340, 0.4696, 0.4269, and 0.3785. Based on the ongoing study, we declare that MQBSO is an outstanding solver.
To clearly illustrate the experimental findings of MQBSO and other four rivals, boxplots are drawn to visualize the performance of all algorithms across nine instances with varying sizes. Figure 7 and Figure 8, respectively, display the boxplots of MQBSO and its competitors with respect to IGD- and HV-metric. According to Figure 7, we find that MQBSO achieves a significantly lower median value than its rivals across all selected instances concerning IGD-metric. Similarly, as can be observed from Figure 8, MQBSO owns a greater median value than its rivals in all the chosen instances in terms of the HV-metric. Therefore, the collected experimental results and subsequent analysis state that MQBSO has an obvious advantage over its rivals in handling the problem under study.
For a further comparison, Figure 9 shows the distribution graphs of the nondominated solutions achieved by MQBSO and its peers in settling six instances with different sizes. From Figure 9, it is apparent that the nondominated solutions obtained by MQBSO are consistently much better than its rivals on the chosen test instances. Specifically, we see that MQBSO produces uniformly distributed and better approximated solutions. Hence, MQBSO can be taken as an outstanding optimizer.
Moreover, this work applies statistical tests to verify whether the performance difference between MQBSO and its peers is significantly different. The Friedman and Wilcoxon signed rank tests [53,54,55] are carried out on each instance of the IGD- and HV-metrics at α = 0.05 level of significance. The average rank values of MQBSO, MOWSA, MOBSO, NSGA-II, and MOEA/D with respect to the IGD-metric are 1.3333, 2.5417, 2.7083, 3.7917, and 4.6250, respectively, and those of the five algorithms regarding the HV-metric are 1.5000, 2.1667, 3.1667, 3.7500, and 4.4167, respectively. Obviously, MQBSO obtains a minimum average rank value on the two metrics. Furthermore, the test statistic indices T F regarding the IGD- and HV-metrics are calculated, i.e., 39.8483 and 8.0714. These values surpass the critical value of 2.53. Accordingly, the performance difference among all algorithms is statistically significant. The statistical results of the Wilcoxon signed rank test are displayed in Table 8, with respect to the IGD- and HV-metrics. The acquired R + values are greater than the R values, and the obtained p -values are lower than 0.05. As a consequence, we state that MQBSO significantly outperforms its peers.

6.5. Effectiveness of the Q-Learning Process in MQBSO

To demonstrate the effect of the Q-learning process in MQBSO, a variant without using it (named MBSO) is constructed. In MBSO, the generation strategy is selected at random to produce new individuals. Table 9 displays the comparison results of MQBSO and MBSO regarding the C-, IGD-, and HV-metrics. MQBSO acquires better results than MBSO on 21 instances of the C- and HV-metrics and performs better than MBSO in 20 instances with respect to the IGD-metric. In addition, MQBSO has an advantage over MBSO in terms of the average value for three metrics. Thus, the Q-learning process is effective in MQBSO when solving the investigated problem.

6.6. Comparison of Five Algorithms in CPU Time

To examine the computation efficiency of MQBSO, this work records the CPU time of MQBSO and its competitors in handling all the employed test instances. The instances are organized into groups based on their job quantity, and the average CPU time is computed for each group, as provided in Table 10. The average CPU time of MOWSA and MOBSO is longer than that of MQBSO in all groups. Thus, MQBSO has better abilities to tackle the considered problem by utilizing less time. The average CPU time of NSGA-II and MOEA/D exhibits a slight advantage over MQBSO in all groups. However, the difference between MQBSO and NSGA-II, along with MQBSO and MOEA/D is not very large. Thus, MQBSO is still regarded as an outstanding solver.

6.7. Comparison of MQBSO and CPLEX

To verify MQBSO’s performance for seeking optimal solutions, CPLEX is used for comparisons. The experiments are conducted as follows: this work develops three instances based on the first 8 jobs, the first 10 jobs, and all jobs of a test problem with 12 jobs, respectively. Table 11 and Table 12 provide the information of the used test problem.
Considering that CPLEX cannot cope with a multi-objective optimization problem, we employ the weighted sum method to develop a variant of the problem under study. For each instance, three weight vectors are used, i.e., (1.0, 0.0), (0.5, 0.5), and (0.0, 1.0), to construct three sub-problems. There are two factories, and each factory contains two machines and two vehicles. The maximum load capacity of vehicles is 60, 80, and 100 when the number of jobs is 8, 10, and 12, respectively. The service time of customers is 90. This work sets the maximum running time of CPLEX as 3600 s. In scenarios where CPLEX fails to identify the global optimal solution in the allowed time, it outputs an approximate optimal value (AOV) instead. Table 13 shows the comparison results.
From Table 13, we can see that CPLEX achieves an optimal solution on two sub-problems and obtains the same value as MQBSO on one sub-problem when solving the instance with 8 jobs. On the contrary, MQBSO shows better than CPLEX on two sub-problems for solving the instances with 10 and 12 jobs, and it acquires the same value as CPLEX on one sub-problem. Moreover, the running time of MQBSO is significantly shorter than CPLEX. Thus, we can infer that MQBSO exhibits clear advantages over CPLEX in handling the proposed problem if the problem scale is larger.
Through analyzing the achieved experimental and statistical results in the previous sections, we verify that MQBSO has a powerful ability to gain better results in handling the problem under consideration. In MQBSO, a dynamic clustering method is designed in the clustering phase. A global search strategy, a local search strategy, and a simulated annealing strategy are used in the generating phase. A Q-learning process is conducted to dynamically choose the generation strategy. It includes four actions defined as the combinations of these strategies, four states described by convergence and uniformity metrics, a reward function, and an improved ε -greedy method. A selection method is employed in the selecting phase. By employing these methods iteratively, MQBSO has a superior trade-off between its abilities to explore and exploit the solution space, thus rendering it an outstanding solver for the problem being examined.

7. Conclusions

This work, for the first time, addresses a multi-objective integrated distributed flow shop and distribution scheduling to minimize makespan and total weighted earliness and tardiness. A mathematical model is given to define it and an MQBSO is developed to cope with it. In MQBSO, a double-string representation method is applied, and a dynamic clustering method is designed. Q-learning is applied to choose the generation strategy, where four actions, four states, a reward function, and an improved ε -greedy approach are designed. To assess the performance of MQBSO, numerical experiments are implemented on test instances by comparing with MOWSA, MOBSO, NSGA-II, MOEA/D, and CPLEX. The results suggest that MQBSO is superior.
The future directions are pointed out, which are encouraged to expand our study in several aspects, e.g., inventory, multiple time windows, stochastic models, flexible distributed scheduling, and energy-related objectives [56,57]. In addition, we will focus on RL to design effective solution methods.

Author Contributions

Conceptualization, J.X. and S.Z.; methodology, S.Z.; software, S.Z. and Y.Q.; validation, S.Z. and Y.Q.; formal analysis, S.Z. and Y.Q.; investigation, S.Z.; resources, J.X.; data curation, Y.Q.; writing—original draft preparation, S.Z. and Y.Q.; writing—review and editing, S.Z., J.X. and Y.Q.; visualization, S.Z.; supervision, J.X.; project administration, J.X.; funding acquisition, J.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part, by the National Natural Science Foundation of China: grant number 72271048.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rafiei, H.; Safaei, F.; Rabbani, M. Integrated production-distribution planning problem in a competition-based four-echelon supply chain. Comput. Ind. Eng. 2018, 119, 85–99. [Google Scholar] [CrossRef]
  2. Ganji, M.; Kazemipoor, H.; Molana, S.M.H.; Sajadi, S.M. A green multi-objective integrated scheduling of production and distribution with heterogeneous fleet vehicle routing and time windows. J. Clean. Prod. 2020, 259, 120824. [Google Scholar] [CrossRef]
  3. Chandra, P.; Fisher, M.L. Coordination of production and distribution planning. Eur. J. Oper. Res. 1994, 72, 503–517. [Google Scholar] [CrossRef]
  4. Liu, L.; Liu, S. Integrated production and distribution problem of perishable products with a minimum total order weighted delivery time. Mathematics 2020, 8, 146. [Google Scholar] [CrossRef]
  5. Moons, S.; Ramaekers, K.; Caris, A.; Arda, Y. Integrating production scheduling and vehicle routing decisions at the operational decision level: A review and discussion. Comput. Ind. Eng. 2017, 104, 224–245. [Google Scholar] [CrossRef]
  6. Wang, J.J.; Wang, L. A bi-population cooperative memetic algorithm for distributed hybrid flow-shop scheduling. IEEE Trans. Emerg. Top. Comput. Intell. 2020, 5, 947–961. [Google Scholar] [CrossRef]
  7. Shao, W.S.; Shao, Z.S.; Pi, D.C. Modeling and multi-neighborhood iterated greedy algorithm for distributed hybrid flow shop scheduling problem. Knowl.-Based Syst. 2020, 194, 105527. [Google Scholar] [CrossRef]
  8. Wang, J.J.; Wang, L. A knowledge-based cooperative algorithm for energy-efficient scheduling of distributed flow-shop. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 1805–1819. [Google Scholar] [CrossRef]
  9. Lu, C.; Gao, L.; Gong, W.Y.; Hu, C.Y.; Yan, X.S.; Li, X.Y. Sustainable scheduling of distributed permutation flow-shop with non-identical factory using a knowledge-based multi-objective memetic optimization algorithm. Swarm Evol. Comput. 2021, 60, 100803. [Google Scholar] [CrossRef]
  10. Shao, W.S.; Pi, D.C.; Shao, Z.S. Optimization of makespan for the distributed no-wait flow shop scheduling problem with iterated greedy algorithms. Knowl.-Based Syst. 2017, 137, 163–181. [Google Scholar] [CrossRef]
  11. Gong, D.W.; Han, Y.Y.; Sun, J.Y. A novel hybrid multi-objective artificial bee colony algorithm for blocking lot-streaming flow shop scheduling problems. Knowl.-Based Syst. 2018, 148, 115–130. [Google Scholar] [CrossRef]
  12. Zheng, J.; Wang, L.; Wang, J.J. A cooperative coevolution algorithm for multi-objective fuzzy distributed hybrid flow shop. Knowl.-Based Syst. 2020, 194, 105536. [Google Scholar] [CrossRef]
  13. Fu, Y.P.; Hou, Y.S.; Wang, Z.F.; Wu, X.W.; Gao, K.Z.; Wang, L. A review of distributed scheduling problems in intelligent manufacturing systems. Tsinghua Sci. Technol. 2021, 26, 625–645. [Google Scholar] [CrossRef]
  14. Zhao, F.Q.; Shao, D.Q.; Wang, L.; Xu, T.P.; Zhu, N.N.; Jonrinal, D. An effective water wave optimization algorithm with problem-specific knowledge for the distributed assembly blocking flow-shop scheduling problem. Knowl.-Based Syst. 2022, 243, 108471. [Google Scholar] [CrossRef]
  15. Li, H.X.; Gao, K.Z.; Duan, P.Y.; Li, J.Q.; Zhang, L. A novel shuffled frog-leaping algorithm with reinforcement learning for distributed assembly hybrid flow shop scheduling. Int. J. Prod. Res. 2022, 61, 1233–1251. [Google Scholar] [CrossRef]
  16. Li, H.X.; Gao, K.Z.; Duan, P.Y.; Li, J.Q.; Zhang, L. An improved artificial bee colony algorithm with Q-learning for solving permutation flow-shop scheduling problems. IEEE Trans. Syst. Man Cybern. Syst. 2022, 53, 2684–2693. [Google Scholar] [CrossRef]
  17. Luo, B.; Liu, D.; Huang, T.; Wang, D. Model-free optimal tracking control via critic-only Q-learning. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 2734–2744. [Google Scholar] [CrossRef] [PubMed]
  18. Cheng, L.X.; Tang, L.X.; Zhang, L.P.; Zhang, Z.K. Multi-objective Q-learning-based hyper-heuristic with Bi-criteria selection for energy-aware mixed shop scheduling. Swarm Evol. Comput. 2022, 69, 100985. [Google Scholar] [CrossRef]
  19. Bdeir, A.; Boeder, S.; Dernedde, T.; Tkachuk, K.; Falkner, J.K.; Schmidt-Thieme, L. RP-DQN: An application of Q-learning to vehicle routing problems. Adv. Artif. Intell. 2021, 12873, 3–16. [Google Scholar] [CrossRef]
  20. Shi, Y.H. Brain storm optimization algorithm. In International Conference in Swarm Intelligence; Springer: Berlin/Heidelberg, Germany, 2011; pp. 303–309. [Google Scholar]
  21. Potts, C.N. Analysis of a heuristic for one machine sequencing with release dates and delivery times. Oper. Res. 1980, 28, 1436–1441. [Google Scholar] [CrossRef]
  22. Chen, Z.L. Integrated production and outbound distribution scheduling: Review and extensions. Oper. Res. 2010, 58, 120–148. [Google Scholar] [CrossRef]
  23. Roberto, F.T.N.; Marcelo, S.N. An iterated greedy approach to integrate production by multiple parallel machines and distribution by a single capacitated vehicle. Swarm Evol. Comput. 2019, 44, 612–621. [Google Scholar] [CrossRef]
  24. Jia, Z.H.; Cui, Y.F.; Li, K. An ant colony-based algorithm for integrated scheduling on batch machines with non-identical capacities. Appl. Intell. 2022, 52, 1752–1769. [Google Scholar] [CrossRef]
  25. Yagmur, E.; Kesen, S.E. A memetic algorithm for joint production and distribution scheduling with due dates. Comput. Ind. Eng. 2020, 142, 106342. [Google Scholar] [CrossRef]
  26. Mohammadi, S.; Al-e-Hashem, S.M.; Rekik, Y. An integrated production scheduling and delivery route planning with multi-purpose machines: A case study from a furniture manufacturing company. Int. J. Prod. Econ. 2020, 219, 347–359. [Google Scholar] [CrossRef]
  27. Gharaei, A.; Jolai, F. A multi-agent approach to the integrated production scheduling and distribution problem in multi-factory supply chain. Appl. Soft Comput. 2018, 65, 577–589. [Google Scholar] [CrossRef]
  28. Fu, Y.P.; Hou, Y.S.; Chen, Z.H.; Pu, X.J.; Gao, K.Z.; Sadollah, A. Modelling and scheduling integration of distributed production and distribution problems via black widow optimization. Swarm Evol. Comput. 2021, 68, 101015. [Google Scholar] [CrossRef]
  29. Hou, Y.S.; Fu, Y.P.; Gao, K.Z.; Zhang, H.; Sadollah, A. Modelling and optimization of integrated distributed flow shop scheduling and distribution problems with time windows. Expert Syst. Appl. 2021, 187. [Google Scholar] [CrossRef]
  30. Qin, H.; Li, T.; Teng, Y.; Wang, K. Integrated production and distribution scheduling in distributed hybrid flow shops. Memetic Comput. 2021, 13, 185–202. [Google Scholar] [CrossRef]
  31. Wang, L.; Pan, Z.X.; Wang, J.J. A review of reinforcement learning based intelligent optimization for manufacturing scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  32. Zhao, F.; Di, S.; Wang, L. A hyperheuristic with Q-learning for the multiobjective energy-efficient distributed blocking flow shop scheduling problem. IEEE Trans. Cybern. 2022, 53, 3337–3350. [Google Scholar] [CrossRef] [PubMed]
  33. Li, R.; Gong, W.Y.; Lu, C. A reinforcement learning based RMOEA/D for bi-objective fuzzy flexible job shop scheduling. Expert Syst. Appl. 2022, 203, 117380. [Google Scholar] [CrossRef]
  34. Wang, H.X.; Sarker, B.R.; Li, J.; Li, J. Adaptive scheduling for assembly job shop with uncertain assembly times based on dual Q-learning. Int. J. Prod. Res. 2020, 59, 5867–5883. [Google Scholar] [CrossRef]
  35. Xu, P.; Luo, W.; Lin, X. BSO20: Efficient brain storm optimization for real-parameter numerical optimization. Complex Intell. Syst. 2021, 7, 2415–2436. [Google Scholar] [CrossRef]
  36. Cheng, S.; Zhang, M.; Ma, L.; Lu, H.; Wang, R.; Shi, Y.H. Brain storm optimization algorithm for solving knowledge spillover problems. Neural Comput. Appl. 2021, 35, 12247–12260. [Google Scholar] [CrossRef]
  37. Hao, J.H.; Li, J.Q.; Du, Y.; Song, M.X.; Duan, P.; Zhang, Y.Y. Solving distributed hybrid flowshop scheduling problems by a hybrid brain storm optimization algorithm. IEEE Access 2019, 7, 66879–66894. [Google Scholar] [CrossRef]
  38. Zhao, F.Q.; Hu, X.T.; Wang, L.; Xu, T.P.; Zhu, N.N.; Jonrinaldi. A reinforcement learning-driven brain storm optimisation algorithm for multi-objective energy-efficient distributed assembly no-wait flow shop scheduling problem. Int. J. Prod. Res. 2023, 61, 2854–2872. [Google Scholar] [CrossRef]
  39. Ma, X.M.; Fu, Y.P.; Gao, K.Z.; Zhu, L.H.; Sadollah, A. A multi-objective scheduling and routing problem for home health care services via brain storm optimization. Complex Syst. Model. Simul. 2023, 3, 32–46. [Google Scholar] [CrossRef]
  40. Ke, L. A brain storm optimization approach for the cumulative capacitated vehicle routing problem. Memetic Comput. 2018, 10, 411–421. [Google Scholar] [CrossRef]
  41. Watkins, C.J.C.H.; Dayan, P. Technical note: Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  42. Deb, K.; Pratap, A.; Agarwal, S. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  43. Wang, G.C.; Gao, L.; Li, X.Y.; Li, P.G.; Tasgetiren, M.F. Energy-efficient distributed permutation flow shop scheduling problem using a multi-objective whale swarm algorithm. Swarm Evol. Comput. 2020, 57, 100716. [Google Scholar] [CrossRef]
  44. Fu, Y.P.; Tian, G.D.; Fard, A.M.H.; Ahmadi, A.; Zhang, C.Y. Stochastic multi-objective modelling and optimization of an energy-conscious distributed permutation flow shop scheduling problem with the total tardiness constraint. J. Clean. Prod. 2019, 226, 515–525. [Google Scholar] [CrossRef]
  45. Lu, C.; Gao, L.; Yi, J.; Li, X. Energy-efficient scheduling of distributed flow shop with heterogeneous factories: A real-world case from automobile industry in China. IEEE Trans. Ind. Inform. 2020, 17, 6687–6696. [Google Scholar] [CrossRef]
  46. Kuidi, M. A memetic algorithm with novel semi-constructive evolution operators for permutation flowshop scheduling problem. Appl. Soft Comput. 2020, 94, 106458. [Google Scholar] [CrossRef]
  47. Zitzler, E.; Deb, K.; Thiele, L. Comparison of multiobjective evolutionary algorithms: Empirical results. Evol. Comput. 2000, 8, 173–195. [Google Scholar] [CrossRef]
  48. Zhang, Q.F.; Li, H. A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  49. Vallada, E.; Ruiz, R.; Framinan, J.M. New hard benchmark for flowshop scheduling problems minimising makespan. Eur. J. Oper. Res. 2015, 240, 666–677. [Google Scholar] [CrossRef]
  50. Gehring, H.; Homberger, J. A parallel hybrid evolutionary metaheuristic for the vehicle routing problem with time windows. In Proceedings of EUROGEN99; Springer: Berlin/Heidelberg, Germany, 1999; Volume 2, pp. 57–64. [Google Scholar]
  51. Karna, S.K.; Sahai, R. An overview on Taguchi method. Int. J. End. Math. Sci. 2012, 1, 1–7. [Google Scholar]
  52. Zitzler, E.; Thiele, L. Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach. IEEE Trans. Evol. Comput. 1999, 3, 257–271. [Google Scholar] [CrossRef]
  53. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  54. Wilcoxon, F.; Katti, S.K.; Wilcox, R.A. Critical values and probability levels for the Wilcoxon rank sum test and the Wilcoxon signed rank test. Sel. Tables Math. Stat. 1970, 1, 171–259. [Google Scholar]
  55. Hou, Y.S.; Wang, H.F.; Fu, Y.P.; Gao, K.Z.; Zhang, H. Multi-objective brain storm optimization for integrated scheduling of distributed flow shop and distribution with maximal processing quality and minimal total weighted earliness and tardiness. Comput. Ind. Eng. 2023, 179, 109217. [Google Scholar] [CrossRef]
  56. Li, F.; Chen, Z.L.; Tang, L. Integrated production, inventory and delivery problems: Complexity and algorithms. INFORMS J. Comput. 2017, 29, 232–250. [Google Scholar] [CrossRef]
  57. Wang, J.; Song, G.; Liang, Z.; Demeulemeester, E.; Hu, X.; Liu, J. Unrelated parallel machine scheduling with multiple time windows: An application to earth observation satellite scheduling. Comput. Oper. Res. 2023, 149, 106010. [Google Scholar] [CrossRef]
Figure 1. A visual example of the devised problem.
Figure 1. A visual example of the devised problem.
Mathematics 11 04306 g001
Figure 2. Illustration of the solution representation method.
Figure 2. Illustration of the solution representation method.
Mathematics 11 04306 g002
Figure 3. Illustration of a new individual generation process.
Figure 3. Illustration of a new individual generation process.
Mathematics 11 04306 g003
Figure 4. Examples of five types of neighborhood structures.
Figure 4. Examples of five types of neighborhood structures.
Mathematics 11 04306 g004
Figure 5. Framework of an action set in MQBSO.
Figure 5. Framework of an action set in MQBSO.
Mathematics 11 04306 g005
Figure 6. The occurrence percentage on each case of C ( U , W ) and D ( U , W ) on four instances.
Figure 6. The occurrence percentage on each case of C ( U , W ) and D ( U , W ) on four instances.
Mathematics 11 04306 g006
Figure 7. Boxplots for the results of nine test instances regarding the IGD-metric.
Figure 7. Boxplots for the results of nine test instances regarding the IGD-metric.
Mathematics 11 04306 g007
Figure 8. Boxplots for the results of nine test instances regarding the HV-metric.
Figure 8. Boxplots for the results of nine test instances regarding the HV-metric.
Mathematics 11 04306 g008
Figure 9. Nondominated solutions obtained by MQBSO and its peers.
Figure 9. Nondominated solutions obtained by MQBSO and its peers.
Mathematics 11 04306 g009
Table 1. The parameters and variables.
Table 1. The parameters and variables.
SymbolDescription
F Set of f factories, F = 1 , 2 , , f .
M Set of m machines, M = 1 , 2 , , m .
N Set of n jobs, N = 0 , 1 , 2 , , n , where 0 means a virtual job.
V Set of v vehicles at a factory, V = 1 , 2 , , v .
g The factory index, g F .
i The machine index, i M .
k , j , l The job index, k , j , l N .
h The vehicle index, h V .
α i j The production time of job j on machine i .
β k j The drive time between customers k and j .
γ g j The drive time between factory g and customer j .
ψ j The load of job j .
φ j The service time at customer j .
d j , d j The delivery time window of customer j .
E j The unit earliness weight of customer j .
T j The unit tardiness weight of customer j .
ϕ The maximum load capacity of the vehicle.
G An extremely positive integer.
X k j g = 1, if job j is directly processed after job k at factory g , 0, otherwise.
Y j g = 1, if job j is assigned to factory g , 0, otherwise.
Z k j h = 1, if customer j is directly delivered after customer k by vehicle h , 0, otherwise.
C i j The completion time of job j on machine i .
S h The delivery start time of vehicle h .
A j The arrival time at customer j .
L j The departure time from customer j .
C m a x The makespan criterion.
T W E T The total weighted earliness and tardiness criterion.
Table 2. Abbreviation information of 24 test instances.
Table 2. Abbreviation information of 24 test instances.
Instance f ProductionDistributionInstance f ProductionDistribution
2-5-302VFR30_5_1C1_2_12-10-302VFR30_10_1C1_2_1
2-5-602VFR60_5_1C1_2_12-10-602VFR60_10_1C1_2_1
2-5-902VFR90_5_1C1_2_12-10-902VFR90_10_1C1_2_1
2-5-1202VFR120_5_1C1_2_12-10-1202VFR120_10_1C1_2_1
3-5-303VFR30_5_1C1_2_13-10-303VFR30_10_1C1_2_1
3-5-603VFR60_5_1C1_2_13-10-603VFR60_10_1C1_2_1
3-5-903VFR90_5_1C1_2_13-10-903VFR90_10_1C1_2_1
3-5-1203VFR120_5_1C1_2_13-10-1203VFR120_10_1C1_2_1
4-5-304VFR30_5_1C1_2_14-10-304VFR30_10_1C1_2_1
4-5-604VFR60_5_1C1_2_14-10-604VFR60_10_1C1_2_1
4-5-904VFR90_5_1C1_2_14-10-904VFR90_10_1C1_2_1
4-5-1204VFR120_5_1C1_2_14-10-1204VFR120_20_1C1_2_1
Table 3. Orthogonal array and AIGD results of MQBSO.
Table 3. Orthogonal array and AIGD results of MQBSO.
No. p s r g r o r t AIGD
1200.200.200.200.1016
2200.400.400.400.0927
3200.600.600.600.0904
4200.800.800.800.0861
5400.200.400.600.1056
6400.400.200.800.0852
7400.600.800.200.0909
8400.800.600.400.1232
9600.200.600.800.1005
10600.400.800.600.0952
11600.600.200.400.0939
12600.800.400.200.0984
13800.200.800.400.0931
14800.400.600.200.0956
15800.600.400.800.0868
16800.800.200.600.0867
Table 4. Parameter settings of four compared algorithms.
Table 4. Parameter settings of four compared algorithms.
AlgorithmParameterLevelsValue
MOWSApopulation size20, 40, 60, 8020
crossover probability0.80, 0.85, 0.90, 0.950.85
mutation probability0.10, 0.15, 0.20, 0.250.15
MOBSO p s 20, 40, 60, 8080
r g 0.20, 0.40, 0.60, 0.800.80
r o 0.20, 0.40, 0.60, 0.800.20
r t 0.20, 0.40, 0.60, 0.800.60
NSGA-IIpopulation size20, 40, 60, 8080
mutation probability0.10, 0.15, 0.20, 0.250.15
MOEA/Dpopulation size20, 40, 60, 8040
neighborhood size10, 15, 20, 2525
Table 5. Results of MQBSO and its peers concerning C-metric.
Table 5. Results of MQBSO and its peers concerning C-metric.
InstanceC(QB,WS)C(WS,QB)C(QB,BS)C(BS,QB)C(QB,GA)C(GA,QB)C(QB,EA)C(EA,QB)
2-5-300.79510.01610.99080.00000.98890.00000.97090.0031
2-5-600.98100.00000.97290.00001.00000.00001.00000.0000
2-5-900.97350.00001.00000.00001.00000.00001.00000.0000
2-5-1200.93890.00000.97390.00180.98180.00001.00000.0000
2-10-300.00000.68100.00650.57850.04390.23610.06960.2504
2-10-600.94810.00001.00000.00001.00000.00001.00000.0000
2-10-900.92970.00001.00000.00001.00000.00001.00000.0000
2-10-1200.95690.00001.00000.00001.00000.00001.00000.0000
3-5-300.97820.00000.95190.00000.80820.12500.68680.2206
3-5-600.93060.00000.97310.00001.00000.00000.97500.0000
3-5-900.98160.00001.00000.00001.00000.00001.00000.0000
3-5-1200.96870.00240.99640.00001.00000.00001.00000.0000
3-10-300.10770.58230.21630.50480.47050.20680.35650.2504
3-10-600.99210.00001.00000.00001.00000.00001.00000.0000
3-10-900.96580.00001.00000.00001.00000.00001.00000.0000
3-10-1200.97640.00001.00000.00001.00000.00001.00000.0000
4-5-300.89910.00000.90640.00000.15310.77970.06400.8341
4-5-600.98200.00000.97530.00000.99520.00001.00000.0000
4-5-900.97920.00000.99570.00001.00000.00001.00000.0000
4-5-1200.94640.00000.99180.00001.00000.00001.00000.0000
4-10-300.68490.02250.94000.00380.18590.41200.11720.4602
4-10-600.96050.00001.00000.00001.00000.00001.00000.0000
4-10-900.97960.00001.00000.00001.00000.00001.00000.0000
4-10-1200.97960.00000.99810.00001.00000.00001.00000.0000
Average0.86820.05430.91200.04540.85950.07330.84330.0841
Table 6. Results of MQBSO and its peers about IGD-metric.
Table 6. Results of MQBSO and its peers about IGD-metric.
InstanceMQBSOMOWSAMOBSONSGA-IIMOEA/D
2-5-300.15950.25850.27880.37840.3512
2-5-600.06680.44560.41320.50730.4948
2-5-900.06950.45870.42520.54850.5790
2-5-1200.09460.29110.28130.42240.4837
2-10-300.42010.17770.23110.31870.3576
2-10-600.12460.37080.38070.52540.5820
2-10-900.11470.36760.47780.65400.9409
2-10-1200.10230.36860.43770.61610.7933
3-5-300.10380.36380.35310.27810.2269
3-5-600.09630.49560.47840.51850.5287
3-5-900.04660.42700.42630.47110.4883
3-5-1200.05580.20640.25350.30550.3883
3-10-300.17990.14210.16260.23800.2398
3-10-600.05610.47020.44620.50000.5307
3-10-900.06030.43440.42240.47970.5345
3-10-1200.05060.20670.27480.33020.4071
4-5-300.23780.48510.48450.11610.1381
4-5-600.06160.50630.52010.54650.5696
4-5-900.06460.44120.48880.51840.5581
4-5-1200.08420.18630.29550.31240.3523
4-10-300.18820.23830.23710.20160.2659
4-10-600.05110.42220.45710.48070.5007
4-10-900.04710.40300.41500.44800.4806
4-10-1200.04770.22450.27200.31130.3809
Average0.10770.34970.37140.41780.4655
Table 7. Results of MQBSO and its peers regarding HV-metric.
Table 7. Results of MQBSO and its peers regarding HV-metric.
InstanceMQBSOMOWSAMOBSONSGA-IIMOEA/D
2-5-300.91060.73310.66550.52220.5543
2-5-600.95370.51260.49920.39140.3804
2-5-900.94740.48250.46250.32920.2711
2-5-1200.86970.59550.56780.38480.3024
2-10-300.32080.80710.70520.52910.5064
2-10-600.94890.60540.52720.35150.3027
2-10-900.96150.64760.41960.26370.0894
2-10-1200.90620.57970.40670.23970.1076
3-5-300.84220.53790.49820.65790.6860
3-5-600.93020.40180.38220.33120.2990
3-5-900.91770.42600.40110.34640.3141
3-5-1200.81110.55190.50740.43830.3359
3-10-300.55480.74810.66400.58270.5695
3-10-600.95830.46430.42650.37140.3171
3-10-900.93850.47560.41930.35250.2742
3-10-1200.78020.51970.44770.36800.2821
4-5-300.78390.44640.41050.88640.8946
4-5-600.93050.37770.32680.29510.2602
4-5-900.90970.39700.35960.31660.2671
4-5-1200.80980.55210.44770.41210.3497
4-10-300.65320.56630.52910.80530.8131
4-10-600.91670.45440.39520.35780.3262
4-10-900.90360.40830.34590.31100.2636
4-10-1200.79580.52520.45580.40190.3179
Average0.84400.53400.46960.42690.3785
Table 8. Results of the Wilcoxon signed rank test.
Table 8. Results of the Wilcoxon signed rank test.
MQBSO vs.Metric R + R z p
MOWSAIGD290104.000.0000
HV279213.690.0001
MOBSOIGD29464.110.0000
HV288123.940.0000
NSGA-IIIGD29374.090.0000
HV289113.970.0000
MOEA/DIGD29464.110.0000
HV288123.940.0000
Table 9. Results of MQBSO and MBSO in the C-, IGD-, and HV-metrics.
Table 9. Results of MQBSO and MBSO in the C-, IGD-, and HV-metrics.
InstanceCIGDHV
C(MQBSO,MBSO)C(MBSO,MQBSO)MQBSOMBSOMQBSOMBSO
2-5-300.62300.13840.17160.21670.80120.6947
2-5-600.97210.00000.12100.36980.90960.4643
2-5-900.97460.00000.09970.38490.89300.4789
2-5-1200.70140.18480.10850.15250.75410.6730
2-10-300.00000.81760.52890.13920.27690.8436
2-10-600.71920.02950.18880.17920.80350.5949
2-10-900.67000.12050.19490.25770.81910.6260
2-10-1200.62060.28570.14640.18880.76120.6732
3-5-300.97760.00000.11420.31740.91770.5385
3-5-600.95640.00000.09050.47090.91890.3773
3-5-900.98710.00000.05640.36880.87240.4316
3-5-1200.75710.14620.06210.09110.71480.6578
3-10-300.04500.68330.32100.16400.47510.7644
3-10-600.98780.00000.06960.3819z0.5006
3-10-900.99390.00000.06510.36970.91190.5338
3-10-1200.72030.18120.05190.09060.71150.6547
4-5-300.96080.00000.06370.46830.91800.3940
4-5-600.97860.00000.06370.51510.89050.3329
4-5-900.98820.00000.07080.41260.87150.3920
4-5-1200.57610.25950.07870.09610.71740.6761
4-10-300.32180.37460.15250.11830.68510.7460
4-10-600.99040.00000.06320.36450.88120.4630
4-10-900.99340.00000.04870.35380.86650.4476
4-10-1200.78380.11820.05400.09660.72240.6465
Average0.76250.13910.12440.27370.79280.5669
Table 10. Average CPU time of all the used algorithms in different groups.
Table 10. Average CPU time of all the used algorithms in different groups.
n Average CPU Time (s)
MQBSOMOWSAMOBSONSGA-IIMOEA/D
3048.338 44.14448.62421.07233.675
60143.127 152.184170.26867.990106.256
90275.020 358.952347.227142.793250.765
120470.585 638.195582.547249.867474.062
Average234.268 298.369287.166120.430216.189
Table 11. Information for the test problem.
Table 11. Information for the test problem.
JobProcessing TimeTime WindowLoadUnit Earliness and Tardiness Weight
α 1 j α 2 j d j d j ψ j e l j t d j
J 1 3645105130200.20.4
J 2 7982193253200.30.2
J 3 2119162192300.20.3
J 4 1682190294100.40.3
J 5 2357141160200.20.3
J 6 3173157192100.30.2
J 7 2990234300100.20.3
J 8 953311345400.10.2
J 9 4530308353200.30.2
J 10 5093278364200.20.4
J 11 5486332378200.30.5
J 12 5420371400100.40.5
Table 12. The drive time for the numerical example.
Table 12. The drive time for the numerical example.
F 1 F 2 J 1 J 2 J 3 J 4 J 5 J 6 J 1 J 8 J 9 J 10 J 11 J 12
F 1 09995270557694542345441294113
F 2 9990434753708660355318443746
J 1 52430899411312924337522487670
J 2 7047890353746107374522675375
J 3 5553943502139105121816295370
J 4 76701133721018125432736933337
J 5 94861294639180143283947563553
J 6 5460241071051251430253648692335
J 7 233533371243282501233445375
J 8 4553754518273936120233410737
J 9 441822221636474833230841822
J 10 1244486729935669443484024107
J 11 9437765353333523531071824017
J 12 11346707570375335753722107170
Table 13. Comparison results of MQBSO and CPLEX.
Table 13. Comparison results of MQBSO and CPLEX.
n WeightCPLEXMQBSO
OutputTime (s)OutputTime (s)
8 1.0,0.0 246.00025.140246.0002.828
0.5,0.5 292.0502242300.5002.213
0.0,1.0 322.700245324.0002.203
10 1.0,0.0 307.000(AOV)3600307.0003.383
0.5,0.5 428.800(AOV)3600419.5002.474
0.0,1.0 499.300(AOV)3600498.0002.894
12 1.0,0.0 360.000(AOV)3600360.0002.187
0.5,0.5 583.650(AOV)3600571.5003.347
0.0,1.0 767.800(AOV)3600723.0004.079
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, S.; Xu, J.; Qiao, Y. Multi-Objective Q-Learning-Based Brain Storm Optimization for Integrated Distributed Flow Shop and Distribution Scheduling Problems. Mathematics 2023, 11, 4306. https://doi.org/10.3390/math11204306

AMA Style

Zhang S, Xu J, Qiao Y. Multi-Objective Q-Learning-Based Brain Storm Optimization for Integrated Distributed Flow Shop and Distribution Scheduling Problems. Mathematics. 2023; 11(20):4306. https://doi.org/10.3390/math11204306

Chicago/Turabian Style

Zhang, Shuo, Jianyou Xu, and Yingli Qiao. 2023. "Multi-Objective Q-Learning-Based Brain Storm Optimization for Integrated Distributed Flow Shop and Distribution Scheduling Problems" Mathematics 11, no. 20: 4306. https://doi.org/10.3390/math11204306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop