1. Introduction
With the rapid acceleration of global urbanization, the effective management of municipal solid waste has become a critical environmental and socio-economic challenge. Waste incineration plants have emerged as a predominant solution. They not only effectively reduce environmental pollution but also generate substantial amounts of energy, achieving waste resource utilization. As the global demand for renewable energy grows, incineration plants are undergoing a significant digital transformation to enhance operational stability and energy recovery efficiency [
1,
2,
3]. The waste storage warehouse, a key transit hub in incineration plants, relies on multiple co-rail waste cranes to perform tasks such as waste transferring, turning, and feeding into furnaces. Consequently, developing an efficient and automated scheduling strategy for multiple co-rail waste cranes is no longer merely a technical enhancement but a strategic necessity for the modern smart waste-to-energy industry.
Due to the unique characteristics of the waste-handling environment, applying existing crane scheduling methodologies to waste incineration plants presents significant challenges. Previous research on crane scheduling has primarily focused on quays [
4,
5], general warehouses [
6], container terminals [
7], or factories [
8]. These scenarios typically involve homogeneous tasks (loading/unloading) within structured storage environments, where goods have fixed coordinates and cranes operate in relatively independent zones with simple interference constraints. In contrast, waste crane systems operate in an unstructured, continuous environment characterized by three distinct complexities that existing methods fail to address:
- (1)
Task heterogeneity and process complexity: Unlike standard material handling, waste cranes must perform diverse tasks (transfer, turning, and feeding) with distinct operational procedures.
- (2)
High-frequency spatial interference: Multiple cranes share a single pair of rails over a limited operational space. The non-deterministic nature of waste accumulation leads to frequent, dynamic trajectory conflicts that simple zoning strategies or static collision penalties cannot resolve.
- (3)
Time-window constraints: The feeding tasks are subject to rigid time windows imposed by the incinerator’s operational state. Failure to meet these windows risks furnace shutdowns, making timeliness far more critical than in general warehousing.
Our previous study [
9] focused on this problem and established a basic scheduling framework using a simulation-based Genetic Algorithm (GA) to minimize makespan. However, it overlooked critical operational realities. First, it formulated only a simulation-based mathematical model rather than a standard Mixed-Integer Linear Programming (MILP) model, which precludes the use of exact algorithms. Second, neglecting time window constraints renders the schedule impractical for real-world continuous incineration. Third, the simulation module, which relied on a time-step-based simulation mechanism, incurred excessive computational costs, thereby hindering the feasibility of real-time application. Fourth, the GA struggles to escape local optima when solving such highly constrained problems.
To address these unresolved issues, this paper tackles the research problem of coordinated task assignment and scheduling for multiple co-rail waste cranes to reduce energy consumption, considering task heterogeneity, dynamic spatial interference, and time window constraints. The novelty of this study and the main contributions related to crane scheduling are summarized as follows:
- (1)
Mathematical Modeling: We construct the first standardized MILP model for the multiple waste crane scheduling problem. It establishes a complete theoretical benchmark for this NP-hard problem, and its relaxed form for dynamic interference provides a rigorous theoretical lower bound (LB) to quantify the performance of various metaheuristic algorithms.
- (2)
Problem-solving Framework: A scheduling scheme is represented by task sequencing and task assignment. A metaheuristic algorithm is employed to optimize the task sequences, while the corresponding task assignment for each sequence is determined by the proposed heuristic method.
- (3)
Algorithm Design: To search for high-quality task sequences, the Ivy algorithm (IVYA) [
10] is first discretized and combined with the Simulated Annealing (SA) algorithm, forming our two-stage Discrete Ivy-Simulated Annealing (DIVY-SA) algorithm. Compared with classical metaheuristics like PSO and GA, it effectively overcomes the premature convergence of PSO and the limited local search capability of GA when handling the high-dimensional, discrete constraints of waste crane scheduling.
- (4)
Simulation: A discrete-event simulation (DES) module is constructed to precisely calculate the objective function values for every task sequence under its corresponding assignment. This allows for accurate objective evaluation while accounting for dynamic spatial interference, which is often neglected in traditional mathematical models.
- (5)
Practical Validation: The effectiveness of the proposed method is validated through extensive experiments using real-world data from an incineration plant in Guangzhou, demonstrating a 22.19% reduction in the total waste crane traveling distance compared with manual scheduling.
3. Problem Description and Modeling
3.1. Problem Description
This paper considers a waste storage warehouse (
Figure 1) with a waste pool, unloading/feeding ports, and waste cranes that perform feeding, turning, and transferring tasks.
- (1)
Waste Pool
The rectangular waste pool (L × W × H) contains s feeding ports and p unloading ports (s, p > 1). Waste is stored in three zones: the raw waste zone (newly unloaded waste), the fermenting waste zone (draining/fermenting waste), and the fermented waste zone (incinerable waste). The zones rotate sequentially and are gridded to facilitate crane operation. Once the waste in the fermented waste zone is completely incinerated and the waste in the fermenting waste zone has fully fermented, the fermenting waste zone is converted into a fermented waste zone, and the raw waste zone will also be converted into a fermenting waste zone. Meanwhile, the original location of the fermented waste zone serves as a raw waste zone to receive newly unloaded waste.
- (2)
Waste Crane
Warehouses typically have r (r > 1) waste cranes spanning the waste pool. Each crane has a gantry (moving along the X-axis within a defined range), a trolley (moving along the Y-axis), and a grab bucket (hoisting on the Z-axis). All cranes utilize a common rail design, and a safe distance must be kept between their gantries.
- (3)
Waste Operation Process
In daily management, cranes are typically required to perform the following three tasks:
Feeding task: Grab waste from the fermented waste zone and move it to the left side of the furnace feeding port, then evenly discharge it into the port from left to right. Each feeding task should be performed within a designated time window.
Transferring task: Transfer newly unloaded waste at the waste unloading ports to the raw waste zone for stockpiling.
Turning task: Loosen surface waste in the fermenting waste zone by repeatedly grabbing and releasing it to facilitate fermentation.
In scheduling, waste cranes should feed furnaces within the required time windows, and complete the transferring and turning tasks in other time periods. The aim of the scheduling is to minimize the total traveling distance of cranes and the total delay time of tasks.
- (4)
Task Interaction Analysis
In the waste crane scheduling problem, task interactions do not originate from logical precedence or procedural dependencies, as the three types of tasks are operationally independent. Instead, the interaction is characterized by an indirect coupling caused by the shared rail system and limited operational space. As illustrated in
Figure 2, since multiple cranes operate on a common rail, the execution of one task inevitably creates spatial-temporal interference, potentially delaying others or making additional travel distances to avoid. Therefore, it is essential to optimize task assignment and sequencing to minimize the total crane traveling distances caused by interference and task delays.
3.2. Model Assumptions
This paper investigates a static scheduling model for waste cranes, based on the following assumptions:
- (1)
The acceleration and deceleration processes are ignored, and the gantries are assumed to move at a constant average speed.
- (2)
The operations of the grab bucket (including trolley travel in the Y-direction, hoisting in the Z-direction, and grabbing/releasing waste) can be completed within a predetermined time.
- (3)
The quantity and information of tasks are known and given by the upper-level system.
- (4)
The model excludes unexpected events like crane failures or demand changes.
These assumptions define the scope of this model as a deterministic optimization approach. These simplifications may not fully capture the dynamic uncertainties of a real-world waste incineration plant, but they are necessary to ensure the computational feasibility of the MILP model. Therefore, in practical engineering applications, this model is intended to be deployed within a rolling horizon strategy. The static scheduling model is iteratively re-optimized at fixed intervals or triggered by specific dynamic events. At the start of each rescheduling window, the cranes’ current physical coordinates and any unfinished tasks are fed into the model as the new initial state, ensuring continuous adaptation to operational uncertainties and compensating for the limitations of the static model.
3.3. Mathematical Model
To address the multiple waste crane scheduling problem, we construct a MILP model as a relaxation of the original scheduling problem. This model aims to minimize the total traveling distance of cranes and the total delay time of tasks. It incorporates various complex constraints, including crane operating ranges, safety distances, and task time windows. The model’s parameters and variables are defined in
Table 1 and
Table 2.
subject to:
Equation (1) is the objective function, aiming to minimize the total traveling distance of cranes and the total delay time of tasks. Considering the actual crane operations, strict hard time window constraints often lead to infeasibility. Therefore, we adopt a soft time window strategy. This approach allows tasks to be executed outside the time window at the cost of a linearly increasing penalty. Additionally, a severe penalty is introduced to prevent excessive task delays.
The assignment of equal weights to traveling distance and delay time is justified by both physical and algorithmic considerations. First, given the typical crane speed of 1 m/s, one second of delay is numerically equivalent to one meter of additional travel distance, allowing both metrics to be unified into a generalized ‘distance-based cost’ without dimensional bias. Second, due to the existence of task time windows, the delay for each task can typically be reduced to zero in an optimal solution. Even if not zero, the magnitude of these delays remains comparable to that of the traveling distance, resulting in a limited impact on the incineration plants.
Constraint (2) enforces the fundamental logic that each task must be assigned to exactly one crane. Constraint (3) incorporates the physical limitations, ensuring that task i can only be assigned to crane k if the task’s location falls within crane k’s operating range. Constraints (4) to (6) define the logical sequence of tasks. Specifically, Constraints (4) and (5) guarantee that each task has at most one predecessor and one successor on the same crane. Constraint (6) ensures that a sequence link ykij can only exist if both tasks i and j are assigned to crane k. Constraints (7) identify the first task for each crane. Constraint (8) guarantees that each crane starts with a single unique task.
Constraints (9) to (11) govern the timeline of operations. Specifically, Constraint (9) ensures that the start time of each task is no earlier than the sum of the completion time of its preceding task and the empty travel time required to reach it. Constraint (10) guarantees that the start time of the initial task for each crane is no earlier than the empty travel time from the crane’s initial position to that task. Constraint (11) enforces that the start time of a task must satisfy the lower bound of its predefined time window.
Constraints (12) to (14) define the soft time window mechanism for feeding tasks. Constraints (12) and (13) calculate the delay time. Constraint (14) identifies tasks that incur excessive delays. Constraints (15) and (16) force task pairs with overlapping physical spaces to be separated in the time dimension to prevent collisions during task execution. However, they do not account for dynamic trajectory interference—such as crossing paths during empty travel between task zones. Modeling such interference in a linear programming framework need to split the whole process timespan into short time intervals and may yield an overly granular programming. Consequently, the MILP can provide a rigorous theoretical LB for the original problem. Constraint (17) formulates the calculation of the total traveling distance for each crane.
This multi-waste crane scheduling model is analogous to the Parallel Machine Scheduling Problem (PMSP), which is a typical NP-hard problem. However, distinct from standard PMSP, this problem involves spatial interference between cranes, task time windows, and dynamic setup times. These unique characteristics render existing parallel machine scheduling methods inapplicable. Moreover, considering the relaxed constraints of crane interference (15–16), this mathematical model can provide a theoretical LB for the original problem. To mitigate this limitation and enhance computational efficiency for large-scale problems, we propose a hybrid scheduling strategy combining metaheuristics with simulation.
4. Solution Method
A two-stage DIVY-SA algorithm is proposed to solve the problem. The flowchart of the algorithm is shown in
Figure 3. We first generate the initial task sequence population and conduct feasibility processing on it. Then, we assign cranes to each task in the individuals of the population according to the task assignment scheme, with the goal of minimizing the execution and avoidance distances for all waste cranes. Due to the complex operation of waste cranes, it is difficult to obtain the fitness value through mathematical calculation. Therefore, we utilize a simulation module for this purpose. After that, we utilize the DIVY algorithm’s broad-range search capability to identify high-quality task sequences, and then employ the SA algorithm’s fine-grained search capability to perform a refined search.
4.1. Population Encoding and Initialization
We use integer encoding for population individuals. For a given task set , each task is represented as , where the parameters respectively denote the task type, start/end positions, start/end times, and threshold of excessive delay. Specifically, denotes a feeding task, represents a transferring task, and indicates a turning task. For feeding tasks, the end position is defined as the center coordinate of the corresponding feeding port. Regarding time constraints, only feeding tasks are subject to specific time windows; for all other tasks, the earliest start time is set to 0, and the latest start time is set to M. Therefore, each population individual can represent a task sequence, expressed as where indicates the index of the j-th task to be executed and .
The encoding generation and preprocessing are shown in
Figure 4. First, a continuous encoded sequence (a) consisting of random numbers is generated. Then, these numbers are sorted in ascending order based on their magnitude, and an integer encoding (c) is derived by mapping each value to its rank in the ascending order. Finally, to improve the initial population’s quality, we perform feasibility processing: tasks with time-window constraints are extracted, sorted by their start times, and reinserted into the original positions vacated by the extraction. This feasibility processing helps improve the quality of the initial task sequence population.
4.2. Allocation of Tasks to Cranes
Directly searching for both task sequences and crane assignments using metaheuristics would result in an enormously large solution space. Consequently, we introduce a heuristic task assignment method. Although this approach may theoretically lead to local optima, it is indispensable for effectively narrowing down the search space to a manageable size.
In the waste crane scheduling system, multiple cranes face issues related to their operating range constraints and mutual interference. Therefore, during task assignment, it is necessary to consider not only whether each crane can complete the task but also the additional traveling distance caused by cranes’ avoidance behaviors. We propose a task assignment scheme based on minimizing the execution and avoidance distances of all waste cranes. This method assumes that all cranes execute the tasks in the task sequence serially. Assuming that crane k () has an initial position , the detailed steps of task assignment for the tasks in a task sequence (denoted as ) are as follows.
Step 1. Set task . Set the positions of all cranes to their initial positions: .
Step 2. For each feasible crane k with Rki = 1, calculate the total traveling distance dki of all cranes when crane k executes task i and other cranes avoid it, which consists of the following three parts.
(1) The loaded travel distance for task
i:
where
w denotes half the width of the feeding port.
(2) The distance from the current position of crane
k to the start position of task
i:
(3) The total avoidance distance travelled by other cranes to avoid crane
k:
Herein, is the avoidance position of crane j when crane k executes task i.
When
j <
k (crane
j is on the left of crane
k), it is determined as:
When
j >
k (crane
j is on the right of crane
k), it is determined as:
where
dsafe is the safety distance between adjacent cranes’ gantries.
The total traveling distance
dki is the sum of three parts:
Step 3. Select the crane
k* with the minimum total traveling distance
dki for task
i. If multiple cranes have the same minimum distance, select the one with the smallest index, expressed as:
Step 4. Update the position of crane
k* as:
Update the position of crane
j as:
Step 5. Set i as the next task in the task sequence and repeat Step 2 to 4 until all tasks have been assigned.
This assignment method tends to choose cranes that minimize interference, thereby reducing the frequency of interference and further decreasing the total traveling distance of all cranes. Meanwhile, minimizing distance implicitly contributes to reducing makespan and delays by shortening operation time.
4.3. Fitness Calculation Using Simulation Module
Considering that the mathematical calculation of the fitness is extremely complex due to the cranes’ complex operation, a DES module is proposed. The module simulates the scheduling scheme (which consists of task sequence and assignment) to obtain the operation trajectories of cranes, and thus calculates the fitness. To reduce unnecessary movements of cranes caused by interference between cranes, the following crane operation rules are established.
Rule 1. When crane
k is performing its current task
i, its gantry will occupy a range along the X-axis, which is a forbidden range for other cranes. This range is determined by three points: the position
Pk of crane
k when it starts to execute task
i, the start position
, and the end position
of current task
i. Formally, the occupied range
Zk is expressed as:
Rule 2. Before crane k intends to execute the next task, it needs to determine whether the range it will occupy overlaps with the ranges currently being occupied by other cranes. If no overlap is detected (), it can start executing the task directly. Otherwise, the task is delayed until the interfering crane(s) complete their current tasks and release their occupied ranges.
Rule 3. When crane k confirms that it can start executing its next task, any other cranes located within its occupied range must move to avoidance positions determined by Equations (21) and (22).
Rule 4. A crane that enters the avoidance position will send a request to execute its next task, with the feasibility determined in accordance with Rule 2.
Rule 5. To prevent deadlocks, the crane that requests to execute its next task first is given higher priority. Simultaneous requests are resolved by crane ID (lower ID has priority).
Figure 5 illustrates the flowchart of the DES. The simulation comprises three types of events, representing “Task Evaluation”, “Task Finish”, and “Active Avoidance Execution”, respectively. At the onset of the simulation, a “Task Evaluation” event is scheduled for all cranes. Starting from these initial events, the simulation module continuously triggers subsequent events until all tasks are completed.
Take an example involving three cranes and four tasks. The trajectories of the cranes’ gantries and their corresponding Gantt charts are shown in
Figure 6. Task 1 is a feeding task, task 2 is a transferring task, and tasks 3 and 4 are turning tasks. When crane 2 intends to execute task 2, since its occupied range overlaps with that of crane 1 (when crane 1 is executing task 1) and crane 2’s current position lies within crane 1’s occupied range, crane 2 first needs to perform avoidance operations and wait until crane 1 completes task 1 before starting its own execution. However, the occupied ranges of crane 3 (when crane 3 is executing tasks 3 and 4) do not overlap with those of the other two cranes, so tasks 3 and 4 can be executed directly without interference.
4.4. First Stage: Broad-Range Search via the DIVY Algorithm
The IVYA is a novel metaheuristic optimization algorithm. It exhibits strong optimization capabilities and low computational costs, making it suitable for solving the considered problem. The IVYA simulates ivy growth, where each individual is characterized by position (a solution in the search space), fitness (the corresponding objective function value), and growth velocity (the rate of solution variation). However, the IVYA is designed for continuous problems. Therefore, we discretize it by redefining the operators in its population update formula and changing the growth rate of individuals to a random perturbation. The procedure of the proposed DIVY algorithm is shown in Algorithm 1, following the structure of the original IVYA.
| Algorithm 1. The DIVY algorithm |
Input: The dimension of the solution space D, maximum number of iterations ItDIVY, population size Npop. Output: The best task sequence.
Generate the initial population according to Section 4.1. For each individual, task assignment is performed in accordance with the method described in Section 4.2, and the fitness is calculated by utilizing the simulation module described in Section 4.3. Sort the individuals from best to worst based on their fitness values, and set best task sequence Xbest = X1. While It < It DIVY do for i = 1 to Npop do Generate a new individual according to Equation (29). Perform task assignment for and calculate the fitness. end for Merge X with Xnew and sort the individuals from best to worst. Retain the top Npop individuals with the best fitness values, and set Xbest = X1. It = It + 1. end while
|
In the original IVYA, the most critical aspect is the method for generating new individuals. During the discretization of the IVYA, we mainly focus on modifying the formula for generating new individuals in the original algorithm, which is shown in Equation (28).
Here, Xj refers to the individual preceding Xi in the population sorted by fitness from best to worst. If Xi is the individual with the best fitness, let Xj be Xbest. β is a random number within the range (1, 1.5). N (1, D) denotes a D-dimensional random vector, whose components are random numbers following a normal distribution. is the growth velocity of Xi. Operators and stand for element-wise multiplication and element-wise division between vectors.
To discretize the IVYA, we modify Equation (28) to (29).
Here, and are two randomly generated D-dimensional sequences. Pa and Pb are two parameters to be adjusted. The detailed explanations of the operators in the formula are as follows.
(1) Operator
denotes sequence subtraction, which is used to obtain the swap sequence (SS) between two sequences. Each SS consists of one or more swap operations (SOs), where each SO refers to swapping the positions of two numbers in the sequence. For example, SO (1, 3) denotes swapping the positions of the values 1 and 3 in the sequence. Assuming there are two n-dimensional sequences X
= (
x1,
x2, …,
xn), Y
= (
y1,
y2, …,
yn), then
can be expressed as:
Figure 7 shows an example of the subtraction between two sequences X and Y.
(2) Operator represents the multiplication operation between a number and an SS. For example, indicates retaining the SOs in SS with a probability of P.
(3) Operator
denotes the composition operation between SSs, which combines all the SOs in two SSs into a new SS. For instance, if SS
1 = (SO
11, SO
12, …, SO
1m) and SS
2 = (SO
21, SO
22, …, SO
2n), their composition can be expressed as:
(4) Operator
represents the addition operation between a sequence and an SS, which means executing the SS operations on the sequence. Assuming a sequence X
= (
x1,
x2, …,
xn) and SS = (SO
1, SO
2, …, SO
m), then
can be expressed as:
Figure 8 shows a schematic diagram of executing SS = [SO(1, 4), SO(3, 5)] on the sequence X = (4, 2, 1, 3, 5).
The DIVY algorithm exhibits strong global search performance, enabling it to efficiently locate high-quality population regions. It also exhibits advantages in parameter design: only two core parameters (Pa and Pb) are required for experimental tuning to effectively regulate the search process. Specifically, Pa denotes the probability that individuals in the population approach the current optimal individual, which governs the algorithm’s directional exploration of target high-quality solution regions. Pb denotes the random search probability, which prevents the algorithm from falling into local optima.
4.5. Second Stage: Fine-Grained Search via the SA Algorithm
The DIVY algorithm utilizes a swarm-based optimization approach, thus enabling extensive exploration within the solution space. However, it lacks refined local search capabilities, and this swarm-based optimization incurs high computational costs. Given these characteristics, the SA algorithm, with its efficient local refinement performance and relatively lower computational costs, acts as an ideal complement to the DIVY algorithm in turn. Therefore, we incorporate the SA algorithm as the second stage. When the DIVY algorithm reaches the preset maximum number of iterations, or when the optimal solution it has found remains unchanged for 10 consecutive iterations, the algorithm will switch to the SA algorithm. The SA algorithm adopts the result obtained by the DIVY algorithm as its initial solution and continuously updates its solution. The steps of the SA algorithm are shown in Algorithm 2.
| Algorithm 2. The SA algorithm |
Input: Initial solution X = the best solution obtained by DIVY; length of the Markov chain LM; initial temperature T; maximum number of iterations ItSA; cooling rate Output: The best task sequence.
for It = 1 to ItSA do for j = 1 to LM do Generate a new solution Xnew by randomly swapping two segments. Perform task assignment for Xnew and calculate the fitness f(Xnew). Accept new solution based on Metropolis criteria. end for T = T × end while
|
The SA algorithm exhibits a strong ability for fine-grained search when tackling highly complex optimization problems, owing to its temperature decay mechanism. However, its performance is relatively sensitive to parameter configurations: reasonable parameter configurations enable it to ensure optimization accuracy while balancing computational efficiency, whereas inappropriate parameter configurations may lead to slow convergence or degradation in solution quality. By using the high-quality solution from DIVY as the starting point, we overcome SA’s initialization sensitivity.
4.6. Computational Complexity Analysis
In this section, we systematically analyze the computational complexity of the proposed algorithm as follows.
- (1)
Task Allocation Method
For a given task sequence of length n and r cranes, the time complexity is .
- (2)
Simulation Module
In the worst case, each task corresponds to discrete events, and the interference judgment for each event requires traversing r cranes. Therefore, the time complexity of the simulation module for a complete scheduling scheme is: .
- (3)
DIVY Algorithm
Generating the initial population with Npop individuals of n-dimensional task sequences has a time complexity of . For each individual, the algorithm generates a new sequence (O(n)), completes task assignment (), and performs fitness evaluation via the simulation module (). After the new population is generated, merging and sorting the population has a complexity of . Since and in the problem scenario, the time complexity of a single iteration is simplified to: . Therefore, with a maximum of ItDIVY iterations, the total time complexity of the DIVY algorithm is: .
- (4)
SA Algorithm
Similarly to the DIVY algorithm, the total time complexity of the SA algorithm is , where LM is the length of the Markov chain and ItSA is the maximum iterations.
In summary, the time complexity of the entire algorithm is . In the practical industrial scenario of waste incineration plants, the number of cranes r is a fixed small constant. Meanwhile, the population size and iteration times are adjusted sub-linearly with the task scale, and their growth rate is lower than that of n. Therefore, the proposed algorithm fully meets the computational requirements of near real-time scheduling when deployed with a rolling horizon strategy.
5. Experimental Analysis
5.1. Simulation Environment and Instance Description
In this section, a series of experiments are conducted to verify the effectiveness of the proposed algorithm. All experiments are conducted on a PC (AMD Ryzen 5 3600 6-Core Processor, 16 GB RAM; AMD, Santa Clara, CA, USA) in MATLAB (Version R2021b, The MathWorks, Inc., Natick, MA, USA). The MILP was solved using Gurobi Optimizer (Version 11.0, Gurobi Optimization, LLC, Beaverton, OR, USA). To ensure the practical validity of the experiment, all simulation parameters, warehouse layouts, and task cases are derived from real-world operational data collected from a waste incineration plant located in Guangzhou, China. The specific settings are as follows: the number of cranes (r), feeding ports (s), and unloading ports (p) is set to 3, 3, and 9, respectively; the length, width, and height of the waste pit are 76, 25, and 30, respectively; the operating range of crane 1 along the rail direction is 0–60 m, that of crane 2 covers the entire rail, and that of crane 3 is 15–76 m; the moving speed of the cranes’ gantries is v = 1.0 m/s; the time required to grab waste for the feeding task is 50 s; the time required to grab and place waste for the transfer task is 20 s; the turning task takes 40 s; and the safety distance between cranes’ gantries is dsafe = 3 m.
5.2. Validity Verification and Sensitivity Analysis of the DES Module
The DES module is the sole basis for fitness calculation, and its output accuracy and logical consistency directly determine the reliability of all experimental conclusions. To fully verify the validity of the simulation module, we conduct deterministic benchmark verification and sensitivity analysis of the simulation output as follows.
- (1)
Deterministic Benchmark Case Verification
We construct three minimal-scale deterministic benchmark cases with known theoretical true values, to verify the consistency between the DES module output and the theoretical calculation results. The verification results in
Table 3 and
Figure 9 show that the output of the DES module is completely consistent with the theoretical true value for all cases. It directly proves that the module can accurately reproduce the crane operation rules, spatial interference constraints, and objective function calculation logic defined in this paper.
- (2)
Sensitivity Analysis of Simulation Output
To further verify the logical consistency of the DES module, we conduct a sensitivity analysis of the simulation output with respect to the safety distance
dsafe that directly affects the simulation operation. We test an instance with 25 tasks, varying the safety distance from 2 m to 6 m. Each data point represents the mean of 20 independent runs, as shown in
Figure 10. It shows a stable upward trend with the increase in safety distance, which is completely consistent with the theoretical expectation: a larger safety distance will lead to more frequent spatial interference and longer avoidance travel distance. This result further verifies that the spatial interference judgment and avoidance position calculation logic of the DES module are implemented correctly.
5.3. Algorithmic Parameter Tuning and Sensitivity Analysis
The main parameters to be determined for the DIVY-SA algorithm include: Pa and Pb (for the DIVY component), initial temperature, terminal temperature, Markov chain length, and cooling rate (for the SA component). For the DIVY component, we employ a comprehensive testing approach. Using a case with 25 tasks, the population size of the DIVY algorithm is set to 30, and the maximum number of iterations is set to 100. We evaluate parameter combinations based on the Average Deviation Rate (ADR), Optimal Deviation Rate (ODR), and Standard Deviation (SD) over 20 independent runs. For the calculation of the deviation rate, the optimal solution is represented by the LB obtained by solving the relaxed problem with Gurobi.
Initially, we tested
Pa and
Pb within [0.1, 0.5]. The heatmap of experimental data (
Figure 11) shows that all three metrics worsen as
Pa and
Pb increase, indicating that smaller parameter values are preferable. We therefore narrowed the search to the [0.01, 0.1] interval (
Figure 12). Within this range, while performance differences were minor, the ADR heatmap indicated that lower values were concentrated in the lower-right area. Synthesizing these results, we ultimately determined the optimal parameters:
Pa = 0.03 and
Pb = 0.07.
Regarding the parameters for the SA component, the Markov chain length is set to four times the DIVY’s population size. This ratio is selected to strike a balance between solution quality and computational efficiency. Since
LM and
Npop are adjusted according to the task scale, linking
LM to
Npop ensures that the intensity of the fine-grained search scales adaptively with the problem complexity. This allows the SA component to reach a quasi-equilibrium state at each temperature without incurring excessive computational costs. The initial temperature is set to 1000, a value chosen to be of the same order of magnitude as the objective function. To determine the optimal cooling parameter, we employed the same tuning methodology used for the DIVY’s parameters. The results are presented in
Table 4, which indicates that the optimal value for the cooling parameter is 0.9.
The results indicate that the algorithm’s performance is minimally affected when Pa is within 0.1, Pb falls within [0.06, 0.09], and α remains in the range of [0.8, 0.95]. This demonstrates that the DIVY-SA algorithm is not highly sensitive to its hyper-parameter configurations, ensuring stable performance in practical industrial applications.
5.4. Performance Comparison with Other Algorithms
To comprehensively evaluate the performance of the proposed DIVY-SA algorithm, we conducted comparative experiments. The experimental task scales range from 25 to 200 tasks, corresponding to a real-world scheduling time horizon of 15 min to 2 h. Due to the specificity of the problem, there are currently no specialized algorithms available for direct comparison. Therefore, we selected several representative general algorithms as benchmarks: GA, SA, GWO and Discrete Particle Swarm Optimization (DPSO). In addition, to validate the effectiveness of the specific improvements, we included GA-SA, IVYA, and DIVY in the comparison. The GA-SA algorithm is a two-stage variant replacing DIVY with GA. GWO and IVYA are designed for continuous optimization problems. Instead of discretizing them, we only map continuous values to discrete task sequences when calculating fitness.
To ensure a fair comparison among the meta-heuristic algorithms, we unified the computational budget for all algorithms. Since the simulation-based fitness calculation is the most time-consuming part of the solving process, we controlled the total number of fitness evaluations to be identical across all algorithms. Therefore, the computational complexity of all comparative algorithms is theoretically of the same order. The detailed settings for population size and maximum iterations for different task sizes are listed in
Table 5. The specific parameter settings for the comparative algorithms, chosen based on their effectiveness in similar combinatorial problems, are detailed in
Table 6.
To evaluate the stability and statistical significance of the proposed DIVY-SA algorithm, each experimental instance was executed independently for
m = 20 runs. The experimental results are reported as the mean value with a 95% Confidence Interval (CI):
where
Avg represents the average objective value;
SD denotes the sample standard deviation;
is the
t-value from the
t-distribution.
For the Gurobi solver, since the objective function does not optimize for makespan, the solver may push tasks without time window constraints to be executed at a very late time, which is impermissible in practical operations. To address this issue, a negligible regularization term (with = 10−10) was added to the objective function in the solver implementation. Due to its minimal weight, it does not alter the optimality of the primary objectives. Furthermore, since the MILP model relaxes certain interference constraints, the solution obtained by Gurobi may be infeasible. Therefore, these solutions were processed through the simulation module to obtain corrected results, denoted as ‘Gurobi (Cor.)’.
Table 7 summarizes the comparative performance of the proposed algorithm and other benchmark methods across various task sizes. Key observations from the experimental data are listed below.
- (1)
Comparison with Exact Solver
For small-scale instances, Gurobi provides a tight LB based on the relaxed model. However, the corrected solution yields a higher objective value. The best solution found by the DIVY-SA algorithm is inferior to the LB but superior to the corrected solution. The gap to LB is minimal (e.g., only 0.48% for the 25-task case, 2.17% for the 50-task case). It not only demonstrates the superior search capability of the DIVY-SA algorithm but also validates the effectiveness of the proposed heuristic assignment method. As the problem scale expands to 100 tasks, with the solver restricted to a 5% optimality gap, the corrected solution proves infeasible as it induces excessive delays for certain tasks. For the large-scale instance of 200 tasks, Gurobi fails to obtain a valid solution within the time limit. In contrast, the proposed algorithm exhibits significantly higher computational efficiency and meets the requirements for near real-time scheduling.
- (2)
Comparison with General Metaheuristics Algorithms
The proposed DIVY-SA demonstrates superior robustness and efficacy. It consistently achieves the best objective values and the smallest SD across all cases. For example, in the 50-task instance, DIVY-SA achieves a gap to LB of 2.17%, significantly outperforming GA (10.84%), SA (15.79%) and DPSO (13.00%). Furthermore, while the DIVY yields a slightly inferior best solution compared with GA and SA in the 200-task instance, it consistently achieves superior average performance and stability, confirming the framework’s effectiveness. The IVYA outperforms GWO, demonstrating the strong search capability of the underlying ivy-based mechanism. Moreover, all calculated p-values are less than 0.05, demonstrating that DIVY-SA significantly outperforms other algorithms.
- (3)
Validation of Improvement Strategies
The ablation study confirms the effectiveness of the specific improvement strategies. The results validate the discretization strategy, as DIVY significantly narrows the optimality gap compared with the continuous IVYA, reducing it from 23.22% to 7.74% in the 50-task instance. Moreover, the hybrid two-stage mechanism proves superior to single-stage approaches, as DIVY-SA outperforms both DIVY and SA individually by further reducing the gap to 0.48% for 25 tasks and 2.17% for 50 tasks. All calculated p-values are less than 0.05, demonstrating that DIVY-SA significantly outperforms DIVY and SA.
- (4)
Computational Time Analysis
Since the simulation module is the most time-consuming part of the algorithm and all metaheuristics perform the same number of objective function evaluations, their runtimes are nearly identical for small-scale tasks. In large-scale cases, higher population quality obtained during the search process accelerates the simulation, making DIVY-SA faster than other algorithms. Furthermore, the runtimes of metaheuristic algorithms are significantly shorter than exact algorithms. The runtime of the algorithm is maintained within 10% of the scheduling window. When combined with a rolling horizon strategy, it meets the requirements for near real-time scheduling in the waste incineration plant.
5.5. Comparison with Existing Manual Scheduling Algorithms
The waste incineration plant in Guangzhou, China, which provided the data for this study, currently relies on a manual scheduling method for its daily operations. The specific scheduling rules adopted by this plant are defined as follows:
- (1)
The middle crane (i.e., Crane 2, in the three-crane configuration) is primarily responsible for handling feeding tasks. Meanwhile, the cranes on both sides (i.e., Crane 1 and Crane 3) are assigned to handle either transferring tasks or turning tasks, depending on the positions of their respective raw waste areas and fermenting waste areas.
- (2)
For feeding tasks, tasks are sorted into an execution sequence following the first-come-first-served (FCFS) principle, sorted by the tasks’ respective start times.
- (3)
For transferring and turning tasks, tasks are sorted into an execution sequence following the proximity-first principle (i.e., prioritizing tasks closer to the crane’s current position), according to the tasks’ start positions in relation to the crane’s current position.
This manual scheduling method simplifies the scheduling complexity but tends to cause excessive interference between cranes, increasing energy waste due to unnecessary movement. To validate the superiority of the proposed scheduling scheme, we conduct experiments to compare the total traveling distances of cranes under different scheduling algorithms. The five test cases have a total task scale of 10, 15, 25, 35, and 50, respectively, with a roughly balanced proportion of feeding, transfer and turning tasks in each case. As the number of tasks increases, the spatial distribution of tasks becomes more uniform. The experimental results are presented in
Table 8.
It shows that the proposed scheme outperforms the manual scheduling method in all five cases, achieving an average reduction of 22.19% in the total traveling distance of cranes. Taking Case 3 as a representative example, the trajectories of the cranes’ gantries under the two scheduling methods are shown in
Figure 13. The makespan of the proposed scheme is shorter, indicating that it not only reduces energy consumption but also improves efficiency.
5.6. Crane Quantity Selection
We also compare the objective function values under different configurations with varying numbers of cranes. When more cranes are deployed, interference between cranes becomes more frequent, potentially leading to an increase in unnecessary crane movement. Conversely, when fewer cranes are deployed, the efficiency of waste disposal decreases, which may lead to some tasks not being processed in time and increasing the tasks’ delay times. Therefore, using the five cases previously discussed in
Section 5.5, we evaluate the crane travel distance across varying numbers of cranes (
Table 9).
The experimental results indicate that among the five cases, the optimal total crane traveling distance for two cases is achieved under the two-crane configuration, while the optimal value for the remaining three cases is achieved under the three-crane configuration. This result indicates that the selection of the optimal crane quantity depends on the specific characteristics of task scenarios. For scenarios with fewer tasks and relatively concentrated operating areas, two cranes can not only effectively avoid task backlogs in the single-crane configuration but also minimize the additional energy consumption induced by mutual interference in the three-crane configuration. For scenarios with more tasks and relatively uniform task locations, although three cranes may lead to a slight increase in energy consumption due to mutual interference, they greatly reduce tasks’ delay times by sharing tasks, which outperforms the two-crane configuration.
6. Conclusions
This study addresses the multiple waste crane scheduling problem in waste incineration plants, which is constrained by task heterogeneity, dynamic spatial interference, and time windows of feeding tasks. To minimize the total crane traveling distance and task delay, we first establish an MILP model that provides a theoretical benchmark for this NP-hard problem. Then, a two-stage DIVY-SA hybrid algorithm is proposed, which combines the global broad-range search capability of the discretized IVYA and the fine-grained local optimization capability of the SA. A heuristic task assignment scheme and a DES module are also developed to accurately evaluate scheduling performance while accounting for dynamic crane interference.
Experiments based on real-world operational data from a waste incineration plant verify the effectiveness of the proposed method. The results show that DIVY-SA consistently outperforms representative metaheuristic algorithms across task scales from 25 to 200, corresponding to a scheduling horizon of 15 min to 2 h. The algorithm runtime ranges from 15.04 s to 652.81 s, which fully meets the requirements of near-real-time industrial scheduling via a rolling horizon strategy. Compared with the manual scheduling method widely used in practice, the proposed scheme reduces the average total traveling distance of cranes by 22.19%, achieving significant energy savings and operational efficiency improvement. The model and algorithm can be directly deployed in waste incineration plants with different pool sizes, crane quantities, and production scales with minimal parameter tuning.
However, this study is based on deterministic static scheduling assumptions and lacks flexibility. It must integrate a rolling horizon strategy to handle practical uncertainties, such as crane failures, urgent task arrivals, and fluctuating incineration feeding demand. Future research will extend this framework to dynamic stochastic scheduling scenarios to enhance the robustness of the scheduling system against unforeseen disruptions. Additionally, reinforcement learning can be utilized in the future to develop rescheduling algorithms, enabling effective scheduling management over longer time horizons. Moving beyond the current offline DES module, future research will develop a digital twin-driven platform that integrates the proposed models and algorithms for real-time, closed-loop scheduling. This system will enable a seamless cycle from data acquisition and virtual verification to physical execution and feedback optimization.