You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

28 October 2025

Emergency Relief Material Distribution Path Optimization Under Multiple Constraints

,
,
,
,
and
1
School of Electronic and Information Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
The Sixty-Third Research Institute, National University of Defense Technology, Nanjing 210007, China
3
Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.

Abstract

To overcome the limitations of traditional methods in emergency response scenarios—such as limited adaptability during the search process and a tendency to fall into local optima, which reduce the overall efficiency of emergency supply distribution—this study develops a Vehicle Routing Problem (VRP) model that incorporates multiple constraints, including service time windows, demand satisfaction, and fleet size. A multi-objective optimization function is formulated to minimize the total travel time, reduce distribution imbalances, and maximize demand satisfaction. To solve this problem, a hybrid deep reinforcement learning framework is proposed that integrates an Adaptive Large Neighborhood Search (ALNS) with Proximal Policy Optimization (PPO). In this framework, ALNS provides the baseline search, whereas the PPO policy network dynamically adjusts the operator weights, acceptance criteria, and perturbation intensities to achieve adaptive search optimization, thereby improving global solution quality. Experimental validation of benchmark instances of different scales shows that, compared with two baseline methods—the traditional Adaptive Large Neighborhood Search (ALNS) and the Improved Ant Colony Algorithm (IACA)—the proposed algorithm reduces the average objective function value by approximately 23.6% and 25.9%, shortens the average route length by 7.8% and 11.2%, and achieves notable improvements across multiple performance indicators.

1. Introduction

The optimization of emergency supply distribution routes is a critical challenge in response to natural disasters, public health crises, and industrial accidents. Under limited resources and strict time constraints, the efficient scheduling of emergency supplies is essential to ensure the orderly progress of relief operations. This problem can be abstracted as a Vehicle Routing Problem (VRP), which is a classic topic in combinatorial optimization [1,2]. With the introduction of multiple constraints in complex environments, the VRP has become an even more challenging NP-hard problem. As the scale of the problem grows and the constraints increase in complexity, the solution difficulty rises accordingly, underscoring the urgent need for efficient optimization strategies.
In recent years, the role of VRP in emergency response has become increasingly critical. For example, during the catastrophic floods in Henan Province in 2021, 12 national highways were blocked, leaving several regions isolated [3]. The traditional scheduling methods faced significant delays in material delivery, highlighting the need for more responsive routing strategies. Similarly, during the 2023 Turkey earthquake, surface ruptures extended over 300 km, severely disrupting traditional route planning models for emergency supply transportation [4]. These cases clearly illustrate the limitations of traditional routing methods under complex constraints and emphasize the need for more intelligent and adaptive optimization approaches.
Against this background, traditional routing methods in emergency environments face three major challenges. (1) Limited adaptability: conventional approaches typically rely on fixed heuristic strategies in complex environments and lack global search mechanisms and dynamic parameter adjustments. As a result, they often become trapped in local optima during natural disasters, thereby reducing the overall scheduling effectiveness. (2) Response delays: Traditional methods depend on predefined routing strategies and lack mechanisms for adaptive adjustments to dynamic conditions. Consequently, they cannot promptly reconstruct transportation plans, which leads to insufficient responsiveness. (3) Low resource utilization: Conventional approaches struggle to handle multi-constraint optimization problems, often resulting in redundant services, material waste, and uneven distribution, which ultimately undermine transportation efficiency.
Previous studies have proposed various algorithms for VRP in both emergency and general contexts, which can broadly be categorized into heuristic methods and machine learning–based strategy optimization approaches. Among traditional algorithms, Dorigo et al. [5] first introduced the ant colony optimization (ACO) algorithm and successfully applied it to the classical Traveling Salesman Problem (TSP), laying the theoretical foundation for subsequent VRP research. Building on this, Kubil et al. [6] developed a multi-objective ACO algorithm for multi-depot VRP variants, while Miao et al. [7] incorporated adaptive mechanisms to enhance global search performance. Zhang et al. [8] proposed the MSALNS algorithm to solve the flexible time window and delivery location VRP of customer satisfaction, and introduced backtracking and correlation operators and simulated annealing acceptance mechanism to improve search performance. Cao et al. [9] addressed cold-chain distribution under traffic congestion by formulating a time-window-constrained routing model and designing an improved genetic algorithm to enhance solution stability and scheduling efficiency. Liu et al. [10] proposed a multi-objective emergency distribution model that considered the urgency of supplies during epidemic outbreaks and introduced a hybrid intelligent optimization algorithm to improve computational efficiency, demonstrating advantages in cost reduction and vehicle utilization. Yu et al. [11] and Tu et al. [12] applied simulated annealing and its improved variants to VRP in omni-channel retail distribution and under multiple time-window constraints, respectively. Wei et al. [13] modeled the routing problem for emergency distribution during epidemic lockdowns as a generalized VRP with complex constraints and proposed an improved ACO algorithm. Compared with standard ACO, their method achieved significantly better routing performance, verifying its effectiveness in disaster scenarios.
With the advancement of machine learning, an increasing number of studies have integrated machine learning into the VRP framework, leveraging adaptive decision-making capabilities to enhance intelligence and generalization. James et al. [14] tackled the route selection problem in last-mile delivery by employing reinforcement learning to model vehicle–environment interactions, thereby achieving optimal routing. Zong et al. [15] proposed an end-to-end reinforcement learning framework for the VRPTW, incorporating feature encoding and a time penalty mechanism to improve solution quality. Silva et al. [16] applied a multi-agent reinforcement learning framework to VRP with time-window constraints, enabling agents to dynamically adjust strategies based on environmental feedback. Kang et al. [17] addressed the routing of electric vehicles providing emergency power supply to shelters after earthquakes and designed a deep reinforcement learning algorithm with a multi-head attention mechanism, which demonstrated strong responsiveness and routing performance in complex environments. Arishi et al. [18] formulated a multi-agent VRP framework for multi-depot distribution scenarios, thereby enhancing system-level collaborative scheduling. Ahamed et al. [19] and Liu et al. [20] employed deep Q-networks and decision models, respectively, for supply chain route optimization and logistics efficiency improvement, both of which demonstrated strong policy-learning capabilities.
Existing research on VRP shows a trend toward the diversification of algorithm design. Some researchers have integrated metaheuristics with other algorithms to improve the adaptability and solution efficiency. Therefore, hybrid ALNS algorithms have become popular research topics. Recent research has focused on embedding learning mechanisms or other metaheuristic methods into the ALNS framework to improve the operator selection and search strategies. Reijnen et al. [21] proposed an adaptive large-neighborhood search algorithm (DR-ALNS) based on deep reinforcement learning. By learning the search state, it dynamically selects operators, adjusts parameters, and controls acceptance criteria, thereby adaptively optimizing the search process. Ma et al. [22] proposed an improved adaptive large neighborhood search algorithm (IALNS-SA) combined with simulated annealing for the vehicle routing problem with simultaneous delivery and pickup time windows. The algorithm improves the robustness by introducing supplementary damage and repair operators, and adopts the SA acceptance criterion to avoid falling into local optimality. Liu et al. [23] proposed an adaptive hybrid genetic and large neighborhood search algorithm (AHGSLNS) for a multi-attribute vehicle routing problem. By combining a genetic search with the ALNS, the parameters and acceptance criteria were adaptively adjusted to improve the stability and convergence of the algorithm.
These studies have improved the intelligence and adaptability of ALNS at the algorithmic level, but have mainly focused on improving operator selection and search mechanisms. In contrast, the ALNS-PPO algorithm proposed in this paper not only integrates the adaptive mechanism into the search strategy but also combines it with demand point satisfaction-oriented target modeling, expanding the research scope of the hybrid ALNS from both the model and algorithm dimensions.
Although existing VRP research has performed well in general scenarios, in complex environments and multi-constraint emergency scenarios, traditional algorithms that rely on fixed-domain search strategies are prone to falling into local optimality and have difficulty obtaining high-quality solutions. They are also highly dependent on manual experience for parameter settings and have limited adaptability and generalization capabilities. To overcome these limitations, this study aims to establish a multi-constrained emergency material scheduling and routing optimization model and proposes a hybrid optimization framework that integrates Adaptive Large Neighborhood Search (ALNS) with Proximal Policy Optimization (PPO) to enhance both the global search capability and adaptive decision-making ability in emergency logistics scheduling. By leveraging deep reinforcement learning mechanisms, the framework dynamically adjusts parameters to enable route reconstruction and adaptive optimization, thereby significantly enhancing the adaptability in complex disaster environments. Regarding the practical implications of this research, from an economic perspective, this model can significantly reduce transportation costs and material waste during large-scale relief operations by improving vehicle utilization and reducing redundant trips. From a social perspective, improved route efficiency can help deliver essential supplies to disaster-stricken areas more quickly and equitably, thereby improving the overall responsiveness and equity of disaster relief efforts. The results suggest that this approach has the potential to provide theoretical and practical value to real-world emergency logistics systems. The main contributions of this study are as follows.
(1)
Multi-constraint model construction: A VRP model was developed that incorporated multiple depots, vehicles, demand points, and real-world constraints, providing a theoretical foundation and modeling framework for route planning in complex emergency scenarios.
(2)
Collaborative hybrid optimization framework combining ALNS and PPO. Unlike existing reinforcement learning-based VRP research, using reinforcement learning as a standalone solver, this study establishes a dynamic interaction mechanism in which the PPO adaptively learns and adjusts the neighborhood search strategy in ALNS. Collaborative integration enhances global exploration capabilities and supports adaptive decision-making under multiple operational constraints.
(3)
Comprehensive evaluation and validation: The proposed approach was systematically evaluated using multiple public VRP benchmark datasets and compared with traditional heuristic algorithms. Experimental results demonstrate that it achieves superior routing performance and significant improvements across multiple performance indicators, underscoring its strong generalization capability and practical application value.

2. Problem Description and Modeling

This section presents a detailed description of the modeling process for the emergency supply distribution vehicle routing optimization problem, including the formulation of objective functions, the definition of constraints, and the specification of relevant parameters. The objective is to ensure that the model comprehensively and accurately captures the complexity and requirements of real-world emergency distribution scenarios.

2.1. Problem Description

In the aftermath of sudden disasters such as earthquakes, floods, or epidemic outbreaks, the rapid and efficient distribution of relief materials has become a critical component of emergency response operations. Unlike conventional logistics, emergency relief distribution is subject to multiple constraints. Its primary objective is not only to minimize transportation time but also to ensure the prompt and reliable delivery of critical supplies to affected regions in order to support life-saving operations. Therefore, to better characterize the logistical features of such emergency scenarios, it is necessary to establish a multi-constraint path optimization model for emergency relief material distribution.
Given several emergency supply depots and multiple affected demand points, the objective is to plan vehicle distribution routes under the constraints of vehicle capacity, road accessibility, and delivery time windows, so that materials are delivered efficiently to target areas in the shortest possible time, producing an optimal or near-optimal emergency distribution plan.
Model Assumptions:
  • The demand quantities and location coordinates of all the demand points are known prior to scheduling and remain fixed throughout the optimization process.
  • Each demand point has a strict service time window. If the delivery cannot be completed within the specified time, the task is considered unsuccessful and the distribution is invalid.
  • All vehicles have identical loading capacities and travel speeds.
  • Each vehicle is prohibited from returning to the depot for replenishment during a single trip; therefore, the entire loading and delivery processes must be completed in one operation.
  • The inventory of each depot is known and non-replenished, and the depot locations are randomly distributed.
  • The demand at each demand point must be fully served by a single vehicle and cannot be split among multiple vehicles.
  • Each vehicle can carry multiple types of supplies. However, the total load must not exceed its maximum capacity.
  • Both waiting time and service time at demand points are considered during transportation.
These simplified assumptions are adopted to preserve the model’s tractability and emphasize the core decision logic of the proposed optimization framework. Although this abstraction inevitably reduces the level of real-world detail, it allows for a clearer interpretation of the key routing relationships and constraint interactions.
Figure 1 shows a typical schematic of vehicle distribution route planning for emergency relief materials. The diagram illustrates the core workflow of emergency material distribution. Multiple vehicles depart from the depots, sequentially deliver materials to demand points while satisfying the constraints, and return to the depot, providing a visual representation of the VRP scenario addressed in this study.
Figure 1. Schematic diagram of vehicle distribution route planning for emergency relief materials.

2.2. Modeling

To effectively address the practical problem of emergency supply vehicle routing optimization, a mathematical model is constructed based on the above assumptions, systematically describing and constraining the key factors in the emergency distribution process. By optimizing vehicle routes, the model minimizes a multi-objective function to ensure the efficient delivery of supplies from depots to demand points. In the modeling process, real-world constraints are comprehensively incorporated to allocate supply and vehicle resources effectively, thereby ensuring the timely delivery of materials and enhancing emergency response capability under disaster scenarios.
The parameters used in the model are defined as follows in Table 1:
Table 1. Meaning of parameters in the model.
All parameters are expressed in dimensionless form, and time and distance can be interpreted in minutes and kilometers, respectively, when applied to real emergency logistics cases.
Considering the problem description and the definitions of the relevant parameters, the objective functions and constraints are formulated as follows:
  • Objective Function
min Z = α k K ( i , j ) E T i j x i j k + γ i N L k 1 K i N f i + δ i N f i k K c R k f c
The proposed optimization model aims to improve the efficiency and rationality of emergency supply distribution by considering multiple objectives. In particular, the objective function integrates three key components:
  • k K ( i , j ) E T i j x i j k , which denotes the total travel time of all vehicles. Optimizing delivery routes while minimizing unnecessary detours and empty trips helps improve both overall transportation efficiency and emergency response speed.
  • i N L k 1 K i N f i denotes the total average deviation in the load distribution among vehicles. Here 1 K i N f i is the expected load capacity per vehicle, which ensures that every demand point receives a sufficient supply of relief materials.
  • i N f i k K c R k f c denotes the degree of demand satisfaction. The difference between the total demand of i N f i and the total number of demand points served k K c R k f c . The objective of maximizing the fulfillment of actual demand at each point and ensuring that material distribution adequately meets the required needs.
To balance the model’s multiple objectives, weighting coefficients α, γ, and δ are introduced. These coefficients are normalized such that α + γ + δ = 1 to ensure comparability across different indicators. Considering that, in emergency scenarios, “the needs of the disaster area must be met first” while also balancing transportation efficiency and vehicle resource waste, the weights are set to α = 0.3, γ = 0.2, and δ = 0.5.
The proposed model considers transportation efficiency, balance of material allocation, and demand satisfaction in route optimization. It integrates three indicators—total vehicle travel time, load distribution deviation, and unsatisfied demand—into a single aggregated objective function via a weighted summation, aiming to achieve overall optimization and a balanced trade-off among multiple objectives.
2.
Constraints
The model must comply with the following fundamental constraints to guarantee the rationality and feasibility of emergency material distribution:
  • Constraint ensuring that each vehicle route forms a closed loop
    i N D x d i k = j N D x i j k = 1 , ( i , j ) E , k K
  • Constraint ensuring that the load carried by each vehicle does not exceed its capacity
    i N f i y i k Q k , f D ,   k K
  • Constraint ensuring that each delivery is made within the specified time window
    e i t i k l i , i N , k K
  • Constraint ensuring that the quantity of materials supplied does not exceed the available inventory
    k K i N f i y i k S d , f D , i N
  • Constraint ensuring that each demand point is served by exactly one vehicle
    k K y i k = 1 ,   i N
In the proposed model, Equation (2) ensures that each vehicle departs from and returns to the distribution center, maintaining route closure and operational feasibility. Equation (3) enforces that the vehicle load does not exceed its capacity, reflecting practical transportation limitations. Equation (4) requires that service at each demand point be completed within the specified time window, thereby satisfying the timeliness requirements of emergency logistics. Equation (5) limits the quantity of materials delivered in a single trip to the available inventory, ensuring supply feasibility. Equation (6) guarantees that each demand point is served by exactly one vehicle, promoting efficient resource utilization. Collectively, these constraints establish a multi-constraint vehicle routing optimization model for emergency material distribution, providing a robust theoretical foundation for the subsequent application of optimization algorithms.

3. Algorithm Design

This study proposes a hybrid deep reinforcement learning framework that integrates Adaptive Large Neighborhood Search (ALNS) with Proximal Policy Optimization (PPO), thereby enhancing both the adaptability and solution efficiency of the algorithm. The theoretical basis for the combination of ALNS and PPO stems from their complementary strengths: ALNS overcomes PPO’s inherent limitation of low solution feasibility in complex constraint scenarios, while PPO addresses ALNS’s flaw of rigid parameter settings leading to local optima. PPO, based on the policy gradient theory in deep reinforcement learning, leverages solution quality feedback to update its policy, which aligns with the iterative “break-and-repair” mechanism of ALNS, enabling ALNS to adjust core search parameters in real time based on the current solution state.
Figure 2 presents the overall flowchart, illustrating the main steps and execution process. The framework primarily consists of four stages: initial solution generation, feature extraction, PPO-based policy parameter output, and subsequent parameter update and solution optimization. Specifically, initial solutions are generated using the ALNS algorithm, while PPO is employed for parameter adjustment. The solution is iteratively refined until the termination criteria are satisfied, at which point the final results are obtained.
Figure 2. Overall flowchart of the ALNS-PPO hybrid algorithm.

3.1. Basic Algorithm Modules

3.1.1. Adaptive Large Neighborhood Search (ALNS)

The ALNS algorithm employs a destroy–repair mechanism in combination with a neighborhood generation strategy [24]. This study introduces three types of destruction operators: random removal (based on 10% removal), long-term conflict removal, and distance cluster removal, and three types of repair operators: minimum-redundant-time insertion, greedy insertion, and nearest-neighbor insertion. Each destruction–repair pair D u , R v constitutes a candidate neighborhood. During the search process, the PPO module dynamically adjusts its selection probability based on recent performance feedback. The repair process enforces the vehicle capacity and time window constraints by recalculating the earliest and latest feasible arrival times after each insertion. This adaptive operator mechanism enables ALNS to explore diverse neighborhood structures while maintaining feasibility, thereby improving overall search capabilities.
As shown in Figure 3, during each iteration, the destroy operator first selects an appropriate destruction operator to remove a subset of nodes from the current solution, resulting in a partial solution. The repair operator then selects an appropriate repair operator to reinsert the removed nodes into the partial solution, thereby generating a new candidate solution. By collaboratively utilizing multiple destroy and repair operators, the ALNS algorithm can dynamically switch neighborhood structures throughout the search process. In particular, the destroy operator can adaptively adjust the removal strategy according to the real-time system requirements, whereas the repair operator considers multiple constraints when reinserting nodes.
Figure 3. Schematic of the Destroy–Repair mechanism in the ALNS algorithm.

3.1.2. Proximal Policy Optimization (PPO)

Proximal Policy Optimization (PPO) is an efficient and stable policy-gradient deep reinforcement learning algorithm that employs a dual policy–value network architecture to improve the stability of policy updates in complex routing optimization scenarios [25].
In this study, the PPO module primarily functions as an adaptive controller within the ALNS framework rather than as a standalone end-to-end deep policy network. The PPO module employs a simplified actor–critic dual-network architecture. Both subnetworks consist of two fully connected layers with 48 and 24 hidden nodes, and the activation function used is ReLU. The output layer of the actor network generates five continuous control parameters, α = [ r d , w r , w d , β , δ ] , corresponding to the key regulatory variables in the ALNS algorithm, and the critic network outputs a scalar state value that is used to evaluate the expected return of the current policy. To avoid gradient instability caused by inconsistent numerical scales, all the input features are normalized to the [0, 1] interval before entering the network. For action parameterization, the continuous actions produced by the PPO are represented in a bounded continuous space and mapped to the feasible range of ALNS parameters using a sigmoid function and linear scaling, ensuring that each action update corresponds to a valid search control variable.
The PPO algorithm implements an enhanced policy-gradient method that includes the following key steps:
Step 1: Data Collection. The interaction data are collected by implementing the current policy in the environment. The dataset includes states, actions, rewards, and the corresponding next states.
Step 2: Advantage Estimation. The GAE method is employed to evaluate the relative value of each action with respect to the average policy, facilitating subsequent gradient optimization. The core GAE formula is as follows:
A ^ t GAE ( γ , λ ) = l = 0 ( γ λ ) l δ t + l
γ denotes the discount factor that reflects the importance of long-term rewards.
λ is the GAE parameter that controls the bias-variance trade-off and δ t + l represents the temporal-difference error.
Step 3: Objective Function Optimization. The PPO algorithm utilizes a specifically designed objective function to guide policy updates in reinforcement learning. Its primary purpose is to constrain the step size of policy updates and prevent abrupt policy changes, thereby enhancing training stability and efficiency. The PPO objective is defined as follows:
L C L I P ( θ )   = E ^ t min r t ( θ ) A ^ t , clip ( r t ( θ ) , 1 ϵ , 1 + ϵ ) A ^ t
Here, r t ( θ ) denotes the probability ratio, that is, the ratio of the action probability under the current policy to that under the old policy. A ^ t represents the estimated advantage function, which indicates the relative value of taking a specific action compared to the average action in the current state.
Step 4: Policy Update. In the PPO algorithm, the policy parameters are updated using gradient descent. The overall loss function is first computed, after which the parameters θ are iteratively adjusted through gradient-based optimization, gradually approaching the optimal policy while maintaining update stability.

3.2. Design of the ALNS-PPO Hybrid Framework

The existing hybrid ALNS algorithms still exhibit two key limitations: first, the passive nature of operator selection and parameter adjustment, which creates a disconnect between the search direction and the core requirements of emergency tasks; and second, the single-dimensional handling of constraints, which limits adaptability to the complex and interdependent constraints typical of disaster scenarios. To overcome these challenges, the ALNS-PPO framework proposed in this study introduces two major innovations: (1) a multi-constraint collaborative control mechanism, which ensures the satisfaction of key constraints while dynamically balancing vehicle capacity and inventory limitations; and (2) a scenario-aware policy network, which extends the PPO feedback signal from pure search performance to multi-objective comprehensive benefits. This design enables the adaptive adjustment of operator weights and perturbation intensity to directly serve the core objectives of emergency logistics optimization.
To further illustrate the implementation logic of this collaborative optimization process, this study presents a framework design from two perspectives: a ‘hybrid architecture coordination mechanism’ and a ‘perception and feedback mechanism’.

3.2.1. Hybrid Architecture Coordination Mechanism

The proposed ALNS-PPO hybrid algorithm uses a hierarchical architecture to integrate traditional heuristic search methods with deep reinforcement learning. The ALNS layer quickly generates high-quality routing solutions through a destroy–repair mechanism, while the PPO layer outputs policy parameters via the policy network to guide the neighborhood search direction and scope within ALNS.
Figure 4 illustrates the core coordination mechanism of the proposed ALNS-PPO hybrid optimization framework, which systematically demonstrates the collaboration between deep reinforcement learning and heuristic search algorithms. The framework builds the main routing search structure using ALNS and integrates the PPO policy network as a control module. This framework adaptively guides the search process through state perception and reward feedback mechanisms.
Figure 4. ALNS-PPO hybrid optimization architecture diagram.
(1)
Input phase:
Within this framework, the ALNS functions as the core component for route optimization, combining destruction and repair operators to iteratively perturb and reconstruct the current delivery routes. The system first extracts key emergency logistics information, including the number and geographic locations of warehouses, vehicle capacities, demand data for each demand point, and their corresponding time windows. Based on this information, ALNS utilizes its embedded route construction logic to generate an initial feasible delivery plan, which serves as the input for subsequent optimization within the framework.
(2)
Decision support phase:
Before each iteration, the PPO policy network perceives and analyzes the current solution state, encoding it into a state vector embedded within the Actor–Critic-based search framework. The Actor module extracts key routing features and outputs policy parameters—such as perturbation intensity and operator weights—that regulate the adaptive behavior of the ALNS process. These parameters dynamically adjust the execution strategies of the destruction and repair operators, thereby enhancing the effectiveness of exploration and guidance throughout the search.
After each path update, ALNS transmits the new solution structure to the evaluation module, which assesses both the feasibility and the quality of the solution. Based on the observed improvements, the system generates feedback signals used for subsequent deep reinforcement learning updates. The evaluation outcomes, together with the corresponding states and actions, form experience data that are stored in the experience pool for Critic network training.
The Critic module evaluates the current policy using a value function and provides feedback to improve the Actor’s decision making. Through this continuous cycle of state evaluation, policy refinement, and feedback integration, the PPO network offers dynamic and adaptive decision support, effectively guiding the ALNS toward high-quality solutions throughout the search process.
The overall framework illustrates the tight integration between deep reinforcement learning and heuristic search. The PPO module perceives the search state, generates control signals, and updates the policy based on feedback from the results, continuously guiding ALNS to adaptively explore high-quality solutions within the search space. This mechanism not only enhances the efficiency of route optimization but also improves the algorithm’s adaptability to complex and constrained environments.

3.2.2. Perception and Feedback Mechanism

In the ALNS-PPO hybrid optimization framework, the effectiveness of the policy network relies on an accurate perception of the search state and the proper construction of feedback information. The algorithm integrates state perception and a feedback mechanism to extract feature information from routing solutions and generate adaptive reward signals based on optimization outcomes.
This mechanism primarily consists of three core functions: (1) state-vector construction, which extracts multidimensional indicators from the current routing solution to represent the search state; (2) action-space output, which maps the key control parameters of the ALNS algorithm into the action space of the policy, enabling the policy network to dynamically adjust the search behavior in response to the current state; and (3) reward signal design, which provides feedback to the policy network based on the trend of objective value changes between the current solution and the historical best solution, guiding the optimization direction.
  • State Space: In emergency material scheduling scenarios, route optimization must balance timeliness and coverage. To enable the PPO policy network to accurately perceive the dynamic environment, the state space is designed as an eight-dimensional vector, comprising the objective function value, vehicle utilization, proportion of customers violating time-window constraints, total delivery load, number of customers, average network distance, distance standard deviation, and number of distribution centers.
  • Action Space: In the ALNS-PPO hybrid framework, the action space comprises five key dimensions, two of which control the intensity of ALNS destroy operations and multi-objective weighting in repair operations. The continuous action vector α = [ r d , w r , w d , β , δ ] output by the PPO undergoes nonlinear transformations to regulate the ALNS parameters. Specifically, r d denotes the destroy ratio; w r represents the selection weight for repair methods, adjusting the usage probability of multiple repair operators; w d is the selection weight for destroy methods; β serves as a solution acceptance probability adjustment factor; and δ is an additional exploration factor that controls the perturbation intensity or magnitude.
  • Reward Function: This study designs a reward mechanism based on improvements in the objective function to effectively guide the PPO policy network in learning high-quality search strategies. Given that the objective function already integrates multiple factors, including the total vehicle travel time, material allocation deviation, and demand satisfaction, the reward function is defined as the scaled difference between the objective function values of the current and previous solutions, and it is expressed as follows:
    r t = C p r e v C c u r r k
Here, C p r e v denotes the best objective function value obtained in the previous iteration, C c u r r represents the objective function value of the new solution generated under the guidance of the PPO in the current iteration, and k is a scaling factor used to balance the magnitude of the reward value to prevent training instability.
To ensure that the continuous action output of the PPO module can be effectively mapped to the discrete decision parameters of the ALNS, this study develops the corresponding action transformation and normalization functions within the hybrid framework. Each continuous action α i output from the PPO is first mapped to the interval [0, 1] using a sigmoid function to ensure boundedness. The normalized results are scaled linearly. For the operator selection weights, because the sum of multiple operator weights must be one, a softmax function is applied for normalization. Through this transformation and normalization mechanism, the continuous action output of the PPO module can be smoothly and stably mapped to the control parameters of the ALNS, thereby achieving effective coupling between deep reinforcement learning and heuristic search.
Through the synergistic interaction of these three components, the deep reinforcement learning strategy continuously refines the search policy in complex environments, thereby effectively guiding the ALNS algorithm’s behavior.

4. Experiments and Analysis

4.1. Experimental Setup

The experiments were conducted on a Windows 10 platform equipped with an Intel Core i5-9300H @ 2.40 GHz processor and an NVIDIA GeForce GTX 1650 GPU. The algorithms were implemented in Python 3.8 and developed and debugged in the PyCharm (2024.1.4) environment. The classic Solomon and Homberger benchmark datasets were selected for the experiments to comprehensively evaluate the performance and adaptability of the proposed ALNS-PPO algorithm. To better reflect a real-world emergency logistics scenario, the original single-warehouse setting in the dataset was modified to a multi-warehouse configuration, and the vehicle capacity was adjusted. Examples of different scales and complexities were constructed to verify the solution effect and robustness of the algorithm under multiple constraints.
To ensure the reliability and reproducibility of the experiments, the parameters of the ALNS-PPO algorithm were appropriately configured. In the PPO module, the discount factor gamma was set to 0.985 to balance short-term and long-term rewards; strategy updates employed eps_clip with a weight of 0.3 to limit policy fluctuations and enhance training stability; the number of epochs was set to 4, the learning rate (lr) to 0.0003, and the Adam optimizer was adopted with a batch size of 32 to balance gradient estimation accuracy and computational efficiency. In the ALNS module, the degree parameter was set to 3 to control solution diversity and prevent premature convergence to local optima, while the number of iterations was set to 10,000 to ensure sufficient convergence of the algorithm.

4.2. Validation of Algorithm Effectiveness

To evaluate the performance of the proposed ALNS-PPO algorithm in emergency relief vehicle routing optimization, experiments were conducted from two perspectives: optimization effectiveness and parameter sensitivity analysis. These evaluations aimed to assess the algorithm’s solution quality and adaptability while also providing data for subsequent comparative studies. Twelve benchmark instances were selected for testing, including C101, C203, R102, R205, RC103, and RC207 from the Solomon dataset, and C121, C221, R122, R223, RC123, and RC225 from the Homberger dataset. In these instances, C-type cases feature clustered customer locations; R-type cases have dispersed customer locations with diverse time-window constraints; and RC-type cases combine both characteristics with more complex time-window restrictions, thereby closely reflecting the scheduling requirements of practical emergency distribution tasks.
For all experiments, the maximum number of iterations was set to 10,000. In the Solomon dataset, the problem scale is relatively small, with no more than 100 customer nodes, 3 depots, and 30 vehicles. In contrast, the Homberger dataset extends the Solomon dataset to larger scales, with up to 200 customer nodes, 5 depots, and 50 vehicles. Each instance was independently executed 10 times. The main evaluation metrics included total travel distance, number of vehicles used, demand fulfillment rate, and objective function value, to assess the algorithm’s robustness and optimization effectiveness under different demand distributions and time-window scenarios.

4.2.1. Parameter Sensitivity Analysis

In this section, a sensitivity analysis is performed on the key parameters of the algorithm to investigate their impact on solution quality. To assess the influence of these parameters, experiments were performed on the C101 instance of the Solomon dataset, and the convergence curves of the objective function values and key statistical indicators were recorded.
  • Analysis of the Sensitivity to Destruction Intensity
In the damage intensity sensitivity analysis, the C101 instance from the Solomon dataset and the R221 instance from the Homberger dataset were selected as representative test cases. The damage intensity was set to 3, 10, and 20, respectively, while the other parameters remained constant, to evaluate the influence on the convergence of the algorithm. The convergence trend of the objective function value with respect to the number of iterations is illustrated in Figure 5, and comparative performance data for different damage intensities are summarized in Table 2.
Figure 5. Convergence trend of the objective function under different damage intensities.
Table 2. Comparison of indicators at different damage intensities.
The convergence curves indicate that, when the destruction intensity is set to 20, the algorithm converged the fastest; however, the objective function value remained relatively high, stabilizing after approximately 1900 iterations. With a destruction intensity of 3, the algorithm showed a rapid initial decrease, yet the final convergence value is slightly higher than that for an intensity of 10. At an intensity of 10, the convergence speed is slower than that at 20; however, the objective function value underwent a greater reduction. These results suggest that an excessively high destruction intensity can substantially disrupt the solution structure, and although convergence may occur faster, the local optimization capability may be reduced, making it more difficult for the algorithm to identify superior solutions.
  • Analysis of Sensitivity to the Learning Rate
In the sensitivity analysis of the learning rate, the C101 instance from the Solomon dataset and the R221 instance from the Homberger dataset were chosen as representative test cases. The learning rate was set to 0.003, 0.0003, and 0.00003, respectively, while keeping other parameters constant, to assess its effect on the convergence behavior of the algorithm.
Figure 6 and Figure 7 show the sensitivity analysis results for different learning rates. They show variations in objective function values and algorithm runtimes across different settings. As can be seen, the learning rate has a clear influence on both convergence behavior and computational performance. A moderate learning rate (0.0003) provided a good balance between stability, convergence speed, and runtime efficiency. Detailed observations are shown in Figure 6 and Figure 7, and the corresponding quantitative results are summarized in Table 3.
Figure 6. Convergence trend of the objective function under different learning rates.
Figure 7. Algorithm runtime trends under different learning rates.
Table 3. Comparison of indicators at different learning rates.
For a relatively high learning rate (e.g., 0.003), the optimization process proceeded quickly, and the runtime increased moderately with the number of iterations. This indicates that a higher learning rate can accelerate convergence but may lead to unstable results and lower optimization precision per iteration. When the learning rate is reduced to 0.0003, the iteration precision improves. However, the overall convergence slows noticeably, resulting in a longer runtime. Further decreasing the learning rate to 0.00003 leads to an even slower convergence, with a significant increase in the required number of iterations and, consequently, in the total runtime. These results suggest that, although a smaller learning rate can enhance the solution accuracy, it requires considerably more computational time. Therefore, selecting an appropriate learning rate is critical for balancing computational efficiency and optimization precision.
  • Analysis of the Sensitivity to Weight Coefficient
To assess the influence of the objective function’s weighting scheme on model performance, a sensitivity analysis of the weight coefficients was conducted using the C101 instance from the Solomon dataset. All other parameters were kept constant, and eight alternative combinations of weight values were tested to evaluate their effect on the experimental results.
Table 4 shows the comparison of indicators under different weight coefficient combinations. The results indicate that the multi-objective optimization results are sensitive to the weight coefficients (α, γ, δ), and different combinations reflect the multi-objective trade-off among “transportation efficiency, load balance, and demand satisfaction”.
Table 4. Comparison of indicators under different weight coefficients.
When the demand satisfaction weight δ is dominant, the model gives priority to ensuring 100% demand satisfaction, but the total route is longer. Increasing the total transportation time weight α can shorten the route and optimize the objective function value, among which the combination (0.3, 0.2, 0.5) achieves the optimal balance among efficiency, service level, and resource utilization. Increasing the load balance weight γ can significantly reduce the average load deviation, and the insufficient vehicle utilization when δ is dominant is a reasonable adaptive performance of the algorithm to maintain service reliability.
Therefore, the proposed optimization framework exhibits notable robustness and interpretability, successfully capturing the inherent trade-offs among multiple objectives while maintaining stable solution performance.

4.2.2. Result Stability Analysis

To further validate the stability of the proposed ALNS-PPO hybrid optimization framework, we performed ten independent runs on selected instances from the Solomon and Homberger benchmark datasets. The resulting changes in the objective function values are shown in Figure 8 and Figure 9, respectively. The algorithm maintained relatively stable performance across multiple runs, with only minor fluctuations observed across different scenarios. This consistency demonstrates the robustness of the proposed framework under varying data distributions and constraint settings. Furthermore, the observed dispersion across runs serves as an implicit statistical validation of the algorithm’s reproducibility and generalization capabilities, providing supporting evidence for the reliability of the reported results.
Figure 8. Objective function values for the Solomon dataset.
Figure 9. Objective function values for the Homberger dataset.
Furthermore, based on the stability verification, the algorithm’s performance was systematically evaluated across multiple independent runs. The mean, standard deviation, and coefficient of variation (CV) of the objective function were calculated for several representative examples. Given the randomness inherent in the algorithm’s initial solution generation and search path selection, analyzing fluctuations in the objective function value is crucial for evaluating the consistency of the algorithm’s performance. CV quantifies the relative variability of the results; smaller values indicate more stable algorithm performance.
As shown in Table 5, the ALNS-PPO algorithm demonstrates consistent performance in terms of the mean and standard deviation of the objective function across different problem instances. In the Solomon dataset, although the CV values for some small and medium-scale instances, such as R205 and RC207, are slightly higher, they remained within an acceptable range, reflecting the strong adaptability and stability of the algorithm across various problem structures. This is mainly related to the search characteristics of random heuristic algorithms in low-dimensional solution spaces. When the solution space is small and the number of feasible solutions is limited, random perturbations may cause the volatility of local search paths to be amplified, making the results deviate more from the mean. For large-scale and more complex instances from the Homberger dataset, the algorithm exhibits even more stable optimization results. For example, the CV values for RC225 and R223 are 10.38% and 11.44%, respectively, indicating a noticeably reduced fluctuation range. These results suggest that, in higher-dimensional search spaces and under more complex constraints, the deep reinforcement learning mechanism, combined with the global search strategy, effectively guides the search process and mitigates the risk of being trapped in local optima.
Table 5. Indicators for evaluating the stability of the objective function.

4.3. Algorithm Comparison

To evaluate the effectiveness and superiority of the proposed ALNS-PPO algorithm in solving the multi-depot, multi-constraint emergency logistics VRP, a series of systematic comparative experiments were conducted. The comparison algorithms included the traditional Adaptive Large Neighborhood Search (ALNS) and a representative heuristic method—the Improved Ant Colony Algorithm (IACA)—to comprehensively assess differences in solution quality and computational efficiency [26]. The experiments were performed on the standard Solomon and Homberger datasets, which encompass instances of varying scales, thereby thoroughly testing the algorithm’s adaptability and stability under multiple constraints and ensuring the generality and objectivity of the evaluation results.
Table 6 presents a comparison of the experimental results obtained using each algorithm under identical environmental and constraint settings across multiple independent runs. Taking the R205 and RC207 instances as examples, the objective function values of ALNS-PPO were reduced by 39.9% and 31.2%, respectively, compared with those of the traditional ALNS, and also show significant improvements over the IACA. These results demonstrate that the dynamic strategy adjustment enabled by the PPO effectively overcomes the local optimum limitations of the ALNS. Although ALNS-PPO uses slightly more vehicles than IACA in the C101 instance, it achieved a substantial reduction in the total route length through more refined route planning. This intelligent trade-off between resource utilization and route optimization highlights the decision-making advantages of deep reinforcement learning in multi-constraint optimization.
Table 6. Experimental comparison results on Solomon and Homberger datasets.
Although the proposed ALNS-PPO algorithm may occasionally increase the number of vehicles compared with traditional heuristic methods (e.g., in instances C101 and C221), this adjustment reflects a deliberate trade-off between delivery timeliness and vehicle utilization. In emergency material transportation, the operational objective is to prioritize the rapid and complete delivery of relief supplies to the affected areas. Increasing the number of vehicles significantly shortens delivery distances and ensures that all demands are fully met within strict service time windows. This behavior is operationally feasible in real disaster response scenarios, where vehicle deployment costs are often secondary to rescue efficiency and timeliness. Therefore, in some cases, an enlarged fleet size indicates a more realistic and effective allocation strategy rather than optimization inefficiency.
The experimental results further indicate that the proposed ALNS-PPO algorithm exhibits significant advantages in large-scale and highly complex instances of the Homberger dataset (e.g., RC225 and R224), achieving lower objective function values and a 100% demand satisfaction rate. This suggests that, in multi-constraint scenarios characterized by large task sizes and dispersed demand distributions, the integration of deep reinforcement learning strategies enables the algorithm to demonstrate a stronger global search capability and adaptability. In contrast, traditional algorithms such as ALNS and IACA perform reasonably well in small-scale and structurally regular instances of the Solomon dataset (e.g., C101 and C203) but are prone to local optima when addressing large-scale, multi-constraint problems, exhibiting weaker stability and generalization ability.
Therefore, the ALNS-PPO algorithm is better suited for large-scale emergency scheduling problems characterized by high complexity and stringent constraints, whereas traditional algorithms remain effective for medium- and small-scale problems with relatively regular structures.
To comprehensively evaluate the efficiency-quality trade-off of the proposed ALNS-PPO algorithm, it is crucial to address the key requirement of emergency logistics—rapidly generating feasible solutions during the golden post-disaster rescue period is crucial. However, a single metric cannot address both metrics. Therefore, this study introduced the “Unit-Time Improvement Rate (UTIR)” metric. The UTIR metric measures the improvement in the objective function value achieved by a given algorithm relative to the ALNS per unit computational time, and is defined as:
U T I R = Z A L N S Z T A L N S + T × 100 %
Here, Z denotes the objective function value and T represents the computational time. The ALNS algorithm serves as the baseline for evaluation. A larger UTIR value indicated a more significant performance improvement per unit time. Furthermore, since both the numerator and denominator are expressed as dimensionless percentages, UTIR is a dimensionless metric. As the denominator approaches zero, UTIR exhibits large positive values, representing an “efficient trade-off”—a significant improvement in quality at a minimal cost in time.
As shown in Table 7, the proposed ALNS-PPO algorithm is further validated in terms of its overall performance in balancing solution quality and computational efficiency. Across multiple instances, ALNS-PPO exhibits relatively high UTIR values, mostly ranging from 24% to 31%, indicating a favorable balance between optimization effectiveness and efficiency. For example, in instance C221, the UTIR of ALNS-PPO reaches 31.53%, demonstrating that the algorithm maintains strong optimization capability per unit time even in large-scale problems. In contrast, although the IACA achieves shorter computational times in some instances, its optimization performance is generally insufficient, resulting in lower UTIR values and even negative values in C221, R224, and RC225. This indicates that IACA is less efficient than ALNS-PPO in terms of performance per unit time. Overall, ALNS-PPO achieves more stable performance in both solution quality and efficiency, making it well-suited for highly complex multi-constraint routing optimization scenarios.
Table 7. UTIR metrics vs. calculation time.
In addition to its numerical advantages, the proposed framework also has potential implications for real-world emergency logistics. However, the current model assumes static demand information and simplified road accessibility, which may limit its direct deployment in highly dynamic disaster environments. Future extensions could integrate dynamic updates to further enhance its applicability in real-world emergency scenarios.
Furthermore, compared to existing research such as reinforcement learning and heuristic algorithms, the proposed method demonstrates stronger solution stability and adaptability under multiple constraints, validating its effectiveness and scalability in complex path planning environments.

5. Conclusions

This study addressed the emergency material distribution path optimization problem under multiple constraints. A multi-constraint optimization model was developed by considering vehicle capacity, service time windows, distribution balance, and number of vehicles. The Adaptive Large Neighborhood Search (ALNS) algorithm was integrated with Proximal Policy Optimization (PPO) to enhance global exploration and adaptive decision-making capabilities. The experiments were performed on several representative instances from the Solomon and Homberger benchmark datasets. The results indicate that, compared with traditional heuristic methods, the proposed ALNS-PPO algorithm effectively reduces the total delivery distance and improves demand satisfaction. It also demonstrates strong stability and robustness in complex emergency logistics scenarios.
The findings provide valuable decision support for emergency logistics managers under time-critical conditions. The proposed method can assist in planning rapid-response operations, optimizing the balance between supply coverage and transportation efficiency, and reducing delivery delays. Furthermore, the hybrid framework offers scalability and practical potential for large-scale disaster response networks.
Despite these promising results, this study has several limitations. The current model assumes fixed demand quantities and single-trip vehicle routes, which may simplify the real-world logistics dynamics. In addition, iterative learning and repair mechanisms increased the computational cost as the problem size increased. Future work will consider dynamic or uncertain demand, multi-trip routing, and the integration of real-time traffic and terrain data. These improvements are expected to enhance the adaptability and practical applicability of the proposed method for large-scale and highly uncertain emergency response scenarios.

Author Contributions

Conceptualization, H.H. and X.Z. (Xiaoxiong Zhang).; Writing—original draft preparation, H.H.; Writing—review and editing, X.Z. (Xiaolei Zhou). and Q.F.; Supervision, J.Y. and B.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant Nos. 72471236 and 62402510 and by the Research Program of National University of Defense Technology under Grant No. ZK23-58.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, J.; Cai, J.; Sun, T.; Zhu, Q.; Lin, Q. Multitask-based evolutionary optimization for vehicle routing problems in autonomous transportation. IEEE Trans. Autom. Sci. Eng. 2023, 21, 2400–2411. [Google Scholar] [CrossRef]
  2. Jain, S.; Bharti, K.K. A combinatorial optimization model for post-disaster emergency resource allocation using meta-heuristics. Soft Comput. 2023, 27, 13595–13611. [Google Scholar] [CrossRef]
  3. Wu, P.; Clark, R.; Furtado, K.; Xiao, C.; Wang, Q.; Sun, R. A case study of the July 2021 Henan extreme rainfall event: From weather forecast to climate risks. Weather Clim. Extrem. 2023, 40, 100571. [Google Scholar] [CrossRef]
  4. Ozkula, G.; Dowell, R.K.; Baser, T.; Lin, J.; Numanoglu, O.; Iihan, O.; Olgun, C.G.; Huang, C.; Uludag, T. Field reconnaissance and observations from the February 6, 2023, Turkey earthquake sequence. Nat. Hazards 2023, 119, 663–700. [Google Scholar] [CrossRef]
  5. Dorigo, M.; Caro, G.D.; Gambardella, L.M. Ant algorithms for discrete optimization. Artif. Life 1999, 5, 137–172. [Google Scholar] [CrossRef] [PubMed]
  6. Kubil, V.N.; Mokhov, V.A.; Grinchenkov, L.M. Multi-Objective ant colony optimization for Multi-Depot heterogeneous vehicle routing problem. In Proceedings of the International Conference on Industrial Engineering, Applications and Manufacturing, Sochi, Russia, 15–18 May 2018; pp. 1–6. [Google Scholar]
  7. Miao, C.; Chen, G.; Yan, C.; Wu, W. Path planning optimization of indoor mobile robot based on adaptive ant colony algorithm. Comput. Ind. Eng. 2021, 156, 107230. [Google Scholar] [CrossRef]
  8. Zhang, X.; Dai, X.; Lou, P.; Hu, J. A multi-strategy ALNS for the VRP with flexible time windows and delivery locations. Appl. Sci. 2025, 15, 4995. [Google Scholar] [CrossRef]
  9. Cao, J.; Wei, J.; Lei, A.; Han, P.; Feng, Z.; Wang, M. Dynamic optimization of cold chain distribution path and construction considering traffic congestion. Appl. Res. Comput. 2025, 42, 2364–2373. [Google Scholar]
  10. Liu, H.; Sun, Y.; Pan, N.; Li, Y.; An, Y.; Pan, D. Study on the optimization of urban emergency supplies distribution paths for epidemic outbreaks. Comput. Oper. Res. 2022, 146, 105912. [Google Scholar] [CrossRef]
  11. Yu, V.F.; Lin, C.H.; Maglasang, R.S.; Lin, S.; Chen, K. An efficient simulated annealing algorithm for the vehicle routing problem in omnichannel distribution. Mathematics 2024, 12, 3664. [Google Scholar] [CrossRef]
  12. Tu, S.; Li, S.; Liu, Z.; Zeng, D. Improved simulated annealing algorithm for vehicle routing problem with multiple time windows using column generation. In Proceedings of the IEEE: The 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 1649–1654. [Google Scholar]
  13. Wei, M. Optimization of emergency material logistics supply chain path based on improved ant colony algorithm. Informatica 2025, 49, 215–246. [Google Scholar] [CrossRef]
  14. James, J.Q.; Yu, W.; Gu, J. Online vehicle routing with neural combinatorial optimization and deep reinforcement learning. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3806–3817. [Google Scholar] [CrossRef]
  15. Zong, Z.; Tong, X.; Zhang, M.; Li, Y. Reinforcement learning for solving multiple vehicle routing problem with time window. ACM Trans. Intell. Syst. Technol. 2024, 15, 1–19. [Google Scholar] [CrossRef]
  16. Silva, M.A.; Souza, S.R.; Souza, M.J.; Bazzan, A.L. A reinforcement learning-based multi-agent framework applied for solving routing and scheduling problems. Expert Syst. Appl. 2019, 131, 148–171. [Google Scholar] [CrossRef]
  17. Kang, M.; Tian, H.; Zhang, J.; Zheng, D.; Wu, Y. A deep reinforcement learning algorithm for dynamic EV routing problem. In Proceedings of the IECON the 50th Annual Conference of the IEEE Industrial Electronics Society, Chicago, IL, USA, 3–6 November 2024; pp. 1–6. [Google Scholar]
  18. Arishi, A.; Krishnan, K. A multi-agent deep reinforcement learning approach for solving the multi-depot vehicle routing problem. J. Manag. Anal. 2023, 10, 493–515. [Google Scholar] [CrossRef]
  19. Ahamed, T.; Zou, B.; Farazi, N.P.; Tulabandhula, T.; Lin, Q. Deep reinforcement learning for crowdsourced urban delivery. Transp. Res. Part B Methodol. 2021, 152, 227–257. [Google Scholar] [CrossRef]
  20. Liu, Q. Improved deep reinforcement learning for intelligent logistics supply chain transportation decision model. In Proceedings of the International Conference on Integrated Intelligence and Communication Systems (ICIICS), Kalaburagi, India, 24–25 November 2023; pp. 1–6. [Google Scholar]
  21. Reijnen, R.; Zhang, Y.; Lau, H.C.; Bukhsh, Z. Online control of adaptive large neighborhood search using deep reinforcement learning. In Proceedings of the International Conference on Automated Planning and Scheduling, Banff, AB, Canada, 1–6 June 2024; pp. 475–483. [Google Scholar]
  22. Ma, H.; Yang, T. Improved adaptive large neighborhood search combined with simulated annealing (IALNS-SA) algorithm for vehicle routing problem with simultaneous delivery and pickup and time windows. Electronics 2025, 14, 2375. [Google Scholar] [CrossRef]
  23. Liu, W.; Luo, Y.; Yu, Y. An adaptive hybrid genetic and large neighborhood search approach for multi-attribute vehicle routing problems. arXiv 2024, arXiv:1603.00278. [Google Scholar]
  24. Wang, S.; Sun, W.; Huang, M. An adaptive large neighborhood search for the multi-depot dynamic vehicle routing problem with time windows. Comput. Ind. Eng. 2024, 191, 110122. [Google Scholar] [CrossRef]
  25. Chen, W.; Zhang, Z.; Tang, D.; Liu, C.; Gui, Y.; Zhao, Z. Probing an LSTM-PPO-Based reinforcement learning algorithm to solve dynamic job shop scheduling problem. Comput. Ind. Eng. 2024, 197, 110633. [Google Scholar] [CrossRef]
  26. Deng, H.; Wu, J. Study on optimization of vehicle routing problem by improved ant colony algorithm. In Proceedings of the China Automation Congress, Chongqing, China, 17–19 November 2023; pp. 399–404. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.