Next Article in Journal
An Elegant Multi-Agent Gradient Descent for Effective Optimization in Neural Network Training and Beyond
Next Article in Special Issue
Path Navigation for Multi-Agent Flight System Based on Distributed Model Predictive Control and Reinforcement Learning
Previous Article in Journal
A Study on Model Training Strategies for Speaker-Independent and Vocabulary-Mismatched Dysarthric Speech Recognition
Previous Article in Special Issue
A UAV Coverage Path Planning Method Based on a Diameter–Height Model for Mountainous Terrain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning for Efficient Drone-Assisted Vehicle Routing

by
Aigerim Bogyrbayeva
,
Bissenbay Dauletbayev
and
Meraryslan Meraliyev
*
Department of Computer Science, SDU University, Kaskelen 040900, Kazakhstan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(4), 2007; https://doi.org/10.3390/app15042007
Submission received: 18 December 2024 / Revised: 25 January 2025 / Accepted: 29 January 2025 / Published: 14 February 2025
(This article belongs to the Special Issue Advances in Unmanned Aerial Vehicle (UAV) System)

Abstract

:
Many exact algorithms, heuristics, and metaheuristics have been proposed to solve the Vehicle Routing Problem with Drones, which involves using a fleet of trucks and drones to fulfil customer orders in last-mile delivery. In this study, the problem is formulated using the Markov Decision Process, and a Reinforcement Learning (RL) based solution is proposed. The proposed RL model is based on an attention-encoder and a recurrent neural network-decoder architecture. This approach enhances coordination by determining which vehicles should visit specific customers and where vehicles can rendezvous, effectively leveraging drones and reducing the overall completion time. The RL model has demonstrated competitive performance compared to benchmark algorithms through extensive experiments.

1. Introduction

Last-mile delivery represents the critical final stage of the logistics process, encompassing the transportation of parcels from distribution centres to customers’ destinations via various vehicles such as trucks, vans, bicycles, or drones. Optimizing this process is essential to providing a seamless and satisfying delivery experience, achieving cost reductions for logistics companies, enhancing operational efficiency, and fostering sustainability and business growth. With a projected 36% increase in delivery vehicles on roads between 2019 and 2030, urban areas face heightened congestion, leading to a 21% rise in average commute times. This surge also brings an additional 6 million tonnes of CO2 emissions, underscoring the urgency of addressing last-mile delivery challenges effectively [1].
Emerging technologies, particularly drones, have gained traction as innovative solutions for overcoming last-mile delivery challenges. Initially utilized in military and disaster management sectors [2,3], drones are now being explored for logistics applications due to their lightweight, battery-operated nature [4]. The recent real-world pilot projects, such as Amazon Prime or Google Wings, illustrate how drones can reduce delivery times, bypass traffic congestion, and serve remote locations or areas with poor infrastructure. By offering faster door-to-door services and potentially lower operational costs, drones can improve the efficiency of last-mile logistics, decrease the reliance on large fleets of traditional vehicles, and reduce fuel consumption and emissions. Drones also produce zero direct emissions and can complement truck-based methods by reaching difficult-to-access regions. While trucks excel at delivering heavy packages over long distances, drones are well-suited for transporting small parcels over short distances. Using drones along with traditional trucks for last-mile delivery can save time, lower emissions, and help reduce traffic congestion [5].
The potential for parallel customer service emerges when multiple trucks and drones are incorporated into last-mile delivery systems, leading to shorter delivery completion times and higher overall operational efficiency. For instance, if a drone travels 50% faster than a truck, and each truck is equipped with up to two drones, delivery times can be reduced by up to 75% in optimal scenarios [6]. Beyond these efficiency gains, drones can also support urgent and time-sensitive deliveries such as medical supplies or high-priority goods, further enhancing service quality. However, the combined deployment of drones and trucks introduces complex computational challenges, such as coordinating their routes and identifying optimal launch and retrieval locations.
The optimization problem associated with finding efficient routes for multiple drones and trucks in last-mile delivery is known as the Vehicle Routing Problem with Drones (VRPD). This problem is classified as an NP-hard combinatorial optimization problem (COP) [7] due to the intricate constraints involved, including drone battery limitations and diverse launch and retrieval points. Despite existing research efforts [8,9], current methods often suffer from prolonged computational times and lack adaptability to dynamic customer distribution patterns, particularly for large problem instances.
This study proposes a novel heuristic method to address the VRPD by formulating it as a Markov Decision Process (MDP) and leveraging Reinforcement Learning (RL) to train a Neural Network (NN) for routing decisions. RL, a machine learning approach, enables an agent to learn optimal decisions through interaction with the environment and iterative improvements based on cumulative rewards. Recent advancements in machine learning have shown promise in addressing challenging COPs [10,11,12].
Among prior studies, the closest work to ours is [13], which applied RL to the Traveling Salesman Problem with Drone (TSPD), a special case of the VRPD involving a single truck and drone without battery limitations. However, the TSPD lacks the complexity of coordinating multiple drones and trucks, which requires addressing both intra- and inter-group decision-making challenges. Additionally, ref. [13] assumes unlimited drone battery life, which is unrealistic for real-world applications. Our study focuses on routing policies for multiple drones and trucks while accounting for drone battery constraints, necessitating novel MDP formulations and RL-based frameworks to deliver effective solutions.
This paper contributes the first machine learning-based heuristic capable of delivering near-instantaneous solutions for large VRPD instances while generalizing effectively to data from the same distribution. Experimental results demonstrate that our RL approach outperforms certain state-of-the-art methods for specific instances and produces results with significantly reduced computational times.
The key contributions of this study include:
  • A novel MDP formulation of the VRPD that explicitly considers multiple vehicles and drones with limited battery life. To the best of our knowledge, no prior work has simultaneously addressed multi-vehicle coordination and drone battery constraints in this manner.
  • An RL-based solution approach utilizing a centralized controller to coordinate routing decisions within and across groups of drones and trucks in urban networks.
  • An efficient training algorithm for RL agents, validated through comprehensive computational analyses comparing our model to state-of-the-art approaches.
The remainder of this paper is organized as follows: Section 2 reviews relevant literature on the VRPD and machine learning methods for routing problems. Section 3 formally defines the VRPD and its MDP formulation. Section 4 presents the proposed solution methods. Section 5 provides computational experiment results, and Section 6 concludes with final remarks and future research directions.

2. Literature Review

This section presents recent studies focusing on solving the VRPD, while comprehensive reviews and comparisons of combined drone-truck routing and problem variants can be found in [4]. Additionally, recent advancements in learning methods for solving VRPs will be discussed.

2.1. The VRPD

The literature on the VRPD and its extensions has explored various aspects of optimizing time and cost savings. Ref. [6] conducted a worst-case analysis for the VRPD, proposing bounds for achieving maximum benefits in terms of time savings. Their analysis considered the same distance metrics for both trucks and drones, with no limitations on drone batteries and the ability to dispatch and retrieve drones from the same node. The results revealed that savings depended on the number of drones per truck and the drones’ speed. Building upon [6]’s work, ref. [8] extended the analysis by considering limited batteries for drones, different distance metrics for trucks and drones, and demonstrated maximum time and cost savings based on bounds. Ref. [14] introduced two heuristics for the VRPD, involving construction and improvement phases to minimize makespan.
In a further study, ref. [9] introduced en route operations, enabling drones and trucks to rendezvous at discrete points along the truck route and schedule new deliveries with negligible drone recharging time. They proposed a new Mixed-Integer Linear Programming (MILP) model and introduced valid inequalities to solve small instances for optimality. Additionally, they suggested a heuristic based on Variable Neighborhood Search (VNS) and Tabu Search (TS) concepts for large instances, which numerically proved the benefits of en route operations in reducing makespan. Notably, they assumed that drones could not be retrieved at the location where they were launched. Ref. [7] expanded on the work of [6] by proposing an arc-based integer programming model for the VRPD, reformulating it as a path-based model, and developing a branch-and-price algorithm for finding minimum logistics costs for up to 15 instances. Through sensitivity analyses, they concluded that a 10% logistics cost reduction could be achieved with drones with higher flying durations.
In another study, ref. [15] proposed a new MILP model for solving the VRPD, allowing cyclic operations for drones, which could be launched and retrieved at the exact customer location, increasing complexity. The authors introduced valid inequalities and proposed a metaheuristic to divide problems into components, such as allocation and sequencing, which were further solved by the classical savings heuristic. They addressed problems with up to 100 customers. Beyond using heuristics and MILP models, ref. [16] tackled a similar problem to the VRPD using an encoder-decoder framework combined with RL. In their approach, deliveries were made exclusively by drones, while trucks acted as portable warehouses and charging stations. Their extensive experiments demonstrated that RL could produce satisfactory solutions in a significantly shorter time than heuristics.
Moreover, many existing studies focus on specific operational scenarios, such as single-drone configurations or simplified battery constraints, and do not always consider advanced real-world applications like partial recharge or multi-objective routes. Drones have been extensively used in diverse fields, including telecommunications, healthcare, and urban surveillance outside classic truck-drone logistics. For example, ref. [17] present a MILP-based framework for drone routing and recharging as flying base stations, demonstrating how UAVs can serve mobile networks with grid-connected microgeneration constraints. In a broader survey, ref. [18] provides an exhaustive overview of optimization approaches for civil applications of UAVs, highlighting how routing variants extend to firefighting, emergency response, and beyond. In healthcare logistics, ref. [19] introduce a bi-objective optimization model for simultaneous pickup and delivery of medical supplies with drones, employing a modified NSGA-II algorithm to balance conflicting objectives. Likewise, ref. [20] proposes a decomposition-based approach to coordinate UAV tours for city-scale video surveillance, where drones even ride public buses for energy-efficient mobility.
Despite these advancements, certain limitations remain unaddressed—for instance, many works either assume ample battery resources or neglect dynamic recharging schedules, thus not fully capturing the complexity of multi-vehicle VRPD settings. By incorporating insights from these broader contexts, our study positions itself to tackle more complex constraints (e.g., multiple drones, limited battery, or real-time route adjustments) and thereby offers novel solutions that could feasibly transfer to sectors like telecommunications or urban security. This perspective also clarifies how our research extends the current literature, focusing on machine learning-based methods to achieve instantaneous or near-instantaneous routing decisions.

2.2. Learning Methods to Solve VRPs

There are two main ML methods used to solve various VRPs. In Supervised Learning, a model undergoes training using a dataset comprising input problems and their corresponding solutions. On the other hand, RL treats VRPs as sequential decision-making challenges, relying on the MDP. While a comprehensive overview of ML techniques for VRPs can be found in [12,21], we summarize the most notable studies.
The pioneering work by [22] introduced the Neural Combinatorial Optimization framework to tackle the TSP by applying NNs and RL. They achieved this by training Pointer Networks, a concept introduced by [23], with RL techniques. Notably, their results were on par with traditional heuristic approaches. Building upon this foundation, ref. [24] further improved the methodology by incorporating Graph Neural Networks and proposing a fusion of RL and graph embeddings to address TSP. This novel approach aimed to enhance the optimization process.
Motivated by the work of [22], ref. [11] presented an end-to-end framework for solving VRPs using RL, achieving comparable results with well-known VRP solvers. Similarly to [24], ref. [10] introduced an attention-based encoder-decoder framework specifically designed to solve the VRP. Their innovative approach achieved high efficiency during training and surpassed previous methods in terms of overall performance. In recent studies by [25,26,27], the authors proposed entirely machine learning-based and hybrid methods combining ML with optimization heuristics to derive near-optimal solutions for VRPs.

3. The Problem Formulation

In this section, the VRPD is formally defined, and its MDP formulation is presented.

3.1. The VRPD Definition

A complete graph G is considered, consisting of a set of nodes N = { 1 , 2 , , N } , where node N is represented as a depot. All other nodes in the graph represent known customer locations with their orders. Each customer must be visited by a drone, a truck, or both, only once. To achieve this, a set of homogeneous trucks K with the corresponding set of identical and equal number of drones D k for all k K leaves the depot loaded with customer orders. The distance between nodes i and j is defined as d i , j such that d i , j = d j , i for all nodes i and j in N . Given the constant velocities of drones and trucks, let v d r and v t r denote the velocities of drones and trucks, respectively, where v d r = α v t r and α 1 , as drones are not subjected to traffic congestion. Then the time to traverse from node i to node j is τ i , j d r and τ i , j t r for drones and trucks, respectively, where τ i , j t r = α τ i , j d r .
A drone can only fly a distance of ε due to its battery limits and must be retrieved by its corresponding truck to be recharged. As established in the literature [8,9,28], drones can serve only a single customer per launch. Additionally, they can be launched and retrieved only at customer locations. Negligible times for drone launching, retrieval, charging, and customer serving are assumed. Each drone can be launched and retrieved only by its corresponding truck. Both drones and trucks are allowed to wait at rendezvous nodes. In general, the decision variables for the VRDP include the sequence of nodes to be visited by each truck. For the drones, one needs to decide whether a drone will serve a customer, including at which node it will be launched to serve the selected customer and at which node the drone will be retrieved from its truck. In modeling the VRPD, a number of constraints must be satisfied. First, the flow preservation constraints ensure that all the drones and trucks leave the depot once and return back only once. Second, each customer node must be served only once. Third, to ensure the synchronization of the routes of drones and trucks, the launching nodes must be visited by trucks before drones serve the selected customers for drone service, followed by retrieval nodes. The objective function of the VRPD is to minimize the time to serve all customers measured from the moment vehicles (both drones and trucks) leave the depot until they all return. The full mathematical model of the VRPD can be found in Appendix A, that follows [29]’s formulation. It is also noted that the obstacle avoidance problem by drones is not considered by the VRPD.

3.2. MDP Formulation of the VRPD

Given the deterministic nature of the VRPD and the common goal of all vehicles to serve customers in minimal time, we assume that a central controller observes the entire graph and makes routing decisions for all vehicles. This problem is formalized as an MDP with discrete steps. An MDP is represented by the tuple M = s , a , P , R , T , where s denotes states, a denotes actions, P denotes transition probabilities, R denotes the reward function, and T denotes the time horizon, with the details outlined below.
Given the set of trucks K and drones D , the state of the MDP for the VRPD is represented by the vector s t : = ( V t , c t t r , c t d r , r t t r , r t d r , ε t d r , l t d r ) . The customers already served up to time step t must be kept track of, denoted as V t N . The current destination node of truck k and its drone d is denoted by c t k and c t k , d , respectively. However, at any time step t, not all vehicles may arrive at their destination nodes. If a truck (or a drone) is still traversing between two nodes at step t, then c t k (or c t k , d ) is set as the destination node of the vehicle. If a truck (or a drone) arrives at a node at time step t, then c t k (or c t k , d ) is set as the node index. Consequently, vectors c t t r : = ( c t k , k K ) and c t d r : = ( c t k , d , k K , d D ) can be constructed. The remaining times to arrive at destinations c t k and c t k , d of truck k and its drone d are denoted as r t k and r t k , d , respectively. When a truck (or a drone) is not in transit, r t k = 0 (or r t k , d = 0 ). r t t r is defined as ( r t k , k K ) and r t d r is defined as ( r t k , d , k K , d D ) . Let vectors ε t d r : = ( ε t k , d , k K , d D ) , and l t d r : = ( l t k , d , k K , d D ) represent the remaining battery and load of drones at step t, and let ε m a x be the maximum battery level for drones. At t = 0 , since all vehicles are located at the depot, we have V 0 = { N } , c 0 k = N , c 0 k , d = N , r 0 k = 0 , r 0 k , d = 0 , ε 0 k , d = ε m a x , l 0 k , d = 1 for all k K and d D .
At each step, the central controller must decide on the next destinations of the trucks, a t t r : = ( a t k N , k K ) , and drones a t d r : = ( a t k , d N , k K , d D ) . When the previously assigned node has not been reached by a drone d of truck k yet, i.e., r t k , d > 0 , we set a t k , d = c t k , d . Analogously, a t k = c t k if truck k is in transit. For convenience, a t : = ( a t t r , a t d r ) is referred to as the action at step t.
Due to the deterministic nature of the problem, the state transition probabilities are also deterministic. We revise the remaining times for each vehicle after its action has been determined and denote them as r ^ t k = r t k + τ c t k , a t k t r and r ^ t k , d = r t k , d + τ c t k , d , a t k , d d r for truck k and drone ( k , d ) respectively. Note that τ i , i t r = 0 and τ i , i d r = 0 for any i N . The time step is evolved when at least one vehicle arrives at its assigned node. Then, the time taken to finish the previous step by all vehicles is measured and denoted as C t = min ( r ^ t k , r ^ t k , d ) among all k K and d D . Correspondingly, the remaining times at the new step are updated as r t + 1 k = r ^ t k C t and r t + 1 k , d = r ^ t k , d C t for all k K and d D . Similarly, the drone’s battery is updated as follows:
ε t + 1 k , d = ε m a x , if c t k = c t k , d and r t k = 0 and r t k , d = 0 ε t k , d v d r C t , otherwise .
If a drone arrives at node n, which has demand q t n at time step t, we have:
l t + 1 k , d = max { 0 , l t k , d q t n } k K , d D .
The demand is updated as follows for all k K , d D , assuming that the maximum demand at each node is equal one:
q t + 1 n = max { 0 , q t n l t k , d } , if n = c t k , d and r t k , d = 0 max { 0 , q t n 1 } , if n = c t k and r t k = 0 .
The set of visited nodes is updated as V t + 1 = V t { a t e r t + 1 e = 0 , e { k K , ( k , d ) , k K , d D } . Also, c t + 1 e = a t e for e { k K , ( k , d ) , k K , d D } .
Let T be the index of the step when all drones and all trucks return to the depot after serving all customers. Then, our cost function (negative reward) is the total time spent in the system or the makespan:
R = C = t = 0 T C t .

4. Solution Methods

This section presents the deep learning model for coordinating the routing of multiple drones and trucks, followed by the training algorithm.

4.1. The Deep Learning Model

To route multiple drones and trucks for the last mile delivery, the stochastic policy π θ with parameters θ needs to be learned, defined as a product of conditional probabilities.
π θ ( a 1 , a 2 , , a T | G ) = t = 1 T p θ ( a t | s t ) .
As demonstrated in the literature [10,23,30], a routing policy for a single vehicle can be learned through encoder-decoder-based deep learning models. The encoder is an information extractor from a graph, encoding its structure and the problem settings. The decoder aims to learn the policy π θ directly using the embeddings from the encoder. Furthermore, ref. [13] has demonstrated the HM’s effectiveness, consisting of an attention-based encoder and an LSTM-based decoder, in routing a single drone and truck. However, the structure of the above model does not consider the interactions between groups when routing multiple drones and trucks, where a truck with its set of drones can be viewed as a group. Within a group, the model must consider coordinating actions taken by a truck and several drones to launch and retrieve them efficiently. On the other hand, one group’s decisions affect another group’s decisions since drones can be launched and retrieved only by their assigned trucks. To serve all customers effectively with a diverse set of locations in minimal time, the model must consider the selected actions of the first group to make decisions about routing the second group.
The HM presented in [13] is modified by adjusting inputs to the decoder, allowing for the efficient incorporation of coordination within a group and interactions between groups within a shared graph. This model is referred to as the attention-encoder group-based LSTM-decoder. Figure 1 gives an overview of the model.
ENCODER. A multi-head attention-based encoder is deployed, which has been shown to be effective in embedding graphs for routing problems [10,25]. In the context of a given graph G with nodes in N and corresponding coordinates denoted as y n = ( y n 1 , y n 2 ) R 2 for each node n N , the process of embedding a fully connected graph begins with an initial transformation of node embeddings. The initial embeddings, represented as h n 0 = W 0 y n + b 0 for all n N , involve learnable parameters W 0 and b 0 .
These initial node embeddings h n 0 : n N undergo L attention layers, where each layer, denoted as l, comprises two sublayers: a multi-head attention (MHA) layer and a fully connected feed-forward (FF) layer. The MHA layer takes input from the previous layer and facilitates message passing between nodes. For each input to the MHA layer, queries q , keys k , and values v are computed by projecting the input h n using trainable parameters W Q , W K , and W V .
Compatibility between nodes i and j is determined based on the fully connected graph structure, expressed as u i , j = d i m , where d i m represents the parameter size. The compatibility values u i , j are then used to compute attention weights, a i , j [ 0 , 1 ] , using the softmax function. Messages received by node n are formed as convex combinations of messages from all nodes.
An MHA layer with head size M is employed to enhance the attention mechanism’s expressive power. Messages from each head are computed and combined for each node n. The output of the MHA sublayer, along with skip connections, undergoes batch normalization.
The resulting batch-normalized output is processed through a fully connected FF network with the ReLU activation function. Skip connections and batch normalization are applied again to the output of the FF layer, resulting in the final representation denoted as h l after l attention layers.
DECODER. The encoder outputs representing embeddings for each node in the graph are used to denote the graph encoding, represented as h ¯ , which is defined as the mean of all node embeddings in the graph. For each group in the problem setting, where groups are defined by the number of trucks k, the LSTM is initialized consisting of a tuple ( z t k , c t k ) for all k K .
To determine an action for a decision-maker, either a drone or the truck in group k at time t, the LSTM is provided with the embedding of the last node selected by the last decision-maker in the group, labeled as h i L , where i corresponds to the node:
z t + 1 k , c t + 1 k = LSTM ( h i L , ( z t k , c t k ) ) .
Moving to the decoder, the hidden state z t + 1 k and cell state c t + 1 k at time t + 1 are subject to dropout with a probability of p:
z t + 1 k = Dropout ( z t + 1 k , p ) , c t + 1 k = Dropout ( c t + 1 k , p ) .
To represent the state of the decision-maker at time t, binary values x t 1 and x t 2 are constructed to indicate whether the current decision-maker is in transit or not, and whether it has a load or not, respectively. The current battery level of a drone is normalized by the maximum battery level, ε m a x , and is denoted as x t 3 . Let vector x t denote the concatenation of x t 1 , x t 2 , and x t 3 .
The current state of the decision-maker x t , the embedding of the graph h ¯ , and the hidden state of the LSTM z t + 1 k are inputs to an attention mechanism, where the attention vector is computed using trainable parameters v a , W a , and W x , and its elements are denoted as a i , j .
a i , j · = v a tanh W a [ h ¯ ; z t + 1 k ; W x x t ] .
Subsequently, the attention vector undergoes softmax to generate probabilities for visiting the next node, as the following equation describes. Its value is set to zero if a node has been visited by any vehicle before:
p θ ( a t v = j | s t ) = exp ( a i , j ) j N V t exp ( a i , j ) : j N V t , 0 : j V t .

4.2. The Training Algorithm

Efficient training of the attention-encoder group-based LSTM-decoder model involves exploring actions and receiving feedback as rewards. The model aims to determine the probability distribution, π θ ( V T | G ) , that, given a graph, generates the sequence of nodes to be visited by both the drones and the trucks in the VRPD. The objective during training is to minimize the makespan C in Equation (4).
The training objective function, denoted as J ( θ | G ) , is defined as:
J ( θ | G ) = E π θ ( V T | G ) [ ( C b ( G ) ) log π θ ( V T | G ) ] .
Here, b ( G ) serves as a baseline for variance reduction. A critic network estimates b ( G ) , while an actor-network is utilized to learn the policy π θ . The gradients of the objective function, θ J ( θ | G ) , are computed using the REINFORCE algorithm [31]. Algorithm 1 outlines the training process of actor and critic networks by sampling a batch of data from a predefined distribution and actions for all drones and trucks at each time step. Then, the final reward resulting at the end of the episode is used as an empirical mean to train both networks. To efficiently explore action spaces of drones and trucks, a masking scheme is employed that prevents some nodes from being visited based on the current state of the problem and the routing constraints, similar to existing literature [10,11]. In particular, a mask is applied during the action selection process, filtering out actions that violate constraints as shown in Algorithm 1.
Algorithm 1: Rollout
Applsci 15 02007 i001

5. Computational Studies

In this section, the proposed model is focused on through computational studies.

5.1. Training and Evaluation Configurations

Data Generation. To demonstrate the computational efficiency of the proposed approach, results are reported on the well-known benchmark dataset first proposed in [28]. In particular, ref. [28] randomly generated customer locations using a uniform distribution within the [1, 100] × [1, 100] area, while the depot location is generated with a uniform distribution within the [0, 1] × [0, 1] range, ensuring it always resides in a corner. Other studies have widely used the same benchmark dataset, including [9,30]. Throughout the experiments, it is assumed that the drone’s speed is double that of the truck, thus α = 2, consistent with existing literature.
Gap. To evaluate the quality of solutions produced by different approaches to solve the VRPD, the GAP is measured and defined as:
Gap i = Z i Z i * Z i * × 100 %
where Z i is the cost of a solution method for each instance i and Z i * is the cost of the best-performing solution method among all methods compared for instance i. In Table 1, Table 2 and Table 3, the average GAP for each setting of the problem as the mean of GAPs for 10 instances of the problem is reported.
Baselines. The efficiency of the proposed neural architecture is analyzed using [9] as a benchmark. In [9], the authors proposed the MILP formulation of the problem, solving the small instances exactly with and without lazy constraints, denoted as Solver and Solver Lazy, respectively in Table 1, Table 2 and Table 3. Beyond the exact method, the authors suggested a more practical strategy involving VNS and TS methods. This approach entails incorporating drones into delivery routes and a divide-and-conquer strategy. VNS is a method of exploring various possibilities by altering the neighbourhood structure, while TS is another approach that learns from previous solutions to avoid retracing steps. The results of VNS and TS-based heuristics are denoted as Heuristic in Table 1, Table 2 and Table 3. The computational time and costs for each instance of [9] have been kindly shared by the authors, where the best solution obtained by their heuristic in five runs and its corresponding computational time are reported.
Decoding Strategies. Three approaches were used to extract solutions from the trained model. The first method, referred to as “greedy”, involves consistently choosing nodes with the highest visitation probability at each time step. In the second method, “sampling”, multiple solutions are independently generated from the trained model with a sampling size of 4800. Subsequently, the solution with minimal cost among the samples is identified, and its results are reported. In the third method, “Ensemble”, multiple models saved throughout the training process are utilized. In particular, the models’ weights in the interval [ 400 k , , 1000 k ] are selected with an increment of 100k. Solutions are sampled using a sample size of 4800 for each model, and the best results are reported.
Computational Resources and Hyperparameters. Computers with the following specifications were utilized: Intel Core i9 12900K and PCI-E 24576Mb Palit RTX 3090 GamingPro OC, GeForce RTX 3090 video card were used in two of the systems, while a Core i9 12900K with RTX 4090 24GB was employed on another computer for training. An Intel Core i9 12900K and PCI-E 24576Mb Palit RTX 3090 GamingPro OC computer were used to conduct testing and report the computational time.
The same hyperparameters as those in [13] were used to train the neural networks, including 256 as the hidden dimension for embedding, 0.0001 as the learning rate, and 1 million iterations for training.

5.2. Results on the VRPD with Unlimited Flying Range

  • Single vehicle with two drones
  • Two vehicles with one drone assigned to each vehicle
  • Two vehicles with two drones assigned to each vehicle
For a small-scale problem with 8 nodes, the RL sampling model exhibits a 1.45% optimality gap. It demonstrates an almost 70-fold increase in speed within the first configuration. It outperforms all other methods in the second configuration and achieves a 2.14% optimality gap in the third configuration.
In a small-scale problem involving 10 nodes, the RL sampling model surpasses all alternative methods in the initial two configurations. The third configuration achieves a minimal 0.92% gap compared to exact methods. Since no results were reported for big-scale problems with 20 nodes, we only reported the RL model results.
The fast results from the RL sampling approach mostly come from its learned policies that use the problem’s structure effectively, such as grouping similar tasks. The proposed model adjusts when multiple drones are used by balancing their routes with the trucks. This adjustment shows in the nearly optimal solutions we see in the 8- and 10-node cases, indicating that our RL method can find efficient orders for dispatching without needing to check every possibility. However, we notice that as the problem size grows (for example, with 20 nodes), the quality of the solutions may drop a bit (1–2% higher gap) compared to smaller cases. This happens because tracking the coordination between drones and trucks gets more complicated, highlighting the balance between quick solutions and ensuring the best overall result.
In summary, findings show that the RL model can quickly create high-quality routes for different situations, offering clear advantages over traditional solvers in terms of speed. This speed can be crucial in real-world operations where route decisions need to change quickly.

5.3. Results on the VRPD with Limited Flying Range

The impact of restricted flying ranges was systematically explored by adjusting a crucial parameter denoted as β . This parameter plays a vital role in defining the operational capabilities of the drone, specifically determining the maximum distance it can cover with a fully charged battery. The outcomes of these scenarios, detailed in Table 2 and Table 3, provide detailed insight into how the system performs under limited flying distances.
This specific variable is adopted based on empirical setups from [9]. In that study, β = 0.6 was tested as a moderately restrictive constraint, limiting drones to 60% of the largest distance between any two nodes, while β = 1 represented a more lenient scenario where drones could cover the entire maximum inter-node distance.
By using these two values, conducted experiments align with established benchmarks, allowing for direct comparisons with previous results. Additionally, these β levels highlight how the proposed method adapts to different drone battery constraints: β = 0.6 simulates stricter capacity limitations, whereas β = 1 reflects drones capable of covering any route segment in a single trip. This setup helps evaluate how varying operational flexibility influences solution quality, route feasibility, and computational efficiency.
Scenario 1: β = 1
In the first scenario, where β is set to 1, indicating the drone’s maximum flight distance aligns with the largest distance between nodes, the results are showcased in Table 2.
In smaller scenarios with 10 nodes, the Ensemble RL model outperforms Solver and Solver Lazy in terms of time and quality across all configurations. Compared to the Heuristic approach, Ensemble RL achieves a 1.04% gap in the first setup, showcasing nearly 60 times faster performance. The second configuration reveals a 2.03% gap with approximately a 16-fold speed boost, while the third setup shows a 2.3% gap with around 15 times faster speed.
In larger-scale problems involving 20 and 50 nodes, Ensemble RL demonstrates superior solution efficiency and time performance compared to the Heuristic method in the initial configuration. However, in the second and third configurations for N = 20 , the gaps are 3.09% and 3.58%, respectively. For N = 50 , Ensemble RL outperforms the Heuristic method in all configurations.
Scenario 2: β = 0.6
Transitioning to the second scenario is characterized by setting β to 0.6, thus constricting the drone’s maximum flight distance to 60% of the largest distance between nodes. The results in Table 3 illuminate additional insights into the system’s adaptability and efficiency under different operational constraints.
Ensemble RL consistently outperformed Solver and Solver Lazy in our experiments with smaller instances across all configurations, demonstrating its reliable efficiency. Ensemble RL achieved a 1.28% gap compared to the Heuristic in the first setup. In the second configuration, Ensemble RL Sampling reduced the gap to 0.08%. In the third configuration, the Ensemble model showcased its ability to find a solution with only a 0.12% gap, and it did so with a significant approximately 10-fold increase in speed.
Shifting to more significant instances with 20 nodes, Ensemble RL sustained its superiority over the Heuristic in the first setup, excelling in speed and achieving the best solutions. This outstanding performance extended to instances with 50 nodes. Ensemble RL consistently outperformed the Heuristic in both the first and second setups, highlighting its scalability and effectiveness in handling more intricate problem scenarios. Considering the third scenario, RL Sampling outperformed the Heuristic with only a 0.18% optimality gap.
Under limited flight ranges (like when β = 0.6 ), data shows that drones tend to stay closer to the trucks. This limits their ability to carry out independent missions. However, the RL-based approaches still keep competitive gaps of less than 2–3% in most cases. This means the learned policy successfully adapts to shorter drone routes. In practice, this suggests that logistics providers can effectively use drones even in challenging environments or when battery life is limited by taking advantage of adaptive route generation to ensure coverage. Additionally, the speed of RL-based sampling and Ensemble solutions is still a significant advantage. This speed is crucial in situations that require quick adjustments, such as when new customer orders come in or traffic conditions change.

5.4. Ablation Studies

To demonstrate the value of differentiating the group’s information in Equation (6), ablation studies are conducted comparing the performance of the RL greedy model with and without LSTM updates in Table 4, where the proposed group information updates are used as an input to the LSTM. As shown in the table, such a group differentiates and minimizes costs in all instances for various values of β .
The ablation studies highlight the importance of updates at the group level for LSTM. When these updates are omitted, a consistent increase in the average cost is observed, typically between 2% and 3% for certain problem sizes. This shows that understanding interactions within the group, especially how each drone-truck pair works together, provides real benefits in performance. Therefore, the system’s ability to capture interactions between multiple agents is key to developing effective routing solutions.

5.5. Routing Solutions

Figure 2 demonstrates the results produced by the RL sampling approach on the first instance on a graph with 10 nodes from [28] with β = 0.6 . As can be seen from the solutions, drones’ utilisation for customers is limited due to the small flying range corresponding to β = 0.6 , where drones and trucks travel together to serve customers. In the case involving two trucks, Figure 2b,c show how the RL approach naturally solves the clustering problem of splitting the networks into two groups even though it has not been explicitly outlined in the MDP problem formulation. Instead, based on trial and error with the RL training, the proposed neural network architectures learn implicitly to cluster nodes to be served by two separate groups while ensuring the minimization of costs. Also, Figure 2 illustrates that increasing the number of drones doesn’t always improve the total serving time, given the small number of customers to serve and the small flying range of drones.

6. Conclusions

In this study, the main focus was on learning and coordinating multiple vehicles with multiple drones to optimize routing dynamics in scenarios involving both vehicles and drones. The exploration and evaluation, especially when testing the model on large and small instances using well-known benchmarks, provided valuable insights into its adaptability and efficiency.
The study, centred around various configurations manipulating factors like drone flight range, the number of vehicles, and drone allocations, consistently demonstrated the comparable performance of RL approaches with traditional methods. Notably, it demonstrated adaptability in scenarios with limited drone flight range, delivering solutions that balance quality and computational speed.
The findings reinforce that the RL model holds significant promise for addressing challenges and optimizing efficiency in drone-assisted vehicle routing. Further exploration of this model, particularly in scenarios involving an increased number of vehicles and drones, presents exciting opportunities to enhance its optimality and robustness in real-world logistics applications.
Additionally, exploring scenarios with an increased number of nodes could provide valuable insights and contribute to the continued advancement of this research field. In conclusion, this research suggests that RL-based models are valuable for planning routes involving vehicles and drones. Exploring its capabilities more thoroughly, the promising potential is seen to improve how deliveries are managed in the future.
Despite these promising directions, the current study has certain limitations that warrant further investigation. First, complex recharging or maintenance schedules for drones are not explicitly modeled, which may limit the applicability of the approach in extended delivery routes or highly variable environments. Second, although strong adaptability is demonstrated by the proposed model in benchmark settings, the robustness would be better validated through additional testing with real-world datasets and unpredictable factors (e.g., weather conditions and traffic disruptions). Finally, while multiple drones and vehicles are considered, exploring larger-scale deployments and more intricate constraints (e.g., multi-objective objectives balancing cost, time, and environmental impact) could uncover new challenges and optimization opportunities. Future work could address these aspects by integrating dynamic recharging strategies, employing real-time data streams for adaptive routing, and evaluating the performance across a broader range of industrial and urban scenarios. It is believed that the proposed reinforcement learning framework can evolve into a fully operational tool for last-mile logistics and beyond by being implemented.

Author Contributions

Conceptualization, B.D. and M.M.; methodology, A.B.; software, B.D.; validation, M.M. and A.B.; formal analysis, M.M.; investigation, A.B.; resources, A.B.; data curation, B.D.; writing—original draft preparation, B.D.; writing—review and editing, A.B.; visualization, M.M.; supervision, A.B.; project administration, M.M.; funding acquisition, A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Higher Education of the Republic of Kazakhstan, IRN AP19575607.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available upon a request from the authors.

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A

We consider a limited set K of vehicles. Furthermore, each vehicle k K holds an equal number of D Z drones assigned to it. Each drone may travel a maximum range of E distance units per operation, where a drone-operation is characterized by a triple ( i , w , j ) as follows: the drone is launched from a vehicle at a node i V L , delivers a parcel to w V N , and is retrieved by the vehicle from which it was launched at node j V R ( i w j and i j ).
We use the following decision variables for modeling the VRPD:
  • x i j k : i , j V , k K – Binary variables that state whether edge ( i , j ) belongs to the route of vehicle k.
  • y i w j k : i , w , j V , k K – Binary variables that specify whether drone delivery ( i , w , j ) is performed by a drone associated with vehicle k.
  • p i j k : i , j V , k K – Binary variables that indicate whether node i is served before (but not necessarily consecutive to) node j in the route of vehicle k.
  • u i k : i V , k K – Integer variables with a lower bound of 0 that state the position of node i in the route of vehicle k, if customer i is served by vehicle k.
  • z a i w j e k : a , i , w , j , e V , k K – Binary variables that specify whether the drone delivery ( a , w , e ) is performed by a drone associated with vehicle k and node a precedes node i while node e follows node j in the route of vehicle k.
  • a i k : i V , k K – Continuous variables with a lower bound of 0 that indicate the earliest time at which customer i is served by either the vehicle k or a drone associated with vehicle k. If i is used as a retrieval node, a i k is the earliest time at which both the (latest) drone and the vehicle have arrived at node i.
The MILP formulation of the VRPD is given in (A1)–(A19), where (A1) is the objective function and the constraints are given by (A2)–(A19):
min τ
s . t . τ a n + 1 k k K
j V N x 0 j k i V N x i , n + 1 k = 0 k K
i V L x i h k j V R x h j k = 0 k K , h V N
j V N x 0 j k 1 k K
j V N x i j k + l V L m V R y l j m k = 1 j V N
2 y i w j k h V L x h i k + l V N x l j k k K , i V N , w V N , j V R
y 0 w j k h V L x h j k k K , w V N , j V R
u i k u j k ( n + 1 ) ( 1 x i j k ) k K , i V L , j V R
u j k u i k 1 ( n + 1 ) ( 1 y i w j k ) k K , i V L , w V N , j V R
u i k u j k 1 ( n + 1 ) ( 1 p i j k ) k K , i V L , j V R
u i k u j k 1 + ( n + 1 ) ( 1 p i j k ) k K , i V L , j V R
3 z a i w j e k y a w e k + p a i k + p j e k k K , a V L , i , w , j V N , e V R
2 + z a i w j e k y a w e k + p a i k + p j e k k K , a V L , i , w , j V N , e V R
M ( x i j k 1 ) + a i k + t i j a j k k K , i V L , j V R
M ( y i w j k 1 ) + a i k + t i w a w k k K , i V L , w V N , j V R
M ( y i w j k 1 ) + a w k + t w j a j k k K , i V L , w V N , j V R
( d i w + d w j ) y i w j k E k K , i V L , j V R , w V N
M ( x i j k 1 ) + v V N y i v j k + l V N m V R y i l m k + p V L , p i q V N y p q j k + a V L w V N e V R z a i w j e k D k K , i V L , j V R
The constraints (A3)–(A19) ensure the flow preservation, synchronization between vehicles and drones, and maximum range of drone operations, as well as satisfying other operational restrictions. The objective function minimizes the arrival time a n + 1 k , which represents the time when the last vehicle or its associated drones return to the depot.

References

  1. Deloison, T.; Hannon, E.; Huber, A.; Heid, B.; Klink, C.; Sahay, R.; Wolff, C. The future of the last-mile ecosystem. In Proceedings of the World Economic Forum, Davos, Switzerland, 21–24 January 2020; Volume 1, pp. 1–28. [Google Scholar]
  2. Mahadevan, P. The military utility of drones. CSS Anal. Secur. Policy 2010, 78. [Google Scholar]
  3. Restas, A. Drone applications for supporting disaster management. World J. Eng. Technol. 2015, 3, 316. [Google Scholar] [CrossRef]
  4. Macrina, G.; Pugliese, L.D.P.; Guerriero, F.; Laporte, G. Drone-aided routing: A literature review. Transp. Res. Part C Emerg. Technol. 2020, 120, 102762. [Google Scholar] [CrossRef]
  5. Borghetti, F.; Caballini, C.; Carboni, A.; Grossato, G.; Maja, R.; Barabino, B. The use of drones for last-mile delivery: A numerical case study in Milan, Italy. Sustainability 2022, 14, 1766. [Google Scholar] [CrossRef]
  6. Wang, X.; Poikonen, S.; Golden, B. The vehicle routing problem with drones: Several worst-case results. Optim. Lett. 2017, 11, 679–697. [Google Scholar] [CrossRef]
  7. Wang, Z.; Sheu, J.B. Vehicle routing problem with drones. Transp. Res. Part B Methodol. 2019, 122, 350–364. [Google Scholar] [CrossRef]
  8. Poikonen, S.; Wang, X.; Golden, B. The vehicle routing problem with drones: Extended models and connections. Networks 2017, 70, 34–43. [Google Scholar] [CrossRef]
  9. Schermer, D.; Moeini, M.; Wendt, O. A hybrid VNS/Tabu search algorithm for solving the vehicle routing problem with drones and en route operations. Comput. Oper. Res. 2019, 109, 134–158. [Google Scholar] [CrossRef]
  10. Kool, W.; van Hoof, H.; Welling, M. Attention, Learn to Solve Routing Problems! arXiv 2019, arXiv:1803.08475. [Google Scholar]
  11. Nazari, M.; Oroojlooy, A.; Snyder, L.; Takác, M. Reinforcement learning for solving the vehicle routing problem. In Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montréal, QC, Canada, 3–5 December 2018. [Google Scholar]
  12. Bogyrbayeva, A.; Meraliyev, M.; Mustakhov, T.; Dauletbayev, B. Machine Learning to Solve Vehicle Routing Problems: A Survey. IEEE Trans. Intell. Transp. Syst. 2024, 25, 4754–4772. [Google Scholar] [CrossRef]
  13. Bogyrbayeva, A.; Yoon, T.; Ko, H.; Lim, S.; Yun, H.; Kwon, C. A deep reinforcement learning approach for solving the traveling salesman problem with drone. Transp. Res. Part C Emerg. Technol. 2023, 148, 103981. [Google Scholar] [CrossRef]
  14. Schermer, D.; Moeini, M.; Wendt, O. Algorithms for solving the vehicle routing problem with drones. In Proceedings of the Intelligent Information and Database Systems: 10th Asian Conference, ACIIDS 2018, Dong Hoi City, Vietnam, 19–21 March 2018; Proceedings, Part I 10. Springer: Berlin/Heidelberg, Germany, 2018; pp. 352–361. [Google Scholar]
  15. Schermer, D.; Moeini, M.; Wendt, O. A matheuristic for the vehicle routing problem with drones and its variants. Transp. Res. Part C Emerg. Technol. 2019, 106, 166–204. [Google Scholar] [CrossRef]
  16. Wu, G.; Fan, M.; Shi, J.; Feng, Y. Reinforcement learning based truck-and-drone coordinated delivery. IEEE Trans. Artif. Intell. 2021, 4, 754–763. [Google Scholar] [CrossRef]
  17. Chiaraviglio, L.; D’andreagiovanni, F.; Choo, R.; Cuomo, F.; Colonnese, S. Joint optimization of area throughput and grid-connected microgeneration in UAV-based mobile networks. IEEE Access 2019, 7, 69545–69558. [Google Scholar] [CrossRef]
  18. Otto, A.; Agatz, N.; Campbell, J.; Golden, B.; Pesch, E. Optimization approaches for civil applications of unmanned aerial vehicles (UAVs) or aerial drones: A survey. Networks 2018, 72, 411–458. [Google Scholar] [CrossRef]
  19. Shi, Y.; Lin, Y.; Li, B.; Li, R.Y.M. A bi-objective optimization model for the medical supplies’ simultaneous pickup and delivery with drones. Comput. Ind. Eng. 2022, 171, 108389. [Google Scholar] [CrossRef]
  20. Trotta, A.; Andreagiovanni, F.D.; Di Felice, M.; Natalizio, E.; Chowdhury, K.R. When UAVs ride a bus: Towards energy-efficient city-scale video surveillance. In Proceedings of the IEEE Infocom 2018-IEEE Conference on Computer Communications, Honolulu, HI, USA, 16–19 April 2018; pp. 1043–1051. [Google Scholar]
  21. Mazyavkina, N.; Sviridov, S.; Ivanov, S.; Burnaev, E. Reinforcement learning for combinatorial optimization: A survey. Comput. Oper. Res. 2021, 134, 105400. [Google Scholar] [CrossRef]
  22. Bello, I.; Pham, H.; Le, Q.V.; Norouzi, M.; Bengio, S. Neural combinatorial optimization with reinforcement learning. arXiv 2016, arXiv:1611.09940. [Google Scholar]
  23. Vinyals, O.; Fortunato, M.; Jaitly, N. Pointer networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  24. Khalil, E.; Dai, H.; Zhang, Y.; Dilkina, B.; Song, L. Learning combinatorial optimization algorithms over graphs. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  25. Kwon, Y.D.; Choo, J.; Kim, B.; Yoon, I.; Gwon, Y.; Min, S. Pomo: Policy optimization with multiple optima for reinforcement learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21188–21198. [Google Scholar]
  26. Kim, M.; Park, J. Learning Collaborative Policies to Solve NP-hard Routing Problems. In Proceedings of the Advances in Neural Information Processing Systems 34 (NIPS 2021), Online, 6–14 December 2021. [Google Scholar]
  27. Ma, Y.; Li, J.; Cao, Z.; Song, W.; Zhang, L.; Chen, Z.; Tang, J. Learning to Iteratively Solve Routing Problems with Dual-Aspect Collaborative Transformer. In Proceedings of the Advances in Neural Information Processing Systems 34 (NIPS 2021), Online, 6–14 December 2021. [Google Scholar]
  28. Agatz, N.; Bouman, P.; Schmidt, M. Optimization approaches for the traveling salesman problem with drone. Transp. Sci. 2018, 52, 965–981. [Google Scholar] [CrossRef]
  29. Schermer, D.; Moeini, M.; Wendt, O. A Variable Neighborhood Search Algorithm for Solving the Vehicle Routing Problem with Drones; Technacal Report; Technische Universität Kaiserslautern: Kaiserslautern, Germany, 2018. [Google Scholar]
  30. Bogyrbayeva, A.; Jang, S.; Shah, A.; Jang, Y.J.; Kwon, C. A reinforcement learning approach for rebalancing electric vehicle sharing systems. IEEE Trans. Intell. Transp. Syst. 2021, 23, 8704–8714. [Google Scholar] [CrossRef]
  31. Williams, R.J. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 1992, 8, 229–256. [Google Scholar] [CrossRef]
Figure 1. An overview of the attention-encoder group-based LSTM-decoder model.
Figure 1. An overview of the attention-encoder group-based LSTM-decoder model.
Applsci 15 02007 g001
Figure 2. Drones’ and trucks’ routes produced by the RL sampling approach with β = 0.6 on a graph with 10 nodes. The solid lines in black and blue colors correspond to the first and second truck’s routes, and the dashed lines correspond to their assigned drones’ routes.
Figure 2. Drones’ and trucks’ routes produced by the RL sampling approach with β = 0.6 on a graph with 10 nodes. The solid lines in black and blue colors correspond to the first and second truck’s routes, and the dashed lines correspond to their assigned drones’ routes.
Applsci 15 02007 g002
Table 1. The VRPD results with Infinite Flying Range of Drones. Averages of 10 problem instances. ‘Cost’ refers to the average cost value. ‘Gap’ is the mean relative difference to the cost of the best algorithm for each instance. ‘Time’ is the average solution time of the algorithm for a single instance.
Table 1. The VRPD results with Infinite Flying Range of Drones. Averages of 10 problem instances. ‘Cost’ refers to the average cost value. ‘Gap’ is the mean relative difference to the cost of the best algorithm for each instance. ‘Time’ is the average solution time of the algorithm for a single instance.
| K | = 1 , | D | = 2 | K | = 2 , | D | = 1 | K | = 2 , | D | = 2
Methods Costs Gaps, % Time, s Costs Gaps, % Time, s Costs Gaps, % Time, s
| N | = 8 Solver158.530.51175.53149.531.10601.33125.551.09481.62
Solver Lazy158.530.5176.97147.600.00148.52123.90.00161.35
Heuristic158.530.51154.55150.321.9461.84125.51.0665.06
RL greedy161.692.380.02148.680.900.03126.982.250.21
RL sampling160.231.600.87148.320.591.31126.872.144.00
Ensemble RL159.951.456.02147.600.009.27126.872.1428.18
| N | = 10 Solver196.4811.25601.29193.7516.42601.52171.7830.89601.69
Solver Lazy195.2610.78556.72200.9120.42600.67176.6334.64602.05
Heuristic178.931.42308.29167.590.8263.68131.360.2399.1
RL greedy185.424.740.01168.401.220.40136.063.580.03
RL sampling180.261.890.43167.570.711.43132.400.922.06
Ensemble RL177.010.283.05167.560.709.38134.202.2314.65
| N | = 20 RL greedy243.063.860.02219.026.910.02185.888.170.32
RL sampling234.450.281.31209.652.124.24174.881.718.01
Ensemble RL234.090.128.94205.670.1529.42172.640.3955.60
Table 2. The VRPD results with Limited Flying Range of Drones with β = 1 . Averages of 10 problem instances.
Table 2. The VRPD results with Limited Flying Range of Drones with β = 1 . Averages of 10 problem instances.
| K | = 1 , | D | = 2 | K | = 2 , | D | = 1 | K | = 2 , | D | = 2
Methods Costs Gaps, % Time, s Costs Gaps, % Time, s Costs Gaps, % Time, s
| N | = 10 Solver238.459.55601.18223.5413.66602.9217.4515.23601.86
Solver Lazy246.8412.89601.27223.5813.95602.1215.8114.33601.22
Heuristic218.010.28304.48196.100.0062.82188.960.0081.43
RL greedy232.246.510.03205.704.770.07198.475.140.02
RL sampling221.722.040.63202.763.300.50195.463.520.80
Ensemble RL219.781.044.54200.162.033.88193.172.305.44
| N | = 20 Heuristic264.5010.14268.97213.900.00326.69196.590.94498.83
RL greedy252.185.090.02231.808.300.03218.3612.020.03
RL sampling244.321.681.40222.744.061.60201.793.731.98
Ensemble RL240.010.009.67220.633.0911.36201.633.5811.96
| N | = 50 Heuristic475.8735.46141.93320.666.51128.64295.6815.84151.92
RL greedy391.5811.370.04318.295.710.05276.618.320.05
RL sampling357.771.734.76305.681.486.79261.612.497.66
Ensemble RL351.630.0033.18301.950.2547.65255.740.2054.45
Table 3. The VRPD results with Limited Flying Range of Drones with β = 0.6 . Averages of 10 problem instances.
Table 3. The VRPD results with Limited Flying Range of Drones with β = 0.6 . Averages of 10 problem instances.
| K | = 1 , | D | = 2 | K | = 2 , | D | = 1 | K | = 2 , | D | = 2
Methods Costs Gaps, % Time, s Costs Gaps, % Time, s Costs Gaps, % Time, s
| N | = 10 Solver324.7516.96567.69260.9812.44601.02247.048.20600.91
Solver Lazy282.903.95476.00240.474.91601.71238.003.98601.68
Heuristic272.920.00123.89229.390.0162.33228.580.0163.45
RL greedy281.621.750.01231.230.800.02230.991.130.02
RL sampling276.931.400.61230.420.450.68230.430.880.98
Ensemble RL276.321.284.47229.530.084.79228.840.126.80
| N | = 20 Heuristic290.083.28229.68236.830.00203.88230.410.34290.28
RL greedy292.834.320.03253.667.130.02248.327.810.04
RL sampling286.132.011.39249.185.281.36237.763.832.94
Ensemble RL283.430.969.79242.892.479.47235.592.6615.74
| N | = 50 Heuristic469.9333.13152.57309.956.32139.44296.2811.41159.92
RL greedy376.766.610.06306.955.240.07279.955.300.07
RL sampling355.710.706.12294.771.066.90266.390.187.72
Ensemble RL353.270.0043.15291.640.0048.74266.680.3084.59
Table 4. The ablation studies on | N | = 10 comparing RL Greedy models with and without the LSTM group updates. The averages of 10 instances were reported for costs and gaps. The relative gap is computed for each instance with respect to the best model result among all benchmark models.
Table 4. The ablation studies on | N | = 10 comparing RL Greedy models with and without the LSTM group updates. The averages of 10 instances were reported for costs and gaps. The relative gap is computed for each instance with respect to the best model result among all benchmark models.
| K | = 1 , | D | = 2 | K | = 2 , | D | = 1 | K | = 2 , | D | = 2
β Methods Costs Gaps, % Costs Gaps, % Costs Gaps, %
1RL Greedy w/t group LSTM234.427.26209.646.83214.9913.79
RL Greedy232.246.24205.704.77198.475.14
2RL Greedy w/t group LSTM185.814.75170.712.40137.634.71
RL Greedy185.424.50168.400.98136.063.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bogyrbayeva, A.; Dauletbayev, B.; Meraliyev, M. Reinforcement Learning for Efficient Drone-Assisted Vehicle Routing. Appl. Sci. 2025, 15, 2007. https://doi.org/10.3390/app15042007

AMA Style

Bogyrbayeva A, Dauletbayev B, Meraliyev M. Reinforcement Learning for Efficient Drone-Assisted Vehicle Routing. Applied Sciences. 2025; 15(4):2007. https://doi.org/10.3390/app15042007

Chicago/Turabian Style

Bogyrbayeva, Aigerim, Bissenbay Dauletbayev, and Meraryslan Meraliyev. 2025. "Reinforcement Learning for Efficient Drone-Assisted Vehicle Routing" Applied Sciences 15, no. 4: 2007. https://doi.org/10.3390/app15042007

APA Style

Bogyrbayeva, A., Dauletbayev, B., & Meraliyev, M. (2025). Reinforcement Learning for Efficient Drone-Assisted Vehicle Routing. Applied Sciences, 15(4), 2007. https://doi.org/10.3390/app15042007

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop