Previous Article in Journal
Organ-on-a-Chip and Lab-on-a-Chip Technologies in Cardiac Tissue Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Bio-Inspired Reactive Approaches for Automated Guided Vehicle Path Planning: A Review

1
School of Computer Engineering, Jimei University, Xiamen 361000, China
2
Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia
3
School of IT and Engineering, Melbourne Institute of Technology, Sydney Campus, Sydney, NSW 2000, Australia
*
Author to whom correspondence should be addressed.
Biomimetics 2026, 11(1), 17; https://doi.org/10.3390/biomimetics11010017 (registering DOI)
Submission received: 22 November 2025 / Revised: 22 December 2025 / Accepted: 28 December 2025 / Published: 30 December 2025
(This article belongs to the Section Biological Optimisation and Management)

Abstract

Automated guided vehicle (AGV) path planning aims to obtain an optimal path from the start point to the target point. Path planning methods are generally divided into classical algorithms and reactive algorithms, and this paper focuses on reactive algorithms. Reactive algorithms are classified into swarm intelligence algorithms and artificial intelligence algorithms, and this paper reviews relevant studies from the past six years (2019–2025). This review involves 123 papers: 81 papers are about reactive algorithms, 44 are based on the swarm intelligence algorithm, and 37 are based on artificial intelligence algorithms. The main categories of swarm intelligence algorithms include particle swarm optimization, ant colony optimization, and genetic algorithms. Neural networks, reinforcement learning, and fuzzy logic represent the main trends in artificial intelligence–based algorithms. Among the cited papers, 45.68% achieve online implementations, and 33.33% address multi-AGV systems. Swarm intelligence algorithms are suitable for static or simplified dynamic environments with a low computational complexity and fast convergence, as 79.55% of papers are based on a static environment and 22.73% achieve online path planning. Artificial intelligence algorithms are effective for dealing with dynamic environments, which contribute 72.97% to online implementation and 54.05% to dynamic environments, while they face the challenge of robustness and the sim-to-real problem.

Graphical Abstract

1. Introduction

Automated guided vehicles (AGVs) are widely used in logistics, manufacturing systems, port terminals, and industrial automation due to safety, reliability, flexibility, efficiency, and scalability [1,2,3]. The AGV system consists of an embedded controller, vehicle chassis, communication devices, battery, sensors, guidance facilities, and a load transfer device [3]. Path planning obtains a continuous curve that starts from a starting point and ends at a target point, including the defined positions in the path [4]. Optimal path planning attempts to optimize the cost function to optimize the path, considering the time or distance [4].
Optimal path planning can be beneficial for safe and efficient transport in the production process, and it requires low cost and latency, precise positioning and remote control [2,5]. Path planning approaches are classified as online or offline implementation [6]. Online path planning considers dynamic environments or real-time applications, while offline path planning is aimed at static obstacle avoidance. The real-time path planning of AGVs in an unknown environment that still remains a challenge in smart logistics applications [7].
The path planning algorithms can be classified as classical algorithms and reactive algorithms, as shown in Figure 1. The classical algorithms include Dijkstra [8,9,10,11,12,13,14], A* [15,16,17,18,19,20], D* [21,22,23,24], Rapidly-exploring random tree (RRT) or RRT* [25,26,27,28,29,30,31,32,33], artificial potential field (APF) [34,35,36,37,38,39], probabilistic roadmap (PRM) [40,41,42,43], etc.
However, reactive algorithms have drawn attention due to their robust learning capabilities [44]. This review focuses on the reactive algorithms developed over the past six years (2019–2025), including swarm intelligence algorithms and artificial intelligence algorithms, as shown in Figure 2. The swarm intelligence algorithms are inspired by natural behavior to gain the optimal solution based on the fitness functions. Artificial intelligence algorithms, such as neural networks, reinforcement learning, and fuzzy logic, are developed due to their learning abilities and adaptability to dynamic environments.
This review focuses on the reactive path planning algorithms, and the literature was collected through Google Scholar, which supports broad coverage of conference proceedings and journal articles, with the published year restricted to 2019–2015, as shown in Figure 3. The keyword combinations include “AGV path planning”, “AGV reactive path planning algorithms”, “AGV path planning swarm intelligence”, “AGV path planning AI”, and the particular algorithm family, such as “AGV path planning RL”.
Papers are included in the analysis of this review if they implement reactive path planning algorithms, including swarm intelligence algorithms or artificial intelligence algorithms, in the AGV’s scenario or the automated vehicle’s scenario with simulated or experimental validation, and are restricted to the conference proceedings or journal articles. Papers are excluded if they employ classical approaches, address only task assignment problems, or involve scenarios such as underwater vehicles or in-flight drones.
This review includes a total of 123 papers, with 81 papers about reactive algorithms and 44 papers based on swarm intelligence algorithms, including 11 papers about particle swarm optimization (PSO), 18 papers about ant colony optimization algorithm (ACO), 9 papers about genetic algorithms (GA), and 6 papers about other swarm intelligence algorithms. For the artificial intelligence algorithms, there are 37 papers, including reinforcement learning (27), neural networks (5), fuzzy logic (4), and other algorithms (1). Most papers are published by IEEE and Elsevier, as shown in Figure 4.
Several survey papers are related to robot or vehicle path planning, such as [6,45]. Lin et al. [6] provides path planning approaches based on the perspective of robots and multi-robot systems, considering the centralized and decentralized decision-making systems and the classification of the algorithms. Reda et al. [45] reviews the models of autonomous driving systems; path planning is treated as part of the autonomous driving systems, and others are about perception, sensors, localization, control, and assessment. However, this paper aims to provide a comprehensive analysis of recent studies, reviewing the research after previous review papers were published (2023–2025). Also, this paper compares the cited studies under the consideration of dynamic environments, online implementation, and multi-AGV coordination, focusing on AGV scenarios.
Section 2 reviews swarm intelligence algorithms, and Section 3 provides the analysis of artificial intelligence algorithms. Section 4 lists the algorithms that are not included in Section 2 and Section 3. This paper compares the algorithms and concludes in Section 5.

2. Swarm Intelligence Algorithms

2.1. Particle Swarm Optimization (PSO)

The hybrid PSO-SA algorithm is proposed for AGV path planning in the warehouse to avoid the local optimum problem with less time consumption and faster convergence and minimize the path length and obtain a smooth path, which is inspired by the simulated annealing algorithm (SA) [5]. It indicates that a dynamic environment, multiple AGVs, or moving obstacles would be future improvements. Qiuyun et al. [46] introduced an improved PSO for a one-line production line for the shortest transportation time, designing a crossover operation and mutation mechanism to avoid falling into the local optimum, while the scenario is simple, and the algorithm cannot be applied to multiple AGVs.
To balance the performance of exploitation and exploration, Lin et al. [47] presents a hybrid optimization algorithm with probability based on PSO and the cultural algorithm, updating the inertia weight according to the improved Metropolis rule, aiming for multi-AGV path planning. However, it lacks the practical implementation and considerations of dynamics. Gul et al. [48] developed a PSO-GWO optimization algorithm based on PSO and the grey wolf optimizer (GWO), which integrated with a local search technique, considering the path length and smoothness. Although it considers two objectives, the problem formation is not a multi-objective optimization. It cannot be implemented in real-time; the moving goal and multiple robots are considered as future work.
Multi-objective PSO (MOPSO) is combined with the dynamic window approach (DWA) to address the complex environment considering collision avoidance, travel time, and smoothness, while the real environment is not considered, such as environmental uncertainties or real experiments [49]. Zhang et al. [2] presented an energy-efficient path planning algorithm based on the PSO, considering multi-objective optimization, including energy consumption and total execution time. But it lacks considerations of energy consumption data, the transport task execution, or the multi-AGV system.
Ahmad et al. [50] presents a global path planning based on an improved PSO algorithm, which introduces alpha and beta as coefficients to adjust movements and balances safety, time, and distance in path planning. However, it is only implemented in a static simulated environment; path prediction and learning capabilities should be improved. The PSO algorithm is combined with the human learning optimal algorithm to enhance search efficiency and convergence speed, but it is applied only to the single robot system in the static environment [51].
Song [52] presents a global path planning algorithm based on PSO with Levy flight and an inductive steering algorithm, and it considers speed control for safety, while the dynamic situation is a simple scenario. An improved PSO is developed based on ant colony optimization, which presents a collision avoidance factor to optimize the node waiting time and AGV path planning [53] while the situation is offline. The PSO algorithm is utilized to adjust the initial parameters of the ant colony optimization algorithm and investigates the expelling behavior and the elite ant principle for updating the pheromone, while it only provides a simulation of a static environment [54].
Table 1 compares the cited PSO-based algorithms from the perspectives of the properties, considerations, environment modeling, and online operations. Most studies use PSO to optimize path length, smoothness, collision avoidance, or multi-objective optimization with MOPSO. The PSO-based algorithms achieve optimization by updating particles iteratively. The PSO-based algorithms are usually employed for static or simplified dynamic environments (36.36%), and most experiments are achieved with simulation, which reaches 90.91%. The experiments are mostly achieved by simulation with a single-robot scenario, with a proportion of 81.82%, and the main environmental models are represented by the grip map. The percentage of online implementation remains restricted at 36.36%, and the algorithms are suitable for structured manufacturing or warehouse scenarios with predicted obstacles.

2.2. Ant Colony Optimization Algorithm (ACO)

An improved ACO algorithm is optimized for a multi-AGV production workshop based on job similarity, multi-objective programming, and a pheromone matrix, which can achieve a faster convergence speed and a shorter maximum time span [55]. It can consider further scenarios, such as flexible job shops or flow shop manufacturing environments in future work. Wang et al. [56] developed an improved ACO for the intelligent parking system with the fallback strategy, valuation function, and the reward/penalty mechanism for the pheromone update strategy, but the efficiency of the algorithm would be reduced if the size of nodes exceeds 1000.
A novel ACO is detailed in [57] by adding a penalty strategy to enhance the exploration of unknown areas with the worst value; however, it lacks a real-world experiment. Modified adaptive ACO employs an improved heuristic function, state transition probability rule, and distribution of initial pheromone concentration to improve the swarm diversity and search efficiency, reducing the path length and turn times [58]. However, the execution time is not competitive when compared with direct search algorithms, and the three-dimensional and multi-objective optimization problems should be paid attention to.
Li et al. [59] introduces grey wolf optimization (GWO) into ACO for improving the pheromone model and adds corner constraints for path smoothness to accelerate the convergence, but it lacks a comparison analysis. Zhou and Huang [60] combines ACO and Dijkstra for the baggage pickup sequencing and AGV path planning for the airport AGV, but it needs to consider multi-AGV conflicts.
ACO is improved with A* multi-directional algorithm to obtain the path, and uses the Markov Decision Process (MDP) trajectory evaluation model to filter and generate the smooth global path in [61]. However, dynamically moving obstacles can be a future direction. Based on the improved ACO and rolling window method, a dynamic path planning algorithm and a second-level safety distance determination rule are proposed in a complex environment [62], while the convergence performance and obstacle avoidance could be further improved.
Wang et al. [63] incorporates fast-scaling RRT* into the ACO algorithm, and it uses dynamic step size strategies, heuristic dynamic sampling, and the two-way search to accelerate speed, but it only focuses on a static environment. Step optimization and path simplification methods are designed to improve the ACO algorithm to avoid low search efficiency, and an adaptive pheromone volatilization coefficient and load balancing strategies are presented for multi-AGV systems [64]. However, it needs to consider conflict resolution in the future.
A hexagonal grid map model is presented in [65], which is used in ACO path planning with a regenerated heuristic factor and a bidirectional search strategy for an intelligent manufacturing system. The future work should concentrate on the robustness of the grid map, search abilities, and real-world experiments. Chen and Yu [66] implements Q-value to adjust the parameters of ACO to enhance algorithm convergence and obstacle avoidance ability, while it is only implemented in static and simple environments and lacks a comparison analysis.
Li et al. [67] designs quantum ACO for optimal and feasible paths based on Bloch coordinates of pheromones and uses a repulsion factor for the space–time distance in an automated container terminal. However, this approach has not been validated in an uncertain environment or real logistics systems. For weak optimization ability and slow convergence of ACO, Ref. [68] employs the fruit fly optimization algorithm (FOA) for pre-searching and the original pheromone distribution anduses ACO for global path planning, but it is only for static, simple environments and lacks comparisons.
To address the problems of path runs and convergence speed, Xiao et al. [69] combines ACO and DWA for indoor AGV global path planning, but it focuses on a static environment and a single AGV. Wu et al. [70] adds the information of the nodes into the heuristic information and the dynamic adjustment factor to guide the information, and introduces the Laplace distribution for the pheromone. However, it is only concerned with static job scheduling.
To improve the optimization effect and search efficiency, non-uniform, and directed distribution of initial pheromones, the adaptive adjustment of iterations and the optimization of parameters by GA are presented to improve the ACO, but it lacks a solid comparison [71]. For operating environments, prior time is introduced, the pheromone increment of the ACO algorithm is modified to minimize running time, and the overall task execution time and distance factors are considered in the pheromone update stage. However, it is not suitable for large-scale or frequently changing tasks [72].
The ACO-based algorithms are compared in Table 2, which usually consider distance, obstacle, and turning angles. The environments can be modeled as a grid or raster map. The ACO-based algorithms face the challenges of limited robustness and difficulty adapting to moving obstacles or changing tasks. The proportions of online implementation and dynamic environments are both 5.56%. The pheromone-based iterations also result in a low rate of multi-AGV scenarios, which reaches only 16.67%. Most cited ACO-based papers are validated through simulations; only 5.56% conduct the experiments.

2.3. Genetic Algorithm (GA)

Lyu et al. [73] proposed an integrated scheduling approach with conflict-free path planning based on GA and Dijkstra with a time window, which optimized the number of AGVs, but it does not consider dynamic scheduling and job sequencing problems. Zhong et al. [74] described a conflict-free multi-AGV path planning, which combines GA and PSO with a fuzzy logic controller for efficiency and reliability in automated container terminals, but it took lengthy computation time and cannot support real-time scheduling.
For a logistics system, a three-stage model is designed for task assignment and speed control based on the GA algorithm and simulation, but it lacks comparison with other algorithms or research on AGV charging [75]. A hybrid GA/heuristic is offered in [76] for a cellular manufacturing system to minimize the intercellular movements and the makespan costs in which cell formation problems are designed as a fuzzy mixed-integer linear programming model. Nonetheless, the article only optimizes several problems and is not related to the real case.
Wu et al. [77] considered the blocking of buildings and flight heights when performing the surveillance tasks and outlined a hybrid EDA-GA algorithm for the cooperative path planning, then applied an online local adjustment strategy for the changes of the requirements. The approach may be used for more applications in the future. GA incorporates ACO to improve the initial population, and it considers path smoothness in [78]. It introduces a three-stage mutation operation inspired by the SA algorithm, while it lacks comparison with the SOTA algorithms.
GA is improved for dual-AGVs to ensure efficient and safe actions with a fitness function, and conducted an experiment in leader–follower ROS AGVs [79], but it only focuses on static obstacles. For forklift and latent AGVs, the A* algorithm and the GA are designed as a two-stage optimization model with cyclic rules and a penalty function [80]. However, its scenario only includes static obstacles, and the charging problem is not considered. Farooq et al. [81] presents an improved GA within spinning drawing frames for multi-AGV decision-making and path planning to meet the real-time requirement, which uses time-dependent and time-independent variables as decision variables to minimize the path length, but it is only compared with the traditional GA.
Table 3 presents the GA-based algorithms that implement selection, mutation, and crossover operators to determine the AGV paths. Additionally, 33.33% of the cited GA-based papers achieve online implementation, and 88.89% are based on multi-AGV scenarios. They handle the optimization problem with completion time, makespan-related costs, safe movement, path length, and smoothness. GA-based approaches are also mostly validated by the simulation in a static environment, with 22.22% of both dynamic environments and experimental validation. Moreover, 77.78% of the GA-based algorithms integrate with other classical approaches, swarm intelligence algorithms, or fuzzy logic controllers to achieve better performance.

3. Artificial Intelligence Algorithms

3.1. Neural Network (NN)

Sung et al. [82] outlined a neural network with offline training and online path planning, which uses the Bellman–Ford algorithm and a quadratic program for offline neural network training in a grid-based graph to minimize the sum of the distances. Despite this, it is hard to acquire ideal situational awareness, and the large size of trained data and the increased dimensionality are the weaknesses.
A recurrent deep neural network with long short-term memory (LSTM) is utilized for the AGV parking maneuver, and it uses an adaptive learning tracking control algorithm for controlling the motions, considering the shortest time, collision avoidance, and the process and terminal costs [83].
For warehouse logistics, Zhang et al. [84] integrates advanced neural networks within the ACO model with a congestion-aware loss function and an adaptive attention mechanism, but the environment model is not clear. Sun et al. [85] combines the A* algorithm and the NAR neural network in the 2D maps, which uses real-time and historical data to establish the NAR neural network, but the success rate is a little low.
Zhang et al. [44] designs a three-layer structure, with the first layer employing the target area adaptive RRT* for data collection, the second layer usning a deep neural network to train the model for learning the relationships between sampling and states, and the third layer using the model to guide RRT* sampling. Its future research focuses on kinematic information, transfer learning, and 3D scenarios.
Table 4 lists the cited model based on the neural network. These models usually combine with the global path planning approach and learn the mapping between the environment and decisions. The NN-based models consider path length, obstacles, or motion planning, and can model the environment as a graph or a grid map. Most models are validated through simulation and can be implemented online. Forty percent of the models consider a dynamic environment, with a single robot scenario.

3.2. Reinforcement Learning (RL)

DQN is integrated with a state-dynamic network model to improve the convergence speed in [86], and it uses a distributed training framework for decision-making, while the collaborative navigation still needs to be improved. Yang et al. [87] used the A* algorithm for path planning in a static environment as a priori knowledge, then used the improved Deep-Q network (DQN) algorithm on a semi-known environment to address the problem of excessive randomness and slow convergence, but the improvements can be further based on the local path planning ability.
Xiao et al. [88] designs an improved DQN algorithm based on a dynamic temperature adjustment mechanism and the priority experience replay mechanism and uses a refined multi-objective reward function to guide the path. Sensor noise and dynamic obstacle prediction modules in the real-world experiment should be considered in future research.
Dueling DQN is integrated with prioritized experience replay, which considers position, velocity, and target [89]. It processes multimodal sensing information, but it could consider multi-agent reinforcement learning for simultaneous path planning in the future. A digital twin model is introduced based on an improved dueling double deep Q network (D3QN) at vertical and horizontal levels for resource allocation settings, which implements count-based exploration [90]. Its future work would consider synchronization of manufacturing systems’ activities and multi-resource production scheduling.
Deep reinforcement learning (DRL) and recurrent neural network (RNN) are combined for multi-AGV path planning in [91], employing LSTM and proximal policy optimization (PPO). Although it can deal with sudden failures or temporary changes, the computational time could be further reduced and the model evaluated for dynamic conflict avoidance strategies.
Nie et al. [92] improves the PPO algorithm with sample regularization and adaptive learning rate, which adjusts the action probability density and learning rate to enhance the stability and convergence speed. However, it lacks a real experiment, and its future work would focus on global path planning. The curiosity mechanism is integrated in the PPO method to consider the sparse external rewards and dynamic obstacles, while it cannot guarantee safety in the training [93].
For collaborative multi-AGV systems, Shi et al. [94] presents a framework based on multi-agent PPO and GNN to improve decision-making and local perception, and it uses an RRT-guided mechanism for training. However, it focuses on the simple dynamic simulation environment. The intrinsic curiosity module (ICM) and LSTM are introduced into the PPO algorithm, but obstacle avoidance is effected by the speed of moving obstacles or by obstacle not following regular patterns [95].
Yu et al. [96] uses A* for generating global paths to guide the MAPPO algorithm for solving the problem of deadlock and conflicts, and MAPPO is for local path planning. Its reward function accumulates penalties on movement steps, boundaries, and obstacles collisions, while it is only applied on the single-robot system. Ref. [97] uses accepted–rejected sampling to generate points to be the states of Q-learning, but it lacks the modification of Q-learning.
Q-learning is combined with a Kohonen network, as a Kohonen Q-learning algorithm, and integrates the improved GA into the scheduling policy for global path planning [98]. However, it is only suitable for a simple task scheduling scenario. To improve the efficiency, Guo et al. [99] adds a learning process into the Q-learning algorithm for faster path planning than the traditional Q-learning algorithm, but it only investigates static obstacles.
Gao et al. [100] combines Q-learning and a contract net protocol for multi-AGV dispatching problems, but its comparison analysis is weak, and it only implements the traditional Q-learning method. According to the dynamic real environment, digital twin-driven Q-learning is proposed to solve the path planning problem on production logistics systems, with locations and destinations of all AGVs [101]. However, it is not suitable for complex scenarios.
For Industry 4.0, Hu et al. [102] presents a self-adaptive traffic control model based on Q-learning and behavior trees to prevent collisions at intersections, but it is only suitable for simple simulation circumstances. Huang and Wang [103] employs a beetle antennae search algorithm for initiating the Q table to get rid of the local optimum and introduces a gradual Epsilon–Green algorithm, but it cannot be adapted to dynamic obstacles.
In the shared charging system, a hybrid model to obtain optimal paths, forecast channel flow, and recognize congested regions is introduced in [104] based on ACO and Q-learning and adds a positive ant colony feedback mechanism to maximize efficiency. However, it is only compared with the traditional algorithms in a static environment. Tian and Yang [105] implements a distributed Q-learning for multi-AGV planning. It combines action replanning and map training, considering turning rewards and dynamic priority. However, it only focuses on the simple static environment and lacks considerations of dynamics.
For the intelligent manufacturing workshops, deep Q-learning achieves AGV path planning based on a neighborhood weighted grid modeling method, experience replay pool, and the direction reward function in an unknown environment, but it only concerns the static environment [106]. A path optimization model is presented for the port environment based on the APF and twin-delayed deep deterministic policy gradient (DDPG) framework to guarantee the safety and smoothness of the path, while it cannot deal with the dynamic or real environment [107].
MADDPG is improved with an epsilon-greedy policy to avoid obstacles and minimize energy consumption, which balances exploitation and exploration, but it has not considered optimal values [108]. Guo et al. [109] presents a composite auxiliary reward for a soft actor–critic-based RL model, and it utilizes sum-tree prioritized experience replay for real-time control, but it is only validated in simulation. Wang et al. [110] combines an improved ACO algorithm and the Dnya-Q method to improve execution and path planning efficiency, based on the improved heuristic method, but it does not have a comparison analysis.
Guo et al. [111] presents a pioneering decentralized path planning model to address the scalability limitations of the traditional algorithms. It uses local observations and designs a reward function and state space to avoid collisions. However, the high density of obstacles would affect the algorithm. An improved Dyna-Q method is designed for AGV path planning, and it uses a global path guidance to reduce the path search space [112], but it only compares the model with the traditional reinforcement learning algorithms.
Table 5 summarizes the cited reinforcement learning model. The popular RL-based models are based on Q-learning, PPO, and DQN, and the agents learn the policies by interacting with the environment. Here, 66.67% of the cited papers consider online decision-making, and 55.56% perform in dynamic environments. Additionally, 37.04% of these models could be applied to multi-robot systems, such as a logistics system, and 14.81% are validated by experiments, while the sim-to-real problem remains a challenge.
From Table 5, value-based models, such as Q-learning and DQN, are frequently implemented in low-dimensional or discretized spaces with reasonable sample efficiency, while their scalability to highly dynamic environments is limited. Policy-gradient and actor–critic models, such as PPO and SAC, are the recent trend in RL-based studies. They are suitable for continuous control problems, and most of them can perform online path planning. However, these models are sensitive to the reward design functions; the safety cannot be guaranteed during the training, and the deployment in real-world systems remains a challenge, especially the sim-to-real transfer problem.

3.3. Fuzzy Logic (FL)

The elite strategy and the rank-based ant system are utilized to improve ACO and integrate fuzzy logic for dynamic environments such as the FLACO, selecting the optimal path based on travel time and distance, pollutant emissions, and fuel cost [113]. FLACO can be further optimized and extended for a group of vehicles. Considering lane lines, obstacles, and velocities, Ref. [114] presents a hybrid APF-model predictive controller (MPC) based on a fuzzy logic system to adjust the coefficients in the port environments, while precise AGV modeling would be required.
Zhou et al. [115] adopts fuzzy controllers for adjustment coefficients, security, and direction, with an improved ACO and DWA algorithms, and the improved ACO involves a reward and punishment mechanism, but it is a static environment. Ambuj and Machavaram [116] presents a hybrid control strategy based on an improved A* algorithm, which is combined with DWA, which reduces the average path search time, path length, and search grid size. It integrates the PID controller with the adaptive neuro-fuzzy inference system, while it needs to improve the applicability and robustness of the algorithm, and conduct real-world experiments.
Table 6 lists the cited FL model. FL-based approaches integrate with global path planning methods, such as ACO, DWA, GA [74], etc., with 75% of the cited papers considering the dynamic environment, and treating travel cost, velocities, safety, and distance as the objective function. FL-based approaches are evaluated in simulation environments, with online performance to handle uncertainty.

4. Others

For AGV sorting systems, Wang et al. [117] proposes an SVM-based model and a temporary target selection algorithm to enhance dynamic path planning, while the model transfer methodology needs to be further developed. BDE-Jaya is presented in [118] for multiple AGVs in a matrix manufacturing workshop to minimize transportation cost, total tardiness, early service penalty with the developed key-task shift, an insertion-based repair method, and three offspring generation methods to improve exploration and exploitation capability. Practical constraints, production environments, multi-objective optimization, and reinforcement learning are treated as future work.
To overcome the local minimum problem, the dynamic enhanced firework algorithm is presented in [119] to enhance the performance of APF, whose optimization objects are path smoothness and safety; incorporating personalized driving style could be a future improvement.
Zhang et al. [120] developed an improved sparrow search algorithm to consider the risk degree, path acquisition time, distance value, and the total rotation angle value based on the linear path strategy, a new neighborhood search strategy, and a multi-index evaluation method. However, it lacks real experimentation, the application of a multi-robot system, and dynamic obstacles. Guo et al. [121] combines GWO and a Kalman filter and uses partially matched crossover mutation operations and an elite strategy for optimization, but it is not validated in dynamic or real-time environments.
An integrated framework is presented in [122] to address AGV path planning, machine scheduling, and process route selection, and it is based on a hybrid variable neighborhood differential evolution (DE) to maximize make span and ensure collision-free operation. However, the AGVs’ speed and multi-objective optimization are not considered. Zhou et al. [123] introduced an artificial fish swarm algorithm for global path planning and applied Markov chain into a trail-based forward search for unforeseen obstacles, using a multiconstrained model predictive controller to calculate command signals, but lacks comparison with state-of-the-art algorithms. The comparison of the above algorithms is listed in Table 7.

5. Discussion and Conclusions

5.1. Discussion

This review analyzes the cited reactive AGV path planning algorithms, including swarm intelligence and artificial intelligence algorithms published in 2019–2025. The distribution of the publishers is shown in Figure 5. More precisely, Figure 6 indicates the yearly distribution of the cited papers included per algorithm category, which demonstrates the increasing trend of AI-based approaches.
Compared with the previous survey work [6,45], this review focuses on AGV path planning algorithms, rather than the algorithm classification, decision-making strategies, or system-level review. It provides the perspective from scenario properties, environment settings, experimental validation, and multi-AGV coordination. Table 8 and Table 9 compare swarm intelligence and artificial intelligence algorithms for AGV path planning from the aspect of the papers’ contributions and limitations or future research.
From the literature, PSO, ACO, and GA are the main swarm intelligence algorithms used in AGV path planning. Only 22.73% of the swarm intelligence algorithms in the literature achieve online path planning. Meanwhile, 79.55% of these papers’ environmental properties are based on a static environment, and 34.09% consider the multi-AGV system. Additionally, 65.91% of the algorithms are presented as a hybrid approach, which integrates with other algorithms, such as DWA [49,69,73], SA [5,78], GWO [48,59], A* [61,80], RRT* [63], etc.
The scenarios include manufacturing workshops [2,73,76,118], warehouse [5], production workshops [46,55], airports [60], automated container terminals [67,68], and urban environments [77]. The considerations of AGV path planning achieved by the swarm intelligence algorithms mainly concern path length [5,56,57,60,65], energy consumption [2,49,80], transportation/completion time [46,55,74,75,80], turning times and angles [61,68,69], path smoothness [78,119,121], and obstacle avoidance [66,67,121].
For the AI-based approaches, DQN, PPO, and Q-learning are the most popular models in the cited papers. The percentage of papers that achieve online implementation reaches 72.97%, and 54.05% of properties are in a dynamic environment. However, 32.43% of the cited papers are presented for the multiple AGVs, and 94.59% combine other approaches for better performance, including LSTM [83,91,92,95], ACO [84,104,110,113,115], A* [85,87,116], DWA [115,116], etc.
The implementational scenarios mainly involve intelligent storage systems [87,91], automated terminals [114], logistic systems [89,101], and manufacturing workshops [106]. Distance [82,86,91,92,100], path length [84,87,98,116,117], collision avoidance [44,83,87,95,102,108], process costs [83,90,103,113], motion [85,89,94,105,114], and smoothness [97,107,115,116] are the considerations.
The percentage of papers on AI-based approaches achieving online path planning in dynamic environments is higher than that of swarm intelligence algorithms, while the implementation of multi-AGV systems shows no significant difference. We found that 45.68% of the swarm intelligence and AI-based approaches achieve online implementation, and 33.33% are presented for multi-AGV systems. Figure 7 compares the quantitative results across the algorithm families, including their online implementation rate, proportion of multi-AGV scenarios, proportion of dynamic environments, degree of hybridization, and frequency of experimental validation.
The swarm intelligence algorithms, such as PSO, ACO, and GA, obtain the optimal path for a complex search space by considering problems independently [45]. However, one of the limitations of these algorithms is that they tend to be restricted in the static environment or simple dynamic environment and most of them cannot handle environmental uncertainties or changing environmental conditions [5,49,51], such as dynamic moving obstacles [5,61], dynamic scheduling [70,73,74], AGV conflict resolution [60,64], and moving goals [48].
The PSO-based algorithms enhance path planning ability by combining with other algorithms or introducing new factors, but they face the challenges of environmental uncertainties or a dynamic environment. The PSO-based approaches are suited for static or simplified dynamic AGV planning tasks. The ACO-based methods implement a penalty strategy, the pheromone-guided mechanism, or a search strategy, which results in limited robustness. The GA-based approaches enable global and local search, optimize the multiple AGV scenario, and integrate with some AI-based approaches, but slow convergence and static environments remain limitations. Therefore, they are more suitable for offline optimization.
The AI-based models, such as DQN, PPO, Q-learning, and neural networks, require training data to build the model or learn the policy from the environment [45]. The limitations of these models include hard-to-obtain perfectly trained data [82] or modeling the environment [84,106,107,114], and the training/computational time is long [91,113]. Also, robustness or applicability could not be measured for AI-based approaches [88,116], and they cannot always guarantee safety/completeness [93,108].
The NN-based models implement global path planning methods to enhance efficiency, but their performance relies on the quality of training data, and the unpredictable or new environment model would be a challenge. The RL-based models are popular approaches for the current AGV path planning. They learn policies from the environment and can deal with uncertain environments; however, the sim-to-real transfer and training efficiency are the major problems. The FL-based approaches also combine with the classical algorithms for inference or prediction, while their adaptability and scalability are limited.
From the perspective of the algorithm families, based on the Table 1, Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7, this review could be considered from the application scenario. The cited papers can be classified into four different scenarios, including (1) single AGV path planning, (2) multi-AGV path planning, (3) dynamic environment, and (4) static environment.
From the perspective of the application scenario, swarm intelligence algorithms tend to be applied in a structured environment or without unpredictable moving obstacles, while artificial intelligence algorithms are more applied to a dynamic environment. The Multi-AGV scenario is frequently solved using GA and RL among the cited papers. Figure 7 presents the online implementation rates of the algorithm families, the proportion of multi-AGV scenarios, and the proportion of dynamic environments.
For the AGV path planning, sim-to-real transfer is a significant problem that needs to be considered. The computational load of these reactive algorithms is quite large for onboard computation, and the adaptability to dynamic scenarios is limited [94,101]. Moreover, the dynamics of AGVs are not considered [105], and the problems of sensor noise, bias, and real experiments rise. Most papers use simulation for validation, even though the real experiment is limited to a small scenario, which is hard for real industrial implementation. Transfer learning [44] or model transfer methodology [117] could be considered as a future direction. Also, the embodied intelligence with environmental perception would be helpful for real-time interaction.

5.2. Conclusions

From Table 8 and Table 9, AI-based models have become a major trend in current AGV path planning research. Recently, AGV systems have been increasingly deployed in dynamic and uncertain environments, and collaborative AGVs are employed for large-scale tasks. As a result, online capability, scalability, and adaptability are required.
Swarm intelligence algorithms are commonly used approaches in AGV path planning because they demonstrate fast convergence and low computational complexity when generating optimal paths and can effectively optimize objective functions. They are suitable for static or simple dynamic environments; however, their adaptability to environmental changes remains limited.
By contrast, AI-based models, especially reinforcement learning, have recently attracted increasing attention in AGV path planning research. These methods support online path planning, multi-AGV systems, and dynamic obstacle avoidance, as the proportion of online implementations reaches 72.97%, and the percentage of studies considering dynamic environments reaches 54.05%.
From the literature, only 13.58% have real-world experiments for both AI-based and swarm intelligence-based algorithms, as inferred from Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9. Most studies rely on simulated validation, while experiment-driven research or real-world system deployment is lacking. Moreover, 35.80% of the literature aims to enhance adaptability to dynamic environments, while the environment uncertainty is simplified, and robustness or completeness cannot be guaranteed. AI-based approaches face the challenge of safety during training, generalization ability, and the sim-to-real problems.
Motivated by the analysis, some research questions could be considered for future AGV path planning research, as follows:
  • How to reduce the gap between the simulation environment and the real-world AGV operation environment, or how to enhance the realism of the simulation environment when validating the algorithms?
  • How to address environmental uncertainty and unpredictable obstacles when maintaining the online implementation of the algorithms with the safety and completeness constraints of path planning?
  • How to improve the sim-to-real transfer or generalization ability of the AGV path planning algorithm through embodied intelligence, transfer learning, or other approaches?

Author Contributions

Conceptualization, S.L.; methodology, S.L.; software, S.L.; validation, S.L.; formal analysis, S.L.; investigation, S.L.; resources, S.L.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L., X.K., and J.W.; visualization, S.L.; supervision, X.K. and J.W.; project administration, J.W.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Fujian Province, China, under Grant 2024J01723, Grant 2024J01721, and Grant 2024J01115 and in part by the Startup Fund of Jimei University under Grant ZQ2024002 and Grant ZQ2024034.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AGVAutomated guided vehicle
RRTRapidly-exploring random tree
APFArtificial potential field
PRMProbabilistic roadmap
PSOParticle swarm optimization
SASimulated annealing
GWOGrey wolf optimizer
MOPSOMulti-objective particle swarm optimization
DWADynamic window approach
FOAFruit fly optimization algorithm
EDAEstimation of distribution algorithm
LSTMLong short-term memory
DQNDeep-Q network
D3QNDueling double deep-Q network
DRLDeep reinforcement learning
RNNRecurrent neural network
PPOProximal policy optimization
ICMIntrinsic curiosity module
DDPGDeep deterministic policy gradient
MPCModel predictive controller
DEDifferential evolution
FLFuzzy logic

References

  1. De Ryck, M.; Versteyhe, M.; Debrouwere, F. Automated guided vehicle systems, state-of-the-art control algorithms and techniques. J. Manuf. Syst. 2020, 54, 152–173. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Wu, L.; Zhang, W.; Peng, T.; Zheng, J. Energy-efficient path planning for a single-load automated guided vehicle in a manufacturing workshop. Comput. Ind. Eng. 2021, 158, 107397. [Google Scholar] [CrossRef]
  3. Moshayedi, A.J.; Jinsong, L.; Liao, L. AGV (automated guided vehicle) robot: Mission and obstacles in design and performance. J. Simul. Anal. Nov. Technol. Mech. Eng. 2019, 12, 5–18. [Google Scholar]
  4. Madridano, A.; Al-Kaff, A.; Martín, D.; de la Escalera, A. Trajectory planning for multi-robot systems: Methods and applications. Expert Syst. Appl. 2021, 173, 114660. [Google Scholar] [CrossRef]
  5. Lin, S.; Liu, A.; Wang, J.; Kong, X. An intelligence-based hybrid PSO-SA for mobile robot path planning in warehouse. J. Comput. Sci. 2023, 67, 101938. [Google Scholar] [CrossRef]
  6. Lin, S.; Liu, A.; Wang, J.; Kong, X. A Review of Path-Planning Approaches for Multiple Mobile Robots. Machines 2022, 10, 773. [Google Scholar] [CrossRef]
  7. Julius Fusic, S.; Kanagaraj, G.; Hariharan, K.; Karthikeyan, S. Optimal path planning of autonomous navigation in outdoor environment via heuristic technique. Transp. Res. Interdiscip. Perspect. 2021, 12, 100473. [Google Scholar] [CrossRef]
  8. Kim, S.; Jin, H.; Seo, M.; Har, D. Optimal Path Planning of Automated Guided Vehicle using Dijkstra Algorithm under Dynamic Conditions. In Proceedings of the 2019 7th International Conference on Robot Intelligence Technology and Applications (RiTA), Daejeon, Republic of Korea, 1–3 November 2019; pp. 231–236. [Google Scholar] [CrossRef]
  9. Sun, Y.; Fang, M.; Su, Y. AGV path planning based on improved Dijkstra algorithm. J. Phys. Conf. Ser. 2024, 1746, 012052. [Google Scholar] [CrossRef]
  10. Dharmasiri, P.; Kavalchuk, I.; Akbari, M. Novel implementation of multiple automated ground vehicles traffic real time control algorithm for warehouse operations: Djikstra approach. Oper. Supply Chain Manag. Int. J. 2020, 13, 396–405. [Google Scholar] [CrossRef]
  11. Zhong, M.; Yang, Y.; Sun, S.; Zhou, Y.; Postolache, O.; Ge, Y.E. Priority-based speed control strategy for automated guided vehicle path planning in automated container terminals. Trans. Inst. Meas. Control 2020, 42, 3079–3090. [Google Scholar] [CrossRef]
  12. Guo, K.; Zhu, J.; Shen, L. An Improved Acceleration Method Based on Multi-Agent System for AGVs Conflict-Free Path Planning in Automated Terminals. IEEE Access 2021, 9, 3326–3338. [Google Scholar] [CrossRef]
  13. Li, Y. Research on AGV Real-Time Path Planning and Obstacle Detection Based on Machine Vision. In Proceedings of the 2025 IEEE 5th International Conference on Electronic Technology, Communication and Information (ICETCI), Changchun, China, 23–25 May 2025; pp. 1454–1459. [Google Scholar] [CrossRef]
  14. Wang, J.; Huang, H.; Li, J.; Jiang, L.; Li, J.; Jiang, F. AGV path planning algorithm based on improved Dijkstra algorithm. In Proceedings of the 2024 6th International Conference on Internet of Things, Automation and Artificial Intelligence (IoTAAI), Guangzhou, China, 26–28 July 2024; pp. 568–574. [Google Scholar] [CrossRef]
  15. Fransen, K.; van Eekelen, J. Efficient path planning for automated guided vehicles using A*(Astar) algorithm incorporating turning costs in search heuristic. Int. J. Prod. Res. 2023, 61, 707–725. [Google Scholar] [CrossRef]
  16. Zhang, D.; Chen, C.; Zhang, G. AGV path planning based on improved A-star algorithm. In Proceedings of the 2024 IEEE 7th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 15–17 March 2024; Volume 7, pp. 1590–1595. [Google Scholar]
  17. Lian, Y.; Yang, Q.; Xie, W.; Zhang, L. Cyber-Physical System-Based Heuristic Planning and Scheduling Method for Multiple Automatic Guided Vehicles in Logistics Systems. IEEE Trans. Ind. Inform. 2021, 17, 7882–7893. [Google Scholar] [CrossRef]
  18. Wu, B.; Chi, X.; Zhao, C.; Zhang, W.; Lu, Y.; Jiang, D. Dynamic Path Planning for Forklift AGV Based on Smoothing A* and Improved DWA Hybrid Algorithm. Sensors 2022, 22, 7079. [Google Scholar] [CrossRef]
  19. Guo, S.; Pan, X.; Liu, Z. AGV Path Planning Algorithm Based on Fusion of Improved A* and DWA. In Proceedings of the 2024 43rd Chinese Control Conference (CCC), Kunming, China, 28–31 July 2024; pp. 1782–1787. [Google Scholar] [CrossRef]
  20. Lian, Y.; Xie, W.; Zhang, L. A Probabilistic Time-Constrained Based Heuristic Path Planning Algorithm in Warehouse Multi-AGV Systems. IFAC-PapersOnLine 2020, 53, 2538–2543. [Google Scholar] [CrossRef]
  21. Deng, X.; Li, R.; Zhao, L.; Wang, K.; Gui, X. Multi-obstacle path planning and optimization for mobile robot. Expert Syst. Appl. 2021, 183, 115445. [Google Scholar] [CrossRef]
  22. Sebastian, B.; Ben-Tzvi, P. Physics based path planning for autonomous tracked vehicle in challenging terrain. J. Intell. Robot. Syst. 2019, 95, 511–526. [Google Scholar] [CrossRef]
  23. Zheng, W.; Yu, H. Research on the Safety of AGV Path Planning Based on D* Algorithm. In Proceedings of the 2024 6th International Conference on Artificial Intelligence and Computer Applications (ICAICA), Dalian, China, 28–30 November 2024; pp. 216–220. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Luo, J.; Cai, X.; Chen, Y.; Peng, E.; Zou, X. AGV Path Planning for Logistics Warehouse by Using an Improved D*Lite Algorithm. In Proceedings of the TEPEN 2022; Zhang, H., Ji, Y., Liu, T., Sun, X., Ball, A.D., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 1018–1027. [Google Scholar]
  25. Wang, H.; Li, G.; Hou, J.; Chen, L.; Hu, N. A Path Planning Method for Underground Intelligent Vehicles Based on an Improved RRT* Algorithm. Electronics 2022, 11, 294. [Google Scholar] [CrossRef]
  26. Li, Y.; Wei, W.; Gao, Y.; Wang, D.; Fan, Z. PQ-RRT*: An improved path planning algorithm for mobile robots. Expert Syst. Appl. 2020, 152, 113425. [Google Scholar] [CrossRef]
  27. Chi, W.; Ding, Z.; Wang, J.; Chen, G.; Sun, L. A Generalized Voronoi Diagram-Based Efficient Heuristic Path Planning Method for RRTs in Mobile Robots. IEEE Trans. Ind. Electron. 2022, 69, 4926–4937. [Google Scholar] [CrossRef]
  28. Yuan, Z.; Yang, Z.; Lv, L.; Shi, Y. A Bi-Level Path Planning Algorithm for Multi-AGV Routing Problem. Electronics 2020, 9, 1351. [Google Scholar] [CrossRef]
  29. Wu, B.; Zhang, W.; Chi, X.; Jiang, D.; Yi, Y.; Lu, Y. A Novel AGV Path Planning Approach for Narrow Channels Based on the Bi-RRT Algorithm with a Failure Rate Threshold. Sensors 2023, 23, 7547. [Google Scholar] [CrossRef]
  30. Wang, J.; Luo, Y.; Tan, X. Path Planning for Automatic Guided Vehicles (AGVs) Fusing MH-RRT with Improved TEB. Actuators 2021, 10, 314. [Google Scholar] [CrossRef]
  31. Zhou, Y.; Lu, Y.; Lv, L. GVP-RRT: A grid based variable probability Rapidly-exploring Random Tree algorithm for AGV path planning. Complex Intell. Syst. 2024, 10, 8273–8286. [Google Scholar] [CrossRef]
  32. Zhang, R.; Chai, R.; Chai, S.; Xia, Y.; Tsourdos, A. Design and Practical Implementation of a High Efficiency Two-Layer Trajectory Planning Method for AGV. IEEE Trans. Ind. Electron. 2024, 71, 1811–1822. [Google Scholar] [CrossRef]
  33. Liu, N.; Hu, Z.; Wei, M.; Guo, P.; Zhang, S.; Zhang, A. Improved A* algorithm incorporating RRT* thought: A path planning algorithm for AGV in digitalised workshops. Comput. Oper. Res. 2025, 177, 106993. [Google Scholar] [CrossRef]
  34. Szczepanski, R.; Tarczewski, T.; Erwinski, K. Energy Efficient Local Path Planning Algorithm Based on Predictive Artificial Potential Field. IEEE Access 2022, 10, 39729–39742. [Google Scholar] [CrossRef]
  35. Ji, Y.; Ni, L.; Zhao, C.; Lei, C.; Du, Y.; Wang, W. TriPField: A 3D potential field model and its applications to local path planning of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 24, 3541–3554. [Google Scholar] [CrossRef]
  36. Szczepanski, R.; Bereit, A.; Tarczewski, T. Efficient Local Path Planning Algorithm Using Artificial Potential Field Supported by Augmented Reality. Energies 2021, 14, 6642. [Google Scholar] [CrossRef]
  37. Chen, X.; Chen, C.; Wu, H.; Postolache, O.; Wu, Y. An improved artificial potential field method for multi-AGV path planning in ports. Intell. Robot. 2025, 5, 19–33. [Google Scholar] [CrossRef]
  38. Chen, Z.; Xu, B. AGV Path Planning Based on Improved Artificial Potential Field Method. In Proceedings of the 2021 IEEE International Conference on Power Electronics, Computer Applications (ICPECA), Shenyang, China, 22–24 January 2021; pp. 32–37. [Google Scholar] [CrossRef]
  39. Wu, H. Research on AGV Path Planning Algorithm Integrating Adaptive A* and Improved APF Algorithm. In Proceedings of the 2025 8th International Conference on Advanced Algorithms and Control Engineering (ICAACE), Shanghai, China, 21–23 March 2025; pp. 764–769. [Google Scholar] [CrossRef]
  40. Zhou, Y.; Lu, Y.; Lv, L. Grid-Based Non-Uniform Probabilistic Roadmap-Based AGV Path Planning in Narrow Passages and Complex Environments. Electronics 2024, 13, 225. [Google Scholar] [CrossRef]
  41. Žužek, T.; Vrabič, R.; Zdešar, A.; Škulj, G.; Banfi, I.; Bošnak, M.; Zaletelj, V.; Klančar, G. Simulation-Based Approach for Automatic Roadmap Design in Multi-AGV Systems. IEEE Trans. Autom. Sci. Eng. 2024, 21, 6190–6201. [Google Scholar] [CrossRef]
  42. Stenzel, J.; Schmitz, L. Automated Roadmap Graph Creation and MAPF Benchmarking for Large AGV Fleets. In Proceedings of the 2022 8th International Conference on Automation, Robotics and Applications (ICARA), Prague, Czech Republic, 18–20 February 2022; pp. 146–153. [Google Scholar] [CrossRef]
  43. Jathunga, T.; Rajapaksha, S. Improved Path Planning for Multi-Robot Systems Using a Hybrid Probabilistic Roadmap and Genetic Algorithm Approach. J. Robot. Control (JRC) 2025, 6, 715–733. [Google Scholar] [CrossRef]
  44. Zhang, R.; Chai, R.; Chen, K.; Zhang, J.; Chai, S.; Xia, Y.; Tsourdos, A. Efficient and Near-Optimal Global Path Planning for AGVs: A DNN-Based Double Closed-Loop Approach with Guarantee Mechanism. IEEE Trans. Ind. Electron. 2025, 72, 681–692. [Google Scholar] [CrossRef]
  45. Reda, M.; Onsy, A.; Haikal, A.Y.; Ghanbari, A. Path planning algorithms in the autonomous driving system: A comprehensive review. Robot. Auton. Syst. 2024, 174, 104630. [Google Scholar] [CrossRef]
  46. Tao, Q.; Sang, H.; Guo, H.; Wang, P. Improved particle swarm optimization algorithm for AGV path planning. IEEE Access 2021, 9, 33522–33531. [Google Scholar] [CrossRef]
  47. Lin, S.; Liu, A.; Wang, J.; Kong, X. An improved fault-tolerant cultural-PSO with probability for multi-AGV path planning. Expert Syst. Appl. 2024, 237, 121510. [Google Scholar] [CrossRef]
  48. Gul, F.; Rahiman, W.; Alhady, S.S.N.; Ali, A.; Mir, I.; Jalil, A. Meta-heuristic approach for solving multi-objective path planning for autonomous guided robot using PSO–GWO optimization algorithm with evolutionary programming. J. Ambient Intell. Humaniz. Comput. 2021, 12, 7873–7890. [Google Scholar] [CrossRef]
  49. Dao, T.K.; Ngo, T.G.; Pan, J.S.; Nguyen, T.T.T.; Nguyen, T.T. Enhancing Path Planning Capabilities of Automated Guided Vehicles in Dynamic Environments: Multi-Objective PSO and Dynamic-Window Approach. Biomimetics 2024, 9, 35. [Google Scholar] [CrossRef]
  50. Ahmad, J.; Wahab, M.N.A.; Ramli, A.; Misro, M.Y.; Ezza, W.Z.; Hasan, W.Z.W. Enhancing performance of global path planning for mobile robot through Alpha–Beta Guided Particle Swarm Optimization (ABGPSO) algorithm. Measurement 2025, 257, 118633. [Google Scholar] [CrossRef]
  51. Wang, Q.; Tang, S. Improved Human Optimisation Algorithm for Global Path Planning in AGV. In Proceedings of the 2024 5th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 12–14 April 2024; pp. 56–62. [Google Scholar] [CrossRef]
  52. Song, J. Automatic Guided Vehicle Global Path Planning Considering Multi-objective Optimization and Speed Control. Sensors Mater. 2021, 33, 1999. [Google Scholar] [CrossRef]
  53. Xu, L.; Wang, N.; Ling, X. Study on Conflict-free AGVs Path Planning Strategy for Workshop Material Distribution Systems. Procedia CIRP 2021, 104, 1071–1076. [Google Scholar] [CrossRef]
  54. Li, X.; Zhang, B.; Wang, D. Research on AGV path planning based on PSO-IACO algorithm. In Proceedings of the 2021 IEEE International Conference on Electrical Engineering and Mechatronics Technology (ICEEMT), Qingdao, China, 2–4 July 2021; pp. 709–713. [Google Scholar] [CrossRef]
  55. Yi, G.; Feng, Z.; Mei, T.; Li, P.; Jin, W.; Chen, S. Multi-AGVs path planning based on improved ant colony algorithm. J. Supercomput. 2019, 75, 5898–5913. [Google Scholar] [CrossRef]
  56. Wang, X.; Shi, H.; Zhang, C. Path planning for intelligent parking system based on improved ant colony optimization. IEEE Access 2020, 8, 65267–65273. [Google Scholar] [CrossRef]
  57. Yue, L.; Chen, H. Unmanned vehicle path planning using a novel ant colony algorithm. EURASIP J. Wirel. Commun. Netw. 2019, 2019, 136. [Google Scholar] [CrossRef]
  58. Wu, L.; Huang, X.; Cui, J.; Liu, C.; Xiao, W. Modified adaptive ant colony optimization algorithm and its application for solving path planning of mobile robot. Expert Syst. Appl. 2023, 215, 119410. [Google Scholar] [CrossRef]
  59. Li, H.; Chen, F.; Luo, W.; Liu, Y.; Li, J.; Sun, Z. Research on AGV Path Planning Based on Gray wolf Improved Ant Colony Optimization. In Proceedings of the 2022 5th International Conference on Robotics, Control and Automation Engineering (RCAE), Changchun, China, 28–30 October 2022; pp. 221–226. [Google Scholar]
  60. Zhou, Y.; Huang, N. Airport AGV path optimization model based on ant colony algorithm to optimize Dijkstra algorithm in urban systems. Sustain. Comput. Inform. Syst. 2022, 35, 100716. [Google Scholar] [CrossRef]
  61. Ali, H.; Gong, D.; Wang, M.; Dai, X. Path planning of mobile robot with improved ant colony algorithm and MDP to produce smooth trajectory in grid-based environment. Front. Neurorobot. 2020, 14, 44. [Google Scholar] [CrossRef]
  62. Jin, Q.; Tang, C.; Cai, W. Research on Dynamic Path Planning Based on the Fusion Algorithm of Improved Ant Colony Optimization and Rolling Window Method. IEEE Access 2022, 10, 28322–28332. [Google Scholar] [CrossRef]
  63. Wang, W.; Li, J.; Bai, Z.; Wei, Z.; Peng, J. Toward Optimization of AGV Path Planning: An RRT*-ACO Algorithm. IEEE Access 2024, 12, 18387–18399. [Google Scholar] [CrossRef]
  64. Liu, Y.; Hou, Z.; Tan, Y.; Liu, H.; Song, C. Research on Multi-AGVs Path Planning and Coordination Mechanism. IEEE Access 2020, 8, 213345–213356. [Google Scholar] [CrossRef]
  65. Zhang, Z.; Zhang, J.; Zhao, Y.; Zhu, Z.; Guo, J. A novel hexagonal grid map model and regenerated heuristic factor based strategy for intelligent manufacturing system’s AGV path planning problem solving. Comput. Ind. Eng. 2024, 192, 110154. [Google Scholar] [CrossRef]
  66. Chen, Y.; Yu, K. Research on AGV Path Planning Design Based on Reinforcement Learning-Ant Colony Algorithm. In Proceedings of the 2024 IEEE 4th International Conference on Electronic Technology, Communication and Information (ICETCI), Changchun, China, 24–26 May 2024; pp. 1148–1153. [Google Scholar] [CrossRef]
  67. Li, J.; Xu, B.; Yang, Y.; Wu, H. Quantum ant colony optimization algorithm for AGVs path planning based on Bloch coordinates of pheromones. Nat. Comput. 2020, 19, 673–682. [Google Scholar] [CrossRef]
  68. Sun, J.; Yu, Y.; Xin, L. Research on Path Planning of AGV Based on Improved Ant Colony Optimization Algorithm. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021; pp. 7567–7572. [Google Scholar] [CrossRef]
  69. Xiao, J.; Yu, X.; Sun, K.; Zhou, Z.; Zhou, G. Multiobjective path optimization of an indoor AGV based on an improved ACO-DWA. Math. Biosci. Eng. 2022, 19, 12532–12557. [Google Scholar] [CrossRef]
  70. Wu, C.; Xiao, Y.; Zhu, X. Research on Optimization Algorithm of AGV Scheduling for Intelligent Manufacturing Company: Taking the Machining Shop as an Example. Processes 2023, 11, 2606. [Google Scholar] [CrossRef]
  71. Yang, Z.; Liu, H.; Xie, R. Improved Ant Colony Algorithm Based on Parameters Optimization for AGV Path Planning. In Proceedings of the 2021 2nd International Symposium on Computer Engineering and Intelligent Communications (ISCEIC), Nanjing, China, 6–8 August 2021; pp. 87–92. [Google Scholar] [CrossRef]
  72. Zhu, Y.; Song, Q.; Li, M. Multi-AGV multitask collaborative scheduling based on an improved ant colony algorithm. Int. J. Adv. Robot. Syst. 2025, 2025, 1–11. [Google Scholar] [CrossRef]
  73. Lyu, X.; Song, Y.; He, C.; Lei, Q.; Guo, W. Approach to integrated scheduling problems considering optimal number of automated guided vehicles and conflict-free routing in flexible manufacturing systems. IEEE Access 2019, 7, 74909–74924. [Google Scholar] [CrossRef]
  74. Zhong, M.; Yang, Y.; Dessouky, Y.; Postolache, O. Multi-AGV scheduling for conflict-free path planning in automated container terminals. Comput. Ind. Eng. 2020, 142, 106371. [Google Scholar] [CrossRef]
  75. Liang, C.; Zhang, Y.; Dong, L. A Three Stage Optimal Scheduling Algorithm for AGV Route Planning Considering Collision Avoidance under Speed Control Strategy. Mathematics 2023, 11, 138. [Google Scholar] [CrossRef]
  76. Goli, A.; Tirkolaee, E.B.; Aydın, N.S. Fuzzy integrated cell formation and production scheduling considering automated guided vehicles and human factors. IEEE Trans. Fuzzy Syst. 2021, 29, 3686–3695. [Google Scholar] [CrossRef]
  77. Wu, Y.; Wu, S.; Hu, X. Cooperative Path Planning of UAVs & UGVs for a Persistent Surveillance Task in Urban Environments. IEEE Internet Things J. 2021, 8, 4906–4919. [Google Scholar] [CrossRef]
  78. Niu, Q.; Fu, Y.; Dong, X. Omnidirectional AGV Path Planning Based on Improved Genetic Algorithm. World Electr. Veh. J. 2024, 15, 166. [Google Scholar] [CrossRef]
  79. Cai, Y.; Liu, H.; Li, M.; Ren, F. A Method of Dual-AGV-Ganged Path Planning Based on the Genetic Algorithm. Appl. Sci. 2024, 14, 7482. [Google Scholar] [CrossRef]
  80. Jiang, Z.; Zhang, X.; Wang, P. Grid-Map-Based Path Planning and Task Assignment for Multi-Type AGVs in a Distribution Warehouse. Mathematics 2023, 11, 2802. [Google Scholar] [CrossRef]
  81. Farooq, B.; Bao, J.; Raza, H.; Sun, Y.; Ma, Q. Flow-shop path planning for multi-automated guided vehicles in intelligent textile spinning cyber-physical production systems dynamic environment. J. Manuf. Syst. 2021, 59, 98–116. [Google Scholar] [CrossRef]
  82. Sung, I.; Choi, B.; Nielsen, P. On the training of a neural network for online path planning with offline path planning algorithms. Int. J. Inf. Manag. 2021, 57, 102142. [Google Scholar] [CrossRef]
  83. Chai, R.; Liu, D.; Liu, T.; Tsourdos, A.; Xia, Y.; Chai, S. Deep learning-based trajectory planning and control for autonomous ground vehicle parking maneuver. IEEE Trans. Autom. Sci. Eng. 2022, 20, 1633–1647. [Google Scholar] [CrossRef]
  84. Zhang, B.; Liang, X.; Song, W.; Chen, Y. Multi-dimensional AGV Path Planning in 3D Warehouses Using Ant Colony Optimization and Advanced Neural Networks. In Proceedings of the Advanced Intelligent Computing Technology and Applications; Huang, D.S., Zhang, Q., Zhang, C., Chen, W., Eds.; Springer Nature: Singapore, 2025; pp. 180–191. [Google Scholar]
  85. Sun, M.; Lu, L.; Ni, H.; Wang, Y.; Gao, J. Research on dynamic path planning method of moving single target based on visual AGV. SN Appl. Sci. 2022, 4, 86. [Google Scholar] [CrossRef]
  86. Xie, T.; Yao, X.; Jiang, Z.; Meng, J. AGV Path Planning with Dynamic Obstacles Based on Deep Q-Network and Distributed Training. Int. J. Precis. Eng. Manuf.-Green Technol. 2025, 12, 1005–1021. [Google Scholar] [CrossRef]
  87. Yang, Y.; Li, J.; Peng, L. Multi-robot path planning based on a deep reinforcement learning DQN algorithm. CAAI Trans. Intell. Technol. 2020, 5, 177–183. [Google Scholar] [CrossRef]
  88. Xiao, Q.; Pan, T.; Wang, K.; Cui, S. Research on AGV Path Planning Based on Improved DQN Algorithm. Sensors 2025, 25, 4685. [Google Scholar] [CrossRef]
  89. Guo, X.; Ren, Z.; Wu, Z.; Lai, J.; Zeng, D.; Xie, S. A Deep Reinforcement Learning Based Approach for AGVs Path Planning. In Proceedings of the 2020 Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020; pp. 6833–6838. [Google Scholar] [CrossRef]
  90. Zhang, L.; Yang, C.; Yan, Y.; Cai, Z.; Hu, Y. Automated guided vehicle dispatching and routing integration via digital twin with deep reinforcement learning. J. Manuf. Syst. 2024, 72, 492–503. [Google Scholar] [CrossRef]
  91. Lin, Y.; Hue, G.; Wang, L.; Li, Q.; Zhu, J. A Multi-AGV Routing Planning Method Based on Deep Reinforcement Learning and Recurrent Neural Network. IEEE/CAA J. Autom. Sin. 2024, 11, 1720–1722. [Google Scholar] [CrossRef]
  92. Nie, J.; Zhang, G.; Lu, X.; Wang, H.; Sheng, C.; Sun, L. Reinforcement learning method based on sample regularization and adaptive learning rate for AGV path planning. Neurocomputing 2025, 614, 128820. [Google Scholar] [CrossRef]
  93. Yin, H.; Lin, Y.; Yan, J.; Meng, Q.; Festl, K.; Schichler, L.; Watzenig, D. AGV Path Planning Using Curiosity-Driven Deep Reinforcement Learning. In Proceedings of the 2023 IEEE 19th International Conference on Automation Science and Engineering (CASE), Auckland, New Zealand, 26–30 August 2023; pp. 1–6. [Google Scholar] [CrossRef]
  94. Shi, H.; Yu, Z.; Huang, J.; Ao, T.; Li, W.; Zhou, Y. Graph-reinforcement-learning-based distributed path planning for collaborative multi-AGV systems. Knowl.-Based Syst. 2025, 328, 114255. [Google Scholar] [CrossRef]
  95. Zhang, Q.; Ma, W.; Zheng, Q.; Zhai, X.; Zhang, W.; Zhang, T.; Wang, S. Path Planning of Mobile Robot in Dynamic Obstacle Avoidance Environment Based on Deep Reinforcement Learning. IEEE Access 2024, 12, 189136–189152. [Google Scholar] [CrossRef]
  96. Yu, L.; Wang, Y.; Sheng, Z.; Xu, P.; He, Z.; Du, H. Path Planning for Multi-AGV Systems Based on Globally Guided Reinforcement Learning Approach. In Proceedings of the 2024 IEEE International Conference on Unmanned Systems (ICUS), Nanjing, China, 18–20 October 2024; pp. 819–825. [Google Scholar] [CrossRef]
  97. Gao, P.; Liu, Z.; Wu, Z.; Wang, D. A Global Path Planning Algorithm for Robots Using Reinforcement Learning. In Proceedings of the 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), Dali, China, 6–8 December 2019; pp. 1693–1698. [Google Scholar] [CrossRef]
  98. Bai, Y.; Ding, X.; Hu, D.; Jiang, Y. Research on Dynamic Path Planning of Multi-AGVs Based on Reinforcement Learning. Appl. Sci. 2022, 12, 8166. [Google Scholar] [CrossRef]
  99. Guo, H.; Tan, M.K.; Lim, K.G.; Chuo, H.S.E.; Yang, B.; Teo, K.T.K. Improved Q-Learning Algorithm for Path Planning of an Automated Guided Vehicle (AGV). In Proceedings of the 2023 IEEE International Conference on Artificial Intelligence in Engineering and Technology (IICAIET), Kota Kinabalu, Malaysia, 12–14 September 2023; pp. 376–381. [Google Scholar] [CrossRef]
  100. Gao, Y.; Chen, C.H.; Chang, D. A Machine Learning-Based Approach for Multi-AGV Dispatching at Automated Container Terminals. J. Mar. Sci. Eng. 2023, 11, 1407. [Google Scholar] [CrossRef]
  101. Lee, D.; Kang, Y.S.; Do Noh, S.; Kim, J.; Kim, H. Digital Twin-Driven Reinforcement Learning for Dynamic Path Planning of AGV Systems. In Proceedings of the Advances in Production Management Systems. Production Management Systems for Volatile, Uncertain, Complex, and Ambiguous Environments; Thürer, M., Riedel, R., von Cieminski, G., Romero, D., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 351–365. [Google Scholar]
  102. Hu, H.; Jia, X.; Liu, K.; Sun, B. Self-Adaptive Traffic Control Model with Behavior Trees and Reinforcement Learning for AGV in Industry 4.0. IEEE Trans. Ind. Inform. 2021, 17, 7968–7979. [Google Scholar] [CrossRef]
  103. Huang, Y.; Wang, C. Improved Q-Learning Algorithm for AGV Path Optimization. In Proceedings of the Advanced Manufacturing and Automation XIII; Wang, Y., Yu, T., Wang, K., Eds.; Springer Nature: Singapore, 2024; pp. 55–60. [Google Scholar]
  104. Mao, Y. Path planning and decision algorithm design of intelligent AGV in electric vehicle overhead channel shared charging system. In Proceedings of the 2024 IEEE 4th International Conference on Electronic Technology, Communication and Information (ICETCI), Changchun, China, 24–26 May 2024; pp. 986–991. [Google Scholar] [CrossRef]
  105. Tian, S.; Yang, S. Research on multi-AGV path planning based on map training and action replanning. Intell. Serv. Robot. 2025, 18, 821–838. [Google Scholar] [CrossRef]
  106. Zhang, J.; Zhang, X.; Han, M.; Ren, N.; Hou, Y. AGV path planning method and intelligent obstacle avoidance strategy for intelligent manufacturing workshops. J. Comput. 2024, 35, 137–151. [Google Scholar] [CrossRef]
  107. Chen, X.; Liu, S.; Zhao, J.; Wu, H.; Xian, J.; Montewka, J. Autonomous port management based AGV path planning and optimization via an ensemble reinforcement learning framework. Ocean Coast. Manag. 2024, 251, 107087. [Google Scholar] [CrossRef]
  108. Ye, X.; Deng, Z.; Shi, Y.; Shen, W. Toward Energy-Efficient Routing of Multiple AGVs with Multi-Agent Reinforcement Learning. Sensors 2023, 23, 5615. [Google Scholar] [CrossRef]
  109. Guo, H.; Ren, Z.; Lai, J.; Wu, Z.; Xie, S. Optimal navigation for AGVs: A soft actor–critic-based reinforcement learning approach with composite auxiliary rewards. Eng. Appl. Artif. Intell. 2023, 124, 106613. [Google Scholar] [CrossRef]
  110. Wang, H.; Zhang, A.; Dong, S. IACO-DQ Path Planning for AGV in Complex and Dynamic Environments. In Proceedings of the 2024 6th International Conference on Robotics and Computer Vision (ICRCV), Wuxi, China, 20–22 September 2024; pp. 232–236. [Google Scholar] [CrossRef]
  111. Guo, D.; Ji, S.; Yao, Y.; Chen, C. A decentralized path planning model based on deep reinforcement learning. Comput. Electr. Eng. 2024, 117, 109276. [Google Scholar] [CrossRef]
  112. Liu, Y.; Yan, S.; Zhao, Y.; Song, C.; Li, F. Improved Dyna-Q: A Reinforcement Learning Method Focused via Heuristic Graph for AGV Path Planning in Dynamic Environments. Drones 2022, 6, 365. [Google Scholar] [CrossRef]
  113. Song, Q.; Zhao, Q.; Wang, S.; Liu, Q.; Chen, X. Dynamic Path Planning for Unmanned Vehicles Based on Fuzzy Logic and Improved Ant Colony Optimization. IEEE Access 2020, 8, 62107–62115. [Google Scholar] [CrossRef]
  114. Feng, J.; Yang, Y.; Zhang, H.; Sun, S.; Xu, B. Path Planning and Trajectory Tracking for Autonomous Obstacle Avoidance in Automated Guided Vehicles at Automated Terminals. Axioms 2024, 13, 27. [Google Scholar] [CrossRef]
  115. Zhou, Z.; Geng, C.; Qi, B.; Meng, A.; Xiao, J. Research and experiment on global path planning for indoor AGV via improved ACO and fuzzy DWA. Electron. Res. Arch. 2023, 31, 19152–19173. [Google Scholar] [CrossRef] [PubMed]
  116. Ambuj; Machavaram, R. Intelligent path planning for autonomous ground vehicles in dynamic environments utilizing adaptive Neuro-Fuzzy control. Eng. Appl. Artif. Intell. 2025, 144, 110119. [Google Scholar] [CrossRef]
  117. Wang, K.; Liang, W.; Shi, H.; Zhang, J.; Wang, Q. Optimal time reuse strategy-based dynamic multi-AGV path planning method. Complex Intell. Syst. 2024, 10, 7089–7108. [Google Scholar] [CrossRef]
  118. Chi, H.; Sang, H.Y.; Zhang, B.; Duan, P.; Zou, W.Q. BDE-Jaya: A binary discrete enhanced Jaya algorithm for multiple automated guided vehicle scheduling problem in matrix manufacturing workshop. Swarm Evol. Comput. 2024, 89, 101651. [Google Scholar] [CrossRef]
  119. Li, H.; Liu, W.; Yang, C.; Wang, W.; Qie, T.; Xiang, C. An Optimization-Based Path Planning Approach for Autonomous Vehicles Using the DynEFWA-Artificial Potential Field. IEEE Trans. Intell. Veh. 2022, 7, 263–272. [Google Scholar] [CrossRef]
  120. Zhang, Z.; He, R.; Yang, K. A bioinspired path planning approach for mobile robots based on improved sparrow search algorithm. Adv. Manuf. 2022, 10, 114–130. [Google Scholar] [CrossRef]
  121. Guo, Z.; Xia, Y.; Li, J.; Liu, J.; Xu, K. Hybrid Optimization Path Planning Method for AGV Based on KGWO. Sensors 2024, 24, 5898. [Google Scholar] [CrossRef]
  122. Liu, Q.; Yang, N.; He, X.J. Integrated Optimization of Process Planning and Scheduling Considering Agv Path Planning. SSRN 2025, 5281476. [Google Scholar]
  123. Zhou, X.; Yu, X.; Zhang, Y.; Luo, Y.; Peng, X. Trajectory Planning and Tracking Strategy Applied to an Unmanned Ground Vehicle in the Presence of Obstacles. IEEE Trans. Autom. Sci. Eng. 2021, 18, 1575–1589. [Google Scholar] [CrossRef]
Figure 1. Classifications of path planning algorithms.
Figure 1. Classifications of path planning algorithms.
Biomimetics 11 00017 g001
Figure 2. Classification of reactive algorithms.
Figure 2. Classification of reactive algorithms.
Biomimetics 11 00017 g002
Figure 3. Yearly distribution of the papers.
Figure 3. Yearly distribution of the papers.
Biomimetics 11 00017 g003
Figure 4. Yearly distribution of the papers by publisher.
Figure 4. Yearly distribution of the papers by publisher.
Biomimetics 11 00017 g004
Figure 5. Distribution of papers by publisher across all algorithm families.
Figure 5. Distribution of papers by publisher across all algorithm families.
Biomimetics 11 00017 g005
Figure 6. Yearly distribution of the papers across all algorithm families.
Figure 6. Yearly distribution of the papers across all algorithm families.
Biomimetics 11 00017 g006
Figure 7. The quantitative results across all algorithm families.
Figure 7. The quantitative results across all algorithm families.
Biomimetics 11 00017 g007
Table 1. Comparison of the PSO-based algorithms.
Table 1. Comparison of the PSO-based algorithms.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[2]MOPSOMulti-objective optimization: energy consumption, total execution timeGraphNoDynamic conditionsSingle robot, manufacturing workshopNoSimulation
[49]MOPSO, DWAEnergy consumption, collisions, travel time, smoothnessGridYesDynamicSingle robotYesSimulation
[46]PSOShortest transportation time-NoStaticSingle robot, one-line production lineNoSimulation
[5]PSO, SAPath length and smoothness, collision avoidanceBinary mapYesStaticSingle robot, warehouseYesSimulation, Experiment
[48]PSO, GWOPath length and smoothness2D mapNoStaticSingle robotYesSimulation
[50]PSOSafety, time, and distance2D mapNoStaticSingle robotNoSimulation
[52]PSOSmoothness, path length2D mapYesDynamicSingle robotYesSimulation
[53]PSO, ACOConflict avoidance, total driving timeNodeNoStaticMulti-robots, Workshop Material Distribution SystemYesSimulation
[47]PSO, GALength, collisionGrid spaceYesDynamicMulti-robotsYesSimulation
[54]PSO, ACOLength, collisionGrid spaceNoStaticSingle robotYesSimulation
[51]PSO, Human optimization algorithmConvergence, lengthRaster mapNoStaticSingle robotYesSimulation
Table 2. Comparison of the ACO-based algorithms.
Table 2. Comparison of the ACO-based algorithms.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[55]ACOTotal completion time, transportation time, time for processing the jobGrid spaceNoStaticMulti-robots, production workshopNoSimulation
[56]ACOPath lengthTopological mapNoStaticSingle robot, AGV-based intelligent parking systemNoSimulation
[57]ACOPath lengthGrid spaceNoDynamicSingle robotNoSimulation
[58]ACOPath length, turn timesGrid spaceNoStaticSingle robotNoSimulation
[60]ACO, DijkstraPath lengthGrid spaceNoStaticSingle robot, airportYesSimulation
[61]ACO, A* Multi-DirectionalDistance, turning times and angleGrid spaceNoStaticSingle robotYesSimulation
[62]ACO, rolling windowPath length, energy consumptionGrid spaceYesStaticSingle robot, complex dynamic environmentYesSimulation
[63]ACO, RRT*Path length, iterations, runtimeGrid spaceNoStaticSingle robotYesSimulation
[64]ACODistanceGrid spaceNoStaticMulti-robotsNoSimulation
[65]ACOPath lengthGridNoStaticSingle robotYesSimulation
[66]ACOIterations, obstacle avoidance, path smoothnessGridNoStaticSingle robotYesSimulation
[67]ACODistance, obstacleGridNoStaticSingle robot, automated container terminalYesSimulation
[68]ACOPath length, turning anglesMatrix yard storage mode, gridNoStaticSingle robot, automatic container terminalYesSimulation
[59]ACO, GWOPath smoothness, convergenceGridNoStaticSingle robotYesSimulation
[69]ACO, DWATurns, path lengthGridNoStaticSingle robot, indoor environmentYesSimulation, Experiment
[70]ACOMaterial flow and path lengthRaster mapNoStaticSingle robot, job shopNoSimulation
[71]ACO, GADistance, iterationsGrid mapNoStaticSingle robotNoSimulation
[72]ACODistance factors, task execution time, waiting timeGrid mapNoStaticMulti-robots, factory environmentNoSimulation
Table 3. Comparison of the GA-based algorithms.
Table 3. Comparison of the GA-based algorithms.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[73]GA, Dijkstra, time windowMinimize the make span, the number of AGVsGrid spaceNoStaticMulti-robots, flexible manufacturing systemYesSimulation
[74]GA, PSO, fuzzy logic controllerDelayed completion time, deadlocksGrid spaceNoStaticMulti-robots, automated container terminalsYesSimulation
[76]GA, heuristicIntercellular transportation and makespan-related costsGrid spaceNoStaticMulti-robots, cellular manufacturing systemYesSimulation
[77]GA, EDAFlight heights, blocking of buildingsGrid spaceYesDynamicMulti-robots, cooperative, surveillance, urban environmentYesSimulation
[78]GA, SAPath smoothnessGridNoStaticSingle robotYesSimulation, Experiment
[79]GASmooth and safe movementGridNoStaticMulti-robots, CooperativeNoSimulation, Experiment
[80]GA, A*Task completion time, energy consumptionRaster map, GridYesStaticMulti-robotsYesSimulation
[75]GACompletion timeRoad network modelNoStaticMulti-robotsYesSimulation
[81]GAPath length2D mapYesDynamicMulti-robotsNoSimulation
Table 4. Comparison of the NN-based models.
Table 4. Comparison of the NN-based models.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[82]Neural network, the Bellman–Ford algorithm, a quadratic programThe sum of the distanceGrid-based graphYesStaticSingle robotYesSimulation
[83]RDNN, LSTMCollision, time, process and terminal costs-YesStaticSingle robot, parkingYesSimulation, Experiment
[84]Neural network, ACOPath lengthGridYesStaticSingle robotYesSimulation
[85]NAR neural network, A*Velocity, motion path2D mapYesDynamicMoving single targetYesSimulation
[44]Deep neural network (DNN)Path length, target, obstaclesGridYesDynamicSingle robotYesSimulation
Table 5. Comparison of the RL-based models.
Table 5. Comparison of the RL-based models.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[91]DRL, RNN, PPO, LSTMPosition, obstacles, distance, spacingGridYesDynamicMultiple robots, automated storage and retrieval system (AS/RS)YesSimulation
[97]Q-LearningPath length and smoothnessGraphYesStaticSingle robotNoSimulation, Experiment
[102]Q-learningCollision, terminal stateGridNoStaticMultiple robotsYesSimulation
[100]Q-learningDistanceGridNoStaticMultiple robotsYesSimulation
[105]Q-learningTurning rewards, dynamic priority, action replanningGridNoStaticMultiple robotsYesSimulation
[98]Q-learningConvergence, path lengthGridYesDynamicMultiple robotsYesSimulation, Experiment
[99]Q-learningTarget, obstaclesGridNoStaticSingle robotNoSimulation
[101]Q-learningLocations, destinationsGridYesDynamicMultiple robots, production logistics systemYesSimulation, Experiment
[104]Q-learning, ACO, GADistance, congestion time, charging priorityGridNoStaticSingle robot, shared charging systemYesSimulation
[103]Q-learning, beetle antennae search (BAS)Path length, average timeGridNoStaticSingle robotYesSimulation
[106]Deep Q-learningObstacles, targetGridNoStaticSingle robot, intelligent manufacturing workshopsYesSimulation
[95]PPO, LSTMDistance, heading angle, collision, target pointGridYesDynamicSingle robotYesSimulation, Experiment
[93]PPOStatic and dynamic obstaclesGridYesDynamicSingle robotYesSimulation
[92]PPO, LSTMDistance, collisionGridYesDynamicSingle robotYesSimulation
[94]MAPPO, GNNPosition, velocity, obstacleGridYesDynamicMultiple robotsYesSimulation
[96]MAPPOMovement, obstacles, global path, target, boundaryGridYesDynamicSingle robotYesSimulation
[107]DDPG, APFSmoothness and safetyGraphNoStaticSingle robotYesSimulation
[111]DRLCollision, movement, finish taskGridYesStaticMultiple robotsYesSimulation
[112]Dyna-QGoalGridYesStaticSingle robotYesSimulation
[110]Dyna-Q, ACOObstacle, targetGridYesDynamicSingle robotYesSimulation
[109]SACObstacle, distance, target and timeGridYesDynamicSingle robotYesSimulation
[108]MADDPGPosition, collision, speedGridNoStaticMultiple robotsYesSimulation
[90]D3QN, A*Average tardiness and energy consumptionGridYesDynamicMultiple robotsYesSimulation
[88]DQNDirection, steps, end pointGridYesDynamicSingle robotYesSimulation
[89]Dueling DQNPosition, velocity, targetGridYesDynamicSingle robot, intelligent logistics systemsYesSimulation
Table 6. Comparison of the FL-based models.
Table 6. Comparison of the FL-based models.
PaperAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[113]Fuzzy logic, ACOPollutant emissions, fuel cost, travel time, and distanceGridYesDynamicSingle robotYesSimulation
[114]Fuzzy logic, APFObstacles, velocities, lane lines2D mapYesDynamicSingle robot, automated terminals, portYesSimulation
[115]Fuzzy control, ACO, DWASafety, smoothness, distance, directionGridYesDynamicSingle robotYesSimulation
[116]Fuzzy control, A*, DWAPath length, path search, smoothnessGridYesStaticSingle robotYesSimulation
Table 7. Comparison of other algorithms.
Table 7. Comparison of other algorithms.
PaperClassificationAlgorithmConsiderationModelOnlinePropertiesScenarioHybridExperiment
[118]Swarm intelligenceJayaMinimize transportation cost, total tardiness, early service penaltyGrid spaceNoStaticMulti-robots, matrix manufacturing workshopNoSimulation
[123]Swarm intelligenceArtificial fish swarm algorithmSafety, fuel economy, trajectory smoothnessGrid spaceYesDynamicSingle robotYesSimulation, Experiment
[119]Swarm intelligenceFireworks algorithm, APFSafety, path smoothness-YesDynamicSingle robot, drivingYesSimulation, Experiment
[120]Swarm intelligenceSparrow search algorithmRisk degree, path acquisition time, distance value, total rotation angle valueGrid spaceNoStaticSingle robotNoSimulation
[121]Swarm intelligenceGWO, Kalman filterPath smoothness and length, obstacle avoidanceGrid spaceNoStaticSingle robotYesSimulation
[122]Swarm intelligenceDEMake span, collisionWorkshop diagramNoStaticMulti-robotsYesSimulation
[117]Artificial intelligenceSVMPath lengthGridNoStaticMultiple robotsYesSimulation
Table 8. Comparison of swarm intelligence algorithms.
Table 8. Comparison of swarm intelligence algorithms.
PaperAlgorithmContributionLimitation/Future Research
[2]MOPSOFormulate energy-efficient AGV path planning model, two solution methodsEnergy consumption data acquisition, integration of transport task execution, multi-AGV system
[49]MOPSO, DWACombines MOPSO and DWA for optimization challenges and dynamicsEnvironmental uncertainties, changing environmental conditions, real-world experiments
[46]PSOCrossover operation, mutation mechanism, local optimum problemMulti-AGV system
[5]PSO, SAGet rid of local optima, accept new solution, and update local-oriented best value with a probabilityDynamic environment, multiple robots, moving obstacles
[48]PSO, GWOLocal search techniqueNot multi-objective optimization, real-time implementation, multi-robots, moving goal
[50]PSOAlpha and beta as two coefficientsPath prediction and learning capabilities, only static simple environment
[52]PSOLevy flight, inductive steering algorithmDynamic situation is simple
[53]PSO, ACOA collision avoidance factor, avoid road-section and node conflictsOnly static environment
[47]PSO, CAThe cultural-PSO algorithm, dynamic adjust inertial weightReal-world experiment
[54]PSO, ACOPSO-IACO, PSO optimizes initial parameters of ACOOnly static environment, lack of real-world experiment
[51]PSO, Human optimization algorithmPSO combines HLOMulti robots, dynamic environment
[55]ACOHeuristic information, compare the similarity of the job, path planning and schedulingLimited robustness, other manufacturing environments (flexible job-shop or flow shop)
[56]ACOFallback strategy, valuation function, reward/penalty mechanismThe efficiency of the algorithm
[57]ACOPenalty strategyMultiple robots, experiment
[58]ACOInitial pheromone concentration, improved state transition probability ruleThree-dimensional problem, multi-objective optimization, execution time
[60]ACO, DijkstraACO-DAMulti-AGV conflicts
[61]ACO, A* Multi-Directional algorithmReward policyDynamic moving obstacles
[62]ACO, rolling windowThe pheromone concentrationOptimization, convergence performance, the scope of application
[63]ACO, RRT*Fast-scaling RRT*-ACOOnly static environment
[64]ACOStep length, adaptive pheromone volatilization coefficientMulti-AGVs’ conflict resolution
[65]ACOHexagonal grid map model, the bidirectional search strategyGlobal search optimization, grid map’s robustness, real-world application, efficiency
[66]ACORL configures ACO parametersLack comparison analysis
[67]ACOBloch coordinates of pheromones; a repulsion factorUncertain environments, task assignment, real automated logistics systems
[68]ACOCombines FOA and ACOLack comparison analysis
[59]ACO, GWOA modified ACO based on GWO, heuristic information, the pheromone model, and transfer rulesOnly static environment, lack comparison analysis
[69]ACO, DWACombine ACO and DWAFocus on global path planning, and the static environment is not complex
[70]ACOAdditional heuristic information, dynamic adjustment factor, Laplace distributionDynamic simulation and scheduling
[71]ACO, GANon-uniform and directed distribution of initial pheromone, adaptive adjustment, parameter optimization by GALack comparison
[72]ACOPrior time, the pheromone incrementLarge-scale and changing tasks
[73]GA, Dijkstra, time windowGlobal, local and random search strategies, optimize the number of AGVsDynamic scheduling and job sequencing problem
[74]GA, PSO, fuzzy logic controllerIntegrated scheduling and path planning, adaptive auto tuningComputation time, dynamic real-time scheduling
[76]GA, heuristicApplying the fuzzy linear programming, hybrid approachComplicate AGVs’ constraints, not real case
[77]GA, EDACooperative path planning model, online adjustment strategyMore possible applications
[78]GA, SAPath smoothness constraints, crossover stage, mutation operationLacks comparison with state-of-the-art techniques
[79]GAFitness functionOnly consider static obstacles
[80]GA, A*A* combines cyclic rules, GA with penalty functionOnly static obstacle, AGV charging problem in the future
[75]GAA three-stage optimal scheduling algorithmLacks comparison analysis, AGV charging, collision avoidance route
[81]GAImproved GA, two decision variablesLacks comparison analysis
[118]JayaThe key-task shift method, initialization methods, offspring generation methods, insertion-based repair methodConsiders more practical constraints and production environments, the use of multi-objective optimization problem and new techniques
[123]Artificial fish swarm algorithmTrail-based forward search algorithm, command signalsLacks comparison with state-of-the-art techniques
[119]Fireworks algorithm, APFDynEFWA-APFIncorporates personalized driving style
[120]Sparrow search algorithmLocation update formula, neighborhood search strategy, linear path strategyExperiment, multi-robots, dynamic obstacles
[121]GWO, Kalman filterRefine with KF correctionsOnly static environment
[122]DEHybrid variable neighborhood DEAGVs’ speed, multi-objective optimization
Table 9. Comparison of artificial intelligence algorithms.
Table 9. Comparison of artificial intelligence algorithms.
PaperAlgorithmContributionLimitation/Future Research
[82]Neural network, the Bellman–Ford algorithm, a quadratic programOffline training, and online path planningHard to acquire perfect situational awareness, trained data, dimensionality
[83]RDNN, LSTMRNDD-based motion planning, transfer learning strategiesMulti-robot environment
[84]Neural network, ACOCombines ACO with neural networksThe environmental model is not clear
[85]NAR neural network, A*Reduced and non-reduced pointThe success rate is fair
[44]Deep neural network (DNN),Target area adaptive RRT*, optimal path backward generation, DNNConsider kinematic information, 3D scenarios, and transfer learning  in future studies
[113]Fuzzy logic, ACOFLACO, local optimum trap, global optimal pathReducing the computing time, multiple vehicles
[114]Fuzzy logic, APFHybrid APF-fuzzy model prediction controllerAGV modeling
[115]Fuzzy control, ACO, DWAImproved ACO and DWA with fuzzy controllers Only static obstacles
[116]Fuzzy control, A*, DWAAdapative neuro-fuzzy inference system, enhanced A* with DWARobustness, applicability, real-world environments
[87]RL DQN, A*Slow convergence and excessive randomnessLocal path planning
[86]DQNState-dynamic network modelMulti-AGV environment
[91]DRL, RNN, PPO, LSTMTemporary changesReduce the computational time, dynamic conflict avoidance strategies
[97]Q-LearningGlobal Q-learning path planningLack the modification of Q-learning
[102]Q-learningBehavior treesNot considering completed situations, AGV scheduling, or real-world system
[100]Q-learningContract net protocolThe comparison analysis is weak; it only uses traditional Q-learning
[105]Q-learningMap training and action replanningDynamics of AGVs are not considered; the environment is simple
[98]Q-learningKohonen Q-learningTask scheduling and assignment
[99]Q-learningA deep learning factorStatic obstacle environment
[101]Q-learningDigital Twin-driven Q-learningMore complex situations, task allocation
[104]Q-learning, ACO, GAQ-learning and ACO, positive ant colony feedback mechanismOnly compared with Dijkstra and A* algorithm, static environment
[103]Q-learning, beetle antennae search (BAS)BAS-QLStatic obstacles
[106]Deep Q-learningExperience replay pool, network structure, neighborhood weighted grid modelingDynamic environments should be studied
[95]PPO, LSTMIntroduce ICM and LSTM into PPOThe success rate decreases when dynamic obstacles moving fast or not follow regular patterns
[93]PPOAdditional intrinsic rewardsCannot guarantee safety in the training
[92]PPO, LSTMSample regularization, adaptive learning rateLack environmental experiments
[94]MAPPO, GNNGNN with MADRLComplex interactions and dynamic environment
[96]MAPPOA* for global guidance, MAPPO for local planningMulti-robot scenario
[107]DDPG, APFAPF, twin delayed DDPGLack environmental perception and testing, real experiment, and hard to implement in complex environment
[111]DRLLocal observationsHigh density of obstacles
[112]Dyna-QHeuristic planningLacks comparison with SOTA methods
[110]Dyna-Q, ACOImproved heuristic function of ACO, combines with Dyna-QLacks comparison analysis
[109]SACSum-tree replayLacks experiments
[108]MADDPGϵ-GreedyOptimal value has not been established
[90]D3QN, A*Digital twin, prevent deadlock and congestionMulti-resource production scheduling problems
[88]DQNA refined multi-objective reward function, the priority experience replay mechanismRobust training methods, dynamic obstacle prediction modules, experimental design
[89]Dueling DQNMultimodal sensing information, prioritized experience replyMARL
[117]SVMSVM-based model, replanning periodModel transfer methodology
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, S.; Wang, J.; Kong, X. Bio-Inspired Reactive Approaches for Automated Guided Vehicle Path Planning: A Review. Biomimetics 2026, 11, 17. https://doi.org/10.3390/biomimetics11010017

AMA Style

Lin S, Wang J, Kong X. Bio-Inspired Reactive Approaches for Automated Guided Vehicle Path Planning: A Review. Biomimetics. 2026; 11(1):17. https://doi.org/10.3390/biomimetics11010017

Chicago/Turabian Style

Lin, Shiwei, Jianguo Wang, and Xiaoying Kong. 2026. "Bio-Inspired Reactive Approaches for Automated Guided Vehicle Path Planning: A Review" Biomimetics 11, no. 1: 17. https://doi.org/10.3390/biomimetics11010017

APA Style

Lin, S., Wang, J., & Kong, X. (2026). Bio-Inspired Reactive Approaches for Automated Guided Vehicle Path Planning: A Review. Biomimetics, 11(1), 17. https://doi.org/10.3390/biomimetics11010017

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop