This section analyzes the experimental methodologies and results from the 71 records included in this review, focusing on key factors such as the number of robots, evaluation metrics, and tested scenarios. By examining these aspects, we aim to identify trends, strengths, and gaps in the current research on task scheduling for mobile robots. This analysis provides a deeper understanding of how different scheduling strategies are tested and evaluated and highlights areas where future studies can further contribute to the field.
5.8.2. Small-Scale Systems Experiments
Small-scale systems (2–5 robots) account for all real-world experimental studies in the reviewed literature. These setups are primarily used to validate proof-of-concept algorithms in controlled environments, typically involving simple layouts, limited task diversity, and a small number of agents. While these conditions facilitate focused evaluation and repeatability, they often fall short of capturing the complexities present in larger, real-world deployments. Simulation environments are frequently employed to extend the scalability of these algorithms beyond what is feasible with physical systems.
Table 13 organizes the experiments conducted in small-scale systems according to whether they were performed in real or simulated environments and according to the simulator used in each case.
Real-world experiments
Luo et al. [
33] tested their ChEA algorithm in two scenarios. The first used a 6 × 6 instance of the Mousavi example [
53] (six jobs and six machines), with 10 runs of 20 iterations each and a population size of 30. The second experiment used data from Bilge [
91], which included four machines and different AGV and job-task combinations based on geographical layouts. The results were compared against evolutionary algorithms like MAS, GAHA, and FMAS. Zhang et al. [
39] evaluated their DRL-based algorithm through both simulations and a real-world case study. The simulations involved 500 dynamic tasks, with release times following a Poisson distribution, and the real-world case was conducted in a production logistics system where AGVs operated on planned paths with speed adjustments and collision avoidance. Their method was compared to traditional scheduling approaches like MIWT-, MIEC-, and DQN-based methods. Gu et al. [
45] built an experimental platform with physical entities such as AS/RS, AGVs, and robots connected via Wi-Fi. The system’s performance was tested using a bio-inspired swarm optimization algorithm (BSOA), comparing it to static (HA) and dynamic (MAS) scheduling approaches in a setting with two AGVs and four machines. Mudrova and Hawes [
56] validated their scheduler in simulated and real-world environments using mobile robots performing assistance tasks. The scheduler, implemented in C++ and distributed as a ROS package, showed improved performance compared to existing scheduling algorithms. Yi and Luo [
67] tested their approach in a real robotic job shop (RJS) system with multiple machines, robotic arms, and AGVs. Their method, compared to a genetic algorithm (GA), demonstrated scalability and effectiveness for varying job loads.
These examples illustrate that while real-world testing provides valuable insights, it remains largely restricted to small-scale systems due to complexity, cost, and hardware limitations. No studies involving medium- or large-scale real-world systems were found in the reviewed records.
MATLAB simulation experiments
Samsuria et al. [
27] used MATLAB to test their scheduling algorithms on benchmark datasets, including the EX44 instance from Bilge and Ulusoy [
91] (four machines, with 19 operations across five jobs) and the ‘la01’ dataset from Lawrence [
92] (10 jobs × 5 operations). Their approach was benchmarked against a genetic algorithm (GA) and a hybrid GA-Simulated Annealing method (HGASA), as well as other Bilge instances. Zhang et al. [
31] used MATLAB to test their integrated AGV scheduling model on datasets from Brandimarte [
93], including MK11–MK15 (middle-scale: 20 jobs, with 100–200 operations) and MK21–MK25 (large-scale: 50 jobs, with 250–500 operations). Processing times followed a uniform distribution. They also measured AGV energy consumption using a battery safety factor of 0.05 and compared the results with a standard GA. Chen and Liu [
19] evaluated their decentralized method in MATLAB using Monte Carlo simulations. The scenarios involved 18 to 30 tasks with robots’ initial positions randomized within a 100 m × 50 m area. Robot speed was set at 2 m/s, and static obstacles were not considered. They compared their algorithm to CBBA and a centralized RHTA (C-RHTA). Xiao et al. [
42] tested their improved Dynamic Particle Swarm Optimization (DPSO) algorithm using 10 examples from the PSPLIB library (J10, J12, J16, J18). They evaluated adaptability by simulating priority changes and increasing task load (including new workstations and AGVs). DPSO was reapplied in each scenario to re-optimize scheduling, and the results were benchmarked against PSO. Qu et al. [
44] built a Jobset20-based case study including four machines, four tools, and 10 parts with 24 operations. Mobile robots were responsible for inter-machine transport. A hybrid GA (population: 80; crossover: 0.9; mutation: 0.2) generated schedules within 24.23 s. The results included both Gantt charts and spatiotemporal trajectories of mobile robots. Wang et al. [
51] used MATLAB to simulate three 2D workshop layouts: chaotic, linear, and obstacle-interrupted. The layouts involved 11–15 workstations, with four–five AGVs and scenarios with both single and multiple depots. Their method, incorporating a map complexity coefficient, outperformed the base model in layout-sensitive scheduling performance. Saidi-Mehrabad et al. [
52] combined GAMS and MATLAB for simulations involving two jobs, two AGVs, and three machines across 13 test problems. MATLAB was used to overcome the limitations of GAMS in handling large or dynamic configurations. Their Ant Colony Optimization method was compared with heuristic baselines. Al-Momani and Al-Aubidv [
58] evaluated a fuzzy-based multi-robot scheduling system involving four programmable machines, a load/unload station, a charging station, and three mobile robots. Performance was benchmarked against Bilge [
91], focusing on cost and time reduction. Samsuria et al. [
64] ran three test cases in MATLAB (small and large scale) using Bilge and Ulusoy benchmarks. They compared the Standard GA (SGA), Particle Swarm Optimization (PSO), Static Time Window (STW), and multiple hybrid/modified GAs, including UGA, AGA, RGA, and IGA. Additional methods like Binary PSO, TS, ALS, HGA, MILP, SLSVNS, SOSA, and SBGA were also evaluated. Li et al. [
65] tested a PSO-based scheduling system on a 50 × 28 grid simulation of a warehouse with three robots and one sorting station. Four task sets (12, 18, 16, and 9 tasks) were processed using PSO with 50 particles and 200 iterations. The model used an inertia weight of 0.7 and acceleration coefficients of 2. The results were presented in Gantt charts and compared against GA. Sun et al. [
73] tested SSA and chaotic-enhanced SSA (CSSA) in MATLAB. The simulations used three and five AGVs, with each test repeated 20 times. The parameters included 100 iterations, a population of 100, and calibrated probabilities for crossover, mutation, and learning. CSSA consistently outperformed SSA in convergence and robustness. Abidi et al. [
76] evaluated a metaheuristic method in MATLAB using 25 iterations and a population of 10. The method was tested against PSO, GWO, WOA, and LA, each combined with classifiers like SVM, Neural Networks (NNs), Deep Belief Networks (DBN), and Fuzzy Deep Belief Networks (FDBN). Pan et al. [
80] tested Learning-based Multi-objective Evolutionary Optimizer (LMEO) using small-scale FJSPT1–10 instances and MK01–10 for medium/large-scale benchmarks. The number of AGVs was scaled with problem size, and machine coordinates were randomly assigned. Transportation time was computed using Euclidean distance. The results were evaluated using Relative Percentage Deviation (RPD) of Cmax. Wang and Xin [
81] simulated a system with 10 jobs, three stages, and three AGVs, with four machines per stage. A genetic algorithm was applied to balance makespan and load distribution among AGVs.
Gurobi simulation experiments
Atik et al. [
63] used the HydraOne AMR platform, integrating Jetson AGX and RealSense cameras with HydraNet for navigation and MobileNet-SSD for object detection. They simulated small, medium, and large instances, varying the number of AMRs, tasks, stations, and maintenance periods. Gurobi solved both an MINLP baseline and the LP version of their proposed MTC algorithm. Performance was compared against three baselines: RMA, RM, and MINLP, under varying maintenance requirements. Gurobi’s academic license ensured reproducibility. Leet et al. [
68] evaluated their approach with two real industrial scenarios: a Kiva (now Amazon Robotics) automated warehouse with two maps, WAREHOUSE1 (280 shelves and four stations), WAREHOUSE2 (240 shelves and 10 stations), and a package sorting center with 28 chutes and four bins (SORTINGCENTER map). The robot used in the simulations was a tricycle model. CCMP implementation was carried out using Python and the Gurobi solver to generate traffic plans. A 2 min time limit was set for planners to start robot movements, with a total execution time defined as 3.6 times the number of products in the workload (in minutes).
CPLEX simulation experiments
Han et al. [
29] tested their dual-resource scheduling method on 20 instances using CPLEX. Yao et al. [
3] evaluated their MILP model on 82 benchmark instances divided by task-to-machine ratios, with representative examples like EX74 and EX740, and implemented the model in C++ using CPLEX. The results were compared to three existing MILP models. Dang et al. [
59] used CPLEX Optimization Studio to solve an MIP model built in OPL, testing 21 uniformly randomized instances.
Flexsim simulation experiments
To validate the model, Mousavi et al. [
53] used two numerical examples. The first involved six jobs processed on six machines, with each job consisting of between two and five operations. The second example involved 15 jobs processed on 10 machines, with each job having between one and five operations. These were followed by experiments conducted in a simulated FMS environment using FlexSim software. The paper compared the performance of the proposed hybrid algorithm with those of the GA and PSO algorithms.
Other simulation experiments
Avhad et al. [
26] tested a Swarm Production System (SPS) in a Software-in-the-Loop (SiL) simulation using Visual Components. A case study with 3 Transportation Robots (TRs) and 10 Workstation Robots (WRs) demonstrated the Swarm Manager’s ability to coordinate planning, scheduling, and control. The impact of three topologies—Linear, Matrix, and Swarm—on performance was evaluated. Integration with the simulation was achieved via OPC-UA. The authors simulated a display manufacturing facility using real-world data from two DRIS units (N1 and N2) in South Korea, incorporating robot velocity, zone lengths, and FIFO job selection rules. The specific simulator was not mentioned. Mareddy et al. [
40] evaluated the adapted SOSA algorithm in a Flexible Manufacturing System (FMS) setup with four CNCs, two AGVs, a central tool magazine, and four layout configurations. A total of 10 job sets were tested with varying transport-to-processing time ratios. A real-world case from an automotive supplier was also included and compared with the Jaya algorithm. Durst et al. [
35] modeled a simplified job shop in SimPy, integrating a Deep Reinforcement Learning-based AGV scheduler using TensorForce. The setup had three AGVs, sources/sinks, and machines with buffers. The study compared FIFO, STD, NJF, LWT, SRPT, and EDD dispatching rules. Geng et al. [
74] simulated a machining workshop with 13 IoT-enabled stations. Real-time data guided a control center running the MAPPO-GC algorithm. Its seven hyperparameters were optimized using Optuna and manual tuning. Xiaoting et al. [
75] tested DQNG on 10 FJSP benchmark instances, each repeated 10 times. Comparisons were made with ILS, TS, GA, and BRKGA. Agrawal et al. [
43] used Unity ML Agents Toolkit to validate agent adaptability under dynamic conditions such as breakdowns and processing delays in a virtual factory. Baruwa and Piera [
54] used the TCPN tool TIMSPAT with ALS and A* algorithms coded in C++ to test 82 problem instances from Bilge and Ulusoy [
91]. Instances were grouped by ratio size (>0.25 and <0.25), and model performance (MCSS vs. VCSS) was evaluated. Kim et al. [
41] evaluated their proposed algorithm in Siemens Plant Simulation, using three baselines: (i) Minimum Time Until Processing, (ii) Shortest Traveling Time, and (iii) Random. The battery charging of the AMRs followed a threshold-based rule with a CONWIP input control policy. Performance was assessed using the average cycle time of car bodies. The small-scale layout included 19 workstations and one charging station, with each workstation performing 4–14 operations. The layout and processing data were based on the specifications of an industry partner. Dang et al. [
57] evaluated their process on a set of 82 benchmark instances. The instances were divided into two groups based on the ratio of average traveling time to average processing time (t/p). This allows for the analysis of the algorithm’s performance under different scenarios. Cechinel and De Pieri [
70] evaluated their approach in a simulated environment created in the Gazebo simulator. This simulated environment represented a warehouse measuring approximately 1500 m², with 8 parking stations, 5 delivery locations, and 192 pickup locations. A grid map was used to represent the environment of the robots’ navigation system. The authors encountered navigation problems due to the detection of other robots as obstacles leading to task failures.
Across the literature, a strong preference for small-scale systems is evident, particularly in real-world implementations, due to constraints in cost, complexity, and scalability. Simulations, especially in MATLAB and optimization environments like Gurobi and CPLEX, dominate experimental setups and allow for broader scalability testing and parameter exploration. Additionally, while many approaches show promising performance improvements, comparative consistency across platforms and standardized benchmarks is lacking. Future research should address scalability in physical deployments, enhance reproducibility, and promote the use of standardized, open-access benchmarks across both real and simulated environments.
5.8.3. Medium-Scale Systems Experiments
Medium-scale systems (6–10 robots) are predominantly evaluated through simulation studies, offering a balance between realism and scalability. These setups typically use standardized benchmarks or custom scenarios to test optimization algorithms under controlled yet varied conditions. Common platforms include MATLAB and Python with Gurobi, along with proprietary or competition-based simulators. Experiments often explore scheduling complexity, AGV coordination, and layout-specific constraints while analyzing performance metrics such as makespan, energy consumption, and system robustness. These studies bridge the gap between small-scale physical implementations and theoretical large-scale models.
Table 14 organizes the experiments conducted in medium-scale systems according to simulated environments and the simulator used in each case.
MATLAB simulation experiments
Samsuria et al. [
27] and Xu et al. [
28] used benchmark datasets from Brandimarte [
93] and Bilge and Ulusoy [
91], implementing their optimization algorithms in MATLAB. He et al. [
2] evaluated 39 benchmark instances across different scales (4 × 4, 6 × 6, 10 × 10, 20 × 5, 20 × 10, and 30 × 10), varying AGV numbers from 2 to mm. The shop floor had a U-shaped layout, and AGVs moved clockwise. Experiments defined processing times, speeds, energy consumption, deterioration rates, and due dates using uniform distributions. Their method was compared with NSGA-II, NSGA-III, MOEA/D, and HACO. Liu et al. [
34] tested their SLGA-D algorithm on seven examples (EX1–EX7), comparing it to GA-D, AGA-D, and HPSO-D in terms of solution quality, stability, and runtime. Additional experiments tested dispatching strategies (DSDS vs. FDS) using EX8 and EX9 and analyzed makespan sensitivity with varying AMR counts (1–10 jobs on four machines, EX10) and job counts (1–15 jobs with four machines and four AMRs, EX11). Krishnamoorthy et al. [
77] used a spinning mill case study to validate their model under two machine configurations (8 and 12 machines). MATLAB simulations were run 50 times with a population of 100 over 100 iterations. The results were benchmarked against existing models, including improved PSO, GA + PSO hybrids, harmony search-based scheduling, and GA + ACO methods tailored for AGV environments.
Gurobi simulation experiments
The study by Qin et al. [
25] employed Python with Gurobi to evaluate experimental instance groups of small, medium, and large scales. The workload balance ratios were varied from 0 to 1 across 11 different scenarios to analyze their impact on performance. The study also examined the effect of order characteristics by considering different order numbers. Large-scale scenarios, including LI1, LI2, LI3, and LI4, were compared to assess the system’s efficiency. The analysis included configurations with 2 to 10 robots, focusing on how varying robot numbers and workload distributions influenced overall performance.
Other simulation experiments
Zhang et al. [
20] tested their method in three warehouse scenarios from a competition, varying grid sizes (18 × 18 and 20 × 20), AGV counts (8 and 10), and berth/obstacle layouts. Order arrival rates were set at 70 and 120 orders. Additional parameters included AGV speed, loading/unloading time, simulation duration, and a penalty term in the ACT. The code was published, but the simulator was not specified. Mayer et al. [
48] simulated six parameter sets to study the impact of AGV count, buffer sizes, task rescheduling, and task release mechanisms (random vs. fixed) across different WIP levels. The metrics included makespan variability, deadlock occurrence, and performance against an idealized optimal solution. Khosiawan et al. [
50] evaluated their method in both lab- and industrial-scale environments with varying task sizes and precedence constraints. Agent teams (three UAVs + two or three AGVs) were used to test coordination effectiveness. The results were compared with standard Differential Evolution (DE) and Particle Swarm Optimization (PSO) algorithms. Vivaldini et al. [
55] simulated two warehouse layouts: Layout 1, with 5 shelves, and Layout 2, with 10. Each included three docks and six depots, with AGVs navigating bidirectional paths. Two orders of 27 unloading tasks (three trucks × nine pallets) were tested per layout, comparing the results with Nearest Neighbor and Tabu Search algorithms. Dou et al. [
66] tested their approach using eight mobile robots in a smart grid warehouse environment. It includes path planning but no coordination between robots, so the system fails when collisions happen. Zhao et al. [
72] tested their approach with 10 robots, and just like the previous authors [
66], the approach only includes path planning, making the solution less scalable. The system developed by Ho et al. [
85] was simulated for an e-commerce scenario involving three geographically distributed warehouses with heterogeneous configurations, including varying block layouts and robot speeds. Task arrivals followed a Poisson distribution (1–6 tasks/min), and task locations were uniformly distributed across rack blocks. Robot speeds varied per time slot using uniform distributions, differing by warehouse. The Neural Networks were trained using PyTorch.
Simulation-based experiments in medium-scale systems provide valuable insights into the performance and adaptability of AGV and AMR scheduling methods. While most studies demonstrate comparative advantages over baseline algorithms, the reliance on synthetic benchmarks and the absence of standardized platforms limit cross-study comparability. Moreover, the lack of real-world physical validation in this scale category indicates an area for future development.
5.8.4. Large-Scale Systems Experiments
Table 15 presents a classification of studies that evaluated robotic and scheduling systems in large-scale systems (>10 robots), distinguishing them by the simulation tools used, including MATLAB, Gurobi, FlexSim, and other custom or hybrid frameworks. These experiments reflect a growing research focus on scaling up task planning, scheduling, and coordination mechanisms in complex warehouse and logistics domains, often featuring dynamic, heterogeneous agents and environments.
MATLAB simulation experiments
Pradhan et al. [
32] considered a dynamic 2D environment in their problem formulation, and simulations were conducted in MATLAB to validate their proposed method. These simulations used accurate models of mobile robots performing various tasks with different robot configurations. To present the results of the queuing model, the MAPLE 18 and MATHEMATICA 9 software packages were used for further analysis and illustration. Bakshi et al. [
60] tested the QCR algorithm on a simple 10-node symmetric TSP to analyze its performance. The results were compared to the orthogonal relaxation method and the case without regularization (a = 0). In this experiment, the QCR algorithm was evaluated before the weight matching step, where the weight of each edge in the solution was represented in the STTS matrix.
Gurobi simulation experiments
Fu et al. [
38] evaluated the CTAS framework using two simulated scenarios. The first, a Capture the Flag Game, featured heterogeneous agents (differing in speed, health, and ammunition) and compared the task assignment of CTAS with STRATA and random assignment. The second scenario, Robotic Services During a Pandemic, involved subtasks like delivery, disinfection, testing, and treatment, modeled across a city using the M3500 road map and a viral exposure-based cost map. Agent capabilities and task demands followed Gaussian distributions. CTAS components (task assignment, scheduling, and generality) were assessed with Gurobi used for optimization. Boysen et al. [
69] conducted experiments on three datasets: ABC (skewed SKU demand), EQ (uniform SKU distribution), and real-world e-commerce data from South Asia. The instances varied in terms of robot and order counts. Gurobi (300 s time limit) was used for solving. The tested methods included rule-based (RULE), random (RND), and decomposition-based approaches (DEC), combining MIPs and heuristics for subproblems: DEC(MIP,MIP) and DEC(HEU,HEU). A single-step MSA variant was also tested against the full two-step MSA (60 scenarios, 120-step horizon) to assess the benefit of the full approach.
Flexsim simulation experiments
To evaluate the proposed dynamic scheduling method, Hu et al. [
36] conducted computational experiments in a dynamic scenario involving varying fleet sizes and task numbers. They used warehouse production data from Changsha, China, with a layout that included 12 buffer area depots, 12 shop depots, 15 automatic vertical warehouse depots, and five charging stations. A distance matrix was employed to compute AGV travel times, and a FlexSim-based digital simulation system was utilized to analyze AGV operations. The dynamic scheduling involved 60 tasks, with new tasks added at specific times, necessitating rescheduling. Gantt charts illustrated the initial and updated schedules, showing task execution and pre-emption during interruptions. The HDSTA algorithm, integrated with the adaptive large neighborhood search (HALNS) algorithm, was compared to the preplanning algorithm (PPA).
Other simulation experiments
Lu et al. [
37] simulated a Robotic Mobile Fulfillment System (RMFS) with 20 storage blocks in a 2 × 5 layout. Picking tasks were randomly generated, with picking and lifting/releasing times of 8 s and 3 s per pod. AGVs moved bidirectionally at 100 cm/s, except near the picking table. A Simulated Annealing (SA) algorithm was applied with defined temperature parameters and cooling rates. Experiments varied the number of AGVs (5, 8, 10, and 14), workstations (two, three, and five), and task sizes (15, 30, 50, and 100). A class-based storage scenario was also tested, where 80% of picking demand came from specific zones. The results were compared with RAR, CAR, OTAR, TS, and SA-I. Bolu and Korcak [
46] developed a digital twin using a Warehouse Execution System (WES) simulation that incorporated real-world factors like robot collisions and energy usage. Scenarios tested included varying order sizes, robot workload balancing, and SKU configurations. Riazi and Lennartson [
47] evaluated their algorithm in Python using Z3 and IBM ILOG CP solvers, with parallelization disabled for fairness. The tests included 8–12 tasks and 4–12 AGVs, simulating 45 min of Volvo Cars production, with a 40 min scheduling horizon and a 3600 s time limit. Dang et al. [
87] implemented a DES model of a COPS warehouse in Python using SimPy. Three scenarios were tested: (1) AMR and picker coordination, tracking congestion and disturbances; (2) dynamic picker assignment using AMR availability; and (3) AMR overtaking, where distances for safe passing were computed based on warehouse layout. Key components (pickruns, AMRs, pickers, and racks) were initialized before simulation execution. Lin et al. [
84] used a small-scale cloud-edge setup with two servers and two PCs. The tests evaluated task scheduling (local vs. cloud PVNS), path planning (PCBS with varying AGVs), and overall system performance in a 20 × 20 warehouse with 12 AGVs and 40 tasks. The study by Feo-Flushing et al. [
82] involved simulations on 2D grid environments of various sizes (5 × 5 to 20 × 20) with heterogeneous agent teams (4–20 agents across four performance classes).
A notable trend across the reviewed studies is the increasing use of heterogeneous agent models, realistic spatial layouts, and multi-objective optimization frameworks to better simulate industrial-scale environments. Tools like Gurobi and custom simulations in Python or C++ were favored for flexibility and computational power, while FlexSim and MATLAB provided visual analysis and quick prototyping.
However, several gaps persist. Many simulations lack real-time constraints or omit inter-agent communication failures, which are critical in operational settings. Furthermore, few studies integrate energy consumption, maintenance scheduling, or unexpected disruptions, which are all vital for deployment in real-world warehouse and production systems.
Looking forward, future research should aim to develop standardized benchmarks for large-scale systems, incorporate hybrid physical-digital experimentation (e.g., digital twins), and address scalability in more detail. Integrating reinforcement learning with formal optimization, as well as simulating uncertainty in demand and supply, also represents promising directions to better bridge simulation and deployment in real-world automated systems.
5.8.5. Evaluation Metrics
During the data extraction phase in
Section 3.4, one of the items defined for data extraction was #DE6, which refers to the evaluation metrics used in each paper to assess the performance and efficiency of the systems under study. To make the visualization of this analysis clearer, a word cloud, as shown in
Figure 7, was created in Python, highlighting the most commonly mentioned evaluation metrics across the reviewed literature.
The makespan, defined as the total time to complete all tasks, emerged as the most frequently evaluated metric, appearing 51 times in the records. This is expected, as makespan is a critical performance measure in many optimization and scheduling problems. Other frequently mentioned metrics include utilization of robots, which appeared 11 times, and computation time, appearing 33 times. The generated schedule is also a strong evaluation metric since many studies include the Gantt chart of the generated schedule in their resultsfor better visualization. These metrics are central to understanding the operational efficiency of systems, especially in contexts involving automation and task scheduling.
Additional important metrics include energy consumption (eight mentions), as many included works aim for energy efficiency, and travel time (seven mentions), related to works including path planning with the aim for the shortest path. Both of these metrics are crucial in evaluating the sustainability and efficiency of automated systems. In relation to operational performance, average waiting time and number of tasks completed were also frequently assessed. These metrics provide insights into system responsiveness and task throughput, which are critical in evaluating the effectiveness of scheduling algorithms.
In terms of solution quality, optimality gap, representing the deviation from the best possible solution, appeared seven times, reflecting the importance of solution accuracy in optimization tasks. fitness value, another key measure of solution quality, was mentioned 6 times.
Other notable metrics include number of deadlocks, which indicate potential system inefficiencies, and robot charging time, relevant in the context of systems that rely on battery-powered robots. On the statistical side, metrics like precision, sensitivity, and mean flow time were discussed but less frequently.
This analysis through a word cloud provides a clear picture of the metrics that are most relevant in the context of the reviewed works, offering a snapshot of the areas of performance that researchers focus on when evaluating system effectiveness.