Next Article in Journal
Vojta Therapy and Conservative Physical Therapy versus Physical Therapy Only for Lumbar Disc Protrusion: A Comparative Cohort Study from Romania
Previous Article in Journal
Relation between Muscle Force of Knee Extensors and Flexor Muscles and Sleep Quality of Women Soccer Players: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Teaching–Learning-Based Optimization Algorithm for the Mobile Robot Path Planning Problem

School of Business and Management, Liaoning Technical University, Huludao 125105, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(4), 2291; https://doi.org/10.3390/app13042291
Submission received: 28 December 2022 / Revised: 7 February 2023 / Accepted: 8 February 2023 / Published: 10 February 2023

Abstract

:
This research proposes an enhanced teaching–learning based optimization (ETLBO) algorithm to realize an efficient path planning for a mobile robot. Four strategies are introduced to accelerate the teaching–learning based optimization (TLBO) algorithm and optimize the final path. Firstly, a divide-and-conquer design, coupled with the Dijkstra method, is developed to realize the problem transformation so as to pave the way for algorithm deployment. Secondly, the interpolation method is utilized to smooth the traveling route as well as to reduce the problem dimensionality. Thirdly, an opposition-based learning strategy is embedded into the algorithm initialization to create initial solutions with high qualities. Finally, a novel, individual update method is established by hybridizing the TLBO algorithm with differential evolution (DE). Simulations on benchmark functions and MRPP problems are conducted, and the proposed ELTBO is compared with some state-of-the-art algorithms. The results show that, in most cases, the ELTBO algorithm performs better than other algorithms in both optimality and efficiency.

1. Introduction

In modern society, mobile robots have been widely introduced into the harsh manufacturing environment and some other industries to reduce labor intensity for people. Many topics associated with mobile robots have been studied, including navigation, path planning, trajectory tracking, and simultaneous localization and mapping (SLAM) [1,2,3,4]. In this regard, MRPP plays a significant role in the field of mobile robots. It is mainly responsible for finding a collision-free and optimal (or approximately optimal) traveling path [5]. This work is motivated by such a context and aims to explore a fast and efficient methodology for MRPP problems.
At present, MRPP-related algorithms mainly include the artificial potential field method, grid method, ant colony algorithm, genetic algorithm and neural network method. Due to its simple mathematical model, small computation time, and strong, real-time nature, the artificial potential field method is widely used [6]. Nevertheless, it is a local path-planning approach in nature that, considering its weaknesses of target unreachability and local minimum points, may lead to the failure of path planning. The grid method is used to model the mobile environment and then to convert the robot path into connections between grids, paving the way for algorithm deployment [7]. It is a global path-planning approach in nature. However, the number of grids increases dramatically when the mobile environment becomes larger. Thus, it is difficult to achieve the desired routes in terms of large and complex environments. In addition, some bionic intelligent optimization algorithms have been proposed and successfully applied to MRPP-related problems. Representative approaches include the ant colony algorithm (ACO), particle swarm optimization (PSO), genetic algorithm (GA), and the neural network (NN). However, it should be noted that some weaknesses of these approaches restrict their applications in MRPP, including the local optimum, low search efficiency, and large burden of computation [8,9,10]. Therefore, this work focuses on designing a more efficient procedure and methodology to generate solutions for MRPP problems.
In real-world applications, major efforts have concentrated on strengthening by hybridizing one candidate bionic intelligent optimization algorithm with other sophisticated skills. The ACO is one example: some of the most recent ACO-based path-planning algorithms are reviewed as follows. To reduce search blindness, Zhao et al. [11] embedded the directional, heuristic information and unevenly distributed initial pheromone into the ACO architecture. The rollback strategy was introduced by Wu et al. [12] to make more ants reach the goal, and the death strategy was also utilized to reduce the influence of an invalid pheromone on the algorithm evolution. Wang et al. [13] designed an improved, pseudo-random proportional rule for the ant state transition which greatly improved the convergence speed. A novel penalty strategy to volatilize the poor path pheromone was embedded into the ACO by Yue et al. [14], which efficiently improved the utilization rate of the ant exploration. Chen et al. [15] tried to improve the visibility of heuristic information by virtue of the principle of infinite step length. In addition, the pheromone update rules, as well as the evaporation rate, were dynamically adjusted to accelerate the convergence speed. The evaluation function of the A-star algorithm and the curvature suppression operator were introduced into the ACO by Dai et al. [16]; these additions were in favor of accelerating the convergence and improving the smoothness of the final path. Luo et al. [17] proposed an attenuation-factor-based heuristic information mechanism, and adjusted the pseudo-random ratio of state selection by means of an adaptive function. The adaptive adjustment factor and pheromone volatilization factor were embedded into the pheromone update rule by Miao et al. [18]. Such modifications efficiently balanced the global search ability and convergence. As can be noted from the above literature, different skills have been embedded into the basic algorithm with respect to the aspects of path formulation, path adjustment, solution modification, individual update, etc. The proposed ETLBO is motivated and designed by these observations and is developed by virtue of TLBO.
Many practical applications have proven TLBO to be superior to some conventional metaheuristics [19,20,21]. TLBO simulates the behaviors of the teacher and students in the classroom, where the so-called teacher and learner phases are utilized to update the population in the algorithm evolution. Some modified, TLBO-related algorithms have already been proposed by many researchers; please refer to the book by Rao [22] for the most recent studies. However, some challenges still exist when applying the TLBO approach to practical optimization problems [23]. First, the adaptation of TLBO to a specified application requires an appropriate problem conversion and solution representation. Second, TLBO tends to get trapped in a local optimum in real-world applications, which is a common weakness of swarm-based metaheuristics. In this context, this paper develops an efficient, TLBO-based algorithm for MRPP.
This work is motivated by the great demand for a fast and efficient path-planning approach for a mobile robot in a complex environment. In the current research, an effort is made to address such an issue with a TLBO-version approach. Four main contributions associated with the methodology design are stated as follows:
  • For the algorithm deployment, a divide-and-conquer design, coupled with the Dijkstra method, is developed to realize the problem transformation;
  • The interpolation method is embedded into the proposed algorithm to smooth the traveling route as well as to reduce the problem dimensionality;
  • An opposition-based learning strategy is utilized to modify the algorithm initialization process;
  • To balance between exploitation and exploration, a novel, individual update method is established by hybridizing TLBO with DE.
The rest of this work is organized as follows. Section 2 formulates the MRPP, and Section 3 presents a brief introduction of TLBO and DE. The proposed methodology is formally described in Section 4. Computational studies are designed and carried out in Section 5. Finally, Section 6 displays conclusions and future work.

2. Problem Formulation

This section models the investigated MRPP problem by virtue of the grid environment map, and then formulates it with the objective of minimizing the route length.

2.1. Problem Description

As is illustrated in Figure 1, a two-dimension grid map was utilized to model the environment so that the coordinate information of each grid can be easily computed with the predefined grid size, a . In this regard, the robot path between the origin and destination can be expressed by a series of connections between different grids.
Figure 2 is employed here to better illustrate the move rules of the grid map. When placed at a specified grid, the robot can only move to, at most, eight adjacent workable grids. Let notation g i represent the robot’s current position and set Ω to denote all workable girds. All possible reachable positions of grid g i can be stated as follows:
N ( g i ) = g j a d ( g i , g j ) 2 a g j Ω
where d ( g i , g j ) represents the distance value between any two grids, g i and g j .

2.2. Mathematical Model

The investigated MRPP aims to find a set of optimal grid connections between the origin and destination so that some numerical metrics can be optimized or improved. Two constraints should be obeyed during the whole decision process: namely, the continuity constraint and the obstacle-avoidance constraint. The previous constraint requires that the robot must start from the origin, move to adjacent grids step by step, and finally reach its destination. The latter states that obstacle grids are not allowed to be included in the robot path.
This research formulated the considered problem with the objective of minimizing the total route length. Given a robot route S = g i | i = 1 , 2 , , n , the objective definition is as follows:
min L = i = 1 n d ( g i , g i + 1 )
where n is the total number of grids in the robot route.

3. TLBO and DE Algorithms

The proposed ETLBO under study capitalizes on the strengths of both TLBO and DE to form a novel and efficient hybrid algorithm.

3.1. TLBO

TLBO simulates the teaching–learning phenomenon in a classroom to solve complicated optimization problems with appreciable efficiency. It is a swarm-based evolutionary algorithm that utilizes a group of so-called individuals to produce a global optimum. In TLBO, the current best solution is considered to be the teacher individual, and the other solutions are treated as student individuals. A group of solutions are created in a random way. They then evolve by two fundamental modes of learning: namely, the teacher phase and the learner phase. The classical TLBO is executed through the following steps:
  • Step 1. Initialize the parameter settings, including the population size, P , maximum iteration, G , problem dimension, D , etc.;
  • Step 2. Create initial individuals then calculate their objective values;
  • Step 3. Employ the teacher phase to update the current population;
  • Step 4. Employ the learner phase to update the current population;
  • Step 5. If any algorithm termination criterion is satisfied, terminate the search process and output the current best individual; otherwise, go to Step 3.
Detailed mathematical formulations of TLBO are stated as follows.
(1)
Initial phase
The solution in TLBO is expressed by an array, x = ( x 1 , x 2 , , x d , , x D ) , in which d { 1 , 2 , , D } . Given a variable range, [ x d l , x d u ] , P random individuals are initialized by the following equation:
x d = x d l + r a n d ( 0 , 1 ) x d u x d l
where r a n d ( 0 , 1 ) is a uniform random number in the interval [0, 1].
(2)
Teacher phase
The mean position of the current population is calculated. Let x ¯ = ( x ¯ 1 , , x ¯ d , , x ¯ D ) be the mean solution, and x represents the current best solution. The teacher phase is implemented based on the position difference between x ¯ and x . In this regard, the equation is stated as follows:
x = x + r a n d ( 0 , 1 ) x x ¯
where parameter is the teaching factor that lies in the set {1, 2}. If the mutant individual performs better than its parent, a greedy selection is then introduced, and x would be replaced by x .
(3)
Learner phase
The learner phase is implemented based on the position difference between the current solution and one randomly selected individual. Let f ( ) represent the objective function formula. The offspring is produced as follows:
x = x + r a n d ( 0 , 1 ) ( x 1 x ) if f ( x 1 ) < f ( x 2 ) x + r a n d ( 0 , 1 ) ( x 2 x ) otherwise
where x 1 and x 2 denote two randomly selected individuals. In addition, the greedy selection is applied in an analogous way to the teacher phase.

3.2. DE

The DE approach is a stochastic and swarm-based optimization algorithm [24]. It begins with a random initial population consisting of a number of target individuals. The population evolves by means of mutation and crossover and selection operators. The detailed operations of a classical DE are as follows:
(1)
Mutation operation
The mutation operation in the DE approach is utilized to generate a mutant vector associated with each target individual in the current population. Given a candidate solution, x , its associated mutant vector, v , can be produced by taking one of the following strategies.
(6) DE / rand / 1   v = x r 1 + F ( x r 2 x r 3 ) (7) DE / rand / 2   v = x r 1 + F ( x r 2 x r 3 ) + F ( x r 4 x r 5 ) (8) DE / best / 1   v = x b e s t + F ( x r 2 x r 3 ) (9) DE / best / 2   v = x b e s t + F ( x r 2 x r 3 ) + F ( x r 4 x r 5 ) (10) DE / current to best / 1   v = x + F ( x b e s t x ) + F ( x r 1 x r 2 )
where x r 1 ~ x r 5 are randomly selected individuals and x b e s t represents the current best solution. Meanwhile, the scaling factor, F , is used to control the differential variation between two individuals.
(2)
Crossover operation
The DE algorithm employs a crossover operation to generate candidate offspring in order to the increase population diversity. In this regard, a candidate offspring, u , is produced by combing solution information from each pair of a target vector, x , and its associated mutant vector, v . The scheme can be described as follows:
u d = v d r a n d ( 0 , 1 ) C R x d otherwise
where d { 1 , 2 , , D } , and r a n d ( 0 , 1 ) is a random number in the interval [ 0 , 1 ] . Symbols x d , v d , and u d represent the d - th gene in the individual vectors x , v , and u , respectively. The crossover probability, C R , is utilized to control the proportion of genes copied from target and mutant vectors.
(3)
Selection operation
The selection operation in the DE algorithm employs a one-to-one greedy selection to determine whether the candidate offspring, u , can replace its parent individual, x . Solution x will be replaced by u once that u performs equally or even better than x .

4. Proposed ETLBO for MRPP

This section presents a detailed introduction of the proposed methodology for the MRPP problem.

4.1. Divide-and-Conquer Design for MRPP

As is illustrated in Figure 3, a divide-and-conquer design is embedded into the algorithm deployment. In this regard, the investigated MRPP problem is broken down into multiple simple sub-problems, for which an appropriate divide-and-conquer grid in each column between the origin and destination must be optimized to form a simple route [25]. Then, local optimal routes between any two divide-and-conquer grids are generated by fast and sophisticated methods to form a solution to the original MRPP problem.
Two decisions must be made when dealing with the converted problem. First, an appropriate divide-and-conquer grid in each column between the origin and destination must be determined in order to construct the simple route. Second, local optimal routes between any two divide-and-conquer grids must be obtained to form a solution to the original MRPP problem. The prior decision is to be tackled by the proposed ETLBO, while the latter decision is optimized by the Dijkstra approach. More details are presented here to better explain the application of the Dijkstra approach. Let us denote the grid in a (row, column) format. For example, (1, 3) represents the grid located at row 1 and column 3. On this basis, all workable grids in column 3 and column 4 can be expressed by sets A = {(1, 3), (2, 3), (3, 3), (6, 3), (7, 3), (8, 3), (9, 3), (10, 3)} and B = {(1, 4), (3, 4), (6, 4), (7, 4), (8, 4), (9, 4), (10, 4)}, respectively. The Dijkstra approach is then utilized to obtain optimal routes between any two positions from sets A and B, respectively. Since these positions are located in columns adjacent to each other, the decision process of the Dijkstra approach is time-saving and always ensures optimality.

4.2. Interpolation Application in ETLBO

As stated in the previous subsection, ETLBO is used to determine the locations of every divide-and-conquer grid at each column between the origin and destination. To reduce the problem dimensionality and to smooth the traveling routes, the interpolation method is embedded into the application of the ETLBO. The corresponding illustrative schematic diagram is depicted in Figure 4.
It can be observed from Figure 4 that some divide-and-conquer grids are to be optimized by the solution array of the ETLBO, such as grids ④ and ⑦, while the other divide-and-conquer grids are obtained by the interpolation method. In the case in which an interpolated grid is located at an obstacle grid, the nearest workable grid in that column is selected as the divide-and-conquer grid. In this regard, the B-Spline interpolation method is selected to pave the way for ELTBO deployment according to some existing MRPP-related studies [26].

4.3. Pyramid of ETLBO

ETLBO is established by virtue of the outstanding performance of TLBO. Meanwhile, the opposition-based learning mechanism and the DE-related offspring method are utilized to enhance the algorithm performance.

4.3.1. Opposition-Based Learning Initialization Method

The opposition-based learning mechanism is a simple but efficient method to strengthen the algorithm’s performance in swarm computation [27]. In ETLBO, the opposition-based learning mechanism is used to create intimal solutions with high qualities. In this regard, P solutions are initialized in a random way. Then, P opposition-based solutions are generated based on the opposition-based learning concept. Finally, P individuals with better evaluations are selected to form the initial population.
As stated previously, TLBO is run by virtue of P individuals, and each candidate can be marked as an array x = ( x 1 , x 2 , , x d , , x D ) in which d { 1 , 2 , , D } . In this regard, the search space is defined by [ x d l , x d u ] . In the initialization, let us denote the i - th random solution and its associated, opposition-based solution by x p , i = ( x 1 p , i , , x d p , i , , x D p , i ) and x o , i = ( x 1 o , i , , x d o , i , , x D o , i ) , respectively. According to “Definition 2” stated by Mahdavi et al. [27], the following formula holds true:
x d o , i = x d l + x d u x d p , i
Based on the above concepts, the initialization process in ETLBO is implemented as follows:
  • Step 1. Set i 1 and d 1 . Then, go to step 2.
  • Step 2. If condition i P holds, go to step 3; otherwise, go to step 7;
  • Step 3. If condition d D holds, go to step 4; otherwise, go to step 5;
  • Step 4. Set x d p , i = x d l + r a n d ( 0 , 1 ) ( x d u x d l ) and x d o , i = x d l + x d u x d p , i . Then, go to step 5;
  • Step 5. Set d d + 1 , if condition d > D holds, go to step 6; otherwise, go to step 3;
  • Step 6. Set i i + 1 ; then, go to step 2;
  • Step 7. Select P best individuals from { x p , 1 , , x p , P , x o , 1 , , x o , P } to form the initial population.

4.3.2. Hybrid Offspring Generation Method

An efficient swarm-based metaheuristic should make full use of both the local and global information on the solutions found. Local information on solutions can be helpful for exploitation, while global information tends to guide the search for exploring promising areas. The evolution of TLBO mainly depends on the global information, and thus TLBO has an obvious advantage in that learner individuals will converge at a single attractor/teacher individual. It can be observed that DE evolves mainly based on the distance and direction information, which are types of local information in nature. In other words, DE has the advantage of not being biased toward any prior defined guide, and such an attribute makes DE able to maintain population diversity and explore local searches.
The key reason for employing the hybridization of different metaheuristics is that the hybrid approach can make full use of the strengths of each algorithm while simultaneously overcoming the associated limitations. In this regard, instead of employing a single updated learner individual in the TLBO, the proposed ETLBO makes use of the integration of two offspring individuals. The first offspring individual is computed by the TLBO, and the second one is obtained by the DE operators. Detailed mathematical expressions are described as follows.
(1)
Teacher phase
Given a candidate solution, x , the corresponding offspring individual, x , is calculated by the hybridization of the teacher phase in TLBO and differential learning in DE:
x = u 1 x T L B O , T e a c h + ( 1 u 1 ) x D E , T e a c h
where x T L B O , T e a c h represents the offspring individual obtained by Equation (4) in TLBO, while x D E , T e a c h represents the offspring individual computed by Equations (6)–(11) in DE. The hybridization factor u 1 is a random number that lies in the interval [0, 1].
(2)
Learner phase
Analogous to the teacher phase, the offspring individual x of a candidate solution x is generated by taking the following equation:
x = u 2 x T L B O , L e a r n + ( 1 u 2 ) x D E , L e a r n
where x T L B O , L e a r n represents the offspring individual obtained by Equation (5) in TLBO, while x D E , L e a r n is the offspring individual computed by Equations (6)–(11) in DE. The hybridization factor u 2 is a random number that lies on interval [0, 1].

4.4. Framework of Proposed Methodology for MRPP

By virtue of the previous statements, the flowchart of the proposed ETLBO for the MRPP problem is illustrated in Figure 5. Simulations are to be conducted in the next section to validate the algorithm performance.

5. Experimental Studies

Simulation experiments on the benchmark functions and MRPP problems were designed and carried out to validate the proposed methodology. All experiments were performed on a MATLAB 2016b platform, using a personal computer with a 2.4 GHz Intel Core i5 CPU and 4 GB of RAM.

5.1. Benchmark Function Problem

In this subsection, the ETLBO is validated by virtue of 20 benchmark function problems which are commonly used to evaluate new optimization methods. These functions can be classified according to several criteria, such as their linearity (or nonlinearity), convexity (or non-convexity), and continuousness (or non-continuousness). However, unimodal (or multimodal), separable (or non-separable), and dimensionality characteristics are of great importance for benchmark functions to evaluate the algorithm efficiency. Unimodal functions increase and decrease monotonically, while multimodal functions have two or more vague peaks. Separable functions are easier to resolve when compared to non-separable functions. Regarding the problem dimensionality, its increase also raises the optimization difficulty. In this context, this work employed benchmark functions with combinations of different characteristics to evaluate the proposed ETLBO.

5.1.1. Preliminaries

A. 
Unimodal functions
Table 1 summarizes the unimodal functions utilized in this study. In this regard, the search space, problem dimension, and optimum for 12 benchmark functions, F1–F12, are presented [28]. Functions F1–F3 were separable, and functions F4–F12 were non-separable. The complexity expanded with increasing problem dimensions, which varied up to a maximum value of 100.
B. 
Multimodal functions
Multimodal functions F13–F20 were employed for the algorithm evaluation in this work [29]. Functions F13–F14 were separable, while functions F15–F20 were non-separable. The multimodal functions employed here are detailed in Table 2 with their related search space, problem dimension, and optimum.
C. 
Evaluation metrics
To evaluate the algorithm performance, some metrics were calculated based on 50 independent runs. The metrics utilized in this subsection included the best, worst, median, average, and standard deviation (SD) values, computed according to 50 independent runs of each benchmark function.
D. 
Non-parametric tests
Non-parametric tests are statistical analysis experiments for sample data that do not require the underlying population for assumptions. This work employed the Wilcoxon rank-sum test to validate the algorithm performance for different test functions [30]. In this regard, the rankings of the sample data were allotted by combining two populations. Since all functions were considered minimization problems, observations in two populations were first sorted in ascending order, and the Wilcoxon rank-sum statistics were then calculated in the following manner:
w A = r R A r
where w A is the sum of rank values in population A.
In addition, in a Wilcoxon rank-sum test for the comparison of two candidate algorithms, the null hypothesis and the alternate hypothesis can be stated as follows: H0: population A ≥ population B; Ha: population A < population B. In this regard, the related p value is given by:
p = Pr ( W A w A ) H a : A > B Pr ( W A w A ) H a : A < B
where W A represents the distribution of the rank sum. The resulting p is compared to the so-call significance level, which is set to 0.05 in the current study.

5.1.2. Results and Discussion

Simulations were carried out to compare the solution performance of ETLBO with some well-known and state-of-art methods, including TLBO, TLBO(S) (an enhanced TLBO proposed by Shukla et al.) [31], PSO, DE, and GA. According to some related studies and tentative experiments, the population size and maximum iteration of all test algorithms were set as 50 and 5000, respectively. For ETLBO, the scaling factor, F , was set to 0.4 and the crossover probability, C R , was fixed at 0.8. Other parameters of compared algorithms were set to values recommended by the related literature. For the sake of constancy, the simulation of a test algorithm for each benchmark function was exacted 50 independent times. The results are presented and discussed as follows.
A. 
Results of unimodal functions
Table 3 presents the simulation results of all unimodal functions, for which the best, worst, median, average, and SD metrics are calculated according to independent runs of each algorithm. In this regard, the best values are emphasized in bold. The following comments can be drawn based on these results:
  • ETLBO obtained the optimum for functions F2, F4, F6, and F9-F12 in all 50 runs; thus, the SD values were zero. In terms of functions F1, F3, and F7, ETLBO obtained near-optimal solutions and exceeded all other algorithms;
  • For function 5, the optimum was obtained by PSO and DE. ETLBO performed third with respect to the previous two algorithms, and obtained a higher accuracy than the other three compared algorithms;
  • For function F8, the DE algorithm yielded the best solution performance, followed by PSO. GA performed third with respect to the previous algorithms. ETLBO obtained more accurate solutions than TLBO and TLBO(S) in terms of the average value metric.
Based on the above presentations, it can be concluded that the proposed ETLBO performed better and obtained more consistent accurate solutions for most unimodal functions.
B. 
Results of multimodal functions
In Table 4, the simulation results of all tested multimodal functions are collected. The resulting best, worst, median, average, and SD values of each algorithm were computed by virtue of independent runs. The best values are emphasized in bold, and the following comments can be provided:
  • ETLBO obtained the optimum of functions F13 and F16 in all 50 independent runs; the resulting SD values were zero;
  • For functions F15 and F19, ETLBO obtained near-optimal solutions and performed better than all compared algorithms. Regarding function F19, the result provided by ETLBO was a close approximation to the optimum and the corresponding SD value was very close to zero;
  • The six algorithms all performed very well for function 14, and no significant difference was found regarding stability;
  • For functions 18 and 20, TLBO(S) and DE yielded the best solution performances of all compared algorithms, respectively. Despite the fact that ETLBO failed to obtain the best solution performance, its metrics were acceptable in a relative manner since those values were located in an acceptable range;
  • For function 17, ETLBO failed to obtain a good solution in 50 independent runs. The resulting metrics of ETLBO were higher when compared to the simulation results obtained by the compared algorithms, to a certain extent.
In summary, the proposed ETLBO performed better and provided more accurate results for most multimodal functions.
C. 
Nonparametric test results
The Wilcoxon rank-sum test was employed here to independently compare ETLBO with the other tested algorithms. Since all algorithms were executed for 50 independent runs, the ranking value varies from 1 to 100 for each sample data in the combination of ETLBO and the compared algorithm. Thus, the possible sum of ranking ( w A ) falls in the range [1275, 3775].
The minimum value of the metric w A occurred when the ETLBO completely outperformed the compared algorithm in all 50 runs. On the other hand, the maximum value of the metric w A was observed when the ETLBO underperformed its compared algorithm in all 50 runs. In Table 5, the w A metric and the p values of the Wilcoxon rank-sum test associated with each compared algorithm are calculated and displayed.
For functions F1, F3–F13, F15–F16, F18, and F20, the associated rank sum of the ETLBO (compared to the basic TLBO) is 1275, and the resulting p value is almost zero. Since the significance level is set to 0.05 for this study, it can be concluded that the ETLBO dominates TLBO for all 50 runs when they are applied to address each of the tested functions. Thus, H0 can be rejected, confirming that ETLBO provides minimum values for these benchmark functions. For functions F2 and F17, the resulting p values are less than (or equal to) the predefined significance level. In other words, TLBO performs equally or even better when compared to ETLBO. For functions F14 and F19, the rank sum of ETLBO lies between the maximum and minimum rank threshold values, implying that ETLBO and TLBO produce a mixed ranking. In addition, since the p value is less than 0.05, H0 can be rejected, and ETLBO obtains more excellent solutions in most cases. A similar analysis is conducted and compared between ETLBO and other test algorithms (namely, TLBO(S), PSO, DE, and GA). A win for ETLBO over the compared algorithm is marked as “√”; similar performance is denoted by ‘‘=’’; and failures are represented by ‘‘×’’.
In summary, the proposed ETLBO outperforms other compared algorithms in a statistical manner when applied to address most of the above-mentioned benchmark functions.

5.2. MRPP Problem

MRPP problems were utilized to compare the proposed methodology’s accuracy and performance with some other path-planning metaheuristics. The experimental design in this subsection was arranged as follows: First, instances with different problem sizes were generated in a random manner. The proposed methodology was then compared with two representative and state-of-art algorithms: ACO(L) and ACO(G) (two enhanced ACO algorithms proposed by Luo et al. and Gao) [17,32]. To reduce statistical errors, one algorithm was independently simulated 50 times for a function.

5.2.1. Preliminaries

A. 
Instance generation
MRPP problems in the current study were carried out on four randomly generated obstacle maps. The grid size was set to 5, and the map sizes were set as 20 × 20, 30 × 30, 40 × 40, and 50 × 50. Such an experimental design reveals the simulations of algorithms for middle-scale and even large-scale instances that reflect the size of real-world applications.
B. 
Evaluation metrics
To evaluate the algorithm performance, some metrics were calculated based on 50 independent runs. Metrics utilized in this subsection included the best, worst, average, and related deviation values, computed according to 50 independent runs of each instance. The deviation is calculated as follows [33]:
G A P = ( L L best ) L best × 100 %
where L best is the optimal route length obtained by three algorithms in all runs. Meanwhile, L represents the best, worst, and average route length by a selected approach in 50 independent runs.

5.2.2. Results and Discussion

Simulations were carried out to compare the solution performance of ETLBO with ACO(L) and ACO(G). By virtue of the related studies and tentative experiments, the population size and maximum iteration of all test algorithms were set as 20 and 500, respectively. Regarding ETLBO, the scaling factor, F , was set to 0.4, and the crossover probability, C R , was fixed at 0.8. Other parameters of ACO(L) and ACO(G) algorithms were set to values recommended by the related literature. To reduce statistical errors, the simulation of a test algorithm for one instance was exacted for 50 independent instances. The results are presented and discussed as follows.
A. 
Presentation of best routes
The best routes of information of different algorithms for all instances are detailed in Figure 6, Figure 7, Figure 8 and Figure 9 in order to vividly state the algorithm performance. In this regard, the best robot path for a candidate algorithm for one instance was described by virtue of a single environment map. Meanwhile, objective values versus the number of iterations are illustrated in each figure by means of convergence curves, which represent the best results among 50 independent trials of each algorithm for the corresponding instance. The following comments can be drawn based on these results.
  • For instance 20 × 20, it can be seen from Figure 6 that ETLBO yielded the best solution performance of the three test algorithms, with the best route length of 151.92 (Figure 6d). ACO(G) performed second to ETLBO, with the best route length of 154.85 (Figure 6c). ACO(L) showed the weakest performance in the best route length, with an objective value of 156.07 (Figure 6b);
  • For the 30 × 30 instance, ETLBO obtained more accurate solutions than ACO(L) and ACO(G) in terms of the best route length metric. The objective value of the proposed ETLBO amounted to 225.56 (see Figure 7d), which is superior to 232.63 (given by ACO(L) in Figure 7b) and 231.42 (provided by ACO(G) in Figure 7c);
  • Similar conclusions can be drawn from Figure 8 and Figure 9, in which ETLBO yielded the best solution performance. The corresponding objective values for instances 40 × 40 and 50 × 50 were 291.27 (see Figure 8d) and 378.70 (see Figure 9d), respectively. ACO(G) performed second to ETLBO, with metric values of 291.27 (see Figure 8c) and 378.70 (see Figure 9c). ACO(L) ranked at the bottom, with an objective value of 302.99 (see Figure 8b) and 400.42 (see Figure 9b).
  • The convergence curves illustrated in four figures (Figure 6a, Figure 7a, Figure 8a and Figure 9a) indicate that the proposed ETLBO not only had a faster convergence in solving all test instances when compared to the state-of-art ACO(L) and ACO(G) but the final convergence accuracy of the proposed approach was better than that of the compared algorithms.
In summary, among 50 independent trials, the proposed ETLBO exceeded other compared algorithms in terms of the best route length metric.
B. 
Statistical comparison
Table 6 collects the simulation results obtained for all considered problem instances and algorithms. The resulting best, worst, and average objective values, as well as the corresponding deviations, were computed according to 50 independent runs. The best values among the three compared algorithms are emphasized in bold, and the following comments can be drawn according to these results.
  • In terms of the optimum, ETLBO was able to find the best solutions for all considered problem instances of different problem scales. However, ACO(L) and ACO(G) failed to obtain the best or near-optimal solutions especially for large-scale problem instances. In this regard, the values of the best deviation metric amounted to 5.73% (given by ACO(L)) and 4.51% (given by ACO(G));
  • ETLBO yielded the best solution performance when it was applied to address different problem instances in terms of best-, worst-, and average-related metrics. ACO(G) performed second to ETLBO, while ACO(L) showed the weakest solution performance;
  • The average metrics versus the problem size indicated that the increasing problem dimensionality raised the optimization difficulty. In this regard, the average deviations of ACO(L), ACO(G), and ETLBO for the 20 × 20 instance were 5.42%, 4.52%, and 3.10%, respectively. For the large-scale instance with a 50 × 50 size, the average deviations amounted to 8.81%, 8.20%, and 5.17%, respectively;
  • Despite the diversity of problem instances, it is evident that all generated values from the proposed ETLBO were more acceptable when compared to the two other algorithms compared.
To reveal the statistical significance of the simulation results, the Wilcoxon rank-sum test was utilized to independently compare the ETLBO with the ACO(L) and ACO(G). As stated in the previous subsection, the value of w A fell in range [1275, 3775] when comparing ETLBO with another robot-path-planning algorithm. The minimal w A occurred when ETLBO completely outperformed the compared algorithm in all 50 runs, while the maximal w A was observed in the circumstance that ETLBO underperformed its rival in all executions. The w A metric and the p values of the Wilcoxon rank-sum test associated with each compared algorithm were calculated and presented in Table 7, where mark “√” represents a win for ETLBO over its rival.
In respect to the comparison between ETLBO and ACO(L), the following comments can be made:
  • For the 50 × 50 instance, the associated rank sum of ETLBO was 1275, and the resulting p value was very close to zero. As is stated in previous subsections, the significance level was fixed at 0.05 for this study. Therefore, it can be deduced that ETLBO dominates the ACO(L) in all executions when they are adopted to this instance.
  • For the instances 20 × 20, 30 × 30, and 40 × 40, the rank sum of ETLBO lay between the maximum and minimum rank threshold values, indicating that ETLBO and ACO(L) produced a mixed ranking. Nonetheless, the associated p values were less than the predefined significance level. In other words, ETLBO performs equally or even better when compared to ACO(L).
  • Despite the increasing problem dimensionality raising the optimization difficulty, the resulting p values decreased with the enlargement of the problem scale. Thus, it can be deduced that ETLBO is more competitive than ACO(L) for a real-world-scale problem.
A similar comparison between ETLBO and ACO(G) was conducted and illustrated. In summary, the proposed ETLBO outperforms the compared algorithms in a statistical manner and provides consistent and more accurate solutions for MRPP problems.

6. Conclusions

This research proposes a so-called ETLBO algorithm to realize an efficient path planning for a mobile robot. Four strategies are introduced to accelerate the basic TLBO algorithm and optimize the final path. Firstly, a divide-and-conquer design coupled with the Dijkstra method is developed to realize the problem transformation so as to pave the way for the algorithm deployment. Then, the interpolation method is utilized to smooth the traveling route as well as to reduce the problem dimensionality. Thirdly, the algorithm initialization process is modified by an opposition-based learning strategy. Finally, a hybrid individual update method is prosed to strike a balance between exploitation and exploration.
Simulation experiments are designed to validate the proposed algorithm. In terms of benchmark function problems, the proposed ETLBO performs better than the compared algorithms in a statistical manner and obtains consistent and more accurate solutions for most optimization function problems. With respect to MRPP problems, ETLBO outperforms some state-of-are algorithms in both optimality and efficiency, and thus can be implemented on the real-world robot for path planning.
Since the local optimal path planning in current algorithm is solved by the Dijkstra method, it is important to modify the existing divide-and-conquer design in future research work to reduce the computation burden when faced with more complicated practical problems. In addition, the proposed methodology can be adjusted and adapted to the dynamic environment in which certain real-time factors are seriously considered.

Author Contributions

S.L.: Conceptualization, Methodology; D.L. (Danyang Liu): Data Curation, Writing—Original Draft; D.L. (Dan Li): Visualization; X.S.: Writing—Review & Editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Liaoning province Department of Education fund (21-A817).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Friudenberg, P.; Koziol, S. Mobile robot rendezvous using potential fields combined with parallel navigation. IEEE Access 2018, 6, 16948–16957. [Google Scholar] [CrossRef]
  2. Contreras-Cruz, M.A.; Ayala-Ramirez, V.; Hernandez-Belmonte, U.H. Mobile robot path planning using artificial bee colony and evolutionary programming. Appl. Soft Comput. 2015, 30, 319–328. [Google Scholar] [CrossRef]
  3. Binh, N.T.; Tung, N.A.; Nam, D.P.; Quang, N.H. An adaptive backstepping trajectory tracking control of a tractor trailer wheeled mobile robot. Int. J. Control Autom. Syst. 2019, 17, 465–473. [Google Scholar] [CrossRef]
  4. Kellalib, B.; Achour, N.; Coelen, V.; Nemra, A. Towards simultaneous localization and mapping tolerant to sensors and software faults: Application to omnidirectional mobile robot. Proc. Inst. Mech. Eng. Part I J. Syst. Control Eng. 2020, 235, 269–288. [Google Scholar] [CrossRef]
  5. Ab Wahab, M.N.; Nefti-Meziani, S.; Atyabi, A. A comparative review on mobile robot path planning: Classical or meta-heuristic methods? Annu. Rev. Control 2020, 50, 233–252. [Google Scholar] [CrossRef]
  6. Orozco-Rosas, U.; Montiel, O.; Sepúlveda, R. Mobile robot path planning using membrane evolutionary artificial potential field. Appl. Soft Comput. 2019, 77, 236–251. [Google Scholar] [CrossRef]
  7. Liu, J.; Yang, J.; Liu, H.; Tian, X.; Gao, M. An improved ant colony algorithm for robot path planning. Soft Comput. 2017, 21, 5829–5839. [Google Scholar] [CrossRef]
  8. Akka, K.; Khaber, F. Mobile robot path planning using an improved ant colony optimization. Int. Ournal Adv. Robot. Syst. 2018, 15, 1729881418774673. [Google Scholar] [CrossRef]
  9. Nazarahari, M.; Khanmirza, E.; Doostie, S. Multi-objective multi-robot path planning in continuous environment using an enhanced genetic algorithm. Expert Syst. Appl. 2019, 115, 106–120. [Google Scholar] [CrossRef]
  10. Wang, G.; Zhou, J. Dynamic robot path planning system using neural network. J. Intell. Fuzzy Syst. 2021, 40, 3055–3063. [Google Scholar] [CrossRef]
  11. Zhao, J.; Cheng, D.; Hao, C. An improved ant colony algorithm for solving the path planning problem of the omnidirectional mobile vehicle. Math. Probl. Eng. 2016, 2016, 7672839. [Google Scholar] [CrossRef]
  12. Wu, X.; Wei, G.; Song, Y.; Huang, X. Improved ACO-based path planning with rollback and death strategies. Syst. Sci. Control Eng. 2018, 6, 102–107. [Google Scholar] [CrossRef]
  13. Wang, H.; Guo, F.; Yao, H.; He, S.; Xu, X. Collision Avoidance Planning Method of USV Based on Improved Ant Colony Optimization Algorithm. IEEE Access 2019, 7, 52964–52975. [Google Scholar] [CrossRef]
  14. Yue, L.; Chen, H. Unmanned vehicle path planning using a novel ant colony algorithm. EURASIP J. Wirel. Commun. Netw. 2019, 2019, 136. [Google Scholar] [CrossRef]
  15. Chen, G.; Liu, J. Mobile Robot Path Planning Using Ant Colony Algorithm and Improved Potential Field Method. Comput. Intell. Neurosci. 2019, 2019, 1932812. [Google Scholar] [CrossRef]
  16. Dai, X.; Long, S.; Zhang, Z.; Gong, D. Mobile Robot Path Planning Based on Ant Colony Algorithm with A Heuristic Method. Front. Neurorobotics 2019, 13, 15. [Google Scholar] [CrossRef]
  17. Luo, Q.; Wang, H.; Zheng, Y.; He, J. Research on path planning of mobile robot based on improved ant colony algorithm. Neural Comput. Appl. 2019, 32, 1555–1566. [Google Scholar] [CrossRef]
  18. Miao, C.; Chen, G.; Yan, C.; Wu, Y. Path planning optimization of indoor mobile robot based on adaptive ant colony algorithm. Comput. Ind. Eng. 2021, 156, 107230. [Google Scholar] [CrossRef]
  19. Rao, R.V.; Savsani, V.; Vakharia, D.P. Teaching–learning-based optimization: A novel method for constrained mechanical design optimization problems. Comput.-Aided Des. 2011, 43, 303–315. [Google Scholar] [CrossRef]
  20. Rao, R.V.; Patel, V. An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems. Sci. Iran. 2013, 20, 710–720. [Google Scholar] [CrossRef] [Green Version]
  21. Camp, C.V.; Farshchin, M. Design of space trusses using modified teaching–learning based optimization. Eng. Struct. 2014, 62, 87–97. [Google Scholar] [CrossRef]
  22. Rao, R.V. Teaching-learning-based optimization algorithm. Teaching Learning Based Optimization Algorithm; Springer: Cham, Switzerland, 2016; pp. 9–39. [Google Scholar]
  23. Zou, F.; Chen, D.; Xu, Q. A survey of teaching–learning-based optimization. Neurocomputing 2019, 335, 366–383. [Google Scholar] [CrossRef]
  24. Das, S.; Suganthan, P.N. Differential evolution: A survey of the state-of-the-art. IEEE Trans. Evol. Comput. 2010, 15, 4–31. [Google Scholar] [CrossRef]
  25. Diwan, S.P.; Deshpande, S.S. Fast Nonlinear Model Predictive Controller using Parallel PSO based on Divide and Conquer Approach. Int. J. Control 2022, 1–10. [Google Scholar] [CrossRef]
  26. Lin, D.; Shen, B.; Liu, Y.; Alsaadi, F.E.; Alsaedi, A. Genetic algorithm-based compliant robot path planning: An improved Bi-RRT-based initialization method. Assem. Autom. 2017, 37, 261–270. [Google Scholar] [CrossRef]
  27. Mahdavi, S.; Rahnamayan, S.; Deb, K. Opposition based learning: A literature review. Swarm Evol. Comput. 2018, 39, 1–23. [Google Scholar] [CrossRef]
  28. Xu, L.; Cao, X.; Du, W.; Li, Y. Cooperative path planning optimization for multiple UAVs with communication constraints. Knowl.-Based Syst. 2023, 260, 110164. [Google Scholar] [CrossRef]
  29. Alorf, A. A survey of recently developed metaheuristics and their comparative analysis. Eng. Appl. Artif. Intell. 2023, 117, 105622. [Google Scholar] [CrossRef]
  30. Hashim, F.A.; Mostafa, R.R.; Hussien, A.G.; Mirjalili, S.; Sallam, K.M. Fick’s Law Algorithm: A physical law-based algorithm for numerical optimization. Knowl.-Based Syst. 2023, 260, 110146. [Google Scholar] [CrossRef]
  31. Shukla, A.K.; Singh, P.; Vardhan, M. An adaptive inertia weight teaching-learning-based optimization algorithm and its applications. Appl. Math. Model. 2020, 77, 309–326. [Google Scholar] [CrossRef]
  32. Gao, W.; Tang, Q.; Ye, B.; Yang, Y.; Yao, J. An enhanced heuristic ant colony optimization for mobile robot path planning. Soft Comput. 2020, 24, 6139–6150. [Google Scholar] [CrossRef]
  33. Yin, L.; Li, X.; Lu, C.; Gao, L. Energy-efficient scheduling problem using an effective hybrid multi-objective evolutionary algorithm. Sustainability 2016, 8, 1268. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Grid environment map.
Figure 1. Grid environment map.
Applsci 13 02291 g001
Figure 2. Possible movements of the robot.
Figure 2. Possible movements of the robot.
Applsci 13 02291 g002
Figure 3. Divide-and-conquer design for MRPP.
Figure 3. Divide-and-conquer design for MRPP.
Applsci 13 02291 g003
Figure 4. Schematic diagram of interpolation method.
Figure 4. Schematic diagram of interpolation method.
Applsci 13 02291 g004
Figure 5. Proposed ETLBO for the MRPP.
Figure 5. Proposed ETLBO for the MRPP.
Applsci 13 02291 g005
Figure 6. Best solution information of instance 20 × 20.
Figure 6. Best solution information of instance 20 × 20.
Applsci 13 02291 g006
Figure 7. Best solution information of instance 30 × 30.
Figure 7. Best solution information of instance 30 × 30.
Applsci 13 02291 g007
Figure 8. Best solution information of instance 40 × 40.
Figure 8. Best solution information of instance 40 × 40.
Applsci 13 02291 g008
Figure 9. Best solution information of instance 50 × 50.
Figure 9. Best solution information of instance 50 × 50.
Applsci 13 02291 g009
Table 1. Unimodal benchmark functions.
Table 1. Unimodal benchmark functions.
LabelNameFormulaRangeDimensionOptimum
F1Sphere f 1 ( x ) = i = 1 n x i 2 [−100, 100]1000
F2Step f 6 ( x ) = x i + 0.5 2 [−100, 100]300
F3Quartic f 7 ( x ) = i = 1 n i x i 4 + r a n d 0 , 1 [−1.28, 1.28]300
F4Schwefel 2.21 f 4 ( x ) = max x i x i , 1 i n [−100, 100]600
F5Booth f 5 ( x ) = ( x 1 + 2 x 2 7 ) 2 + ( 2 x 1 + x 2 5 ) 2 [−10, 10]20
F6Schwefel 2.22 f 2 ( x ) = i = 1 n x i + i = 1 n x i [−10, 10]400
F7Schwefel 1.2 f 3 ( x ) = i = 1 n j 1 i x j 2 [−100, 100]500
F8Easom f 8 ( x ) = cos ( x 1 ) cos ( x 2 ) exp ( ( x 1 π ) 2 ( x 2 π ) 2 ) [−100, 100]2−1
F9Bohachevsky f 14 ( x ) = x 1 2 + 2 x 2 2 0.3 cos ( 3 π x 1 ) 0.4 cos ( 4 π x 2 ) + 0.7 [−100, 100]20
F10Matyas f 17 ( x ) = 0.26 x 1 2 + x 2 2 0.48 x 1 x 2 [−10, 10]20
F11Three-hump camel f 18 ( x ) = 2 x 2 1.05 x 4 + x 6 6 + x y + y 2 [−5, 5]20
F12Schaffer N.2 f 20 ( x ) = 0.5 + sin 2 x 2 y 2 0.5 1 + 0.001 x 2 + y 2 2 [−100, 100]20
Table 2. Multimodal benchmark functions.
Table 2. Multimodal benchmark functions.
LabelNameFormulaRangeDimensionOptimum
F13Rastrigin f 9 ( x ) = i = 1 n x i 2 10 cos 2 π x i + 10 [−5.12, 5.12]200
F14Branin f 15 ( x ) = x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 2 + 10 1 8 π cos x 1 + 10 [−5, 15]20.398
F15Ackley f 10 ( x ) = 20 exp 0.2 1 n i = 1 n x i 2 exp 1 n i = 1 n cos 2 π x i + 20 + e [−32.32]300
F16Griewank f 11 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos x i i + 1 [−600, 600]300
F17Penalized f 12 ( x ) = π n 10 sin 2 π y i + i = 1 n 1 y i 1 2 1 + 10 sin 2 π y i + 1 + y i 1 2 + i = 1 n u x i , 10 , 100 , 4 y i = 1 + x i + 1 4 u x i . a , k , m = k x i a m ; x i > a 0 ; a < x i < a k x i a m ; x i < a [−50, 50]500
F18Penalized2 f 13 ( x ) = 0.1 sin 2 3 π x 1 + i = 1 n 1 x i 1 2 1 + sin 2 3 π x i 1 + x n 1 2 1 + sin 2 2 π x n + i = 1 n u x i , 10 , 100 , 4 u x i . a , k , m = k x i a m ; x i > a 0 ; a < x i < a k x i a m ; x i < a [−50, 50]300
F19Goldstein-Price f 16 ( x ) = 1 + x 1 + x 2 + 1 2 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 30 + 2 x 1 3 x 2 2 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 [−2, 2]23
F20Levi f 19 ( x ) = sin 2 3 π x + x 1 2 1 + sin 2 3 π y + y 1 2 1 + sin 2 2 π y [−10, 10]20
Table 3. Simulation results of unimodal functions.
Table 3. Simulation results of unimodal functions.
AlgorithmMetricETLBOTLBOTLBO(S)PSODEGA
F1Best1.41 × 10−1341.11 × 10−117.31 × 10−201.44 × 10014.61 × 10004.07 × 1002
Worst3.35 × 10−1291.47 × 10−112.81 × 10−111.59 × 10026.83 × 10009.34 × 1002
Median1.92 × 10−1321.40 × 10−111.72 × 10−141.59 × 10015.73 × 10005.59 × 1002
Average1.19 × 10−1301.32 × 10−111.16 × 10−122.79 × 10015.72 × 10005.65 × 1002
SD5.04 × 10−1308.61 × 10−134.21 × 10−123.36 × 10015.11 × 10−019.93 × 1001
F2Best0.00 × 10000.00 × 10000.00 × 10002.70 × 10−230.00 × 10000.00 × 1000
Worst0.00 × 10000.00 × 10000.00 × 10004.46 × 10−150.00 × 10000.00 × 1000
Median0.00 × 10000.00 × 10000.00 × 10001.90 × 10−200.00 × 10000.00 × 1000
Average0.00 × 10000.00 × 10000.00 × 10009.13 × 10−170.00 × 10000.00 × 1000
SD0.00 × 10000.00 × 10000.00 × 10006.23 × 10−100.00 × 10000.00 × 1000
F3Best1.29 × 10−074.80 × 10−017.70 × 10−052.45 × 10−022.50 × 10−021.43 × 10−02
Worst1.75 × 10−046.76 × 10−011.01 × 10−025.40 × 10−028.61 × 10−021.13 × 10−01
Median2.95 × 10−056.17 × 10−011.41 × 10−035.23 × 10−055.20 × 10−023.90 × 10−02
Average4.40 × 10−056.12 × 10−012.15 × 10−038.61 × 10−025.54 × 10−024.22 × 10−02
SD3.85 × 10−053.66 × 10−021.86 × 10−031.30 × 10−021.48 × 10−021.84 × 10−02
F4Best0.00 × 10004.72 × 10−091.00 × 10−162.19 × 10−011.00 × 10011.18 × 1001
Worst0.00 × 10006.11 × 10−092.62 × 10−131.56 × 10001.61 × 10013.07 × 1001
Median0.00 × 10005.37 × 10−091.60 × 10−147.17 × 10−011.38 × 10011.88 × 1001
Average0.00 × 10005.35 × 10−093.57 × 10−148.01 × 10−011.35 × 10011.97 × 1001
SD0.00 × 10003.27 × 10−105.11 × 10−143.82 × 10−011.38 × 10003.76 × 1000
F5Best0.00 × 10002.00 × 10008.37 × 10−080.00 × 10000.00 × 10002.07 × 10−03
Worst2.15 × 10−042.00 × 10002.30 × 10−010.00 × 10000.00 × 10001.91 × 1000
Median3.74 × 10−162.00 × 10001.64 × 10−020.00 × 10000.00 × 10006.00 × 10−01
Average8.05 × 10−062.00 × 10002.30 × 10−020.00 × 10000.00 × 10009.45 × 10−01
SD3.00 × 10−054.35 × 10−094.97 × 10−020.00 × 10000.00 × 10008.29 × 10−01
F6Best0.00 × 10009.76 × 10−031.31 × 10−085.01 × 10−061.72 × 10−031.39 × 10−02
Worst0.00 × 10003.79 × 10−011.19 × 10−051.36 × 10−013.69 × 10−039.86 × 10−02
Median0.00 × 10001.35 × 10−016.33 × 10−074.43 × 10−042.57 × 10−033.73 × 10−02
Average0.00 × 10001.60 × 10−012.01 × 10−064.13 × 10−032.69 × 10−034.04 × 10−02
SD0.00 × 10008.19 × 10−022.71 × 10−061.81 × 10−024.34 × 10−041.72 × 10−02
F7Best2.69 × 10−1702.50 × 10−191.93 × 10−111.27 × 10037.34 × 10045.14 × 1003
Worst1.65 × 10−1552.12 × 10−108.55 × 10−219.14 × 10031.25 × 10052.20 × 1004
Median4.25 × 10−1632.13 × 10−141.78 × 10−113.99 × 10031.08 × 10059.40 × 1003
Average3.29 × 10−1576.85 × 10−129.30 × 10−114.43 × 10031.03 × 10051.05 × 1004
SD2.17 × 10−1563.24 × 10−119.32 × 10−111.79 × 10031.24 × 10043.66 × 1003
F8Best−9.99 × 10−01−3.07 × 10−010.00 × 1000−9.99 × 10−01−1.00 × 1000−1.00 × 1000
Worst−4.71 × 10−07−2.20 × 10−010.00 × 1000−2.46 × 10−08−1.00 × 1000−4.27 × 10−01
Median−8.43 × 10−01−2.48 × 10−010.00 × 1000−9.86 × 10−06−1.00 × 1000−1.00 × 1000
Average−4.79 × 10−01−2.51 × 10−010.00 × 1000−1.83 × 10−01−1.00 × 1000−9.88 × 10−01
SD4.80 × 10−011.69 × 10−020.00 × 10003.77 × 10−010.00 × 10008.26 × 10−02
F9Best0.00 × 10004.15 × 10−014.86 × 10−130.00 × 10000.00 × 10000.00 × 1000
Worst0.00 × 10004.22 × 10−011.10 × 10−070.00 × 10000.00 × 10004.80 × 10−01
Median0.00 × 10006.08 × 10−011.45 × 10−090.00 × 10000.00 × 10000.00 × 1000
Average0.00 × 10004.13 × 10−017.18 × 10−090.00 × 10000.00 × 10009.38 × 10−03
SD0.00 × 10002.82 × 10−021.80 × 10−080.00 × 10000.00 × 10006.72 × 10−02
F10Best0.00 × 10001.64 × 10−108.19 × 10−223.09 × 10−1353.95 × 10−1011.24 × 10−52
Worst0.00 × 10001.08 × 10−033.72 × 10−138.89 × 10−1091.36 × 10−931.18 × 10−04
Median0.00 × 10003.20 × 10−049.11 × 10−151.23 × 10−1181.24 × 10−978.97 × 10−11
Average0.00 × 10003.75 × 10−043.75 × 10−142.00 × 10−1104.04 × 10−955.00 × 10−06
SD0.00 × 10003.01 × 10−046.93 × 10−141.27 × 10−1092.05 × 10−941.95 × 10−05
F11Best0.00 × 10003.83 × 10−097.28 × 10−193.70 × 10−2454.13 × 10−1013.20 × 10−78
Worst0.00 × 10009.47 × 10−043.08 × 10−125.38 × 10−2321.40 × 10−938.07 × 10−06
Median0.00 × 10007.89 × 10−054.01 × 10−142.71 × 10−2401.26 × 10−971.37 × 10−21
Average0.00 × 10002.17 × 10−043.06 × 10−131.13 × 10−2334.01 × 10−951.67 × 10−07
SD0.00 × 10002.51 × 10−046.46 × 10−130.00 × 10002.02 × 10−941.17 × 10−06
F12Best0.00 × 10009.30 × 10−080.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Worst0.00 × 10003.44 × 10−012.01 × 10−130.00 × 10000.00 × 10002.99 × 10−03
Median0.00 × 10001.07 × 10−016.55 × 10−150.00 × 10000.00 × 10000.00 × 1000
Average0.00 × 10001.19 × 10−012.01 × 10−140.00 × 10000.00 × 10001.44 × 10−04
SD0.00 × 10008.05 × 10−023.83 × 10−140.00 × 10000.00 × 10005.66 × 10−04
Table 4. Simulation results of multimodal functions.
Table 4. Simulation results of multimodal functions.
FunctionStatisticETLBOTLBOTLBO(S)PSODEGA
F13Best0.00 × 10008.10 × 10−052.58 × 10−043.53 × 10−151.91 × 10−040.00 × 1000
Worst0.00 × 10005.01 × 10−025.46 × 10−022.22 × 10−032.35 × 10−022.40 × 10−05
Median0.00 × 10008.70 × 10−036.52 × 10−031.27 × 10−063.67 × 10−030.00 × 1000
Average0.00 × 10001.16 × 10−021.37 × 10−021.11 × 10−045.68 × 10−031.45 × 10−06
SD0.00 × 10001.05 × 10−021.30 × 10−023.62 × 10−045.50 × 10−034.87 × 10−06
F14Best3.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−01
Worst3.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−01
Median3.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−01
Average3.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−01
SD4.84 × 10−089.98 × 10−094.02 × 10−078.54 × 10−073.34 × 10−108.46 × 10−08
F15Best8.83 × 10−162.04 × 10013.49 × 10−051.51 × 10−123.68 × 10−034.11 × 1000
Worst8.83 × 10−162.00 × 10013.63 × 10003.19 × 10001.01 × 10−029.22 × 1000
Median8.83 × 10−162.04 × 10014.30 × 10−031.15 × 10006.10 × 10−037.36 × 1000
Average8.83 × 10−162.00 × 10011.51 × 10001.00 × 10006.34 × 10−037.13 × 1000
SD0.00 × 10002.14 × 10−011.72 × 10008.29 × 10−011.19 × 10−021.14 × 1000
F16Best0.00 × 10001.00 × 10003.92 × 10−113.22 × 10−023.50 × 10003.08 × 1000
Worst0.00 × 10001.03 × 10002.64 × 10−072.23 × 10005.89 × 10002.06 × 1001
Median0.00 × 10001.02 × 10006.76 × 10−096.09 × 10−014.33 × 10009.51 × 1000
Average0.00 × 10001.02 × 10002.22 × 10−086.25 × 10−014.37 × 10001.05 × 1001
SD0.00 × 10006.00 × 10−034.38 × 10−084.26 × 10−015.12 × 10−013.96 × 1000
F17Best9.31 × 10−012.00 × 10−011.43 × 10−092.31 × 10−011.09 × 10005.49 × 1000
Worst1.34 × 10007.92 × 10−011.77 × 10−051.20 × 10012.91 × 10006.70 × 1002
Median1.06 × 10005.14 × 10−018.22 × 10−073.15 × 10001.68 × 10001.25 × 1001
Average1.11 × 10005.23 × 10−012.16 × 10−063.61 × 10001.73 × 10003.54 × 1001
SD1.00 × 10−011.35 × 10−013.23 × 10−062.30 × 10004.33 × 10−011.08 × 1002
F18Best2.59 × 10004.46 × 10017.64 × 10−113.44 × 10014.22 × 10011.91 × 1001
Worst2.98 × 10001.16 × 10011.62 × 10−051.17 × 10021.03 × 10021.24 × 1004
Median2.98 × 10002.03 × 10011.51 × 10−075.93 × 10015.71 × 10015.52 × 1001
Average2.75 × 10001.23 × 10011.53 × 10−066.09 × 10015.77 × 10019.39 × 1002
SD4.42 × 10−023.89 × 10003.48 × 10−061.86 × 10011.12 × 10012.61 × 1003
F19Best3.00 × 10003.00 × 10003.28 × 10013.00 × 10003.00 × 10003.00 × 1000
Worst3.00 × 10001.66 × 10013.27 × 10013.00 × 10003.00 × 10003.00 × 1000
Median3.00 × 10003.00 × 10003.30 × 10013.00 × 10003.00 × 10003.00 × 1000
Average3.00 × 10003.30 × 10003.24 × 10013.00 × 10003.00 × 10003.00 × 1000
SD1.41 × 10−141.95 × 10007.04 × 10−081.24 × 10−034.24 × 10−083.10 × 10−05
F20Best3.02 × 10−011.67 × 10−031.24 × 10−031.33 × 10−311.30 × 10−311.35 × 10−31
Worst3.02 × 10−011.14 × 10011.00 × 10001.38 × 10−311.31 × 10−312.26 × 10−07
Median3.02 × 10−011.52 × 10002.03 × 10−011.33 × 10−311.36 × 10−311.35 × 10−31
Average3.02 × 10−012.25 × 10005.28 × 10−011.32 × 10−311.40 × 10−318.01 × 10−06
SD4.20 × 10−052.43 × 10004.86 × 10−011.82 × 10−401.80 × 10−461.26 × 10−06
Table 5. Wilcoxon rank-sum test for benchmark function problems.
Table 5. Wilcoxon rank-sum test for benchmark function problems.
FunctionTLBOTLBO(S)PSODEGA
w E T L B O p Win w E T L B O p Win w E T L B O p Win w E T L B O p Win w E T L B O p Win
F112753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F225255.00 × 10−01=25255.00 × 10−01=12753.42 × 10−1825255.00 × 10−01=25255.00 × 10−01=
F312753.42 × 10−1813421.74 × 10−1612753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F412753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F512753.42 × 10−1813994.16 × 10−1537751.00 × 1000×37751.00 × 1000×12753.42 × 10−18
F612753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F712753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F812753.42 × 10−1812753.42 × 10−1835801.00 × 1000×37751.00 × 1000×37311.00 × 1000×
F912753.42 × 10−1812753.42 × 10−1825255.00 × 10−01=25255.00 × 10−01=18531.80 × 10−06
F1012753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F1112753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F1212753.42 × 10−1813994.16 × 10−1525255.00 × 10−01=25255.00 × 10−01=13001.52 × 10−17
F1312753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1813421.74 × 10−16
F1421504.86 × 10−0325255.00 × 10−01=35801.00 × 1000×35801.00 × 1000×25255.00 × 10−01=
F1512753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F1612753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F1737751.00 × 1000×37751.00 × 1000×13001.52 × 10−1713994.16 × 10−1512753.42 × 10−18
F1812753.42 × 10−1837751.00 × 1000×12753.42 × 10−1812753.42 × 10−1812753.42 × 10−18
F1914383.35 × 10−1412753.42 × 10−1818531.80 × 10−0637311.00 × 1000×37691.00 × 1000×
F2012753.42 × 10−1812753.42 × 10−1837751.00 × 1000×37751.00 × 1000×37751.00 × 1000×
Table 6. Simulation results for MRPP problems.
Table 6. Simulation results for MRPP problems.
InstanceAlgorithmBestAverageWorst
LengthDeviation (%)LengthDeviation (%)LengthDeviation (%)
20 × 20ACO(L)156.072.73%160.165.42%163.647.71%
ACO(G)154.851.93%158.784.52%161.926.58%
ETLBO151.920.00%156.643.10%160.715.78%
30 × 30ACO(L)232.633.13%240.066.43%245.568.87%
ACO(G)231.422.60%238.565.76%244.358.33%
ETLBO225.560.00%233.783.64%240.216.49%
40 × 40ACO(L)302.994.02%313.527.64%322.9910.89%
ACO(G)301.273.43%309.566.28%318.859.47%
ETLBO291.270.00%303.874.33%312.497.28%
50 × 50ACO(L)400.425.73%412.078.81%421.1311.20%
ACO(G)395.774.51%409.778.20%418.7010.56%
ETLBO378.700.00%398.275.17%410.428.37%
Table 7. Wilcoxon rank-sum test for MRPP problems.
Table 7. Wilcoxon rank-sum test for MRPP problems.
InstanceACO(L)ACO(G)
w E T L B O p Win w E T L B O p Win
20 × 2015905.75 × 10−1116313.56 × 10−10
30 × 3013994.16 × 10−1515752.89 × 10−11
40 × 4013001.52 × 10−1714383.35 × 10−14
50 × 5012753.42 × 10−1813421.74 × 10−16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, S.; Liu, D.; Li, D.; Shao, X. Enhanced Teaching–Learning-Based Optimization Algorithm for the Mobile Robot Path Planning Problem. Appl. Sci. 2023, 13, 2291. https://doi.org/10.3390/app13042291

AMA Style

Lu S, Liu D, Li D, Shao X. Enhanced Teaching–Learning-Based Optimization Algorithm for the Mobile Robot Path Planning Problem. Applied Sciences. 2023; 13(4):2291. https://doi.org/10.3390/app13042291

Chicago/Turabian Style

Lu, Shichang, Danyang Liu, Dan Li, and Xulun Shao. 2023. "Enhanced Teaching–Learning-Based Optimization Algorithm for the Mobile Robot Path Planning Problem" Applied Sciences 13, no. 4: 2291. https://doi.org/10.3390/app13042291

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop