Next Article in Journal
Open-Circuit Fault Diagnosis for T-Type Three-Level Inverter via Improved Adaptive Threshold Sliding Mode Observer
Previous Article in Journal
Beyond Traceability: Decentralised Identity and Digital Twins for Verifiable Product Identity in Agri-Food Supply Chains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UAV Path Planning Using a State Transition Simulated Annealing Algorithm Based on Integrated Destruction Operators and Backward Learning Strategies

School of Electrical and Power Engineering, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(11), 6064; https://doi.org/10.3390/app15116064
Submission received: 22 March 2025 / Revised: 3 May 2025 / Accepted: 8 May 2025 / Published: 28 May 2025

Abstract

:
This study introduces a state transition simulated annealing algorithm that incorporates integrated destruction operators and backward learning strategies (DRSTASA) to address complex challenges in UAV path planning within multidimensional environments. UAV path planning is a critical optimization problem that requires smooth flight paths, obstacle avoidance, moderate angle changes, and minimized flight distance to conserve fuel and reduce travel time. Traditional algorithms often become trapped in local optima, preventing them from finding globally optimal solutions. DRSTASA improves global search capabilities by initializing the population with Latin hypercube sampling, combined with destruction operators and backward learning strategies. Testing on 23 benchmark functions demonstrates that the algorithm outperforms both traditional and advanced metaheuristic algorithms in solving single and multimodal problems. Furthermore, in eight engineering design optimization scenarios, DRSTASA exhibits superior performance compared to the STASA and SNS algorithms, highlighting the significant advantages of this method. DRSTASA is also successfully applied to UAV path planning, identifying optimal paths and proving the practical value of the algorithm.

1. Introduction

Unmanned aerial vehicles (UAVs) are widely used in various fields such as military operations, agriculture, disaster relief, environmental monitoring, and communications due to their low cost, high maneuverability, and adaptability [1,2,3]. However, path planning in complex and dynamic environments remains a significant challenge, as it must ensure obstacle avoidance, flight safety, and energy efficiency. Finding the optimal path has become a key issue in UAV applications [4].
Traditional path planning methods such as the A* algorithm [5,6], ant colony optimization (ACO) [7], and particle swarm optimization (PSO) [8] have been widely used but often struggle to balance real-time performance, global optimality, and obstacle avoidance. Therefore, researchers have worked to refine or integrate these methods with new approaches to improve their performance.
For instance, Jie Chen, Fang Ye, and Tao Jiang proposed a UAV trajectory planning strategy that leverages the rapid optimization capabilities of ant colony optimization (ACO), though it did not account for complex factors like flight angles [9]. Luji Guo, Chenbo Zhao, and Jiacheng Li optimized the cost function weights of the A* algorithm and introduced virtual target points in the artificial potential field (APF) method, improving both path smoothness and obstacle avoidance [10]. Hongbo Xiang, Xiaobo Liu, and their team combined enhanced particle swarm optimization (EPSO) with genetic algorithms, adaptively tuning the acceleration coefficients of EPSO based on fitness values, which enhanced global search [11]. Dongcheng Li, Wangping Yin, and W. Eric Wong applied Q-learning to dynamically adjust factors such as step size and cost function in the A* algorithm, creating a hybrid method that effectively balances global and local search [12]. Bo Li, Xiaogang Qi, and Baoguo Yu incorporated the Metropolis criterion from simulated annealing into ACO, reducing the likelihood of getting trapped in local optima. They also employed an inscribed circle smoothing method to optimize trajectories, enhancing both feasibility and effectiveness [13].
In recent years, the rapid development of artificial intelligence and intelligent optimization algorithms has led to the increasing application of advanced algorithms and deep learning methods in UAV path planning. Hui Li, Teng Long, and Guangtong Xu proposed a coupling degree-based heuristic priority planning method (CDH-PP), which enhances the efficiency of cooperative path planning for UAV swarms through coupling degree heuristics [14]. Ronglei Xie, Zhijun Meng, and Lifeng Wang addressed path planning challenges in complex dynamic environments using deep learning, introducing the RQ method and an adaptive sampling mechanism to improve obstacle avoidance capabilities [15]. Liguo Tan, Yaohua Zhang, and Jianwen Huo combined the rapidly-exploring random tree (RRT) algorithm with driver visual behavior to propose an RRT path planning method that simulates driver vision. They optimized the path using a greedy algorithm; however, while this method is effective in complex scenarios, it exhibits high computational complexity [16]. Xing Wang, Jeng-Shyang Pan, and Qingyong Yang introduced an improved mayfly algorithm (modMA), which optimizes UAV layout and reduces overall costs through exponentially decreasing inertia weights (EDIW), adaptive Cauchy mutations, and enhanced crossover operators [17]. Jingfan Tian, Yankai Wang, and Dongdong Yuan improved the global optimization capability of the elastic rope algorithm by enhancing the node update mechanism and introducing paths composed of m-dimensional nodes to increase its applicability [18].
Additionally, other studies have proposed innovative strategies for algorithm improvement. Xiaohui Cui, Yu Wang, and Shijie Yang effectively addressed two-point and multi-point path planning problems by employing a chaotic initialization method combined with genetic algorithms (GA) and simulated annealing (SA) [19]. Xiaobing Yu, Chenliang Li, and Jiafang Zhou introduced an adaptive selection mutation-constrained differential evolution algorithm, which shows promising application prospects in disaster scenarios [20]. Xiangyin Zhang, Shuang Xia, and Xiuzhi Li developed a quantum theory-improved firefly algorithm (QFOA), utilizing wave functions to replace random search, thereby enhancing population diversity and outperforming the traditional firefly algorithm (FQA) in terms of search capability, stability, and robustness [21]. The combination of simplified grey wolf optimization (SGWO) and the mutual symbiotic organism search (MSOS) algorithm has also demonstrated potential in UAV path planning [22]. To improve path planning efficiency and ensure the smoothness and safety of UAV operations, Yaqing Chen, Qizhou Yu, and Dan Han proposed a hybrid algorithm that integrates grey wolf optimization (GWO) with the artificial potential field method (APF), significantly enhancing path planning capabilities, particularly in scenarios with short distances and high safety requirements [23]. Despite the widespread application of intelligent algorithms and neural networks in UAV path planning, most algorithms remain limited to two-dimensional path planning, failing to adequately consider factors such as obstacle height and flight slope, while some methods exhibit high computational complexity.
The state transition simulated annealing algorithm (STASA), a hybrid intelligent optimization algorithm proposed by Han et al. [24]. It combines the strengths of the state transition algorithm (STA) [25] and simulated annealing (SA) [26], achieving significant results in various application areas, such as PM2.5 prediction [27] and the optimization of methane production conditions in coal-to-gas processes [28]. Compared to traditional algorithms such as genetic algorithm (GA) [29], particle swarm optimization (PSO) [30], and grey wolf optimizer (GWO) [31], STASA demonstrates greater efficiency and flexibility. STASA offers greater efficiency and flexibility. However, its performance declines when addressing complex, high-dimensional problems.
To address the shortcomings of the state transition simulated annealing algorithm (STASA), this paper introduces the state transition simulated annealing algorithm with destructive perturbation operators and reverse learning strategies (DRSTASA). This algorithm enhances population diversity using Latin hypercube sampling and incorporates destructive perturbation and reverse learning strategies to improve global search, speed up convergence, and avoid local optima. Experimental results show that DRSTASA outperforms other algorithms on 23 benchmark test functions and 8 engineering optimization problems. It has also been successfully applied to three-dimensional UAV path planning, demonstrating strong practical feasibility [32].
The rest of this paper is structured as follows: Section 2 covers the objective function for UAV path planning, Section 3 explains the original STASA algorithm, Section 4 describes the enhancements made by DRSTASA, and Section 5 presents a comparison with other algorithms and experimental results. Finally, a summary is provided.

2. UAV Three-Dimensional Path Planning

With the rapid advancement of unmanned aerial vehicle (UAV) technology, these systems have found widespread applications in military operations, agriculture, logistics, and environmental monitoring. To ensure that UAVs can safely and effectively execute tasks in complex environments, path planning has emerged as a critical issue. The objective of UAV path planning is to identify an optimal route in three-dimensional space that enables the UAV to travel from a starting point to a destination while avoiding obstacles and adhering to safety and energy consumption requirements. The UAV three-dimensional path planning problem aims to find an optimal trajectory that satisfies the following objectives:
  • The Shortest Path: The total flight distance from the starting point to the endpoint should be minimized.
  • Obstacle Avoidance: The flight path must navigate around all obstacles to ensure safety.
  • Flight Altitude: The flight path must comply with the specified altitude restrictions.
  • Path Smoothness: The trajectory should be as smooth as possible, avoiding sharp turns and steep inclines.
Therefore, this problem primarily considers four key factors. The specific cost function and objective function for UAV path planning are defined in the following subsection.

2.1. Path Optimality

Path planning must achieve optimality under specific criteria to ensure the efficient operation of UAVs. For applications such as aerial photography, surveying, and surface inspection, minimizing the path length is a primary objective.
The cost function for path length F 1 can be expressed as follows:
F ( X i ) = j = 1 n 1 P l j P l , j + 1
X i represents the list of n waypoints that the flight path needs to traverse, with the coordinates of each waypoint denoted as P i j = X i j , Y i j , Z i j .

2.2. Safety and Feasibility Constraints

In path planning, ensuring the safe operation of the UAV is crucial. This involves avoiding obstacles in the environment while maintaining flight within a specified altitude range. Furthermore, the flight path should be as smooth as possible to prevent abrupt ascents and descents. Consequently, the remaining cost functions include threat cost, flight altitude cost, and path smoothness cost.
Threat Cost:
F 2 ( X i ) = j = 1 n 1 K = 1 K T k ( P ι j P ι , j + 1 )
where
T k ( P i j P i , j + 1 ) 0 i f   d k > S + D + R k ( S + D + R k ) d k i f   D + R k < d k S + D + R k i f   d k D + R k
In the problem, it is assumed that the set K contains several cylindrical obstacles, each defined by a center coordinate C k and a radius R k . The diameter of the UAV is represented by D. The vertical distance from adjacent path nodes to the origin is denoted as d k , and S represents the danger zone of the obstacles. The experimental parameters were set as k = 6, R k = 80, P start = [200,100,150], and P end = [800;800;150]. The obstacle layout is illustrated in Figure 1.
In practice, there are usually specific altitude requirements for UAVs, which stipulate that the UAV must operate within a defined range between a minimum and maximum height. The cost associated with flight altitude can be calculated using the following formula:
H i j = h i j h m a x + h m i n 2 i f   h m i n h i j h m a x o t h e r w i s e
The total altitude cost F 3 can be calculated using the following formula:
F 3 ( X i ) = j = 1 n H i j
where h i j represents the current altitude of the UAV, and h m a x and h m i n denote the maximum and minimum allowable flight altitudes, respectively. The experimental parameters were configured as h max = 200; h min = 100.
Smoothness Cost:
The smoothness cost evaluates the turning and ascent angles of the path. The turning angle ϕ i j and the ascent angle ψ i j are calculated using the following formulas:
Turning Angle:
ϕ i j = arctan P l J P l , j + 1 × P l , j + 1 P l , j + 2 P l J P l , j + 1 P l , j + 1 P l , j + 2
Ascent Angle:
ψ i j = arctan z i , j + 1 z i j P l j P l , j + 1
The total smoothness cost can be calculated using the following formula:
F 4 ( X i ) = a 1 j = 1 n 2 ϕ i j + a 2 j = 1 n 1 | ψ i j ψ i , j 1 |
a 1 and a 2 are the penalty coefficients for the horizontal turning angle and the vertical pitch angle, respectively. The experimental parameters were set as a 1 = 1; a 2 = 1.

3. STASA

The state transition simulated annealing algorithm (STASA) integrates discrete and continuous state transition operators derived from the state transition algorithm (STA), thereby improving its search capabilities. In both STA and STASA, the solution to an optimization problem is represented as a “state”, and the process of refining the solution is analogous to a “state transition”. This refinement occurs through local and global searches using update operators, followed by the application of a criterion to determine whether the solution has improved.
The general framework for generating candidate solutions using the state transition algorithm is as follows:
x k + 1 = A k x k + B k u k y k + 1 = f ( x k + 1 )
where x k , x k + 1 represent the current and the new generation solution, respectively, u k is a function that relates the current state to historical states, and A k and B k are matrices for random transformations. f ( ) denotes the fitness function for the problem being solved, and y k represents the fitness value of x k .

3.1. State Transition Operators

In continuous optimization problems, STASA employs four types of state transition operators: rotation, translation, scaling, and axis transformation. Each operator plays a distinct role in improving the search process.
(a)
Rotation Operator
x k + 1 = x k + ε α 1 n x k 2 R r x k
where ε α is a positive constant known as the rotation factor, and R r R n × n is an n-dimensional random matrix with elements uniformly distributed in the range of [−1, 1]. Additionally, 2 denotes the Euclidean norm (L2 norm). The rotation operator allows the generated candidate solutions to fall within a hypersphere of radius ε α .
(b)
Translation Operator
x k + 1 = x k + ε β R t x k x k 1 x k x k 1 2
where ε β is a positive constant known as the scaling factor, and R t is a random variable defined within the range of [0, 1]. The translation operator starts from the initial point x k and searches in the direction from point x k 1 to point x k , with a maximum search length of ε β . This operator possesses local search capabilities.
(c)
Scaling Operator
x k + 1 = x k + ε γ R e x k
where ε γ is a positive constant known as the scaling factor, and R e R n × n is an n-dimensional random diagonal matrix following a Gaussian distribution with a mean of 0 and a variance of 1. The scaling operator has global search capabilities.
(d)
Axis Transformation Operator
x k + 1 = x k + ε δ R α x k
where ε δ is a positive constant known as the axis transformation factor, and R a R n × n is an n-dimensional random diagonal matrix following a Gaussian distribution in the range [0, 1], with only one non-zero element. The axis transformation operator enhances the search capability in a single dimension.

3.2. Update Strategy of the STASA

STASA does not rely on a greedy update criterion. Instead, it uses the Metropolis criterion, similar to the simulated annealing algorithm, to refine solutions. The update process is as follows:
x k = x k + 1 ,     i f   f ( x k + 1 ) < f ( x k ) x k + 1 ,   i f   f ( x k + 1 ) f ( x k )   a n d exp ( ( ( f ( x k + 1 ) f ( x k ) ) / T k ) η x k , i f   f ( x k + 1 ) f ( x k )   a n d exp ( ( ( f ( x k + 1 ) f ( x k ) ) / T k ) η
where η is a random number in the range [0, 1]. When the fitness value of the next-generation solution is less than that of the current solution, the new solution is accepted. When the next-generation solution is not better than the current solution, the simulated annealing temperature is calculated using formula: exp ( ( ( f ( x K + 1 ) f ( x k ) ) / T k ) ) . If the temperature is greater than η, the new solution is accepted; otherwise, the current solution is retained.

4. DRSTASA

4.1. Population Initialization Based on Latin Hypercube Sampling

In intelligent algorithms, the initialization of the population plays a crucial role in determining the performance and convergence speed of the algorithm. A well-initialized population ensures diversity, which facilitates a more efficient exploration of the search space and accelerates convergence towards the global optimal solution. As a result, this improves the algorithm’s stability and efficiency. Traditional approaches, such as STA and STASA, typically rely on random initialization functions. However, this randomness can hinder the uniform distribution of solutions within the search space, ultimately reducing search efficiency.
To overcome this issue, the Latin hypercube sampling (LHS) method [33] is employed. LHS divides the multidimensional parameter space into non-overlapping intervals, randomly selecting samples from each interval to ensure uniform distribution and better coverage of the parameter space. The key steps of LHS involve determining the number of samples, dividing the parameter range for each dimension into NNN intervals, selecting values from each interval for every dimension, mapping these values to the desired distribution, and shuffling the order to maintain randomness.
Figure 2 illustrates a comparison between the population distributions generated using random initialization and Latin hypercube sampling (LHS) for two- and three-dimensional populations. The comparison demonstrates that unlike random initialization, LHS uniformly partitions each dimension and independently selects sample points from each partition, ensuring an even distribution of points throughout the sample space while maintaining uniqueness across all dimensions. This approach enhances initial population coverage, effectively mitigates the “curse of dimensionality” commonly encountered in high-dimensional spaces with conventional methods, and increases initial population diversity. As a result, it significantly improves the algorithm’s global exploration capability within the solution space.

4.2. Disruption Operator

Inspired by the gravitational destruction phenomenon observed in astrophysics, the interference operator (DO) was first introduced in the gravitational search algorithm (GSA) [34,35]. Its main objective is to improve exploration and exploitation capabilities that will prevent premature convergence and maintain population diversity. Over time, DO has been widely used in heuristic algorithms, such as elite particle swarm optimization algorithms [36].
The DO simulates the disruption of smaller-mass particles as they approach a larger-mass object. Here, the current optimal solution is treated as the large-mass object, and the remaining solutions are considered smaller-mass particles. When specific conditions are met, disruptive perturbations are applied to these solutions. Formula (15) describes the activation of the DO, which occurs when the Euclidean distance between neighboring solutions exceeds a predefined threshold C. The distance between a candidate solution and the current optimal solution is used to determine the degree of disruption.
R i , j R i , best C
The algorithm activates the disruption operator to update the current solution only when the condition in Formula (15) is satisfied, where R i , j = X i X j represents the Euclidean distance between neighboring candidate solutions i and j, and R i , b e s t = X i X b e s t is the Euclidean distance between the particle and the current optimal solution. C is a preset threshold.
To improve the efficacy of the DO, the threshold C is dynamically adjusted throughout the algorithm’s iterations. A larger initial C value facilitates broad exploration early in the process, while a gradually decreasing C encourages convergence during the later stages. Formula (16) details this dynamic threshold adjustment.
C = C 0 ( 1 i t e r / T )
C 0 is the manually defined initial threshold, typically set between 1 and 3. Iter represents the current iteration count, and T denotes the total number of iterations set for the algorithm. It is important to note that, aside from the current optimal solution, all other solutions will be evaluated according to Formula (16). Only the solutions that meet the criteria of this formula will activate the disruption operator, which is defined as follows:
D = r a n d ( R i , j 2 , R i , j 2 ) i f   R i , b e s t 1 R i , j + r a n d ( R i , j 2 , R i , j 2 ) o t h e r w i s e
Here, r a n d ( R i , j 2 , R i , j 2 ) is a randomly distributed number, with its range defined based on the Euclidean distance between neighboring candidate solutions i and j. When R i , b e s t 1 , it indicates that the solution is far from the optimal solution, and the disruption operator explores a broader range. In other cases, when the distance is closer, the disruption operator conducts a more localized search around the current solution.
Finally, for solutions that meet the conditions, a disruptive perturbation is applied, and the solution is updated according to the following formula:
x i , j t + 1 = t T x i , j t + 1 t T x i , j t D
In Formula (18), x i , j t + 1 represents the solution after applying the disruptive perturbation. The first part of the formula retains some information from the current solution, while the second part incorporates the influence of the disruption operator. The impact of the disruption operator gradually diminishes as the algorithm progresses, allowing for broader exploration of the search space in the early stages to avoid local optima, and accelerating convergence in the later stages.
In STASA, the disruption operator is executed after the application of the four update operators. This sequence allows the algorithm to conduct extensive exploration during the early stages and prevent falling into local optima in later stages, ultimately increasing the chances of finding the global optimal solution.

4.3. Reverse Learning Strategy

To enhance optimization performance and accelerate convergence, a dynamic reverse learning strategy is introduced. Reverse learning compares the current solution X = [ x 1 , x 2 , x D ] with its reverse point X ¯ = [ x 1 ¯ , x 2 ¯ x D ¯ ] , selecting the better option to improve the algorithm’s search capability and exploration range [37]. Here, x [ u b , l b ] , u b and l b represent the upper and lower bounds of the variables, respectively. The formula for calculating the reverse point is as follows:
x i , j ¯ = l b + u b x i , j
Reverse learning expands the search space of the algorithm, improving both convergence accuracy and speed. To enhance flexibility, a dynamic reverse learning strategy is employed, where the upper and lower bounds of the reverse point change with each iteration. The formula for calculating the dynamic reverse point is as follows:
x ι , J t ¯ = r 5 ( a ¯ ι ( t ) + b ¯ ι ( t ) x i , j t
In the equation, x i , j t represents the position of the j-th solution in the i-th dimension during the t-th iteration.
a ¯ t ( t ) and b ¯ ι ( t ) are the minimum and maximum values of the current population in the i-th dimension, respectively, and r 5 is a random number in the range [0, 1]. The decision to perform reverse learning is made by comparing with the control parameter p , and the specific form is as follows:
x i , j t + 1 = ( 1 r 6 ) x i , j + t r 6 x i , j t ¯   i f   r a n d p
where t represents the number of iterations, x i , j t is the current solution, and x i , j t ¯ is the corresponding reverse solution. Both r 6 and rand are random numbers in the range [0, 1], and x i , j t + 1 is the population in the t+1th iteration. In the subsequent experiments, the control parameter p is set to 0.5. The opposition-based learning strategy activates when meeting predefined stochastic conditions, utilizing opposition solutions to expand search space diversity and balance exploration versus exploitation capabilities. Through fitness comparison between original and opposition solutions, this approach accelerates convergence while enhancing solution quality. Moreover, the opposition mechanism assists in escaping premature stagnation, thereby markedly improving algorithm robustness and optimization efficiency.

4.4. Basic Process of DRSTASA

The overall structure and pseudocode for DRSTASA are presented in Figure 3 and Algorithm 1. The algorithm begins by initializing the population using the Latin hypercube sampling method and evaluating the initial solutions. The best population members and optimal objective values are selected. During each iteration, four transformation operators are applied first to update the population, followed by the Metropolis criterion from simulated annealing to determine whether to replace the current solution. After assessing the distance between current solutions, the disruption operator is applied as needed, and the reverse learning strategy is probabilistically executed to further refine the search process.
Algorithm 1: The Proposed Algorithm DRSTASA
1: Set the initial parameters
2: LHS generate the initial population u, Evaluate the fitness value: f, the current optimal solution: Best
3: while the set stop temperature value is not reached
4: Update the xk through four transformation operators by Equations (2)–(5) and obtain xk+1.
5: Calculate the fitness value of the current solution and new solution: f(xk), f(xx+1).
6: Update Best and fBest with Metropolis Criteria by Equation (6).
7: Calculate threshold C using Equation (16).
8: Judging by Equation (15), the condition is updated using Equation (17).
9: if rand > p
   Use Equation (21) to update.
10: T = α*T, k = k + 1 and return to step 4.

5. Experimental Testing and Analysis

To evaluate the performance of DRSTASA, a total of 23 standard benchmark functions were selected for experimentation. The results were compared with those of nine other recent algorithms, including STA, STASA, AOA, SCA, AHA, FDA, AVOA, GWO, and the improved ASTASA, to analyze the testing performance of each algorithm. The experiments were conducted on a standard PC running Windows 10 64-bit, equipped with a 12th Gen Intel® Core™ i5-12490F processor operating at 3.00 GHz and 16 GB of RAM. The implementation was based on MATLAB 2020a. For all algorithms, the population size and search intensity were set to 50, with parameter configurations either consistent with the original publications or based on classical settings, as detailed in Table 1.

5.1. Benchmark Test Functions

In this section, 23 fundamental benchmark functions are used to evaluate DRSTASA and eight comparison algorithms. The test functions cover both unimodal and multimodal problems. Specifically, F 1 ~ F 7 functions are unimodal, and F 8 ~ F 23 are bimodal. The dimensionality of the functions is variable; for these tests, it is set to 30. Detailed definitions can be found in Table 2 and Table 3.
Each algorithm was independently executed 30 times, during which the optimal values, mean values, and standard deviations (std) were recorded. The optimal values of each algorithm were compared: if an algorithm outperformed the comparison algorithm, it was assigned a score of 1; if the results were equal, it received a score of 0; and if it underperformed, it was assigned a score of −1. In cases where the optimal values were identical, the mean values and standard deviations were sequentially compared. Ultimately, the total scores were tallied for each algorithm across all test functions, clearly demonstrating the performance of DRSTASA in relation to other algorithms.
The results of DRSTASA, compared to nine other algorithms, are presented in Table 4. Apart from its weaker performance on the multimodal functions F21, F22, and F23—where the mean values were suboptimal and stability was lacking—DRSTASA exhibited stable performance on the other test functions, with minimal differences between the minimum and mean values. In functions F6, F12, F13, F16, F21, F22, and F23, DRSTASA did not achieve the optimal value; however, it reached the theoretical optimal values in other test functions, demonstrating adequate solution accuracy.
Overall, DRSTASA shows advantages over other algorithms in most tests, with results surpassing those of recent algorithms. Specifically, when compared to STASA and ASTASA, which scored 13 and 11 points, respectively, it is evident that DRSTASA achieved superior results in the majority of test cases. Although DRSTASA successfully escaped local optima in the multimodal test functions F21, F22, and F23, there remains a significant gap when compared to more robust algorithms such as AHA, FDA, and AVOA. While DRSTASA performed exceptionally well in unimodal tests, it slightly underperformed in multimodal tests relative to these stronger algorithms.

5.1.1. Convergence Analysis

Due to space limitations, Figure 4 illustrates the convergence behavior of various algorithms on selected test functions, highlighting the best results from each algorithm. The blue curve represents the convergence trajectory of the DRSTASA algorithm, with the horizontal axis indicating the number of iterations and the vertical axis representing the optimal objective function value. In the F3 and F4 tests, although DRSTASA achieved the theoretical optimal value, the convergence speed was slower compared to other algorithms, especially in unimodal test functions where the AVOA algorithm demonstrated the fastest convergence. In the F7 and F9 tests, DRSTASA also reached the minimum objective function value; however, significant oscillations were observed during the early iterations. This suggests that the algorithm likely employed techniques such as simulated annealing and perturbation to escape local optima and search for the global optimum. In the F12 and F13 tests, while DRSTASA did not surpass the FDA algorithm in finding the optimal value, it still exhibited a faster convergence speed compared to the other algorithms.

5.1.2. Time Complexity Analysis

Time complexity reflects the efficiency of an algorithm, and as the problem size increases, it directly influences both time and resource costs. The time complexity of the proposed algorithm is primarily determined by the population size, the number of iterations, the dimensionality of the problem, and the time required to update the operators. Let N represent the population size, T denote the number of iterations, and D signify the problem dimensionality. Ignoring the time complexity of the operators, the overall time complexity of the algorithm is as follows:
O ( D R S T A S A ) = O ( N D T )
In a single iteration, let the time complexity of evaluating the objective function for a single solution be t 1 , and the time complexity of updating a solution using the four operators be t 2 . Without incorporating any improvement strategies, the time complexity of the algorithm is as follows:
O ( D R S T A S A ) = O ( N D T ( t 1 + t 2 ) )
When introducing destruction operators and opposition-based learning strategies, let the time complexities of the destruction operator and opposition-based learning be denoted as t 3 and t 4 , respectively. Consequently, the time complexity of the algorithm incorporating these improvements becomes the following:
O ( D R S T A S A ) = O ( N D T ( t 1 + t 2 + t 3 + t 4 ) )
From the analysis, it can be concluded that the time complexity of DRSTASA does not increase by an order of magnitude compared to the original STASA. Additionally, the destruction operator and opposition-based learning strategies enhance the efficiency of the algorithm, resulting in faster convergence.
Figure 5 shows the average runtime of nine algorithms across 23 test functions. It is observed that the SCA, AOA, and AHA algorithms have the shortest runtimes, but they exhibit slightly lower accuracy compared to the others. AVOA and GWO fall into the second tier, characterized by higher time costs but with improved accuracy. STA, STASA, and DRSTASA are in the third tier, where DRSTASA has a slightly longer runtime than STASA but achieves significantly higher accuracy, making the additional time acceptable. The FDA requires the longest runtime but offers greater stability, albeit at a considerable time.
To evaluate the effectiveness and stability of DRSTASA, eight engineering design optimization problems were selected for testing. The DRSTASA, STASA, and social network search (SNS) algorithms were employed to solve each problem, with 30 independent runs and 2000 iterations per run. The performance of each algorithm was assessed by comparing the optimal values, mean values, and standard deviations. Table 5 provides details of the eight engineering problems, while Table 6 presents the comprehensive results of each algorithm.
From Table 6, it is clear that the performance of DRSTASA in the string design problem is not as strong as that of the other two algorithms; however, the algorithm excels in the remaining engineering challenges. In the cantilever beam, tubular column, and welded beam design problems, although DRSTASA achieves optimal values, the average performance and standard deviation still show a gap compared to the other algorithms, suggesting that further improvements in stability are needed. The comparison results indicate that DRSTASA demonstrates significant improvements over STASA in various engineering problems and surpasses the SNS algorithm, highlighting the potential of this approach for solving real-world engineering challenges.

5.2. Experimental Results

By considering path optimality, safety, and feasibility constraints, specific weights are assigned to each cost function to determine the overall objective function value for the problem.
F ( X i ) = k = 1 4 b k F k ( X i )
The variables b 1 ~ b 4 represent the weight coefficients, while F 1 ( X i ) to F 4 ( X i ) denote the values of the objective function corresponding to the aforementioned constraints. In the experiment, the parameters were configured as b 1 = 5, b 2 = 1, b 3 = 1, and b 4 = 5, while the remaining parameters are given in Section 2.
In this study, the DRSTASA, STASA, SNS, and PSO algorithms were each executed 30 times within the hardware system environment previously described, using a population size of 500 and a maximum of 2000 iterations. Table 7 presents the optimal, worst, and average fitness values obtained from these 30 runs. Figure 6 shows the three-dimensional view, top view, and side view of the optimal path planning results achieved by the algorithms, while Figure 7 illustrates the fitness iteration curves.
From Table 7, it is clear that this algorithm demonstrates relatively high stability across multiple runs, with both the optimal and worst fitness values being lower than those achieved by other algorithms. This suggests a strong capability for escaping local optima. Figure 5 shows that the optimal path generated by the algorithm successfully avoids obstacles, ensuring a smooth transition from the starting point to the destination. Compared to the other three algorithms, it identifies a safer and faster flight route. Figure 6 illustrates the rapid convergence of the algorithm during iterations, allowing it to find the optimal solution efficiently.

6. Summary

This paper presents an enhancement to the STASA algorithm by introducing a state transition simulated annealing algorithm (DRSTASA) that incorporates the use of destruction perturbation operators and opposition-based learning strategies. During the initialization phase, the super-Latin hypercube sampling method is employed to ensure a more uniform distribution of initial solutions. Subsequently, the integration of destruction operators and opposition-based learning strategies helps to accelerate the convergence process and improve the algorithm’s capacity for escaping local optima. The effectiveness of these enhancements is validated through comparisons with eight other algorithms across 23 standard test functions and eight classical engineering design problems.
DRSTASA is further applied to the UAV path planning problem, demonstrating greater stability in the solutions compared to three other algorithms. The paths produced by DRSTASA are more feasible, safer, and more efficient, thus showcasing the practical value of the algorithm. However, the comparative results indicate that there is room for improvement in terms of reducing time costs for complex problems and enhancing stability in multimodal scenarios. Future research will focus on increasing the convergence speed of the algorithm and reducing computational time.

Author Contributions

Conceptualization, J.L. and X.H.; methodology, J.L.; software, J.L.; validation, F.L., J.W. and W.Z.; formal analysis, J.L.; investigation, F.L.; resources, J.W.; data curation, W.Z.; writing—original draft preparation, J.L.; writing—review and editing, X.H.; visualization, F.L.; supervision, J.W.; project administration, W.Z.; funding acquisition, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China Funding Project grant number (62176176).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author, Xiaoxia Han, upon reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Fang, Z.; Savkin, A.V. Strategies for Optimized UAV Surveillance in Various Tasks and Scenarios: A Review. Drones 2024, 8, 193. [Google Scholar] [CrossRef]
  2. Torrero, L.; Seoli, L.; Molino, A.; Giordan, D.; Manconi, A.; Allasia, P.; Baldo, M. The Use of Micro-UAV to Monitor Active Landslide Scenarios. In Engineering Geology for Society and Territory—Volume 5; Lollino, G., Manconi, A., Guzzetti, F., Culshaw, M., Bobrowsky, P., Luino, F., Eds.; Springer: Cham, Switzerland, 2015. [Google Scholar] [CrossRef]
  3. Yan, C.; Fu, L.; Zhang, J.; Wang, J. A Comprehensive Survey on UAV Communication Channel Modeling. IEEE Access 2019, 7, 107769–107792. [Google Scholar] [CrossRef]
  4. Shakhatreh, H.; Sawalmeh, A.H.; Al-Fuqaha, A.; Dou, Z.; Almaita, E.; Khalil, I.; Othman, N.S.; Khreishah, A.; Guizani, M. Unmanned Aerial Vehicles (UAVs): A Survey on Civil Applications and Key Research Challenges. IEEE Access 2019, 7, 48572–48634. [Google Scholar] [CrossRef]
  5. Hart, P.E.; Nilsson, N.J.; Raphael, B. A Formal Basis for the Heuristic Determination of Minimum Cost Paths. IEEE Trans. Syst. Sci. Cybern. 1968, 4, 100–107. [Google Scholar] [CrossRef]
  6. Korf, R.E. Depth-First Iterative-Deepening: An Optimal Admissible Tree Search. Artif. Intell. 1985, 27, 97–109. [Google Scholar] [CrossRef]
  7. Cen, Y.; Song, C.; Xie, N.; Wang, L. Path planning method for mobile robot based on ant colony optimization algorithm. In Proceedings of the 2008 3rd IEEE Conference on Industrial Electronics and Applications, Singapore, 3–5 June 2008; pp. 298–301. [Google Scholar] [CrossRef]
  8. Qin, Y.-Q.; Sun, D.-B.; Li, N.; Cen, Y.-G. Path planning for mobile robot using the particle swarm optimization with mutation operator. In Proceedings of the 2004 International Conference on Machine Learning and Cybernetics (IEEECat. No. 04EX826), Shanghai, China, 26–29 August 2004; Volume 4, pp. 2473–2478. [Google Scholar] [CrossRef]
  9. Chen, J.; Ye, F.; Jiang, T. Path planning under obstacle-avoidance constraints based on ant colony optimization algorithm. In Proceedings of the 2017 IEEE 17th International Conference on Communication Technology (ICCT), Chengdu, China, 27–30 October 2017; pp. 1434–1438. [Google Scholar] [CrossRef]
  10. Guo, L.; Zhao, C.; Li, J.; Yan, Q.; Li, W.; Chen, P. UAV path planning based on improved A * and artificial potential field algorithm. In Proceedings of the 2024 36th Chinese Control and Decision Conference (CCDC), Xi’an, China, 25–27 May 2024; pp. 5471–5476. [Google Scholar] [CrossRef]
  11. Xiang, H.; Liu, X.; Song, X.; Zhou, W. UAV Path Planning Based on Enhanced PSO-GA. In Artificial Intelligence; Fang, L., Pei, J., Zhai, G., Wang, R., Eds.; CICAI 2023; Lecture Notes in Computer Science; Springer: Singapore, 2024; Volume 14474. [Google Scholar] [CrossRef]
  12. Li, D.; Yin, W.; Wong, W.E.; Jian, M.; Chau, M. Quality-Oriented Hybrid Path Planning Based on A* and Q-Learning for Unmanned Aerial Vehicle. IEEE Access 2022, 10, 7664–7674. [Google Scholar] [CrossRef]
  13. Li, B.; Qi, X.; Yu, B.; Liu, L. Trajectory Planning for UAV Based on Improved ACO Algorithm. IEEE Access 2020, 8, 2995–3006. [Google Scholar] [CrossRef]
  14. Li, H.; Long, T.; Xu, G.; Wang, Y. Coupling-Degree-Based Heuristic Prioritized Planning Method for UAV Swarm Path Generation. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 3636–3641. [Google Scholar] [CrossRef]
  15. Xie, R.; Meng, Z.; Wang, L.; Li, H.; Wang, K.; Wu, Z. Unmanned Aerial Vehicle Path Planning Algorithm Based on Deep Reinforcement Learning in Large-Scale and Dynamic Environments. IEEE Access 2021, 9, 24884–24900. [Google Scholar] [CrossRef]
  16. Tan, L.; Zhang, Y.; Huo, J.; Song, S. UAV Path Planning Simulating Driver’s Visual Behavior with RRT algorithm. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 219–223. [Google Scholar] [CrossRef]
  17. Wang, X.; Pan, J.-S.; Yang, Q.; Kong, L.; Snášel, V.; Chu, S.-C. Modified Mayfly Algorithm for UAV Path Planning. Drones 2022, 6, 134. [Google Scholar] [CrossRef]
  18. Tian, J.; Wang, Y.; Yuan, D. An Unmanned Aerial Vehicle Path Planning Method Based on the Elastic Rope Algorithm. In Proceedings of the 2019 IEEE 10th International Conference on Mechanical and Aerospace Engineering (ICMAE), Brussels, Belgium, 22–25 July 2019; pp. 137–141. [Google Scholar] [CrossRef]
  19. Cui, X.; Wang, Y.; Yang, S.; Liu, H.; Mou, C. UAV path planning method for data collection of fixed-point equipment in complex forest environment. Front. Neurorobot. 2022, 16, 1105177. [Google Scholar] [CrossRef]
  20. Yu, X.; Li, C.; Zhou, J. A constrained differential evolution algorithm to solve UAV path planning in disaster scenarios. Knowl.-Based Syst. 2020, 204, 106209. [Google Scholar] [CrossRef]
  21. Zhang, X.; Xia, S.; Li, X. Quantum Behavior-Based Enhanced Fruit Fly Optimization Algorithm with Application to UAV Path Planning. Int. J. Comput. Intell. Syst. 2020, 13, 1315–1331. [Google Scholar] [CrossRef]
  22. Qu, C.; Gai, W.; Zhang, J.; Zhong, M. A novel hybrid grey wolf optimizer algorithm for unmanned aerial vehicle (UAV) path planning. Knowl.-Based Syst. 2020, 194, 105530. [Google Scholar] [CrossRef]
  23. Chen, Y.; Yu, Q.; Han, D.; Jiang, H. UAV path planning: Integration of grey wolf algorithm and artificial potential field. Concurr. Comput. Pract. Exp. 2024, 36, e8120. [Google Scholar] [CrossRef]
  24. Han, X.; Dong, Y.; Yue, L.; Xu, Q. State Transition Simulated Annealing Algorithm for Discrete-Continuous Optimization Problems. IEEE Access 2019, 7, 44391–44403. [Google Scholar] [CrossRef]
  25. Zhou, X.; Yang, C.; Gui, W. State Transition Algorithm. J. Ind. Manag. Optim. 2012, 8, 1039–1056. [Google Scholar] [CrossRef]
  26. Aarts, E.H.L. Simulated Annealing: Theory and Applications; Springer Nature: Dordrecht, The Netherlands, 1987. [Google Scholar]
  27. Chu, J.; Dong, Y.; Han, X.; Xie, J.; Xu, X.; Xie, G. Short-term prediction of urban PM2. 5 based on a hybrid modified variational mode decomposition and support vector regression model. Environ. Sci. Pollut. Res. 2021, 28, 56–72. [Google Scholar] [CrossRef]
  28. Shen, Y.; Dong, Y.; Han, X.; Wu, J.; Xue, K.; Jin, M.; Xie, G.; Xu, X. Prediction model for methanation reaction conditions based on a state transition simulated annealing algorithm optimized extreme learning machine. Int. J. Hydrog. Energy 2023, 48, 24560–24573. [Google Scholar] [CrossRef]
  29. Alhijawi, B.; Awajan, A. Genetic algorithms: Theory, genetic operators, solutions, and applications. Evol. Intel. 2024, 17, 1245–1256. [Google Scholar] [CrossRef]
  30. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61, ISSN 0965-9978. [Google Scholar] [CrossRef]
  31. Phung, M.D.; Ha, Q.P. Safety-enhanced UAV path planning with spherical vector-based particle swarm optimization. Appl. Soft Comput. 2021, 107, 107376. [Google Scholar] [CrossRef]
  32. Dige, N.; Diwekar, U. Efficient sampling algorithm for large-scale optimization under uncertainty problems. Comput. Chem. Eng. 2018, 115, 431–454. [Google Scholar] [CrossRef]
  33. Rashedi, E.; Nezamabadi-pour, H.; Saryazdi, S. GSA: A Gravitational Search Algorithm. Inf. Sci. 2009, 179, 2232–2248. [Google Scholar] [CrossRef]
  34. Sarafrazi, S.; Nezamabadi-pour, H.; Saryazdi, S. Disruption: A new operator in gravitational search algorithm. Sci. Iran. 2011, 18, 539–548. [Google Scholar] [CrossRef]
  35. Liu, H.; Ding, G.; Wang, B. Bare-bones particle swarm optimization with disruption operator. Appl. Math. Comput. 2014, 238, 106–122. [Google Scholar] [CrossRef]
  36. Shekhawat, S.; Saxena, A. Development and applications of an intelligent crow search algorithm based on opposition based learning. ISA Trans. 2020, 99, 210–230. [Google Scholar] [CrossRef]
  37. Gupta, S.; Deep, K. A hybrid self-adaptive sine cosine algorithm with opposition based learning. Expert Syst. Appl. 2019, 119, 210–230. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of obstacles.
Figure 1. Schematic diagram of obstacles.
Applsci 15 06064 g001
Figure 2. Initialize population comparisons.
Figure 2. Initialize population comparisons.
Applsci 15 06064 g002
Figure 3. DRSTASA algorithm flowchart.
Figure 3. DRSTASA algorithm flowchart.
Applsci 15 06064 g003
Figure 4. Convergence curves for each algorithm.
Figure 4. Convergence curves for each algorithm.
Applsci 15 06064 g004aApplsci 15 06064 g004bApplsci 15 06064 g004c
Figure 5. Algorithmic average time.
Figure 5. Algorithmic average time.
Applsci 15 06064 g005
Figure 6. Optimal path planning diagram.
Figure 6. Optimal path planning diagram.
Applsci 15 06064 g006
Figure 7. Fitness iteration curve.
Figure 7. Fitness iteration curve.
Applsci 15 06064 g007
Table 1. Compare algorithm parameter settings.
Table 1. Compare algorithm parameter settings.
AlgorithmName and Year of PublicationParameter Settings
STAState Transition Algorithm/2012SE = 50, T = 1010, beta = 1, gamma = 1, delta = 1;
STASAState Transition Simulated Annealing Algorithm/2019SE = 50, T = 1010, beta = 1, gamma = 1, delta = 1, a = 0.93;
AOAArithmetic Optimization Algorithm/2021Max_iteration = 2000, PopSize = 50;
SCASine Cosine Algorithm/2016Pop_size = 50, Max_iter = 2000, a = 2;
AHAArtificial Hummingbird Algorithm/2021Max_iteration = 2000, Pop_size = 50;
FDAFlow Direction Algorithm/2021alpha = 50, beta = 1, Max_iteration = 2000;
AVOAAfrican Vultures Optimization Algorithm/2021Pop_size = 50, Max_iter = 2000, p1 = 0.6, p2 = 0.4, p3 = 0.6, alpha = 0.8, betha = 0.2, gamma = 2.5;
GWOGrey Wolf Optimizer Algorithm/2014SearchAgents_no = 50, Max_iteration = 2000;   a m a x = 2 ,   a m i n = 0 ;
ASTASAAdaptive State Transition Simulated Annealing Algorithm/2023SE = 50,Tp = 10, a1 = a2 = 0.5, T = 1010, a = 0.93,
Ω = {0, 0.5, 0.1, 0.1, 1 × 10−3, 1 × 10−6, 1 × 10−9};
Table 2. Unimodal benchmarking function definition.
Table 2. Unimodal benchmarking function definition.
FunctionValue RangeBest
f 1 = i = 1 n x i 2 [−100, 100]0
f 2 = i = 1 n | x i | + i = 1 n | x i | [−10, 10]0
f 3 ( x ) = i = 1 n j = 1 i x j 2 [−100, 100]0
f 4 ( x ) = max i x i , 1 i n [−100, 100]0
f 5 ( x ) = i = 1 n 1 100 x i + 1 x i 2 2 + x i 1 2 [−30, 30]0
f 6 ( x ) = i = 1 n x i + 0.5 2 [−100, 100]0
f 7 ( x ) = i = 1 n i x i 4 + rand [ 0 , 1 ) [−1.28, 1.28]0
Table 3. Multimodal benchmarking function definition.
Table 3. Multimodal benchmarking function definition.
FunctionDimensionValue RangeBest
f 8 ( x ) = i = 1 n x i sin | x i | d[−500, 500] 418.9829 × d
f 9 ( x ) = i = 1 n x i 2 10 cos 2 π x i + 10 d[−5.12, 5.12]0
f 10 ( x ) = 20 exp 0.2 1 n i = 1 n x i 2 exp 1 n i = 1 n cos 2 π x i + 20 + e d[−32, 32]0
f 11 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos x i i + 1 d[−600, 600]0
f 12 ( x ) = π n 10 sin π y 1 + i = 1 n 1 y i 1 2 1 + 10 sin 2 π y i + 1 + y n 1 2 + i = 1 n u x i , 10 , 100 , 4 y i = 1 + x i + 1 4 u x i , a , k , m = k x i a m x i > a 0 a < x i < a k x i a m x i < a d[−50, 50]0
f 13 ( x ) = 0.1 sin 2 ( 3 π x 1 ) + i = 1 n ( x i 1 ) 2 1 + sin 2 ( 3 π x i + 1 ) + ( x n 1 ) 2 1 + sin 2 ( 2 π x n ) + i = 1 n u ( x i , 5 , 100 , 4 ) d[−50, 50]0
f 14 ( x ) = ( 1 500 + j = 1 25 1 j + i = 1 2 ( x i a i j ) 6 ) 1 2[−65.53,
65.53]
1
f 15 = i = 1 11 [ a i x 1 b i 2 + b i x 2 b i 2 + b i x 3 + x 4 ] 2 4[−5, 5]0.00030
f 16 ( x ) = 4 x 1 2 2.1 x 1 4 + x 1 6 / 3 + x 1 x 2 4 x 2 2 + 4 x 2 4 2[−5, 5]−1.0316
f 17 ( x ) = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) cos x 1 + 10 2[−5, 5]0.398
f 18 ( x ) = [ 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ] × [ 30 + ( 2 x 1 3 x 2 ) 2 × ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ] 2[−2, 2]3
f 19 ( x ) = i = 1 4 c i exp ( j = 1 3 a i j ( x j p i j ) 2 ) 3[1, 3]−3.86
f 20 ( x ) = i = 1 4 c i exp ( j = 1 6 a i j ( x j p i j ) 2 ) 6[0, 1]−3.32
f 21 = i = 1 5 X a i X a i T + c i 1 4[0, 10]−10.1532
f 22 = i = 1 7 X a i X a i T + c i 1 4[0, 10]−10.4028
f 23 = i = 1 10 X a i X a i T + c i 1 4[0, 10]−10.5363
Table 4. Test results of nine algorithms on benchmark functions (Among them, the values in bold are the optimal values).
Table 4. Test results of nine algorithms on benchmark functions (Among them, the values in bold are the optimal values).
DSTASASTASASCAAOASTAAHAFDAAVOAGWOASTASA
F1min002.23826 × 10−50002.49859 × 10−2706.545097 × 10−1270
mean002.22950 × 10−50008.20562 × 10−2501.11579 × 10−1210
std004.41827 × 10−50001.69896 × 10−2402.81725 × 10−1210
R 010001010
F2min001.36292 × 10−80005.12704 × 10−2003.19172 × 10−722.73 × 10−298
mean002.18929 × 10−3009.75350 × 10−3055.80501 × 10−1909.43916 × 10−715.78 × 10−285
std005.07143 × 10−30009.87080 × 10−1901.50523 × 10−700
R 010011011
F3min005.34185 × 10−201.67556 × 10−16508.44382 × 10−501.66451 × 10−402.45 × 10−39
mean06.39864 × 10−101.64976 × 10202.11171 × 10−1001.19773 × 10−302.20731 × 10−328.25 × 10−13
std02.32239 × 10−94.05739 × 10204.67669 × 10−1001.54680 × 10−308.16378 × 10−322.68 × 10−12
R 110101011
F4min01.40860 × 10−3200.01267912802.75719 × 10−2942.76011 × 10−2915.85814309409.51328 × 10−335.47 × 10−125
mean02.78906 × 10−3053.5419375184.11638 × 10−33.20126 × 10−2523.20126 × 10−25211.5833053409.71144 × 10−309.77 × 10−10
std005.8069301641.25609 × 10−3002.47263966902.56898 × 10−293.66 × 10−9
R-111111011
F5min1.68646 × 10−821.0827045039.9890906425.8015631020.7250908623.258913699.56426 × 10−54.08905 × 10−825.1999110419.22748051
mean17.888921.66482354415349.140327.1882916723.3089255223.7682897213.809405477.48172 × 10−726.3391054419.60826434
std8.13780.2738335667.79634 × 1050.6840531439.6350975720.30779596216.792423688.34461 × 10−70.7240283750.185495286
R-111111111
F6min6.39082 × 10−146.56884 × 10−145.2613950821.3088482944.734708 × 10−142.29367 × 10−104.61674 × 10−283.31762 × 10−120.2489941788.02 × 10−22
mean1.13264 × 10−131.10163 × 10−13149.86069051.8155372199.077552 × 10−142.91864 × 10−83.53723 × 10−252.03156 × 10−110.7070499934.52 × 10−19
std3.30438 × 10−143.42395 × 10−14595.31568870.2438289263.397547 × 10−146.42005 × 10−81.20000 × 10−241.36005 × 10−110.3526467939.58 × 10−19
R-111−11−111−1
F7min9.09943 × 10−85.78305 × 10−40.0407135603.95305 × 10−74.29126 × 10−52.52943 × 10−69.43934 × 10−31.50881 × 10−63.73270 × 10−50.000105144
mean1.73036 × 10−66.17465 × 10−40.8011069825.62328 × 10−65.52640 × 10−42.64510 × 10−59.43934 × 10−33.52845 × 10−53.39785 × 10−40.010289159
std1.54743 × 10−66.61753 × 10−41.6053386524.83192 × 10−64.07464 × 10−42.16516 × 10−51.01640 × 10−23.79440 × 10−51.85178 × 10−40.008127468
R-111111111
F8min−12,569.4866−12,569.4866−5476.55146−8318.39435−12,569.4866−12,569.4865−10,316.1050−12,569.4866−7925.33057−12,569.4866
mean−12,569.4866−12,464.6926−4321.69164−7716.85829−12,569.4866−12,541.8054−8984.9918−12,538.5609−6031.63602−12,290.9614
std3.46119 × 10−12197.5076886297.4265754315.75643526.47089 × 10−12105.2953070677.69594980.25593103709.6064865291.5008075
R-111111111
F9min006.72810911500027.85883344000
mean0082.7625999600051.6716276100.14194345480
std0051.5006457100014.9847108400.77745632080
R-010001010
F10min8.88178 × 10−168.88178 × 10−160.2159510108.88178 × 10−164.440892 × 10−158.88178 × 10−162.1200533618.88178 ×10−167.99360 × 10−158.88178 × 10−16
mean8.88178 × 10−168.88178 × 10−1617.683533218.88178 × 10−164.440892 × 10−158.88178 × 10−163.5221615818.88178 × 10−168.82257 × 10−158.88178 × 10−16
std005.8661020580001.08152623602.01908 × 10−150
R-010101010
F11min000.2916756430000000
mean002.0784030090.031828867000.09830074506.82138 × 10−40
std002.8037967440.031945879000.02864571000.0026792380
R-011001010
F12min3.21421 × 10−151.47667 × 10−150.8477854040.0807142352.36383 × 10−151.88023 × 10−116.24752 × 10−242.99681 × 10−130.0119887897.06 × 10−24
mean5.81410 × 10−155.67626 × 10−153.49168 × 1070.1232404925.75915 × 10−151.25750 × 10−90.4118700571.38250 × 10−120.0300248751.07 × 10−20
std1.77067 × 10−158.06953 × 10−151.18813 × 1080.0432684061.39051 × 10−154.33242 × 10−90.5455873771.04906 × 10−120.0119558392.70 × 10−20
R--111-11-111-1
F13min3.18192 × 10−143.67286 × 10−147.1236472552.4080099504.22109 × 10−143.76686 × 10−93.37417 × 10−241.61800 × 10−110.1000478831.96 × 10−22
mean7.94253 × 10−147.95419 × 10−145.48897 × 1072.6963935918.33771 × 10−140.3058325780.0135402131.51531 × 10−100.4453470410.045889875
std2.43030 × 10−141.46694 × 10−131.27678 × 1080.1437918362.36018 × 10−140.2355066200.0283546951.14064 × 10−100.1839780240.102558824
R-11111−111−1
F14min0.9980038372.9821051560.9980038370.9980038370.9980038370.9980038370.9980038370.9980038370.9980038370.998003838
mean4.4997544295.4385236550.9980371947.7708960364.7795972600.9980038370.9980038370.9980038373.6785430205.692655933
std5.4404551174.1657932160.0000610964.4332863614.522047715001.93398 × 10−163.7134267104.557993224
R-1−111−1−1−1−11
F15min0.0003074850.0003074850.0003571340.0003203760.0003074850.0003074850.0003.074850.0003074860.0003074860.000307486
mean0.0003074850.0004397130.0008844660.0053133770.0003794560.0003074855.51669 × 10−40.0003075040.0076613020.000514668
std3.62847 × 10−140.0003150200.0003086070.0128727820.0002370061.66935 × 10−134.11854 × 10−45.88603 × 10−80.0098300230.00039831
R-111111111
F16min−1.03162845−1.03162845−1.03162807−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845
mean−1.03162845−1.03162845−1.03160669−1.03162842−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845
std4.38309 × 10−166.15734 × 10−162.45274 × 10−52.16194 × 10−84.75518 × 10−160.3978873576.77521 × 10−165.13342 × 10−161.12689 × 10−94.52 × 10−16
R-1−1111−1111
F17min0.3978873570.3978873570.3978958130.3978873570.3978873570.3978873570.3978873570.3978873570.3978873580.397887358
mean0.3978873570.3978873570.3983178760.3978873730.3978873570.3978873570.3978873570.3978873570.3978874180.397887358
std009.39071 × 10−131.39785 × 10−800007.21333 × 10−80
R-011000011
F18min3333333333
mean33324.2262613733338.4000023793
std1.90209 × 10−142.00653 × 10−141.40481 × 10−529.416643912.48436 × 10−141.27488× 10−152.01660 × 10−156.61248 × 10−820.550358684.95 × 10−16
R-1111−1−1111
F19min−3.86278214−3.86278214−3.86151460−3.86111628−3.86278214−3.86278214−3.86278214−3.86278214−3.86278214−3.86278215
mean−3.86278214−3.86278214−3.85507954−3.85687242−3.86278214−3.86278214−3.86278214−3.86278214−3.86184716−3.86278215
std1.45290 × 10−143.78218 × 10−140.0019854132.89580 × 10−35.20156 × 10−142.71008 × 10−152.71008 × 10−152.14885 × 10−150.0022597032.31 × 10−15
R-111111111
F20min−3.32199517−3.32199517−3.15647379−3.25828015−3.32199517−3.32199517−3.32199517−3.32199517−3.321994517−3.32199517
mean−3.22688067−3.22291757−2.98017193−3.17153577−3.26254861−3.32199517−3.30614275−3.28632723−3.26180981−3.24669620
std0.0483702510.0450663210.1663032200.0355383150.0604628141.34243 × 10−150.0411068090.0554150850.0684848690.058273385
R −1111−11111
F21min−10.1531996−5.05519772−5.71194546−8.30382836−10.15319967−10.1531996−10.1531996−10.1531996−10.1531868−10.1531997
mean−5.90486472−5.05519772−3.33330583−4.75221370−7.032031215−9.90418836−9.90418836−10.1531996−9.47487572−8.04555502
std1.9323926526.53716 × 10−151.6492058811.2482579712.62803321841.3638911121.3638911124.51078 × 10−151.7586562682.659975707
R-111−1−1−1−1−11
F22min−10.4029405 −5.08767182−5.86881454−9.03871310−10.40294056−10.4029405 −10.4029405 −10.4029405 −10.4029243−10.4029406
mean−6.32790119−5.08767182 −3.47931164 −5.64706181−8.915907821−10.4029405 −9.12752888 −10.4029405−10.4028323−9.1468558
std2.2865386104.49568 × 10−151.7547314641.5947637832.54172021491.58195 × 10−152.6222212641.27754 × 10−157.26464 × 10−52.37780875
R-111−1−1−1−1−11
F23min−10.5364098−5.12848078−5.69960702−8.60971414−10.53640981−10.5364098−10.5364098−10.5364098−10.5364073−10.5364098
mean−6.39033089−5.30874508−3.84238656−5.16251028−9.815352612−10.5364098−9.95408731−10.5364098−10.5363015−9.1018401
std2.3263994970.9873482391.1655621371.8507346161.86976930931.89490 × 10−151.7880220773.55271 × 10−155.50112 × 10−52.41893659
R-111−1−1−1−1−1−1
R -13191885471511
Table 5. Eight classic engineering design problems.
Table 5. Eight classic engineering design problems.
Engineering IssuesAttribute
1. Compression spring design3 variables, 4 constraints
2. I-beam design4 variables, 2 constraints
3. Welded Beam Design4 variables, 7 constraints
4. Cantilever design issues5 variables, 1 constraint
5. String design issues2 variables, 6 constraint
6. Three-bar truss design2 variables, 3 constraint
7. Reducer design issues7 variables, 11 constraint
8. Piston rod optimization 4 variables, 4 constraint
Table 6. Comparison of the three algorithms.
Table 6. Comparison of the three algorithms.
DSTSASTASASNS
CSDmin0.0126878070.0126652540.012667240
mean0.0127213150.0127241390.012751065
std1.425476 × 10−53.686603 × 10−51.333570 × 10−4
R-1-1
I-BDmin0.0066259580.0066259580.013074118
mean0.0066259580.0066259580.013074128
std2.148971 × 10−132.272516 × 10−133.415683 × 10−8
R-01
WBDmin1.5569762741.5575394941.724852322
mean1.56843587851.5766625931.724884703
std0.01912363420.0260830909.51866 × 10−5
R-11
CDMin1.3399563751.3399563751.339959827
mean1.3399564941.3399564721.340040862
std9.837556 × 10−87.401418 × 10−81.221619 × 10−4
R-11
SDmin26.4863615226.4863615226.48636147
mean26.4863617026.4863618026.48636147
std1.808759 × 10−72.758648 × 10−77.226896 × 10−15
R-1-1
3-BTmin263.89584341263.89584347263.89584345
mean263.89584439263.89587123263.89587848
std7.925477 × 10−74.292542 × 10−55.3584038 × 10−5
R-1-1
RDmin2994.42457802994.42487402994.4246658
mean2994.42451192994.44244642994.4404264
std8.3096763 × 10−51.4695980 × 10−21.3590637 × 10−2
R-1-1
PROmin8.41269822908.41269833548.4126983231
mean8.41269839588.412698414856.130713531
std4.4586517 × 10−81.0172303 × 10−774.136541109
R-11
Table 7. Algorithm running results.
Table 7. Algorithm running results.
AlgorithmOptimal FitnessWorst FitnessAverage Fitness
DRSTASA4673.51745195.89194737.651
STASA4793.84315485.77534902.3325
SNS4811.46885934.15975005.4432
PSO4966.26535857.13265211.4123
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Han, X.; Liu, F.; Wu, J.; Zhang, W. UAV Path Planning Using a State Transition Simulated Annealing Algorithm Based on Integrated Destruction Operators and Backward Learning Strategies. Appl. Sci. 2025, 15, 6064. https://doi.org/10.3390/app15116064

AMA Style

Liu J, Han X, Liu F, Wu J, Zhang W. UAV Path Planning Using a State Transition Simulated Annealing Algorithm Based on Integrated Destruction Operators and Backward Learning Strategies. Applied Sciences. 2025; 15(11):6064. https://doi.org/10.3390/app15116064

Chicago/Turabian Style

Liu, Jianping, Xiaoxia Han, Fengyi Liu, Jinde Wu, and Wenjie Zhang. 2025. "UAV Path Planning Using a State Transition Simulated Annealing Algorithm Based on Integrated Destruction Operators and Backward Learning Strategies" Applied Sciences 15, no. 11: 6064. https://doi.org/10.3390/app15116064

APA Style

Liu, J., Han, X., Liu, F., Wu, J., & Zhang, W. (2025). UAV Path Planning Using a State Transition Simulated Annealing Algorithm Based on Integrated Destruction Operators and Backward Learning Strategies. Applied Sciences, 15(11), 6064. https://doi.org/10.3390/app15116064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop