Next Article in Journal
CBCT Comparison of Dentoskeletal Effects of Haas-Type and Hyrax-Type Expanders Using Deciduous Teeth as Anchorage: A Randomized Clinical Trial
Previous Article in Journal
Impact of Chelating Agent Salt Type on the Enhanced Oil Recovery from Carbonate and Sandstone Reservoirs
Previous Article in Special Issue
Intelligent Control Systems in Urban Planning Conflicts: Social Media Users’ Perception
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Approaches to Numerical Solution of Optimal Control Problem Using Evolutionary Computations

1
Federal Research Center “Computer Science and Control”, Russian Academy of Sciences, 119333 Moscow, Russia
2
Department of Mechanics and Mechatronics, RUDN University, 117198 Moscow, Russia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(15), 7096; https://doi.org/10.3390/app11157096
Submission received: 1 June 2021 / Revised: 26 July 2021 / Accepted: 27 July 2021 / Published: 31 July 2021
(This article belongs to the Special Issue 14th International Conference on Intelligent Systems (INTELS’20))

Abstract

:
Two approaches to the numerical solution of the optimal control problem are studied. The direct approach is based on the reduction of the optimal control problem to a nonlinear programming problem. Another approach is so-called synthesized optimal control, and it includes the solution of the control synthesis problem and stabilization at some point in the state space, followed by the search of stabilization points and movement of the control object along these points. The comparison of these two approaches was carried out as the solution of the optimal control problem as a time function cannot be directly used in the control system, although the obtained discretized control can be embedded. The control object was a group of interacting mobile robots. Dynamic and static constraints were included in the quality criterion. Implemented methods were evolutionary algorithms and a random parameter search of piecewise linear approximation and coordinates of stabilization points, along with a multilayer network operator for control synthesis.

1. Introduction

The focus of this research was the study of numerical methods for solving the optimal control problem. A group of objects should move from given initial states to terminal ones while avoiding obstacles in a minimum time. The problem belongs to the class of infinite-dimensional optimization. There are two approaches to solve it numerically [1]. A direct approach is based on a discretization of the control function and reduction to the finite-dimensional optimization. An indirect approach is based on the Pontryagin maximum principle, which transforms the original optimization problem into a boundary one, which is numerically solved by shooting methods or as a finite-dimensional optimization problem [2].
A complex control object (a group of mobile robots) is considered. The main property of the object is the presence of phase constraints to avoid collisions between robots, which significantly complicates the development of a numerical method. An attempt to solve this problem, for example, by introducing additional control components (so-called epsilon control) was considered in [3,4]. Other methods that deal with constraints are the potential field method [5], the vector field histogram [6], the gap method [7], or artificial intelligence methods such as swarm intelligence algorithms [8], artificial neural networks [9], and fuzzy logic [10].
The presence of phase constraints in the optimal control problem also leads to the absence of convexity and unimodality of the integral quality criterion. Despite the huge number of works in this area, no effective numerical method has been obtained for its solution [11,12,13,14]. All the main solutions were obtained analytically for low-dimensional models, mainly with differential equations that allow finding a general solution. There are more theoretical results devoted to proving convergence [15] rather than developed algorithms.
Although there is no exact mathematical theory available today for studying the unimodality of quality criteria, it is intuitively obvious that, for example, collision avoidance of two robots moving on a plane is possible using at least two methods, i.e., bypassing the robot to the right or left. Both methods will provide the local minima of the quality criterion. This shows that the quality criterion is not unimodal.
It follows from the above that it is more expedient to develop numerical methods based on the direct approach for solving the optimal control problem, whereby phase constraints are included in the quality criterion as penalties. In this case, global optimization methods, including evolutionary algorithms, can be used to solve the optimal control problem [16]. An indirect approach based on the Pontryagin maximum principle requires the functional to be convex, which cannot be ensured in the presence of phase constraints. The presence of phase constraints requires supplementary variables in addition to the conjugate variables.
In this paper, two numerical methods for the solution of the optimal control problem are compared. The first method uses a piecewise linear control approximation as a time function [17]. The second method is synthesized optimal control [18], in which the control synthesis problem is initially solved, and the stability of the control object to some point in the state space is ensured. Then, the search of the optimal location of stabilization points is performed. The points are switched at a given time interval, which ensures that the object reaches the goal with the optimal value of the quality criterion. As a result of solving the optimal control problem using the second method, we obtain the control function as a function of the coordinates of the state space and a piecewise constant function of time that ensures switching of the stabilization points.
The application of the second method of synthesized optimal control is caused by the fact that the solution of the optimal control problem in the form of time function cannot be directly used in a real object. Such a solution is an open-loop control system, and any deviations from the optimal trajectory caused by external influences or inaccuracies in the model will lead to terminal conditions being missed. To implement the control system, it is necessary to develop a stabilization system to move the object along the optimal trajectory.
The presence of the stabilization system changes the mathematical model of the control object. Thus, we can obtain different mathematical models of the control object for the optimal control problem and its implementation. On the other hand, the stabilization system relative to the programmed path depends on the path itself; therefore, it cannot be included in the mathematical model when solving the optimal control problem.
The lack of unimodality of the quality criterion is caused by either phase constraints or the properties of the quality criterion itself or the model. Ensuring the stability of the control object before solving the optimal control problem is an obvious stage in the development of control systems, which is often used in practice, especially in robotics. Usually, this stage involves the study of the properties of the control object and is based on the experience and intuition of the developer of control systems. In this article, a relatively new symbolic regression method, a multilayer network operator method, is used to solve the problem of control synthesis and ensure stability.
In this paper, we continue a study of evolutionary algorithms for solving complex optimal control problems [17,19] with phase constraints. Well-known evolutionary and population methods, namely, genetic algorithm (GA) [20], particle swarm optimization (PSO) [21,22,23], bee algorithm (BA) [24], and gray wolf optimizer (GWO) [25], which proved their effectiveness in solving optimal control problems [19], were selected and compared to the classic random search (RS) when searching for parameters in piecewise linear approximation and coordinates of stabilization points.
The contribution of this paper is that we show that the optimal control of a group of robots under phase constraints is not unimodal, and it is necessary to use methods of global optimization, for example, evolutionary algorithms. Machine learning is used in the second approach, when solving the problem of stabilization system synthesis. Direct control and synthesized control are compared for the effectiveness of using evolutionary algorithms.
The rest of the article is organized as follows: the optimal control problem for a group of robots with phase constraints is presented in Section 2. Section 3 describes the method of synthesized control. Section 4 introduces a multilayer network operator method. Section 5 contains a short review of evolutionary algorithms. Simulation results are given in Section 6, followed by a conclusion.

2. Optimal Control Problem for Group of Robots

The optimal control problem for a group of mobile robots is considered. Robots should move from given initial states to terminal states in a minimum time. They must not collide with each other or with static obstacles. The coordinates of obstacles in the state space are given.
An essential feature of this problem is the mandatory presence of phase constraints that significantly complicate the search for its solution. To solve the optimal control problem with phase constraints, due to the lack of convexity and unimodality of the quality function, it is suggested to use mainly evolutionary algorithms.
The second obstacle here is the feasibility of the obtained solution. It is obvious that the found optimal control function as a time function cannot be applied to a real object. All researchers in the field of optimal control argue that, to implement the found optimal solution, it is necessary to develop a stabilization system for the optimal program path. In this paper, we propose to solve the optimal control problem using the synthesized control method, which is widely used in practice.
Formal studies on this method in mathematical scientific papers have not been carried out because this method requires solving the control synthesis problem, which is more complicated than the optimal control problem since it requires finding a solution in the form of a mathematical expression of the control function, the argument of which is the state vector of the control object. Substitution of the found control function into the model of the control object ensures the stability of the control object in the neighborhood of a point in the state space. The approaches to solving control synthesis problems are often limited by analytical methods, backstepping [26], and the analytical design of aggregated regulators [27]. The successful application of these methods depends on the mathematical model of the control object. In this work, the numerical method of symbolic regression, a multilayer network operator, was used to solve the synthesis problem. After solving the synthesis problem, the object was controlled by switching from one stabilization point to another.
Evolutionary algorithms were used to implement the synthesized control method. Initially, when solving the synthesis problem, a specialized genetic algorithm was used to find an encoded mathematical expression for the optimal control function. Then, using another evolutionary algorithm, the coordinates of stable equilibrium points for all control objects were found. To analyze the effectiveness of this approach, we also investigated a direct solution to the optimal control problem using various algorithms.
Consider the optimal control problem below for a group of N identical mobile robots. A mathematical model of a mobile robot is given as follows [28]:
x ˙ 1 j = u 1 j cos ( x 3 j ) , x ˙ 2 j = u 1 j sin ( x 3 j ) , x ˙ 3 j = u 2 j ,
where j is an index of a robot in a group, j = 1 , , N , x 1 ,   x 2 are coordinates of the mass center of the robot, x 3 is a rotation angle of robot axis, and u 1 ,   u 2 are control signals on rotors.
The control signals are limited by
u i u i j u i + ,   i = 1 , 2 ,   j = 1 , N ¯ .
Initial states of each mobile robot j are given as
x i j ( 0 ) = x i 0 , j ,   i = 1 , 2 , 3 ,   j = 1 , N ¯ .
Static phase constraints in the state space are
β i ( x j ) = r i 2 ( x 1 * , i x 1 j ) 2 ( x 2 * , i x 2 j ) 2 0 , j = 1 , N ¯ , i = 1 , B ¯ ,
where r i represents distances between the center of static phase constraint and center of the robot that cannot be violated, x 1 * , i ,   x 2 * , i are coordinates of the center of static phase constraint i , and B is the number of static constraints.
Dynamic phase constraints that consider collisions between robots are the following:
δ ( x i , x i + k ) = r 0 2 ( x 1 i x 1 i + k ) 2 ( x 2 i x 2 i + k ) 2 0 ,   i = 1 , N 1 ¯ ,   k = 1 , N i ¯ ,
where r 0 is the radius of dynamic phase constraints, i.e., outside dimensions of the robot, that cannot be violated.
Terminal states of robots are
x i j x i f , j = 0 ,   i = 1 , 2 , 3 ,   j = 1 , N ¯ .
The quality function is
J = t f min ,
where
t f = { t ,     if     t < t + and   max { Δ j : j = 1 ,   N ¯ } ε t + ,     otherwise ,
where Δ j = i = 1 3 ( x i j ( t ) x i f , j ) , j = 1 ,   N ¯ , t + is the maximal control time, and ε is a small positive value.
Phase constraints are included in the quality criterion using the Heaviside function.
J ˜ = t f + i = 1 B j = 1 N 0 t f ϑ ( β i ( x j ) ) d t + i = 1 N 1 k = i + 1 N i 0 t f ϑ ( δ ( x i , x i + k ) ) d t min .
When we need to obtain a differentiable quality function, we substitute the Heaviside function with a sigmoid function as follows:
J ˜ ˜ = t f + i = 1 B j = 1 N 0 t f d t 1 + exp ( A β i ( x j ) ) + i = 1 N 1 k = i + 1 N i 0 t f d t 1 + exp ( A δ ( x i , x i + k ) ) min ,
where A is a large positive value.
The solution of the optimal control problem is a control vector u ˜ ( · ) = [ u ˜ 1 ( · ) u ˜ m ( · ) ] T . An optimal control problem can be transformed into a finite-dimensional optimization one by discretization of control in time. To apply nonlinear programming methods, we approximated u ˜ i ( · ) , i = 1 , m ¯ by the functional dependences with a finite number of parameters using piecewise linear approximations.
Let us introduce a constant time interval Δ t = t i + 1 t i , i = 0 , , d = t + / Δ t . The control is searched for in the form of piecewise linear functions between parameters q i , i = 0 , d ¯ . The control is constrained by
u ( t ) = { u + ,     if   u ˜ u + u ,     if   u ˜ u u ˜     ,   otherwise .
The task is to find a vector of parameters,
q = [ q 1 q ( d + 1 ) 2 N ] T ,
where q q i q + , i = 1 , , ( d + 1 ) 2 N .
For the search of parameters, we used evolutionary and population methods, as well as a random search.

3. Synthesized Control

The synthesized control problem was solved in two steps. First, we found a stability system for each robot to guarantee its stability near a given point in the state space. To solve the synthesis problem, we used one of the symbolic regression methods, a multilayer network operator method.
A stabilization problem involves synthesizing a control function.
u j = g j ( x ˜ j x j ) ,   j = 1 , N ¯ ,
where x ˜ j is a point in the state space R n j or state vector of robot j .
Secondly, we found a set of points in the state space and parameters of the switch.
X ˜ j = ( x ˜ j , 1 , x ˜ j , 2 , , x ˜ j , K j , ε j ) ,   j = 1 , N ¯ .
Points found in Equation (12) represented stabilization points of robots.
u j = g j ( x ˜ j , p x j ) ,
where index p increases its value by 1 when reaching the given stability point,
p p + ( 1 ϑ ( | | x ˜ j , p x j | | ε j ) ) ,
where ϑ ( A ) = { 1 ,     if     A > 0 , 0   ,   otherwise .
Coordinates of stability points were searched simultaneously for all robots. The objective function includes the penalty for phase constraint violation.
J = J ˜ + ω 1 h 1 + ω 2 h 2 ,
where
h 1 = 0 t f k = 1 r j = 1 N ϑ ( α k ( x i ( t ) ) d t ,
h 2 = 0 t f k = 1 s w = 1 W ϑ ( β k ( x j 1 ( t ) , x j 2 ( t ) ) d t .
It is necessary to solve the control synthesis problem in Equation (11) and find a multidimensional nonlinear function u j = g j ( x ˜ j x j ) that guarantees the stability of ODE.
x j = f j ( x j , g j ( x ˜ j x j ) ) .

4. Multilayer Network Operator Method

A network operator method encodes mathematical expressions in the form of directed graphs [29,30]. A multilayer network operator method is a development of the network operator method that encodes mathematical expressions in the form of directed graphs which consist of several subgraphs. Let us consider an example of coding a mathematical expression using the multilayer network operator method.
Let a mathematical expression be given as
y = { q 1 x 1 + x 2 exp ( q 2 x 2 ) cos ( q 1 x 1 ) ,   if   x 1 2 x 2 2 0 q 1 x 1 + x 2 sin ( q 2 x 2 + q 3 ) ,   otherwise .
To code this expression, it is enough to have the following basic sets:
a   set   of   arguments   F 0 = ( f 0 , 1 = q 1 , f 0 , 2 = q 2 , f 0 , 3 = q 3 , f 0 , 4 = x 1 , f 0 , 5 = x 2 ) ;
a   set   of   unary   functions   F 1 = ( f 1 , 1 ( z ) = z , f 1 , 2 ( z ) = z , f 1 , 3 ( z ) = exp ( z ) , f 1 , 4 ( z ) = z 2 ,
f 1 , 5 ( z ) = sin ( z ) , f 1 , 6 ( z ) = cos ( z ) , f 1 , 7 ( z ) = ϑ ( z ) , f 1 , 8 ( z ) = 1 ϑ ( z ) ) ,
where
ϑ ( z ) = { 1 , if   z > 0 0 ,   otherwise ;
a set of binary functions F 2 = ( f 2 , 1 ( z 1 , z 2 ) = z 1 + z 2 , f 2 , 2 ( z 1 , z 2 ) = z 1 z 2 ) .
If some element of the set has two indices, then the first one shows the number of arguments, whereas the second one is an index of the element in the set. The Heaviside function ϑ ( z ) is used for conditional operator IF.
Let A = q 1 x 1 + x 2 exp ( q 2 x 2 ) cos ( q 1 x 1 ) , B = x 1 2 x 2 2 , and C = q 1 x 1 + sin ( q 2 x 2 + q 3 ) ; then,
y = ( 1 ϑ ( B ) ) A + ϑ ( B ) C = f 1 , 8 ( B ) A + f 1 , 7 ( B ) C .
Suppose that a multilayer network operator has K = 4 layers or subgraphs. Suppose that the number of nodes is equal to 8, and four of them are source nodes, n 0 = 4 . Each layer has the same number of source nodes, some of which may not be used. The graph of a multilayer network operator for Equation (11) is shown in Figure 1.
In Figure 1, the source nodes of the graph contain the indices of binary functions; next to the edges are the indices of unary functions. To calculate a mathematical expression, it is necessary to specify the matrix of connections D . This matrix has dimensions of K × 2 n 0 . Each row of the matrix is associated with a specific layer and indicates the source nodes of this layer. The odd element of the row contains the index of the layer. If the index is equal to 0 then it is a set of arguments. The even element of the row indicates the node number of this layer or the element number in the argument set. For the multilayer network operator in Figure 1, the matrix of connections has the form
D = [ 0 1 0 2 0 4 0 5 1 5 1 6 1 8 0 3 2 6 1 8 2 8 0 5 1 5 3 5 3 7 3 8 ] .
The network operator matrices of the four layers are
Ψ 1 = [ 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 4 0 0 0 0 0 1 4 0 0 0 0 0 2 0 0 0 0 0 0 0 0 2 0 0 0 0 0 0 0 0 1 2 0 0 0 0 0 0 0 1 ] ,   Ψ 2 = [ 0 0 0 0 0 6 0 0 0 0 0 0 2 0 1 0 0 0 0 0 0 8 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 3 0 0 0 0 0 0 0 2 0 0 0 0 0 0 0 0 1 5 0 0 0 0 0 0 0 1 ] , Ψ 3 = [ 0 0 0 0 0 6 0 0 0 0 0 0 7 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 2 1 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 2 1 0 0 0 0 0 0 0 1 ] ,   Ψ 4 = [ 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 2 1 0 0 0 0 0 0 0 2 1 0 0 0 0 0 0 0 1 ] .
Note that, when solving the synthesis problem using the symbolic regression method, direct coding is not required. In the search process, variations of codes and calculations of mathematical expressions are performed, i.e., decoding.

5. Evolutionary Methods

Evolutionary methods form a class of modern optimization algorithms that work with a set of randomly created possible solutions and apply certain modifications of these solutions. All evolutionary methods have the following common features:
-
The search process is iterative. For each search iteration, a set of possible solutions is considered.
-
Each solution takes part in avoiding local extrema and navigating to promising areas of search space.
-
Not only the best solution but also other information about the search space is preserved over the iterations and used to form new possible solutions.
-
The evolution of solutions is based on the inheritance property, in which better solutions have a greater chance of being included in next search iteration.
The efficiency of evolutionary methods is due to the right balance between exploration and exploitation search. Exploration allows investigating the whole search space for the promising areas. This part of the search process should be applied to the search space as broadly as possible. On the other hand, exploitation involves a local search in promising areas to ensure higher accuracy of the solution. The methods of performing exploration and exploitation searches and the balance between them differ from one evolutionary method to another.
The first and the most well-known of the evolutionary methods is the genetic algorithm (GA) [20]. This method was inspired by Darwin evolution concepts. In the genetic algorithm, operators of crossover and mutation were introduced to modify a set of possible solutions in search for a best one. In this method, each possible solution is encoded with a Gray binary code and called a chromosome, whereas each search iteration is called a generation. Analogous to Darwin’s theory, better solutions have a higher probability of participating in generating new chromosomes for the next generation. However, taking into account only the best solution may lead to premature convergence to some local extremum. To prevent this, similar solutions should not be used during crossover. The mutation operator also helps not to get stuck in a local extremum. The probability of mutation, the probability of crossing depending on the quality of solutions, and their proximity are the main tuning parameters of the genetic algorithm.
Another well-known and popular evolutionary method is particle swarm optimization (PSO) [21,22,23]. As the method’s name suggests, it mimics the social behavior of some creatures in nature which exist in swarms. PSO was proposed in 1995 and gained popularity due to its effective way of combining exploration and exploitation searches during the solution modification. In this method, possible solutions are called particles. These particles are distributed in the search space and seek better positions by taking into account the best particle and the best position in previous iterations. This technique can be seen in flocks of birds or in schools of fish and is called social intelligence. The movement toward the global best solution is part of the exploitation search. The inertial movement and the movement toward the best position is part of the exploration search. The balance among these three components of the particle movement is determined by the corresponding tuning parameters of the method.
The observation of the collective behavior of bees in finding nectar sources prompted the creation of the bee algorithm (BA) [24]. In this evolutionary method, the exploration part of the search is represented by a random search of the whole search space. The analogous process among bees is the process of looking for promising places with high nectar concentration by scout bees. The found solutions can be divided into highly interesting, interesting, and not interesting. Not interesting solutions are substituted by new randomly created ones. The subdomains of given radii of highly interesting and interesting solutions are investigated more intensively. A given number of randomly created solutions are distributed within these subdomains. It is obvious that highly interesting subdomains get more attention. Best found solutions within each subdomain succeed to the next iteration. The radii of highly interesting and interesting subdomains decrease with each iteration. This ensures the success of the exploitation search. The main tuning parameters of the bee algorithm are the number of additional solutions in subdomains and their initial radii.
The gray wolf optimizer (GWO) is the most recent evolutionary method used in the study [25]. It appeared in 2014. This optimization method was inspired by the social behavior of a pack of wolves during hunting. The leader of the pack is called the alpha. The alpha always participates in the hunting process and coordinates it. Next in the wolf pack hierarchy are betas and deltas, which help in attacking prey. In a mathematical model of this hierarchy, the alpha is the current best solution, whereas the second and third best solutions are the beta and delta, respectively. For each search iteration, the three best possible solutions (alpha, beta, and delta) are selected from the set of all possible solutions. The modification of each possible solution is performed with respect to the positions of alpha, beta, and delta. The balance between exploration and exploitation search is achieved by a special component that is linearly decreased over iterations. The main advantage of the gray wolf optimizer with respect to the methods discussed above is that it is free of any tuning parameters.
When developed, all evolutionary algorithms are tested on benchmark functions that may have many local minima. An algorithm is considered satisfactory if it finds a global optimum. To increase the probability of finding the global optimum, it is necessary to enlarge parameters such as the population size and the number of generations.

6. Simulation Results

In both considered approaches, we needed to solve the finite-dimensional problem of nonlinear programming. The methods recommended in [18] for solving nonlinear programming problems require the quality criterion to satisfy the conditions of smoothness, convexity, and unimodality. In most applied optimal control problems, satisfaction of these conditions cannot be fulfilled. The application of evolutionary methods for solving nonlinear programming problems does not impose the abovementioned conditions on the quality criterion. Recent research has shown the high efficiency of evolutionary methods for solving optimal control problems [19].
To estimate the complexity of algorithms, we used the average number of objective function calculations required to obtain a result, Y a v g . Parameters of the algorithms were selected such that the number of objective function calculations Y a v g for different algorithms was nearly the same.
The following parameters were used in computational experiments: N = 4 , t + = 2.8 , ε = 0.01 , x 1 1 ( 0 ) = 0 , x 2 1 ( 0 ) = 0 , x 3 1 ( 0 ) = 0 , x 1 2 ( 0 ) = 0 , x 2 2 ( 0 ) = 5 , x 3 2 ( 0 ) = 0 , x 1 3 ( 0 ) = 0 , x 2 3 ( 0 ) = 10 , x 3 3 ( 0 ) = 0 , x 1 4 ( 0 ) = 0 , x 2 4 ( 0 ) = 15 , x 3 4 ( 0 ) = 0 , x 1 f , 1 = 15 , x 2 f , 1 = 10 , x 3 f , 1 = 0 , x 1 f , 2 = 15 , x 2 f , 2 = 15 , x 3 f , 2 = 0 , x 1 f , 3 = 15 , x 2 f , 3 = 0 , x 3 f , 3 = 0 , x 1 f , 4 = 15 , x 2 f , 4 = 5 , x 3 f , 4 = 0 , x 1 * , 1 = 5 , x 2 * , 1 = 5 , x 1 * , 2 = 10 , x 2 * , 2 = 10 , r 1 = r 2 = 3 , r 0 = 2 , u 1 = 10 , u 1 + = 10 , u 2 = 10 , u 2 + = 10 , q = 20 , q + = 20 , and Δ t = 0.25 .
A detailed description of used evolutionary algorithms for both approaches was provided in [19]. For a group of four robots, we searched using a vector of 12 parameters × 2 controls × 4 robots = 96 elements.
The results of computational experiments for the optimal control problem solution using the direct approach are given in Table 1. Table 1 contains the mean values of objective function J a v g and average complexity estimation Y a v g , i.e., the number of functional calculations. Note that, in Table 1, each row from 1 to 10 contains the results of one experiment for each algorithm.
Results presented in Table 1 show that all evolutionary algorithms performed better than the random search. Furthermore, PSO and GA solved the problem more effectively.
Figure 2 shows the trajectories of robots that bypassed the obstacles (circles) with control obtained using PSO, J = 2.81 . As can be seen from Figure 2, PSO found the optimal solution that ensured the attainment of terminal conditions with given accuracy without violation of the phase constraints and without collisions. The trajectories of the robots had a certain excess length.
In the second approach with synthesized control, we first solved a control synthesis problem relative to the point in the state space. Since all robots were identical, then the synthesized control function in Equation (13) was used for each robot.
Then, we solved a finite-dimensional optimization problem and searched for stabilization points x ˜ i j , k , where i is an index of the element in the state vector, i = 1 , 3 ¯ , j is a robot index, j = 1 , 4 ¯ , and k is a point index.
The switch from one point to another was performed after each time interval Δ t = 0.7 . For each robot, we searched three points and one known terminal point. Thus, we searched using a vector of 3 · 4 · ( 2.8 / 0.7 1 ) = 36 elements.
The following parameters were used in computational experiments: u 1 = 10 , u 1 + = 10 , u 2 = 10 , u 2 + = 10 , q = 20 , q + = 20 , q 1 = 1 , q 2 = 1 , q 3 = 1.57 , q 1 + = 16 , q 2 + = 16 , and q 3 + = 1.57 . Additional parameters of the algorithms were as follows GA: H = 256, p_m = 0.7, W = 288; PSO: H = 32, W = 2048, α = 0.72 , β = 0.5 , γ = 0.1 , δ = 1 ; BA: H = 30, W = 198, α e = 0.95 , α s = 0.95 r e = 4 , r s = 4 ; GWO: H = 32, W = 2048, where p_m is the probability of mutation, H is the size of the population, and W is the number of generations.
The average number of objective function calculations Y a v g was set smaller than in the direct approach because the simulation of synthesized control was very time-consuming. The results of computational experiments are given in Table 2.
Figure 3 shows the trajectories of robots with control obtained using GWO, J = 2.49 . Here, black squares show the identified stabilization points, whereas red circles are obstacles. The paths of the robots crossed, but at different time moments; thus, they avoided collisions and did not come across obstacles. Stabilization points “attracted” the robots but did not necessarily lie on their paths. All robots reached the terminal conditions with given accuracy without violation of the phase constraints and without collisions. The trajectories of robots were smoother in comparison to those shown in Figure 2.
Despite the fact that the optimal control search using the synthesized control method required fourfold less calculation of the objective function for each method in comparison to the direct approach, the search results turned out to be much better. The best result was obtained using GWO, on average, followed by PSO.
All evolutionary techniques in both approaches performed better than the random search. We may suppose that evolutionary algorithms use certain features of intelligent search.

7. Conclusions

The optimal control problem of a group of robots with phase constraints was solved using two numerical approaches. Robots had to achieve the terminal points without collisions among themselves or with obstacles in a minimal time. The presence of phase constraints complicated the search. Evolutionary computations were applied to cope with the nonconvex and multimodal quality criterion.
In the first approach, a robotic group was treated as one object. The optimal control problem was reduced to the nonlinear programming problem. In the second approach, the problem of control synthesis was first solved separately for each robot relative to the points in the state space. To solve the synthesis problem, the multilayer network operator method was used, although this could also be achieved using other symbolic regression methods that can derive mathematical expressions.
Then, the optimal control problem was considered in the original statement, where the coordinates of the stabilization points of robots were used as control. A search was carried out for three stabilization points for each of the robots. Switching between points occurred at a specified time interval. A search of the coordinates of stability points was performed using evolutionary algorithms (genetic algorithm, particle swarm optimization, bee algorithm, and gray wolf optimizer) and a random search.
After a series of 10 tests for each algorithm, they were evaluated by the average value of functional for the best solution identified. Experiments showed that, on average, the PSO algorithm was the most effective in terms of the search of parameters in the direct method and coordinates of stabilization points in synthesized control.
With respect to the application of evolutionary algorithms to the compared approaches, a synthesized control approach was proven to perform better. In the direct approach, only GA and PSO gave satisfactory results on average, whereas, in the second approach, all the evolutionary algorithms performed well and gave, on average, approximately the same results.
Currently, there are no universal numerical methods for solving optimal control problems. It is proposed to continue the study of evolutionary algorithms and consider other new evolutionary algorithms, including hybrid ones. To use evolutionary algorithms in optimal control problems, it is necessary to include phase constraints in the functional, discretize the problem, and reduce it to a nonlinear programming problem. For this purpose, the time of the control process is divided into intervals in which the control function is approximated by polynomials depending on a finite number of parameters. Furthermore, the search for the values of the parameters of the approximating function can be carried out by evolutionary algorithms.
A distinctive feature of the application of evolutionary algorithms for solving optimal control problems is that, when calculating the value of the quality criterion, it is necessary to integrate a system of differential equations that describe the mathematical model of the control object with an approximating control function.

Author Contributions

Conceptualization, A.D.; methodology, A.D.; investigation, A.D., E.S., S.K.; software, A.D., E.S., S.K.; supervision, A.D.; writing—original draft preparation, A.D., E.S.; writing—review and editing, E.S., S.K. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the Ministry of Science and Higher Education of the Russian Federation, project No. 075-15-2020-799.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors thank the unknown reviewers for their useful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rao, A.V. A Survey of Numerical Methods for Optimal Control; Preprint AAS 09-334; American Astronautical Society by Univelt: Escondido, CA, USA, 2015; pp. 1–32. [Google Scholar]
  2. Diveev, A.; Sofronova, E.; Zelinka, I. Optimal Control Problem Solution with Phase Constraints for Group of Robots by Pontryagin Maximum Principle and Evolutionary Algorithm. Mathematics 2020, 8, 2105. [Google Scholar] [CrossRef]
  3. Chertovskih, R.; Karamzin, D.; Khalil, N.T.; Pereira, F. Regular path-constrained time-optimal control problems in three-dimensional flow fields. Eur. J. Control 2020, 56, 98–106. [Google Scholar] [CrossRef] [Green Version]
  4. Arutyunov, A.; Karamzin, D. A Survey on Regularity Conditions for State-Constrained Optimal Control Problems and the Non-degenerate Maximum Principle. J. Optim. Theory Appl. 2020, 184, 697–723. [Google Scholar] [CrossRef]
  5. Koren, Y.; Borenstein, J. Potential field methods and their inherent limitations for mobile robot navigation. In Proceedings of the 1991 IEEE International Conference on Robotics and Automation, Sacramento, CA, USA, 7–12 April 1991; pp. 1398–1404. [Google Scholar]
  6. Borenstein, J.; Koren, Y. The vector field histogram-fast obstacle avoidance for mobile robots. IEEE Trans. Robot. Autom. 1991, 7, 278–288. [Google Scholar] [CrossRef] [Green Version]
  7. Sezer, V.; Gokasan, M. A novel obstacle avoidance algorithm: Follow the gap method. Robot. Auton. Syst. 2012, 60, 1123–1134. [Google Scholar] [CrossRef]
  8. Diep, Q.B.; Zelinka, I.; Senkerik, R. An Algorithm for Swarm Robot to Avoid Multiple Dynamic Obstacles and to Catch the Moving Target. In Proceedings of the Transactions on Petri Nets and Other Models of Concurrency XV, Zakopane, Poland, 16–20 June 2019; pp. 666–675. [Google Scholar]
  9. Duguleana, M.; Mogan, G. Neural networks based reinforcement learning for mobile robots obstacle avoidance. Expert Syst. Appl. 2016, 62, 104–115. [Google Scholar] [CrossRef]
  10. Reignier, P. Fuzzy logic techniques for mobile robot obstacle avoidance. Robot. Auton. Syst. 1994, 12, 143–153. [Google Scholar] [CrossRef]
  11. Fedorenko, R.P. Approximate Solution of Optimal Control Problems; Nauka: Moscow, Russian, 1978; 488p. (In Russian) [Google Scholar]
  12. Grachev, I.; Evtushenko, Y. A library of programs for solving optimal control problems. USSR Comput. Math. Math. Phys. 1979, 19, 99–119. [Google Scholar] [CrossRef]
  13. Frego, D.M. Numerical Methods for Optimal Control Problems with Application to Autonomous Vehicles. Ph.D. Thesis, University of Trento, Trento, Italy, 2014. [Google Scholar]
  14. Tyatyushkin, A.; Zarodnyuk, T. Numerical method for solving optimal control problems with phase constraints. Numer. Algebra Control Optim. 2017, 7, 481–492. [Google Scholar] [CrossRef]
  15. Pytlak, R. Numerical Methods for Optimal Control Problems with State Constraints; Springer: Berlin/Heidelberg, Germany, 1999; p. 216. [Google Scholar]
  16. Esposito, W.R.; Floudas, C.A. Deterministic Global Optimization in Nonlinear Optimal Control Problems. J. Glob. Optim. 2000, 17, 97–126. [Google Scholar] [CrossRef]
  17. Diveev, A.I.; Konstantinov, S.V.; Sofronova, E.A. A Comparison of Evolutionary Algorithms and Gradient-based Methods for the Optimal Control Problem. In Proceedings of the 2018 5th International Conference on Control, Decision and Information Technologies (CoDIT), Thessaloniki, Greece, 10–13 April 2018; pp. 259–264. [Google Scholar]
  18. Diveev, A.I. Numerical Method of Synthesized Control for Solution of the Optimal Control Problem. In Intelligent Systems and Computing; Kohei, A., Supriya, K., Rahul, B., Eds.; Springer: Berlin, Germany, 2020; Volume 1228, pp. 137–157. [Google Scholar]
  19. Diveev, A.I.; Konstantinov, S.V. Study of the Practical Convergence of Evolutionary Algorithms for the Optimal Program Control of a Wheeled Robot. J. Comput. Syst. Sci. Int. 2018, 57, 561–580. [Google Scholar] [CrossRef]
  20. Goldberg, D.E. Genetic Algorithms in Search Optimization and Machine Learning; Addison-Wesley: Boston, MA, USA, 1989; 432p. [Google Scholar]
  21. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95 International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  22. Yu, Y.; Li, Y.; Li, J. Parameter Identification of a Novel Strain Stiffening Model for Magnetorheological Elastomer Base Isola-tor Utilizing Enhanced Particle Swarm Optimization. J. Intell. Mater. Syst. Struct. 2015, 26, 2446–2462. [Google Scholar] [CrossRef]
  23. Yu, Y.; Zhang, C.; Gu, X.; Cui, Y. Expansion prediction of alkali aggregate reactivity-affected concrete structures using a hybrid soft computing method. Neural Comput. Appl. 2019, 31, 8641–8660. [Google Scholar] [CrossRef]
  24. Pham, D.T.; Ghanbarzadeh, A.; Koc, E.; Otri, S. The bees algorithm a novel tool for complex optimization problems. In Intel-ligent Production Machines and Systems, Proceedings of the 2nd I*PROMS Virtual International Conference, 3–14 July 2006; Elsevier: Amsterdam, The Netherlands, 2006; pp. 454–459. [Google Scholar]
  25. Mirjalili, S.A.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  26. Kokotovic, P.; Sussmann, H. A positive real condition for global stabilization of nonlinear systems. Syst. Control Lett. 1989, 13, 125–133. [Google Scholar] [CrossRef]
  27. Kolesnikov, A.A.; Al Kolesnikov, A.; Kuz’menko, A.A. Backstepping and ADAR Method in the Problems of Synthesis of the Nonlinear Control Systems. Mekhatronika Avtom. Upr. 2016, 17, 435–445. [Google Scholar] [CrossRef] [Green Version]
  28. De Luca, A.; Oriolo, G.; Samson, C. Feedback control of a nonholonomic car-like robot. In Sensing and Control for Autonomous Vehicles; Laumond, J.P., Ed.; Robot Motion Planning and Control, Lecture Notes in Control and Information Sciences; Springer: Berlin/Heidelberg, Germany, 1998; Volume 229, pp. 171–253. [Google Scholar]
  29. Diveev, A.I.; Sofronova, E.A. Numerical method of network operator for multi-objective synthesis of optimal control system. In Proceedings of the 7th International Conference on Control and Automation (ICCA’09) 2009, Christchurch, New Zealand, 9–11 December 2009; pp. 701–708. [Google Scholar]
  30. Diveev, A.I. A numerical method for network operator for synthesis of a control system with uncertain initial values. J. Comput. Syst. Sci. Int. 2012, 51, 228–243. [Google Scholar] [CrossRef]
Figure 1. The graph of a multilayer network operator.
Figure 1. The graph of a multilayer network operator.
Applsci 11 07096 g001
Figure 2. Solution obtained using a direct approach with PSO, J = 2.81 .
Figure 2. Solution obtained using a direct approach with PSO, J = 2.81 .
Applsci 11 07096 g002
Figure 3. The solution obtained using synthesized control with GWO, J = 2.49 .
Figure 3. The solution obtained using synthesized control with GWO, J = 2.49 .
Applsci 11 07096 g003
Table 1. The results of experiments: direct approach.
Table 1. The results of experiments: direct approach.
No.GAPSOBAGWORS
15.484.5212.747.7114.77
25.064.459.8211.4317.25
35.295.389.5810.7113.8
46.645.7811.758.8218.67
55.542.8112.959.413.56
67.036.2513.739.5116.6
74.433.279.976.5518.02
83.73.5611.6710.0819.9
94.435.99.6711.2312.81
107.213.6913.899.7612.86
J a v g 5.54.5711.599.5315.83
Y a v g 260.512250.102233.490241.730247.503
Table 2. The results of experiments: synthesized control.
Table 2. The results of experiments: synthesized control.
No.GAPSOBAGWORS
12.82.812.982.897.75
22.692.864.246.296.86
32.732.843.722.838.77
42.862.633.132.836.2
52.592.753.282.8610.24
62.742.863.722.876.7
72.792.643.232.858.59
82.932.624.222.499.74
92.792.663.72.655.56
102.672.874.172.8810.29
J a v g 2.762.753.643.158.07
Y a v g 63.47765.60265.83865.02663.939
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Diveev, A.; Sofronova, E.; Konstantinov, S. Approaches to Numerical Solution of Optimal Control Problem Using Evolutionary Computations. Appl. Sci. 2021, 11, 7096. https://doi.org/10.3390/app11157096

AMA Style

Diveev A, Sofronova E, Konstantinov S. Approaches to Numerical Solution of Optimal Control Problem Using Evolutionary Computations. Applied Sciences. 2021; 11(15):7096. https://doi.org/10.3390/app11157096

Chicago/Turabian Style

Diveev, Askhat, Elena Sofronova, and Sergey Konstantinov. 2021. "Approaches to Numerical Solution of Optimal Control Problem Using Evolutionary Computations" Applied Sciences 11, no. 15: 7096. https://doi.org/10.3390/app11157096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop