Next Article in Journal
Development of a Recognition System for Spraying Areas from Unmanned Aerial Vehicles Using a Machine Learning Approach
Previous Article in Journal
Dynamic Load Balancing of Software-Defined Networking Based on Genetic-Ant Colony Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Model-Free Minimum-Seeking Autotuning Method for Unmanned Aerial Vehicle Controllers Based on Fibonacci-Search Algorithm

1
Institute of Control, Robotics and Information Engineering, Poznan University of Technology, Piotrowo 3a, 60-965 Poznan, Poland
2
Department of Cybernetics, Faculty of Electrical Engineering, Czech Technical University in Prague, Technicka 2, 166 27 Prague 6, Czech Republic
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(2), 312; https://doi.org/10.3390/s19020312
Submission received: 14 November 2018 / Revised: 30 December 2018 / Accepted: 9 January 2019 / Published: 14 January 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
The paper presents a novel autotuning approach for finding locally-best parameters of controllers on board of unmanned aerial vehicles (UAVs). The controller tuning is performed fully autonomously during flight on the basis of predefined ranges of controller parameters. Required controller properties may be simply interpreted by a cost function, which is involved in the optimization process. For example, the sum of absolute values of the tracking error samples or performance indices, including weighed functions of control signal samples, can be penalized to achieve very precise position control, if required. The proposed method relies on an optimization procedure using Fibonacci-search technique fitted into bootstrap sequences, enabling one to obtain a global minimizer for a unimodal cost function. The approach is characterized by low computational complexity and does not require any UAV dynamics model (just periodical measurements from basic onboard sensors) to obtain proper tuning of a controller. In addition to the theoretical background of the method, an experimental verification in real-world outdoor conditions is provided. The experiments have demonstrated a high robustness of the method to in-environment disturbances, such as wind, and its easy deployability.

1. Introduction

Controller tuning precision significantly influences in-flight properties of any architecture used to control multirotor unmanned aerial vehicles (UAVs). Effective and precise tracking of a desired state during flight, precise positioning, and autonomous landing can be ensured by nonlinear controllers with complicated structure and adaptation mechanisms [1]. However, they usually require knowledge about UAV dynamics, which results in a large computational burden when calculating a control law, made additionally more severe, by the introduction of optimality criteria. Alternatively, fixed-parameter controllers of PD (proportional-derivative) or PID (proportional-integral-derivative) types are nowadays widely used [2,3,4,5,6].
When these controllers are tuned appropriately, they can ensure sufficient performance in most of the required UAV applications, including demanding active interaction with the environment. Furthermore, they neither require the knowledge of the analytical model of UAVs nor they require performing any online estimation of the parameters. The introduction of an appropriate optimality criterion in the form of a cost function streamlines the tuning procedure and synthesis of a controller. In the problem considered in this paper, it is assumed, in addition, that the tuning procedure of a controller (or performing fine-tuning for rough initial estimates of controller parameters) is performed in real time during flight of a UAV. This is especially useful in applications such as transporting load by a single flying unit [7,8] or by multiple cooperating UAVs [9] that require even better precision of trajectory tracking due to the coupling between the robots, e.g., in grasping tasks (see Figure 1). When the load changes, it is important to tune controller parameters to the new dynamics of the system. In the presented approach, it is also possible in the ideal case of no model-UAV mismatch (perfect modeling), to obtain an optimal controller parameters for various flight scenarios, e.g., in acrobatic maneuvres (with dynamic acceleration phase) or maximization of flight time by avoiding sudden moves using smooth and feasible reference trajectories.
The work presented in this paper has been motivated by the Mohamed Bin Zayed International Robotics Challenge 2017 (MBZIRC) in Abu Dhabi, where a precise altitude controller has shown to be one of the key aspects of proper system behavior in both tasks designed for UAVs. In the first challenge, landing on a fast moving vehicle, fast descent and precise acquirement of the required altitude are obviously crucial for a safe approach and soft touch-down onto the helipad. Overshooting the set point can result in a too fast descent and consequently failing to attach the device due to a too aggressive impact. In the third challenge, a treasure hunt, in which a set of static and dynamic objects has to be located, tracked, picked and placed, a precisely tuned controller is even more important to achieve the desired performance.
The necessity of frequent controller tuning has been highlighted in the MBZIRC competition, which required a fleet of multiple UAVs with different sensory equipment and actuators, but its importance seems as big in any deployment of UAV systems. Very precise manual tuning of one set of controller parameters takes days of flying outdoor in different environment conditions. Moreover, each change of the platform (mainly those that influence UAV weight) requires obtaining a new set of parameters and therefore additional time-consuming on flight parameter analyses is needed. Therefore, the aim of this paper is to design an advanced method for parameter tuning that achieves even better results than manual tuning in a fraction of time required by current approaches.
In the paper, a novel method for real-time model-free minimum-seeking tuning of controllers is presented (FGT, Fibonacci-search Gain Tuning), including verification of its applicability in a sample application of in-flight tuning of an altitude controller of the PD type in a hexacopter UAV. This method is universal to such an extent that it can be as well adapted for tuning of any other controller (in general, of different types too) used in multirotor and fixed-wing UAVs. The effective tuning of a controller is ensured by the Fibonacci-search method, which rapidly reduces the search space, which is especially important in the field of UAVs from practical reasons, i.e., a limited flight duration, in which both the voltage of the batteries and thrust forces of flying robots decrease in time [10]. The proposed approach considers possible constraints imposed on a control signal as well as the measurement noise and external disturbances the effect of which is compensated using a low-pass filter. The output signal of a closed-loop system is assumed to originate from a single or multiple sources (data fusion), e.g., various on-board sensors, GPS or any external motion capture system.

2. State of the Art

PD and PID type controllers with fixed parameters mentioned in previous subsection are widely used in UAVs and are a good starting point to characterize the current state of the art considering their tuning and techniques to enhance their performance. From a historical point of view, early methods, such as [11,12], enable one to obtain controller parameters for plants in various structures of different nature and dynamics, usually leading to stable closed-loop systems. However, the obtained parameters are not optimal since there is no optimality criterion introduced. Over the years, multiple model-based or model-free approaches have appeared, and an in-depth review of 51 practice-oriented methods for a PID controller design for specified performance criteria can be found in [13]. The authors would like to focus readers’ attention on the third edition of the book [14] in which an exhaustive review of tuning methods used between 1935 and 2008 for PI and PID controllers has been made along with the proceedings of the cyclic conference IFAC Conference on Advances in Proportional-Integral-Derivative Control.
In parallel to the increase in computational power of computers, batch optimization, soft-computing, and methods of optimal controller tuning have appeared, which are still developing up to this date. Such methods require cost functions to be defined, expressing designer demands with the mathematical language, imposed on control systems, such as stability, robustness against disturbances, predefined dynamics of selected signals, either in steady or transient state, and overshoot requirements or expected steady-state properties. In the simplest and widely-encountered applications, the cost functions have been constructed using an absolute or squared tracking error or control signal, leading to integral performance indices, such as IAE, ISE, ITAE, ITSE, etc. The problem of parameter tuning, however, does not require in general any knowledge concerning plant dynamics, which can be identified as the greatest advantage of this approach. Later in the text, the focus turns to model-free tuning procedures. Optimization methods based on fuzzy logic [15,16] or neural networks [17,18] used to implement and tune control algorithms have been treated as out of scope in this paper due to the reasons analogous to these pointed out in [19]. In the latter reference, a pure definition of a model-free approach is stated as follows: "However, we question most of these claims on the basis of the hidden or implicit system modeling required. (…) For example, fuzzy controllers are frequently claimed as model-free; nevertheless, in all fuzzy controllers there is a requirement for a rules base (or associative memory matrix) that describes the dynamics of the system in a linguistic-type fashion." The authors decided to characterize current state of the art and give the background to the proposed original approach in the context of two main optimization strands: bio-inspired metaheuristics and global deterministic optimization methods.
In the past few years, there has been a noticeable increase in interest in bio-inspired methods in various applications [20]. These metaheuristic methods independently of their major drawbacks (no guarantee to find the global optimum of the cost function, necessary prototyping of solutions using computer simulations, and—usually—a long calculation time) are used nowadays in both model-based and model-free controller tuning procedures. As an example, Ref. [21] proposed fixed-gain controllers in PID type for a four-rotor coaxial mathematical model with multiple additional effects taken into account, such as drag or friction forces in a PSO (Particle Swarm Optimization) framework. The idea and the principles of this metaheuristics, as well as others, such as Ant Colony Optimization, Artificial Bee Colony, Glowworm Swarm Optimization, Bacteria Foraging Optimization, Bat-inspired algorithm, have been presented in [22], which, in a wide context, presents computational bio-inspired algorithms applied in the field of UAVs.
In this paper, the authors propose an alternative approach to the methods listed in the above paragraph to perform automated tuning of controllers, using deterministic optimization methods. It is based on the assumption that the obtained solution for a specific type of a cost function is a global minimizer with prescribed tolerance, calculated in a defined time regime. From multiple available approaches to tuning problems, easily-implementable iterative optimization methods have been selected and expected to allow either tuning or fine-tuning of controllers during flight of UAVs. Their computational complexity and conceptual simplicity is an advantage in comparison with multi-agent based optimization techniques, where time necessary to conduct all computations is proportional to the number of agents. The approach in this paper is inspired by a classic Iterative Feedback Tuning method [23,24,25] in which optimal controller parameters are sought in reference to some cost function calculated using a current output signal of a closed-loop system. The references mention the latter method among such approaches as Fictitious Reference Iterative Tuning [26,27,28], Iterative Learning Approach [29,30] or Virtual Reference Feedback Tuning [31]. The latter methods enable direct tuning based on experimental input-output data. Direct methods can be divided into classes such as zero- or first-order optimization techniques. The first class uses only zero-order interval arithmetic, as in the bisection, golden-search or Fibonacci-search method, and the second one uses interval gradients or interval slopes, as in IFT or ILA, Steepest Descent Method, Reinforcement Learning [32] or methods proposed in [30,33].
The literature terms zero-order optimization methods alternatively as derivative-free optimization methods.
The analysis of the literature indicates that there are just few papers so far which discuss the use of golden-search and Fibonacci-search methods in the context of derivative-free optimization used for control purposes. This might be due to the reason that these methods are considered to be slower in comparison with gradient-based approaches [34]. In addition, as stated in [35], "there has recently been renewed interest in optimization problems with only functional (zero-order) information available rather than first-order gradient information in optimization, machine learning, and statistics", where the reference to [36,37,38] is given.
Ref. [39] presents a novel strategy of maximum power point tracking for photovoltaic power generation systems based on the Fibonacci-search technique. The aim of the paper has been to realize a simple control system to track the real maximum power point even under non-uniform or for rapidly changing insolation conditions. In the article [40], a practical PI/PID controller tuning method for integrating processes with dead time and inverse response based on a model has been proposed. Optimum tuning of the parameter for disturbance rejection based on the model and the minimum of the IAE criterion has been successfully obtained via adaptation of the golden-search technique. However, the approach presented in the current paper aims to result in obtaining an iterative tuning scheme of controller parameters independently of the availability of the model, as shown in [41], where, to the best knowledge of the authors, there is only a single reference available in the literature referring to the concept of using Fibonacci-search method in iterative model-based PID controller tuning. Ref. [41] unfortunately delivers only a blurry and general procedure for possible search in parameter space stating that the reciprocated Fibonacci sequence which has been used to adjust the controller parameters has been chosen arbitrarily because it is a convenient, convergent sequence. The paper does not present in full the potential and possible advantages of the use of this algorithm in controller tuning procedures, which is the basic reason for writing this paper and formulating further research.
In the presented approach, the authors went beyond the topics of the above-mentioned papers in several aspects. Ordering the novel ideas presented in the paper: the added value to this work is the iterative tuning procedure, seeking for the minimum of a function, of a time-invariant controller with small number of parameters, i.e., 1–3, for either slow or fast processes, where availability of a plant model is not required. The controller is assumed to have a constant structure, with its parameters changed when the tuning algorithm runs. It results in a modified behaviour of the closed-loop system, though selecting the gains in an orderly manner, i.e., according to the presented algorithm, a similar search procedure is obtained as one of a human, who seeks for the best gains. The considered tuning procedure is, however, fully automated. As it does not need indeed in the ideal case any model (see comments to Remark 4—the optimality will be henceforth used to present optimality of a function, not control), the final decision in the worst case is locally optimal, in the best case—globally optimal, depending on the characteristics of the environment and the proper method chosen. The model free feature of the presented tuning method is, actually, its main novelty, and the greatest difference from the existing tuning methods. Knowing safe ranges of parameters (ensuring stability), this is simply a plug-and-tune method. The method is designed to enable tuning with averaging the already-considered pairs of gains along multiple iterations, which results in diminishing the outlier impact on the final tuning result, or to include a low-pass filter to the signal which is used to build the cost function. Both of the above modifications of the method allow one to fit it to the point of its application. Sample tuning of a PD controller has been theoretically described and verified by means of experiments in controlling altitude in flight conditions of a hexacopter UAV, where the Fibonacci-search method is used together with a bootstrap approach in obtaining optimal controller parameters in a duration time-limited experiment. In addition, three modifications of the method have also been proposed and verified, aiming at increasing both the speed and accuracy of the procedure of obtaining controller parameters. During implementation, a low-pass filtering has been used and tested to verify robustness of the procedure against measurement noise both in simulations and during experiments using a hexacopter UAV in real-world conditions.
The paper is structured as follows: a brief statement of the considered practical tuning problems regarding UAVs controllers is summarized in Section 3. In Section 4, the Fibonacci-search Gain Tuning method is described. Subsequent subsections present a full description of the theoretical basis necessary to understand and implement the method both in flight simulators and systems embedded in UAVs. The testing platform is presented in Section 5. Numerical and hardware experiments are presented in Section 6 followed by the authors’ conclusions in Section 7. Finally, the Appendix gives a detailed message printed out of the console obtained during the experiment of minimum-seeking tuning of controller parameters described in the paper, concerning real-world hexacopter UAV robot using the FGT method.

3. UAV Applications—Practical Problems

The authors of the paper are mainly interested in applicability of the FGT method in the field of multirotor UAVs operating in various tasks and on different missions [42,43,44,45,46,47], including multi-UAV applications, where the highest level of reliability, safety and maneuvering precision is required. To shed a light on the scope of possible applications, a general idea has been presented in Figure 1 comprising a set of current applications. Each of the applications requires the use of effective position and orientation controllers. Furthermore, prior experience of the authors gained from the MBZIRC 2017 (Figure 1b,c) stresses the need to be able to stabilize the robot at a given altitude and to control its autonomous vertical maneuvers and landing, especially in conditions with outdoor-like disturbances, at the highest level of precision. This is the core problem that has been encountered in the field of robotics in UAVs for years, both in RC-controlled flights of simple, low-cost constructions as well as in advanced autonomous systems. An application of fixed-parameter controllers, such as PDs or PIDs, is very often connected to tuning which mirrors a compromise to find parameters for drastically different flying conditions accompanied by dynamically changing altitude during flight. For example, one type of a controller response is needed during the take-off phase in acrobatic contests, another one for precise landing and yet another one when recording smooth video footage or during manipulating with objects lifted by the UAV. This explains why it is desired to ensure the possibility to tune or to change controller parameters of UAV during flight in as short time as possible, and for the problems connected with prototyping new machines—their fine-tuning.
The problem considered in the paper is confined to finding the locally-best parameters of a PD controller based on a given reference altitude primitive and measured altitude in a predefined accumulative cost function which is calculated in flight conditions for a hexacopter UAV with given safe ranges of controller parameters. This enables the procedure to find controller parameters, taking preferred dynamics and initial controller parameters into account, for the given tolerance of the solution. The tuning method is used in a control system with a convex performance index in a form of a sum of absolute values of the tracking error. In the usual case, and when averaged over a longer time horizon, the majority of controller design tasks present their convex nature. This is the main point of applicability of the presented method. However, when run a single time in a highly-disturbed environment, the method will definitely result in optimal, though locally optimal solution. As the presented algorithm revolves around certain points in a search space, cropping the ranges from two opposite sides alternating the new ending point, a single missed decision, when impeded by disturbances, is compensated by the decisions to follow, which will be presented in due course of the paper.

4. Fibonacci-Search Gain Tuning

4.1. Preliminaries and Notation

The optimization procedure is based on the following assumptions:
  • the model/mathematical description of the dynamical system is unknown/imprecise or does not simply allow tuning the controller parameters by analytical calculations;
  • there is no formula describing the cost function based on which it can be verified whether the solution to the problem, i.e., controller parameters, is optimal;
  • the value of the cost function is based on observing a performance index in a given time horizon, and can be obtained repeatedly by performing consecutive experiments during the same flight;
  • the optimal solution (minimizer) in the form of optimal parameters of a controller is obtained in an iterative manner;
  • a zero-order iterative algorithm (a branch and bound algorithm) has been chosen to find the solution.
The notation used in the paper to generally describe the FGT algorithm, is summarized in Table 1.

4.2. The Idea behind Zero-Order Iterative Methods for Finding Minimum Points

The following two definitions are considered here.
Definition 1.
A function f ( x ) is unimodal if for a value m, it is monotonically decreasing for x m and monotonically increasing for x m . In that case, the minimum value of f ( x ) is f ( m ) and are no other local minima exist.
Definition 2.
The presented method applies to finding the minimum of a function f : R R expected to lie in the range x ( 0 ) , x ( 0 + ) , where x ( 0 + ) > x ( 0 ) , f is expected to be a unimodal function, and the argument x of a function f can be, e.g., a parameter of a controller.
If the conditions from the above definition are violated, and the function has multiple minima, the method will find a local minimum of a multimodal function. In order to obtain the global minimum, the method should be initialized from different starting points, with cost functions compared between local minima. On the basis of the performed experiments, and observations of their results from several runs of the method, it is possible. Please refer to the case presented in Appendix A, where, in the case of a two-parameter controller with fixed parameters, the method has been initialized from various starting points (Par2). The same can be done for Par1, and on the basis of expert knowledge, by changing a range for Par1 (see Par(−) and Par(+)). It is recommended to repeat the tuning procedure an arbitrary number of times to make the result closer to the global one.
Theorem 1.
Let a unimodal function f be given (see Figure 2) and considered in the following range of its arguments: x ( 0 ) , x ( 0 + ) . The information concerning current value of f can be obtained for any value of its argument with no use/no information concerning a possible reduction of the considered range, where the minimum lies.
Proof. 
In the procedure, two intermediate points to evaluate f are taken to symmetrically reduce the range of possible arguments, i.e.,
x ( 1 ) x ( 0 ) = x ( 1 + ) x ( 0 + ) = ρ ( x ( 0 + ) x ( 0 ) ) ,
where a reduction factor ρ < 1 2 , as in Figure 3.
When the values of f at the two intermediate points are known, the following cases apply:
  • f ( x ( 1 ) ) < f ( x ( 1 + ) ) → the minimum is in the range x ( 0 ) , x ( 1 + ) (see Figure 4);
  • f ( x ( 1 ) ) f ( x ( 1 + ) ) → the minimum is in the range x ( 1 ) , x ( 0 + ) .
 ☐
Proposition 1.
In the second iteration, two intermediate points are chosen, i.e.,  x ( 2 ) and x ( 2 + ) , function values are evaluated again, and the algorithm runs in a loop on the basis of the branch-and-bound procedure. The function f does not have to be neither differentiable nor continuous in the range x ( 0 ) , x ( 0 + ) , but can be simply obtained, e.g., by measuring signals and incrementing some performance index.
Proposition 2.
For the initial range x D ( 0 ) = x ( 0 ) , x ( 0 + ) , the branch-and-bound algorithm is as follows:
  • evaluate the minimal number of iterations N for which the difference between true minimum x * and iterative solution x ^ * (it is assumed that it is in the middle of the range D ( N ) ) does not exceed the prescribed relative accuracy ϵ, where
    | x * x ^ * | ϵ ( x ( 0 + ) x ( 0 ) ) ,
  • for k = 1 , , N :
    (1) 
    pick two intermediate points x ^ ( k ) , x ^ ( k + ) ( x ^ ( k ) < x ^ ( k + ) , x ^ ( k ) , x ^ ( k + ) D ( k 1 ) ) from the current range D ( k 1 ) ;
    (2) 
    obtain the new range D ( k ) selecting its limits as:
    (a) 
    for f ( x ^ ( k ) ) < f ( x ^ ( k + ) ) , x ( k + 1 ) D ( k ) = x ( k 1 ) , x ^ ( k + ) ;
    (b) 
    for f ( x ^ ( k ) ) f ( x ^ ( k + ) ) , x ( k + 1 ) D ( k ) = x ^ ( k ) , x ( k 1 + ) ;
    (3) 
    put k : = k + 1 ;
  • assume that x ^ * = 1 2 ( x ( N + ) + x ( N ) ) is the solution to the problem.

4.3. Fibonacci-Search Method Concept

Assumption A1.
Let it hold that in the current range D ( 0 ) f ( x ( 0 ) ) < f ( x ( 0 + ) ) ; then, x * D ( 1 ) = [ x ( 0 ) , x ( 1 + ) ] .
Proposition 3.
Since the point x ( 1 ) is in the new range D ( 1 ) , the value of f ( x ( 1 ) ) is already known from the previous iteration (a potential reduction of the computational burden of the method), then x ( 1 ) should be chosen to be at x ( 2 + ) . In such a case, it is necessary to evaluate f at a single point only, i.e.,  x ( 2 ) .
This can be achieved by a special choice of ρ k and ρ k + 1 between iterations, which is shown in Figure 5 and Figure 6 and explained in the further parts of the text.
Definition 3.
Let us choose the reduction factor which at k-th iteration is ρ k , and in the next one, becomes ρ k + 1 , with 0 ρ k 1 2 .
According to the above,
ρ k + 1 ( 1 ρ k ) = 1 2 ρ k ,
and
ρ k + 1 = 1 ρ k 1 ρ k .
Remark 1.
There are many sequences ρ 1 , ρ 2 , …, that satisfy the above-mentioned conditions and 0 ρ k 1 2 , however:
ρ 1 = F N 1 F N + 1 , ρ 2 = F N 2 F N , ρ k = F N k F N k + 2 , ρ N = F 0 F 2 = 1 2 ,
where F 0 = 1 , F 1 = 1 , F 2 = 2 , F 3 = 3 , F 4 = 5 , F 5 = 8 , F 6 = 13 , F 7 = 21 , … are Fibonacci numbers with F k = F k 1 + F k 2   for the optimal sequence, resulting in maximal range reduction [48,49].
Remark 2.
At the N-th iteration, it holds that ρ N = 1 2 , and the two intermediate points are chosen to be the same, making a decision regarding where the optimum lies impossible. To avoid this problem, the N-th reduction factor is modified to ρ N = F 0 F 2 δ = 1 2 δ , where 0 < δ 1 .
Proposition 4.
The required number of iterations N to achieve selected accuracy ϵ satisfies F N + 1 1 ϵ , and the two intermediate points are selected as below [50]:
x ^ ( k ) = x ( k 1 ) + ρ k ( x ( k 1 + ) x ( k 1 ) ) ,
x ^ ( k + ) = x ( k 1 ) + ( 1 ρ k ) ( x ( k 1 + ) x ( k 1 ) ) ,
where
ρ k = F N k F N k + 2 .

4.4. Sketch of the Algorithm for Optimal Tuning of a Two-Parameter Controller

Assumption 2.
Suppose we need to tune two parameters of a controller, i.e.,  P 1 and P 2 , as an example k P and k D in a continuous-time PD controller.
Proposition 5.
The algorithm in the case of a two-parameter controller is composed of the following steps:
(0) 
calculate/estimate allowable ranges of P 1 and P 2 for the given dynamical system, ensuring stability of the closed-loop system (see Remark 3);
(1) 
define initial value of P 2 ( 0 ) ;
(2) 
using a sequence of Fibonacci-search iterations for defined tolerance ϵ and k = 0 , implement the followingbootstrappingtechnique (put P 2 ( k + 1 ) = P 2 ( k ) ):
(2a) 
starting with the initial range for P 1 and fixed P 2 ( k + 1 ) , find by means of the Fibonacci method the optimal P ^ 1 * ( k + 1 ) , and proceed to the step 2b;
(2b) 
starting with the initial range for P 2 and fixed P 1 ( k + 1 ) = P ^ 1 * ( k + 1 ) , find by means of the Fibonacci method the optimal P ^ 2 * ( k + 1 ) , and proceed to the step 2c;
(2c) 
if the updated point P ^ 1 * ( k + 1 ) , P ^ 2 * ( k + 1 ) has already been found in the past iterations, stop the algorithm (no improvement is possible anymore); otherwise, enter k : = k + 1 and proceed to the step 2a.
The step 2c does not have to be implemented when the number of bootstraps is assumed and the number of iterations N of the Fibonacci method is known (see Figure 7). When 2c is omitted, the algorithm has a deterministic running time.
Remark 3.
Recalling research motivation from Section 3, the proposed method to perform auto-tuning would be even more suitable in the case of fine-tuning of nominal controller parameters aiming at improving its responsive capabilities to different dynamics requirements than to its prototyping from a scratch. The latter is the results of no need to have an analytical criterion using knowledge from a model of UAV, allowing for obtaining a good estimate of the search space for the given controller, ensuring stability of the control system. From the implementation point of view, and in order to ensure the safety of the tuning procedure, it is recommended to start the tuning from narrow deviations with respect to the nominal controller parameters and widen then along the way, repeating the procedure, if necessary.
Remark 4.
On obtaining initial ranges for the considered parameters.
The step where allowable ranges for the considered parameters is of prime importance and requires the expert knowledge obtained either from a series of flights with a set of different gains, and conclusions drawn from them, or any analytical model of the UAV available. At this point, however, the method would require any initial knowledge concerning the model of the used UAV unit. When fine-tuning is considered, this model is usually not needed, which has been verified by our experiments.
In order to obtain a range for the considered parameters, the following approaches might be adopted:
(1) 
obtain a nominal model of the UAV, e.g., for altitude changes only, as one of [51], include uncertainty information, see [52,53,54,55], and perform stability analysis of the uncertain system, yielding estimates of safe ranges, stabilising the closed-loop system,
(2) 
for a simulation-developed model of the UAV, as used in the paper in ROS-Gazebo environment, obtained, e.g., by designing a model of the UAV in such software as Autodesk Inventor Professional 2015 (Dyson, 9.0.23.0, Autodesk, San Rafael, CA, USA), perform a series of simulations to observe the behaviour of the model in a simulation environment, as in [56],
(3) 
obtain a linearized model of the UAV for the expected motion, assume possible disturbances, such as wind gusts, as bounded, an obtain stability analysis as in [57] considering ranges for controller parameters,
(4) 
heuristic approach: set the ranges close to the nominal parameters of the controller, which proved to stabilise the UAV in real-world experiments.
The stability analysis is a major requirement to fly the UAV safely in any environment, and should be performed prior to performing any experiments. As stated, however, above, it does require any knowledge about the model, in such an approach.
On the other hand, from the adaptive control viewpoint, based on the works of [58,59], one can use dynamic programming techniques, presented also in [60], to perform optimality and stability analyses.

4.5. Stopping Criteria for Iterative Algorithms

Proposition 6.
As a part of the step 2c from the previous algorithm, the following sample stopping criteria might be potentially considered:
  • theoretical tests (in general, for x ̲ R n )
    | f ( x ̲ ( k ) ) f ( x ̲ * ) | < ϵ 1 , x ̲ ( k ) x ̲ * < ϵ 2 ,
  • practical stopping criteria
    | x i ( k + 1 ) x i ( k ) | ϵ i
    or
    | f ( x ̲ ( k + 1 ) ) f ( x ̲ ( k ) ) | < ϵ 1 ,
where x ̲ ( k ) denotes the solution at k-th iteration of the algorithm, and x ̲ * is the true minimum of the cost function  f ( x ̲ ) .
However, in this approach, the number of iterations is limited by the expected number of bootstraps and expected accuracy of the result, which are functions of the duration of the experiment.

4.6. Outline of the Implementation of the Algorithm

The method is iterative-based and collects information about the performance index (on incremental cost function value) at sample time instants, equally spaced every T S seconds. Assuming the sampling period is sufficiently small, a single main iteration of the method is initialized N max times where:
  • for i = 1 , , N c 1 and with the controller parameters updated in the previous iteration, the performance index is evaluated as
    J ( i ) = J ( i 1 ) + Δ J ( i ) ,
    where Δ J ( i ) might be, e.g., chosen as Δ J ( i ) = | e i | corresponding to the i-th sample of the tracking error at time t = i T S or Δ J ( i ) = e i 2 + u i 2 corresponding to the sum squared values of the tracking error and control input to the plant;
  • for i = N c , a single iteration of the Fibonacci method is initialized—the cost function (performance index) value is stored, and if it is possible to compare the values of cost functions in a given range, i.e., either if the two intermediate points have been evaluated, the range has been reduced or the optimal solution has been found and bootstrapping takes place—either way at this point controller parameters change (are updated), which results in transient behavior of the dynamical system,
  • for i = N c + 1 , , N max , no action is taken (the parameters of the controller have been updated, no performance index is collected), and these steps are intended to allow the closed-loop system to stabilize at some point (to decay the transients), in order to allow performance index evaluation during the next main iteration (starting again from i = 1 ).
Thus, the experiment-based tuning process takes N max · T S seconds, which should be correlated with the reference signal changes. Having assumed that a complete run for the Fibonacci method takes N steps and two parameters are changes, every bootstrap stage takes 2 N · N max · T S seconds. This time horizon can be reduced if past the first iteration of the Fibonacci method, a single intermediate point is evaluated only, with the other taken from one of the previous iterations.

5. Experimental Platform

The experiments have been carried out using a customized micro aerial vehicle platform (see Figure 1). The considered UAV composed of a DJI hexacopter frame F550 (Shenzhen, China) and E310 DJI motors, UAV PixHawk Autopilot, Mobius ActionCam Lens C2 and Matrix Vision mvBlueFOX-MLC200w cameras, Real Time Kinematic GPS, Rangefinder TeraRanger One sensor and an onboard Intel NUC-i7 PC (Santa Clara, CA, USA) has been developed for the purposes of the Mohamed Bin Zayed International Robotic Challenge 2017, where its high performance has been proven in Treasure Hunt and Landing tasks [8,61].
In the present experiments, the Gripper module is detached. Its use in carrying load will be a subject of further, separate, tests. It is also planned in the future to apply the FGT method to automatically tune the parameters of the altitude controller and the parameters of orientation controllers (roll, pitch, and yaw angles) during flight with attached load with different masses. However, this work focuses on automatic and minimum-seeking tuning of parameters of an altitude controller in a UAV during flight to track reference signals with different dynamics, named, respectively, slow (S), medium (M), fast (F)—see Figure 8. As a result, and having applied the FGT method, three sets of controller parameters are obtained which can be used to translate between them to obtain smooth and stable altitude changes. This is particularly useful when a sharp picture has to be obtained from video cameras to, subsequently, analyze it or to perform fast autonomous landings from a set altitude, e.g., on a moving vehicle, as in Figure 1b.
By performing the fusion from on-board sensors, an estimation of the UAV states can be performed, which is an important part in this UAV system to enable its precise and reliable control at high speeds. The current architecture is based on a combination of an Extended Kalman Filter (EKF) which is implemented in PixHawk, gathering the data from GPS and the Inertial Measurement Unit (IMU), magnetometer, and barometer, as well as the output from the Rangefinder using a Linear Kalman Filter (LKF) in the onboard computer.
Since the control system must be capable of ensuring good performance at large deviations from the hover configuration or to compensate heavy wind, a nonlinear controller is used. The UAV has torque m ̲ R 3 and thrust force control f R as inputs, thus based on [62,63]
m ̲ = k R e ̲ R k Ω e ̲ Ω + Ω ̲ × J Ω ̲ J Ω ̲ ^ R T R c Ω ̲ c R T R c Ω ̲ ˙ c ,
f = k P e ̲ x k i b R 0 t R ( τ ) T e ̲ x d τ k i w 0 t e ̲ x d τ k D e ̲ v m g e ̲ 3 + m x ̲ ¨ d R e ̲ 3 ,
where J R 3 × 3 is the inertia matrix with respect to the body frame, R R 3 × 3 is the orientation of the center of mass matrix, R c R 3 × 3 is the commanded value of the inertia matrix with respect to the body frame, Ω ̲ R 3 is the angular velocity vector in the body fixed frame, Ω ̲ c R 3 is the commanded angular velocity vector in the body fixed frame, x ̲ ¨ d the desired acceleration, and k P , k D , k R , k Ω positive definite terms.
The e x , e v , e R , and e Ω denote, respectively, position, velocity, orientation, and angular rate errors, with the subscript · C denoting the commanded value. In our approach, the gains k P and k D are subject to tuning.
The control design, i.e., the controller, is a well-known structure, compensating multiple nonlinear effects known from UAV dynamics, and previously described in multiple papers of the authors, e.g., [8,61]. It has been assumed that, in order to stress the main point of the paper, and its main novelty, the section presenting the controller has been reduced to the necessary minimum. The controller design does not result from control point of view, but from the nonlinear dynamics of the plant.
The combination of ROS and the Gazebo environments has also been used at the stage of prototyping the solutions and at the stage of simulation tests of the FGT methods. Gazebo under ROS is a respected robotic simulator, which can be used for simulation-in-the-loop (SITL) together with a firmware from PixHawk. The results of simulation and experimental tests selected and presented in the paper, to the best knowledge of the authors, are most representative, and are described later in the text. The experiment with a real robot and a computational time analysis reflects the applicability of the FGT method to real problems. A video reporting one of the experiments is attached to this paper and can be found in the Mathematical Models Database (http://mathematicalmodels.put.poznan.pl) [64]. In addition, it is planned in the near future to make the overall system with implementation of all method variants public in the MMD together with numerous data sets available as ROS-files for testing the system in a Gazebo simulator. Further information regarding hardware used is available in the previous author’s papers, i.e., [8,65] and in previously self-cited papers, therein. A detailed description of hardware and software solutions is presented there, including path planning, low- and high-level control, image recognition, etc. In addition, one may find therein an extended description of the implemented and tested precise multi-level estimator of the altitude of the UAV, which fuses data from GPS and a laser Rangefinder sensor to enable flying smoothly above a vehicle and landing on it. Finally, the authors provided a description of how they integrated an estimator of velocity and position of the vehicle into the MPC controller of the UAV to be able to achieve fast landing on a moving object. The developed approaches and gained experience during the MBZIRC 2017 provide the starting point for gaining results from the tests in further parts of the text.

6. Experimental Results

6.1. Introduction

In this section, the performance of the FGT method as well as its experimental verification are presented (for various scenarios), referring to the theoretical results from Section 4.

6.2. ROS Implementation of the FGT Algorithm

For the model of the DJI F550 hexacopter developed in Gazebo, the described FGT algorithm has been implemented under the ROS control for tuning of the two parameters, namely k P and k D , given their ranges and the initial value of the second parameter. These values have been taken on the basis of examination of the results of manual tuning of the UAV in manual flight conditions.
In real world applications, in a disturbed environment, and when control signals are constrained, as well as the feasible dynamics of the UAV, the most appealing control law from the theoretical point of view might give unacceptable results when the gusts of wind will be greater than those claimed to be in safe ranges. The authors do not perform here any stability analysis, making an assumption, that the ranges for sought parameters are ’safe’, in the sense that any combination of them does not cause the closed-loop system to become unstable. This is a realistic assumption, as operators of UAVs usually try to fly their machines selecting a set of controller parameters, and based on them, a smallest set of gains can be formulated, leading to searching the locally-best gains between the stabilizing parameters of the controller.
The main loop of the program is executed with the frequency of 100 Hz , whereas the tuning algorithm with 5 Hz (i.e., every 200 ms ). The syntax of the method is as follows:
update_parameters_FIB(input,main_iteration_counter,P1_range,P2_range,N_c,N_max,Par_initial,method,output)
with the following parameters:
  • input—the current value of the increment of the performance index, i.e., the tracking error or the low-pass filtered tracking error;
  • main_iteration_counter—the variable referring to the number of iterations of the algorithm, here: 48;
  • P1_range—the range for the first tuned parameter, here [ 4 , 12 ] ;
  • P2_range—the range for the second tuned parameter, here [ 2 , 7 ] ;
  • N_c—the number of samples when the performance index is collected (see Figure 9), here 50;
  • N_max—the number of samples corresponding to the length of the trajectory primitive, here 60;
  • Par_initial—the initial value for the second parameter, here treated as variable;
  • method:
    • 0—the performance index is calculated at every step;
    • 1—the performance index is averaged over all past measurements with the same parameters;
    • 2—the performance index is evaluated only for parameters that have been unconsidered before (the length of the tuning procedure is reduced);
  • output—the vector of output variables, including tuned gains.
The designed method has the following features:
  • comprises two bootstraps;
  • ϵ = 0.05 (see Figure 10);
  • the number of iterations in a single bootstrap is 2 × 2 × 6 = 24 (each iteration is composed of two evaluations of performance indices);
  • δ = 0.01 (Fibonacci–search);
  • requires 48 loops of a single trajectory primitive.
In order to change the trajectory primitive, the N_max multiplicity of the time interval (multiplicity of the inverse of the frequency of executions of the tuning method) must be equal to the duration of the trajectory primitive. In this way, it can be changed, e.g., extended, for other regimes of work of the UAV, such as tuning at high speeds.
For example, when the trajectory primitive has 12 s duration and the experiment consists of 48 iterations (two bootstraps in two parameters), the total length of the experiment is 576 s . For the sampling period T S = 0.2 s , there are 60 samples for one period of the trajectory primitive ( N c = 50 , N max = 60 ). The sample experimental result can be found in the Appendix A. In Appendix B, a detailed explanation of the tuning process from the first iterations is included.

6.3. ROS Gazebo Simulation Results

To mimic the test presented in Section 6.2 (see Figure 11), a few hundred simulation results have been gained in various configurations of the FGT algorithm, with results of 90 tests selected (presented in Table 2). It must be noted that some initial values of the parameters have been intentionally selected as out of range to show the proper and desired behavior of the FGT method. These results form a base to formulate comments concerning the procedure and draw conclusion, and serve as a good starting point for conducting experiments on a real UAV. The FGT algorithm has been tested in different configurations and variants with great attention paid to the efficiency of obtaining PD controller parameters and the performance of the obtained tracking properties with the reference altitude supplied to the control system. The following have been tested:
(i)
accuracy of the obtained controller parameters for reference primitives with different dynamics: fast ( v max = 10 m s , a max = 100 m s 2 ), medium ( v max = 3.33 m s , a max = 11.11 m s 2 ), slow ( v max = 1.67 m s , a max = 2.78 m s 2 ) reference primitives,
(ii)
the impact of the initial value of the second parameter of the controller (Par_initial) on the convergence of the FGT procedure,
(iii)
ability to reduce tuning errors using a low-pass first-order filter in the variants listed above where a low-pass filter recursive equation is proposed to diminish the impact of the measurement noise on the results of optimization.
Each version of the FGT procedure has been tested in two ways: with and without low-pass filtration, which gives six independent results (versions) of the FGT method. The low-pass filter is assumed to have the transfer function
G ( s ) = 1 1 + 0.1 s ,
and to be discretized with sampling period T S = 0.01 s . The resulting recursive equation of the low-pass filter is given as
y n = a n 1 + 1 a u n 1 ,
where the constant parameter
a = exp T S / 0.1 ,
with y ( n ) and u ( n ) as filtered and pure errors at sample n, respectively.
In all of the 90 presented results (remaining configuration parameters as in Section 6.2), the incremental cost function has been the sum of absolute values of the tracking error:
J = i = 1 N c ( | e i | ) .
Summing up cost function values from Table 2, the lowest J is equal to 9.227 and has been obtained in the configuration 1–F–1 (method–reference primitive–filtration) for all three reference primitives and different initial points for the second parameter. This configuration of the method has superior performance in comparison with the others. The second best is the 2–F–1 configuration, when both dynamics M and F have given better tuning results than S.
The configurations listed in the above paragraph perform much worse in the case of no low-pass filtration (the highest total cost function values are 12.041 and 11.261, respectively). By performing low-pass filtration, the results (J) improve by 16.80%, giving 20 superior results in the case of filtration in comparison with two superior results in the case of no filtration. Similarly, seven worst results in comparison with 16 worse results in the case of no filtration. It is to be noted that omitting the filter deteriorates the results for F dynamics, and, in addition, the algorithm is more sensitive to the case when the initial point is out-of-defined-range.
On the basis of histograms of the final values of the following tuned parameters k P and k D presented in Figure 12, Figure 13 and Figure 14, it can be said that the controller parameters for the defined reference trajectory primitives result in similar values of tuned gains, where small differences between the parameters inform us that either the true minimum of the cost function is relatively flat, and extend some of the parameter ranges, or that the function has multiple minima. Differences between cost functions J from the last iterations result from disturbances acting on the system.
Using a low-pass filter for the measured tracking error signal, the method is less vulnerable to measurement noise, and the results might be treated as more accurate.
Overall results are close to those obtained in the configuration with a low-pass filter. Using unfiltered measurement data results in flat histograms, and instead of a single pair of controller parameters repeatedly obtained among simulations, neighboring parameters tend to be seen for every experiment. A normal-like density function (Gauss-like) can be observed in the case of values for k D , which could also be true for k P ; however, the upper value of the admissible range for this parameter turned out to be too small.
In further tests, the following configuration of parameters has been chosen:
  • method = 1,
  • filter = 1,
  • Par_initial = 1,
with the remaining conditions as presented in Section 6.2 and in Figure 11. Additional 30 tests have been carried out where every case has been tested 10 times for all three trajectory primitives (S, M, F). Table 3 presents the results of analysis of these tests based on the obtained data from ROS-Gazebo. Figure 15 depicts the results obtained from all three series of experiments, i.e., plots of functions J = f ( k P , k D ) and their surface approximations.

6.4. Tests in Flight Conditions

For the purpose of this publication, sample results have been obtained during one of the verification experiments carried out to autotune altitude controller parameters using the FGT method of the DJI F550 hexacopter in flight conditions. Real outdoor scenario experiments, analogous to these presented in Section 6.3 (see Figure 11), have been carried out using a real machine this time during an intensive experimental camp in the countryside of South Bohemia (the Czech Republic) (see the website: http://mrs.felk.cvut.cz/projects/mbzirc which is systematically updated by its authors with videos and photos documenting the experiments carried out during these camps). In the outlined experiment, the same configuration of the FGT as in Section 6.3 has been used, where in the following case:
  • method = 1,
  • filter = 1,
  • Par_initial = 1,
  • trajectory primitive (M).
The total time of a single experiment is 11 min, with 9.6 min taken by two full bootstraps of the Fibonacci-search algorithm. From all the results presented in the paper, it can be seen that the choice of the number of bootstraps, the number of main iterations and the choice of the accuracy of the Fibonacci-search method are not random and can be determined from the constraints resulting from the choice of voltage source. Based on the electrical capacity of the batteries used in the selected hexacopter UAVs, the maximum time of flight has been estimated for the planned tests. This allowed to perform two full bootstrap sequences. It has been assumed that the voltage level drop during the experiments is negligible, if the batteries are initially fully charged. It is a natural simplification leading to the assumption that the thrust force of the driving units, dependent on the voltage level, is constant during the flight. Due to the fact that in the selected optimization method, the time of termination of the optimization procedure is deterministic, the autotuning experiment can be divided into stages (bootstraps), which can be carried out independently. This allowed for minimizing the impact of the voltage drop on the tuning results.
In the discussed trial, the main point has been the evaluation of the convergence of the algorithm during the experiment, expressed by a reduction of the cost function after every iteration of the Fibonacci-search algorithm after the tuned parameter has been amended. To depict the performance of the FGT algorithms, the complete data set is presented in Figure 16, Figure 17 and Figure 18 obtained from the considered trial. The first bootstrap is particularly interesting (see Figure 16 and the output presented in the Appendix) where one can see that starting from iteration no. 10, the cost function value reduces to the obtained minimum every few iterations. In addition, it is obtained with totally different PD controller parameters, which supports previous statements concerning the properties of the cost function J = f ( k P , k D ) (Figure 15).
A set of pictures taken during the real-world experiment, in windy conditions, is presented in Figure 19. A sample clip from this tuning can be found in the Mathematical Models Database (http://mathematicalmodels.put.poznan.pl) [64].

7. Conclusions

In the paper, a new real-time model-free minimum-seeking autotuning method for controllers based on the Fibonacci-search algorithm is presented. This gain tuning strategy is introduced in the context of improving reference signal tracking during flight in multirotor UAVs. As discussed in the introduction, due to the variety of tasks or missions of UAVs, one can distinguish multiple dynamic conditions of flight. In these conditions, controller parameters must ensure superior performance of the control system. Simple, fixed-value controllers can be successfully used, provided that they are well-tuned. In this context, the FGT method should be of the greatest use. The following may be included among the most important properties of this algorithm:
  • it allows automatic tuning, in predefined ranges, of parameters of a controller that is widely used with a low number of parameters, such as of PD or PID type (if the method is extended to 3-parameter framework, which is not discussed in the paper, though, is the topic of the current research);
  • the resulting solution from the FGT algorithm is optimal in the sense of the optimality of a function, with respect to the prescribed accuracy, and the cost function might be, in general, arbitrarily chosen;
  • with respect to multirotor UAV altitude controllers, the optimal tuning process is carried out in real time and can be conducted either using a model of a UAV, or using a real robot—to the best knowledge of the authors, there are no such results reported in the literature up until now;
  • two-parameter controller tuning due to the use of the proposed Fibonacci-search algorithm is characterized with rapid convergence, as has been observed during experiments even past several iterations of the algorithm; the obtained controller parameters ensure visible improvement in the reference altitude tracking task in comparison to nominal controller parameters; the time required to obtain optimal parameters is considerably shorter in comparison to bio-inspired batch methods;
  • due to the structure of the algorithm, the FGT method can be used in a single run or performed in stages, in separate flights taking the maximum flight time limited by capabilities of the voltage source in a particular UAV; from authors’ observations, it can be said that provided the UAV has a maximum time of flight approximately equal to 20 min, and the tuning experiment takes about 9.5 min to terminate ( 576 s ) for smooth reference altitude tracking, the experiment can be performed in a single run; however, if a visible voltage drop is observed, resulting in a loss in the effective thrust force of the used driving units, the FGT approach should be used in stages, e.g., the first bootstrap in the first stage, and the second one in the next flight with new batteries;
  • the last and obvious feature of the FGT algorithm that has appeared as the result of the experiments is the safety of this method during autotuning process of the altitude controller of the hexacopter UAV in flight conditions; in all the tests, independently of the initial value of the second parameter, no stability loss of the UAV has been noted, which could be a potential threat for the operator, environment or the very robot; naturally, any initial/expert knowledge concerning nominal parameters of the controller that can be used during flight is of importance. These parameters can be fine-tuned in selected ranges using the FGT method; intuitively, to obtain these ranges for parameters can be extended in consecutive experiments or to use any rapid prototyping methods prior to the experimental stage.
This work has successfully demonstrated the capabilities, simplicity, utility and applicability of the FGT method. The above list presents the main contributions of this paper from the perspectives of UAV control. Furthermore, the conducted experiments show the performance of the FGT method and verify its basic properties in simulated and real environment. Challenges identified during the flight-tests have been outlined and discussed. Considering the six proposed variations of the method, the superior case is with the method equal to 1, using a low-pass filtration and where the performance index is averaged over all past measurements with the same parameters. This case is widely-discussed in the paper and presented in a statistical analysis based on the results obtained from ROS-Gazebo simulations, and with representative results obtained in real experiments. The latter results have been obtained in real outdoor scenarios during flight of a real hexacopter UAV. This variant of the FGT method has been equipped with a low-pass filter recursive equation to diminish the impact of the measurement noise on the results of optimization. During the experiments and simulation tests, the superior performance has been obtained for this particular configuration which highlighted a number of directions for future research. In particular, it is planned to compare this approach to the FGT method with other optimization methods based on golden-search and bisection algorithms.
This work was motivated by the MBZIRC 2017 competition and mainly by the challenges requiring precise altitude control, such as autonomous landing and object grasping in demanding outdoor conditions. In addition to the competition tasks, the results of this research are important in further exploitation of the MBZIRC system in multi-UAV tasks and in applications where a manual tuning approach cannot be used (for example, see [66]).

Supplementary Materials

Exemplary recorded video (with ROS bag) is available online at the webpage (http://mathematicalmodels.put.poznan.pl).

Author Contributions

Conceptualization, W.G. and D.H.; Methodology, W.G., D.H. and M.S.; Software, W.G., D.H., and T.B.; Validation, W.G. and D.H.; Formal Analysis, M.S.; Investigation, W.G., D.H.; Resources, W.G., D.H.; Data Curation, W.G., D.H.; Writing—Original Draft Preparation, W.G.; Writing—Review and Editing, W.G., D.H.; Visualization, W.G., D.H.; Supervision, M.S.; Project Administration, W.G., D.H.; Funding Acquisition, W.G. and D.H.

Funding

This research was financially supported as a statutory work of Poznan University of Technology (04/45/DSPB/0184). The work has been supported by CTU grant no. SGS17/187/OHK3/3T/13 and the Grant Agency of the Czech Republic under grant no. 17-16900Y.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
EKFExtended Kalman Filter
FGTFibonacci-search Gain Tuning
GPSGlobal Positioning System
IAEIntegral Absolute Error
IFTIterative Feedback Tuning
ILAIterative Learning Approach
IMUInertial Measurement Unit
ISEIntegral Squared Error
ITAEIntegral Time-weighted Absolute Error
ITSEIntegral Time-weighted Squared Error
LKFLinear Kalman Filter
MBZIRCMohamed Bin Zayed International Robotics Challenge
MPCModel Predictive Control
MMDMathematical Models Database
PDProportional-Derivative type controller
PIProportional-Integral controller
PIDProportional-Integral-Derivative type controller
PSOParticle Swarm Optimization
SITLSimulation-In-The-Loop
UAVUnmanned Aerial Vehicle

Appendix A

*************************************************************
  (c) Model-free autotuner (2 parameters)
      based on zero-order Fibonacci--search method
                             Dariusz Horla, Wojciech Giernacki
                             Poznan University of Technology
                             [email protected]
 *************************************************************
  accuracy of calculations (epsilon) = 0.05
  no. of bootstrap cycles            = 2
  total number of main iterations    = 48 (for method==0 or 1)
                                     < 48 (for method==2)
  *** WARNING: some performance indices might be averaged (method==1)
 Iter.   Par  |      Par(-)      Par(+) |        Par1        Par2 |      J
 -------------------------------------------------------------------------------
     1      1 |      4.0000     12.0000 |      7.0476      1.0000 |      3.2797
     2      1 |      4.0000     12.0000 |      8.9524      1.0000 |      2.0056
     3      1 |      7.0476     12.0000 |      8.9524      1.0000 |      1.8226
     4      1 |      7.0476     12.0000 |     10.0952      1.0000 |      1.6950
     5      1 |      8.9524     12.0000 |     10.0952      1.0000 |      1.7295
     6      1 |      8.9524     12.0000 |     10.8571      1.0000 |      1.2114
     7      1 |     10.0952     12.0000 |     10.8571      1.0000 |      1.3400
     8      1 |     10.0952     12.0000 |     11.2381      1.0000 |      0.8479
     9      1 |     10.8571     12.0000 |     11.2381      1.0000 |      0.9715
    10      1 |     10.8571     12.0000 |     11.6190      1.0000 |      0.5721
    11      1 |     11.2381     12.0000 |     11.6114      1.0000 |      0.6537
    12      1 |     11.2381     12.0000 |     11.6267      1.0000 |      0.9414
 -------------------------------------------------------------------------------
    13      2 |      2.0000      7.0000 |     11.4734      3.9048 |      0.5626
    14      2 |      2.0000      7.0000 |     11.4734      5.0952 |      0.7648
    15      2 |      2.0000      5.0952 |     11.4734      3.1905 |      0.7458
    16      2 |      2.0000      5.0952 |     11.4734      3.9048 |      0.5389
    17      2 |      3.1905      5.0952 |     11.4734      3.9048 |      0.5103
    18      2 |      3.1905      5.0952 |     11.4734      4.3810 |      0.7227
    19      2 |      3.1905      4.3810 |     11.4734      3.6667 |      0.6390
    20      2 |      3.1905      4.3810 |     11.4734      3.9048 |      0.5790
    21      2 |      3.6667      4.3810 |     11.4734      3.9048 |      0.5618
    22      2 |      3.6667      4.3810 |     11.4734      4.1429 |      0.7059
    23      2 |      3.6667      4.1429 |     11.4734      3.9000 |      0.8414
    24      2 |      3.6667      4.1429 |     11.4734      3.9095 |      0.5088
 -------------------------------------------------------------------------------
    25      1 |      4.0000     12.0000 |      7.0476      3.9958 |      0.9767
    26      1 |      4.0000     12.0000 |      8.9524      3.9958 |      2.0221
    27      1 |      4.0000      8.9524 |      5.9048      3.9958 |      1.6874
    28      1 |      4.0000      8.9524 |      7.0476      3.9958 |      0.7902
    29      1 |      5.9048      8.9524 |      7.0476      3.9958 |      0.8292
    30      1 |      5.9048      8.9524 |      7.8095      3.9958 |      0.5437
    31      1 |      7.0476      8.9524 |      7.8095      3.9958 |      0.6552
    32      1 |      7.0476      8.9524 |      8.1905      3.9958 |      1.3442
    33      1 |      7.0476      8.1905 |      7.4286      3.9958 |      0.6063
    34      1 |      7.0476      8.1905 |      7.8095      3.9958 |      0.6940
    35      1 |      7.0476      7.8095 |      7.4210      3.9958 |      0.8598
    36      1 |      7.0476      7.8095 |      7.4362      3.9958 |      0.6038
 -------------------------------------------------------------------------------
    37      2 |      2.0000      7.0000 |      7.5742      3.9048 |      0.7084
    38      2 |      2.0000      7.0000 |      7.5742      5.0952 |      1.6431
    39      2 |      2.0000      5.0952 |      7.5742      3.1905 |      0.8694
    40      2 |      2.0000      5.0952 |      7.5742      3.9048 |      0.7246
    41      2 |      3.1905      5.0952 |      7.5742      3.9048 |      0.9771
    42      2 |      3.1905      5.0952 |      7.5742      4.3810 |      0.7666
    43      2 |      3.9048      5.0952 |      7.5742      4.3810 |      0.8260
    44      2 |      3.9048      5.0952 |      7.5742      4.6190 |      1.0384
    45      2 |      3.9048      4.6190 |      7.5742      4.1429 |      0.9957
    46      2 |      3.9048      4.6190 |      7.5742      4.3810 |      1.2615
    47      2 |      3.9048      4.3810 |      7.5742      4.1381 |      1.3135
    48      2 |      3.9048      4.3810 |      7.5742      4.1476 |      0.5283
 -------------------------------------------------------------------------------
 Tuning finished!!
The final results are:
gain1: 7.57, gain2: 4.26
Symbols Par1 and Par2 denote k P and k D , respectively, whereas Par defines the number of tuned parameter, and Par(−) and Par(+) denote x ( k ) and x ( k + ) , respectively, and, finally, J is the performance index value.

Appendix B

Consecutive calculations for the auto-tuning procedure presented in Appendix A
 -------------------------------------------------------------------------------
 Iteration I
 -------------------------------------------------------------------------------
 Range                          |    [4, 12]
 Reduction factor (rho)         |    8/21
 Left candidate point           |    4 + 8/21 * (12-4) = 7.0476
 Left-side cost function value  |    3.2797
 -------------------------------------------------------------------------------
 Iteration II
 -------------------------------------------------------------------------------
 Range                          |    [4, 12]
 Reduction factor (rho)         |    8/21
 Left candidate point           |    12 - 8/21 * (12-4) = 8.9524
 Left-side cost function value  |    2.0056
 -------------------------------------------------------------------------------
 DECISION STAGE: As 3.2797 > 2.0056, the range is modified to [7.0476, 12]
 -------------------------------------------------------------------------------
 Iteration III
 -------------------------------------------------------------------------------
 Range                          |    [7.0476, 12]
 Reduction factor (rho)         |    5/13
 Left candidate point           |    7.0476 + 5/13 * (12-7.0476) = 8.9524
 Left-side cost function value  |    1.8226
 -------------------------------------------------------------------------------
 Iteration IV
 -------------------------------------------------------------------------------
 Range                          |    [7.0476, 12]
 Reduction factor (rho)         |    5/13
 Left candidate point           |    12 - 5/13 * (12-7.0476) = 10.0952
 Left-side cost function value  |    1.6950
 -------------------------------------------------------------------------------
 DECISION STAGE: As 1.8226 > 1.6950 the range is modified to [8.9524, 12]
 -------------------------------------------------------------------------------
Remark A1.
At every stage, the estimated optimal point might be treated as the point in the middle of the considered interval. Thus, after the iterations 1–2 it is 8, after iters 3–4: 9.5238, after iters 5–6: 10.4762, after iters 7–8: 11.0476, after iters 9–10: 11.4285, and after iters 11–12 the point from the range [11.2381, 11.6267]: 11.4734.

References

  1. Liu, Y.; Rajappa, S.; Montenbruck, J.M.; Stegagno, P.; Bülthoff, H.; Allgöwer, F.; Zell, A. Robust nonlinear control approach to nontrivial maneuvers and obstacle avoidance for quadrotor UAV under disturbances. Robot. Auton. Syst. 2017, 98, 317–332. [Google Scholar] [CrossRef]
  2. Pounds, P.E.I.; Bersak, D.R.; Dollar, A.M. Stability of small-scale UAV helicopters and quadrotors with added payload mass under PID control. Auton. Robot. 2012, 33, 129–142. [Google Scholar] [CrossRef]
  3. Mahony, R.; Kumar, V.; Corke, P. Multirotor Aerial Vehicles: Modeling, Estimation, and Control of Quadrotor. IEEE Robot. Autom. Mag. 2012, 19, 20–32. [Google Scholar] [CrossRef]
  4. Chovancová, A.; Fico, T.; Hubinský, P.; Duchoň, F. Comparison of various quaternion-based control methods applied to quadrotor with disturbance observer and position estimator. Robot. Auton. Syst. 2016, 79, 87–98. [Google Scholar] [CrossRef]
  5. Espinoza, T.; Dzul, A.; Llama, M. Linear and nonlinear controllers applied to fixed-wing UAV. Int. J. Adv. Robot. Syst. 2013, 10, 1–10. [Google Scholar] [CrossRef]
  6. Espinoza-Fraire, T.; Dzul, A.; Cortés-Martínez, F.; Giernacki, W. Real-Time Implementation and Flight Tests using Linear and Nonlinear Controllers for a Fixed-Wing Miniature Aerial Vehicle (MAV). Int. J. Control Autom. Syst. 2017, 16, 392–396. [Google Scholar] [CrossRef]
  7. Loianno, G.; Spurny, V.; Thomas, J.; Baca, T.; Thakur, D.; Hert, D.; Penicka, R.; Krajnik, T.; Zhou, A.; Cho, A.; et al. Localization, grasping, and transportation of magnetic objects by a team of MAVs in challenging desert like environments. IEEE Robot. Autom. Lett. 2018, 3, 1576–1583. [Google Scholar] [CrossRef]
  8. Spurný, V.; Báča, T.; Saska, M.; Pěnička, R.; Krajník, T.; Loianno, G.; Thomas, J.; Thakur, D.; Kumar, V. Cooperative Autonomous Search, Grasping and Delivering in Treasure Hunt Scenario by a Team of UAVs. J. Field Robot. 2018. [Google Scholar] [CrossRef]
  9. Kohout, P. A system for Autonomous Grasping and Carrying of Objects by a Pair of Helicopters. Master’s Thesis, Czech Technical University in Prague, Prague, Czech Republic, 2017. [Google Scholar]
  10. Gąsior, P.; Bondyra, A.; Gardecki, S.; Giernacki, W.; Kasiński, A. Thrust estimation by fuzzy modeling of coaxial propulsion unit for multirotor UAVs. In Proceedings of the 2016 IEEE International Conference in Multisensor Fusion and Integration for Intelligent Systems (MFI), Baden-Baden, Germany, 19–21 September 2016; pp. 418–423. [Google Scholar] [CrossRef]
  11. Ziegler, J.G.; Nichols, N.B. Optimum settings for automatic controllers. Trans. ASME 1942, 64, 759–768. [Google Scholar] [CrossRef]
  12. Cohen, G.; Coon, G. Theoretical consideration of retarded control. Trans. ASME 1953, 75, 827–834. [Google Scholar]
  13. Panda, R. Introduction to PID Controllers—Theory, Tuning and Application to Frontier Areas; In-Tech: Rijeka, Croatia, 2012. [Google Scholar]
  14. O’Dwyer, A. Handbook of PI and PID Controller Tuning Rules, 3rd ed.; Imperial College Press: London, UK, 2009. [Google Scholar]
  15. Gautam, D.; Ha, C. Control of a Quadrotor Using a Smart Self-Tuning Fuzzy PID Controller. Int. J. Adv. Robot. Syst. 2013, 10, 1–9. [Google Scholar] [CrossRef]
  16. Zemalache, K.M.; Beji, L.; Maaref, H. Control of a Drone: Study and Analysis of the Robustness. J. Autom. Mob. Robot. Intell. Syst. 2008, 2, 33–42. [Google Scholar]
  17. Huang, S.N.; Tan, K.K.; Lee, T.H. Adaptive friction compensation using neural network approximations. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2000, 30, 551–557. [Google Scholar] [CrossRef]
  18. Huang, S.N.; Tan, K.K.; Lee, T.H. Adaptive motion control using neural network approximations. Automatica 2002, 38, 227–233. [Google Scholar] [CrossRef]
  19. Spall, J.C.; Cristion, J.A. Model-Free Control of Nonlinear Stochastic Systems with Discrete-Time Measurements. IEEE Trans. Autom. Control 1998, 43, 1198–1210. [Google Scholar] [CrossRef]
  20. Fister, D.; Fister, I., Jr.; Fister, J.; Šafarič, R. Parameter tuning of PID controller with reactive nature-inspired algorithms. Robot. Auton. Syst. 2016, 84, 64–75. [Google Scholar] [CrossRef]
  21. Giernacki, W.; Gośliński, J. Coaxial Quadrotor—From the most important physical aspects to a useful mathematical model for control purposes. In review.
  22. Duan, H.; Li, P. Bio-inspired Computation in Unmanned Aerial Vehicles; Springer-Verlag: Berlin, Germany, 2014. [Google Scholar]
  23. Hjalmarsson, H.; Gunnarsson, S.; Gevers, M. A convergent iterative restricted complexity control design scheme. In Proceedings of the 33rd IEEE Conference on Decision and Control, Lake Buena Vista, FL, USA, 14–16 December 1994; pp. 1735–1740. [Google Scholar] [CrossRef]
  24. Hjalmarsson, H.; Gevers, M.; Gunnarsson, S.; Lequin, O. Iterative feedback tuning: Theory and applications. IEEE Control Syst. Mag. 1998, 18, 26–41. [Google Scholar] [CrossRef]
  25. Lequin, O. Optimal closed-loop PID tuning in the process industry with the “Iterative Feedback Tuning” scheme. In Proceedings of the 1997 European Control Conference (ECC), Brussels, Belgium, 1–7 July 1997; pp. 3931–3936. [Google Scholar] [CrossRef]
  26. Soma, S.; Kaneko, O.; Fujii, T. A new method of a controller parameter tuning based on input-output data Fictitious Reference Iterative Tuning. In Proceedings of the 2nd IFAC Workshop on Adaptation and Learning in Control and Signal Processing (ALCOSP 2004), Yokohama, Japan, 30 August–1 September 2004; pp. 789–794. [Google Scholar] [CrossRef]
  27. Julkananusart, A.; Nilkhamhang, I. Quadrotor Tuning for Attitude Control based on Double-Loop PID Controller using Fictitious Reference Iterative Tuning (FRIT). In Proceedings of the 41st Annual Conference of the IEEE Industrial Electronics Society IECON 2015, Yokohama, Japan, 9–12 November 2015; pp. 4865–4870. [Google Scholar] [CrossRef]
  28. Julkananusart, A.; Nilkhamhang, I.; Vanijjirattikhan, R.; Takahashi, A. Quadrotor Tuning for Attitude Control based on PID Controller using Fictitious Reference Iterative Tuning (FRIT). In Proceedings of the 2015 6th International Conference of Information and Communication Technology for Embedded Systems (IC-ICTES), Hua-Hin, Thailand, 22–24 March 2015; pp. 56–60. [Google Scholar] [CrossRef]
  29. Bien, Z.; Xu, J.X. Iterative Learning Control—Analysis, Design, Integration and Applications; Kluwer Academic Press: Boston, MA, USA, 1998. [Google Scholar]
  30. Xu, J.X.; Huang, D. Optimal Tuning of PID Parameters Using Iterative Learning Approach. In Proceedings of the IEEE Multi-conference on Systems and Control, Singapore, 1–3 October 2007; pp. 226–231. [Google Scholar] [CrossRef]
  31. Campi, M.C.; Lecchini, A.; Savaresi, S.M. Virtual Reference Feedback Tuning (VRFT): A direct method for the design of feedback controllers. Automatica 2002, 18, 1337–1346. [Google Scholar] [CrossRef]
  32. Junell, J.; Mannucci, T.; Zhou, Y.; Kampen, E.J. Self-tuning Gains of a Quadrotor using a Simple Model for Policy Gradient Reinforcement Learning. In Proceedings of the AIAA Guidance, Navigation, and Control Conference, San Diego, CA, USA; pp. 1–15. [CrossRef]
  33. Killingsworth, N.J.; Krstic, M. PID Tuning Using Extremum Seeking. Model-free for online performance optimization. IEEE Control Syst. 2006, 26, 70–79. [Google Scholar] [CrossRef]
  34. Skoglar, P.; Nygards, J.; Ulvklo, M. Concurrent Path and Sensor Planning for a UAV—Towards an Information Based Approach Incorporating Models of Environment and Sensor. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 2436–2442. [Google Scholar] [CrossRef]
  35. Duchi, J.C.; Jordan, M.I.; Wainwright, M.J.; Wibisono, A. Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations. IEEE Trans. Inf. Theory 2015, 61, 2788–2806. [Google Scholar] [CrossRef] [Green Version]
  36. Agarwal, A.; Foster, D.P.; Hsu, D.; Kakade, S.M.; Rakhlin, A. Stochastic convex optimization with bandit feedback. SIAM J. Optim. 2013, 23, 213–240. [Google Scholar] [CrossRef]
  37. Ghadimi, S.; Lan, G. Stochastic First- and Zeroth-Order Methods for Nonconvex Stochastic Programming; University of Florida, Department of Industrial and Systems Engineering: Gainesville, FL, USA, 2013. [Google Scholar]
  38. Flaxman, A.D.; Kalai, A.T.; McMahan, H.B. Online convex optimization in the bandit setting: Gradient descent without a gradient. In Proceedings of the Sixteenth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Vancouver, BC, Canada; 2005; pp. 385–394. [Google Scholar]
  39. Ahmed, N.A.; Miyatake, M. A novel maximum power point tracking for photovoltaic applications under partially shaded insolation conditions. Electr. Power Syst. Res. 2008, 78, 777–784. [Google Scholar] [CrossRef]
  40. Pai, N.S.; Chang, S.C.; Huang, C.T. Tuning PI/PID controllers for integrating processes with deadtime and inverse response by simple calculations. J. Process Control 2010, 20, 726–733. [Google Scholar] [CrossRef]
  41. Litt, J. An expert system to perform online controller tuning. IEEE Control Syst. 1991, 11, 18–23. [Google Scholar] [CrossRef]
  42. Pěnička, R.; Faigl, J.; Vaňa, P.; Saska, M. Dubins Orienteering Problem. IEEE Robot. Autom. Lett. 2017, 2, 1210–1217. [Google Scholar] [CrossRef]
  43. Saska, M.; Báča, T.; Thomas, J.; Chudoba, J.; Preucil, L.; Krajník, T.; Faigl, J.; Loianno, G.; Kumar, V. System for deployment of groups of unmanned micro aerial vehicles in GPS-denied environments using onboard visual relative localization. Auton. Robot. 2017, 41, 919–944. [Google Scholar] [CrossRef]
  44. Saska, M.; Vonásek, V.; Chudoba, J.; Thomas, J.; Loianno, G.; Kumar, V. Swarm Distribution and Deployment for Cooperative Surveillance by Micro-Aerial Vehicles. J. Intell. Robot. Syst. 2016, 84, 469–492. [Google Scholar] [CrossRef]
  45. Saska, M.; Vonásek, V.; Krajník, T.; Přeučil, L. Coordination and navigation of heterogeneous MAV-UGV formations localized by a ‘hawk-eye’-like approach under a model predictive control scheme. Int. J. Robot. Res. 2014, 33, 1393–1412. [Google Scholar] [CrossRef]
  46. Saska, M.; Krajník, T.; Vonásek, V.; Kasl, Z.; Spurný, V.; Přeučil, L. Fault-Tolerant Formation Driving Mechanism Designed for Heterogeneous MAVs-UGVs Groups. J. Intell. Robot. Syst. 2014, 73, 603–622. [Google Scholar] [CrossRef]
  47. Maher, A.; Taha, H.; Zhang, B. Realtime multi-aircraft tracking in aerial scene with deep orientation network. J. Real-Time Image Proc. 2018, 15, 495–507. [Google Scholar] [CrossRef]
  48. Chong, E.K.P.; Zak, S.H. An Introduction to Optimization, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2001. [Google Scholar]
  49. Horla, D. Computational Methods in Optimization, 2nd ed.; Publishing House of Poznan University of Technology: Poznan, Poland, 2016. (In Polish) [Google Scholar]
  50. Horla, D. Performance evaluation of iterative methods to unconstrained single variable minimization problems. Stud. Autom. Inf. Eng. 2013, 38, 7–34. [Google Scholar]
  51. Lee, K.U.; Yun, Y.H.; Chang, W.; Park, J.B.; Choi, Y.H. Modeling and altitude control of quad-rotor UAV. In Proceedings of the 2011 11th International Conference on Control, Automation and Systems, Gyeonggi-do, South Korea, 26–29 October 2011; pp. 1897–1902. [Google Scholar]
  52. Ackermann, J. Robust Control—Systems with Uncertain Physical Parameters; Springer-Verlag: London, UK, 1993. [Google Scholar]
  53. Barmish, B.R. New Tools for Robustness of Linear Systems; Macmillan: New York, NY, USA, 1994. [Google Scholar]
  54. Bhattacharyya, S.P.; Chapellat, H.; Keel, L.H. Robust Control: The Parametric Approach; Prentice Hall: Englewood Cliffs, NJ, USA, 1995. [Google Scholar]
  55. Sanchez-Pena, R.; Sznaier, M. Robust Systems: Theory and Applications; John Wiley & Sons: New York, NY, USA, 1998. [Google Scholar]
  56. Zietkiewicz, J.; Horla, D.; Owczarkowski, A. Sparse in the Time Stabilization of a Bicycle Robot Model: Strategies for Event- and Self-Triggered Control Approaches. Robotics 2018, 77. [Google Scholar] [CrossRef]
  57. Tarbouriech, S.; Gomes da Silva, J.M.; Garcia, J. Stability and disturbance tolerance for linear systems with bounded controls. In Proceedings of the 2001 European Control Conference (ECC), Porto, Portugal, 4–7 September 2001; pp. 3219–3224. [Google Scholar] [CrossRef]
  58. Vrabie, D.; Pastravanu, O.; Abu-Khalaf, M. Adaptive optimal control for continuous-time linear systems based on policy iteration. Automatica 2009, 45, 477–484. [Google Scholar] [CrossRef]
  59. Vrabie, D.; Lewis, F.L. Adaptive optimal control algorithm for continuous-time nonlinear systems based on policy iteration. In Proceedings of the 47th IEEE Conference on Decision and Control, Cancun, Mexico, 9–11 December 2008; pp. 73–79. [Google Scholar] [CrossRef]
  60. Lewis, F.L.; Vrabie, D.; Syrmos, V.L. Optimal Control, 3rd ed.; John Wiley & Sons: New York, NY, USA, 2012. [Google Scholar]
  61. Báča, T.; Štepán, P.; Spurný, V.; Saska, M.; Pěnička, R.; Loianno, G.; Thomas, J.; Kumar, V. Autonomous Landing on a Moving Vehicle with Unmanned Aerial Vehicle. J. Fied Robot. 2017. in review. Available online: http://mrs.felk.cvut.cz/data/papers/mbzirc-landing2017.pdf (accessed on 13 January 2019).
  62. Lee, T.; Leok, M.; McClamroch, N.H. Nonlinear Robust Tracking Control of a Quadrotor UAV on SE(3). Asian J. Control 2013, 15, 391–408. [Google Scholar] [CrossRef]
  63. Mellinger, D.; Kumar, V. Minimum snap trajectory generation and control for quadrotors. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation (ICRA), Shanghai, China, 9–13 May 2011; pp. 2520–2525. [Google Scholar] [CrossRef]
  64. Giernacki, W.; Horla, D.; Sadalla, T. Mathematical models database (MMD ver. 1.0) non-commercial proposal for researchers. In Proceedings of the 21st International Conference on Methods and Models in Automation and Robotics (MMAR), Miedzyzdroje, Poland, 29 August–1 September 2016; pp. 555–558. [Google Scholar] [CrossRef]
  65. Saska, M.; Baca, T.; Spurný, V.; Loianno, G.; Thomas, J.; Krajník, T.; Stepan, P.; Kumar, V. Vision-based high-speed autonomous landing and cooperative objects grasping—Towards the MBZIRC competition. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems—Vision-based High Speed Autonomous Navigation of UAVs (Workshop), Daejeon, South Korea, 9–14 October 2016; pp. 1–5. [Google Scholar]
  66. Saska, M.; Krátký, V.; Spurný, V.; Báča, T. Documentation of Dark Areas of Large Historical Buildings by a Formation of Unmanned Aerial Vehicles using Model Predictive Control. In Proceedings of the IEEE Conference on Emerging Technologies and Factory Automation (ETFA’17), Limassol, Cyprus, 12–15 September 2017; pp. 1–8. [Google Scholar] [CrossRef]
Figure 1. Six-rotor UAV used during experiments: (a) cooperative transport of a large object motivated by the third challenge of the MBZIRC 2017 competition (https://youtu.be/rTlZQ1x93h0); (b) autonomous landing on a driving vehicle (https://youtu.be/zHs5LtLyBsc); (c) robots cooperating in a transport task (https://youtu.be/ogmQSjkqqp0?t=205); (d) surveillance mission of a UAV formation; (e) recording a video inside a historical building by a pair of UAV (a cameraman robot, and a lightning robot) (https://youtu.be/-sTUwzFf_Mk); (f) outdoor experiment in the countryside of South Bohemia (Czech Republic) during one of the testing days for the MBZIRC 2017 competition.
Figure 1. Six-rotor UAV used during experiments: (a) cooperative transport of a large object motivated by the third challenge of the MBZIRC 2017 competition (https://youtu.be/rTlZQ1x93h0); (b) autonomous landing on a driving vehicle (https://youtu.be/zHs5LtLyBsc); (c) robots cooperating in a transport task (https://youtu.be/ogmQSjkqqp0?t=205); (d) surveillance mission of a UAV formation; (e) recording a video inside a historical building by a pair of UAV (a cameraman robot, and a lightning robot) (https://youtu.be/-sTUwzFf_Mk); (f) outdoor experiment in the countryside of South Bohemia (Czech Republic) during one of the testing days for the MBZIRC 2017 competition.
Sensors 19 00312 g001
Figure 2. Examples of unimodal functions.
Figure 2. Examples of unimodal functions.
Sensors 19 00312 g002
Figure 3. Two intermediate points in the considered range.
Figure 3. Two intermediate points in the considered range.
Sensors 19 00312 g003
Figure 4. The case for f ( x ( 1 ) ) < f ( x ( 1 + ) ) , where the minimum x * x ( 0 ) , x ( 1 + ) .
Figure 4. The case for f ( x ( 1 ) ) < f ( x ( 1 + ) ) , where the minimum x * x ( 0 ) , x ( 1 + ) .
Sensors 19 00312 g004
Figure 5. Obtaining ρ to enable evaluation of f at a single point only.
Figure 5. Obtaining ρ to enable evaluation of f at a single point only.
Sensors 19 00312 g005
Figure 6. Selection method of the intermediate points, ensuring property discussed in Proposition 3.
Figure 6. Selection method of the intermediate points, ensuring property discussed in Proposition 3.
Sensors 19 00312 g006
Figure 7. Number of iterations in a Fibonacci-search algorithm vs. ϵ .
Figure 7. Number of iterations in a Fibonacci-search algorithm vs. ϵ .
Sensors 19 00312 g007
Figure 8. Trajectory primitives used during experiments (see the text for further information).
Figure 8. Trajectory primitives used during experiments (see the text for further information).
Sensors 19 00312 g008
Figure 9. Values of N c and N max must be chosen in the way allowing transients to decay, after controller parameters are changed by the FGT method (depicted: reference vs. actual output, evolution in time).
Figure 9. Values of N c and N max must be chosen in the way allowing transients to decay, after controller parameters are changed by the FGT method (depicted: reference vs. actual output, evolution in time).
Sensors 19 00312 g009
Figure 10. Visualization of ϵ = 0.05 accuracy for sample arbitrary ranges [ 0 , 1 ] of the both parameters.
Figure 10. Visualization of ϵ = 0.05 accuracy for sample arbitrary ranges [ 0 , 1 ] of the both parameters.
Sensors 19 00312 g010
Figure 11. Sample snapshots of one of the experiments conducted in ROS: (a,b) UAV takes off from the ground and flies to the desired position in an autonomous mode; (c) safe altitude is reached (in desired coordinates); (d) decreasing the height down to 2 m ; (e,f) altitude controller (PD) autotuning: repeated 48 cycles of raising up to 3 m and going down again the level of 2 m ( 12 s per single iteration); (g) termination of the autotuning experiment with autonomous landing of the UAV on the ground (http://mathematicalmodels.put.poznan.pl).
Figure 11. Sample snapshots of one of the experiments conducted in ROS: (a,b) UAV takes off from the ground and flies to the desired position in an autonomous mode; (c) safe altitude is reached (in desired coordinates); (d) decreasing the height down to 2 m ; (e,f) altitude controller (PD) autotuning: repeated 48 cycles of raising up to 3 m and going down again the level of 2 m ( 12 s per single iteration); (g) termination of the autotuning experiment with autonomous landing of the UAV on the ground (http://mathematicalmodels.put.poznan.pl).
Sensors 19 00312 g011
Figure 12. Histogram of the final values for the following parameters in the case without filter (F): (a) k P ; (b) k D , for all three methods (15 values).
Figure 12. Histogram of the final values for the following parameters in the case without filter (F): (a) k P ; (b) k D , for all three methods (15 values).
Sensors 19 00312 g012
Figure 13. Histogram of the final values for the following parameters in the case with a low-pass filter (F): (a) k P ; (b) k D , for all three methods (15 values).
Figure 13. Histogram of the final values for the following parameters in the case with a low-pass filter (F): (a) k P ; (b) k D , for all three methods (15 values).
Sensors 19 00312 g013
Figure 14. Overall histogram of the final values for the following parameters in the case without filter (F): (a) k P ; (b) k D , for all three methods and both configurations of the filter (30 values).
Figure 14. Overall histogram of the final values for the following parameters in the case without filter (F): (a) k P ; (b) k D , for all three methods and both configurations of the filter (30 values).
Sensors 19 00312 g014
Figure 15. Results obtained from 30 additional experiments carried out in ROS-Gazebo: the plots of functions J = f ( k P , k D ) and their surface approximations for fast (left column), medium (central) and slow (right) trajectory primitives.
Figure 15. Results obtained from 30 additional experiments carried out in ROS-Gazebo: the plots of functions J = f ( k P , k D ) and their surface approximations for fast (left column), medium (central) and slow (right) trajectory primitives.
Sensors 19 00312 g015
Figure 16. Tuning of controller parameters: (a) gains; (b) performance index.
Figure 16. Tuning of controller parameters: (a) gains; (b) performance index.
Sensors 19 00312 g016
Figure 17. (a) pitch θ ; (b) roll ϕ ; (c) yaw ψ angles in ( ).
Figure 17. (a) pitch θ ; (b) roll ϕ ; (c) yaw ψ angles in ( ).
Sensors 19 00312 g017
Figure 18. Position in the world frame ( m ) , in the axis: (a) X (East); (b) Y (North); (c) Z (UP) relative to a fixed point in space. Phases of flight: (1) arming the UAV; (2) reaching the desired position; (3) attaining the desired altitude; (4) tuning starts; (5) tuning stops; (6) landing; (d,e)—error between commanded and actual Z (UP) positions.
Figure 18. Position in the world frame ( m ) , in the axis: (a) X (East); (b) Y (North); (c) Z (UP) relative to a fixed point in space. Phases of flight: (1) arming the UAV; (2) reaching the desired position; (3) attaining the desired altitude; (4) tuning starts; (5) tuning stops; (6) landing; (d,e)—error between commanded and actual Z (UP) positions.
Sensors 19 00312 g018
Figure 19. Sample photos from the outdoor experiment in windy conditions.
Figure 19. Sample photos from the outdoor experiment in windy conditions.
Sensors 19 00312 g019
Table 1. Notation in the paper.
Table 1. Notation in the paper.
x ( k ) lower bound for a parameter at k-th iteration
x ( k + ) upper bound for a parameter at k-th iteration
x ^ candidate point in branch-and-bound procedure
f ( · ) cost function
D ( k ) considered range of a parameter at k-th iteration
x ^ * iterative estimate of the optimal solution
Table 2. Tuning results from 90 experiments carried out in ROS-Gazebo.
Table 2. Tuning results from 90 experiments carried out in ROS-Gazebo.
Slow (S)Medium (M)Fast (F)
methodfilterPar_init k P k D J k P k D J k P k D J
0018.345.451.0711.762.360.759.863.550.61
0038.344.501.019.482.360.6710.624.030.88
0054.509.860.8510.624.030.8710.625.450.78
00710.626.160.6411.004.500.739.862.350.68
0099.965.690.739.954.260.709.863.550.73
Total J for S, M, F: 11.713
0117.573.550.808.043.210.539.866.160.68
01310.622.120.525.762.841.159.864.500.75
0156.434.260.7810.622.120.4610.334.020.56
0179.484.740.739.953.780.5610.623.550.60
01911.765.930.6610.334.020.729.865.450.85
Total J for S, M, F: 10.368
1018.724.740.968.722.590.5510.622.360.59
10310.243.070.7610.334.260.885.294.260.83
1056.052,590.9110.242.350.6310.623.790.79
1079.102.830.608.815.931.169.103.780.80
1098.722.600.949.482.600.769.483.790.86
Total J for S, M, F: 12.041
1117.284.500.9410.334.020.6110.624.260.63
11310.622.120.4310.243.310.469.103.310.66
11510.242.600.618.722.600.6310.244.500.55
1178.722.600.7211.382.120.6410.622.120.55
11910.332.120.5610.245.690.667.193.550.59
Total J for S, M, F: 9.227
20110.334.020.8311.472.350.688.714.740.94
20311.763.310.749.194.260.739.483.070.69
2059.863.310.829.952.840.629.104.020.73
20710.332.840.679.194.020.627.574.261.00
2098.342.360.799.193.550.669.194.500.74
Total J for S, M, F: 11.261
2118.332.590.809.195.450.7910.622.830.47
21310.243.070.549.863.780.5410.243.070.54
2157.663.550.9010.333.070.478.434.020.58
21711.382.840.4410.332.350.428.345.221.14
21911.093.310.428.042.350.526.144.500.92
Total J for S, M, F: 9.490
Table 3. Tuning results from 30 additional experiments carried out in ROS-Gazebo (statistics).
Table 3. Tuning results from 30 additional experiments carried out in ROS-Gazebo (statistics).
Slow (S)Medium (M)Fast (F)
k P k D J k P k D J k P k D J
mean value9.2213.3350.6409.2303.0950.6708.6893.4780.740
std deviation1.1901.1000.1701.2301.0500.2101.1301.0900.200
min value7.1902.1200.3797.1902.1200.4776.4302.3600.572
max value11.0004.9800.93510.6205.2101.15310.2404.9801.278

Share and Cite

MDPI and ACS Style

Giernacki, W.; Horla, D.; Báča, T.; Saska, M. Real-Time Model-Free Minimum-Seeking Autotuning Method for Unmanned Aerial Vehicle Controllers Based on Fibonacci-Search Algorithm. Sensors 2019, 19, 312. https://doi.org/10.3390/s19020312

AMA Style

Giernacki W, Horla D, Báča T, Saska M. Real-Time Model-Free Minimum-Seeking Autotuning Method for Unmanned Aerial Vehicle Controllers Based on Fibonacci-Search Algorithm. Sensors. 2019; 19(2):312. https://doi.org/10.3390/s19020312

Chicago/Turabian Style

Giernacki, Wojciech, Dariusz Horla, Tomáš Báča, and Martin Saska. 2019. "Real-Time Model-Free Minimum-Seeking Autotuning Method for Unmanned Aerial Vehicle Controllers Based on Fibonacci-Search Algorithm" Sensors 19, no. 2: 312. https://doi.org/10.3390/s19020312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop