You are currently viewing a new version of our website. To view the old version click .
Entropy
  • Article
  • Open Access

4 December 2017

An Improved Chaotic Optimization Algorithm Applied to a DC Electrical Motor Modeling

and
1
Dipartimento di Ingegneria dell’Informazione, Università Politecnica delle Marche, 60131 Ancona, Italy
2
Master’s Program Systems and Control, Technische Universiteit Eindhoven, 5612 Eindhoven, The Netherlands
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Probabilistic Methods for Inverse Problems

Abstract

The chaos-based optimization algorithm (COA) is a method to optimize possibly nonlinear complex functions of several variables by chaos search. The main innovation behind the chaos-based optimization algorithm is to generate chaotic trajectories by means of nonlinear, discrete-time dynamical systems to explore the search space while looking for the global minimum of a complex criterion function. The aim of the present research is to investigate the numerical properties of the COA, both on complex optimization test-functions from the literature and on a real-world problem, to contribute to the understanding of its global-search features. In addition, the present research suggests a refinement of the original COA algorithm to improve its optimization performances. In particular, the real-world optimization problem tackled within the paper is the estimation of six electro-mechanical parameters of a model of a direct-current (DC) electrical motor. A large number of test results prove that the algorithm achieves an excellent numerical precision at a little expense in the computational complexity, which appears as extremely limited, compared to the complexity of other benchmark optimization algorithms, namely, the genetic algorithm and the simulated annealing algorithm.

1. Introduction

A chaotic systems is a complex system that shows sensitivity to initial conditions, such as an economy, a stock market, the earth’s weather system, the behavior of water boiling on a stove, migratory patterns of birds, the spread of vegetation across a continent, an astronomical or a social system. In such systems, any uncertainty—no matter how small—in the beginning will produce rapidly escalating errors in the prediction of the system’s future behavior. To make an accurate prediction of the long-term behavior of such systems, the initial conditions must be known in their entirety and to an infinite level of accuracy. Even very simple or small systems can exhibit very complex behaviors. Typical features of chaotic systems include [1]:
  • Nonlinearity: A linear system cannot be chaotic; therefore, only nonlinear systems may evolve in time forming chaotic trajectories.
  • Determinism: A chaotic system is deterministic, as it does not include any random or stochastic element. Given an initial state for the system, it evolves according to deterministic rules. Chaos theory is the study of nonlinear dynamics, in which seemingly random events are actually predictable from deterministic equations.
  • Sensitivity to initial conditions: Small changes in the initial state of a chaotic system may lead to radically different trajectories (this is commonly referred to as the ”butterfly effect”).
  • Irregularity: Hidden order including a large or infinite number of unstable periodic trajectories forms the infrastructure of irregular chaotic systems. Chaos refers to an apparent lack of order in a system that nevertheless obeys particular laws or rules.
  • Long-term unpredictability: Long-term prediction of a trajectory of a chaotic system is extremely hard due to its sensitivity to initial conditions, which can be known only up to a finite degree of precision.
Historically, the study of chaos started in mathematics and physics and expanded into engineering, as well as into information and social sciences. Information entropy is tightly related to chaoticity in real-world discrete-time sequences, as underlined, for example, in Hagmair et al. [2]. There exist different kinds of potential commercial and industrial applications of chaotic systems, which, for simplicity, are classified into stabilization, synthesis, and analysis [3]. A typical application of chaotic systems is found in the secure transmission of information, where a meaningful message is superimposed to a chaotic signal that makes it non-understandable to someone who neither knows the system that produces the chaotic trajectory nor its initial state. To decode the received signal, it is necessary to synchronize the chaotic oscillators at the transmitter and at the receiver [4].
In the present research study, chaotic systems are applied to constrained global optimization of possibly non-smooth, nonlinear complex functions of several variables. Although non-conventional optimization techniques based on stochastic methods have a long hystory (see, for example, [5]), optimization techniques based on chaos theory have been less investigated.
The present study took its moves from the seminal contribution [6] that proposed how to optimize complex functions by chaos search. The main idea behind the resulting chaos-based optimization algorithm (COA) [6] is to generate chaotic trajectories by means of nonlinear dynamical systems to explore the search space while looking for the global minimum of a complex criterion function. The main reasons to choose a chaotic search over a deterministic search may be summarized as follows:
  • The criterion function to optimize may be non-smooth and non-differentiable, hence, standard search methods such as gradient steepest descent may not be applied in its optimization.
  • A chaotic system may produces complex, non-repeating trajectories that hardly visit the same state more than once, hence a chaos-based search algorithm is able to visit the whole space efficiently while looking for the optimal configuration.
  • A multivariate criterion function may include a large number of variables to optimize simultaneously, therefore, a simple mechanism to generate multivariate candidate solutions over a high-dimensional search space makes the search algorithm fast and easy to implement.
Despite its success in the optimization of complex functions, to the best of our knowledge, the potentiality of the chaos-based optimization algorithm has only been marginally investigated in the literature [7,8]. The aim of the present research study is to investigate the numerical properties of the COA, both on complex optimization test-functions from the literature and on a real-world problem, in order to contribute to the understanding of its global-search features. Through a large number of tests and numerical evaluations, it was found that the algorithm may be further fine-tuned in order to improve its numerical precision at a little expense in computational complexity, which is nevertheless extremely limited, compared to the complexity of other optimization algorithms, namely, the genetic algorithm and the simulated annealing algorithm.
The present paper is organized as follows. Section 2 summarizes the chaos-based optimization algorithm and tests a number of possible chaotic maps over a large number of test functions drawn from the specialized literature. Section 2 also suggests a fine-tuning of the original COA algorithm and compares the modified algorithm with the genetic algorithm and the simulated-annealing based optimization algorithm. Section 3 illustrates the performances of the refined COA algorithm in the modeling of a direct-current electrical motor. Section 4 concludes the paper.

2. The Chaos-Based Optimization Algorithm and Its Fine-Tuning

Traditional algorithms are unable to solve some optimization problems since these algorithms can get trapped into local minima or need too much search time. The properties of chaotic systems allow the development of an efficient chaos-based optimization algorithm. Unlike traditional stochastic optimization algorithms that escape from local minima by accepting some sub-optimal solutions according to a certain probability, the COA’s search takes advantage of chaotic motions’ regularity. A chaotic movement can go through every state in a definite area thanks to its own regularity and every state is obtained only once.
This Section illustrates the chaos-based optimization algorithm in Section 2.1 and describes in Section 2.2 several chaotic maps, which are indispensable for the development and the implementation of the COA; it also shows in Section 2.3 the performances of the COA in the optimization of a large set of benchmark functions. The results obtained from the large set of test experiments suggest a refinement of the original chaos-based optimization method, which is explained and numerically tested in Section 2.4. Section 2.5 presents a comparison between the refined COA and both the genetic optimization algorithm and the simulated-annealing-based optimization algorithm.

2.1. The Chaos-Based Optimization Algorithm

If a given optimization problem is described by a criterion f : R n R and the constraints that it is subjected to prescribe its variables x i to belong to specified intervals, the optimization problem can be described as min f ( x i ) , i = 1 , , n , subjected to a i x i b i , with a i , b i R . In the COA, thanks to the carrier wave method, the optimization variables follow the trajectories of specific chaotic functions whose ergodic areas are sized according to the constraints of the problem. The COA is outlined as follows:
  • Generate n chaotic variables x i , k from the same nonlinear chaotic oscillator, where the index i denotes the variable’s index and k denotes a discrete-time index. There will be n initial states x i , 0 that are chosen very close one to another (for instance 0.1 , 0.1001 , 0.1002 ). These initial states, because of chaotic systems properties, will give rise to completely different trajectories.
  • Launch the first carrier wave: This method allows us to obtain the optimization variables with the amplification of the ergodic areas of the n chaotic variables by the equation x i , k = c i + d i x i , k , where c i and d i denote amplification constants. Perform the rough search: Let x ¯ i = x i , 0 , calculate the value of the objective function f at the initial state and save its value f ¯ = f ( x i , 0 ) . Then start to iterate over the time k: Increase the time-index (iteration counter) k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 1 search steps, where the integer constant K 1 determines the complexity of the search method. The first carrier wave will produce a rough solution ( x ¯ 1 , x ¯ 2 , , x ¯ n ) .
  • Launch the second carrier wave: The second carrier wave explores the search space in a narrower interval around the rough solution, in fact, the trajectories of the optimization variables are generated according to the rule x i , k = x ¯ i + e i + f i x i , k , where the x i , k denote again the states of n discrete-time chaotic systems, while e i and f i are the new amplification constants to perform the search within small ergodic areas around x ¯ i . Perform the fine search: Start to iterate over the time k: Increase the time-index k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 2 search iterations.
  • The process ends and the values of parameters ( x ¯ 1 , x ¯ 2 , , x ¯ n ) , together with the criterion value f ¯ , will be the best approximation of the global solution of the given optimization problem as found by the chaos-based optimization algorithm.

2.2. Chaotic Maps

To generate the carrier waves, the COA makes use of a chaotic map that controls the features of the chaotic system used to generate the variables that are essential for the optimization.
There are some known universal properties of the chaotic maps, for instance:
  • They may present different kinds of attractors: An attractor is a set of numerical values towards which a dynamical system evolves after a sufficiently long time. A set of numerical values can be defined as an attractor if the trajectories that are related to them stay close to each other even if they are slightly perturbed.
  • They may present different values of the Lyapunov coefficient: This coefficient measures how much the system’s orbits are dependent from the initial state. Specifically, the Lyapunov coefficient measures the average parting speed of two orbits that are initially close and then drift apart.
Four chaotic maps have been surveyed. The characteristics that are more interesting for the optimization problem are the state’s trajectory variation (assuming that the initial states x i , 0 are very close to each other) and the ergodic area of each chaotic system, which is measured by pooling numerically the distribution of its states.
  • Logistic map. The “logistic map” is a second order polynomial function, described by:
    x k + 1 = r x k ( 1 x k ) , k = 0 , 1 , 2 , 3 , ,
    where r is the control parameter. It is supposed that 0 x 0 1 and that 0 r 4 . The Equation (1) represents a discrete-time, deterministic dynamical system without any stochastic interference. The long-term behavior of said system cannot be predicted because it changes completely as the control parameter r varies. In particular, when r = 4 , the system becomes chaotic. As we can see in Figure 1, from k > 10 , two waves originating from very close initial states start to behave in a different way and the system’s initial information is completely lost. It may be observed that the values generated by the logistic map are condensed around its ergodic area’s bounds, which are 0 and 1.
    Figure 1. Logistic map. Left-hand panel: Comparison between system’s output with two different initial conditions. (The wave with initial value x 0 = 0.1 is represented by the red line and the other with initial value x 0 = 0.1001 by the blue line.) Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (1).
  • Quadratic map. The “quadratic map” arose as the real-valued version of the Mandelbrot set complex map and is described by:
    x k + 1 = x k 2 c , k = 0 , 1 , 2 , 3 , ,
    where the constant c is the control parameter. For c = 2 , the system acquires interesting complicated dynamics. The ergodic area is included between 2 and 2, as may be readily observed in Figure 2.
    Figure 2. Quadratic map. Left-hand panel: Comparison between system’s output with two different, albeit very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (2).
  • Tent map. The “tent map” is defined by the following dynamical system:
    x k + 1 = { μ x k if x k < 1 2 , μ ( 1 x k ) if x k 1 2 ,
    where μ is the control parameter and k = 0 , 1 , 2 , 3 , . For μ 1.5 , the waves start to show bifurcations. The value that has been used in the tests is μ = 1.9999 . Unlike the two previous maps, the tent map has an uniform distribution in its ergodic area [ 0 , 1 ] , as it may be readily appreciated in Figure 3.
    Figure 3. Tent map. Left-hand panel: Comparison between system’s output with two different, although very close, initial conditions. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (3).
  • Sine map. The “sine map” gives rise to the following discrete-time dynamical system:
    x k + 1 = λ sin ( π x k ) , k = 0 , 1 , 2 , 3 , ,
    where λ denotes the control parameter. This map generates a unique bifurcations diagram, symmetrical under both plane’s axis λ , x . The value that has been used in the tests is λ = 1 . Figure 4 shows the behavior of the sine map and that its ergodic area is again [ 0 , 1 ] .
    Figure 4. Sine map. Left-hand panel: Comparison between system’s output with two different initial conditions, very close to one another. Right-hand panel: Occurrence histogram of the values of the states of the dynamical system (4).

2.3. Optimization of Test Functions

Seven benchmark functions have been taken into consideration to test and to further develop the COA. These functions are usually used in the literature [9] because they show a wide spectrum of characteristics, namely, they give rise to continuous/discontinuous, convex/nonconvex, unimodal/multimodal, quadratic/nonquadratic optimization problems.
  • F1: Generalized Rosenbrock function. This function of n variables is defined as
    f 1 ( x ) = i = 1 n 1 100 ( x i 2 x i + 1 ) 2 + ( 1 x i ) 2 .
    It was proposed by Rosenbrock in 1960 and is commonly used to test optimization algorithms since, for n = 2 , the search for the minimum is problematic in its parabolic area x 2 = x 1 2 ; furthermore, it becomes multimodal for n > 3 . The constraints and the global minimum of the generalized Rosenbrock function are recalled in the Table 1 and its graphical rendering is given in the Figure 5.
    Table 1. Constraints and global minimum of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
    Figure 5. Rendering of the generalized Rosenbrock function (benchmark function “F1”) for the case of n = 2 independent variables.
  • F2: Parabolic function. This function is defined as
    f 2 ( x ) = i = 1 n x i 2 .
    It was introduced to test optimization by genetic adaptive systems. The global minimum is not difficult to find because it is the only minimum of the function, but it can provide some information about the precision of the algorithm. The constraints that this optimization problem is subjected to are 5.12 x i 5.12 .
  • F3: Goldstein–Price function. This function is defined as
    f 3 ( x ) = ( 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ) × ( ( 30 + ( 2 x 1 3 x 2 ) 2 ) ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ) .
    The Goldstein–Price Function only has two variables but it exhibits several local minima. The constraints and the global minimum of this function within the considered domain are recalled in the Table 2, while its graphical rendering is given in the Figure 6.
    Table 2. Constraints and global minimum of the Goldstein–Price function (benchmark function “F3”).
    Figure 6. Rendering of the Goldstein–Price function (benchmark function “F3”).
  • F4: Schaffer function. This function is defined as
    f 4 ( x ) = 0.5 sin 2 x 1 2 + x 2 2 0.5 1 + 0.001 ( x 1 2 + x 2 2 ) 2 .
    The Schaffer function (number 4) is quite interesting because its global minimum is not unique, but it is given by all the points situated on a circumference with a well-defined radius r * . The constraints and the global minimum (in terms of radius) of this function are recalled in the Table 3 and its graphical rendering is given in the Figure 7.
    Table 3. Constraints and global minimum of the Schaffer function (benchmark function “F4”).
    Figure 7. Rendering of the Schaffer function (benchmark function “F4”).
  • F5: Step function. This function is defined as
    f 5 ( x ) = i = 1 n x i ,
    where the symbol · denotes rounding down to the nearest integer. This function tests an optimization algorithm’s ability to overcome discontinuities. The constraints and the global minimum of this function are recalled in Table 4 and its graphical rendering is given in Figure 8. Since there is not any unique global minimum, an optimization algorithm is supposed to stop when x i 5 .
    Table 4. Constraints and global minimum of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
    Figure 8. Rendering of the “step” function (benchmark function “F5”) for the case of n = 2 independent variables.
  • F6: Normalized Schwefel function. This function is defined as
    f 6 ( x ) = 1 n i = 1 n x i sin | x i | .
    The Schwefel Function has its global minimum far away from the best local minimum: For this reason, optimization algorithms could converge on a wrong direction. The constraints and the global minimum of this function are recalled in the Table 5 and its graphical rendering is given in Figure 9.
    Table 5. Constraints and global minimum of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
    Figure 9. Rendering of the Schwefel function (benchmark function “F6”) for the case of n = 2 independent variables.
  • F7: Rastring function. This function of n independent variables is defined as
    f 7 ( x ) = 10 n + i = 1 n x i 2 10 cos ( 2 π x i ) .
    The Rastring function is an example of a nonlinear, multimodal function. Its optimization is considered as a complicated problem because of its extended search space and its several local minima. The constraints and the global minimum of this function are recalled in Table 6 and its graphical rendering is given in the Figure 10.
    Table 6. Constraints and global minimum of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
    Figure 10. Rendering of the Rastring function (benchmark function “F7”) for the case of n = 2 independent variables.
The COA algorithm has been tested over the above seven benchmark functions, initially with only the first carrier wave and subsequently with the full algorithm. All the recalled four types of chaotic maps have been used to generate the optimization variables trajectories. The parameters of the algorithm are recalled below for the convenience of the reader:
  • K 1 : Stop criterion for the “rough search”.
  • K 2 : Stop criterion for the “fine search”.
  • c i , d i : Amplifiers for the first carrier wave.
  • e i , f i : Amplifiers for the second carrier wave.
We recall that the amplifiers c i , d i are constants that allow the chaotic maps’ ergodic areas to adapt to the specific search intervals (constraints) of the test function s. In the following tables, the figure Total Loops stands for the number of iterations taken to run and Error is the percentage distance (compared to the search interval) from the coordinates of the global minimum. The rough search has been run several times for all the functions and for all the chaotic maps with K 1 = 100 , 200 , 400 , 600 , 800 , 1000 . Only the results that correspond to an error under 10% and a limited number of Total Loops have been reported.
Table 7 illustrates the results obtained by applying a rough search to the benchmark functions F1 and F2 with n = 2 variables.
Table 7. Results of the application of a rough search (only first carrier wave) on the benchmark functions F1 and F2.
Only the quadratic map achieves a reasonable value but with too many iterations. With regard to the function F2, the algorithm, during the first 100 loops, could not find any better results than the initial states (that have been set at x 1 , 0 = 0.1 and x 2 , 0 = 0.1001 ).
Table 8 illustrates the results obtained by applying a rough search to the benchmark functions F3 and F4 with n = 2 variables. F3 function’s error is overall low, the logistic map is the one that exhibits better results. The optimization of the F4 function could be resolved by an infinite number of points that belongs to the circumference of radius r * = 1.5692 . Even in this case, the quadratic map needs the highest number of iterations.
Table 8. Results of the application of a rough search (only first carrier wave) on the benchmark functions F3 and F4.
The algorithm could easily overcome the discontinuity of the F5 function and it could find immediately values lower than 5 , therefore it is not necessary to show the relative table.
Table 9 illustrates the results obtained by applying a rough search to the benchmark functions F6 and F7 with n = 2 variables. The results about the optimization of the functions F6 and F7 do not provide any further information about the algorithm’s behavior but they confirm its excellent optimization ability.
Table 9. Results of the application of a rough search (only first carrier wave) on the benchmark functions F6 and F7.
In summary, the logistic map is the chaotic map that allows the first carrier wave to obtain better outcomes, namely, lower number of Total Loops, stop criterion to about 200 iterations and a final value close to any global minimum (in the worst case, with an Error of 8%). Moreover, the tent map provide good results. The quadratic map proves to be the worst in almost all tests (the cause may be its ergodic area included between 2 and 2). As it can be seen, the rough search does not need a high value of K 1 ; in fact, since the initial search area is rather wide, increasing the number of Total Loops does not always imply a proportional Error’s reduction.
It is now worth recalling that the amplifiers e i , f i of the second carrier wave are the algorithm’s most important parameters because they influence substantially its precision. These parameters are problem-dependent: specific constants were used, that could modify the ergodic area of the chaotic waves. In some cases, this area was changed to be 3% of the total search area, in others it was change to 2% or 8%. Thanks to its good performance during the “rough search”, the logistic map was used to guide the second carrier wave. The amplifiers were calculated in the following way:
e i = ε | α i | , f i = 2 ε | β i α i | ,
where α i and β i are, respectively, the lower bound and the upper bound of the constraints that were associated to the variable x i , ε is the area’s shrinkage percentage (for example, 2% or 3% or 8%).
Several tests were effected by changing at every test the stopping parameter K 2 . The obtained results are displayed in Table 10. In the table, x ¯ 1 and x ¯ 2 denote the values found by the COA, whereas P is a precision index that shows the ratio between the errors’ average on the individual variables and a normalization constant proportional to the width of the search area, namely:
P = i = 1 n | x ¯ i x i * | i = 1 n | β i α i | .
Table 10. Results of the application of a fine search (both first and second carrier waves) on the benchmark functions F1–F7.
In all experiments, n = 2 variables were chosen.
The results pertaining to some selected values of the parameters K 2 are included in the table to point out the precision that can be obtained with either a low or a high number of iterations. The substantial Error of the “rough search” affects the performance of the “fine search” applied to the F1 function, which is the most troublesome for both stopping criteria. Therefore, it is important to use, for the purpose of optimization, a limited neighborhood for the second search. The values of the performance index P lay between 10 5 and 10 4 in most cases, which proves the quality of the chaos-based optimization algorithm. It can be seen that almost for all functions, excluding F4 and F5, the values x ¯ 1 and x ¯ 2 are close to the coordinates of the global minimum but only after a relatively large number of iterations.

2.4. Fine-Tuning of the COA Method (R-COA)

Specific problems that have dozens of variables in their objective functions or dynamic systems that need the resolution of differential equations could see their computing time exponentially increased by an high number of iterations. In the present research, to avoid the need for a large number of iterations, it has been proposed the introduction of a third carrier wave, which corresponds to an additional search stage, hereafter denoted as refined search. The whole refined chaos-based optimization algorithm (R-COA) may be outlined as follows:
  • Generation of chaotic variables.
  • Launch the first carrier wave and perform the rough search (identical to that of the COA).
  • Launch the second carrier wave and perform the fine search (identical to that of the COA).
  • Launch the third carrier wave: It is described by the equation x i , k = x ¯ i + g i + h i x i , k , where x ¯ i is the current best solution as found by the second carrier wave, g i and h i are new amplification constants that allow the search in small ergodic areas around x ¯ i , and x i , k is a new chaotic wave. Perform the refined search: Start to iterate over the time k, increase the time-index (iteration counter) k, generate a new attempted solution x i , k , calculate the value of the objective function f = f ( x i , k ) and evaluate the solution: If f f ¯ , then set f ¯ = f and x ¯ i = x i , k . Continue with the iteration until f ¯ does not improve in K 3 search loops.
  • The process ends and the parameter values ( x ¯ 1 , x ¯ 2 , , x ¯ n ) together with the criterion value f ¯ will be the best global solution to the optimization problem, as found by the refined COA.
For the new analysis of the test functions, a lower value of the constant K 2 was used. In addition, different K 3 values were tried. The amplifiers g i , h i are defined as:
g i = δ | α i | , h i = 2 δ | β i α i | ,
where δ represents the shrinkage percentage of analyzed area (2.5%, 1.5% or 0.5%).
The obtained results are displayed in Table 11. In the table, x ¯ 1 and x ¯ 2 again denot the values found by the R-COA, whereas P is the precision index defined before. Comparing the results displayed in Table 11 with the results displayed in Table 10, it can be noticed that the COA endowed with a third carrier wave always achieves a better (or comparable) precision, albeit with a lower number of Total Loops, than its original version. The values x ¯ i found by the algorithm, excluded the ones pertaining to the F6 function (since it has a wide search interval), displays an error in the fourth or fifth decimal digit.
Table 11. Results of the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7 (Since the functions F4 and F5 did not need a better optimization, they were not tested with the new algorithm).
Table 12 displays the values f ¯ of the test function achieved by the refined COA: The differences between the value found by the optimization algorithm and the actual minimum are around 10 4 and 10 5 .
Table 12. Values of the test functions resulting from the application of a refined search (first, second and third carrier waves) on the benchmark functions F1–F3, F6, F7. (The functions F4 and F5 were not tested).

2.5. Comparison with the Genetic and the Simulated-Annealing Algorithms

The performances of the chaos-based optimization algorithm R-COA have been compared with those exhibited by Genetic Algorithm and Simulated Annealing algorithm.
Genetic algorithms (GAs) are computer programs that mimic the processes of biological evolution in order to solve problems and to model evolutionary systems (see, e.g., [10,11], for an overview). The continuing performance improvements of computational systems has made them attractive for some types of optimization. In particular, genetic algorithms work well on mixed continuous/discrete combinatorial problems. GAs are less susceptible to getting stuck at local optima than gradient search methods, although they tend to be computationally expensive. To run a genetic algorithm, one must represent a solution to a given problem as a genome (or chromosome) and the quality of an individual in relation to a given optimization problem is measured by a fitness function. The genetic algorithm then creates a population of solutions and applies genetic operators such as mutation and crossover to evolve the solutions in order to find the best ones. Although it is not guaranteed that a GA will find a global optimal solution, in general, a GA is able to find good solutions within an acceptable time span. Genetic algorithms present several parameters to set and operators to tune in order to optimize their optimization performances. In the present research, the GA solver of MATLAB’s Global Optimization Toolbox (version 2015a) was utilized maintaining without changes its already optimized options. See, e.g., [12], for a recent application to system identification and [13] for an application to magnetic resonance image segmentation.
Simulated annealing (SA) is a method for finding a good (albeit not necessarily optimal) solution to an optimization problem [14]. Simulated annealing’s strength is that it avoids getting caught at local optima, namely, at solutions that are better than any others nearby, and yet are not the global optima. In fact, generally speaking, an optimization algorithm searches for the best solution by generating a random initial solution and “exploring” the area nearby: If a neighboring solution is better than the current one, then it moves to it, otherwise the algorithm stays put. This basic mechanism can lead to situations where it gets stuck at a sub-optimal place. Simulated annealing injects some sort of randomness into the basic mechanism to escape local optima early in the process without getting off when an optimal solution is nearby. The simulated annealing algorithm used in the experiments is the one implemented within MATLAB (version 2015a). See, e.g., [15] for an application to indoor wireless local area network (WLAN) localization.
These algorithms were run 10 times on independent trials. Table 13 shows the results obtained by comparing the R-COA algorithm and the simulated annealing algorithm, while Table 14 shows the results obtained by comparing the COA algorithm and the genetic algorithm. In particular, for the SA and the GA, the tables show the average values x i over the independent trials and the standard deviation of the found global minimum’s coordinates, the average number of iterations required to converge and the number of times that the SA or the GA algorithm achieved convergence.
Table 13. Results of a comparison between the refined COA and the SA (Simulated Annealing Algorithm) on the benchmark functions F1–F7.
Table 14. Results of a comparison between the refined COA and the GA (Genetic Algorithm) on the benchmark functions F1–F7.
When tested over the functions F1–F5, the SA always converges towards the global minimum and obtains the exact solution for the benchmark functions F2, F3 and F5 (as it can be seen from the values of the standard deviation). For the functions F1, F4 and F7, the SA is less accurate than the R-COA, moreover, the average number of iterations for five functions out of seven is higher than the number of total loops required by the R-COA. The search for the test function F6 did not converge (the algorithm probably was trapped in a local minima due to the large search’s area).
GA’s results are worse compared with those of SA, even if, generally, the number of iterations is lower. The GA algorithm, not only does not always converge when tested with the benchmark functions F6 and F7, but for seven times out of ten, with F3, it achieves values that lay far apart from the global minimum. GA and R-COA exhibit the same accuracy for the problem F4. The number of Total Loops is higher than the average number of iterations only in two cases.
In conclusion, the chaos optimization algorithm exhibits two important properties that distinguish it from the optimization algorithms commonly invoked in the scientific literature:
  • Convergence: The chaos-based optimization algorithm always converges towards the global minimum since: (a) it does not use variables related to probability; and (b) it is able to examine the whole search area, thanks to its mapping action;
  • Low computation time: Since it does not utilize gradient’s calculation to evaluate the tendency of the objective function, it exhibits a computational complexity equal to Θ ( n ) (the notation Θ ( f ( n ) ) (commonly referred to as “big-Theta of f ( n ) ”) is used to define the computational complexity of a problem: It expresses how fast an algorithm temporal cost increases as a function of The problem size n).

2.6. Generic Refined Chaos-Based Optimization Algorithm

The chaos-based algorithm, similar to other optimization algorithms, has parameters (ergodic areas’ amplifiers and stopping criteria) that must be adapted according to the faced problem. It was possible to model these parameters in order to create a generic algorithm, thanks to the large amount of data that were collected during the above tests.
The Euclidean distance between the minimum approximately found at the end of the rough search and the actual global minimum could be considered as a circle’s radius. This radius represents the sub-area where the second carrier wave will perform a second search. The same measure can be used for the fine search and the third carrier wave. Some general values were found while comparing all the sub-areas (regarded as a percentage of the total search area) and their relative stopping criteria. These values allow solving a generic optimization problem, even if it will perform less efficiently than a R-COA tailored to a specific problem.
The suggested amplifiers for the second carrier wave are:
e i = 0.025 d i 2 π , f i = 2 | e i | ,
and the suggested amplifiers for the third carrier wave are:
g i = 0.00015 d i 2 π , h i = 2 | g i | ,
where the d i ’s are the amplifiers for the first carrier wave. The stopping criteria are:
K 1 = 400 , K 2 = 500 .
Table 15 shows the outcomes of the optimization performed by the generic R-COA versus the results obtained with a specific R-COA. The global minimum’s coordinates are, in general, less accurate compared with those of specific R-COA (except for the first function). Moreover, the number of Total Loops is higher. However, since the error stays confined to the third or fourth decimal digit, the results can be considered acceptable.
Table 15. Results of a comparison between the specific refined COA and the generic refined COA on the benchmark functions F1–F7.

3. Application to the Modeling of a Direct-Current Electrical Motor

The refined chaos-based optimization algorithm was applied to a real-world physical problem, namely, the estimation of the electro-mechanical parameters of a model of a direct-current (DC) electrical motor.
Direct-current motors found wide applications in industrial systems because they are easy to control and to model. For analytic control system design and optimization, a precise model of a DC motor may be desirable, since the specifications provided by the motor manufacturer may not be considered adequate, especially for cheaper DC motors which tend to exhibit relatively large tolerances in their electrical and mechanical parameters [16].
The present section recalls the fundamental equations describing the electro-mechanical dynamics of a direct-current electrical motor in Section 3.1. Section 3.2 casts the identification of the parameters of a DC motor model as an optimization problem. Section 3.3 describes and comments the results about the motor’s parameters identification when the motor is driven by a step signal.

3.1. Direct-Current Electrical Motors

Figure 11 shows an electro-mechanical model of a DC electrical motor. It is assumed that the stator has only a pair of polar expansions, characterized by an inductance L e associated to its relative winding and by a resistance R e associated to conductor’s leakage. The analytic equation that describes the stator-equivalent electrical circuit is:
v e ( t ) = L e d i e ( t ) d t + R e i e ( t ) ,
where i e denotes the stator current intensity. Likewise, it is assumed that the rotor has only a pair of polar expansions, characterized by an inductance L a and by a resistance R a . Furthermore, the model takes into account the electromotive force’s effect, e, that correspond to an induced voltage proportional to the rotation speed. The following equation describes the rotor’s electrical circuit:
v a ( t ) = L a d i a ( t ) d t + R a i a ( t ) + e ( t ) ,
where i a denotes the rotor current intensity. According to the physical properties of the motor, while the motor rotates with angular velocity ω m , a proportional electromotive force is generated in the armature circuit, whose intensity is given by
e ( t ) = K ω m ( t ) ,
where K is a proportionality constant whose value depends on the physical design of the DC motor.
Figure 11. An electro-mechanical model of the direct-current electrical motor (we assumed no angular speed change after gear).
Whenever an electrical current of intensity i a flows through the rotor winding, magnetic attraction/repulsion forces appear between the stator’s and rotor’s windings, which cause the motor shaft to rotate. The magnetic field inside the motor transfers a mechanical torque to the drive shaft. If we denote the rotor inertia by J, the viscous friction constant by B m and the torque load applied to the drive shaft by T L (that is supposed to be constant), the equation relating the torque T m available at the drive shaft may be written as
T m ( t ) = J d ω m ( t ) d t + B m ω m ( t ) + T L .

3.2. Formulation of the Motor’s Parameters Identification as an Optimization Problem

To estimate, through an optimization algorithm, the values of the parameters of the above model of a DC motor, it is necessary to know its output values corresponding to a given input drive signal [17]. A DC motor was simulated by a Simulink model, assuming constant stator voltage, controlled by the armature voltage. Figure 12 shows the Simulink model block-diagram.
Figure 12. A Simulink model of an electrical direct-current motor with constant stator voltage.
The armature voltage v a is the only reference signal, whereas the constant load torque T L is regarded as a system disturbance. In the Simulink model, they are multiplied by V r and T r , respectively, to amplify their values. In the two sections of the diagram that represent the electrical dynamics and the mechanical dynamics, there is an integrator block with initial conditions i a 0 and ω m 0 . The outputs of the model are the armature current i a and the angular speed ω m of the drive shaft.
The state-space representation of the model is
d i a ( t ) d t = v a ( t ) R a i a ( t ) K ω m ( t ) L a , d ω m ( t ) d t = T m ( t ) B m w m ( t ) T L J .
The initial value of the armature current and of the angular speed of the drive shaft are i a 0 = 22.8284 A and ω m 0 = 122.3259 rad/s.
A step input response was analyzed and its corresponding outputs are shown in Figure 13.
Figure 13. Example of output values of the model of a direct-current electrical motor driven by a step input voltage. Left-hand panel: Armature current i a versus time. Right-hand panel: Drive shaft angular velocity ω m versus time.
To estimate the best values of the electro-mechanical parameters of a DC motor model, the problem is formulated in terms of model performance, to be optimized by means of the R-COA.
In [18], two performance criteria were tested, namely, a quadratic criterion and an absolute criterion. A comparison reported in the paper [18] reveals that the results obtained though the optimization of the quadratic criterion appear as better than those obtained by means of the absolute criterion. However, in the study [19], it was observed how these criteria perform nearly the same. In the present research, the following absolute criterion was selected as objective function:
C = j = 1 D ( k a | i a , j i a s , j | + k b | ω m , j ω m s , j | ) ,
where i a , j and ω m , j represent the values of the variables computed by means of the Simulink model, upon time discretization, while i a s , j and ω m s , j denote the values of the variables obtained by the optimization algorithm, upon time discretization, at discrete-time step j = 1 , 2 , , D . The constants k a > 0 and k b > 0 are weights, which attribute a different importance to the error in the estimate of the armature current and to the error in the estimate of the drive shaft angular speed, that will be discussed later.
Table 16 shows the actual values of the electro-mechanical parameters of the considered DC motor.
Table 16. Application to DC motor model parameters estimation: Actual values of motor’s physical electro-mechanical parameters.
When launching the R-COA, it is necessary to take into account that each parameter is constrained to lie in a given interval, summarized in Table 17. Therefore, the R-COA must adapt the ergodic areas of six chaotic waves to these constraints.
Table 17. Application to DC motor model parameters estimation: Search range for each parameter based on physical constrains.

3.3. Results for an Electrical Motor Driven by a Step Armature Voltage

In the first set of experiments, a reference step signal of 240 V of amplitude was applied as armature voltage. In this test condition, the velocity of the drive shaft ω m rises until it reaches the steady-state value of 124.9103 rad/s, while the armature current intensity decreases until it stabilizes at a value of 14.8600 A. The transient lasts approximately 0.5 second.

3.3.1. Separated Electrical and Mechanical Model Parameters Estimation

As a first test, only the electrical dynamics of the DC motor was considered as unknown, whereas the mechanical parameters were considered as exactly known. Namely, the armature resistance R a and armature inductance L a were taken as the only parameters to optimize for. In this test, the exact values of the torque load T L and of the torque coefficient K were made use of. In Table 18, the Error represents the difference between the value of the physical characteristic found by the algorithm and the value of the actual one, whereas the Total Error is the value of the objective function C in (23).
Table 18. Motor driven by a step armature voltage: Test on estimating the electrical model parameters only.
As a second test, only the mechanical dynamics of the DC motor was considered as unknown, whereas the electrical parameters were considered as exactly known. Namely, in the second case, the optimization variables are the viscous friction B m and the rotor inertia J. In this test, the exact values of the torque load T L and of the torque coefficient K were again made use of. In Table 19, the Error represents the difference between the value of the physical characteristic found by the algorithm and the value of the actual one, whereas the Total Error is the value of the objective function C in (23).
Table 19. Motor driven by a step armature voltage: Test on estimating the mechanical model parameters only.
As can be readily seen in the above tables, the COA can obtain values that are very close to the actual ones, even without a refined search. The number of iterations needed to achieve such results appears as low.

3.3.2. Joint Electrical and Mechanical Model Parameters Estimation and Weights Analysis

In the definition of the absolute criterion C (23), the weights k a and k b are assigned to the armature current mismatch and to the drive shaft velocity mismatch, respectively. In the third test, different values for such weights were considered in order to achieve a joint electrical and mechanical model parameters estimation. In this test, the torque load T L was assumed equal to zero.
In Table 20, the parameters’ values are shown as obtained with both a low and an high number of iterations. In the latter case, it was evaluated the way that the quality of results improves while increasing the computation time. In the first two cases, even if slightly different results were obtained with few loops (521), it can be seen that both runs converge to the same values by increasing the iterations number. In the third case, the estimations are worse compared to the first two. Therefore, the values k a = 1 and k b = 1 will be used for the subsequent tests.
Table 20. Motor driven by a step armature voltage: Weights analysis for the joint electrical and mechanical model parameters estimation.
As a side note, one might consider that the weights k a and k b may be regarded as constants that represent the measurements’ quality. For instance, if it is known in advance that the samples of the drive shaft’s velocity are affected by a measurement error, the Total Error could be reduced by modifying the constant k b . That is to say, if it is known that a sensor is not particularly sophisticated or accurate, the weight related to the corresponding physical quantity could be adjusted. As a numerical example, Table 21 shows results obtained by adding a Gaussian noise drawn from a normal distribution to the drive shaft velocity and by adjusting the coefficient k b . The Total Error obtained by setting k b = 1 increases drastically due to the noises added to measurements. By decreasing the weight k b by 70 % (namely, to k b = 0.3 ), the Total Error is reduced.
Table 21. Motor driven by a step armature voltage: Joint electrical and mechanical model parameters estimation in the presence of a disturbance on the drive shaft velocity measurements.

3.3.3. Full DC Motor Modeling

After the preliminary tests on modeling separated electrical/mechanical dynamics and after the evaluation of the criterion’s weights, the R-COA was used to obtain a solution to the full DC motor model’s parameter estimation. Several ergodic areas amplifiers were tested. Table 22 illustrates the best solutions found with a low and an high number of iterations.
Table 22. Motor driven by a step armature voltage: Full motor modeling results.
In the table, Total Error is defined as the sum of the differences between samples from the actual motors outputs and the outputs corresponding to the inferred parameters values. The Parameters Error is a performance index that allows one to evaluate the quality of the found parameters values and is defined as a weighted sum, namely
Param . Error = | R a R a s | + | B m B m s | + 10 | J J s | + 100 | L a L a s | + 1 10 | T L T L s | .
Note that the weights of each term in the above sum were adjusted to balance the different values-ranges of the parameters.
The R-COA always converges towards the correct set of parameters, and the results are satisfactory. Comparing the obtained values of Total Error to the Parameters Error, two observations arise:
  • These errors do not decrease proportionally: From the second and third carrier waves, the Parameters Error drops by 62.3 % and by 26.4 %, while the Total Error drops by 31.4 % and by 58.9 %.
  • An improvement in the matching of the model responses to the actual responses does not always imply an improvement towards the correct parameters values. This may be due to the fact that there exists an infinite number of combinations of the parameters that correspond to the same model. In fact, from the state-space representation (22), it can be noticed that the model depends on a ratio between variables, namely, L a divides the electrical parameters while J divides the mechanical parameters: For this reason, infinite potential combinations are possible.
In Figure 14, a comparison between the response of the motor model, in terms of the armature current and of the velocity of the drive shaft, is shown. It is readily observed how the the steady-state responses, also reported in Table 23 for the convenience of the reader, look identical. The small differences of the two motors’ transient compose the Total Error.
Figure 14. Motor driven by a step armature voltage: Comparison of the output curves. Blue-solid line: Actual motor response. Red-crosshatched line: Inferred motor model response.
Table 23. Motor driven by a step armature voltage: Actual steady-state values versus predicted values.

3.3.4. Expansion of the Search Areas

To challenge the R-COA by increasing the problem’s difficulty, additional tests were conducted on expanded search areas of the electric motor’s physical characteristics compared to the ranges indicated in Table 17. Namely, in order to test the behavior of the R-COA when the search areas are badly estimated, the actual ranges were doubled and quadrupled, making them as indicated in Table 24. The R-COA always converges in spite of the enlarged search areas. The refined chaos-based optimization algorithm could find solutions that lay slightly apart from the actual ones and yet that minimize the absolute criterion. Table 25 presents the results of parameter estimation obtained after doubling the search areas, while Table 26 presents the results of parameter estimation obtained after quadrupling the search areas. Both Total and Parameters Error raise in the first and second carrier waves, while they decrease in the third one. If the number of iterations is quadrupled, the curves of the motor’s response get close to one another although the parameters’ precision does not improve. It can be observed that, even if the search areas increase, the number of iterations remains almost the same. In this case, the performance degrades, especially in the identification of the armature resistance R a , with a minimum error of 0.0762 on the actual value, and in the identification of the load torque T L . The Total Error is 6.7405 , while in the previous tests did not exceed 5.
Table 24. Motor driven by a step armature voltage: Ranges for the parameters search areas doubled and quadrupled with respect to the values indicated in Table 17.
Table 25. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on doubled search areas.
Table 26. Motor driven by a step armature voltage: Results of parameters estimation by R-COA on quadrupled search areas.
The two panels of Figure 15 correspond to the values of the parameters reported in Table 25 and Table 26: The shown results seem to suggest that the more the search areas increases, the more the dynamics tend to differ, especially in the transient, while they always converge towards the same steady state.
Figure 15. Motor driven by a step armature voltage: Comparison of responses corresponding to the values of the parameters reported in Table 25 and Table 26. Blue-solid lines: Actual response. Red-crosshatched line: Response corresponding to the values of the parameters inferred upon doubling the search areas. Black dot-dashed lines: Response corresponding to the values of the parameters inferred upon quadrupling the search areas.

3.4. Comparison with the Genetic Algorithm and the Simulated-Annealing Based Algorithm

As a last test, the solutions computed by means of the R-COA were compared to the solutions found by means of a Genetic Algorithm and a Simulated Annealing algorithm. Each optimization algorithm was run 10 times for each input type. Table 27 illustrates the average results. In particular, the bold-faced values are the best estimations for every parameter (e.g., for the Total loops, the parameter R a and the parameter L a ). In this way, it is also clear what algorithm performs the best on a particular parameter or figure.
Table 27. Estimation of a DC motor parameters: Comparison of the R-COA with the SA algorithm and the GA.
The GA is always able to keep the number of total loops low, whereas the SA algorithm in a few cases gets over 3000 iterations. On the other hand, the R-COA exhibits considerably reduced computing time. Quantitatively, it took about 120 min to run the GA- and SA-based experiments, while it took only 10 min to run the R-COA-based experiments (all the computer codes were written in MATLAB and were run on a 2.40 GHz, 4 GB RAM, Intel-based PC).
Genetic algorithm and Simulated Annealing Algorithm rarely converge towards the correct solution; in fact, R-COA’s Total Error is equal to the 5 % of the other two algorithms’ error. These two algorithms, due to the parameters’ imprecision, show a mean-square error, for every measurement, equal to 3.7808 for the GA and equal to 4.4943 for the SA.
Figure 16 illustrates the motor responses modeled by the three optimization algorithms.
Figure 16. Estimation of a DC motor parameters: A comparison of motor responses between models instantiated with parameters estimated by the R-COA, GA and SA algorithm. In the two panels, the solid (blue) lines denote the expected responses, the crosshatched (red) lines denote the results corresponding to the R-COA-based optimization, the dashed (black) lines denote the results pertaining to the GA-based optimization, while the dot-dashed (purple) lines correspond to the SA-based optimization.
In general, COA and R-COA manage, on the one hand, to map the whole constrained area (maintaining a low computational complexity thanks to algorithm design) and, on the other hand, to explore gradually smaller sections by the chaotic wave mechanism that allow to refine the solution without getting lost in parameter space and in local minima. This can be an intuitive explanation of why it outperforms both GA and SA in the benchmark function analysis and in the motor modeling problem. For these reasons, chaotic methods have successfully been exploited in other areas, such as neural-network training, thanks to their excellent optimization abilities (see, for example, [20]).

4. Conclusions

The aim of the research endeavor summarized in the present paper was to investigate the COA, a global optimization algorithm based on the properties of chaotic waves.
The optimization features of the COA were first tested by means of a series of numerical experiments performed on seven benchmark functions, drawn from the scientific literature, that afforded getting an insight into the fundamental properties of this algorithm, which is based on the notion of two carrier waves. Such preliminary numerical tests also led to an interesting refinement that consisted in introducing a third carrier wave, which significantly improves the algorithm’s efficacy as it allows refining the solution at a little expense in terms of optimization loops.
The obtained R-COA algorithm was then applied to a real-world optimization problem, namely, a direct-current electric motor modeling. Several experiments were carried out, where the electrical and mechanical dynamics were analyzed separately or the search areas of the carrier waves were expanded. The same tests were performed using two different optimization algorithms: the genetic algorithm and the simulated annealing algorithm. The comparison with GA and SA highlights COA’s good features: convergence and low computing time.
It is important to emphasize that the effectiveness of COA and R-COA is highly related to the choice of its more important operator, the chaotic map. Further investigations could be carried out to find maps that would grant even better performances. Moreover, for future research, it could be interesting to carry out a more extensive study about both evolutionary strategies to set the ergodic areas’ amplifiers and about the extension of the modeling problem to control.

Acknowledgments

The authors wish to gratefully thank G. Ippoliti for the discussions on the modeling of a DC motor and for suggesting some interesting numerical tests. The authors also wish to express their gratitude to the Academic Editor, A. Mohammad-Djafari, and to the anonymous reviewers for providing a quick and accurate review turnaround.

Author Contributions

The authors have contributed equally in this research and in the writing of this paper. Simone Fiori and Ruben Di Filippo conceived and designed the experiments; Ruben Di Filippo prepared the computer codes and performed the numerical experiments; Simone Fiori and Ruben Di Filippo analyzed the numerical results; Simone Fiori and Ruben Di Filippo wrote the paper. All authors have read and approved the final version of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ditto, W.; Munakata, T. Principles and applications of chaotic systems. Commun. ACM 1995, 38, 96–102. [Google Scholar] [CrossRef]
  2. Hagmair, S.; Bachler, M.; Braunisch, M.C.; Lorenz, G.; Schmaderer, C.; Hasenau, A.-L.; von Stülpnagel, L.; Bauer, A.; Rizas, K.D.; Wassertheurer, S.; et al. Challenging recently published parameter sets for entropy measures in risk prediction for end-stage renal disease patients. Entropy 2017, 19, 582. [Google Scholar] [CrossRef]
  3. Ott, E.; Sauer, T.; Yorke, J.A. Coping with Chaos—Analysis of Chaotic Data and Exploitation of Chaotic Systems. In Wiley Series in Nonlinear Science, 1st ed.; John Wiley: New York, NY, USA, 1994. [Google Scholar]
  4. Pecora, L.M.; Carroll, T.L. Synchronization in chaotic systems. Phys. Rev. Lett. 1990, 64, 821. [Google Scholar] [CrossRef] [PubMed]
  5. Rubinstein, R.Y.; Kroese, D.P. The Cross-Entropy Method—A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning; Springer: New York, NY, USA, 2004. [Google Scholar]
  6. Li, B.; Jiang, W. Optimizing complex functions by chaos search. Cybern. Syst. Int. J. 1998, 29, 409–419. [Google Scholar]
  7. Yang, D.; Li, G.; Cheng, G. On the efficiency of chaos optimization algorithms for global optimization. Chaos Solitons Fractals 2007, 34, 1366–1375. [Google Scholar] [CrossRef]
  8. Yang, D.; Liu, Z.; Zhou, J. Chaos optimization algorithms based on chaotic maps with different probability distribution and search speed for global optimization. Commun. Nonlinear Sci. Numer. Simul. 2014, 19, 1229–1246. [Google Scholar]
  9. Brachem, C. Implementation and Test of a Simulated Annealing Algorithm in the Bayesian Analysis Toolkit (BAT). Bachelor’s Thesis, Georg-August-Universität Göttingen, Göttingen, Germany, 2009. [Google Scholar]
  10. Mitchell, M. Genetic algorithms: An overview. Complexity 1995, 1, 31–39. [Google Scholar] [CrossRef]
  11. Goldberg, D.E. Genetic Algorithms in Search, Optimization & Machine Learning; Addison-Wesley: Boston, MA, USA, 1989. [Google Scholar]
  12. Zhou, S.; Cao, J.; Chen, Y. Genetic algorithm-based identification of fractional-order systems. Entropy 2013, 15, 1624–1642. [Google Scholar] [CrossRef]
  13. Zang, W.; Zhang, W.; Zhang, W.; Liu, X. A kernel-based intuitionistic fuzzy C-means clustering using a DNA genetic algorithm for magnetic resonance image segmentation. Entropy 2017, 19, 578. [Google Scholar] [CrossRef]
  14. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, M.; Qiu, F.; Tian, Z.; Wu, H.; Zhang, Q.; He, W. An information-based approach to precision analysis of indoor WLAN localization using location fingerprint. Entropy 2015, 17, 8031–8055. [Google Scholar] [CrossRef]
  16. Wu, W. DC motor parameter identification using speed step responses. Model. Simul. Eng. 2012, 2012, 30. [Google Scholar] [CrossRef]
  17. Alonge, F.; D’Ippolito, F.; Raimondi, F.M. Least squares and genetic algorithms for parameter identification of induction motors. Control Eng. Pract. 2001, 9, 647–657. [Google Scholar] [CrossRef]
  18. Razik, H.; Rezzoug, A. Identification of electrical parameters of an induction motor, a comparison between R.L.S. and genetic algorithm. In Proceedings of the International Conference on Electrical Machines (ICEM 2000), Espoo, Finland, 28–30 August 2000; pp. 1433–1437. [Google Scholar]
  19. Benaïdja, N. Softcomputing identification techniques of asynchronous machine parameters: Evolutionary strategy and chemotaxis algorithm. Turk. J. Electr. Eng. Comput. Sci. 2009, 17, 69–85. [Google Scholar]
  20. Wang, S.; Zhang, Y.; Ji, G.; Yang, J.; Wu, J.; Wei, L. Fruit classification by wavelet-entropy and feedforward neural network trained by fitness-scaled chaotic ABC and biogeography-based optimization. Entropy 2015, 17, 5711–5728. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.