Rotations for Uniform Adaptive Control and Soft Modeling of Mechanical Devices

: The model-based controllers generally suffer from the lack of precise dynamic models. Making reliable analytical models can be evaded by soft modeling techniques, while the consequences of modeling imprecisions are tackled by either robust or adaptive techniques. In robotics, the prevailing adaptive techniques are based on Lyapunov’s “direct method” that normally uses special error metrics and adaptation rules containing fragments of the Lyapunov function. The soft models and controllers need massive parallelism and suffer from the curse of dimensionality. A different adaptive approach based on Banach’s ﬁxed point theorem and using special abstract rotations was recently suggested. Similar rotations were suggested to develop particular neural network-like soft models, too. Presently, via integrating these approaches, a uniform adaptive controlling and modeling methodology is suggested with especial emphasis on the effects of the measurement noises. Its applicability is investigated via simulations for a two degree of freedom mechanical system in which one of the generalized coordinates is under control, while the other one belongs to a coupled parasite dynamical system. The results are promising for allowing the development of relatively coarse soft models and a simple adaptive rule that can be implemented in embedded systems.


Introduction
The "Model Predictive Controller (MPC)" takes into consideration often contradictory requirements by minimizing a composite cost function under the constraints that precisely describe the dynamic model (i.e., the "abilities") of the controlled system. The mathematical background of this approach corresponds to optimization of functionals for the implementation of which the computationally very greedy "dynamic programming" was suggested in the 1950s [1,2]. By the application of a finite time grid and Lagrange's "reduced gradient method" [3] referred to as "nonlinear programming", the "Receding Horizon Controller (RHC)" [4] was introduced in 1978. Though the resolution of the grid must be fine enough to allow Euler integration over it, RHC normally needs considerably less computational resources than dynamic programming. It was successfully applied for the control of relatively slow processes as e.g., crystallization [5], and biomedical processes in the artificial pancreas [6]. The increase in computational power of modern computers later allowed various fields for its application (e.g., [7][8][9]); however, in robotics, where normally fast motion of the robot arms is required, the optimal control framework was found to be too complicated, and the available dynamic model of the robot was directly applied for computing the necessary driving torque in the "Computed Torque Control (CTC)" approach (e.g., [10][11][12]). In the 1990s, it became clear that it is difficult to obtain precise dynamic robot models (e.g., [13]). The friction effects that are not very well modeled by classical mechanics generally mean problems in model identification even with the relatively simple Striebeck model [14], while more sophisticated friction approaches as the "LuGre" model have to introduce new, dynamically coupled system variables [15].
Since the CTC controller can guarantee asymptotically zero tracking error for arbitrary initial conditions only in the possession of exact dynamic model, to achieve precise tracking, more sophisticated methods were developed. The fundamental idea came from Lyapunov's Ph.D. thesis [16,17] in which he considered the stability of the equilibrium points of certain physical systems that were described by coupled nonlinear systems of ordinary differential equations. Since these equations of motion normally do not have closed form analytical solutions, their subtle details were not discovered; however, by Lyapunov's ingenious method it became possible to define and prove various stability properties for these equilibrium points. In control applications, the "zero tracking error" was placed in the role of the equilibrium point, and in the 1990s, the first paradigms as the "Adaptive Inverse Dynamics Controller (AIDC)" and the "Slotine-Li Adaptive Robot Controller" appeared [18]. This approach works with the introduction of a special metric tensor in the space of the tracking error, its time-derivative and time-integral, by the use of which the Lyapunov function as a square of an error metrics can stagnate, monotonically or strictly monotonically decrease, or asymptotically converge to zero. This metric tensor depends on the feedback gains applied in the control rule. If the exact dynamical model is available, it is possible to prove that this special metrics converges to zero by solving the Lyapunov equation. In this case, the "arbitrary components" of this metric tensor are not present in the control law. If the available dynamical model is imprecise, in the adaptive feedback, the arbitrary components of this metric tensor appear, too. The result normally is not a particular, but rather a whole set of stable adaptive controllers the elements of which produce different error relaxation features. Since in several applications as e.g., in life sciences, the pure fact of stability cannot guarantee the avoidance of "lethal states", complementary tuning of the parameters of the stable solutions became necessary as e.g., in [6]. The main problem is that only the original components of the tracking error have clear phenomenological interpretation and physical significance, as a result, guaranteeing the behavior of these components would be really significant; however, by guaranteeing the decrease of the norm obtained by the use of the particular metric tensor cannot guarantee the decrease of the physically interpreted error components because they are "mixed" in this metric. In the practice, the decrease of the individual components would be desirable.
The adaptive controllers can be divided into two big subsets: the "parameter adaptive approach" as in [18] utilizes the ab ovo exactly known formal properties of the dynamic model and realizes continuous refinement of the model parameters. In the "signal adaptive approach" as in the "Model Reference Adaptive Controller (MRAC)" [19] fast complementary control signals are applied that make the behavior of the controlled system identical to that of the reference model. This reference model normally can be chosen as a stable "Linear Time-Invariant (LTI)" system that easily can be controlled. This approach also uses the Lyapunov function in its design. Though for different problem classes, appropriate "candidate Lyapunov functions" are available (e.g., [20]), this approach is mathematically very difficult and needs well educated, innovative control designers. The failure of finding an appropriate Lyapunov function to a given particular problem does not mean any conclusion for its stability.
To tackle the problem of the "transient behavior" of the controlled system that is not clearly addressed in the approaches that guarantee only the stability of the solution, in [21] a novel iterative controller was suggested in which the task of computing the appropriate control signal was transformed to finding the fixed point of a contractive map. Its structure for a second-order physical system is sketched in Figure 1.
It is a "flexible framework" that can be filled in with various particular contents. In the "Kinematic Block", an arbitrary tracking strategy can be formulated that can drive the trajectory tracking error to 0 as t → ∞. For instance, in a "Proportional, Integral, Derivative (PID)" design the following quantities can be introduced with a positive constant Λ in (1). Its stability can be proved without the use of any Lyapunov function.

Realized Trajectory
The integration is realized by the Physics of the controlled systeṁ q(t ) q(t ) Figure 1. The structure of the fixed point iteration-based adaptive controller for a second-order dynamical system (after [21]).
It is evident that ∀k ∈ {0, 1, 2, . . .} the functions g k (t) : therefore the linear space of the general solution of the LTI system in (1c) can be spanned by three basis functions of which each converges to 0 if t → ∞ in the form given in (3) in which the constants {c 0 , c 1 , c 2 } are determined by the arbitrary initial conditions. The basis function g 0 is mapped to zero by the operator Λ + d dt , g 1 is mapped to zero by , and finally g 2 is mapped to zero by Λ + d dt 3 . It is evident that not only e Int (t) converges to zero: e(t) andė(t) also have to converge to zero because a sequence of implications can be obtained that "after a while" (practically a few times Λ −1 ) various components will become 0 as in (4) e int (t) = 0 "after a while" .
Evidently, (4b) can be rewritten in the form of (5) that corresponds to a stable inhomogeneous system in which due to (4d) the inhomogeneous "driving term" will vanish after a while, therefore e(t) → 0 as t → ∞ In similar manner (4a) can be rewritten as (6) 3Λ that again is a stable inhomogeneous LTI system in which the inhomogeneous driving terms vanish after a while, thereforeė(t) → 0 as t → ∞.
In the case of digital controllers, the box "Delay" normally corresponds to the cycle time δt of the digital controllers. In the box "Deformation", the "Response Function" of the system can be introduced asq = f q De f orm (t),q(t), q(t) ≈ f q De f orm (t), . . . in which the variables in the position denoted by the symbol ". . ." only slowly can vary while the controller very quickly can modifyq De f orm (t). The box contains the functionq De f orm (t) = that is, the solution of the control task is the fixed point of function G. For the construction of G various proposals were put forward and investigated in [21][22][23] by the use of functions in analytical form. For proving the convergence of the iteration Banach's fixed point theorem [24] was used according to which in a linear, normed, complete metric space B ("Banach Space") the iterative sequence created by the contractive mapping Ψ : B → B as {x 0 , x 1 = Ψ(x 0 ), . . . , x n+1 = Ψ(x n ), . . .} converges to the unique fixed point of the function x = Ψ(x ) ∈ B. For this purpose the response function f q Deorm generally has to meet the condition that the real part of each eigenvalue of ∂ f (q De f orm ) ∂q De f orm must be negative or positive simultaneously in the vicinity of the fixed point. In this case, it is possible to introduce appropriate parameters in the function G that can guarantee its contractivity [25]. Though during one digital control step only one step of this iteration can be performed, since q(t) andq(t) vary only very slowly in comparison withq(t), according to ample simulation investigations, this method showed good convergence properties in many cases. The proofs were based on Taylor series approximation of the function values in the vicinity of the fixed point. By realizing similar deformation in the field of the control forces the fixed point iteration-based method was found applicable in a novel design of MRAC controllers [26]. It was shown that this novel technique could be interpreted from the point of view of the Lyapunov function, too [27,28].
In [29] a simple, geometrically interpreted method was introduced. The arraysq Des , q De f orm ∈ R n can be so augmented to arrays X Des , X De f orm ∈ R n+1 that they obtain a common Frobenius norm: X Des = X De f orm = R > 0. If these vectors are not parallel to each other they span a 2 dimensional hyperplane in R n+1 that is also spanned by the orthogonal unit vectors e A := X Des X Des and e B that is made of the component of X De f orm that is orthogonal to X Des . In this case the skew-symmetric matrix Ω := e A e T B − e B e T A and the angle ϕ ∈ R determine an Orthogonal Matrix O(Ω, ϕ) that makes rotations in the R n+1 space so that the "axis of rotation" is the n − 2 dimensional orthogonal subspace of the vectors e A , e B O(Ω, ϕ) = exp(ϕΩ) = that corresponds to the generalization of the Rodrigues formula [30] (this fact easily can be proved by utilizing that Ω 3 = −Ω). If ϕ is calculated from the scalar product as X Des T X De f orm = X Des · X De f orm cos ϕ, O(Ω, ϕ) just rotates the augmented vector X De f orm into the other augmented vector X Des . Consequently, the physically interpreted projection of X De f orm ,q De f orm is exactly transformed intoq Des . By the introduction of an interpolation factor λ a ∈ [0, 1] the rotation with O(Ω, λ a ϕ) makesq De f orm approachq Des . When this function is applied in the block called Deformation, via setting a great value to R and a small one to λ a , the convergence of the iteration can be guaranteed if the angle formed between δ f := f (x + δx) − f (x) and δx is acute. The steering systems of bicycles and cars work accordingly: if the driver wishes to achieve "sharper turn" to the tune of δ f , the necessary modification of the rotational angle of the steering wheel, i.e., δx, forms an acute angle with δ f . In other words: if the driver wishes to turn to the left, the steering wheel has to be turned to the left, too. This simple property can be well utilized in the practice in the development of steering algorithms. With regard to modeling issues, the pioneering discovery by Weierstraß has to be mentioned. In a lecture in 1872 he gave the first example for an everywhere continuous function that nowhere was differentiable [31]. He highlighted the fact that the class of continuous functions is so complicated that we cannot even "imagine" the graph of a general continuous function. We can "see" only the graphs of "smooth" functions the derivatives of which may have some jumps at a few discrete points only. His discovery anticipated difficulties in the field of approximation of continuous functions; however, in his inaugural lecture at the Academy of Berlin in 1885 he has shown that polynomials can be regarded as universal approximators of single variable continuous functions [32]. His theorem was extended from polynomials to other approximators by Stone in 1948 [33]. The approximation of multivariable continuous functions was found to be a more difficult task. In around 1900, Hilbert in one of his conjectures "guessed" that it was impossible to construct continuous multivariable functions by the use of single variable ones [34]. Though in 1927 Volterra introduced a series model [35] that is a sequence of approximations for continuous functions using a polynomial functional expansion [36] for dealing with integro-differential equations, the first rigorous rebuttals of Hilbert's conjecture were published only in 1957 by Arnold for the functions of three variables [37], and by Kolmogorov for continuous functions of many variables [38]. Kolmogorov's constructive proof in the 1960s was simplified and made more elegant by Sprecher in 1965 [39] and Lorentz in 1966 [40,41], and served as the mathematical background of the feedforward neural networks that can be regarded as technical realizations of the universal approximators that appeared in the 1990s [42,43]. In 1990, the idea of "Convolutional Neural Networks (CNN)" was suggested for image recognition applications by Le Cun et al. [44] in which a convolutional layer was the fundamental building block. The early convolutional layers were linear systems: their outputs were affine transformations of their inputs. To take into account nonlinear effects in face recognition, the Volterra polynomials appeared in 2009 in [45], and later were built in the CNNs [36], too. While the feedforward neural networks can be taught by examples, Kohonen's "self-organizing map" [46] was able to automatically find categories in samples. For modeling dynamic effects, recurrent neural networks appeared [47,48] and were introduced in the CNN's convolutional layer [49], too.
For the mathematically rigorous tackling of uncertainties of non-statistical nature in 1965, Zadeh introduced the concept of "fuzzy sets" [50]. By our time, the concept was extended to "type 3" sets [51]. In the 1990s, it became clear that the fuzzy systems can be regarded as universal approximators, too [52,53], in which the Weierstraß-Stone approximation theorem can also be used [54]. Practically satisfactory description of the physical state of certain machines as e.g., turbojet engines, various parameters as temperature, pressure, speed, and vibrations are used that can be revealed by complicated diagnostic methods (e.g., [55][56][57]); however, for control purposes, only certain aspects of the "complete" physical model can be used (e.g., [58,59]) in the form of very "incomplete" models.
From the point of view of function approximation, the various realizations of the universal approximators can be regarded as huge structures that contain numerous free parameters that have to be fitted. For this purpose, numerous methods, such as the gradient descent type "error backpropagation" [60] that can be made more efficient by using appropriate activation functions in the neurons, its combination with genetic algorithms [61], other stochastic optimization methods as simulated annealing [62], memetic algorithms and bacterial memetic algorithms [63][64][65][66], simplex algorithm [67][68][69], "particle swarm algorithm" [70,71], etc., can be mentioned.
Rigorous theoretical investigations soon revealed the phenomenon called "curse of dimensionality" due to which the "universal approximator" ability of the soft computing tools was doubted at least from practical point of view (e.g., [72][73][74]). Much of these problems in terms of "nowhere denseness" stem from the "irregular nature" of the continuous functions discovered by Weierstraß in [31]. It can be expected that the situation is not so hard in the case of "smooth functions". In this view, the idea of "polytopic models" was introduced in the 1910s [75,76] in which the function values were sampled over some grid points of a multidimensional space (i.e., in the vertices of polytopes). By the use of the higher-order generalization of Golub's and Kahan's "Singular Value Decomposition (SVD)" [77] by Lathauwer et al. in 2000 [78], these models can be simplified by finding the most important contributions belonging to the larger singular values, under well controlled conditions. Further, by also taking into account control possibilities via "Linear Matrix Inequalities (LMI)", the controller design methodology of the "tensor product model" was proposed in [79]. This approach renders the program announced by Boyd et al. in 1994 [80] generally applicable. By the use of LMIs a wide set of control problems can be solved by a Lyapunov function-based design, for which efficient MATLAB packages were developed [81]. In this approach, in each cell different metric tensor can be applied in the Lyapunov function. Passing the cells' limits can make certain problems arise that can be tackled by the use of either "switching controllers" (e.g., [82][83][84]) or some redundant system of coordinates can be used from the vertices of a polytope within a convex hull to deal with a continuous problem (e.g., [85]).
To tackle the problem of the curse of dimensionality, to evade the need for massive parallelism and sophisticated data synchronization in the learning and operating phases of the traditional structures, in [86], a novel soft computing structure was suggested that is more or less akin to a coarse resolution grid or a fuzzy model in which very approximate and simple rules are applied for control purposes. It broke with the Lyapunov functionbased design that generally was present in the switching controllers and in the "Linear Parameter Varying (LPV)" or "Quasi LPV" design by the application of simple smoothed tracking of the jumps in the control force. At first, its modeling ability was checked for the description of the behavior of the free and the controlled van der Pol oscillator [87] that is a popular benchmark system. It makes nonlinear oscillations in a limit cycle. For modeling the motion of the free system, the function R q(q,q), and for control purposes, the function R Q q,q,q Des , were approximated in a dynamic range that was "filled in" during the control process. It had the main properties as follows: 1.
For the free system, the two dimensional vectors [q, 0] T and [q,q] T were augmented to three-dimensional ones of identical Frobenius norms [q, 0, D 1 ] T and [q,q, D 2 ] T exactly as it was performed in the rotations-based abstract deformations applied in [29]. In the case of the controlled model the three-dimensional vectors [Q, 0, 0] T and [q,q,q Des ] T were similarly augmented to produce the four dimensional ones of identical norms as [Q, 0, 0, D 1 ] T and [q,q,q Des , D 2 ] in which D 1 and D 2 were the "dummy components" without any physical interpretation. Their role was to guarantee equal norms.

2.
Coarse resolution grids were introduced in R 2 for the q,q values, and in R 3 for the q,q,q Des values. In the center points of the grids, the appropriateq and Q values were computed from the available exact model of the van der Pol oscillator. Following that, the abstract rotations defined in (7) were calculated that rotated [q,q, D 2 ] T into [q, 0, D 1 ] T , and transformed [q,q,q Des , D 2 ] into [Q, 0, 0, D 1 ] T , respectively. Each grid cell was associated with a "neuron" that had the "activation function". It executed the rotations according to (7), and had the following parameters: • Its cell-limits as {q min , q max }, {q min ,q max } for the free motion, and {q min , q max }, {q min ,q max }, {q Des min ,q Des max } for the controlled system, respectively; • The orthogonal unit vectors e A and e B of which the generator Ω of the rotation in (7) can be computed, and the angle of the necessary rotation, ϕ.

3.
These neurons were arranged in a single layer in which each neuron obtained its input value for the "teaching process" as {q,q,q} for the free system, and {q,q,q Des , Q} for the dynamic model. If the input signal belonged to the "range of competence" of the given neuron, e A , e B , and ϕ were computed. During the "normal operation" the neuron used the input for the free system modeling as {q,q} and {q,q,q Des } for "the use for control mode". If the input data belonged to its range of competence, it computed O (Ω, ϕ), computed the rotated vector O(Ω, ϕ)[q,q, D 1 ] T for the free system, and O(Ω, ϕ)[q,q,q Des , D 1 ] T for the control application, and as its output, it provided the first component of the rotated vector that corresponded to the modeled value ofq and Q, respectively. 4.
The last layer of the novel neural structure consisted of a single neuron that summarized the calculated outputs. Since the cells' limits were determined in a way that the model had only disjoint cells, the output of the summarizing layer was the result of the "soft model".

5.
To reduce the effects of the jumps in the control signal Q m at the cell boundaries, the really applied generalized force Q s was smoothed by the tracking rule based on a positive constant λ > 0 in (8) that, for a stationary "driving term" Q m ≡ const, has the stationary solution Q s ≡ Q m , furthermore, for time-varying Q m (t), two different solutions of (8), i.e., Q s (t) and Q s (t) + δQ s (t), can differ from each other by δQ s (t) that satisfies the differential equation λ + d dt δQ s (t) ≡ 0, that is, the differences can stem from the initial conditions and converge to 0 as t → ∞. Consequently, the smoothed signal tracks well the actual Q m (t) if the variation of which is not significant during the time-interval of duration λ −1 , and "smooths well" the signals that have faster variation. 6.
The data representation made it possible to apply real-time modification ("step-bystep learning") of the neuron's previously learned parameters as the unit vectors e A , e B and ϕ were refreshed according to a learning rule determined by the parameter α ∈ [0, 1] as  can generate rota-tions in the form exp ϕ New Ω New , but, because Ω New 3 = −Ω New , instead of (7) we can state only that i.e., by maintaining the right-hand side of (10) that is easy to compute, instead of exact rotations, linear transformations can be used that are good approximations of rotations.
In the present paper the above outlined idea is extended and it contains the novelties as follows: 1.
The controlled system is an underactuated two degree of freedom construction in which the directly controlled subsystem is dynamically coupled with a noncontrollable one acting as "parasite dynamics".

2.
Instead of the simple CTC control and its robust variable structure/sliding modebased correction (e.g., [88][89][90]), the "fixed point iteration-based adaptive control scheme" depicted in Figure 1 is applied to compensate the effects of the imprecisions of the coarse grid-based model with the application of the rotations-based adaptive controller announced in [29]. 3.
The effects of the measurement noises are investigated and reduced by a smoothing technique that is similar to the solution published in [91].

4.
The computation time of the controller was measured for the hardware and software environment that was used in the simulations.
In Section 2, the dynamic model of the controlled system is discussed. Section 3 presents the rotational network-based soft model; in Section 4, simulation results are presented. The paper is closed with Section 5 in which the results are discussed and further possible research is outlined.

The Dynamic Model of the Controlled System
The controlled system is a wheel that is rotated around a horizontal axle. In one of its spokes, a mass point is built in. It is located between two springs and its motion is damped by viscous friction. The Euler-Lagrange equations of motion of this model are completed with the friction term and are given in (11) that can be used for simulation purposes, in which q 1 denotes the rotational angle of the wheel, q 2 is the radial distance of the mass point along the spoke, measured from the rotational axle, r is the zero force position of the spring connected the axle with the mass point, k is the spring constant, d denotes the viscous damping coefficient of the mass point as it moves along the spoke, m is its inertia, and Θ corresponds to the inertia momentum of the wheel, Q 1 denotes the driving torque. For control purposes, the necessary control torque can be obtained by rearranging (11a) as The numerical data used in the simulations are given in Table 1.
With regard to noise issues, it was assumed that q 1 was measurable in each digital cycle with a Gaussian noise of σ 1 = 10 −3 rad and q 2 was measurable with a Gaussian noise of σ 2 = 10 −3 m. For filtering the measurement noises, the measured noisy signal x was tracked by the filtered onex according to the tracking rule given in (13) ...
and the filtered valuesx,ẋ,ẍ were used in the "Kinematic Block" and in the fixed point iteration with the filtering parameter λ n = 500.0 s −1 . The operation of this filter can be analyzed exactly as it was performed in the case of (8).

Simulation Results
Numerical simulations were performed by the use of the Julia language Version 1.5.1 (25 August 2020) working under the operating system Linux 5.10.53-1-MANJARO x86_64 21.1.0 Pahvo on a DELL inspiron 15R laptop using the program code that is available on the Web given in Section "Sample Availability": "Wheel_Applied_Sciences_Noisy.jl".

Comparative Analysis of the Performances of the Neural and the Exact Models
In this subsection, the figures also contain simulation results obtained when the exact dynamic model was in use to support comparative analysis. In Figure 3, two 2-dimensional sub-spaces of the stored "soft dynamic model" in R 5 are exemplified by giving the stored abstract rotational angles of the rotations in the augmented R 6 space. The non-adaptive control corresponds to the simple CTC controller. In Figure 4, the trajectory tracking properties can be compared.      According to Figure 6, the adaptation makes the phase trajectory of the controlled system "more canonical" and reduces the excitation of the coupled parasite dynamical subsystem. According to Figure 7, the adaptive approach causes more even control torque contributions in spite of the noisy signals. In the non-adaptive results, the effects of changing the cell that is competent to fire can be identified. The dynamic range of the adaptive control forces is considerably narrower than that of the non-adaptive control. Figure 8 reveals the operation of the adaptive controller: due to the adaptive deformation, the realized 2nd time-derivative values well approximate the desired ones. This effect can be observed better in the lack of measurement noises in Figure 9.  To reveal the behavior of other cells, nominal motion of smaller amplitude and higher circular frequency was investigated in Figures 10-13.
Finally, simulation results are displayed for nominal motion of small amplitude and small circular frequency in Figures 14-16.

Estimation of the Computational Time of the Operations in the Control Cycles
Computational complexity of the suggested method and the duration of the necessary computations within the control cycles is an interesting practical question. It cannot be estimated by counting the necessary mathematical operations, because it depends on various external factors, such as the operating system that manages the controller, and the properties of the program that realizes the control task. Normally, a multitasking controller is available that uses a single central processor unit. Its capacities are shared by various cooperating processes. If the operating system is not definitely designed to manage realtime tasks, the actual duration of a given computation depends on the other duties of the computer. For instance, if during the calculation of the simulations some video player is in use, its effects can be well observed in the duration of each actually executed computation.
In a similar manner, if the program language makes "garbage collection" in the stack, such an activity may need considerably longer time than that of a "normal computation" within the control cycle. These longer sessions, though not very frequently, sometimes appear during the simulation. The Julia language is a typical garbage collecting construction that generates such effects.
However, it offers a simple solution to measure the time that is needed for the operations defined in a given command line. For instance, the command "a+=@elapsed" b=myfunc(-12.1)" adds to the content of variable "a" the computational time (in seconds) of the following steps: the function "myfunc" is called with the input argument -12.1 and its output is stored in variable "b". The time-need of a bigger unit as e.g., "a+=@elapsed for...end", "a+=@elapsed while...end", etc. can be measured in similar manner.
A program code "Wheel_Applied_Sciences_Noisy_Computing_Time_New.jl" that utilizes this possibility (also available in the Web) was made in which the time needed for the essential computations within each control cycle was summarized. In these investigations, the parameter settings λ a = 5 × 10 −2 and Λ = 0.9 s −1 were used, all the other control and model parameters remained invariant. In this program, no computation was made for the operation of the "exact system"-based control. The nominal trajectory had time-dependent, increasing amplitude. During the calculations, only the console running Julia was in operation, a text editor in which the program file was kept opened, and the standard Internet connection was kept running. In the sequel, results obtained for the adaptive and non-adaptive controllers are discussed. Figure 17 reveals a few huge peaks that certainly cannot belong to the essential computations. The important information can be seen in the "zoomed-in excerpts" of the graphs displayed in Figure 18.
It is evident that during one digital control step of duration of 1.0 ms the non-adaptive computations needed less than 0.065 ms, while the adaptive ones consumed up approximately maximum 0.1 ms. This observation indicates that the suggested method can be implemented by the use of common hardware/software tools. Figures 19 and 20 compare the trajectory tracking performances. In Figures 21 and 22 the phase trajectory of the controlled and that of the coupled "parasite system" can be seen. These figures well reveal the effects of the necessary centripetal force that increases with the amplitude of the nominal motion.

Discussion
In this paper, the idea of "rotational neural networks" introduced in [86] was applied to bring about a coarse grid-based soft dynamic model for control applications. The jumps in the model output at the cell boundaries are smoothed by a first-order tracking controller, and the effects of the model inaccuracies are compensated by a fixed point iteration-based adaptive controller. Both the adaptive and modeling mechanisms are based on the same mathematical background that applies simple rotations in higher dimensional spaces. It was shown that the network is able to work in a "step-by-step" learning mode in which the information stored in a visited cell can be refined depending on the fact that various parts of the coarse resolution cells can be visited during the motion of the controlled system.
The suggested network has very simple topological structure and it is easy to simulate it, too. It was shown that the modeling and control methods can be combined with simple noise filtering techniques. The operation of this controller seems to be simpler than that of the traditional fuzzy controller that has to compute ample number of "generalized" maximum and minimum operations over wide ranges.
The most important advantage of this approach is that it breaks with the Lyapunov function-based design in which the phenomenologically well interpreted error components, their integrals and derivatives are "mixed" by the use of very special metric tensors so that these metrics as they are do not have clear interpretation and physical meaning. In the novel approach for each error component kinematically well interpreted behavior can be prescribed.
The applicability of the suggested method was demonstrated in the case of a two degree of freedom underactuated paradigm in which the control of the motion of a wheel is perturbed by a directly not controllable, dynamically coupled parasite system within one of its spokes. The sum of the duration of the necessary computations within each control cycle was calculated for the hardware/software system that run the simulation. It was concluded at the level of our prevailing possibilities the suggested approach is realistic.
In the simulations, special PID-like kinematic behavior was required that does not result in monotonic decrease of the error components; however, this method can be combined with various fractional order derivatives-based kinematic requirements (e.g., [92][93][94]) that can produce monotonic variation.
In the future, we wish to compare the operation of this idea with that of traditional fuzzy controllers based on approximate and "coarse" rules. For this purpose an electromechanical testbed was built and used for testing the operation of the traditional PID and CTC controllers in [95]. The first successful preliminary results with regard to the implementation of the fixed point iteration-based adaptive controller using abstract rotations on the same testbed were recently obtained by Árpád Varga. The planned future investigations are based on this instrument.